Pre-screened and vetted.
Mid-level Data Engineer specializing in healthcare data platforms and MLOps
“ML/NLP practitioner with healthcare payer experience at HCSC, focused on connecting messy unstructured clinical notes to structured claims/provider data to improve fraud-analytics workflows. Has hands-on experience fine-tuning transformers in AWS SageMaker, building large-scale embedding search with FAISS, and implementing robust entity resolution using golden datasets, precision/recall calibration, and production monitoring for drift.”
Intern Data Scientist specializing in robotics localization and SLAM
“Robotics/embodied-AI practitioner who built a TurtleBot3 LiDAR-fingerprint localization pipeline end-to-end (autonomous data collection + multi-head NN) achieving ~30 cm error in a 10x10 m space. Also has industry experience at Infineon building large-scale production data/AI pipelines and rapidly fixing a deployed recommendation system by correcting upstream data normalization, improving accuracy by 20%+.”
Junior AI/ML Engineer specializing in LLM agents and RAG systems
“Backend/data engineer who built a production-ready multi-agent financial intelligence system (Mycroft) that orchestrates specialized AI agents to analyze real-time market data using FastAPI and Pinecone vector search. Brings strong security/reliability instincts (rate limiting, JWT/OAuth2, retries/backoff, health checks) and has caught high-impact data integrity issues in financial migrations (timezone normalization across global legacy systems).”
Mid-level Full-Stack Java Engineer specializing in banking microservices and AI backends
“Backend-focused software engineer building distributed, event-driven Java/Spring Boot microservices with Kafka for low-latency, high-frequency processing. Has hands-on experience modernizing a legacy Java system into containerized microservices deployed on Kubernetes with GitHub Actions CI/CD, and has integrated retrieval-based AI components into production workflows; no ROS/robot hardware experience yet.”
Junior Machine Learning Engineer specializing in LLMs and RAG systems
“Production-focused applied ML/LLM engineer who has deployed an LLM-powered RAG assistant and improved reliability through rigorous retrieval evaluation (recall/MRR), reranking, and guardrails that prevent confident wrong answers. Experienced running containerized ML/LLM services on Kubernetes (including AWS-managed layers) with CI/CD and observability, and has delivered a real-time predictive maintenance system using streaming sensor data and time-series anomaly detection in close partnership with maintenance teams.”
Mid-level AI & Machine Learning Engineer specializing in Generative AI and MLOps
“Built a production GPT-4/LangChain/Pinecone RAG “AI Copilot” at Northern Trust to automate financial report generation and analyst Q&A over internal structured (SQL warehouse) and unstructured policy data. Focused on real-world production challenges—grounding and latency—achieving major speed gains (seconds to milliseconds) via MiniLM embedding optimization and Redis caching, and implemented rigorous testing/evaluation with MLflow-backed metrics while aligning compliance and finance stakeholders for deployment.”
Mid-level Machine Learning Engineer specializing in LLMs, GenAI, and Computer Vision
“LLM/agent engineer who built a production multi-agent research automation system using LangGraph (planner, retriever with FAISS, supervisor, evaluator) with structured outputs and citation tracking for traceable reports. Emphasizes reliability and operations—LangSmith-based observability, multi-level testing, hallucination mitigation, and latency/cost controls—plus prior experience as a Computer Vision Software Engineer at Deepsight AI Labs working directly with non-technical customers.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and multimodal AI on AWS
“Built and deployed a production RAG-based enterprise document intelligence platform for financial/compliance/operational documents on AWS (Spark/Glue ingestion, embeddings + vector DB, LangChain orchestration, REST APIs on Docker/Kubernetes). Deep hands-on experience orchestrating multi-step and multi-agent LLM workflows (LangChain, LangGraph, CrewAI) with strong focus on grounding, evaluation, observability, and cost/latency optimization, and has partnered closely with non-technical finance/compliance teams to drive adoption.”
Mid-level AI/ML Software Engineer specializing in data pipelines, BI dashboards, and computer vision
“Graduate Assistant Intern at Friends University who built and deployed a GenAI-driven requirement understanding system that automates extraction and semantic grouping of technical requirements from large unstructured documents. Demonstrates strong LLM engineering rigor (golden datasets, regression testing, post-processing validation) and production-minded delivery using LangChain/LlamaIndex orchestration, FastAPI microservices, Docker, and cloud deployment.”
Senior Data Scientist/Software Engineer specializing in ML systems and cloud DevOps
“AI software engineer with experience spanning LLM/RAG production systems and regulated fintech infrastructure. Built an end-to-end natural-language-to-SQL analytics assistant (Weaviate + GPT-4 + Supabase) shipped as an API with 92% accuracy and major time savings for non-technical users, and also owned demand-forecasting and CI/CD/containerization improvements for a Bank of America core banking deployment at Infosys.”
Mid-level AI/ML Engineer specializing in GenAI and cloud MLOps
“Applied LLMs to high-stakes domains (wildfire risk for emergency teams and loan approval via a fine-tuned IBM Granite model), with a strong focus on reliability—using RAG-based cross-validation to reduce hallucinations and continuous ingestion pipelines (MODIS satellite imagery via AWS Lambda) to keep data current. Experienced in production orchestration and MLOps-style workflows using Airflow, AWS Step Functions, and SageMaker Pipelines, and collaborates closely with analysts on KPI-driven evaluation.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM/RAG system at CVS to automate clinical documents, addressing PHI compliance, retrieval accuracy, and latency; achieved a 35–40% reduction in review effort through chunking and FP16/INT8 optimization. Also has experience translating AI outputs into actionable insights for non-technical stakeholders (sports analysts).”
Mid-level AI/ML Engineer specializing in fraud detection and healthcare predictive analytics
“ML/AI engineer with production experience in high-scale banking fraud detection at Truist, building an end-to-end pipeline (Airflow/AWS Glue/Snowflake, PyTorch/sklearn) with automated retraining and Kubernetes-based deployment; delivered measurable gains (22% fewer false positives, 15% higher recall) and reduced manual ops ~40%. Also partnered with clinicians at Kellton to deploy an LLM system for summarizing/classifying clinical notes, improving review time and decision speed.”
Mid-level AI/ML Engineer specializing in MLOps, NLP, and scalable model deployment
“Built and deployed a production autonomous AI data analyst agent (LangChain + GPT + Streamlit on AWS) that turns natural-language questions into validated SQL, visualizations, and insights, cutting manual analysis time by ~50%. Emphasizes reliability and MLOps: schema-aware validation/guardrails to prevent hallucinations, scalable large-data processing, and Azure DevOps CI/CD + MLflow for automated deployment and experiment tracking.”
Mid-level Machine Learning Engineer specializing in deep learning and generative AI
“ML/NLP engineer with hands-on experience building production systems for unstructured insurance claims and customer data linking. Delivered measurable impact at scale (millions of documents), combining transformer-based NLP, vector search (FAISS/Pinecone), and human-in-the-loop validation, and has strong production workflow/observability practices (Airflow, AWS Batch, Grafana/Prometheus).”
Mid-level AI/ML Engineer specializing in fraud detection and NLP
“Built production AI/RAG-style systems for message Q&A and insurance claims workflows, combining data ingestion, indexing/retrieval, and LLM integration with fallback modes. Has hands-on orchestration experience (Airflow, Prefect, LangChain) and cites large operational gains (claims processing reduced to ~45 seconds; manual review -50%; false alerts -30%) through automated, monitored pipelines and close collaboration with non-technical stakeholders.”
“AI/ML engineer with banking domain experience (M&T Bank) who built a production credit-risk prediction and reporting platform combining ML models (XGBoost/TensorFlow) with a RAG pipeline (LangChain + GPT-4) over compliance documents. Delivered measurable impact (≈20% better risk detection/precision, 50% less manual reporting) and productionized workflows on Vertex AI/Kubeflow with CI/CD and monitoring; also implemented embedding-based semantic search using FAISS/Pinecone.”
Mid-level AI/ML Engineer specializing in healthcare ML and generative AI
“AI/LLM engineer at Humana who built and deployed a HIPAA-aware RAG system for clinical record retrieval, cutting search time dramatically and improving retrieval efficiency by 30%. Experienced with Spark-scale data preprocessing, QLoRA fine-tuning, LangChain orchestration, and MLflow+SageMaker integration, with a strong testing/evaluation discipline (A/B tests, human eval) to hit 95%+ accuracy and production latency targets.”
Mid-level AI & Data Scientist specializing in LLMs, RAG, and healthcare NLP
“Built a production LLM/RAG solution for healthcare operations teams to query large policy and care-guideline repositories in natural language. Improved domain alignment using vector retrieval plus parameter-efficient fine-tuning and prompt optimization, validated through internal user testing and metrics, cutting manual lookup time by ~40%. Also has hands-on experience orchestrating automated ML pipelines with Apache Airflow.”
Intern Data Analyst specializing in data pipelines and LLM/RAG applications
“Built and deployed LLM-powered analytics and reporting systems, including a RAG-based assistant over Snowflake that let business users ask questions in plain English instead of writing SQL. Experienced orchestrating LLM agents (LangChain) and serverless reporting pipelines (AWS Lambda/S3/RDS), with a strong focus on grounded outputs, monitoring/evaluation, and data quality—used daily by non-technical finance and operations teams at Cigna.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps for financial services
“Built and deployed a production Llama 3-based RAG document Q&A system using FAISS, addressing context-window limits through chunking and keeping retrieval accurate by regularly refreshing embeddings. Has hands-on orchestration experience with LangChain and LlamaIndex for multi-step LLM workflows (including memory management) and collaborates with non-technical teams (e.g., marketing) to deliver AI solutions like recommendation systems.”
Senior Full-Stack Software Engineer specializing in IIoT, Edge AI, and real-time analytics
“Full-stack engineer who built an end-to-end low-code/no-code IDE for creating AI/ML workflows for industrial IoT sensors using Next.js/TypeScript and NestJS microservices. Focused on scaling high-volume sensor dashboards—improved UX and performance via WebSockets, debouncing, pagination, and API payload reduction—validated with profiling tools and user feedback in a startup environment.”
Mid-Level Software Engineer specializing in backend, distributed systems, and AI/LLM platforms
“Built and shipped AI-powered workflow automation at Oracle, including an MCP-based agentic workflow with tool-calling and guardrails, plus Grafana monitoring and Confluence documentation. Also led a Django monolith-to-microservices migration at Chamsmobile using blue-green deployment and load balancer traffic splitting to avoid regressions while modernizing production systems.”
Mid-Level AI Backend Engineer specializing in Python, LLM/RAG, and healthcare/insurance platforms
“AI Backend Engineer in MetLife’s claims technology group who built and deployed a production LLM-based decision support system that helps claim adjusters quickly find relevant policy rules from long PDFs and historical notes. Designed it as multiple production-grade services with retrieval-first guardrails, continuous validation, and Airflow-orchestrated pipelines for ingestion, embeddings, and vector index updates to keep the system reliable as policies and data evolve.”