Pre-screened and vetted.
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and MLOps
“Built and deployed a production RAG system for financial/compliance teams using GPT-4, Claude, and local models to retrieve and summarize thousands of internal documents with strong security controls (role-based retrieval, PII masking). Drove significant operational gains (30+ hours/week saved, ~35% productivity lift, ~45% faster responses) and orchestrated end-to-end ingestion/embedding/index refresh pipelines with Airflow, S3, and SageMaker while partnering closely with compliance stakeholders on auditability and traceability.”
“ML/LLM engineer with production experience building a RAG-based LLM support assistant (FastAPI, Redis, Kafka) with multi-layer validation and human-in-the-loop feedback loops to improve accuracy over time. Has orchestration and MLOps depth using Airflow and Kubeflow on Kubernetes (autoscaling, alerting, monitoring) and delivered measurable ops impact (40% ticket efficiency improvement) by partnering closely with customer support teams.”
Mid-level AI/ML Engineer specializing in MLOps, NLP, and real-time ML pipelines
“Built a production, real-time insurance claims document-understanding and fraud-detection pipeline using TensorFlow + fine-tuned BERT, deployed on AWS (SageMaker/Lambda/API Gateway) with automated retraining via MLflow and Jenkins. Addressed noisy documents and latency using augmentation and model distillation (3x faster), cutting claims ops manual review by ~50% and reducing fraudulent payouts.”
“ML engineer/data scientist who deployed a production credit risk + insurance claims triage platform at Hartford Financial, combining XGBoost default prediction with BERT-based document classification. Demonstrated strong MLOps by cutting inference latency to sub-500ms and building drift monitoring plus automated retraining/deployment pipelines (MLflow, CloudWatch, GitHub Actions, SageMaker) with human-in-the-loop review and SHAP-based explainability for underwriting adoption.”
Mid-level Data Scientist/MLOps Engineer specializing in NLP, GenAI, and cloud ML platforms
“AI/ML engineer who led production deployment of a multimodal (text/video/image) RAG system on GCP using Gemini 2.5 + Vertex AI Vector Search, scaling to 10M+ documents with sub-second latency and +40% retrieval accuracy. Strong MLOps/orchestration background (Kubernetes, CI/CD, Airflow, MLflow) with proven impact on reliability (75% fewer incidents) and deployment speed (92% faster), plus experience delivering explainable ML (XGBoost + SHAP + Tableau) to non-technical retail stakeholders.”
Mid-level Data Scientist specializing in cloud ML, MLOps, and predictive analytics
“NLP/ML engineer with hands-on healthcare and support-ticket text experience, building clinical-note structuring and semantic linking systems using spaCy, BERT clinical embeddings, and FAISS. Emphasizes production-grade delivery (Airflow/Databricks, PySpark, Docker, AWS/FastAPI/Lambda) and rigorous validation via clinician-labeled datasets, retrieval metrics, and user feedback.”
Junior AI/ML Engineer specializing in NLP, LLMs, and MLOps deployment
“Built and deployed NeuroDoc, a production-grade RAG system for PDF Q&A that delivers citation-backed answers with strong anti-hallucination guardrails. Experienced in orchestrating and scaling ML/LLM pipelines with Kubernetes, Airflow/Prefect, and PyTorch Distributed, and in building rigorous evaluation and citation-verification tooling to ensure reliability in production.”
Mid-level Data Engineer specializing in healthcare data platforms and MLOps
“ML/NLP practitioner with healthcare payer experience at HCSC, focused on connecting messy unstructured clinical notes to structured claims/provider data to improve fraud-analytics workflows. Has hands-on experience fine-tuning transformers in AWS SageMaker, building large-scale embedding search with FAISS, and implementing robust entity resolution using golden datasets, precision/recall calibration, and production monitoring for drift.”
Mid-level Machine Learning Engineer specializing in LLM apps, RAG pipelines, and MLOps
“Software engineer with connected-car/automotive production experience who owned an end-to-end remote door lock/unlock feature and introduced unit testing (GTest) plus rig/simulator validation. Also built and productionized an AI-native AWS cloud cost assistant (Lex + GPT-based LLM + Lambda + RAG/vector DB) with guardrails and achieved 94% evaluation accuracy. Helped replace a third-party solution with an in-house build, saving the company ~€9M.”
Senior Full-Stack AI Engineer specializing in Generative AI and FinTech
“Backend engineer who built and owned an AI-powered financial research product end-to-end, using a typed NestJS/GraphQL backend with LangGraph-style agent routing to produce sourced, structured financial analysis. Emphasizes finance-grade correctness (Zod validation, metric registries, unit/empty-result guardrails) while keeping latency low via batching, caching, and fast token streaming, and has led incremental migrations using strangler/feature-flag/shadow traffic patterns.”
Junior Software Engineer specializing in full-stack web and cloud systems
“Co-op engineer at EnFi who built and maintained a multi-tenant prompt library and LLM workflow tooling used by internal teams and external enterprise clients. Led TypeScript/React package design and standardized a typed workflow abstraction across disparate implementations (React, Go, JSON), improving reliability and developer adoption. Delivered measurable performance gains (~25% latency reduction) and owned end-to-end execution including docs, demos, debugging, and deployment.”
Intern Data Scientist specializing in robotics localization and SLAM
“Robotics/embodied-AI practitioner who built a TurtleBot3 LiDAR-fingerprint localization pipeline end-to-end (autonomous data collection + multi-head NN) achieving ~30 cm error in a 10x10 m space. Also has industry experience at Infineon building large-scale production data/AI pipelines and rapidly fixing a deployed recommendation system by correcting upstream data normalization, improving accuracy by 20%+.”
Senior Software Engineer specializing in Python microservices, cloud platforms, and ML-powered APIs
“Backend/data engineer focused on AWS-native Python systems: built a FastAPI microservice on ECS/Fargate serving real-time analytics at millions of daily requests with strong reliability (OAuth2/JWT, retries/timeouts, correlation IDs) and autoscaling. Also delivered Glue/PySpark ETL pipelines to curated S3 Parquet/Athena with schema evolution + data quality controls, owned Airflow pipeline incidents, and has a track record of measurable performance and cost optimizations (e.g., ~80%+ query latency reduction; reduced logging/NAT/Fargate spend).”
Junior Machine Learning Engineer specializing in LLMs and RAG systems
“Production-focused applied ML/LLM engineer who has deployed an LLM-powered RAG assistant and improved reliability through rigorous retrieval evaluation (recall/MRR), reranking, and guardrails that prevent confident wrong answers. Experienced running containerized ML/LLM services on Kubernetes (including AWS-managed layers) with CI/CD and observability, and has delivered a real-time predictive maintenance system using streaming sensor data and time-series anomaly detection in close partnership with maintenance teams.”
Mid-level AI & Machine Learning Engineer specializing in Generative AI and MLOps
“Built a production GPT-4/LangChain/Pinecone RAG “AI Copilot” at Northern Trust to automate financial report generation and analyst Q&A over internal structured (SQL warehouse) and unstructured policy data. Focused on real-world production challenges—grounding and latency—achieving major speed gains (seconds to milliseconds) via MiniLM embedding optimization and Redis caching, and implemented rigorous testing/evaluation with MLflow-backed metrics while aligning compliance and finance stakeholders for deployment.”
Mid-Level AI/Full-Stack Engineer specializing in agentic LLM systems and RAG
“Built and deployed Clyra.AI, an AI-driven daily scheduling product that uses a LangGraph-based multi-agent LLM pipeline (task extraction, verification, reflection) grounded with strict RAG over emails/documents/calendars and real-world signals like health metrics. Designed a custom agent orchestrator with bounded loops/termination conditions and a self-auditing verification/reflection layer to reduce hallucinations while controlling latency and cost via caching and model distillation.”
Mid-level Machine Learning Engineer specializing in LLMs, GenAI, and Computer Vision
“LLM/agent engineer who built a production multi-agent research automation system using LangGraph (planner, retriever with FAISS, supervisor, evaluator) with structured outputs and citation tracking for traceable reports. Emphasizes reliability and operations—LangSmith-based observability, multi-level testing, hallucination mitigation, and latency/cost controls—plus prior experience as a Computer Vision Software Engineer at Deepsight AI Labs working directly with non-technical customers.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and multimodal AI on AWS
“Built and deployed a production RAG-based enterprise document intelligence platform for financial/compliance/operational documents on AWS (Spark/Glue ingestion, embeddings + vector DB, LangChain orchestration, REST APIs on Docker/Kubernetes). Deep hands-on experience orchestrating multi-step and multi-agent LLM workflows (LangChain, LangGraph, CrewAI) with strong focus on grounding, evaluation, observability, and cost/latency optimization, and has partnered closely with non-technical finance/compliance teams to drive adoption.”
Junior Full-Stack Software Engineer specializing in React, Kubernetes, and AI-powered apps
“Backend/DevOps-leaning engineer managing multiple customer service platforms end-to-end (requirements through deployment). Built an in-house Python monitoring/alerting solution for Salesforce-to-Java contact sync jobs (Snowflake dependencies) that increased uptime ~60%, and helped modernize delivery by moving the team from manual releases to automated Jenkins-based deployments while coordinating an Oracle EBS→Fusion transition with business/data/IT stakeholders.”
Intern AI Researcher specializing in NLP, LLMs, and knowledge graphs
“Built and shipped “LabMate,” a production AI assistant specialized in laboratory hardware, using a weighted multi-source RAG pipeline with reranking and reasoning-focused query decomposition to handle complex user questions. Deployed on a local GPU cluster with vLLM and NVIDIA MPS (plus OCR/VLM components), and established evaluation using synthetic + public reasoning datasets while collaborating weekly with non-technical admins to align requirements and resource constraints.”
Mid-level Full-Stack Python Developer specializing in Healthcare IT
“Backend/AI engineer with Johnson & Johnson experience building data-heavy payer/claims analytics services (Python/FastAPI, PostgreSQL, AWS) and optimizing them under peak ingestion load via indexing/query tuning and caching. Also shipped an end-to-end RAG feature for clinicians to extract insights from unstructured clinical notes, using constrained prompts and retrieval-confidence guardrails to prevent hallucinations.”
Senior Data Scientist/Software Engineer specializing in ML systems and cloud DevOps
“AI software engineer with experience spanning LLM/RAG production systems and regulated fintech infrastructure. Built an end-to-end natural-language-to-SQL analytics assistant (Weaviate + GPT-4 + Supabase) shipped as an API with 92% accuracy and major time savings for non-technical users, and also owned demand-forecasting and CI/CD/containerization improvements for a Bank of America core banking deployment at Infosys.”
Mid-level AI/ML Engineer specializing in GenAI and cloud MLOps
“Applied LLMs to high-stakes domains (wildfire risk for emergency teams and loan approval via a fine-tuned IBM Granite model), with a strong focus on reliability—using RAG-based cross-validation to reduce hallucinations and continuous ingestion pipelines (MODIS satellite imagery via AWS Lambda) to keep data current. Experienced in production orchestration and MLOps-style workflows using Airflow, AWS Step Functions, and SageMaker Pipelines, and collaborates closely with analysts on KPI-driven evaluation.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM/RAG system at CVS to automate clinical documents, addressing PHI compliance, retrieval accuracy, and latency; achieved a 35–40% reduction in review effort through chunking and FP16/INT8 optimization. Also has experience translating AI outputs into actionable insights for non-technical stakeholders (sports analysts).”