Pre-screened and vetted.
Senior Backend Software Engineer specializing in distributed systems and cloud microservices
“Backend engineer with NTT Data experience building Java/Spring Boot services for product-data ingestion, including Kafka-based asynchronous pipelines and Redis read-through caching. Also built a personal RAG system deployed on Google Kubernetes Service using FastAPI, LangChain, and Pinecone with multi-tenant data isolation; holds a Master’s background in Machine Learning.”
“ML engineer/data scientist who deployed a production credit risk + insurance claims triage platform at Hartford Financial, combining XGBoost default prediction with BERT-based document classification. Demonstrated strong MLOps by cutting inference latency to sub-500ms and building drift monitoring plus automated retraining/deployment pipelines (MLflow, CloudWatch, GitHub Actions, SageMaker) with human-in-the-loop review and SHAP-based explainability for underwriting adoption.”
Mid-level Data Scientist/MLOps Engineer specializing in NLP, GenAI, and cloud ML platforms
“AI/ML engineer who led production deployment of a multimodal (text/video/image) RAG system on GCP using Gemini 2.5 + Vertex AI Vector Search, scaling to 10M+ documents with sub-second latency and +40% retrieval accuracy. Strong MLOps/orchestration background (Kubernetes, CI/CD, Airflow, MLflow) with proven impact on reliability (75% fewer incidents) and deployment speed (92% faster), plus experience delivering explainable ML (XGBoost + SHAP + Tableau) to non-technical retail stakeholders.”
Junior AI/ML Engineer specializing in NLP, LLMs, and MLOps deployment
“Built and deployed NeuroDoc, a production-grade RAG system for PDF Q&A that delivers citation-backed answers with strong anti-hallucination guardrails. Experienced in orchestrating and scaling ML/LLM pipelines with Kubernetes, Airflow/Prefect, and PyTorch Distributed, and in building rigorous evaluation and citation-verification tooling to ensure reliability in production.”
Mid-level Data Engineer specializing in cloud data platforms and AI/ML analytics
“Backend/data engineer in healthcare who built an AWS-based clinical analytics platform from scratch (DynamoDB/S3/Airflow/dbt) with sub-second clinician query goals, 99.9% uptime, and HIPAA-grade controls (KMS encryption, IAM RBAC, audit trails). Also modernized ML delivery by replacing a manual 4-hour deployment with a 30-minute Docker/GitHub Actions CI/CD pipeline using parallel runs, parity testing, and rollback, and caught critical EHR data edge cases (date formats/timezones) that could have impacted patient care.”
Intern Data Scientist specializing in robotics localization and SLAM
“Robotics/embodied-AI practitioner who built a TurtleBot3 LiDAR-fingerprint localization pipeline end-to-end (autonomous data collection + multi-head NN) achieving ~30 cm error in a 10x10 m space. Also has industry experience at Infineon building large-scale production data/AI pipelines and rapidly fixing a deployed recommendation system by correcting upstream data normalization, improving accuracy by 20%+.”
Junior Machine Learning Engineer specializing in LLMs and RAG systems
“Production-focused applied ML/LLM engineer who has deployed an LLM-powered RAG assistant and improved reliability through rigorous retrieval evaluation (recall/MRR), reranking, and guardrails that prevent confident wrong answers. Experienced running containerized ML/LLM services on Kubernetes (including AWS-managed layers) with CI/CD and observability, and has delivered a real-time predictive maintenance system using streaming sensor data and time-series anomaly detection in close partnership with maintenance teams.”
Mid-level AI & Machine Learning Engineer specializing in Generative AI and MLOps
“Built a production GPT-4/LangChain/Pinecone RAG “AI Copilot” at Northern Trust to automate financial report generation and analyst Q&A over internal structured (SQL warehouse) and unstructured policy data. Focused on real-world production challenges—grounding and latency—achieving major speed gains (seconds to milliseconds) via MiniLM embedding optimization and Redis caching, and implemented rigorous testing/evaluation with MLflow-backed metrics while aligning compliance and finance stakeholders for deployment.”
Entry-level AI/ML Engineer specializing in AWS MLOps and computer vision
“Built and shipped a production RAG question-answering system using LangChain/OpenAI, Docker, and FastAPI, then reduced hallucinations through disciplined retrieval tuning and constrained prompting. Also implemented a custom evaluation framework (QA-pair dataset) to measure faithfulness/relevance and deployed containerized ML microservices on AWS ECS/Fargate with ALB and rolling, zero-downtime updates.”
Mid-level Machine Learning Engineer specializing in LLMs, GenAI, and Computer Vision
“LLM/agent engineer who built a production multi-agent research automation system using LangGraph (planner, retriever with FAISS, supervisor, evaluator) with structured outputs and citation tracking for traceable reports. Emphasizes reliability and operations—LangSmith-based observability, multi-level testing, hallucination mitigation, and latency/cost controls—plus prior experience as a Computer Vision Software Engineer at Deepsight AI Labs working directly with non-technical customers.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and multimodal AI on AWS
“Built and deployed a production RAG-based enterprise document intelligence platform for financial/compliance/operational documents on AWS (Spark/Glue ingestion, embeddings + vector DB, LangChain orchestration, REST APIs on Docker/Kubernetes). Deep hands-on experience orchestrating multi-step and multi-agent LLM workflows (LangChain, LangGraph, CrewAI) with strong focus on grounding, evaluation, observability, and cost/latency optimization, and has partnered closely with non-technical finance/compliance teams to drive adoption.”
Intern AI Researcher specializing in NLP, LLMs, and knowledge graphs
“Built and shipped “LabMate,” a production AI assistant specialized in laboratory hardware, using a weighted multi-source RAG pipeline with reranking and reasoning-focused query decomposition to handle complex user questions. Deployed on a local GPU cluster with vLLM and NVIDIA MPS (plus OCR/VLM components), and established evaluation using synthetic + public reasoning datasets while collaborating weekly with non-technical admins to align requirements and resource constraints.”
Mid-level AI/ML Engineer specializing in GenAI and cloud MLOps
“Applied LLMs to high-stakes domains (wildfire risk for emergency teams and loan approval via a fine-tuned IBM Granite model), with a strong focus on reliability—using RAG-based cross-validation to reduce hallucinations and continuous ingestion pipelines (MODIS satellite imagery via AWS Lambda) to keep data current. Experienced in production orchestration and MLOps-style workflows using Airflow, AWS Step Functions, and SageMaker Pipelines, and collaborates closely with analysts on KPI-driven evaluation.”
Principal Data Scientist specializing in cybersecurity ML and MLOps
“ML/NLP engineer (Beyond Identity) who built production semantic search and entity-resolution systems over internal security documentation, using LDA + BERT embeddings with FAISS/Pinecone to cut search time by 30%. Also scaled a real-time anomaly detection pipeline to millions of events/day with Spark and AWS Lambda, with strong emphasis on measurable validation (Precision@k, MRR, F1, ARI).”
Mid-level AI/ML Engineer specializing in fraud detection and NLP
“Built production AI/RAG-style systems for message Q&A and insurance claims workflows, combining data ingestion, indexing/retrieval, and LLM integration with fallback modes. Has hands-on orchestration experience (Airflow, Prefect, LangChain) and cites large operational gains (claims processing reduced to ~45 seconds; manual review -50%; false alerts -30%) through automated, monitored pipelines and close collaboration with non-technical stakeholders.”
Mid-level AI & Data Scientist specializing in LLMs, RAG, and healthcare NLP
“Built a production LLM/RAG solution for healthcare operations teams to query large policy and care-guideline repositories in natural language. Improved domain alignment using vector retrieval plus parameter-efficient fine-tuning and prompt optimization, validated through internal user testing and metrics, cutting manual lookup time by ~40%. Also has hands-on experience orchestrating automated ML pipelines with Apache Airflow.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps for financial services
“Built and deployed a production Llama 3-based RAG document Q&A system using FAISS, addressing context-window limits through chunking and keeping retrieval accurate by regularly refreshing embeddings. Has hands-on orchestration experience with LangChain and LlamaIndex for multi-step LLM workflows (including memory management) and collaborates with non-technical teams (e.g., marketing) to deliver AI solutions like recommendation systems.”
Mid-Level Software Engineer specializing in backend, distributed systems, and AI/LLM platforms
“Built and shipped AI-powered workflow automation at Oracle, including an MCP-based agentic workflow with tool-calling and guardrails, plus Grafana monitoring and Confluence documentation. Also led a Django monolith-to-microservices migration at Chamsmobile using blue-green deployment and load balancer traffic splitting to avoid regressions while modernizing production systems.”
Junior Robotics & Reinforcement Learning Engineer specializing in autonomous systems
“Robotics/ML candidate building an individual pedestrian trajectory forecasting system by adapting a GAN-style Social-GN training architecture from LSTM to a transformer-based AgentFormer design. Also has hands-on embedded robotics experience debugging lane-following behavior on a JetBot by tuning PID control, and uses Docker for reproducible training environments.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Built a production RAG-based healthcare chatbot to retrieve patient medical documents spread across multiple platforms, reducing manual and error-prone searching. Implemented semantic search with custom embeddings (Hugging Face) and Pinecone, deployed via FastAPI/Docker on AWS SageMaker with MLflow tracking, and optimized fine-tuning cost using LoRA while orchestrating retraining pipelines in Airflow.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG
“GenAI/LLM engineer who built and deployed a production RAG system for enterprise document search and decision support, cutting manual lookup time by 40%+. Experienced with LangChain/LangGraph agent orchestration plus Airflow/Prefect for ingestion and incremental reindexing, with a strong focus on reliability (testing, observability) and stakeholder-driven metrics.”
Mid-level AI/ML Engineer specializing in GenAI, MLOps, and anomaly detection
“LLM/MLOps engineer who has shipped a production RAG-based technical documentation assistant (FastAPI) cutting manual review by 45%, with deep hands-on retrieval optimization in Pinecone/LangChain (HNSW, hybrid + multi-query search, caching). Also brings healthcare domain experience—building Airflow-orchestrated EHR pipelines and delivering FDA-auditability-friendly predictive maintenance solutions using SHAP/LIME explainability surfaced in Power BI.”
Mid-level AI Engineer & Data Scientist specializing in LLMs, RAG, and multimodal systems
“LLM/GenAI engineer who built a production AI-powered credit risk policy summarization and compliance alerting platform at HCL Tech, focused on factual accuracy and auditability for a financial client. Implemented a multi-retriever LangChain RAG architecture with citations-only prompting, fallback agents, and human-in-the-loop legal review—cutting manual review time by 35% and scaling to 12 teams.”
Mid-level Machine Learning Engineer specializing in NLP, recommender systems, and MLOps
“ML/LLM engineer with production experience at General Motors building Transformer-based search and recommendation personalization for a high-traffic vehicle platform. Delivered significant KPI gains (17% conversion lift, 14% bounce-rate reduction) and optimized real-time inference via ONNX Runtime and INT8 quantization while implementing robust MLOps (Airflow/MLflow, monitoring, drift-triggered retraining) and stakeholder-facing explainability/dashboards.”