Pre-screened and vetted.
Intern Full-Stack/ML Engineer specializing in LLM applications and mobile development
“Backend engineer who built a serverless AWS Lambda microservices backend for a parenting assistance mobile app, including a personalized recommendation system optimized to sub-500ms via precomputed scoring and DynamoDB caching. Demonstrates strong production pragmatism: CloudWatch-driven performance tuning (provisioned concurrency), zero-downtime phased schema migrations, and robustness patterns like optimistic locking and request deduplication. Also led a refactor of an LLM RAG pipeline to improve retrieval quality and cut latency from ~5s to ~3s.”
Mid-level AI/ML Engineer specializing in LLM agents, RAG, and enterprise ML systems
“Built a production multi-agent recommendation/RAG system for internal data analysts to speed up weekly report creation by improving document discovery and automating report/SQL generation. Implemented LangGraph-based orchestration with deterministic agent routing, robust error handling (interrupt/resume), and metadata-driven semantic chunking for diverse PDF/document formats, plus monitoring for latency, throughput, and token/cost efficiency.”
“Built and deployed a production Retrieval-Augmented Generation (RAG) platform in a healthcare setting to automate clinical documentation review and summarization, targeting near-real-time, explainable outputs. Emphasizes grounded generation to reduce hallucinations, latency optimizations (chunking/embedding reuse), and PHI-safe workflows with access controls, plus strong orchestration experience using Apache Airflow.”
Mid-level Machine Learning Engineer specializing in NLP, LLMs, and multimodal modeling
“Built and productionized a telecom-focused RAG assistant by LoRA fine-tuning LLaMA-2 and integrating LangChain+FAISS behind a FastAPI service, with dashboards and a human feedback UI for engineers. Demonstrated measurable impact (≈40% faster document lookup, +8–10% retrieval precision) and strong MLOps rigor via Airflow orchestration, CI/CD, and monitoring for drift and failures.”
Director-level AI & Data Science leader specializing in GenAI, LLMs, and MLOps
“ML/NLP engineer currently working in NYC on a system that connects complex unstructured data sources to deliver personalized insights, using embeddings + vector DB retrieval and a RAG architecture (LangChain, Pinecone/OpenSearch). Strong focus on production constraints—especially low-latency retrieval—using FAISS/ANN, PCA, index partitioning, and Redis caching, plus PEFT fine-tuning (LoRA/QLoRA) and KPI/SLA-driven promotion to production.”
Principal Data Scientist specializing in healthcare analytics and medical imaging AI
“Developed an LLM-driven recommendation agent in Azure Databricks to triage oncology patients and trigger second-opinion case creation using medical claims and EHR data. Uses ICD-10/CPT/J-code features in prompts, embeddings + vector DB similarity, and a backtesting framework emphasizing recall to avoid missing clinically relevant cases while supporting business revenue.”
Junior AI/ML Engineer specializing in real-time computer vision and tracking systems
“Full-stack engineer who built and owned a production real-time computer-vision inference platform at Credence, spanning Next.js App Router/TypeScript frontend with SSE/WebSocket streaming, a Flask backend, and Postgres analytics. Demonstrated measurable performance wins (70% fewer re-renders; latency cut to ~40–50ms) and strong production rigor (durable orchestration, idempotency, observability, AWS EC2 + CI/CD) with tight post-launch UX iteration based on analyst feedback.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“AI engineer who built a production RAG-based internal analyst tool at BlackRock, fine-tuning an LLM on proprietary financial data and adding four layers of guardrails (input/retrieval/generation/output) to improve grounding and reduce hallucinations. Implemented a LangChain-based multi-agent orchestration (7 major agents) deployed on AWS ECS, with reliability measured via internal human evaluation, LLM-as-judge, and RLHF/drift monitoring.”
“LLM engineer who has deployed production RAG systems for regulated document QA (PDFs/knowledge bases), emphasizing grounded answers with citations, RBAC, monitoring, and continuous feedback. Demonstrates deep practical expertise in retrieval quality (semantic chunking, hybrid BM25+embeddings, re-ranking), reliability (guardrails, deterministic workflows), and measurable evaluation (golden sets, log replay, A/B tests) while partnering closely with compliance/operations stakeholders.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Intern AI/ML Engineer specializing in RAG, multimodal AI, and LLM systems
“Built and shipped 'PetPulse,' a production AI pet-health note system that records voice notes, transcribes them, converts transcripts into structured symptom/event data, and supports grounded Q&A over a user’s notes and vet PDFs. Demonstrates full-stack LLM product execution (FastAPI + GPT-4 + Firebase), with concrete reliability/performance work (async endpoints, caching, RAG/embeddings, function calling) and user-centered iteration with a non-technical product stakeholder.”
Mid-level Machine Learning Engineer specializing in LLMs and NLP classification systems
“Internship experience building a production RAG+LLM pipeline to map messy card transaction descriptions to merchant brands, including a custom modified-ROUGE evaluation approach for weak/variant ground truth. Improved scalability and cost by moving from a managed LLM endpoint (e.g., Bedrock) to self-hosted vLLM, and orchestrated massive embedding backfills (5,000+ files, 10B+ rows) using an Airflow-triggered SQS + ECS worker architecture with robust retry/DLQ handling.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
Entry Machine Learning Engineer specializing in NLP, computer vision, and recommender systems
“Built and shipped an end-to-end podcast recommendation system exposed via a Flask API and React UI, explicitly balancing relevance, diversity (MMR), and safety constraints while meeting ~200ms latency targets. Also implemented a production-style RAG/information-extraction pipeline using web retrieval, spaCy NER, and fine-tuned SpanBERT with guardrails and evaluation loops (precision/recall/F1) to tune confidence thresholds and improve reliability.”
Mid-level Machine Learning Engineer specializing in MLOps, NLP, and production ML systems
“Backend/founding-engineer-style builder who designed and evolved a near-real-time customer churn prediction platform (FastAPI + AWS SageMaker/Lambda + Redis + MLflow) to enable real-time retention actions, reporting ~18% churn reduction. Demonstrates strong production engineering in secure API design, incremental migrations with data integrity safeguards, and robustness improvements in async pipelines (idempotency, DLQs, retry visibility).”
Entry-Level Software Engineer specializing in ML and backend systems
“Built and deployed a production LLM-based real-time stance detection system for social media, fine-tuning LLaMA 3.1 on A100s with DeepSpeed ZeRO/FSDP and iteratively refining data to handle sarcasm and context-dependent meaning. Also has Kubernetes operations experience (Kafka/Logstash/Elasticsearch observability pipeline) and delivered an OCR automation project during a Worley India internship that saved 20+ hours/week for on-site energy safety stakeholders.”
Mid-level Full-Stack & GenAI Engineer specializing in RAG and LLM applications
“Software engineer working on an e-commerce platform, currently building a RAG-based recommendation system with a team new to the technology. Has delivered an end-to-end React/TypeScript website for a local car dealer and built an internal "encryption as a service" tool to secure sensitive data across repositories and through release/UAT, with experience debugging microservices integration issues.”
Mid-level Data/ML Engineer specializing in NLP, GenAI, and scalable data pipelines
“AI/ML engineer with production experience building LLM-powered document intelligence and customer support systems in healthcare/insurance, emphasizing high-accuracy RAG, long-document processing, and robust monitoring/fallback mechanisms. Also automates and scales ML lifecycle workflows using Apache Airflow and Kubeflow, and partners closely with non-technical operations stakeholders to drive adoption.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMOps, and MLOps
“Built and deployed an AWS-based LLM/RAG ticket triage and knowledge retrieval system (Pinecone/FAISS + Step Functions + MLflow) that cut support resolution time by 20%. Demonstrates strong production focus on hallucination reduction, PII security, and low-latency orchestration, with measurable evaluation improvements (e.g., ~25% grounding accuracy gain via re-ranking) and proven collaboration with support operations stakeholders.”
Mid-level Software Engineer specializing in AI/ML and data platforms
“AI/ML engineer who built a production agentic system to automate computational research experiments (simulation execution, parameter exploration, and numerical analysis) and mitigated context-window failures using constrained tool-calling/prompt-chaining patterns in LangChain with OpenAI tool-enabled models. Also has adtech/big-data pipeline experience at InMobi, orchestrating Spark jobs in Airflow to filter bot-like user IDs and publish clean IDs to an online NoSQL store for live serving, plus Apache open-source collaboration experience.”
Intern Machine Learning Engineer specializing in LLMs, MLOps, and NLP
“Built and deployed a production LLM-driven Dungeons & Dragons game where the model acts as a dungeon master, adding a structured combat system and a macro-state tree to ensure campaigns converge to a clear ending. Fine-tuned Gemini 2.5 Flash on Vertex AI and deployed on GCP with Kubernetes, using RAG over DnD rules/spells plus multi-agent orchestration (intent-based routing between narrative and combat agents) to reduce hallucinations and improve reliability.”
Senior AI/ML Engineer specializing in Generative AI and LLM platforms
“Backend engineer focused on multi-tenant enterprise AI personalization and recommendation platforms, combining ML/LLM intent extraction with deterministic policy guardrails for compliance and auditability. Has hands-on AWS experience (ECS/Lambda/DynamoDB/S3) and led a careful DynamoDB single-table migration using dual write/read, canary + feature-flag rollouts, and strong observability/security (JWT/OAuth2, RBAC, Postgres RLS).”