Pre-screened and vetted.
Intern Machine Learning Engineer specializing in NLP, RAG, and deepfake detection
“Early-career (fresher) candidate who built and deployed a production AI medical document chatbot using a RAG architecture (LangChain + Hugging Face LLM + Pinecone) with a Flask backend on AWS EC2 via Docker. Has experience troubleshooting real deployment constraints (model dependencies, disk space, container stability) and setting up continuous-style evaluation with fixed query test sets tracking relevance, latency, and error rate.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and RAG systems
“Machine learning/NLP engineer who built a production-oriented retrieval-based AI system at Morgan Stanley for healthcare use cases, combining RAG over unstructured patient records with deep-learning medical image segmentation (U-Net/Mask R-CNN). Strong in end-to-end pipelines and MLOps (Spark/MongoDB, AWS SageMaker, CI/CD, monitoring, automated retraining) and in entity resolution/data quality validation for noisy clinical data.”
Mid-level Machine Learning Engineer specializing in forecasting, NLP, and GenAI
“GenAI/ML engineer with production experience building multilingual LLM systems (English/Spanish) and RAG-based clinical documentation summarization at Walgreens, combining prompt engineering, structured output validation, and rigorous evaluation (ROUGE + pharmacist review). Also orchestrated end-to-end ML pipelines for demand forecasting using Apache Airflow, PySpark, and MLflow with scheduled retraining and production monitoring.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
Entry Machine Learning Engineer specializing in NLP, computer vision, and recommender systems
“Built and shipped an end-to-end podcast recommendation system exposed via a Flask API and React UI, explicitly balancing relevance, diversity (MMR), and safety constraints while meeting ~200ms latency targets. Also implemented a production-style RAG/information-extraction pipeline using web retrieval, spaCy NER, and fine-tuned SpanBERT with guardrails and evaluation loops (precision/recall/F1) to tune confidence thresholds and improve reliability.”
Mid-level Full-Stack Software Engineer specializing in Java/Spring microservices and React
“Uber engineer who has owned internal products end-to-end across backend (Spring Boot microservices, MySQL) and frontend (React), including performance optimization and secure JWT-based auth. Also shipped a production internal RAG/embeddings LLM support assistant over policy docs and support tickets, with guardrails (confidence thresholds, human review) and an evaluation loop that directly reduced hallucinations.”
Mid-level AI/ML Engineer specializing in MLOps, NLP/LLMs, and computer vision
“Built and shipped a production LLM/RAG risk-case summarization and triage system used by fraud/compliance analysts, with strong grounding controls (evidence-cited outputs and refusal on low confidence). Demonstrates end-to-end ownership across retrieval quality, Airflow-orchestrated indexing pipelines, and compliance-grade privacy (PII redaction, RBAC, encrypted redacted logging, and auditable prompt/model versioning) plus a tight feedback loop with non-technical domain experts.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Mid-level Software Engineer specializing in AI and full-stack healthcare platforms
“Built and deployed a RAG-based clinical knowledge assistant at GE Healthcare to help clinicians query large volumes of messy, unstructured clinical documents with grounded, cited answers. Hands-on across the full stack (OCR/ETL, de-identification for PHI, Azure OpenAI embeddings, Cosmos DB indexing, FastAPI/Django) with production monitoring via LangSmith and performance tuning through batching and index optimization.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Mid-level Supply Chain Analyst specializing in global logistics automation and forecasting
“Built and shipped a production LLM-powered recruiting workflow that ranks resumes against job descriptions, generates evidence-based justifications, and finds "hidden fit" candidates using embeddings + RAG. Demonstrates strong production engineering around hallucination control, latency, and predictable LLM cost management (budget checks, top-K pruning, tenant caps), plus orchestration experience with Airflow/Prefect/Kubernetes and a structured evaluation/monitoring methodology for AI agents.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMOps, and MLOps
“Built and deployed an AWS-based LLM/RAG ticket triage and knowledge retrieval system (Pinecone/FAISS + Step Functions + MLflow) that cut support resolution time by 20%. Demonstrates strong production focus on hallucination reduction, PII security, and low-latency orchestration, with measurable evaluation improvements (e.g., ~25% grounding accuracy gain via re-ranking) and proven collaboration with support operations stakeholders.”
Senior Data Scientist / ML Engineer specializing in cloud ML pipelines and GenAI
“ML/NLP practitioner with experience building a transformer-failure prediction system that combines sensor signals with unstructured maintenance comments using LLM-based extraction and similarity validation. Strong emphasis on production readiness—data leakage controls, SQL-driven data quality tiers, and rigorous bias/fairness validation (including contract/spec evaluation across diverse company profiles).”
Mid-level Data Scientist specializing in NLP/LLMs, time series forecasting, and MLOps
“Data/ML practitioner with hands-on experience building NLP systems from prototype to production: delivered a Twitter sentiment classifier with robust preprocessing, SVM modeling, and Power BI reporting, and built entity-resolution pipelines for messy multi-source customer data (reporting ~95% improvement in unique entity identification). Also implemented semantic linking/search using SBERT embeddings with FAISS vector retrieval and domain fine-tuning (reported ~15% precision lift), and applies production workflow best practices (Airflow/Prefect, Docker, Azure ML/Databricks, Great Expectations).”
Mid-level Full-Stack Developer specializing in cloud microservices and internal tooling
“LLM/RAG engineer who has shipped production systems in high-stakes domains (fraud analytics at Mastercard and security compliance as a CI/CD gate). Strong focus on reliability: hybrid retrieval for latency, citation-backed outputs for trust, and code-driven eval/regression pipelines using golden datasets. Also built scalable OCR-based ingestion for messy classroom artifacts (handwriting, PDFs, whiteboard photos) using Go/Python and cloud services.”
Mid-level AI/ML Engineer specializing in GenAI, LLMs, RAG, and MLOps
“Built and deployed a production LLM-powered RAG document intelligence/Q&A system for healthcare prior authorization, reducing manual medical document review time and improving decision efficiency. Strong in end-to-end LLM application engineering (LangChain/LangGraph), retrieval quality improvements (hybrid search, embedding tuning, chunking strategies), and rigorous evaluation/monitoring for reliability.”
Junior Software Engineer specializing in ML, distributed systems, and LLM applications
“Interned at Zonda where he built an AI-driven semantic search solution over ~280M housing/builder records. Iterated from local LLMs via llama.cpp quantization to a vector-embedding retrieval system, then boosted semantic accuracy with a custom spaCy NER layer and re-ranking, optimizing for latency through precomputation. Collaborated with economics-focused stakeholders to reduce manual document/paperwork time by enabling natural-language search over internal data.”
Mid-level Machine Learning Engineer specializing in NLP and cloud MLOps
“Built and deployed a production LLM-powered internal documentation assistant using embeddings, a vector database, and a RAG pipeline to reduce time spent searching PDFs/manuals. Experienced in orchestrating end-to-end LLM workflows with Airflow/LangChain, improving reliability via monitoring/error handling, and driving measurable quality through retrieval and hallucination-focused evaluation metrics.”
“Built and productionized an LLM-powered PDF document Q&A system to eliminate manual searching through long documents, focusing on scalability and answer reliability. Implemented semantic chunking (using headings/paragraphs/tables), overlap, and preprocessing/quality checks to reduce hallucinations, and orchestrated the end-to-end pipeline with Airflow using retries, alerts, and parallel tasks.”
Junior Software Engineer specializing in AI, security, and cloud systems
“Built and deployed an LLM + RAG + memory system on a Furhat social robot, adding continuous face/voice recognition embeddings over WebSockets to enable persistent, natural conversations across sessions. Experienced working around real-world hardware/latency constraints and uses Datadog plus structured debugging/rollback practices for stabilizing customer-facing LLM workflows.”
Principal Software Engineer specializing in AI/ML and cloud-native backend systems
“McKinsey data/ML practitioner who led production deployment of an entity resolution + semantic search platform for unstructured finance and healthcare data, integrating with legacy systems under HIPAA constraints. Deep hands-on stack across transformers (spaCy/HF BERT), embeddings + FAISS, and production MLOps/workflow tooling (Airflow, Docker, CI/CD, Prometheus/Grafana), with reported gains of +30% decision speed and +25% search relevance.”
Mid-Level Software Engineer specializing in LLM-powered developer tools
“Built and owned "Cortex," an AI agent that helps users understand large GitHub repositories by mapping architecture and relationships between files/folders in minutes. Implemented an agentic, multi-stage prompt decomposition approach and validated it across open-source repos, while also doing legacy service modernization work involving dependency upgrades and refactors.”