Pre-screened and vetted.
Intern Machine Learning Engineer specializing in NLP, RAG, and deepfake detection
“Early-career (fresher) candidate who built and deployed a production AI medical document chatbot using a RAG architecture (LangChain + Hugging Face LLM + Pinecone) with a Flask backend on AWS EC2 via Docker. Has experience troubleshooting real deployment constraints (model dependencies, disk space, container stability) and setting up continuous-style evaluation with fixed query test sets tracking relevance, latency, and error rate.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and RAG systems
“Machine learning/NLP engineer who built a production-oriented retrieval-based AI system at Morgan Stanley for healthcare use cases, combining RAG over unstructured patient records with deep-learning medical image segmentation (U-Net/Mask R-CNN). Strong in end-to-end pipelines and MLOps (Spark/MongoDB, AWS SageMaker, CI/CD, monitoring, automated retraining) and in entity resolution/data quality validation for noisy clinical data.”
Mid-level Full-Stack Software Engineer specializing in Java/Spring microservices and React
“Uber engineer who has owned internal products end-to-end across backend (Spring Boot microservices, MySQL) and frontend (React), including performance optimization and secure JWT-based auth. Also shipped a production internal RAG/embeddings LLM support assistant over policy docs and support tickets, with guardrails (confidence thresholds, human review) and an evaluation loop that directly reduced hallucinations.”
Mid-level GenAI/ML Engineer specializing in LLM applications and enterprise automation
“Built and shipped a production LLM-powered healthcare support agent at UnitedHealthGroup, using LangChain + FAISS RAG on AWS SageMaker with CloudWatch monitoring and human-in-the-loop fallbacks for safety. Strong focus on reliability engineering (confidence gating, retries/timeouts, caching) and continuous evaluation loops; reported ~40% improvement in query resolution efficiency while reducing manual support workload.”
Junior Machine Learning Engineer specializing in LLMs, RAG, and medical imaging
“At Fileread, the candidate built and deployed an LLM-powered legal document classification and retrieval layer for an agentic extraction system that turns unstructured legal PDFs into structured tables with line-level citations. They productionized a RAG-style pipeline (ingestion, embeddings, retrieval, reranking, generation) and report 95%+ F1 across 70+ legal categories, emphasizing rigorous evaluation and close collaboration with legal domain experts for high-stakes precision.”
Mid-level Machine Learning Engineer specializing in forecasting, NLP, and GenAI
“GenAI/ML engineer with production experience building multilingual LLM systems (English/Spanish) and RAG-based clinical documentation summarization at Walgreens, combining prompt engineering, structured output validation, and rigorous evaluation (ROUGE + pharmacist review). Also orchestrated end-to-end ML pipelines for demand forecasting using Apache Airflow, PySpark, and MLflow with scheduled retraining and production monitoring.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
Entry Machine Learning Engineer specializing in NLP, computer vision, and recommender systems
“Built and shipped an end-to-end podcast recommendation system exposed via a Flask API and React UI, explicitly balancing relevance, diversity (MMR), and safety constraints while meeting ~200ms latency targets. Also implemented a production-style RAG/information-extraction pipeline using web retrieval, spaCy NER, and fine-tuned SpanBERT with guardrails and evaluation loops (precision/recall/F1) to tune confidence thresholds and improve reliability.”
Mid-level Machine Learning Engineer specializing in AI/LLM systems
“ML/LLM systems engineer who has owned AI support automation products end-to-end, including ServiceNow-integrated incident routing, RAG-based resolution suggestion systems, and production stabilization. Stands out for combining hands-on platform work across PySpark, AWS Glue, FastAPI, Kubernetes, and Pinecone with measurable operational impact, including 30-35% MTTR reduction and 25-30% improvement in first-touch resolution.”
Mid-level AI/ML Engineer specializing in Generative AI and financial services
“ML/AI engineer with hands-on experience shipping regulated financial AI systems at JPMC and Capgemini, spanning credit risk, fraud detection, and generative AI assistants. Stands out for combining modern LLM/RAG architectures with strong MLOps, real-time infrastructure, and explainability/compliance practices, while delivering measurable business impact in latency, accuracy, cost, and risk reduction.”
Junior Software Engineer specializing in full-stack and AI-powered web systems
“Built the backend for “codeGuard,” an AI-powered static code analysis platform, using FastAPI and Docker. Structured the system into API/service/execution layers and addressed heavy-workload container resource/cleanup issues via strict CPU/memory limits and a queued execution model.”
Mid-level AI/ML Engineer specializing in MLOps, NLP/LLMs, and computer vision
“Built and shipped a production LLM/RAG risk-case summarization and triage system used by fraud/compliance analysts, with strong grounding controls (evidence-cited outputs and refusal on low confidence). Demonstrates end-to-end ownership across retrieval quality, Airflow-orchestrated indexing pipelines, and compliance-grade privacy (PII redaction, RBAC, encrypted redacted logging, and auditable prompt/model versioning) plus a tight feedback loop with non-technical domain experts.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Mid-level Software Engineer specializing in AI and full-stack healthcare platforms
“Built and deployed a RAG-based clinical knowledge assistant at GE Healthcare to help clinicians query large volumes of messy, unstructured clinical documents with grounded, cited answers. Hands-on across the full stack (OCR/ETL, de-identification for PHI, Azure OpenAI embeddings, Cosmos DB indexing, FastAPI/Django) with production monitoring via LangSmith and performance tuning through batching and index optimization.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Senior Data Scientist / ML Engineer specializing in cloud ML pipelines and GenAI
“ML/NLP practitioner with experience building a transformer-failure prediction system that combines sensor signals with unstructured maintenance comments using LLM-based extraction and similarity validation. Strong emphasis on production readiness—data leakage controls, SQL-driven data quality tiers, and rigorous bias/fairness validation (including contract/spec evaluation across diverse company profiles).”
Mid-level Data Scientist specializing in NLP/LLMs, time series forecasting, and MLOps
“Data/ML practitioner with hands-on experience building NLP systems from prototype to production: delivered a Twitter sentiment classifier with robust preprocessing, SVM modeling, and Power BI reporting, and built entity-resolution pipelines for messy multi-source customer data (reporting ~95% improvement in unique entity identification). Also implemented semantic linking/search using SBERT embeddings with FAISS vector retrieval and domain fine-tuning (reported ~15% precision lift), and applies production workflow best practices (Airflow/Prefect, Docker, Azure ML/Databricks, Great Expectations).”
Mid-level Full-Stack Developer specializing in cloud microservices and internal tooling
“LLM/RAG engineer who has shipped production systems in high-stakes domains (fraud analytics at Mastercard and security compliance as a CI/CD gate). Strong focus on reliability: hybrid retrieval for latency, citation-backed outputs for trust, and code-driven eval/regression pipelines using golden datasets. Also built scalable OCR-based ingestion for messy classroom artifacts (handwriting, PDFs, whiteboard photos) using Go/Python and cloud services.”
Mid-level AI/ML Engineer specializing in GenAI, LLMs, RAG, and MLOps
“Built and deployed a production LLM-powered RAG document intelligence/Q&A system for healthcare prior authorization, reducing manual medical document review time and improving decision efficiency. Strong in end-to-end LLM application engineering (LangChain/LangGraph), retrieval quality improvements (hybrid search, embedding tuning, chunking strategies), and rigorous evaluation/monitoring for reliability.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMOps, and MLOps
“Built and deployed an AWS-based LLM/RAG ticket triage and knowledge retrieval system (Pinecone/FAISS + Step Functions + MLflow) that cut support resolution time by 20%. Demonstrates strong production focus on hallucination reduction, PII security, and low-latency orchestration, with measurable evaluation improvements (e.g., ~25% grounding accuracy gain via re-ranking) and proven collaboration with support operations stakeholders.”
Mid-level Supply Chain Analyst specializing in global logistics automation and forecasting
“Built and shipped a production LLM-powered recruiting workflow that ranks resumes against job descriptions, generates evidence-based justifications, and finds "hidden fit" candidates using embeddings + RAG. Demonstrates strong production engineering around hallucination control, latency, and predictable LLM cost management (budget checks, top-K pruning, tenant caps), plus orchestration experience with Airflow/Prefect/Kubernetes and a structured evaluation/monitoring methodology for AI agents.”
Junior Software Engineer specializing in AI/ML and cloud platforms
“LLM/agent engineer who shipped a production "Memory Assistant" at HydroX AI, building a LangChain/LlamaIndex RAG memory pipeline on ChromaDB/FAISS with robust fallbacks (BERT/BART), prompt-injection mitigation, and 99.9% uptime monitoring. Also built a multi-step customer support agent using Rasa + OpenAI Assistants API with structured tool calling, guardrails, and human-in-the-loop escalation, and has experience hardening agents against messy ERP data via Pydantic validation, idempotency, and transactional outbox patterns.”