Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
“GenAI/data engineering practitioner with production experience across Equinix, Optum, and Citibank—built an Azure OpenAI (GPT-4) + LangChain document intelligence platform processing 1.5M+ docs/month and a HIPAA-compliant Airflow healthcare pipeline handling 5M+ claims/day. Also delivered a real-time fraud detection + explainability system using LightGBM and a fine-tuned T5 NLG component, improving fraud accuracy by 15%+ while partnering closely with compliance stakeholders.”
Mid-level Machine Learning Engineer specializing in LLMs and NLP classification systems
“Internship experience building a production RAG+LLM pipeline to map messy card transaction descriptions to merchant brands, including a custom modified-ROUGE evaluation approach for weak/variant ground truth. Improved scalability and cost by moving from a managed LLM endpoint (e.g., Bedrock) to self-hosted vLLM, and orchestrated massive embedding backfills (5,000+ files, 10B+ rows) using an Airflow-triggered SQS + ECS worker architecture with robust retry/DLQ handling.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and RAG systems
“Machine learning/NLP engineer who built a production-oriented retrieval-based AI system at Morgan Stanley for healthcare use cases, combining RAG over unstructured patient records with deep-learning medical image segmentation (U-Net/Mask R-CNN). Strong in end-to-end pipelines and MLOps (Spark/MongoDB, AWS SageMaker, CI/CD, monitoring, automated retraining) and in entity resolution/data quality validation for noisy clinical data.”
Mid-level Machine Learning Engineer specializing in forecasting, NLP, and GenAI
“GenAI/ML engineer with production experience building multilingual LLM systems (English/Spanish) and RAG-based clinical documentation summarization at Walgreens, combining prompt engineering, structured output validation, and rigorous evaluation (ROUGE + pharmacist review). Also orchestrated end-to-end ML pipelines for demand forecasting using Apache Airflow, PySpark, and MLflow with scheduled retraining and production monitoring.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
“ServiceNow engineer who built and launched a production LLM-powered ticket resolution/knowledge assistant using RAG (LangChain + Hugging Face embeddings + vector search) integrated into internal support dashboards via REST APIs. Optimized the system from ~6–8s to ~2–3s latency while improving usability with concise, cited answers and guardrails (grounding + similarity thresholds), delivering ~30–35% reduction in manual ticket investigation effort.”
Mid-level Machine Learning Engineer specializing in MLOps, NLP, and production ML systems
“Backend/founding-engineer-style builder who designed and evolved a near-real-time customer churn prediction platform (FastAPI + AWS SageMaker/Lambda + Redis + MLflow) to enable real-time retention actions, reporting ~18% churn reduction. Demonstrates strong production engineering in secure API design, incremental migrations with data integrity safeguards, and robustness improvements in async pipelines (idempotency, DLQs, retry visibility).”
Mid-level AI/ML Engineer specializing in MLOps, NLP/LLMs, and computer vision
“Built and shipped a production LLM/RAG risk-case summarization and triage system used by fraud/compliance analysts, with strong grounding controls (evidence-cited outputs and refusal on low confidence). Demonstrates end-to-end ownership across retrieval quality, Airflow-orchestrated indexing pipelines, and compliance-grade privacy (PII redaction, RBAC, encrypted redacted logging, and auditable prompt/model versioning) plus a tight feedback loop with non-technical domain experts.”
Mid-level AI/ML Engineer specializing in MLOps, computer vision, and NLP
“GenAI/ML engineer from Lucid Motors who built and productionized an LLM-powered RAG diagnostic assistant for manufacturing and maintenance teams, deployed on AWS with Docker/Kubernetes and MLflow. Demonstrates end-to-end ownership from retrieval/prompt design to scalability, monitoring, and workflow integration via APIs, plus production ML pipeline orchestration with Kubeflow (Spark/Kafka + TensorFlow) for predictive maintenance use cases.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Mid-level Data/ML Engineer specializing in NLP, GenAI, and scalable data pipelines
“AI/ML engineer with production experience building LLM-powered document intelligence and customer support systems in healthcare/insurance, emphasizing high-accuracy RAG, long-document processing, and robust monitoring/fallback mechanisms. Also automates and scales ML lifecycle workflows using Apache Airflow and Kubeflow, and partners closely with non-technical operations stakeholders to drive adoption.”
Junior Machine Learning Engineer specializing in multimodal AI and audio deepfakes detection
“Internship experience building production-oriented AI systems, including a real-time voice scam/spoof detector (RawNet + AASIST) hardened for noisy audio via aggressive augmentation and Zoom-based noise simulation, evaluated with EER on clean and wild datasets. Also built an LLM-driven UI automation agent using Playwright for apps like Linear/Notion with modular tool design, unit tests, and replayable scripted scenarios, and has AWS Step Functions experience orchestrating Lambda/Cognito workflows.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMOps, and MLOps
“Built and deployed an AWS-based LLM/RAG ticket triage and knowledge retrieval system (Pinecone/FAISS + Step Functions + MLflow) that cut support resolution time by 20%. Demonstrates strong production focus on hallucination reduction, PII security, and low-latency orchestration, with measurable evaluation improvements (e.g., ~25% grounding accuracy gain via re-ranking) and proven collaboration with support operations stakeholders.”
Senior Applied AI/ML Engineer specializing in GenAI, LLMs, RAG and agents
“Applied AI/ML Engineer at JPMorgan Chase who led a banker-facing LLM chatbot from an OpenAI-API POC to a production RAG workflow, including hallucination mitigation, automated evaluation in SageMaker, and operational monitoring with Dynatrace. Also delivers external technical education—hosted a hands-on Grace Hopper Celebration 2025 workshop teaching LangChain/LangGraph agentic workflows.”
Junior Machine Learning Engineer specializing in LLMs and applied data science
“Built and shipped multiple production AI systems, including Auto DocGen (LLM-generated OpenAPI docs kept in sync via AST diffs, schema-constrained generation, and CI/CD on Render) and a multimodal sign-language recognition pipeline at USC orchestrated with FastAPI, MediaPipe, and PyTorch. Also partnered with Esri’s non-technical community team to fine-tune an LLaMA-based spam classifier with a review UI, cutting moderation time by 70%.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Red Hat ML/LLM engineer who designed and deployed a production LLM-powered customer support automation system using RAG, improving latency by 30% via PEFT and vector search optimization. Built security and governance into retrieval (access-level filtering, encrypted Pinecone/ChromaDB) and delivered SHAP-based explainability via a dashboard for non-technical stakeholders. Experienced orchestrating distributed ML/RAG pipelines across AWS SageMaker and OpenShift with Airflow/Prefect, plus multi-agent workflows using CrewAI and LangGraph.”
Intern Software Engineer specializing in backend systems, cloud infrastructure, and ML/LLM tooling
“Infrastructure-leaning engineer who has built real-time ML systems end-to-end: a Jetson-deployed adaptive Whisper ASR service (Flask + WebSockets, React/TS UI) and a high-throughput Postgres schema for live transcription. Also delivered customer-facing AI billing/OCR improvements for a dental startup (Dentite), boosting OCR performance by 38%, and has experience instrumenting open-source ML deployment stacks to add infrastructure visibility.”
Mid-level AI/ML Engineer specializing in GenAI, LLMs, RAG, and MLOps
“Built and deployed a production LLM-powered RAG document intelligence/Q&A system for healthcare prior authorization, reducing manual medical document review time and improving decision efficiency. Strong in end-to-end LLM application engineering (LangChain/LangGraph), retrieval quality improvements (hybrid search, embedding tuning, chunking strategies), and rigorous evaluation/monitoring for reliability.”
Mid-level AI/ML Engineer specializing in risk, fraud detection, and Generative AI
“Built and deployed an LLM-powered RAG document intelligence/search platform for banking risk & compliance teams, emphasizing sensitive-data handling, traceability, and conservative fallback logic to minimize hallucinations; deployed via Docker/REST on AWS and cut manual review effort by 35%. Also partnered with TD Bank marketing to deliver an AI customer segmentation solution that improved targeted campaign engagement by 18%.”
Senior AI/ML Engineer specializing in Generative AI and LLM platforms
“Backend engineer focused on multi-tenant enterprise AI personalization and recommendation platforms, combining ML/LLM intent extraction with deterministic policy guardrails for compliance and auditability. Has hands-on AWS experience (ECS/Lambda/DynamoDB/S3) and led a careful DynamoDB single-table migration using dual write/read, canary + feature-flag rollouts, and strong observability/security (JWT/OAuth2, RBAC, Postgres RLS).”