Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in GenAI and predictive modeling
“Built and deployed a GPT-4-powered medical assistant for clinical staff to reduce time spent searching guidelines and EHR information, with a strong emphasis on safety and compliance. Uses strict RAG, confidence thresholds, and fallback behaviors to prevent hallucinations, and runs production-grade workflows orchestrated with LangChain/LangGraph plus Docker/Kubernetes/MLflow and monitoring for reliability and cost.”
Mid-level Machine Learning Engineer specializing in Generative AI and RAG systems
“LLM/ML engineer who has shipped an enterprise RAG-based Q&A system (LangChain/LlamaIndex, FAISS + Azure Cognitive Search, GPT-3.5/4 via OpenAI/Azure OpenAI) to production on Docker + Kubernetes/OpenShift, tackling hallucinations, retrieval quality, latency/cost, and RBAC/IAM security. Also partnered with operations leaders to turn manual reporting into an LLM-powered summarization and forecasting dashboard driven by real KPIs and iterative stakeholder feedback.”
Mid-level Data Scientist specializing in AI/ML, MLOps, and LLM-powered analytics
“Built and deployed a production LLM-powered document Q&A system enabling natural-language querying of large PDFs, focusing on retrieval quality (overlapped chunking) and low-latency performance (optimized embeddings + vector search). Experienced with scaling ML/LLM workflows using async/batch processing, caching, cloud storage, and orchestration via Apache Airflow with robust testing, monitoring, and failure handling.”
Mid-level AI/ML Engineer specializing in healthcare, fraud detection, and recommender systems
“Healthcare-focused applied ML/LLM engineer who has deployed production systems including an LLM medical documentation assistant that summarizes unstructured EHR notes into physician-ready structured outputs. Experienced building secure, compliant pipelines (PHI minimization, RBAC, encryption) and scaling via Docker/Kubernetes/Azure ML, plus orchestrating ETL/ML workflows with Airflow and Kubeflow; also built an LLM-driven clinical coding assistant at Centene with measurable performance metrics.”
Mid-level AI/ML Engineer specializing in Generative AI and healthcare data
“Built and deployed a production RAG-based document Q&A system on Azure OpenAI to help business teams search thousands of PDFs/Word files, using Qdrant vector search, MongoDB, and a Flask API. Demonstrates strong production engineering (streaming large-file ingestion, parallel preprocessing, monitoring/retries) plus systematic prompt/embedding/chunking experimentation to improve accuracy and reduce hallucinations, and has hands-on orchestration experience with ADF/Airflow/Databricks/Synapse.”
Mid-level Data & Machine Learning Engineer specializing in production ML and data platforms
“Built and deployed a production LLM system that scraped Google Maps menu photos, extracted structured prices via OpenAI, and cross-validated them against website-scraped data to automate data-quality verification at scale (replacing costly manual contractor checks). Demonstrates strong reliability instincts—precision-first prompting, output gating with image-quality metadata, and fuzzy matching/RAG techniques—plus solid orchestration (Dagster/Airflow) and observability (Sentry, Prometheus/Grafana).”
Senior AI/ML Engineer specializing in Generative AI and RAG
“ML/NLP practitioner at Morf Health focused on unifying fragmented healthcare data by linking structured patient/encounter records with unstructured clinical notes. Has hands-on experience with transformer embeddings, vector databases, and domain fine-tuning, plus rigorous evaluation (precision/recall) and human-in-the-loop validation with clinical SMEs to make pipelines production-grade.”
Mid-level AI/ML Engineer specializing in healthcare analytics and MLOps
“AI/ML engineer at Cigna Healthcare building a production, HIPAA-compliant LLM-powered clinical insights platform that summarizes unstructured medical notes using a fine-tuned transformer + RAG on AWS. Demonstrates strong end-to-end MLOps and cloud optimization (distillation, Spot/Lambda/Auto Scaling) with quantified outcomes (~28% accuracy lift, ~40% less manual review, ~25% lower ops cost) and strong clinician-facing explainability via SHAP and dashboards.”
Mid-level Generative AI Engineer specializing in LLM systems and RAG
“Currently at Huntington Bank, built a production-grade RAG system that helps business/operations teams get grounded answers from large volumes of internal enterprise documents. Owns ingestion and FastAPI backend, tuned hybrid BM25+vector retrieval and chunking for relevance, and evaluates reliability with metrics and observability (LangSmith, CloudWatch, Prometheus/Grafana) while partnering closely with non-technical stakeholders.”
Mid-level Backend Software Engineer specializing in distributed microservices
“Internship at ActiveVM where they tackled large-scale Spring Boot 2→3/library migrations across hundreds of downstream products by combining OpenRewrite (AST-based recipes) with an LLM/RAG-based classifier that routed risky files to human experts. Reported ~70% reduction in manual effort and 90%+ accuracy after testing across multiple branches and cutovers; also built a CTR-driven book recommendation capstone showcased at the Google office in Cambridge.”
Junior Software Engineer specializing in Full-Stack and ML for FinTech
“Full-stack engineer with fintech trading-platform experience who shipped and operated a real-time portfolio P&L/performance feature end-to-end (React + Node/WebSockets + MongoDB) on AWS, including significant performance tuning under peak trading load. Also built a Spark-based trading analytics pipeline with idempotency and reconciliation for auditability, and has a personal React/TS + Node/Express project (Artsy) with JWT auth and schema-evolution practices.”
Mid-level Machine Learning Engineer specializing in LLM-powered products
“Verizon engineer who productionized an LLM-based personalization capability for a customer-facing digital platform, owning the path from success metrics through scalable APIs, A/B validation, and post-launch monitoring (latency/accuracy/drift). Experienced in diagnosing and fixing real-time LLM/RAG workflow issues under peak load, and in enabling adoption via tailored technical demos/workshops and sales support materials.”
“Backend engineer focused on productionizing LLM systems: built a FastAPI-based RAG and multi-agent automation platform deployed with Docker/Kubernetes, prioritizing safe execution and reduced hallucinations. Experienced in refactoring monolithic ML services with feature-flagged incremental rollouts, and implementing JWT/RBAC plus row-level security (e.g., Supabase) for secure, scalable APIs.”
Mid-level Robotics Software Engineer specializing in autonomous perception and sensor fusion
“Robotics engineer with Honeywell and Tata Motors experience deploying ROS/ROS2 autonomous mobile robot fleets into live factory environments, integrating sensors, safety PLCs, and on-prem services. Known for solving end-to-end latency and stability issues (including network spikes under load) using gRPC, Docker, and improved diagnostics—cutting diagnosis time from hours to minutes and achieving sub-150 ms control response.”
Senior GenAI/ML Engineer specializing in LLMs, RAG, and multimodal generative AI
“LLM/RAG engineer with production deployments in highly regulated domains (Frost Bank and GE Healthcare). Built secure, explainable document-grounded Q&A systems using LoRA fine-tuning, strict RAG with confidence thresholds, and citation-based responses; also established evaluation/monitoring (golden QA sets, hallucination tracking, drift) and achieved ~40% latency reduction through retrieval/prompt tuning.”
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and Clinical AI
“Built and productionized a HIPAA-compliant LLM+RAG Clinical AI assistant at Optum, fine-tuning GPT/LLaMA on de-identified patient notes and integrating FAISS/Pinecone for sub-second retrieval; reported to cut diagnosis time by ~20 minutes per case. Experienced in orchestrating ML pipelines (Airflow, AWS Step Functions, Azure Data Factory) and in reliability techniques for LLM systems (grounding, citations, confidence filters, monitoring) while partnering closely with clinicians and compliance teams.”
Mid-level AI/ML Engineer specializing in Generative AI and LLMOps
“Built and deployed a GPT-based RAG enterprise search system for healthcare clinicians, emphasizing low-latency performance and reduced hallucinations while maintaining end-to-end HIPAA compliance. Demonstrates deep applied experience with PHI-safe data governance (detection/redaction/de-identification), secure Azure ML deployment patterns, and orchestration of production LLM workflows using LangChain and Airflow.”
Senior Software Engineer specializing in AI, cloud infrastructure, and full-stack development
“ML/NLP engineer who built a production system that converts large-scale unstructured text into a connected, searchable knowledge base using spaCy + Sentence Transformers/FAISS and a Neo4j knowledge graph, with BERTopic and XGBoost for organization/labeling. Strong focus on production-grade Python workflows (FastAPI/Celery, Pydantic validation, Docker, AWS ECS/Lambda) and robust entity resolution with measurable precision/recall and human review for low-confidence matches.”
Senior ML Engineer & Data Scientist specializing in LLM agents, retrieval/ranking, and MLOps
“Machine Learning Engineer currently at Webster Bank building an enterprise-scale LLM agent for Temenos Journey Manager/Maestro, using RAG-style multi-stage retrieval with FAISS/Pinecone, hybrid dense+sparse search, and LoRA fine-tuning optimized via NDCG/MAP and A/B testing. Previously handled messy incident/telemetry data at Deuta Werke GmbH with deterministic + fuzzy entity resolution, and has strong production data engineering experience across Spark/Hadoop and Python ETL systems.”
Mid-level AI/ML Engineer specializing in enterprise ML, MLOps, and Generative AI
“ML/LLM engineer who has shipped production RAG systems (LangChain + HF Transformers + FAISS) with hybrid retrieval and cross-encoder re-ranking, deployed via FastAPI/Docker/Kubernetes and monitored with MLflow. Also partnered with wealth advisors at Edward Jones to deliver a client retention model with SHAP-driven explanations and a dashboard that improved trust, adoption, and reduced high-value client churn.”
Mid-level AI/ML Engineer specializing in Generative AI, RAG, and real-time fraud detection
“GenAI/ML engineer who has shipped production agentic systems in highly regulated and high-throughput environments, including an AWS Bedrock-based fraud/compliance workflow at U.S. Bank with PII redaction and hallucination detection that cut investigation time by 50%+. Also built and evaluated RAG and recommendation systems at Target, using RAGAS-driven testing, hybrid retrieval with re-ranking, and SHAP explainability dashboards to align model behavior with merchandising business KPIs.”
Mid-level GenAI Engineer specializing in production AI agents and evaluation pipelines
“Built and shipped a production LLM-powered internal operations automation platform using LangChain RAG (Pinecone) and FastAPI microservices, deployed on AWS EKS, serving 10k+ daily interactions. Implemented a rigorous evaluation/observability stack (golden datasets, prompt regression tests, MLflow, retrieval metrics, hallucination monitoring) that drove hallucinations below 2% and improved reliability, and partnered closely with non-technical ops leaders to cut manual lookup work by 60%+.”
Mid-level Data & GenAI Engineer specializing in lakehouse, streaming, and RAG platforms
“Built a production internal LLM-powered knowledge assistant using a RAG architecture (Python, LLM APIs, cloud services) that answers employee questions with sourced, grounded responses from internal documents. Demonstrates strong practical depth in retrieval tuning (chunking/metadata filters), orchestration with LangChain, and production reliability practices (latency optimization, automated embedding refresh, evaluation metrics, logging/monitoring) while partnering closely with non-technical operations teams.”