Pre-screened and vetted.
Staff Software Engineer specializing in distributed systems, cloud platforms, and IoT
“CTO/Chief Architect who rebuilt an IoT platform from a fragile legacy stack into an AWS-based, multi-tenant cloud-native system supporting 50k+ connected devices and 10M+ monthly events, then layered in real-time data pipelines and ML anomaly detection. Known for tightly aligning roadmaps and OKRs to business KPIs (onboarding speed, uptime, velocity) and for scaling teams into domain-focused pods; previously led a shift from LAMP to event-driven Node.js microservices using MQTT and message queues.”
Intern AI Engineer specializing in LLMs, NLP, and conversational search
“Student building a production trip-planning LLM agent (LangChain + Streamlit) that routes user queries across multiple tools (maps/places/Wikipedia). Implemented zero-shot multi-label intent detection with priority rules to handle multi-intent requests, and collaborates with a startup product manager to shape tone, features, and user experience.”
Intern Software Engineer specializing in ML applications and LLM platform engineering
“Full-stack engineer who builds and scales customer-facing and internal AI products end-to-end (React/TypeScript/FastAPI/MongoDB) with strong product instrumentation and rapid MVP iteration. Built an AI-powered code review assistant adopted across teams and integrated into CI/CD, reducing manual review time by 30%+, and has hands-on experience with LLM retrieval/reasoning systems (LangChain + FAISS) and microservices scaling using RabbitMQ, Docker, and AWS.”
Senior AI/ML Engineer specializing in LLMs, RAG, and VR/XR multimodal systems
“PhD researcher (University of Utah) who built a production RAG-powered Virtual Reality Research Assistant to answer lab research questions with concrete citations. Implemented an end-to-end LangChain pipeline using PyPDFLoader, chunking strategies, OpenAI embeddings, and ChromaDB, with emphasis on grounding to reduce hallucinations and ensure research-grade accuracy. Collaborated closely with a non-technical PhD advisor to scope requirements, manage cost constraints, and demo iterative progress.”
Junior Machine Learning Engineer specializing in NLP, data pipelines, and LLM workflows
“Built and shipped a production LLM-powered decision system that replaced a slow, inconsistent manual review process by turning messy text into structured, auditable outputs behind an API. Demonstrates strong end-to-end ownership of reliability and operations (schema validation, retries/fallbacks, latency/cost controls, monitoring for drift) and a disciplined approach to evaluation and regression testing. Experienced collaborating with non-technical reviewers to define success criteria and deliver interpretable outputs that get adopted.”
Intern AI/Software Engineer specializing in RAG, LLM agents, and cloud-deployed search
“Built and deployed a production AI document Q&A (RAG) platform that lets non-technical users query hundreds of PDFs/Word files, cutting search time from hours to seconds. Experienced with scaling retrieval pipelines (chunking, embeddings, vector search, batching/caching) and orchestrating reliable workflows using AWS Step Functions/Airflow with robust retries, monitoring, and fallbacks.”
Mid-level AI Engineer specializing in NLP, computer vision, and MLOps
“AI Engineer at DXC Technology who has shipped production LLM/NLP systems on AWS (SageMaker, FastAPI) and optimized them for real-time latency and unpredictable traffic using quantization, batching, and autoscaling. Strong MLOps and monitoring discipline (MLflow, CloudWatch, SageMaker Model Monitor) and proven business impact—delivered models with 92% predictive accuracy and cut enterprise decision-making time by 30% through close collaboration with product managers.”
Senior Machine Learning & Computer Vision Researcher specializing in vision-language models
“Developed and deployed CaptionFace, a production vision-language system that boosts low-resolution/surveillance face recognition by generating discriminative natural-language captions (ViT encoder + GPT-2 decoder) and enabling text-to-face retrieval and zero-shot recognition. Orchestrated distributed training on Kubernetes with MLflow tracking, mixed-precision optimization, and comprehensive evaluation including out-of-domain robustness; collaborated with non-technical NSF project stakeholders via demos, visualization, and clear documentation.”
Mid-level AI/ML Engineer & Data Scientist specializing in NLP and Generative AI
“Built and deployed an agentic RAG platform at Centene Health to support healthcare claims and complaints workflows (Q&A for claims agents, executive complaint summarization, and compliance triage/classification). Experienced in LangChain/LangGraph orchestration, production deployment on AWS with FastAPI/Docker/Kubernetes, and implementing HIPAA-compliant guardrails to reduce hallucinations and ensure explainable outputs.”
Junior Machine Learning Engineer specializing in MLOps and real-time systems
“Built and shipped a production GPT-4 + RAG customer support chatbot that materially improved support operations (response time 4 hours to <3 minutes; ~65% tier-1 ticket automation). Demonstrates strong end-to-end LLM engineering across retrieval (Sentence Transformers/Pinecone), safety (multi-layer moderation), cost/latency optimization (caching/streaming, Celery/Redis), and rigorous evaluation/monitoring (shadow deploys, Datadog, 500+ test cases), plus proven stakeholder buy-in leading to 80% adoption.”
Mid-level Machine Learning & AI Engineer specializing in Generative AI, NLP, and MLOps
“Built and deployed production LLM systems for summarizing sensitive legal and financial documents, emphasizing GDPR-aligned privacy controls and scalable hybrid cloud architecture. Experienced with Kubernetes/Airflow orchestration and rigorous testing/monitoring practices, and has delivered measurable business impact (18% conversion lift) by translating AI outputs for non-technical marketing stakeholders.”
Junior AI/ML Engineer specializing in healthcare and financial risk modeling
“Built and productionized a clinical NLP + patient risk stratification platform at Dermanture, combining Spark/PySpark pipelines with BERT/BioBERT for entity extraction and text classification and downstream risk models in TensorFlow/scikit-learn. Experienced running regulated, auditable ML workflows with Airflow and AWS SageMaker, emphasizing data validation (Great Expectations), drift monitoring, and explainability (SHAP) to drive clinician trust and adoption.”
Intern Software Engineer specializing in backend systems and Generative AI
“Built and deployed a scalable, production-ready LLM knowledge assistant using a RAG architecture (LangChain + vector store/FAISS) to replace keyword search for internal documents. Demonstrates hands-on expertise in hallucination reduction and retrieval quality improvements through semantic chunking, similarity tuning, prompt design, and human-in-the-loop validation, plus strong stakeholder communication via demos and visual explanations.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and LLM systems
“Built a production multi-agent cybersecurity defense simulator orchestrated with CrewAI, combining Red/Blue team LLM agents, a RAG runbook retriever, and an RL remediation agent trained via state-space simplification and reward shaping for rapid incident response. Also partnered with quant analysts and fund managers to deliver an automated trading and portfolio management system using statistical methods plus CNN/LSTM models, reporting up to 15% weekly ROI.”
“At Liberty Mutual, built a production underwriting decision assistant combining LLM reasoning with quantitative models and strong auditability. Implemented a claims-based response verification pipeline that cut hallucinations from 18% to 3% and materially improved user trust/validation scores. Experienced orchestrating ML/LLM workflows end-to-end with Airflow, Kubeflow Pipelines, and Jenkins, including SLA-focused pipeline hardening.”
Senior Data Scientist specializing in LLM applications, RAG systems, and production ML
“Senior Data Scientist in consulting who has built production RAG systems for insurance/annuity document search at large scale (100K+ PDF pages), emphasizing grounded answers, guardrails, and low-latency retrieval. Experienced in end-to-end MLOps for LLM apps—monitoring, evaluation sets, drift handling, and safe rollouts—and in orchestrating complex pipelines with Prefect/Airflow and deploying services on Kubernetes.”
Mid-level AI/ML Engineer specializing in data engineering, LLM/RAG pipelines, and recommender systems
“Research assistant at St. Louis University who built and deployed a production document-intelligence RAG system (Python/TensorFlow, vector DB, FastAPI) on AWS, focusing on grounding to reduce hallucinations and latency optimization via caching/async/batching. Also developed a personalized recommendation system for the Frenzy social platform and partnered closely with product/UX to define metrics and iterate on hybrid recommenders and cold-start handling.”
Mid-level AI Engineer specializing in NLP and production ML systems
“AI/LLM engineer who has shipped production RAG chatbots using LangChain/OpenAI with FAISS and FastAPI, focusing on real-world constraints like context windows, concurrency, and latency (reported ~40% latency reduction and <2s average response). Experienced orchestrating AI pipelines with Celery and fault-tolerant long-running workflows with Temporal, and has applied NLP model tradeoff testing (Word2Vec vs BERT) to drive measurable accuracy gains.”
Mid-level AI Engineer specializing in ML, NLP, and Generative AI
“AI/LLM engineer with production experience building an LLM-powered investment recommendation system using RAG and chatbots, deployed via Docker/CI/CD and scaled on Kubernetes. Demonstrated measurable performance wins (sub-200ms latency) through QLoRA fine-tuning and TensorRT INT8/INT4 quantization, plus strong MLOps/orchestration background (Airflow ETL + scoring, MLflow monitoring) and stakeholder-facing delivery using demos and Tableau dashboards.”
Senior Machine Learning Engineer specializing in LLMs, RAG, and agentic AI systems
“LLM/RAG practitioner who has taken a support-ticket triage automation system from prototype to production, building the full pipeline (fine-tuned models, FastAPI inference services, vector storage, monitoring) and delivering measurable impact (~40% reduction in triage time). Demonstrates strong operational troubleshooting of LLM/agentic workflows (observability-driven debugging, fixing agent routing/looping) and supports adoption through tailored demos and sales-aligned technical communication.”
Mid-level Software Engineer specializing in AI-driven distributed systems
“Backend engineer who built a high-stakes, privacy-first platform at be Still Analytics for survivors of domestic violence, emphasizing anonymity, security, and reliability. Experienced with GenAI backends (LangChain + AWS Bedrock) including RAG to prevent hallucinations, plus cloud-native scaling (Docker/Kubernetes) and cost-saving migrations from legacy VMs to serverless (30% reduction).”
“Built and deployed a production LLM-powered internal AI assistant using a RAG pipeline to help teams search internal PDFs/knowledge bases and generate grounded summaries/answers. Demonstrates strong end-to-end ownership (ingestion through APIs) plus production rigor (monitoring/logging/CI-CD, evaluation metrics) and practical optimizations for hallucination, latency, and answer quality (thresholding, fallbacks, caching, async, re-ranking, two-tier model routing).”
Mid-level AI/ML Engineer specializing in NLP, computer vision, and MLOps
“Built and deployed a production LLM/RAG intelligent document understanding platform for healthcare clinical documents (notes, discharge summaries, diagnostic reports), integrating spaCy entity extraction, Pinecone vector search, and a Spring Boot API on AWS with monitoring and guardrails. Demonstrates strong MLOps/orchestration (LangChain, Airflow, Kubeflow/Kubernetes) and a metrics-driven evaluation approach, and partnered with a healthcare operations manager to cut manual review time by 80%.”