Pre-screened and vetted.
Mid-level AI Engineer specializing in LLM agents and production ML systems
Mid-level Machine Learning Engineer specializing in LLM alignment and applied reinforcement learning
“AI/LLM engineer who has shipped production systems end-to-end, including a note-taking product (Notey) combining audio/image capture, ASR, summarization, and a semantic chat agent over past notes. Also has applied ML experience in healthcare, collaborating directly with doctors to validate an EEG seizure-detection pipeline, and uses Kubernetes to optimize GPU usage for LLM training.”
Senior AI/ML Engineer specializing in LLMs, RAG, and VR/XR multimodal systems
“PhD researcher (University of Utah) who built a production RAG-powered Virtual Reality Research Assistant to answer lab research questions with concrete citations. Implemented an end-to-end LangChain pipeline using PyPDFLoader, chunking strategies, OpenAI embeddings, and ChromaDB, with emphasis on grounding to reduce hallucinations and ensure research-grade accuracy. Collaborated closely with a non-technical PhD advisor to scope requirements, manage cost constraints, and demo iterative progress.”
Junior Machine Learning Engineer specializing in NLP, data pipelines, and LLM workflows
“Built and shipped a production LLM-powered decision system that replaced a slow, inconsistent manual review process by turning messy text into structured, auditable outputs behind an API. Demonstrates strong end-to-end ownership of reliability and operations (schema validation, retries/fallbacks, latency/cost controls, monitoring for drift) and a disciplined approach to evaluation and regression testing. Experienced collaborating with non-technical reviewers to define success criteria and deliver interpretable outputs that get adopted.”
Intern AI/Software Engineer specializing in RAG, LLM agents, and cloud-deployed search
“Built and deployed a production AI document Q&A (RAG) platform that lets non-technical users query hundreds of PDFs/Word files, cutting search time from hours to seconds. Experienced with scaling retrieval pipelines (chunking, embeddings, vector search, batching/caching) and orchestrating reliable workflows using AWS Step Functions/Airflow with robust retries, monitoring, and fallbacks.”
Mid-level ML Engineer specializing in LLMs, Generative AI, and MLOps
“AI/ML engineer with production experience building an enterprise network-fault prediction assistant that combines anomaly detection (Isolation Forest + LSTM) with an LLM layer for incident diagnosis and recommended resolutions. Hands-on with orchestration (Airflow, Prefect, Dagster) to run ETL/ELT and automated training/fine-tuning workflows, and has delivered AI solutions with non-technical stakeholders (retail customer support ticket categorization/response suggestions).”
Principal DevOps Architect specializing in cloud platform engineering and SRE
“End-to-end engineer focused on AI-native enterprise systems, including a production generative knowledge platform using RAG + semantic search over internal documentation (React, Python/Flask, GPU-hosted NLP models, Pinecone) with strong CI/CD and observability. Reports concrete outcomes including 40% faster knowledge access and ~75% employee adoption, and has led incremental cloud-native modernization using feature flags, parallel runs, canary releases, and regression testing.”
Junior Software Engineer specializing in backend systems and AI data pipelines
“Backend engineer with fintech/AI startup experience who built an Azure serverless, event-driven pipeline for large-scale crypto sentiment analysis and semantic search (OCR/NLP to vector search) and integrated LLM + blockchain data for predictive insights. Demonstrated measurable impact (25% lower retrieval latency, 10% fewer data errors, 15% higher engagement) and has led safe microservices migrations with strong security and reliability practices.”
Mid-level AI Engineer specializing in LLM apps, RAG pipelines, and multi-agent systems
“AI Engineer at Humanitarian AI who has built and productionized both a LangGraph-based multi-agent workflow system and a RAG pipeline (OpenAI embeddings + vector DB) with rigorous evaluation/guardrails. Reports strong measurable impact (60% faster workflow delivery, 40% fewer incidents, 70% reduced research time) and has prior enterprise modernization experience at Infosys migrating ETL to microservices with zero production incidents.”
Mid-level AI/ML Engineer & Data Scientist specializing in NLP and Generative AI
“Built and deployed an agentic RAG platform at Centene Health to support healthcare claims and complaints workflows (Q&A for claims agents, executive complaint summarization, and compliance triage/classification). Experienced in LangChain/LangGraph orchestration, production deployment on AWS with FastAPI/Docker/Kubernetes, and implementing HIPAA-compliant guardrails to reduce hallucinations and ensure explainable outputs.”
Junior Machine Learning Engineer specializing in MLOps and real-time systems
“Built and shipped a production GPT-4 + RAG customer support chatbot that materially improved support operations (response time 4 hours to <3 minutes; ~65% tier-1 ticket automation). Demonstrates strong end-to-end LLM engineering across retrieval (Sentence Transformers/Pinecone), safety (multi-layer moderation), cost/latency optimization (caching/streaming, Celery/Redis), and rigorous evaluation/monitoring (shadow deploys, Datadog, 500+ test cases), plus proven stakeholder buy-in leading to 80% adoption.”
Intern Software Engineer specializing in backend systems and Generative AI
“Built and deployed a scalable, production-ready LLM knowledge assistant using a RAG architecture (LangChain + vector store/FAISS) to replace keyword search for internal documents. Demonstrates hands-on expertise in hallucination reduction and retrieval quality improvements through semantic chunking, similarity tuning, prompt design, and human-in-the-loop validation, plus strong stakeholder communication via demos and visual explanations.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and LLM systems
“Built a production multi-agent cybersecurity defense simulator orchestrated with CrewAI, combining Red/Blue team LLM agents, a RAG runbook retriever, and an RL remediation agent trained via state-space simplification and reward shaping for rapid incident response. Also partnered with quant analysts and fund managers to deliver an automated trading and portfolio management system using statistical methods plus CNN/LSTM models, reporting up to 15% weekly ROI.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and agentic systems
“Built a production "Mini RAG Assistant" for internal document Q&A, focusing on grounded answers (anti-hallucination), retrieval quality, and latency/cost optimization. Uses LangChain/LangGraph for orchestration and applies a metrics-driven evaluation loop (including reranking and semantic chunking improvements) while collaborating closely with product stakeholders.”
“Built and deployed a production LLM-powered RAG knowledge system to unify operational/policy information across PDFs, wikis, and databases, emphasizing auditability and low-latency/cost performance. Improved answer relevance at scale by moving from pure vector search to hybrid retrieval with metadata filtering and reranking, and partnered closely with healthcare operations/compliance to define acceptance criteria and human-in-the-loop guardrails.”
Mid-level GenAI Engineer specializing in RAG, LLM agents, and enterprise automation
“Accenture engineer who built and shipped a production RAG-based automation/chatbot for SAP incident triage and troubleshooting, embedding thousands of runbooks/logs/tickets into a semantic search pipeline and integrating it into Teams/Slack. Reported major productivity gains (30–60% time reduction), >90% validated answer accuracy, and sub-2-second responses, with strong orchestration (Airflow/Prefect/LangGraph) and reliability practices (guardrails, testing, monitoring).”
Senior Data Scientist specializing in LLM applications, RAG systems, and production ML
“Senior Data Scientist in consulting who has built production RAG systems for insurance/annuity document search at large scale (100K+ PDF pages), emphasizing grounded answers, guardrails, and low-latency retrieval. Experienced in end-to-end MLOps for LLM apps—monitoring, evaluation sets, drift handling, and safe rollouts—and in orchestrating complex pipelines with Prefect/Airflow and deploying services on Kubernetes.”
Mid-level AI/ML Engineer specializing in data engineering, LLM/RAG pipelines, and recommender systems
“Research assistant at St. Louis University who built and deployed a production document-intelligence RAG system (Python/TensorFlow, vector DB, FastAPI) on AWS, focusing on grounding to reduce hallucinations and latency optimization via caching/async/batching. Also developed a personalized recommendation system for the Frenzy social platform and partnered closely with product/UX to define metrics and iterate on hybrid recommenders and cold-start handling.”
Mid-level Software Engineer specializing in Python backend and LLM/ML systems
“Backend/AI engineer who has shipped production LLM systems end-to-end, including an AI request-routing service (FastAPI + BART MNLI + OpenAI/Gemini) that improved accuracy ~25% after launch via eval-driven prompt/category iteration. Also built an enterprise document intelligence/RAG platform on Azure (Blob/SharePoint/Teams ingestion, OCR/NLP chunking, embeddings in Azure Cognitive Search) with PII guardrails (Presidio), confidence gating, and scalable event-driven pipelines handling millions of documents.”
Senior Full-Stack Developer specializing in React, Node.js, and AWS
“Backend/data engineer with hands-on production experience across Python/Flask microservices and AWS serverless/data platforms (Lambda, DynamoDB, S3, Glue/PySpark). Demonstrated strong reliability and operations mindset (JWT/RBAC, retries/timeouts/circuit breakers, CloudWatch/SNS alerting) and measurable performance wins (SQL report runtime cut from 10 minutes to 30 seconds). Seeking ~$150k base and cannot travel for onsite meetings for the next 5–6 months due to family medical constraints.”
Intern Data Scientist specializing in Generative AI and NLP
“Backend/AI engineer with internship experience building an AI-powered financial insights platform (FastAPI, Redis, BigQuery) and prior HCL experience leading a monolith-to-microservices refactor (Flask, Kafka) using blue-green deployments. Demonstrates strong performance/security focus (OAuth/JWT/RBAC, encryption) and measurable impact on latency, downtime, and ML model reliability; MVP was submitted to Google’s accelerator program.”