Pre-screened and vetted.
Junior Machine Learning Engineer specializing in MLOps and real-time systems
“Built and shipped a production GPT-4 + RAG customer support chatbot that materially improved support operations (response time 4 hours to <3 minutes; ~65% tier-1 ticket automation). Demonstrates strong end-to-end LLM engineering across retrieval (Sentence Transformers/Pinecone), safety (multi-layer moderation), cost/latency optimization (caching/streaming, Celery/Redis), and rigorous evaluation/monitoring (shadow deploys, Datadog, 500+ test cases), plus proven stakeholder buy-in leading to 80% adoption.”
Intern Software Engineer specializing in backend systems and Generative AI
“Built and deployed a scalable, production-ready LLM knowledge assistant using a RAG architecture (LangChain + vector store/FAISS) to replace keyword search for internal documents. Demonstrates hands-on expertise in hallucination reduction and retrieval quality improvements through semantic chunking, similarity tuning, prompt design, and human-in-the-loop validation, plus strong stakeholder communication via demos and visual explanations.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and agentic systems
“Built a production "Mini RAG Assistant" for internal document Q&A, focusing on grounded answers (anti-hallucination), retrieval quality, and latency/cost optimization. Uses LangChain/LangGraph for orchestration and applies a metrics-driven evaluation loop (including reranking and semantic chunking improvements) while collaborating closely with product stakeholders.”
Mid-level GenAI Engineer specializing in RAG, LLM agents, and enterprise automation
“Accenture engineer who built and shipped a production RAG-based automation/chatbot for SAP incident triage and troubleshooting, embedding thousands of runbooks/logs/tickets into a semantic search pipeline and integrating it into Teams/Slack. Reported major productivity gains (30–60% time reduction), >90% validated answer accuracy, and sub-2-second responses, with strong orchestration (Airflow/Prefect/LangGraph) and reliability practices (guardrails, testing, monitoring).”
Senior Data Scientist specializing in LLM applications, RAG systems, and production ML
“Senior Data Scientist in consulting who has built production RAG systems for insurance/annuity document search at large scale (100K+ PDF pages), emphasizing grounded answers, guardrails, and low-latency retrieval. Experienced in end-to-end MLOps for LLM apps—monitoring, evaluation sets, drift handling, and safe rollouts—and in orchestrating complex pipelines with Prefect/Airflow and deploying services on Kubernetes.”
Mid-level Software Engineer specializing in Python backend and LLM/ML systems
“Backend/AI engineer who has shipped production LLM systems end-to-end, including an AI request-routing service (FastAPI + BART MNLI + OpenAI/Gemini) that improved accuracy ~25% after launch via eval-driven prompt/category iteration. Also built an enterprise document intelligence/RAG platform on Azure (Blob/SharePoint/Teams ingestion, OCR/NLP chunking, embeddings in Azure Cognitive Search) with PII guardrails (Presidio), confidence gating, and scalable event-driven pipelines handling millions of documents.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and LLM systems
“Built a production multi-agent cybersecurity defense simulator orchestrated with CrewAI, combining Red/Blue team LLM agents, a RAG runbook retriever, and an RL remediation agent trained via state-space simplification and reward shaping for rapid incident response. Also partnered with quant analysts and fund managers to deliver an automated trading and portfolio management system using statistical methods plus CNN/LSTM models, reporting up to 15% weekly ROI.”
Mid-level AI/ML Engineer specializing in data engineering, LLM/RAG pipelines, and recommender systems
“Research assistant at St. Louis University who built and deployed a production document-intelligence RAG system (Python/TensorFlow, vector DB, FastAPI) on AWS, focusing on grounding to reduce hallucinations and latency optimization via caching/async/batching. Also developed a personalized recommendation system for the Frenzy social platform and partnered closely with product/UX to define metrics and iterate on hybrid recommenders and cold-start handling.”
Mid-level Full-Stack Product Engineer specializing in SaaS and AI search
“Two-time founder and former CTO/CPO who has shipped and operated full-stack products solo, including a real-time community platform (Twitter/Slack-like) with Next.js/TypeScript, WebSockets, Redis, and strong post-launch analytics (PostHog/Sentry). Also built durable multi-step AI-agent workflows using Inngest with state machines, checkpointing, and validation gates, and has hands-on Postgres performance tuning experience validated via EXPLAIN ANALYZE.”
“ML engineer with hands-on experience building banking AI systems end-to-end, including a customer-targeting model that improved campaign response rates by about 10%. Also shipped a RAG-based banking FAQ/support feature with safety guardrails and production optimizations around retrieval quality, latency, and cost, plus reusable Python services that reduced duplicate work for other engineers.”
Senior Full-Stack Developer specializing in React, Node.js, and AWS
“Backend/data engineer with hands-on production experience across Python/Flask microservices and AWS serverless/data platforms (Lambda, DynamoDB, S3, Glue/PySpark). Demonstrated strong reliability and operations mindset (JWT/RBAC, retries/timeouts/circuit breakers, CloudWatch/SNS alerting) and measurable performance wins (SQL report runtime cut from 10 minutes to 30 seconds). Seeking ~$150k base and cannot travel for onsite meetings for the next 5–6 months due to family medical constraints.”
Intern Data Scientist specializing in Generative AI and NLP
“Backend/AI engineer with internship experience building an AI-powered financial insights platform (FastAPI, Redis, BigQuery) and prior HCL experience leading a monolith-to-microservices refactor (Flask, Kafka) using blue-green deployments. Demonstrates strong performance/security focus (OAuth/JWT/RBAC, encryption) and measurable impact on latency, downtime, and ML model reliability; MVP was submitted to Google’s accelerator program.”
Mid-level AI Engineer specializing in NLP and production ML systems
“AI/LLM engineer who has shipped production RAG chatbots using LangChain/OpenAI with FAISS and FastAPI, focusing on real-world constraints like context windows, concurrency, and latency (reported ~40% latency reduction and <2s average response). Experienced orchestrating AI pipelines with Celery and fault-tolerant long-running workflows with Temporal, and has applied NLP model tradeoff testing (Word2Vec vs BERT) to drive measurable accuracy gains.”
Junior Software Engineer specializing in backend, cloud, and LLM-powered search
“Python backend engineer (BetterWorld Technology) who owns microservice systems end-to-end on Azure, including Kubernetes deployments, CI/CD, and production monitoring/alerting. Has hands-on experience integrating SQL/NoSQL (including Cosmos DB with vector search/graph workflow) and has built a Kafka + Spark Streaming pipeline to Snowflake with a reported 40% latency reduction.”
Mid-level Data Scientist specializing in insurance, healthcare, and cloud analytics
“Built a production-style LLM document summarization/generation workflow that mitigates token limits and reduces hallucinations using semantic chunking, FAISS-based embedding retrieval (top-k via cosine similarity), and section-wise generation. Orchestrated the end-to-end pipeline with AWS Step Functions and aligned outputs with sales stakeholders through demos, visuals, and documentation.”
Senior Machine Learning Engineer specializing in LLMs, RAG, and agentic AI systems
“LLM/RAG practitioner who has taken a support-ticket triage automation system from prototype to production, building the full pipeline (fine-tuned models, FastAPI inference services, vector storage, monitoring) and delivering measurable impact (~40% reduction in triage time). Demonstrates strong operational troubleshooting of LLM/agentic workflows (observability-driven debugging, fixing agent routing/looping) and supports adoption through tailored demos and sales-aligned technical communication.”
Mid-level AI/ML Engineer specializing in Generative AI and RAG systems
“LLM/RAG engineer who has built and shipped production assistants, including a RAG-based teaching assistant (Marvel AI) using LangChain/LlamaIndex/ChromaDB with OpenAI embeddings and Redis vector search, achieving ~30% accuracy gains and ~35% latency reduction. Also deployed FastAPI services on Google Cloud Run with observability and prompt-level monitoring, and partnered with non-technical ops stakeholders to deliver an internal policy-document RAG assistant.”
Mid-level AI/ML Engineer specializing in Generative AI and LLM-powered NLP
“LLM/AI engineer who built a production automated document-understanding pipeline on Azure using a grounded RAG layer, designed to reduce manual review time for unstructured financial documents. Demonstrates strong real-world scaling and reliability practices (Service Bus queueing, Kubernetes autoscaling, observability, retries/circuit breakers) plus rigorous evaluation (shadow testing, replaying traffic, multilingual edge-case suites) and stakeholder-friendly, evidence-based explainability.”
Mid-level Full-Stack AI Engineer specializing in healthcare and enterprise SaaS
“Full-stack product engineer who has built AI-assisted CRM and agent workflows in Project SARA and operational systems like payroll for a staffing platform. Stands out for combining React/TypeScript, Django/Postgres, real-time systems, and LLM orchestration with strong product instincts—delivering measurable gains in response time, conversion, and engineering leverage.”
Mid-level AI/Data Engineer specializing in agentic AI and data platforms
“AI/LLM engineer who built a production resume-parsing and candidate-matching platform at Quadrant Technologies, combining agentic LangChain workflows, VLM-based document template extraction (~85% accuracy), and a hybrid RAG backend for resume-to-JD search. Notably integrated automated LLM evals and metric-based CI/CD quality gates to catch silent prompt/model regressions, and led a 3-person team across frontend/backend/testing.”
Junior AI Engineer specializing in LLMs, RAG, and MLOps
“At ReferU.AI, designed and deployed an agentic RAG pipeline that automates multi-jurisdiction legal document drafting, emphasizing hallucination reduction through hybrid retrieval, validation agents, guardrails, and iterative regeneration. Experienced with orchestration frameworks (especially CrewAI) and rigorous testing/evaluation practices including human-in-the-loop review, adversarial testing, and production metrics/logging.”
Mid-level Full-Stack Engineer specializing in cloud-native FinTech analytics
“Full-stack/ML-leaning engineer who has shipped production-grade real-time analytics and an internal AI support assistant using RAG over enterprise documentation. Demonstrates strong systems thinking across scalability, reliability, observability, and LLM safety/evaluation (thresholded retrieval, RBAC, response validation, regression-gated evals), with concrete iteration based on performance metrics and user feedback.”
Mid-level Software Engineer specializing in LLM agents and cloud-native systems
“Built and shipped production LLM agents in compliance-sensitive environments (FERPA), emphasizing reliability via structured outputs, state-graph orchestration (LangGraph), and CI-driven eval/regression testing. Also has experience hardening messy ERP ingestion pipelines at scale (50K monthly orders) with normalization, idempotency/deduplication, and robust failure handling using AWS (SQS/CloudWatch) and PostgreSQL.”