Pre-screened and vetted.
Mid-level Generative AI Engineer specializing in LLMs, RAG, and agentic systems
“Built a production "Mini RAG Assistant" for internal document Q&A, focusing on grounded answers (anti-hallucination), retrieval quality, and latency/cost optimization. Uses LangChain/LangGraph for orchestration and applies a metrics-driven evaluation loop (including reranking and semantic chunking improvements) while collaborating closely with product stakeholders.”
Mid-level Software Engineer specializing in Java microservices and cloud-native systems
“Enterprise workflow/product engineer (DXC) who owned a customer-facing workflow application for 500+ users and improved performance ~30% through API/SQL optimization, caching, and CI/CD-backed iteration. Experienced designing React/TypeScript + Java/Spring Boot systems and operating microservices with RabbitMQ/Kafka-style messaging, emphasizing reliability via DLQs, backpressure, and strong observability. Also built an internal automation dashboard adopted by support/ops teams to cut manual work and reduce SLA misses.”
“Built and deployed a production LLM-powered RAG knowledge system to unify operational/policy information across PDFs, wikis, and databases, emphasizing auditability and low-latency/cost performance. Improved answer relevance at scale by moving from pure vector search to hybrid retrieval with metadata filtering and reranking, and partnered closely with healthcare operations/compliance to define acceptance criteria and human-in-the-loop guardrails.”
Mid-level GenAI Engineer specializing in RAG, LLM agents, and enterprise automation
“Accenture engineer who built and shipped a production RAG-based automation/chatbot for SAP incident triage and troubleshooting, embedding thousands of runbooks/logs/tickets into a semantic search pipeline and integrating it into Teams/Slack. Reported major productivity gains (30–60% time reduction), >90% validated answer accuracy, and sub-2-second responses, with strong orchestration (Airflow/Prefect/LangGraph) and reliability practices (guardrails, testing, monitoring).”
“At Liberty Mutual, built a production underwriting decision assistant combining LLM reasoning with quantitative models and strong auditability. Implemented a claims-based response verification pipeline that cut hallucinations from 18% to 3% and materially improved user trust/validation scores. Experienced orchestrating ML/LLM workflows end-to-end with Airflow, Kubeflow Pipelines, and Jenkins, including SLA-focused pipeline hardening.”
Mid-level Software Engineer specializing in cloud-native microservices and AI/ML
“Full-stack engineer with healthcare/AI platform experience (Humana), owning an end-to-end high-risk patient prediction feature from React dashboards through FastAPI/TensorFlow real-time inference to AWS EKS operations. Emphasizes production reliability and contract-driven APIs (OpenAPI + generated TS types), plus strong data integration patterns (Kafka, idempotency, DLQs, backfills) in regulated, high-traffic environments.”
Junior AI/ML Developer specializing in GenAI, LLM agents, and RAG systems
“Built and shipped an agentic RAG chatbot module for NexaCLM to answer questions across large volumes of contracts while minimizing hallucinations and incorrect legal interpretations. Implemented routing between vector retrieval and ReAct-style agent retrieval plus an automated grading/validation layer (cosine-similarity thresholds, retries) and deployed via GitHub Actions to Azure Container Apps, partnering closely with legal stakeholders to define risk/clause-focused objectives.”
Mid-level AI Engineer specializing in causal inference and LLM research
“LLM engineer who has deployed a production system combining LLMs with causal inference (DoWhy) to enable counterfactual “what-if” analysis for experimental research, including a robust variable-mapping/validation layer to reduce hallucinations. Also partnered with non-technical operations leadership at Irriion Technologies to deliver an AI-assisted onboarding workflow that cut onboarding time by 50% and reduced manual errors by ~40%.”
Entry-Level Data Scientist specializing in ML, Azure, and LLM applications
“ML/computer-vision practitioner who shipped a CycleGAN-based bilingual handwriting translation demo (English↔Telugu) for low-resource scripts using unpaired datasets, focusing on preserving handwriting style and real-time deployment via Gradio. Also delivered a medical imaging pipeline by fine-tuning ResNet-50 and ViT-B/16 for pneumonia detection, emphasizing reproducibility, measurable evaluation, and stakeholder-friendly iteration.”
Mid-level Data Scientist specializing in insurance, healthcare, and cloud analytics
“Built a production-style LLM document summarization/generation workflow that mitigates token limits and reduces hallucinations using semantic chunking, FAISS-based embedding retrieval (top-k via cosine similarity), and section-wise generation. Orchestrated the end-to-end pipeline with AWS Step Functions and aligned outputs with sales stakeholders through demos, visuals, and documentation.”
Mid-level AI/ML Engineer specializing in Generative AI and LLM-powered NLP
“LLM/AI engineer who built a production automated document-understanding pipeline on Azure using a grounded RAG layer, designed to reduce manual review time for unstructured financial documents. Demonstrates strong real-world scaling and reliability practices (Service Bus queueing, Kubernetes autoscaling, observability, retries/circuit breakers) plus rigorous evaluation (shadow testing, replaying traffic, multilingual edge-case suites) and stakeholder-friendly, evidence-based explainability.”
Senior Machine Learning Engineer specializing in LLMs, RAG, and agentic AI systems
“LLM/RAG practitioner who has taken a support-ticket triage automation system from prototype to production, building the full pipeline (fine-tuned models, FastAPI inference services, vector storage, monitoring) and delivering measurable impact (~40% reduction in triage time). Demonstrates strong operational troubleshooting of LLM/agentic workflows (observability-driven debugging, fixing agent routing/looping) and supports adoption through tailored demos and sales-aligned technical communication.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and agentic workflows
“Applied AI/ML engineer with hands-on production experience building a RAG-based AI assistant for pharmaceutical maintenance troubleshooting using LangChain + FAISS/Pinecone, including a custom normalization layer to handle inconsistent terminology and duplicate document revisions. Also built Airflow-orchestrated pipelines for document ingestion/embeddings and predictive maintenance workflows (SCADA ETL, drift-based retraining), and partnered closely with production supervisors/quality engineers via Power BI dashboards and real-time alerts.”
Mid-level AI/ML Engineer specializing in Generative AI and RAG systems
“LLM/RAG engineer who has built and shipped production assistants, including a RAG-based teaching assistant (Marvel AI) using LangChain/LlamaIndex/ChromaDB with OpenAI embeddings and Redis vector search, achieving ~30% accuracy gains and ~35% latency reduction. Also deployed FastAPI services on Google Cloud Run with observability and prompt-level monitoring, and partnered with non-technical ops stakeholders to deliver an internal policy-document RAG assistant.”
Mid-level AI Engineer specializing in AI agents, RAG pipelines, and LLM evaluation
“Built and shipped production LLM systems at Founderbay, including a low-latency voice agent and a graph-based multi-agent research assistant. Strong focus on reliability in real workflows—hybrid SERP + full-site scraping RAG, grounding guardrails, validation checkpoints, and transcript-driven evaluation—plus performance tuning with async FastAPI, Redis caching, and containerization. Also partnered with a non-technical ops lead to automate post-call follow-ups via call summarization, field extraction, and tool-triggered actions.”
“Built and deployed a production LLM-powered internal AI assistant using a RAG pipeline to help teams search internal PDFs/knowledge bases and generate grounded summaries/answers. Demonstrates strong end-to-end ownership (ingestion through APIs) plus production rigor (monitoring/logging/CI-CD, evaluation metrics) and practical optimizations for hallucination, latency, and answer quality (thresholding, fallbacks, caching, async, re-ranking, two-tier model routing).”
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and MLOps
“Built and shipped a production real-time content moderation platform for Zoom/WebEx-style meetings, combining Whisper speech-to-text with fast NLP classifiers and REST APIs to flag hate speech, bias, and HIPAA-related content under strict latency constraints. Demonstrates strong MLOps/infra depth (Airflow, Kubernetes, Terraform/Helm, observability) and a pragmatic approach to reducing false positives via threshold tuning, context validation, and hard-negative data—while partnering closely with compliance and product stakeholders.”
Mid-level AI/ML Engineer specializing in healthcare ML, MLOps, and LLM/RAG systems
“Healthcare-focused ML/LLM engineer who built a production hybrid RAG workflow to automate prior authorization by retrieving from medical guidelines/historical cases (FAISS) and generating grounded rationales for clinicians. Strong in operationalizing ML with Airflow/Kubeflow/MLflow on SageMaker, optimizing latency (ONNX/quantization/async), and reducing hallucinations via evidence-only prompting; also partnered closely with clinical ops to deploy a readmission prediction tool used in daily rounds.”
Entry-level AI Engineer specializing in LLM agents, RAG, and computer vision
“Robotics/AV-focused candidate who contributed to an F1TENTH autonomous vehicle college project, building key autonomy components from raw sensor data to driving commands. Strong in perception and state estimation (visual odometry, particle-filter localization), plus mapping (occupancy grids) and planning/control (RRT, Gap Follow, PID), with hands-on ROS tooling and simulation validation in Gazebo/RViz and ROS environment containerization using Docker.”
Junior Machine Learning Engineer specializing in LLM fine-tuning and semantic retrieval
“Backend engineer with legal-tech and AI workflow experience: built JurisAI, an end-to-end legal research system using OCR + embeddings + Pinecone vector search to deliver citation-grounded LLM answers with safe failure modes (~90% recall@K). Also led a GW Law metadata migration into Caspio with batch validation and parallel rollout, and has strong FastAPI/GCP production reliability and observability practices.”
Mid-level AI/ML Engineer specializing in production ML, MLOps, and NLP
“Built and deployed a transformer-based clinical document classification system that processes unstructured clinical notes in a HIPAA-compliant healthcare setting, served via FastAPI on AWS and integrated into an Airflow/S3 pipeline. Demonstrates strong end-to-end MLOps skills (data quality remediation, low-latency inference optimization, monitoring with MLflow/CloudWatch) and effective collaboration with clinicians to drive adoption.”
Mid-level Software & ML Engineer specializing in agentic LLM systems and ML infrastructure
“Built and deployed an LLM-to-SQL automation system in a closed/internal environment, using a retriever–reranker–validator architecture on Kubernetes with strong security controls (semantic + rule-based validation and RBAC), achieving 99% uptime and cutting manual query time ~40%. Also worked on genomic sequence classification and semantic search workflows, orchestrating data prep with Airflow, tracking/deploying with MLflow, and optimizing distributed multi-GPU training on a university Kubernetes cluster.”
Mid-level Data Scientist specializing in NLP, recommender systems, and ML deployment
“At Provenbase, built and shipped a production LLM-powered semantic search and candidate matching platform (RAG with GPT-4/Gemini, multi-agent orchestration, Elasticsearch vector search) to scale sourcing across 10M+ candidate records and 1000+ data sources. Drove sub-second performance, cut LLM spend 30% with routing/caching, and improved recruiting outcomes (+45% sourcing accuracy; +38% visibility of underrepresented talent) through bias-aware ranking and tight collaboration with recruiting stakeholders.”
Junior Machine Learning Engineer specializing in multimodal systems and LLMs
“Built and productionized a domain-specific LLM-powered RAG knowledge assistant at JerseyStem for answering questions over large internal document corpora, owning the full stack from FAISS retrieval and LoRA/QLoRA fine-tuning to AWS autoscaling GPU deployment. Drove measurable gains (28% accuracy lift, 25% latency reduction) and improved reliability through hybrid retrieval, grounded decoding, preference-model reranking, and Airflow-orchestrated pipelines (35% faster runtime), while partnering closely with non-technical stakeholders to define success metrics and ensure adoption.”