Pre-screened and vetted.
Executive IT and Operations leader specializing in digital transformation and security
“Candidate is very familiar with the venture capital and broader investment landscape, but is not interested in founding a company. They have worked with several TPG-backed or TPG-owned organizations, helping drive business scaling, cost reduction, and execution against investor governance requirements.”
Mid-level Full-Stack AI Engineer specializing in agentic systems and scalable platforms
“AI-focused full-stack/DevOps engineer who goes beyond using copilots and has built production-oriented LLM systems such as natural-language-to-SQL and structured insight extraction pipelines. Stands out for treating AI as an accelerator rather than a replacement, with a strong emphasis on guardrails, validation, observability, and safe deployment practices in agent-based and distributed systems.”
Principal AI/ML Leader specializing in Generative AI, MLOps, and NLP
“Founding member of Tausight, building AI systems to detect and protect PHI for healthcare organizations; helped take the company through post–Series A funding and exited after ~6 years. Drove a strategic collaboration with Intel’s OpenVINO team—becoming the first to deploy it in a real production system and improving model performance by ~30% on customer Intel-CPU machines.”
Mid-level Generative AI Engineer specializing in LLMs and RAG systems
“Built and shipped a production RAG-based enterprise knowledge assistant to replace slow/inaccurate search across millions of documents, using LangChain orchestration with GPT-4/LLaMA and vector databases. Strong focus on production constraints—latency, hallucination control, and cost—using hybrid retrieval, guardrails, LLM-as-judge validation, and model routing, and has experience translating non-technical stakeholder pain points into measurable outcomes.”
Mid-level Data Engineer specializing in Lakehouse, Streaming, and ML/LLM data systems
“Built and productionized an enterprise retrieval-augmented generation platform for internal knowledge over large unstructured corpora, emphasizing trust via strict citation/grounding and hybrid retrieval (BM25 + FAISS + cross-encoder re-ranking). Demonstrates strong scaling and cost/latency optimization through incremental indexing/embedding and index partitioning, plus disciplined evaluation/observability practices. Has experience operationalizing pipelines with Airflow/Databricks/GitHub Actions and partnering closely with risk & compliance stakeholders on auditability requirements.”
Mid-level AI/ML Engineer specializing in Generative AI and production ML systems
“Built and deployed a production SecureAIChatBot (RAG-based) for secure internal information retrieval, using embeddings/vector search, GPT models, monitoring, and safety filters. Focused on real-world production challenges like latency and output consistency, applying caching, retrieval scoping, smaller models, and controlled prompting, and used LangChain to orchestrate the end-to-end workflow.”
Mid-level AI/ML Engineer specializing in MLOps and LLM applications
“BNY Mellon engineer who has built and operated production AI systems end-to-end: a LangChain/Pinecone RAG platform scaled via FastAPI + Kubernetes to 1000 RPM with 99.9% uptime, supported by monitoring and data-drift detection. Also deep in data/infra orchestration (Airflow, Dagster, Terraform on AWS/EMR/EC2), processing 500GB+ daily and delivering measurable reliability and performance gains, plus strong compliance-facing model explainability using SHAP and Tableau.”
Mid-level AI Engineer specializing in GenAI, LLM integration, and RAG pipelines
“Built and led deployment of an autonomous, self-correcting multi-agent knowledge retrieval and validation system at HCA Healthcare to reduce heavy manual research/validation in clinical/compliance documentation. Deeply focused on production reliability and cost—used LangGraph StateGraph orchestration plus ONNX/CUDA/quantization to cut GPU costs by 25%, and partnered with the Compliance VP using real-time contradiction-rate dashboards to hit a 40% automation goal without compromising compliance.”
Mid-level AI/ML Engineer specializing in cloud data engineering and GenAI
“AI/LLM engineer with production experience in legal tech: built a GPT-4 + LangChain RAG summarization system at Govpanel that reduced legal case-file review time by 50%+. Previously at LexisNexis, orchestrated end-to-end Airflow data/AI pipelines processing 5M+ legal documents daily, improving ETL runtime by 35% with robust validation, monitoring, and SLAs.”
Intern Full-Stack Software Engineer specializing in AI/ML and cloud
“Built a Python-based geospatial machine learning backend for PFAS contamination risk mapping, including reproducible feature pipelines, ensemble modeling, and a FastAPI layer for visualization/analysis. Emphasizes data integrity and robustness (CRS/coverage checks, fail-fast validation) and has led safe backend refactors using feature flags, idempotent backfills, and Postgres RLS for secure, queryable results delivery.”
Mid-level AI/ML Engineer specializing in generative AI, RAG platforms, and LLM agents
“AI/LLM engineer who has shipped 10+ production applications, including InvestIQ on GCP—a production-grade RAG due-diligence engine that ethically scrapes web/PDF sources, builds a ChromaDB knowledge base, and delivers analyst-style dashboards plus a citation-backed chat copilot. Deep focus on reliability (evidence-only answers, hard citations, refusal gating), retrieval tuning, and orchestration (Airflow/Cloud Composer), plus multi-agent systems (CrewAI with 7 specialized finance agents).”
“Backend-focused intern who built and refactored the backend for an LLM-driven gifting mobile app using FastAPI, tackling high-latency LLM + product-API workflows. Implemented async worker-pool/queue processing with Redis caching plus retries/fallbacks, cutting end-to-end suggestion latency from ~4–5 seconds to ~1 second while improving reliability and rollout safety via staged migrations and testing.”
Mid-level Full-Stack .NET Developer specializing in cloud-native microservices
“Full-stack .NET engineer with cloud and applied GenAI experience who shipped a real-time policy status tracking module at Lincoln Financial using ASP.NET Core/.NET 8, Kafka, Angular, SQL Server, Redis, and AKS autoscaling. Also delivered a production internal LLM+RAG support assistant at Honeywell with strong security/guardrails (PII masking, RBAC) and a rigorous eval/regression loop built on a 200-question gold set.”
Senior Machine Learning Engineer specializing in LLMs, speech AI, and RAG systems
“AI engineer with production experience building multilingual speech-to-speech translation pipelines (ASR + LLM) for enterprise/media, focused on reliability at scale. Has hands-on orchestration experience (including IBM Watson contexts) and emphasizes production evaluation/monitoring using a mix of traditional metrics and LLM-based evaluators to catch quality regressions while balancing latency and cost.”
Mid-level AI Engineer specializing in LLMs, MLOps, and healthcare NLP
“Built a production, real-time clinical documentation system at HCA that converts doctor–patient conversations into structured clinical summaries using speech-to-text, LLM summarization, and RAG. Demonstrated measurable gains from medical-domain fine-tuning (clinical concept recall +18%, ROUGE-L 0.62 to 0.74) while meeting HIPAA constraints via PHI anonymization and encryption, and deployed via Docker/FastAPI with CI/CD and monitoring.”
Senior Software Engineer specializing in AI systems and platform engineering
“Backend/AI engineer with experience owning production systems in fintech and product startups, including a predictive scaling platform that cut AWS spend by 40% and an ambiguous social-intelligence feature that doubled MRR from $50K to $100K. Also building AI search and document-processing workflows, with reported 99.7% extraction accuracy and hands-on use of both classical forecasting and modern LLM stacks.”
“Software engineer currently building AI-powered backend systems for interview analysis, with end-to-end ownership of an LLM-based monitoring platform. Stands out for combining practical product delivery in an ambiguous early-stage environment with measurable impact: over 40% reduction in manual review effort and roughly 20% lower inference cost.”
Mid-level GenAI Engineer specializing in LLM fine-tuning, RAG, and MLOps
“Healthcare-focused LLM engineer who deployed a production triage and clinical knowledge retrieval assistant using RAG and LangGraph-orchestrated multi-agent workflows. Emphasizes clinical safety and compliance with robust hallucination controls, HIPAA/PHI protections (tokenization, encryption, audit logging, zero-retention), and human-in-the-loop escalation; reports a 75% latency reduction in a healthcare agent system.”
Junior Machine Learning Engineer specializing in Generative AI and analytics automation
“AI/LLM engineer who built a production intelligent support system using RAG over a vectorized documentation library, addressing real-world issues like lost-in-the-middle context failures and doc freshness via automated GitHub-driven re-embedding pipelines. Emphasizes rigorous agent evaluation (component/E2E/ops) and prefers lightweight, decoupled workflow automation using message brokers (Redis/RabbitMQ) over heavyweight orchestration frameworks.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMs, and NLP
“AI/ML engineer with forensic analytics and healthcare claims experience (Optum), building production LLM/RAG systems to surface context-driven fraud patterns from unstructured claim notes and explain risk to investigators. Strong in large-scale retrieval performance tuning, legacy API integration with reliability patterns (SQS, circuit breakers), and MLOps orchestration on Airflow/Kubernetes with rigorous testing, monitoring, and stakeholder-friendly interpretability.”
Mid-level AI/ML Engineer specializing in Generative AI and data engineering
“IBM engineer who built and deployed a production RAG-based LLM assistant using LangChain/FAISS with a fine-tuned LLaMA model, served via FastAPI microservices on Kubernetes, achieving 99%+ uptime. Demonstrates strong practical expertise in reducing hallucinations (semantic chunking + metadata-driven retrieval) and managing latency, plus mature MLOps practices (Airflow/dbt pipelines, MLflow tracking, monitoring, A/B and shadow deployments) and effective collaboration with non-technical stakeholders.”
Mid-level AI/ML Engineer specializing in Generative AI and production ML systems
“At CVS Health, the candidate productionized a RAG-based LLM solution in a regulated healthcare setting, emphasizing reliable data pipelines, LoRA fine-tuning, monitoring, safety guardrails, and A/B testing. They have hands-on experience troubleshooting real-time RAG failures (e.g., chunking/embedding issues) and regularly lead developer-focused demos/workshops while translating technical architecture into business value for stakeholders.”
Mid-level Full-Stack Engineer specializing in enterprise AI systems
“Built and productionized an AI NL-to-SQL capability inside legacy accounts receivable software (React + Spring Boot + Postgres/pgvector RAG), adding semantic caching and a SELECT-only validation layer to satisfy infosec. Achieved measurable impact (3 days to seconds turnaround, 60% token cost reduction, 50% latency reduction) with strong adoption (40 analysts, 50+ queries/week) and documented/monitored via Confluence + logging and user feedback loops.”
Mid-level Backend Software Developer specializing in cloud-native microservices
“LLM-focused engineer who has shipped multiple production-grade AI reliability systems: an LLM output validation/monitoring service (FastAPI) with prompt versioning and failure analytics, plus a RAG feature using embeddings/vector DBs with retrieval thresholds, schema/context validation, and safe fallbacks. Strong in evaluation loops (groundedness, schema accuracy, human review) and scalable pipelines for messy document ingestion with observability and early detection of data quality issues.”