Pre-screened and vetted.
Mid-level Data Scientist specializing in GenAI, LLM orchestration, and MLOps
Senior Data Scientist specializing in ML, fraud risk, and Generative AI (RAG/LLMs)
Mid-level AI/ML Engineer specializing in conversational AI, NLP, and LLM-powered RAG systems
Junior Software Engineer specializing in AI, LLM systems, and full-stack development
“Product-focused full-stack engineer at startup (Zippy) who shipped a production multi-agent AI system for restaurant operations plus payments workflows. Built end-to-end: RAG grounded on a Notion knowledge base, structured function-calling task routing, FastAPI/JWT multi-tenant backend, and a polished React+TypeScript owner dashboard. Has real production incident experience (duplicate Stripe webhooks) and reports ~94% task-routing accuracy under load.”
Mid-level Machine Learning Engineer specializing in Generative AI and RAG systems
“LLM/ML engineer who has shipped an enterprise RAG-based Q&A system (LangChain/LlamaIndex, FAISS + Azure Cognitive Search, GPT-3.5/4 via OpenAI/Azure OpenAI) to production on Docker + Kubernetes/OpenShift, tackling hallucinations, retrieval quality, latency/cost, and RBAC/IAM security. Also partnered with operations leaders to turn manual reporting into an LLM-powered summarization and forecasting dashboard driven by real KPIs and iterative stakeholder feedback.”
Mid-level Forward Deployed Engineer specializing in AI automation for finance and data platforms
“LLM/agentic workflow specialist with healthcare deployment experience who has taken LLM-based automation from prototype to production using operator-in-the-loop validation, RAG-style retrieval, RBAC, and monitoring for sensitive data compliance. Demonstrated real-time incident resolution (retrieval timeouts due to network/proxy misconfig) and strong GTM support—hands-on developer workshops and sales demos translating technical safeguards and real-time ETL into measurable ROI (70% ops reduction, ~$200K/year savings).”
Senior AI/ML Engineer specializing in Generative AI and RAG
“ML/NLP practitioner at Morf Health focused on unifying fragmented healthcare data by linking structured patient/encounter records with unstructured clinical notes. Has hands-on experience with transformer embeddings, vector databases, and domain fine-tuning, plus rigorous evaluation (precision/recall) and human-in-the-loop validation with clinical SMEs to make pipelines production-grade.”
Mid-Level Software Engineer specializing in backend systems and cloud-native platforms
“Software engineer with experience across TCS, Rakuten, and USC who has owned production integrations and data pipelines end-to-end. Notably improved a trading platform payment flow by replacing fragile polling with a webhook-driven status system with robust fallbacks, and has shipped LLM-assisted design-to-webpage automation plus evaluation-driven prompt iteration (NYT Connections).”
Mid-level Full-Stack Software Engineer specializing in Generative AI
“Full-stack engineer who shipped an end-to-end speech capability for an LLM chatbot UI, integrating OpenAI APIs and publishing via Google Apigee with client documentation. Has experience operating deployments with Jenkins/Kubernetes/Docker and monitoring with Datadog, and has worked in an innovation-center environment building rapid prototypes under ambiguity with tight stakeholder feedback loops.”
“Backend engineer focused on productionizing LLM systems: built a FastAPI-based RAG and multi-agent automation platform deployed with Docker/Kubernetes, prioritizing safe execution and reduced hallucinations. Experienced in refactoring monolithic ML services with feature-flagged incremental rollouts, and implementing JWT/RBAC plus row-level security (e.g., Supabase) for secure, scalable APIs.”
Senior GenAI/ML Engineer specializing in LLMs, RAG, and multimodal generative AI
“LLM/RAG engineer with production deployments in highly regulated domains (Frost Bank and GE Healthcare). Built secure, explainable document-grounded Q&A systems using LoRA fine-tuning, strict RAG with confidence thresholds, and citation-based responses; also established evaluation/monitoring (golden QA sets, hallucination tracking, drift) and achieved ~40% latency reduction through retrieval/prompt tuning.”
Mid-level Full-Stack Java Developer specializing in cloud-native microservices
“Full-stack engineer focused on enterprise, cloud-native microservices—building Spring Boot backends and React/Angular front ends with strong security (OAuth/JWT), AWS infrastructure (RDS/S3), and containerized deployments (Docker/Kubernetes). Has delivered data-heavy order/account/transaction platforms and healthcare solutions including EHR integrations for secure patient data exchange, with emphasis on testing, performance tuning, and reliability (load testing).”
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and Clinical AI
“Built and productionized a HIPAA-compliant LLM+RAG Clinical AI assistant at Optum, fine-tuning GPT/LLaMA on de-identified patient notes and integrating FAISS/Pinecone for sub-second retrieval; reported to cut diagnosis time by ~20 minutes per case. Experienced in orchestrating ML pipelines (Airflow, AWS Step Functions, Azure Data Factory) and in reliability techniques for LLM systems (grounding, citations, confidence filters, monitoring) while partnering closely with clinicians and compliance teams.”
Mid-level AI/ML Engineer specializing in Generative AI and LLMOps
“Built and deployed a GPT-based RAG enterprise search system for healthcare clinicians, emphasizing low-latency performance and reduced hallucinations while maintaining end-to-end HIPAA compliance. Demonstrates deep applied experience with PHI-safe data governance (detection/redaction/de-identification), secure Azure ML deployment patterns, and orchestration of production LLM workflows using LangChain and Airflow.”
Mid-level GenAI Engineer specializing in production AI agents and evaluation pipelines
“Built and shipped a production LLM-powered internal operations automation platform using LangChain RAG (Pinecone) and FastAPI microservices, deployed on AWS EKS, serving 10k+ daily interactions. Implemented a rigorous evaluation/observability stack (golden datasets, prompt regression tests, MLflow, retrieval metrics, hallucination monitoring) that drove hallucinations below 2% and improved reliability, and partnered closely with non-technical ops leaders to cut manual lookup work by 60%+.”
Mid-level Data & GenAI Engineer specializing in lakehouse, streaming, and RAG platforms
“Built a production internal LLM-powered knowledge assistant using a RAG architecture (Python, LLM APIs, cloud services) that answers employee questions with sourced, grounded responses from internal documents. Demonstrates strong practical depth in retrieval tuning (chunking/metadata filters), orchestration with LangChain, and production reliability practices (latency optimization, automated embedding refresh, evaluation metrics, logging/monitoring) while partnering closely with non-technical operations teams.”
Junior Data Scientist specializing in healthcare ML and clinical NLP/LLMs
“Healthcare-focused LLM engineer who has built two production clinical applications: an automated structured clinical report generator from physician-patient conversations and a RAG-based chatbot for retrieving patient history (procedures, allergies, etc.). Demonstrates strong applied RAG expertise (overlapping chunking, entity dependency graphs, temporal filtering, graph RAG) to reduce hallucinations/omissions and partners closely with clinicians to automate hospital workflows.”
Mid-level AI Engineer specializing in multi-agent LLM systems and multimodal tutoring
“LLM/agentic systems builder who has deployed multi-agent educational chatbots using LangChain + LangGraph, with LangFuse-based tracing and FastAPI hosting. Focused on production reliability and performance (latency reduction via agent decomposition and caching) and on evaluation/testing (routing test scenarios, LLM-as-judge). Partnered with product to add image understanding by parsing and storing images in S3, expanding chatbot coverage to 30+ books with images.”
Mid-level AI/ML Engineer specializing in GenAI, LLMs, and computer vision
“Built and productionized a multi-agent, LLM-powered document understanding system to replace manual review of long documents, using LangGraph orchestration plus RAG to reduce hallucinations. Implemented layered reliability controls (structured templates, checker agent, and human-in-the-loop feedback) and reported ~40% speed improvement after orchestration; also has hands-on Airflow experience for scheduled data pipelines.”
Mid-level AI/ML Engineer specializing in LLM reasoning and inference optimization
“Former Field Application Engineer at TestTorrent supporting proprietary hardware deployments. Built and integrated customer-requested benchmarking (including MLPerf-style work) by adapting GPU benchmark implementations and validating performance against NVIDIA baselines, plus delivered a backend fix that enabled customer performance testing and helped close a deal. Also has experience running models on AWS/GCP with multi-instance cost optimization techniques like memory offloading.”
Mid-level AI/ML Engineer specializing in fraud detection and risk analytics in Financial Services
“Finance-domain ML/LLM engineer who has shipped production systems including a RAG-based financial insights assistant with a custom post-generation validation layer that verifies atomic claims against retrieved source text to prevent hallucinations in compliance-critical workflows. Also built large-scale MLOps automation on AWS using Kubeflow + MLflow + CI/CD for fraud detection and credit risk models processing 500M+ transactions/day with a 99.99% uptime goal, and partnered closely with JP Morgan risk/compliance stakeholders on NLP-driven compliance monitoring.”
Mid-level Generative AI Engineer specializing in LLM apps, RAG, and MLOps
“LLM/GenAI engineer with US Bank experience building a production financial-document intelligence platform using LangChain/LangGraph, GPT-4, and Amazon OpenSearch. Delivered a RAG-based assistant for compliance/audit teams with grounded, cited answers, focusing on reducing hallucinations and latency, and deployed securely on AWS (SageMaker/EKS) with CI/CD and evaluation tooling (LangSmith, RAGAS).”