Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in LLM agents, RAG, and enterprise ML systems
“Built a production multi-agent recommendation/RAG system for internal data analysts to speed up weekly report creation by improving document discovery and automating report/SQL generation. Implemented LangGraph-based orchestration with deterministic agent routing, robust error handling (interrupt/resume), and metadata-driven semantic chunking for diverse PDF/document formats, plus monitoring for latency, throughput, and token/cost efficiency.”
Mid-level AI Engineer specializing in GenAI, NLP, and MLOps
“LLM/agentic-systems engineer with PayPal experience hardening an LLM-powered fraud support assistant from prototype to production, focusing on low-latency distributed architecture, rigorous evaluation/testing, and security/compliance. Comfortable in customer-facing and GTM contexts—runs technical demos/workshops, builds tailored pilots, and aligns sales/CS with engineering to close deals and drive adoption.”
“LLM/agent workflow engineer with healthcare experience (CVS/CBS Health) who built and deployed a production call-insights platform using Azure OpenAI + LangChain/LangGraph, including sentiment and compliance checks. Demonstrates deep HIPAA/PHI handling (tenant-contained processing, redaction, RBAC/encryption/audit logging) and production rigor (testing, eval sets, validation/retries, autoscaling) to scale to thousands of transcripts.”
“Built and deployed a production Retrieval-Augmented Generation (RAG) platform in a healthcare setting to automate clinical documentation review and summarization, targeting near-real-time, explainable outputs. Emphasizes grounded generation to reduce hallucinations, latency optimizations (chunking/embedding reuse), and PHI-safe workflows with access controls, plus strong orchestration experience using Apache Airflow.”
Mid-level Machine Learning Engineer specializing in NLP, LLMs, and multimodal modeling
“Built and productionized a telecom-focused RAG assistant by LoRA fine-tuning LLaMA-2 and integrating LangChain+FAISS behind a FastAPI service, with dashboards and a human feedback UI for engineers. Demonstrated measurable impact (≈40% faster document lookup, +8–10% retrieval precision) and strong MLOps rigor via Airflow orchestration, CI/CD, and monitoring for drift and failures.”
Mid-level AI/ML Engineer specializing in Generative AI, NLP, and Computer Vision
“Built an LLM-powered learning assistant (EduQuizPro/EduCrest Pro) that uses RAG over URLs and PDFs to generate quizzes, notes, and explanations for students/professors. Emphasizes production robustness—implemented dependency fallbacks (FAISS/Sentence Transformers/Gradio), CLI-safe mode, and NumPy-based indexing—along with a custom orchestration layer to keep multi-step AI workflows reliable.”
Mid-level Data Scientist / AI-ML Engineer specializing in Generative AI and LLM applications
“Built a production GenAI-powered analytics assistant to reduce reliance on data analysts by enabling natural-language Q&A over Databricks/Power BI dashboards, backed by vector search (Pinecone/Milvus) and a Neo4j knowledge graph, including multimodal support via OpenAI Vision. Demonstrates strong real-world LLM reliability engineering with strict RAG, LangGraph multi-step verification, and Guardrails/custom validators, plus broad orchestration and production monitoring experience (Airflow, ADF, Step Functions, Kubernetes, Prometheus/CloudWatch).”
Director-level AI & Data Science leader specializing in GenAI, LLMs, and MLOps
“ML/NLP engineer currently working in NYC on a system that connects complex unstructured data sources to deliver personalized insights, using embeddings + vector DB retrieval and a RAG architecture (LangChain, Pinecone/OpenSearch). Strong focus on production constraints—especially low-latency retrieval—using FAISS/ANN, PCA, index partitioning, and Redis caching, plus PEFT fine-tuning (LoRA/QLoRA) and KPI/SLA-driven promotion to production.”
Staff Full-Stack Engineer specializing in AI platforms and infrastructure automation
“Backend/full-stack engineer building complex internal platforms and customer-facing demos at the intersection of infrastructure and product. Shipped a no-code Product Lifecycle Manager for manufacturing (3 manufacturers, 1000+ evolving tests) using AWS S3/SQS ingestion and extensible Postgres (EAV+JSONB) with end-to-end traceability. Also built a FastAPI-based company data intelligence platform with Okta-secured RBAC and an LLM/MCP layer for ChatGPT-like analytics over enterprise data sources.”
Mid-level AI/ML Engineer specializing in deep learning, NLP/LLMs, and MLOps
“Built and shipped a real-time oncology risk prediction system used by doctors during patient visits, trained on clinical data in AWS SageMaker and deployed via FastAPI with sub-second responses. Emphasizes clinician-trust features (SHAP explainability, validation checks) and HIPAA-compliant controls (encryption, RBAC, audit logging), plus Kubernetes-based production operations with autoscaling, monitoring, and drift/retraining workflows; collaborated closely with oncologists at Flatiron Health.”
Principal Data Scientist specializing in healthcare analytics and medical imaging AI
“Developed an LLM-driven recommendation agent in Azure Databricks to triage oncology patients and trigger second-opinion case creation using medical claims and EHR data. Uses ICD-10/CPT/J-code features in prompts, embeddings + vector DB similarity, and a backtesting framework emphasizing recall to avoid missing clinically relevant cases while supporting business revenue.”
Mid-level AI/ML Engineer specializing in Generative AI, RAG, and Conversational AI
“Built a production RAG-based GenAI copilot backend at Aetna using Python/FastAPI, GPT-4, LangChain, and Azure AI Search, deployed on AKS with Prometheus/Grafana observability. Owned the system end-to-end (ingestion through deployment) and improved peak-time reliability by addressing vector search and embedding bottlenecks with Redis caching, index optimization, and async processing, plus added anti-hallucination guardrails via retrieval confidence thresholds.”
Mid-level GenAI Engineer specializing in production RAG and LLM fine-tuning
“LLM engineer who built a production seller-support RAG system at eBay using hybrid retrieval (BM25 + Pinecone vectors) with Cohere reranking, LangGraph orchestration, and citation-grounded answers. Strong focus on reliability: semantic/structure-aware chunking, automated Ragas-based evaluation with nightly regressions, and production observability (LangSmith) plus drift monitoring (Arize). Also implemented a multi-agent fraud pipeline with AutoGen using JSON-schema contracts and explicit termination conditions.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“AI engineer who built a production RAG-based internal analyst tool at BlackRock, fine-tuning an LLM on proprietary financial data and adding four layers of guardrails (input/retrieval/generation/output) to improve grounding and reduce hallucinations. Implemented a LangChain-based multi-agent orchestration (7 major agents) deployed on AWS ECS, with reliability measured via internal human evaluation, LLM-as-judge, and RLHF/drift monitoring.”
Mid-level Data Scientist / Machine Learning Engineer specializing in fraud, risk, and MLOps
“AI/ML practitioner with Northern Trust experience who has shipped production LLM systems (internal support assistant) using RAG, vector databases, orchestration (LangChain/custom pipelines), and rigorous monitoring/feedback loops. Also built AI-driven fraud detection/risk monitoring solutions in a regulated financial environment, emphasizing explainability (SHAP), audit readiness, and stakeholder trust through dashboards and clear communication.”
Mid-level AI/ML Engineer specializing in fraud detection and risk analytics in Financial Services
“At JP Morgan Chase, built and deployed a production LLM-powered RAG knowledge assistant to help fraud investigators and risk analysts quickly navigate regulatory updates and internal policies, reducing investigation delays and compliance risk. Strong focus on secure retrieval (RBAC filtering), reliability (layered testing + observability), and production constraints (latency/SLOs), with Airflow-orchestrated, auditable ML pipelines.”
Mid-level GenAI/ML Engineer specializing in LLM agents and RAG for Financial Services & Healthcare
“Built and deployed a production GenAI internal support agent at Bank of America (“Ask GPS/AskGPT”) using RAG on Azure, focused on reducing escalations and improving response quality for repetitive knowledge-based queries. Demonstrates strong production LLM engineering: custom LangChain orchestration, retrieval tuning to reduce hallucinations, rigorous offline/online evaluation, and model benchmarking with dynamic routing (e.g., GPT-4 vs Claude).”
Mid-level AI/ML Engineer specializing in NLP, MLOps, and scalable data pipelines
“Built and shipped a production LLM-powered personalized client engagement assistant in the financial domain, balancing real-time recommendations with strict privacy/compliance requirements. Demonstrates strong MLOps/LLMOps depth (Airflow + MLflow, containerized microservices, drift monitoring) and a privacy-by-design approach validated in collaboration with risk and compliance teams.”
Mid-level Software Engineer specializing in cloud-native microservices and workflow automation
“Enterprise platform engineer/product owner who led end-to-end delivery of customer-facing ServiceNow Service Catalog/workflow solutions, emphasizing reliability, security, and fast iteration. Built React/TypeScript portals with Node.js and Spring Boot backends, and improved microservices reliability at scale using Kafka, monitoring, and robust retry/timeout patterns.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
Mid-Level Software Engineer specializing in FinTech payments and fraud detection
“Backend/platform engineer with payments domain experience, having owned core services for MasterCard’s global card tokenization and settlement platform. Built Django/Celery microservices plus Kafka/Redis real-time fraud streaming, delivering 27% latency improvement, sub-100ms fraud checks, and 18% fewer false positives. Strong DevOps/IaC background across Kubernetes, AWS ECS, Terraform, GitHub Actions, and GitOps practices for high-scale transaction systems (including UPI at PhonePe).”
“GenAI/data engineering practitioner with production experience across Equinix, Optum, and Citibank—built an Azure OpenAI (GPT-4) + LangChain document intelligence platform processing 1.5M+ docs/month and a HIPAA-compliant Airflow healthcare pipeline handling 5M+ claims/day. Also delivered a real-time fraud detection + explainability system using LightGBM and a fine-tuned T5 NLG component, improving fraud accuracy by 15%+ while partnering closely with compliance stakeholders.”
Mid-Level Software Engineer specializing in cloud infrastructure and microservices
“Backend engineer who has led major platform evolution to cloud-native microservices (Spring Boot on AWS with Terraform) and built scalable, secure FastAPI APIs. Demonstrates strong production rigor with metric-driven validation, canary/phased rollouts, and incremental migrations using shadow traffic/feature flags/parallel writes—achieving faster deployments, fewer incidents, and zero-downtime traffic spikes and migrations.”