Pre-screened and vetted.
Mid-Level AI/ML Software Engineer specializing in agentic LLM systems
“Built and deployed a production LLM-powered multi-agent compliance copilot (life sciences/finance) using LangChain/LangGraph + RAG over vector databases, delivered via async FastAPI on Kubernetes. Emphasizes audit-ready, deterministic outputs with schema constraints and citations, plus rigorous evaluation/monitoring; reports 60%+ reduction in manual research time and successful production adoption.”
Mid-level AI Engineer specializing in GenAI, NLP, and MLOps
“LLM/agentic-systems engineer with PayPal experience hardening an LLM-powered fraud support assistant from prototype to production, focusing on low-latency distributed architecture, rigorous evaluation/testing, and security/compliance. Comfortable in customer-facing and GTM contexts—runs technical demos/workshops, builds tailored pilots, and aligns sales/CS with engineering to close deals and drive adoption.”
Mid-level AI/ML Engineer specializing in Generative AI, NLP, and Computer Vision
“Built an LLM-powered learning assistant (EduQuizPro/EduCrest Pro) that uses RAG over URLs and PDFs to generate quizzes, notes, and explanations for students/professors. Emphasizes production robustness—implemented dependency fallbacks (FAISS/Sentence Transformers/Gradio), CLI-safe mode, and NumPy-based indexing—along with a custom orchestration layer to keep multi-step AI workflows reliable.”
Mid-level Data Scientist / AI-ML Engineer specializing in Generative AI and LLM applications
“Built a production GenAI-powered analytics assistant to reduce reliance on data analysts by enabling natural-language Q&A over Databricks/Power BI dashboards, backed by vector search (Pinecone/Milvus) and a Neo4j knowledge graph, including multimodal support via OpenAI Vision. Demonstrates strong real-world LLM reliability engineering with strict RAG, LangGraph multi-step verification, and Guardrails/custom validators, plus broad orchestration and production monitoring experience (Airflow, ADF, Step Functions, Kubernetes, Prometheus/CloudWatch).”
Director-level AI & Data Science leader specializing in GenAI, LLMs, and MLOps
“ML/NLP engineer currently working in NYC on a system that connects complex unstructured data sources to deliver personalized insights, using embeddings + vector DB retrieval and a RAG architecture (LangChain, Pinecone/OpenSearch). Strong focus on production constraints—especially low-latency retrieval—using FAISS/ANN, PCA, index partitioning, and Redis caching, plus PEFT fine-tuning (LoRA/QLoRA) and KPI/SLA-driven promotion to production.”
Principal Data Scientist specializing in healthcare analytics and medical imaging AI
“Developed an LLM-driven recommendation agent in Azure Databricks to triage oncology patients and trigger second-opinion case creation using medical claims and EHR data. Uses ICD-10/CPT/J-code features in prompts, embeddings + vector DB similarity, and a backtesting framework emphasizing recall to avoid missing clinically relevant cases while supporting business revenue.”
Mid-level AI/ML Engineer specializing in Generative AI, RAG, and Conversational AI
“Built a production RAG-based GenAI copilot backend at Aetna using Python/FastAPI, GPT-4, LangChain, and Azure AI Search, deployed on AKS with Prometheus/Grafana observability. Owned the system end-to-end (ingestion through deployment) and improved peak-time reliability by addressing vector search and embedding bottlenecks with Redis caching, index optimization, and async processing, plus added anti-hallucination guardrails via retrieval confidence thresholds.”
“Built and productionized an AI-native, agentic appeals decisioning system for health insurance operations, automating 500k+ scanned appeals/year. Delivered measurable impact by cutting review time from 12–15 minutes to ~3 minutes and auto-resolving ~85% of cases with strong auditability, evaluations, and human-in-the-loop guardrails, deployed as containerized microservices on Azure AKS.”
Mid-level Backend Software Engineer specializing in distributed cloud-native systems
“Backend/AI workflow engineer who built production-grade orchestration systems for hardware security verification at Silicon Assurance (Nextflow/Python/Postgres) and a multi-agent LLM-driven regulatory code checking system at the University of Florida. Emphasizes reliability: strict plan/execute/verify boundaries, queue-based isolation, and strong observability/auditability with Prometheus/Grafana and persisted prompts/tool calls.”
“LLM engineer who has deployed production RAG systems for regulated document QA (PDFs/knowledge bases), emphasizing grounded answers with citations, RBAC, monitoring, and continuous feedback. Demonstrates deep practical expertise in retrieval quality (semantic chunking, hybrid BM25+embeddings, re-ranking), reliability (guardrails, deterministic workflows), and measurable evaluation (golden sets, log replay, A/B tests) while partnering closely with compliance/operations stakeholders.”
Mid-level Data Scientist / Machine Learning Engineer specializing in fraud, risk, and MLOps
“AI/ML practitioner with Northern Trust experience who has shipped production LLM systems (internal support assistant) using RAG, vector databases, orchestration (LangChain/custom pipelines), and rigorous monitoring/feedback loops. Also built AI-driven fraud detection/risk monitoring solutions in a regulated financial environment, emphasizing explainability (SHAP), audit readiness, and stakeholder trust through dashboards and clear communication.”
Mid-level Python & AI/ML Engineer specializing in backend APIs and MLOps
“Built and deployed a production LLM/RAG document automation system for business documents (contracts/claim forms) that extracts schema-validated JSON, generates grounded summaries/Q&A, and integrates into transaction systems via APIs. Emphasizes real-world reliability: hallucination controls, layout-aware parsing with OCR fallback, Step Functions-orchestrated workflows with retries/timeouts, and human-in-the-loop review designed in close partnership with operations and claims stakeholders.”
Mid-level Data Scientist specializing in Generative AI, LLMOps, and clinical data pipelines
“LLM/RAG engineer who has built and deployed corporate-scale systems at Novartis and Johnson & Johnson, including a healthcare AI agent that generates day-to-day treatment schedules. Recently handled a high-stakes safety incident (LLM suggesting overdose) by tightening model instructions and validating with ~200 test prompts, and has strong end-to-end data/embedding/vector DB pipeline experience (PySpark, FAISS, Pinecone) plus SME-in-the-loop evaluation (RLHF).”
Junior Data Scientist specializing in Generative AI and applied machine learning
“At Evoke Tech, built a production LLM "Testbench" to quickly compare LLMs/embedding models and RAG strategies (semantic, hybrid BM25, re-ranking, HyDE, query expansion) to select optimal architectures for different client needs. Also developed a multi-agent, multimodal (voice/text) RAG system for live catalog retrieval and safe product recommendations using LangGraph/LangChain with LangSmith monitoring, and regularly translated PM/UX goals into concrete agent behaviors via demos and flowcharts.”
Junior Machine Learning Engineer specializing in LLMs, RAG, and medical imaging
“At Fileread, the candidate built and deployed an LLM-powered legal document classification and retrieval layer for an agentic extraction system that turns unstructured legal PDFs into structured tables with line-level citations. They productionized a RAG-style pipeline (ingestion, embeddings, retrieval, reranking, generation) and report 95%+ F1 across 70+ legal categories, emphasizing rigorous evaluation and close collaboration with legal domain experts for high-stakes precision.”
Mid-level Software Engineer specializing in AI agents, backend systems, and data engineering
“Amazon engineer who built a production AI agent platform (Python/AWS Strands on Bedrock) that lets teams create tool-using, multi-agent workflows—e.g., agents that auto-triage and resolve customer support tickets by reading internal documentation and collaborating with a research agent. Previously worked in Deloitte on IAM using Ping Identity/Ping DaVinci orchestration, and applies orchestration thinking plus structured evaluation (LLM-as-judge, surveys, automated tests) to improve agent reliability.”
Senior Software Engineer specializing in full-stack systems, data pipelines, and ML
“Built and productionized an autonomous research agent (AutoGPT) in a Docker/Kubernetes environment with Pinecone-based long-term memory and custom Python tools for analysis, visualization, and report drafting. Implemented layered guardrails (prompt templates, automated validation, self-critique loops, and monitoring) and achieved ~25% reduction in manual report generation time while scaling the workflow to support multiple concurrent users.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Mid-level Software Engineer specializing in FinTech full-stack and AI applications
“Built and productionized an NLP-powered customer support assistant at JPMorgan Chase for digital banking, focused on reducing response time for repetitive client queries. Strong in real-world AI deployment challenges—sensitive data handling, low-latency FastAPI services, and AWS/Kubernetes operations with CI/CD—plus a metrics- and guardrails-driven approach to reliable AI workflows.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
“GenAI/data engineering practitioner with production experience across Equinix, Optum, and Citibank—built an Azure OpenAI (GPT-4) + LangChain document intelligence platform processing 1.5M+ docs/month and a HIPAA-compliant Airflow healthcare pipeline handling 5M+ claims/day. Also delivered a real-time fraud detection + explainability system using LightGBM and a fine-tuned T5 NLG component, improving fraud accuracy by 15%+ while partnering closely with compliance stakeholders.”
Mid-level Machine Learning Engineer specializing in LLMs and NLP classification systems
“Internship experience building a production RAG+LLM pipeline to map messy card transaction descriptions to merchant brands, including a custom modified-ROUGE evaluation approach for weak/variant ground truth. Improved scalability and cost by moving from a managed LLM endpoint (e.g., Bedrock) to self-hosted vLLM, and orchestrated massive embedding backfills (5,000+ files, 10B+ rows) using an Airflow-triggered SQS + ECS worker architecture with robust retry/DLQ handling.”
“Backend/AI engineer who built a real-time vector database system for high-frequency financial data using Kafka/Flink on Kubernetes, achieving sub-100ms similarity search at 10k+ concurrent load and resolving tricky duplication issues with idempotency/versioning. Also shipped an end-to-end LLM-based travel itinerary feature (profiling + prompt workflows + APIs) with a focus on quality consistency and low latency.”
Mid-level AI Engineer specializing in Generative AI, MLOps, and NLP for finance and healthcare
“Built and deployed a secure, production LLM-based document summarization and risk-highlighting tool for financial auditors, running inside a private Azure environment to protect confidential data. Focused on reliability (hallucination mitigation via retrieval-based prompts and source citations) and validated performance through comparisons to auditor summaries plus a user pilot, cutting review time by about half.”