Pre-screened and vetted.
Executive AI Architect specializing in enterprise cloud and FinTech solutions
“Candidate brings an operator-to-founder profile with leadership experience in IT and Business Systems and a strong grasp of how ideas become venture-backable products. They speak fluently about startup evaluation criteria such as TAM, technical defensibility, speed to scale, and AI differentiation, and appear especially motivated by building solutions end-to-end in startup or venture studio environments.”
Director-level AI & Data Science leader specializing in GenAI, LLMs, and MLOps
“ML/NLP engineer currently working in NYC on a system that connects complex unstructured data sources to deliver personalized insights, using embeddings + vector DB retrieval and a RAG architecture (LangChain, Pinecone/OpenSearch). Strong focus on production constraints—especially low-latency retrieval—using FAISS/ANN, PCA, index partitioning, and Redis caching, plus PEFT fine-tuning (LoRA/QLoRA) and KPI/SLA-driven promotion to production.”
Mid-level AI/ML Engineer specializing in Generative AI, RAG, and Conversational AI
“Built a production RAG-based GenAI copilot backend at Aetna using Python/FastAPI, GPT-4, LangChain, and Azure AI Search, deployed on AKS with Prometheus/Grafana observability. Owned the system end-to-end (ingestion through deployment) and improved peak-time reliability by addressing vector search and embedding bottlenecks with Redis caching, index optimization, and async processing, plus added anti-hallucination guardrails via retrieval confidence thresholds.”
Intern Software Engineer specializing in edge AI deployment and distributed systems
“Full-stack engineer who built an enterprise search platform (Codlens) delivering natural-language Q&A over Jira/Slack using embeddings, vector DB search, re-ranking (RRF), and LLM responses with source grounding. Also designed and benchmarked a distributed IAM system with Postgres transaction-log replication and Raft-based quorum consistency, reporting ~253 TPS at ~60ms latency in a multi-node setup. Experience spans early-stage startups (Zetic AI, Sagwara Capital) and large-scale orgs (Akamai, Atlassian).”
Mid-level AI Engineer specializing in GenAI, NLP, and MLOps
“LLM/agentic-systems engineer with PayPal experience hardening an LLM-powered fraud support assistant from prototype to production, focusing on low-latency distributed architecture, rigorous evaluation/testing, and security/compliance. Comfortable in customer-facing and GTM contexts—runs technical demos/workshops, builds tailored pilots, and aligns sales/CS with engineering to close deals and drive adoption.”
Mid-level AI/ML Engineer specializing in NLP, Generative AI, and MLOps
“Internship experience shipping production AI systems: built an end-to-end RAG platform (Python/FastAPI + LangChain/LangGraph + vector search) to answer support questions from unstructured internal docs, with a strong focus on hallucination prevention through confidence gating and rigorous offline/online evaluation. Also delivered an AI-driven personalization/analytics feature using an unsupervised clustering pipeline, iterating with PMs to align statistically strong clusters with actionable business segmentation.”
Mid-level Data Scientist specializing in Generative AI, LLMOps, and clinical data pipelines
“LLM/RAG engineer who has built and deployed corporate-scale systems at Novartis and Johnson & Johnson, including a healthcare AI agent that generates day-to-day treatment schedules. Recently handled a high-stakes safety incident (LLM suggesting overdose) by tightening model instructions and validating with ~200 test prompts, and has strong end-to-end data/embedding/vector DB pipeline experience (PySpark, FAISS, Pinecone) plus SME-in-the-loop evaluation (RLHF).”
Mid-level GenAI/ML Engineer specializing in LLM agents and RAG for Financial Services & Healthcare
“Built and deployed a production GenAI internal support agent at Bank of America (“Ask GPS/AskGPT”) using RAG on Azure, focused on reducing escalations and improving response quality for repetitive knowledge-based queries. Demonstrates strong production LLM engineering: custom LangChain orchestration, retrieval tuning to reduce hallucinations, rigorous offline/online evaluation, and model benchmarking with dynamic routing (e.g., GPT-4 vs Claude).”
“ServiceNow engineer who built and launched a production LLM-powered ticket resolution/knowledge assistant using RAG (LangChain + Hugging Face embeddings + vector search) integrated into internal support dashboards via REST APIs. Optimized the system from ~6–8s to ~2–3s latency while improving usability with concise, cited answers and guardrails (grounding + similarity thresholds), delivering ~30–35% reduction in manual ticket investigation effort.”
Mid-level AI/ML Engineer specializing in GenAI, RAG, and healthcare ML
“Built an end-to-end GenAI/RAG platform for financial compliance and research at BlackRock, focused on safe, auditable answers in a highly regulated environment. Combines strong LLM engineering depth with production platform skills and delivered clear business impact, including reducing research/compliance turnaround from hours to seconds, improving retrieval relevance by 22%, and cutting inference costs by 75%.”
Mid-level Software Development Engineer specializing in cloud-native AI/ML systems
“AI/ML-focused engineer with practical experience building RAG-based and multi-agent systems, including architectures for retrieval, reasoning, context processing, and response generation. Stands out for combining LLM productivity gains with disciplined software engineering practices like validation, monitoring, and reproducibility.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Mid-Level Software Engineer specializing in secure cloud microservices and FinTech
“Built and owned major parts of a real-time distributed AI fraud-detection pipeline (ingestion, inference microservice integration, and automated action layer), optimizing latency and observability and reducing false positives by ~35%. Understands ROS/ROS2 concepts (nodes/topics/services) and planned hands-on ramp-up via ROS2 pub/sub exercises and Gazebo simulation, but has not worked on physical robots or ROS in production.”
Mid-level Data/ML Engineer specializing in NLP, GenAI, and scalable data pipelines
“AI/ML engineer with production experience building LLM-powered document intelligence and customer support systems in healthcare/insurance, emphasizing high-accuracy RAG, long-document processing, and robust monitoring/fallback mechanisms. Also automates and scales ML lifecycle workflows using Apache Airflow and Kubeflow, and partners closely with non-technical operations stakeholders to drive adoption.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Red Hat ML/LLM engineer who designed and deployed a production LLM-powered customer support automation system using RAG, improving latency by 30% via PEFT and vector search optimization. Built security and governance into retrieval (access-level filtering, encrypted Pinecone/ChromaDB) and delivered SHAP-based explainability via a dashboard for non-technical stakeholders. Experienced orchestrating distributed ML/RAG pipelines across AWS SageMaker and OpenShift with Airflow/Prefect, plus multi-agent workflows using CrewAI and LangGraph.”
Mid-level Data Scientist specializing in NLP/LLMs, time series forecasting, and MLOps
“Data/ML practitioner with hands-on experience building NLP systems from prototype to production: delivered a Twitter sentiment classifier with robust preprocessing, SVM modeling, and Power BI reporting, and built entity-resolution pipelines for messy multi-source customer data (reporting ~95% improvement in unique entity identification). Also implemented semantic linking/search using SBERT embeddings with FAISS vector retrieval and domain fine-tuning (reported ~15% precision lift), and applies production workflow best practices (Airflow/Prefect, Docker, Azure ML/Databricks, Great Expectations).”
Junior Machine Learning Engineer specializing in LLMs and applied data science
“Built and shipped multiple production AI systems, including Auto DocGen (LLM-generated OpenAPI docs kept in sync via AST diffs, schema-constrained generation, and CI/CD on Render) and a multimodal sign-language recognition pipeline at USC orchestrated with FastAPI, MediaPipe, and PyTorch. Also partnered with Esri’s non-technical community team to fine-tune an LLaMA-based spam classifier with a review UI, cutting moderation time by 70%.”
Mid-level Software Engineer specializing in autonomous driving simulation and 3D mapping
“Founding software engineer who built an autonomous-vehicle 3D digital twin using Unreal Engine 5 and CARLA, owning core simulator logic (traffic/scenarios/weather) and a ROS 2-based pipeline to record synchronized multi-sensor data (RGB/depth/segmentation/LiDAR/IMU/GPS). Also implemented distributed synchronization patterns (server + client prediction) using FastAPI and WebSockets; seeking roles with H1B transfer and targeting ~$110k.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMOps, and MLOps
“Built and deployed an AWS-based LLM/RAG ticket triage and knowledge retrieval system (Pinecone/FAISS + Step Functions + MLflow) that cut support resolution time by 20%. Demonstrates strong production focus on hallucination reduction, PII security, and low-latency orchestration, with measurable evaluation improvements (e.g., ~25% grounding accuracy gain via re-ranking) and proven collaboration with support operations stakeholders.”
Junior Software Engineer specializing in AI/ML and cloud platforms
“LLM/agent engineer who shipped a production "Memory Assistant" at HydroX AI, building a LangChain/LlamaIndex RAG memory pipeline on ChromaDB/FAISS with robust fallbacks (BERT/BART), prompt-injection mitigation, and 99.9% uptime monitoring. Also built a multi-step customer support agent using Rasa + OpenAI Assistants API with structured tool calling, guardrails, and human-in-the-loop escalation, and has experience hardening agents against messy ERP data via Pydantic validation, idempotency, and transactional outbox patterns.”
Mid-level AI/ML Engineer specializing in GenAI, NLP, and MLOps
“Built and deployed an enterprise GenAI knowledge assistant over thousands of internal PDFs/reports using a RAG stack (GPT-4 + Hugging Face embeddings + vector DB) to reduce manual search and SME escalations. Uses LangGraph/LangChain to orchestrate modular agent workflows with relevance filtering and fallback handling, and applies rigorous evaluation (golden datasets, edge cases, A/B tests) with production monitoring metrics.”
Junior Software Engineer specializing in ML, distributed systems, and LLM applications
“Interned at Zonda where he built an AI-driven semantic search solution over ~280M housing/builder records. Iterated from local LLMs via llama.cpp quantization to a vector-embedding retrieval system, then boosted semantic accuracy with a custom spaCy NER layer and re-ranking, optimizing for latency through precomputation. Collaborated with economics-focused stakeholders to reduce manual document/paperwork time by enabling natural-language search over internal data.”
Junior Machine Learning & Quant Research Engineer specializing in low-latency data and trading systems
“Applied ML to physical EV fleet systems at ST Labs, building a real-time CNN-LSTM fault prediction pipeline from streaming vehicle telemetry and addressing live data alignment issues via resampling/interpolation and buffered inference. Also developed a V2G/G2V energy transfer algorithm to automate charging/discharging for profit optimization, and made high-impact low-latency pipeline decisions at Astera Holdings using profiling, replay testing, and live A/B validation.”