Pre-screened and vetted.
Staff Full-Stack & AI Engineer specializing in LLM platforms and scalable cloud systems
Mid-level AI/ML Engineer specializing in Generative AI agents and FinTech risk systems
Senior AI/ML Engineer specializing in LLMs, RAG, and multimodal recommendation systems
Mid-level AI/ML Engineer specializing in LLMs, RAG, and multi-agent systems
Mid-level AI/ML Engineer specializing in Generative AI, RAG, and MLOps
“AI/LLM engineer with production experience at NVIDIA and Microsoft, including building a RAG-based enterprise knowledge assistant that improved accuracy by 42% and scaled to thousands of queries. Deep in inference optimization (TensorRT-LLM, Triton, quantization, speculative decoding) and MLOps/observability (Prometheus/Grafana, MLflow, LangSmith), plus orchestration with Kubeflow/Airflow across multi-cloud.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and distributed MLOps
Senior Full-Stack AI/ML Engineer specializing in cloud data platforms and GenAI
Executive Engineering Leader specializing in data platforms, cloud modernization, and AI
Principal Machine Learning Scientist specializing in GenAI, LLMs, and RAG
Mid-level AI/ML Engineer specializing in LLMs, RAG, and production MLOps
Mid-level Software Engineer specializing in Python, distributed systems, and AI backend services
Senior Backend Software Engineer specializing in healthcare platforms and AI/ML tooling
“Built a chatbot for a learning management system during a Deep Atlas bootcamp by mapping an end-to-end RAG architecture (document ingestion, Qdrant-based retrieval scoring, and LLM response synthesis). Previously at Rally Health/UnitedHealthcare, diagnosed load-related memory spikes with JMeter and improved stability by migrating caching from Guava to Redis, and also supported adoption through UI A/B testing in a technical marketing engineer rotation.”
Intern Software Engineer specializing in AI agents, RAG pipelines, and semiconductor systems
“Built a web-based interface that connects an internal bug system to an LLM for initial debugging and issue classification, aiming to boost QA and software engineer efficiency while balancing latency and accuracy. Worked as a one-person project and managed constraints like limited hardware and difficulty extracting team debugging context, relying on manager communication and rapid modeling to validate direction.”
Junior Software Engineer specializing in backend systems and AI/ML pipelines
“Robotics-focused engineer with ROS 2 experience who has built and debugged real-time, distributed control/orchestration systems under production-like latency and safety constraints. Led platform changes at Persona for a real-time verification orchestration system using deterministic state machines and async workers, and has hands-on experience stabilizing multi-robot navigation/SLAM behavior using rosbag, RViz, and stress testing in simulation (Gazebo).”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and scalable inference
“Backend/retrieval-focused engineer with production experience at Perplexity building a large-scale real-time Q&A system using retrieval-augmented generation, emphasizing low-latency, high-quality answers through ranking, context optimization, and caching. Also has orchestration experience from both product-facing LLM pipelines and large-scale infrastructure workflows at Meta, and has partnered with non-technical stakeholders to align AI trade-offs with business goals.”
Staff-level Machine Learning Engineer specializing in LLMs and MLOps for Financial Services
“Machine learning/NLP practitioner at J.P. Morgan who led development of a production RAG system and an entity resolution pipeline for complex financial data. Deep hands-on experience with embeddings (Sentence-BERT), vector search (FAISS/pgvector), LLM fine-tuning (LoRA/PEFT), and rigorous evaluation (human-in-the-loop + A/B testing) backed by strong MLOps on AWS (Docker/Kubernetes, MLflow, Prometheus/Datadog).”
Mid-level AI/ML Engineer specializing in LLM alignment, safety, and scalable inference
“Built and productionized an AWS-hosted, Kubernetes-orchestrated RAG assistant that enables natural-language Q&A over internal document repositories with grounded answers and citations. Demonstrates strong applied LLM engineering: hallucination mitigation, hybrid retrieval + re-ranking, and rigorous evaluation via benchmarks and A/B testing, plus real-world scaling of compute-heavy inference with dynamic batching and monitoring.”
Executive CTO and Founder specializing in AI platforms and hyper-scale SaaS
“CTO-minded builder seeking to join a startup; previously created an AI-driven platform that abstracted away DevOps and infrastructure for drug discovery researchers. Emphasizes high-leverage, zero-to-one execution with managed cloud/open-source tooling, and a strong reliability/reproducibility mindset validated against existing scientific pipelines.”
Junior AI Engineer specializing in LLM systems, RAG, and full-stack automation
“Built and deployed an AI receptionist product for field-service businesses (HVAC/electrician), including real-time Jobber scheduling integrations and Twilio-based calling. Combines hands-on customer/operator shadowing with strong production engineering (queueing to handle API limits, rigorous testing/mocking, mirrored prod environment) and cross-layer troubleshooting, driving user adoption through review/override workflows.”