Pre-screened and vetted in the Bay Area.
Mid-level AI/ML Engineer specializing in fraud detection and risk analytics in Financial Services
“Finance-domain ML/LLM engineer who has shipped production systems including a RAG-based financial insights assistant with a custom post-generation validation layer that verifies atomic claims against retrieved source text to prevent hallucinations in compliance-critical workflows. Also built large-scale MLOps automation on AWS using Kubeflow + MLflow + CI/CD for fraud detection and credit risk models processing 500M+ transactions/day with a 99.99% uptime goal, and partnered closely with JP Morgan risk/compliance stakeholders on NLP-driven compliance monitoring.”
Mid-level AI/ML & MLOps Engineer specializing in cloud AI infrastructure and GenAI
“At HPE, led and deployed an enterprise-grade LLM document intelligence platform for an insurance client, automating extraction from highly variable PDFs/scans/emails and raising field accuracy from 74% to 93%. Built a LangChain/Pinecone/OpenSearch RAG framework to cut hallucinations by 37% and operationalized LangSmith evals in CI, driving a 41% triage accuracy lift and >33% fewer incorrect resolutions while partnering closely with claims operations via HITL workflows.”
Junior ML Engineer specializing in energy forecasting and battery optimization
“Backend/ML engineer working on a battery energy storage system operations dashboard: built a Flask backend integrated with OAuth and a separate FastAPI optimization/simulation service, deployed via Docker CI/CD to Azure Container Apps. Strong in productionizing ML (AzureML to batch endpoints) and in performance/scalability patterns (Postgres indexing/JSONB, per-unit data isolation, async throttling + caching for year-long CPU-intensive simulations across 40+ scenarios).”
Mid-level AI/ML Engineer specializing in MLOps and LLM-powered applications
“AI/ML engineer with production experience building a RAG-based internal analytics assistant (Databricks + ADF ingestion, Pinecone vector store, LangChain orchestration) deployed via Docker on AWS SageMaker with CI/CD and MLflow. Strong focus on real-world constraints—latency/cost optimization (LoRA ~60% compute reduction), hallucination control with citation grounding, and enterprise security/governance. Previously at Intuit, delivered an interpretable churn prediction system (PySpark/Databricks, Airflow/Azure ML) that improved retention targeting ~12%.”
Senior Machine Learning Engineer specializing in MLOps and production AI systems
Junior ML Infrastructure Engineer specializing in low-latency LLM and inference serving
Mid-level Machine Learning Engineer specializing in robotics and autonomous driving
Mid-level Machine Learning Engineer specializing in GenAI, RAG, and computer vision
Mid-level Software/Data Engineer specializing in AI-driven data platforms and cloud ETL
Mid-Level Software/ML Engineer specializing in cloud-native ML systems
Senior AI Solutions Engineer specializing in LLM-powered enterprise systems
Junior Software Engineer specializing in ML systems and cloud deployment
Entry-level Machine Learning Engineer specializing in LLM systems and RAG
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and MLOps
Mid-level AI/ML Engineer specializing in GenAI, MLOps, and anomaly detection
Junior AI Engineer specializing in RAG pipelines and agentic AI systems
“Built and shipped production RAG/agentic systems in high-stakes domains (biomedical and legal), including an enterprise biomedical document retrieval platform over ~10k scientific docs and a multilingual African-law assistant at the World Bank. Deep hands-on experience with LangChain/LangGraph/LlamaIndex and evaluation tooling (LLM-as-a-judge, safety/hallucination detection), with measurable gains in retrieval quality and hallucination reduction.”
Mid-level Software Engineer specializing in ML, LLM apps, and cloud data systems
“Built a production SQL chatbot for access-log analytics that replaced manual custom report requests with natural-language querying, using LangGraph and a ChromaDB-backed RAG pipeline for grounded, consistent answers. Implemented a privacy-preserving design where the LLM never sees raw customer data (only query metadata) and has experience building multi-agent/tool-calling systems with LangGraph (DeepAgents), including solving sub-agent communication drift via self-reflection.”
Mid-level Machine Learning Engineer specializing in deep learning and generative AI
“AI/ML engineer who has deployed transformer-based NLP systems to production via Python REST APIs and Kubernetes on AWS/Azure, with a strong focus on latency optimization (p95), reliability, and scalable orchestration. Demonstrates pragmatic model tradeoff decision-making and strong stakeholder collaboration—improving adoption by making outputs more actionable with summaries, extracted fields, and confidence indicators.”
Mid-level Machine Learning Engineer specializing in LLMs, agentic AI, and risk/fraud modeling
“Built and productionized an agentic LLM workflow during a summer internship to transform unstructured clinical reports into analytics-ready structured data, using a LangChain multi-agent design plus an LLM-as-a-judge layer to control quality in a regulated setting. Also has experience orchestrating ML pipelines at Piramal Capital using AWS Step Functions/EventBridge/CloudWatch, with strong emphasis on observability, evaluation rigor, and measurable impact (80–90% reduction in manual data entry).”
Junior Robotics Data Engineer specializing in multi-sensor perception datasets
“Robotics software engineer focused on perception data pipelines and multi-robot coordination. Built ROS 2 (rclpy) nodes for synchronized RGB/ToF/pose processing and scaled a perception training data generation pipeline from single-object to multi-object while preserving backward compatibility. Also has strong DevOps experience deploying containerized APIs on Kubernetes with Kustomize and automated releases via GitHub Actions.”
Intern Data Scientist specializing in ML engineering and LLM agentic workflows
“Built an agentic, multi-step LLM system that generates full-stack code for API integrations using LangChain orchestration, Pinecone/SentenceBERT RAG, and a human-in-the-loop feedback loop for iterative code refinement. Also collaborated with non-technical content writers and PMs during a Contentstack internship to deliver a Slack-based AI workflow that generates and brand-checks articles with one-click approvals.”
Intern Machine Learning Engineer specializing in forecasting, NLP, and RAG systems
“Intern who built and deployed a production LLM-powered contract analysis system for finance teams: Azure Document Intelligence for text/table extraction plus Gemini prompting to surface key terms and risks via an async API and simple UI. Emphasizes reliability in production with fallbacks, guardrails against hallucinations, and operational concerns like latency/cost/versioning, delivering summaries in under 30 seconds instead of hours.”