Pre-screened and vetted.
Junior Machine Learning Engineer specializing in LLM systems and GPU inference
“LLM/agent engineer who shipped a production RAG-based recommendation + explanation system that replaced a traditional recommender stack, delivering ~20% CTR lift (and +8% after a reliability iteration) with strong cold-start performance. Demonstrates strong production rigor: schema-constrained generation, typed tool calling, explicit state/orchestration, deep monitoring/feedback loops, and safe integration with messy ERP inventory/order data using normalization, idempotency, and conflict-resolution guardrails.”
Intern Software Engineer specializing in edge AI deployment and distributed systems
“Full-stack engineer who built an enterprise search platform (Codlens) delivering natural-language Q&A over Jira/Slack using embeddings, vector DB search, re-ranking (RRF), and LLM responses with source grounding. Also designed and benchmarked a distributed IAM system with Postgres transaction-log replication and Raft-based quorum consistency, reporting ~253 TPS at ~60ms latency in a multi-node setup. Experience spans early-stage startups (Zetic AI, Sagwara Capital) and large-scale orgs (Akamai, Atlassian).”
Mid-level Software Engineer specializing in embedded AI and full-stack systems
“Robotics software engineer who built and owned core navigation components for a TurtleBot in ROS/ROS2 and Gazebo, including an RRT-based planner, waypoint-to-velocity motion planning, and PID trajectory tracking. Demonstrates strong real-time debugging skills (control-loop timing under CPU load), costmap/occupancy-grid tuning, and distributed ROS2 communication design using DDS/QoS, plus Docker and CI/CD automation experience from Keysight.”
Engineering Leader specializing in Digital Health, AI, and Cloud Platforms
“Senior Engineering Manager at Roche leading two Scrum teams building internally shared (“inner-sourced”) tools and libraries for a healthcare enterprise. Has led security/compliance-first architecture decisions (e.g., Python AI modules running inside a Java container) and front-end modularization (Angular monorepo to module federation), with a strong focus on developer experience via automated Swagger/OpenAPI documentation and robust testing/versioning practices.”
Junior Full-Stack AI Engineer specializing in LLM apps and RAG systems
“Built and shipped a production LLM-powered “Vet agent” that automates pet symptom intake across multimodal inputs (images/files/text/speech) and provides analysis/home-care guidance, reaching thousands of daily active users within two months. Demonstrates strong agent engineering fundamentals: state-machine orchestration with structured JSON, tool/schema validation, high-availability routing/failover, and rigorous offline/online evaluation loops with trace-driven reliability improvements.”
Mid-level AI/ML Engineer specializing in NLP, Generative AI, and MLOps
“Internship experience shipping production AI systems: built an end-to-end RAG platform (Python/FastAPI + LangChain/LangGraph + vector search) to answer support questions from unstructured internal docs, with a strong focus on hallucination prevention through confidence gating and rigorous offline/online evaluation. Also delivered an AI-driven personalization/analytics feature using an unsupervised clustering pipeline, iterating with PMs to align statistically strong clusters with actionable business segmentation.”
Mid-level Software Developer specializing in backend microservices for healthcare and FinTech
“Built and deployed an AI-powered insurance claims fraud platform end-to-end using Java/Spring Boot, Kafka, OpenAI, pgvector, and AWS EKS. Stands out for combining LLM/RAG architecture with production-grade scalability and observability, delivering measurable impact including 62% less manual review, 40% better fraud precision, 37% higher throughput, and 99.95% uptime.”
Mid-level AI/ML Engineer specializing in NLP, MLOps, and scalable data pipelines
“Built and shipped a production LLM-powered personalized client engagement assistant in the financial domain, balancing real-time recommendations with strict privacy/compliance requirements. Demonstrates strong MLOps/LLMOps depth (Airflow + MLflow, containerized microservices, drift monitoring) and a privacy-by-design approach validated in collaboration with risk and compliance teams.”
Mid-level AI Engineer specializing in Generative AI, MLOps, and NLP for finance and healthcare
“Built and deployed a secure, production LLM-based document summarization and risk-highlighting tool for financial auditors, running inside a private Azure environment to protect confidential data. Focused on reliability (hallucination mitigation via retrieval-based prompts and source citations) and validated performance through comparisons to auditor summaries plus a user pilot, cutting review time by about half.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and RAG systems
“Machine learning/NLP engineer who built a production-oriented retrieval-based AI system at Morgan Stanley for healthcare use cases, combining RAG over unstructured patient records with deep-learning medical image segmentation (U-Net/Mask R-CNN). Strong in end-to-end pipelines and MLOps (Spark/MongoDB, AWS SageMaker, CI/CD, monitoring, automated retraining) and in entity resolution/data quality validation for noisy clinical data.”
Mid-level Machine Learning Engineer specializing in LLMs and NLP classification systems
“Internship experience building a production RAG+LLM pipeline to map messy card transaction descriptions to merchant brands, including a custom modified-ROUGE evaluation approach for weak/variant ground truth. Improved scalability and cost by moving from a managed LLM endpoint (e.g., Bedrock) to self-hosted vLLM, and orchestrated massive embedding backfills (5,000+ files, 10B+ rows) using an Airflow-triggered SQS + ECS worker architecture with robust retry/DLQ handling.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
Mid-level Software Engineer specializing in backend, AI, and full-stack systems
“Built and shipped production LLM agents including an internal RAG-based compliance classification system at SAIL (FastAPI/Redis/Docker) designed to handle real failure modes and scale to ~10k LLM calls/hour, achieving ~93% pipeline accuracy with reduced hallucination risk via multi-model orchestration and strict grounding. Also architected “Elara,” a state-machine-driven conversational appointment booking agent using structured JSON outputs and backend function execution for reliability, and has experience normalizing messy OTA/PMS data at RateGain.”
Mid-level Machine Learning Engineer specializing in AI/LLM systems
“ML/LLM systems engineer who has owned AI support automation products end-to-end, including ServiceNow-integrated incident routing, RAG-based resolution suggestion systems, and production stabilization. Stands out for combining hands-on platform work across PySpark, AWS Glue, FastAPI, Kubernetes, and Pinecone with measurable operational impact, including 30-35% MTTR reduction and 25-30% improvement in first-touch resolution.”
Mid-level AI/ML Engineer specializing in FinTech risk and fraud systems
“Senior AI/ML engineer focused on production LLM systems, combining RAG, fine-tuning, distributed training, and AI safety to ship scalable real-time moderation and conversational AI platforms. Stands out for pairing deep AWS/Kubernetes MLOps expertise with measurable impact: 40% lower latency/cost, 30-50% fewer hallucinations, and major reliability gains through observability and automation.”
Mid-level AI/ML Engineer specializing in Generative AI and financial services
“ML/AI engineer with hands-on experience shipping regulated financial AI systems at JPMC and Capgemini, spanning credit risk, fraud detection, and generative AI assistants. Stands out for combining modern LLM/RAG architectures with strong MLOps, real-time infrastructure, and explainability/compliance practices, while delivering measurable business impact in latency, accuracy, cost, and risk reduction.”
Entry-Level Software Engineer specializing in ML and backend systems
“Built and deployed a production LLM-based real-time stance detection system for social media, fine-tuning LLaMA 3.1 on A100s with DeepSpeed ZeRO/FSDP and iteratively refining data to handle sarcasm and context-dependent meaning. Also has Kubernetes operations experience (Kafka/Logstash/Elasticsearch observability pipeline) and delivered an OCR automation project during a Worley India internship that saved 20+ hours/week for on-site energy safety stakeholders.”
Mid-level AI/ML Engineer specializing in MLOps, computer vision, and NLP
“GenAI/ML engineer from Lucid Motors who built and productionized an LLM-powered RAG diagnostic assistant for manufacturing and maintenance teams, deployed on AWS with Docker/Kubernetes and MLflow. Demonstrates end-to-end ownership from retrieval/prompt design to scalability, monitoring, and workflow integration via APIs, plus production ML pipeline orchestration with Kubeflow (Spark/Kafka + TensorFlow) for predictive maintenance use cases.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Mid-level Software Engineer specializing in AI and full-stack healthcare platforms
“Built and deployed a RAG-based clinical knowledge assistant at GE Healthcare to help clinicians query large volumes of messy, unstructured clinical documents with grounded, cited answers. Hands-on across the full stack (OCR/ETL, de-identification for PHI, Azure OpenAI embeddings, Cosmos DB indexing, FastAPI/Django) with production monitoring via LangSmith and performance tuning through batching and index optimization.”
Junior AI/Backend Software Engineer specializing in ML and scalable systems
“Backend engineer with strong AWS/CI/CD experience (multi-repo deployments, Lambda + core app, immutable ECR and image promotion) and a published master’s thesis building an ML framework for Solar PV energy prediction and CO2 reduction impact modeling using ensemble and meta-learning approaches benchmarked against SAM.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Red Hat ML/LLM engineer who designed and deployed a production LLM-powered customer support automation system using RAG, improving latency by 30% via PEFT and vector search optimization. Built security and governance into retrieval (access-level filtering, encrypted Pinecone/ChromaDB) and delivered SHAP-based explainability via a dashboard for non-technical stakeholders. Experienced orchestrating distributed ML/RAG pipelines across AWS SageMaker and OpenShift with Airflow/Prefect, plus multi-agent workflows using CrewAI and LangGraph.”