Pre-screened and vetted.
Intern AI/Data Scientist specializing in LLMs, RAG, and MLOps
“Internship project at Builder Market: built an end-to-end production multimodal LLM application that estimates renovation/replacement costs from appliance photos (CLIP embeddings) or text descriptions, combining fine-tuning with agentic RAG. Focused heavily on real-world performance constraints—latency and cost—using parallel agent workflows, model routing to smaller/open-source models, re-ranking, and retrieval chunking, and collaborated closely with CEO/co-founders to deliver the solution.”
Mid-level AI Engineer specializing in Generative AI, RAG systems, and fraud analytics
“Built and deployed a RAG-based student/faculty support chatbot at a university that answers from official syllabus/policy documents and now supports 4,000+ students while reducing repetitive support requests. Hands-on with LangChain, LangGraph, and CrewAI to orchestrate reliable agentic workflows, with a strong focus on testing/monitoring in production and cross-functional delivery (e.g., marketing analytics automation at Steve Madden).”
Mid-level Machine Learning Engineer specializing in NLP, Generative AI, and RAG systems
“Built and deployed a production LLM-powered phone assistant for a healthcare clinic, combining streaming STT/TTS with RAG over approved clinic documents and strict safety guardrails to prevent unverified medical advice, plus seamless human handoff. Also has hands-on Apache Airflow experience building robust daily ML/data pipelines with data validation, retries/timeouts, monitoring, and metric-gated model deployment, and iterates closely with clinic staff using real call reviews.”
Mid-level Full-Stack Developer specializing in Python/Java and cloud-native web apps
“Robotics-focused full-stack engineer with hands-on ROS experience building sensor-processing and control nodes, plus a track record of debugging and optimizing real-time robot responsiveness via profiling and message-timing analysis. Uses Webots for pre-hardware validation and Docker/CI/CD to standardize deployments and catch issues early.”
“ML/LLM engineer with production experience building a RAG-based LLM support assistant (FastAPI, Redis, Kafka) with multi-layer validation and human-in-the-loop feedback loops to improve accuracy over time. Has orchestration and MLOps depth using Airflow and Kubeflow on Kubernetes (autoscaling, alerting, monitoring) and delivered measurable ops impact (40% ticket efficiency improvement) by partnering closely with customer support teams.”
“ML engineer/data scientist who deployed a production credit risk + insurance claims triage platform at Hartford Financial, combining XGBoost default prediction with BERT-based document classification. Demonstrated strong MLOps by cutting inference latency to sub-500ms and building drift monitoring plus automated retraining/deployment pipelines (MLflow, CloudWatch, GitHub Actions, SageMaker) with human-in-the-loop review and SHAP-based explainability for underwriting adoption.”
Mid-Level AI/Full-Stack Engineer specializing in agentic LLM systems and RAG
“Built and deployed Clyra.AI, an AI-driven daily scheduling product that uses a LangGraph-based multi-agent LLM pipeline (task extraction, verification, reflection) grounded with strict RAG over emails/documents/calendars and real-world signals like health metrics. Designed a custom agent orchestrator with bounded loops/termination conditions and a self-auditing verification/reflection layer to reduce hallucinations while controlling latency and cost via caching and model distillation.”
Mid-level Machine Learning Engineer specializing in LLMs, GenAI, and Computer Vision
“LLM/agent engineer who built a production multi-agent research automation system using LangGraph (planner, retriever with FAISS, supervisor, evaluator) with structured outputs and citation tracking for traceable reports. Emphasizes reliability and operations—LangSmith-based observability, multi-level testing, hallucination mitigation, and latency/cost controls—plus prior experience as a Computer Vision Software Engineer at Deepsight AI Labs working directly with non-technical customers.”
Intern AI Researcher specializing in NLP, LLMs, and knowledge graphs
“Built and shipped “LabMate,” a production AI assistant specialized in laboratory hardware, using a weighted multi-source RAG pipeline with reranking and reasoning-focused query decomposition to handle complex user questions. Deployed on a local GPU cluster with vLLM and NVIDIA MPS (plus OCR/VLM components), and established evaluation using synthetic + public reasoning datasets while collaborating weekly with non-technical admins to align requirements and resource constraints.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM/RAG system at CVS to automate clinical documents, addressing PHI compliance, retrieval accuracy, and latency; achieved a 35–40% reduction in review effort through chunking and FP16/INT8 optimization. Also has experience translating AI outputs into actionable insights for non-technical stakeholders (sports analysts).”
Intern Software Engineer specializing in full-stack and LLM/RAG systems
“Full-stack engineer who built "Workstream AI," an AI-powered engineering visibility product that converts GitHub activity into real-time insights using an event-driven microservices stack (RabbitMQ/Postgres/Express) and GPT-4 with a React frontend. Previously a Founding SWE at a health & wellness startup, building data-driven user management tooling, and also delivered a real-time shuttle tracking/ride request system using Java Spring Boot/Hibernate + React; comfortable owning production deployment details (AWS EC2, DNS, SSL).”
Mid-level AI/ML Engineer specializing in MLOps, NLP, and scalable model deployment
“Built and deployed a production autonomous AI data analyst agent (LangChain + GPT + Streamlit on AWS) that turns natural-language questions into validated SQL, visualizations, and insights, cutting manual analysis time by ~50%. Emphasizes reliability and MLOps: schema-aware validation/guardrails to prevent hallucinations, scalable large-data processing, and Azure DevOps CI/CD + MLflow for automated deployment and experiment tracking.”
Mid-level AI/ML Engineer specializing in healthcare ML and generative AI
“AI/LLM engineer at Humana who built and deployed a HIPAA-aware RAG system for clinical record retrieval, cutting search time dramatically and improving retrieval efficiency by 30%. Experienced with Spark-scale data preprocessing, QLoRA fine-tuning, LangChain orchestration, and MLflow+SageMaker integration, with a strong testing/evaluation discipline (A/B tests, human eval) to hit 95%+ accuracy and production latency targets.”
Intern Data Analyst specializing in data pipelines and LLM/RAG applications
“Built and deployed LLM-powered analytics and reporting systems, including a RAG-based assistant over Snowflake that let business users ask questions in plain English instead of writing SQL. Experienced orchestrating LLM agents (LangChain) and serverless reporting pipelines (AWS Lambda/S3/RDS), with a strong focus on grounded outputs, monitoring/evaluation, and data quality—used daily by non-technical finance and operations teams at Cigna.”
Mid-level Full-Stack Software Engineer specializing in cloud-native microservices and FinTech
“At Delta Airlines, built and shipped a production LLM-powered semantic search/troubleshooting assistant over maintenance logs and operational documentation using OpenAI embeddings and a vector database. Implemented hybrid ranking, query enrichment, and structured filters to improve relevance ~35% while optimizing latency via caching and vector tuning. Also designed a scalable Kafka + AWS (Lambda/SQS) ingestion pipeline with strong reliability/observability and an eval loop using real engineer queries and human review.”
Junior AI Engineer specializing in ML, LLM systems, and RAG
“Built and deployed an LLM/applied-ML system enabling efficient extraction of useful information from large unstructured multimodal datasets, owning the full pipeline from ingestion to inference and APIs with a strong emphasis on production reliability, latency, and monitoring. Also delivered a voice-based AI workflow for Hindi policy document access for the Election Commission of India by translating non-technical usability needs into iterative demos and a successful implementation.”
Mid-level AI/ML & Full-Stack Engineer specializing in LLM agents and generative AI
“LLM/agent builder who shipped a live consumer AI-agent app (kalpa.chat) that visualizes complex reasoning as interactive graphs and abstracts multi-provider model usage via a unified wallet. Professionally has applied LangChain/LangGraph to IVR parsing and to scaling a football video-generation pipeline at DAZN, including shipping a VAR-specific retrieval/order fix via SQL after iterating with a non-technical PM.”
Junior Machine Learning Engineer specializing in NLP, computer vision, and MLOps
“ML/LLM engineer with Meta experience building production AI systems for near real-time user-report classification and summarization under strict latency (<250ms), safety, cost, and privacy constraints. Has hands-on MLOps/orchestration experience (Airflow, Spark, MLflow, Kubernetes, Docker, GitHub Actions) plus observability (Prometheus/Grafana) and applies rigorous evaluation, staged rollouts, and A/B testing to keep agent workflows reliable in production.”
Mid-level AI Engineer specializing in multi-agent systems and RAG
“Built and shipped a production LangGraph-based multi-agent LLM analytics/decision copilot that answers questions across SQL/BI systems and unstructured docs, emphasizing grounded, tool-verified outputs with citations and confidence gating. Deep hands-on experience with orchestration (LangGraph, CrewAI, OpenAI Assistants, MCP) plus real-world latency/cost optimization (vLLM batching/KV caching, speculative decoding, quantization) and rigorous eval/observability. Partnered closely with business/ops stakeholders to deliver explainable reporting automation, cutting manual reporting time by 50%+.”
Mid-level AI/ML Engineer specializing in predictive modeling and cloud ML pipelines
“LLM engineer/data engineer who has deployed production RAG systems for internal-document Q&A, building end-to-end ingestion, embedding, vector search, and FastAPI serving while actively reducing hallucinations and latency through rigorous retrieval tuning and caching. Also experienced in orchestrating cloud data pipelines (Airflow, AWS Glue, Azure Data Factory) and partnering with non-technical business teams to deliver AI solutions like automated document review.”
Mid-Level Software Engineer specializing in cloud-native microservices on AWS and Kubernetes
“Backend engineer who built a stateless Python/Flask service supporting a healthcare-document ETL pipeline, offloading heavy processing to Celery workers and adding strong observability (metrics, structured logs, audits). Demonstrates practical performance/reliability work: batch chunking, priority queues, autoscaling by queue depth/CPU, DLQ routing, and PostgreSQL tuning (indexes, pagination) to cut slow API responses. Also has experience deploying real-time ML classification via TensorFlow Serving behind a FastAPI wrapper and integrating models via REST/gRPC.”
Junior Full-Stack Engineer specializing in backend systems and agentic AI
“Founding/early engineer experience across Asante and a Series A startup (Adgency), shifting from data science/ML into owning production full-stack systems end-to-end. Built core product flows (registration, business profiles, map service), AWS-deployed gRPC microservices with CI/CD, and operated low-latency agent/video ad generation workflows with retries/fallbacks and PostHog-based observability.”
Mid-level ML & Data Engineer specializing in GenAI, graph modeling, and fraud/risk analytics
“Built a production AI fraud/risk scoring platform at BlueArc that ingests web business/product/site data, generates text+image embeddings, and connects entities in a graph to detect reuse patterns and links to known bad actors. Optimized for scale with incremental graph re-scoring and delivered investigator-friendly explainability by surfacing the exact signals/relationships behind each score; orchestrated workflows with Airflow and GCP event-driven components (Pub/Sub, Dataflow, Cloud Run) and has recent LLM workflow orchestration experience (retrieval, prompting, scoring).”
Mid-level AI/ML Engineer specializing in financial risk, fraud analytics, and forecasting
“Built and productionized an LLM-powered financial intelligence and forecasting platform at Northern Trust using a RAG architecture (LangChain + Hugging Face + FAISS) with end-to-end MLOps (Docker/Kubernetes, Airflow, MLflow). Emphasized regulatory-grade explainability (SHAP/Power BI) and hallucination control (retrieval-only grounding), achieving ~30% forecasting accuracy improvement and ~65% reduction in analyst research time, with sub-second inference and 95% uptime on EKS/AKS.”