Pre-screened and vetted.
Mid-level Data Scientist / GenAI & ML Engineer specializing in LLM apps and MLOps
Senior Full-Stack Python Engineer specializing in cloud microservices and AI/LLM systems
Mid-level AI/ML Engineer specializing in LLMs, RAG, and scalable MLOps
Mid-level Software Engineer specializing in backend APIs, data pipelines, and cloud microservices
Mid-level Applied AI Engineer specializing in LLMs, MLOps, and real-time AI systems
Mid-level AI/ML Engineer specializing in LLMs, multilingual NLP, and low-latency MLOps
Senior AI/ML Engineer specializing in LLM agents, RAG, and production ML systems
Mid-level Machine Learning Engineer specializing in Generative AI and LLM applications
Mid-level AI/ML Engineer specializing in Generative AI, LLM alignment, and RAG
“Built and productionized a real-time enterprise RAG pipeline to improve factual accuracy and reduce LLM hallucinations by grounding responses in constantly changing internal knowledge bases (policies, manuals, FAQs). Experienced in orchestrating end-to-end ML workflows (Airflow/Kubernetes), handling messy multi-format data with schema enforcement (Pydantic/Hydra), and maintaining freshness via streaming incremental embeddings plus batch refresh. Also delivers applied ML solutions with non-technical teams (marketing/CRM) for segmentation and personalized engagement.”
Intern/Junior Software Engineer specializing in AI/ML and cloud-based systems
“Embedded/robotics software engineer with Hyundai Motors experience who owned an AI-driven perception validation pipeline using a Transformer-based approach to generate stable synthetic in-cabin audio for autonomy/ASR testing, cutting downstream testing time by 50%+. Has hands-on ROS integration (IMU sensor streaming, inference, control nodes), MQTT-based distributed messaging, and cloud/container deployment experience (Docker, Node/Express, AWS, CI/CD).”
Intern Applied Scientist / ML Engineer specializing in NLP and conversational AI
“LLM/Conversational AI engineer who built a production multi-turn dialogue system using LoRA fine-tuning on LLaMA, cutting training compute/memory by 90%+ while maintaining low-latency inference via quantization and streaming generation. Experienced in orchestrating end-to-end ML workflows with Prefect/Airflow/Kubeflow (including hyperparameter sweeps and W&B tracking) and improving agent reliability through benchmark-driven testing, shadow-mode rollouts, and stakeholder-informed guardrails.”
Staff Full-Stack Engineer specializing in Healthcare AI and FinTech payments
“Backend/data engineer from Oscar Health specializing in healthcare claims systems on AWS. Built HIPAA-compliant real-time services (FastAPI/Postgres/Kafka on EKS) and serverless ingestion pipelines, and led modernization of a legacy SAS claims pricing system to Python/Spark with rigorous parity validation. Demonstrated measurable impact with high uptime/low latency services and major Snowflake performance and cost reductions.”
Mid-level Machine Learning Engineer specializing in LLMs, generative AI, and MLOps
“Built and shipped a production LLM-powered medical scribe that generates structured clinical visit summaries using RAG, strict JSON schemas, and post-generation validation to reduce hallucinations. Experienced in making LLM workflows deterministic and observable (structured logging/metrics/tracing) and in evaluation-driven iteration with metrics like schema pass rate and edit rate; collaborated closely with clinicians and policy stakeholders at Scale AI to drive adoption.”
Mid-level Backend & ML Engineer specializing in LLM systems and scalable AI pipelines
“Built and shipped a real-time AI phone agent for small businesses that handles bookings/FAQs/messages using streaming ASR, an LLM with tool-calling, and TTS; deployed to production for multiple paying customers. Demonstrates strong applied LLM reliability practices (tool-first grounding, retrieval, hard-negative testing, and production monitoring) and experience orchestrating multi-step AI workflows with Airflow, Prefect, and AWS Step Functions.”
Mid-level Python Backend Developer specializing in cloud-native microservices and AI/ML platforms
“Backend/AI engineer who built a production GPU-backed real-time inference API at Nvidia and debugged burst-induced tail latency, cutting P95 by ~29% through dynamic batching and backpressure. Also shipped an end-to-end RAG + agentic operational diagnostics assistant with strict tool controls, evidence citation, confidence gating, and strong production guardrails, plus demonstrated hands-on Postgres optimization (900ms to 40–60ms).”
Intern Software Engineer specializing in AI agents, RAG, and full-stack web development
Director-level AI/ML Technology Leader specializing in healthcare and life sciences
Junior Software Engineer specializing in cloud infrastructure and automation testing
Mid-level Data Scientist specializing in NLP, MLOps, and semiconductor manufacturing analytics