Pre-screened and vetted.
Intern Applied Scientist / ML Engineer specializing in NLP and conversational AI
“LLM/Conversational AI engineer who built a production multi-turn dialogue system using LoRA fine-tuning on LLaMA, cutting training compute/memory by 90%+ while maintaining low-latency inference via quantization and streaming generation. Experienced in orchestrating end-to-end ML workflows with Prefect/Airflow/Kubeflow (including hyperparameter sweeps and W&B tracking) and improving agent reliability through benchmark-driven testing, shadow-mode rollouts, and stakeholder-informed guardrails.”
Mid-level Machine Learning Engineer specializing in LLMs, generative AI, and MLOps
“Built and shipped a production LLM-powered medical scribe that generates structured clinical visit summaries using RAG, strict JSON schemas, and post-generation validation to reduce hallucinations. Experienced in making LLM workflows deterministic and observable (structured logging/metrics/tracing) and in evaluation-driven iteration with metrics like schema pass rate and edit rate; collaborated closely with clinicians and policy stakeholders at Scale AI to drive adoption.”
Entry-Level Software Engineer specializing in systems, networking, and ML
“Robotics software candidate with hands-on experience building controllers for an Autonomous Underwater Vehicle, including dual-PID control in Python with state-space modeling and a planned path to LQR. Developed ROS nodes for odometry-based localization, waypoint planning, and control command publishing, validated through a custom Gazebo/ROS simulation workflow with control-metric-driven testing. Also worked on F1Tenth simulation and scan-matching localization (PL-ICP), with additional cloud deployment experience using Docker/Kubernetes and CI/CD.”
Mid-level Machine Learning Engineer specializing in search ranking and NLP
Senior AI/ML Engineer specializing in LLMs, RAG, and multimodal recommendation systems
Senior Software Engineer specializing in cross-platform rendering and game UI systems
“Game UI engineer with shipped experience on Paladins (Unreal Engine 3/C++), including an in-game store with automated grid-based recommendation layout and an inventory UI built for PC/PlayStation/Xbox/Switch. Emphasizes production readiness through clean separation of UI/data, strong documentation for engineering and content workflows, and memory-conscious UI patterns (pooled entries/paged navigation) for constrained platforms like Switch.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and distributed MLOps
Mid-level AI/ML Engineer specializing in LLMs, RAG, and production MLOps
Staff Software Engineer specializing in commerce, checkout, and payments platforms
Director-level Product-Led CTO specializing in AI/MarTech, retail media, and omnichannel platforms
Senior Machine Learning Engineer specializing in LLM inference and GPU infrastructure
Mid-level AI/ML Engineer specializing in LLM training, RAG, and low-latency inference
Mid-level AI/ML Engineer specializing in LLMs, RAG, and scalable inference
“Backend/retrieval-focused engineer with production experience at Perplexity building a large-scale real-time Q&A system using retrieval-augmented generation, emphasizing low-latency, high-quality answers through ranking, context optimization, and caching. Also has orchestration experience from both product-facing LLM pipelines and large-scale infrastructure workflows at Meta, and has partnered with non-technical stakeholders to align AI trade-offs with business goals.”
Staff-level Machine Learning Engineer specializing in LLMs and MLOps for Financial Services
“Machine learning/NLP practitioner at J.P. Morgan who led development of a production RAG system and an entity resolution pipeline for complex financial data. Deep hands-on experience with embeddings (Sentence-BERT), vector search (FAISS/pgvector), LLM fine-tuning (LoRA/PEFT), and rigorous evaluation (human-in-the-loop + A/B testing) backed by strong MLOps on AWS (Docker/Kubernetes, MLflow, Prometheus/Datadog).”
Mid-level AI/ML Engineer specializing in LLM alignment, safety, and scalable inference
“Built and productionized an AWS-hosted, Kubernetes-orchestrated RAG assistant that enables natural-language Q&A over internal document repositories with grounded answers and citations. Demonstrates strong applied LLM engineering: hallucination mitigation, hybrid retrieval + re-ranking, and rigorous evaluation via benchmarks and A/B testing, plus real-world scaling of compute-heavy inference with dynamic batching and monitoring.”
“Backend/full-stack engineer (Amazon experience) who built an AWS-based integration testing platform using Flask, ECS, Docker, and CloudWatch—cutting 1000+ test cases from ~5 hours to ~30 minutes while improving log visibility for non-engineering users. Also led a zero-downtime EU region migration with rigorous ORR testing, and built a Kinesis/Firehose/S3 + Glue/Spark replay mechanism for resilient data recovery. Side project: reproducible, cost-efficient LLM hosting platform on EKS using CDK and Karpenter for scale-to-zero.”
Intern AI/ML Engineer specializing in NLP, LLMs, and semantic search
“Built and deployed a production RAG-based semantic search and summarization system for large legal/technical document sets, owning the full backend (embeddings, vector store, chunking, prompting) and driving a reported 40–60% reduction in manual review time. Experienced with LangChain/LlamaIndex plus Airflow/Temporal-style orchestration, and applies rigorous evaluation/monitoring (A/B tests, drift detection, staged rollouts) to keep agentic systems reliable. Also partnered with a supply-chain manager at TE Connectivity to deliver an AI inventory recommendation tool projected to drive millions in value.”
Senior Backend Engineer specializing in distributed microservices and event-driven systems
“Backend engineer with production experience building a high-scale notification pipeline (~20M/day) using Java/Dropwizard with Kafka and Azure Queue, including DLQ/poison-message handling and the outbox pattern for reliability. Also led a batch-based migration of Yammer Messaging user data from PostgreSQL to Azure Cosmos DB for global multi-region scale, addressing throttling and network failures via retries, escalation policies, and dynamic throughput tuning.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“ML/LLM engineer who built a production RAG system (GPT-4 + FAISS + FastAPI) to deliver fast, grounded answers from proprietary documents, optimizing for sub-200ms latency and high-concurrency scale. Strong MLOps/observability background: drift monitoring with Prometheus + Streamlit, automated retraining via Airflow, Kubernetes autoscaling, and MLflow-managed model lifecycle, plus inference cost reduction through quantization and structured pruning.”