Pre-screened and vetted.
Senior Software Engineer specializing in AI for Healthcare and Enterprise SaaS
Executive CTO specializing in AI, cloud platforms, and scaling SaaS products
“NYC-based startup founder/CTO who sold products to Omnicom and Sprinklr, then built an AI-powered cultural insights engine inside Omnicom using AWS Lambda + ML to process ~1M items/day and reached ~$1MM ARR in year one. Former senior leader at Sprinklr managing 200+ people globally, delivering enterprise martech solutions with SLAs and high-reliability social data pipelines (Twitter firehose).”
Junior Data Scientist specializing in LLM agents, RAG, and reinforcement learning
“McKinsey practitioner who built and deployed production LLM systems for consultants/clients, including a Power BI-integrated multi-agent chatbot (RAG + text-to-SQL + formatting) with custom Python orchestration, verification loops, and a 100+ case eval set achieving ~95% consistency. Also delivered a taxonomy-mapper agent that standardized inconsistent labeling for C-suite stakeholders, cutting a process from >2 weeks to <30 minutes through demos and business-focused communication.”
Mid-level AI/ML Engineer specializing in Generative AI, LLM alignment, and RAG
“Built and productionized a real-time enterprise RAG pipeline to improve factual accuracy and reduce LLM hallucinations by grounding responses in constantly changing internal knowledge bases (policies, manuals, FAQs). Experienced in orchestrating end-to-end ML workflows (Airflow/Kubernetes), handling messy multi-format data with schema enforcement (Pydantic/Hydra), and maintaining freshness via streaming incremental embeddings plus batch refresh. Also delivers applied ML solutions with non-technical teams (marketing/CRM) for segmentation and personalized engagement.”
Entry-level Supply Chain & Test Engineer specializing in warehouse automation and robotics
“P&G operator who is also building and selling an AI receptionist (voice agent) SaaS for healthcare/service clinics, using EHR + calendar API compatibility to target accounts and letting the Voice AI run parts of the demo to prove value. Has already closed and deployed to two clients in the last two months, with production impact via reduced front-desk overhead and automated scheduling/FAQs, and brings a structured, scalable deployment/process mindset from global WMS rollouts.”
Intern Software/AI Engineer specializing in LLM fine-tuning and agentic RAG systems
“Built and shipped an end-to-end LLM agent during an AT&T internship to automate network troubleshooting, with production-style reliability safeguards (timeouts/retries/fallbacks) and structured, state-machine orchestration; project won 3rd place in AT&T’s nationwide intern innovation challenge and was demoed to leadership. Also handled messy multi-partner data at Tencent by implementing schema validation/normalization, confidence-threshold fallbacks, and idempotent Python/ORM-based pipelines.”
Senior AI Research Engineer specializing in LLM agents and large-scale ML
“AT&T Labs builder who deployed a production multi-agent LLM system that lets engineers ask natural-language questions and automatically generates deterministic, schema-grounded Snowflake SQL (200–400 lines) to detect anomalies in massive wireless/network event data (~11B events/day). Experienced with LangChain and Palantir Foundry orchestration, RAG-based result interpretation, and rigorous evaluation/monitoring loops to continuously improve reliability.”
Junior Software Engineer specializing in distributed systems and machine learning
“Google backend engineer with strong experience in large-scale identity, membership, and access-control systems. Notable work includes reconciling customer IDs across 2B+ roster records and leading a 0-to-1 Drive sharing feature to classify external users as crossover members, with a strong emphasis on correctness, rollout safety, and low-latency service design.”
Mid-level Data Engineer specializing in AI/ML platforms and cloud data pipelines
“Built and shipped an LLM-powered data quality assistant that generates maintainable validation checks from metadata while executing validations via Great Expectations, exposed through FastAPI and integrated into Airflow-managed pipelines. Emphasizes production reliability (structured outputs, guardrails, monitoring, versioning, human review) and works closely with compliance/operations teams to deliver clear, auditable, user-friendly AI outputs.”
Intern/Junior Software Engineer specializing in AI/ML and cloud-based systems
“Embedded/robotics software engineer with Hyundai Motors experience who owned an AI-driven perception validation pipeline using a Transformer-based approach to generate stable synthetic in-cabin audio for autonomy/ASR testing, cutting downstream testing time by 50%+. Has hands-on ROS integration (IMU sensor streaming, inference, control nodes), MQTT-based distributed messaging, and cloud/container deployment experience (Docker, Node/Express, AWS, CI/CD).”
Intern Machine Learning & AI Engineer specializing in computer vision and ML systems
“Robotics/ML engineer with internship experience at Valeo building a deep-learning prototype to replace parts of a legacy SLAM backend for autonomous parking, focused on making models run reliably in real time on embedded hardware (quantization/distillation + TensorRT). Also brings strong MLOps/deployment experience (Docker, Kubernetes on AWS EKS, CI via GitHub Actions) and has supported patent filing by explaining the technical approach to legal stakeholders.”
Mid-Level Software Engineer specializing in data pipelines, observability, and analytics
“Meta engineer who improved a critical revenue estimation dataset pipeline that was arriving ~6 days late—diagnosed via raw logs/lineage, redesigned legacy scans to only process the needed window, and shipped validation plus freshness/lag dashboards. Delivered ~50% latency reduction (to ~3 days) and regained adoption by running old/new pipelines in parallel with gated cutover and evidence-based customer communication. Applies incident-response rigor to real-time LLM/agentic workflow debugging and regularly runs developer demos/workshops.”
Intern Applied Scientist / ML Engineer specializing in NLP and conversational AI
“LLM/Conversational AI engineer who built a production multi-turn dialogue system using LoRA fine-tuning on LLaMA, cutting training compute/memory by 90%+ while maintaining low-latency inference via quantization and streaming generation. Experienced in orchestrating end-to-end ML workflows with Prefect/Airflow/Kubeflow (including hyperparameter sweeps and W&B tracking) and improving agent reliability through benchmark-driven testing, shadow-mode rollouts, and stakeholder-informed guardrails.”
Intern Machine Learning Engineer specializing in RAG systems and AWS cloud infrastructure
“Internship at BlueFoxLabs building and deploying an AI/ML RAG system for a biopharma client on top of LibreChat, including an AWS Textract ingestion pipeline and PGVector retrieval deployed to AWS EKS. Demonstrated production-minded scalability work by moving from a vertically scaled EC2 setup to a horizontally scaling Kubernetes/EKS deployment, using CI/CD to safely incorporate requirement changes like tabular document data.”
Senior Machine Learning Engineer specializing in production ML and predictive analytics
“ML/AI engineering leader who has owned end-to-end production systems from experimentation through deployment, monitoring, and iteration at meaningful scale. They describe running a 1M+ records/day prediction platform with 99.9% availability, shipping a RAG-based conversational AI feature for 50,000 active users, and consistently improving precision, latency, reliability, and cost with measurable business impact.”
Staff Full-Stack Engineer specializing in Healthcare AI and FinTech payments
“Backend/data engineer from Oscar Health specializing in healthcare claims systems on AWS. Built HIPAA-compliant real-time services (FastAPI/Postgres/Kafka on EKS) and serverless ingestion pipelines, and led modernization of a legacy SAS claims pricing system to Python/Spark with rigorous parity validation. Demonstrated measurable impact with high uptime/low latency services and major Snowflake performance and cost reductions.”
Mid-level Machine Learning Engineer specializing in LLMs, generative AI, and MLOps
“Built and shipped a production LLM-powered medical scribe that generates structured clinical visit summaries using RAG, strict JSON schemas, and post-generation validation to reduce hallucinations. Experienced in making LLM workflows deterministic and observable (structured logging/metrics/tracing) and in evaluation-driven iteration with metrics like schema pass rate and edit rate; collaborated closely with clinicians and policy stakeholders at Scale AI to drive adoption.”
Mid-level DevOps Engineer specializing in cloud-native infrastructure on AWS and Azure
“DevOps/SRE focused on cloud-based distributed systems, with strong hands-on Kubernetes production experience (microservices deployments, Helm, probes, resource tuning, CI/CD and Docker build standardization). Demonstrated end-to-end troubleshooting across application, infrastructure, and networking layers—e.g., isolating degraded storage via node disk I/O metrics and restoring performance by draining the node and replacing the volume. Builds Python automation for operational reliability, including scheduled Kubernetes secrets rotation integrated with an external secret manager.”
Mid-level Backend & ML Engineer specializing in LLM systems and scalable AI pipelines
“Built and shipped a real-time AI phone agent for small businesses that handles bookings/FAQs/messages using streaming ASR, an LLM with tool-calling, and TTS; deployed to production for multiple paying customers. Demonstrates strong applied LLM reliability practices (tool-first grounding, retrieval, hard-negative testing, and production monitoring) and experience orchestrating multi-step AI workflows with Airflow, Prefect, and AWS Step Functions.”
Junior Robotics & Embedded Software Engineer specializing in autonomous systems and RF software
“Robotics/embedded engineer with hands-on experience building real-time control systems on RP2040 (hydroponics automation, 1-DOF helicopter stabilization) and full ROS 2 navigation stacks in simulation (URDF, TF, PID, A* in RViz/Gazebo). Demonstrates strong low-level protocol work (timing-sensitive one-wire in C) and rigorous debugging across hardware and software using UART instrumentation and oscilloscope verification, plus reproducible workflows with Docker and CI/CD (GitHub Actions/GitLab, incl. Sandia National Labs).”
Director-level Data Architecture & Governance leader specializing in cloud analytics platforms
“Technology/architecture leader with Accenture experience delivering data- and AI/ML-driven products, including a legal contract search solution and customer sales analytics for AWS. Known for scaling distributed teams (onshore/offshore), making pragmatic architecture decisions, and solving hard data problems (proprietary sources, data quality) while implementing scalable integrations like Redshift-to-Salesforce via parallelized pipelines.”
Mid-level Robotics Software Engineer specializing in teleoperation, simulation, and autonomy
“Robotics engineer who helped bootstrap Meta’s humanoid robotics effort, building simulation training and deployment infrastructure for vision-language-action (VLA) models. Evaluated multiple physics backends (Bullet, MuJoCo, Isaac, internal) to minimize sim-to-real gap and addressed control-loop frequency mismatches via sequence optimization/MPC-like approaches and trajectory-output modifications. Published research that contributed a new addition to ROS 2 and has built ROS2 node stacks spanning control, perception, teleop, tactile sensing, and imaging.”
Mid-level Python Backend Developer specializing in cloud-native microservices and AI/ML platforms
“Backend/AI engineer who built a production GPU-backed real-time inference API at Nvidia and debugged burst-induced tail latency, cutting P95 by ~29% through dynamic batching and backpressure. Also shipped an end-to-end RAG + agentic operational diagnostics assistant with strict tool controls, evidence citation, confidence gating, and strong production guardrails, plus demonstrated hands-on Postgres optimization (900ms to 40–60ms).”