Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in GenAI agents and production ML systems
Senior Staff Full-Stack Engineer specializing in AI copilots and cloud platforms
Mid-level Data Scientist/ML Engineer specializing in LLMs, NLP, and recommender systems
Senior Full-Stack Python Developer specializing in cloud, data platforms, and GenAI
Mid-level Agentic AI & ML Engineer specializing in LLM agents and RAG systems
Mid-level Data Scientist specializing in GenAI, LLMs, and MLOps
Mid-level AI/ML Engineer specializing in cloud MLOps and GenAI for fraud detection
Senior AI Engineer specializing in healthcare and FinTech AI systems
Director of Software Engineering specializing in enterprise Data, ML & AI platforms
“Former Walmart Director of Software Engineering who left in March 2025 to build products for clients. Recently delivered an LLM/RAG-based UNSPSC classification solution for an MRO client using a multi-stage retrieval + web search + prompt-engineering workflow, and has led large-scale retail forecasting initiatives and high-severity cloud-migration incidents end-to-end.”
Junior Machine Learning Engineer specializing in LLMs and data pipelines
“Research Extern at Google DeepMind and former AWS Software Development Engineer Intern with a strong focus on practical, trustworthy AI engineering. Built a multi-agent RAG system for personalized news headline generation using a fine-tuned Flan-T5 model, parallel critic agents, FAISS retrieval, and style embeddings, while also leading a 3-person team on the project.”
Intern Software Engineer specializing in data engineering and LLM/RAG systems
“Built and productionized enterprise LLM/RAG systems, including a Boeing internal solution that gave 400+ program managers conversational access to 1M+ rows of schedule data, with strong emphasis on governance, reliability, and reducing hallucinations in tabular domains. Also has experience running developer-focused workshops (UC Berkeley computer architecture) and partnering with customer-facing stakeholders to drive adoption of a compliance-sensitive NLP product (SEC-aligned) at Penserra.”
Staff Software Engineer specializing in Healthcare platforms and AI data pipelines
“Backend/data engineer with hands-on production AWS experience spanning serverless APIs (Chalice/Lambda/API Gateway/Cognito) and data pipelines (Glue PySpark + Step Functions). Has modernized a legacy SAS reporting system into AWS microservices and implemented schema-drift detection and incident prevention for ETL workflows, plus measurable SQL tuning wins (30 min to <10 min runtime).”
Intern Data Scientist specializing in GenAI (LLMs, RAG) and ML model optimization
“Built and deployed a production LLM-powered risk assistant for KPMG and Freddie Mac that lets analysts query a confidential Neo4j risk graph in natural language (no Cypher), turning multi-day analysis into minutes with traceable, cited answers. Implemented rigorous guardrails, deterministic verification, RBAC/security controls, and a full eval/observability stack, cutting query error rate by ~50% and iterating through weekly UAT with non-technical risk analysts.”
Junior AI Software Engineer specializing in LLM pipelines, OCR, and RAG
“Built and shipped a production LLM pipeline for nursing home Medicare reimbursement (PDF OCR + fact extraction + keyword RAG + QA) that reportedly increased payouts by ~$1K/month per patient. Strong in LLM ops/benchmarking (ground truth, LLM-as-judge, cost/I-O tracking) and pragmatic optimization—swapped retrieval approaches, fine-tuned a small model to cut OCR cost 90%, and migrated workloads to Azure/Temporal to scale nightly processing 10x.”
Mid-Level Software Development Engineer specializing in GenAI and full-stack cloud systems
“Full-stack engineer with experience across Magna, C3.ai, and Amazon, building GenAI-enabled products and finance transaction systems. Has shipped Next.js (App Router) + TypeScript features backed by Go/Python RAG pipelines, and emphasizes production quality via load testing, Selenium regression coverage, LLM-aware integration testing, and Azure observability. Also built LangGraph-orchestrated multi-step content generation workflows with robust retry/idempotency strategies.”
Junior Backend/Platform Engineer specializing in AI microservices and cloud-native systems
“Cofounder at MeowyAI who shipped a production multimodal (vision/voice/text) AI task manager using Gemini, tackling real-world issues like hallucinations, tool-calling safety, and RAG-based preference memory. Also built a production multi-agent RAG system orchestrated with LangGraph (and contributes to LangChain), with strong emphasis on latency optimization, observability (OpenTelemetry), and rigorous testing/evaluation including A/B tests and adversarial prompting.”
Staff Software Engineer specializing in FinTech and AI-powered customer support
“Technical lead who shipped a production GPT-4-powered customer support agent for Square, serving a large fintech customer base through a React chat interface with tool-using orchestration, guardrails, and live handoff paths. Brings strong real-world experience in agent reliability, evaluation, observability, and workflow orchestration using Temporal, Sidekiq, Pinecone, Datadog, and Snowflake.”
Junior ML Engineer specializing in Generative AI and LLM applications
“Built a production internal knowledge assistant using a RAG pipeline over large spreadsheets, PDFs, and support documents, using transformer embeddings stored in FAISS. Focused on real-world production challenges—format normalization, retrieval quality, hallucination reduction (context-only + citations), and latency—using hybrid retrieval, quantization, and containerized deployment, and communicated the workflow to non-technical stakeholders using simple analogies.”
Mid-Level Software Engineer specializing in Generative AI and RAG systems
“Built a production RAG-based natural-language-to-SQL system at Global Atlantic to replace slow, expensive manual analytics ticket workflows, focusing heavily on retrieval quality and measurable evaluation (200-question ground-truth set; recall@5 improved 0.65→0.78 via semantic chunking). Also built a custom MCP-style agent orchestrator for a personal project (arxiv-ai) to improve flexibility and Langfuse-aligned observability, and has hands-on experience with LangGraph, CrewAI, and n8n.”