Pre-screened and vetted.
Senior Software Engineer specializing in backend microservices and data platforms
Senior Full-Stack Software Engineer specializing in cloud microservices and data platforms
Mid-level AI/ML Engineer specializing in conversational AI, NLP, and LLM-powered RAG systems
Senior Machine Learning Engineer specializing in MLOps and Generative AI
Mid-Level Software Engineer specializing in Python microservices and scalable web APIs
“Backend engineer who replaced an Excel-heavy forecasting workflow with a secure, auditable FastAPI system (React UI + relational model + async workers), emphasizing deterministic processing, idempotency, and versioned ledger-style ingestion. Led a monolith-to-FastAPI migration at Bounteous using a strangler approach, feature-flagged incremental rollout, and data reconciliation/shadow-compare to protect integrity while scaling onboarding workflows.”
Mid-Level Full-Stack Engineer specializing in AI and 3D computer vision
“Built and productionized an LLM-driven document verification workflow for a construction firm’s submittals process, moving from a Vercel/Next.js prototype to a FastAPI + LangChain/LangGraph backend with background workers and multi-server deployment. Uses LLM tools (e.g., OpenAI Codex/Cloud Code) for rapid development and log-driven root cause analysis, and partners with customer teams on evaluation metrics and iterative improvements.”
Mid-level AI/ML Engineer specializing in LLM systems and cloud MLOps
“Built a production LLM-powered fraud detection platform at Wells Fargo, combining OpenAI/Hugging Face models with RAG-based explanations to make flagged transactions interpretable for risk and compliance teams. Delivered low-latency, real-time inference at high scale on AWS (SageMaker + EKS), with strong observability and security controls, reducing manual reviews and false positives in a regulated environment.”
Mid-level AI/ML Engineer specializing in LLMs, NLP, and analytics automation
“AI/ML Engineer (TCS) who built and deployed a production LLM-powered audit transaction validation service to reduce manual review of unstructured transaction records and comments. Implemented a LangChain/Python pipeline for extraction/normalization and discrepancy detection, with strong production reliability practices (decision logging, dashboards, labeled eval sets) and a human-in-the-loop auditor feedback loop to improve precision/recall under strict data-sensitivity and near-real-time constraints.”
Director-level Technology Leader specializing in cloud-native platforms, AI/ML, and SaaS
“Engineering leader (Director/VP level) who has repeatedly aligned product and engineering through ROI-driven quarterly roadmaps and strong stakeholder communication, including board presentations. Built a parallel cloud team to migrate an on-prem product to the cloud, credited with delivering $9M ARR, and led a Python monolith-to-serverless event-driven microservices transformation. Currently manages distributed teams across Mexico, India, and the US using pod-based structures, clear KPIs, and a supportive accountability culture.”
Mid-level Software Engineer specializing in Agentic AI and RAG systems
“Built and shipped a production AI-powered Q&A/RAG onboarding assistant at One Community Global that unified knowledge across Notion, Google Docs, and Slack, cutting volunteer onboarding time by 45%. Demonstrates strong end-to-end ownership: LangChain agent orchestration integrated into a FastAPI backend, rigorous evaluation (200-query dataset, ~85% accuracy), and production feedback/monitoring with source-attributed answers to build user trust.”
Mid-level AI Engineer specializing in GenAI and RAG systems
“AI engineer who built a production e-commerce system that analyzes product images alongside sales and demographic data to generate actionable creative recommendations, now used by 20+ clients. Also built orchestrated document/agent pipelines (Airflow, LangGraph) including a compliance drift detector auditing 401 compliance documents, with an emphasis on traceability, logging, and production integration.”
Mid-level Software Engineer specializing in SRE, observability, and LLM-powered automation
Intern Software Engineer specializing in full-stack development and applied AI
“Internship experience building an end-to-end medical AI pipeline that extracts and normalizes messy medical PDFs, fine-tunes BioBERT to classify tumor-related statements (including negation/ambiguity handling), and integrates image-model outputs (MedSAM/GroundingDINO) for tumor localization and classification. Also worked on an LLM/RAG system to draft IPO prospectuses using retrieved regulatory/financial sources (including SEC EDGAR) with structured prompts to reduce hallucinations.”
Mid-level AI/ML Engineer specializing in MLOps and production ML systems
“Backend/ML engineer who has shipped high-scale real-time systems across e-commerce and healthcare: built a PharmEasy real-time recommendation engine for ~2M monthly users (cut feature latency 5 min→30 sec; +15% cross-sell) and architected a HIPAA-compliant multimodal clinical diagnostic workflow (DICOM+EHR) with XAI, MLOps (MLflow/Airflow/K8s), and drift/monitoring guardrails supporting 10k+ daily predictions.”
Junior Software Engineer specializing in AI, LLM systems, and full-stack development
“Product-focused full-stack engineer at startup (Zippy) who shipped a production multi-agent AI system for restaurant operations plus payments workflows. Built end-to-end: RAG grounded on a Notion knowledge base, structured function-calling task routing, FastAPI/JWT multi-tenant backend, and a polished React+TypeScript owner dashboard. Has real production incident experience (duplicate Stripe webhooks) and reports ~94% task-routing accuracy under load.”
Intern LLM/GenAI Engineer specializing in RAG, agentic systems, and low-latency inference
“Interned at Larsen & Toubro where they built and deployed an agentic RAG document question-answering system to reduce time spent searching documents and improve trustworthiness. Implemented ReAct-style multi-step orchestration with LangChain/LlamaIndex plus evidence-bounded generation, grounding/citations, and rigorous evaluation—cutting latency ~40%, hallucinations ~35%, and unsafe outputs ~40% while collaborating closely with non-technical business/ops stakeholders.”
Mid-level Full-Stack Software Engineer specializing in Java microservices and cloud-native systems
“Backend engineer with experience building and modernizing high-volume healthcare transaction systems, including migrating Java services to Spring Boot microservices and adopting Kafka-based event-driven architectures. Strong focus on production reliability and operability (observability, CI/CD, standardized patterns) plus security (OAuth/JWT, RBAC, Postgres/Supabase RLS) and resilient stream processing (idempotency, DLQs).”
Intern Software Engineer specializing in AI/LLMs and full-stack development
“AI/ML infrastructure-focused engineer who has built production RAG systems from scratch (Supabase/pgvector + OpenAI embeddings) and iterated using formal eval metrics to improve retrieval quality. Also debugged real-time audio issues in a LiveKit-based pipeline by correlating packet loss with VAD behavior, and has deep experience building brittle, customer-specific financial platform integrations in Python/Playwright (2FA, redirects, token refresh, rate limits).”
Intern Data Scientist specializing in ML, NLP, and MLOps for healthcare and enterprise AI
“Built a production multi-cloud LLM-driven IT ticket automation system using LangGraph, Azure + Pinecone RAG, and an Ollama-hosted LLM on AWS, with Terraform-managed infra and PostgreSQL audit/state tracking for reliability. Also partnered with UW School of Medicine & Public Health students to deliver a glioma survival risk-ranking model, translating clinical feedback into practical pipeline improvements (imputation, site harmonization) and stakeholder-friendly visualizations.”
Mid-level Machine Learning Engineer specializing in Generative AI and RAG systems
“LLM/ML engineer who has shipped an enterprise RAG-based Q&A system (LangChain/LlamaIndex, FAISS + Azure Cognitive Search, GPT-3.5/4 via OpenAI/Azure OpenAI) to production on Docker + Kubernetes/OpenShift, tackling hallucinations, retrieval quality, latency/cost, and RBAC/IAM security. Also partnered with operations leaders to turn manual reporting into an LLM-powered summarization and forecasting dashboard driven by real KPIs and iterative stakeholder feedback.”
Mid-level Data Scientist specializing in AI/ML, MLOps, and LLM-powered analytics
“Built and deployed a production LLM-powered document Q&A system enabling natural-language querying of large PDFs, focusing on retrieval quality (overlapped chunking) and low-latency performance (optimized embeddings + vector search). Experienced with scaling ML/LLM workflows using async/batch processing, caching, cloud storage, and orchestration via Apache Airflow with robust testing, monitoring, and failure handling.”
Mid-level Full-Stack Software Engineer specializing in cloud-native microservices
“Cloud-native integration engineer (Oracle/OCI) with strong production deployment and incident-response experience, including API gateway rollouts, observability (Prometheus/Grafana), and multi-layer debugging for payments systems. Built Python/FastAPI microservices and automation for customer-specific reporting and data sync, and has delivered major performance gains (45 min to <10) plus reliability improvements (MTTD reduced 40%+) through monitoring, playbooks, and resilient integration patterns (streaming/queuing, retries, secure tokens, VPC peering).”
Mid-level Forward Deployed Engineer specializing in AI automation for finance and data platforms
“LLM/agentic workflow specialist with healthcare deployment experience who has taken LLM-based automation from prototype to production using operator-in-the-loop validation, RAG-style retrieval, RBAC, and monitoring for sensitive data compliance. Demonstrated real-time incident resolution (retrieval timeouts due to network/proxy misconfig) and strong GTM support—hands-on developer workshops and sales demos translating technical safeguards and real-time ETL into measurable ROI (70% ops reduction, ~$200K/year savings).”