Pre-screened and vetted.
Executive VP of Engineering specializing in FinTech platforms, cloud modernization, and AI/ML
Senior Data Engineer specializing in cloud lakehouse platforms and healthcare data
Mid-level AI Data Engineer specializing in real-time streaming and LLM-powered fraud analytics
Mid-level Data Engineer specializing in analytics engineering, ML forecasting, and modern data stacks
Mid-level Generative AI Engineer specializing in LLMs, NLP, and multimodal systems
Senior Software Engineer specializing in distributed systems and agentic AI platforms
Mid-level DevOps Engineer specializing in cloud-native CI/CD and Kubernetes
Mid-level Software Engineer specializing in cloud-native platforms and healthcare systems
“Backend engineer with healthcare-domain experience building a security-critical RBAC identity/authentication/authorization microservice suite used across hospital imaging platforms (X-Ray, Ultrasound, etc.). Demonstrates strong security mindset (mTLS, cert hygiene, JWT, pen-testing collaboration) and pragmatic scaling/reliability practices (Nginx load balancing, Redis caching, automated tests, canary rollouts).”
Mid-level AI/ML Engineer specializing in LLM applications and cloud-native systems
“LLM engineer who has shipped production AI systems, including an RFP requirements extraction platform (OpenAI o4-mini + Azure AI Search + FastAPI) achieving 90%+ accuracy and ~5x throughput through grounding, structured outputs, parallelization, and caching. Also partnered with legal/compliance stakeholders at Nexteer Automotive to deliver an AI document comparison tool with traceability and confidence indicators, adopted by non-technical users and saving ~2 FTEs of review time.”
Mid-level Software Engineer specializing in ML platforms and cloud-native backend systems
“Software engineer with experience at Google and the City and County of San Francisco building production AI systems, including a RAG-based internal support chatbot and ML-driven ticket priority tagging. Has scaled data/ML platforms with Airflow on GCP (1M+ records/day, 99.9% SLA) and deployed multi-component systems with Docker and Kubernetes (GKE), using modern LLM tooling (LangChain/CrewAI, Claude/OpenAI, Pinecone/ChromaDB, Bedrock/Ollama).”
Staff/Lead Data Scientist specializing in Generative AI, NLP/LLMs, and MLOps
“Lead Data Scientist (10+ years) with recent work in healthcare data: built production pipelines that unify EHR, genomics, and clinical notes using NLP (spaCy/BERT/BioBERT) and scalable Spark-based processing. Also led development of domain-specific LLM/NLP systems for chatbots and semantic search, deploying models via FastAPI/Flask and improving retrieval with FAISS-backed, fine-tuned clinical embeddings and RAG-style workflows.”
Senior Data Scientist / ML Engineer specializing in GenAI, LLMs, and NLP
“ML/NLP engineer focused on production GenAI and data linking systems: built a large-scale RAG pipeline over millions of support docs using LangChain/Pinecone and added a LangGraph-based validation layer to cut hallucinations ~40%. Also built scalable PySpark entity resolution (95%+ accuracy) and fine-tuned Sentence-BERT embeddings with contrastive learning for ~30% relevance lift, with strong CI/CD and observability practices (OpenTelemetry, Prometheus/Grafana).”
Intern Full-Stack/AI Software Engineer specializing in GenAI and cloud microservices
“Backend engineer who owned the AI/data pipeline layer for an EV-charging management platform (Ampure Intelligence), ingesting real-time charger telemetry via OCPP and serving FastAPI APIs to web/mobile clients. Strong in production reliability for asynchronous systems (state reconciliation, idempotency), Kubernetes GitOps (ArgoCD), Kafka streaming, and zero-downtime cloud-to-on-prem migrations; also improved LSTM-based forecasting through targeted preprocessing.”
Mid-level Data Scientist/ML Engineer specializing in GenAI agents and MLOps
“AI/LLM engineer at Capital One who deployed a production RAG-powered fraud analysis and document intelligence platform using LangChain, OpenAI, Pinecone, Kafka, and AWS. Focused on reliability in real-time investigations via hybrid retrieval, schema-validated outputs, and LLM verification loops, reporting review-time reduction from hours to minutes and ~99% fraud detection precision.”
Junior Data & Machine Learning Engineer specializing in MLOps and NLP
“ML/LLM practitioner with production experience building a healthcare review sentiment pipeline (RateMDs) using Hugging Face Transformers plus a LangChain+FAISS RAG layer for interactive querying. Also led orchestration-driven optimization of Nike’s Fusion ETL pipeline, improving runtime efficiency by 20%, and has experience translating ML outputs into Tableau dashboards for non-technical healthcare stakeholders (e.g., readmission risk).”
Senior Data Engineer specializing in cloud lakehouse and real-time streaming pipelines
“Senior data engineer with experience in both healthcare (CVS Health) and financial services (Bank of America), building large-scale Azure lakehouse pipelines (30+ EHR sources, ~5TB) and real-time streaming services (Event Hubs/Kafka) for patient vitals. Strong focus on reliability and data quality (Great Expectations, monitoring/alerting, schema drift automation), with measurable outcomes like 50% runtime reduction and 99%+ uptime for regulatory reporting pipelines.”
Senior AI/ML Engineer specializing in Generative AI and agentic multi-agent systems
“Built and shipped a production LLM-powered multi-agent RAG system to automate complex internal support workflows, integrating tool execution (SQL/APIs) with validation guardrails to reduce hallucinations. Optimized for real-world latency and cost via model routing, caching, and async parallel tool calls, and enforced reliability with CI-gated golden test sets derived from anonymized production queries.”
Junior AI Engineer specializing in fraud detection, credit risk, and LLMs in FinTech
“AI engineer with production experience building a high-accuracy (98%) fraud detection system operating at real-time latency (1–2s) over millions of transactions, using a multi-model pipeline approach to meet performance constraints. Also implemented Airflow-orchestrated workflows (DAGs, retries, alerts) to replace brittle cron scripts and is currently pursuing a master’s project on real-time ASL-to-text conversion.”
Mid-level Full-Stack Product Engineer specializing in data-driven web apps and healthcare systems
“Full-stack engineer with production experience shipping a healthcare-focused web app (Pregnancy-Pal) using Next.js/TypeScript on GCP, integrating a Python/Flask middleware and FHIR server for patient/practitioner dashboards and messaging. Former Wikimedia Foundation Android engineer who led the end-to-end 'Year in Review' feature and built robust automated testing/CI practices (Espresso, GitHub Actions matrix). Strong emphasis on reliability via rigorous validation, comprehensive Postman testing, and detailed API documentation.”
Intern Full-Stack Software Engineer specializing in cloud data pipelines and internal tools
“Built an internal Meta tool (HiVA Bot) for notification customization and end-to-end task tracking around advertiser-reported issues, including chat-thread creation, org-hierarchy opt-ins, SLA reminders, and search/typeahead features. Implemented the system with a Java/Spring Boot microservices approach and asynchronous patterns, and supported adoption via internal wiki documentation.”
Mid-level Software Engineer specializing in cloud platforms, data engineering, and distributed systems
“Full-stack engineer who built and owned an AI-assisted job-matching dashboard in Next.js App Router/TypeScript, keeping LLM logic server-side and improving performance via deduplication, caching/revalidation, and streaming (35% fewer duplicate LLM calls; 40% faster first render). Also has strong data/backend chops: designed Postgres models and optimized queries at million-record scale (1.8s to 120ms) and built durable AWS multi-region telemetry workflows with idempotency, retries, and monitoring.”
Senior Data Scientist / Generative AI Engineer specializing in fraud, risk, and MLOps
“Built and deployed a production LLM/RAG fraud investigation system to replace manual investigator workflows, combining transaction data, historical cases, and policy documents with agent-style steps and LoRA fine-tuning. Demonstrates strong reliability engineering (grounding, citations, abstention paths), performance optimization (retrieval/indexing/caching), and end-to-end MLOps orchestration using Azure ML Pipelines/MLflow plus Kubernetes/Argo with canary and rollback deployments.”