Pre-screened and vetted.
Junior AI Integration Engineer specializing in LLM agents and RAG on cloud platforms
“Built and deployed LLM-powered features for a startup organizational management application, focusing on real-world deployment constraints like latency and cost. Implemented RAG with FAISS and improved retrieval quality by switching embedding models (OpenAI/Hugging Face) and fine-tuning embeddings on medical corpora for a medical-report UI feature. Uses LangChain and LangGraph to orchestrate multi-node LLM API workflows and evaluates systems with metrics like latency, cost per request, and error taxonomy.”
Senior AI/ML Engineer & Data Scientist specializing in LLMs, RAG, and MLOps
“ML/NLP practitioner who has delivered production systems in regulated domains, including a healthcare compliance pipeline using RAG (GPT-4/Claude) plus TF-IDF retrieval that increased document review throughput 4.5x. Also has hands-on experience improving fraud detection data quality via entity resolution (Levenshtein, Dedupe.py) validated with A/B testing, and building scalable, monitored workflows with Airflow, CI/CD, and AWS SageMaker.”
Entry AI Engineer specializing in LLMs, RAG, and MLOps
“Built and shipped a production Python-based agentic RAG document retrieval system over 80K records using FastAPI, OCR, vector search, and AWS infrastructure, with a strong emphasis on reliability, testing, and observability. Stands out for treating AI failures like production incidents—turning hallucinations, retrieval misses, and OCR issues into regression tests—and for quantifiably reducing document lookup time from about 12 minutes to under 90 seconds.”
Mid-level AI/ML Engineer specializing in MLOps, NLP, and Generative AI
“Built and deployed a production LLM-powered text-to-SQL/document intelligence chatbot on AWS that lets non-technical business users query complex enterprise databases in plain English. Demonstrates deep practical expertise in schema-aware prompting, embeddings-based schema retrieval, SQL safety/validation guardrails, and rigorous offline/online evaluation with human-in-the-loop approvals for risky queries.”
Entry-Level Software Engineer specializing in AI APIs and RAG systems
“Junior/entry-level AI/LLM engineer who built a production-oriented RAG onboarding and knowledge assistant that ingests GitHub repos and internal sources (e.g., Confluence/Jira) using ChromaDB, with reliability features like retrieval fallbacks, retries, caching, and monitoring. Currently implementing a LangGraph-based multi-agent workflow with intent routing and Pydantic/Magentic-validated structured outputs, plus CI/CD offline evals and online metrics (Grafana/Prometheus) to improve predictability and reliability.”
Mid-level Data Scientist specializing in Generative AI and LLMOps
“Built a production-grade, semi-automated document recognition and classification system for large volumes of scanned PDFs, starting from little/no labeled data and handling highly variable scan quality. Deployed on AWS using SageMaker + Docker and orchestrated on EKS with a microservices design that scales CPU-heavy OCR separately from GPU inference, with strong reliability controls (validation, fallbacks, retries, readiness probes).”
Mid-level Software Engineer specializing in full-stack development, data engineering, and GenAI
“Built and deployed an LLM product called "Content Craft" combining BART-based summarization with a RAG Q&A chatbot using LangChain, embeddings, and a vector database. Has hands-on MLOps experience containerizing and serving models with FastAPI and running them on Kubernetes with monitoring, self-healing, and autoscaling, and has practical experience reducing hallucinations through structured prompting.”
Mid-Level Software Engineer specializing in distributed systems and AI agent workflows
“Software engineer with enterprise CPQ/CRM/ERP integration experience (Argano) who owned an end-to-end pricing preview capability deployed on AWS Kubernetes with Jenkins CI/CD and full observability (Prometheus/Grafana). Also built an AI-native research agent using LangChain + Chroma to filter academic papers, reporting ~15 hours/week saved for a professor.”
Mid-level AI Engineer specializing in RAG, conversational AI, and agentic systems
“Built and deployed a production RAG-based clinical decision support assistant at MedLib, focused on fast, trustworthy answers from large medical documents. Demonstrates deep practical experience improving retrieval accuracy (semantic chunking + metadata-aware search), controlling hallucinations with grounded generation and thresholds, and adding clinician-requested citations using chunk metadata, with evaluation driven by healthcare professional review.”
Junior AI Data Engineer specializing in Azure Databricks lakehouse and GenAI RAG systems
“Backend/applied AI engineer from Cloud Rack Systems who built production GenAI/RAG and data platforms on Azure/Databricks at enterprise scale (2.5M records/day). Known for making LLM systems behave like deterministic services via strict retrieval contracts, citation-based validation, and strong observability—shipping a knowledge assistant used daily by 50+ users while driving hallucinations near zero and materially improving latency and cost.”
Junior Software Engineer specializing in backend systems and machine learning
“Independent builder of production-grade systems: shipped an end-to-end URL shortener with JWT auth, Redis rate limiting/caching, Postgres, Docker, and real-time analytics, and separately architected a Redis-backed distributed task queue handling 1000+ tasks/min. Demonstrates strong distributed-systems instincts (atomicity, retries/DLQ, idempotency, heartbeats) plus a focus on maintainable code and self-documenting APIs (FastAPI/OpenAPI, versioned routes).”
Director-level growth marketer specializing in international SaaS, e-commerce, and AI-native acquisition
“Founder of distribb.io, an 8-month-old AI SEO software company already at $10k MRR. Brings 15 years of startup ecosystem exposure and a pragmatic, market-driven approach to building businesses by targeting proven demand rather than chasing novelty.”
Senior AI Engineer specializing in LLMs, RAG, and production ML systems
“Built GynAI, an end-to-end maternal clinical decision support platform for OB/GYN practices and hospitals in North America, combining predictive ML with RAG-based LLM explainability. The candidate emphasizes real production ownership across experimentation, deployment, monitoring, and iteration, with reported impact including fewer delayed interventions in high-risk pregnancies and a 15-20% reduction in false positives.”
Mid-level Software Engineer specializing in backend systems and AI-powered platforms
“Backend engineer who built a production retrieval-augmented narrative analysis platform for 100-page screenplays using a Node/Express orchestrator and a Python/FastAPI AI engine, including a key redesign from disk-based uploads to in-memory streaming to eliminate Windows file-lock failures. Also led a refactor of a municipal vehicle tracking system into a C-based distributed engine handling 4M+ daily packets with 99.99% data integrity and automation that reduced manual ops by 50%.”
Mid-level Full-Stack AI Engineer specializing in deployed LLM agents and RAG systems
“Built a real-time AI meeting assistant using a Chrome extension that streams audio to a backend LLM workflow with transcription and RAG, then hardened it for production with queue-based streaming, async pipelines, security controls, and full observability. Also has hands-on startup sales experience, partnering with customers to define measurable technical win conditions (latency/accuracy) to close deals and drive adoption.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and MLOps
“LLM engineer/data analyst who built a production RAG QA assistant over the Jurafsky & Martin NLP textbook to reduce hallucinations and provide explainable, source-grounded answers. Experienced with LangChain/LangGraph orchestration, retrieval optimization (embeddings, vector DBs, caching), and rigorous evaluation/monitoring (Retrieval@K, A/B tests, telemetry/drift). Previously communicated analytics insights to non-technical stakeholders at GS Analytics using Power BI and simplified reporting.”
Intern AI & Machine Learning Engineer specializing in computer vision and edge deployment
“Built and shipped a real-time AI robotic inspection system, using a synthetic data generation pipeline to address rare edge cases—cutting data collection costs ~60% and boosting hard-scenario accuracy ~20%. Experienced in productionizing ML on constrained Jetson hardware and orchestrating end-to-end ML workflows with Airflow/Docker/Kubernetes, with a metrics-driven approach to reliability, evaluation, and stakeholder communication.”
Junior Backend/Platform Engineer specializing in cloud-native APIs and data systems
“Startup-style full-stack/backend engineer with hands-on AWS architecture experience who shipped an LLM-driven assessment-question automation feature (Python microservice calling AWS Bedrock via SQS, deployed on Lambda) with strong validation/guardrails and retry strategies. Also improved production scalability by moving a CPU/IO-heavy file upload path out of a Go API into a queue/Lambda design monitored with CloudWatch, and has React+TypeScript experience optimizing analytics dashboards.”
Mid-Level Full-Stack/Product Engineer specializing in B2B SaaS and AI search systems
“Full-stack engineer operating in early-stage, high-velocity environments (OpGov.AI/UST Calibrate) who ships production Next.js App Router features end-to-end (RSC, Server Actions, SEO, RBAC, caching) and owns performance post-launch. Demonstrates strong data/infra depth—designed Postgres JSONB-based event models for DevOps/DORA analytics and tuned queries from ~2s to <50ms, plus built durable ingestion workflows with retries and idempotency on Azure.”
“Forward Deployed Engineer at EasyBee AI who productionized a self-storage customer’s multi-agent LLM system end-to-end—rebuilding it with LangGraph/CrewAI, integrating with real property management + CRM systems via an MCP server, and adding observability/guardrails for reliable daily use. Experienced in live troubleshooting of agentic workflows, developer demos/workshops (including an open-source project, MerryQuery), and partnering with sales to close deals through customer-specific technical demos and fast integration feedback loops.”
Mid-level AI Engineer specializing in Generative AI, LLMs, and RAG
“Internship at Discovery Education building a production LLM/RAG chatbot that let marketing and sales teams query and interpret Looker/BI dashboards in natural language, with responses grounded in compliance and state education standards. Emphasizes rigorous evaluation (faithfulness/precision/recall/latency) plus user-feedback analytics, and used LangChain for orchestration, chunking/context-window control, and integration with enterprise sources like SharePoint.”
Senior Full-Stack Software Engineer specializing in AI-powered web and mobile applications
“Backend/full-stack TypeScript engineer who has owned end-to-end, production-oriented systems including an AI property management platform (NestJS/Postgres/WebSockets on Google Cloud using Gemini Vision) and an AI logistics platform (Node/Redis queues/Postgres) focused on low-latency, correctness, and observability. Also designed a public GraphQL API and TypeScript SDK for education partners at StudyFetch, citing 40+ partner integrations in the first quarter.”
Junior Machine Learning Engineer specializing in data science and automation
“Built and shipped an end-to-end AI-powered portfolio chatbot, owning the React frontend, FastAPI backend, and FAISS-based retrieval layer. Demonstrates hands-on full-stack product thinking with attention to UI performance, TypeScript maintainability, and post-launch iteration on response relevance and speed.”
Mid-level AI/ML Software Engineer specializing in GPU-optimized LLM inference and cloud microservices
“Built and deployed a production RAG-based multilingual analytics assistant for healthcare operations, enabling non-technical teams to query claims/EHR and risk metrics with grounded explanations. Demonstrates strong end-to-end LLM system engineering (retrieval tuning, re-ranking, hallucination controls, verification layers) plus workflow orchestration (Airflow/Composer/Step Functions) and stakeholder-driven iteration via prototypes and dashboards.”