Pre-screened and vetted.
Mid-level AI Engineer and Data Scientist specializing in LLM agents and RAG systems
“Built a production-grade LLM evaluation and regression system that stress-tests models across hundreds of iterations, combining LLM-as-judge, semantic similarity, statistical metrics, and rule-based checks, with results delivered via stakeholder-friendly HTML reports and dashboards. Experienced orchestrating multi-agent RAG workflows using LangChain/LangGraph and event-driven GenAI pipelines in n8n integrating OCR, speech-to-text, and external APIs, with strong emphasis on reliability, observability, and explainable failures.”
Mid-level Data Scientist specializing in Generative AI and LLMOps
“Built a production-grade, semi-automated document recognition and classification system for large volumes of scanned PDFs, starting from little/no labeled data and handling highly variable scan quality. Deployed on AWS using SageMaker + Docker and orchestrated on EKS with a microservices design that scales CPU-heavy OCR separately from GPU inference, with strong reliability controls (validation, fallbacks, retries, readiness probes).”
Mid-level Software Engineer specializing in full-stack development, data engineering, and GenAI
“Built and deployed an LLM product called "Content Craft" combining BART-based summarization with a RAG Q&A chatbot using LangChain, embeddings, and a vector database. Has hands-on MLOps experience containerizing and serving models with FastAPI and running them on Kubernetes with monitoring, self-healing, and autoscaling, and has practical experience reducing hallucinations through structured prompting.”
Mid-level AI Data Engineer specializing in GenAI, RAG, and cloud data pipelines
“LLM/agentic AI builder who deployed a production ITSM automation agent on Google ADK integrating ServiceNow and FreshService, with strong safety guardrails (human-approval gating and runbook-only command execution) and rigorous evaluation (500 synthetic tickets; 80%+ false-positive reduction). Also partnered with finance to deliver an AI agent that automated invoice/SOW retrieval and monthly reporting to account managers, reducing manual back-and-forth.”
Mid-level AI Engineer specializing in Generative AI and LLM systems
“Built and deployed a production-grade, multi-agent Text-to-SQL assistant that lets non-technical stakeholders query large enterprise databases in natural language. Uses Pinecone-based schema retrieval + LLM reasoning (Gemini/Claude/GPT) with a dedicated validation agent (schema/syntax checks and safe dry runs) to reduce hallucinations and improve reliability, while optimizing latency and cost via async execution and embedding caching.”
Mid-level AI Engineer specializing in RAG, conversational AI, and agentic systems
“Built and deployed a production RAG-based clinical decision support assistant at MedLib, focused on fast, trustworthy answers from large medical documents. Demonstrates deep practical experience improving retrieval accuracy (semantic chunking + metadata-aware search), controlling hallucinations with grounded generation and thresholds, and adding clinician-requested citations using chunk metadata, with evaluation driven by healthcare professional review.”
Mid-Level Software Engineer specializing in backend systems, cloud, and applied LLM/NLP
“Applied LLMs to classify long nonprofit mission statements into 8 segments without labeled data, using an ensemble of clustering/embedding methods plus zero-shot RoBERTa/BART and a Tree-of-Thought prompting pipeline with LLM-as-judge evaluation (Gemma). Also built LangChain/LlamaIndex agentic RAG workflows including a text-to-SQL data analysis assistant grounded on DB schema with retries and performance optimizations on an HPC cluster.”
Junior AI Data Engineer specializing in Azure Databricks lakehouse and GenAI RAG systems
“Backend/applied AI engineer from Cloud Rack Systems who built production GenAI/RAG and data platforms on Azure/Databricks at enterprise scale (2.5M records/day). Known for making LLM systems behave like deterministic services via strict retrieval contracts, citation-based validation, and strong observability—shipping a knowledge assistant used daily by 50+ users while driving hallucinations near zero and materially improving latency and cost.”
Junior Software Engineer specializing in backend systems and machine learning
“Independent builder of production-grade systems: shipped an end-to-end URL shortener with JWT auth, Redis rate limiting/caching, Postgres, Docker, and real-time analytics, and separately architected a Redis-backed distributed task queue handling 1000+ tasks/min. Demonstrates strong distributed-systems instincts (atomicity, retries/DLQ, idempotency, heartbeats) plus a focus on maintainable code and self-documenting APIs (FastAPI/OpenAPI, versioned routes).”
“Built a production ad-spend optimization system that combined deterministic audit logic with LLM-generated explanations, surfacing severe inefficiencies including 70-90% wasted spend in some Google Ads accounts. Stands out for pairing measurable business impact with pragmatic AI safety and usability decisions, including approval-gated execution and structured, human-readable recommendations.”
Senior AI Engineer specializing in LLMs, RAG, and production ML systems
“Built GynAI, an end-to-end maternal clinical decision support platform for OB/GYN practices and hospitals in North America, combining predictive ML with RAG-based LLM explainability. The candidate emphasizes real production ownership across experimentation, deployment, monitoring, and iteration, with reported impact including fewer delayed interventions in high-risk pregnancies and a 15-20% reduction in false positives.”
Mid-level AI Engineer specializing in Python, LLMs, and production ML systems
“Production-focused ML/AI engineer with hands-on ownership across classical ML and GenAI systems, from CV/NLP services to enterprise RAG. Stands out for combining research-to-production execution with measurable business impact: 40% processing-efficiency gains, 35% fewer support tickets, 5x latency improvement, and 3x throughput gains while maintaining safety and quality.”
Junior Full-Stack & LLM Engineer specializing in AI agents and cloud document intelligence
“Backend engineer specializing in event-driven/serverless systems and Python/FastAPI APIs. Built a scalable PDF-to-structured-data pipeline on AWS (S3, Lambda, Step Functions, Textract, DynamoDB, SNS) with strong observability (p50/p90/p99) and reliability patterns (idempotency, retries/DLQs), and has led zero-downtime migrations using feature flags, dual writes, and incremental rollouts.”
Mid-level Full-Stack AI Engineer specializing in deployed LLM agents and RAG systems
“Built a real-time AI meeting assistant using a Chrome extension that streams audio to a backend LLM workflow with transcription and RAG, then hardened it for production with queue-based streaming, async pipelines, security controls, and full observability. Also has hands-on startup sales experience, partnering with customers to define measurable technical win conditions (latency/accuracy) to close deals and drive adoption.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and MLOps
“LLM engineer/data analyst who built a production RAG QA assistant over the Jurafsky & Martin NLP textbook to reduce hallucinations and provide explainable, source-grounded answers. Experienced with LangChain/LangGraph orchestration, retrieval optimization (embeddings, vector DBs, caching), and rigorous evaluation/monitoring (Retrieval@K, A/B tests, telemetry/drift). Previously communicated analytics insights to non-technical stakeholders at GS Analytics using Power BI and simplified reporting.”
Mid-Level Full-Stack/Product Engineer specializing in B2B SaaS and AI search systems
“Full-stack engineer operating in early-stage, high-velocity environments (OpGov.AI/UST Calibrate) who ships production Next.js App Router features end-to-end (RSC, Server Actions, SEO, RBAC, caching) and owns performance post-launch. Demonstrates strong data/infra depth—designed Postgres JSONB-based event models for DevOps/DORA analytics and tuned queries from ~2s to <50ms, plus built durable ingestion workflows with retries and idempotency on Azure.”
“Forward Deployed Engineer at EasyBee AI who productionized a self-storage customer’s multi-agent LLM system end-to-end—rebuilding it with LangGraph/CrewAI, integrating with real property management + CRM systems via an MCP server, and adding observability/guardrails for reliable daily use. Experienced in live troubleshooting of agentic workflows, developer demos/workshops (including an open-source project, MerryQuery), and partnering with sales to close deals through customer-specific technical demos and fast integration feedback loops.”
Junior Full-Stack Software Engineer specializing in Python APIs, React, and cloud AI integrations
“Customer-facing software engineer who builds and deploys practical AI/RAG solutions (e.g., an AI assistant for searching billing PDFs) by deeply understanding support workflows and iterating with users. Demonstrates strong production instincts—quickly stabilizing peak-traffic API timeouts with caching/background jobs, then implementing durable fixes with proper monitoring and maintainable code practices.”
Mid-level AI Engineer specializing in Generative AI, LLMs, and RAG
“Internship at Discovery Education building a production LLM/RAG chatbot that let marketing and sales teams query and interpret Looker/BI dashboards in natural language, with responses grounded in compliance and state education standards. Emphasizes rigorous evaluation (faithfulness/precision/recall/latency) plus user-feedback analytics, and used LangChain for orchestration, chunking/context-window control, and integration with enterprise sources like SharePoint.”
Junior Machine Learning Engineer specializing in data science and automation
“Built and shipped an end-to-end AI-powered portfolio chatbot, owning the React frontend, FastAPI backend, and FAISS-based retrieval layer. Demonstrates hands-on full-stack product thinking with attention to UI performance, TypeScript maintainability, and post-launch iteration on response relevance and speed.”
Mid-level Full-Stack & AI Engineer specializing in LLM applications
“Full-stack engineer who has shipped and operated generative-AI chat/QA features end-to-end, including a RAG-based pipeline with guardrails and cost/latency monitoring in production. Experienced with React/TypeScript + Node/Postgres architectures, Dockerized deployments to AWS (EC2) via GitHub Actions CI/CD, and building reliable ingestion/ETL systems with idempotency, backfills, and reconciliation.”
Junior Software Engineer specializing in ML, RAG systems, and safety-critical risk modeling
“Backend/cloud engineer from Resilient Tech with hands-on experience deploying REST APIs and database migrations into a live ERP used by real customers while maintaining 99% uptime. Has debugged intermittent AWS container timeouts down to security group/load balancer misconfigurations, and has extended Python in an ERPNext system to meet GST/e-invoicing compliance requirements with strong customer collaboration.”
Intern Data Scientist specializing in GenAI agents, RAG, and ML platforms
“LLM/agent systems builder who deployed a production hybrid router for immerso.ai that dynamically selects retrieval vs reasoning vs generative pathways, achieving an 82% factual-accuracy lift. Deep hands-on experience optimizing local Mistral 7B inference (4–5 bit GGUF quantization, KV-cache reuse) and building reliable RAG/agent workflows with LangChain/LangGraph/AutoGen across GCP Cloud Run and AWS (ECS/Lambda).”
Junior Full-Stack Software Engineer specializing in cloud-native web apps and AI tooling
“Software engineer with experience across edtech, live gaming, and an AI document intelligence platform, delivering end-to-end customer-facing features and production backends. Built secure, automated live-session scheduling integrating Zoom and TalentLMS (JWT/RBAC, idempotency, transactions) cutting setup time from ~3 minutes to under 1 minute, and optimized real-time gaming dashboards/APIs with query tuning, caching, and CDN improvements (~60% latency reduction under peak load) on AWS.”