Pre-screened and vetted.
Mid-Level Machine Learning Engineer specializing in LLMs and RAG systems
Junior Machine Learning Engineer specializing in deep learning and healthcare AI
Junior AI/ML Engineer specializing in RAG and multi-agent LLM systems
Mid-Level Full-Stack Software Engineer specializing in cloud microservices and mobile apps
Mid-level AI/ML Engineer specializing in cloud AI, MLOps, and NLP
Mid-level AI/ML Engineer specializing in MLOps, streaming data, and NLP/CV
Mid-level Backend Software Engineer specializing in AI-powered microservices and cloud infrastructure
Mid-level AI Engineer specializing in LLMs, RAG, and enterprise analytics
Mid-level AI & Backend Engineer specializing in RAG systems and scalable APIs
“Built and deployed a production LLM-powered document Q&A system using a strict RAG pipeline (LangChain-style orchestration + FAISS) to help users query large internal document sets. Demonstrates strong reliability focus through hallucination mitigation, curated offline evaluation with grounding checks, and production monitoring (latency/fallback rates) plus stakeholder alignment via demos and business metrics.”
Junior AI/ML Software Engineer specializing in LLM agents and RAG systems
“AI/back-end engineer at Canon who helped build and operate an internal production LLM platform that acts as a secure middle layer between users and models, defending against jailbreaks/prompt injection while enabling RAG, memory, and grounded responses over company data. Experienced with LangChain/LangGraph orchestration, vector DB retrieval, and reliability practices (testing, monitoring, adversarial prompts) to run high-throughput, low-latency AI workflows in production.”
Junior AI Integration Engineer specializing in LLM agents and RAG on cloud platforms
“Built and deployed LLM-powered features for a startup organizational management application, focusing on real-world deployment constraints like latency and cost. Implemented RAG with FAISS and improved retrieval quality by switching embedding models (OpenAI/Hugging Face) and fine-tuning embeddings on medical corpora for a medical-report UI feature. Uses LangChain and LangGraph to orchestrate multi-node LLM API workflows and evaluates systems with metrics like latency, cost per request, and error taxonomy.”
Senior AI/ML Engineer & Data Scientist specializing in LLMs, RAG, and MLOps
“ML/NLP practitioner who has delivered production systems in regulated domains, including a healthcare compliance pipeline using RAG (GPT-4/Claude) plus TF-IDF retrieval that increased document review throughput 4.5x. Also has hands-on experience improving fraud detection data quality via entity resolution (Levenshtein, Dedupe.py) validated with A/B testing, and building scalable, monitored workflows with Airflow, CI/CD, and AWS SageMaker.”
Mid-level AI/ML Engineer specializing in MLOps, NLP, and Generative AI
“Built and deployed a production LLM-powered text-to-SQL/document intelligence chatbot on AWS that lets non-technical business users query complex enterprise databases in plain English. Demonstrates deep practical expertise in schema-aware prompting, embeddings-based schema retrieval, SQL safety/validation guardrails, and rigorous offline/online evaluation with human-in-the-loop approvals for risky queries.”
Senior Software Developer specializing in backend, distributed systems, and IoT
“Backend engineer who built a production retrieval-augmented narrative analysis platform for 100-page screenplays using a Node/Express orchestrator and a Python/FastAPI AI engine, including a key redesign from disk-based uploads to in-memory streaming to eliminate Windows file-lock failures. Also led a refactor of a municipal vehicle tracking system into a C-based distributed engine handling 4M+ daily packets with 99.99% data integrity and automation that reduced manual ops by 50%.”
Mid-level Data Scientist specializing in Generative AI and LLMOps
“Built a production-grade, semi-automated document recognition and classification system for large volumes of scanned PDFs, starting from little/no labeled data and handling highly variable scan quality. Deployed on AWS using SageMaker + Docker and orchestrated on EKS with a microservices design that scales CPU-heavy OCR separately from GPU inference, with strong reliability controls (validation, fallbacks, retries, readiness probes).”
Mid-level Software Engineer specializing in full-stack development, data engineering, and GenAI
“Built and deployed an LLM product called "Content Craft" combining BART-based summarization with a RAG Q&A chatbot using LangChain, embeddings, and a vector database. Has hands-on MLOps experience containerizing and serving models with FastAPI and running them on Kubernetes with monitoring, self-healing, and autoscaling, and has practical experience reducing hallucinations through structured prompting.”
Junior AI Data Engineer specializing in Azure Databricks lakehouse and GenAI RAG systems
“Backend/applied AI engineer from Cloud Rack Systems who built production GenAI/RAG and data platforms on Azure/Databricks at enterprise scale (2.5M records/day). Known for making LLM systems behave like deterministic services via strict retrieval contracts, citation-based validation, and strong observability—shipping a knowledge assistant used daily by 50+ users while driving hallucinations near zero and materially improving latency and cost.”
Mid-Level Software Engineer specializing in distributed systems and AI agent workflows
“Software engineer with enterprise CPQ/CRM/ERP integration experience (Argano) who owned an end-to-end pricing preview capability deployed on AWS Kubernetes with Jenkins CI/CD and full observability (Prometheus/Grafana). Also built an AI-native research agent using LangChain + Chroma to filter academic papers, reporting ~15 hours/week saved for a professor.”
Mid-level AI Engineer specializing in RAG, conversational AI, and agentic systems
“Built and deployed a production RAG-based clinical decision support assistant at MedLib, focused on fast, trustworthy answers from large medical documents. Demonstrates deep practical experience improving retrieval accuracy (semantic chunking + metadata-aware search), controlling hallucinations with grounded generation and thresholds, and adding clinician-requested citations using chunk metadata, with evaluation driven by healthcare professional review.”
Mid-level Full-Stack AI Engineer specializing in deployed LLM agents and RAG systems
“Built a real-time AI meeting assistant using a Chrome extension that streams audio to a backend LLM workflow with transcription and RAG, then hardened it for production with queue-based streaming, async pipelines, security controls, and full observability. Also has hands-on startup sales experience, partnering with customers to define measurable technical win conditions (latency/accuracy) to close deals and drive adoption.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and MLOps
“LLM engineer/data analyst who built a production RAG QA assistant over the Jurafsky & Martin NLP textbook to reduce hallucinations and provide explainable, source-grounded answers. Experienced with LangChain/LangGraph orchestration, retrieval optimization (embeddings, vector DBs, caching), and rigorous evaluation/monitoring (Retrieval@K, A/B tests, telemetry/drift). Previously communicated analytics insights to non-technical stakeholders at GS Analytics using Power BI and simplified reporting.”