Pre-screened and vetted.
Junior Full-Stack AI Engineer specializing in LLMs and RAG systems
Junior Software Engineer specializing in APIs, data pipelines, and LLM/RAG systems
Mid-level Applied AI Engineer specializing in LLMs, Prompt Engineering, and RAG
Mid-level AI Engineer specializing in Generative AI, LLMs, and RAG on AWS
“Built and deployed an LLM-powered clinical decision support and risk monitoring platform for mental health at Valuai.io, emphasizing low-latency, evidence-grounded responses and crisis-safe behavior with clinician escalation. Strong production agent-orchestration background (LangChain/CrewAI) plus rigorous evaluation (clinician-in-the-loop + evaluator agent) and large-scale synthetic testing; also applied multi-agent workflows to document verification and fraud detection during an AI internship at Nixacom.”
Junior AI/ML Software Engineer specializing in LLM agents and RAG systems
“AI/back-end engineer at Canon who helped build and operate an internal production LLM platform that acts as a secure middle layer between users and models, defending against jailbreaks/prompt injection while enabling RAG, memory, and grounded responses over company data. Experienced with LangChain/LangGraph orchestration, vector DB retrieval, and reliability practices (testing, monitoring, adversarial prompts) to run high-throughput, low-latency AI workflows in production.”
Junior AI Integration Engineer specializing in LLM agents and RAG on cloud platforms
“Built and deployed LLM-powered features for a startup organizational management application, focusing on real-world deployment constraints like latency and cost. Implemented RAG with FAISS and improved retrieval quality by switching embedding models (OpenAI/Hugging Face) and fine-tuning embeddings on medical corpora for a medical-report UI feature. Uses LangChain and LangGraph to orchestrate multi-node LLM API workflows and evaluates systems with metrics like latency, cost per request, and error taxonomy.”
Senior AI/ML Engineer & Data Scientist specializing in LLMs, RAG, and MLOps
“ML/NLP practitioner who has delivered production systems in regulated domains, including a healthcare compliance pipeline using RAG (GPT-4/Claude) plus TF-IDF retrieval that increased document review throughput 4.5x. Also has hands-on experience improving fraud detection data quality via entity resolution (Levenshtein, Dedupe.py) validated with A/B testing, and building scalable, monitored workflows with Airflow, CI/CD, and AWS SageMaker.”
Mid-level Data Scientist specializing in Generative AI and Healthcare Analytics
“Built a LangGraph-based, tool-routing LLM chatbot to deliver fast, trustworthy investment-stock insights (including tariff impact) and deployed it to production on Snowflake after initially developing in Azure with AI Search and the Microsoft Agent Framework. Improved routing robustness by moving from LLM-based decisions to a deterministic router backed by schema-relationship graphs and YAML metadata, and ran the project iteratively with non-technical stakeholders over an 8-month engagement.”
Junior AI/ML & Mobile Engineer specializing in LLMs, synthetic data, and React Native
“Currently at Uplift AI shipping production LLM features that generate personalized growth insights from user reflections using BERT + embeddings + RAG, with strong safety/guardrail practices for sensitive contexts. Also built an end-to-end React Native UGC challenge submission/moderation system that improved repeat submissions and 7-day retention, and has applied rigorous clinical-style evaluation methods on a dental X-ray disease detection project to reduce false negatives.”
Mid-level Data Scientist specializing in Generative AI and LLMOps
“Built a production-grade, semi-automated document recognition and classification system for large volumes of scanned PDFs, starting from little/no labeled data and handling highly variable scan quality. Deployed on AWS using SageMaker + Docker and orchestrated on EKS with a microservices design that scales CPU-heavy OCR separately from GPU inference, with strong reliability controls (validation, fallbacks, retries, readiness probes).”
Mid-level Software Engineer specializing in full-stack development, data engineering, and GenAI
“Built and deployed an LLM product called "Content Craft" combining BART-based summarization with a RAG Q&A chatbot using LangChain, embeddings, and a vector database. Has hands-on MLOps experience containerizing and serving models with FastAPI and running them on Kubernetes with monitoring, self-healing, and autoscaling, and has practical experience reducing hallucinations through structured prompting.”
Junior AI Data Engineer specializing in Azure Databricks lakehouse and GenAI RAG systems
“Backend/applied AI engineer from Cloud Rack Systems who built production GenAI/RAG and data platforms on Azure/Databricks at enterprise scale (2.5M records/day). Known for making LLM systems behave like deterministic services via strict retrieval contracts, citation-based validation, and strong observability—shipping a knowledge assistant used daily by 50+ users while driving hallucinations near zero and materially improving latency and cost.”
Mid-level AI Engineer specializing in RAG, conversational AI, and agentic systems
“Built and deployed a production RAG-based clinical decision support assistant at MedLib, focused on fast, trustworthy answers from large medical documents. Demonstrates deep practical experience improving retrieval accuracy (semantic chunking + metadata-aware search), controlling hallucinations with grounded generation and thresholds, and adding clinician-requested citations using chunk metadata, with evaluation driven by healthcare professional review.”
Intern AI & Machine Learning Engineer specializing in computer vision and edge deployment
“Built and shipped a real-time AI robotic inspection system, using a synthetic data generation pipeline to address rare edge cases—cutting data collection costs ~60% and boosting hard-scenario accuracy ~20%. Experienced in productionizing ML on constrained Jetson hardware and orchestrating end-to-end ML workflows with Airflow/Docker/Kubernetes, with a metrics-driven approach to reliability, evaluation, and stakeholder communication.”
Mid-level AI Engineer specializing in Generative AI, LLMs, and RAG
“Internship at Discovery Education building a production LLM/RAG chatbot that let marketing and sales teams query and interpret Looker/BI dashboards in natural language, with responses grounded in compliance and state education standards. Emphasizes rigorous evaluation (faithfulness/precision/recall/latency) plus user-feedback analytics, and used LangChain for orchestration, chunking/context-window control, and integration with enterprise sources like SharePoint.”
Senior Data Scientist / AI Engineer specializing in LLMs, RAG, and production ML
“Data science professional who has built a production RAG-based LLM question-answering system ("Flash Query") to deliver fast, accurate answers over large document collections, focusing on retrieval quality and grounded responses. Also collaborates with non-technical retail/jewelry stakeholders to turn business questions into predictive models and dashboards for decision-making.”
Senior Software Engineer specializing in Backend Systems and Generative AI (RAG)
“Backend engineer with experience building an end-to-end civic tech AI platform that ingests city council meeting videos, transcribes them with Whisper, and enables natural-language Q&A via a LangChain/FAISS RAG pipeline. Demonstrated strong systems thinking by tuning retrieval for accuracy/latency/memory (cutting response time ~3s→1s and memory ~500MB→25MB) and by safely migrating an ERP from monolith toward services using dual writes, reconciliation, and idempotency to protect financial workflows.”
Junior Full-Stack/AI Engineer specializing in web platforms and LLM applications
“Backend engineer from FoodSupply.ai who built and evolved a scalable restaurant/supplier product and order management platform using Node.js and REST APIs. Implemented a hybrid MySQL+MongoDB data architecture, optimized performance with Redis/Prisma, and led a phased migration with feature flags and a temporary sync layer to maintain data consistency. Strong focus on production security (OAuth2, RBAC, row-level security, AWS IAM) and reliability practices (testing with Pytest, Docker/AWS pipelines).”
Mid-level GenAI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production RAG-based LLM assistant that answers day-to-day operational questions from internal PDFs/SOPs, with strong emphasis on data consistency (metadata versioning, confidence thresholds, conflict handling) and low-latency retrieval at scale. Experienced designing and orchestrating multi-agent LLM workflows (retrieval/validation/generation) and pipeline orchestration for ingestion/embedding/vector-store updates, plus iterative delivery with non-technical operations/business stakeholders.”
Junior Machine Learning & Backend Engineer specializing in LLM systems and ML infrastructure
“Built and deployed production RAG-based document search/Q&A systems (DocChat and an internship marketing RAG), using a React + FastAPI stack on GCP with docs stored in GCP buckets and retrieval via embeddings/vector DB. Emphasizes cost/performance tradeoffs (reported ~40% cost reduction) and ships via Docker (Railway), with load/API testing using JMeter and Swagger; regularly collaborates with a CEO stakeholder to iterate and push changes to production.”
Junior Software Engineer specializing in full-stack web development and test automation
“Full-stack engineer who built and owned a production workflow/kanban-style drag-and-drop system in Next.js (App Router) with Postgres/Prisma, including reusable component abstractions, Cypress E2E coverage, and post-launch performance/bug ownership. Notable for measurable impact (25% faster UI dev, ~30% query perf improvement) and for leading an incremental Express→NestJS migration that reduced technical debt (~40%) through better structure, docs, and team enablement.”