Pre-screened and vetted.
Mid-level Software/Data Engineer specializing in LLM apps, RAG pipelines, and cloud microservices
“Backend/data engineer who built an enterprise LLM assistant (AI Genie) at Broadband Insights using a LangChain + GPT-4 + Pinecone RAG pipeline to automate broadband analytics reporting. Developed Python/Dagster ETL processing 10M+ records/day and improved data freshness by 60%, with production-grade scalability patterns (async workers, containerized microservices, Kubernetes) and strong multi-tenant isolation practices.”
Principal DevOps Architect specializing in cloud platform engineering and SRE
“End-to-end engineer focused on AI-native enterprise systems, including a production generative knowledge platform using RAG + semantic search over internal documentation (React, Python/Flask, GPU-hosted NLP models, Pinecone) with strong CI/CD and observability. Reports concrete outcomes including 40% faster knowledge access and ~75% employee adoption, and has led incremental cloud-native modernization using feature flags, parallel runs, canary releases, and regression testing.”
Junior Full-Stack AI Developer specializing in LLMs and RAG applications
“Product-minded software engineer who owned a Shopify POS app end-to-end at Swym, shipping an MVP and then scaling iteration speed with E2E automation and CI/CD—resulting in a Shopify Badge, Top-5 App Store ranking, and +40% new user acquisition. Also built an ESG insights tool using React/TypeScript + FastAPI with Snowflake and a RAG pipeline, plus microservices patterns (async jobs, queues, DLQs, autoscaling) and internal Metabase/SQL analytics dashboards.”
Senior AI/ML Engineer specializing in LLMs, RAG, and VR/XR multimodal systems
“PhD researcher (University of Utah) who built a production RAG-powered Virtual Reality Research Assistant to answer lab research questions with concrete citations. Implemented an end-to-end LangChain pipeline using PyPDFLoader, chunking strategies, OpenAI embeddings, and ChromaDB, with emphasis on grounding to reduce hallucinations and ensure research-grade accuracy. Collaborated closely with a non-technical PhD advisor to scope requirements, manage cost constraints, and demo iterative progress.”
Senior Full-Stack & AI Engineer specializing in scalable web platforms and LLM automation
“Built a production agentic AI assistant in Python using Playwright plus Google Gemini’s vision capabilities to automatically document and execute UI workflows step-by-step, reducing developer time spent on trivial documentation/knowledge transfer. Also built an Apache Airflow ETL pipeline and has experience evaluating AI agents with human-in-the-loop methods, plus successfully communicated a vision-model-based CMS analytics PoC to non-technical university stakeholders and proposed it to Academic Technology with cost-savings rationale.”
Mid-level AI/ML Engineer specializing in Generative AI and LLM systems
“Senior AI/ML engineer with hands-on experience building production LLM systems in healthcare, including RAG-based clinical question answering and end-to-end MLOps on Vertex AI and Kubernetes. They combine strong platform engineering with applied GenAI work, citing a 35% improvement in factual accuracy and a 30% boost in internal team productivity through modular Python services and CI/CD.”
Senior Machine Learning Engineer specializing in LLMs, computer vision, and cloud AI
“Healthcare-focused ML/AI engineer who has built clinical note summarization and medical image annotation solutions using LLMs, RAG, and multimodal models. They combine experimentation across major model providers with practical production concerns like monitoring, drift detection, and latency/cost tradeoffs, and also earned 2nd place in a Google hackathon for a medical AI assistant.”
Junior Software Engineer specializing in backend systems and AI data pipelines
“Backend engineer with fintech/AI startup experience who built an Azure serverless, event-driven pipeline for large-scale crypto sentiment analysis and semantic search (OCR/NLP to vector search) and integrated LLM + blockchain data for predictive insights. Demonstrated measurable impact (25% lower retrieval latency, 10% fewer data errors, 15% higher engagement) and has led safe microservices migrations with strong security and reliability practices.”
Mid-level AI Engineer specializing in LLM apps, RAG pipelines, and multi-agent systems
“AI Engineer at Humanitarian AI who has built and productionized both a LangGraph-based multi-agent workflow system and a RAG pipeline (OpenAI embeddings + vector DB) with rigorous evaluation/guardrails. Reports strong measurable impact (60% faster workflow delivery, 40% fewer incidents, 70% reduced research time) and has prior enterprise modernization experience at Infosys migrating ETL to microservices with zero production incidents.”
Junior Machine Learning Engineer specializing in MLOps and real-time systems
“Built and shipped a production GPT-4 + RAG customer support chatbot that materially improved support operations (response time 4 hours to <3 minutes; ~65% tier-1 ticket automation). Demonstrates strong end-to-end LLM engineering across retrieval (Sentence Transformers/Pinecone), safety (multi-layer moderation), cost/latency optimization (caching/streaming, Celery/Redis), and rigorous evaluation/monitoring (shadow deploys, Datadog, 500+ test cases), plus proven stakeholder buy-in leading to 80% adoption.”
Intern Software Engineer specializing in backend systems and Generative AI
“Built and deployed a scalable, production-ready LLM knowledge assistant using a RAG architecture (LangChain + vector store/FAISS) to replace keyword search for internal documents. Demonstrates hands-on expertise in hallucination reduction and retrieval quality improvements through semantic chunking, similarity tuning, prompt design, and human-in-the-loop validation, plus strong stakeholder communication via demos and visual explanations.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and agentic systems
“Built a production "Mini RAG Assistant" for internal document Q&A, focusing on grounded answers (anti-hallucination), retrieval quality, and latency/cost optimization. Uses LangChain/LangGraph for orchestration and applies a metrics-driven evaluation loop (including reranking and semantic chunking improvements) while collaborating closely with product stakeholders.”
Junior Full-Stack Software Engineer specializing in React/Node, cloud, and LLM-powered automation
“Master’s program project lead who built and deployed a real-time sound recognition system (Flask + React Native + ML) that was adopted by 200+ university students. Demonstrates strong production engineering and cross-layer debugging—solving latency, unreliable uploads, and observability gaps using microservice separation, chunked/idempotent transfers, and packet-capture-driven network diagnosis—plus AWS/on-prem and IoT edge-to-cloud integration experience.”
Mid-level GenAI Engineer specializing in RAG, LLM agents, and enterprise automation
“Accenture engineer who built and shipped a production RAG-based automation/chatbot for SAP incident triage and troubleshooting, embedding thousands of runbooks/logs/tickets into a semantic search pipeline and integrating it into Teams/Slack. Reported major productivity gains (30–60% time reduction), >90% validated answer accuracy, and sub-2-second responses, with strong orchestration (Airflow/Prefect/LangGraph) and reliability practices (guardrails, testing, monitoring).”
Senior Data Scientist specializing in LLM applications, RAG systems, and production ML
“Senior Data Scientist in consulting who has built production RAG systems for insurance/annuity document search at large scale (100K+ PDF pages), emphasizing grounded answers, guardrails, and low-latency retrieval. Experienced in end-to-end MLOps for LLM apps—monitoring, evaluation sets, drift handling, and safe rollouts—and in orchestrating complex pipelines with Prefect/Airflow and deploying services on Kubernetes.”
Mid-level Software Engineer specializing in Java/Spring Boot microservices
“Full-stack AI engineer who built Skillmatch AI, an LLM/RAG-based job matching platform using FastAPI microservices, Airflow-orchestrated async pipelines, and Pinecone vector search (sub-second retrieval across 50k+ vectors) deployed on GCP with autoscaling. Also partnered directly with a cancer researcher to automate SEER + PubMed-driven report generation via an AI pipeline, emphasizing rapid prototyping and outcome-focused communication.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and LLM systems
“Built a production multi-agent cybersecurity defense simulator orchestrated with CrewAI, combining Red/Blue team LLM agents, a RAG runbook retriever, and an RL remediation agent trained via state-space simplification and reward shaping for rapid incident response. Also partnered with quant analysts and fund managers to deliver an automated trading and portfolio management system using statistical methods plus CNN/LSTM models, reporting up to 15% weekly ROI.”
Senior Full-Stack Engineer specializing in AI, cloud infrastructure, and DevOps
“Frontend engineer focused on building and scaling data-heavy, real-time dashboards with React/Next.js/TypeScript. Emphasizes performance and reliability at scale through modular architecture, centralized state (Zustand/Redux), strict API contracts, automated testing, and production monitoring (Grafana/CloudWatch), and has experience shipping quickly with feature-flagged rollouts and rapid iteration from user feedback.”
Junior Machine Learning Engineer specializing in LLMs, RAG, and on-device AI
“Built an "Offline Study Assistant" that runs LLM inference locally on a 5-year-old Android device using Llama.cpp and the Android NDK, achieving a 27x speedup and cutting time-to-first-token from 11 minutes to 30 seconds. Also has applied backend/API experience with FastAPI, Supabase (Auth + RLS), and production hardening of a RAG system at Hashmint using Celery and Redis to eliminate PDF-processing-related query failures.”
Mid-level Machine Learning & AI Engineer specializing in Generative AI, NLP, and MLOps
“Built and deployed production LLM systems for summarizing sensitive legal and financial documents, emphasizing GDPR-aligned privacy controls and scalable hybrid cloud architecture. Experienced with Kubernetes/Airflow orchestration and rigorous testing/monitoring practices, and has delivered measurable business impact (18% conversion lift) by translating AI outputs for non-technical marketing stakeholders.”
Mid-level Software/AI Engineer specializing in GenAI, AWS, and microservices
“Built a production AI pipeline at EyCrowd to automatically grade shaky outdoor user-submitted brand videos using CV + CLIP/BLIP and a LangChain RAG layer per brand, with GPT-4 generating structured JSON explanations and grades. Optimized for latency and cost (batch PyTorch inference, caching), cutting review time from ~8 minutes to <2 minutes while reaching ~90% alignment with human graders and supporting thousands of videos/day.”
Mid-level Full-Stack Product Engineer specializing in SaaS and AI search
“Two-time founder and former CTO/CPO who has shipped and operated full-stack products solo, including a real-time community platform (Twitter/Slack-like) with Next.js/TypeScript, WebSockets, Redis, and strong post-launch analytics (PostHog/Sentry). Also built durable multi-step AI-agent workflows using Inngest with state machines, checkpointing, and validation gates, and has hands-on Postgres performance tuning experience validated via EXPLAIN ANALYZE.”
Entry Software Engineer specializing in cloud backend and microservices
“Built production-oriented LLM agent systems for incident investigation and CRM workflows using LangGraph, FastAPI, AWS, and retrieval grounding. Stands out for treating agents like real software systems—adding schema enforcement, retries, fallbacks, monitoring, and eval loops—and tying that work to measurable gains in accuracy, latency, and analysis speed.”
“ML engineer with hands-on experience building banking AI systems end-to-end, including a customer-targeting model that improved campaign response rates by about 10%. Also shipped a RAG-based banking FAQ/support feature with safety guardrails and production optimizations around retrieval quality, latency, and cost, plus reusable Python services that reduced duplicate work for other engineers.”