Pre-screened and vetted.
Junior AI/ML Software Engineer specializing in LLM agents and RAG systems
“AI/back-end engineer at Canon who helped build and operate an internal production LLM platform that acts as a secure middle layer between users and models, defending against jailbreaks/prompt injection while enabling RAG, memory, and grounded responses over company data. Experienced with LangChain/LangGraph orchestration, vector DB retrieval, and reliability practices (testing, monitoring, adversarial prompts) to run high-throughput, low-latency AI workflows in production.”
Junior Data Engineer specializing in Azure, CRM data pipelines, and marketing personalization
“LLM/AI engineer who has deployed production RAG conversational analytics and Text-to-SQL systems over Snowflake and curated data marts, emphasizing enterprise-grade guardrails for accuracy, security, and cost. Notable for a structured approach to reducing hallucinations (curated metric/table registry, SQL validation, RBAC, and citation-backed responses) and for building resilient, observable multi-step agent workflows using LangChain/LlamaIndex and Airflow.”
Junior Data & AI Engineer specializing in cloud AI and analytics
“Built production AI backend systems in healthcare and e-commerce, including a healthcare agent that automated clinical workflows like medication refills, immunizations, and scheduling using FHIR APIs and cloud-native infrastructure. Strong in end-to-end backend ownership, LLM orchestration, and adding guardrails/validation for high-stakes and customer-facing AI workflows.”
Mid-level Data Scientist specializing in Generative AI and Healthcare Analytics
“Built a LangGraph-based, tool-routing LLM chatbot to deliver fast, trustworthy investment-stock insights (including tariff impact) and deployed it to production on Snowflake after initially developing in Azure with AI Search and the Microsoft Agent Framework. Improved routing robustness by moving from LLM-based decisions to a deterministic router backed by schema-relationship graphs and YAML metadata, and ran the project iteratively with non-technical stakeholders over an 8-month engagement.”
Entry AI Engineer specializing in LLMs, RAG, and MLOps
“Built and shipped a production Python-based agentic RAG document retrieval system over 80K records using FastAPI, OCR, vector search, and AWS infrastructure, with a strong emphasis on reliability, testing, and observability. Stands out for treating AI failures like production incidents—turning hallucinations, retrieval misses, and OCR issues into regression tests—and for quantifiably reducing document lookup time from about 12 minutes to under 90 seconds.”
Mid-level AI Engineer and Data Scientist specializing in LLM agents and RAG systems
“Built a production-grade LLM evaluation and regression system that stress-tests models across hundreds of iterations, combining LLM-as-judge, semantic similarity, statistical metrics, and rule-based checks, with results delivered via stakeholder-friendly HTML reports and dashboards. Experienced orchestrating multi-agent RAG workflows using LangChain/LangGraph and event-driven GenAI pipelines in n8n integrating OCR, speech-to-text, and external APIs, with strong emphasis on reliability, observability, and explainable failures.”
Mid-level AI Engineer specializing in Generative AI and LLM systems
“Built and deployed a production-grade, multi-agent Text-to-SQL assistant that lets non-technical stakeholders query large enterprise databases in natural language. Uses Pinecone-based schema retrieval + LLM reasoning (Gemini/Claude/GPT) with a dedicated validation agent (schema/syntax checks and safe dry runs) to reduce hallucinations and improve reliability, while optimizing latency and cost via async execution and embedding caching.”
Junior AI Data Engineer specializing in Azure Databricks lakehouse and GenAI RAG systems
“Backend/applied AI engineer from Cloud Rack Systems who built production GenAI/RAG and data platforms on Azure/Databricks at enterprise scale (2.5M records/day). Known for making LLM systems behave like deterministic services via strict retrieval contracts, citation-based validation, and strong observability—shipping a knowledge assistant used daily by 50+ users while driving hallucinations near zero and materially improving latency and cost.”
Director-level growth marketer specializing in international SaaS, e-commerce, and AI-native acquisition
“Founder of distribb.io, an 8-month-old AI SEO software company already at $10k MRR. Brings 15 years of startup ecosystem exposure and a pragmatic, market-driven approach to building businesses by targeting proven demand rather than chasing novelty.”
“Built a production ad-spend optimization system that combined deterministic audit logic with LLM-generated explanations, surfacing severe inefficiencies including 70-90% wasted spend in some Google Ads accounts. Stands out for pairing measurable business impact with pragmatic AI safety and usability decisions, including approval-gated execution and structured, human-readable recommendations.”
Mid-level Software Engineer specializing in backend systems and AI-powered platforms
“Backend engineer who built a production retrieval-augmented narrative analysis platform for 100-page screenplays using a Node/Express orchestrator and a Python/FastAPI AI engine, including a key redesign from disk-based uploads to in-memory streaming to eliminate Windows file-lock failures. Also led a refactor of a municipal vehicle tracking system into a C-based distributed engine handling 4M+ daily packets with 99.99% data integrity and automation that reduced manual ops by 50%.”
Mid-level AI Engineer specializing in Python, LLMs, and production ML systems
“Production-focused ML/AI engineer with hands-on ownership across classical ML and GenAI systems, from CV/NLP services to enterprise RAG. Stands out for combining research-to-production execution with measurable business impact: 40% processing-efficiency gains, 35% fewer support tickets, 5x latency improvement, and 3x throughput gains while maintaining safety and quality.”
Mid-level AI Engineer specializing in NLP, computer vision, and healthcare analytics
“Data scientist who has built production LLM agents (GPT-4o + LangChain + RAG) to automate analyst-style ad hoc CSV analysis with guardrails and GPT-as-a-judge evaluation. Also delivered an explainable healthcare NLP system for ICD code classification by collaborating closely with clinicians, using a hybrid rule-based decision tree + BERT model to reach 97% accuracy and cut manual review time.”
Mid-level AI Engineer specializing in Generative AI, LLMs, and RAG
“Internship at Discovery Education building a production LLM/RAG chatbot that let marketing and sales teams query and interpret Looker/BI dashboards in natural language, with responses grounded in compliance and state education standards. Emphasizes rigorous evaluation (faithfulness/precision/recall/latency) plus user-feedback analytics, and used LangChain for orchestration, chunking/context-window control, and integration with enterprise sources like SharePoint.”
Junior Software Engineer specializing in distributed systems and ML platforms
“Built and deployed real-world systems end-to-end across security and healthcare contexts: led a 3-person team delivering a university vehicle tracking system with 30% cost savings and 1-year post-launch monitoring. Also implemented a healthcare RAG chatbot with adaptive query routing that cut LLM costs by 40% while maintaining answer accuracy, and has experience debugging non-deterministic LLM behavior in DevOps pipeline automation.”
Mid-level Full-Stack & AI Engineer specializing in LLM applications
“Full-stack engineer who has shipped and operated generative-AI chat/QA features end-to-end, including a RAG-based pipeline with guardrails and cost/latency monitoring in production. Experienced with React/TypeScript + Node/Postgres architectures, Dockerized deployments to AWS (EC2) via GitHub Actions CI/CD, and building reliable ingestion/ETL systems with idempotency, backfills, and reconciliation.”
Mid-level AI/ML Software Engineer specializing in GPU-optimized LLM inference and cloud microservices
“Built and deployed a production RAG-based multilingual analytics assistant for healthcare operations, enabling non-technical teams to query claims/EHR and risk metrics with grounded explanations. Demonstrates strong end-to-end LLM system engineering (retrieval tuning, re-ranking, hallucination controls, verification layers) plus workflow orchestration (Airflow/Composer/Step Functions) and stakeholder-driven iteration via prototypes and dashboards.”
Mid-level Full-Stack Engineer specializing in AI-powered and cloud-native systems
“Product-minded engineer who has owned features end-to-end, including a full onboarding redesign that lifted completion ~25% and a production LLM/RAG report-generation system with strong guardrails (schema-constrained JSON, confidence gating, logging) and an automated eval/regression loop built from real user queries. Also built a scalable research data pipeline ingesting messy PDFs/JSON/CSVs with normalization, idempotent reruns, observability, and cost/latency tradeoffs.”
Senior Full-Stack Software Engineer specializing in cloud-native web, mobile, and AI features
“Frontend lead for a consumer-facing social platform, owning architecture through release. Built scalable React/TypeScript systems (Redux Toolkit, Remix) with a shared Storybook component library and strong quality gates (CI, Jest/Cypress). Experienced modernizing legacy codebases incrementally with feature flags and shipping major dashboard features with staged rollouts and close QA collaboration.”
Junior AI/ML Software Engineer specializing in Generative AI and scalable data pipelines
“Built and operated large-scale biodiversity/ecological research platforms, integrating 50+ heterogeneous global datasets into a unified BIEN 3 schema on PostgreSQL/PostGIS and improving data consistency by 35%. Strong production engineering background (Linux monitoring, CI/CD performance gates, Docker on AWS/Azure) plus applied AI work building a Python RAG system (0.90 precision) and halving latency with Elasticsearch.”
Mid-level GenAI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production RAG-based LLM assistant that answers day-to-day operational questions from internal PDFs/SOPs, with strong emphasis on data consistency (metadata versioning, confidence thresholds, conflict handling) and low-latency retrieval at scale. Experienced designing and orchestrating multi-agent LLM workflows (retrieval/validation/generation) and pipeline orchestration for ingestion/embedding/vector-store updates, plus iterative delivery with non-technical operations/business stakeholders.”
Junior Data/AI Engineer specializing in MLOps, real-time pipelines, and LLM applications
“Built an LLM-driven MLOps agent at SBD Technologies that automated an EV-charging prediction workflow end-to-end, integrating with real-time Kafka/FastAPI systems supporting 120K+ chargers at 99.99% event delivery. Addressed frequent schema drift by implementing SQLAlchemy/Flyway validation (60% reduction in drift issues) and deployed as Kubernetes microservices with GitHub Actions CI/CD; also has Airflow-based ingestion/crawling experience into Snowflake and stakeholder-facing delivery via a Fleetcharge PWA.”
Mid-level Python Full-Stack Engineer specializing in AI microservices and cloud data platforms
“Backend-leaning full-stack engineer in fintech/payments who shipped an end-to-end Stripe payments + webhook system for a financial microservices platform, emphasizing ledger accuracy via idempotency, transactional writes, retries, and DLQs. Also delivered a real-time React/TypeScript payment status dashboard informed by user interviews, and improved production performance by 35% p95 latency through PostgreSQL tuning and Redis caching on AWS.”
Intern Machine Learning Engineer specializing in Generative AI and RAG systems
“Early-career AI/LLM builder who created and deployed a multi-agent news analysis agent (Patrakarita) using CrewAI, coordinating researcher/analyst roles to turn noisy article URLs into structured, prioritized outputs (claims, tone, verification questions, opposing views). Strong focus on orchestration debugging and reliability evaluation, including measuring hallucination/redundancy and improving reasoning by refactoring pipeline sequencing.”