Pre-screened and vetted.
Mid-level AI Engineer specializing in LLMs, RAG, and content automation
“AI/LLM engineer who built a production autonomous GenAI content ecosystem that generates short-form scripts, extracts viral highlights from long-form video, and dubs content into 33+ languages. Focused on making LLM outputs production-safe via schema enforcement, token-to-time alignment, critic-agent verification, and scalable async orchestration—cutting manual workflows by ~90% and saving $200k+ annually.”
Junior Data Scientist specializing in healthcare ML and clinical NLP/LLMs
“Healthcare-focused LLM engineer who has built two production clinical applications: an automated structured clinical report generator from physician-patient conversations and a RAG-based chatbot for retrieving patient history (procedures, allergies, etc.). Demonstrates strong applied RAG expertise (overlapping chunking, entity dependency graphs, temporal filtering, graph RAG) to reduce hallucinations/omissions and partners closely with clinicians to automate hospital workflows.”
Mid-level Machine Learning Engineer specializing in GenAI, LLMs, and real-time ML systems
“Built and deployed a production long-form article summarization system using BART/T5/PEGASUS, tackling real-world constraints like token limits, latency/quality tradeoffs, and factual drift via chunking/merge logic and constrained decoding. Uses pragmatic Python-based pipeline orchestration (scheduled jobs, modular scripts, logging/retries) and iterates with stakeholder feedback to make outputs genuinely useful for content workflows.”
Intern AI/ML Engineer specializing in LLMs, MLOps, and distributed training
“Founding AI engineer (June 2024) at Talon Labs who built and productionized an LLM-powered chatbot for interacting with proprietary supply-chain documents, deployed at large scale (25–100,000 users). Experienced with RAG/LLM orchestration (LangChain, LlamaIndex, Groq AI) and production ops tooling (Kubernetes, Docker, Kubeflow, Airflow), with a metrics-driven approach to evaluation, observability, and stakeholder alignment.”
Mid-level Generative AI Engineer specializing in LLM apps, RAG, and MLOps
“LLM/GenAI engineer with US Bank experience building a production financial-document intelligence platform using LangChain/LangGraph, GPT-4, and Amazon OpenSearch. Delivered a RAG-based assistant for compliance/audit teams with grounded, cited answers, focusing on reducing hallucinations and latency, and deployed securely on AWS (SageMaker/EKS) with CI/CD and evaluation tooling (LangSmith, RAGAS).”
Mid-level Data Analyst specializing in procurement, supply chain analytics, and applied machine learning
“Strategic sourcing professional specializing in seasonal apparel supply chains, combining Coupa/JD Edwards analytics with Excel/Python modeling and Power BI dashboards to drive cost reduction and OTIF gains. Notable for rapid mitigation of a 10-day factory delay affecting 12 holiday SKUs (preserved 95% of revenue) and for automating PO workflows to cut cycle time by 4.2 days and improve OTIF by 15%.”
Mid-level AI/ML Engineer specializing in NLP, computer vision, and Generative AI
“Built and deployed a production LLM-powered clinical insights/summarization assistant for healthcare teams, including a Spark+Airflow pipeline, fine-tuned transformer models, and a FastAPI Docker service on AWS. Demonstrates strong MLOps/LLMOps depth (Airflow on Kubernetes, custom AWS operators/IAM, MLflow, CloudWatch) and practical reliability work like hallucination mitigation, confidence scoring, and retrieval-backed evaluation with shadow deployments.”
Mid-level ML/AI Engineer specializing in NLP, RAG pipelines, and financial risk & fraud systems
“Built and shipped LLM/RAG systems in finance and startup settings, including a Goldman Sachs document intelligence platform that indexed ~8TB of regulatory filings and delivered cited, conversational answers with <2s latency—cutting compliance research by ~4.5 hours per batch. Also developed LangChain-based agent workflows at Finta to automate CRM enrichment and investor lookup with strong testing, tracing (LangSmith), privacy guardrails, and auditability.”
Junior Full-Stack & AI/ML Engineer specializing in LLMs and multimodal document processing
“Built a production RAG-based NBA player scouting assistant that embeds player profiles into FAISS, orchestrates retrieval and LLM recommendations with LangChain, and surfaces results via embedded Tableau dashboards. Demonstrates strong focus on evaluation/monitoring (batch tests, LLM-as-judge, latency/failure/token metrics) and has experience translating non-technical founder goals into DAPT + fine-tuning plans on curated data.”
Mid-level AI/ML Engineer specializing in MLOps, LLMs, and real-time inference in FinTech
“ML/LLM engineer who has deployed a production LLM-powered assistant for intent classification and query routing (order recommendation/support deflection), combining BERT fine-tuning with an embedding-based retrieval layer and optimizing for low-latency inference. Experienced with end-to-end reliability practices—Airflow-orchestrated ETL, data validation/alerting, MLflow experiment tracking, and iterative improvements driven by user feedback and monitoring.”
Mid-level Data Engineer specializing in cloud data platforms, Spark, and streaming pipelines
“Data/MLOps engineer (Cognizant background) who owned an AWS/Airflow/Snowflake healthcare transactions pipeline processing ~8–10M records/day and cut pipeline/data-quality incidents by ~33%. Also built and deployed a production FastAPI model-inference service on Kubernetes (Docker, HPA) with strong observability (Prometheus/Grafana), versioned endpoints, and resilient backfill/idempotent external data ingestion patterns.”
“Built and deployed a production RAG-based internal knowledge assistant that let analysts query company documents in natural language, using LangChain/LangGraph with Pinecone and a FastAPI service for integration. Emphasizes reliability in production through hallucination mitigation (retrieval tuning + prompt guardrails) and measurable evaluation/monitoring (accuracy, latency, task completion, hallucination rate), iterating based on user feedback.”
Mid-level Full-Stack Developer specializing in AI-powered analytics platforms
“Backend/DevOps engineer pivoting into robotics/space, building hands-on ROS2 (Humble) skills via Gazebo simulations and experimenting with Nav2 and slam_toolbox. Brings strong distributed-systems and real-time debugging practices (profiling, instrumentation, QoS/retry patterns) and is actively learning perception and control fundamentals to transition into autonomous robotics.”
Mid-level Generative AI & Machine Learning Engineer specializing in agentic LLM systems
“Built and deployed a production agentic LLM knowledge assistant that answers complex questions over internal documents, APIs, and databases using a RAG architecture (FAISS/Pinecone) and LangChain/LangGraph orchestration. Emphasizes production-grade reliability and hallucination control through grounding, confidence thresholds, validation, retries/fallbacks, and full observability (logging/metrics/traces) with continuous evaluation and feedback loops.”
Mid-level Data Scientist specializing in Generative AI, MLOps, and cloud data platforms
“GenAI/ML engineer (CitiusTech) who has deployed production RAG systems for compliance/operations document Q&A, using Pinecone + FastAPI microservices on Kubernetes with strong monitoring and guardrails. Also built a GenAI-powered incident triage/routing solution in collaboration with non-technical stakeholders, achieving 35% faster response times and 40% fewer misclassified tickets, and has hands-on orchestration experience with Airflow and AutoSys.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and multimodal generation
“Open-source JavaScript contributor focused on performance and maintainability in data visualization libraries—refactored legacy ES5 into modular ES6, added tests/docs, and delivered ~30% faster load times with positive community adoption. Also optimized a React dashboard (~40% load-time reduction) and took ownership in an ambiguous AI product initiative by setting milestones, standing up an initial ML pipeline, and shipping a prototype in ~6 weeks that became the basis for production.”
Mid-level AI/ML Engineer specializing in LLM agents, RAG, and ML systems
“At Inertia Systems, built a production LLM-powered ingestion pipeline that converts heterogeneous sources (PDF/JSON/IFC/SQL and financial tables) into standardized text and uses GraphRAG to construct a knowledge graph with verified dependency relationships. Also has hands-on HPC orchestration experience with SLURM, including creating a custom wrapper process manager to improve resource utilization under restrictive scheduling policies.”
Junior AI/ML Engineer specializing in LLM applications and RAG systems
“Built and deployed LLM-powered agentic systems including a multi-agent travel planning assistant using LangChain, RAG (FAISS), real-time APIs, and a supervisor agent to manage coordination and reduce hallucinations. Also developed a Text-to-SQL system with schema-aware validation guardrails, and collaborated with drilling domain experts at CNPC USA to build an ML model predicting rate of penetration (ROP).”
“ML/GenAI engineer with recent CVS Health experience building a production RAG system over unstructured financial/research documents using LangChain, FAISS, and Pinecone, plus LoRA/PEFT fine-tuning of GPT/LLaMA for domain-aware summarization. Demonstrates strong applied MLOps and data engineering skills (Airflow/Prefect, Docker/Kubernetes, CI/CD, MLflow) and measurable impact (sub-second retrieval, ~40% better context retrieval, ~25% entity matching improvement).”
Mid-level Data Scientist & AI Engineer specializing in RAG, agentic AI, and production ML
“AI/data engineer who built a production LLM-powered schema drift detection system (LangChain/LangGraph) to catch semantic data changes before they break downstream analytics/ML. Deployed on AWS with Docker/S3 and implemented an LLM-as-a-judge evaluation framework to improve trust, reduce hallucinations, and control false positives/alert fatigue. Collaborated with non-technical risk/business analytics stakeholders at EY by delivering human-readable drift explanations that improved confidence in financial analytics dashboards.”
Senior Data & Backend Engineer specializing in cloud data pipelines and LLM/RAG systems
“Data engineer with end-to-end ownership of large-scale retail and clinical data ingestion/processing on AWS, including real-time streaming and batch pipelines. Delivered measurable outcomes: 20M daily transactions processed, latency cut from 4 hours to 5 minutes, ~70% fewer failures, and 120+ pipelines running at 99.8% reliability with full audit compliance.”
Senior AI/ML Engineer specializing in Generative AI, RAG, and agentic systems
“GenAI/LLM ML engineer (currently at Webprobo) building an enterprise GenAI platform with document intelligence and automation on AWS and blockchain. Has hands-on experience with RAG, LLM evaluation tooling, and orchestrating production LLM workflows with Apache Airflow, plus deep exposure to reliability challenges in globally distributed/edge deployments. Also partnered with business/marketing stakeholders at a banking client to deliver an AI-driven customer retention insights solution.”
Senior Full-Stack & GenAI Engineer specializing in healthcare and financial services
“Built and deployed a production LLM-powered customer support assistant using a RAG backend in Python, focused on deflecting repetitive Tier-1 tickets and reducing resolution time. Demonstrates strong production engineering instincts around reliability (confidence scoring + human fallback), scalability/cost optimization (multi-stage pipelines), and workflow orchestration/observability (LangChain, custom DAGs, structured logging, step metrics).”
Senior Software Engineer specializing in AI-driven marketing and data platforms
“Backend/data engineer who builds production FastAPI microservices and AWS serverless/Glue pipelines for SMS analytics and marketing segmentation. Led a legacy batch modernization into modular services (FastAPI + Glue/Athena + ClickHouse) using shadow-mode parity checks, feature flags, and incremental rollout. Demonstrated measurable performance wins (12s to sub-second SQL; ~40% CPU reduction) and strong incident ownership with proactive schema-drift prevention.”