Pre-screened and vetted.
Principal Software Architect specializing in AI/ML and cloud-native full-stack platforms
“AI/LLM engineer who built a production content-generation system for nursing education, combining multimodal RAG over proprietary PDFs (including images) with structured Cosmos DB data and external sources. Strong focus on production reliability—prompt-chaining with LangChain, validation/guardrails, and Azure-based monitoring/observability—plus experience designing Azure AI agents with tool integrations like Bing Search.”
Mid-level AI Engineer specializing in multi-agent LLM systems and multimodal tutoring
“LLM/agentic systems builder who has deployed multi-agent educational chatbots using LangChain + LangGraph, with LangFuse-based tracing and FastAPI hosting. Focused on production reliability and performance (latency reduction via agent decomposition and caching) and on evaluation/testing (routing test scenarios, LLM-as-judge). Partnered with product to add image understanding by parsing and storing images in S3, expanding chatbot coverage to 30+ books with images.”
Mid-level AI/ML Engineer specializing in GenAI, LLMs, and computer vision
“Built and productionized a multi-agent, LLM-powered document understanding system to replace manual review of long documents, using LangGraph orchestration plus RAG to reduce hallucinations. Implemented layered reliability controls (structured templates, checker agent, and human-in-the-loop feedback) and reported ~40% speed improvement after orchestration; also has hands-on Airflow experience for scheduled data pipelines.”
Mid-level Machine Learning Engineer specializing in GenAI, LLMs, and real-time ML systems
“Built and deployed a production long-form article summarization system using BART/T5/PEGASUS, tackling real-world constraints like token limits, latency/quality tradeoffs, and factual drift via chunking/merge logic and constrained decoding. Uses pragmatic Python-based pipeline orchestration (scheduled jobs, modular scripts, logging/retries) and iterates with stakeholder feedback to make outputs genuinely useful for content workflows.”
Intern AI/ML Engineer specializing in LLMs, MLOps, and distributed training
“Founding AI engineer (June 2024) at Talon Labs who built and productionized an LLM-powered chatbot for interacting with proprietary supply-chain documents, deployed at large scale (25–100,000 users). Experienced with RAG/LLM orchestration (LangChain, LlamaIndex, Groq AI) and production ops tooling (Kubernetes, Docker, Kubeflow, Airflow), with a metrics-driven approach to evaluation, observability, and stakeholder alignment.”
Junior Data Scientist specializing in fraud analytics and cloud data platforms
“Built and deployed production LLM-powered document summarization/classification systems using embeddings, vector databases (RAG-style retrieval), and automated evaluation (BERTScore/ROUGE), with a focus on monitoring and scalable cloud pipelines. Also partnered with a fraud analytics team to deliver a transaction anomaly detection solution, translating model outputs into Power BI dashboards and actionable KPIs while iterating on thresholds and alerts based on stakeholder feedback.”
Senior Software Engineer specializing in Generative AI product development
“AI product builder at Padlet who shipped multiple production LLM features for education workflows, including an AI document generator (AI Recipes) and a RAG-enabled in-product chat assistant. Built an AI microservice layer (LangChain) to swap model providers easily and created automated + human-in-the-loop evaluation systems (including ~100-test runs) to iterate on prompts and quality.”
Mid-level Full-Stack Software Engineer specializing in cloud and AI-enabled applications
“Product-focused full-stack engineer (70/30 app vs infra) with Accenture experience and recent AI workflow work, shipping end-to-end systems from React/TypeScript UIs through FastAPI backends to Postgres. Built an AI-driven data extraction platform with async job APIs, strict schema validation, and strong observability, and has operated AWS ECS-based deployments with real incident mitigation (DB connection exhaustion/latency under traffic spikes).”
Mid-level Data Analyst specializing in healthcare and financial analytics
“Built and productionized an LLM-powered clinical documentation and insights pipeline at Cardinal Health using LangChain + GPT-4 with RAG to summarize long clinical notes, extract medication/dosage entities, and generate structured SQL-ready outputs for downstream analytics. Emphasizes clinical reliability via labeled benchmarking (precision/recall/F1), shadow deployments, clinician human-in-the-loop review, and ongoing monitoring/orchestration with Airflow, Lambda, S3, Postgres, and Power BI.”
Mid-level Full-Stack/Backend Engineer specializing in Java microservices and cloud platforms
“PayPal ML/AI practitioner who built and productionized a hybrid recommendation engine (BERT/LLM embeddings + collaborative filtering + XGBoost ranking) on AWS with end-to-end MLOps and orchestration. Addressed real-world issues like cold start and embedding latency (ONNX, clustering, caching, PySpark/Delta Lake) and drove a 27% lift in upsell conversion via A/B testing and stakeholder collaboration with marketing.”
Junior Backend Software Engineer specializing in microservices and API platforms
“Backend engineer with strong performance and security instincts: built a Flask API for readability metrics with clean, testable modular design; optimized SQLAlchemy/Postgres to eliminate N+1 issues (800ms to 120ms). Also implemented an LLM-powered natural-language travel search using Claude Sonnet + Amadeus with RAG and anti-exploitation safeguards, plus multi-tenant isolation via Postgres RLS and Redis caching that cut search latency from ~20s to ~4–5s while reducing storage costs.”
Mid-level Generative AI Engineer specializing in LLM apps, RAG, and MLOps
“LLM/GenAI engineer with US Bank experience building a production financial-document intelligence platform using LangChain/LangGraph, GPT-4, and Amazon OpenSearch. Delivered a RAG-based assistant for compliance/audit teams with grounded, cited answers, focusing on reducing hallucinations and latency, and deployed securely on AWS (SageMaker/EKS) with CI/CD and evaluation tooling (LangSmith, RAGAS).”
Mid-level AI/ML & MLOps Engineer specializing in cloud AI infrastructure and GenAI
“At HPE, led and deployed an enterprise-grade LLM document intelligence platform for an insurance client, automating extraction from highly variable PDFs/scans/emails and raising field accuracy from 74% to 93%. Built a LangChain/Pinecone/OpenSearch RAG framework to cut hallucinations by 37% and operationalized LangSmith evals in CI, driving a 41% triage accuracy lift and >33% fewer incorrect resolutions while partnering closely with claims operations via HITL workflows.”
Junior Full-Stack/AI Engineer specializing in enterprise AI agents and web platforms
“Forward Deployed Engineer focused on taking enterprise LLM voice agents from prototype to production. Led a turnaround on a high churn-risk account by building a custom nested-API integration and preprocessing layer that enabled the LLM to reason over complex order hierarchies, cutting call handle time from 15 minutes to 2 minutes and driving expansions. Strong in real-time agent/workflow debugging, developer workshops, and sales partnership for adoption.”
Mid-level ML/AI Engineer specializing in NLP, RAG pipelines, and financial risk & fraud systems
“Built and shipped LLM/RAG systems in finance and startup settings, including a Goldman Sachs document intelligence platform that indexed ~8TB of regulatory filings and delivered cited, conversational answers with <2s latency—cutting compliance research by ~4.5 hours per batch. Also developed LangChain-based agent workflows at Finta to automate CRM enrichment and investor lookup with strong testing, tracing (LangSmith), privacy guardrails, and auditability.”
Mid-level AI/ML Engineer specializing in NLP, computer vision, and Generative AI
“Built and deployed a production LLM-powered clinical insights/summarization assistant for healthcare teams, including a Spark+Airflow pipeline, fine-tuned transformer models, and a FastAPI Docker service on AWS. Demonstrates strong MLOps/LLMOps depth (Airflow on Kubernetes, custom AWS operators/IAM, MLflow, CloudWatch) and practical reliability work like hallucination mitigation, confidence scoring, and retrieval-backed evaluation with shadow deployments.”
Mid-level AI/ML Engineer specializing in MLOps, LLMs, and real-time inference in FinTech
“ML/LLM engineer who has deployed a production LLM-powered assistant for intent classification and query routing (order recommendation/support deflection), combining BERT fine-tuning with an embedding-based retrieval layer and optimizing for low-latency inference. Experienced with end-to-end reliability practices—Airflow-orchestrated ETL, data validation/alerting, MLflow experiment tracking, and iterative improvements driven by user feedback and monitoring.”
Mid-level Full-Stack Java Engineer specializing in microservices, cloud, and event-driven systems
“Software engineer at Procter & Gamble focused on warehouse/operations systems, building near-real-time order/inventory visibility using Java/Spring Boot, React, Kafka, PostgreSQL, and Redis with measurable latency and load-time gains. Also shipped internal LLM/RAG knowledge assistants grounded in company runbooks and workflows, implementing guardrails and an evaluation loop that drove concrete retrieval improvements (document chunking) and regression prevention.”
Junior Full-Stack & AI/ML Engineer specializing in LLMs and multimodal document processing
“Built a production RAG-based NBA player scouting assistant that embeds player profiles into FAISS, orchestrates retrieval and LLM recommendations with LangChain, and surfaces results via embedded Tableau dashboards. Demonstrates strong focus on evaluation/monitoring (batch tests, LLM-as-judge, latency/failure/token metrics) and has experience translating non-technical founder goals into DAPT + fine-tuning plans on curated data.”
Mid-level Machine Learning Engineer specializing in MLOps and GenAI analytics
“ML/LLM practitioner who has deployed a production RAG-based trouble-call identifier using multiple datasets (device, network, past complaints). Experienced in end-to-end MLOps (FastAPI + Docker + Kubernetes with HPA) and in evaluating/monitoring LLM behavior to reduce hallucinations, with additional applied work in forecasting/anomaly detection and churn prediction for retention campaigns.”
“Built and deployed a production RAG-based internal knowledge assistant that let analysts query company documents in natural language, using LangChain/LangGraph with Pinecone and a FastAPI service for integration. Emphasizes reliability in production through hallucination mitigation (retrieval tuning + prompt guardrails) and measurable evaluation/monitoring (accuracy, latency, task completion, hallucination rate), iterating based on user feedback.”
Mid-level QA Automation Engineer / SDET specializing in Financial Services and Healthcare
“Fintech-focused engineer who built an end-to-end KYC verification pipeline for advisor onboarding using Flask microservices, Celery/Redis, and AWS (Lambda/ECS/EC2) with CloudWatch-driven scaling and latency optimizations. Also shipped a production internal knowledge assistant using RAG + embeddings/vector search with guardrails (similarity-based fallback, prompt-injection protections) and an evaluation loop with compliance specialist review that drove measurable retrieval improvements.”
Intern Software Engineer specializing in cloud, big data, and test automation
“Internship experience at Qualitest building and deploying an LLM-powered test automation system that reduced manual test creation and improved efficiency (~40%). Demonstrates strong production engineering for LLM systems (timeouts/retries/monitoring/caching, prompt optimization, batching) and has scaled workflows to 100+ concurrent jobs; also has orchestration experience with AWS Step Functions and Kubernetes.”
Mid-Level Software Development Engineer specializing in full-stack and cloud-native systems
“Backend engineer who has shipped production LLM-powered features, including an AI-assisted developer tool on AWS (Spring Boot) and a blog platform capability using embeddings + Elasticsearch for semantic retrieval and LLM-generated summaries/recommendations. Demonstrates practical tradeoff management (quality/latency/cost), guardrails to reduce hallucinations, and evaluation-driven iteration using real user queries and observability via ELK.”