Pre-screened and vetted.
Mid-level Data Scientist / AI-ML Engineer specializing in Generative AI and LLM applications
“Built a production GenAI-powered analytics assistant to reduce reliance on data analysts by enabling natural-language Q&A over Databricks/Power BI dashboards, backed by vector search (Pinecone/Milvus) and a Neo4j knowledge graph, including multimodal support via OpenAI Vision. Demonstrates strong real-world LLM reliability engineering with strict RAG, LangGraph multi-step verification, and Guardrails/custom validators, plus broad orchestration and production monitoring experience (Airflow, ADF, Step Functions, Kubernetes, Prometheus/CloudWatch).”
Mid-level GenAI Engineer specializing in production RAG and LLM fine-tuning
“LLM engineer who built a production seller-support RAG system at eBay using hybrid retrieval (BM25 + Pinecone vectors) with Cohere reranking, LangGraph orchestration, and citation-grounded answers. Strong focus on reliability: semantic/structure-aware chunking, automated Ragas-based evaluation with nightly regressions, and production observability (LangSmith) plus drift monitoring (Arize). Also implemented a multi-agent fraud pipeline with AutoGen using JSON-schema contracts and explicit termination conditions.”
“LLM engineer who has deployed production RAG systems for regulated document QA (PDFs/knowledge bases), emphasizing grounded answers with citations, RBAC, monitoring, and continuous feedback. Demonstrates deep practical expertise in retrieval quality (semantic chunking, hybrid BM25+embeddings, re-ranking), reliability (guardrails, deterministic workflows), and measurable evaluation (golden sets, log replay, A/B tests) while partnering closely with compliance/operations stakeholders.”
Intern Machine Learning Engineer specializing in NLP, RAG, and deepfake detection
“Early-career (fresher) candidate who built and deployed a production AI medical document chatbot using a RAG architecture (LangChain + Hugging Face LLM + Pinecone) with a Flask backend on AWS EC2 via Docker. Has experience troubleshooting real deployment constraints (model dependencies, disk space, container stability) and setting up continuous-style evaluation with fixed query test sets tracking relevance, latency, and error rate.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
“GenAI/data engineering practitioner with production experience across Equinix, Optum, and Citibank—built an Azure OpenAI (GPT-4) + LangChain document intelligence platform processing 1.5M+ docs/month and a HIPAA-compliant Airflow healthcare pipeline handling 5M+ claims/day. Also delivered a real-time fraud detection + explainability system using LightGBM and a fine-tuned T5 NLG component, improving fraud accuracy by 15%+ while partnering closely with compliance stakeholders.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
“ServiceNow engineer who built and launched a production LLM-powered ticket resolution/knowledge assistant using RAG (LangChain + Hugging Face embeddings + vector search) integrated into internal support dashboards via REST APIs. Optimized the system from ~6–8s to ~2–3s latency while improving usability with concise, cited answers and guardrails (grounding + similarity thresholds), delivering ~30–35% reduction in manual ticket investigation effort.”
Mid-level Software Engineer specializing in autonomous driving simulation and 3D mapping
“Founding software engineer who built an autonomous-vehicle 3D digital twin using Unreal Engine 5 and CARLA, owning core simulator logic (traffic/scenarios/weather) and a ROS 2-based pipeline to record synchronized multi-sensor data (RGB/depth/segmentation/LiDAR/IMU/GPS). Also implemented distributed synchronization patterns (server + client prediction) using FastAPI and WebSockets; seeking roles with H1B transfer and targeting ~$110k.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Junior AI/Backend Software Engineer specializing in ML and scalable systems
“Backend engineer with strong AWS/CI/CD experience (multi-repo deployments, Lambda + core app, immutable ECR and image promotion) and a published master’s thesis building an ML framework for Solar PV energy prediction and CO2 reduction impact modeling using ensemble and meta-learning approaches benchmarked against SAM.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMOps, and MLOps
“Built and deployed an AWS-based LLM/RAG ticket triage and knowledge retrieval system (Pinecone/FAISS + Step Functions + MLflow) that cut support resolution time by 20%. Demonstrates strong production focus on hallucination reduction, PII security, and low-latency orchestration, with measurable evaluation improvements (e.g., ~25% grounding accuracy gain via re-ranking) and proven collaboration with support operations stakeholders.”
Senior Applied AI/ML Engineer specializing in GenAI, LLMs, RAG and agents
“Applied AI/ML Engineer at JPMorgan Chase who led a banker-facing LLM chatbot from an OpenAI-API POC to a production RAG workflow, including hallucination mitigation, automated evaluation in SageMaker, and operational monitoring with Dynatrace. Also delivers external technical education—hosted a hands-on Grace Hopper Celebration 2025 workshop teaching LangChain/LangGraph agentic workflows.”
Senior Data Scientist / ML Engineer specializing in cloud ML pipelines and GenAI
“ML/NLP practitioner with experience building a transformer-failure prediction system that combines sensor signals with unstructured maintenance comments using LLM-based extraction and similarity validation. Strong emphasis on production readiness—data leakage controls, SQL-driven data quality tiers, and rigorous bias/fairness validation (including contract/spec evaluation across diverse company profiles).”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Red Hat ML/LLM engineer who designed and deployed a production LLM-powered customer support automation system using RAG, improving latency by 30% via PEFT and vector search optimization. Built security and governance into retrieval (access-level filtering, encrypted Pinecone/ChromaDB) and delivered SHAP-based explainability via a dashboard for non-technical stakeholders. Experienced orchestrating distributed ML/RAG pipelines across AWS SageMaker and OpenShift with Airflow/Prefect, plus multi-agent workflows using CrewAI and LangGraph.”
Mid-level Data Scientist specializing in NLP/LLMs, time series forecasting, and MLOps
“Data/ML practitioner with hands-on experience building NLP systems from prototype to production: delivered a Twitter sentiment classifier with robust preprocessing, SVM modeling, and Power BI reporting, and built entity-resolution pipelines for messy multi-source customer data (reporting ~95% improvement in unique entity identification). Also implemented semantic linking/search using SBERT embeddings with FAISS vector retrieval and domain fine-tuning (reported ~15% precision lift), and applies production workflow best practices (Airflow/Prefect, Docker, Azure ML/Databricks, Great Expectations).”
Mid-level AI/ML Engineer specializing in GenAI, LLMs, RAG, and MLOps
“Built and deployed a production LLM-powered RAG document intelligence/Q&A system for healthcare prior authorization, reducing manual medical document review time and improving decision efficiency. Strong in end-to-end LLM application engineering (LangChain/LangGraph), retrieval quality improvements (hybrid search, embedding tuning, chunking strategies), and rigorous evaluation/monitoring for reliability.”
Junior Full-Stack Software Engineer specializing in React and AI-powered applications
“Full-stack/AI-focused builder who shipped a production Career Advisor app using LLMs + RAG + vector DB (React/Node/MongoDB/Claude API) and grew it to 2000+ users, handling real deployment issues and CI/CD on Vercel/Render. Also developing an AI-powered iOS “3D World Explorer” (text-to-3D) and has cloud experience across Azure and AWS (S3/SageMaker/EC2).”
Mid-level Data Scientist specializing in LLMs, MLOps, and predictive analytics in healthcare and finance
“Built and deployed a production LLM/RAG clinical decision support system that enables real-time semantic search over unstructured EHR notes and delivers patient risk insights. Strong in healthcare-grade MLOps and compliance (HIPAA, PHI handling, encryption, RBAC, audit logs) and scaled embedding/retrieval pipelines using Spark/Databricks and Airflow. Partnered with clinicians via Power BI dashboards and explainability, contributing to an 18% reduction in patient readmissions.”
Mid-level AI/ML Engineer specializing in GenAI, NLP, and MLOps
“Built and deployed an enterprise GenAI knowledge assistant over thousands of internal PDFs/reports using a RAG stack (GPT-4 + Hugging Face embeddings + vector DB) to reduce manual search and SME escalations. Uses LangGraph/LangChain to orchestrate modular agent workflows with relevance filtering and fallback handling, and applies rigorous evaluation (golden datasets, edge cases, A/B tests) with production monitoring metrics.”
Mid-Level Software Engineer specializing in distributed systems and cloud-native backends
“AI/LLM engineer with production experience at Charles Schwab building a RAG-based assistant to help 5,000+ reps answer complex financial policy questions. Implemented a multi-layer anti-hallucination approach (GNN-driven ontology/graph retrieval + citation-only answers) and compliance-focused guardrails (Azure AI Content Safety) in partnership with audit/compliance stakeholders.”
Junior Software Engineer specializing in ML, distributed systems, and LLM applications
“Interned at Zonda where he built an AI-driven semantic search solution over ~280M housing/builder records. Iterated from local LLMs via llama.cpp quantization to a vector-embedding retrieval system, then boosted semantic accuracy with a custom spaCy NER layer and re-ranking, optimizing for latency through precomputation. Collaborated with economics-focused stakeholders to reduce manual document/paperwork time by enabling natural-language search over internal data.”
Mid-level AI/ML Data Scientist specializing in NLP, computer vision, and risk analytics
“ML/AI engineer with Capital One experience building production-grade customer segmentation and fraud detection systems combining NLP (transformers) and anomaly detection. Strong MLOps and orchestration background (PySpark ETL, MLflow, Airflow, Docker/Kubernetes, Azure ML) with real-time monitoring/alerting and performance optimizations like quantization and caching, plus proven ability to deliver business-facing insights through Power BI/Tableau for marketing stakeholders.”
Mid-level Machine Learning Engineer specializing in financial AI, NLP, and MLOps
“AI/ML engineer with experience at Accenture and Morgan Stanley, building production LLM systems (GPT-3 summarization) and finance-focused ML models (credit risk and trading anomaly detection). Combines MLOps depth (Docker/Kubernetes, AWS SageMaker/Glue/Lambda, MLflow, A/B testing, drift monitoring) with practical domain adaptation techniques like few-shot prompting and RAG/knowledge-base integration.”