Pre-screened and vetted.
Senior AI/ML Engineer specializing in Generative AI and RAG
“ML/NLP practitioner at Morf Health focused on unifying fragmented healthcare data by linking structured patient/encounter records with unstructured clinical notes. Has hands-on experience with transformer embeddings, vector databases, and domain fine-tuning, plus rigorous evaluation (precision/recall) and human-in-the-loop validation with clinical SMEs to make pipelines production-grade.”
Mid-level Generative AI Engineer specializing in LLM systems and RAG
“Currently at Huntington Bank, built a production-grade RAG system that helps business/operations teams get grounded answers from large volumes of internal enterprise documents. Owns ingestion and FastAPI backend, tuned hybrid BM25+vector retrieval and chunking for relevance, and evaluates reliability with metrics and observability (LangSmith, CloudWatch, Prometheus/Grafana) while partnering closely with non-technical stakeholders.”
Junior Machine Learning Engineer specializing in computer vision and LLM applications
“Built and led an autonomous driving software effort for Formula Student, owning the full autonomy stack (perception, planning, control) orchestrated in ROS. Implemented stereo depth + YOLO object detection, RRT/RRT* planning, and a robust SLAM pipeline (Kalman filter, submapping) while leveraging Gazebo simulation and modern deployment tooling (Docker/Kubernetes, AWS, GitHub Actions CI/CD).”
Mid-level Machine Learning & Data Infrastructure Engineer specializing in MLOps on AWS
“Built and deployed a fine-tuned Qwen 2.5 14B model into production at Dextr.ai as the backbone for hotel-operations agentic workflows, running on AWS EKS with Triton and TensorRT-LLM. Demonstrates strong cost-aware LLM engineering (QLoRA, FP8/BF16 on H100) plus rigorous benchmarking/observability (Prometheus, LangSmith) with reported sub-30ms TTNT. Previously handled long-running ETL orchestration with Airflow at GE Healthcare and Lowe's.”
Senior GenAI/ML Engineer specializing in LLMs, RAG, and multimodal generative AI
“LLM/RAG engineer with production deployments in highly regulated domains (Frost Bank and GE Healthcare). Built secure, explainable document-grounded Q&A systems using LoRA fine-tuning, strict RAG with confidence thresholds, and citation-based responses; also established evaluation/monitoring (golden QA sets, hallucination tracking, drift) and achieved ~40% latency reduction through retrieval/prompt tuning.”
Mid-level AI/ML Engineer specializing in LLMs, NLP, and MLOps
“AI/ML engineer with healthcare domain depth who led a HIPAA-compliant, production LLM system at McKesson to automate clinical document understanding—extracting entities, summarizing provider notes, and supporting authorization decisions. Hands-on across Spark/Python ETL, Hugging Face + LoRA/QLoRA fine-tuning, RAG, and cloud-native MLOps (Airflow/Kubernetes/Step Functions, MLflow, blue-green on EKS/GKE), with explicit work on PHI handling and hallucination reduction.”
Senior ML Engineer & Data Scientist specializing in LLM agents, retrieval/ranking, and MLOps
“Machine Learning Engineer currently at Webster Bank building an enterprise-scale LLM agent for Temenos Journey Manager/Maestro, using RAG-style multi-stage retrieval with FAISS/Pinecone, hybrid dense+sparse search, and LoRA fine-tuning optimized via NDCG/MAP and A/B testing. Previously handled messy incident/telemetry data at Deuta Werke GmbH with deterministic + fuzzy entity resolution, and has strong production data engineering experience across Spark/Hadoop and Python ETL systems.”
Mid-level AI/ML Engineer specializing in enterprise ML, MLOps, and Generative AI
“ML/LLM engineer who has shipped production RAG systems (LangChain + HF Transformers + FAISS) with hybrid retrieval and cross-encoder re-ranking, deployed via FastAPI/Docker/Kubernetes and monitored with MLflow. Also partnered with wealth advisors at Edward Jones to deliver a client retention model with SHAP-driven explanations and a dashboard that improved trust, adoption, and reduced high-value client churn.”
Mid-Level Software Development Engineer specializing in full-stack and LLM/AI systems
“AI engineer with hands-on production experience building an end-to-end RAG system that reduced document-answering time from hours to minutes, improving accuracy through chunk overlap and hybrid BM25+semantic retrieval. Also built a LangGraph-based agent that researches company financial news via web search (Google Serper), using Pydantic structured outputs and checkpointing for reliability; experienced collaborating with non-technical stakeholders at JPMC and communicating ROI.”
Mid-level Machine Learning Engineer specializing in GenAI, LLMs, and real-time ML systems
“Built and deployed a production long-form article summarization system using BART/T5/PEGASUS, tackling real-world constraints like token limits, latency/quality tradeoffs, and factual drift via chunking/merge logic and constrained decoding. Uses pragmatic Python-based pipeline orchestration (scheduled jobs, modular scripts, logging/retries) and iterates with stakeholder feedback to make outputs genuinely useful for content workflows.”
Mid-level Generative AI Engineer specializing in LLM apps, RAG, and MLOps
“LLM/GenAI engineer with US Bank experience building a production financial-document intelligence platform using LangChain/LangGraph, GPT-4, and Amazon OpenSearch. Delivered a RAG-based assistant for compliance/audit teams with grounded, cited answers, focusing on reducing hallucinations and latency, and deployed securely on AWS (SageMaker/EKS) with CI/CD and evaluation tooling (LangSmith, RAGAS).”
Mid-level AI/ML & MLOps Engineer specializing in cloud AI infrastructure and GenAI
“At HPE, led and deployed an enterprise-grade LLM document intelligence platform for an insurance client, automating extraction from highly variable PDFs/scans/emails and raising field accuracy from 74% to 93%. Built a LangChain/Pinecone/OpenSearch RAG framework to cut hallucinations by 37% and operationalized LangSmith evals in CI, driving a 41% triage accuracy lift and >33% fewer incorrect resolutions while partnering closely with claims operations via HITL workflows.”
Executive Technology Leader (CTO) specializing in IoT sensing, AI/ML, and RF/embedded systems
“Currently a startup CTO who thrives on building new technology stacks and rapidly turning technical ideas into products. Interested in partnering with a CEO/business team to commercialize embedded/edge concepts such as multi-sensor drone localization (video/audio/RF with SDR), low-cost solar+battery power nodes networked via LoRa, and an Amazon Sidewalk/LoRa connectivity device with cloud management.”
Junior Data Scientist / ML Engineer specializing in LLMs and Computer Vision
“Currently working in CoRAL Lab, built and deployed IntegrityShield—a document-layer PDF watermarking system that keeps assessments visually identical while disrupting LLM-based solving; validated in a real classroom where it helped catch 12 AI-cheating cases. Also built MALDOC, a modular red-teaming platform for document-processing AI agents using LangGraph to run reproducible, deterministic adversarial trials across OCR/text/vision routes.”
Mid-level Generative AI & Machine Learning Engineer specializing in agentic LLM systems
“Built and deployed a production agentic LLM knowledge assistant that answers complex questions over internal documents, APIs, and databases using a RAG architecture (FAISS/Pinecone) and LangChain/LangGraph orchestration. Emphasizes production-grade reliability and hallucination control through grounding, confidence thresholds, validation, retries/fallbacks, and full observability (logging/metrics/traces) with continuous evaluation and feedback loops.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and multimodal generation
“Open-source JavaScript contributor focused on performance and maintainability in data visualization libraries—refactored legacy ES5 into modular ES6, added tests/docs, and delivered ~30% faster load times with positive community adoption. Also optimized a React dashboard (~40% load-time reduction) and took ownership in an ambiguous AI product initiative by setting milestones, standing up an initial ML pipeline, and shipping a prototype in ~6 weeks that became the basis for production.”
Mid-level AI/ML Engineer specializing in LLM agents, RAG, and ML systems
“At Inertia Systems, built a production LLM-powered ingestion pipeline that converts heterogeneous sources (PDF/JSON/IFC/SQL and financial tables) into standardized text and uses GraphRAG to construct a knowledge graph with verified dependency relationships. Also has hands-on HPC orchestration experience with SLURM, including creating a custom wrapper process manager to improve resource utilization under restrictive scheduling policies.”
“ML/GenAI engineer with recent CVS Health experience building a production RAG system over unstructured financial/research documents using LangChain, FAISS, and Pinecone, plus LoRA/PEFT fine-tuning of GPT/LLaMA for domain-aware summarization. Demonstrates strong applied MLOps and data engineering skills (Airflow/Prefect, Docker/Kubernetes, CI/CD, MLflow) and measurable impact (sub-second retrieval, ~40% better context retrieval, ~25% entity matching improvement).”
Senior Full-Stack & GenAI Engineer specializing in healthcare and financial services
“Built and deployed a production LLM-powered customer support assistant using a RAG backend in Python, focused on deflecting repetitive Tier-1 tickets and reducing resolution time. Demonstrates strong production engineering instincts around reliability (confidence scoring + human fallback), scalability/cost optimization (multi-stage pipelines), and workflow orchestration/observability (LangChain, custom DAGs, structured logging, step metrics).”
Junior MLOps Engineer specializing in LLMs and cloud infrastructure
“Built a production multimodal LLM system (Gemini on GCP) to automate behavioral coding of family-involved science experiment videos, including preprocessing for inconsistent lighting/audio and LangGraph-orchestrated parallel workflows. Also developed rubric-based AI grading workflows and partnered closely with non-technical education stakeholders through explainability-focused walkthroughs and manual-vs-AI evaluation alignment.”
Mid-level Data Scientist specializing in NLP, LLMs, and RAG systems
“Built and deployed a production-style vision-language pipeline that generates structured medical reports from chest X-rays using BioViLT embeddings, an image-text alignment module, and BiGPT fine-tuned with LoRA, delivered via Streamlit and hosted on AWS EC2. Also collaborating experience presenting EDA findings, feature importance, and model performance to Ford managers while working with vehicle parts data at Bimcon.”
Senior Software Developer specializing in AI/ML automation and cloud-native systems
“ML/MLOps practitioner who built production systems for telecom network analytics, including an automated labeling + multi-label Random Forest solution that cut labeling effort by 90% and sped up RCA. Led an Ericsson auto-deployment platform using Airflow, Azure IoT Hub, Docker, and Celery to orchestrate 120+ containerized ML/rule-based deployments, saving ~80 hours of setup per deployment.”
Mid-level AI/ML Engineer specializing in healthcare ML and LLM/RAG systems
“AI/LLM engineer with recent production experience at UnitedHealth Group building an end-to-end RAG system over structured EMR data and unstructured clinical notes, including evidence retrieval, GPT/LLaMA-based reasoning, and a validation layer for reliability. Strong in orchestration (Kubeflow/Airflow/MLflow), prompt engineering for noisy healthcare text, and rigorous evaluation/monitoring with gold-standard benchmarking, plus close collaboration with clinical operations stakeholders.”
Mid-level AI/ML Engineer specializing in MLOps and LLM-powered applications
“AI/ML engineer with production experience building a RAG-based internal analytics assistant (Databricks + ADF ingestion, Pinecone vector store, LangChain orchestration) deployed via Docker on AWS SageMaker with CI/CD and MLflow. Strong focus on real-world constraints—latency/cost optimization (LoRA ~60% compute reduction), hallucination control with citation grounding, and enterprise security/governance. Previously at Intuit, delivered an interpretable churn prediction system (PySpark/Databricks, Airflow/Azure ML) that improved retention targeting ~12%.”