Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in NLP, LLMs, and MLOps
Mid-Level Software Engineer specializing in full-stack, data engineering, and ML
Senior Data Scientist specializing in ML, fraud risk, and Generative AI (RAG/LLMs)
Mid-level AI/ML Engineer specializing in GenAI, computer vision, and real-time ML pipelines
Mid-level Data Scientist / ML Engineer specializing in LLMs and predictive analytics
Senior Data Scientist specializing in healthcare analytics and scalable ML pipelines
Senior Machine Learning Engineer specializing in MLOps and Generative AI
Mid-level Data Scientist specializing in financial ML, NLP, and MLOps
Mid-level AI/ML Engineer specializing in LLMs, NLP, and analytics automation
“AI/ML Engineer (TCS) who built and deployed a production LLM-powered audit transaction validation service to reduce manual review of unstructured transaction records and comments. Implemented a LangChain/Python pipeline for extraction/normalization and discrepancy detection, with strong production reliability practices (decision logging, dashboards, labeled eval sets) and a human-in-the-loop auditor feedback loop to improve precision/recall under strict data-sensitivity and near-real-time constraints.”
Mid-level Data Scientist specializing in LLM development and scalable ML pipelines
“Built and deployed production LLM pipelines for evidence-based scoring in two domains: biomedical literature mining (scoring ~2700 drug compounds vs gene targets/mechanisms) and long-horizon news analytics (35 years of Chinese articles). Emphasizes reliability at scale (retries/checkpointing/validation), rigorous empirical model benchmarking (GPT-4o/mini/5), and translating results into stakeholder-friendly visual narratives.”
Mid-level Machine Learning Engineer specializing in healthcare NLP and MLOps
“ML/AI practitioner in healthcare (Syneos Health) who has deployed production clinical NLP and risk models. Built a BERT-based physician-note information extraction system on Docker + AWS SageMaker (reported ~42% retrieval improvement) and automated retraining/deployment with Airflow and drift detection, while partnering closely with clinicians to drive adoption (reported ~18% readmission reduction).”
Mid-level AI/ML Engineer specializing in GenAI and predictive modeling
“Built and deployed a GPT-4-powered medical assistant for clinical staff to reduce time spent searching guidelines and EHR information, with a strong emphasis on safety and compliance. Uses strict RAG, confidence thresholds, and fallback behaviors to prevent hallucinations, and runs production-grade workflows orchestrated with LangChain/LangGraph plus Docker/Kubernetes/MLflow and monitoring for reliability and cost.”
Intern Data Scientist specializing in ML, NLP, and MLOps for healthcare and enterprise AI
“Built a production multi-cloud LLM-driven IT ticket automation system using LangGraph, Azure + Pinecone RAG, and an Ollama-hosted LLM on AWS, with Terraform-managed infra and PostgreSQL audit/state tracking for reliability. Also partnered with UW School of Medicine & Public Health students to deliver a glioma survival risk-ranking model, translating clinical feedback into practical pipeline improvements (imputation, site harmonization) and stakeholder-friendly visualizations.”
Mid-level Machine Learning Engineer specializing in Generative AI and RAG systems
“LLM/ML engineer who has shipped an enterprise RAG-based Q&A system (LangChain/LlamaIndex, FAISS + Azure Cognitive Search, GPT-3.5/4 via OpenAI/Azure OpenAI) to production on Docker + Kubernetes/OpenShift, tackling hallucinations, retrieval quality, latency/cost, and RBAC/IAM security. Also partnered with operations leaders to turn manual reporting into an LLM-powered summarization and forecasting dashboard driven by real KPIs and iterative stakeholder feedback.”
Mid-level Data Scientist specializing in AI/ML, MLOps, and LLM-powered analytics
“Built and deployed a production LLM-powered document Q&A system enabling natural-language querying of large PDFs, focusing on retrieval quality (overlapped chunking) and low-latency performance (optimized embeddings + vector search). Experienced with scaling ML/LLM workflows using async/batch processing, caching, cloud storage, and orchestration via Apache Airflow with robust testing, monitoring, and failure handling.”
Mid-level AI/ML Engineer specializing in fraud detection, NLP, and MLOps
“Built a production real-time fraud detection and customer-support automation platform at Citibank, tackling extreme class imbalance (reported ~1:5000) and strict latency constraints. Combines hands-on MLOps (Airflow, Kubernetes, MLflow; Snowflake/Spark/S3 integrations; CI/CD model promotion) with cross-functional delivery to Risk & Compliance focused on interpretability and reducing false positives.”
Mid-level AI/ML Engineer specializing in MLOps, NLP, and Computer Vision
“Built and deployed a production LLM-powered text extraction/classification system that converts messy unstructured reports into searchable insights, running on AWS SageMaker with automated retraining and monitoring. Strong in orchestration (Step Functions/Kubernetes/Airflow patterns) and reliability practices (gold datasets, prompt/tool unit tests, shadow/canary/A-B testing, guardrails/rollback), and has experience translating non-technical stakeholder needs into an NLP workflow plus dashboard.”
Junior Applied AI Engineer specializing in LLMs, RAG, and agentic systems
“Co-founded a healthcare AI startup building and deploying software directly with end users, emphasizing rapid shipping, deep user interviews, and workflow-first adoption. Has hands-on production deployment experience on AWS (including diagnosing a silent AWS App Runner failure caused by an ARM vs amd64 Docker build mismatch) and is motivated by customer-facing, travel-heavy roles to keep engineering tightly connected to real-world usage.”
Mid-level AI/ML Engineer specializing in healthcare analytics and MLOps
“AI/ML engineer at Cigna Healthcare building a production, HIPAA-compliant LLM-powered clinical insights platform that summarizes unstructured medical notes using a fine-tuned transformer + RAG on AWS. Demonstrates strong end-to-end MLOps and cloud optimization (distillation, Spot/Lambda/Auto Scaling) with quantified outcomes (~28% accuracy lift, ~40% less manual review, ~25% lower ops cost) and strong clinician-facing explainability via SHAP and dashboards.”
Mid-level Generative AI Engineer specializing in LLM systems and RAG
“Currently at Huntington Bank, built a production-grade RAG system that helps business/operations teams get grounded answers from large volumes of internal enterprise documents. Owns ingestion and FastAPI backend, tuned hybrid BM25+vector retrieval and chunking for relevance, and evaluates reliability with metrics and observability (LangSmith, CloudWatch, Prometheus/Grafana) while partnering closely with non-technical stakeholders.”
Junior Machine Learning Engineer specializing in computer vision and LLM applications
“Built and led an autonomous driving software effort for Formula Student, owning the full autonomy stack (perception, planning, control) orchestrated in ROS. Implemented stereo depth + YOLO object detection, RRT/RRT* planning, and a robust SLAM pipeline (Kalman filter, submapping) while leveraging Gazebo simulation and modern deployment tooling (Docker/Kubernetes, AWS, GitHub Actions CI/CD).”
Senior GenAI/ML Engineer specializing in LLMs, RAG, and multimodal generative AI
“LLM/RAG engineer with production deployments in highly regulated domains (Frost Bank and GE Healthcare). Built secure, explainable document-grounded Q&A systems using LoRA fine-tuning, strict RAG with confidence thresholds, and citation-based responses; also established evaluation/monitoring (golden QA sets, hallucination tracking, drift) and achieved ~40% latency reduction through retrieval/prompt tuning.”
Mid-level Data Analyst specializing in financial risk and healthcare analytics
“AI/ML engineer focused on real-time, production-grade LLM systems, with a robotics-adjacent mindset around latency/accuracy tradeoffs and modular pipelines. Built a scalable RAG-based assistant orchestrated as microservices on Kubernetes with Kafka async messaging, ONNX/quantization optimizations, and monitoring (Prometheus/Grafana), citing a ~35% hallucination reduction; has also experimented with ROS Noetic/Gazebo to understand ROS concepts.”