Pre-screened and vetted.
Senior Data Scientist specializing in NLP, MLOps, and cloud ML platforms
Mid-level AI/ML Engineer specializing in healthcare, risk modeling, and MLOps
“Robotics software engineer who built a ROS Noetic-based perception-to-control stack for a pick-and-place robotic arm, integrating OpenCV/TensorFlow vision with motion planning and PID tuning. Demonstrated strong real-time debugging skills (rosbag, queue/latency fixes) and experience deploying reproducible robotics environments with Gazebo simulation, Docker, and GitLab CI.”
Mid-level Machine Learning Engineer specializing in deep learning and generative AI
“AI/ML engineer who has deployed transformer-based NLP systems to production via Python REST APIs and Kubernetes on AWS/Azure, with a strong focus on latency optimization (p95), reliability, and scalable orchestration. Demonstrates pragmatic model tradeoff decision-making and strong stakeholder collaboration—improving adoption by making outputs more actionable with summaries, extracted fields, and confidence indicators.”
Mid-level Data Analyst specializing in AWS-based ETL, churn analytics, and BI dashboards
“Data/ML practitioner with experience at Airtel and Lincoln Financial delivering measurable business outcomes: improved retention 15% via NLP sentiment analysis and cut response time ~25% using sentence-BERT + FAISS semantic linking. Strong in data quality/identity resolution (SQL + fuzzy matching) and in building production-grade Python workflows orchestrated with Airflow/AWS Glue, including validation and dashboard integration in Power BI.”
Senior Data Scientist/ML Engineer specializing in scalable ML and LLM systems
“Built and deployed an end-to-end product that brings a research-paper approach into production for large-scale time-series clustering, with attention to partitioning, latency, and scalability. Also designed a Python-based backend validation service (comparing outputs to database ground truths) and handled production reliability issues by reproducing dataset-specific crashes and hardening corner-case behavior with client-friendly errors.”
Mid-level Data Scientist specializing in MLOps and Generative AI
“Robotics software/ML engineer who built perception and navigation-related ML systems for autonomous supermarket carts, including object detection, shelf recognition, and obstacle avoidance. Strong ROS/ROS2 practitioner who optimized real-time performance (reported 50% latency reduction) and deployed containerized ROS/ML pipelines at scale using Docker, Kubernetes, and CI/CD.”
Mid-level Data Scientist specializing in ML, NLP, and Generative AI
“GenAI/ML engineer with production experience at Cognizant and Ally Financial, building end-to-end LLM/RAG systems and ML pipelines. Delivered a domain chatbot trained from 90k tickets and 45k docs, improving intent accuracy (65%→83%), scaling to 800+ concurrent users with 99.2% uptime and sub-150ms latency, and driving +14% customer satisfaction. Strong in Azure ML + DevOps CI/CD, Dockerized deployments, and explainable/PII-safe modeling using SHAP/LIME to satisfy stakeholder trust and GDPR needs.”
Principal AI/ML Leader specializing in Generative AI, MLOps, and NLP
“Founding member of Tausight, building AI systems to detect and protect PHI for healthcare organizations; helped take the company through post–Series A funding and exited after ~6 years. Drove a strategic collaboration with Intel’s OpenVINO team—becoming the first to deploy it in a real production system and improving model performance by ~30% on customer Intel-CPU machines.”
Mid-level Data Scientist specializing in Generative AI, NLP, and MLOps
“Built and deployed an LLM-powered claims-document summarization system (insurance domain) that cut agent review time from 4–5 minutes to under 2 minutes and saved 1,200+ hours per quarter. Hands-on across orchestration and production infrastructure (Airflow retraining DAGs, Kubernetes, SageMaker endpoints, FastAPI) and recent RAG workflows using n8n + Pinecone, with a strong focus on reliability, cost, and explainability for non-technical stakeholders.”
Mid-level AI/ML Engineer specializing in deep learning, MLOps, and LLM applications
“Built and deployed production LLM assistants for internal Q&A and customer-feedback summarization, emphasizing reliability (RAG, prompt tuning, validation/whitelisting) and privacy safeguards. Improved adoption by adding explainable outputs and a user feedback mechanism, and has hands-on orchestration experience with Aflow and Azure Logic Apps.”
Intern Machine Learning & Full-Stack Engineer specializing in computer vision and healthcare AI
“AI/ML-focused backend engineer who shipped two production systems: PersonaPal (agentic LLM chatbot with RAG, FAISS-based retrieval, and Redis semantic caching) and CervixScan (clinical diagnostics platform with PostgreSQL data modeling and human-in-the-loop safety for low-confidence predictions). Demonstrates strong performance/reliability work (indexed vector search, caching, query optimization to ~200ms) and end-to-end ownership from orchestration design through deployment.”
Mid-level QA Engineer specializing in AI/ML model validation and data quality
“ML practitioner with a QA background who has built end-to-end ML pipelines for a health risk prediction use case (lifestyle + demographics), emphasizing robustness through strict data validation, leakage prevention, and cross-validation. Collaborated with a dietician to sanity-check predictions and refine feature interpretation for real-world practicality; has not yet deployed LLM/AI systems to production and has no hands-on orchestration framework experience but is willing to learn.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMs, and NLP
“AI/ML engineer with forensic analytics and healthcare claims experience (Optum), building production LLM/RAG systems to surface context-driven fraud patterns from unstructured claim notes and explain risk to investigators. Strong in large-scale retrieval performance tuning, legacy API integration with reliability patterns (SQS, circuit breakers), and MLOps orchestration on Airflow/Kubernetes with rigorous testing, monitoring, and stakeholder-friendly interpretability.”
Mid-level AI/ML Engineer specializing in GenAI and financial risk & compliance analytics
“Built and deployed a production LLM-powered financial risk and compliance platform to reduce manual trade exception handling and speed up insights from regulatory documents. Implemented a LangChain multi-agent workflow with structured/unstructured data integration (Redshift + vector DB) and emphasized hallucination reduction for regulatory safety using Amazon Bedrock. Strong MLOps/orchestration background across Kubernetes, Airflow, Jenkins, and monitoring/testing with MLflow, Evidently AI, and PyTest.”
Mid-level Data Scientist specializing in predictive modeling, NLP/LLMs, and RAG search systems
“Built production LLM/RAG platforms for financial services to enable natural-language Q&A over large policy/compliance document sets stored in Snowflake and SharePoint. Strong in MLOps and orchestration (Airflow, ADF, Step Functions, MLflow) and in solving real production issues like stale embeddings and model performance, including an incremental Snowflake Streams sync that cut processing time from hours to minutes.”
Senior AI Engineer specializing in Generative AI, NLP, and applied deep learning
“Built a production multi-agent LLM system at Live Nation on Databricks (LangGraph/LangChain) that let venue/event teams ask questions in Slack, auto-generated optimized route schedules, and produced inventory/stocking recommendations from historical SQL data and venue trends. Improved reliability by tightening prompts with strict JSON schemas, providing sample questions/SQL, and adding guardrails plus synthetic/edge-case testing, while iterating with event managers and senior VPs via prototypes and feedback loops.”
Intern Full-Stack/Backend Software Engineer specializing in test automation and web systems
“Backend/ML engineer who built an end-to-end greenwashing detection system for corporate ESG reports: Python preprocessing pipeline, logistic regression + fine-tuned DistilBERT models, and a Dockerized FastAPI inference service optimized for latency. Internship experience maintaining GitLab CI/CD for TypeScript services (Jest/Playwright), improving pipeline stability and test determinism; familiar with Kubernetes/GitOps concepts and AWS CLI/SSO.”
Mid-level Data Scientist & AI/ML Engineer specializing in GenAI and cloud ML
“GenAI/LLM engineer who recently built a production compliance assistant at State Farm for KYC/AML and regulatory teams, using AWS Bedrock + LangChain with Textract/Lambda pipelines to extract fields, tag risk, and summarize long documents. Implemented RAG, strict structured outputs, and human-in-the-loop guardrails, and reports automating ~80% of documentation work while reducing review time by ~40%.”
Senior AI/ML Engineer and Data Scientist specializing in Generative AI and MLOps
“ML/NLP practitioner focused on financial-services document intelligence and compliance workflows—built an end-to-end pipeline to classify documents and extract financial entities from loan applications, emails, and statements stored in S3/internal databases. Strong in entity resolution/record linkage and in productionizing pipelines with GitHub Actions CI/CD, testing, data validation, and Docker, plus semantic search using OpenAI embeddings and a vector database.”
Mid-level Data Analyst specializing in cloud ETL, BI, and machine learning
“Data/ML practitioner with experience at UnitedHealth Group building a fraud claims detection solution combining structured claims data and unstructured notes, validated with compliance stakeholders to improve actionable accuracy. Also applied embeddings, vector databases, and fine-tuned language models in a Bank of America capstone to detect threats/anomalies in financial documents, with production-minded Python ETL workflows using Airflow.”
Principal Data Scientist specializing in Generative AI, NLP, and MLOps
“ML/NLP practitioner with banking experience (M&T Bank) who has built a GPT-4 RAG system using LangChain and Pinecone to connect unstructured customer data with internal knowledge bases, improving accuracy and reducing manual lookup time by 50%+. Strong in entity resolution and productionizing scalable Python data workflows, including major performance wins by migrating bottleneck joins from Pandas to Dask.”