Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in NLP and Generative AI
“Built and deployed a production LLM-powered RAG assistant for healthcare teams (care managers/support) to answer questions from clinical and policy documentation, emphasizing trustworthiness via improved retrieval, reranking, and strict grounding prompts to reduce hallucinations. Also has hands-on orchestration experience with Apache Airflow for end-to-end ETL/ML workflows and applies rigorous testing/metrics (hallucination rate, tool-call accuracy, latency, cost) to ensure reliable AI agent behavior.”
Mid-level Data Scientist specializing in LLMs, RAG, and document intelligence
“LLM/ML engineer who has shipped production systems in legal/financial-risk domains at Wolters Kluwer, including a hybrid OCR+deterministic+LLM extraction pipeline that structured UCC filings at massive scale and drove $6M+ in revenue. Also built LangGraph-based multi-agent “Deep Research” workflows with model routing, tool calls (MCP), persistence, and human-in-the-loop review, and partnered closely with policy writers to deliver LLM summarization that cut writing time by ~60%.”
Mid-level Data Scientist specializing in predictive analytics and LLM-powered data pipelines
“Early-career engineer from BNP Paribas who drove a large-scale observability modernization—selecting and implementing Prometheus/Grafana for a 2000+ server estate, then productionizing it on Kubernetes via Docker/Jenkins. Known for hands-on demos, strong documentation/templates, and pragmatic troubleshooting (including custom Python metrics) that improved visibility and cut debugging time by ~60%.”
Junior AI/ML Engineer specializing in cloud-native LLM systems and RAG
“AI/LLM engineer who has shipped production RAG copilots and multi-agent workflows, including a real-time Llama3 (Ollama) copilot backend handling 12k+ concurrent queries at 99.9% uptime. Deep on orchestration (Langflow/Airflow/Kubernetes), reliability evaluation (hallucination detection, p95 latency, token cost), and monitoring (Prometheus/Grafana), with demonstrated stakeholder-facing analytics delivery via Tableau.”
Senior Data Engineer specializing in cloud data platforms and ML pipelines
“Built and deployed AcademiQ Ai, a production LLM-based teaching assistant using GPT/BERT with RAG (LangChain + Pinecone) to handle large student notes and generate adaptive explanations/quizzes. Demonstrated measurable retrieval-quality gains (18% precision improvement, 22% less irrelevant context) by tuning similarity thresholds and chunking based on user satisfaction signals. Also orchestrated terabyte-scale, real-time demand forecasting pipelines using Airflow and Kubeflow on GCP with strong monitoring, shadow deployment, and feedback-loop practices.”
Mid-level AI/ML Engineer specializing in healthcare NLP and MLOps
“ML/AI engineer with healthcare payer experience (Signal Healthcare, Cigna) who has shipped production fraud/claims prediction systems using Python/TensorFlow and exposed them via FastAPI/Flask microservices integrated with EHR and Salesforce. Emphasizes operational reliability and trust—Airflow-orchestrated pipelines with data quality gates plus SHAP-based interpretability, A/B testing, and drift/debug workflows—backed by reported outcomes of 22% lower false payouts and 17% higher model accuracy.”
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and MLOps
“LLM/agentic systems engineer who built a production "Agentic AI Diagnostic Assistant" for network engineers, using a multi-agent Llama 2 + LangChain architecture with RAG over telemetry/incident data in DynamoDB and confidence-based deferrals to reduce hallucinations. Also has strong MLOps/orchestration experience (Airflow, EventBridge, Spark, Docker, SageMaker/ECS) at multi-terabyte/day scale and delivered multilingual NLP analytics (fine-tuned BERT/spaCy) for support operations through hands-on stakeholder workshops.”
Mid-level Full-Stack Developer specializing in web platforms and cloud (AWS)
“Full-stack engineer with financial services experience (Lincoln Financial) who owned a customer-facing financial portal end-to-end using TypeScript/React and Node/Express. Has hands-on microservices and RabbitMQ event-driven workflows, addressing scale issues like retries/duplicates with idempotency and traceable logging, and built an internal real-time ops/support dashboard to improve monitoring and incident response.”
Mid-level Data Scientist specializing in predictive modeling, NLP/LLMs, and RAG search systems
“Built production LLM/RAG platforms for financial services to enable natural-language Q&A over large policy/compliance document sets stored in Snowflake and SharePoint. Strong in MLOps and orchestration (Airflow, ADF, Step Functions, MLflow) and in solving real production issues like stale embeddings and model performance, including an incremental Snowflake Streams sync that cut processing time from hours to minutes.”
Mid-level Machine Learning Engineer specializing in NLP, LLMs, and MLOps
“Built a production internal LLM/RAG assistant at CVS Health to cut time spent searching long policy and clinical guideline PDFs, combining fine-tuned BERT/GPT models with FAISS retrieval and a FastAPI service on AWS. Demonstrates strong real-world reliability work (document cleanup, hallucination controls, monitoring/drift tracking with MLflow) and close collaboration with non-technical clinical operations teams via demos and feedback-driven iteration.”
Mid-level AI/ML Engineer specializing in Generative AI, NLP, and healthcare RAG systems
“Built and deployed a production clinical claim validation RAG system at GE HealthCare that automated nurses’ patient-history/claims checks, cutting manual review time by ~65%. Designed the full stack (retrieval, embeddings, Pinecone, prompt/verification guardrails, FastAPI backend) with PHI-compliant anonymization via NER and orchestrated pipelines using Airflow, Azure ML Pipelines, and MLflow with drift monitoring.”
Mid-level AI Engineer specializing in healthcare claims analytics and RAG copilots
“Built a production "appeals co-pilot" for a healthcare claims appeals team, combining an XGBoost/logistic ranking model with a Python/LangChain RAG stack (FAISS + Mistral 7B) to surface high-probability appeal wins and speed policy-grounded drafting. Emphasizes reliability and trust: hybrid retrieval with metadata routing, citation/eval scripts, guardrails, and an explainability layer that non-technical stakeholders could understand and override.”
Junior ML Data Associate specializing in AI training data and LLM prompt evaluation
“Applied ML/embodied AI practitioner who built an on-device gesture-control system for smart-home lights using Raspberry Pi + camera, focusing on privacy-preserving real-time inference and hardware-constrained optimization (async pipeline + TF Lite INT8). Also made a high-impact architecture decision for an ML content evaluation/QA pipeline processing millions of annotated text samples weekly, reducing batch runtime from ~6 hours to ~40 minutes while lowering compute cost.”
Mid-level AI/ML Engineer specializing in computer vision, NLP/LLMs, and MLOps
“ML/AI engineer with defense and commercial analytics experience: deployed a real-time aerial object detection system at Dynetics (YOLOv5 + TorchServe in Docker on AWS EC2) with drift-triggered retraining and 99.5% uptime, tackling ambiguous targets and weather degradation. Previously at Fractal Analytics, built and explained a churn prediction model for marketing stakeholders using SHAP and delivered it via a Flask API into dashboards, driving a reported 22% attrition reduction.”
Senior AI/ML Engineer and Data Scientist specializing in Generative AI and MLOps
“ML/NLP practitioner focused on financial-services document intelligence and compliance workflows—built an end-to-end pipeline to classify documents and extract financial entities from loan applications, emails, and statements stored in S3/internal databases. Strong in entity resolution/record linkage and in productionizing pipelines with GitHub Actions CI/CD, testing, data validation, and Docker, plus semantic search using OpenAI embeddings and a vector database.”
Mid-level Data Analyst specializing in cloud ETL, BI, and machine learning
“Data/ML practitioner with experience at UnitedHealth Group building a fraud claims detection solution combining structured claims data and unstructured notes, validated with compliance stakeholders to improve actionable accuracy. Also applied embeddings, vector databases, and fine-tuned language models in a Bank of America capstone to detect threats/anomalies in financial documents, with production-minded Python ETL workflows using Airflow.”
Senior Data Scientist specializing in NLP and explainable machine learning
“NLP/ML practitioner who built an explainable, clinician-aligned system to detect cognitive decline (Alzheimer’s/stroke-related) from audio responses, achieving 97% accuracy on only a few hundred data points. Also has experience with healthcare claims entity resolution and prototyped a word2vec-based patent search vector database in Elasticsearch, with strong emphasis on testing, interpretability, and scalable Python data workflows.”
Principal Data Scientist specializing in Generative AI, NLP, and MLOps
“ML/NLP practitioner with banking experience (M&T Bank) who has built a GPT-4 RAG system using LangChain and Pinecone to connect unstructured customer data with internal knowledge bases, improving accuracy and reducing manual lookup time by 50%+. Strong in entity resolution and productionizing scalable Python data workflows, including major performance wins by migrating bottleneck joins from Pandas to Dask.”
Mid-Level Software Engineer specializing in data engineering and cloud platforms
“Backend-leaning full-stack engineer who has shipped production-critical data/reporting features at Walmart and built an end-to-end workflow automation product (FastAPI + React/TypeScript + PostgreSQL) deployed on AWS. Strong in performance/reliability engineering (parallel ETL, batch DB operations, indexing via EXPLAIN ANALYZE), secure API design (JWT/RBAC), and pragmatic incident-driven scaling (separating workers from API layer).”
Mid-level Full-Stack Java Developer specializing in FinTech and real-time systems
“Backend/full-stack engineer with finance domain experience (State Street) who built and shipped a Kafka-based real-time trade validation system handling 50k+ trades/sec and cut latency by 42%. Also delivered real-time React dashboards (Redux Toolkit/React Query/WebSockets) and operates AWS EKS microservices with GitOps/ArgoCD; has built a FastAPI + LangChain/GPT-4 intelligent document processing service with JWT/RBAC.”
Mid-level AI Software Engineer specializing in LLM systems and cloud APIs
“Built and productionized an LLM-powered support/knowledge pipeline using embeddings and retrieval (RAG) to deliver more grounded, higher-quality responses while reducing manual effort. Focused on real-world reliability and performance—adding structured validation/guardrails, optimizing vector search and context size for latency/scale, and monitoring failure patterns in production. Experienced with orchestration via LangChain for LLM workflows and Airflow for production data/ML pipelines, and iterates closely with operations stakeholders through demos and feedback.”
Mid-level Data Scientist / ML Engineer specializing in FinTech and Healthcare ML systems
“AI/LLM engineer who has shipped production RAG systems (including a 250K-document compliance knowledge tool on AWS) and focuses on reliability via citations, guardrails, and rigorous evaluation (Ragas/Opik/DeepEval). Also built a LangGraph-orchestrated webcrawler agent that cut research paper extraction from hours to minutes, and collaborated with clinical teams to deliver patient volume forecasting with an optimization layer for staffing.”
Mid-level Data Engineer specializing in AWS cloud data platforms
“Data engineer with Charter Communications experience modernizing large-scale AWS data lake pipelines: ingesting S3 data, validating against legacy systems, transforming with PySpark/Spark SQL, and serving via Iceberg/Delta tables. Worked at 50M–300M record scale, delivered >99.5% data match, and built monitoring/alerting (CloudWatch/SNS) plus retry orchestration (Step Functions) and data quality gates (Great Expectations).”
Junior Software Engineer specializing in backend, cloud, and data engineering