Pre-screened and vetted.
Junior Data Scientist specializing in fraud analytics and cloud data platforms
“Built and deployed production LLM-powered document summarization/classification systems using embeddings, vector databases (RAG-style retrieval), and automated evaluation (BERTScore/ROUGE), with a focus on monitoring and scalable cloud pipelines. Also partnered with a fraud analytics team to deliver a transaction anomaly detection solution, translating model outputs into Power BI dashboards and actionable KPIs while iterating on thresholds and alerts based on stakeholder feedback.”
Senior Data Analyst specializing in cloud data platforms, experimentation, and predictive analytics
“Healthcare data/ML practitioner with experience at UnitedHealth Group building production ETL and streaming pipelines (Python, BigQuery, Kafka) that unify EHR, IoT device, and lab data for patient risk prediction. Also implemented embedding-based semantic search/linking for noisy clinical notes via domain adaptation and rigorous validation with clinical stakeholders; previously built churn prediction at DirecTV using XGBoost.”
Mid-level Data Engineer specializing in cloud data platforms, Spark, and streaming pipelines
“Data/MLOps engineer (Cognizant background) who owned an AWS/Airflow/Snowflake healthcare transactions pipeline processing ~8–10M records/day and cut pipeline/data-quality incidents by ~33%. Also built and deployed a production FastAPI model-inference service on Kubernetes (Docker, HPA) with strong observability (Prometheus/Grafana), versioned endpoints, and resilient backfill/idempotent external data ingestion patterns.”
Mid-level Data Engineer specializing in cloud data pipelines and streaming
“Data engineer with experience at Wells Fargo and Accenture owning end-to-end production pipelines processing hundreds of millions of transactional/risk records daily. Strong focus on data quality and reliability (reconciliation checks, schema drift detection, CloudWatch alerting) plus Spark performance tuning and idempotent backfills using Delta Lake/merge logic across AWS (S3/EMR/Databricks/Redshift) and Azure (ADF/Azure DevOps/Azure Monitor).”
Mid-level Data Engineer specializing in AWS/Azure pipelines and streaming analytics
“Data engineer with experience across healthcare and geospatial risk systems, owning end-to-end pipelines from ingestion through serving on AWS/Azure stacks. Built HIPAA-compliant data quality gates and CDC for millions of daily claims, and also delivered a real-time wildfire risk platform with 20-minute refresh cycles and a 60% data accuracy lift. Strong in streaming (Kafka), Spark performance tuning, and production-grade orchestration/CI/CD (Airflow, Docker, Jenkins, GitHub Actions, Terraform).”
Mid-level AI/ML Engineer specializing in fraud detection and risk analytics in Financial Services
“Finance-domain ML/LLM engineer who has shipped production systems including a RAG-based financial insights assistant with a custom post-generation validation layer that verifies atomic claims against retrieved source text to prevent hallucinations in compliance-critical workflows. Also built large-scale MLOps automation on AWS using Kubeflow + MLflow + CI/CD for fraud detection and credit risk models processing 500M+ transactions/day with a 99.99% uptime goal, and partnered closely with JP Morgan risk/compliance stakeholders on NLP-driven compliance monitoring.”
Mid-level Data Analyst specializing in healthcare and financial analytics
“Built and productionized an LLM-powered clinical documentation and insights pipeline at Cardinal Health using LangChain + GPT-4 with RAG to summarize long clinical notes, extract medication/dosage entities, and generate structured SQL-ready outputs for downstream analytics. Emphasizes clinical reliability via labeled benchmarking (precision/recall/F1), shadow deployments, clinician human-in-the-loop review, and ongoing monitoring/orchestration with Airflow, Lambda, S3, Postgres, and Power BI.”
Mid-level Data Engineer specializing in cloud ETL and financial data platforms
“Data engineer with experience at Capital One and HSBC building and operating GCP-based data platforms. Led an end-to-end Oracle-to-BigQuery migration processing ~200–300GB/day using Dataflow/Beam, Airflow, Dataproc/PySpark, and Looker, achieving ~99.5% pipeline success and ~30% fewer data quality issues. Strong in production reliability, schema drift handling for external APIs, and BigQuery performance/serving patterns (materialized views, authorized views, versioned datasets).”
Mid-level Data Analyst specializing in healthcare and financial analytics
“Analytics-focused candidate with hands-on experience turning messy CRM, e-commerce, payments, and support data into trusted reporting datasets using SQL and Python. They have owned end-to-end churn and retention analytics work, including RFM-based segmentation, dashboard delivery, and metric standardization across sales, marketing, and finance.”
Senior Software Engineer specializing in full-stack SaaS and cloud systems
“Built and owned Laserfiche's Schedule Custom Report platform end-to-end, spanning Angular frontend, backend orchestration, and production observability for an enterprise reporting workflow. Stands out for integrating multiple independent microservices with RabbitMQ-based event orchestration while also improving UX on complex migration tooling.”
Mid-level Generative AI Engineer specializing in LLMs, RAG, and multimodal generation
“Open-source JavaScript contributor focused on performance and maintainability in data visualization libraries—refactored legacy ES5 into modular ES6, added tests/docs, and delivered ~30% faster load times with positive community adoption. Also optimized a React dashboard (~40% load-time reduction) and took ownership in an ambiguous AI product initiative by setting milestones, standing up an initial ML pipeline, and shipping a prototype in ~6 weeks that became the basis for production.”
“ML/GenAI engineer with recent CVS Health experience building a production RAG system over unstructured financial/research documents using LangChain, FAISS, and Pinecone, plus LoRA/PEFT fine-tuning of GPT/LLaMA for domain-aware summarization. Demonstrates strong applied MLOps and data engineering skills (Airflow/Prefect, Docker/Kubernetes, CI/CD, MLflow) and measurable impact (sub-second retrieval, ~40% better context retrieval, ~25% entity matching improvement).”
Mid-level Data Engineer specializing in cloud ETL/ELT and healthcare analytics
“Healthcare-focused data engineer/ML practitioner with experience at Lightbeam Health Solutions and Humana building production entity-resolution and semantic similarity pipelines across EMR, lab, and claims data. Uses NLP/ML (spaCy, scikit-learn, BioBERT/LightGBM) plus Snowflake/Airflow and vector search (Pinecone) to improve linkage accuracy (reported 90%) and semantic match quality (reported +12–15%), while reducing manual cleanup by 40%+.”
Senior Site Reliability Engineer specializing in multi-cloud Kubernetes and DevSecOps
“Cloud/Kubernetes-focused production engineer with experience running 99.95% uptime platforms across AWS/Azure/GCP. Strong in incident response and performance troubleshooting (including a 30% MTTR reduction), and in building secure CI/CD and Terraform-based IaC for AKS/GKE microservices with robust change controls and rollback practices. Notably does not have direct IBM Power/AIX/VIOS/HMC or PowerHA/HACMP ownership.”
Senior AI/ML Engineer specializing in Generative AI, RAG, and agentic systems
“GenAI/LLM ML engineer (currently at Webprobo) building an enterprise GenAI platform with document intelligence and automation on AWS and blockchain. Has hands-on experience with RAG, LLM evaluation tooling, and orchestrating production LLM workflows with Apache Airflow, plus deep exposure to reliability challenges in globally distributed/edge deployments. Also partnered with business/marketing stakeholders at a banking client to deliver an AI-driven customer retention insights solution.”
Engineering leader specializing in FinTech ML/AI platforms
“Engineering Manager/player-coach leading Data Infrastructure, ML/DS, and AI Engineering pods who recently shipped multiple production agentic GenAI features. Built privacy-preserving LLM workflows (PII redaction via Microsoft Presidio) and drove an AI expense-approval agent from ambiguous ask to GA, cutting approval time from ~2.5 days to <4 hours with >85% accuracy. Also owned a major LLM cost overrun incident and implemented cost observability plus circuit breakers to prevent runaway agent loops.”
Senior Data Analyst specializing in data pipelines, web scraping, and legal data enrichment
“Data engineer focused on reliable, scalable analytics pipelines and external data collection. Has owned end-to-end pipelines processing 5–10M records/day, serving Snowflake data marts to Power BI/Tableau, and reports ~99% reliability through strong validation/monitoring. Also shipped versioned REST APIs for curated data with query optimization and caching.”
Mid-level Java Full-Stack Developer specializing in banking and telecom platforms
“Frontend-focused engineer with experience at T-Mobile and U.S. Bank who maintained a TypeScript utility library (types, tests, build pipeline, and docs) adopted by multiple teams, and improved React workflow performance by refactoring components and optimizing data fetching. Known for pragmatic cross-team support—reproducing issues quickly, shipping well-tested fixes, and managing changes carefully to avoid breaking downstream apps.”
Mid-level Data Scientist specializing in Generative AI, MLOps, and cloud data platforms
“GenAI/ML engineer (CitiusTech) who has deployed production RAG systems for compliance/operations document Q&A, using Pinecone + FastAPI microservices on Kubernetes with strong monitoring and guardrails. Also built a GenAI-powered incident triage/routing solution in collaboration with non-technical stakeholders, achieving 35% faster response times and 40% fewer misclassified tickets, and has hands-on orchestration experience with Airflow and AutoSys.”
Mid-level Data Scientist / ML Engineer specializing in secure GenAI and financial compliance
“Built a production "sentinel insight engine" to tame information overload from millions of product reviews and support transcripts, combining Azure OpenAI (GPT-3.5) zero-shot classification with a fine-tuned T5 summarizer to generate weekly actionable product insights. Demonstrated strong MLOps/production engineering by adding drift monitoring with embedding-based detection, integrating REST with legacy SOAP/queue-based CRM via FastAPI middleware, and scaling reliably on Kubernetes with HPA.”
Mid-level Data Scientist & AI Engineer specializing in RAG, agentic AI, and production ML
“AI/data engineer who built a production LLM-powered schema drift detection system (LangChain/LangGraph) to catch semantic data changes before they break downstream analytics/ML. Deployed on AWS with Docker/S3 and implemented an LLM-as-a-judge evaluation framework to improve trust, reduce hallucinations, and control false positives/alert fatigue. Collaborated with non-technical risk/business analytics stakeholders at EY by delivering human-readable drift explanations that improved confidence in financial analytics dashboards.”
Senior Data & Backend Engineer specializing in cloud data pipelines and LLM/RAG systems
“Data engineer with end-to-end ownership of large-scale retail and clinical data ingestion/processing on AWS, including real-time streaming and batch pipelines. Delivered measurable outcomes: 20M daily transactions processed, latency cut from 4 hours to 5 minutes, ~70% fewer failures, and 120+ pipelines running at 99.8% reliability with full audit compliance.”
Mid-level Data Engineer specializing in big data pipelines and real-time streaming
“Data engineer who has owned end-to-end production pipelines processing a few million records/day, using Python/Airflow/SQL/PySpark with Snowflake serving to BI (Power BI). Built resilient external web data collection systems (anti-bot, schema-change detection, backfills) and shipped versioned REST APIs for internal consumers, improving pipeline success rates to 99% through monitoring, retries, and idempotent design.”
Mid-Level Data Engineer specializing in cloud data platforms and governed analytics
“Data engineer with Optum experience building end-to-end healthcare data pipelines for HL7/FHIR, processing millions of records daily across Kafka streaming and Databricks/Spark batch. Strong focus on data quality (schema enforcement/validations), reliability (Airflow monitoring/alerts), and analytics-ready serving in Snowflake powering Power BI/Tableau, with CI/CD via Git and Jenkins.”