Pre-screened and vetted.
Mid-level Data Engineer specializing in Analytics & AI/ML
“Data engineer with experience at Sony and Walmart building high-volume, near-real-time analytics and ingestion systems. Has owned end-to-end pipelines from Kafka/Spark streaming through S3/Parquet and Redshift/Looker, emphasizing data quality (Great Expectations), observability (CloudWatch/Azure Monitor), and reliability (Airflow SLAs, retries, checkpointing), including measurable performance and latency improvements.”
Mid-level Data Engineer specializing in Azure ETL/ELT and data warehousing
“Data engineer who has owned end-to-end production pipelines for customer transaction data (~2–5 GB/day) using Python/PySpark/SQL and Airflow, delivering major reliability and speed gains (70% faster reporting; 60–70% fewer data issues). Also built a daily external web-scraping system with anti-bot handling and safe, idempotent Airflow-driven backfills, plus a Python data API optimized with indexing/caching and tested for correctness.”
Mid-level Data Engineer specializing in real-time analytics and regulated domains
“Data platform engineer focused on large-scale, real-time fraud systems, with hands-on ownership of streaming architectures using Kafka, Spark, Snowflake, and Databricks. Stands out for combining performance tuning and platform automation with LLM/RAG-based enrichment, delivering measurable gains in latency, fraud accuracy, false positives, and analyst decision speed.”
Mid-level Data Engineer specializing in scalable pipelines, Spark, and cloud data warehousing
“Backend/data platform engineer who recently owned an end-to-end large-scale financial data platform delivering real-time decision support for finance and operations. Has hands-on experience modernizing legacy batch pipelines into AWS cloud-native ELT with parallel-run cutovers, strong data quality controls (dbt-style tests, reconciliation), and measurable improvements in runtime, cost, and SLA compliance. Also builds scalable, secure FastAPI microservices using Docker, ALB-based horizontal scaling, Redis caching, and managed auth with Cognito/Supabase plus Postgres RLS.”
“Built and productionized an LLM-powered PDF document Q&A system to eliminate manual searching through long documents, focusing on scalability and answer reliability. Implemented semantic chunking (using headings/paragraphs/tables), overlap, and preprocessing/quality checks to reduce hallucinations, and orchestrated the end-to-end pipeline with Airflow using retries, alerts, and parallel tasks.”
Mid-level Data Engineer specializing in real-time pipelines and cloud analytics
“Researcher from the University of South Dakota who built a production medical RAG system to help interpret model predictions by retrieving relevant clinical notes and medical literature, overcoming retrieval accuracy and imaging-dataset challenges through semantic chunking and metadata-driven indexing. Also has hands-on orchestration experience with Airflow and Azure Data Factory, plus a pragmatic approach to LLM evaluation and stakeholder-driven iteration.”
Senior Data Engineer specializing in Azure Lakehouse, Databricks/Spark, and Snowflake
“Data engineer/platform builder with experience across PwC and Liberty Mutual delivering high-volume, production-grade pipelines and real-time data services. Has owned end-to-end streaming + batch architectures on AWS and Azure, including web scraping systems, with quantified reliability gains (99.9% availability, 90%+ error reduction, 30% latency reduction) and strong observability/CI-CD practices.”
Mid-level AI/ML Data Scientist specializing in NLP, computer vision, and risk analytics
“ML/AI engineer with Capital One experience building production-grade customer segmentation and fraud detection systems combining NLP (transformers) and anomaly detection. Strong MLOps and orchestration background (PySpark ETL, MLflow, Airflow, Docker/Kubernetes, Azure ML) with real-time monitoring/alerting and performance optimizations like quantization and caching, plus proven ability to deliver business-facing insights through Power BI/Tableau for marketing stakeholders.”
Mid-level Data Engineer specializing in cloud data pipelines and real-time streaming
“Data engineer with PNC Bank experience owning high-volume financial transaction pipelines end-to-end (Kafka/REST ingestion through Spark/Glue transformations to Redshift serving) for risk and fraud analytics. Built strong reliability and data quality practices (Great Expectations, reconciliation, Airflow alerting, idempotent retries, incremental/windowed processing), reporting 40% ingestion efficiency gains and ~99.9% data accuracy.”
Mid-level Business Analyst specializing in healthcare and data analytics
“Analytics candidate with hands-on experience at BCBS building HIPAA-compliant SQL/Snowflake/Tableau pipelines across fragmented legacy healthcare systems. Stands out for turning a 5-day claims reporting process into a near real-time 10-minute dashboard and for pairing strong data engineering discipline with reproducible Python-based churn modeling that drove measurable retention outcomes.”
Mid-level Business Data Analyst specializing in healthcare analytics
“Analytics-focused candidate with strong SQL, Excel, Python, and Tableau skills who supports payroll-, compensation-, and finance-adjacent processes through rigorous data validation and reconciliation. Stands out for uncovering a duplicate-record mapping issue that exposed roughly $250K in revenue leakage and for building repeatable controls, dashboards, and automated checks to improve reporting accuracy.”
Executive technology leader specializing in model risk and regulatory technology
“Candidate is pursuing a CTO role and has helped multiple startups turn early technology concepts into concrete, real-world technical requirements. They cite a systems science and mathematics background, along with experience at JPMorgan Chase, and appear strongest in technical strategy, concept fleshing, and identifying strong people to help teams succeed.”
Mid-level AI/ML Engineer specializing in LLM agents and workflow automation
“AI/LLM engineer with strong healthcare domain depth who has shipped production-grade agents for care coordination and clinical workflow automation. Stands out for combining Knowledge Graph RAG, LangGraph orchestration, and rigorous eval/guardrail systems to improve reliability in high-stakes environments, with measurable gains in review time, hallucination reduction, latency, and clinician adoption.”
Senior Data Engineer specializing in cloud data platforms and regulated analytics
“Data engineer at Capital One building AWS-based real-time and batch pipelines and backend data services for financial/fraud use cases. Has owned end-to-end pipelines processing millions of records/day, implemented dbt/Great Expectations quality gates, and tuned Redshift/Snowflake workloads (cutting query latency ~22–25% and reducing pipeline failures ~30–40%) while supporting 15+ downstream consumers.”
Mid-level Data Engineer specializing in cloud ETL pipelines (Azure, AWS, GCP)
“Data engineer/backend developer who owned end-to-end pipelines and external data collection systems, including API ingestion and large-scale web scraping. Worked at ~50M records/month scale, improving processing speed by 20% and reducing reporting errors by 15%, and shipped a Rust-based internal data API with versioning, caching, and strong validation/observability practices.”
Mid-level Data Engineer specializing in lakehouse ETL and analytics engineering
“Data engineer with strong end-to-end ownership of production lakehouse pipelines (Snowflake + Databricks + Airflow + dbt + Great Expectations), handling 8M+ records/month and 500K+ daily CDC updates. Delivered measurable reliability and efficiency gains (41% cost reduction, freshness improved from 4h to 30m, 35% fewer downstream incidents) and has experience building a lakehouse platform from scratch across 12 source systems.”
Junior Data Engineer specializing in cloud ETL and big data platforms
“Data engineer focused on transit/transportation datasets, building Spark-based pipelines that ingest from Oracle/APIs, apply PySpark data-quality fixes, and publish star-schema fact tables to Azure Data Lake. Experienced troubleshooting complex Spark failures (using checkpointing to manage long lineage) and operating Airflow-driven backfills and GitLab CI deployments for production DAGs.”
Mid-level AI/ML & GenAI Engineer specializing in LLMs, RAG, and MLOps
“LLM/agent engineer with production experience in healthcare claims automation, delivering large operational impact (cut case handling from ~8–10 minutes to ~3 minutes, ~2,000 staff hours saved/month at ~3,000 claims/month). Built resilient Azure-based deployments (Azure DevOps CI/CD, Docker/FastAPI, Redis caching, autoscaling, observability) and improved reliability via safety/evaluation frameworks that reduced hallucinations by 32%.”
Senior Business Analytics Consultant specializing in BI, data engineering, and predictive analytics
“Healthcare analytics candidate with hands-on experience turning messy claims, enrollment, and reference data into trusted SQL reporting layers and reproducible Python workflows. They emphasize metric standardization, stakeholder alignment, and operational impact, including ~40% reduction in manual reporting effort and improved forecasting/resource prioritization through high-risk patient segmentation.”
Mid-level Data Analyst specializing in banking and product analytics
“Analytics engineer/data analyst with Bank of America experience turning fragmented financial data across SQL Server, PostgreSQL, Kafka, and flat files into trusted Snowflake/dbt reporting models. Stands out for unifying disputed business definitions like churn and payment success rate, automating manual analysis in Python, and pairing strong data quality rigor with stakeholder adoption through self-service dashboards.”
Mid-level AI/ML Engineer specializing in healthcare and financial ML systems
“ML/AI engineer with hands-on experience shipping both predictive healthcare models and clinical GenAI assistants into production. They combine strong MLOps depth across Azure and AWS with healthcare-specific safety thinking, including PHI guardrails, retrieval grounding, and production monitoring, and they also built internal Python tooling for fraud ML workflows at Capital One.”
Mid-level Data Analyst specializing in financial services analytics
Mid-level Business Analyst specializing in financial data and supply chain analytics
Mid-level AI/ML Engineer specializing in cloud MLOps and real-time ML pipelines