Pre-screened and vetted.
Mid-level Data Engineer specializing in streaming and cloud data platforms for financial services
“Data engineering-focused candidate (internship/project experience) who built end-to-end pipelines processing a few million transactional records/day for fraud detection and reporting, using Airflow, Python/SQL, and PySpark with strong emphasis on data quality gates, idempotency, and monitoring. Also implemented an external web/API data collection system with anti-bot tactics and schema-change quarantine, and shipped a versioned Flask API to serve curated warehouse data.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Mid-level Full-Stack Engineer specializing in AI/ML data platforms for biotech and FinTech
“AI/ML full-stack practitioner in a small-scale manufacturing/lab operations environment who deployed a production ML system to improve blood cell order fulfillment by predicting yield/success from donor characteristics. Experienced building custom multi-agent orchestration (Python, LangChain/LangGraph, MCP) and balancing reliability, data quality constraints, and token/ROI economics while communicating tradeoffs to VP-level business stakeholders.”
Senior Data Engineer specializing in Databricks, Spark, and AWS for government healthcare data systems
“Python/AWS engineer focused on batch-processing and data workflows, including building reusable S3/boto3 utilities with reliability features and IAM-based auth. Has led low-risk legacy modernizations using parity testing plus a month of parallel production runs, and has owned production issues end-to-end (including fixing a client-side Excel macro) while contributing to significant AWS cost reductions (~$10k/month).”
Mid-level Machine Learning Engineer specializing in forecasting, NLP, and GenAI
“GenAI/ML engineer with production experience building multilingual LLM systems (English/Spanish) and RAG-based clinical documentation summarization at Walgreens, combining prompt engineering, structured output validation, and rigorous evaluation (ROUGE + pharmacist review). Also orchestrated end-to-end ML pipelines for demand forecasting using Apache Airflow, PySpark, and MLflow with scheduled retraining and production monitoring.”
Mid-level Data Engineer specializing in cloud data warehousing and analytics
“Data engineer at American Express who owned end-to-end pipelines for transaction and customer data used in finance reporting and risk analytics, processing ~5–8M records/day. Built Airflow-orchestrated ingestion (including external APIs/web sources) with strong data quality controls, monitoring/alerts, and resilient backfill/retry patterns, and also shipped a versioned REST API serving aggregated metrics to analytics teams.”
Senior Big Data Engineer specializing in AML/KYC compliance and cloud data platforms
“Data engineer with experience delivering an end-to-end pipeline handling ~3.5TB in a star-schema setup (fact + dimensions) and producing business-facing tables in Hive/Spark. Identified and resolved UAT-reported duplicate issues caused by joins through root-cause analysis, and also built automation to run Spark SQL metrics on weekly/monthly/quarterly cadences and distribute results to users.”
Senior Data Engineer specializing in cloud lakehouse platforms and streaming analytics
“Data engineer focused on fraud and banking analytics who has owned end-to-end batch + streaming pipelines at very large scale (hundreds of millions of records/day). Built robust data quality/observability layers (schema validation, anomaly detection, alerting) and delivered low-latency serving via AWS Lambda/API Gateway with DynamoDB + Redis, plus external data ingestion/scraping pipelines orchestrated in Airflow with anti-bot protections.”
Mid-level Data Engineer specializing in Azure ETL/ELT and data warehousing
“Data engineer who has owned end-to-end production pipelines for customer transaction data (~2–5 GB/day) using Python/PySpark/SQL and Airflow, delivering major reliability and speed gains (70% faster reporting; 60–70% fewer data issues). Also built a daily external web-scraping system with anti-bot handling and safe, idempotent Airflow-driven backfills, plus a Python data API optimized with indexing/caching and tested for correctness.”
Mid-level Data Engineer specializing in scalable pipelines, Spark, and cloud data warehousing
“Backend/data platform engineer who recently owned an end-to-end large-scale financial data platform delivering real-time decision support for finance and operations. Has hands-on experience modernizing legacy batch pipelines into AWS cloud-native ELT with parallel-run cutovers, strong data quality controls (dbt-style tests, reconciliation), and measurable improvements in runtime, cost, and SLA compliance. Also builds scalable, secure FastAPI microservices using Docker, ALB-based horizontal scaling, Redis caching, and managed auth with Cognito/Supabase plus Postgres RLS.”
Mid-level Data Engineer specializing in real-time pipelines and cloud analytics
“Researcher from the University of South Dakota who built a production medical RAG system to help interpret model predictions by retrieving relevant clinical notes and medical literature, overcoming retrieval accuracy and imaging-dataset challenges through semantic chunking and metadata-driven indexing. Also has hands-on orchestration experience with Airflow and Azure Data Factory, plus a pragmatic approach to LLM evaluation and stakeholder-driven iteration.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and MLOps
“AI/ML engineer who has shipped production AI systems end-to-end, including an automated multi-channel (Gmail/WhatsApp/voice) candidate interviewing workflow and an enterprise RAG knowledge search platform. Demonstrates strong production rigor (monitoring, A/B tests, guardrails, schema validation, shadow testing) with quantified impact: ~60–70% reduction in interview evaluation time and ~20–30% relevance gains in RAG retrieval.”
Senior Data Engineer specializing in Azure Lakehouse, Databricks/Spark, and Snowflake
“Data engineer/platform builder with experience across PwC and Liberty Mutual delivering high-volume, production-grade pipelines and real-time data services. Has owned end-to-end streaming + batch architectures on AWS and Azure, including web scraping systems, with quantified reliability gains (99.9% availability, 90%+ error reduction, 30% latency reduction) and strong observability/CI-CD practices.”
Senior AI/ML Engineer specializing in Generative AI and LLM platforms
“Backend engineer focused on multi-tenant enterprise AI personalization and recommendation platforms, combining ML/LLM intent extraction with deterministic policy guardrails for compliance and auditability. Has hands-on AWS experience (ECS/Lambda/DynamoDB/S3) and led a careful DynamoDB single-table migration using dual write/read, canary + feature-flag rollouts, and strong observability/security (JWT/OAuth2, RBAC, Postgres RLS).”
Mid-level Data Engineer specializing in cloud data pipelines and real-time streaming
“Data engineer with PNC Bank experience owning high-volume financial transaction pipelines end-to-end (Kafka/REST ingestion through Spark/Glue transformations to Redshift serving) for risk and fraud analytics. Built strong reliability and data quality practices (Great Expectations, reconciliation, Airflow alerting, idempotent retries, incremental/windowed processing), reporting 40% ingestion efficiency gains and ~99.9% data accuracy.”
Mid-level Business Analyst specializing in healthcare and data analytics
“Analytics candidate with hands-on experience at BCBS building HIPAA-compliant SQL/Snowflake/Tableau pipelines across fragmented legacy healthcare systems. Stands out for turning a 5-day claims reporting process into a near real-time 10-minute dashboard and for pairing strong data engineering discipline with reproducible Python-based churn modeling that drove measurable retention outcomes.”
Mid-level Data Engineer specializing in cloud data platforms and AI/ML pipelines
“Data-engineering-oriented candidate with hands-on experience building an agentic AI product and operational automation workflows. They described automating inventory-to-ERP discrepancy reconciliation with anomaly detection and daily reporting, and also have practical scraping/automation experience dealing with Cloudflare-protected sites using Selenium and Puppeteer.”
Executive technology leader specializing in model risk and regulatory technology
“Candidate is pursuing a CTO role and has helped multiple startups turn early technology concepts into concrete, real-world technical requirements. They cite a systems science and mathematics background, along with experience at JPMorgan Chase, and appear strongest in technical strategy, concept fleshing, and identifying strong people to help teams succeed.”
Junior Full-Stack Software Engineer specializing in AI data systems
“Full-stack engineer with strong DevOps/AWS production experience who builds and operates multi-agent AI systems end-to-end (Streamlit/Python through Docker/Kubernetes and ECS/Fargate). Has delivered measurable outcomes: sub-2s latency and ~92% routing accuracy for an AI wellness assistant, shipped an AI-for-BI prototype in under 6 weeks cutting analysis time ~40%, and improved pipeline iteration speed ~35% via modularization and CI/regression checks.”
Senior Data Engineer specializing in cloud data platforms and regulated analytics
“Data engineer at Capital One building AWS-based real-time and batch pipelines and backend data services for financial/fraud use cases. Has owned end-to-end pipelines processing millions of records/day, implemented dbt/Great Expectations quality gates, and tuned Redshift/Snowflake workloads (cutting query latency ~22–25% and reducing pipeline failures ~30–40%) while supporting 15+ downstream consumers.”
Senior Engineering Manager specializing in platform, data/ML, and identity/access systems
“Senior engineering leader from Goodyear’s AndGo startup-like division who scaled the org from 12 to 30+ across pod-based teams and introduced an Architect Guild/ARD governance model. Led a 4-month Europe launch requiring AWS regional infrastructure, GDPR compliance, i18n/l10n, and new EMEA reporting pipelines, and has hands-on depth in API performance, incident response, and GraphQL/Hasura adoption to boost product velocity.”
Mid-level Software Engineer specializing in backend systems, cloud-native apps, and AI platforms
“Backend/full-stack engineer who has owned production systems end-to-end, including a Dockerized Node.js/TypeScript probabilistic fault-tree analysis service for nuclear safety research deployed on AWS. Also built and operated a FastAPI-based RAG pipeline over 200+ PDFs using FAISS, focusing on low-latency, idempotent workflows and strong observability; experienced with API design and Playwright E2E automation across React/Angular projects.”
Senior Business Analytics Consultant specializing in BI, data engineering, and predictive analytics
“Healthcare analytics candidate with hands-on experience turning messy claims, enrollment, and reference data into trusted SQL reporting layers and reproducible Python workflows. They emphasize metric standardization, stakeholder alignment, and operational impact, including ~40% reduction in manual reporting effort and improved forecasting/resource prioritization through high-risk patient segmentation.”