Pre-screened and vetted.
Senior Data Engineer specializing in cloud data platforms and ML pipelines
“Data engineer focused on AWS-based enterprise data platforms, owning end-to-end pipelines from multi-source batch/stream ingestion (Glue/Kinesis/StreamSets/Airflow) through PySpark transformations into curated datasets for Redshift/Snowflake. Emphasizes production reliability with strong monitoring/observability and data quality gates, and reports ~30% performance improvement plus improved SLAs and latency after optimization.”
Mid-level Backend Python Engineer specializing in APIs, microservices, and data pipelines
“Backend engineer (Marsh McLennan) who evolved a high-volume claims automation pipeline in Python, emphasizing thin APIs with background job processing, strong validation/retries, and production-grade observability. Experienced in secure FastAPI API design (centralized JWT/RBAC), multi-tenant Postgres/Supabase-style row-level security, and low-risk refactors using parallel runs and feature flags; targeting founding-engineer scope roles.”
Mid-level Data Engineer specializing in cloud data pipelines for healthcare and financial services
“Data engineer with ~4 years of experience (Cigna) building and operating Azure Data Factory pipelines for healthcare claims/member/provider data at 2–3M records/day. Emphasizes reliability and downstream safety via schema/data-quality validation, quarantine workflows, idempotent processing, and backfills; also improved runtime ~20% through SQL optimization and served curated datasets through versioned views and well-documented, analyst-friendly interfaces.”
Mid-level Data Engineer specializing in cloud-native healthcare and enterprise data platforms
“Data Engineer (TCS) who owned an end-to-end CRM analytics pipeline for Bayer’s eSalesWeb integration, ingesting from Salesforce APIs/databases/S3 and serving analytics-ready datasets via PostgreSQL/S3 for Tableau. Drove measurable outcomes: ~60% reduction in manual data-quality effort, ~30% lower latency through SQL optimization, and ~35% improved stability via monitoring, retries, and idempotent processing.”
Senior Python Backend Engineer specializing in scalable APIs and cloud-native microservices
“Backend/data platform engineer who has built and operated a cloud-native media ingestion/processing platform in Python (Django/DRF, FastAPI) with Kafka, Postgres, and Redis, emphasizing multi-tenant security and reliability. Delivered AWS production systems combining EKS and Lambda with Terraform + GitHub Actions/Helm, and built Glue-based ETL pipelines with strong schema-evolution and data-quality practices; also modernized SAS analytics into Python on AWS. Seeking fully remote roles with a $120K–$140K base range.”
Mid-level Machine Learning Engineer specializing in Generative AI and LLM applications
“GenAI engineer who has deployed production LLM/RAG chatbots for internal document search, focusing on reliability (hallucination reduction via prompt guardrails + retrieval filtering) and performance (latency improvements via caching). Experienced with LangChain/LangGraph orchestration for multi-step agent workflows and iterates using monitoring/logs and benchmark-driven evaluation while partnering closely with product and business teams.”
Mid-level Full-Stack Java Developer specializing in cloud-native microservices
“Software engineer with deep healthcare claims domain experience who has owned customer-facing portals end-to-end (Java/Spring Boot + React/TypeScript) and improved usability/performance based on real user feedback. Built microservices using REST and RabbitMQ with strong observability (Splunk/cloud metrics), and delivered an internal claims investigation dashboard that streamlined operations through centralized data, search, and filtering.”
Mid-level Data Scientist specializing in real-time fraud detection and MLOps
“ML/NLP engineer with experience at Charles Schwab building an NLP + graph (Neo4j) entity-resolution system to unify fragmented user/device/transaction data and improve downstream model quality and analyst querying. Has applied embeddings (SentenceTransformers + FAISS) with domain fine-tuning to boost hard-case matching recall by ~12% while maintaining precision, and has a track record of hardening scalable Python/Spark pipelines and productionizing fraud models via A/B tests and shadow-mode monitoring.”
Senior Data & Platform Engineer specializing in cloud-native streaming and distributed systems
“Financial data engineer who has built and operated high-volume batch + streaming pipelines (200–300 GB/day; 5–10k events/sec) using AWS, Spark/Delta, Airflow, Kafka, and Snowflake, with strong emphasis on data quality and reliability. Demonstrated measurable impact via 99.9% SLA adherence, major reductions in bad records/nulls, MTTR improvements, and significant latency/runtime/query performance gains; also built a distributed web-scraping system processing 5–10M records/day with anti-bot and schema-drift defenses.”
Mid-level Data Engineer specializing in cloud ETL and real-time streaming
“Data engineer focused on AWS + Spark/Databricks pipelines, including an end-to-end nightly loan-data ingestion flow (~2.2M records) from Postgres/S3 through Glue and Databricks into a DWH with layered validation and alerting. Also built real-time streaming with Kafka + Spark Structured Streaming and a master’s project streaming Reddit data for sentiment analysis under ambiguous requirements and tight budget constraints.”
Senior Software Engineer specializing in backend systems, AI/LLM integration, and cloud infrastructure
“Backend engineer with experience in highly regulated and high-stakes systems, including an airline crew messaging platform requiring near-zero-error real-time operations and a HIPAA-compliant mental health application built from an early-stage concept. They also show strong operational maturity, having owned a GoDaddy production incident through resolution and then led deployment pipeline improvements that reduced build failures by 40% and doubled deployment frequency.”
Mid-level Data Engineer specializing in Lakehouse, Streaming, and ML/LLM data systems
“Built and productionized an enterprise retrieval-augmented generation platform for internal knowledge over large unstructured corpora, emphasizing trust via strict citation/grounding and hybrid retrieval (BM25 + FAISS + cross-encoder re-ranking). Demonstrates strong scaling and cost/latency optimization through incremental indexing/embedding and index partitioning, plus disciplined evaluation/observability practices. Has experience operationalizing pipelines with Airflow/Databricks/GitHub Actions and partnering closely with risk & compliance stakeholders on auditability requirements.”
Mid-level Data Engineer specializing in cloud data platforms and real-time analytics
“Customer-facing data engineering professional who builds and deploys real-time reporting/dashboard solutions, gathering reporting and compliance requirements through direct stakeholder engagement. Experienced with Google Cloud IAM governance, secure integrations (encryption, audit logging), and fast production troubleshooting of ETL/pipeline failures with follow-on monitoring and automated recovery improvements; motivated by hands-on, travel-oriented customer work.”
Senior Data Engineer specializing in data infrastructure and marketing/CRM analytics
“Salesforce-focused implementation/solutions engineer from Full Circle Insights who owned end-to-end campaign attribution and reporting deployments for multiple customers at once (3–5 concurrently), including sandbox testing, KPI monitoring, and rollback-safe migrations from legacy reporting. Also builds personal multi-agent workflows and uses Claude Code to rapidly scaffold data/analytics scripts like an advertising optimization parser over CSV/XLSX inputs.”
Senior Analytics and Business Intelligence professional specializing in e-commerce and digital analytics
“Analytics professional with hands-on experience unifying marketing-platform data through Fivetran and Snowflake, building reporting views, and catching source-to-report issues like timezone-driven spend discrepancies. They also owned subscription LTV/cohort analysis and engagement tracking initiatives, partnering with e-commerce, product, and senior leadership to turn behavioral and demographic data into dashboards, lead-qualification metrics, and lifecycle marketing insights.”
Mid-level Data Engineer specializing in cloud ETL/ELT and lakehouse architecture
“Data engineer focused on sales/marketing analytics pipelines, owning ingestion from CRMs/ad platforms through warehouse serving and dashboards at ~hundreds of thousands of records/day. Built reliability-focused systems including dbt/SQL/Python data quality gates with alerting, a resilient web-scraping pipeline (retries/backoff, anti-bot tactics, schema-change detection, backfills), and a versioned internal REST API with caching and strong developer usability.”
Intern Data Scientist specializing in ML engineering and LLM agentic workflows
“Built an agentic, multi-step LLM system that generates full-stack code for API integrations using LangChain orchestration, Pinecone/SentenceBERT RAG, and a human-in-the-loop feedback loop for iterative code refinement. Also collaborated with non-technical content writers and PMs during a Contentstack internship to deliver a Slack-based AI workflow that generates and brand-checks articles with one-click approvals.”
Mid-level Data Analyst specializing in business intelligence and cloud data platforms
“Healthcare analytics professional with TCS/Humana experience turning messy claims and eligibility data into reliable reporting assets using SQL and Python. They combine strong data engineering and analytics execution with stakeholder management, including automating monthly claims reporting from half a day to under 5 minutes and driving a provider outreach effort that reduced claim rejection rates by about 20%.”
Mid-Level Software Engineer specializing in cloud-native microservices and data platforms
“Robotics software engineer focused on multi-robot fleet orchestration in ROS 2, owning the fleet manager and task dispatch layer for pick/drop workflows. Strong in real-world reliability and safety (heartbeats, idempotent tasking, E-stop/localization confidence gates) and in debugging timing/state issues via telemetry alignment and rosbag replay, with experience in simulation, CI/CD, Docker, and Kubernetes-based deployments.”
Mid-level Machine Learning Engineer specializing in NLP, LLMs, and MLOps
“Built a production internal LLM/RAG assistant at CVS Health to cut time spent searching long policy and clinical guideline PDFs, combining fine-tuned BERT/GPT models with FAISS retrieval and a FastAPI service on AWS. Demonstrates strong real-world reliability work (document cleanup, hallucination controls, monitoring/drift tracking with MLflow) and close collaboration with non-technical clinical operations teams via demos and feedback-driven iteration.”
Mid-level Data Analyst specializing in cloud ETL, BI, and machine learning
“Data/ML practitioner with experience at UnitedHealth Group building a fraud claims detection solution combining structured claims data and unstructured notes, validated with compliance stakeholders to improve actionable accuracy. Also applied embeddings, vector databases, and fine-tuned language models in a Bank of America capstone to detect threats/anomalies in financial documents, with production-minded Python ETL workflows using Airflow.”
Mid-level Data Engineer specializing in AWS cloud data platforms
“Data engineer with Charter Communications experience modernizing large-scale AWS data lake pipelines: ingesting S3 data, validating against legacy systems, transforming with PySpark/Spark SQL, and serving via Iceberg/Delta tables. Worked at 50M–300M record scale, delivered >99.5% data match, and built monitoring/alerting (CloudWatch/SNS) plus retry orchestration (Step Functions) and data quality gates (Great Expectations).”
Mid-level Full-Stack Developer specializing in React and scalable web applications
“Backend/data engineer with hands-on production experience across FastAPI microservices and AWS data platforms. Has delivered serverless and Glue/EMR-based ETL pipelines with strong observability (Prometheus/Grafana/Sentry, CloudWatch/SNS), schema-evolution resilience, and measurable SQL performance wins (5 min to <30 sec). Open to onsite meetings in the Bethesda, MD area and flexible on remote arrangements.”
Mid-level Solutions Architect/Engineer specializing in AI and data integrations
“Solutions Engineer specializing in taking LLM copilots from demo to production, with a strong emphasis on enterprise security (RBAC/OAuth), grounded RAG behavior (cite-or-refuse), and operational readiness (eval loops, logging, runbooks). Experienced in real-time diagnosis of agentic/LLM workflow failures and in partnering with Sales/CS to run integration-first POCs that clear security and reliability concerns and accelerate rollout.”