Pre-screened and vetted in New Jersey.
Senior Data Engineer specializing in cloud data platforms and large-scale ETL
“Data engineer focused on large-scale ETL/ELT pipelines across cloud stacks (GCP and AWS), including Spark-based transformations and orchestration with Airflow. Has experience loading up to ~2TB per BigQuery target table and designing atomic loads to multiple downstream systems (Elasticsearch + Kafka), with Kubernetes deployment and Jenkins CI/CD.”
Senior Data Engineer specializing in cloud data platforms and real-time pipelines
Mid-level Data Engineer specializing in streaming and cloud data platforms for financial services
“Data engineering-focused candidate (internship/project experience) who built end-to-end pipelines processing a few million transactional records/day for fraud detection and reporting, using Airflow, Python/SQL, and PySpark with strong emphasis on data quality gates, idempotency, and monitoring. Also implemented an external web/API data collection system with anti-bot tactics and schema-change quarantine, and shipped a versioned Flask API to serve curated warehouse data.”
Mid-level Data Engineer specializing in cloud lakehouse and streaming platforms
Mid-level GenAI/Data Engineer specializing in LLM agents and RAG systems
Junior Data Scientist specializing in analytics automation and BI dashboards
Mid-level Data Engineer specializing in real-time analytics and FinTech data platforms
Mid-level Data Engineer specializing in financial data engineering and scalable pipelines
Mid-level Data Engineer specializing in multi-cloud real-time and batch data pipelines
“Data engineer with healthcare domain experience who owned 100M+ record pipelines end-to-end (Kafka/Kinesis/ADF → PySpark/dbt validation → Spark SQL transforms → Snowflake/Power BI serving). Built production-grade reliability practices (Airflow orchestration, CloudWatch/Grafana monitoring, pytest + contract/regression tests, idempotent ingestion/backfills) and delivered measurable improvements: 35% lower latency and 40% better query performance.”
Mid-level Data Engineer specializing in capital markets post-trade data platforms
“Data/streaming engineer in capital markets who led an end-to-end trade settlement data product (Kafka→MongoDB→data lake) with rigorous data-quality logic and ~$175K first-year operational impact. Also built a low-latency Go-based CME market data engine feeding SOFR curve generation, using MSK on EKS with performance tuning (idempotency, compression, partitioning) to achieve sub-100ms delivery.”
Mid-level Data Engineer specializing in financial data pipelines on AWS and Databricks
Mid-level Azure Data Engineer specializing in Databricks lakehouse and Spark pipelines
Mid-level Data Engineer specializing in cloud ETL, big data, and analytics
Mid-level Data Engineer specializing in cloud ETL/ELT and lakehouse architecture
“Data engineer focused on sales/marketing analytics pipelines, owning ingestion from CRMs/ad platforms through warehouse serving and dashboards at ~hundreds of thousands of records/day. Built reliability-focused systems including dbt/SQL/Python data quality gates with alerting, a resilient web-scraping pipeline (retries/backoff, anti-bot tactics, schema-change detection, backfills), and a versioned internal REST API with caching and strong developer usability.”
Mid-level Data Engineer specializing in cloud data pipelines for banking analytics
Mid-level Data Engineer specializing in real-time streaming and cloud data platforms
Senior AI & GenAI Data Engineer specializing in RAG pipelines and lakehouse data platforms
Senior Lead Data Engineer specializing in cloud data platforms and real-time ML pipelines
Mid-level sales and data professional specializing in FinTech, telecom, and insurance
Mid-Level Data Engineer specializing in cloud data pipelines and big data platforms
“Data engineer with ~4 years of experience building Python-based data ingestion/processing services and real-time streaming pipelines (Kafka/PubSub + Spark Structured Streaming). Has deployed containerized data applications on Kubernetes with GitLab CI/Jenkins pipelines and applied GitOps to cut deployment time ~40% while reducing config drift. Also supported a legacy on-prem data warehouse/backend migration to GCP using phased migration and parallel validation to meet strict reliability/SLA needs.”
Junior Data Engineer specializing in cloud ETL/ELT and lakehouse platforms
Mid-level Software/Data Engineer specializing in cloud ETL pipelines and data infrastructure
“Backend/data engineer who built a production analytics data service (Python/FastAPI on AWS/Postgres with PySpark ETL) handling millions of records per day and drove major latency improvements (10–15s to <2s) via indexing, Redis caching, and shifting aggregations into ETL. Also shipped an LLM-based natural-language-to-SQL assistant end-to-end with strong guardrails (schema restrictions, read-only validation, RBAC, masking) and designed a multi-step agent workflow with verification and fallback logic.”