Pre-screened and vetted in New Jersey.
Intern Data Scientist specializing in machine learning and trustworthy AI
Mid-level Data Scientist / ML Engineer specializing in NLP, recommender systems, and insurance analytics
Principal Data Scientist specializing in Generative AI and MLOps
Senior Data Engineer specializing in cloud data platforms and large-scale ETL
“Data engineer focused on large-scale ETL/ELT pipelines across cloud stacks (GCP and AWS), including Spark-based transformations and orchestration with Airflow. Has experience loading up to ~2TB per BigQuery target table and designing atomic loads to multiple downstream systems (Elasticsearch + Kafka), with Kubernetes deployment and Jenkins CI/CD.”
Senior Data Scientist specializing in LLM products, voice agents, and FinTech risk modeling
Mid-level analytics professional specializing in pricing and survey analytics
Mid-level Data Scientist specializing in financial risk, fraud detection, and GenAI NLP
Senior Data Engineer specializing in cloud data platforms and real-time pipelines
Mid-level Business Analyst specializing in BI, reporting, and data analytics
“Finance data and reporting professional with PwC experience who bridges accounting and technology, especially around GL-related reconciliations, reporting accuracy, and close support. While not a direct PeopleSoft GL owner, they bring strong SQL-driven troubleshooting, ETL/data mapping remediation, and process automation experience that helped shorten close cycles and improve audit readiness.”
Mid-level Data Engineer specializing in streaming and cloud data platforms for financial services
“Data engineering-focused candidate (internship/project experience) who built end-to-end pipelines processing a few million transactional records/day for fraud detection and reporting, using Airflow, Python/SQL, and PySpark with strong emphasis on data quality gates, idempotency, and monitoring. Also implemented an external web/API data collection system with anti-bot tactics and schema-change quarantine, and shipped a versioned Flask API to serve curated warehouse data.”
Mid-level Data Scientist specializing in LLMs, MLOps, and predictive analytics in healthcare and finance
“Built and deployed a production LLM/RAG clinical decision support system that enables real-time semantic search over unstructured EHR notes and delivers patient risk insights. Strong in healthcare-grade MLOps and compliance (HIPAA, PHI handling, encryption, RBAC, audit logs) and scaled embedding/retrieval pipelines using Spark/Databricks and Airflow. Partnered with clinicians via Power BI dashboards and explainability, contributing to an 18% reduction in patient readmissions.”
Mid-level Data Analyst specializing in BI, ETL, and operational analytics
Mid-level Data Engineer specializing in cloud lakehouse and streaming platforms
Mid-level Data Analyst specializing in healthcare and financial risk analytics
Junior Product & Business Analyst specializing in analytics, dashboards, and go-to-market strategy
Mid-level GenAI/Data Engineer specializing in LLM agents and RAG systems
Principal/Lead Data Engineer specializing in large-scale pipelines, NLP, and graph databases
Junior Data Scientist specializing in analytics automation and BI dashboards
Mid-level Data Engineer specializing in real-time analytics and FinTech data platforms
Mid-level Data Engineer specializing in financial data engineering and scalable pipelines
Mid-level Data Analyst specializing in healthcare and finance analytics
“Built an end-to-end Alexa smart-home IoT application controlling a Wi-Fi bulb, including ESP32 firmware (MQTT) and an AWS serverless backend (IoT Core/Device Shadow, Lambda, DynamoDB) with a REST API. Demonstrates strong real-time scalability patterns (streaming ingestion, stateless processing, partition-key design) and full-stack delivery with Spring Boot + React (JWT auth, CORS, data-heavy dashboards).”
Mid-level Data Engineer specializing in multi-cloud real-time and batch data pipelines
“Data engineer with healthcare domain experience who owned 100M+ record pipelines end-to-end (Kafka/Kinesis/ADF → PySpark/dbt validation → Spark SQL transforms → Snowflake/Power BI serving). Built production-grade reliability practices (Airflow orchestration, CloudWatch/Grafana monitoring, pytest + contract/regression tests, idempotent ingestion/backfills) and delivered measurable improvements: 35% lower latency and 40% better query performance.”