Pre-screened and vetted.
Mid-level Data Engineer specializing in cloud ETL, big data, and analytics
Mid-level Data Scientist specializing in NLP, time-series forecasting, and GenAI
Senior Data Scientist specializing in AWS ML solutions for healthcare, telecom, and e-commerce
Senior Data Engineer specializing in AWS cloud data platforms and streaming analytics
Mid-level Data Engineer specializing in lakehouse architectures and cloud ELT
Mid-level Data Scientist / ML Engineer specializing in NLP, GenAI, and cloud ML deployment
Senior Cloud & DevOps Engineer specializing in AWS, Kubernetes, IaC, and cloud security
Senior Data Engineer specializing in multi-cloud lakehouse architectures and privacy/AI governance
Senior Data Scientist specializing in NLP, MLOps, and cloud ML platforms
Mid-level Software Engineer specializing in ML, LLM apps, and cloud data systems
“Built a production SQL chatbot for access-log analytics that replaced manual custom report requests with natural-language querying, using LangGraph and a ChromaDB-backed RAG pipeline for grounded, consistent answers. Implemented a privacy-preserving design where the LLM never sees raw customer data (only query metadata) and has experience building multi-agent/tool-calling systems with LangGraph (DeepAgents), including solving sub-agent communication drift via self-reflection.”
Mid-level Data Scientist & Generative AI Engineer specializing in LLMs and RAG
“ML/NLP practitioner who built a retrieval-augmented generation (RAG) system for large financial and operational document sets using Sentence-Transformers (all-mpnet-base-v2) and a vector DB (e.g., Pinecone), with a strong focus on retrieval evaluation and chunking strategy optimization. Experienced in entity resolution (rules + embedding similarity with type-specific thresholds) and in productionizing scalable Python data workflows using Airflow/Dagster and Spark.”
Mid-level Data Analyst specializing in AWS-based ETL, churn analytics, and BI dashboards
“Data/ML practitioner with experience at Airtel and Lincoln Financial delivering measurable business outcomes: improved retention 15% via NLP sentiment analysis and cut response time ~25% using sentence-BERT + FAISS semantic linking. Strong in data quality/identity resolution (SQL + fuzzy matching) and in building production-grade Python workflows orchestrated with Airflow/AWS Glue, including validation and dashboard integration in Power BI.”
Senior Data Engineer specializing in cloud data platforms and ML pipelines
“Data engineer focused on AWS-based enterprise data platforms, owning end-to-end pipelines from multi-source batch/stream ingestion (Glue/Kinesis/StreamSets/Airflow) through PySpark transformations into curated datasets for Redshift/Snowflake. Emphasizes production reliability with strong monitoring/observability and data quality gates, and reports ~30% performance improvement plus improved SLAs and latency after optimization.”
Mid-level Backend Python Engineer specializing in APIs, microservices, and data pipelines
“Backend engineer (Marsh McLennan) who evolved a high-volume claims automation pipeline in Python, emphasizing thin APIs with background job processing, strong validation/retries, and production-grade observability. Experienced in secure FastAPI API design (centralized JWT/RBAC), multi-tenant Postgres/Supabase-style row-level security, and low-risk refactors using parallel runs and feature flags; targeting founding-engineer scope roles.”
Mid-level Data Engineer specializing in cloud data pipelines for healthcare and financial services
“Data engineer with ~4 years of experience (Cigna) building and operating Azure Data Factory pipelines for healthcare claims/member/provider data at 2–3M records/day. Emphasizes reliability and downstream safety via schema/data-quality validation, quarantine workflows, idempotent processing, and backfills; also improved runtime ~20% through SQL optimization and served curated datasets through versioned views and well-documented, analyst-friendly interfaces.”
Senior Python Backend Engineer specializing in scalable APIs and cloud-native microservices
“Backend/data platform engineer who has built and operated a cloud-native media ingestion/processing platform in Python (Django/DRF, FastAPI) with Kafka, Postgres, and Redis, emphasizing multi-tenant security and reliability. Delivered AWS production systems combining EKS and Lambda with Terraform + GitHub Actions/Helm, and built Glue-based ETL pipelines with strong schema-evolution and data-quality practices; also modernized SAS analytics into Python on AWS. Seeking fully remote roles with a $120K–$140K base range.”
Mid-level Machine Learning Engineer specializing in Generative AI and LLM applications
“GenAI engineer who has deployed production LLM/RAG chatbots for internal document search, focusing on reliability (hallucination reduction via prompt guardrails + retrieval filtering) and performance (latency improvements via caching). Experienced with LangChain/LangGraph orchestration for multi-step agent workflows and iterates using monitoring/logs and benchmark-driven evaluation while partnering closely with product and business teams.”
Mid-level Full-Stack Java Developer specializing in cloud-native microservices
“Software engineer with deep healthcare claims domain experience who has owned customer-facing portals end-to-end (Java/Spring Boot + React/TypeScript) and improved usability/performance based on real user feedback. Built microservices using REST and RabbitMQ with strong observability (Splunk/cloud metrics), and delivered an internal claims investigation dashboard that streamlined operations through centralized data, search, and filtering.”
Mid-level Data Scientist specializing in real-time fraud detection and MLOps
“ML/NLP engineer with experience at Charles Schwab building an NLP + graph (Neo4j) entity-resolution system to unify fragmented user/device/transaction data and improve downstream model quality and analyst querying. Has applied embeddings (SentenceTransformers + FAISS) with domain fine-tuning to boost hard-case matching recall by ~12% while maintaining precision, and has a track record of hardening scalable Python/Spark pipelines and productionizing fraud models via A/B tests and shadow-mode monitoring.”
Senior Data & Platform Engineer specializing in cloud-native streaming and distributed systems
“Financial data engineer who has built and operated high-volume batch + streaming pipelines (200–300 GB/day; 5–10k events/sec) using AWS, Spark/Delta, Airflow, Kafka, and Snowflake, with strong emphasis on data quality and reliability. Demonstrated measurable impact via 99.9% SLA adherence, major reductions in bad records/nulls, MTTR improvements, and significant latency/runtime/query performance gains; also built a distributed web-scraping system processing 5–10M records/day with anti-bot and schema-drift defenses.”
Mid-level Data Engineer specializing in cloud ETL and real-time streaming
“Data engineer focused on AWS + Spark/Databricks pipelines, including an end-to-end nightly loan-data ingestion flow (~2.2M records) from Postgres/S3 through Glue and Databricks into a DWH with layered validation and alerting. Also built real-time streaming with Kafka + Spark Structured Streaming and a master’s project streaming Reddit data for sentiment analysis under ambiguous requirements and tight budget constraints.”
Mid-level Data Engineer specializing in Lakehouse, Streaming, and ML/LLM data systems
“Built and productionized an enterprise retrieval-augmented generation platform for internal knowledge over large unstructured corpora, emphasizing trust via strict citation/grounding and hybrid retrieval (BM25 + FAISS + cross-encoder re-ranking). Demonstrates strong scaling and cost/latency optimization through incremental indexing/embedding and index partitioning, plus disciplined evaluation/observability practices. Has experience operationalizing pipelines with Airflow/Databricks/GitHub Actions and partnering closely with risk & compliance stakeholders on auditability requirements.”