Pre-screened and vetted.
Mid-level Data Scientist & Generative AI Engineer specializing in LLMs and RAG
“ML/NLP practitioner who built a retrieval-augmented generation (RAG) system for large financial and operational document sets using Sentence-Transformers (all-mpnet-base-v2) and a vector DB (e.g., Pinecone), with a strong focus on retrieval evaluation and chunking strategy optimization. Experienced in entity resolution (rules + embedding similarity with type-specific thresholds) and in productionizing scalable Python data workflows using Airflow/Dagster and Spark.”
Mid-level Data Analyst specializing in AWS-based ETL, churn analytics, and BI dashboards
“Data/ML practitioner with experience at Airtel and Lincoln Financial delivering measurable business outcomes: improved retention 15% via NLP sentiment analysis and cut response time ~25% using sentence-BERT + FAISS semantic linking. Strong in data quality/identity resolution (SQL + fuzzy matching) and in building production-grade Python workflows orchestrated with Airflow/AWS Glue, including validation and dashboard integration in Power BI.”
Senior Data Engineer specializing in cloud data platforms and ML pipelines
“Data engineer focused on AWS-based enterprise data platforms, owning end-to-end pipelines from multi-source batch/stream ingestion (Glue/Kinesis/StreamSets/Airflow) through PySpark transformations into curated datasets for Redshift/Snowflake. Emphasizes production reliability with strong monitoring/observability and data quality gates, and reports ~30% performance improvement plus improved SLAs and latency after optimization.”
Mid-level Backend Python Engineer specializing in APIs, microservices, and data pipelines
“Backend engineer (Marsh McLennan) who evolved a high-volume claims automation pipeline in Python, emphasizing thin APIs with background job processing, strong validation/retries, and production-grade observability. Experienced in secure FastAPI API design (centralized JWT/RBAC), multi-tenant Postgres/Supabase-style row-level security, and low-risk refactors using parallel runs and feature flags; targeting founding-engineer scope roles.”
Mid-level Data Engineer specializing in cloud data platforms and lakehouse architectures
“Data engineer in a banking context who has owned end-to-end Azure lakehouse pipelines ingesting financial/vendor data from APIs, Azure SQL, and flat files into Databricks/Delta (bronze-silver-gold). Emphasizes production reliability via schema-drift validation, data quality controls, monitoring/alerting, retries/checkpointing, and Spark/Delta performance tuning, with outputs served to BI/reporting teams (e.g., Tableau).”
Mid-Level Software Engineer specializing in cloud-native microservices on AWS
“Backend engineer with experience across healthcare and fintech platforms (Anthem, Citia) building high-throughput Python microservices with strong compliance/security focus (HIPAA, tenant isolation). Has integrated ML workflows into production systems (ResNet embedding-based image similarity) using async pipelines (Celery/Redis) and AWS (Lambda/S3/ECS), delivering measurable performance and fraud/content-integrity improvements at scale.”
Mid-level AI/ML Engineer specializing in NLP, LLMs, and RAG for finance and healthcare
“Built an AI lending assistant (RAG + DeBERTa) used by credit analysts to retrieve policies and past loan decisions, tackling real production issues like hallucinations, document quality, and sub-second latency. Deployed a modular, Dockerized AWS architecture (ECS/EMR + load balancer) with load testing, caching/precomputed embeddings, and CloudWatch monitoring, and used Airflow to automate scheduled data/embedding/vector DB refresh pipelines with retries and alerts.”
Director-level Data Science & Analytics Leader specializing in cloud data platforms and AI/ML
“Candidate states they are very familiar with the venture capital/studio/accelerator landscape and expresses strong willingness to pursue entrepreneurship "at all costs," but did not provide details on a current startup, business plan, fundraising, or prior accelerator/VC involvement during the interview.”
Mid-level Data Engineer specializing in cloud data pipelines for healthcare and financial services
“Data engineer with ~4 years of experience (Cigna) building and operating Azure Data Factory pipelines for healthcare claims/member/provider data at 2–3M records/day. Emphasizes reliability and downstream safety via schema/data-quality validation, quarantine workflows, idempotent processing, and backfills; also improved runtime ~20% through SQL optimization and served curated datasets through versioned views and well-documented, analyst-friendly interfaces.”
Mid-level Data Engineer specializing in cloud-native healthcare and enterprise data platforms
“Data Engineer (TCS) who owned an end-to-end CRM analytics pipeline for Bayer’s eSalesWeb integration, ingesting from Salesforce APIs/databases/S3 and serving analytics-ready datasets via PostgreSQL/S3 for Tableau. Drove measurable outcomes: ~60% reduction in manual data-quality effort, ~30% lower latency through SQL optimization, and ~35% improved stability via monitoring, retries, and idempotent processing.”
Executive Technology Leader (CTO) specializing in cloud, AI/ML, and scalable product platforms
“Technical leader and hands-on engineer with 20+ years of experience who has previously raised funding and exited a venture. Currently bootstrapping a new AI-direction startup with personal and family capital, leveraging structured financial planning and a relationship-driven approach to investor outreach.”
Mid-level Data Analyst specializing in financial and customer analytics
“Analytics professional with experience at KPMG and Robosoft Technologies, working across financial and customer engagement data. They combine SQL, Python, experimentation, and BI dashboards to turn messy multi-source data into decision-ready insights, including a pricing test that improved conversion rates by 9%.”
Mid-level Software Engineer specializing in Python backend and AI applications
“ML engineer at CGI who built demand forecasting models end-to-end, from feature engineering and training through AWS deployment. Stands out for a production-first mindset and strong skepticism of AI-generated code, including catching a Copilot-generated SQL query that would have caused a costly full table scan in production.”
Mid-level Data Scientist specializing in MLOps and Generative AI
“Robotics software/ML engineer who built perception and navigation-related ML systems for autonomous supermarket carts, including object detection, shelf recognition, and obstacle avoidance. Strong ROS/ROS2 practitioner who optimized real-time performance (reported 50% latency reduction) and deployed containerized ROS/ML pipelines at scale using Docker, Kubernetes, and CI/CD.”
Senior Python Backend Engineer specializing in scalable APIs and cloud-native microservices
“Backend/data platform engineer who has built and operated a cloud-native media ingestion/processing platform in Python (Django/DRF, FastAPI) with Kafka, Postgres, and Redis, emphasizing multi-tenant security and reliability. Delivered AWS production systems combining EKS and Lambda with Terraform + GitHub Actions/Helm, and built Glue-based ETL pipelines with strong schema-evolution and data-quality practices; also modernized SAS analytics into Python on AWS. Seeking fully remote roles with a $120K–$140K base range.”
Senior AI/ML & Full-Stack Engineer specializing in GenAI, RAG, and MLOps platforms
“Backend/data platform engineer who owned end-to-end production services for a fleet analytics/GenAI platform, spanning FastAPI microservices on Kubernetes and AWS (EKS + Lambda) event-driven workloads. Strong in reliability/observability (OpenTelemetry, circuit breakers, idempotency), data pipelines (Glue/Airflow/Snowflake), and measurable performance/cost wins (SQL 10s to <800ms P95; ~30% compute cost reduction).”
Mid-level Full-Stack Java Developer specializing in cloud-native microservices
“Software engineer with deep healthcare claims domain experience who has owned customer-facing portals end-to-end (Java/Spring Boot + React/TypeScript) and improved usability/performance based on real user feedback. Built microservices using REST and RabbitMQ with strong observability (Splunk/cloud metrics), and delivered an internal claims investigation dashboard that streamlined operations through centralized data, search, and filtering.”
Mid-level Data Scientist specializing in real-time fraud detection and MLOps
“ML/NLP engineer with experience at Charles Schwab building an NLP + graph (Neo4j) entity-resolution system to unify fragmented user/device/transaction data and improve downstream model quality and analyst querying. Has applied embeddings (SentenceTransformers + FAISS) with domain fine-tuning to boost hard-case matching recall by ~12% while maintaining precision, and has a track record of hardening scalable Python/Spark pipelines and productionizing fraud models via A/B tests and shadow-mode monitoring.”
Senior Data & Platform Engineer specializing in cloud-native streaming and distributed systems
“Financial data engineer who has built and operated high-volume batch + streaming pipelines (200–300 GB/day; 5–10k events/sec) using AWS, Spark/Delta, Airflow, Kafka, and Snowflake, with strong emphasis on data quality and reliability. Demonstrated measurable impact via 99.9% SLA adherence, major reductions in bad records/nulls, MTTR improvements, and significant latency/runtime/query performance gains; also built a distributed web-scraping system processing 5–10M records/day with anti-bot and schema-drift defenses.”
Mid-level Data Engineer specializing in multi-cloud data platforms for healthcare and finance
“Data engineer with Cigna experience building and operating an end-to-end AWS-based healthcare claims pipeline processing ~2TB/day, using Glue/Kafka/PySpark/SQL into Redshift. Strong focus on data quality and reliability (schema validation, monitoring/alerting, retries/checkpointing/backfills), reporting improved accuracy (~99%) and reduced latency, plus experience serving real-time Kafka/Spark data to downstream analytics with documented data contracts.”
Mid-Level Full-Stack Software Engineer specializing in cloud-native and GenAI solutions
“Built and shipped production RAG-based LLM agents automating multi-step document query workflows, emphasizing reliability via monitoring, retries, structured exception handling, and fallback retrieval (alternative embeddings/keyword search). Demonstrated measurable gains (18% latency improvement, 25% retrieval efficiency, 12% precision) and has experience integrating agents with messy tax and transaction data at RSM using validation/cleaning and idempotent design.”
Mid-level Data Engineer specializing in cloud ETL and real-time streaming
“Data engineer focused on AWS + Spark/Databricks pipelines, including an end-to-end nightly loan-data ingestion flow (~2.2M records) from Postgres/S3 through Glue and Databricks into a DWH with layered validation and alerting. Also built real-time streaming with Kafka + Spark Structured Streaming and a master’s project streaming Reddit data for sentiment analysis under ambiguous requirements and tight budget constraints.”
Senior Integration Developer specializing in MuleSoft API-led connectivity
“Backend/integration-focused engineer in the Maryland area with production experience building FastAPI REST services secured with OAuth2.1/JWT and reliability patterns (timeouts, selective retries, idempotency, centralized error handling). Has delivered AWS-integrated MuleSoft/CloudHub solutions and supported AWS Glue ETL workflows, plus demonstrated strong SQL tuning with a 30–40s to 3–5s performance improvement.”
Mid-level Data Engineer specializing in Lakehouse, Streaming, and ML/LLM data systems
“Built and productionized an enterprise retrieval-augmented generation platform for internal knowledge over large unstructured corpora, emphasizing trust via strict citation/grounding and hybrid retrieval (BM25 + FAISS + cross-encoder re-ranking). Demonstrates strong scaling and cost/latency optimization through incremental indexing/embedding and index partitioning, plus disciplined evaluation/observability practices. Has experience operationalizing pipelines with Airflow/Databricks/GitHub Actions and partnering closely with risk & compliance stakeholders on auditability requirements.”