Pre-screened and vetted.
Mid-level Machine Learning Engineer specializing in LLM agents, RAG, and MLOps
“Built a production AI-driven contract/document extraction system combining OCR, normalization, and LLM schema-guided extraction, orchestrated with PySpark and Azure Data Factory and loaded into PostgreSQL for analytics. Emphasizes reliability at scale—using strict JSON schemas, confidence scoring, targeted retries, and multi-layer validation to control hallucinations while processing thousands of PDFs per hour—and partners closely with non-technical business teams to refine fields and deliver usable dashboards.”
Mid-level Data Analyst specializing in AI/ML and advanced analytics
“Accenture data/ML practitioner who deployed a retail churn prediction and BERT-based sentiment analysis system to production, integrating behavioral + feedback data and operationalizing it with ETL automation, orchestration, and CI/CD. Experienced managing 2TB+ multi-source data, monitoring drift in Databricks, and translating results into Power BI dashboards for marketing teams (including K-means customer segmentation).”
Senior DevOps & Release Engineer specializing in CI/CD automation and AWS IaC
“Infrastructure/DevOps engineer (Vidmob) focused on AWS + containers, owning GitLab CI/CD and Terraform-managed environments. Led a high-impact CI incident by correlating runner queue time, Docker pull latency, and NAT egress; implemented ECR pull-through caching and VPC endpoints to restore performance and then standardized the fix in Terraform for future scale-ups.”
Mid-level Sales Engineer & Solution Architect specializing in cloud and data platforms
“LLM-focused customer-facing technical leader with experience productionizing LLM workflows in financial services (State Street), including guardrails, retrieval tuning, and reliability improvements. Also partners closely with sales and executives—at Payoneer helped drive enterprise-wide adoption for a $10M ARR global account through technical discovery, demos, and pilots.”
Mid-level Data Engineer specializing in cloud data pipelines and analytics platforms
“Data engineer with healthcare and enterprise experience (Molina Healthcare, Dell Technologies) building and operating high-volume batch + streaming pipelines across AWS and Azure. Strong focus on data quality (schema validation, fail-fast checks), reliability (monitoring/alerts, retries), and performance tuning (Spark/partitioning), with measurable runtime reduction and improved downstream trust.”
Mid-level Data Engineer specializing in cloud data pipelines and financial services warehousing
“Data engineer (Charles Schwab) who took ownership of an unstable, ambiguous nightly financial data pipeline and rebuilt it into a reliable, incremental AWS Glue/Airflow/Redshift system feeding Power BI. Created a custom Python data-quality framework with hard-stop gating and schema drift detection, improving integrity (99.9%), cutting runtime (~20%), and reducing incidents/tickets (35% fewer schema-related dashboard incidents; 30% fewer investigations).”
Senior ML Engineer & Data Scientist specializing in LLM agents, retrieval/ranking, and MLOps
“Machine Learning Engineer currently at Webster Bank building an enterprise-scale LLM agent for Temenos Journey Manager/Maestro, using RAG-style multi-stage retrieval with FAISS/Pinecone, hybrid dense+sparse search, and LoRA fine-tuning optimized via NDCG/MAP and A/B testing. Previously handled messy incident/telemetry data at Deuta Werke GmbH with deterministic + fuzzy entity resolution, and has strong production data engineering experience across Spark/Hadoop and Python ETL systems.”
Director-level Engineering Leader specializing in enterprise SaaS and cloud-native platforms
“Engineering leader/player-coach who modernized a legacy C#/SQL Server system to Snowflake + Python on GCP, enabling ~30x scale and supporting hundreds of millions of transactions per day per customer. Strong in architecture tradeoffs (Snowflake vs Databricks), production reliability (New Relic, logging/alerting), and lightweight process improvements like a rigorous Definition of Done and structured PR reviews.”
Mid-level Data Engineer specializing in cloud lakehouse and streaming platforms
“Data engineer focused on building production-grade pipelines on AWS (Kafka/Kinesis/Glue/S3) through to curated serving layers in Snowflake and Delta Lake. Emphasizes automated data quality validation (PySpark + CI/CD), modular dbt transformations for analytics (customer spending, risk metrics), and operational reliability with CloudWatch and DLQs; data consumed by BI tools and ML pipelines for fraud detection and risk analytics.”
Mid-level Data Engineer specializing in real-time pipelines and cloud data platforms
“Backend engineer with hands-on experience building secure Python/Flask services (sessions, JWT, RBAC) and optimizing PostgreSQL/SQLAlchemy performance, including custom SQL using CTEs/window functions profiled via EXPLAIN ANALYZE. Also integrates LLM features via OpenAI/Azure into backend systems and improves scalability with RabbitMQ-driven async processing, caching, and multi-tenant data isolation patterns.”
Mid-level Data Engineer specializing in capital markets post-trade data platforms
“Data/streaming engineer in capital markets who led an end-to-end trade settlement data product (Kafka→MongoDB→data lake) with rigorous data-quality logic and ~$175K first-year operational impact. Also built a low-latency Go-based CME market data engine feeding SOFR curve generation, using MSK on EKS with performance tuning (idempotency, compression, partitioning) to achieve sub-100ms delivery.”
Senior Data Analyst specializing in healthcare and financial analytics
“Healthcare analytics candidate with hands-on experience turning messy claims data in Redshift and S3 into validated reporting tables, plus automating KPI workflows in Python. They’ve owned end-to-end operational analytics projects, including a claims delay analysis that improved processing efficiency by about 20%, and have experience driving stakeholder adoption of standardized metrics across dashboards.”
Junior Data Analyst specializing in financial and operational analytics
“Analytics professional with experience at KPMG turning messy operational and financial data from SQL Server and AWS S3 into clean reporting datasets and automated Python workflows. They combine SQL, Python, Power BI, and experimentation methods to deliver stakeholder-aligned KPI dashboards and marketing performance insights with a strong focus on data integrity and reproducibility.”
Junior Business & Data Analyst specializing in analytics and AI-driven insights
“Master’s in Business Analytics candidate with hands-on project experience spanning FMCG sales analytics, insurance risk modeling, and HR attrition analysis. Demonstrates strong SQL and Python fundamentals, including advanced CTE/window-function work, reproducible modeling workflows, and Power BI dashboards that translate analysis into clear business actions.”
Intern AI/ML Engineer specializing in full-stack and data systems
“Built an LLM-powered customer segmentation agent during a Chewy internship, consolidating Snowflake data into a knowledge graph so non-technical marketing users could query customer cohorts in natural language. Stands out for combining agent/tooling design with rigorous data engineering practices, including schema audits, imputation, validation layers, and idempotent pipelines on messy large-scale datasets.”
Mid-level Data Analyst specializing in financial and healthcare analytics
“Analytics professional with experience at Franklin Templeton and IQVIA India, focused on turning messy cross-system data into trusted reporting and actionable business insights. Stands out for combining SQL, Python, AWS ETL, and BI dashboards to solve data quality issues, improve investor engagement analysis, and standardize commercial reporting in financial services and pharma contexts.”
Senior Data Scientist and AI/ML Engineer specializing in GenAI and cloud ML
“ML/AI engineer with hands-on experience owning systems from experimentation through deployment and monitoring, including a Bank of Montreal project that improved timely interventions by 12%. Also brings GenAI/RAG experience with evaluation and safety guardrails, plus clinical NLP pipeline work extracting medication data from notes for patient risk prediction.”
Senior Full-Stack & Mobile Engineer specializing in Node.js and React
“Backend engineer with TaskRabbit experience building and operating payment/booking services in Python/Django on AWS (ECS + Lambda) with Kafka/SQS eventing. Demonstrates strong production reliability and incident ownership in high-stakes payment flows (idempotency, strict timeouts, retries, monitoring/alerting) plus data/ETL work in AWS Glue and measurable SQL performance wins.”
Junior Data Scientist specializing in fraud analytics and cloud data platforms
“Built and deployed production LLM-powered document summarization/classification systems using embeddings, vector databases (RAG-style retrieval), and automated evaluation (BERTScore/ROUGE), with a focus on monitoring and scalable cloud pipelines. Also partnered with a fraud analytics team to deliver a transaction anomaly detection solution, translating model outputs into Power BI dashboards and actionable KPIs while iterating on thresholds and alerts based on stakeholder feedback.”
Mid-level Generative AI Engineer specializing in LLM apps, RAG, and MLOps
“LLM/GenAI engineer with US Bank experience building a production financial-document intelligence platform using LangChain/LangGraph, GPT-4, and Amazon OpenSearch. Delivered a RAG-based assistant for compliance/audit teams with grounded, cited answers, focusing on reducing hallucinations and latency, and deployed securely on AWS (SageMaker/EKS) with CI/CD and evaluation tooling (LangSmith, RAGAS).”
Mid-level Data Engineer specializing in cloud data platforms, Spark, and streaming pipelines
“Data/MLOps engineer (Cognizant background) who owned an AWS/Airflow/Snowflake healthcare transactions pipeline processing ~8–10M records/day and cut pipeline/data-quality incidents by ~33%. Also built and deployed a production FastAPI model-inference service on Kubernetes (Docker, HPA) with strong observability (Prometheus/Grafana), versioned endpoints, and resilient backfill/idempotent external data ingestion patterns.”
Mid-level Data Engineer specializing in cloud data pipelines and streaming
“Data engineer with experience at Wells Fargo and Accenture owning end-to-end production pipelines processing hundreds of millions of transactional/risk records daily. Strong focus on data quality and reliability (reconciliation checks, schema drift detection, CloudWatch alerting) plus Spark performance tuning and idempotent backfills using Delta Lake/merge logic across AWS (S3/EMR/Databricks/Redshift) and Azure (ADF/Azure DevOps/Azure Monitor).”
Mid-level Data Engineer specializing in AWS/Azure pipelines and streaming analytics
“Data engineer with experience across healthcare and geospatial risk systems, owning end-to-end pipelines from ingestion through serving on AWS/Azure stacks. Built HIPAA-compliant data quality gates and CDC for millions of daily claims, and also delivered a real-time wildfire risk platform with 20-minute refresh cycles and a 60% data accuracy lift. Strong in streaming (Kafka), Spark performance tuning, and production-grade orchestration/CI/CD (Airflow, Docker, Jenkins, GitHub Actions, Terraform).”
“Data engineer/backend engineer with experience in healthcare (Cardinal Health provider enrollment) and finance (Northern Trust) building and stabilizing data pipelines and REST services. Worked with APIs and Kafka at ~200k–300k records/day, improving data quality (DLQ + validation), performance (SQL/indexing), and reliability/observability (logging, alerts, consumer lag metrics), and stood up an early-stage financial data service with Jenkins-based CI/CD.”