Pre-screened and vetted.
Mid-level Data Engineer specializing in real-time pipelines and cloud analytics
“Researcher from the University of South Dakota who built a production medical RAG system to help interpret model predictions by retrieving relevant clinical notes and medical literature, overcoming retrieval accuracy and imaging-dataset challenges through semantic chunking and metadata-driven indexing. Also has hands-on orchestration experience with Airflow and Azure Data Factory, plus a pragmatic approach to LLM evaluation and stakeholder-driven iteration.”
Junior Machine Learning & Quant Research Engineer specializing in low-latency data and trading systems
“Applied ML to physical EV fleet systems at ST Labs, building a real-time CNN-LSTM fault prediction pipeline from streaming vehicle telemetry and addressing live data alignment issues via resampling/interpolation and buffered inference. Also developed a V2G/G2V energy transfer algorithm to automate charging/discharging for profit optimization, and made high-impact low-latency pipeline decisions at Astera Holdings using profiling, replay testing, and live A/B validation.”
Executive Economist specializing in macro-fiscal policy and international development
“Academic and international development leader (economics/public policy) who has advised SMEs and government institutions across the Middle East, including USAID-funded higher education and private sector development work in Palestine and fiscal management leadership across Iraqi ministries. Known for systems-thinking operational frameworks that align strategy, processes, and talent, plus large-scale training/ToT delivery (10,000+ trainees) and policy/regulatory collaboration to enable market development.”
Senior Data Scientist specializing in machine learning and customer analytics
“Data/ML practitioner with experience applying NLP and classical ML to large-scale customer data (2B+ records) for segmentation, prediction, and survey-text classification, delivering measurable business impact (~18% engagement efficiency). Has hands-on entity resolution across multi-source datasets and has built embedding-based semantic search using SentenceBERT + a vector database with domain fine-tuning (~20% relevance improvement), plus production workflow experience with Spark/Airflow and cloud tooling (AWS/Azure).”
Senior QA & Localization Specialist in Video Games and Language Services
“Localization QA tester with 4 years of experience testing Xbox AAA titles (including Dead Rising 3, Forza Motorsport, and Halo 5), focused on Spanish localization accuracy and UI/HUD issues. Experienced logging bugs in Jira, tracking them through regression in subsequent builds, and supporting testing through gold certification.”
Senior Data Engineer specializing in Azure Lakehouse, Databricks/Spark, and Snowflake
“Data engineer/platform builder with experience across PwC and Liberty Mutual delivering high-volume, production-grade pipelines and real-time data services. Has owned end-to-end streaming + batch architectures on AWS and Azure, including web scraping systems, with quantified reliability gains (99.9% availability, 90%+ error reduction, 30% latency reduction) and strong observability/CI-CD practices.”
Junior Full-Stack & Data Scientist specializing in ML/NLP and analytics products
“Built and deployed profitprops.io, a sports betting player-props prediction product using ML/AI. Implemented backend APIs with FastAPI/Express.js and Supabase, trained models on AWS GPU (P3) using Docker + RAPIDS, and set up CI/CD with GitHub Actions while working around cost constraints and data-collection hurdles (EC2 proxy rotation/rate limits).”
Mid-level Data Scientist specializing in predictive and generative AI
“AI/ML engineer with production LLM experience in regulated financial services (J.P. Morgan Chase), building a customer response engine to automate first-contact resolution while addressing privacy, bias, compliance, and scale. Strong MLOps/orchestration background (Airflow, Docker/Kubernetes, AWS Step Functions, Azure ML/SageMaker) plus proven ability to integrate with legacy systems and drive stakeholder adoption through dashboards, auditability, and training.”
Mid-level AI/ML Data Scientist specializing in NLP, computer vision, and risk analytics
“ML/AI engineer with Capital One experience building production-grade customer segmentation and fraud detection systems combining NLP (transformers) and anomaly detection. Strong MLOps and orchestration background (PySpark ETL, MLflow, Airflow, Docker/Kubernetes, Azure ML) with real-time monitoring/alerting and performance optimizations like quantization and caching, plus proven ability to deliver business-facing insights through Power BI/Tableau for marketing stakeholders.”
Intern Data Scientist specializing in computer vision and LLM agents
“Software engineering candidate with hands-on experience building and shipping LLM agents: created a production AI enrichment/coding agent at Covalent Metrology using Apollo.io + OpenAI, and built a Mistral hackathon router that dynamically selects among models to reduce token cost while maintaining quality. Also developed a real-time financial margin analysis agent that emails actionable insights and iterated on reliability issues (e.g., fixing misrouted emails, improving news relevance filtering).”
Mid-level Data Engineer specializing in cloud data pipelines and real-time streaming
“Data engineer with PNC Bank experience owning high-volume financial transaction pipelines end-to-end (Kafka/REST ingestion through Spark/Glue transformations to Redshift serving) for risk and fraud analytics. Built strong reliability and data quality practices (Great Expectations, reconciliation, Airflow alerting, idempotent retries, incremental/windowed processing), reporting 40% ingestion efficiency gains and ~99.9% data accuracy.”
Entry-Level Software Engineer specializing in data engineering and ML systems
“Built an end-to-end Next.js/TypeScript LLM-based scientific PDF analyzer using local Ollama/Llama inference to prioritize privacy and cost, producing structured research artifacts (e.g., authors/methods/findings) with ~92% extraction accuracy. At Qualtrics, helped replace a batch pipeline with a real-time, low-latency ML inference service (Python/Go on Kubernetes) using Redis caching, Grafana-based observability, and graceful fallbacks to protect UX during failures.”
Intern Data Analyst specializing in business intelligence and financial analytics
“Analytics candidate with hands-on experience in both fraud and churn use cases, including SQL-based preparation of 6.5M transaction records and reproducible Python modeling workflows. Stands out for combining technical rigor in data quality, feature engineering, and imbalance handling with strong stakeholder alignment, metric definition, and dashboard adoption.”
Mid-level Business Analyst specializing in healthcare and data analytics
“Analytics candidate with hands-on experience at BCBS building HIPAA-compliant SQL/Snowflake/Tableau pipelines across fragmented legacy healthcare systems. Stands out for turning a 5-day claims reporting process into a near real-time 10-minute dashboard and for pairing strong data engineering discipline with reproducible Python-based churn modeling that drove measurable retention outcomes.”
Mid-level Business Data Analyst specializing in healthcare analytics
“Analytics-focused candidate with strong SQL, Excel, Python, and Tableau skills who supports payroll-, compensation-, and finance-adjacent processes through rigorous data validation and reconciliation. Stands out for uncovering a duplicate-record mapping issue that exposed roughly $250K in revenue leakage and for building repeatable controls, dashboards, and automated checks to improve reporting accuracy.”
Mid-level Data Engineer specializing in cloud data platforms and AI/ML pipelines
“Data-engineering-oriented candidate with hands-on experience building an agentic AI product and operational automation workflows. They described automating inventory-to-ERP discrepancy reconciliation with anomaly detection and daily reporting, and also have practical scraping/automation experience dealing with Cloudflare-protected sites using Selenium and Puppeteer.”
Mid-level Data Scientist specializing in GenAI, customer insights, and forecasting
“ML/AI practitioner with hands-on experience shipping production time-series forecasting and RAG-based customer insights platforms in an enterprise setting. At BASF, he improved seed sales forecasting beyond naive baselines using model selection tailored by brand size, and he also led a RAG solution over Salesforce reports, complaints, and surveys that reached 2,000+ users with strong daily engagement.”
Junior AI/ML Software Engineer specializing in backend systems and cloud deployment
“Built multiple end-to-end automation and data systems, including an Accio RAG pipeline combining PDF parsing, FastAPI, Neo4j, and vector search, plus Selenium-based scraping for a virtual try-on product. Stands out for reliability-minded engineering: automated testing, structured logging, validation layers, and a data-driven approach to debugging flaky automation that improved CI pass rates to over 98%.”
Junior Data Scientist specializing in ML research, NLP, and healthcare analytics
“Completed an Amazon externship building a GPT-4 + RAG pipeline to summarize themes from hundreds of employee reviews for workforce analytics aimed at improving warehouse retention. Emphasizes production-readiness through labeled-data evaluation, source attribution for explainability, human-in-the-loop review, and rigorous data cleaning/observability to debug real-world LLM workflow issues.”
Junior AI & ML Engineer specializing in agentic systems and full-stack AI products
“Won a machine learning contest and was placed onto a Kaiser data science team, where they built ML models for hospital bottleneck prediction and resource allocation. They later built and deployed a full-stack LLM-based “data analyst agent” (with custom orchestration plus LangChain/OpenAI Agents experience) that generates analysis code, answers questions, and produces dashboards from uploaded datasets, emphasizing rigorous evaluation sets, robustness, and healthcare security/compliance integration.”
Senior Data Engineer specializing in cloud data platforms and regulated analytics
“Data engineer at Capital One building AWS-based real-time and batch pipelines and backend data services for financial/fraud use cases. Has owned end-to-end pipelines processing millions of records/day, implemented dbt/Great Expectations quality gates, and tuned Redshift/Snowflake workloads (cutting query latency ~22–25% and reducing pipeline failures ~30–40%) while supporting 15+ downstream consumers.”
Mid-level Data Engineer specializing in cloud data platforms and big data pipelines
“Healthcare data engineer with hands-on ownership of claims/member data pipelines on a cloud analytics platform, spanning batch and streaming ingestion (Airflow/Kafka/Spark/Databricks) through serving for reporting. Emphasizes reliability and data quality via embedded validation, schema-drift detection, deduplication, and operational monitoring/incident response, plus pragmatic CI/CD and observability setup in early-stage/ambiguous projects.”
Mid-level Data Engineer specializing in cloud ETL pipelines (Azure, AWS, GCP)
“Data engineer/backend developer who owned end-to-end pipelines and external data collection systems, including API ingestion and large-scale web scraping. Worked at ~50M records/month scale, improving processing speed by 20% and reducing reporting errors by 15%, and shipped a Rust-based internal data API with versioning, caching, and strong validation/observability practices.”
Mid-level Data Engineer specializing in lakehouse ETL and analytics engineering
“Data engineer with strong end-to-end ownership of production lakehouse pipelines (Snowflake + Databricks + Airflow + dbt + Great Expectations), handling 8M+ records/month and 500K+ daily CDC updates. Delivered measurable reliability and efficiency gains (41% cost reduction, freshness improved from 4h to 30m, 35% fewer downstream incidents) and has experience building a lakehouse platform from scratch across 12 source systems.”