Pre-screened and vetted.
Intern Software Engineer specializing in backend systems and data engineering
“Backend/AI engineer who has built and shipped two products: Know Founder (Python/SQL/AWS) scaling to 2,000+ users in the first month, and Unifr (unifr.online), an AI search visibility engine that queries multiple LLMs and turns responses into structured brand insights. Strong in production reliability/performance (Redis caching, indexing, precomputation) and in designing agentic workflows with guardrails, validation, retries, and human escalation.”
Mid-level Data Engineer specializing in real-time pipelines and cloud data platforms
“Backend engineer with hands-on experience building secure Python/Flask services (sessions, JWT, RBAC) and optimizing PostgreSQL/SQLAlchemy performance, including custom SQL using CTEs/window functions profiled via EXPLAIN ANALYZE. Also integrates LLM features via OpenAI/Azure into backend systems and improves scalability with RabbitMQ-driven async processing, caching, and multi-tenant data isolation patterns.”
Mid-level Data Engineer specializing in capital markets post-trade data platforms
“Data/streaming engineer in capital markets who led an end-to-end trade settlement data product (Kafka→MongoDB→data lake) with rigorous data-quality logic and ~$175K first-year operational impact. Also built a low-latency Go-based CME market data engine feeding SOFR curve generation, using MSK on EKS with performance tuning (idempotency, compression, partitioning) to achieve sub-100ms delivery.”
Senior Data Analyst specializing in healthcare and financial analytics
“Healthcare analytics candidate with hands-on experience turning messy claims data in Redshift and S3 into validated reporting tables, plus automating KPI workflows in Python. They’ve owned end-to-end operational analytics projects, including a claims delay analysis that improved processing efficiency by about 20%, and have experience driving stakeholder adoption of standardized metrics across dashboards.”
Junior Data Analyst specializing in financial and operational analytics
“Analytics professional with experience at KPMG turning messy operational and financial data from SQL Server and AWS S3 into clean reporting datasets and automated Python workflows. They combine SQL, Python, Power BI, and experimentation methods to deliver stakeholder-aligned KPI dashboards and marketing performance insights with a strong focus on data integrity and reproducibility.”
Junior Software Engineer specializing in full-stack, data engineering, and mobile apps
“Built production LLM agents at Hivenue and Amazon, spanning consumer booking automation and internal data-query/reporting workflows. Stands out for combining conversational UX with strong reliability engineering—strict tool use, state machines, schema validation, idempotency, and evaluation pipelines—and can point to measurable impact including a 21% reduction in time to book and a 12% conversion lift.”
Junior Business & Data Analyst specializing in analytics and AI-driven insights
“Master’s in Business Analytics candidate with hands-on project experience spanning FMCG sales analytics, insurance risk modeling, and HR attrition analysis. Demonstrates strong SQL and Python fundamentals, including advanced CTE/window-function work, reproducible modeling workflows, and Power BI dashboards that translate analysis into clear business actions.”
Mid-level Data Analyst specializing in financial and healthcare analytics
“Analytics professional with experience at Franklin Templeton and IQVIA India, focused on turning messy cross-system data into trusted reporting and actionable business insights. Stands out for combining SQL, Python, AWS ETL, and BI dashboards to solve data quality issues, improve investor engagement analysis, and standardize commercial reporting in financial services and pharma contexts.”
Mid-level Business Analyst specializing in finance, insurance, and data analytics
“Business/data analyst with experience at KPMG and Liberty Mutual, focused on financial reporting, data quality, and analytics automation. Has built SQL and Python workflows for large transaction datasets, reduced manual reporting effort by 15+ hours per week, and translated ambiguous business questions into standardized KPIs and Power BI dashboards used for decision-making.”
Senior Data Scientist and AI/ML Engineer specializing in GenAI and cloud ML
“ML/AI engineer with hands-on experience owning systems from experimentation through deployment and monitoring, including a Bank of Montreal project that improved timely interventions by 12%. Also brings GenAI/RAG experience with evaluation and safety guardrails, plus clinical NLP pipeline work extracting medication data from notes for patient risk prediction.”
Junior Data Scientist specializing in fraud analytics and cloud data platforms
“Built and deployed production LLM-powered document summarization/classification systems using embeddings, vector databases (RAG-style retrieval), and automated evaluation (BERTScore/ROUGE), with a focus on monitoring and scalable cloud pipelines. Also partnered with a fraud analytics team to deliver a transaction anomaly detection solution, translating model outputs into Power BI dashboards and actionable KPIs while iterating on thresholds and alerts based on stakeholder feedback.”
Senior Data Analyst specializing in cloud data platforms, experimentation, and predictive analytics
“Healthcare data/ML practitioner with experience at UnitedHealth Group building production ETL and streaming pipelines (Python, BigQuery, Kafka) that unify EHR, IoT device, and lab data for patient risk prediction. Also implemented embedding-based semantic search/linking for noisy clinical notes via domain adaptation and rigorous validation with clinical stakeholders; previously built churn prediction at DirecTV using XGBoost.”
Mid-level Data Engineer specializing in cloud data platforms, Spark, and streaming pipelines
“Data/MLOps engineer (Cognizant background) who owned an AWS/Airflow/Snowflake healthcare transactions pipeline processing ~8–10M records/day and cut pipeline/data-quality incidents by ~33%. Also built and deployed a production FastAPI model-inference service on Kubernetes (Docker, HPA) with strong observability (Prometheus/Grafana), versioned endpoints, and resilient backfill/idempotent external data ingestion patterns.”
Intern Data Scientist specializing in ML systems and LLM-powered analytics
“Built an autonomous decision analytics LLM agent for end-to-end tabular binary classification, using RAG (FAISS) to retain context across multi-step queries. Deployed as a FastAPI service with production-style reliability features (schema-aware validation, fallbacks, retries, structured outputs) plus offline/online evaluation and monitoring to reduce analysis time and improve consistency versus stateless approaches.”
Mid-level Data Engineer specializing in cloud data pipelines and streaming
“Data engineer with experience at Wells Fargo and Accenture owning end-to-end production pipelines processing hundreds of millions of transactional/risk records daily. Strong focus on data quality and reliability (reconciliation checks, schema drift detection, CloudWatch alerting) plus Spark performance tuning and idempotent backfills using Delta Lake/merge logic across AWS (S3/EMR/Databricks/Redshift) and Azure (ADF/Azure DevOps/Azure Monitor).”
Mid-level Data Engineer specializing in AWS/Azure pipelines and streaming analytics
“Data engineer with experience across healthcare and geospatial risk systems, owning end-to-end pipelines from ingestion through serving on AWS/Azure stacks. Built HIPAA-compliant data quality gates and CDC for millions of daily claims, and also delivered a real-time wildfire risk platform with 20-minute refresh cycles and a 60% data accuracy lift. Strong in streaming (Kafka), Spark performance tuning, and production-grade orchestration/CI/CD (Airflow, Docker, Jenkins, GitHub Actions, Terraform).”
Senior Data Engineer specializing in cloud data platforms and automated data quality
“Data engineer at CenterPoint Energy who built and operated multiple production-grade GCP data systems: a daily Snowflake→BigQuery replication framework (150+ tables) with Monte Carlo/Atlan-driven observability and schema-drift protection, plus a FastAPI metrics service for pipeline health. Demonstrated measurable impact (40% faster dashboard queries, 70% less manual refresh work, zero data loss) and strong operational rigor (scaling Cloud Run jobs, SAP SLT reconciliation, quarantine patterns, CI/CD via GitHub Actions + Terraform).”
Mid-level Data Analyst specializing in procurement, supply chain analytics, and applied machine learning
“Strategic sourcing professional specializing in seasonal apparel supply chains, combining Coupa/JD Edwards analytics with Excel/Python modeling and Power BI dashboards to drive cost reduction and OTIF gains. Notable for rapid mitigation of a 10-day factory delay affecting 12 holiday SKUs (preserved 95% of revenue) and for automating PO workflows to cut cycle time by 4.2 days and improve OTIF by 15%.”
Mid-level Data Analyst specializing in healthcare and financial analytics
“Built and productionized an LLM-powered clinical documentation and insights pipeline at Cardinal Health using LangChain + GPT-4 with RAG to summarize long clinical notes, extract medication/dosage entities, and generate structured SQL-ready outputs for downstream analytics. Emphasizes clinical reliability via labeled benchmarking (precision/recall/F1), shadow deployments, clinician human-in-the-loop review, and ongoing monitoring/orchestration with Airflow, Lambda, S3, Postgres, and Power BI.”
Junior Data Scientist specializing in healthcare ML and clinical NLP/LLMs
“Healthcare-focused LLM engineer who has built two production clinical applications: an automated structured clinical report generator from physician-patient conversations and a RAG-based chatbot for retrieving patient history (procedures, allergies, etc.). Demonstrates strong applied RAG expertise (overlapping chunking, entity dependency graphs, temporal filtering, graph RAG) to reduce hallucinations/omissions and partners closely with clinicians to automate hospital workflows.”
Mid-level Data Engineer specializing in cloud ETL and financial data platforms
“Data engineer with experience at Capital One and HSBC building and operating GCP-based data platforms. Led an end-to-end Oracle-to-BigQuery migration processing ~200–300GB/day using Dataflow/Beam, Airflow, Dataproc/PySpark, and Looker, achieving ~99.5% pipeline success and ~30% fewer data quality issues. Strong in production reliability, schema drift handling for external APIs, and BigQuery performance/serving patterns (materialized views, authorized views, versioned datasets).”
Senior Business Analyst specializing in AI and commercial banking analytics
“Analytics candidate with hands-on experience supporting a workforce system transformation from symplr to Oracle Fusion Time and Labor, using SQL and Python to turn operational HR, attendance, and payroll data into reporting-ready datasets. They emphasize performance optimization, reusable analytics pipelines, and metric consistency across dashboards, with project work focused on overtime reduction, workforce efficiency, and retention trends by department.”
Junior Data Analyst specializing in ML, NLP, and cloud data pipelines
“Built and deployed a GenAI-powered PhD career intelligence platform at NYU that maps academic backgrounds to career paths and converts long academic CVs into job-ready resumes. Stands out for treating LLM systems as structured production pipelines—combining NLP extraction, embeddings, orchestration, and AWS deployment—to improve recommendation quality and cut resume preparation time by 70%.”
Mid-level Data Analyst specializing in healthcare and financial analytics
“Analytics-focused candidate with hands-on experience turning messy CRM, e-commerce, payments, and support data into trusted reporting datasets using SQL and Python. They have owned end-to-end churn and retention analytics work, including RFM-based segmentation, dashboard delivery, and metric standardization across sales, marketing, and finance.”