Pre-screened and vetted.
Mid-level Data Engineer specializing in real-time pipelines and cloud data platforms
“Backend engineer with hands-on experience building secure Python/Flask services (sessions, JWT, RBAC) and optimizing PostgreSQL/SQLAlchemy performance, including custom SQL using CTEs/window functions profiled via EXPLAIN ANALYZE. Also integrates LLM features via OpenAI/Azure into backend systems and improves scalability with RabbitMQ-driven async processing, caching, and multi-tenant data isolation patterns.”
Senior Data Analyst specializing in healthcare and financial analytics
“Healthcare analytics candidate with hands-on experience turning messy claims data in Redshift and S3 into validated reporting tables, plus automating KPI workflows in Python. They’ve owned end-to-end operational analytics projects, including a claims delay analysis that improved processing efficiency by about 20%, and have experience driving stakeholder adoption of standardized metrics across dashboards.”
Junior Data Analyst specializing in financial and operational analytics
“Analytics professional with experience at KPMG turning messy operational and financial data from SQL Server and AWS S3 into clean reporting datasets and automated Python workflows. They combine SQL, Python, Power BI, and experimentation methods to deliver stakeholder-aligned KPI dashboards and marketing performance insights with a strong focus on data integrity and reproducibility.”
Junior Business & Data Analyst specializing in analytics and AI-driven insights
“Master’s in Business Analytics candidate with hands-on project experience spanning FMCG sales analytics, insurance risk modeling, and HR attrition analysis. Demonstrates strong SQL and Python fundamentals, including advanced CTE/window-function work, reproducible modeling workflows, and Power BI dashboards that translate analysis into clear business actions.”
Mid-level Data Analyst specializing in financial and healthcare analytics
“Analytics professional with experience at Franklin Templeton and IQVIA India, focused on turning messy cross-system data into trusted reporting and actionable business insights. Stands out for combining SQL, Python, AWS ETL, and BI dashboards to solve data quality issues, improve investor engagement analysis, and standardize commercial reporting in financial services and pharma contexts.”
Mid-level Cloud DevOps/SRE Engineer specializing in Google Cloud
“SRE-oriented infrastructure engineer who built an internal Vertex AI/Gemini knowledge chatbot to centralize product and development documentation, cutting routine support questions from 10+ daily to roughly 2. Also brings hands-on experience debugging Kubernetes production incidents and monitoring ETL/data-quality issues in Dataflow-based pipelines.”
Mid-level Business Analyst specializing in finance, insurance, and data analytics
“Business/data analyst with experience at KPMG and Liberty Mutual, focused on financial reporting, data quality, and analytics automation. Has built SQL and Python workflows for large transaction datasets, reduced manual reporting effort by 15+ hours per week, and translated ambiguous business questions into standardized KPIs and Power BI dashboards used for decision-making.”
Mid-level Software Engineer specializing in AI, full-stack systems, and platform engineering
“Full-stack/AI engineer with experience spanning supply-chain product deployments, biomedical agentic search, and research-grade RAG evaluation. Stands out for owning customer-facing migrations at scale (including 216,000 historical shipments), building measurable LLM systems, and pairing AI experimentation with rigorous evals, rollout controls, and auditability.”
Mid-level Full-Stack Software Engineer specializing in AI and document automation
“Backend/AI infrastructure engineer focused on production-ready LLM systems and distributed workflows. They described building a RAG-based multi-step agent with strong reliability controls, evaluation loops, and graceful degradation that improved latency by 30%, retrieval accuracy by 15%, and reduced support workload by 40%.”
Mid-level Python Backend Engineer specializing in cloud-native and AI-powered systems
“Backend/AI engineer who has shipped an LLM-powered enterprise support-ticket agent at Comcast, building a production-grade microservices pipeline (FastAPI, SQS, Redis) with strong observability (OpenTelemetry/Splunk/Prometheus/Grafana) and reliability patterns (async, caching, circuit breakers, idempotency). Demonstrated quantified impact at scale—processing 10k+ tickets/day while improving response SLAs and routing accuracy through evaluation and human feedback loops.”
Senior Software Engineer specializing in backend systems and data platforms
“Software developer who uses AI pragmatically across the full stack to accelerate coding, testing, debugging, and documentation while maintaining strong human oversight. Stands out for treating AI output like any other code source—reviewing for architecture fit, security risks, performance, and standards before integration—and for coordinating multiple AI tools across backend, frontend, and test workflows.”
Mid-Level Full-Stack Software Engineer specializing in React, Node.js, and cloud-native systems
“Data engineer/backend engineer with healthcare domain experience at Centene, where they owned an end-to-end claims processing pipeline handling over 1 million monthly records. They combine Python/SQL pipeline work with API and event-driven service development, and cite a measurable 35% reduction in incident detection time through automated monitoring and validation.”
Mid-level Full-Stack Java Developer specializing in enterprise web applications
“Backend engineer who built and scaled a transaction-processing microservice (150K+ records/day) in a microservices ecosystem, debugging peak-load latency/timeouts via CloudWatch/Grafana, Kafka lag analysis, and DB query tuning (indexes, Redis caching, batching). Also shipped an LLM-powered document assistant end-to-end with prompt/response validation plus retries/fallbacks for production reliability.”
“Backend engineer focused on productionizing LLM systems: built a FastAPI-based RAG and multi-agent automation platform deployed with Docker/Kubernetes, prioritizing safe execution and reduced hallucinations. Experienced in refactoring monolithic ML services with feature-flagged incremental rollouts, and implementing JWT/RBAC plus row-level security (e.g., Supabase) for secure, scalable APIs.”
Junior Data Scientist specializing in fraud analytics and cloud data platforms
“Built and deployed production LLM-powered document summarization/classification systems using embeddings, vector databases (RAG-style retrieval), and automated evaluation (BERTScore/ROUGE), with a focus on monitoring and scalable cloud pipelines. Also partnered with a fraud analytics team to deliver a transaction anomaly detection solution, translating model outputs into Power BI dashboards and actionable KPIs while iterating on thresholds and alerts based on stakeholder feedback.”
Intern Machine Learning & AI Automation Engineer specializing in ML workflows and AI hardware
“ML practitioner with hands-on experience adapting diffusion models (DDPM + U-Net in PyTorch) to improve low-dose CT medical imaging quality via denoising and upsampling against high-dose ground truth. Also built a RAG workflow during a recent internship by cleaning client survey data, embedding with OpenAI text-embedding-3-large, and indexing in Pinecone with MD5 deduplication, alongside a strong emphasis on production-grade Python practices.”
Mid-level Instrumentation & Controls Engineer specializing in SCADA and industrial automation
“Operations/industrial automation engineer with several years supporting and upgrading controls, PLCs, networks, and IoT across 300+ North American sites. Led a zero-downtime IoT safety-device integration into an existing plant control/SCADA environment by building a parallel secure network and a Python/Flask + AWS/SQL telemetry pipeline, avoiding a major outage and saving ~$300K. Also co-founded an IoT + ML flood monitoring pilot shaped through direct collaboration with urban planners, emphasizing geospatial flood mapping for decision-making.”
Senior Data Analyst specializing in cloud data platforms, experimentation, and predictive analytics
“Healthcare data/ML practitioner with experience at UnitedHealth Group building production ETL and streaming pipelines (Python, BigQuery, Kafka) that unify EHR, IoT device, and lab data for patient risk prediction. Also implemented embedding-based semantic search/linking for noisy clinical notes via domain adaptation and rigorous validation with clinical stakeholders; previously built churn prediction at DirecTV using XGBoost.”
Mid-Level Software Engineer specializing in AI-enabled backend and full-stack web systems
“Backend/AI workflow engineer with experience at AirKitchenz, Uber, and Vivma Software, building production systems on AWS (Lambda, DynamoDB, Step Functions). Has a track record of major performance wins (DynamoDB latency 2s to <150ms; Postgres query 2s to ~180ms) and shipping LLM-powered onboarding and ticket-routing workflows with strong guardrails (schema validation, confidence thresholds, human-in-the-loop escalation).”
Junior Data Scientist / ML Engineer specializing in LLMs and Computer Vision
“Currently working in CoRAL Lab, built and deployed IntegrityShield—a document-layer PDF watermarking system that keeps assessments visually identical while disrupting LLM-based solving; validated in a real classroom where it helped catch 12 AI-cheating cases. Also built MALDOC, a modular red-teaming platform for document-processing AI agents using LangGraph to run reproducible, deterministic adversarial trials across OCR/text/vision routes.”
Mid-level Full-Stack Engineer specializing in cloud-native microservices (FinTech/Healthcare)
“Built and shipped production systems spanning real-time operational dashboards and an LLM-powered internal documentation assistant using RAG (embeddings + vector DB). Demonstrates strong focus on reliability and iteration: implemented guardrails and evaluation loops (human review, hallucination tracking, regression prevention) and improved performance/scalability through query optimization, caching, and retrieval tuning.”
Mid-level Data Engineer specializing in cloud data platforms, Spark, and streaming pipelines
“Data/MLOps engineer (Cognizant background) who owned an AWS/Airflow/Snowflake healthcare transactions pipeline processing ~8–10M records/day and cut pipeline/data-quality incidents by ~33%. Also built and deployed a production FastAPI model-inference service on Kubernetes (Docker, HPA) with strong observability (Prometheus/Grafana), versioned endpoints, and resilient backfill/idempotent external data ingestion patterns.”
“Data engineer/backend engineer with experience in healthcare (Cardinal Health provider enrollment) and finance (Northern Trust) building and stabilizing data pipelines and REST services. Worked with APIs and Kafka at ~200k–300k records/day, improving data quality (DLQ + validation), performance (SQL/indexing), and reliability/observability (logging, alerts, consumer lag metrics), and stood up an early-stage financial data service with Jenkins-based CI/CD.”
Senior Data Engineer specializing in cloud data platforms and automated data quality
“Data engineer at CenterPoint Energy who built and operated multiple production-grade GCP data systems: a daily Snowflake→BigQuery replication framework (150+ tables) with Monte Carlo/Atlan-driven observability and schema-drift protection, plus a FastAPI metrics service for pipeline health. Demonstrated measurable impact (40% faster dashboard queries, 70% less manual refresh work, zero data loss) and strong operational rigor (scaling Cloud Run jobs, SAP SLT reconciliation, quarantine patterns, CI/CD via GitHub Actions + Terraform).”