Pre-screened and vetted.
Mid-level Data Engineer specializing in cloud ETL and big data pipelines
“Data engineer focused on building reliable, production-grade pipelines and data services end-to-end, including a 50+ GB/day pipeline ingesting from APIs/files into Snowflake with PySpark/SQL transformations. Emphasizes strong data quality controls, monitoring/retries, and performance optimization, and has also shipped a Python data API with caching and backward-compatible versioning.”
Mid-level Machine Learning Engineer specializing in real-time AI and data platforms
“ML/NLP engineer who has built production systems end-to-end: a real-time recommendation platform (100k+ profiles) using BERTopic-style clustering and a RAG-based news summarization/recommendation stack with ChromaDB. Strong focus on scaling and reliability (GPU batching, Redis caching, Kafka ingestion, Docker/Kubernetes, Prometheus/Grafana) and on maintaining model quality over time via drift monitoring and retraining triggers.”
Mid-Level Software Engineer specializing in Healthcare Data Platforms
“Backend/ML engineer with healthcare domain experience building secure Medicare/Medicaid data APIs and real-time patient risk scoring. Shipped an end-to-end ML pipeline (scikit-learn/XGBoost) served via SageMaker and integrated into Flask APIs, with strong production reliability practices (Kafka schema validation, regression replay, observability, drift monitoring, and human-in-the-loop guardrails).”
Junior Data Analyst & Business Analyst specializing in BI, analytics, and process optimization
Intern Data Scientist specializing in LLM agents, RAG, and real-time ML pipelines
Mid-level Data Scientist specializing in GenAI, MLOps, and computer vision for robotics
Mid-level Data Engineer specializing in cloud-native batch and streaming pipelines
Mid-level Data Engineer specializing in cloud ELT pipelines and analytics engineering
“Data engineer who has owned end-to-end ELT pipelines on Airflow + AWS (S3/Glue/Lambda) with Snowflake/Redshift, processing millions of records per day and tens of GBs via PySpark. Built strong data quality and reliability practices (40% quality improvement, 99%+ uptime), and also designed a resilient web-scraping system with anti-bot defenses and schema-change versioning plus REST APIs for serving curated data.”
Intern AI/ML & Data Engineer specializing in deep learning, NLP, and cloud data pipelines
“AI/ML practitioner with production experience building a RAG-powered contextual customer support agent, optimizing for low latency using vector databases and smaller LLMs. Also deployed a fraud detection model on Kubernetes with auto-scaling for heavy transactional loads, and improved chatbot accuracy by 15% through metric-driven testing and evaluation. Partners with Marketing on personalization/recommendation initiatives with measurable outcomes tied to customer feedback.”
Junior Full-Stack Software Engineer specializing in cloud-native microservices and data pipelines
Junior Full-Stack Data Engineer specializing in data pipelines and analytics