Pre-screened and vetted.
Mid-Level Software Engineer specializing in AI/ML and distributed systems
“Software engineer with production experience building a serverless monolith and multi-layer video pipeline at easyML, plus hands-on integration of multiple LLM providers (Grok/Claude/OpenAI) into a full-stack app. Interested in robotics via computer vision (OpenCV/OpenMMLab), with a strong real-time systems mindset around SLOs, latency, determinism, and reliability; also has low-level OS experience writing a keyboard device driver.”
Intern AI/ML Engineer specializing in agentic systems and full-stack development
“Built and scaled a multi-agent LLM automation pipeline during a fintech internship, growing from a rapid 1-week proof-of-concept to a 15+ agent hierarchical system that cut market brief report generation time from ~5 hours to under 30 minutes. Hands-on with agent frameworks (Haystack, CrewAI, LangChain) and experienced in debugging agent communication issues via sandboxed modular testing and context/token management; also regularly gives architecture-first technical demos at multiple hackathons and university events.”
Mid-level Sales Engineer & Solution Architect specializing in cloud and data platforms
“LLM-focused customer-facing technical leader with experience productionizing LLM workflows in financial services (State Street), including guardrails, retrieval tuning, and reliability improvements. Also partners closely with sales and executives—at Payoneer helped drive enterprise-wide adoption for a $10M ARR global account through technical discovery, demos, and pilots.”
Mid-level Data Scientist specializing in MLOps, LLM/RAG applications, and deep learning
“Built and deployed a production compliance automation RAG system (at Citi) that generates citation-backed, schema-validated risk summaries for regulatory document review. Emphasizes regulated-environment reliability with retrieval-only grounding, abstention, confidence thresholds, and immutable audit logging, plus orchestration using LangChain/LangGraph and Airflow. Reported ~60% reduction in compliance review effort while maintaining high precision and traceability.”
Mid-level GenAI Engineer specializing in production AI agents and evaluation pipelines
“Built and shipped a production LLM-powered internal operations automation platform using LangChain RAG (Pinecone) and FastAPI microservices, deployed on AWS EKS, serving 10k+ daily interactions. Implemented a rigorous evaluation/observability stack (golden datasets, prompt regression tests, MLflow, retrieval metrics, hallucination monitoring) that drove hallucinations below 2% and improved reliability, and partnered closely with non-technical ops leaders to cut manual lookup work by 60%+.”
Mid-Level Backend Software Engineer specializing in Java/Spring microservices and AWS
“Backend-focused engineer with production experience building Spring Boot services for automated workflow and data-processing platforms, using queues plus retry and idempotency patterns. Also uses Python to automate data processing; emphasizes testing and peer review for maintainability.”
Senior Full-Stack Java Developer specializing in cloud-native microservices
“Backend/platform engineer with production ownership of high-volume transaction analytics and fraud monitoring services built in Java/Spring Boot. Has scaled data processing platforms (including healthcare datasets) and operated Kafka-based event pipelines with schema versioning, deduplication, and replay/backfill workflows, using strong observability via CloudWatch/Grafana and CI/CD with Jenkins.”
Mid-level Data & GenAI Engineer specializing in lakehouse, streaming, and RAG platforms
“Built a production internal LLM-powered knowledge assistant using a RAG architecture (Python, LLM APIs, cloud services) that answers employee questions with sourced, grounded responses from internal documents. Demonstrates strong practical depth in retrieval tuning (chunking/metadata filters), orchestration with LangChain, and production reliability practices (latency optimization, automated embedding refresh, evaluation metrics, logging/monitoring) while partnering closely with non-technical operations teams.”
Mid-level Data Engineer specializing in cloud data pipelines and financial services warehousing
“Data engineer (Charles Schwab) who took ownership of an unstable, ambiguous nightly financial data pipeline and rebuilt it into a reliable, incremental AWS Glue/Airflow/Redshift system feeding Power BI. Created a custom Python data-quality framework with hard-stop gating and schema drift detection, improving integrity (99.9%), cutting runtime (~20%), and reducing incidents/tickets (35% fewer schema-related dashboard incidents; 30% fewer investigations).”
Mid-level Data Analyst specializing in financial risk and healthcare analytics
“AI/ML engineer focused on real-time, production-grade LLM systems, with a robotics-adjacent mindset around latency/accuracy tradeoffs and modular pipelines. Built a scalable RAG-based assistant orchestrated as microservices on Kubernetes with Kafka async messaging, ONNX/quantization optimizations, and monitoring (Prometheus/Grafana), citing a ~35% hallucination reduction; has also experimented with ROS Noetic/Gazebo to understand ROS concepts.”
Senior ML Engineer & Data Scientist specializing in LLM agents, retrieval/ranking, and MLOps
“Machine Learning Engineer currently at Webster Bank building an enterprise-scale LLM agent for Temenos Journey Manager/Maestro, using RAG-style multi-stage retrieval with FAISS/Pinecone, hybrid dense+sparse search, and LoRA fine-tuning optimized via NDCG/MAP and A/B testing. Previously handled messy incident/telemetry data at Deuta Werke GmbH with deterministic + fuzzy entity resolution, and has strong production data engineering experience across Spark/Hadoop and Python ETL systems.”
Mid-level Full-Stack Java Developer specializing in FinTech and Healthcare platforms
“Software engineer who built internal operations/monitoring dashboards for real-time trading and money-movement systems, emphasizing auditability and rapid iteration. Deep experience with microservices on Azure using Kafka/RabbitMQ, plus strong testing discipline (JUnit/Mockito/Testcontainers, contract/E2E) and observability patterns (correlation IDs, centralized logging, distributed tracing) to reduce incident triage time and improve resilience.”
Mid-Level Software Engineer specializing in Java microservices and AWS cloud-native systems
“Full-stack engineer who has owned customer-critical analytics and course intelligence platforms end-to-end (React/TypeScript + Node/Express + SQL), including an internal self-serve Reporting & Analytics Center adopted by 1,000+ users. Demonstrates strong systems thinking across performance (2× faster heavy reports), reliability (feature flags, testing), and distributed architecture (RabbitMQ microservices with idempotency, DLQs, and correlation-ID observability).”
Director-level Engineering Leader specializing in enterprise SaaS and cloud-native platforms
“Engineering leader/player-coach who modernized a legacy C#/SQL Server system to Snowflake + Python on GCP, enabling ~30x scale and supporting hundreds of millions of transactions per day per customer. Strong in architecture tradeoffs (Snowflake vs Databricks), production reliability (New Relic, logging/alerting), and lightweight process improvements like a rigorous Definition of Done and structured PR reviews.”
Mid-level Data Engineer specializing in cloud lakehouse and streaming platforms
“Data engineer focused on building production-grade pipelines on AWS (Kafka/Kinesis/Glue/S3) through to curated serving layers in Snowflake and Delta Lake. Emphasizes automated data quality validation (PySpark + CI/CD), modular dbt transformations for analytics (customer spending, risk metrics), and operational reliability with CloudWatch and DLQs; data consumed by BI tools and ML pipelines for fraud detection and risk analytics.”
Mid-level Data Engineer specializing in multi-cloud real-time and batch data pipelines
“Data engineer with healthcare domain experience who owned 100M+ record pipelines end-to-end (Kafka/Kinesis/ADF → PySpark/dbt validation → Spark SQL transforms → Snowflake/Power BI serving). Built production-grade reliability practices (Airflow orchestration, CloudWatch/Grafana monitoring, pytest + contract/regression tests, idempotent ingestion/backfills) and delivered measurable improvements: 35% lower latency and 40% better query performance.”
Mid-level Data Engineer specializing in real-time pipelines and cloud data platforms
“Backend engineer with hands-on experience building secure Python/Flask services (sessions, JWT, RBAC) and optimizing PostgreSQL/SQLAlchemy performance, including custom SQL using CTEs/window functions profiled via EXPLAIN ANALYZE. Also integrates LLM features via OpenAI/Azure into backend systems and improves scalability with RabbitMQ-driven async processing, caching, and multi-tenant data isolation patterns.”
Senior Data Analyst specializing in healthcare and financial analytics
“Healthcare analytics candidate with hands-on experience turning messy claims data in Redshift and S3 into validated reporting tables, plus automating KPI workflows in Python. They’ve owned end-to-end operational analytics projects, including a claims delay analysis that improved processing efficiency by about 20%, and have experience driving stakeholder adoption of standardized metrics across dashboards.”
Junior Data Analyst specializing in financial and operational analytics
“Analytics professional with experience at KPMG turning messy operational and financial data from SQL Server and AWS S3 into clean reporting datasets and automated Python workflows. They combine SQL, Python, Power BI, and experimentation methods to deliver stakeholder-aligned KPI dashboards and marketing performance insights with a strong focus on data integrity and reproducibility.”
Intern AI/ML Engineer specializing in full-stack and data systems
“Built an LLM-powered customer segmentation agent during a Chewy internship, consolidating Snowflake data into a knowledge graph so non-technical marketing users could query customer cohorts in natural language. Stands out for combining agent/tooling design with rigorous data engineering practices, including schema audits, imputation, validation layers, and idempotent pipelines on messy large-scale datasets.”
Mid-level Business Analyst specializing in finance, insurance, and data analytics
“Business/data analyst with experience at KPMG and Liberty Mutual, focused on financial reporting, data quality, and analytics automation. Has built SQL and Python workflows for large transaction datasets, reduced manual reporting effort by 15+ hours per week, and translated ambiguous business questions into standardized KPIs and Power BI dashboards used for decision-making.”
Senior Data Scientist and AI/ML Engineer specializing in GenAI and cloud ML
“ML/AI engineer with hands-on experience owning systems from experimentation through deployment and monitoring, including a Bank of Montreal project that improved timely interventions by 12%. Also brings GenAI/RAG experience with evaluation and safety guardrails, plus clinical NLP pipeline work extracting medication data from notes for patient risk prediction.”
Mid-level AI/ML Engineer specializing in FinTech and retail ML systems
“ML-focused candidate with strong Wells Fargo experience building production fraud systems and internal GenAI tools for fraud analysts. Stands out for measurable impact in fraud detection—raising recall from 71% to 88%—while also demonstrating hands-on depth across streaming infrastructure, MLOps, LLM/RAG implementation, and Python service architecture.”
Senior Software Engineer specializing in backend systems and data platforms
“Software developer who uses AI pragmatically across the full stack to accelerate coding, testing, debugging, and documentation while maintaining strong human oversight. Stands out for treating AI output like any other code source—reviewing for architecture fit, security risks, performance, and standards before integration—and for coordinating multiple AI tools across backend, frontend, and test workflows.”