Pre-screened and vetted.
Mid-level Software Engineer specializing in data pipelines and backend APIs
“Data engineer with Webster Bank experience owning end-to-end pipelines (APIs + databases) processing millions of records/day, improving data quality (25–30% fewer issues) and reliability (~99.9% successful runs). Built resilient external data ingestion/scraping systems (schema-change validation, idempotent backfills, monitoring/alerts) and shipped a FastAPI service exposing curated datasets with versioning and consistently low latency.”
Senior Full-Stack Java Developer specializing in capital markets and trading systems
“Backend/data engineer with production experience in payment initiation/processing services built in Python/FastAPI, emphasizing reliability patterns (JWT/RBAC, timeouts, retries, circuit breakers). Has delivered AWS deployments on ECS (ALB, autoscaling, CI/CD to ECR) plus Lambda-based reporting, and built AWS Glue ETL pipelines with schema evolution and CloudWatch monitoring. Also modernized a legacy SAS reporting platform to Python/PostgreSQL with regression parity testing and parallel-run migration, and achieved a 70% SQL performance improvement.”
Mid-level Data Analyst/Data Engineer specializing in BI, ETL pipelines, and cloud analytics
“Data engineer focused on marketing/web analytics and external API pipelines, handling ~10M records/week. Built Azure-based ingestion and PySpark transformations with rigorous data quality checks, then served curated datasets into Synapse/Redshift for Power BI. Also designed an Airflow-orchestrated crypto REST API pipeline with monitoring, retries/exponential backoff, schema-change detection, and backfill-friendly reprocessing.”
Mid-level Data Analyst specializing in financial risk and data automation
“Analytics professional from Capital One with strong experience automating risk, reconciliation, and regulatory reporting workflows in financial services. They combine deep SQL/Python pipeline skills with stakeholder-facing dashboard and KPI design, delivering measurable impact like 30% performance gains, sub-24-hour anomaly detection, and 100% data integrity for regulatory filings.”
Senior AI/ML Engineer specializing in LLMs, generative AI, and applied research
“Research-heavy ML/AI candidate with a PhD/publications background who translated LLM evaluation and clinical summarization techniques into production at ModMed. They owned an end-to-end healthcare GenAI pipeline that cut clinician documentation time from ~22 minutes to ~7-8 minutes, reduced token costs by ~30%, and built an internal evaluation framework later adopted by multiple teams.”
Mid-level Forward Deployed Engineer specializing in backend systems and FinTech
“Backend-focused engineer with experience at Charles Schwab owning financial workflow deployments end-to-end, including API/database design, SQL optimization, Python automation, and AWS-based production stabilization. Also brings applied AI quality experience through building LLM/agent validation pipelines focused on scenario testing, edge-case detection, and reducing production risk.”
“Software engineer currently building AI-powered backend systems for interview analysis, with end-to-end ownership of an LLM-based monitoring platform. Stands out for combining practical product delivery in an ambiguous early-stage environment with measurable impact: over 40% reduction in manual review effort and roughly 20% lower inference cost.”
Senior AI/ML Engineer specializing in Generative AI, LLMs, and MLOps
“Telecom (Verizon) AI/ML practitioner who built a production multimodal system that ingests messy customer issue reports (calls, chats, emails, screenshots, videos) and turns them into confidence-scored incident summaries with reproducible steps and evidence links. Also built KPI/alarm-to-ticket correlation to rank likely root-cause domains (RAN/Core/Transport), cutting triage from hours to minutes and improving MTTR.”
Junior Data Scientist/Data Engineer specializing in ML pipelines and analytics
“Machine Learning Intern at Docsumo who delivered a customer-facing fraud-detection solution end-to-end: rebuilt the pipeline, deployed a Random Forest model, and shipped a Python/Flask microservice on AWS SageMaker. Drove measurable production impact (precision +30%, processing time cut in half, manual review -60%, customer satisfaction +15%) and demonstrated strong customer integration and live-incident response skills.”
Mid-level Backend Software Developer specializing in cloud-native microservices
“LLM-focused engineer who has shipped multiple production-grade AI reliability systems: an LLM output validation/monitoring service (FastAPI) with prompt versioning and failure analytics, plus a RAG feature using embeddings/vector DBs with retrieval thresholds, schema/context validation, and safe fallbacks. Strong in evaluation loops (groundedness, schema accuracy, human review) and scalable pipelines for messy document ingestion with observability and early detection of data quality issues.”
Mid-Level Backend Engineer specializing in Java/Spring Boot and LLM-integrated microservices
“Built and deployed a live production LLM document Q&A platform (DocumindAI) with an adaptive RAG pipeline (Claude + Cohere embeddings + pgvector), source-cited structured outputs, and engineered fallbacks for reliability and sub-2s latency. Also has enterprise integration experience at Tech Mahindra working with messy IFS ERP XML integrations, using validation/normalization and JTA transactions to prevent partial writes and data corruption.”
Mid-level Data Engineer specializing in real-time streaming and cloud data platforms
“Data engineer with Wells Fargo experience owning an end-to-end lakehouse ETL pipeline on Databricks/Azure Data Factory, processing ~480GB daily and implementing robust data quality/reconciliation across 40+ tables to reach ~99.3% reliability. Strong in performance optimization (cut runtime 5.5h→3.8h), CI/CD and monitoring, and resilient external/API ingestion with retries, schema validation, and backfills.”
Intern Machine Learning Engineer specializing in forecasting, NLP, and RAG systems
“Intern who built and deployed a production LLM-powered contract analysis system for finance teams: Azure Document Intelligence for text/table extraction plus Gemini prompting to surface key terms and risks via an async API and simple UI. Emphasizes reliability in production with fallbacks, guardrails against hallucinations, and operational concerns like latency/cost/versioning, delivering summaries in under 30 seconds instead of hours.”
Senior Data Engineer specializing in Spark, Kafka, and Databricks Lakehouse platforms
“Data engineer at Fidelity who built and operated a real-time financial transactions lakehouse on AWS/Databricks, processing millions of records daily with Kafka streaming. Demonstrated strong reliability and data quality practices (watermarking, idempotent Delta writes, validation/reconciliation, observability) and delivered measurable improvements (~30% faster jobs and ~30% fewer data issues) while enabling trusted gold-layer analytics for downstream teams.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and healthcare ML systems
“Healthcare ML/AI engineer at Cigna who has owned a clinical RAG pipeline from prototype through production, monitoring, compliance, and iteration. Stands out for combining LLM product delivery with healthcare-grade safety and explainability, driving a 38% retrieval precision gain, 42% hallucination reduction, and meaningful improvements in team velocity and system reliability.”
“Senior AI/ML engineer focused on production ML, LLMs, and MLOps, with concrete experience shipping fraud detection and enterprise RAG systems. They combine strong deployment and monitoring discipline with measurable business impact, including 31% precision improvement in fraud detection and 37% better answer relevance in a financial-document QA system.”
Senior AI/ML Engineer specializing in Generative AI, NLP, and regulated industries
“Built end-to-end ML and GenAI systems at Northern Trust, including a production RAG-based document intelligence platform for financial reports and contracts. Stands out for combining strong MLOps execution with practical product judgment—improving forecast accuracy by 22%, document review accuracy by 38%, and cutting deployment time by 45% while keeping latency and reliability production-ready.”
Mid-level Backend Software Engineer specializing in cloud-native microservices
“Backend/platform engineer with experience across Cigna, Cognizant, and a university environment, focused on reliability, distributed systems, and regulated-domain workflows. Stands out for combining Kubernetes/Kafka/AWS infrastructure expertise with a production RAG-based healthcare compliance assistant that cut manual reporting work from 30-45 minutes to under 2 minutes while maintaining strong uptime and data-quality controls.”
“Engineer with a thoughtful, hands-on approach to AI-assisted software development, treating AI as a force multiplier for debugging, prototyping, and large-codebase work rather than a substitute for judgment. Particularly strong in multi-agent coding workflows, contract-driven development, and maintaining consistency across backend, frontend, and testing through shared schemas and OpenAPI-based coordination.”
Mid-level AI/ML Engineer specializing in NLP and Generative AI
“Built and deployed a production LLM-powered RAG assistant for healthcare teams (care managers/support) to answer questions from clinical and policy documentation, emphasizing trustworthiness via improved retrieval, reranking, and strict grounding prompts to reduce hallucinations. Also has hands-on orchestration experience with Apache Airflow for end-to-end ETL/ML workflows and applies rigorous testing/metrics (hallucination rate, tool-call accuracy, latency, cost) to ensure reliable AI agent behavior.”
Junior Software Engineer specializing in AI agents and full-stack cloud systems
“Backend-focused engineer who has built and refactored FastAPI services backed by MongoDB, emphasizing async concurrency, stateless design for horizontal scaling, and performance tuning via indexing and request-level timing. Has implemented production authentication patterns (JWT, SSO, OAuth2 + PKCE) and user/org-scoped access controls, and improved reliability of LLM document-extraction APIs with fallback mechanisms.”
Mid-Level Software Developer specializing in full-stack, cloud-native microservices and AI integrations
“Backend/AI engineer who has built production Spring Boot APIs on AWS (JWT auth, Redis/MySQL) and solved a real-world silent data integrity issue by implementing idempotency keys plus DB constraints/transactions. Also shipped an LLM-based document Q&A feature using a RAG pipeline with evaluation + human review, and designed multi-step agent workflows with verification, retries, and escalation guardrails.”
Junior Software Engineer specializing in backend microservices and GenAI automation
“Built and productionized an LLM/RAG-based credit case triage system that indexes credit bureau narratives and delivers structured, source-linked summaries for analysts. Emphasizes production reliability (data masking, human-in-the-loop review, abstention/fallbacks, monitoring) and reports >99% uptime plus ~30% efficiency gains, with iterative improvements driven by offline/online evaluation and schema-aware validation.”