Pre-screened and vetted.
Junior Software Engineer specializing in cloud-native microservices
“Backend engineer (Nokia) who designs and migrates cloud-native microservices at scale, including a secure low-latency system handling 500k+ daily transactions. Strong in Kubernetes/OpenShift operations, CI/CD standardization, and production security (OAuth2/JWT/RBAC) with SOC2-aligned controls and zero critical security incidents. Demonstrated expertise in safe migrations (canary/blue-green, dual writes, reconciliation) and concurrency correctness in real-time systems.”
Mid-level Data Engineer specializing in cloud lakehouse and streaming platforms
“Data engineer focused on building production-grade pipelines on AWS (Kafka/Kinesis/Glue/S3) through to curated serving layers in Snowflake and Delta Lake. Emphasizes automated data quality validation (PySpark + CI/CD), modular dbt transformations for analytics (customer spending, risk metrics), and operational reliability with CloudWatch and DLQs; data consumed by BI tools and ML pipelines for fraud detection and risk analytics.”
Mid-level Data Engineer specializing in multi-cloud real-time and batch data pipelines
“Data engineer with healthcare domain experience who owned 100M+ record pipelines end-to-end (Kafka/Kinesis/ADF → PySpark/dbt validation → Spark SQL transforms → Snowflake/Power BI serving). Built production-grade reliability practices (Airflow orchestration, CloudWatch/Grafana monitoring, pytest + contract/regression tests, idempotent ingestion/backfills) and delivered measurable improvements: 35% lower latency and 40% better query performance.”
Senior Full-Stack Java Developer specializing in cloud-native microservices
“Full-stack engineer with experience building secure, cloud-native document/workflow platforms handling high-volume customer and medical data across microservices on Kubernetes. Demonstrated impact improving performance via event-driven AWS architectures (Lambda + DynamoDB Streams) and strengthening compliance/security for S3-stored documents using IAM and KMS. Has delivered end-to-end APIs and UIs using Java/Spring Boot with Angular/React, plus Docker and CI/CD.”
Mid-level Data Engineer specializing in real-time pipelines and cloud data platforms
“Backend engineer with hands-on experience building secure Python/Flask services (sessions, JWT, RBAC) and optimizing PostgreSQL/SQLAlchemy performance, including custom SQL using CTEs/window functions profiled via EXPLAIN ANALYZE. Also integrates LLM features via OpenAI/Azure into backend systems and improves scalability with RabbitMQ-driven async processing, caching, and multi-tenant data isolation patterns.”
Mid-Level Full-Stack Software Engineer specializing in cloud-native MERN microservices
“Full-stack engineer who built an internal user-activity tracking and reporting system end-to-end using React/TypeScript, Node/Express, and Postgres, deployed on AWS (EC2/ALB, S3/CloudFront) with CloudWatch observability. Emphasizes reliability and data correctness via idempotent ingestion, retries with exponential backoff, backfills/reconciliation, and performance tuning as data scales, and has experience shipping quickly in ambiguous early-stage startup conditions.”
Senior Software Engineer specializing in Python automation and hybrid cloud integration
“Embodied AI / robotics-focused ML engineer with experience at JPMorgan and EY building language-to-robot control systems that connect transformer/LLM intent to safe real-world robotic actions. Designed production-grade, low-latency architectures (Kafka/Redis, monitoring, CI/CD) and applied sim-to-real and model distillation to make research ideas deployable on physical systems.”
Mid-level Full-Stack Java Developer specializing in FinTech microservices
“Backend-focused Python/Flask engineer with strong performance and scalability experience across PostgreSQL/SQLAlchemy optimization, caching, and async processing. Has implemented multi-tenant data isolation (schema/db per tenant with RBAC and encryption) and integrated TensorFlow-based ML inference behind a Flask REST API using Redis caching, batching, and async execution; reports measurable wins like cutting endpoints from 6–8s to ~2s and increasing throughput 3–4x via Celery queues.”
Mid-level Data Engineer specializing in cloud data platforms, Spark, and streaming pipelines
“Data/MLOps engineer (Cognizant background) who owned an AWS/Airflow/Snowflake healthcare transactions pipeline processing ~8–10M records/day and cut pipeline/data-quality incidents by ~33%. Also built and deployed a production FastAPI model-inference service on Kubernetes (Docker, HPA) with strong observability (Prometheus/Grafana), versioned endpoints, and resilient backfill/idempotent external data ingestion patterns.”
Mid-level Data Engineer specializing in cloud data pipelines and streaming
“Data engineer with experience at Wells Fargo and Accenture owning end-to-end production pipelines processing hundreds of millions of transactional/risk records daily. Strong focus on data quality and reliability (reconciliation checks, schema drift detection, CloudWatch alerting) plus Spark performance tuning and idempotent backfills using Delta Lake/merge logic across AWS (S3/EMR/Databricks/Redshift) and Azure (ADF/Azure DevOps/Azure Monitor).”
Senior Full-Stack Java Developer specializing in microservices and cloud platforms
“Backend engineer focused on scalable Python/Flask services and high-performance PostgreSQL/SQLAlchemy systems, with demonstrated wins like reducing N+1-driven response times to under 200ms and cutting P95 latency below 1s via background queues and caching. Has production experience operationalizing ML models as Dockerized APIs on AWS (S3/Lambda) with monitoring (CloudWatch/ELK), plus robust multi-tenant isolation using JWT-driven tenant context and row-level security.”
Senior Security Engineer specializing in detection engineering, cloud security, and DFIR
“LLM workflow/agentic systems practitioner who has helped customers harden an LLM-based incident triage prototype into a trusted daily-use production system by adding observability, audits, confidence gating, and deterministic fallbacks. Brings an SRE-style approach to real-time debugging (trace replay, rollback/canary, safe toggles) and is experienced running developer-centric demos/workshops and partnering with sales on technical qualification and security/architecture artifacts.”
Mid Software Engineer specializing in machine learning and real-time data systems
“Hands-on implementation-focused candidate with experience owning cloud deployments and putting LLM/RAG workflows into production. They stand out for combining customer-facing deployment ownership with practical AI systems work, including retrieval tuning, hallucination mitigation, production incident response, and document-processing pipelines for messy real-world inputs.”
Mid-level Data Analyst specializing in healthcare and financial analytics
“Analytics-focused candidate with hands-on experience turning messy CRM, e-commerce, payments, and support data into trusted reporting datasets using SQL and Python. They have owned end-to-end churn and retention analytics work, including RFM-based segmentation, dashboard delivery, and metric standardization across sales, marketing, and finance.”
Mid-Level Data Engineer specializing in cloud data platforms and governed analytics
“Data engineer with Optum experience building end-to-end healthcare data pipelines for HL7/FHIR, processing millions of records daily across Kafka streaming and Databricks/Spark batch. Strong focus on data quality (schema enforcement/validations), reliability (Airflow monitoring/alerts), and analytics-ready serving in Snowflake powering Power BI/Tableau, with CI/CD via Git and Jenkins.”
Mid-level Software Engineer specializing in backend systems and real-time analytics
“Full-stack engineer at BigCommerce who combines customer-facing deployment ownership with hands-on AI/LLM systems work. Built and launched merchant analytics and predictive inventory workflows using React, TypeScript, FastAPI, Kafka, AWS, and RAG-style architectures, and has real production experience debugging non-deterministic AI issues caused by data pipeline freshness and event-ordering problems.”
Mid-level Full-Stack Java Developer specializing in cloud-native enterprise systems
“Backend/full-stack engineer with Blue Cross Blue Shield experience building a reactive, event-driven claims processing microservice platform on AWS (ECS, SNS/SQS) with Terraform-based IaC and strong observability (Dynatrace/CloudWatch). Demonstrated measurable production impact (32% less downtime, 24% higher processing efficiency) and deep database performance/migration expertise across MongoDB and Postgres.”
Mid-level Full-Stack Software Engineer specializing in FinTech and backend platforms
“Built an AI-native legal research platform that automated analysis across 100,000+ dense legal documents, combining LLM workflows, async backend architecture, and conversational retrieval in production. Also brings cross-domain experience in investment-analysis agents and healthcare claims/billing systems, with a strong emphasis on reliability, deterministic orchestration, and safe handling of messy operational data.”
Mid-Level Full-Stack Software Engineer specializing in healthcare, cloud, and data platforms
“Backend/platform engineer who owned a real-time customer analytics microservice stack in Python/FastAPI with Kafka streaming into PostgreSQL, including schema enforcement (Avro) and high-throughput optimizations. Strong Kubernetes + GitOps practitioner (EKS/GKE, Helm, Argo CD) who has handled CI/CD reliability issues with automated pre-deploy checks and rollbacks, and supported major migrations (on-prem to AWS; VM to EKS) with blue-green cutover planning.”
Mid-level Data & AI Engineer specializing in healthcare data pipelines and MLOps
“Built and deployed a production LLM-powered clinical note summarization system used by care managers to speed review of 5–20 page unstructured medical records. Implemented safety-focused validation (prompt constraints, rule-based and section-level checks, human-in-the-loop) to reduce hallucinations while maintaining low latency and meeting privacy/regulatory constraints, integrating via APIs into existing clinical tools.”
Mid-level Data Scientist/ML Engineer specializing in healthcare AI and MLOps
“Designed and deployed an enterprise LLM-powered clinical/pharmacy policy knowledge assistant at CVS Health, replacing manual searches across PDFs/Word/SharePoint with a HIPAA-compliant RAG system. Built end-to-end ingestion and orchestration (Airflow + Azure ML/Data Lake + vector index) with PHI masking, versioned re-embedding, and production monitoring (Prometheus/Grafana), and partnered closely with clinicians/compliance to ensure policy-grounded, auditable answers.”
Mid-level AI/ML Engineer specializing in MLOps and LLM-powered applications
“AI/ML engineer with production experience building a RAG-based internal analytics assistant (Databricks + ADF ingestion, Pinecone vector store, LangChain orchestration) deployed via Docker on AWS SageMaker with CI/CD and MLflow. Strong focus on real-world constraints—latency/cost optimization (LoRA ~60% compute reduction), hallucination control with citation grounding, and enterprise security/governance. Previously at Intuit, delivered an interpretable churn prediction system (PySpark/Databricks, Airflow/Azure ML) that improved retention targeting ~12%.”
Senior Data Engineer specializing in cloud-native data platforms for finance and healthcare
“Data engineer/backend data services practitioner with Bank of America experience building real-time and batch transaction-monitoring pipelines and APIs (Kafka + databases, REST/GraphQL). Highlights include a reported 45% response-time improvement through performance optimizations and use of Delta Lake schema evolution plus CI/CD (GitHub Actions/Jenkins) and operational reliability patterns like CloudWatch monitoring and dead-letter queues.”
Mid-level AI/ML Engineer specializing in LLM, RAG/GraphRAG, and fraud analytics
“LLM/agent engineer who has deployed a production internal assistant to reduce employee inquiry resolution time while maintaining regulatory compliance. Experienced with RAG, hallucination risk triage, and graph-based orchestration (LangGraph) for enterprise/banking-style workflows, emphasizing schema-validated, citation-backed, tool-constrained agent designs and tight collaboration with non-technical business/compliance stakeholders.”