Pre-screened and vetted.
Mid-Level Full-Stack & Cloud Engineer specializing in scalable distributed systems
Mid-Level Software Engineer specializing in microservices and cloud data pipelines
“Full-stack engineer with end-to-end ownership across React/TypeScript frontends, Spring Boot/Node microservices, and production ops on Docker/Kubernetes and AWS (ECS/CloudWatch). Built real-time healthcare eligibility and analytics systems at Cigna and an early-stage seller onboarding platform at Flipkart, driving measurable performance gains (35–40% latency/throughput improvements) through event-driven Kafka pipelines, Redis caching, and strong reliability/observability practices.”
“Built and deployed a production Retrieval-Augmented Generation (RAG) platform in a healthcare setting to automate clinical documentation review and summarization, targeting near-real-time, explainable outputs. Emphasizes grounded generation to reduce hallucinations, latency optimizations (chunking/embedding reuse), and PHI-safe workflows with access controls, plus strong orchestration experience using Apache Airflow.”
Mid-level Data Scientist / AI-ML Engineer specializing in Generative AI and LLM applications
“Built a production GenAI-powered analytics assistant to reduce reliance on data analysts by enabling natural-language Q&A over Databricks/Power BI dashboards, backed by vector search (Pinecone/Milvus) and a Neo4j knowledge graph, including multimodal support via OpenAI Vision. Demonstrates strong real-world LLM reliability engineering with strict RAG, LangGraph multi-step verification, and Guardrails/custom validators, plus broad orchestration and production monitoring experience (Airflow, ADF, Step Functions, Kubernetes, Prometheus/CloudWatch).”
Mid-level Python Developer specializing in AWS microservices and cloud automation
“Backend engineer focused on Python/FastAPI microservices running on Kubernetes (AWS EKS) with strong GitOps/CI/CD ownership (GitHub Actions + ArgoCD). Demonstrated measurable performance wins (p95 latency cut from >1s to <200ms) and production reliability work across Kafka/Redis streaming and cloud-to-on-prem migrations (RDS/S3 to Postgres/MinIO) using parallel validation and checksum-based consistency checks.”
Senior Python Full-Stack Developer specializing in cloud-native microservices and data platforms
“Backend/data engineer from Oliver Wyman who built and ran production Python (FastAPI) services on AWS (ECS/Lambda/API Gateway) supporting risk modeling and regulatory reporting. Strong in reliability/observability, Glue-based ETL with data quality controls, and legacy SAS-to-Python modernization with rigorous parity validation; also demonstrated measurable SQL performance wins and cost-control improvements in serverless scaling. Based in Raleigh, NC and can travel onsite for important Bethesda-area meetings.”
Senior Python Full-Stack Developer specializing in cloud, data engineering, and ML/GenAI
“Backend/data engineer with hands-on production experience building FastAPI services on AWS and implementing strong reliability/observability (CloudWatch, ELK, correlation IDs, alarms). Has delivered serverless + container solutions with IaC (CloudFormation/Terraform) and Jenkins CI/CD, and built AWS Glue/PySpark pipelines into S3/Redshift with schema-evolution and data-quality safeguards; demonstrated large-scale SQL tuning (45 min to 3 min on a 500M-row workload).”
Principal Cloud & Infrastructure Engineer specializing in reliability and regulated data platforms
“Founder/CTO-type startup leader who has built cloud-native data and AI platforms from scratch while owning both technical vision and product direction. Brings rare end-to-end startup experience spanning zero-to-one building, growth-stage execution, and fundraising from early stage through exit, with a strong ability to translate technical complexity into clear investor narratives.”
Mid-level Data Scientist / Machine Learning Engineer specializing in fraud, risk, and MLOps
“AI/ML practitioner with Northern Trust experience who has shipped production LLM systems (internal support assistant) using RAG, vector databases, orchestration (LangChain/custom pipelines), and rigorous monitoring/feedback loops. Also built AI-driven fraud detection/risk monitoring solutions in a regulated financial environment, emphasizing explainability (SHAP), audit readiness, and stakeholder trust through dashboards and clear communication.”
Mid-level AI/ML Engineer specializing in fraud detection and risk analytics in Financial Services
“At JP Morgan Chase, built and deployed a production LLM-powered RAG knowledge assistant to help fraud investigators and risk analysts quickly navigate regulatory updates and internal policies, reducing investigation delays and compliance risk. Strong focus on secure retrieval (RBAC filtering), reliability (layered testing + observability), and production constraints (latency/SLOs), with Airflow-orchestrated, auditable ML pipelines.”
Mid-level Python & AI/ML Engineer specializing in backend APIs and MLOps
“Built and deployed a production LLM/RAG document automation system for business documents (contracts/claim forms) that extracts schema-validated JSON, generates grounded summaries/Q&A, and integrates into transaction systems via APIs. Emphasizes real-world reliability: hallucination controls, layout-aware parsing with OCR fallback, Step Functions-orchestrated workflows with retries/timeouts, and human-in-the-loop review designed in close partnership with operations and claims stakeholders.”
Mid-level AI/ML Engineer specializing in NLP, MLOps, and scalable data pipelines
“Built and shipped a production LLM-powered personalized client engagement assistant in the financial domain, balancing real-time recommendations with strict privacy/compliance requirements. Demonstrates strong MLOps/LLMOps depth (Airflow + MLflow, containerized microservices, drift monitoring) and a privacy-by-design approach validated in collaboration with risk and compliance teams.”
Mid-level Software Engineer specializing in cloud-native microservices and data platforms
“Backend engineer with experience at Comcast and in healthcare/pharmacy automation (PrimeRx), building Python/Flask services that orchestrate large-scale batch workflows (Airflow) and high-throughput event processing (Kafka). Demonstrated measurable performance wins (cut provisioning latency to ~150–200ms) and strong multi-tenant isolation strategies (Postgres RLS, partitioning), plus practical integration of ML model outputs into production systems with validation and fallback controls.”
Senior Full-Stack Software Engineer specializing in Python and AWS
“Backend/data engineer who has built production Python microservices (FastAPI) and AWS-native platforms for event ingestion and analytics, combining ECS/Fargate + Lambda with CloudFormation-driven environments and strong secrets/IAM practices. Experienced modernizing legacy logic with parallel-run parity validation and safe phased cutovers, and has demonstrated measurable SQL tuning wins (20–30s down to 1–2s) plus incident ownership in Glue/Step Functions ETL pipelines.”
Mid-level Data Engineer specializing in streaming and cloud data platforms for financial services
“Data engineering-focused candidate (internship/project experience) who built end-to-end pipelines processing a few million transactional records/day for fraud detection and reporting, using Airflow, Python/SQL, and PySpark with strong emphasis on data quality gates, idempotency, and monitoring. Also implemented an external web/API data collection system with anti-bot tactics and schema-change quarantine, and shipped a versioned Flask API to serve curated warehouse data.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Mid-level Data Engineer specializing in cloud data warehousing and analytics
“Data engineer at American Express who owned end-to-end pipelines for transaction and customer data used in finance reporting and risk analytics, processing ~5–8M records/day. Built Airflow-orchestrated ingestion (including external APIs/web sources) with strong data quality controls, monitoring/alerts, and resilient backfill/retry patterns, and also shipped a versioned REST API serving aggregated metrics to analytics teams.”
Mid-level Machine Learning Engineer specializing in AI/LLM systems
“ML/LLM systems engineer who has owned AI support automation products end-to-end, including ServiceNow-integrated incident routing, RAG-based resolution suggestion systems, and production stabilization. Stands out for combining hands-on platform work across PySpark, AWS Glue, FastAPI, Kubernetes, and Pinecone with measurable operational impact, including 30-35% MTTR reduction and 25-30% improvement in first-touch resolution.”
Mid-level AI/ML Engineer specializing in Generative AI and financial services
“ML/AI engineer with hands-on experience shipping regulated financial AI systems at JPMC and Capgemini, spanning credit risk, fraud detection, and generative AI assistants. Stands out for combining modern LLM/RAG architectures with strong MLOps, real-time infrastructure, and explainability/compliance practices, while delivering measurable business impact in latency, accuracy, cost, and risk reduction.”
Senior Full-Stack/Backend Software Engineer specializing in cloud-native automation and microservices
“Backend/data engineer with strong AWS production experience across containers (ECS) and serverless (API Gateway/Lambda/SQS), plus Glue-based ETL to Parquet for Athena/Redshift. Demonstrates hands-on reliability and security depth (Cognito OAuth2/JWT with JWKS rotation, idempotency/DLQs, monitoring) and measurable performance wins (Redis caching + query tuning), along with legacy-to-services modernization using parallel-run parity and feature-flagged cutovers.”
Mid-level AI/ML Engineer specializing in MLOps, NLP/LLMs, and computer vision
“Built and shipped a production LLM/RAG risk-case summarization and triage system used by fraud/compliance analysts, with strong grounding controls (evidence-cited outputs and refusal on low confidence). Demonstrates end-to-end ownership across retrieval quality, Airflow-orchestrated indexing pipelines, and compliance-grade privacy (PII redaction, RBAC, encrypted redacted logging, and auditable prompt/model versioning) plus a tight feedback loop with non-technical domain experts.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Senior Data Scientist / ML Engineer specializing in cloud ML pipelines and GenAI
“ML/NLP practitioner with experience building a transformer-failure prediction system that combines sensor signals with unstructured maintenance comments using LLM-based extraction and similarity validation. Strong emphasis on production readiness—data leakage controls, SQL-driven data quality tiers, and rigorous bias/fairness validation (including contract/spec evaluation across diverse company profiles).”