Pre-screened and vetted.
Mid-level Python Developer specializing in AWS microservices and cloud automation
“Backend engineer focused on Python/FastAPI microservices running on Kubernetes (AWS EKS) with strong GitOps/CI/CD ownership (GitHub Actions + ArgoCD). Demonstrated measurable performance wins (p95 latency cut from >1s to <200ms) and production reliability work across Kafka/Redis streaming and cloud-to-on-prem migrations (RDS/S3 to Postgres/MinIO) using parallel validation and checksum-based consistency checks.”
Junior Software Engineer specializing in full-stack, cloud serverless, and AI systems
“SDE who worked on an MGICS Lab robotics project building a multi-agent model to help agents understand tasks and generate robot instructions, emphasizing task-splitting, checking, and a reflection agent to improve accuracy. Also has experience using GitHub with automated CI/CD pipelines.”
Mid-Level Software Engineer specializing in full-stack web and cloud systems
“Full-stack engineer with strong data engineering and privacy-domain experience, having owned an automated Data Subject Rights (DSR) processing pipeline end-to-end across Azure SQL and GCP (GCS/BigQuery). Emphasizes production reliability via idempotency, validation checkpoints, structured logging/monitoring, and safe CI/CD-driven deployments, and has also built React+TypeScript + Node/Postgres web apps with scalable, maintainable architecture.”
Senior Python Full-Stack Developer specializing in cloud, data engineering, and ML/GenAI
“Backend/data engineer with hands-on production experience building FastAPI services on AWS and implementing strong reliability/observability (CloudWatch, ELK, correlation IDs, alarms). Has delivered serverless + container solutions with IaC (CloudFormation/Terraform) and Jenkins CI/CD, and built AWS Glue/PySpark pipelines into S3/Redshift with schema-evolution and data-quality safeguards; demonstrated large-scale SQL tuning (45 min to 3 min on a 500M-row workload).”
Intern Full-Stack/ML Engineer specializing in LLM applications and mobile development
“Backend engineer who built a serverless AWS Lambda microservices backend for a parenting assistance mobile app, including a personalized recommendation system optimized to sub-500ms via precomputed scoring and DynamoDB caching. Demonstrates strong production pragmatism: CloudWatch-driven performance tuning (provisioned concurrency), zero-downtime phased schema migrations, and robustness patterns like optimistic locking and request deduplication. Also led a refactor of an LLM RAG pipeline to improve retrieval quality and cut latency from ~5s to ~3s.”
Senior Python Full-Stack Developer specializing in cloud-native microservices and data platforms
“Backend/data engineer from Oliver Wyman who built and ran production Python (FastAPI) services on AWS (ECS/Lambda/API Gateway) supporting risk modeling and regulatory reporting. Strong in reliability/observability, Glue-based ETL with data quality controls, and legacy SAS-to-Python modernization with rigorous parity validation; also demonstrated measurable SQL performance wins and cost-control improvements in serverless scaling. Based in Raleigh, NC and can travel onsite for important Bethesda-area meetings.”
Mid-level AI/ML Engineer specializing in LLM agents, RAG, and enterprise ML systems
“Built a production multi-agent recommendation/RAG system for internal data analysts to speed up weekly report creation by improving document discovery and automating report/SQL generation. Implemented LangGraph-based orchestration with deterministic agent routing, robust error handling (interrupt/resume), and metadata-driven semantic chunking for diverse PDF/document formats, plus monitoring for latency, throughput, and token/cost efficiency.”
Mid-Level AI/ML Software Engineer specializing in agentic LLM systems
“Built and deployed a production LLM-powered multi-agent compliance copilot (life sciences/finance) using LangChain/LangGraph + RAG over vector databases, delivered via async FastAPI on Kubernetes. Emphasizes audit-ready, deterministic outputs with schema constraints and citations, plus rigorous evaluation/monitoring; reports 60%+ reduction in manual research time and successful production adoption.”
Mid-level Full-Stack Developer specializing in AI-powered cloud-native applications
“Full-stack engineer who has owned customer-facing AI recommendation and analytics dashboards end-to-end (backend APIs/data processing through React UI, deployment, and monitoring). Demonstrates strong systems thinking around scaling microservices—using observability, caching, async workflows, and resilience patterns—and also built an internal ops dashboard that became the default tool for on-call incident reviews.”
Mid-level AI Engineer specializing in GenAI, NLP, and MLOps
“LLM/agentic-systems engineer with PayPal experience hardening an LLM-powered fraud support assistant from prototype to production, focusing on low-latency distributed architecture, rigorous evaluation/testing, and security/compliance. Comfortable in customer-facing and GTM contexts—runs technical demos/workshops, builds tailored pilots, and aligns sales/CS with engineering to close deals and drive adoption.”
Mid-Level Full-Stack Software Engineer specializing in API-first microservices and cloud platforms
“Backend-focused engineer who built a resume processing and job application platform using Python/MongoDB/Streamlit, including OpenAI-powered skill/keyword extraction and recruiter-facing search/filtering. Has hands-on cloud deployment experience on AWS/Azure and executed an on-prem reservation portal migration to Azure using a phased trial-and-cutover approach; also automated CI/CD with Jenkins and GitHub Actions.”
“LLM/agent workflow engineer with healthcare experience (CVS/CBS Health) who built and deployed a production call-insights platform using Azure OpenAI + LangChain/LangGraph, including sentiment and compliance checks. Demonstrates deep HIPAA/PHI handling (tenant-contained processing, redaction, RBAC/encryption/audit logging) and production rigor (testing, eval sets, validation/retries, autoscaling) to scale to thousands of transcripts.”
“Built and deployed a production Retrieval-Augmented Generation (RAG) platform in a healthcare setting to automate clinical documentation review and summarization, targeting near-real-time, explainable outputs. Emphasizes grounded generation to reduce hallucinations, latency optimizations (chunking/embedding reuse), and PHI-safe workflows with access controls, plus strong orchestration experience using Apache Airflow.”
Mid-level Machine Learning Engineer specializing in NLP, LLMs, and multimodal modeling
“Built and productionized a telecom-focused RAG assistant by LoRA fine-tuning LLaMA-2 and integrating LangChain+FAISS behind a FastAPI service, with dashboards and a human feedback UI for engineers. Demonstrated measurable impact (≈40% faster document lookup, +8–10% retrieval precision) and strong MLOps rigor via Airflow orchestration, CI/CD, and monitoring for drift and failures.”
Mid-level AI/ML Engineer specializing in Generative AI, NLP, and Computer Vision
“Built an LLM-powered learning assistant (EduQuizPro/EduCrest Pro) that uses RAG over URLs and PDFs to generate quizzes, notes, and explanations for students/professors. Emphasizes production robustness—implemented dependency fallbacks (FAISS/Sentence Transformers/Gradio), CLI-safe mode, and NumPy-based indexing—along with a custom orchestration layer to keep multi-step AI workflows reliable.”
Junior Full-Stack Software Engineer specializing in cloud-native microservices
“Backend engineer with hands-on IoT and AI product work: built a decoupled Raspberry Pi + AWS IoT Core weather monitoring backend and a Dockerized FastAPI LLM service on AWS ECS using OpenAI/HuggingFace with an emerging RAG layer. Also delivered measurable performance gains at DAZN by redesigning event-driven/serverless ingestion (SNS, S3->Lambda->DynamoDB), cutting latency ~30% and boosting throughput ~25% while automating ~90% of manual sync work.”
Junior AI/ML Systems Engineer specializing in LLM infrastructure and distributed training
“Built and shipped a production NMT system translating medical documentation for a rare/low-resource language, tackling data scarcity with retrieval-driven pattern matching plus dictionary/grammar- and LLM-based augmentation and validating quality with a linguistic expert. Also develops agentic LLM workflows with LangChain/LangGraph (including a deep-research style system) and has experience aligning medical AI deployments with clinician-defined risk metrics and human-in-the-loop decision making.”
Mid-level Data Scientist / AI-ML Engineer specializing in Generative AI and LLM applications
“Built a production GenAI-powered analytics assistant to reduce reliance on data analysts by enabling natural-language Q&A over Databricks/Power BI dashboards, backed by vector search (Pinecone/Milvus) and a Neo4j knowledge graph, including multimodal support via OpenAI Vision. Demonstrates strong real-world LLM reliability engineering with strict RAG, LangGraph multi-step verification, and Guardrails/custom validators, plus broad orchestration and production monitoring experience (Airflow, ADF, Step Functions, Kubernetes, Prometheus/CloudWatch).”
Director-level AI & Data Science leader specializing in GenAI, LLMs, and MLOps
“ML/NLP engineer currently working in NYC on a system that connects complex unstructured data sources to deliver personalized insights, using embeddings + vector DB retrieval and a RAG architecture (LangChain, Pinecone/OpenSearch). Strong focus on production constraints—especially low-latency retrieval—using FAISS/ANN, PCA, index partitioning, and Redis caching, plus PEFT fine-tuning (LoRA/QLoRA) and KPI/SLA-driven promotion to production.”
Staff Full-Stack Engineer specializing in AI platforms and infrastructure automation
“Backend/full-stack engineer building complex internal platforms and customer-facing demos at the intersection of infrastructure and product. Shipped a no-code Product Lifecycle Manager for manufacturing (3 manufacturers, 1000+ evolving tests) using AWS S3/SQS ingestion and extensible Postgres (EAV+JSONB) with end-to-end traceability. Also built a FastAPI-based company data intelligence platform with Okta-secured RBAC and an LLM/MCP layer for ChatGPT-like analytics over enterprise data sources.”
Mid-level AI/ML Engineer specializing in deep learning, NLP/LLMs, and MLOps
“Built and shipped a real-time oncology risk prediction system used by doctors during patient visits, trained on clinical data in AWS SageMaker and deployed via FastAPI with sub-second responses. Emphasizes clinician-trust features (SHAP explainability, validation checks) and HIPAA-compliant controls (encryption, RBAC, audit logging), plus Kubernetes-based production operations with autoscaling, monitoring, and drift/retraining workflows; collaborated closely with oncologists at Flatiron Health.”
Mid-level AI/ML Engineer specializing in Generative AI, RAG, and Conversational AI
“Built a production RAG-based GenAI copilot backend at Aetna using Python/FastAPI, GPT-4, LangChain, and Azure AI Search, deployed on AKS with Prometheus/Grafana observability. Owned the system end-to-end (ingestion through deployment) and improved peak-time reliability by addressing vector search and embedding bottlenecks with Redis caching, index optimization, and async processing, plus added anti-hallucination guardrails via retrieval confidence thresholds.”
Mid-Level Software Engineer specializing in Cloud Platform & Automation
“Software engineer at Wrap who built production AWS Lambda services for large-scale Parquet dataset generation (50k+ records) and a synthetic traffic/lead generation system using Python, Playwright, and Jenkins. Also built and deployed a full-stack hobby product (MyAnimeListRanker) that ingests MyAnimeList user data and uses an Elo-based ranking workflow, with operational guardrails like rate limiting and monitoring via Vercel/logs.”
Senior Software Engineer specializing in identity, cloud-native microservices, and reactive web apps
“Product-focused full-stack engineer with Walmart and Dell experience who built and shipped a real-time engagement dashboard end-to-end (Kafka Streams, Spring Boot, React/TypeScript/D3) used daily by business teams, moving them from next-day reports to real-time decisioning. Strong in performance/reliability (Redis caching cut latency ~40%, 90%+ test coverage, Prometheus/CloudWatch monitoring) and production operations on AWS/EKS including handling a cascading failure from a memory leak with zero-downtime rollback and redeploy.”