Pre-screened and vetted.
Senior AI/ML Engineer specializing in production-grade LLM systems for regulated finance
“AI/LLM engineer with published work who built FinVet, a production financial misinformation detection system using multi-pipeline RAG, confidence-based voting, and evidence-backed outputs (F1 0.85, +37% vs baseline). Also built NexusForest-MCP, a Dockerized Model Context Protocol server exposing structured global deforestation/carbon data via SQL tools for reliable LLM tool use. Previously delivered borrower risk-rating (PD) models at BMO Financial Group that were validated and integrated into an enterprise credit system through close collaboration with credit officers and portfolio managers.”
Senior Integration Developer specializing in MuleSoft API-led connectivity
“Backend/integration-focused engineer in the Maryland area with production experience building FastAPI REST services secured with OAuth2.1/JWT and reliability patterns (timeouts, selective retries, idempotency, centralized error handling). Has delivered AWS-integrated MuleSoft/CloudHub solutions and supported AWS Glue ETL workflows, plus demonstrated strong SQL tuning with a 30–40s to 3–5s performance improvement.”
Mid-level Full-Stack Engineer specializing in cloud-native microservices and AI automation
“Software engineer/product owner who has led end-to-end delivery of AI and content-management platforms, including building RAG-based reliability improvements and migrating fragile systems to containerized AWS ECS/Kubernetes with Terraform-managed CI/CD. Experienced designing event-driven microservices (SQS/SNS/RabbitMQ), scaling queue consumers with autoscaling, and creating internal Python tooling to standardize data connectors (e.g., BigQuery/Airtable/internal APIs) to speed iteration.”
“Backend-focused intern who built and refactored the backend for an LLM-driven gifting mobile app using FastAPI, tackling high-latency LLM + product-API workflows. Implemented async worker-pool/queue processing with Redis caching plus retries/fallbacks, cutting end-to-end suggestion latency from ~4–5 seconds to ~1 second while improving reliability and rollout safety via staged migrations and testing.”
Senior Full-Stack Java Developer specializing in capital markets and trading systems
“Backend/data engineer with production experience in payment initiation/processing services built in Python/FastAPI, emphasizing reliability patterns (JWT/RBAC, timeouts, retries, circuit breakers). Has delivered AWS deployments on ECS (ALB, autoscaling, CI/CD to ECR) plus Lambda-based reporting, and built AWS Glue ETL pipelines with schema evolution and CloudWatch monitoring. Also modernized a legacy SAS reporting platform to Python/PostgreSQL with regression parity testing and parallel-run migration, and achieved a 70% SQL performance improvement.”
Mid-level Generative AI Engineer specializing in LLMs and RAG systems
“Built and shipped a production RAG-based enterprise knowledge assistant to replace slow/inaccurate search across millions of documents, using LangChain orchestration with GPT-4/LLaMA and vector databases. Strong focus on production constraints—latency, hallucination control, and cost—using hybrid retrieval, guardrails, LLM-as-judge validation, and model routing, and has experience translating non-technical stakeholder pain points into measurable outcomes.”
Mid-level Data Engineer specializing in Lakehouse, Streaming, and ML/LLM data systems
“Built and productionized an enterprise retrieval-augmented generation platform for internal knowledge over large unstructured corpora, emphasizing trust via strict citation/grounding and hybrid retrieval (BM25 + FAISS + cross-encoder re-ranking). Demonstrates strong scaling and cost/latency optimization through incremental indexing/embedding and index partitioning, plus disciplined evaluation/observability practices. Has experience operationalizing pipelines with Airflow/Databricks/GitHub Actions and partnering closely with risk & compliance stakeholders on auditability requirements.”
Senior Machine Learning Engineer specializing in LLMs, speech AI, and RAG systems
“AI engineer with production experience building multilingual speech-to-speech translation pipelines (ASR + LLM) for enterprise/media, focused on reliability at scale. Has hands-on orchestration experience (including IBM Watson contexts) and emphasizes production evaluation/monitoring using a mix of traditional metrics and LLM-based evaluators to catch quality regressions while balancing latency and cost.”
Junior Full-Stack Engineer specializing in TypeScript/React, Python, and AWS
“Full-stack engineer who built and owned an end-to-end real-time engineering dashboard for Medtronic robotic surgical systems, streaming high-frequency sensor/kinematic data via Python WebSockets to a React/TypeScript UI. Differentiates through performance/reliability practices (stable core vs experimental layer, observability, caching) and high-impact 3D visualization + session playback that became part of engineers' regular bench-testing workflows.”
Mid-level AI/ML Engineer specializing in generative AI, RAG platforms, and LLM agents
“AI/LLM engineer who has shipped 10+ production applications, including InvestIQ on GCP—a production-grade RAG due-diligence engine that ethically scrapes web/PDF sources, builds a ChromaDB knowledge base, and delivers analyst-style dashboards plus a citation-backed chat copilot. Deep focus on reliability (evidence-only answers, hard citations, refusal gating), retrieval tuning, and orchestration (Airflow/Cloud Composer), plus multi-agent systems (CrewAI with 7 specialized finance agents).”
Junior Full-Stack & ML Engineer specializing in research tooling and applied machine learning
“Full-stack engineer and ML assistant in UC Irvine’s CS department who deployed a lab project showcase platform and integrated on-demand execution of computational projects using Docker for isolation. Also built and optimized Linux cloud/cluster test automation for research, diagnosing RAM and network sync bottlenecks, and later led development of a Python-based predictive analytics tool for musicians using probabilistic graphical models and flexible data pipelines.”
Mid-level AI/ML Engineer specializing in MLOps and LLM applications
“BNY Mellon engineer who has built and operated production AI systems end-to-end: a LangChain/Pinecone RAG platform scaled via FastAPI + Kubernetes to 1000 RPM with 99.9% uptime, supported by monitoring and data-drift detection. Also deep in data/infra orchestration (Airflow, Dagster, Terraform on AWS/EMR/EC2), processing 500GB+ daily and delivering measurable reliability and performance gains, plus strong compliance-facing model explainability using SHAP and Tableau.”
Junior Hardware/Product Engineer specializing in PCB design, NPI, and FPGA validation
“Backend/platform engineer who owned a Python-based smart finance assistant backend, building async FastAPI microservices with PostgreSQL/Redis and deploying to AWS EKS via Docker/Helm and CI/CD (GitHub Actions, Jenkins). Strong in production reliability and migrations—implemented observability (Prometheus/Grafana), security (JWT RBAC), and executed a low-downtime monolith-to-microservices migration plus Kafka-based event streaming with ordering/retry/idempotency patterns.”
Intern Full-Stack Software Engineer specializing in web apps, distributed systems, and AI tooling
“Software engineer with experience spanning high-scale backend systems and distributed consensus: led a 6-person team delivering a production data querying/visualization platform with major latency improvements via cursor-based pagination and streamed results. Built a RAFT-based distributed logging tool resilient to partitions and storage constraints, and at Nasuni developed FastAPI services processing multi-terabyte workloads for 500+ enterprise customers with secure API key management.”
Mid-level Full-Stack Software Engineer specializing in AI platforms and microservices
“Backend engineer currently building an AWS Lambda/FastAPI inventory recommendation system using a LangChain + GPT-4 RAG pipeline and MongoDB vector search; drove major cost optimization via Redis caching (60% reduction) while sustaining 10k+ daily requests under 2s latency. Previously deployed Node.js microservices on AWS OpenShift with Jenkins/Helm at UnitedHealth Group and led a zero-downtime monolith-to-microservices migration at Verizon, including RabbitMQ-based real-time messaging with DLQs and idempotency.”
Mid-level Full-Stack Software Engineer specializing in cloud-native microservices
“JavaScript/Node.js engineer who contributes to open-source utilities focused on API integrations and JSON validation, including a 30–35% throughput improvement by profiling and optimizing deep-clone-heavy code paths. Strong in performance tooling (Node performance hooks, Chrome DevTools flame graphs), incremental/test-driven changes, and community-facing issue triage plus developer-friendly documentation.”
Mid-Level Full-Stack Software Engineer specializing in FinTech and cloud-native microservices
“Open-source React dashboard/visualization library maintainer focused on runtime performance and API clarity. Led a significant effort to eliminate severe render lag on large live-updating datasets through profiling-driven refactors (normalized state, memoized selectors) and locked improvements in with CI, linting, and documentation that reduced regressions and improved external contributor onboarding.”
Mid-level Software Engineer specializing in cloud, data engineering, and AI/ML
“Backend/platform engineer who owned an AI-powered resume optimization service end-to-end (FastAPI + Celery + Redis/Postgres) and optimized it for unpredictable LLM task latency. Strong Kubernetes/GitOps practitioner (Helm, autoscaling, probes, ArgoCD rollbacks) with experience in on-prem-to-cloud migrations using Terraform and CDC-based replication, plus real-time Kafka pipelines monitored via Prometheus/Grafana.”
Senior AI Engineer specializing in Agentic AI and distributed systems
“LLM/agentic workflow engineer with healthcare domain experience who built a HIPAA-compliant multi-agent RAG system for clinical review automation at UnitedHealth Group, achieving 92% precision and cutting latency 40% through async orchestration and Redis semantic caching. Also has strong data engineering orchestration background (Airflow on AWS EMR with Great Expectations) and a proven clinician-in-the-loop feedback process that improved model faithfulness by 18%.”
Mid-level AI Engineer specializing in GenAI, LLM integration, and RAG pipelines
“Built and led deployment of an autonomous, self-correcting multi-agent knowledge retrieval and validation system at HCA Healthcare to reduce heavy manual research/validation in clinical/compliance documentation. Deeply focused on production reliability and cost—used LangGraph StateGraph orchestration plus ONNX/CUDA/quantization to cut GPU costs by 25%, and partnered with the Compliance VP using real-time contradiction-rate dashboards to hit a 40% automation goal without compromising compliance.”
AI & Full-Stack Software Engineer specializing in LLM-powered applications
“Full-stack engineer focused on productionizing LLM applications, including an Android privacy-policy risk summarization app (Kotlin/React Native + FastAPI + Ollama) that cut response times from ~10s to ~5–6s via batching, caching, async, and event-driven architecture. Currently at PRGX building an LLM-based legal contract clause extraction system, partnering closely with legal/procurement SMEs to create schemas, labeled datasets, and evaluation pipelines that improved accuracy from 70% to 85%. Also has experience architecting real-time voice/LLM systems with streaming microservices (Kafka, Kubernetes, gRPC/WebSockets) and an avatar chatbot pipeline (TalkingHead, Google TTS, AnythingLLM).”
Mid-level Data Scientist specializing in Generative AI, NLP, and MLOps
“Built and deployed an LLM-powered claims-document summarization system (insurance domain) that cut agent review time from 4–5 minutes to under 2 minutes and saved 1,200+ hours per quarter. Hands-on across orchestration and production infrastructure (Airflow retraining DAGs, Kubernetes, SageMaker endpoints, FastAPI) and recent RAG workflows using n8n + Pinecone, with a strong focus on reliability, cost, and explainability for non-technical stakeholders.”
Mid-level AI/ML Engineer specializing in NLP, RAG systems, and real-time risk modeling
“AI/ML Engineer with 4+ years of experience (Capital One, Odin Technologies) and a master’s in Data Analytics (4.0 GPA) who has deployed LLM/RAG systems to production for compliance/risk and document review. Strong in orchestration and MLOps (Airflow, Kubernetes, MLflow, GitHub Actions) and in tackling real-world LLM constraints like latency, context limits, and data privacy, with measurable impact (20%+ manual review reduction; 33% faster release cycles).”
Mid-level AI/ML Engineer specializing in deep learning, MLOps, and LLM applications
“Built and deployed production LLM assistants for internal Q&A and customer-feedback summarization, emphasizing reliability (RAG, prompt tuning, validation/whitelisting) and privacy safeguards. Improved adoption by adding explainable outputs and a user feedback mechanism, and has hands-on orchestration experience with Aflow and Azure Logic Apps.”