Pre-screened and vetted.
Intern AI/ML Engineer specializing in LLMs, MLOps, and distributed training
“Founding AI engineer (June 2024) at Talon Labs who built and productionized an LLM-powered chatbot for interacting with proprietary supply-chain documents, deployed at large scale (25–100,000 users). Experienced with RAG/LLM orchestration (LangChain, LlamaIndex, Groq AI) and production ops tooling (Kubernetes, Docker, Kubeflow, Airflow), with a metrics-driven approach to evaluation, observability, and stakeholder alignment.”
Senior Software Engineer specializing in Generative AI product development
“AI product builder at Padlet who shipped multiple production LLM features for education workflows, including an AI document generator (AI Recipes) and a RAG-enabled in-product chat assistant. Built an AI microservice layer (LangChain) to swap model providers easily and created automated + human-in-the-loop evaluation systems (including ~100-test runs) to iterate on prompts and quality.”
Mid-level Full-Stack Software Engineer specializing in cloud and AI-enabled applications
“Product-focused full-stack engineer (70/30 app vs infra) with Accenture experience and recent AI workflow work, shipping end-to-end systems from React/TypeScript UIs through FastAPI backends to Postgres. Built an AI-driven data extraction platform with async job APIs, strict schema validation, and strong observability, and has operated AWS ECS-based deployments with real incident mitigation (DB connection exhaustion/latency under traffic spikes).”
Mid-level Full-Stack Developer specializing in FinTech and Healthcare systems
“Open-source contributor who improved React Query’s caching/subscription behavior to reduce unnecessary re-renders via debouncing and batched updates, validated with benchmarking and extensive tests. Also maintained a Flask extension and resolved production background-task hangs by tracing Redis connection handling issues, adding cleanup/retry logic and troubleshooting docs. In a fast-paced startup, owned the design of a Celery+Redis multi-queue background processing system with Prometheus-based observability.”
Executive Technology Leader (CTO) specializing in AI, cloud, and distributed platforms
“Engineering leader who stays hands-on in high-leverage technical areas (architecture, scalability, reliability) while operating at an executive level. Led StagePilot’s shift from a tightly coupled legacy system to a cloud-native, event-driven real-time platform proven at 1M+ concurrent users, and previously scaled multiple SRE teams at McGraw-Hill with SLOs, on-call, and blameless ops practices.”
Entry-Level Frontend Software Developer specializing in React and ML-enabled web apps
“Backend-focused Python/Flask engineer who owned REST APIs for a video analysis system, including preprocessing, ML inference integration, and post-processing into time-aligned predictions consumable by a React UI. Demonstrated practical performance/scalability work by decoupling API request handling from CPU-heavy processing and adding timing instrumentation to identify and optimize bottlenecks.”
Senior Full-Stack Java Engineer specializing in cloud-native microservices and FinTech
“Backend engineer who owned a Python task management API with JWT auth, async notifications, and performance work (DB optimization/caching) to handle high volumes. Led an on-prem to Azure private cloud migration at Morgan Stanley using GitOps and IaC (Terraform/ARM) with phased rollout and rollback planning. Also built a Kafka real-time streaming pipeline with exactly-once/idempotent consumers and Prometheus/Grafana monitoring.”
Mid-level Full-Stack Software Engineer specializing in Java/Spring microservices and AWS
“Backend/platform engineer who has owned a real-time business analytics dashboard backend (Python/Flask/MongoDB) and built Kafka event-streaming pipelines with idempotent processing and DLQs. Strong DevOps/GitOps experience deploying containerized microservices to AWS EKS with CI/CD (Jenkins/GitHub Actions/CodePipeline) and ArgoCD auto-sync/drift detection, plus hands-on support for phased hybrid cloud/on-prem migrations using feature flags and replication.”
Senior Full-Stack Java Developer specializing in microservices and cloud platforms
“Backend engineer focused on scalable Python/Flask services and high-performance PostgreSQL/SQLAlchemy systems, with demonstrated wins like reducing N+1-driven response times to under 200ms and cutting P95 latency below 1s via background queues and caching. Has production experience operationalizing ML models as Dockerized APIs on AWS (S3/Lambda) with monitoring (CloudWatch/ELK), plus robust multi-tenant isolation using JWT-driven tenant context and row-level security.”
Mid-level Full-Stack Engineer specializing in cloud-native microservices and DevOps
“Backend engineer with strong Python/FastAPI microservices ownership, including an ML-serving service with embeddings, async DB access, and Redis caching to reduce latency under high load. Experienced deploying and operating containerized services on Kubernetes using GitOps (Argo CD/Helm) with automated CI/CD, plus hands-on Kafka streaming pipeline tuning and enterprise migration work (Infosys) using blue-green/active-passive strategies.”
Mid-level Full-Stack Java Developer specializing in FinTech microservices
“Backend-focused Python/Flask engineer with strong performance and scalability experience across PostgreSQL/SQLAlchemy optimization, caching, and async processing. Has implemented multi-tenant data isolation (schema/db per tenant with RBAC and encryption) and integrated TensorFlow-based ML inference behind a Flask REST API using Redis caching, batching, and async execution; reports measurable wins like cutting endpoints from 6–8s to ~2s and increasing throughput 3–4x via Celery queues.”
Mid-level Full-Stack/Backend Engineer specializing in Java microservices and cloud platforms
“PayPal ML/AI practitioner who built and productionized a hybrid recommendation engine (BERT/LLM embeddings + collaborative filtering + XGBoost ranking) on AWS with end-to-end MLOps and orchestration. Addressed real-world issues like cold start and embedding latency (ONNX, clustering, caching, PySpark/Delta Lake) and drove a 27% lift in upsell conversion via A/B testing and stakeholder collaboration with marketing.”
Junior Backend Software Engineer specializing in microservices and API platforms
“Backend engineer with strong performance and security instincts: built a Flask API for readability metrics with clean, testable modular design; optimized SQLAlchemy/Postgres to eliminate N+1 issues (800ms to 120ms). Also implemented an LLM-powered natural-language travel search using Claude Sonnet + Amadeus with RAG and anti-exploitation safeguards, plus multi-tenant isolation via Postgres RLS and Redis caching that cut search latency from ~20s to ~4–5s while reducing storage costs.”
Junior Software Engineer specializing in full-stack web development and computer vision
“Backend engineer who built an AI voice-driven calendar assistant using Python/Flask with a multi-agent architecture (CrewAI + OpenAI), including agent-specific memory management and summarization to handle token limits. Also has production performance optimization experience at Trendyol, adding Redis caching to speed up an advertisement retrieval endpoint and improve page load experience.”
Junior Full-Stack/AI Engineer specializing in enterprise AI agents and web platforms
“Forward Deployed Engineer focused on taking enterprise LLM voice agents from prototype to production. Led a turnaround on a high churn-risk account by building a custom nested-API integration and preprocessing layer that enabled the LLM to reason over complex order hierarchies, cutting call handle time from 15 minutes to 2 minutes and driving expansions. Strong in real-time agent/workflow debugging, developer workshops, and sales partnership for adoption.”
Mid-Level Software Engineer specializing in AI-enabled backend and full-stack web systems
“Backend/AI workflow engineer with experience at AirKitchenz, Uber, and Vivma Software, building production systems on AWS (Lambda, DynamoDB, Step Functions). Has a track record of major performance wins (DynamoDB latency 2s to <150ms; Postgres query 2s to ~180ms) and shipping LLM-powered onboarding and ticket-routing workflows with strong guardrails (schema validation, confidence thresholds, human-in-the-loop escalation).”
Junior Full-Stack Software Engineer specializing in Django, AWS, and AI/ML
“Full-stack engineer who built and owned an AI-powered personal statement editor in Next.js (App Router + TypeScript), including dynamic routing, server-side data fetching, and typed API route handlers. Post-launch, they handled production monitoring/debugging and shipped reliability/performance upgrades (rate limiting, retries, rollback, DB indexing), and report a 40% latency reduction using Suspense/streaming and React concurrency patterns. Also implemented a durable Temporal-orchestrated AI document workflow with robust retry/idempotency strategies.”
Mid-level Full-Stack Java Engineer specializing in microservices, cloud, and event-driven systems
“Software engineer at Procter & Gamble focused on warehouse/operations systems, building near-real-time order/inventory visibility using Java/Spring Boot, React, Kafka, PostgreSQL, and Redis with measurable latency and load-time gains. Also shipped internal LLM/RAG knowledge assistants grounded in company runbooks and workflows, implementing guardrails and an evaluation loop that drove concrete retrieval improvements (document chunking) and regression prevention.”
Junior Full-Stack & AI/ML Engineer specializing in LLMs and multimodal document processing
“Built a production RAG-based NBA player scouting assistant that embeds player profiles into FAISS, orchestrates retrieval and LLM recommendations with LangChain, and surfaces results via embedded Tableau dashboards. Demonstrates strong focus on evaluation/monitoring (batch tests, LLM-as-judge, latency/failure/token metrics) and has experience translating non-technical founder goals into DAPT + fine-tuning plans on curated data.”
Junior Data Scientist / ML Engineer specializing in LLMs and Computer Vision
“Currently working in CoRAL Lab, built and deployed IntegrityShield—a document-layer PDF watermarking system that keeps assessments visually identical while disrupting LLM-based solving; validated in a real classroom where it helped catch 12 AI-cheating cases. Also built MALDOC, a modular red-teaming platform for document-processing AI agents using LangGraph to run reproducible, deterministic adversarial trials across OCR/text/vision routes.”
Mid-level Full-Stack Java Developer specializing in enterprise SaaS and FinTech
“Software engineer with fintech/retirement-fund domain experience who led an internal dashboard consolidating fund transactions, approvals, and reporting into a single workflow tool. Strong in full-stack delivery (React + REST APIs + DB optimization) and in scaling/cleaning messy operational data via modular ETL pipelines (Python/Node), iterating post-launch with performance improvements like caching, pagination, and enhanced filtering.”
Mid-level Full-Stack Java Developer specializing in cloud-native microservices and FinTech
“Full-stack Java engineer (4+ years) who led end-to-end modernization of high-latency order management systems into cloud-native reactive microservices (Spring WebFlux) and built real-time React/Redux dashboards, reporting 99.98% uptime and 22% infra cost savings. Also headed a production RAG-based Order Support Bot at Dell Technologies with embeddings + MongoDB semantic search, automated validation and human fallback, plus CI/CD-driven LLM eval loops to reduce hallucinations.”
Mid-level Full-Stack Engineer specializing in cloud-native microservices (FinTech/Healthcare)
“Built and shipped production systems spanning real-time operational dashboards and an LLM-powered internal documentation assistant using RAG (embeddings + vector DB). Demonstrates strong focus on reliability and iteration: implemented guardrails and evaluation loops (human review, hallucination tracking, regression prevention) and improved performance/scalability through query optimization, caching, and retrieval tuning.”
“Built and deployed a production RAG-based internal knowledge assistant that let analysts query company documents in natural language, using LangChain/LangGraph with Pinecone and a FastAPI service for integration. Emphasizes reliability in production through hallucination mitigation (retrieval tuning + prompt guardrails) and measurable evaluation/monitoring (accuracy, latency, task completion, hallucination rate), iterating based on user feedback.”