Pre-screened and vetted.
Mid-Level Software Engineer specializing in backend systems and LLM/RAG applications
“Backend/AI engineer at Intuit who built a production AI-powered case assistant for support agents (FastAPI on AWS EKS) combining Postgres case data, OpenSearch retrieval with embedding reranking, and internal LLMs. Improved peak-season reliability by diagnosing P95/P99 timeout spikes and cutting P95 latency from ~800ms to <400ms via composite indexing, keyset pagination, connection pool tuning, and caching, while adding grounded-generation guardrails (evidence packs, confidence thresholds, fallbacks, human-in-the-loop).”
Intern Software Engineer specializing in distributed systems and backend infrastructure
“Backend engineer with deep experience building event-driven logistics systems (orders, warehouse execution, real-time delivery tracking) using Spring Boot/PostgreSQL/Redis and strong observability (Prometheus/Grafana). Led a zero-downtime migration from monolithic MySQL to a sharded architecture for ~2M users with dual-write, checksum validation, and fast auto-rollback, and has strong security expertise including PostgreSQL RLS for multi-tenant SaaS and robust OAuth/JWT handling.”
Senior Full-Stack Engineer specializing in AI and cloud-native applications
“Built and shipped a production LLM-powered internal developer tool that accelerated code reviews by about 30% while maintaining reliability through modular orchestration, validation, and monitoring. Demonstrates strong practical depth in agent architecture, backend workflow orchestration, and observability for non-deterministic AI systems, with concrete examples of reducing agent errors by 60%.”
Senior AI/ML Data Scientist specializing in NLP, computer vision, and MLOps
“Applied LLMs and a graph-RAG architecture in Neo4j to automate an accounting firm's cross-checking of transactional books against tax regulations, indexing 1,000+ pages into a knowledge graph with vector search. Combines agentic LLM workflows with classical NER (Hugging Face/NLTK) and validates using expert-labeled held-out data plus precision/recall and measured accountant time savings after deployment.”
Entry-level Software Engineer specializing in AI and full-stack data systems
“Backend/AI engineer who has built an offline, citation-grounded RAG system end-to-end with hybrid retrieval, local LLM inference, and quantitative evaluation via RAGAS. Also brings real-time systems experience from an Airbnb-like booking platform and data pipeline/ML quality work from a Bilibili internship, with a strong emphasis on reliability, privacy, and measurable correctness.”
Staff Machine Learning Engineer specializing in LLM agents and ML systems
Senior Data Scientist specializing in GenAI, LLM systems, and production ML
Junior Software Engineer specializing in AI systems, retrieval, and knowledge graphs
Mid-level GenAI & Analytics Engineer specializing in LLM and cloud cost/finance analytics
Entry-Level AI Support Engineer specializing in ML tooling and full-stack debugging
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and document intelligence
Mid-level Applied AI Engineer specializing in reliable LLM agent workflows for regulated domains
Mid-level AI Engineer specializing in Generative AI and LLM/RAG systems
Senior Data Engineer specializing in Cloud Data Platforms and Generative AI
Senior AI/ML Engineer specializing in LLMs and enterprise conversational AI
Mid-Level Software Engineer specializing in Cloud, GenAI, and Federal systems
“Cloud-focused engineer experienced deploying and stabilizing complex production systems that span APIs, infrastructure, and automated workflows, with a strong observability and safe-release mindset (feature flags/canaries/rollbacks). Has hands-on, customer-facing incident leadership, including executing DR regional failover during an AWS us-east-1 outage to maintain service and reportedly save a client ~$10M.”
Mid-level Data Scientist / AI-ML Engineer specializing in Generative AI and LLM applications
“Built a production GenAI-powered analytics assistant to reduce reliance on data analysts by enabling natural-language Q&A over Databricks/Power BI dashboards, backed by vector search (Pinecone/Milvus) and a Neo4j knowledge graph, including multimodal support via OpenAI Vision. Demonstrates strong real-world LLM reliability engineering with strict RAG, LangGraph multi-step verification, and Guardrails/custom validators, plus broad orchestration and production monitoring experience (Airflow, ADF, Step Functions, Kubernetes, Prometheus/CloudWatch).”
Director-level AI & Data Science leader specializing in GenAI, LLMs, and MLOps
“ML/NLP engineer currently working in NYC on a system that connects complex unstructured data sources to deliver personalized insights, using embeddings + vector DB retrieval and a RAG architecture (LangChain, Pinecone/OpenSearch). Strong focus on production constraints—especially low-latency retrieval—using FAISS/ANN, PCA, index partitioning, and Redis caching, plus PEFT fine-tuning (LoRA/QLoRA) and KPI/SLA-driven promotion to production.”
Staff Full-Stack Engineer specializing in AI platforms and infrastructure automation
“Backend/full-stack engineer building complex internal platforms and customer-facing demos at the intersection of infrastructure and product. Shipped a no-code Product Lifecycle Manager for manufacturing (3 manufacturers, 1000+ evolving tests) using AWS S3/SQS ingestion and extensible Postgres (EAV+JSONB) with end-to-end traceability. Also built a FastAPI-based company data intelligence platform with Okta-secured RBAC and an LLM/MCP layer for ChatGPT-like analytics over enterprise data sources.”
Mid-level AI/ML Engineer specializing in Generative AI, RAG, and Conversational AI
“Built a production RAG-based GenAI copilot backend at Aetna using Python/FastAPI, GPT-4, LangChain, and Azure AI Search, deployed on AKS with Prometheus/Grafana observability. Owned the system end-to-end (ingestion through deployment) and improved peak-time reliability by addressing vector search and embedding bottlenecks with Redis caching, index optimization, and async processing, plus added anti-hallucination guardrails via retrieval confidence thresholds.”
Mid-level GenAI Engineer specializing in production RAG and LLM fine-tuning
“LLM engineer who built a production seller-support RAG system at eBay using hybrid retrieval (BM25 + Pinecone vectors) with Cohere reranking, LangGraph orchestration, and citation-grounded answers. Strong focus on reliability: semantic/structure-aware chunking, automated Ragas-based evaluation with nightly regressions, and production observability (LangSmith) plus drift monitoring (Arize). Also implemented a multi-agent fraud pipeline with AutoGen using JSON-schema contracts and explicit termination conditions.”