Pre-screened and vetted.
Intern Software Engineer specializing in cloud, full-stack, and distributed systems
“Interned at SLB and owned an end-to-end GenAI chatbot deployment for a finance team, including invoice PDF data extraction and an LLM-driven layer (LangGraph/LangChain) that translated natural language to SQL and returned results in natural language. Validated LLM JSON outputs against benchmarks using DeepDiff and deployed the solution via Docker to Kubernetes, managing pods with k9s.”
Intern Data Scientist specializing in GenAI (LLMs, RAG) and ML model optimization
“Built and deployed a production LLM-powered risk assistant for KPMG and Freddie Mac that lets analysts query a confidential Neo4j risk graph in natural language (no Cypher), turning multi-day analysis into minutes with traceable, cited answers. Implemented rigorous guardrails, deterministic verification, RBAC/security controls, and a full eval/observability stack, cutting query error rate by ~50% and iterating through weekly UAT with non-technical risk analysts.”
Junior Backend/Platform Engineer specializing in AI microservices and cloud-native systems
“Cofounder at MeowyAI who shipped a production multimodal (vision/voice/text) AI task manager using Gemini, tackling real-world issues like hallucinations, tool-calling safety, and RAG-based preference memory. Also built a production multi-agent RAG system orchestrated with LangGraph (and contributes to LangChain), with strong emphasis on latency optimization, observability (OpenTelemetry), and rigorous testing/evaluation including A/B tests and adversarial prompting.”
Mid-level AI/ML Engineer specializing in MLOps, LLMs, and scalable ML systems
“ML/LLM engineer at Adobe who deployed a transformer-based personalization and campaign-targeting recommender system end-to-end, including PySpark/Airflow pipelines processing 12M+ events/day and containerized inference on AWS SageMaker (Docker/Kubernetes). Also has hands-on LLM workflow experience (RAG, semantic search, prompt optimization, hallucination mitigation) with a metrics-driven approach to reliability, drift monitoring, and reproducible retraining via MLflow.”
Mid-Level Software Development Engineer specializing in GenAI and full-stack cloud systems
“Full-stack engineer with experience across Magna, C3.ai, and Amazon, building GenAI-enabled products and finance transaction systems. Has shipped Next.js (App Router) + TypeScript features backed by Go/Python RAG pipelines, and emphasizes production quality via load testing, Selenium regression coverage, LLM-aware integration testing, and Azure observability. Also built LangGraph-orchestrated multi-step content generation workflows with robust retry/idempotency strategies.”
Junior ML Engineer specializing in Generative AI and LLM applications
“Built a production internal knowledge assistant using a RAG pipeline over large spreadsheets, PDFs, and support documents, using transformer embeddings stored in FAISS. Focused on real-world production challenges—format normalization, retrieval quality, hallucination reduction (context-only + citations), and latency—using hybrid retrieval, quantization, and containerized deployment, and communicated the workflow to non-technical stakeholders using simple analogies.”
Mid-Level Software Engineer specializing in Generative AI and RAG systems
“Built a production RAG-based natural-language-to-SQL system at Global Atlantic to replace slow, expensive manual analytics ticket workflows, focusing heavily on retrieval quality and measurable evaluation (200-question ground-truth set; recall@5 improved 0.65→0.78 via semantic chunking). Also built a custom MCP-style agent orchestrator for a personal project (arxiv-ai) to improve flexibility and Langfuse-aligned observability, and has hands-on experience with LangGraph, CrewAI, and n8n.”
Executive ML/AI Founder specializing in agentic analytics and data infrastructure
“Founder of Photosphere Labs (agentic AI for ecommerce data synthesis/analysis) who worked directly with customers to scope, build, demo, and iterate LLM-based solutions, including an AI chat product for brand owners. Previously at Block, built and explained a nuanced causal inference/propensity model tied to Square POS integrations, translating model specs and outputs into business impact for varied client contexts.”
Mid-level AI/ML Engineer specializing in LLMs, FinTech, and Healthcare IT
“Built production GenAI systems in both healthcare and financial services, including a Verily clinical platform and an Accenture financial Q&A product. Stands out for combining advanced RAG, fine-tuning, safety evaluation, and infrastructure engineering to deliver measurable gains in engagement, groundedness, hallucination reduction, and cost efficiency.”
Mid-level Software Engineer specializing in AI/ML and full-stack systems
“Engineer with Apple experience building LLM-powered internal workflow orchestration systems using Python, LangGraph, FastAPI, Redis, vector search, and Kubernetes. Stands out for a highly pragmatic, production-focused approach to agentic systems: deterministic state management, strong guardrails, observability, and human review for high-risk actions.”
Mid-level AI Engineer specializing in agentic LLM systems
“Built and productionized a dual-agent LLM invoice-processing system for GFI Partners, adding guardrails and audit trails to earn stakeholder trust and drive adoption while cutting operational burden by 75%. Uses LangSmith observability to diagnose real-time workflow regressions and has experience teaching agentic AI concepts (e.g., at Carnegie Mellon) through hands-on, scaffolded demos.”
Intern Applied AI/Software Engineer specializing in computer vision and full-stack platforms
“Built production LLM systems focused on reliability and safety, including a plain-English deployment tool that generates validated plans and provisions to Kubernetes while preventing unsafe actions via schema enforcement and plan/execute separation. Also created multi-LLM workflows (LangGraph) and stakeholder-friendly demos at Bosch, including a PyQt/FastAPI/CUDA app comparing SAM2 vs SAMWISE for on-device object detection with intuitive UX for business users.”
Senior Full-Stack Software Engineer specializing in workflow automation and healthcare AI
“Backend/data engineer who has owned production Python APIs and high-throughput async workflows on AWS (FastAPI, Docker, ECS/EKS/Lambda) with mature reliability practices like idempotency, bounded retries, circuit breakers, and strong observability. Also built AWS Glue ETL into an S3/Redshift lakehouse and modernized legacy batch systems via parallel-run parity testing and feature-flagged migrations, including a SQL tuning win cutting a multi-minute query to under 10 seconds.”
Mid-level Machine Learning Engineer specializing in GPU-accelerated LLM training and inference
“ML/LLM engineer with production experience building a multi-GPU LLM inference platform using TensorRT and vLLM, achieving ~40% p95 latency reduction through batching/KV caching, quantization, and CUDA/runtime tuning. Also has end-to-end orchestration experience (Kubernetes, Airflow) and has delivered real-time fraud detection systems at Accenture in close collaboration with non-technical risk and product stakeholders.”
Junior Machine Learning Engineer specializing in LLMs, computer vision, and robotics
“Built and deployed an agentic, multimodal LLM system that automates privacy redaction pipelines (audio/video/tabular) using LangChain orchestration and a closed-loop self-correction design. Personally implemented and performance-optimized core CV tooling (face blurring with tracking/Kalman filter) achieving >100 FPS on CPU, and validated reliability with golden-dataset benchmarking across 100+ privacy intents and measurable redaction metrics.”
Senior Software Engineer specializing in AI and FinTech platforms
“Built a production LLM pipeline at Walter AI that scans massive user inboxes, identifies financial newsletters, and extracts trading strategies into structured JSON for downstream paper-trading workflows. Stands out for combining agent architecture with strong production discipline—cutting scan time from 20 to 5 minutes, reducing LLM costs by 90%, and achieving 3-second P99 latency while handling messy, inconsistent email data at scale.”
Mid-level Data Science AI/ML Engineer specializing in Generative AI, LLMs, and RAG systems
“Built a production RAG-based "knowledge copilot" for support/ops using LangChain/LangGraph, implementing the full pipeline (ingestion, chunking, embeddings, vector DB retrieval/rerank, guarded generation with citations) and operating it as monitored microservices with CI/CD. Also designed an event-driven, streaming backend for real-time inventory ordering predictions that reduced stockouts by 25%, and has hands-on incident response experience stabilizing LLM API latency/5xx spikes using Datadog/APM and resilience patterns.”
Mid-level AI/ML Engineer specializing in LLMs, MLOps, and recommendation systems
Intern AI/ML Engineer specializing in LLM systems and cloud-native microservices
Senior Engineering Manager specializing in observability platforms and Generative AI
Mid-level NLP Research Engineer specializing in LLM evaluation and retrieval-augmented QA