Reval Logo

Vetted Latency Optimization Professionals

Pre-screened and vetted.

LW

LEQUAN WANG

Screened

Intern Applied Scientist / ML Engineer specializing in NLP and conversational AI

Seattle, WA0y exp
AmazonUC Irvine

LLM/Conversational AI engineer who built a production multi-turn dialogue system using LoRA fine-tuning on LLaMA, cutting training compute/memory by 90%+ while maintaining low-latency inference via quantization and streaming generation. Experienced in orchestrating end-to-end ML workflows with Prefect/Airflow/Kubeflow (including hyperparameter sweeps and W&B tracking) and improving agent reliability through benchmark-driven testing, shadow-mode rollouts, and stakeholder-informed guardrails.

View profile
SM

Mid-level Machine Learning Engineer specializing in LLMs, generative AI, and MLOps

San Francisco, CA5y exp
Scale AIConcordia University Wisconsin

Built and shipped a production LLM-powered medical scribe that generates structured clinical visit summaries using RAG, strict JSON schemas, and post-generation validation to reduce hallucinations. Experienced in making LLM workflows deterministic and observable (structured logging/metrics/tracing) and in evaluation-driven iteration with metrics like schema pass rate and edit rate; collaborated closely with clinicians and policy stakeholders at Scale AI to drive adoption.

View profile
SW

Entry-Level Software Engineer specializing in systems, networking, and ML

Atlanta, GA0y exp
AtlassianGeorgia Tech

Robotics software candidate with hands-on experience building controllers for an Autonomous Underwater Vehicle, including dual-PID control in Python with state-space modeling and a planned path to LQR. Developed ROS nodes for odometry-based localization, waypoint planning, and control command publishing, validated through a custom Gazebo/ROS simulation workflow with control-metric-driven testing. Also worked on F1Tenth simulation and scan-matching localization (PL-ICP), with additional cloud deployment experience using Docker/Kubernetes and CI/CD.

View profile
JT

Mid-level Machine Learning Engineer specializing in search ranking and NLP

San Francisco, CA3y exp
FaireUC San Diego
View profile
KS

Senior AI/ML Engineer specializing in LLMs, RAG, and multimodal recommendation systems

CA6y exp
PerplexityVirginia Tech
View profile
SY

Mid-level AI/ML Engineer specializing in LLMs, multimodal systems, and MLOps

5y exp
MetaEast Texas A&M University
View profile
YT

YUSEN TIAN

Screened

Senior Software Engineer specializing in cross-platform rendering and game UI systems

Hawaii, U.S.9y exp
Catenary GamesUCLA

Game UI engineer with shipped experience on Paladins (Unreal Engine 3/C++), including an in-game store with automated grid-based recommendation layout and an inventory UI built for PC/PlayStation/Xbox/Switch. Emphasizes production readiness through clean separation of UI/data, strong documentation for engineering and content workflows, and memory-conscious UI patterns (pooled entries/paged navigation) for constrained platforms like Switch.

View profile
CP

Mid-level AI/ML Engineer specializing in LLMs, RAG, and distributed MLOps

San Francisco, CA6y exp
PerplexityUniversity of North Texas
View profile
BM

Mid-level AI/ML Engineer specializing in LLMs, RAG, and production MLOps

San Francisco, CA6y exp
Scale AISaint Louis University
View profile
CI

Staff Software Engineer specializing in commerce, checkout, and payments platforms

Beaumont, CA13y exp
ShopifyUC Riverside
View profile
PM

Director-level Product-Led CTO specializing in AI/MarTech, retail media, and omnichannel platforms

San Francisco, CA25y exp
NikeNorthwestern University
View profile
KK

Senior Machine Learning Engineer specializing in LLM inference and GPU infrastructure

San Francisco, CA6y exp
PerplexityStevens Institute of Technology
View profile
NS

Mid-level AI/ML Engineer specializing in LLM training, RAG, and low-latency inference

New York city, NY4y exp
PerplexityCleveland State University
View profile
NT

Mid-level AI/ML Engineer specializing in LLMs, RAG, and scalable inference

San Francisco, CA6y exp
PerplexityUniversity of Nebraska Omaha

Backend/retrieval-focused engineer with production experience at Perplexity building a large-scale real-time Q&A system using retrieval-augmented generation, emphasizing low-latency, high-quality answers through ranking, context optimization, and caching. Also has orchestration experience from both product-facing LLM pipelines and large-scale infrastructure workflows at Meta, and has partnered with non-technical stakeholders to align AI trade-offs with business goals.

View profile
KT

Kenil Tanna

Screened

Staff-level Machine Learning Engineer specializing in LLMs and MLOps for Financial Services

New York, NY7y exp
JPMorgan ChaseIIT Guwahati

Machine learning/NLP practitioner at J.P. Morgan who led development of a production RAG system and an entity resolution pipeline for complex financial data. Deep hands-on experience with embeddings (Sentence-BERT), vector search (FAISS/pgvector), LLM fine-tuning (LoRA/PEFT), and rigorous evaluation (human-in-the-loop + A/B testing) backed by strong MLOps on AWS (Docker/Kubernetes, MLflow, Prometheus/Datadog).

View profile
SS

Sai supriya

Screened

Mid-level AI/ML Engineer specializing in LLM alignment, safety, and scalable inference

St. Louis, MO7y exp
AnthropicSaint Louis University

Built and productionized an AWS-hosted, Kubernetes-orchestrated RAG assistant that enables natural-language Q&A over internal document repositories with grounded answers and citations. Demonstrates strong applied LLM engineering: hallucination mitigation, hybrid retrieval + re-ranking, and rigorous evaluation via benchmarks and A/B testing, plus real-world scaling of compute-heavy inference with dynamic batching and monitoring.

View profile
HL

Hung-Chih Liu

Screened

Mid-level Distributed Systems & AI Infrastructure Engineer

Sunnyvale, CA3y exp
AmazonUCLA

Backend/full-stack engineer (Amazon experience) who built an AWS-based integration testing platform using Flask, ECS, Docker, and CloudWatch—cutting 1000+ test cases from ~5 hours to ~30 minutes while improving log visibility for non-engineering users. Also led a zero-downtime EU region migration with rigorous ORR testing, and built a Kinesis/Firehose/S3 + Glue/Spark replay mechanism for resilient data recovery. Side project: reproducible, cost-efficient LLM hosting platform on EKS using CDK and Karpenter for scale-to-zero.

View profile
AR

Anagha Ram

Screened

Intern AI/ML Engineer specializing in NLP, LLMs, and semantic search

Los Altos, CA2y exp
Columbia UniversityCornell University

Built and deployed a production RAG-based semantic search and summarization system for large legal/technical document sets, owning the full backend (embeddings, vector store, chunking, prompting) and driving a reported 40–60% reduction in manual review time. Experienced with LangChain/LlamaIndex plus Airflow/Temporal-style orchestration, and applies rigorous evaluation/monitoring (A/B tests, drift detection, staged rollouts) to keep agentic systems reliable. Also partnered with a supply-chain manager at TE Connectivity to deliver an AI inventory recommendation tool projected to drive millions in value.

View profile
AP

Amrita Pritam

Screened

Senior Backend Engineer specializing in distributed microservices and event-driven systems

Fremont, CA10y exp
MicrosoftManipal Institute of Technology

Backend engineer with production experience building a high-scale notification pipeline (~20M/day) using Java/Dropwizard with Kafka and Azure Queue, including DLQ/poison-message handling and the outbox pattern for reliability. Also led a batch-based migration of Yammer Messaging user data from PostgreSQL to Azure Cosmos DB for global multi-region scale, addressing throttling and network failures via retries, escalation policies, and dynamic throughput tuning.

View profile
KS

Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps

CA, USA4y exp
AnthropicCalifornia State University, Long Beach

ML/LLM engineer who built a production RAG system (GPT-4 + FAISS + FastAPI) to deliver fast, grounded answers from proprietary documents, optimizing for sub-200ms latency and high-concurrency scale. Strong MLOps/observability background: drift monitoring with Prometheus + Streamlit, automated retraining via Airflow, Kubernetes autoscaling, and MLflow-managed model lifecycle, plus inference cost reduction through quantization and structured pruning.

View profile

Need someone specific?

AI Search