Pre-screened and vetted.
Entry Machine Learning Engineer specializing in NLP, computer vision, and recommender systems
“Built and shipped an end-to-end podcast recommendation system exposed via a Flask API and React UI, explicitly balancing relevance, diversity (MMR), and safety constraints while meeting ~200ms latency targets. Also implemented a production-style RAG/information-extraction pipeline using web retrieval, spaCy NER, and fine-tuned SpanBERT with guardrails and evaluation loops (precision/recall/F1) to tune confidence thresholds and improve reliability.”
Junior Software Engineer specializing in full-stack systems, ML, and robotics perception
“Robotics software engineer with autonomous driving lab experience at UCSD, building and optimizing ROS2 perception and control pipelines (camera-based real-time object detection) with a strong focus on low-latency performance and robust message interfaces. Also brings production deployment experience from Hewlett Packard Enterprise, using Docker and Kubernetes for containerized environments and deployment pipelines.”
Mid-level AI/ML Engineer specializing in MLOps, computer vision, and NLP
“GenAI/ML engineer from Lucid Motors who built and productionized an LLM-powered RAG diagnostic assistant for manufacturing and maintenance teams, deployed on AWS with Docker/Kubernetes and MLflow. Demonstrates end-to-end ownership from retrieval/prompt design to scalability, monitoring, and workflow integration via APIs, plus production ML pipeline orchestration with Kubeflow (Spark/Kafka + TensorFlow) for predictive maintenance use cases.”
Junior AI/ML Engineer specializing in LLM systems and retrieval-augmented generation
“Built and deployed a production LLM-powered market intelligence and decision-support platform for noisy, real-time financial data, using a high-throughput embedding + vector DB RAG architecture to reduce hallucinations while keeping latency and cost low. Operated it at scale with GPU-backed inference (continuous batching/quantization), FastAPI on Kubernetes, and Airflow-orchestrated ingestion/embedding/retraining workflows, with strong schema-based reliability and monitoring.”
Mid-level Machine Learning Engineer specializing in NLP and computer vision
“AI/ML engineer with production experience building an LLM-powered resume-to-job matching and feedback product using RAG, with a strong focus on latency, hallucination control, and scalable deployment. Experienced orchestrating ML inference and backend services on Kubernetes and applying rigorous evaluation/guardrail practices; also partnered with business/product stakeholders at Walmart to improve an NLP-based supplier support system.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Mid-level Data/ML Engineer specializing in NLP, GenAI, and scalable data pipelines
“AI/ML engineer with production experience building LLM-powered document intelligence and customer support systems in healthcare/insurance, emphasizing high-accuracy RAG, long-document processing, and robust monitoring/fallback mechanisms. Also automates and scales ML lifecycle workflows using Apache Airflow and Kubeflow, and partners closely with non-technical operations stakeholders to drive adoption.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Mid-level AI Researcher specializing in LLMs, developer tools, and human-centered AI
“Research-focused AI engineer who built an agentic pipeline to automatically extract Sphinx-based API documentation/changelogs and generate synthetic tasks for a dynamic LLM code benchmark targeting real-world API evolution and deprecations. Experienced with multi-agent orchestration (AutoGen, LangChain, CrewAI) and rigorous evaluation methods, and has prior multi-agent work from a Microsoft Research internship.”
Mid-level AI/ML Engineer specializing in robotics perception and AR/VR systems
“AI engineer with robotics perception experience at Forterra, building and deploying moving-object/obstacle detection models into real-time robot pipelines. Addressed training crashes/latency via sub-batch training and optimizer tuning, and improved debugging using ROS/ROS2 tooling with 3D voxel visualization and color-coded validation.”
Junior Machine Learning Engineer specializing in LLMs and applied data science
“Built and shipped multiple production AI systems, including Auto DocGen (LLM-generated OpenAPI docs kept in sync via AST diffs, schema-constrained generation, and CI/CD on Render) and a multimodal sign-language recognition pipeline at USC orchestrated with FastAPI, MediaPipe, and PyTorch. Also partnered with Esri’s non-technical community team to fine-tune an LLaMA-based spam classifier with a review UI, cutting moderation time by 70%.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Red Hat ML/LLM engineer who designed and deployed a production LLM-powered customer support automation system using RAG, improving latency by 30% via PEFT and vector search optimization. Built security and governance into retrieval (access-level filtering, encrypted Pinecone/ChromaDB) and delivered SHAP-based explainability via a dashboard for non-technical stakeholders. Experienced orchestrating distributed ML/RAG pipelines across AWS SageMaker and OpenShift with Airflow/Prefect, plus multi-agent workflows using CrewAI and LangGraph.”
Mid-level Data Scientist specializing in NLP/LLMs, time series forecasting, and MLOps
“Data/ML practitioner with hands-on experience building NLP systems from prototype to production: delivered a Twitter sentiment classifier with robust preprocessing, SVM modeling, and Power BI reporting, and built entity-resolution pipelines for messy multi-source customer data (reporting ~95% improvement in unique entity identification). Also implemented semantic linking/search using SBERT embeddings with FAISS vector retrieval and domain fine-tuning (reported ~15% precision lift), and applies production workflow best practices (Airflow/Prefect, Docker, Azure ML/Databricks, Great Expectations).”
Junior Data Scientist and ML Researcher specializing in Transformers, multimodal AI, and autonomy
“Autonomous robotics student who built an end-to-end ROS2 semantic goal navigation system as a solo course project, integrating CLIP-based vision-language understanding with SLAM Toolbox and Nav2 to execute natural-language commands in Gazebo/RViz. Also implemented and tuned an RRT planner from scratch in Python and uses Docker plus GitHub workflows for reproducible, tested robotics codebases.”
Junior AI/ML Engineer specializing in multimodal generative models and NLP
“AI/ML engineer who has built a production text-to-image generation system in PyTorch with an AWS-backed inference setup, focusing on GPU-efficient training and embedding-space architectural choices inspired by recent research (e.g., Meta VL-JEPA). Uses both metric-based evaluation (FID) and human testing to validate real-world visual quality, and can translate technical concepts for non-technical stakeholders.”
Senior AI Research Engineer specializing in LLM agents and predictive maintenance
“At Delta Electronics, partnered with automotive firmware teams to productionize an LLM-based coding assistant for identifying safety standard violations and generating bug-fix guidance. Built an agentic workflow with stepwise context extraction, similarity search, and a separate judge model for scoring reasoning/retrieval, and drove internal adoption through pain-point discovery and tailored technical demos using real firmware code.”
Mid-level Software Engineer specializing in NLP and search systems
“Built an AI journaling app at HackCU 2025 featuring a speaking AI avatar with long-term memory via RAG (ChromaDB) and low-latency microservices coordinated through Kafka, including deployment under AMD/non-CUDA constraints using a quantized Llama 8B model. Also has Goldman Sachs experience deploying a Trade UI on Kubernetes with CI/CD rollback automation, plus a healthcare AI internship at CU Anschutz collaborating closely with physicians on diagnostic reasoning and dataset annotation.”
Junior Software Engineer specializing in ML, distributed systems, and LLM applications
“Interned at Zonda where he built an AI-driven semantic search solution over ~280M housing/builder records. Iterated from local LLMs via llama.cpp quantization to a vector-embedding retrieval system, then boosted semantic accuracy with a custom spaCy NER layer and re-ranking, optimizing for latency through precomputation. Collaborated with economics-focused stakeholders to reduce manual document/paperwork time by enabling natural-language search over internal data.”
Junior Full-Stack & Data Scientist specializing in ML/NLP and analytics products
“Built and deployed profitprops.io, a sports betting player-props prediction product using ML/AI. Implemented backend APIs with FastAPI/Express.js and Supabase, trained models on AWS GPU (P3) using Docker + RAPIDS, and set up CI/CD with GitHub Actions while working around cost constraints and data-collection hurdles (EC2 proxy rotation/rate limits).”
Junior Machine Learning Engineer specializing in MLOps and statistical modeling
“Integration engineer at ES Foundry who led deployment of ELsentinel, a production EL image-based solar cell quality monitoring system using a Swin Transformer classifier (>0.8 F1 across 15+ classes) plus a live real-time prediction dashboard. Strong in solving messy labeling/data-quality problems with process-team collaboration and shipping ML systems despite limited compute/infrastructure.”
Mid-level AI/ML Data Scientist specializing in NLP, computer vision, and risk analytics
“ML/AI engineer with Capital One experience building production-grade customer segmentation and fraud detection systems combining NLP (transformers) and anomaly detection. Strong MLOps and orchestration background (PySpark ETL, MLflow, Airflow, Docker/Kubernetes, Azure ML) with real-time monitoring/alerting and performance optimizations like quantization and caching, plus proven ability to deliver business-facing insights through Power BI/Tableau for marketing stakeholders.”
Mid-level Machine Learning Engineer specializing in NLP and cloud MLOps
“Built and deployed a production LLM-powered internal documentation assistant using embeddings, a vector database, and a RAG pipeline to reduce time spent searching PDFs/manuals. Experienced in orchestrating end-to-end LLM workflows with Airflow/LangChain, improving reliability via monitoring/error handling, and driving measurable quality through retrieval and hallucination-focused evaluation metrics.”
Mid-level AI/ML Engineer specializing in Generative AI and Conversational AI
“GenAI Engineer at Infosys who built and deployed a production multi-agent RAG system for a top-tier bank, scaling to ~50,000 queries/day with 99.9% uptime. Drove measurable gains (45% accuracy improvement, 30% API cost reduction) through open-source LLM fine-tuning, Pinecone indexing/retrieval optimization, and AWS-based MLOps/monitoring, and has experience enabling adoption via developer workshops and customer-facing collaboration.”