Pre-screened and vetted.
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and RAG systems
“Machine learning/NLP engineer who built a production-oriented retrieval-based AI system at Morgan Stanley for healthcare use cases, combining RAG over unstructured patient records with deep-learning medical image segmentation (U-Net/Mask R-CNN). Strong in end-to-end pipelines and MLOps (Spark/MongoDB, AWS SageMaker, CI/CD, monitoring, automated retraining) and in entity resolution/data quality validation for noisy clinical data.”
Mid-level Machine Learning Engineer specializing in forecasting, NLP, and GenAI
“GenAI/ML engineer with production experience building multilingual LLM systems (English/Spanish) and RAG-based clinical documentation summarization at Walgreens, combining prompt engineering, structured output validation, and rigorous evaluation (ROUGE + pharmacist review). Also orchestrated end-to-end ML pipelines for demand forecasting using Apache Airflow, PySpark, and MLflow with scheduled retraining and production monitoring.”
Mid-level AI/ML Engineer specializing in MLOps, NLP/LLMs, and computer vision
“Built and shipped a production LLM/RAG risk-case summarization and triage system used by fraud/compliance analysts, with strong grounding controls (evidence-cited outputs and refusal on low confidence). Demonstrates end-to-end ownership across retrieval quality, Airflow-orchestrated indexing pipelines, and compliance-grade privacy (PII redaction, RBAC, encrypted redacted logging, and auditable prompt/model versioning) plus a tight feedback loop with non-technical domain experts.”
Mid-level AI/ML Engineer specializing in MLOps, computer vision, and NLP
“GenAI/ML engineer from Lucid Motors who built and productionized an LLM-powered RAG diagnostic assistant for manufacturing and maintenance teams, deployed on AWS with Docker/Kubernetes and MLflow. Demonstrates end-to-end ownership from retrieval/prompt design to scalability, monitoring, and workflow integration via APIs, plus production ML pipeline orchestration with Kubeflow (Spark/Kafka + TensorFlow) for predictive maintenance use cases.”
Mid-level Machine Learning Engineer specializing in NLP and computer vision
“AI/ML engineer with production experience building an LLM-powered resume-to-job matching and feedback product using RAG, with a strong focus on latency, hallucination control, and scalable deployment. Experienced orchestrating ML inference and backend services on Kubernetes and applying rigorous evaluation/guardrail practices; also partnered with business/product stakeholders at Walmart to improve an NLP-based supplier support system.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Data/ML Engineer specializing in NLP, GenAI, and scalable data pipelines
“AI/ML engineer with production experience building LLM-powered document intelligence and customer support systems in healthcare/insurance, emphasizing high-accuracy RAG, long-document processing, and robust monitoring/fallback mechanisms. Also automates and scales ML lifecycle workflows using Apache Airflow and Kubeflow, and partners closely with non-technical operations stakeholders to drive adoption.”
Mid-level Machine Learning Engineer/Researcher specializing in computer vision and multimodal AI
“Developed a production wildfire smoke detection system where smoke is visually subtle and easily confused with fog/clouds; addressed this with a hybrid CNN+LSTM+ViT model and multimodal weather features to reduce false positives. Experienced running scalable, reproducible ML pipelines on shared GPU infrastructure using Slurm and Kubernetes-style batch jobs with checkpointing, retries, and rigorous error analysis.”
Intern Full-Stack Software Engineer specializing in data pipelines and AI/ML systems
“Software engineer with experience building a Vue.js/TypeScript internal component library (with Jest testing standards) and improving JS runtime performance via profiling, code splitting, and lazy loading. Also led documentation and community support for a Python ML utility library, diagnosing metric-calculation bugs for imbalanced datasets and driving large reductions in support inquiries through targeted docs, tests, and rapid hotfixes in a startup environment.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Red Hat ML/LLM engineer who designed and deployed a production LLM-powered customer support automation system using RAG, improving latency by 30% via PEFT and vector search optimization. Built security and governance into retrieval (access-level filtering, encrypted Pinecone/ChromaDB) and delivered SHAP-based explainability via a dashboard for non-technical stakeholders. Experienced orchestrating distributed ML/RAG pipelines across AWS SageMaker and OpenShift with Airflow/Prefect, plus multi-agent workflows using CrewAI and LangGraph.”
Mid-level Data Scientist specializing in NLP/LLMs, time series forecasting, and MLOps
“Data/ML practitioner with hands-on experience building NLP systems from prototype to production: delivered a Twitter sentiment classifier with robust preprocessing, SVM modeling, and Power BI reporting, and built entity-resolution pipelines for messy multi-source customer data (reporting ~95% improvement in unique entity identification). Also implemented semantic linking/search using SBERT embeddings with FAISS vector retrieval and domain fine-tuning (reported ~15% precision lift), and applies production workflow best practices (Airflow/Prefect, Docker, Azure ML/Databricks, Great Expectations).”
Junior Full-Stack & Data Scientist specializing in ML/NLP and analytics products
“Built and deployed profitprops.io, a sports betting player-props prediction product using ML/AI. Implemented backend APIs with FastAPI/Express.js and Supabase, trained models on AWS GPU (P3) using Docker + RAPIDS, and set up CI/CD with GitHub Actions while working around cost constraints and data-collection hurdles (EC2 proxy rotation/rate limits).”
“Built an LLM multi-agent “ingredient safety” analyzer for cosmetics that cuts consumer research time from ~20+ minutes to minutes, using LangGraph orchestration, hybrid retrieval (Qdrant + Tavily), and safety-focused critic validation (false rejections reduced ~30%→~8%). Also has research-internship experience building computer-vision pipelines to classify emerald color/clarity by translating gem-expert heuristics into quantitative model features.”
Junior Machine Learning Engineer specializing in MLOps and statistical modeling
“Integration engineer at ES Foundry who led deployment of ELsentinel, a production EL image-based solar cell quality monitoring system using a Swin Transformer classifier (>0.8 F1 across 15+ classes) plus a live real-time prediction dashboard. Strong in solving messy labeling/data-quality problems with process-team collaboration and shipping ML systems despite limited compute/infrastructure.”
Mid-Level Software Engineer specializing in backend microservices and FinTech data pipelines
“Backend engineer at Goldman Sachs who built LLM-powered reconciliation/reporting services and high-throughput Kafka pipelines (8M+ events/day). Strong in production-grade Python/FastAPI microservices on Kubernetes with GitOps-style CI/CD, plus experience migrating legacy reporting/settlement services onto an internal Kubernetes platform using shadow deployments and gradual cutovers.”
Mid-level Machine Learning Engineer specializing in financial AI, NLP, and MLOps
“AI/ML engineer with experience at Accenture and Morgan Stanley, building production LLM systems (GPT-3 summarization) and finance-focused ML models (credit risk and trading anomaly detection). Combines MLOps depth (Docker/Kubernetes, AWS SageMaker/Glue/Lambda, MLflow, A/B testing, drift monitoring) with practical domain adaptation techniques like few-shot prompting and RAG/knowledge-base integration.”
Mid-level AI/ML Engineer specializing in healthcare NLP and MLOps
“Healthcare/clinical ML practitioner who built and productionized ClinicalBERT-based pipelines to extract and standardize oncology EHR data, improving downstream model F1 from 0.81 to 0.92 while controlling training cost via LoRA/QLoRA. Experienced orchestrating real-time AWS ETL/ML workflows (Glue, Lambda, SageMaker) and partnering with clinicians using SHAP-based interpretability, contributing to an 18% reduction in readmissions and full adoption.”
Senior Data Scientist specializing in machine learning and customer analytics
“Data/ML practitioner with experience applying NLP and classical ML to large-scale customer data (2B+ records) for segmentation, prediction, and survey-text classification, delivering measurable business impact (~18% engagement efficiency). Has hands-on entity resolution across multi-source datasets and has built embedding-based semantic search using SentenceBERT + a vector database with domain fine-tuning (~20% relevance improvement), plus production workflow experience with Spark/Airflow and cloud tooling (AWS/Azure).”
Mid-level AI Software Engineer specializing in risk and fraud detection
“AI/software engineer with experience at Visa building a real-time transaction fraud/risk scoring microservice in the card authorization path (Python, Kafka, Kubernetes on AWS) with strict 120–150ms latency constraints and reason-code outputs for downstream decisioning. Owns ML backend end-to-end (data/feature engineering, model training, deployment) and has demonstrated production reliability work including latency spike mitigation, SLO-based observability, drift monitoring, and safe fallbacks to rule-based decisions.”
Mid-level AI/ML Engineer specializing in NLP, LLMs, and risk modeling
“GenAI/LLM engineer who architected and deployed a production RAG “research assistant” for JPMorgan Chase’s regulatory compliance team, focused on safety-critical behavior (mandatory citations, refusal when evidence is missing). Deep hands-on experience with LlamaIndex, Pinecone, Hugging Face embeddings, LangGraph agent workflows, and metric-driven evaluation (golden sets, TruLens), including a reported 28% relevancy lift via cross-encoder re-ranking.”
Mid-level Machine Learning Engineer specializing in fraud detection and LLM applications
“Unreal Engine UI engineer focused on scalable, production-ready UI architecture (C++/Slate/UMG/CommonUI) with strong designer enablement via decoupled, interface-driven patterns and MVVM. Demonstrated measurable performance wins: replaced 200+ per-frame Blueprint bindings to cut UI prepass/paint from 4.2ms to 0.5ms and reduced VRAM by ~120MB using texture streaming proxies.”
Mid-level AI/ML Engineer specializing in financial risk, fraud detection, and GenAI
“GenAI/ML engineer in Citigroup’s finance environment who has deployed production RAG systems for investment banking under strict privacy and model-risk constraints. Built an internal-VPC Llama2 + Pinecone + LangChain solution with NER redaction and citation-based verification to prevent hallucinations, delivering major time savings, and also partnered with global finance executives to ship an AI early-warning indicator for treasury/liquidity risk.”
Mid-Level AI Engineer specializing in NLP, computer vision, and LLM applications
“LLM/RAG practitioner who productionized an LLM-driven customer communication and transaction understanding system at PayPal, emphasizing privacy/compliance guardrails and large-scale data normalization. Experienced in real-time debugging of hallucinations via retrieval pipeline tuning and in leading hands-on developer workshops and sales-aligned POCs to drive adoption.”
Junior AI & ML Engineer specializing in agentic systems and full-stack AI products
“Won a machine learning contest and was placed onto a Kaiser data science team, where they built ML models for hospital bottleneck prediction and resource allocation. They later built and deployed a full-stack LLM-based “data analyst agent” (with custom orchestration plus LangChain/OpenAI Agents experience) that generates analysis code, answers questions, and produces dashboards from uploaded datasets, emphasizing rigorous evaluation sets, robustness, and healthcare security/compliance integration.”