Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in fraud detection and NLP
“Built production AI/RAG-style systems for message Q&A and insurance claims workflows, combining data ingestion, indexing/retrieval, and LLM integration with fallback modes. Has hands-on orchestration experience (Airflow, Prefect, LangChain) and cites large operational gains (claims processing reduced to ~45 seconds; manual review -50%; false alerts -30%) through automated, monitored pipelines and close collaboration with non-technical stakeholders.”
“AI/ML engineer with banking domain experience (M&T Bank) who built a production credit-risk prediction and reporting platform combining ML models (XGBoost/TensorFlow) with a RAG pipeline (LangChain + GPT-4) over compliance documents. Delivered measurable impact (≈20% better risk detection/precision, 50% less manual reporting) and productionized workflows on Vertex AI/Kubeflow with CI/CD and monitoring; also implemented embedding-based semantic search using FAISS/Pinecone.”
Mid-level Machine Learning Engineer specializing in cloud, governance automation, and distributed systems
“Governance engineer intern at GSK who built policy-as-code automation using Open Policy Agent/Rego integrated into GitHub CI/CD and Terraform workflows. Also built and shipped a voice-enabled expense tracking app using speech-to-text + LLM structured extraction with strong validation, retries, and semantic guardrails, and designed the supporting PostgreSQL data model with performance-focused indexing.”
Junior Machine Learning Engineer specializing in LLMs and RAG systems
“Production-focused applied ML/LLM engineer who has deployed an LLM-powered RAG assistant and improved reliability through rigorous retrieval evaluation (recall/MRR), reranking, and guardrails that prevent confident wrong answers. Experienced running containerized ML/LLM services on Kubernetes (including AWS-managed layers) with CI/CD and observability, and has delivered a real-time predictive maintenance system using streaming sensor data and time-series anomaly detection in close partnership with maintenance teams.”
Mid-level AI/ML & Backend Engineer specializing in AI platforms and computer vision
“Backend engineer with hands-on experience building real-time, low-latency systems: owned the Python backend for a real-time crowd-monitoring product (top 5% at HackHarvard 2025) using OpenCV, GPU YOLO inference (PyTorch), WebRTC, and OAuth. Also has production Kubernetes/GitOps experience (Helm/Kustomize, GitHub Actions, Argo CD), Kafka-based event pipelines, and executed a minimal-downtime on-prem PostgreSQL migration to AWS EC2.”
Mid-level AI/ML Engineer specializing in GenAI and cloud MLOps
“Applied LLMs to high-stakes domains (wildfire risk for emergency teams and loan approval via a fine-tuned IBM Granite model), with a strong focus on reliability—using RAG-based cross-validation to reduce hallucinations and continuous ingestion pipelines (MODIS satellite imagery via AWS Lambda) to keep data current. Experienced in production orchestration and MLOps-style workflows using Airflow, AWS Step Functions, and SageMaker Pipelines, and collaborates closely with analysts on KPI-driven evaluation.”
Mid-level AI/ML Engineer specializing in MLOps, NLP, and scalable model deployment
“Built and deployed a production autonomous AI data analyst agent (LangChain + GPT + Streamlit on AWS) that turns natural-language questions into validated SQL, visualizations, and insights, cutting manual analysis time by ~50%. Emphasizes reliability and MLOps: schema-aware validation/guardrails to prevent hallucinations, scalable large-data processing, and Azure DevOps CI/CD + MLflow for automated deployment and experiment tracking.”
Mid-level Software Engineer specializing in AI/ML and full-stack development
“Backend Java engineer with strong platform/DevOps experience: modernized an insurance claims legacy monolith into DDD-aligned microservices, deployed containerized services on Kubernetes with Jenkins CI/CD and static analysis gates, and implemented GitOps using ArgoCD. Also led major AWS migration planning with dependency mapping and network monitoring to uncover hidden dependencies, and built Kafka-based real-time event streaming with schema-registry-driven evolution.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Built a production RAG-based healthcare chatbot to retrieve patient medical documents spread across multiple platforms, reducing manual and error-prone searching. Implemented semantic search with custom embeddings (Hugging Face) and Pinecone, deployed via FastAPI/Docker on AWS SageMaker with MLflow tracking, and optimized fine-tuning cost using LoRA while orchestrating retraining pipelines in Airflow.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG
“GenAI/LLM engineer who built and deployed a production RAG system for enterprise document search and decision support, cutting manual lookup time by 40%+. Experienced with LangChain/LangGraph agent orchestration plus Airflow/Prefect for ingestion and incremental reindexing, with a strong focus on reliability (testing, observability) and stakeholder-driven metrics.”
Mid-level AI/ML & Full-Stack Engineer specializing in LLM agents and generative AI
“LLM/agent builder who shipped a live consumer AI-agent app (kalpa.chat) that visualizes complex reasoning as interactive graphs and abstracts multi-provider model usage via a unified wallet. Professionally has applied LangChain/LangGraph to IVR parsing and to scaling a football video-generation pipeline at DAZN, including shipping a VAR-specific retrieval/order fix via SQL after iterating with a non-technical PM.”
Mid-level AI/ML Engineer specializing in GenAI, MLOps, and anomaly detection
“LLM/MLOps engineer who has shipped a production RAG-based technical documentation assistant (FastAPI) cutting manual review by 45%, with deep hands-on retrieval optimization in Pinecone/LangChain (HNSW, hybrid + multi-query search, caching). Also brings healthcare domain experience—building Airflow-orchestrated EHR pipelines and delivering FDA-auditability-friendly predictive maintenance solutions using SHAP/LIME explainability surfaced in Power BI.”
Mid-level AI/ML Engineer specializing in NLP, RAG, and MLOps for FinTech
“ML/LLM engineer with production experience building a compliant RAG-based virtual assistant at Intuit, optimizing embeddings and FAISS retrieval (including PCA) for low-latency, privacy-controlled search and deploying via AWS SageMaker containers. Also built scalable Airflow+MLflow pipelines using Docker and KubernetesExecutor, cutting training cycles by 37%, and partnered with civil engineers/project managers at Aegis Infra to deliver predictive maintenance for construction equipment.”
Junior AI/ML & Full-Stack Engineer specializing in LLMs and RAG systems
“Forward-deployed engineer who built a production AI drone-control chatbot that lets users fly a drone via natural language while viewing a real-time feed. Implemented RAG over drone SDK documentation (vector DB + top-k retrieval) and LoRA fine-tuning, with a focus on latency, token efficiency, and cost reduction, and regularly works with non-technical clients to integrate and explain AI system architecture.”
Mid-level Machine Learning Engineer specializing in NLP, recommender systems, and MLOps
“ML/LLM engineer with production experience at General Motors building Transformer-based search and recommendation personalization for a high-traffic vehicle platform. Delivered significant KPI gains (17% conversion lift, 14% bounce-rate reduction) and optimized real-time inference via ONNX Runtime and INT8 quantization while implementing robust MLOps (Airflow/MLflow, monitoring, drift-triggered retraining) and stakeholder-facing explainability/dashboards.”
Senior Backend/Cloud Engineer specializing in IaC, SaaS platforms, and ML/Computer Vision
“Backend/infrastructure engineer with experience across API development (FastAPI/MySQL/SQLAlchemy), Kubernetes deployments, and large-scale data processing—built a Dockerized Python pipeline to pre-aggregate ~1B Graylog events for efficient querying. Has enterprise infrastructure automation background at Hewlett Packard Enterprise (Datafabric) using Terraform/Ansible with fail-fast and rollback practices, plus Kafka-based sensor streaming prototypes to Google Cloud with Java workers and autoscaling.”
Mid-level AI/ML Engineer specializing in predictive modeling and cloud ML pipelines
“LLM engineer/data engineer who has deployed production RAG systems for internal-document Q&A, building end-to-end ingestion, embedding, vector search, and FastAPI serving while actively reducing hallucinations and latency through rigorous retrieval tuning and caching. Also experienced in orchestrating cloud data pipelines (Airflow, AWS Glue, Azure Data Factory) and partnering with non-technical business teams to deliver AI solutions like automated document review.”
Mid-level Machine Learning Engineer specializing in LLM platforms and robotic perception
“Built and shipped a production multi-agent personal financial assistant at AlphevaAI on AWS ECS, combining FastAPI microservices, Redis/SQS orchestration, and Pinecone-based hybrid RAG (semantic + BM25) to ground financial guidance. Improved routing accuracy with an embedding-based SetFit + logistic regression intent classifier feeding an LLM router, and optimized UX with live streaming plus cost controls via model tiering and caching.”
Junior Machine Learning Engineer specializing in NLP and multimodal transformers
“Built and deployed LLM-powered agentic chatbot and text-to-SQL systems using LangGraph/LangChain (and Bedrock), structuring workflows as DAGs with planning/replanning and validation to improve tool-calling reliability and reduce hallucinations. Operates production feedback loops with online/offline metrics, drift detection, and LangSmith-based evaluation pipelines, and regularly partners with business stakeholders and clinicians using slide decks and visual charts.”
Junior Machine Learning Engineer specializing in NLP, computer vision, and MLOps
“ML/LLM engineer with Meta experience building production AI systems for near real-time user-report classification and summarization under strict latency (<250ms), safety, cost, and privacy constraints. Has hands-on MLOps/orchestration experience (Airflow, Spark, MLflow, Kubernetes, Docker, GitHub Actions) plus observability (Prometheus/Grafana) and applies rigorous evaluation, staged rollouts, and A/B testing to keep agent workflows reliable in production.”
Mid-level AI/ML Engineer specializing in production ML, RAG systems, and MLOps
“Built and shipped a widely adopted, production-grade RAG internal search assistant that unified scattered engineering knowledge, deployed as a FastAPI service on Kubernetes with FAISS + LangChain. Demonstrates deep practical expertise in retrieval tuning (chunking, hybrid search, re-ranking) and in making LLM workflows reliable in production via guardrails, monitoring, and evaluation, plus strong cross-functional delivery with non-technical operations teams.”
Mid-level Software Engineer specializing in AI/ML for FinTech and Healthcare
“Built and deployed an end-to-end fintech product, FinSight, for bank statement analysis and financial Q&A using a production-style RAG architecture. Stands out for combining FastAPI, OpenAI embeddings, FAISS, hybrid SQL/vector retrieval, and practical reliability work like chunking optimization, validation, and low-latency performance tuning.”
Mid-level AI Engineer specializing in LLM agents and RAG for health-tech
“Backend engineer with health-tech AI platform experience who designed a modular FastAPI/PostgreSQL architecture supporting real-time user data and swap-in AI workflows. Has hands-on production experience with observability (CloudWatch, structured logging, LangSmith/LangGraph/LangChain tracing), secure auth (OAuth2/JWT, RBAC, RLS), and careful data-pipeline migrations using parallel runs and rollback planning.”
Mid-level AI/ML Engineer specializing in financial risk, fraud analytics, and forecasting
“Built and productionized an LLM-powered financial intelligence and forecasting platform at Northern Trust using a RAG architecture (LangChain + Hugging Face + FAISS) with end-to-end MLOps (Docker/Kubernetes, Airflow, MLflow). Emphasized regulatory-grade explainability (SHAP/Power BI) and hallucination control (retrieval-only grounding), achieving ~30% forecasting accuracy improvement and ~65% reduction in analyst research time, with sub-second inference and 95% uptime on EKS/AKS.”