Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps for financial services
“Built and deployed a production Llama 3-based RAG document Q&A system using FAISS, addressing context-window limits through chunking and keeping retrieval accurate by regularly refreshing embeddings. Has hands-on orchestration experience with LangChain and LlamaIndex for multi-step LLM workflows (including memory management) and collaborates with non-technical teams (e.g., marketing) to deliver AI solutions like recommendation systems.”
Mid-level Software Engineer specializing in LLM, RAG, and cloud AI
“Recent master’s graduate who led a team project building an LLM-based chatbot with RBAC-controlled information disclosure and a focus on reducing hallucinations. Also has hands-on embedded robotics experience (Arduino obstacle-avoiding robot using ultrasonic sensors) and practical DevOps/cloud deployment exposure with Docker, Terraform, Jenkins, and AWS (EKS/ECS/CodePipeline).”
Mid-level Machine Learning Engineer specializing in LLM platforms and robotic perception
“Built and shipped a production multi-agent personal financial assistant at AlphevaAI on AWS ECS, combining FastAPI microservices, Redis/SQS orchestration, and Pinecone-based hybrid RAG (semantic + BM25) to ground financial guidance. Improved routing accuracy with an embedding-based SetFit + logistic regression intent classifier feeding an LLM router, and optimized UX with live streaming plus cost controls via model tiering and caching.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Built a production RAG-based healthcare chatbot to retrieve patient medical documents spread across multiple platforms, reducing manual and error-prone searching. Implemented semantic search with custom embeddings (Hugging Face) and Pinecone, deployed via FastAPI/Docker on AWS SageMaker with MLflow tracking, and optimized fine-tuning cost using LoRA while orchestrating retraining pipelines in Airflow.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG
“GenAI/LLM engineer who built and deployed a production RAG system for enterprise document search and decision support, cutting manual lookup time by 40%+. Experienced with LangChain/LangGraph agent orchestration plus Airflow/Prefect for ingestion and incremental reindexing, with a strong focus on reliability (testing, observability) and stakeholder-driven metrics.”
Junior Machine Learning Engineer specializing in NLP, computer vision, and MLOps
“ML/LLM engineer with Meta experience building production AI systems for near real-time user-report classification and summarization under strict latency (<250ms), safety, cost, and privacy constraints. Has hands-on MLOps/orchestration experience (Airflow, Spark, MLflow, Kubernetes, Docker, GitHub Actions) plus observability (Prometheus/Grafana) and applies rigorous evaluation, staged rollouts, and A/B testing to keep agent workflows reliable in production.”
Mid-level AI Engineer specializing in multi-agent systems and RAG
“Built and shipped a production LangGraph-based multi-agent LLM analytics/decision copilot that answers questions across SQL/BI systems and unstructured docs, emphasizing grounded, tool-verified outputs with citations and confidence gating. Deep hands-on experience with orchestration (LangGraph, CrewAI, OpenAI Assistants, MCP) plus real-world latency/cost optimization (vLLM batching/KV caching, speculative decoding, quantization) and rigorous eval/observability. Partnered closely with business/ops stakeholders to deliver explainable reporting automation, cutting manual reporting time by 50%+.”
Junior Data Scientist specializing in ML, geospatial analytics, and LLM applications
“Built and deployed a production AI “term explainer” agent that adapts explanations to beginner/intermediate/expert users by combining multi-step LLM reasoning with grounded Wikipedia retrieval. Owns end-to-end agent orchestration (smolagents/Python), reliability patterns (fallback across LLM providers, retries, guardrails), and observability/metrics-driven evaluation; also partnered with a non-technical researcher to deliver a plain-language research assistant agent.”
Mid-level AI/ML Engineer specializing in GenAI, MLOps, and anomaly detection
“LLM/MLOps engineer who has shipped a production RAG-based technical documentation assistant (FastAPI) cutting manual review by 45%, with deep hands-on retrieval optimization in Pinecone/LangChain (HNSW, hybrid + multi-query search, caching). Also brings healthcare domain experience—building Airflow-orchestrated EHR pipelines and delivering FDA-auditability-friendly predictive maintenance solutions using SHAP/LIME explainability surfaced in Power BI.”
Mid-level AI/ML Engineer specializing in NLP, RAG, and MLOps for FinTech
“ML/LLM engineer with production experience building a compliant RAG-based virtual assistant at Intuit, optimizing embeddings and FAISS retrieval (including PCA) for low-latency, privacy-controlled search and deploying via AWS SageMaker containers. Also built scalable Airflow+MLflow pipelines using Docker and KubernetesExecutor, cutting training cycles by 37%, and partnered with civil engineers/project managers at Aegis Infra to deliver predictive maintenance for construction equipment.”
“Backend/data engineer who builds Python (FastAPI) data-processing API services for internal analytics/reporting, emphasizing modular architecture, async performance tuning, and reliability patterns (health checks, retries, observability). Also migrated legacy on-prem ETL pipelines to Azure using ADF/Data Lake/Functions and implemented a near-real-time ingestion flow with Event Hubs plus watermarking to handle late events and deduplication.”
Mid-level Software Engineer specializing in cloud-native data pipelines and ML platforms
“Backend engineer who has owned end-to-end delivery of Python/FastAPI microservices for real-time data processing and alerting, including performance tuning (Postgres optimization, caching, async processing). Strong DevOps/GitOps background: Docker + Kubernetes deployments with GitHub Actions CI/CD and ArgoCD-driven GitOps, plus experience supporting phased on-prem to AWS migrations and building Kafka-based streaming pipelines.”
Mid-level Data Scientist specializing in ML, NLP, and LLM-powered solutions
“AI/NLP-focused practitioner who built a zero-/few-shot LLM event extraction system on the long-tail Maven dataset, combining prompt-structured outputs with LoRA/QLoRA fine-tuning and rigorous F1 evaluation. Also implemented entity resolution/data cleaning pipelines and embedding-based semantic search using Sentence-BERT + FAISS, and has healthcare experience delivering a multilingual speech/translation mobile prototype using HIPAA-compliant Azure Cognitive Services.”
Mid-level Data Scientist specializing in NLP and predictive modeling
“AI/ML practitioner in healthcare/insurance (Blue Cross Blue Shield) who built and deployed a production NLP system to classify patient risk from unstructured clinical notes. Experienced in end-to-end pipeline orchestration (Airflow, AWS Step Functions/Lambda/SageMaker) and real-time optimization (BERT to DistilBERT on AWS GPUs), with strong clinician collaboration to drive adoption.”
Mid-level Machine Learning Engineer specializing in NLP, recommender systems, and MLOps
“ML/LLM engineer with production experience at General Motors building Transformer-based search and recommendation personalization for a high-traffic vehicle platform. Delivered significant KPI gains (17% conversion lift, 14% bounce-rate reduction) and optimized real-time inference via ONNX Runtime and INT8 quantization while implementing robust MLOps (Airflow/MLflow, monitoring, drift-triggered retraining) and stakeholder-facing explainability/dashboards.”
Mid-level Applied AI Engineer specializing in agentic LLM workflows
“AI engineer with production experience building a LangGraph-based, stateful multi-agent system at MetLife to automate complex insurance claims adjudication, integrating document discovery, Azure Document Intelligence OCR/extraction, and health data analysis. Strong in agent orchestration and production deployment (Docker + FastAPI REST APIs), with a structured approach to reliability, evaluation, and stakeholder-driven requirements.”
Mid-level DevOps/Cloud Engineer specializing in multi-cloud CI/CD and Kubernetes
“IBM Power/AIX infrastructure engineer who has owned a sizable production estate (50 Power servers / ~200 LPARs) spanning VIOS/HMC, SAN/NFS, and PowerHA clusters. Demonstrates strong incident leadership (LPAR outage + split-brain recovery) and a process-improvement mindset with measurable reductions in recurrence/MTTR, while also bringing modern DevOps/IaC experience (Jenkins, ArgoCD, Terraform, security scanning, canary/blue-green).”
Mid-Level Full-Stack Software Engineer specializing in cloud-native FinTech systems
“Software engineer with JPMorgan Chase experience delivering end-to-end fintech features (Next.js/React/Node/Postgres on AWS) and measurable performance gains. Built and productionized an AI-native credit decisioning workflow combining LLMs, vector retrieval, and a rules engine with strong governance (bias checks, auditability, human-in-loop), improving precision and cutting underwriting turnaround time by 40%.”
Mid-level AI/ML Engineer specializing in financial risk, fraud analytics, and forecasting
“Built and productionized an LLM-powered financial intelligence and forecasting platform at Northern Trust using a RAG architecture (LangChain + Hugging Face + FAISS) with end-to-end MLOps (Docker/Kubernetes, Airflow, MLflow). Emphasized regulatory-grade explainability (SHAP/Power BI) and hallucination control (retrieval-only grounding), achieving ~30% forecasting accuracy improvement and ~65% reduction in analyst research time, with sub-second inference and 95% uptime on EKS/AKS.”
Senior LLM Engineer specializing in Generative AI, RAG, and multimodal assistants
“GenAI/NLP engineer with experience building classification and summarization pipelines in PyTorch and deploying multimodal GPT-4-style workflows. Has integrated LLM applications across OpenAI, Azure OpenAI, and Amazon Bedrock, and uses LangChain/LlamaIndex/Semantic Kernel to orchestrate RAG and agent workflows with production-focused evaluation metrics like task success rate and groundedness.”
Junior Software Engineer specializing in Full-Stack and GenAI/LLM applications
“LLM/RAG practitioner building clinician-facing AI search and Q&A inside EHR workflows, focused on trust, latency, and safety (grounded answers with citations, PHI controls, encryption/audit logs). Demonstrated real-time incident response for production LLM systems (e.g., fixing a metadata-filter deployment regression to prevent irrelevant results/cross-patient leakage) and strong demo/enablement skills for mixed technical and clinical stakeholders; also shipped a multi-model RAG tool at OrbeX Labs with upload/search/audit features for day-to-day adoption.”
Mid-level AI/ML Engineer specializing in MLOps and cloud-deployed ML systems
“ML/AI engineer who built and productionized an NLP system at PurevisitX, orchestrating end-to-end ML workflows with Airflow (S3 ingestion through auto-retraining) and optimizing for drift and low-latency inference. Also partnered with Citibank risk teams on a fraud detection model, translating results via dashboards and iterating thresholds based on stakeholder feedback.”
Mid-level Machine Learning Engineer specializing in LLMs, NLP, and MLOps
“Built a production LLM-RAG system at McKesson to let internal healthcare operations teams query large volumes of unstructured operational documents via natural language with source-backed answers, designed with HIPAA/FHIR compliance in mind. Demonstrated strong production engineering across hallucination mitigation, retrieval quality tuning, and latency/scalability optimization, using LangChain/LangGraph and Airflow plus rigorous evaluation/monitoring practices.”
Mid-level ML Engineer specializing in NLP and Generative AI
“Healthcare AI/ML engineer with Epic experience who built and deployed a HIPAA-compliant GPT-4 RAG clinical assistant over large medical document sets, emphasizing privacy controls and low-latency performance. Also automated end-to-end retraining and deployment of patient risk models using orchestration/CI-CD (Jenkins, SageMaker, MLflow), cutting deployment time from hours to minutes while improving reliability.”