Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
Entry-Level Software Engineer specializing in ML and backend systems
“Built and deployed a production LLM-based real-time stance detection system for social media, fine-tuning LLaMA 3.1 on A100s with DeepSpeed ZeRO/FSDP and iteratively refining data to handle sarcasm and context-dependent meaning. Also has Kubernetes operations experience (Kafka/Logstash/Elasticsearch observability pipeline) and delivered an OCR automation project during a Worley India internship that saved 20+ hours/week for on-site energy safety stakeholders.”
Mid-level AI/ML Engineer specializing in MLOps, computer vision, and NLP
“GenAI/ML engineer from Lucid Motors who built and productionized an LLM-powered RAG diagnostic assistant for manufacturing and maintenance teams, deployed on AWS with Docker/Kubernetes and MLflow. Demonstrates end-to-end ownership from retrieval/prompt design to scalability, monitoring, and workflow integration via APIs, plus production ML pipeline orchestration with Kubeflow (Spark/Kafka + TensorFlow) for predictive maintenance use cases.”
“Built and deployed a production LLM-powered RAG assistant for semiconductor manufacturing failure analysis, reducing engineer triage effort by grounding outputs in retrieved evidence and gating responses with SPC + ML signals (LSTM anomaly scores, XGBoost probabilities). Experienced with LangChain/LangGraph to ship reliable, observable multi-step agents with branching/fallback logic, and evaluates impact using both technical metrics and business KPIs like mean time to triage and downtime reduction.”
Mid-level Generative AI Engineer specializing in enterprise RAG and multimodal NLP
“Built and deployed a production LLM/RAG chatbot at Wells Fargo for securely querying regulated financial and compliance documents, emphasizing low hallucination rates, explainability, and strict governance. Experienced with LangChain multi-agent orchestration plus Airflow/Prefect pipelines for ingestion, embeddings, evaluation, and retraining, and partnered closely with compliance/operations to drive adoption through demos and feedback-driven retrieval rules.”
Mid-level Software Engineer specializing in AI and full-stack healthcare platforms
“Built and deployed a RAG-based clinical knowledge assistant at GE Healthcare to help clinicians query large volumes of messy, unstructured clinical documents with grounded, cited answers. Hands-on across the full stack (OCR/ETL, de-identification for PHI, Azure OpenAI embeddings, Cosmos DB indexing, FastAPI/Django) with production monitoring via LangSmith and performance tuning through batching and index optimization.”
Junior AI/Backend Software Engineer specializing in ML and scalable systems
“Backend engineer with strong AWS/CI/CD experience (multi-repo deployments, Lambda + core app, immutable ECR and image promotion) and a published master’s thesis building an ML framework for Solar PV energy prediction and CO2 reduction impact modeling using ensemble and meta-learning approaches benchmarked against SAM.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production LLM/RAG knowledge assistant integrating internal docs, wikis, and ticket histories to reduce tribal-knowledge dependency and repetitive questions. Emphasizes reliability via grounding + a validation layer, and achieved major latency gains (>50%) through vector index optimization, caching, quantization, and selective re-validation. Comfortable orchestrating end-to-end LLM/data workflows with Airflow, Prefect, and Dagster, including monitoring and alerting.”
Mid-level Software Engineer specializing in cloud infrastructure and distributed systems
“Backend/platform engineer who built an AI RAG system on FastAPI/Postgres/AWS with 10+ microservices, vector search optimization (ANN + two-stage re-ranking), and GitOps-driven CI/CD that cut deploy time from hours to minutes. Also deployed Java identity services on Kubernetes at TSMC for 200K+ users using ArgoCD/Azure Pipelines, and built a reliable real-time IoT pipeline (MQTT/Node/MongoDB) with strong consistency controls.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMOps, and MLOps
“Built and deployed an AWS-based LLM/RAG ticket triage and knowledge retrieval system (Pinecone/FAISS + Step Functions + MLflow) that cut support resolution time by 20%. Demonstrates strong production focus on hallucination reduction, PII security, and low-latency orchestration, with measurable evaluation improvements (e.g., ~25% grounding accuracy gain via re-ranking) and proven collaboration with support operations stakeholders.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Red Hat ML/LLM engineer who designed and deployed a production LLM-powered customer support automation system using RAG, improving latency by 30% via PEFT and vector search optimization. Built security and governance into retrieval (access-level filtering, encrypted Pinecone/ChromaDB) and delivered SHAP-based explainability via a dashboard for non-technical stakeholders. Experienced orchestrating distributed ML/RAG pipelines across AWS SageMaker and OpenShift with Airflow/Prefect, plus multi-agent workflows using CrewAI and LangGraph.”
Mid-level AI/ML Engineer specializing in GenAI, LLMs, RAG, and MLOps
“Built and deployed a production LLM-powered RAG document intelligence/Q&A system for healthcare prior authorization, reducing manual medical document review time and improving decision efficiency. Strong in end-to-end LLM application engineering (LangChain/LangGraph), retrieval quality improvements (hybrid search, embedding tuning, chunking strategies), and rigorous evaluation/monitoring for reliability.”
Mid-level Full-Stack Developer specializing in cloud-native microservices and event-driven systems
“Software engineer with experience at Molina Healthcare and Target, owning production features end-to-end across backend, data pipelines, and UI. Built an event-driven claims validation system (Python/Java/Spring Boot/Kafka) with strong observability, and shipped embeddings-based semantic product search with evaluation loops (CTR/top-k + human review) and guardrails like keyword-search fallback.”
Junior Full-Stack Software Engineer specializing in React and AI-powered applications
“Full-stack/AI-focused builder who shipped a production Career Advisor app using LLMs + RAG + vector DB (React/Node/MongoDB/Claude API) and grew it to 2000+ users, handling real deployment issues and CI/CD on Vercel/Render. Also developing an AI-powered iOS “3D World Explorer” (text-to-3D) and has cloud experience across Azure and AWS (S3/SageMaker/EC2).”
Mid-level AI/ML Engineer specializing in risk, fraud detection, and Generative AI
“Built and deployed an LLM-powered RAG document intelligence/search platform for banking risk & compliance teams, emphasizing sensitive-data handling, traceability, and conservative fallback logic to minimize hallucinations; deployed via Docker/REST on AWS and cut manual review effort by 35%. Also partnered with TD Bank marketing to deliver an AI customer segmentation solution that improved targeted campaign engagement by 18%.”
Mid-Level Software Engineer specializing in distributed systems and cloud-native backends
“AI/LLM engineer with production experience at Charles Schwab building a RAG-based assistant to help 5,000+ reps answer complex financial policy questions. Implemented a multi-layer anti-hallucination approach (GNN-driven ontology/graph retrieval + citation-only answers) and compliance-focused guardrails (Azure AI Content Safety) in partnership with audit/compliance stakeholders.”
Mid-level Data Scientist specializing in predictive and generative AI
“AI/ML engineer with production LLM experience in regulated financial services (J.P. Morgan Chase), building a customer response engine to automate first-contact resolution while addressing privacy, bias, compliance, and scale. Strong MLOps/orchestration background (Airflow, Docker/Kubernetes, AWS Step Functions, Azure ML/SageMaker) plus proven ability to integrate with legacy systems and drive stakeholder adoption through dashboards, auditability, and training.”
Mid-level AI/ML Engineer specializing in Generative AI and Conversational AI
“GenAI Engineer at Infosys who built and deployed a production multi-agent RAG system for a top-tier bank, scaling to ~50,000 queries/day with 99.9% uptime. Drove measurable gains (45% accuracy improvement, 30% API cost reduction) through open-source LLM fine-tuning, Pinecone indexing/retrieval optimization, and AWS-based MLOps/monitoring, and has experience enabling adoption via developer workshops and customer-facing collaboration.”
Principal Software Engineer specializing in AI/ML and cloud-native backend systems
“McKinsey data/ML practitioner who led production deployment of an entity resolution + semantic search platform for unstructured finance and healthcare data, integrating with legacy systems under HIPAA constraints. Deep hands-on stack across transformers (spaCy/HF BERT), embeddings + FAISS, and production MLOps/workflow tooling (Airflow, Docker, CI/CD, Prometheus/Grafana), with reported gains of +30% decision speed and +25% search relevance.”
Intern AI Engineer specializing in LLM agents, RAG, and applied biostatistics
“Siemens AI engineer who shipped production multi-agent LLM systems across cybersecurity and sustainability, including a vulnerability automation agent that cut manual work 70%. Deep in orchestration (LangGraph supervisor-worker state machines), reliability engineering (async fault tolerance, retries, spike handling), and rigorous evaluation (offline benchmarks, LLM-as-a-Judge improving label agreement 28.9%) with measurable production guardrails.”
Mid-Level Software Engineer specializing in LLM-powered developer tools
“Built and owned "Cortex," an AI agent that helps users understand large GitHub repositories by mapping architecture and relationships between files/folders in minutes. Implemented an agentic, multi-stage prompt decomposition approach and validated it across open-source repos, while also doing legacy service modernization work involving dependency upgrades and refactors.”
Mid-level Generative AI Engineer specializing in decision intelligence and RAG for regulated enterprises
“Healthcare GenAI engineer who built a HIPAA-compliant, auditable RAG-based claims decision support system at Molina Healthcare, processing 3M claims and delivering major impact (48% faster manual reviews, 43% higher decision accuracy). Deep hands-on experience with LangChain orchestration, vector search (ChromaDB/FAISS), embedding fine-tuning, and safety controls (confidence scoring, rule validation, human-in-the-loop escalation) for clinical workflows.”
Mid-level Machine Learning Engineer specializing in Generative AI and RAG systems
“GenAI/LLM engineer with production deployments in both fintech and retail: built an AI-powered mortgage document analysis/automated underwriting pipeline at Fannie Mae (OCR + custom LLM) cutting underwriting review from 3–4 hours to under an hour with privacy-by-design controls. Also helped build Sephora’s GenAI product advisory bot using LangChain-orchestrated RAG (Azure GPT-4, Azure AI Search, MySQL HeatWave vector search), focusing on grounding, evaluation, and compliance-aware architecture choices.”
Junior Full-Stack Software Engineer specializing in AI data systems
“Full-stack engineer with strong DevOps/AWS production experience who builds and operates multi-agent AI systems end-to-end (Streamlit/Python through Docker/Kubernetes and ECS/Fargate). Has delivered measurable outcomes: sub-2s latency and ~92% routing accuracy for an AI wellness assistant, shipped an AI-for-BI prototype in under 6 weeks cutting analysis time ~40%, and improved pipeline iteration speed ~35% via modularization and CI/regression checks.”