Pre-screened and vetted.
Mid-level Generative AI Engineer specializing in LLMs, RAG, and multimodal generation
“Open-source JavaScript contributor focused on performance and maintainability in data visualization libraries—refactored legacy ES5 into modular ES6, added tests/docs, and delivered ~30% faster load times with positive community adoption. Also optimized a React dashboard (~40% load-time reduction) and took ownership in an ambiguous AI product initiative by setting milestones, standing up an initial ML pipeline, and shipping a prototype in ~6 weeks that became the basis for production.”
Mid-level AI/ML Engineer specializing in LLM agents, RAG, and ML systems
“At Inertia Systems, built a production LLM-powered ingestion pipeline that converts heterogeneous sources (PDF/JSON/IFC/SQL and financial tables) into standardized text and uses GraphRAG to construct a knowledge graph with verified dependency relationships. Also has hands-on HPC orchestration experience with SLURM, including creating a custom wrapper process manager to improve resource utilization under restrictive scheduling policies.”
“ML/GenAI engineer with recent CVS Health experience building a production RAG system over unstructured financial/research documents using LangChain, FAISS, and Pinecone, plus LoRA/PEFT fine-tuning of GPT/LLaMA for domain-aware summarization. Demonstrates strong applied MLOps and data engineering skills (Airflow/Prefect, Docker/Kubernetes, CI/CD, MLflow) and measurable impact (sub-second retrieval, ~40% better context retrieval, ~25% entity matching improvement).”
Mid-level Data Engineer specializing in cloud ETL/ELT and healthcare analytics
“Healthcare-focused data engineer/ML practitioner with experience at Lightbeam Health Solutions and Humana building production entity-resolution and semantic similarity pipelines across EMR, lab, and claims data. Uses NLP/ML (spaCy, scikit-learn, BioBERT/LightGBM) plus Snowflake/Airflow and vector search (Pinecone) to improve linkage accuracy (reported 90%) and semantic match quality (reported +12–15%), while reducing manual cleanup by 40%+.”
Mid-level Data Scientist & AI Engineer specializing in RAG, agentic AI, and production ML
“AI/data engineer who built a production LLM-powered schema drift detection system (LangChain/LangGraph) to catch semantic data changes before they break downstream analytics/ML. Deployed on AWS with Docker/S3 and implemented an LLM-as-a-judge evaluation framework to improve trust, reduce hallucinations, and control false positives/alert fatigue. Collaborated with non-technical risk/business analytics stakeholders at EY by delivering human-readable drift explanations that improved confidence in financial analytics dashboards.”
Intern Software Engineer specializing in backend, cloud data platforms, and microservices
“Full-stack engineer who shipped a group scheduling SaaS feature with live availability updates using Next.js App Router + TypeScript, owning production reliability after launch (auth debugging, monitoring, polling/backoff tuning). Has hands-on experience with Postgres schema/index design and query optimization (EXPLAIN ANALYZE) and building durable orchestrated backend workflows with retries and idempotency.”
Senior Data & Backend Engineer specializing in cloud data pipelines and LLM/RAG systems
“Data engineer with end-to-end ownership of large-scale retail and clinical data ingestion/processing on AWS, including real-time streaming and batch pipelines. Delivered measurable outcomes: 20M daily transactions processed, latency cut from 4 hours to 5 minutes, ~70% fewer failures, and 120+ pipelines running at 99.8% reliability with full audit compliance.”
Senior AI/ML Engineer specializing in Generative AI, RAG, and agentic systems
“GenAI/LLM ML engineer (currently at Webprobo) building an enterprise GenAI platform with document intelligence and automation on AWS and blockchain. Has hands-on experience with RAG, LLM evaluation tooling, and orchestrating production LLM workflows with Apache Airflow, plus deep exposure to reliability challenges in globally distributed/edge deployments. Also partnered with business/marketing stakeholders at a banking client to deliver an AI-driven customer retention insights solution.”
Senior Full-Stack & GenAI Engineer specializing in healthcare and financial services
“Built and deployed a production LLM-powered customer support assistant using a RAG backend in Python, focused on deflecting repetitive Tier-1 tickets and reducing resolution time. Demonstrates strong production engineering instincts around reliability (confidence scoring + human fallback), scalability/cost optimization (multi-stage pipelines), and workflow orchestration/observability (LangChain, custom DAGs, structured logging, step metrics).”
Senior Software Engineer specializing in AI-driven marketing and data platforms
“Backend/data engineer who builds production FastAPI microservices and AWS serverless/Glue pipelines for SMS analytics and marketing segmentation. Led a legacy batch modernization into modular services (FastAPI + Glue/Athena + ClickHouse) using shadow-mode parity checks, feature flags, and incremental rollout. Demonstrated measurable performance wins (12s to sub-second SQL; ~40% CPU reduction) and strong incident ownership with proactive schema-drift prevention.”
Mid-level AI/ML Engineer specializing in LLM, RAG/GraphRAG, and fraud analytics
“LLM/agent engineer who has deployed a production internal assistant to reduce employee inquiry resolution time while maintaining regulatory compliance. Experienced with RAG, hallucination risk triage, and graph-based orchestration (LangGraph) for enterprise/banking-style workflows, emphasizing schema-validated, citation-backed, tool-constrained agent designs and tight collaboration with non-technical business/compliance stakeholders.”
Mid-level Applied AI Engineer specializing in agentic LLM workflows
“Master’s-in-Data-Science candidate (UHV) with 4+ years in AI engineering building production LLM and multimodal systems. Designed an LLM-powered workflow automation platform using RAG over vector stores with guardrails (schema/output validation, fallbacks) and a rigorous evaluation/monitoring framework including drift tracking and shadow deployments. Experienced orchestrating large-scale vision-language pipelines with Airflow and Kubernetes (OCR, distributed training) and partnering with non-technical ops stakeholders to cut cycle time and reduce errors.”
Mid-level Data & AI Engineer specializing in healthcare data pipelines and MLOps
“Built and deployed a production LLM-powered clinical note summarization system used by care managers to speed review of 5–20 page unstructured medical records. Implemented safety-focused validation (prompt constraints, rule-based and section-level checks, human-in-the-loop) to reduce hallucinations while maintaining low latency and meeting privacy/regulatory constraints, integrating via APIs into existing clinical tools.”
Mid-level AI/ML Engineer specializing in Generative AI and NLP
“AI/LLM engineer with production experience building secure, scalable compliance-focused generative AI systems (GPT-3/4, BERT) including RAG over internal regulatory document bases. Has delivered end-to-end pipelines on AWS with PySpark/Airflow/Kubernetes/FastAPI, emphasizing privacy controls, monitoring, and iterative evaluation (A/B testing). Also partnered closely with bank compliance officers using prototypes to refine NLP summarization/classification and reduce document review time.”
Mid-level AI/ML Engineer specializing in healthcare ML and LLM/RAG systems
“AI/LLM engineer with recent production experience at UnitedHealth Group building an end-to-end RAG system over structured EMR data and unstructured clinical notes, including evidence retrieval, GPT/LLaMA-based reasoning, and a validation layer for reliability. Strong in orchestration (Kubeflow/Airflow/MLflow), prompt engineering for noisy healthcare text, and rigorous evaluation/monitoring with gold-standard benchmarking, plus close collaboration with clinical operations stakeholders.”
Mid-level AI/ML Engineer specializing in MLOps and LLM-powered applications
“AI/ML engineer with production experience building a RAG-based internal analytics assistant (Databricks + ADF ingestion, Pinecone vector store, LangChain orchestration) deployed via Docker on AWS SageMaker with CI/CD and MLflow. Strong focus on real-world constraints—latency/cost optimization (LoRA ~60% compute reduction), hallucination control with citation grounding, and enterprise security/governance. Previously at Intuit, delivered an interpretable churn prediction system (PySpark/Databricks, Airflow/Azure ML) that improved retention targeting ~12%.”
Mid-level AI/ML Engineer specializing in GenAI agents, RAG pipelines, and MLOps
“AI/ML engineer who built a production RAG-based internal document intelligence assistant (LangChain + Pinecone) to let employees query enterprise reports in natural language. Demonstrated hands-on pipeline orchestration with Apache Airflow and tackled real production issues like retrieval grounding and latency using tuning, caching, and token optimization, while partnering closely with non-technical business stakeholders through iterative demos.”
Intern Full-Stack Software Engineer specializing in AWS serverless and real-time web apps
“New-grad/early-career engineer who led high-stakes modernization of a field-operations platform from Firebase to AWS using an incremental/dual-write strategy, achieving zero downtime and ~30–32% infra cost reduction while improving scalability. Also built and productionized an AI-native code assistant (LangChain + Pinecone RAG) with measurable online metrics and safety guardrails, and has experience working directly with CEO/CTO/CPO and embedded with customer teams to ship enterprise features quickly.”
Mid-Level Software Engineer specializing in cloud, microservices, and AI/ML
“Backend/API engineer with ~4 years experience building production services in .NET Core/PostgreSQL/Redis/Docker and optimizing real-world latency issues (claims ~60% response-time improvement). Also built and owned an end-to-end RAG-based AI assistant using Python/FastAPI, OpenAI APIs, and Pinecone, plus agentic workflows with reliability guardrails (retries, confidence thresholds, monitoring). Currently pursuing a master’s degree and targeting a $150k base salary.”
Mid-level Data Engineer specializing in cloud lakehouse, streaming, and MLOps
“Data engineer at AT&T focused on large-scale telecom (5G/IoT) data platforms, owning end-to-end pipelines from Kafka/Azure ingestion through Databricks/Delta Lake transformations to serving analytics and ML. Has operated at very high volumes (~50+ TB/day) and delivered measurable performance gains (25–30% faster processing) plus improved reliability via Airflow monitoring, robust data quality checks, and resilient external data collection patterns (rate limiting, retries, dynamic schemas).”
Senior Data Scientist specializing in NLP, LLMs, and Computer Vision
“Applied NLP/ML engineer with experience at KeyBank and Novartis building production document intelligence and entity-resolution systems in finance and healthcare. Has delivered end-to-end pipelines (Airflow + AWS) using transformers (DistilBERT/Sentence-BERT), vector search (FAISS/Milvus/Pinecone), and human-in-the-loop labeling to achieve measurable gains (40%+ faster queries; up to 88% F1 and 93% precision/90% recall in entity linking).”
Senior AI/ML & Data Engineer specializing in Generative AI and RAG systems
“GenAI/RAG engineer who has deployed a production policy/regulatory search assistant for a financial client using LangChain + Vertex AI, FastAPI, Docker/Kubernetes, and Airflow-orchestrated data pipelines. Demonstrated measurable impact with 50–60% latency reduction and 70% fewer pipeline failures, plus KPI-driven grounding evaluation (90%+ target) and strong cross-functional collaboration with compliance/business teams.”
Senior AI/ML Engineer specializing in Python, RAG systems, and LLM fine-tuning
“Built and owned an end-to-end RAG-based AI support platform at Mechanize (FastAPI/LangChain/Pinecone/React) with rigorous evals and guardrails, driving 45% fewer support tickets and ~$280K annual savings. Also led a high-risk legacy modernization at Argo AI, incrementally extracting a monolithic Django backend using Strangler Fig + feature flags while supporting 10K+ concurrent users.”