Pre-screened and vetted.
Mid-level Data Scientist specializing in ML, NLP, and Generative AI
“GenAI/ML engineer with production experience at Cognizant and Ally Financial, building end-to-end LLM/RAG systems and ML pipelines. Delivered a domain chatbot trained from 90k tickets and 45k docs, improving intent accuracy (65%→83%), scaling to 800+ concurrent users with 99.2% uptime and sub-150ms latency, and driving +14% customer satisfaction. Strong in Azure ML + DevOps CI/CD, Dockerized deployments, and explainable/PII-safe modeling using SHAP/LIME to satisfy stakeholder trust and GDPR needs.”
Mid-level Machine Learning & Full-Stack Engineer specializing in GenAI platforms
“LLM/agent builder who has shipped production AI systems in the wellness space, including an LLM-powered food tracking product used by 5000+ users and a voice/call-routing onboarding workflow using LangGraph/LangChain with LiveKit and Twilio. Strong focus on practical reliability work: latency reduction, retrieval/embedding tuning, and CI-driven evaluation with simulations and metrics.”
Mid-level Full-Stack Developer specializing in FinTech platforms and cloud-native microservices
“Backend engineer focused on AI-enabled systems, having built a production-style RAG pipeline (vector search + LLM) exposed via Python/Flask endpoints with strong observability and hallucination-reduction techniques. Demonstrates deep performance work in PostgreSQL/SQLAlchemy (5x faster analytics queries) and high-throughput optimization using Celery + Redis (800ms to 120ms latency, 3x throughput), plus schema-per-tenant multi-tenancy with tenant-aware middleware and logging.”
Mid-level AI/ML Engineer specializing in FinTech risk, fraud detection, and GenAI/RAG systems
“Built and productionized Azure-based LLM/RAG systems for regulatory/compliance use cases, including automating analyst research and compliance report generation across large unstructured document sets. Demonstrates strong practical depth in hallucination mitigation, hybrid retrieval tuning (BM25 + embeddings), and production MLOps (Databricks, Cognitive Search, AKS, Airflow/MLflow), plus proven ability to deliver auditable, explainable solutions with non-technical compliance teams.”
Mid-level Data Scientist & Generative AI Engineer specializing in LLMs and RAG
“Built production LLM + hybrid RAG and multi-agent orchestration systems at Wells Fargo to automate complaint document/audio transcript understanding and categorization, addressing vocabulary drift via embedding + vector index updates instead of frequent retraining. Strong in LLM workflow reliability (testing/benchmarks/observability) and stakeholder-facing delivery with explainability (citations/SHAP-style justifications) and Tableau dashboards.”
Mid-level Full-Stack Software Engineer specializing in enterprise web apps and real-time dashboards
“Backend/full-stack engineer from Foxconn Industrial Internet who led development of a production TypeScript/Node.js facility monitoring platform delivering near real-time manufacturing metrics (e.g., downtime and OEE) using MySQL + InfluxDB and a React dashboard. Demonstrates strong production operations mindset with queue-based workers, idempotency/DLQ patterns, structured observability, and automated Docker + GitLab CI/CD deployments.”
Mid-level Full-Stack Software Engineer specializing in microservices and scalable backend systems
“Backend/microservices engineer (Java/Spring Boot, Kafka, Angular microfrontends) with Teradata experience building distributed analytics/query routing platforms and delivering 20–30% latency reductions through event-driven redesign and reliability hardening. Also built and shipped an end-to-end multimodal medical imaging AI feature (LLaVA/Mistral 7B + LoRA) with production guardrails like confidence-based human review, drift monitoring, and audit logs.”
Principal AI/ML Leader specializing in Generative AI, MLOps, and NLP
“Founding member of Tausight, building AI systems to detect and protect PHI for healthcare organizations; helped take the company through post–Series A funding and exited after ~6 years. Drove a strategic collaboration with Intel’s OpenVINO team—becoming the first to deploy it in a real production system and improving model performance by ~30% on customer Intel-CPU machines.”
Mid-level Generative AI Engineer specializing in LLMs and RAG systems
“Built and shipped a production RAG-based enterprise knowledge assistant to replace slow/inaccurate search across millions of documents, using LangChain orchestration with GPT-4/LLaMA and vector databases. Strong focus on production constraints—latency, hallucination control, and cost—using hybrid retrieval, guardrails, LLM-as-judge validation, and model routing, and has experience translating non-technical stakeholder pain points into measurable outcomes.”
Mid-level AI/ML Engineer specializing in Generative AI and production ML systems
“Built and deployed a production SecureAIChatBot (RAG-based) for secure internal information retrieval, using embeddings/vector search, GPT models, monitoring, and safety filters. Focused on real-world production challenges like latency and output consistency, applying caching, retrieval scoping, smaller models, and controlled prompting, and used LangChain to orchestrate the end-to-end workflow.”
Senior Business Analytics Analyst specializing in product and customer analytics
“Darwinbox team member who supported talent/recruiting operations while also driving product improvements across HR modules (recruitment, onboarding, payroll, performance). Led a small team (5–6) and implemented discovery-driven configuration and BI reporting (Power BI/Tableau/Confluence), including a reported 30% reduction in recruitment configuration issues and real-time funnel reporting to support fast hiring.”
Mid-level Data Scientist specializing in Generative AI, NLP, and MLOps
“Built and deployed an LLM-powered claims-document summarization system (insurance domain) that cut agent review time from 4–5 minutes to under 2 minutes and saved 1,200+ hours per quarter. Hands-on across orchestration and production infrastructure (Airflow retraining DAGs, Kubernetes, SageMaker endpoints, FastAPI) and recent RAG workflows using n8n + Pinecone, with a strong focus on reliability, cost, and explainability for non-technical stakeholders.”
Mid-level AI/ML Engineer specializing in cloud data engineering and GenAI
“AI/LLM engineer with production experience in legal tech: built a GPT-4 + LangChain RAG summarization system at Govpanel that reduced legal case-file review time by 50%+. Previously at LexisNexis, orchestrated end-to-end Airflow data/AI pipelines processing 5M+ legal documents daily, improving ETL runtime by 35% with robust validation, monitoring, and SLAs.”
Mid-level Data Scientist specializing in Generative AI and NLP for financial risk
“Built and shipped production generative AI/RAG assistants in regulated financial contexts (S&P Global), automating compliance-oriented Q&A over earnings reports/filings with grounded answers and citations. Experienced across the full stack—AWS-based ingestion (PySpark/Glue), vector retrieval + LangChain agents, GPT-4/Claude model selection, and production reliability (monitoring, caching, retries) plus rigorous evaluation and regression testing.”
Mid-level AI/ML Engineer specializing in deep learning, MLOps, and LLM applications
“Built and deployed production LLM assistants for internal Q&A and customer-feedback summarization, emphasizing reliability (RAG, prompt tuning, validation/whitelisting) and privacy safeguards. Improved adoption by adding explainable outputs and a user feedback mechanism, and has hands-on orchestration experience with Aflow and Azure Logic Apps.”
Junior Software Engineer specializing in AI/ML and Full-Stack Development
“Built production LLM tooling focused on reproducibility and verification by enforcing JSON schemas and using multi-step checks with tools like Firecrawl and Perplexity. Also implemented the containerized infrastructure layer for a 9-agent app on K3s, dealing with rolling updates and uptime, and has experience advising a non-technical builder on search grounding and LLM data-flow design.”
Mid-Level Full-Stack/Backend Engineer specializing in AWS, APIs, and GenAI systems
“Backend engineer who built the core backend for Air Kitchens’ discovery/booking platform on AWS (Node + Python, DynamoDB, SQS/Lambda), optimizing for fast user-facing APIs and scalable async workflows. Introduced an AI matching service with a deterministic pre-filter + LLM ranking approach to balance latency vs quality, and has hands-on experience with production security (JWT/RBAC/RLS), CI/CD, and blue-green, staged migrations from Django to modular services.”
Junior Machine Learning Engineer specializing in Generative AI and analytics automation
“AI/LLM engineer who built a production intelligent support system using RAG over a vectorized documentation library, addressing real-world issues like lost-in-the-middle context failures and doc freshness via automated GitHub-driven re-embedding pipelines. Emphasizes rigorous agent evaluation (component/E2E/ops) and prefers lightweight, decoupled workflow automation using message brokers (Redis/RabbitMQ) over heavyweight orchestration frameworks.”
Mid-level GenAI Engineer specializing in LLM fine-tuning, RAG, and MLOps
“Healthcare-focused LLM engineer who deployed a production triage and clinical knowledge retrieval assistant using RAG and LangGraph-orchestrated multi-agent workflows. Emphasizes clinical safety and compliance with robust hallucination controls, HIPAA/PHI protections (tokenization, encryption, audit logging, zero-retention), and human-in-the-loop escalation; reports a 75% latency reduction in a healthcare agent system.”
“Built a production multi-agent orchestration platform to automate healthcare claims and HR workflows, combining LangChain/CrewAI/AutoGPT with RAG (FAISS/Pinecone) and fine-tuned open-source LLMs (LLaMA/Mistral/Falcon) in private Azure ML environments to meet HIPAA requirements. Emphasizes rigorous agent evaluation/observability (trajectory eval, adversarial testing, LLM-as-judge, drift monitoring) and reports measurable outcomes including 35% faster claims processing and 40% fewer chatbot errors.”
Senior AI/ML Engineer specializing in Generative AI, LLMs, and MLOps
“Telecom (Verizon) AI/ML practitioner who built a production multimodal system that ingests messy customer issue reports (calls, chats, emails, screenshots, videos) and turns them into confidence-scored incident summaries with reproducible steps and evidence links. Also built KPI/alarm-to-ticket correlation to rank likely root-cause domains (RAN/Core/Transport), cutting triage from hours to minutes and improving MTTR.”
Mid-level AI/ML Engineer specializing in Generative AI, LLMs, and NLP
“AI/ML engineer with forensic analytics and healthcare claims experience (Optum), building production LLM/RAG systems to surface context-driven fraud patterns from unstructured claim notes and explain risk to investigators. Strong in large-scale retrieval performance tuning, legacy API integration with reliability patterns (SQS, circuit breakers), and MLOps orchestration on Airflow/Kubernetes with rigorous testing, monitoring, and stakeholder-friendly interpretability.”
Mid-level Software Engineer specializing in full-stack and cloud-native microservices
“Backend engineer who built a Python/Flask system for high-volume healthcare claims processing, using PostgreSQL as the source of truth and RabbitMQ workers for scalable async processing. Experienced in SQLAlchemy/Postgres performance tuning, multi-tenant data isolation (including Postgres RLS), and integrating/versioning ML model services (scikit-learn/PyTorch/Hugging Face) with controlled rollouts. Drove measurable performance gains by batching background jobs and adding Redis caching (40% less workload; response times cut from ~10s to 2–3s).”
Mid-level AI/ML Engineer specializing in Generative AI and data engineering
“IBM engineer who built and deployed a production RAG-based LLM assistant using LangChain/FAISS with a fine-tuned LLaMA model, served via FastAPI microservices on Kubernetes, achieving 99%+ uptime. Demonstrates strong practical expertise in reducing hallucinations (semantic chunking + metadata-driven retrieval) and managing latency, plus mature MLOps practices (Airflow/dbt pipelines, MLflow tracking, monitoring, A/B and shadow deployments) and effective collaboration with non-technical stakeholders.”