Pre-screened and vetted.
Mid-level Machine Learning Engineer specializing in IoT, edge AI, and enterprise ML
“Built and productionized an LLM/RAG question-answering service over technical documentation, focusing on retrieval quality (reranking + IR metrics), latency, and scaling. Experienced orchestrating end-to-end ETL/ML workflows with Airflow/Prefect/AWS Step Functions and improving reliability via parallelism, retries, and shadow testing. Also delivered an explainable healthcare risk-flagging classifier with a stakeholder-friendly dashboard for a non-technical program manager.”
Senior Data Scientist specializing in ML, NLP, and production AI systems
“Machine learning/NLP engineer with deep Azure stack experience (Data Factory, Databricks/Spark, Delta Lake, Azure OpenAI, Azure AI Search) who built end-to-end production systems for semantic clustering, entity resolution, and hybrid search. Demonstrated measurable gains from embedding fine-tuning (~15% retrieval precision, ~10–12% nDCG@10) and designed scalable, quality-checked pipelines with MLOps best practices.”
Mid-level AI/ML Engineer specializing in NLP, LLMs, and RAG for banking and healthcare
“Deployed a real-time LLM-driven call center summarization and agent-assist platform at Fifth Third Bank, combining transformer models (BERT/GPT) with FastAPI inference on AKS and vector storage (ChromaDB/PostgreSQL). Emphasizes production-grade reliability (autoscaling, CI/CD, monitoring) and measurable evaluation (A/B testing), and translates model outputs into business-facing Power BI insights for call center leadership.”
Senior LLM Engineer specializing in Generative AI, RAG, and multimodal assistants
“GenAI/NLP engineer with experience building classification and summarization pipelines in PyTorch and deploying multimodal GPT-4-style workflows. Has integrated LLM applications across OpenAI, Azure OpenAI, and Amazon Bedrock, and uses LangChain/LlamaIndex/Semantic Kernel to orchestrate RAG and agent workflows with production-focused evaluation metrics like task success rate and groundedness.”
Mid-level AI/Machine Learning Engineer specializing in Generative AI, NLP, and MLOps
“Built a production LLM/RAG document analysis system for large financial documents (credit reports/PDFs) to help business analysts extract insights faster. Implemented end-to-end pipeline orchestration with LangChain, vector search (e.g., FAISS), and hallucination controls (context grounding, similarity thresholds, and no-answer fallback), delivered as a Dockerized Python API.”
Junior AI/ML Engineer specializing in deep learning and full-stack ML applications
“Built and operated a production-used RAG-based AI study planner (GPT-4 + FAISS) that handled 250+ concurrent users, with real-world reliability engineering (caching, fallbacks, schema validation, Redis state, monitoring). Also has healthcare data integration experience at Medinet Analytics, standardizing messy EHR/practice-management data with canonical schemas, idempotency hashing, and compliance-grade audit trails.”
Staff Full-Stack & DevOps Engineer specializing in cloud-native platforms and AI
“Backend/data engineer focused on production Python and AWS: built FastAPI REST services and a containerized ECS Fargate + Lambda architecture deployed via Terraform/CI-CD. Strong in data engineering (Glue/S3/Parquet/RDS) and operational reliability (CloudWatch/SNS, retries, schema-evolution handling), with experience modernizing legacy SAS reporting into Python microservices using feature flags and parity validation.”
Mid-level Full-Stack Developer specializing in AI/ML and cloud-native applications
“Full-stack/AI engineer who has shipped production systems spanning real-time analytics dashboards and an internal LLM-powered knowledge assistant. Experienced with RAG pipelines (embeddings/vector DB, semantic retrieval, query rewriting) plus evaluation loops and guardrails, and builds observable Kafka-based data pipelines monitored with Prometheus/Grafana.”
Mid-level Full-Stack Java Developer specializing in microservices and cloud-native systems
“Senior full-stack engineer with strong healthcare domain experience who has shipped an Azure OpenAI RAG-based patient medication support chatbot to production, driving ~10K queries/month and a reported 38% reduction in call center volume. Also builds polished real-time React/TypeScript pharmacy tooling and operates large-scale Python/Spark ETL pipelines (~12M records/day) with strong API design, observability, and cloud deployment experience across Azure/Kubernetes and AWS.”
Senior AI/ML Engineer specializing in financial risk, fraud detection, and GenAI analytics
“AI/ML engineer with experience at Northern Trust and Persistent Systems building production LLM + RAG systems for regulated financial use cases, including liquidity forecasting, anomaly detection, and credit scoring. Emphasizes compliance-first design with explainability (SHAP), traceability (MLflow), and hallucination controls (FAISS + citation-grounded prompting), and has delivered drift-triggered retraining pipelines using Airflow and Kubernetes while translating model outputs into business-ready marketing segments.”
Mid-Level Backend Software Engineer specializing in FinTech and distributed systems
“Backend engineer who built an AI RAG quoting system for the fastener industry, reducing quote turnaround from weeks to ~30 minutes and raising retrieval accuracy to ~90% by solving a semantic-collision issue with a parent-document retrieval design. Strong in production AWS integrations (Cognito auth, S3 pre-signed uploads), performance optimization (multithreading/out-of-core), and real-time streaming (Kafka/Spark Kappa architecture achieving sub-second latency), plus Kubernetes logging and GitHub Actions CI/CD to ECR.”
Mid-level AI Developer & Machine Learning Engineer specializing in LLM and MLOps systems
“Built and deployed an enterprise RAG application at Centene to help clinical teams retrieve insights from large internal policy document sets, cutting manual research by 30–40%. Implemented custom domain-adapted embeddings (SageMaker + BERT transfer learning) and hybrid retrieval (BM25 + Pinecone) to drive a 22% relevance lift, and ran the system in production on AWS EKS with CI/CD, MLflow, and Prometheus monitoring (99% uptime, ~40% latency reduction).”
Mid-level AI/ML & Data Engineer specializing in MLOps and cloud data pipelines
“AI/ML engineer (Merkle) with hands-on experience deploying RAG-based LLM applications and real-time recommendation engines into production. Strong in cloud/on-prem architectures, GPU autoscaling, caching, and network optimization—delivered measurable latency reductions (40–70%) and improved retrieval relevance by systematically benchmarking chunking/embedding configurations and validating pipelines via CI/CD.”
Mid-Level Forward Deployed AI Engineer specializing in RAG systems and backend microservices
“LLM solutions practitioner with SOC/alert-triage experience who takes LLM prototypes to production using RAG (Pinecone), FastAPI services, guardrails, CI/CD, monitoring, and robust fallback logic. Known for rapid real-time debugging of embedding/vector and agent workflow issues, and for driving adoption through code-first workshops and sales-aligned custom demos with measurable improvements (35% faster triage; 40% increase in correct tool usage).”
Junior Full-Stack Software Engineer specializing in cloud web apps and authentication
“Full-stack engineer with Deloitte and CrowdDoing experience shipping production web platforms on AWS (EC2/RDS/S3/Fargate) using React/TypeScript and Node/Express/PostgreSQL. Built customer-facing authentication/SSO flows (OAuth2 + JWT) and state-specific US privacy consent workflows, and also delivered a Python/Flask LLM-based finance document parser chatbot with vector DB integration and latency optimizations.”
Mid-level Software Engineer specializing in automation, AI agents, and full-stack web development
“Full-stack engineer who built and shipped an AI-powered internal knowledge search system for APL Services, including document ingestion into a vector database, a Python backend, and a React/TypeScript chat-style UI with source citations for trust. Improved production reliability by migrating from Streamlit Cloud to GCP with containerization and scaling controls to eliminate cold-start friction; also co-led a Mensa chapter website redesign as Digital Communications Committee co-chair.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and MLOps
“Data professional with ~4 years of experience, most recently at AIG (insurance), building ML/NLP systems for fraud detection and policy automation using transformers, CNNs, and clustering/anomaly detection. Also developed a RAG-based knowledge retrieval system, iterating across embedding models and moving to production based on precision and latency SLAs, then containerizing and deploying with SageMaker and CI/CD.”
Mid-level Data Scientist/Data Analyst specializing in ML, BI dashboards, and ETL pipelines
“Data/ML practitioner with experience at Humana and Hexaware, focused on turning messy, semi-structured datasets into production-ready pipelines. Built an age-prediction model from book ratings using heavy feature engineering and multiple regression models, and has hands-on entity resolution (deterministic + fuzzy matching) plus embeddings/vector DB approaches for linking and search relevance.”
Senior Full-Stack AI Engineer specializing in LLM/RAG agentic systems
“Built and deployed JobMatcher AI, an LLM-driven workflow automation product for job seekers that extracts requirements from job descriptions, matches to user skills, and generates tailored outreach. Demonstrated strong production engineering by cutting per-run cost ~70%, improving reliability with retries/backoff/fallbacks, and reducing hallucinations via schema validation and templating; also orchestrated the system with LangGraph plus Docker Compose across API, vector DB, and workers.”
Mid-level AI Engineer specializing in agentic LLM systems and RAG platforms
“Built and shipped Serrano AI, a multi-tenant SaaS conversational AI platform that automates Odoo ERP workflows and lets ops/finance/supply-chain teams query ERP data in natural language. Implemented a multi-agent architecture (LangChain/LangGraph/CrewAI) with hybrid RAG over ERP schemas, deployed on Heroku/Vercel with production observability, cutting reporting time by ~80% while addressing hallucinations, latency, and schema complexity.”
Staff/Lead Software Engineer specializing in distributed data and ML platforms
“Defense-domain AI engineer who built a production ReAct-style RAG system for military training data/material generation, scaling to ~1000 users and cutting generation time by 50%. Also has experience designing GPU-cluster parallel computation with PyTorch and handling production incidents involving database performance and schema design.”
Mid-Level Data/ML Engineer specializing in Generative AI and cloud data platforms
“Built and productionized an LLM-based financial document analysis system using a RAG pipeline, including robust ingestion/chunking/embedding workflows, vector DB retrieval, and an AWS-deployed FastAPI service containerized with Docker. Demonstrates strong applied expertise in improving retrieval quality and latency at scale, plus hands-on experience debugging agentic/LLM workflows with monitoring and trace-based analysis while supporting demos and customer-facing adoption.”
Junior AI and Backend Engineer specializing in LLM systems
“AI/LLM engineer who has shipped production RAG copilots and multi-agent workflows, including a real-time Llama3 (Ollama) copilot backend handling 12k+ concurrent queries at 99.9% uptime. Deep on orchestration (Langflow/Airflow/Kubernetes), reliability evaluation (hallucination detection, p95 latency, token cost), and monitoring (Prometheus/Grafana), with demonstrated stakeholder-facing analytics delivery via Tableau.”
Senior Machine Learning Engineer specializing in LLMs, NLP, and computer vision
“Built and owned production GenAI systems for both infrastructure automation and customer support. Most notably, they created a self-healing multi-cloud incident response system that automated 65% of tier-1 alerts and reduced application crashes by 75%, and also shipped a hybrid RAG support triage agent that automated 60% of tier-1 inquiries with human escalation guardrails.”