Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“Built an LLM-powered academic research assistant for a professor (LangChain + OpenAI + arXiv) focused on synthesizing papers quickly, with emphasis on reliability (ReAct prompting, citation verification) and cost control (caching). Has production MLOps/orchestration experience at Cisco and HCL Tech using Kubernetes, plus MLflow and GitHub Actions for lifecycle management and CI/CD.”
Mid-level Data Scientist & Generative AI Engineer specializing in LLMs and RAG
“ML/NLP practitioner who built a retrieval-augmented generation (RAG) system for large financial and operational document sets using Sentence-Transformers (all-mpnet-base-v2) and a vector DB (e.g., Pinecone), with a strong focus on retrieval evaluation and chunking strategy optimization. Experienced in entity resolution (rules + embedding similarity with type-specific thresholds) and in productionizing scalable Python data workflows using Airflow/Dagster and Spark.”
Mid-level AI/ML Engineer specializing in NLP, LLMs, and RAG for finance and healthcare
“Built an AI lending assistant (RAG + DeBERTa) used by credit analysts to retrieve policies and past loan decisions, tackling real production issues like hallucinations, document quality, and sub-second latency. Deployed a modular, Dockerized AWS architecture (ECS/EMR + load balancer) with load testing, caching/precomputed embeddings, and CloudWatch monitoring, and used Airflow to automate scheduled data/embedding/vector DB refresh pipelines with retries and alerts.”
Mid-level Data Scientist specializing in Generative AI and multimodal systems
“Recent J&J intern who built a conversational RAG agent and led a shift from a monolithic model to a modular RAG workflow, cutting response time from several days to under a second by tackling data fragmentation, context retention, and embedding/latency optimization. Also worked on a large (7B-parameter) multimodal VQA pipeline for healthcare research and stays current via NeurIPS/ICLR and open-source contributions.”
Junior AI/ML Engineer specializing in RAG systems and cloud-native MLOps
“Built and shipped a production LLM-powered RAG system at Upstart enabling natural-language search across 50k+ scattered internal technical docs. Delivered sub-300ms p95 latency for ~50 active users with strong hallucination safeguards (retrieval-first, thresholds, citations) plus robust testing/monitoring and cost controls (prompt caching cutting API spend ~20%).”
Senior AI/ML & Full-Stack Engineer specializing in GenAI, RAG, and MLOps platforms
“Backend/data platform engineer who owned end-to-end production services for a fleet analytics/GenAI platform, spanning FastAPI microservices on Kubernetes and AWS (EKS + Lambda) event-driven workloads. Strong in reliability/observability (OpenTelemetry, circuit breakers, idempotency), data pipelines (Glue/Airflow/Snowflake), and measurable performance/cost wins (SQL 10s to <800ms P95; ~30% compute cost reduction).”
Mid-level Machine Learning Engineer specializing in Generative AI and LLM applications
“GenAI engineer who has deployed production LLM/RAG chatbots for internal document search, focusing on reliability (hallucination reduction via prompt guardrails + retrieval filtering) and performance (latency improvements via caching). Experienced with LangChain/LangGraph orchestration for multi-step agent workflows and iterates using monitoring/logs and benchmark-driven evaluation while partnering closely with product and business teams.”
Entry-Level AI/ML Engineer specializing in LLM apps, RAG pipelines, and production ML systems
“AI/LLM practitioner at iFrog Marketing Solutions who drove a RAG chatbot from prototype to production in a legacy, AI-resistant environment by validating customer needs and building a business case. Implemented production-grade LLM practices (CI/CD eval gating, rollbacks, prompt/context engineering) and led internal workshops to bring non-AI-native developers up to speed while partnering with sales on tailored demos to drive adoption.”
Mid-Level Software Engineer specializing in Cloud, DevOps, and MLOps
“Built and productionized a recommendation system from notebook prototype into a low-latency, scalable Cloud Run service using Docker, FastAPI, Terraform, CI/CD (GitHub Actions), and MLOps tooling (Vertex AI, MLflow). Experienced diagnosing real-time workflow issues using structured logging/ELK and GCP metrics, including resolving intermittent 504s by fixing unbounded SQL and adding caching. Also partners with sales/customer teams (Wasabi) to deliver tailored demos, troubleshoot, and drive onboarding/adoption.”
Mid-level AI/ML Engineer specializing in NLP and conversational AI
“ML/NLP engineer focused on real-time IT ops analytics, building a predictive maintenance/anomaly detection platform end-to-end (multi-source ETL, streaming, modeling, and production deployment on GCP/Vertex AI). Uses deep learning (LSTMs, autoencoders/VAEs) plus embeddings (SentenceBERT) and vector search to improve incident correlation and search, citing ~40% reduction in duplicate alert noise.”
“Senior data scientist with ~5 years’ experience building production ML/NLP systems in finance (Wells Fargo) and deep learning for sensor analytics in connected vehicles (Medtronic). Has delivered end-to-end platforms combining time-series forecasting with transformer-based NLP, including automated drift monitoring/retraining (MLflow + Airflow) and standardized Docker/CI/CD deployments; achieved a reported 22% precision improvement after domain fine-tuning.”
Mid-level Data Scientist specializing in real-time fraud detection and MLOps
“ML/NLP engineer with experience at Charles Schwab building an NLP + graph (Neo4j) entity-resolution system to unify fragmented user/device/transaction data and improve downstream model quality and analyst querying. Has applied embeddings (SentenceTransformers + FAISS) with domain fine-tuning to boost hard-case matching recall by ~12% while maintaining precision, and has a track record of hardening scalable Python/Spark pipelines and productionizing fraud models via A/B tests and shadow-mode monitoring.”
Intern Full-Stack Software Engineer specializing in AI/ML and cloud
“Built a Python-based geospatial machine learning backend for PFAS contamination risk mapping, including reproducible feature pipelines, ensemble modeling, and a FastAPI layer for visualization/analysis. Emphasizes data integrity and robustness (CRS/coverage checks, fail-fast validation) and has led safe backend refactors using feature flags, idempotent backfills, and Postgres RLS for secure, queryable results delivery.”
Principal AI/ML Leader specializing in Generative AI, MLOps, and NLP
“Founding member of Tausight, building AI systems to detect and protect PHI for healthcare organizations; helped take the company through post–Series A funding and exited after ~6 years. Drove a strategic collaboration with Intel’s OpenVINO team—becoming the first to deploy it in a real production system and improving model performance by ~30% on customer Intel-CPU machines.”
Mid-level AI/ML Engineer specializing in MLOps and LLM applications
“BNY Mellon engineer who has built and operated production AI systems end-to-end: a LangChain/Pinecone RAG platform scaled via FastAPI + Kubernetes to 1000 RPM with 99.9% uptime, supported by monitoring and data-drift detection. Also deep in data/infra orchestration (Airflow, Dagster, Terraform on AWS/EMR/EC2), processing 500GB+ daily and delivering measurable reliability and performance gains, plus strong compliance-facing model explainability using SHAP and Tableau.”
Mid-level AI Engineer specializing in GenAI, LLM integration, and RAG pipelines
“Built and led deployment of an autonomous, self-correcting multi-agent knowledge retrieval and validation system at HCA Healthcare to reduce heavy manual research/validation in clinical/compliance documentation. Deeply focused on production reliability and cost—used LangGraph StateGraph orchestration plus ONNX/CUDA/quantization to cut GPU costs by 25%, and partnered with the Compliance VP using real-time contradiction-rate dashboards to hit a 40% automation goal without compromising compliance.”
Mid-level Data Scientist specializing in Generative AI and NLP for financial risk
“Built and shipped production generative AI/RAG assistants in regulated financial contexts (S&P Global), automating compliance-oriented Q&A over earnings reports/filings with grounded answers and citations. Experienced across the full stack—AWS-based ingestion (PySpark/Glue), vector retrieval + LangChain agents, GPT-4/Claude model selection, and production reliability (monitoring, caching, retries) plus rigorous evaluation and regression testing.”
Mid-level AI/ML Engineer specializing in deep learning, MLOps, and LLM applications
“Built and deployed production LLM assistants for internal Q&A and customer-feedback summarization, emphasizing reliability (RAG, prompt tuning, validation/whitelisting) and privacy safeguards. Improved adoption by adding explainable outputs and a user feedback mechanism, and has hands-on orchestration experience with Aflow and Azure Logic Apps.”
Mid-level AI/ML Engineer specializing in LLM systems, RAG, and MLOps
“Built a production, real-time clinical documentation system at HCA that converts doctor–patient conversations into structured clinical summaries using speech-to-text, LLM summarization, and RAG. Demonstrated measurable gains from medical-domain fine-tuning (clinical concept recall +18%, ROUGE-L 0.62 to 0.74) while meeting HIPAA constraints via PHI anonymization and encryption, and deployed via Docker/FastAPI with CI/CD and monitoring.”
Mid-level GenAI Engineer specializing in LLM fine-tuning, RAG, and MLOps
“Healthcare-focused LLM engineer who deployed a production triage and clinical knowledge retrieval assistant using RAG and LangGraph-orchestrated multi-agent workflows. Emphasizes clinical safety and compliance with robust hallucination controls, HIPAA/PHI protections (tokenization, encryption, audit logging, zero-retention), and human-in-the-loop escalation; reports a 75% latency reduction in a healthcare agent system.”
“Built a production multi-agent orchestration platform to automate healthcare claims and HR workflows, combining LangChain/CrewAI/AutoGPT with RAG (FAISS/Pinecone) and fine-tuned open-source LLMs (LLaMA/Mistral/Falcon) in private Azure ML environments to meet HIPAA requirements. Emphasizes rigorous agent evaluation/observability (trajectory eval, adversarial testing, LLM-as-judge, drift monitoring) and reports measurable outcomes including 35% faster claims processing and 40% fewer chatbot errors.”
Senior AI/ML Engineer specializing in Generative AI, LLMs, and MLOps
“Telecom (Verizon) AI/ML practitioner who built a production multimodal system that ingests messy customer issue reports (calls, chats, emails, screenshots, videos) and turns them into confidence-scored incident summaries with reproducible steps and evidence links. Also built KPI/alarm-to-ticket correlation to rank likely root-cause domains (RAN/Core/Transport), cutting triage from hours to minutes and improving MTTR.”
Mid-level Machine Learning Engineer specializing in NLP, LLMs, and MLOps
“Built and deployed an LLM-powered financial/regulatory document analysis platform at State Street, combining fine-tuned transformer models with a RAG pipeline over internal knowledge bases. Owned the productionization stack (FastAPI, Docker, SageMaker, Terraform, CI/CD) plus monitoring for drift/latency/hallucinations, delivering ~40% faster analyst review and improved reliability through chunking/embeddings and grounding.”
Mid-level Software Engineer specializing in full-stack and cloud-native microservices
“Backend engineer who built a Python/Flask system for high-volume healthcare claims processing, using PostgreSQL as the source of truth and RabbitMQ workers for scalable async processing. Experienced in SQLAlchemy/Postgres performance tuning, multi-tenant data isolation (including Postgres RLS), and integrating/versioning ML model services (scikit-learn/PyTorch/Hugging Face) with controlled rollouts. Drove measurable performance gains by batching background jobs and adding Redis caching (40% less workload; response times cut from ~10s to 2–3s).”