Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in MLOps, NLP, and real-time ML pipelines
“Built a production, real-time insurance claims document-understanding and fraud-detection pipeline using TensorFlow + fine-tuned BERT, deployed on AWS (SageMaker/Lambda/API Gateway) with automated retraining via MLflow and Jenkins. Addressed noisy documents and latency using augmentation and model distillation (3x faster), cutting claims ops manual review by ~50% and reducing fraudulent payouts.”
Mid-level Data Scientist/MLOps Engineer specializing in NLP, GenAI, and cloud ML platforms
“AI/ML engineer who led production deployment of a multimodal (text/video/image) RAG system on GCP using Gemini 2.5 + Vertex AI Vector Search, scaling to 10M+ documents with sub-second latency and +40% retrieval accuracy. Strong MLOps/orchestration background (Kubernetes, CI/CD, Airflow, MLflow) with proven impact on reliability (75% fewer incidents) and deployment speed (92% faster), plus experience delivering explainable ML (XGBoost + SHAP + Tableau) to non-technical retail stakeholders.”
Senior Data Engineer specializing in cloud lakehouse and streaming data platforms
“Data platform/data engineer with cross-industry experience in banking and healthcare, building cloud-native lakehouse architectures across AWS/Azure/GCP. Has owned high-volume (millions of records; TB/day) pipelines with strong data quality automation (dbt/Great Expectations), observability (Grafana/Prometheus), and real-time streaming (Kafka/Spark) for fraud monitoring; also delivered an early-stage migration from SQL Server to BigQuery with 40% batch latency reduction.”
Senior DevSecOps Engineer specializing in Azure cloud infrastructure and CI/CD
“GCP-focused database/infrastructure engineer with hands-on production support for Cloud SQL and Firestore, spanning provisioning, IAM, scaling, backups, and performance tuning. They also described supporting a hybrid GCP architecture for a monolithic on-prem PostgreSQL workload and resolving a major latency incident by tracing cascading failures and fixing indexing issues.”
Senior AI/ML Engineer specializing in Generative AI, LLMs, and RAG systems
“AI/ML engineer with hands-on experience shipping production systems across fintech, travel, and legal use cases. They’ve built end-to-end chatbot, generative content, and RAG solutions on AWS with CI/CD, monitoring, and guardrails, including a loan application platform that generated $3,000 in sales in its first month.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps for financial services
“Built and deployed a production Llama 3-based RAG document Q&A system using FAISS, addressing context-window limits through chunking and keeping retrieval accurate by regularly refreshing embeddings. Has hands-on orchestration experience with LangChain and LlamaIndex for multi-step LLM workflows (including memory management) and collaborates with non-technical teams (e.g., marketing) to deliver AI solutions like recommendation systems.”
Mid-level AI/ML Engineer specializing in healthcare ML and generative AI
“AI/LLM engineer at Humana who built and deployed a HIPAA-aware RAG system for clinical record retrieval, cutting search time dramatically and improving retrieval efficiency by 30%. Experienced with Spark-scale data preprocessing, QLoRA fine-tuning, LangChain orchestration, and MLflow+SageMaker integration, with a strong testing/evaluation discipline (A/B tests, human eval) to hit 95%+ accuracy and production latency targets.”
Intern Data Analyst specializing in data pipelines and LLM/RAG applications
“Built and deployed LLM-powered analytics and reporting systems, including a RAG-based assistant over Snowflake that let business users ask questions in plain English instead of writing SQL. Experienced orchestrating LLM agents (LangChain) and serverless reporting pipelines (AWS Lambda/S3/RDS), with a strong focus on grounded outputs, monitoring/evaluation, and data quality—used daily by non-technical finance and operations teams at Cigna.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Built a production RAG-based healthcare chatbot to retrieve patient medical documents spread across multiple platforms, reducing manual and error-prone searching. Implemented semantic search with custom embeddings (Hugging Face) and Pinecone, deployed via FastAPI/Docker on AWS SageMaker with MLflow tracking, and optimized fine-tuning cost using LoRA while orchestrating retraining pipelines in Airflow.”
Mid-Level Full-Stack .NET Developer specializing in cloud-native microservices and AI integration
“Software engineer with hands-on experience building and maintaining a React accessibility utility/component library (open-source-style) used across university portals, emphasizing WCAG 2.2 compliance, robust focus/keyboard behavior, and Jest/React Testing Library coverage. Also built and maintained .NET Core microservices at the Florida Department of Transportation, including integrating AI-driven features, with strong ownership around observability, incident response, and performance-focused refactoring.”
“Backend/data engineer who builds Python (FastAPI) data-processing API services for internal analytics/reporting, emphasizing modular architecture, async performance tuning, and reliability patterns (health checks, retries, observability). Also migrated legacy on-prem ETL pipelines to Azure using ADF/Data Lake/Functions and implemented a near-real-time ingestion flow with Event Hubs plus watermarking to handle late events and deduplication.”
Mid-level AI Engineer specializing in LLMs, RAG, and data engineering
“AI Engineer Co-Op at Northeastern University who built a production Patient Persona Chat Bot to help nursing students practice clinical interactions, fine-tuning Llama 3 and integrating a LangChain + Pinecone RAG pipeline deployed on Amazon Bedrock. Emphasizes clinical accuracy and reliability with guardrails, retrieval filtering, and continuous evaluation, and also brings strong data engineering/orchestration experience (Airflow, EMR/PySpark, ADF, dbt, Databricks, Snowflake).”
Mid-Level Full-Stack Software Engineer specializing in cloud-native microservices
“Backend engineer with cloud-native Python/Flask experience building high-throughput financial platforms (loan origination intelligent document processing and real-time fraud detection). Has scaled microservices on AKS with event-driven Azure messaging, delivered measurable performance gains (e.g., 700ms→180ms query latency; ~40% API improvements), and implemented strong security controls (OAuth2/JWT, Azure AD RBAC, audit logging, AES-256/TLS) for sensitive regulated data.”
Mid-level AI/ML Engineer specializing in predictive modeling and cloud ML pipelines
“LLM engineer/data engineer who has deployed production RAG systems for internal-document Q&A, building end-to-end ingestion, embedding, vector search, and FastAPI serving while actively reducing hallucinations and latency through rigorous retrieval tuning and caching. Also experienced in orchestrating cloud data pipelines (Airflow, AWS Glue, Azure Data Factory) and partnering with non-technical business teams to deliver AI solutions like automated document review.”
Junior Data & Insights Analyst specializing in BI, dashboards, and automation
“Worked on taking an LLM-based system at Soundmakr from prototype to production by adding prompt constraints, validation/guardrails, deterministic ranking, and robust logging/monitoring with feedback loops. Also partnered with product/marketing during an internship on Thea: Study Smart to analyze onboarding drop-offs and run A/B tests on AI-driven flows, translating results into actions that improved retention and conversion.”
Mid-level Data Engineer specializing in cloud data platforms and AI agents
“Data/Backend engineer who has owned end-to-end merchant analytics systems on AWS: orchestrated multi-source ingestion (FISERV/Shopify/Clover) with Step Functions/Lambda, enforced strong data quality gates, and served curated datasets via Redshift and a FastAPI layer. Also built an early-stage Merchant Insights AI agent that converts natural language questions into SQL using OpenAI models, with full CI/CD and observability.”
Mid-level Data Engineer and Business Analyst specializing in cloud ETL and analytics
“Data analyst with cross-industry experience spanning insurance analytics at L&T Infotech and experimental imaging analytics at Mylyser. Stands out for building scalable SQL/PySpark data pipelines, standardizing business-critical metrics like claims lifecycle and policy retention, and delivering measurable impact such as 50%+ faster query performance and a 15% reduction in claims settlement time.”
Mid-level AI/ML Engineer specializing in MLOps and cloud-deployed ML systems
“ML/AI engineer who built and productionized an NLP system at PurevisitX, orchestrating end-to-end ML workflows with Airflow (S3 ingestion through auto-retraining) and optimizing for drift and low-latency inference. Also partnered with Citibank risk teams on a fraud detection model, translating results via dashboards and iterating thresholds based on stakeholder feedback.”
Mid-level GenAI/ML Engineer specializing in LLM systems and RAG chatbots
“Built and shipped a production agentic LLM analytics platform that lets non-SQL business users query relational databases in plain English via a RAG + LangChain/LangGraph workflow and FastAPI service. Emphasizes safety and reliability with guardrails (validation/access control), testing/evaluation frameworks, and performance optimization (caching, monitoring, Dockerized scalable deployment), reducing dependency on data teams and speeding analytics turnaround.”
Senior Data Scientist specializing in ML, NLP, and production AI systems
“Machine learning/NLP engineer with deep Azure stack experience (Data Factory, Databricks/Spark, Delta Lake, Azure OpenAI, Azure AI Search) who built end-to-end production systems for semantic clustering, entity resolution, and hybrid search. Demonstrated measurable gains from embedding fine-tuning (~15% retrieval precision, ~10–12% nDCG@10) and designed scalable, quality-checked pipelines with MLOps best practices.”
Mid-level Machine Learning Engineer specializing in cloud-native generative AI for healthcare
“AI engineer at Cleveland Clinic building production LLM/NLP systems for radiology documentation, focused on HIPAA-aware, real-time performance across ~298 campuses. Re-architected infrastructure with AWS event-driven services to handle scaling and improved SLA compliance ~40%, and complements this with a personal multi-agent debate system (CrewAI) using local Llama/Mistral plus rigorous evaluation (A/B tests, red teaming, observability).”
Mid-level AI/ML Engineer specializing in healthcare imaging and GenAI/LLM systems
“Built and deployed a production LLM/RAG clinical document understanding and summarization system for healthcare, focused on reducing manual review time while meeting strict accuracy, latency, and compliance needs. Demonstrates strong MLOps/orchestration depth (Airflow, Kubernetes, Azure ML Pipelines) and a rigorous approach to hallucination mitigation through layered, source-grounded safeguards and stakeholder-driven requirements with physicians/compliance teams.”
Mid-level Automation Developer specializing in RPA, test automation, and data/ETL pipelines
“Python backend engineer who owned an end-to-end Django/DRF authentication and account-management module (JWT, RBAC, email verification) and optimized token validation performance. Has hands-on Kubernetes + Helm delivery with GitOps via ArgoCD (multi-environment app-of-apps, drift detection/rollback) and has supported a cloud-to-on-prem migration using staged testing and phased cutover. Also built and scaled a Kafka-based real-time user activity tracking pipeline with reliability and backpressure controls.”
Mid-level AI/ML & Data Engineer specializing in MLOps and cloud data pipelines
“AI/ML engineer (Merkle) with hands-on experience deploying RAG-based LLM applications and real-time recommendation engines into production. Strong in cloud/on-prem architectures, GPU autoscaling, caching, and network optimization—delivered measurable latency reductions (40–70%) and improved retrieval relevance by systematically benchmarking chunking/embedding configurations and validating pipelines via CI/CD.”