Pre-screened and vetted.
Mid-level Full-Stack Java Developer specializing in cloud-native microservices
“Software engineer with deep healthcare claims domain experience who has owned customer-facing portals end-to-end (Java/Spring Boot + React/TypeScript) and improved usability/performance based on real user feedback. Built microservices using REST and RabbitMQ with strong observability (Splunk/cloud metrics), and delivered an internal claims investigation dashboard that streamlined operations through centralized data, search, and filtering.”
Mid-level Full-Stack Java Developer specializing in cloud-native microservices and React
“Full-stack engineer who owned enterprise workflow platforms end-to-end at Northern Trust and Elevance Health—building NestJS/Java Spring Boot APIs, React UIs, and cloud deployments on GCP Cloud Run. Strong in data-heavy applications (hundreds of thousands of records) with proven production performance tuning (indexing/query rewrites, Cloud Run concurrency/min instances) and secure RBAC via Azure AD.”
Mid-level Data Scientist specializing in real-time fraud detection and MLOps
“ML/NLP engineer with experience at Charles Schwab building an NLP + graph (Neo4j) entity-resolution system to unify fragmented user/device/transaction data and improve downstream model quality and analyst querying. Has applied embeddings (SentenceTransformers + FAISS) with domain fine-tuning to boost hard-case matching recall by ~12% while maintaining precision, and has a track record of hardening scalable Python/Spark pipelines and productionizing fraud models via A/B tests and shadow-mode monitoring.”
Senior Data & Platform Engineer specializing in cloud-native streaming and distributed systems
“Financial data engineer who has built and operated high-volume batch + streaming pipelines (200–300 GB/day; 5–10k events/sec) using AWS, Spark/Delta, Airflow, Kafka, and Snowflake, with strong emphasis on data quality and reliability. Demonstrated measurable impact via 99.9% SLA adherence, major reductions in bad records/nulls, MTTR improvements, and significant latency/runtime/query performance gains; also built a distributed web-scraping system processing 5–10M records/day with anti-bot and schema-drift defenses.”
Mid-level Data Engineer specializing in multi-cloud data platforms for healthcare and finance
“Data engineer with Cigna experience building and operating an end-to-end AWS-based healthcare claims pipeline processing ~2TB/day, using Glue/Kafka/PySpark/SQL into Redshift. Strong focus on data quality and reliability (schema validation, monitoring/alerting, retries/checkpointing/backfills), reporting improved accuracy (~99%) and reduced latency, plus experience serving real-time Kafka/Spark data to downstream analytics with documented data contracts.”
Mid-level AI/ML Engineer specializing in NLP and conversational AI
“ML/NLP engineer focused on real-time IT ops analytics, building a predictive maintenance/anomaly detection platform end-to-end (multi-source ETL, streaming, modeling, and production deployment on GCP/Vertex AI). Uses deep learning (LSTMs, autoencoders/VAEs) plus embeddings (SentenceBERT) and vector search to improve incident correlation and search, citing ~40% reduction in duplicate alert noise.”
Mid-level AI/ML Engineer specializing in FinTech risk, fraud detection, and GenAI/RAG systems
“Built and productionized Azure-based LLM/RAG systems for regulatory/compliance use cases, including automating analyst research and compliance report generation across large unstructured document sets. Demonstrates strong practical depth in hallucination mitigation, hybrid retrieval tuning (BM25 + embeddings), and production MLOps (Databricks, Cognitive Search, AKS, Airflow/MLflow), plus proven ability to deliver auditable, explainable solutions with non-technical compliance teams.”
Mid-Level Software Engineer specializing in Cloud, DevOps, and MLOps
“Built and productionized a recommendation system from notebook prototype into a low-latency, scalable Cloud Run service using Docker, FastAPI, Terraform, CI/CD (GitHub Actions), and MLOps tooling (Vertex AI, MLflow). Experienced diagnosing real-time workflow issues using structured logging/ELK and GCP metrics, including resolving intermittent 504s by fixing unbounded SQL and adding caching. Also partners with sales/customer teams (Wasabi) to deliver tailored demos, troubleshoot, and drive onboarding/adoption.”
Mid-level Data Analyst specializing in healthcare and business intelligence
“Healthcare analytics candidate with hands-on experience turning messy EHR, billing, and operational data into validated SQL datasets and automated Python/Airflow pipelines. They appear strongest in hospital KPI reporting—especially length of stay, readmissions, retention, and bed utilization—and have owned projects from metric definition through Power BI delivery and impact measurement.”
Executive IT and Operations leader specializing in digital transformation and security
“Candidate is very familiar with the venture capital and broader investment landscape, but is not interested in founding a company. They have worked with several TPG-backed or TPG-owned organizations, helping drive business scaling, cost reduction, and execution against investor governance requirements.”
Mid-level Business Analyst specializing in operations data and reporting
“Candidate has hands-on project experience in healthcare analytics, using SQL, Python, and Power BI to analyze CMS hospital readmissions and HRRP penalty risk in Florida. Their work centers on turning messy CMS flat files into reporting-ready datasets, benchmarking hospitals against national references, and surfacing financial risk through dashboards.”
Mid-level Data Scientist specializing in machine learning, MLOps, and cloud analytics
“Senior data scientist with ~5 years’ experience building production ML/NLP systems in finance (Wells Fargo) and deep learning for sensor analytics in connected vehicles (Medtronic). Has delivered end-to-end platforms combining time-series forecasting with transformer-based NLP, including automated drift monitoring/retraining (MLflow + Airflow) and standardized Docker/CI/CD deployments; achieved a reported 22% precision improvement after domain fine-tuning.”
Principal AI/ML Leader specializing in Generative AI, MLOps, and NLP
“Founding member of Tausight, building AI systems to detect and protect PHI for healthcare organizations; helped take the company through post–Series A funding and exited after ~6 years. Drove a strategic collaboration with Intel’s OpenVINO team—becoming the first to deploy it in a real production system and improving model performance by ~30% on customer Intel-CPU machines.”
Mid-level Full-Stack Engineer specializing in cloud-native microservices and AI automation
“Software engineer/product owner who has led end-to-end delivery of AI and content-management platforms, including building RAG-based reliability improvements and migrating fragile systems to containerized AWS ECS/Kubernetes with Terraform-managed CI/CD. Experienced designing event-driven microservices (SQS/SNS/RabbitMQ), scaling queue consumers with autoscaling, and creating internal Python tooling to standardize data connectors (e.g., BigQuery/Airtable/internal APIs) to speed iteration.”
Mid-level Data Engineer specializing in Lakehouse, Streaming, and ML/LLM data systems
“Built and productionized an enterprise retrieval-augmented generation platform for internal knowledge over large unstructured corpora, emphasizing trust via strict citation/grounding and hybrid retrieval (BM25 + FAISS + cross-encoder re-ranking). Demonstrates strong scaling and cost/latency optimization through incremental indexing/embedding and index partitioning, plus disciplined evaluation/observability practices. Has experience operationalizing pipelines with Airflow/Databricks/GitHub Actions and partnering closely with risk & compliance stakeholders on auditability requirements.”
Mid-level AI/ML Engineer specializing in MLOps and LLM applications
“BNY Mellon engineer who has built and operated production AI systems end-to-end: a LangChain/Pinecone RAG platform scaled via FastAPI + Kubernetes to 1000 RPM with 99.9% uptime, supported by monitoring and data-drift detection. Also deep in data/infra orchestration (Airflow, Dagster, Terraform on AWS/EMR/EC2), processing 500GB+ daily and delivering measurable reliability and performance gains, plus strong compliance-facing model explainability using SHAP and Tableau.”
Mid-Level Software Developer specializing in Java/Spring microservices and Salesforce
“Backend/AI engineer who built an AI icon-generation SaaS backend (Java/Spring Boot, MongoDB) on AWS, including async job processing with idempotency and S3-based result storage to handle traffic spikes. Also shipped applied AI in finance—an end-to-end fraud detection pipeline with risk scoring—and designed a banking support AI agent with strict guardrails, audit logs, and human-in-the-loop escalation.”
Mid-level Financial/Data Analyst specializing in analytics, forecasting, and healthcare/MarTech data
“Growth/creative marketer from Esleydunn Games who uses Google Analytics to integrate cross-channel performance data (TikTok, YouTube, LinkedIn, Facebook) and run structured A/B tests on video ad length and layout. Reported reducing CPA by 20 per customer when leveraging YouTube and TikTok, and improved CTR through CTA/button placement testing and ongoing user-feedback loops (forum/WeChat topics).”
Mid-level AI/ML Engineer specializing in cloud data engineering and GenAI
“AI/LLM engineer with production experience in legal tech: built a GPT-4 + LangChain RAG summarization system at Govpanel that reduced legal case-file review time by 50%+. Previously at LexisNexis, orchestrated end-to-end Airflow data/AI pipelines processing 5M+ legal documents daily, improving ETL runtime by 35% with robust validation, monitoring, and SLAs.”
Mid-level Data Engineer specializing in cloud data platforms and real-time analytics
“Customer-facing data engineering professional who builds and deploys real-time reporting/dashboard solutions, gathering reporting and compliance requirements through direct stakeholder engagement. Experienced with Google Cloud IAM governance, secure integrations (encryption, audit logging), and fast production troubleshooting of ETL/pipeline failures with follow-on monitoring and automated recovery improvements; motivated by hands-on, travel-oriented customer work.”
Mid-level AI/ML Engineer specializing in generative AI, RAG platforms, and LLM agents
“AI/LLM engineer who has shipped 10+ production applications, including InvestIQ on GCP—a production-grade RAG due-diligence engine that ethically scrapes web/PDF sources, builds a ChromaDB knowledge base, and delivers analyst-style dashboards plus a citation-backed chat copilot. Deep focus on reliability (evidence-only answers, hard citations, refusal gating), retrieval tuning, and orchestration (Airflow/Cloud Composer), plus multi-agent systems (CrewAI with 7 specialized finance agents).”
Mid-level Data Scientist specializing in Generative AI, NLP, and MLOps
“Built and deployed an LLM-powered claims-document summarization system (insurance domain) that cut agent review time from 4–5 minutes to under 2 minutes and saved 1,200+ hours per quarter. Hands-on across orchestration and production infrastructure (Airflow retraining DAGs, Kubernetes, SageMaker endpoints, FastAPI) and recent RAG workflows using n8n + Pinecone, with a strong focus on reliability, cost, and explainability for non-technical stakeholders.”
Senior Machine Learning Engineer specializing in LLMs, speech AI, and RAG systems
“AI engineer with production experience building multilingual speech-to-speech translation pipelines (ASR + LLM) for enterprise/media, focused on reliability at scale. Has hands-on orchestration experience (including IBM Watson contexts) and emphasizes production evaluation/monitoring using a mix of traditional metrics and LLM-based evaluators to catch quality regressions while balancing latency and cost.”
“Analytics-focused candidate with internship and project experience at Recotap and CoUnderscore, combining SQL, Python, and BI dashboards to turn messy marketing and engagement data into decision-ready reporting. Stands out for tying analytics work to business outcomes, including ~15% CTR improvement, identifying ~40% misattributed spend, and enabling a ~$75K budget shift through better targeting.”