Pre-screened and vetted.
Junior Mobile & Full-Stack Software Engineer specializing in Flutter and Java/Spring
“Software engineer with experience at PTC (Onshape cloud CAD) and Firebolt building customer-facing features end-to-end—from user research and prototyping (tooltips, CSV imports) through deployment and usage monitoring. Also handled urgent production crashes in a Flutter mobile app across iOS/Android by diagnosing state-management conflicts and shipping a stabilizing patch; enjoys hands-on, customer-facing work and travel.”
Senior QA Automation Engineer specializing in API and microservices testing
“QA automation engineer who owned an end-to-end automated regression suite for a PlayStation digital store flow (login through checkout/payment), building a hybrid POM/data-driven framework from scratch with Selenium/TestNG/Cucumber and also using Playwright/TypeScript and Cypress. Integrated the suite into Jenkins CI/CD with nightly runs and reporting, improved coverage (happy + negative paths), and reduced release risk by catching critical issues like session timeout and transaction/payment defects before production.”
Senior Data & Platform Engineer specializing in cloud-native streaming and distributed systems
“Financial data engineer who has built and operated high-volume batch + streaming pipelines (200–300 GB/day; 5–10k events/sec) using AWS, Spark/Delta, Airflow, Kafka, and Snowflake, with strong emphasis on data quality and reliability. Demonstrated measurable impact via 99.9% SLA adherence, major reductions in bad records/nulls, MTTR improvements, and significant latency/runtime/query performance gains; also built a distributed web-scraping system processing 5–10M records/day with anti-bot and schema-drift defenses.”
Mid-level Data Engineer specializing in multi-cloud data platforms for healthcare and finance
“Data engineer with Cigna experience building and operating an end-to-end AWS-based healthcare claims pipeline processing ~2TB/day, using Glue/Kafka/PySpark/SQL into Redshift. Strong focus on data quality and reliability (schema validation, monitoring/alerting, retries/checkpointing/backfills), reporting improved accuracy (~99%) and reduced latency, plus experience serving real-time Kafka/Spark data to downstream analytics with documented data contracts.”
Executive Technology Leader specializing in digital transformation, headless e-commerce, and cloud architecture
“Technology leader focused on business-aligned roadmaps and integration-heavy ecommerce platforms. Recently delivered an on-time launch for lutusooking.com (a premium Hamilton Beach brand) by coordinating UX/UI, component-based middleware, BigCommerce, Algolia search, personalization/recommendations, payments, and supply chain integrations, and later improved scalability via a Jitterbit iPaaS approach proven during Black Friday/Cyber Monday traffic.”
Mid-level Data Engineer specializing in cloud ETL and real-time streaming
“Data engineer focused on AWS + Spark/Databricks pipelines, including an end-to-end nightly loan-data ingestion flow (~2.2M records) from Postgres/S3 through Glue and Databricks into a DWH with layered validation and alerting. Also built real-time streaming with Kafka + Spark Structured Streaming and a master’s project streaming Reddit data for sentiment analysis under ambiguous requirements and tight budget constraints.”
Mid-level AI/ML Engineer specializing in FinTech risk, fraud detection, and GenAI/RAG systems
“Built and productionized Azure-based LLM/RAG systems for regulatory/compliance use cases, including automating analyst research and compliance report generation across large unstructured document sets. Demonstrates strong practical depth in hallucination mitigation, hybrid retrieval tuning (BM25 + embeddings), and production MLOps (Databricks, Cognitive Search, AKS, Airflow/MLflow), plus proven ability to deliver auditable, explainable solutions with non-technical compliance teams.”
Mid-level Data Scientist specializing in machine learning, MLOps, and cloud analytics
“Senior data scientist with ~5 years’ experience building production ML/NLP systems in finance (Wells Fargo) and deep learning for sensor analytics in connected vehicles (Medtronic). Has delivered end-to-end platforms combining time-series forecasting with transformer-based NLP, including automated drift monitoring/retraining (MLflow + Airflow) and standardized Docker/CI/CD deployments; achieved a reported 22% precision improvement after domain fine-tuning.”
Mid-level AI/ML Engineer specializing in cloud data engineering and GenAI
“AI/LLM engineer with production experience in legal tech: built a GPT-4 + LangChain RAG summarization system at Govpanel that reduced legal case-file review time by 50%+. Previously at LexisNexis, orchestrated end-to-end Airflow data/AI pipelines processing 5M+ legal documents daily, improving ETL runtime by 35% with robust validation, monitoring, and SLAs.”
Mid-level AI/ML Engineer specializing in NLP, RAG systems, and real-time risk modeling
“AI/ML Engineer with 4+ years of experience (Capital One, Odin Technologies) and a master’s in Data Analytics (4.0 GPA) who has deployed LLM/RAG systems to production for compliance/risk and document review. Strong in orchestration and MLOps (Airflow, Kubernetes, MLflow, GitHub Actions) and in tackling real-world LLM constraints like latency, context limits, and data privacy, with measurable impact (20%+ manual review reduction; 33% faster release cycles).”
Mid-level Data Engineer specializing in cloud data platforms and real-time analytics
“Customer-facing data engineering professional who builds and deploys real-time reporting/dashboard solutions, gathering reporting and compliance requirements through direct stakeholder engagement. Experienced with Google Cloud IAM governance, secure integrations (encryption, audit logging), and fast production troubleshooting of ETL/pipeline failures with follow-on monitoring and automated recovery improvements; motivated by hands-on, travel-oriented customer work.”
Mid-level AI/ML Engineer specializing in Generative AI and data engineering
“IBM engineer who built and deployed a production RAG-based LLM assistant using LangChain/FAISS with a fine-tuned LLaMA model, served via FastAPI microservices on Kubernetes, achieving 99%+ uptime. Demonstrates strong practical expertise in reducing hallucinations (semantic chunking + metadata-driven retrieval) and managing latency, plus mature MLOps practices (Airflow/dbt pipelines, MLflow tracking, monitoring, A/B and shadow deployments) and effective collaboration with non-technical stakeholders.”
Mid-level Data Engineer specializing in cloud data platforms and scalable ETL pipelines
“Data engineer (~4 years) with full-stack delivery experience (Next.js App Router/TypeScript + React) building a real-time operations monitoring dashboard backed by Kafka and orchestrated data pipelines. Strong production focus: Airflow + CloudWatch monitoring, automated Python/SQL validation (99.5% accuracy), and CI/CD with Jenkins/Docker; has delivered measurable improvements in latency, pipeline reliability, and query performance (Postgres/Redshift).”
Mid-level Data Engineer specializing in cloud ETL/ELT and lakehouse architecture
“Data engineer focused on sales/marketing analytics pipelines, owning ingestion from CRMs/ad platforms through warehouse serving and dashboards at ~hundreds of thousands of records/day. Built reliability-focused systems including dbt/SQL/Python data quality gates with alerting, a resilient web-scraping pipeline (retries/backoff, anti-bot tactics, schema-change detection, backfills), and a versioned internal REST API with caching and strong developer usability.”
Mid-level Data Engineer specializing in real-time streaming and cloud data platforms
“Data engineer with Wells Fargo experience owning an end-to-end lakehouse ETL pipeline on Databricks/Azure Data Factory, processing ~480GB daily and implementing robust data quality/reconciliation across 40+ tables to reach ~99.3% reliability. Strong in performance optimization (cut runtime 5.5h→3.8h), CI/CD and monitoring, and resilient external/API ingestion with retries, schema validation, and backfills.”
Mid-Level Full-Stack Software Engineer specializing in AWS cloud and Python/Java
“Accenture consultant who shipped an LLM-based production solution during a client cloud migration to parse application code and identify only the database objects actually used, cutting migration time by 30% and accelerating realization of cloud cost benefits. Emphasizes production robustness with timeouts/retries/fallback routing, validation, observability, and a disciplined eval/monitoring loop that turns failures into regression tests.”
Mid-level Data Scientist specializing in Generative AI and LLM production systems
“Built and deployed a production LLM-powered workflow assistant that automated internal marketing/production business tasks (document summarization, repeated Q&A, status updates). Demonstrates end-to-end applied LLM engineering: modular RAG architecture, hallucination/latency mitigation, automated evals to prevent prompt regressions, and Azure-based orchestration (Functions/Logic Apps) with monitoring and controlled rollouts.”
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and MLOps
“LLM/agentic systems engineer who built a production "Agentic AI Diagnostic Assistant" for network engineers, using a multi-agent Llama 2 + LangChain architecture with RAG over telemetry/incident data in DynamoDB and confidence-based deferrals to reduce hallucinations. Also has strong MLOps/orchestration experience (Airflow, EventBridge, Spark, Docker, SageMaker/ECS) at multi-terabyte/day scale and delivered multilingual NLP analytics (fine-tuned BERT/spaCy) for support operations through hands-on stakeholder workshops.”
Mid-level Data Scientist & AI/ML Engineer specializing in GenAI and cloud ML
“GenAI/LLM engineer who recently built a production compliance assistant at State Farm for KYC/AML and regulatory teams, using AWS Bedrock + LangChain with Textract/Lambda pipelines to extract fields, tag risk, and summarize long documents. Implemented RAG, strict structured outputs, and human-in-the-loop guardrails, and reports automating ~80% of documentation work while reducing review time by ~40%.”
Mid-Level Full-Stack Python Developer specializing in AI and data platforms
“Full-stack engineer who builds TypeScript/React SPAs on Python (Flask/FastAPI) backends and has hands-on experience integrating AI components (Azure OpenAI, LangChain, vector databases) into user workflows. Has built internal AI-enabled dashboards/search tools for analysts and business users, emphasizing typed API contracts, CI/CD-driven quality, and microservices reliability patterns (monitoring, retries, idempotency) at scale.”
Mid-Level Software Engineer specializing in data engineering and cloud platforms
“Backend-leaning full-stack engineer who has shipped production-critical data/reporting features at Walmart and built an end-to-end workflow automation product (FastAPI + React/TypeScript + PostgreSQL) deployed on AWS. Strong in performance/reliability engineering (parallel ETL, batch DB operations, indexing via EXPLAIN ANALYZE), secure API design (JWT/RBAC), and pragmatic incident-driven scaling (separating workers from API layer).”
Mid-level Data Engineer specializing in cloud lakehouse/warehouse pipelines
“Data engineer with HCA Healthcare experience building and operating end-to-end AWS-based pipelines for clinical and operational reporting (50–100 GB/day), serving curated data into Redshift/Snowflake for Power BI/Tableau. Emphasizes production reliability (Airflow SLAs/retries/alerting, logging/observability) and strong data quality controls (reconciliations, schema/null/duplicate checks), and has shipped versioned REST APIs to expose warehouse data to downstream systems.”
Senior Data Engineer specializing in cloud data platforms and real-time analytics
“Data engineer (Credit One) who built and owned real-time financial transaction and credit risk/fraud data systems end-to-end on AWS + Snowflake. Delivered high-scale pipelines (150k events/hour; ~2TB/week), raised data accuracy to 99%, and cut Snowflake costs 42% while adding strong observability, schema-drift handling, and production-grade APIs/documentation.”
Mid-level Data Analyst specializing in analytics, ETL, and cloud data platforms
“Data analyst with 4 years of experience spanning banking and retail/marketing analytics. Has hands-on experience building churn analytics pipelines in SQL and Python, optimizing large-query performance, and turning stakeholder-aligned metrics into recurring dashboards and business actions.”