Pre-screened and vetted.
Mid-level Data Engineer specializing in Analytics & AI/ML
“Data engineer with experience at Sony and Walmart building high-volume, near-real-time analytics and ingestion systems. Has owned end-to-end pipelines from Kafka/Spark streaming through S3/Parquet and Redshift/Looker, emphasizing data quality (Great Expectations), observability (CloudWatch/Azure Monitor), and reliability (Airflow SLAs, retries, checkpointing), including measurable performance and latency improvements.”
Senior Data Engineer specializing in cloud lakehouse platforms and streaming analytics
“Data engineer focused on fraud and banking analytics who has owned end-to-end batch + streaming pipelines at very large scale (hundreds of millions of records/day). Built robust data quality/observability layers (schema validation, anomaly detection, alerting) and delivered low-latency serving via AWS Lambda/API Gateway with DynamoDB + Redis, plus external data ingestion/scraping pipelines orchestrated in Airflow with anti-bot protections.”
Junior Machine Learning Engineer specializing in LLMs and applied data science
“Built and shipped multiple production AI systems, including Auto DocGen (LLM-generated OpenAPI docs kept in sync via AST diffs, schema-constrained generation, and CI/CD on Render) and a multimodal sign-language recognition pipeline at USC orchestrated with FastAPI, MediaPipe, and PyTorch. Also partnered with Esri’s non-technical community team to fine-tune an LLaMA-based spam classifier with a review UI, cutting moderation time by 70%.”
Junior AI/ML Engineer specializing in LLM systems and retrieval-augmented generation
“Built and deployed a production LLM-powered market intelligence and decision-support platform for noisy, real-time financial data, using a high-throughput embedding + vector DB RAG architecture to reduce hallucinations while keeping latency and cost low. Operated it at scale with GPU-backed inference (continuous batching/quantization), FastAPI on Kubernetes, and Airflow-orchestrated ingestion/embedding/retraining workflows, with strong schema-based reliability and monitoring.”
Junior Software Engineer specializing in AI and full-stack development
“Consulting-background AI practitioner who led a production LLM pipeline on Snowflake Cortex to map hundreds of thousands of messy OCR/form-based contract fields into standardized Salesforce fields, including confidence scoring and an LLM-driven feedback loop. Strong focus on real-world constraints—token limits, cost control, and evaluation without ground truth—paired with frequent stakeholder-facing progress reporting.”
Senior Software Engineer specializing in cloud-native microservices and event-driven systems
“Senior engineer/tech lead with 18+ years building large-scale distributed applications, specializing in performance and reliability improvements. Recently owned multiple apps on an email personalization team, shipping major optimizations (including a push-update feature and audience-count architecture redesign) that reportedly lifted system performance from ~50% to ~99% while also leading code standards, reviews, and mentoring.”
Mid-level Data Engineer specializing in Azure ETL/ELT and data warehousing
“Data engineer who has owned end-to-end production pipelines for customer transaction data (~2–5 GB/day) using Python/PySpark/SQL and Airflow, delivering major reliability and speed gains (70% faster reporting; 60–70% fewer data issues). Also built a daily external web-scraping system with anti-bot handling and safe, idempotent Airflow-driven backfills, plus a Python data API optimized with indexing/caching and tested for correctness.”
Mid-level Data Engineer specializing in financial data pipelines and reliability
“Systems/robotics-oriented software engineer focused on real-time orchestration and reliability: built a central control layer coordinating multiple concurrent agents with safe state machines, failure isolation, and recovery. Has hands-on ROS/ROS 2 integration experience in simulation (DDS/QoS, lifecycle, nodes in Python/C++) and emphasizes observability (structured JSON logs, correlation IDs) and low-latency control-loop performance under load.”
Mid-level Data Engineer specializing in real-time analytics and regulated domains
“Data platform engineer focused on large-scale, real-time fraud systems, with hands-on ownership of streaming architectures using Kafka, Spark, Snowflake, and Databricks. Stands out for combining performance tuning and platform automation with LLM/RAG-based enrichment, delivering measurable gains in latency, fraud accuracy, false positives, and analyst decision speed.”
Mid-level Business Analyst specializing in BI, reporting, and data insights
“Healthcare analytics professional with experience at UnitedHealth Group, focused on turning messy claims, eligibility, and provider data into clean reporting datasets and Power BI dashboards. Combines SQL and Python automation with strong stakeholder alignment around KPI definitions, helping operations teams improve claim turnaround visibility and cost efficiency.”
Mid-level AI/ML Engineer specializing in generative AI, NLP, and MLOps
“ML/AI engineer with hands-on ownership of production GenAI and computer vision systems, spanning experimentation, deployment, monitoring, and iterative optimization. Stands out for shipping an enterprise RAG platform that cut manual review by 50% and a defect detection pipeline that reduced report generation from 15 minutes to under 1 second while maintaining high uptime and strong operational discipline.”
Mid-level Full-Stack Engineer specializing in enterprise SaaS and optimization platforms
“Full-stack engineer with strong enterprise delivery experience across manufacturing and semiconductor use cases, owning deployments from discovery through post-launch support. Stands out for combining traditional product engineering with applied GenAI workflows and data pipeline reliability work, including a manufacturing app that reportedly saved a Fortune 500 customer about $6M and an AI chat panel adopted by 70% of pricing analysts.”
Senior Backend Software Engineer specializing in AI, FinTech, and Healthcare
“Founding engineer who has built web products end-to-end in startup settings, spanning FastAPI/React application development, auth, cloud deployment, and Kubernetes-based scaling. Particularly notable for designing custom GPU autoscaling for an AI-style recommendation product and later shipping workflow-driven healthcare support tooling using Temporal, Postgres, and modular backend logic.”
Mid-level Data Engineer specializing in scalable pipelines, Spark, and cloud data warehousing
“Backend/data platform engineer who recently owned an end-to-end large-scale financial data platform delivering real-time decision support for finance and operations. Has hands-on experience modernizing legacy batch pipelines into AWS cloud-native ELT with parallel-run cutovers, strong data quality controls (dbt-style tests, reconciliation), and measurable improvements in runtime, cost, and SLA compliance. Also builds scalable, secure FastAPI microservices using Docker, ALB-based horizontal scaling, Redis caching, and managed auth with Cognito/Supabase plus Postgres RLS.”
“Built an LLM multi-agent “ingredient safety” analyzer for cosmetics that cuts consumer research time from ~20+ minutes to minutes, using LangGraph orchestration, hybrid retrieval (Qdrant + Tavily), and safety-focused critic validation (false rejections reduced ~30%→~8%). Also has research-internship experience building computer-vision pipelines to classify emerald color/clarity by translating gem-expert heuristics into quantitative model features.”
Junior Software Engineer specializing in cloud-native microservices and AI/ML observability
“Engineer with banking and industrial/IoT experience who has deployed a payment-processing microservice with zero downtime, handling Protobuf schema evolution and sensitive data migration via dual-write/checksum techniques. Demonstrates strong cross-stack troubleshooting (pinpointed intermittent distributed timeouts to a failing ToR switch port) and customer-facing Python ETL customization using plugin-based parsers and Pydantic validation, plus hands-on monitoring/alerting improvements with operators.”
Mid-level Machine Learning Engineer specializing in NLP and cloud MLOps
“Built and deployed a production LLM-powered internal documentation assistant using embeddings, a vector database, and a RAG pipeline to reduce time spent searching PDFs/manuals. Experienced in orchestrating end-to-end LLM workflows with Airflow/LangChain, improving reliability via monitoring/error handling, and driving measurable quality through retrieval and hallucination-focused evaluation metrics.”
“Built and productionized an LLM-powered PDF document Q&A system to eliminate manual searching through long documents, focusing on scalability and answer reliability. Implemented semantic chunking (using headings/paragraphs/tables), overlap, and preprocessing/quality checks to reduce hallucinations, and orchestrated the end-to-end pipeline with Airflow using retries, alerts, and parallel tasks.”
Mid-level Data Engineer specializing in real-time pipelines and cloud analytics
“Researcher from the University of South Dakota who built a production medical RAG system to help interpret model predictions by retrieving relevant clinical notes and medical literature, overcoming retrieval accuracy and imaging-dataset challenges through semantic chunking and metadata-driven indexing. Also has hands-on orchestration experience with Airflow and Azure Data Factory, plus a pragmatic approach to LLM evaluation and stakeholder-driven iteration.”
Senior Data Scientist specializing in machine learning and customer analytics
“Data/ML practitioner with experience applying NLP and classical ML to large-scale customer data (2B+ records) for segmentation, prediction, and survey-text classification, delivering measurable business impact (~18% engagement efficiency). Has hands-on entity resolution across multi-source datasets and has built embedding-based semantic search using SentenceBERT + a vector database with domain fine-tuning (~20% relevance improvement), plus production workflow experience with Spark/Airflow and cloud tooling (AWS/Azure).”
Mid-level Machine Learning Engineer specializing in MLOps, NLP, and Computer Vision
“ML/AI engineer with production experience across retail and healthcare: built a real-time computer-vision shelf monitoring system at Walmart and optimized edge inference latency by ~30% using TensorRT/ONNX and pruning. Also partnered with CVS Health clinical/pharmacy teams to deliver a medication-adherence predictive model, using Streamlit explainability dashboards and achieving an 18% adherence improvement.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and MLOps
“AI/ML engineer who has shipped production AI systems end-to-end, including an automated multi-channel (Gmail/WhatsApp/voice) candidate interviewing workflow and an enterprise RAG knowledge search platform. Demonstrates strong production rigor (monitoring, A/B tests, guardrails, schema validation, shadow testing) with quantified impact: ~60–70% reduction in interview evaluation time and ~20–30% relevance gains in RAG retrieval.”
Mid-level Generative AI Engineer specializing in decision intelligence and RAG for regulated enterprises
“Healthcare GenAI engineer who built a HIPAA-compliant, auditable RAG-based claims decision support system at Molina Healthcare, processing 3M claims and delivering major impact (48% faster manual reviews, 43% higher decision accuracy). Deep hands-on experience with LangChain orchestration, vector search (ChromaDB/FAISS), embedding fine-tuning, and safety controls (confidence scoring, rule validation, human-in-the-loop escalation) for clinical workflows.”
Senior Data Engineer specializing in Azure Lakehouse, Databricks/Spark, and Snowflake
“Data engineer/platform builder with experience across PwC and Liberty Mutual delivering high-volume, production-grade pipelines and real-time data services. Has owned end-to-end streaming + batch architectures on AWS and Azure, including web scraping systems, with quantified reliability gains (99.9% availability, 90%+ error reduction, 30% latency reduction) and strong observability/CI-CD practices.”