Pre-screened and vetted.
Mid-Level Full-Stack Software Engineer specializing in cloud-native microservices
“Full-stack engineer with experience at Capital One and Prime Softech owning production systems end-to-end: secure authentication (Java/Spring Security + React/Redux) through AWS ECS deployments with Terraform and CI/CD. Strong reliability/observability focus (Prometheus/Grafana/ELK/CloudWatch) with quantified improvements (15% reliability gain, 30% fewer post-release defects). Also led legacy monolith-to-microservices refactors and built real-time Kafka/Spark ingestion pipelines for analytics/fraud detection.”
Mid-level Data Engineer specializing in cloud data platforms and scalable ETL pipelines
“Data engineer (~4 years) with full-stack delivery experience (Next.js App Router/TypeScript + React) building a real-time operations monitoring dashboard backed by Kafka and orchestrated data pipelines. Strong production focus: Airflow + CloudWatch monitoring, automated Python/SQL validation (99.5% accuracy), and CI/CD with Jenkins/Docker; has delivered measurable improvements in latency, pipeline reliability, and query performance (Postgres/Redshift).”
Mid-level AI Engineer specializing in LLM orchestration, RAG, and multi-agent systems
“Research Assistant at the University of Houston who built and live-deployed a production RAG system for 1000+ research documents, using hybrid retrieval (dense+BM25+RRF) with cross-encoder reranking and RAGAS-based evaluation; reported 66% MRR, 0.85+ faithfulness, and 68% lower LLM inference costs. Also built a deployed LangGraph multi-agent research system (Researcher/Critic/Writer) with tool integrations (Tavily, arXiv) and dual memory (ChromaDB + Neo4j), plus freelance automation work delivering a WhatsApp chatbot and n8n workflows for a wholesale clothing business.”
Senior Data Engineer specializing in Spark, Kafka, and Databricks Lakehouse platforms
“Data engineer at Fidelity who built and operated a real-time financial transactions lakehouse on AWS/Databricks, processing millions of records daily with Kafka streaming. Demonstrated strong reliability and data quality practices (watermarking, idempotent Delta writes, validation/reconciliation, observability) and delivered measurable improvements (~30% faster jobs and ~30% fewer data issues) while enabling trusted gold-layer analytics for downstream teams.”
Mid-level AI/ML Engineer specializing in NLP, LLMs, and RAG systems
“Backend engineer who built and evolved a PHI-compliant RAG system (FastAPI + LangChain + embeddings/FAISS) for internal document search and summarization, delivering <400ms p95 latency at ~2,500 daily requests and measurable impact (30% faster investigations, +17% retrieval relevance). Demonstrates strong security and rollout discipline (RBAC/RLS/JWT, redaction/audits, shadow mode, dual writes, canaries) and a focus on reducing hallucination risk via grounded guardrails and confidence-based fallbacks.”
Mid-Level Software Engineer specializing in cloud-native microservices and data platforms
“Robotics software engineer focused on multi-robot fleet orchestration in ROS 2, owning the fleet manager and task dispatch layer for pick/drop workflows. Strong in real-world reliability and safety (heartbeats, idempotent tasking, E-stop/localization confidence gates) and in debugging timing/state issues via telemetry alignment and rosbag replay, with experience in simulation, CI/CD, Docker, and Kubernetes-based deployments.”
Senior Full-Stack Software Engineer specializing in distributed systems and cloud microservices
“Product-minded full-stack engineer from CouponDunia who owned end-to-end notification and recommendation services at million-user scale. Built internal admin/analytics and operations dashboards in React/TypeScript with typed contracts and scalable Node.js REST APIs, and has deep microservices experience with Kafka/RabbitMQ (idempotency, retries/DLQs, partitioning, consumer tuning, and observability).”
Junior Machine Learning Engineer specializing in geospatial analytics and computer vision
“Built and evolved a geospatial ETL + API platform that processes pixel-wise satellite imagery in PostgreSQL/PostGIS into low-latency farm-level time-series metrics for an interactive dashboard, using precomputed hotspot analysis to reduce latency by 75–80%. Experienced in FastAPI-style API contract design (OpenAPI), caching, server-side filtering/compression, and production-minded security patterns (RBAC, session-derived authorization, password hashing) with disciplined rollback/versioning practices.”
Mid-level Data Scientist specializing in Generative AI and LLM production systems
“Built and deployed a production LLM-powered workflow assistant that automated internal marketing/production business tasks (document summarization, repeated Q&A, status updates). Demonstrates end-to-end applied LLM engineering: modular RAG architecture, hallucination/latency mitigation, automated evals to prevent prompt regressions, and Azure-based orchestration (Functions/Logic Apps) with monitoring and controlled rollouts.”
Mid-level AI/ML Engineer specializing in NLP and Generative AI
“Built and deployed a production LLM-powered RAG assistant for healthcare teams (care managers/support) to answer questions from clinical and policy documentation, emphasizing trustworthiness via improved retrieval, reranking, and strict grounding prompts to reduce hallucinations. Also has hands-on orchestration experience with Apache Airflow for end-to-end ETL/ML workflows and applies rigorous testing/metrics (hallucination rate, tool-call accuracy, latency, cost) to ensure reliable AI agent behavior.”
Senior Data Engineer specializing in cloud data platforms and ML pipelines
“Built and deployed AcademiQ Ai, a production LLM-based teaching assistant using GPT/BERT with RAG (LangChain + Pinecone) to handle large student notes and generate adaptive explanations/quizzes. Demonstrated measurable retrieval-quality gains (18% precision improvement, 22% less irrelevant context) by tuning similarity thresholds and chunking based on user satisfaction signals. Also orchestrated terabyte-scale, real-time demand forecasting pipelines using Airflow and Kubeflow on GCP with strong monitoring, shadow deployment, and feedback-loop practices.”
Mid-Level Full-Stack Python Developer specializing in AI and data platforms
“Full-stack engineer who builds TypeScript/React SPAs on Python (Flask/FastAPI) backends and has hands-on experience integrating AI components (Azure OpenAI, LangChain, vector databases) into user workflows. Has built internal AI-enabled dashboards/search tools for analysts and business users, emphasizing typed API contracts, CI/CD-driven quality, and microservices reliability patterns (monitoring, retries, idempotency) at scale.”
Mid-level Machine Learning Engineer specializing in LLMs, RAG, and MLOps
“LLM/agentic systems engineer who built a production "Agentic AI Diagnostic Assistant" for network engineers, using a multi-agent Llama 2 + LangChain architecture with RAG over telemetry/incident data in DynamoDB and confidence-based deferrals to reduce hallucinations. Also has strong MLOps/orchestration experience (Airflow, EventBridge, Spark, Docker, SageMaker/ECS) at multi-terabyte/day scale and delivered multilingual NLP analytics (fine-tuned BERT/spaCy) for support operations through hands-on stakeholder workshops.”
Mid-level Full-Stack Java Developer specializing in cloud microservices and enterprise apps
“Software engineer/product owner experience at UnitedHealth Group owning a high-volume claims eligibility console end-to-end (React/TypeScript + Spring Boot microservices) processing 1M+ transactions/day. Strong in event-driven architecture (Kafka/RabbitMQ), HIPAA-aligned security (OAuth/JWT/RBAC), and building internal observability tools that improve incident triage and production reliability.”
Mid-level AI/ML Engineer specializing in healthcare NLP and MLOps
“ML/AI engineer with healthcare payer experience (Signal Healthcare, Cigna) who has shipped production fraud/claims prediction systems using Python/TensorFlow and exposed them via FastAPI/Flask microservices integrated with EHR and Salesforce. Emphasizes operational reliability and trust—Airflow-orchestrated pipelines with data quality gates plus SHAP-based interpretability, A/B testing, and drift/debug workflows—backed by reported outcomes of 22% lower false payouts and 17% higher model accuracy.”
Mid-level Full-Stack Developer specializing in web platforms and cloud (AWS)
“Full-stack engineer with financial services experience (Lincoln Financial) who owned a customer-facing financial portal end-to-end using TypeScript/React and Node/Express. Has hands-on microservices and RabbitMQ event-driven workflows, addressing scale issues like retries/duplicates with idempotency and traceable logging, and built an internal real-time ops/support dashboard to improve monitoring and incident response.”
Mid-level Full-Stack Developer specializing in FinTech platforms and cloud-native microservices
“Backend/platform-focused Python engineer who has owned FastAPI services with Postgres/SQLAlchemy and production-grade auth (JWT + RBAC). Experienced deploying and operating microservices on Kubernetes with GitOps (ArgoCD), HPA tuning, and Prometheus/Grafana monitoring, plus hands-on cloud-to-on-prem migrations and Kafka-based real-time streaming pipelines.”
Mid-Level Full-Stack Java Developer specializing in cloud-native microservices
“Full-stack engineer with production experience building Java 17 Spring Boot microservices for high-traffic systems at Verizon and on a JPMC payments platform (funds transfer/validation using ISO 20022), plus modern React/TypeScript dashboards for ops and analytics. Demonstrates strong scalability and reliability chops (Kafka event-driven pipelines, Redis caching, clustering, BullMQ background jobs) and has built real-time apps end-to-end with secure JWT refresh-token auth and Socket.io performance tuning.”
Mid-level Data Scientist specializing in predictive modeling, NLP/LLMs, and RAG search systems
“Built production LLM/RAG platforms for financial services to enable natural-language Q&A over large policy/compliance document sets stored in Snowflake and SharePoint. Strong in MLOps and orchestration (Airflow, ADF, Step Functions, MLflow) and in solving real production issues like stale embeddings and model performance, including an incremental Snowflake Streams sync that cut processing time from hours to minutes.”
Mid-level Machine Learning Engineer specializing in NLP, LLMs, and MLOps
“Built a production internal LLM/RAG assistant at CVS Health to cut time spent searching long policy and clinical guideline PDFs, combining fine-tuned BERT/GPT models with FAISS retrieval and a FastAPI service on AWS. Demonstrates strong real-world reliability work (document cleanup, hallucination controls, monitoring/drift tracking with MLflow) and close collaboration with non-technical clinical operations teams via demos and feedback-driven iteration.”
Mid-level Data Scientist & AI/ML Engineer specializing in GenAI and cloud ML
“GenAI/LLM engineer who recently built a production compliance assistant at State Farm for KYC/AML and regulatory teams, using AWS Bedrock + LangChain with Textract/Lambda pipelines to extract fields, tag risk, and summarize long documents. Implemented RAG, strict structured outputs, and human-in-the-loop guardrails, and reports automating ~80% of documentation work while reducing review time by ~40%.”
Senior AI/ML Engineer and Data Scientist specializing in Generative AI and MLOps
“ML/NLP practitioner focused on financial-services document intelligence and compliance workflows—built an end-to-end pipeline to classify documents and extract financial entities from loan applications, emails, and statements stored in S3/internal databases. Strong in entity resolution/record linkage and in productionizing pipelines with GitHub Actions CI/CD, testing, data validation, and Docker, plus semantic search using OpenAI embeddings and a vector database.”
Junior ML Data Associate specializing in AI training data and LLM prompt evaluation
“Applied ML/embodied AI practitioner who built an on-device gesture-control system for smart-home lights using Raspberry Pi + camera, focusing on privacy-preserving real-time inference and hardware-constrained optimization (async pipeline + TF Lite INT8). Also made a high-impact architecture decision for an ML content evaluation/QA pipeline processing millions of annotated text samples weekly, reducing batch runtime from ~6 hours to ~40 minutes while lowering compute cost.”
Mid-level Data Analyst specializing in cloud ETL, BI, and machine learning
“Data/ML practitioner with experience at UnitedHealth Group building a fraud claims detection solution combining structured claims data and unstructured notes, validated with compliance stakeholders to improve actionable accuracy. Also applied embeddings, vector databases, and fine-tuned language models in a Bank of America capstone to detect threats/anomalies in financial documents, with production-minded Python ETL workflows using Airflow.”