Pre-screened and vetted.
Mid-level Data Engineer specializing in cloud data pipelines and full-stack analytics
Mid-level Data Engineer specializing in cloud data platforms and streaming pipelines
Mid-level Data Engineer specializing in cloud data pipelines and streaming analytics
Mid-level Data Engineer specializing in AWS data platforms and streaming pipelines
Senior Data Engineer specializing in AWS-based data pipelines and multi-tenant SaaS
Junior Data Engineer specializing in data pipelines and streaming ingestion
“Backend/data platform engineer who owned a near-real-time patient feedback ingestion system, building a FastAPI + Kafka service with Snowflake/Airflow orchestration. Demonstrates strong production Kubernetes/GitOps practices on AWS EKS (Helm, Argo CD, Sealed Secrets) and solved real-time data integrity issues via idempotent processing with Redis.”
Mid-Level Data Engineer specializing in cloud data pipelines and big data platforms
“Data engineer with ~4 years of experience building Python-based data ingestion/processing services and real-time streaming pipelines (Kafka/PubSub + Spark Structured Streaming). Has deployed containerized data applications on Kubernetes with GitLab CI/Jenkins pipelines and applied GitOps to cut deployment time ~40% while reducing config drift. Also supported a legacy on-prem data warehouse/backend migration to GCP using phased migration and parallel validation to meet strict reliability/SLA needs.”
Mid-level Data Engineer specializing in cloud data pipelines and analytics engineering
“Built and deployed a production LLM-powered demand and churn forecasting system for an e-commerce client, combining open-source LLMs (LLaMA/Mistral) and Sentence-BERT embeddings to generate business-friendly explanations of forecast drivers. Strong focus on data quality and model trust (validation, baselines, segmented monitoring) and production reliability via Airflow-orchestrated pipelines with readiness checks, retries, and ongoing drift/A-B testing.”
Mid-level Data Scientist specializing in ML, data engineering, and real-time analytics
Senior Full-Stack Software Engineer specializing in web apps, integrations, and data pipelines
Junior Data Engineer specializing in cloud ETL/ELT and lakehouse platforms
Mid-level Generative AI & ML Engineer specializing in production LLM and RAG systems
“AI/ML engineer who shipped a production blood-test report understanding and personalized supplement recommendation product, using a LangGraph multi-agent pipeline on AWS serverless with OCR via Bedrock and RAG over vetted clinical research. Also built end-to-end recommender system pipelines at ASANTe using Airflow (ingestion, embeddings/features, training, registry, batch scoring/monitoring) with KPI reporting to Tableau, with a strong focus on safety, evaluation, and measurable reliability.”
Mid-level Software/Data Engineer specializing in LLM apps, RAG pipelines, and cloud microservices
“Backend/data engineer who built an enterprise LLM assistant (AI Genie) at Broadband Insights using a LangChain + GPT-4 + Pinecone RAG pipeline to automate broadband analytics reporting. Developed Python/Dagster ETL processing 10M+ records/day and improved data freshness by 60%, with production-grade scalability patterns (async workers, containerized microservices, Kubernetes) and strong multi-tenant isolation practices.”
Junior Full-Stack Software Engineer specializing in web apps, data workflows, and AI integrations
“Backend engineer with experience stabilizing data processing/analytics pipelines and refactoring brittle backend APIs. Has hands-on FastAPI work emphasizing strong validation (Pydantic), clear layering, and secure JWT-based auth with role/row-level controls, plus pragmatic migration tactics like parallel runs to protect data integrity.”
Mid-level GenAI/Data Engineer specializing in LLMs, RAG systems, and fraud detection
“ML/NLP engineer with banking domain experience who built a GenAI-powered fraud detection and risk intelligence system at Origin Bank, combining RAG (LangChain + FAISS), fine-tuned BERT NER, and GPT-4/Sentence-BERT embeddings. Delivered measurable impact (25% higher fraud detection accuracy, 40% less manual review) and emphasizes production-grade pipelines on AWS SageMaker/Airflow with strong data validation and scalable PySpark processing.”
Mid-level AI Engineer specializing in AI agents, RAG pipelines, and LLM evaluation
“Built and shipped production LLM systems at Founderbay, including a low-latency voice agent and a graph-based multi-agent research assistant. Strong focus on reliability in real workflows—hybrid SERP + full-site scraping RAG, grounding guardrails, validation checkpoints, and transcript-driven evaluation—plus performance tuning with async FastAPI, Redis caching, and containerization. Also partnered with a non-technical ops lead to automate post-call follow-ups via call summarization, field extraction, and tool-triggered actions.”
Mid-level Data Engineer specializing in AI/ML, RAG systems, and cloud data pipelines
“Built a production lead-generation system using AI agents that researches the internet for relevant leads and integrates RAG-based contact enrichment/shortlisting aligned to existing CRM data, enabling sales reps to focus more on selling. Also has hands-on AWS data orchestration experience (Glue, Step Functions) moving raw data into Redshift and evaluates agent performance with human-in-the-loop plus BLEU/perplexity metrics.”
Mid-level Software Engineer specializing in cloud data platforms and serverless ETL
“Data/ML engineer from HCLTech who modernized enterprise data by linking fragmented financial and supply-chain data across SAP/SQL Server/Snowflake using NLP entity linking and embeddings (FAISS). Delivered measurable impact including ~40% reduction in manual error-log triage and entity-linking accuracy improvements from ~86% to ~93%, with results surfaced in Power BI for real-time analytics.”
Junior Backend Software Engineer specializing in search, data systems, and LLM applications
“Built a contract and customer documentation retrieval solution for Urban Studio, designing a RAG + Elasticsearch hybrid search stack (RRF + cross-encoder reranking) with a strong emphasis on chunking/data quality and hallucination reduction. Experienced in diagnosing LLM workflow issues via observability traces and tailoring technical demos to developer concerns like reliability and high concurrency.”
Mid-level Full-Stack Engineer specializing in cloud-native FinTech analytics
“Full-stack/ML-leaning engineer who has shipped production-grade real-time analytics and an internal AI support assistant using RAG over enterprise documentation. Demonstrates strong systems thinking across scalability, reliability, observability, and LLM safety/evaluation (thresholded retrieval, RBAC, response validation, regression-gated evals), with concrete iteration based on performance metrics and user feedback.”
Mid-level Software & ML Engineer specializing in agentic LLM systems and ML infrastructure
“Built and deployed an LLM-to-SQL automation system in a closed/internal environment, using a retriever–reranker–validator architecture on Kubernetes with strong security controls (semantic + rule-based validation and RBAC), achieving 99% uptime and cutting manual query time ~40%. Also worked on genomic sequence classification and semantic search workflows, orchestrating data prep with Airflow, tracking/deploying with MLflow, and optimizing distributed multi-GPU training on a university Kubernetes cluster.”
Mid-level Software/Data Engineer specializing in cloud ETL pipelines and data infrastructure
“Backend/data engineer who built a production analytics data service (Python/FastAPI on AWS/Postgres with PySpark ETL) handling millions of records per day and drove major latency improvements (10–15s to <2s) via indexing, Redis caching, and shifting aggregations into ETL. Also shipped an LLM-based natural-language-to-SQL assistant end-to-end with strong guardrails (schema restrictions, read-only validation, RBAC, masking) and designed a multi-step agent workflow with verification and fallback logic.”
Junior Data Engineer specializing in cloud data pipelines and warehousing