Pre-screened and vetted.
Mid-level GenAI Engineer specializing in production AI agents and evaluation pipelines
“Built and shipped a production LLM-powered internal operations automation platform using LangChain RAG (Pinecone) and FastAPI microservices, deployed on AWS EKS, serving 10k+ daily interactions. Implemented a rigorous evaluation/observability stack (golden datasets, prompt regression tests, MLflow, retrieval metrics, hallucination monitoring) that drove hallucinations below 2% and improved reliability, and partnered closely with non-technical ops leaders to cut manual lookup work by 60%+.”
Mid-level Data & GenAI Engineer specializing in lakehouse, streaming, and RAG platforms
“Built a production internal LLM-powered knowledge assistant using a RAG architecture (Python, LLM APIs, cloud services) that answers employee questions with sourced, grounded responses from internal documents. Demonstrates strong practical depth in retrieval tuning (chunking/metadata filters), orchestration with LangChain, and production reliability practices (latency optimization, automated embedding refresh, evaluation metrics, logging/monitoring) while partnering closely with non-technical operations teams.”
Entry-Level Software Engineer specializing in AI/ML and Full-Stack Development
“Backend engineer who built an NL-to-SQL system at Target, using a multi-step LLM pipeline with vector-store schema retrieval and SQL validation to safely answer business questions. Strong in production FastAPI systems (async, Pydantic, Docker/Uvicorn, load balancing) and security (OAuth2/JWT, scopes, and database row-level security), with experience migrating Flask apps to FastAPI + PostgreSQL using strangler/feature-flagged canary rollouts.”
Mid-level AI Engineer specializing in LLMs, RAG, and content automation
“AI/LLM engineer who built a production autonomous GenAI content ecosystem that generates short-form scripts, extracts viral highlights from long-form video, and dubs content into 33+ languages. Focused on making LLM outputs production-safe via schema enforcement, token-to-time alignment, critic-agent verification, and scalable async orchestration—cutting manual workflows by ~90% and saving $200k+ annually.”
Principal Software Architect specializing in AI/ML and cloud-native full-stack platforms
“AI/LLM engineer who built a production content-generation system for nursing education, combining multimodal RAG over proprietary PDFs (including images) with structured Cosmos DB data and external sources. Strong focus on production reliability—prompt-chaining with LangChain, validation/guardrails, and Azure-based monitoring/observability—plus experience designing Azure AI agents with tool integrations like Bing Search.”
Junior Data Scientist specializing in fraud analytics and cloud data platforms
“Built and deployed production LLM-powered document summarization/classification systems using embeddings, vector databases (RAG-style retrieval), and automated evaluation (BERTScore/ROUGE), with a focus on monitoring and scalable cloud pipelines. Also partnered with a fraud analytics team to deliver a transaction anomaly detection solution, translating model outputs into Power BI dashboards and actionable KPIs while iterating on thresholds and alerts based on stakeholder feedback.”
Senior Software Engineer specializing in Generative AI product development
“AI product builder at Padlet who shipped multiple production LLM features for education workflows, including an AI document generator (AI Recipes) and a RAG-enabled in-product chat assistant. Built an AI microservice layer (LangChain) to swap model providers easily and created automated + human-in-the-loop evaluation systems (including ~100-test runs) to iterate on prompts and quality.”
Mid-Level Software Engineer specializing in AI-enabled backend and full-stack web systems
“Backend/AI workflow engineer with experience at AirKitchenz, Uber, and Vivma Software, building production systems on AWS (Lambda, DynamoDB, Step Functions). Has a track record of major performance wins (DynamoDB latency 2s to <150ms; Postgres query 2s to ~180ms) and shipping LLM-powered onboarding and ticket-routing workflows with strong guardrails (schema validation, confidence thresholds, human-in-the-loop escalation).”
Mid-level AI/ML Engineer specializing in NLP, computer vision, and Generative AI
“Built and deployed a production LLM-powered clinical insights/summarization assistant for healthcare teams, including a Spark+Airflow pipeline, fine-tuned transformer models, and a FastAPI Docker service on AWS. Demonstrates strong MLOps/LLMOps depth (Airflow on Kubernetes, custom AWS operators/IAM, MLflow, CloudWatch) and practical reliability work like hallucination mitigation, confidence scoring, and retrieval-backed evaluation with shadow deployments.”
Mid-level Data Engineer specializing in cloud data platforms, Spark, and streaming pipelines
“Data/MLOps engineer (Cognizant background) who owned an AWS/Airflow/Snowflake healthcare transactions pipeline processing ~8–10M records/day and cut pipeline/data-quality incidents by ~33%. Also built and deployed a production FastAPI model-inference service on Kubernetes (Docker, HPA) with strong observability (Prometheus/Grafana), versioned endpoints, and resilient backfill/idempotent external data ingestion patterns.”
Intern Software Engineer specializing in Applied AI and LLM systems
“Built and deployed a production RAG-based conversational "Yelp for AI tools" at Search-AI Inc., focused on personalized, explainable AI tool recommendations from thousands of options. Emphasizes production-grade reliability and performance (hybrid retrieval, async two-stage pipelines) and is also building a multi-agent orchestration layer (MAgIc) with typed memory and controlled coordination policies.”
Mid-level Machine Learning Engineer specializing in LLM agents, RAG, and MLOps
“Built production LLM systems including a real-time customer feedback analysis and workflow automation platform using RAG and multi-agent orchestration with confidence-based human escalation, addressing privacy and legacy integration challenges. Also automated ML operations with Airflow/Kubernetes (e.g., daily churn model retraining) cutting retraining time to under 30 minutes, and demonstrates a rigorous testing/monitoring approach plus strong non-technical stakeholder collaboration.”
Mid-level Data Scientist / ML Engineer specializing in secure GenAI and financial compliance
“Built a production "sentinel insight engine" to tame information overload from millions of product reviews and support transcripts, combining Azure OpenAI (GPT-3.5) zero-shot classification with a fine-tuned T5 summarizer to generate weekly actionable product insights. Demonstrated strong MLOps/production engineering by adding drift monitoring with embedding-based detection, integrating REST with legacy SOAP/queue-based CRM via FastAPI middleware, and scaling reliably on Kubernetes with HPA.”
Mid-level Data Scientist & AI Engineer specializing in RAG, agentic AI, and production ML
“AI/data engineer who built a production LLM-powered schema drift detection system (LangChain/LangGraph) to catch semantic data changes before they break downstream analytics/ML. Deployed on AWS with Docker/S3 and implemented an LLM-as-a-judge evaluation framework to improve trust, reduce hallucinations, and control false positives/alert fatigue. Collaborated with non-technical risk/business analytics stakeholders at EY by delivering human-readable drift explanations that improved confidence in financial analytics dashboards.”
Senior Full-Stack & GenAI Engineer specializing in healthcare and financial services
“Built and deployed a production LLM-powered customer support assistant using a RAG backend in Python, focused on deflecting repetitive Tier-1 tickets and reducing resolution time. Demonstrates strong production engineering instincts around reliability (confidence scoring + human fallback), scalability/cost optimization (multi-stage pipelines), and workflow orchestration/observability (LangChain, custom DAGs, structured logging, step metrics).”
Junior AI Software Engineer specializing in LLMs, RAG, and agent workflows
“Backend/ML-leaning engineer who built a content-based event recommender for FlowMingle using embeddings + HNSW vector search on Google Cloud, with Firebase as the backend and a managed recommendation lifecycle (15 recs/user, daily async generation, weekly deletion) now serving 1500+ users. Also led a cost-driven migration of ConvAI services to Azure AI using parallel request testing from a Unity client, with post-migration monitoring via logs and model evals; contributed to a Massachusetts law-enforcement conversation analysis system by expanding ingestion to PDF/TXT/Excel and multi-file inputs.”
Senior AI Engineer specializing in Generative AI and RAG applications
“AI engineer who has shipped production LLM systems across customer service and marketing use cases—building a RAG app on Azure OpenAI and speeding retrieval with Redis caching tied to Okta sessions. Also implemented a LangGraph multi-agent workflow that pulls image context from Figma to generate structured HTML marketing emails, adding a verification agent to improve image-selection accuracy while optimizing solution cost for business stakeholders.”
Mid-level Machine Learning Engineer specializing in LLM systems and healthcare data automation
“React performance-focused engineer who contributed performance patches back to an open-source context+reducer state helper after profiling and fixing excessive re-renders in an enterprise project management platform at Easley Dunn Productions. Also built an end-to-end LLM-driven pipeline at Prime Healthcare to normalize millions of supply-chain records, reducing defects by 80% and saving 160+ hours/month.”
Mid-level AI/ML Engineer specializing in healthcare ML and LLM/RAG systems
“AI/LLM engineer with recent production experience at UnitedHealth Group building an end-to-end RAG system over structured EMR data and unstructured clinical notes, including evidence retrieval, GPT/LLaMA-based reasoning, and a validation layer for reliability. Strong in orchestration (Kubeflow/Airflow/MLflow), prompt engineering for noisy healthcare text, and rigorous evaluation/monitoring with gold-standard benchmarking, plus close collaboration with clinical operations stakeholders.”
Mid-level AI/ML Engineer specializing in Generative AI and NLP
“AI/LLM engineer with production experience building secure, scalable compliance-focused generative AI systems (GPT-3/4, BERT) including RAG over internal regulatory document bases. Has delivered end-to-end pipelines on AWS with PySpark/Airflow/Kubernetes/FastAPI, emphasizing privacy controls, monitoring, and iterative evaluation (A/B testing). Also partnered closely with bank compliance officers using prototypes to refine NLP summarization/classification and reduce document review time.”
Mid-level AI/ML Engineer specializing in GenAI agents, RAG pipelines, and MLOps
“AI/ML engineer who built a production RAG-based internal document intelligence assistant (LangChain + Pinecone) to let employees query enterprise reports in natural language. Demonstrated hands-on pipeline orchestration with Apache Airflow and tackled real production issues like retrieval grounding and latency using tuning, caching, and token optimization, while partnering closely with non-technical business stakeholders through iterative demos.”
Mid-level Data Scientist specializing in Generative AI, RAG systems, and ML engineering
“AI/LLM engineer who built a production QA RAG for a University of Massachusetts faculty success initiative, cutting service tickets by 70%. Strong end-to-end RAG implementation skills (LangChain, Qdrant, hybrid/HyDE retrieval, FastAPI) with rigorous evaluation (RAGAS, LLM-as-judge) and practical handling of constraints like API rate limits and cost. Prior cross-functional delivery experience collaborating with SMEs and business owners at TCS and IBM.”
Mid-Level Software Engineer specializing in cloud, microservices, and AI/ML
“Backend/API engineer with ~4 years experience building production services in .NET Core/PostgreSQL/Redis/Docker and optimizing real-world latency issues (claims ~60% response-time improvement). Also built and owned an end-to-end RAG-based AI assistant using Python/FastAPI, OpenAI APIs, and Pinecone, plus agentic workflows with reliability guardrails (retries, confidence thresholds, monitoring). Currently pursuing a master’s degree and targeting a $150k base salary.”
Senior AI/ML & Data Engineer specializing in Generative AI and RAG systems
“GenAI/RAG engineer who has deployed a production policy/regulatory search assistant for a financial client using LangChain + Vertex AI, FastAPI, Docker/Kubernetes, and Airflow-orchestrated data pipelines. Demonstrated measurable impact with 50–60% latency reduction and 70% fewer pipeline failures, plus KPI-driven grounding evaluation (90%+ target) and strong cross-functional collaboration with compliance/business teams.”