Pre-screened and vetted.
Mid-Level Full-Stack Software Engineer specializing in AI-enabled web platforms
“Backend/AI engineer in construction tech (HyperWater AI) who delivered major production performance wins (analytics API from ~1 hour to 0.5s) and shipped LLM features for parsing subcontractor manifests into CSI divisions with human-in-the-loop review. Also built a freelance agentic document-verification system using OCR + RAG over pgvector with robust retry/escalation logic and user feedback loops.”
Mid-level Full-Stack & AI Engineer specializing in LLM applications
“Full-stack engineer who has shipped and operated generative-AI chat/QA features end-to-end, including a RAG-based pipeline with guardrails and cost/latency monitoring in production. Experienced with React/TypeScript + Node/Postgres architectures, Dockerized deployments to AWS (EC2) via GitHub Actions CI/CD, and building reliable ingestion/ETL systems with idempotency, backfills, and reconciliation.”
Junior Full-Stack/AI Engineer specializing in web platforms and LLM applications
“Backend engineer from FoodSupply.ai who built and evolved a scalable restaurant/supplier product and order management platform using Node.js and REST APIs. Implemented a hybrid MySQL+MongoDB data architecture, optimized performance with Redis/Prisma, and led a phased migration with feature flags and a temporary sync layer to maintain data consistency. Strong focus on production security (OAuth2, RBAC, row-level security, AWS IAM) and reliability practices (testing with Pytest, Docker/AWS pipelines).”
Mid-level AI/ML Software Engineer specializing in GPU-optimized LLM inference and cloud microservices
“Built and deployed a production RAG-based multilingual analytics assistant for healthcare operations, enabling non-technical teams to query claims/EHR and risk metrics with grounded explanations. Demonstrates strong end-to-end LLM system engineering (retrieval tuning, re-ranking, hallucination controls, verification layers) plus workflow orchestration (Airflow/Composer/Step Functions) and stakeholder-driven iteration via prototypes and dashboards.”
Mid-level AI/ML Engineer specializing in LLM agents, RAG retrieval, and IoT ML systems
“Built production LLM-driven products including a job-hunt AI (job ranking + resume optimization) and an InterviewAI agentic pipeline using LangChain. Focused on practical deployment concerns like securing OpenAI usage via rate limiting and tiered quotas, and demonstrates an applied approach to choosing models, retrieval methods (RAG), and prompting strategies.”
Junior Machine Learning Engineer specializing in LLM agents, RAG, and MLOps
“AI/ML engineer who has shipped production systems across computer vision and conversational agents: built a YOLOv8-based wheel fitment pipeline at a Techstars-backed automotive startup, focusing on sub-second latency, monitoring, and robust fallback mechanisms that drove 2–3x page view growth and +5–6k users. Also built a voice-based interview platform orchestrating Deepgram + GPT-4 Mini + OpenAI TTS with FSM-driven reliability, and has hands-on RAG experience (LangChain, hybrid retrieval, cross-encoder reranking, custom pseudo-query generation).”
Mid-level GenAI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production RAG-based LLM assistant that answers day-to-day operational questions from internal PDFs/SOPs, with strong emphasis on data consistency (metadata versioning, confidence thresholds, conflict handling) and low-latency retrieval at scale. Experienced designing and orchestrating multi-agent LLM workflows (retrieval/validation/generation) and pipeline orchestration for ingestion/embedding/vector-store updates, plus iterative delivery with non-technical operations/business stakeholders.”
Junior AI Engineer specializing in LLM evaluation, prompt engineering, and AI orchestration
“LLM workflow builder who has deployed a personalized GPT experience (including Delphi AI-based knowledge ingestion) and built a LangChain/LangGraph job-aggregation pipeline that ingests, normalizes/dedupes, filters, then uses an LLM to rank and summarize matches. Emphasizes production reliability with structured outputs, retries/fallbacks, metric-driven evaluation, logging/prompt versioning, and A/B testing, and collaborates with non-technical stakeholders through demo-driven iteration.”
Junior Machine Learning & Backend Engineer specializing in LLM systems and ML infrastructure
“Built and deployed production RAG-based document search/Q&A systems (DocChat and an internship marketing RAG), using a React + FastAPI stack on GCP with docs stored in GCP buckets and retrieval via embeddings/vector DB. Emphasizes cost/performance tradeoffs (reported ~40% cost reduction) and ships via Docker (Railway), with load/API testing using JMeter and Swagger; regularly collaborates with a CEO stakeholder to iterate and push changes to production.”
Junior AI/ML Software Engineer specializing in Generative AI and scalable data pipelines
“Built and operated large-scale biodiversity/ecological research platforms, integrating 50+ heterogeneous global datasets into a unified BIEN 3 schema on PostgreSQL/PostGIS and improving data consistency by 35%. Strong production engineering background (Linux monitoring, CI/CD performance gates, Docker on AWS/Azure) plus applied AI work building a Python RAG system (0.90 precision) and halving latency with Elasticsearch.”
Senior Software Engineer specializing in full-stack systems, big data, and applied AI
“Built and deployed ForensicLLM, a local domain-specific LLaMA-3.1-8B model for digital forensic investigators using RAFT + RAG over 1000+ curated research papers, with citation-aware responses and rigorous evaluation (BERTScore/G-Eval). Deployed via vLLM and Docker and validated through a chatbot survey with 80+ participants; published at DFRWS EU 2025.”
Junior AI/ML Engineer specializing in Generative AI, NLP, and MLOps
“LLM engineer who has deployed a production RAG system (LangChain/FAISS/FastAPI) for enterprise semantic search, tackling real-world latency by LoRA/PEFT fine-tuning and grounding outputs with retrieval. Brings strong MLOps (Docker, AWS EKS, CI/CD, MLflow) plus stakeholder-facing explainability experience using SHAP to align ML-driven financial guidance with non-technical domain experts.”
Mid-Level Software Engineer specializing in cloud-native microservices and full-stack development
“Full-stack engineer with deep startup experience building products from scratch under ambiguous requirements. Delivered a scalable, admin-configurable notification platform (Spring Boot/Java/Kafka) supporting 50+ notification types across 3 channels for 10k+ users, cutting new notification setup to ~5 minutes. Also built a Tinder-meets-LinkedIn job-swiping app (React/TS + Node/Prisma) and has hands-on AWS production ops (ECS/EKS, RDS, CloudWatch) plus multiple third-party integrations (Stripe, QuickBooks, Twilio).”
Senior Full-Stack AI/ML Engineer specializing in MLOps and GenAI
“Senior backend/data engineer who has built and maintained HIPAA-compliant, real-time clinical FastAPI services on AWS, orchestrating ML/LLM and vector DB calls with strong reliability patterns (auth, timeouts/retries, graceful degradation, idempotency). Also delivered AWS IaC/CI-CD (Terraform/Helm/GitHub Actions) across EKS/Lambda/SageMaker and built Glue/Spark ETL with schema evolution and data quality controls, plus demonstrated large SQL performance wins (15 min to <9 sec) and hands-on incident ownership.”
Mid-Level Applied AI Engineer specializing in LLM services, RAG, and OCR/NLP extraction
“Backend/platform engineer who built and evolved a large-scale healthcare document processing system (OCR + LLM orchestration) in Python/FastAPI on Google Cloud (Cloud Run, GCS, Firestore), processing ~1.5M files per batch and tens of millions overall. Emphasizes reliability and operational safety via deterministic IDs, idempotent state machines, strong observability, and self-healing reconciliation, plus disciplined migrations using dual-run validation and incremental rollouts.”
Entry-Level Computer Vision Research Assistant specializing in medical imaging AI
“New grad who shipped an LLM-powered writing app (“Write-it”) to production on Azure with CI/CD (GitHub Actions + JFrog) and implemented an unconventional RAG pipeline to prevent repetitive prompts using embeddings and cosine similarity. Also participated in a Luma AI image/video generation hackathon, iterating with artist feedback and improving usability by rewriting non-technical prompts via an LLM.”
Junior Backend Engineer specializing in cloud APIs and AI-enabled systems
“Built and shipped "OnCall Copilot," a production Slack-based RAG assistant that answers on-call questions from runbooks and postmortems with citations using a FAISS vector index. Emphasizes reliability and measurable performance via strict guardrails ("no evidence, no answer"), evaluation metrics, drift monitoring, and operational hardening with Docker, logging, health checks, and offline fallback.”
Mid-level Full-Stack & XR Developer specializing in GenAI and immersive AR/VR systems
“Built and deployed a "personal second brain" product (CloneMind) with an end-to-end RAG pipeline for retrieving information across PDFs, URLs, images, and audio using Next.js/Node.js/Postgres/Supabase/Redis. Demonstrates strong practical depth in retrieval quality tuning, latency reduction via caching, and stateful orchestration with LangChain/LangGraph, plus experience persuading a non-technical professor stakeholder by shipping a working prototype.”
Entry-Level GenAI/LLM Engineer specializing in agentic systems and RAG
“LLM/AI agent engineer with consulting/contract experience (Kanhaiya Consulting LLC) who deployed a production AI agent to automate BIM list workflows end-to-end—from database understanding and data cleaning to automated visualizations/dashboards. Worked around restricted real-time data access by generating synthetic data and improving outputs via supervised fine-tuning, and uses AWS-based LLMOps observability (Opic/OPEC) plus hybrid retrieval (vector+BM25 with reranking) to optimize relevance, latency, and cost.”
Junior Full-Stack Software Engineer specializing in AI/ML platforms and microservices
“Graduate-school lab engineer who built and owned the final architecture of a Microservices Hub that integrates REST APIs, issues API keys, monitors 10+ Linux servers, and visualizes service dependencies via a topology graph. Strong in bridging legacy and modern stacks (Dockerized and non-Dockerized services like Apache/screen) using deep Linux/networking knowledge, plus practical real-time audio streaming for STT/TTS and experience mentoring others.”
Senior Full-Stack & AI Developer specializing in Python/React, AWS, and LLM/RAG systems
“Backend Python engineer who owned the full backend build of an AI-driven platform for UK golf clubs, including FastAPI microservices, vector search, and a tuned LangChain+Pinecone RAG pipeline focused on cost and hallucination reduction. Experienced deploying Django/FastAPI/Flask stacks on AWS-backed Kubernetes with GitOps/ArgoCD-style delivery, plus executing legacy-to-AWS migrations and building Kafka-based real-time analytics pipelines.”
Junior Solutions Engineer specializing in full-stack automation and LLM prompt engineering
“Built and productionized an LLM-powered customer support system using a RAG architecture with structured document ingestion, embedding retrieval, and prompt templates for product-specific grounding. Experienced diagnosing live agent/workflow failures (e.g., retrieval regressions after new docs) by refactoring ingestion/chunking and adding grounding constraints plus evaluation benchmarks. Also supports go-to-market by joining discovery calls, shaping MVP workflows into demos/prototypes, and creating post-launch documentation to drive adoption.”
Junior Full-Stack Software Engineer specializing in Node.js, React, and REST APIs
“Full-stack engineer who shipped and owned a production Document Chat feature built with Next.js App Router/TypeScript and a Node/Express RAG backend, including JWT-secured route handlers and streaming responses. Demonstrated strong post-launch ownership by improving latency (~30%) via MongoDB indexing/query optimization and reducing AI costs through caching, backed by profiling with React Profiler and Chrome DevTools.”
Intern Full-Stack Engineer specializing in AI-powered products
“Software engineer (internship experience) who built and owned an AWS serverless multi-user “challenge” feature end-to-end (UI + REST APIs + DynamoDB + deployment), delivering measurable gains in latency (-30%), debugging time (-50%), and join drop-offs (~-30%). Also productionized a multilingual RAG-based QA system with vector retrieval and guardrails, improving accuracy to ~85% and driving ~20% DAU growth.”