Pre-screened and vetted.
Intern AI & Machine Learning Engineer specializing in computer vision and edge deployment
“Built and shipped a real-time AI robotic inspection system, using a synthetic data generation pipeline to address rare edge cases—cutting data collection costs ~60% and boosting hard-scenario accuracy ~20%. Experienced in productionizing ML on constrained Jetson hardware and orchestrating end-to-end ML workflows with Airflow/Docker/Kubernetes, with a metrics-driven approach to reliability, evaluation, and stakeholder communication.”
“Forward Deployed Engineer at EasyBee AI who productionized a self-storage customer’s multi-agent LLM system end-to-end—rebuilding it with LangGraph/CrewAI, integrating with real property management + CRM systems via an MCP server, and adding observability/guardrails for reliable daily use. Experienced in live troubleshooting of agentic workflows, developer demos/workshops (including an open-source project, MerryQuery), and partnering with sales to close deals through customer-specific technical demos and fast integration feedback loops.”
Mid-level AI Engineer specializing in Generative AI, LLMs, and RAG
“Internship at Discovery Education building a production LLM/RAG chatbot that let marketing and sales teams query and interpret Looker/BI dashboards in natural language, with responses grounded in compliance and state education standards. Emphasizes rigorous evaluation (faithfulness/precision/recall/latency) plus user-feedback analytics, and used LangChain for orchestration, chunking/context-window control, and integration with enterprise sources like SharePoint.”
Junior Backend/Platform Engineer specializing in cloud-native APIs and data systems
“Startup-style full-stack/backend engineer with hands-on AWS architecture experience who shipped an LLM-driven assessment-question automation feature (Python microservice calling AWS Bedrock via SQS, deployed on Lambda) with strong validation/guardrails and retry strategies. Also improved production scalability by moving a CPU/IO-heavy file upload path out of a Go API into a queue/Lambda design monitored with CloudWatch, and has React+TypeScript experience optimizing analytics dashboards.”
Mid-Level Full-Stack/Product Engineer specializing in B2B SaaS and AI search systems
“Full-stack engineer operating in early-stage, high-velocity environments (OpGov.AI/UST Calibrate) who ships production Next.js App Router features end-to-end (RSC, Server Actions, SEO, RBAC, caching) and owns performance post-launch. Demonstrates strong data/infra depth—designed Postgres JSONB-based event models for DevOps/DORA analytics and tuned queries from ~2s to <50ms, plus built durable ingestion workflows with retries and idempotency on Azure.”
Senior Software Engineer specializing in Backend Systems and Generative AI (RAG)
“Backend engineer with experience building an end-to-end civic tech AI platform that ingests city council meeting videos, transcribes them with Whisper, and enables natural-language Q&A via a LangChain/FAISS RAG pipeline. Demonstrated strong systems thinking by tuning retrieval for accuracy/latency/memory (cutting response time ~3s→1s and memory ~500MB→25MB) and by safely migrating an ERP from monolith toward services using dual writes, reconciliation, and idempotency to protect financial workflows.”
Mid-level AI/ML Software Engineer specializing in GPU-optimized LLM inference and cloud microservices
“Built and deployed a production RAG-based multilingual analytics assistant for healthcare operations, enabling non-technical teams to query claims/EHR and risk metrics with grounded explanations. Demonstrates strong end-to-end LLM system engineering (retrieval tuning, re-ranking, hallucination controls, verification layers) plus workflow orchestration (Airflow/Composer/Step Functions) and stakeholder-driven iteration via prototypes and dashboards.”
Junior Machine Learning Engineer specializing in LLM agents, RAG, and MLOps
“AI/ML engineer who has shipped production systems across computer vision and conversational agents: built a YOLOv8-based wheel fitment pipeline at a Techstars-backed automotive startup, focusing on sub-second latency, monitoring, and robust fallback mechanisms that drove 2–3x page view growth and +5–6k users. Also built a voice-based interview platform orchestrating Deepgram + GPT-4 Mini + OpenAI TTS with FSM-driven reliability, and has hands-on RAG experience (LangChain, hybrid retrieval, cross-encoder reranking, custom pseudo-query generation).”
Mid-level AI/ML Engineer specializing in Generative AI and RAG systems
“Currently at ProShare and reports building an AI/LLM-powered system deployed to production, aimed at helping with status-related difficulties and reducing misunderstandings across transactions. Also cites prior collaboration at Porsche with marketing teams, focusing on translating marketing goals into technical requirements and communicating solutions clearly to non-technical stakeholders.”
Mid-level GenAI Engineer specializing in LLM agents and RAG systems
“Built and deployed a production RAG-based LLM assistant that answers day-to-day operational questions from internal PDFs/SOPs, with strong emphasis on data consistency (metadata versioning, confidence thresholds, conflict handling) and low-latency retrieval at scale. Experienced designing and orchestrating multi-agent LLM workflows (retrieval/validation/generation) and pipeline orchestration for ingestion/embedding/vector-store updates, plus iterative delivery with non-technical operations/business stakeholders.”
Junior Full-Stack Software Engineer specializing in React, Node.js, AWS, and Generative AI
“Built and production-deployed a Streamlit-based PDF RAG chatbot using LangChain (FAISS, embeddings, prompt templates) and OpenAI, optimizing Streamlit’s stateless behavior by caching vector DB + chat history to cut latency and API cost. Demonstrates a rigorous evaluation mindset (gold datasets, unit tests, LLM-as-judge, groundedness KPIs) and has experience communicating privacy/accuracy safeguards (RBAC, data masking, citations) to a non-technical client at Kalven Technologies.”
Junior Machine Learning & Backend Engineer specializing in LLM systems and ML infrastructure
“Built and deployed production RAG-based document search/Q&A systems (DocChat and an internship marketing RAG), using a React + FastAPI stack on GCP with docs stored in GCP buckets and retrieval via embeddings/vector DB. Emphasizes cost/performance tradeoffs (reported ~40% cost reduction) and ships via Docker (Railway), with load/API testing using JMeter and Swagger; regularly collaborates with a CEO stakeholder to iterate and push changes to production.”
Junior Data/AI Engineer specializing in MLOps, real-time pipelines, and LLM applications
“Built an LLM-driven MLOps agent at SBD Technologies that automated an EV-charging prediction workflow end-to-end, integrating with real-time Kafka/FastAPI systems supporting 120K+ chargers at 99.99% event delivery. Addressed frequent schema drift by implementing SQLAlchemy/Flyway validation (60% reduction in drift issues) and deployed as Kubernetes microservices with GitHub Actions CI/CD; also has Airflow-based ingestion/crawling experience into Snowflake and stakeholder-facing delivery via a Fleetcharge PWA.”
Junior Software Engineer specializing in ML, RAG systems, and safety-critical risk modeling
“Backend/cloud engineer from Resilient Tech with hands-on experience deploying REST APIs and database migrations into a live ERP used by real customers while maintaining 99% uptime. Has debugged intermittent AWS container timeouts down to security group/load balancer misconfigurations, and has extended Python in an ERPNext system to meet GST/e-invoicing compliance requirements with strong customer collaboration.”
Intern Data Scientist specializing in GenAI agents, RAG, and ML platforms
“LLM/agent systems builder who deployed a production hybrid router for immerso.ai that dynamically selects retrieval vs reasoning vs generative pathways, achieving an 82% factual-accuracy lift. Deep hands-on experience optimizing local Mistral 7B inference (4–5 bit GGUF quantization, KV-cache reuse) and building reliable RAG/agent workflows with LangChain/LangGraph/AutoGen across GCP Cloud Run and AWS (ECS/Lambda).”
Junior Full-Stack Software Engineer specializing in cloud-native web apps and AI tooling
“Software engineer with experience across edtech, live gaming, and an AI document intelligence platform, delivering end-to-end customer-facing features and production backends. Built secure, automated live-session scheduling integrating Zoom and TalentLMS (JWT/RBAC, idempotency, transactions) cutting setup time from ~3 minutes to under 1 minute, and optimized real-time gaming dashboards/APIs with query tuning, caching, and CDN improvements (~60% latency reduction under peak load) on AWS.”
Senior Full-Stack AI/ML Engineer specializing in MLOps and GenAI
“Senior backend/data engineer who has built and maintained HIPAA-compliant, real-time clinical FastAPI services on AWS, orchestrating ML/LLM and vector DB calls with strong reliability patterns (auth, timeouts/retries, graceful degradation, idempotency). Also delivered AWS IaC/CI-CD (Terraform/Helm/GitHub Actions) across EKS/Lambda/SageMaker and built Glue/Spark ETL with schema evolution and data quality controls, plus demonstrated large SQL performance wins (15 min to <9 sec) and hands-on incident ownership.”
Junior Backend Engineer specializing in cloud APIs and AI-enabled systems
“Built and shipped "OnCall Copilot," a production Slack-based RAG assistant that answers on-call questions from runbooks and postmortems with citations using a FAISS vector index. Emphasizes reliability and measurable performance via strict guardrails ("no evidence, no answer"), evaluation metrics, drift monitoring, and operational hardening with Docker, logging, health checks, and offline fallback.”
Junior Full-Stack Software Engineer specializing in AI/ML platforms and microservices
“Graduate-school lab engineer who built and owned the final architecture of a Microservices Hub that integrates REST APIs, issues API keys, monitors 10+ Linux servers, and visualizes service dependencies via a topology graph. Strong in bridging legacy and modern stacks (Dockerized and non-Dockerized services like Apache/screen) using deep Linux/networking knowledge, plus practical real-time audio streaming for STT/TTS and experience mentoring others.”
Senior Full-Stack & AI Developer specializing in Python/React, AWS, and LLM/RAG systems
“Backend Python engineer who owned the full backend build of an AI-driven platform for UK golf clubs, including FastAPI microservices, vector search, and a tuned LangChain+Pinecone RAG pipeline focused on cost and hallucination reduction. Experienced deploying Django/FastAPI/Flask stacks on AWS-backed Kubernetes with GitOps/ArgoCD-style delivery, plus executing legacy-to-AWS migrations and building Kafka-based real-time analytics pipelines.”
Mid-level Full-Stack Software Engineer specializing in GenAI (RAG/LLM) systems
“Backend/platform engineer who has owned FastAPI microservices for analytics/ML ingestion and driven measurable performance gains (cut latency ~40%). Experienced deploying to AWS (ECS/EKS) with GitOps (GitHub Actions + ArgoCD + Helm), and has executed an on-prem to AWS migration using Terraform with parallel-run cutover and ~30% runtime improvement. Also built Kafka-based real-time user activity streaming with Prometheus/Grafana observability.”
Intern Full-Stack Engineer specializing in AI-powered products
“Software engineer (internship experience) who built and owned an AWS serverless multi-user “challenge” feature end-to-end (UI + REST APIs + DynamoDB + deployment), delivering measurable gains in latency (-30%), debugging time (-50%), and join drop-offs (~-30%). Also productionized a multilingual RAG-based QA system with vector retrieval and guardrails, improving accuracy to ~85% and driving ~20% DAU growth.”
Mid-level AI Engineer specializing in GenAI, agentic workflows, and RAG systems
“Built a production multi-agent RAG assistant using LangChain/LangGraph with OpenAI embeddings and FAISS, focusing on retrieval quality and latency (Redis caching, parallel retrieval, precomputed embeddings). Experienced orchestrating ETL/ML pipelines with Airflow and Databricks Workflows, and has delivered an AI assistant for business ops to extract insights from policy/compliance documents through close non-technical stakeholder collaboration.”