Pre-screened and vetted.
Intern Data Scientist specializing in robotics localization and SLAM
“Robotics/embodied-AI practitioner who built a TurtleBot3 LiDAR-fingerprint localization pipeline end-to-end (autonomous data collection + multi-head NN) achieving ~30 cm error in a 10x10 m space. Also has industry experience at Infineon building large-scale production data/AI pipelines and rapidly fixing a deployed recommendation system by correcting upstream data normalization, improving accuracy by 20%+.”
Mid-level DevOps/Cloud Engineer specializing in AWS infrastructure automation
“Frontend engineer with experience building a large-scale React + TypeScript administrative dashboard for an e-commerce platform, using Redux Toolkit plus TanStack Query to separate UI and server state. Emphasizes quality at scale through CI/CD automation, Jest/integration testing, and performance techniques like code splitting and caching, with experience coordinating integration across multiple teams.”
Mid-level AI & Machine Learning Engineer specializing in Generative AI and MLOps
“Built a production GPT-4/LangChain/Pinecone RAG “AI Copilot” at Northern Trust to automate financial report generation and analyst Q&A over internal structured (SQL warehouse) and unstructured policy data. Focused on real-world production challenges—grounding and latency—achieving major speed gains (seconds to milliseconds) via MiniLM embedding optimization and Redis caching, and implemented rigorous testing/evaluation with MLflow-backed metrics while aligning compliance and finance stakeholders for deployment.”
Entry-level AI/ML Engineer specializing in AWS MLOps and computer vision
“Built and shipped a production RAG question-answering system using LangChain/OpenAI, Docker, and FastAPI, then reduced hallucinations through disciplined retrieval tuning and constrained prompting. Also implemented a custom evaluation framework (QA-pair dataset) to measure faithfulness/relevance and deployed containerized ML microservices on AWS ECS/Fargate with ALB and rolling, zero-downtime updates.”
Mid-level Machine Learning Engineer specializing in LLMs, GenAI, and Computer Vision
“LLM/agent engineer who built a production multi-agent research automation system using LangGraph (planner, retriever with FAISS, supervisor, evaluator) with structured outputs and citation tracking for traceable reports. Emphasizes reliability and operations—LangSmith-based observability, multi-level testing, hallucination mitigation, and latency/cost controls—plus prior experience as a Computer Vision Software Engineer at Deepsight AI Labs working directly with non-technical customers.”
Mid-level Full-Stack Developer specializing in Java/Spring Boot, React, and cloud microservices
“Backend/platform engineer with hands-on ownership of Kubernetes GitOps delivery (GitHub Actions + Argo CD) on AWS EKS, including progressive rollouts and reliable rollback across interdependent microservices. Built a Python/FastAPI ML-driven document-processing service (PostgreSQL + S3) to complement existing Spring Boot systems, and implemented Kafka streaming pipelines with Schema Registry plus Prometheus/Grafana observability. Also supported a hybrid cloud-to-on-prem migration for compliance/latency with phased rollout and incremental PostgreSQL migration.”
Entry-Level Software Engineer specializing in distributed systems and backend infrastructure
“Built and operated an end-to-end customer-facing "Record Platform" web product as both engineer and primary user, focusing on reliability and correctness in core flows like search and checkout. Implemented a TypeScript/React frontend with a multi-service backend and Kafka-based event-driven architecture, and created internal tooling to automate risky ops like Kubernetes TLS certificate rotation with k6 load/chaos testing (including HTTP/2 and HTTP/3 validation).”
Mid-level AI/ML Software Engineer specializing in data pipelines, BI dashboards, and computer vision
“Graduate Assistant Intern at Friends University who built and deployed a GenAI-driven requirement understanding system that automates extraction and semantic grouping of technical requirements from large unstructured documents. Demonstrates strong LLM engineering rigor (golden datasets, regression testing, post-processing validation) and production-minded delivery using LangChain/LlamaIndex orchestration, FastAPI microservices, Docker, and cloud deployment.”
Mid-level DevOps Engineer specializing in cloud automation and DevSecOps
“Cloud/hybrid infrastructure engineer with McKesson experience migrating tightly coupled healthcare applications to microservices on AWS EKS. Strong in IaC-driven standardization, CI/CD automation, and production observability (CloudWatch/Splunk/Prometheus/tracing), with demonstrated ability to debug complex incidents spanning Kubernetes and cloud networking.”
Senior Data Scientist/Software Engineer specializing in ML systems and cloud DevOps
“AI software engineer with experience spanning LLM/RAG production systems and regulated fintech infrastructure. Built an end-to-end natural-language-to-SQL analytics assistant (Weaviate + GPT-4 + Supabase) shipped as an API with 92% accuracy and major time savings for non-technical users, and also owned demand-forecasting and CI/CD/containerization improvements for a Bank of America core banking deployment at Infosys.”
Mid-level AI/ML Engineer specializing in GenAI and cloud MLOps
“Applied LLMs to high-stakes domains (wildfire risk for emergency teams and loan approval via a fine-tuned IBM Granite model), with a strong focus on reliability—using RAG-based cross-validation to reduce hallucinations and continuous ingestion pipelines (MODIS satellite imagery via AWS Lambda) to keep data current. Experienced in production orchestration and MLOps-style workflows using Airflow, AWS Step Functions, and SageMaker Pipelines, and collaborates closely with analysts on KPI-driven evaluation.”
Mid-level AI/ML Engineer specializing in LLMs, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM/RAG system at CVS to automate clinical documents, addressing PHI compliance, retrieval accuracy, and latency; achieved a 35–40% reduction in review effort through chunking and FP16/INT8 optimization. Also has experience translating AI outputs into actionable insights for non-technical stakeholders (sports analysts).”
Mid-Level Full-Stack Developer specializing in AWS and scalable web platforms
“Software engineer with hands-on AWS experience optimizing an email campaign delivery system—re-architected a monolithic worker into multi-threaded/multi-worker ECS components to boost throughput ~600% (5 to 35 emails/sec). Comfortable debugging production issues (e.g., SQS/EventBridge policy misconfiguration) and emphasizes maintainable delivery via design docs, TDD, versioned APIs, and strong test coverage.”
Senior Backend Engineer specializing in Python microservices and cloud-native systems
“Backend/data platform engineer who owned a FastAPI + Kafka microservice in Verizon’s billing pipeline, handling high-volume usage ingestion/validation/enrichment with strong observability and CI/CD on AWS EKS. Demonstrated measurable performance gains (latency down to ~120–150ms; Kafka throughput +30–40%; DB CPU -25%) and led an on-prem ETL-to-AWS migration using Terraform, parallel validation, and phased cutover with zero downtime.”
Mid-level Applied AI/ML Engineer specializing in agentic systems and LLM automation
“Built a production LLM-powered workflow at Frontier to extract structured signals from messy, high-volume documents and route work to the right teams, replacing a multi-day, error-prone manual process. Emphasizes production reliability with schema/consistency validation, re-prompting and deterministic fallbacks, plus async pipeline optimizations for predictable latency. Experienced with multi-agent orchestration (LangGraph, AutoGen, CrewAI) and AWS workflow tooling (Step Functions, SQS, Lambda), and delivered ~70% safe automation via stakeholder-driven thresholds and human review.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps for financial services
“Built and deployed a production Llama 3-based RAG document Q&A system using FAISS, addressing context-window limits through chunking and keeping retrieval accurate by regularly refreshing embeddings. Has hands-on orchestration experience with LangChain and LlamaIndex for multi-step LLM workflows (including memory management) and collaborates with non-technical teams (e.g., marketing) to deliver AI solutions like recommendation systems.”
Mid-level Machine Learning Engineer specializing in cloud, governance automation, and distributed systems
“Governance engineer intern at GSK who built policy-as-code automation using Open Policy Agent/Rego integrated into GitHub CI/CD and Terraform workflows. Also built and shipped a voice-enabled expense tracking app using speech-to-text + LLM structured extraction with strong validation, retries, and semantic guardrails, and designed the supporting PostgreSQL data model with performance-focused indexing.”
Mid-level DevOps/Cloud Engineer specializing in AWS, GCP, Kubernetes, and CI/CD
“Infrastructure/DevOps engineer (Geico) focused on AWS and Kubernetes at production scale. Has hands-on experience building secure GitHub Actions CI/CD for EKS, provisioning core AWS infrastructure with Terraform/CDK, and leading end-to-end incident response with post-incident automation to prevent recurrence; no direct IBM Power/AIX/PowerHA experience.”
Junior Full-Stack Software Engineer specializing in cloud-native microservices
“Backend/data engineer with experience at Assurant and Capgemini, focused on reliability and performance at scale. Improved high-latency backend APIs by adding and iterating on a Redis caching layer driven by CloudWatch/monitoring metrics, and built scalable BI pipelines that normalize messy multi-source enterprise data with strong observability and error handling. Familiar with LLM/RAG architecture and practical guardrails, though has not yet shipped an LLM feature to production.”
Mid-Level AI Backend Engineer specializing in Python, LLM/RAG, and healthcare/insurance platforms
“AI Backend Engineer in MetLife’s claims technology group who built and deployed a production LLM-based decision support system that helps claim adjusters quickly find relevant policy rules from long PDFs and historical notes. Designed it as multiple production-grade services with retrieval-first guardrails, continuous validation, and Airflow-orchestrated pipelines for ingestion, embeddings, and vector index updates to keep the system reliable as policies and data evolve.”
Mid-level Machine Learning Engineer specializing in LLM platforms and robotic perception
“Built and shipped a production multi-agent personal financial assistant at AlphevaAI on AWS ECS, combining FastAPI microservices, Redis/SQS orchestration, and Pinecone-based hybrid RAG (semantic + BM25) to ground financial guidance. Improved routing accuracy with an embedding-based SetFit + logistic regression intent classifier feeding an LLM router, and optimized UX with live streaming plus cost controls via model tiering and caching.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Built a production RAG-based healthcare chatbot to retrieve patient medical documents spread across multiple platforms, reducing manual and error-prone searching. Implemented semantic search with custom embeddings (Hugging Face) and Pinecone, deployed via FastAPI/Docker on AWS SageMaker with MLflow tracking, and optimized fine-tuning cost using LoRA while orchestrating retraining pipelines in Airflow.”
Mid-level Generative AI Engineer specializing in LLM agents and RAG
“GenAI/LLM engineer who built and deployed a production RAG system for enterprise document search and decision support, cutting manual lookup time by 40%+. Experienced with LangChain/LangGraph agent orchestration plus Airflow/Prefect for ingestion and incremental reindexing, with a strong focus on reliability (testing, observability) and stakeholder-driven metrics.”
Junior Machine Learning Engineer specializing in NLP, computer vision, and MLOps
“ML/LLM engineer with Meta experience building production AI systems for near real-time user-report classification and summarization under strict latency (<250ms), safety, cost, and privacy constraints. Has hands-on MLOps/orchestration experience (Airflow, Spark, MLflow, Kubernetes, Docker, GitHub Actions) plus observability (Prometheus/Grafana) and applies rigorous evaluation, staged rollouts, and A/B testing to keep agent workflows reliable in production.”