Pre-screened and vetted.
Intern AI/ML Engineer specializing in LLM applications, RAG, and model evaluation
“Backend/ML engineer who built production LLM-enabled systems at PRGX, including an interpretable contract opportunity scoring engine (Bradley-Terry pairwise ranking) that reached 0.82 weighted Spearman agreement with SME auditors and was integrated into workflow. Also built a Duke student advisor chatbot and hardened it for real-world reliability/security with schema-driven tool calling, normalization, and off-domain defenses; led staged production rollouts with shadow testing and achieved 0.90 F1 on a new extraction field before shipping.”
Entry-Level Full-Stack Software Engineer specializing in web, mobile, and distributed systems
“Backend engineer who built a Logistics-as-a-Service platform in Go, proactively refactoring a monolithic REST service into gRPC microservices to improve performance and maintainability. Led a 3-person team with disciplined code reviews, Dockerized DB migrations, and a canary-style rollout (5% traffic) monitored for latency and failures; also implemented JWT/OAuth2 RBAC and production-minded edge-case handling in an ordering system.”
Mid-level Data Scientist / Machine Learning Engineer specializing in fraud, risk, and MLOps
“AI/ML practitioner with Northern Trust experience who has shipped production LLM systems (internal support assistant) using RAG, vector databases, orchestration (LangChain/custom pipelines), and rigorous monitoring/feedback loops. Also built AI-driven fraud detection/risk monitoring solutions in a regulated financial environment, emphasizing explainability (SHAP), audit readiness, and stakeholder trust through dashboards and clear communication.”
Mid-level AI/ML Engineer specializing in fraud detection and risk analytics in Financial Services
“At JP Morgan Chase, built and deployed a production LLM-powered RAG knowledge assistant to help fraud investigators and risk analysts quickly navigate regulatory updates and internal policies, reducing investigation delays and compliance risk. Strong focus on secure retrieval (RBAC filtering), reliability (layered testing + observability), and production constraints (latency/SLOs), with Airflow-orchestrated, auditable ML pipelines.”
Mid-level Python & AI/ML Engineer specializing in backend APIs and MLOps
“Built and deployed a production LLM/RAG document automation system for business documents (contracts/claim forms) that extracts schema-validated JSON, generates grounded summaries/Q&A, and integrates into transaction systems via APIs. Emphasizes real-world reliability: hallucination controls, layout-aware parsing with OCR fallback, Step Functions-orchestrated workflows with retries/timeouts, and human-in-the-loop review designed in close partnership with operations and claims stakeholders.”
Mid-level Data Scientist specializing in Generative AI, LLMOps, and clinical data pipelines
“LLM/RAG engineer who has built and deployed corporate-scale systems at Novartis and Johnson & Johnson, including a healthcare AI agent that generates day-to-day treatment schedules. Recently handled a high-stakes safety incident (LLM suggesting overdose) by tightening model instructions and validating with ~200 test prompts, and has strong end-to-end data/embedding/vector DB pipeline experience (PySpark, FAISS, Pinecone) plus SME-in-the-loop evaluation (RLHF).”
Junior Data Scientist specializing in Generative AI and applied machine learning
“At Evoke Tech, built a production LLM "Testbench" to quickly compare LLMs/embedding models and RAG strategies (semantic, hybrid BM25, re-ranking, HyDE, query expansion) to select optimal architectures for different client needs. Also developed a multi-agent, multimodal (voice/text) RAG system for live catalog retrieval and safe product recommendations using LangGraph/LangChain with LangSmith monitoring, and regularly translated PM/UX goals into concrete agent behaviors via demos and flowcharts.”
Junior Machine Learning Engineer specializing in LLMs, RAG, and medical imaging
“At Fileread, the candidate built and deployed an LLM-powered legal document classification and retrieval layer for an agentic extraction system that turns unstructured legal PDFs into structured tables with line-level citations. They productionized a RAG-style pipeline (ingestion, embeddings, retrieval, reranking, generation) and report 95%+ F1 across 70+ legal categories, emphasizing rigorous evaluation and close collaboration with legal domain experts for high-stakes precision.”
Mid-level Full-Stack Engineer specializing in AI/ML data platforms for biotech and FinTech
“AI/ML full-stack practitioner in a small-scale manufacturing/lab operations environment who deployed a production ML system to improve blood cell order fulfillment by predicting yield/success from donor characteristics. Experienced building custom multi-agent orchestration (Python, LangChain/LangGraph, MCP) and balancing reliability, data quality constraints, and token/ROI economics while communicating tradeoffs to VP-level business stakeholders.”
Mid-level Software Engineer specializing in AI agents, backend systems, and data engineering
“Amazon engineer who built a production AI agent platform (Python/AWS Strands on Bedrock) that lets teams create tool-using, multi-agent workflows—e.g., agents that auto-triage and resolve customer support tickets by reading internal documentation and collaborating with a research agent. Previously worked in Deloitte on IAM using Ping Identity/Ping DaVinci orchestration, and applies orchestration thinking plus structured evaluation (LLM-as-judge, surveys, automated tests) to improve agent reliability.”
Intern Machine Learning Engineer specializing in NLP, RAG, and deepfake detection
“Early-career (fresher) candidate who built and deployed a production AI medical document chatbot using a RAG architecture (LangChain + Hugging Face LLM + Pinecone) with a Flask backend on AWS EC2 via Docker. Has experience troubleshooting real deployment constraints (model dependencies, disk space, container stability) and setting up continuous-style evaluation with fixed query test sets tracking relevance, latency, and error rate.”
Mid-level NLP/LLM Researcher specializing in question answering and retrieval-augmented generation
“Built ToolDreamer, a framework for selecting relevant tools for LLM agents by training a retriever on LLM-generated reasoning traces, and has hands-on experience building multi-agent systems in AutoGen (MAG-V) focused on question generation and tool-trajectory verification. Currently works as an AI-guides supervisor at Penn State, regularly communicating AI concepts to non-technical stakeholders.”
Senior Software Engineer specializing in full-stack systems, data pipelines, and ML
“Built and productionized an autonomous research agent (AutoGPT) in a Docker/Kubernetes environment with Pinecone-based long-term memory and custom Python tools for analysis, visualization, and report drafting. Implemented layered guardrails (prompt templates, automated validation, self-critique loops, and monitoring) and achieved ~25% reduction in manual report generation time while scaling the workflow to support multiple concurrent users.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Senior Data Scientist specializing in ML, NLP, and GenAI analytics
“Built and deployed an LLM-powered analytics assistant enabling business users to ask questions in plain English and receive validated Spark SQL executed in Databricks, with a Streamlit/Flask UI. Addressed strict client schema-privacy constraints by implementing a RAG strategy and ultimately leveraging AWS Bedrock and fine-tuned reference docs. Also has production ML pipeline experience using Docker + Airflow and AWS (S3/ECS/EC2) for financial classification models.”
Mid-level Software Engineer specializing in FinTech full-stack and AI applications
“Built and productionized an NLP-powered customer support assistant at JPMorgan Chase for digital banking, focused on reducing response time for repetitive client queries. Strong in real-world AI deployment challenges—sensitive data handling, low-latency FastAPI services, and AWS/Kubernetes operations with CI/CD—plus a metrics- and guardrails-driven approach to reliable AI workflows.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
Intern AI/ML Engineer specializing in RAG, multimodal AI, and LLM systems
“Built and shipped 'PetPulse,' a production AI pet-health note system that records voice notes, transcribes them, converts transcripts into structured symptom/event data, and supports grounded Q&A over a user’s notes and vet PDFs. Demonstrates full-stack LLM product execution (FastAPI + GPT-4 + Firebase), with concrete reliability/performance work (async endpoints, caching, RAG/embeddings, function calling) and user-centered iteration with a non-technical product stakeholder.”
“GenAI/data engineering practitioner with production experience across Equinix, Optum, and Citibank—built an Azure OpenAI (GPT-4) + LangChain document intelligence platform processing 1.5M+ docs/month and a HIPAA-compliant Airflow healthcare pipeline handling 5M+ claims/day. Also delivered a real-time fraud detection + explainability system using LightGBM and a fine-tuned T5 NLG component, improving fraud accuracy by 15%+ while partnering closely with compliance stakeholders.”
Mid-level Machine Learning Engineer specializing in LLMs and NLP classification systems
“Internship experience building a production RAG+LLM pipeline to map messy card transaction descriptions to merchant brands, including a custom modified-ROUGE evaluation approach for weak/variant ground truth. Improved scalability and cost by moving from a managed LLM endpoint (e.g., Bedrock) to self-hosted vLLM, and orchestrated massive embedding backfills (5,000+ files, 10B+ rows) using an Airflow-triggered SQS + ECS worker architecture with robust retry/DLQ handling.”
Mid-level AI Engineer specializing in Generative AI, MLOps, and NLP for finance and healthcare
“Built and deployed a secure, production LLM-based document summarization and risk-highlighting tool for financial auditors, running inside a private Azure environment to protect confidential data. Focused on reliability (hallucination mitigation via retrieval-based prompts and source citations) and validated performance through comparisons to auditor summaries plus a user pilot, cutting review time by about half.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and RAG systems
“Machine learning/NLP engineer who built a production-oriented retrieval-based AI system at Morgan Stanley for healthcare use cases, combining RAG over unstructured patient records with deep-learning medical image segmentation (U-Net/Mask R-CNN). Strong in end-to-end pipelines and MLOps (Spark/MongoDB, AWS SageMaker, CI/CD, monitoring, automated retraining) and in entity resolution/data quality validation for noisy clinical data.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps on AWS
“LLM engineer who built a production document intelligence/RAG pipeline to extract structured data from thousands of unstructured PDFs, cutting manual review time by 60%. Experienced with LangChain and Airflow orchestration plus rigorous evaluation (labeled datasets, prompt testing, HITL review, monitoring) to improve accuracy and reduce hallucinations while partnering closely with non-technical operations stakeholders.”
“ServiceNow engineer who built and launched a production LLM-powered ticket resolution/knowledge assistant using RAG (LangChain + Hugging Face embeddings + vector search) integrated into internal support dashboards via REST APIs. Optimized the system from ~6–8s to ~2–3s latency while improving usability with concise, cited answers and guardrails (grounding + similarity thresholds), delivering ~30–35% reduction in manual ticket investigation effort.”