Pre-screened and vetted.
Mid-level AI/ML Engineer specializing in Generative AI, NLP, and Computer Vision
“Built an LLM-powered learning assistant (EduQuizPro/EduCrest Pro) that uses RAG over URLs and PDFs to generate quizzes, notes, and explanations for students/professors. Emphasizes production robustness—implemented dependency fallbacks (FAISS/Sentence Transformers/Gradio), CLI-safe mode, and NumPy-based indexing—along with a custom orchestration layer to keep multi-step AI workflows reliable.”
Mid-level Data Scientist / AI-ML Engineer specializing in Generative AI and LLM applications
“Built a production GenAI-powered analytics assistant to reduce reliance on data analysts by enabling natural-language Q&A over Databricks/Power BI dashboards, backed by vector search (Pinecone/Milvus) and a Neo4j knowledge graph, including multimodal support via OpenAI Vision. Demonstrates strong real-world LLM reliability engineering with strict RAG, LangGraph multi-step verification, and Guardrails/custom validators, plus broad orchestration and production monitoring experience (Airflow, ADF, Step Functions, Kubernetes, Prometheus/CloudWatch).”
Mid-level Research Engineer specializing in machine learning and computational neuroscience
“Master’s-level ML researcher with hands-on embodied/edge deployment experience: built a Google Glass motion-tracking system at Sandia using MobileNetV1 + LSTM trained in TensorFlow and deployed via TensorFlow Lite. Has reimplemented transformer-based research for a thesis and demonstrated strong judgment adapting quickly when upstream assumptions changed, and stays current through active reading groups and a JEPA collaboration.”
Senior Machine Learning Engineer specializing in optimization, LLMs, and on-device AI
“Engineer with hands-on experience debugging and hardening a fixed-point implementation for an internal PoC, quickly diagnosing overflow/underflow issues that caused intermittent failures across thousands of runs and delivering a code fix. Comfortable presenting technical solutions with layered slide depth and doing follow-up deep dives for interested stakeholders, though has limited direct customer/sales partnership experience.”
Director-level AI & Data Science leader specializing in GenAI, LLMs, and MLOps
“ML/NLP engineer currently working in NYC on a system that connects complex unstructured data sources to deliver personalized insights, using embeddings + vector DB retrieval and a RAG architecture (LangChain, Pinecone/OpenSearch). Strong focus on production constraints—especially low-latency retrieval—using FAISS/ANN, PCA, index partitioning, and Redis caching, plus PEFT fine-tuning (LoRA/QLoRA) and KPI/SLA-driven promotion to production.”
Principal Data Scientist specializing in healthcare analytics and medical imaging AI
“Developed an LLM-driven recommendation agent in Azure Databricks to triage oncology patients and trigger second-opinion case creation using medical claims and EHR data. Uses ICD-10/CPT/J-code features in prompts, embeddings + vector DB similarity, and a backtesting framework emphasizing recall to avoid missing clinically relevant cases while supporting business revenue.”
Mid-level Full-Stack Software Engineer specializing in microservices and cloud platforms
“Software engineer with experience across enterprise (AIG, MSCI) and an early-stage startup (Job Map), owning production systems end-to-end. Built secure insurance microservices on Spring Boot with JWT/RBAC and AWS-based CI/CD/observability, plus Kafka streaming pipelines for financial data. Also shipped a GenAI personalization MVP using FastAPI and LLM APIs in a high-ambiguity startup environment.”
Junior Software Engineer specializing in video streaming and processing systems
“Software engineering intern at China Telecom who built and continuously evolved a real-time transaction platform ("Smart Tangerine") focused on strong consistency and peak-hour concurrency. Implemented microservices with Redis and RabbitMQ to decouple heavy processing and cut latency (~80ms to ~30ms), and led a zero-downtime migration from a monolith using strangler pattern, dual-write, and traffic shadowing.”
Intern AI/ML Engineer specializing in LLM applications, RAG, and model evaluation
“Backend/ML engineer who built production LLM-enabled systems at PRGX, including an interpretable contract opportunity scoring engine (Bradley-Terry pairwise ranking) that reached 0.82 weighted Spearman agreement with SME auditors and was integrated into workflow. Also built a Duke student advisor chatbot and hardened it for real-world reliability/security with schema-driven tool calling, normalization, and off-domain defenses; led staged production rollouts with shadow testing and achieved 0.90 F1 on a new extraction field before shipping.”
Mid-level Backend Software Engineer specializing in FinTech
“Backend engineer with Citigroup experience who built and evolved a self-service user provisioning/identity backend, cutting onboarding from 45 minutes to under 2 minutes. Demonstrates strong production-grade integration and reliability practices (isolated integrations, retries, rollback logic, heavy logging) plus secure API development in Python/FastAPI with OAuth scope-based authorization and incremental, low-risk rollout strategies.”
Intern AI/ML Engineer specializing in robotics and computer vision
“Worked on Sophia the humanoid robot, building production animation pipelines and enhancing human-robot interaction via perception and behavior orchestration. Experienced in stabilizing noisy perception-driven state transitions and designing smooth, user-centered behavioral flows, collaborating closely with artists, animators, and experience designers to translate creative intent into measurable system behavior.”
Mid-level Full-Stack Software Engineer specializing in backend microservices and enterprise AI tools
“Backend/platform engineer with experience across C3.ai (supply chain demand planning) and Amdocs (telecom), working on large-scale data systems and microservices. Has driven first-time adoption experiments of Snowflake + Spark to handle billion-record workloads, built Jenkins-to-Kubernetes delivery pipelines with Nexus artifact management, and implemented Kafka streaming between microservices with HA and retry/error-handling patterns.”
Mid-level Data Scientist / Machine Learning Engineer specializing in fraud, risk, and MLOps
“AI/ML practitioner with Northern Trust experience who has shipped production LLM systems (internal support assistant) using RAG, vector databases, orchestration (LangChain/custom pipelines), and rigorous monitoring/feedback loops. Also built AI-driven fraud detection/risk monitoring solutions in a regulated financial environment, emphasizing explainability (SHAP), audit readiness, and stakeholder trust through dashboards and clear communication.”
Mid-level Data Scientist specializing in Generative AI, LLMOps, and clinical data pipelines
“LLM/RAG engineer who has built and deployed corporate-scale systems at Novartis and Johnson & Johnson, including a healthcare AI agent that generates day-to-day treatment schedules. Recently handled a high-stakes safety incident (LLM suggesting overdose) by tightening model instructions and validating with ~200 test prompts, and has strong end-to-end data/embedding/vector DB pipeline experience (PySpark, FAISS, Pinecone) plus SME-in-the-loop evaluation (RLHF).”
Junior Machine Learning Engineer specializing in LLMs, RAG, and medical imaging
“At Fileread, the candidate built and deployed an LLM-powered legal document classification and retrieval layer for an agentic extraction system that turns unstructured legal PDFs into structured tables with line-level citations. They productionized a RAG-style pipeline (ingestion, embeddings, retrieval, reranking, generation) and report 95%+ F1 across 70+ legal categories, emphasizing rigorous evaluation and close collaboration with legal domain experts for high-stakes precision.”
Intern Machine Learning Engineer specializing in NLP, RAG, and deepfake detection
“Early-career (fresher) candidate who built and deployed a production AI medical document chatbot using a RAG architecture (LangChain + Hugging Face LLM + Pinecone) with a Flask backend on AWS EC2 via Docker. Has experience troubleshooting real deployment constraints (model dependencies, disk space, container stability) and setting up continuous-style evaluation with fixed query test sets tracking relevance, latency, and error rate.”
Senior Data Scientist specializing in ML, NLP, and GenAI analytics
“Built and deployed an LLM-powered analytics assistant enabling business users to ask questions in plain English and receive validated Spark SQL executed in Databricks, with a Streamlit/Flask UI. Addressed strict client schema-privacy constraints by implementing a RAG strategy and ultimately leveraging AWS Bedrock and fine-tuned reference docs. Also has production ML pipeline experience using Docker + Airflow and AWS (S3/ECS/EC2) for financial classification models.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
“Backend/AI engineer who built a real-time vector database system for high-frequency financial data using Kafka/Flink on Kubernetes, achieving sub-100ms similarity search at 10k+ concurrent load and resolving tricky duplication issues with idempotency/versioning. Also shipped an end-to-end LLM-based travel itinerary feature (profiling + prompt workflows + APIs) with a focus on quality consistency and low latency.”
Mid-level AI Engineer specializing in Generative AI, MLOps, and NLP for finance and healthcare
“Built and deployed a secure, production LLM-based document summarization and risk-highlighting tool for financial auditors, running inside a private Azure environment to protect confidential data. Focused on reliability (hallucination mitigation via retrieval-based prompts and source citations) and validated performance through comparisons to auditor summaries plus a user pilot, cutting review time by about half.”
Mid-level Machine Learning Engineer specializing in NLP, computer vision, and RAG systems
“Machine learning/NLP engineer who built a production-oriented retrieval-based AI system at Morgan Stanley for healthcare use cases, combining RAG over unstructured patient records with deep-learning medical image segmentation (U-Net/Mask R-CNN). Strong in end-to-end pipelines and MLOps (Spark/MongoDB, AWS SageMaker, CI/CD, monitoring, automated retraining) and in entity resolution/data quality validation for noisy clinical data.”
Mid-level Machine Learning Engineer specializing in forecasting, NLP, and GenAI
“GenAI/ML engineer with production experience building multilingual LLM systems (English/Spanish) and RAG-based clinical documentation summarization at Walgreens, combining prompt engineering, structured output validation, and rigorous evaluation (ROUGE + pharmacist review). Also orchestrated end-to-end ML pipelines for demand forecasting using Apache Airflow, PySpark, and MLflow with scheduled retraining and production monitoring.”
“ServiceNow engineer who built and launched a production LLM-powered ticket resolution/knowledge assistant using RAG (LangChain + Hugging Face embeddings + vector search) integrated into internal support dashboards via REST APIs. Optimized the system from ~6–8s to ~2–3s latency while improving usability with concise, cited answers and guardrails (grounding + similarity thresholds), delivering ~30–35% reduction in manual ticket investigation effort.”
Entry Machine Learning Engineer specializing in NLP, computer vision, and recommender systems
“Built and shipped an end-to-end podcast recommendation system exposed via a Flask API and React UI, explicitly balancing relevance, diversity (MMR), and safety constraints while meeting ~200ms latency targets. Also implemented a production-style RAG/information-extraction pipeline using web retrieval, spaCy NER, and fine-tuned SpanBERT with guardrails and evaluation loops (precision/recall/F1) to tune confidence thresholds and improve reliability.”