Pre-screened and vetted.
Mid-level Machine Learning Engineer specializing in deep learning and generative AI
“ML/NLP engineer with hands-on experience building production systems for unstructured insurance claims and customer data linking. Delivered measurable impact at scale (millions of documents), combining transformer-based NLP, vector search (FAISS/Pinecone), and human-in-the-loop validation, and has strong production workflow/observability practices (Airflow, AWS Batch, Grafana/Prometheus).”
Mid-level AI/ML Engineer specializing in fraud detection and NLP
“Built production AI/RAG-style systems for message Q&A and insurance claims workflows, combining data ingestion, indexing/retrieval, and LLM integration with fallback modes. Has hands-on orchestration experience (Airflow, Prefect, LangChain) and cites large operational gains (claims processing reduced to ~45 seconds; manual review -50%; false alerts -30%) through automated, monitored pipelines and close collaboration with non-technical stakeholders.”
“AI/ML engineer with banking domain experience (M&T Bank) who built a production credit-risk prediction and reporting platform combining ML models (XGBoost/TensorFlow) with a RAG pipeline (LangChain + GPT-4) over compliance documents. Delivered measurable impact (≈20% better risk detection/precision, 50% less manual reporting) and productionized workflows on Vertex AI/Kubeflow with CI/CD and monitoring; also implemented embedding-based semantic search using FAISS/Pinecone.”
Mid-level AI/ML Engineer specializing in healthcare ML and generative AI
“AI/LLM engineer at Humana who built and deployed a HIPAA-aware RAG system for clinical record retrieval, cutting search time dramatically and improving retrieval efficiency by 30%. Experienced with Spark-scale data preprocessing, QLoRA fine-tuning, LangChain orchestration, and MLflow+SageMaker integration, with a strong testing/evaluation discipline (A/B tests, human eval) to hit 95%+ accuracy and production latency targets.”
Mid-level AI & Data Scientist specializing in LLMs, RAG, and healthcare NLP
“Built a production LLM/RAG solution for healthcare operations teams to query large policy and care-guideline repositories in natural language. Improved domain alignment using vector retrieval plus parameter-efficient fine-tuning and prompt optimization, validated through internal user testing and metrics, cutting manual lookup time by ~40%. Also has hands-on experience orchestrating automated ML pipelines with Apache Airflow.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps for financial services
“Built and deployed a production Llama 3-based RAG document Q&A system using FAISS, addressing context-window limits through chunking and keeping retrieval accurate by regularly refreshing embeddings. Has hands-on orchestration experience with LangChain and LlamaIndex for multi-step LLM workflows (including memory management) and collaborates with non-technical teams (e.g., marketing) to deliver AI solutions like recommendation systems.”
Mid-level Full-Stack Software Engineer specializing in cloud-native microservices and GenAI
“Full-stack engineer with cloud and GenAI experience who has owned production features end-to-end, including a reporting dashboard optimized from 14s to 5s using query/API refactoring and monitored via AWS CloudWatch. Also productionized an OpenAI-powered chatbot using LangChain with prompt design, guardrails, and evaluation via production logs and user feedback, and has led incremental legacy-to-microservices modernization with parallel run to avoid regressions.”
Senior Full-Stack Software Engineer specializing in IIoT, Edge AI, and real-time analytics
“Full-stack engineer who built an end-to-end low-code/no-code IDE for creating AI/ML workflows for industrial IoT sensors using Next.js/TypeScript and NestJS microservices. Focused on scaling high-volume sensor dashboards—improved UX and performance via WebSockets, debouncing, pagination, and API payload reduction—validated with profiling tools and user feedback in a startup environment.”
Junior AI Engineer specializing in ML, LLM systems, and RAG
“Built and deployed an LLM/applied-ML system enabling efficient extraction of useful information from large unstructured multimodal datasets, owning the full pipeline from ingestion to inference and APIs with a strong emphasis on production reliability, latency, and monitoring. Also delivered a voice-based AI workflow for Hindi policy document access for the Election Commission of India by translating non-technical usability needs into iterative demos and a successful implementation.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and MLOps
“Built a production RAG-based healthcare chatbot to retrieve patient medical documents spread across multiple platforms, reducing manual and error-prone searching. Implemented semantic search with custom embeddings (Hugging Face) and Pinecone, deployed via FastAPI/Docker on AWS SageMaker with MLflow tracking, and optimized fine-tuning cost using LoRA while orchestrating retraining pipelines in Airflow.”
Mid-level AI/ML & Full-Stack Engineer specializing in LLM agents and generative AI
“LLM/agent builder who shipped a live consumer AI-agent app (kalpa.chat) that visualizes complex reasoning as interactive graphs and abstracts multi-provider model usage via a unified wallet. Professionally has applied LangChain/LangGraph to IVR parsing and to scaling a football video-generation pipeline at DAZN, including shipping a VAR-specific retrieval/order fix via SQL after iterating with a non-technical PM.”
Junior Machine Learning Engineer specializing in NLP, computer vision, and MLOps
“ML/LLM engineer with Meta experience building production AI systems for near real-time user-report classification and summarization under strict latency (<250ms), safety, cost, and privacy constraints. Has hands-on MLOps/orchestration experience (Airflow, Spark, MLflow, Kubernetes, Docker, GitHub Actions) plus observability (Prometheus/Grafana) and applies rigorous evaluation, staged rollouts, and A/B testing to keep agent workflows reliable in production.”
Mid-level AI Engineer specializing in multi-agent systems and RAG
“Built and shipped a production LangGraph-based multi-agent LLM analytics/decision copilot that answers questions across SQL/BI systems and unstructured docs, emphasizing grounded, tool-verified outputs with citations and confidence gating. Deep hands-on experience with orchestration (LangGraph, CrewAI, OpenAI Assistants, MCP) plus real-world latency/cost optimization (vLLM batching/KV caching, speculative decoding, quantization) and rigorous eval/observability. Partnered closely with business/ops stakeholders to deliver explainable reporting automation, cutting manual reporting time by 50%+.”
Junior Machine Learning Engineer specializing in NLP and multimodal transformers
“Built and deployed LLM-powered agentic chatbot and text-to-SQL systems using LangGraph/LangChain (and Bedrock), structuring workflows as DAGs with planning/replanning and validation to improve tool-calling reliability and reduce hallucinations. Operates production feedback loops with online/offline metrics, drift detection, and LangSmith-based evaluation pipelines, and regularly partners with business stakeholders and clinicians using slide decks and visual charts.”
Mid-level AI/ML Engineer specializing in GenAI, MLOps, and anomaly detection
“LLM/MLOps engineer who has shipped a production RAG-based technical documentation assistant (FastAPI) cutting manual review by 45%, with deep hands-on retrieval optimization in Pinecone/LangChain (HNSW, hybrid + multi-query search, caching). Also brings healthcare domain experience—building Airflow-orchestrated EHR pipelines and delivering FDA-auditability-friendly predictive maintenance solutions using SHAP/LIME explainability surfaced in Power BI.”
Mid-level AI/ML Engineer specializing in NLP, RAG, and MLOps for FinTech
“ML/LLM engineer with production experience building a compliant RAG-based virtual assistant at Intuit, optimizing embeddings and FAISS retrieval (including PCA) for low-latency, privacy-controlled search and deploying via AWS SageMaker containers. Also built scalable Airflow+MLflow pipelines using Docker and KubernetesExecutor, cutting training cycles by 37%, and partnered with civil engineers/project managers at Aegis Infra to deliver predictive maintenance for construction equipment.”
“Backend/data engineer who builds Python (FastAPI) data-processing API services for internal analytics/reporting, emphasizing modular architecture, async performance tuning, and reliability patterns (health checks, retries, observability). Also migrated legacy on-prem ETL pipelines to Azure using ADF/Data Lake/Functions and implemented a near-real-time ingestion flow with Event Hubs plus watermarking to handle late events and deduplication.”
Mid-level Data Scientist specializing in ML, NLP, and LLM-powered solutions
“AI/NLP-focused practitioner who built a zero-/few-shot LLM event extraction system on the long-tail Maven dataset, combining prompt-structured outputs with LoRA/QLoRA fine-tuning and rigorous F1 evaluation. Also implemented entity resolution/data cleaning pipelines and embedding-based semantic search using Sentence-BERT + FAISS, and has healthcare experience delivering a multilingual speech/translation mobile prototype using HIPAA-compliant Azure Cognitive Services.”
Mid-level Machine Learning Engineer specializing in NLP, recommender systems, and MLOps
“ML/LLM engineer with production experience at General Motors building Transformer-based search and recommendation personalization for a high-traffic vehicle platform. Delivered significant KPI gains (17% conversion lift, 14% bounce-rate reduction) and optimized real-time inference via ONNX Runtime and INT8 quantization while implementing robust MLOps (Airflow/MLflow, monitoring, drift-triggered retraining) and stakeholder-facing explainability/dashboards.”
Mid-level Data Scientist specializing in NLP and predictive modeling
“AI/ML practitioner in healthcare/insurance (Blue Cross Blue Shield) who built and deployed a production NLP system to classify patient risk from unstructured clinical notes. Experienced in end-to-end pipeline orchestration (Airflow, AWS Step Functions/Lambda/SageMaker) and real-time optimization (BERT to DistilBERT on AWS GPUs), with strong clinician collaboration to drive adoption.”
Junior Software Engineer and ML Researcher specializing in full-stack and applied deep learning
“LLM engineer who built a production-style educational questionnaire generation system (MCQs/fill-in-the-blanks/short answers) using Hugging Face models (BERT/T5) and implemented grounding, decoding tuning, and post-generation validation to control hallucinations and quality. Also developed a "tech care" assistant chatbot with a custom Python orchestration/router layer (intent classification, context management, multi-step flows) and a structured testing/evaluation approach including expert review and automated checks.”
Junior Robotics & AI/ML Engineer specializing in multi-agent reinforcement learning and computer vision
“Robotics software candidate whose thesis focused on multi-robot warehouse coordination using MAPPO reinforcement learning, trained in simulation (LBF environment, Isaac Sim/RViz) and deployed onto three real-time robots. Built custom ROS 2 Humble nodes for multi-robot control with namespaces, TF broadcasting, and an RL pipeline integrating LiDAR odometry and camera observations.”
Mid-level AI/ML Engineer specializing in financial risk, fraud analytics, and forecasting
“Built and productionized an LLM-powered financial intelligence and forecasting platform at Northern Trust using a RAG architecture (LangChain + Hugging Face + FAISS) with end-to-end MLOps (Docker/Kubernetes, Airflow, MLflow). Emphasized regulatory-grade explainability (SHAP/Power BI) and hallucination control (retrieval-only grounding), achieving ~30% forecasting accuracy improvement and ~65% reduction in analyst research time, with sub-second inference and 95% uptime on EKS/AKS.”
Mid-level AI/ML Engineer specializing in production RAG systems and MLOps
“Built and deployed a GPT-4 + Pinecone RAG system that lets users query large internal document collections with grounded, cited answers. Demonstrates strong applied LLM engineering (chunking experiments, hallucination controls, metadata recency boosting) plus production-minded evaluation/monitoring and performance tuning (rate-limit mitigation via pooling/batching). Also effective at translating complex AI concepts to non-technical stakeholders through prototypes and live demos, helping secure client sponsorship.”