Pre-screened and vetted.
Mid-level AI Engineer specializing in GenAI, NLP, and MLOps
“LLM/agentic-systems engineer with PayPal experience hardening an LLM-powered fraud support assistant from prototype to production, focusing on low-latency distributed architecture, rigorous evaluation/testing, and security/compliance. Comfortable in customer-facing and GTM contexts—runs technical demos/workshops, builds tailored pilots, and aligns sales/CS with engineering to close deals and drive adoption.”
Entry-level Full-Stack Developer specializing in logistics and AI-powered web applications
“Backend engineer who led the end-to-end modernization of FleetView into a scalable, event-driven system supporting 1,000+ users and 13,000+ assets, cutting API latency by ~40%. Also built an AI-powered exit interview analytics pipeline on Azure using GPT-4o with strong guardrails, validation, and evaluation practices, showing a rare mix of production backend rigor and applied LLM workflow experience.”
Mid-level Data Scientist / Machine Learning Engineer specializing in fraud, risk, and MLOps
“AI/ML practitioner with Northern Trust experience who has shipped production LLM systems (internal support assistant) using RAG, vector databases, orchestration (LangChain/custom pipelines), and rigorous monitoring/feedback loops. Also built AI-driven fraud detection/risk monitoring solutions in a regulated financial environment, emphasizing explainability (SHAP), audit readiness, and stakeholder trust through dashboards and clear communication.”
Mid-level AI/ML Engineer specializing in NLP, MLOps, and scalable data pipelines
“Built and shipped a production LLM-powered personalized client engagement assistant in the financial domain, balancing real-time recommendations with strict privacy/compliance requirements. Demonstrates strong MLOps/LLMOps depth (Airflow + MLflow, containerized microservices, drift monitoring) and a privacy-by-design approach validated in collaboration with risk and compliance teams.”
Senior Data Scientist specializing in ML, NLP, and GenAI analytics
“Built and deployed an LLM-powered analytics assistant enabling business users to ask questions in plain English and receive validated Spark SQL executed in Databricks, with a Streamlit/Flask UI. Addressed strict client schema-privacy constraints by implementing a RAG strategy and ultimately leveraging AWS Bedrock and fine-tuned reference docs. Also has production ML pipeline experience using Docker + Airflow and AWS (S3/ECS/EC2) for financial classification models.”
Mid-level Software Engineer specializing in FinTech full-stack and AI applications
“Built and productionized an NLP-powered customer support assistant at JPMorgan Chase for digital banking, focused on reducing response time for repetitive client queries. Strong in real-world AI deployment challenges—sensitive data handling, low-latency FastAPI services, and AWS/Kubernetes operations with CI/CD—plus a metrics- and guardrails-driven approach to reliable AI workflows.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, RAG, and MLOps
“AI/ML engineer with HP experience building and productionizing an LLM-powered document intelligence platform (LangChain + Pinecone) to deliver semantic search and contextual Q&A across millions of enterprise support documents. Demonstrates strong MLOps and scaling expertise (Airflow, Kubernetes autoscaling, Triton GPU inference, monitoring with Prometheus/W&B) plus a structured approach to evaluation (A/B tests, shadow deployments, failover) and effective collaboration with non-technical stakeholders.”
Mid-level AI/ML Engineer specializing in GenAI, RAG pipelines, and cloud MLOps
“Built and deployed a production LLM + vector search clinical decision support system at UnitedHealth Group, retrieving medical evidence and patient context in real time for prior authorization and risk scoring. Strong in end-to-end RAG architecture (Hugging Face embeddings, Pinecone/FAISS, SageMaker, Redis) plus orchestration (Airflow/Kubeflow) and rigorous evaluation/monitoring, with demonstrated ability to align solutions with clinical operations stakeholders.”
Mid-level Data Engineer specializing in cloud data warehousing and analytics
“Data engineer at American Express who owned end-to-end pipelines for transaction and customer data used in finance reporting and risk analytics, processing ~5–8M records/day. Built Airflow-orchestrated ingestion (including external APIs/web sources) with strong data quality controls, monitoring/alerts, and resilient backfill/retry patterns, and also shipped a versioned REST API serving aggregated metrics to analytics teams.”
Mid-level Software Development Engineer specializing in cloud-native AI/ML systems
“AI/ML-focused engineer with practical experience building RAG-based and multi-agent systems, including architectures for retrieval, reasoning, context processing, and response generation. Stands out for combining LLM productivity gains with disciplined software engineering practices like validation, monitoring, and reproducibility.”
Mid-Level Software Engineer specializing in secure cloud microservices and FinTech
“Built and owned major parts of a real-time distributed AI fraud-detection pipeline (ingestion, inference microservice integration, and automated action layer), optimizing latency and observability and reducing false positives by ~35%. Understands ROS/ROS2 concepts (nodes/topics/services) and planned hands-on ramp-up via ROS2 pub/sub exercises and Gazebo simulation, but has not worked on physical robots or ROS in production.”
Mid-level Software Engineer specializing in AI and full-stack healthcare platforms
“Built and deployed a RAG-based clinical knowledge assistant at GE Healthcare to help clinicians query large volumes of messy, unstructured clinical documents with grounded, cited answers. Hands-on across the full stack (OCR/ETL, de-identification for PHI, Azure OpenAI embeddings, Cosmos DB indexing, FastAPI/Django) with production monitoring via LangSmith and performance tuning through batching and index optimization.”
Mid-level Data Engineer specializing in Analytics & AI/ML
“Data engineer with experience at Sony and Walmart building high-volume, near-real-time analytics and ingestion systems. Has owned end-to-end pipelines from Kafka/Spark streaming through S3/Parquet and Redshift/Looker, emphasizing data quality (Great Expectations), observability (CloudWatch/Azure Monitor), and reliability (Airflow SLAs, retries, checkpointing), including measurable performance and latency improvements.”
Mid-level Data Engineer specializing in Azure ETL/ELT and data warehousing
“Data engineer who has owned end-to-end production pipelines for customer transaction data (~2–5 GB/day) using Python/PySpark/SQL and Airflow, delivering major reliability and speed gains (70% faster reporting; 60–70% fewer data issues). Also built a daily external web-scraping system with anti-bot handling and safe, idempotent Airflow-driven backfills, plus a Python data API optimized with indexing/caching and tested for correctness.”
Mid-level Data Engineer specializing in real-time analytics and regulated domains
“Data platform engineer focused on large-scale, real-time fraud systems, with hands-on ownership of streaming architectures using Kafka, Spark, Snowflake, and Databricks. Stands out for combining performance tuning and platform automation with LLM/RAG-based enrichment, delivering measurable gains in latency, fraud accuracy, false positives, and analyst decision speed.”
Mid-level Business Analyst specializing in BI, reporting, and data insights
“Healthcare analytics professional with experience at UnitedHealth Group, focused on turning messy claims, eligibility, and provider data into clean reporting datasets and Power BI dashboards. Combines SQL and Python automation with strong stakeholder alignment around KPI definitions, helping operations teams improve claim turnaround visibility and cost efficiency.”
Mid-level Data Engineer specializing in scalable pipelines, Spark, and cloud data warehousing
“Backend/data platform engineer who recently owned an end-to-end large-scale financial data platform delivering real-time decision support for finance and operations. Has hands-on experience modernizing legacy batch pipelines into AWS cloud-native ELT with parallel-run cutovers, strong data quality controls (dbt-style tests, reconciliation), and measurable improvements in runtime, cost, and SLA compliance. Also builds scalable, secure FastAPI microservices using Docker, ALB-based horizontal scaling, Redis caching, and managed auth with Cognito/Supabase plus Postgres RLS.”
“Built and productionized an LLM-powered PDF document Q&A system to eliminate manual searching through long documents, focusing on scalability and answer reliability. Implemented semantic chunking (using headings/paragraphs/tables), overlap, and preprocessing/quality checks to reduce hallucinations, and orchestrated the end-to-end pipeline with Airflow using retries, alerts, and parallel tasks.”
Mid-level Data Engineer specializing in real-time pipelines and cloud analytics
“Researcher from the University of South Dakota who built a production medical RAG system to help interpret model predictions by retrieving relevant clinical notes and medical literature, overcoming retrieval accuracy and imaging-dataset challenges through semantic chunking and metadata-driven indexing. Also has hands-on orchestration experience with Airflow and Azure Data Factory, plus a pragmatic approach to LLM evaluation and stakeholder-driven iteration.”
Senior Data Scientist specializing in machine learning and customer analytics
“Data/ML practitioner with experience applying NLP and classical ML to large-scale customer data (2B+ records) for segmentation, prediction, and survey-text classification, delivering measurable business impact (~18% engagement efficiency). Has hands-on entity resolution across multi-source datasets and has built embedding-based semantic search using SentenceBERT + a vector database with domain fine-tuning (~20% relevance improvement), plus production workflow experience with Spark/Airflow and cloud tooling (AWS/Azure).”
Senior Data Engineer specializing in Azure Lakehouse, Databricks/Spark, and Snowflake
“Data engineer/platform builder with experience across PwC and Liberty Mutual delivering high-volume, production-grade pipelines and real-time data services. Has owned end-to-end streaming + batch architectures on AWS and Azure, including web scraping systems, with quantified reliability gains (99.9% availability, 90%+ error reduction, 30% latency reduction) and strong observability/CI-CD practices.”
Mid-level Software Engineer specializing in AI/ML and data platforms
“AI/ML engineer who built a production agentic system to automate computational research experiments (simulation execution, parameter exploration, and numerical analysis) and mitigated context-window failures using constrained tool-calling/prompt-chaining patterns in LangChain with OpenAI tool-enabled models. Also has adtech/big-data pipeline experience at InMobi, orchestrating Spark jobs in Airflow to filter bot-like user IDs and publish clean IDs to an online NoSQL store for live serving, plus Apache open-source collaboration experience.”
Mid-level AI/ML Engineer specializing in healthcare NLP and MLOps
“Healthcare/clinical ML practitioner who built and productionized ClinicalBERT-based pipelines to extract and standardize oncology EHR data, improving downstream model F1 from 0.81 to 0.92 while controlling training cost via LoRA/QLoRA. Experienced orchestrating real-time AWS ETL/ML workflows (Glue, Lambda, SageMaker) and partnering with clinicians using SHAP-based interpretability, contributing to an 18% reduction in readmissions and full adoption.”
Mid-level Data Engineer specializing in cloud data pipelines and real-time streaming
“Data engineer with PNC Bank experience owning high-volume financial transaction pipelines end-to-end (Kafka/REST ingestion through Spark/Glue transformations to Redshift serving) for risk and fraud analytics. Built strong reliability and data quality practices (Great Expectations, reconciliation, Airflow alerting, idempotent retries, incremental/windowed processing), reporting 40% ingestion efficiency gains and ~99.9% data accuracy.”