Pre-screened and vetted.
Mid-Level Software Engineer specializing in cloud, microservices, and AI/ML
“Backend/API engineer with ~4 years experience building production services in .NET Core/PostgreSQL/Redis/Docker and optimizing real-world latency issues (claims ~60% response-time improvement). Also built and owned an end-to-end RAG-based AI assistant using Python/FastAPI, OpenAI APIs, and Pinecone, plus agentic workflows with reliability guardrails (retries, confidence thresholds, monitoring). Currently pursuing a master’s degree and targeting a $150k base salary.”
Senior Data Engineer specializing in cloud data platforms and big data pipelines
“Data engineer focused on building reliable, production-grade pipelines and external data collection systems on AWS (S3/Lambda/SQS/Glue/EMR) using PySpark/SQL, serving curated datasets to Snowflake/Redshift for finance and fraud teams. Has operated a large-scale crawler ingesting millions of records/day with anti-bot tactics, schema versioning/quarantine, and CloudWatch/Datadog monitoring, and also shipped a versioned REST API with caching and query optimization.”
Mid-level ML Data Engineer specializing in MLOps and scalable healthcare data pipelines
“Data/ML platform engineer with healthcare (Cigna) experience owning an end-to-end pipeline spanning Airflow + Debezium CDC ingestion, PySpark/SQL transformations, rigorous data quality gates, and feature-store/API serving for ML training and inference. Worked at 10+ TB scale and cites a ~30% latency reduction plus stronger reliability via idempotent design, monitoring, and backfill-safe reprocessing; also built pragmatic early-stage data pipelines at Frankenbuild Ventures.”
Mid-level Software Engineer specializing in cloud microservices and data pipelines
“Data engineer/platform builder who has owned production pipelines end-to-end processing millions of records/day, with strong emphasis on data quality (quarantine workflows) and reliability (monitoring, retries, incremental loads). Also designed large-scale external data collection/crawling with anti-bot handling and backfills, and shipped versioned REST data services optimized for performance and developer usability in an early-stage environment.”
Entry-level AI/ML Engineer specializing in LLMs, RAG, and DevOps automation
“Built and owned a production-scale AI-driven software release/version intelligence platform orchestrated via GitHub Actions that tracks 1000+ upstream repositories and automatically generates SLA-bound JIRA upgrade tickets for hardened container images. Replaced brittle regex/PEP440 parsing with an LLM-based semantic filtering layer plus deterministic validation to handle noisy/inconsistent GitHub tags at scale, with monitoring for coverage, latency, and correctness validated against upstream ground truth.”
Junior MLOps Engineer specializing in LLMs and cloud infrastructure
“Built a production multimodal LLM system (Gemini on GCP) to automate behavioral coding of family-involved science experiment videos, including preprocessing for inconsistent lighting/audio and LangGraph-orchestrated parallel workflows. Also developed rubric-based AI grading workflows and partnered closely with non-technical education stakeholders through explainability-focused walkthroughs and manual-vs-AI evaluation alignment.”
Senior Backend Software Engineer specializing in Java microservices, Kafka, and AWS
“AI engineer who shipped a production chat assistant for a storage company by building the underlying RAG-style knowledge base (document ingestion, chunking/embeddings, FAISS vector store) and an admin update interface to keep content current. Also has full-stack delivery experience (Python REST APIs + React/TypeScript UI) and AWS operations using Terraform/Jenkins, including handling a real production performance incident by optimizing DB queries and adding auto-scaling.”
Mid-level Data Engineer specializing in cloud lakehouse, streaming, and MLOps
“Data engineer at AT&T focused on large-scale telecom (5G/IoT) data platforms, owning end-to-end pipelines from Kafka/Azure ingestion through Databricks/Delta Lake transformations to serving analytics and ML. Has operated at very high volumes (~50+ TB/day) and delivered measurable performance gains (25–30% faster processing) plus improved reliability via Airflow monitoring, robust data quality checks, and resilient external data collection patterns (rate limiting, retries, dynamic schemas).”
Mid-level Full-Stack Software Engineer specializing in cloud-native and AI-integrated systems
“Built and deployed a Virginia Tech CS department blog/archive application using a MERN/Next.js stack and a fully serverless AWS architecture (Lambda, API Gateway, S3, CloudFront, Route 53), including CI/CD via the Serverless Framework. Implemented RBAC for student/faculty/admin users and added an article export feature backed by MongoDB.”
Mid-level Software Engineer specializing in backend systems and data-driven APIs
“Candidate approaches AI-assisted coding like a senior developer supervising junior contributors: they define precise technical requirements, enforce code quality and documentation, and review outputs before approval. They also actively lead multi-agent workflows using OpenClaw and a Kanban-style AI project management setup, coordinating both coding and non-technical agents.”
Mid-level Machine Learning & GenAI Engineer specializing in LLMs, RAG, and NLP
“Built and deployed an LLM-powered customer support assistant (“Notable Assistant”) focused on automating common post-customer queries while maintaining multi-turn context and meeting scalability/latency needs. Experienced with production orchestration and operations using Kubernetes and Apache Airflow (DAG-based ETL, scheduling, monitoring/alerts), and has partnered closely with customer service stakeholders to align chatbot behavior with brand voice through iterative testing.”
Senior Data Scientist specializing in NLP, LLMs, and Computer Vision
“Applied NLP/ML engineer with experience at KeyBank and Novartis building production document intelligence and entity-resolution systems in finance and healthcare. Has delivered end-to-end pipelines (Airflow + AWS) using transformers (DistilBERT/Sentence-BERT), vector search (FAISS/Milvus/Pinecone), and human-in-the-loop labeling to achieve measurable gains (40%+ faster queries; up to 88% F1 and 93% precision/90% recall in entity linking).”
Senior AI/ML Engineer specializing in Python, RAG systems, and LLM fine-tuning
“Built and owned an end-to-end RAG-based AI support platform at Mechanize (FastAPI/LangChain/Pinecone/React) with rigorous evals and guardrails, driving 45% fewer support tickets and ~$280K annual savings. Also led a high-risk legacy modernization at Argo AI, incrementally extracting a monolithic Django backend using Strangler Fig + feature flags while supporting 10K+ concurrent users.”
Junior Full-Stack Software Engineer specializing in web platforms and data-driven search
Junior Full-Stack Software Engineer specializing in cloud-native microservices and web apps
Senior Full-Stack Software Engineer specializing in Healthcare IT and cloud migration
Senior Machine Learning Engineer specializing in LLMs, NLP, and computer vision
Senior Data Scientist specializing in machine learning, MLOps, and generative AI
Junior Machine Learning Engineer specializing in NLP, search, and performance optimization
Director of Analytics & Software Engineering specializing in ML, MLOps, and data platforms
Mid-level Data Engineer specializing in cloud data platforms and real-time pipelines