Pre-screened and vetted.
Senior Software Engineer specializing in Python and AWS cloud backend systems
Staff Software Engineer specializing in Healthcare SaaS and real-time systems
Mid-Level Software Development Engineer specializing in AWS serverless and ML/GenAI
Senior Full-Stack Software Engineer specializing in large-scale streaming platforms
Staff Software Engineer specializing in FinTech and scalable distributed systems
Mid-level Machine Learning Engineer specializing in generative AI, NLP, and MLOps
Senior Machine Learning Engineer specializing in GenAI, NLP, and recommendation systems
Director-level Software Development Manager specializing in large-scale cloud platforms
Intern Software Engineer specializing in distributed systems and cloud infrastructure
“Built and operated a production warehouse metadata collection platform at Sigma Computing, integrating Go/gRPC services with a TypeScript backend and MySQL, with strong emphasis on idempotency, retries, bounded-concurrency job queues, and Datadog-based observability. Also created Kurral (kurral.com), an AI agent runtime security and observability/governance SDK/proxy concept, iterating via pilot-customer feedback and market research; targeting founding engineer roles with $180–200k base and ~2–5% equity.”
Entry-level Software Engineer specializing in full-stack and AI systems
“Frontend-leaning full-stack engineer who described owning an artist search and detail experience across UI, backend integrations, and data modeling. They show practical strength in scalable React architecture, TypeScript safety, and performance tuning, with a product-minded approach to shipping 0→1 features quickly and iterating after launch.”
Mid-Level Software Engineer specializing in cloud infrastructure and full-stack web development
“Backend engineer at Electric Hydrogen who built a serverless device-log ingestion and processing platform in Python/Flask, scaling throughput (4x peak ingestion) while keeping sub-300ms API latency. Strong in Postgres/SQLAlchemy performance (partitioning, materialized views) and production ML integration (ONNX model served via FastAPI microservice with async batch inference, Redis feature caching, and drift monitoring via S3/Lambda). Experienced designing secure multi-tenant systems with schema-per-tenant isolation and KMS-backed encryption.”
Senior Software Engineer specializing in FinTech and distributed systems
“Backend/AI engineer who has built a rule-service platform on AWS and evolved it into an agentic RAG system using LangChain, ReAct, tool calling, and LLM-as-judge review. Notable for combining heavy AI-assisted development with production safeguards like manual CR, CloudWatch monitoring, fallback strategies, benchmark testing, and user-feedback-driven model improvement.”
Junior Software Engineer specializing in backend systems and AI/ML pipelines
“Robotics-focused engineer with ROS 2 experience who has built and debugged real-time, distributed control/orchestration systems under production-like latency and safety constraints. Led platform changes at Persona for a real-time verification orchestration system using deterministic state machines and async workers, and has hands-on experience stabilizing multi-robot navigation/SLAM behavior using rosbag, RViz, and stress testing in simulation (Gazebo).”
Mid-Level Software Engineer specializing in AWS data infrastructure and pipeline automation
“AWS-focused software engineer who built a self-serve ETL pipeline scheduling service for non-engineers, including automated CloudFormation-based onboarding that cut setup time from 2–3 weeks to ~5 minutes. Strong in production reliability and customer-facing data platforms (EMR/DynamoDB/Lambda), with examples spanning pagination at scale, cross-table consistency, and phased rollouts to improve Parquet log SLAs.”
Mid-level AI/ML Engineer specializing in LLMs, RAG, and scalable inference
“Backend/retrieval-focused engineer with production experience at Perplexity building a large-scale real-time Q&A system using retrieval-augmented generation, emphasizing low-latency, high-quality answers through ranking, context optimization, and caching. Also has orchestration experience from both product-facing LLM pipelines and large-scale infrastructure workflows at Meta, and has partnered with non-technical stakeholders to align AI trade-offs with business goals.”
Mid-level AI/ML Engineer specializing in LLM fine-tuning, inference optimization, and AI safety
“AI/LLM engineer with production experience at NVIDIA, where they fine-tuned and deployed a financial-services chatbot and cut latency ~50% using TensorRT + NVIDIA Triton, scaling via Docker/Kubernetes. Also has consulting experience at Accenture delivering a predictive maintenance solution for a logistics network, bridging non-technical stakeholders with actionable dashboards.”
Executive CTO and Founder specializing in AI platforms and hyper-scale SaaS
“CTO-minded builder seeking to join a startup; previously created an AI-driven platform that abstracted away DevOps and infrastructure for drug discovery researchers. Emphasizes high-leverage, zero-to-one execution with managed cloud/open-source tooling, and a strong reliability/reproducibility mindset validated against existing scientific pipelines.”
“Backend/full-stack engineer (Amazon experience) who built an AWS-based integration testing platform using Flask, ECS, Docker, and CloudWatch—cutting 1000+ test cases from ~5 hours to ~30 minutes while improving log visibility for non-engineering users. Also led a zero-downtime EU region migration with rigorous ORR testing, and built a Kinesis/Firehose/S3 + Glue/Spark replay mechanism for resilient data recovery. Side project: reproducible, cost-efficient LLM hosting platform on EKS using CDK and Karpenter for scale-to-zero.”
Mid-level AI/ML Engineer specializing in GPU inference and LLM platforms
“Built and deployed an LLM-powered platform that turns models into scalable REST/gRPC APIs, focusing on keeping GPU-backed inference fast and stable during traffic spikes. Experienced with AWS orchestration (EKS/ECS/Step Functions), safe model rollouts, and production-grade monitoring/testing for reliable AI agents and workflows.”