We’re building Foundry to be the cloud compute platform AI developers actually want—no more battling procurement, limited quotas, or clunky tooling. Our platform gives ML engineers frictionless access to high-performance GPUs, clean APIs, and modern infra primitives to train, fine-tune, and serve state-of-the-art models. Backed by Sequoia, Lightspeed, and founders of Databricks, Google Brain, and Scale.
We’re hiring:
>> General Software (SWE) / Infrastructure Engineers << Build our batch + streaming workload engine for ML. Think: GPU scheduling, fault-tolerant execution, rich job DAGs. 0→1 ownership. Apply here: https://job-boards.greenhouse.io/foundrytechnologiesinc/jobs...
>> SRE, Supply (Site Reliability Engineer) << Manage GPU provisioning, spot bidding, and node pool health across clouds and on-prem. Work on the systems behind our global GPU fleet. Apply here: https://job-boards.greenhouse.io/foundrytechnologiesinc/jobs...
>> Solutions Engineers << Help top ML teams succeed on Foundry. Infra + ML intuition & technical customer facing prowess a must. Apply here: https://job-boards.greenhouse.io/foundrytechnologiesinc/jobs...
>> Founding Product Manager << Define the roadmap for the most advanced ML infra users. Apply here: https://job-boards.greenhouse.io/foundrytechnologiesinc/jobs...
Looking for compute? We support distributed deep learning workloads, including long-running batch jobs, streaming inference, and GPU autoscaling for LLM training and fine-tuning. https://mlfoundry.com/contact-sales
discovery tags: gpu, kubernetes, hpc, terraform, helm, distributed, ml, training, inference, llm, deep learning, fine tuning, customer, product, api, on-demand, infrastructure, elastic, scalable, cloud, platform