San Francisco Compute
@sfcompute
we sell very-large H100 training clusters with IB & fast disk you can buy in short-bursts (ie: 96 H100s for a week) 🇺🇸🏳️⚧️🌁
every time we work with a customer on inference we deeply embed with them, setup an apples-to-apples comparison, and use that to quote a price that beats what they're currently doing while keeping the same accuracy we're starting to productize that whole experience now
buried in @sriramk's America's AI Action Plan is endorsement that the US compute market will financialize with spot and forward contracts. this podcast explains why this is so necessary, not just for speculation one of the most consistent themes with @latentspacepod's GPU…
🆕 SF Compute: Commoditizing Compute latent.space/p/sfcompute We're excited for our latest deep dive into the compute market with @evanjconrad of @sfcompute! It should not be normal for the prices of one of the world’s most important resources right now to swing from $8 to $1…
we, sfcompute, are hiring someone for a short term contract that involves running around the city if you are in sf, want to work in startups, and want to have an adventure, please DM us
I'm very excited to partner with SFCompute - Evan and team are phenomenally driven and built a powerful platform for scaling GPU solutions like never before. Combined with Modular's high-performance inference solutions, they're able to deliver incredible TCO advantages! 👇
We've partnered with Modular to create Large Scale Inference (LSI), a new OpenAI-compatible inference service. It's up to 85% cheaper than other offerings & can handle trillion-token scale. We originally created it at the request of a major AI lab to do large scale multimodal…
We've partnered with Modular to create Large Scale Inference (LSI), a new OpenAI-compatible inference service. It's up to 85% cheaper than other offerings & can handle trillion-token scale. We originally created it at the request of a major AI lab to do large scale multimodal…
coming soon june 2025 cuda crunch
coming soon june 2025 cuda crunch
Alpha users on @sfcompute can inference @deepseek_ai v2 Prover 671B on 24 H100s IB at market rate since a few minutes after downloaded, and cache compiled. @vllm_project project, multi-node LWS ready. $0.44/gpu/hr right now per hour, but market rates apply.
🆕 SF Compute: Commoditizing Compute latent.space/p/sfcompute We're excited for our latest deep dive into the compute market with @evanjconrad of @sfcompute! It should not be normal for the prices of one of the world’s most important resources right now to swing from $8 to $1…
This is fucking _awesome_ and trained on @sfcompute. Lots of intense work into making this product really great, and stuff like this is why I'm so excited. More accessible AI infrastructure. You can't do this slinging YAML alone.
Today, we're releasing a new paper – One-Minute Video Generation with Test-Time Training. We add TTT layers to a pre-trained Transformer and fine-tune it to generate one-minute Tom and Jerry cartoons with strong temporal consistency. Every video below is produced directly by…
Side note everyone - we are hiring a Financial Systems Engineer at @sfcompute! Come help build a market that enables liquidity in a GPU market that otherwise prefers massive up front commitments, and enables researchers and companies to get the resources they need for ML.
r u an engineer that likes finance? @sfcompute is hiring an engineer to build the platform that allows traders trade gpu compute
app banner - no layout shift - doesn't block render - corner radii change to 0 to reduce visual contrast - bezier curves for the notch
This is very cool
Happy to announce what @rami_mmo and I have been working on for the past few months at YC. At Lucid we're building generative simulations powered by fast world models. We trained a minecraft simulator that runs smoothly on gaming gpus, and are now training a real world sim!
kinda wild how easy it is to get access to a supercomputer from anywhere in the world