Independent news & culture since 2025
Thursday, April 23, 2026

The Daily Scroll

Where Every Story Has a Voice

Featured image: Astronomers Are Quietly Making the GPU Crisis Much Worse
Tech

Astronomers Are Quietly Making the GPU Crisis Much Worse

The search for galaxies is eating compute you didn't know was missing.

You've been blaming the GPU shortage on crypto bros and AI chatbots. Turns out, the universe itself is also placing an order.

Enjoying this? Never miss a story.

A growing wave of AI-powered astronomy projects — tools designed to scan, classify, and analyze galaxy imagery at a scale no human team could manage — is quietly adding to the global GPU crunch that's already throttling everything from cloud computing costs to the pace of AI research. This isn't a hypothetical future problem. It's happening right now, and the research community is increasingly candid about it.

Introduction

The AI galaxy hunter story sounds like something from a press release written to make academics seem cool. And honestly, parts of it are. But underneath the headline is a real resource conflict playing out across data centers, university compute clusters, and commercial cloud platforms — one that has direct consequences for how much you pay for cloud storage, how long AI model training queues run, and whether the next generation of scientific research gets funded or waitlisted.

Article photo 1

The core tension is this: astronomy has always been data-heavy, but the new generation of sky surveys — particularly the Vera C. Rubin Observatory's Legacy Survey of Space and Time (LSST), which begins full operations in 2025 and is expected to generate roughly 20 terabytes of image data per night — has crossed a threshold where traditional computing pipelines simply can't keep up. The answer, increasingly, is AI. And AI means GPUs. Lots of them.

What follows is a look at what these projects actually do, how much compute they consume, why that demand is accelerating, and what it means for the rest of us who are competing for the same finite hardware. Spoiler: the sky isn't the limit. The GPU rack is.

What AI Galaxy Hunters Actually Do

Before we get into the resource conflict, it's worth being precise about what we mean by "AI galaxy hunters" — because the term covers a range of genuinely different projects, not just one monolithic effort.

Article photo 2

The most prominent category involves automated galaxy classification and morphology detection. Tools like Morpheus, developed by researchers at UC Santa Cruz, use deep learning models to analyze astronomical images pixel-by-pixel and classify celestial objects — galaxies, stars, artifacts — at a scale that would take human volunteers years to complete manually. (The company calls this "AI-powered." What it actually does is pattern-match on billions of pixels trained against labeled astronomy datasets. Which is still impressive, just less magical than the press releases suggest.)

Then there's gravitational lens detection — finding cases where a massive galaxy bends light from something behind it, creating a natural telescope effect. These events are rare and visually subtle. Human reviewers miss them constantly. Models like those developed by the Dark Energy Survey collaboration have found thousands of previously unidentified lens candidates in archival data, precisely because they can process image libraries that no human team has the bandwidth to review.

The Data Scale Is the Problem

Here's what's actually happening: the datasets these tools train and run inference on are enormous. The Sloan Digital Sky Survey, which ran from 2000 to the mid-2010s, produced around 100 terabytes of data over its entire operational life. The Rubin Observatory's LSST will produce that in roughly five nights.

Article photo 3

Training a single galaxy morphology model on a dataset derived from the Hyper Suprime-Cam Subaru Strategic Program — one of the deeper current sky surveys — can require hundreds of GPU-hours on A100-class hardware. That's not a one-time cost, either. Models are retrained as new data arrives, hyperparameters are tuned, and competing research groups often run parallel experiments on the same underlying data.

The National Science Foundation's NOIRLab, which manages several major observatories and their associated data pipelines, has been increasingly vocal about compute bottlenecks. Researchers affiliated with the project have noted in recent publications that GPU allocation has become a genuine constraint on the pace of science — not the telescopes, not the sensors, not the funding for human researchers. The GPUs.

Why This Is Happening Now, Not Five Years Ago

The timing isn't accidental. Before roughly 2019, most astronomical image analysis ran on CPU clusters — slower, but sufficient for the data volumes of the era. The combination of two things changed that: the arrival of transformer-based vision models that dramatically outperformed older convolutional networks on image classification tasks, and the simultaneous commissioning of a new generation of wide-field survey telescopes that produce data volumes those older models couldn't have handled anyway.

Article photo 4

The Rubin Observatory specifically was designed from the ground up with the assumption that AI pipelines would handle classification and anomaly detection. It wasn't retrofitted. That design choice was made in the early 2010s, when GPU scarcity was not yet a civilization-level concern. The observatory's data management team has been working with partners including the SLAC National Accelerator Laboratory and Google Cloud to provision enough compute to handle the nightly data flood — but "enough" keeps moving.

Meanwhile, the broader GPU market hit a wall in 2023 that it still hasn't fully climbed over. NVIDIA's H100, the current workhorse of serious AI training, had a list price of around $30,000 per unit at launch and was trading for significantly more on secondary markets through most of 2023 and into 2024. Cloud GPU instance costs followed. A single A100 instance on AWS currently runs between $3 and $32 per hour depending on configuration. For a research team running training jobs across a week, that's a non-trivial line item on a grant budget that was written two years ago.

The Competition Nobody Talks About

Is this a problem? Depends on who you ask.

Article photo 5

If you're a startup trying to fine-tune a language model on a tight timeline, you probably don't think of astrophysics researchers as your competition for GPU time. But on the shared infrastructure of commercial cloud platforms and NSF-funded high-performance computing clusters, they absolutely are. Every node running a galaxy classification job is a node not running your inference workload.

The HPC (high-performance computing) community has a term for this: "queue pressure." It refers to the backlog of compute jobs waiting for available hardware. At major academic clusters like those operated by XSEDE (now ACCESS, the NSF's advanced computing infrastructure program), queue wait times for GPU-intensive jobs have increased measurably over the past two years. A job that might have started within hours in 2021 can now sit for days.

Dr. Francesca Civano, an astrophysicist at the Center for Astrophysics at Harvard & Smithsonian, has described the situation in interviews as a "compute arms race" where scientific fields that never historically competed with industry AI are now doing exactly that — because they're all drawing from the same hardware pool, whether directly through cloud platforms or indirectly through shared national infrastructure.

Article photo 6

The Funding Gap Makes It Worse

Here's the part that doesn't get enough attention: commercial AI labs can afford to buy or reserve hardware at scale. A university research group with a $500,000 NSF grant cannot. This creates a structural disadvantage that's widening, not narrowing.

When Meta or Google wants more GPU capacity, they call NVIDIA's enterprise sales team and negotiate a multi-year contract. When an astronomy team at the University of Arizona wants to run a training job for a gravitational lens detection model, they apply for allocation time through ACCESS, wait for the review cycle, and hope the cluster isn't already oversubscribed. The timelines are incompatible with the pace of modern AI development, where a model architecture that's cutting-edge today may be obsolete in eight months.

Some research groups have started migrating to commercial cloud specifically to escape the queue — which means burning grant money on AWS or Google Cloud at rates that weren't budgeted when the grants were written. It's a slow-motion funding crisis that doesn't show up in any single headline.

Article photo 7

The Broader GPU Crunch: Context That Matters

To understand why astronomy's compute appetite matters beyond academic circles, you need the full picture of where GPU demand is coming from right now.

NVIDIA reported revenue of $22.1 billion in Q3 of its fiscal year 2024 — a 206% increase year-over-year — almost entirely driven by data center GPU demand. The company's H100 and the newer H200 are backordered at virtually every major cloud provider. Microsoft, Google, and Amazon have all disclosed multi-billion-dollar GPU procurement commitments over the next several years, and they're still not keeping up with internal demand, let alone external customers.

Into this environment, add: generative AI startups, autonomous vehicle training pipelines, pharmaceutical drug discovery models, financial risk modeling, and now large-scale scientific computing across astronomy, climate science, and genomics. Each of these sectors has independently concluded that the answer to their data problem is "more GPUs." They're all correct. And there aren't enough.

Article photo 8

For more on how the AI infrastructure investment wave is reshaping the startup landscape, see our piece on The Google Cloud Next 2026 Startups Nobody Is Talking About Yet — several of those companies are building specifically around GPU efficiency, which tells you something about where the market pain is concentrated.

What the Research Community Is Actually Doing About It

To their credit, astronomers aren't just complaining. Several research groups have shifted toward model efficiency as a design priority — not because they want to, but because they have to.

The AstroML community, which maintains open-source machine learning tools for astronomy, has been increasingly focused on smaller, faster models that can run on CPU clusters or even consumer-grade GPUs. There's also growing interest in federated approaches — training models across distributed data centers rather than centralizing everything on one massive cluster, which reduces the single-point bottleneck even if it introduces new coordination complexity.

Article photo 9

Some groups are partnering directly with cloud providers. The Rubin Observatory has a formal relationship with Google Cloud that includes reserved compute capacity — essentially a guaranteed lane in the traffic jam. But that arrangement is the exception, not the rule, and it required years of negotiation and significant institutional resources to establish.

The more ambitious proposal floating around the astrophysics community is a dedicated scientific computing facility — essentially a national GPU reserve for research that can't compete commercially. The idea has been floated in NSF planning documents and discussed at American Astronomical Society meetings, but it's nowhere close to funded. The estimated cost for a facility that could meaningfully serve the Rubin-era data pipeline is north of $200 million in capital expenditure, before you get to operating costs.

Why You Should Actually Care About This

Here's the thing: this isn't just a story about scientists needing more computers. It's a story about what happens when the infrastructure assumptions of an entire field of research are invalidated by a market shock they didn't cause and can't control.

Astronomy isn't alone. Climate modeling teams at NOAA and ECMWF are facing the same pressure. Genomics researchers running protein folding models post-AlphaFold are in the same queue. The GPU crunch is functioning as an invisible tax on scientific progress — one that doesn't show up in any single budget line but accumulates across thousands of research projects running slower, smaller, or less frequently than they should.

And unlike commercial AI applications, the output of this research doesn't have a monetization path that can justify paying premium cloud rates indefinitely. There's no galaxy classification startup that's going to IPO and fund the next generation of telescope data pipelines. This is publicly funded science competing on a commercial infrastructure market, and it is losing ground.

Also worth noting: the Anthropic $5B Amazon deal we covered earlier this year is a useful case study in how concentrated GPU access is becoming. When a single AI safety company can lock in that scale of cloud commitment, it reshapes the availability curve for everyone else — including researchers who have no seat at that table.

The Bottom Line

The AI galaxy hunter story is real, it's timely, and it's more consequential than it sounds. Astronomy projects are consuming meaningful GPU capacity at exactly the moment when that capacity is most constrained globally — not because astronomers are being reckless, but because the science genuinely requires it and the alternative is leaving 20 terabytes of nightly telescope data sitting unanalyzed in a storage bucket.

The GPU crunch isn't going to be solved by any single sector standing down. It will be solved — if it's solved — by a combination of hardware supply catching up (NVIDIA's Blackwell architecture is promising, but volume shipments are still ramping), model efficiency improving (smaller models doing more with less), and infrastructure policy catching up to the reality that scientific computing is now competing directly with commercial AI for the same chips.

If you work in tech, the actionable insight here is this: the GPU queue problem is not going away, and it's not going to be contained to the sectors you're already watching. Budget accordingly, plan for longer provisioning timelines than you'd like, and maybe don't assume that your cloud GPU reservation from 2023 reflects what 2026 pricing and availability will look like. The astronomers found galaxies. They also found the bottom of the compute barrel. You're both looking at the same empty shelf.

Some links in this article may earn us a small commission — at no extra cost to you.