NSF and NVIDIA Are Spending $152 Million to Build an Open AI for Science. Here Is Why It Matters.

March 2, 2026 · 8 min read

David Almeida

The week that OpenAI closed its $110 billion funding round at a $730 billion valuation, the National Science Foundation and NVIDIA quietly announced a $152 million investment in something fundamentally different: fully open AI models built specifically for scientific research, with every component — weights, training data, code, and documentation — released for public use, modification, and reproduction.

The contrast is not accidental, and it's not subtle. One vision of AI concentrates capability behind API calls, licensing agreements, and corporate roadmaps that researchers cannot inspect, modify, or reproduce. The other makes transparency and reproducibility structural requirements — not aspirational values, but engineering constraints baked into the funding terms.

The project is called OMAI — Open Multimodal AI Infrastructure to Accelerate Science — and it represents NSF's most significant commitment yet to the proposition that open AI development is not just philosophically preferable but scientifically necessary.

What OMAI Will Build

The $152 million breaks down to $75 million from NSF and $77 million from NVIDIA, channeled through a Mid-Scale Research Infrastructure award led by Dr. Noah A. Smith at the Allen Institute for AI (AI2) in Seattle. AI2 was founded in 2014 by the late Paul Allen with a mandate to conduct open, high-impact AI research. The organization has consistently released models, datasets, and research tools without restrictions — a track record that made it the natural anchor for a project whose entire premise is that openness is a feature, not a concession.

The technical scope is ambitious. OMAI will develop leading multimodal models — systems that process text, images, and other data types — optimized for scientific research applications. Unlike commercial models that are trained on broad internet data and fine-tuned for general-purpose chat, OMAI's models will be designed to accelerate discovery across specific scientific domains.

Critically, every model will ship with "all of the components needed to analyze, modify, and fine-tune them, or even train them from scratch." That's a meaningful distinction from the "open weights" releases that some commercial AI companies offer, where you can download the model but can't see the training data, reproduce the training run, or understand the data filtering decisions that shaped the model's behavior. For scientific research, where reproducibility is foundational, the difference between "open weights" and "fully open" is the difference between a result you can cite and a result you can verify.

Why This Matters for Researchers

The practical problem OMAI addresses is straightforward: most academic researchers cannot afford to use commercial AI at scale, and even when they can, they can't inspect what the models are actually doing.

Consider a biomedical researcher using a commercial language model to analyze thousands of clinical trial abstracts. The model is a black box — trained on unknown data, filtered through unknown processes, making predictions based on patterns the researcher cannot examine. If the model consistently misclassifies a category of abstracts, the researcher has no way to diagnose why. If a journal reviewer asks whether the results would change with a different model or training approach, the researcher has no way to answer. The entire analysis rests on trust in a system designed for commercial applications, not scientific rigor.

OMAI aims to solve this by providing models that researchers can audit end-to-end. If a model produces unexpected results, you can trace the issue to training data, model architecture, or fine-tuning decisions. If you need to adapt a model for a specialized domain — marine biology, materials science, clinical genomics — you can fine-tune it on your own data without licensing restrictions. If a reviewer questions your methodology, you can point to a fully reproducible pipeline.

The compute access component is equally important. Training and running large AI models requires GPU infrastructure that individual research groups rarely possess. The NVIDIA component of the OMAI funding provides hardware access through the partnership, and the project's collaboration with Cirrascale Cloud Services establishes a compute pathway for researchers who don't have their own GPU clusters.

The Institutional Architecture

OMAI is not a single-institution project. The collaboration spans AI2 (lead), the University of Washington (co-PI: Hanna Hajishirzi, one of the most cited researchers in natural language processing), the University of Hawai'i at Hilo, the University of New Hampshire, and the University of New Mexico. The choice of partner institutions is deliberate — it includes both a world-class AI research center and a set of institutions that represent the kind of geographic and institutional diversity that NSF has been increasingly prioritizing.

This matters because one of the persistent critiques of AI research infrastructure is that it concentrates at a handful of elite institutions with existing GPU clusters, industry partnerships, and deep AI expertise. A researcher at the University of New Mexico or the University of Hawai'i at Hilo faces a fundamentally different compute landscape than a researcher at Stanford or MIT. OMAI's distributed architecture is designed to change that equation — not by eliminating the gap entirely, but by creating shared infrastructure that institutions across the capability spectrum can access.

The project fits into a broader NSF strategy that has been taking shape across multiple programs. The NAIRR Operations Center is building a national compute resource for AI research. The NSF Tech Labs initiative is funding independent research organizations with massive multi-year awards. The $100 million quantum and nanotechnology infrastructure program is establishing shared fabrication and characterization facilities. Taken together, these initiatives represent a systematic effort to build national research infrastructure that doesn't depend on any single commercial vendor or institutional host.

Open vs. Closed: The Scientific Stakes

The argument for open AI in science goes beyond ideological preference. It addresses a structural vulnerability in the research enterprise.

When researchers build their work on top of closed commercial models, they create dependencies that compromise both reproducibility and continuity. Commercial AI companies change their models without notice — updating weights, modifying safety filters, adjusting pricing, or deprecating versions entirely. A research pipeline that depends on GPT-4.1 in March 2026 may produce different results in September 2026 if OpenAI updates the model. The researcher may never know the results changed, because the model version is a black box maintained by a third party.

This is not a theoretical concern. Multiple studies have documented "model drift" — the phenomenon where commercial AI models change behavior over time without explicit notification to users. For research that requires precise reproducibility, building on a commercial model is like running an experiment on equipment whose calibration changes overnight without your knowledge.

Open models eliminate this problem by design. When you have the full model — weights, training code, data, documentation — you control the entire pipeline. You can freeze a version for reproducibility, modify it for your specific needs, and share it with collaborators who can reproduce your exact results. The model becomes a research instrument that the researcher controls, not a service that a vendor manages.

The commercial AI industry's response to openness concerns has been mixed. Some companies release model weights while withholding training data — a halfway measure that enables fine-tuning but not full reproducibility. Others argue that fully open releases create safety risks by enabling misuse. OMAI's position, backed by $152 million in public and private investment, is that the scientific benefits of full openness outweigh the risks — and that the scientific community needs at least one fully transparent alternative to commercial offerings.

What This Means for Grant Proposals

If you're writing proposals that involve AI or machine learning, OMAI changes the landscape in several concrete ways.

Reproducibility sections get stronger. Review panels across NSF, NIH, and DOE have been increasingly scrutinizing the reproducibility of AI-dependent research. Proposals that commit to using fully open models — where the entire pipeline can be inspected, reproduced, and shared — address the reproducibility concern more convincingly than proposals built on commercial APIs. When OMAI models become available, referencing them as your AI infrastructure provides a concrete answer to the inevitable "how will you ensure reproducibility?" question.

Budget justifications get easier. One of the persistent challenges in AI-dependent research proposals is justifying compute costs. Commercial API fees scale unpredictably with usage, making multi-year budget projections unreliable. OMAI's public infrastructure model provides a cost structure that can be projected more accurately — and, for some use cases, eliminates the cost entirely.

Broader impacts write themselves. NSF proposals require a broader impacts section, and proposals involving AI frequently struggle to articulate impacts beyond "we'll use AI to do science better." OMAI's emphasis on open, reproducible, democratically accessible AI infrastructure provides a natural broader impacts narrative: your research contributes to an open scientific commons that researchers at under-resourced institutions can access and build upon.

Cross-disciplinary collaboration becomes more feasible. A chemist and a computer scientist working on the same project can both access, modify, and understand the AI models in the pipeline. When the model is a black box, the computer scientist holds all the interpretive authority. When the model is fully open, domain experts can examine how the model handles their specific data — a structural improvement in the quality of interdisciplinary research.

The Bigger Picture

OMAI is a $152 million bet — significant by NSF standards, trivial by comparison to the $110 billion that OpenAI raised in a single round. The scale mismatch is unavoidable. But it would be a mistake to evaluate OMAI purely on dollar terms.

What OMAI represents is the federal government taking a position on a question that the AI industry would prefer to treat as settled: whether the future of AI is open or closed. NSF's answer, backed by real money and a concrete engineering plan, is that at least for science, the future must be open — because the entire point of scientific research is that other people can check your work.

For the hundreds of thousands of researchers whose work increasingly depends on AI tools they didn't build, can't inspect, and don't control, that's not an abstract principle. It's the difference between doing science and doing something that looks like science but can't be verified.

The OMAI models aren't available yet — development timelines suggest initial releases later in 2026 — but the infrastructure decisions you make now will determine whether you can take advantage of them when they arrive. Orienting your research workflows toward open, reproducible AI pipelines is good practice regardless of OMAI, and it positions you to leverage the most significant public AI investment in American research history when it goes live. Granted can help you identify AI-related funding opportunities across NSF, DOE, NIH, and DARPA that align with open research infrastructure — building proposals that ride the wave rather than swimming against it.

Get AI Grants Delivered Weekly

New funding opportunities, deadline alerts, and grant writing tips every Tuesday.

Browse all NSF-AI grants

More NSF-AI Articles

Not sure which grants to apply for?

Use our free grant finder to search active federal funding opportunities by agency, eligibility, and deadline.

Find Grants

Ready to write your next grant?

Draft your proposal with Granted AI. Win a grant in 12 months or get a full refund.

Backed by the Granted Guarantee