FindArticles FindArticles
  • News
  • Technology
  • Business
  • Entertainment
  • Science & Health
  • Knowledge Base
FindArticlesFindArticles
Font ResizerAa
Search
  • News
  • Technology
  • Business
  • Entertainment
  • Science & Health
  • Knowledge Base
Follow US
  • Contact Us
  • About Us
  • Write For Us
  • Privacy Policy
  • Terms of Service
FindArticles © 2025. All Rights Reserved.
FindArticles > News > Technology

Neurophos Secures $110M For Tiny Optical AI Chips

Gregory Zuckerman
Last updated: January 22, 2026 4:18 pm
By Gregory Zuckerman
Technology
6 Min Read
SHARE

Neurophos has raised $110 million to bring photonic computing to AI inference, betting that tiny metasurface modulators can deliver big gains in speed and energy efficiency. Spun out of Duke University and incubator Metacept, the Austin startup draws on metamaterials research that once enabled early “invisibility cloak” experiments and is now steering it toward the data center.

How Metamaterials Are Powering Next‑Gen AI Inference

The company’s core device is a metasurface modulator engineered to perform matrix–vector multiplications, the linear algebra at the heart of transformer inference. In essence, the modulator behaves like an optical tensor core: thousands of units are tiled onto a chip to process many operations in parallel using light rather than electrons.

Table of Contents
  • How Metamaterials Are Powering Next‑Gen AI Inference
  • Performance Claims and the Power Math Behind Them
  • Why This Matters for Data Centers and AI Scale-Out
  • Backers and the Competitive Landscape in Photonics
  • Roadmap and What to Watch as Neurophos Scales Up
A black server chassis with multiple circuit boards and processors, presented on a professional flat design background with soft patterns.

Neurophos argues that photons offer three advantages for AI workloads: minimal heat generation, high propagation speed, and resilience to electromagnetic noise. The challenge has always been density and manufacturability—optical components tend to be bulky, analog–digital conversions burn power, and yields can be unforgiving. The company claims its metasurface is roughly 10,000x smaller than traditional optical elements, enabling dense arrays that fit within standard chip footprints.

Performance Claims and the Power Math Behind Them

On paper, the performance targets are eye-catching. Neurophos says its optical processing unit runs at 56 GHz with a peak 235 peta-operations per second (POPS) while drawing 675 watts. By comparison, Nvidia’s B200 is cited at about 9 POPS at 1,000 watts. That implies a performance-per-watt uplift on the order of 39x (235/675 vs. 9/1000), assuming comparable operation definitions and low-precision arithmetic common in inference.

The company also projects roughly 50x gains in speed and energy efficiency over Blackwell-class silicon by its planned launch window, asserting an advantage that outpaces incremental node shrinks. It points to the historical cadence of leading-edge foundries—often mid-teens % efficiency improvements per node over multi-year intervals—as evidence that purely silicon-based progress is increasingly incremental.

Real-world throughput will hinge on more than raw POPS: memory bandwidth, nonlinear activation handling, model sparsity, and interconnect fabric all shape end-to-end latency. In most photonic accelerators, the optical array executes the linear layers while electronics handle activation functions and control. Neurophos says its approach is designed with this hybrid flow in mind.

Why This Matters for Data Centers and AI Scale-Out

Inference is becoming the dominant cost driver for AI services as deployed models serve billions of requests. The International Energy Agency has warned that data center electricity demand could roughly double over the next few years, with AI a material contributor. Any accelerator that delivers large efficiency gains without sacrificing accuracy stands to unlock capacity and cap operating costs for hyperscalers.

A man in a pink shirt working at a standing desk with multiple monitors in a modern office environment.

Neurophos contends its chips can be fabricated using standard silicon foundry materials, tools, and processes, a critical claim given the history of photonics startups constrained by bespoke manufacturing. If validated, that compatibility should simplify supply chain integration, packaging, and cost curves, especially as co-packaged optics and advanced substrates become mainstream in AI systems.

Backers and the Competitive Landscape in Photonics

The $110 million Series A is led by Gates Frontier with participation from Microsoft’s M12, Carbon Direct, Aramco Ventures, Bosch Ventures, Tectonic Ventures, Space Capital, and others. Microsoft technical leaders have publicly emphasized the need for step-change efficiency in core AI infrastructure, and the company is said to be evaluating the technology closely.

Neurophos enters a crowded arena anchored by Nvidia, whose GPUs power most of today’s training and inference. Several photonics players have shifted toward optical interconnects rather than full compute, including Lightmatter, while others such as Ayar Labs focus on optical I/O. Meanwhile, hyperscalers continue to push custom silicon for inference efficiency (e.g., TPU and Inferentia). To stand out, Neurophos must translate lab metrics into repeatable system-level gains on mainstream workloads.

Roadmap and What to Watch as Neurophos Scales Up

The fresh capital will fund an integrated photonic compute system spanning data center–ready OPU modules, a full software stack, and early-access developer hardware. The company is expanding in Austin and opening a San Francisco engineering site, signaling a push to recruit compiler, systems, and photonics talent.

Key milestones to monitor include developer toolchains that slot into existing workflows (think ONNX and PyTorch bridges), MLPerf Inference results, memory and I/O integration, and cost per deployed tera-op. Equally important will be evidence that the analog–digital boundaries do not erase the promised efficiency gains when scaled to model sizes used in production.

If Neurophos can deliver even a fraction of the projected performance and power advantages at volume, photonic compute could move from research curiosity to a practical accelerator tier. The metamaterials journey from cloaking experiments to AI chips underscores a broader trend in computing: the next leap may come not from a smaller transistor, but from rethinking the physics that do the math.

Gregory Zuckerman
ByGregory Zuckerman
Gregory Zuckerman is a veteran investigative journalist and financial writer with decades of experience covering global markets, investment strategies, and the business personalities shaping them. His writing blends deep reporting with narrative storytelling to uncover the hidden forces behind financial trends and innovations. Over the years, Gregory’s work has earned industry recognition for bringing clarity to complex financial topics, and he continues to focus on long-form journalism that explores hedge funds, private equity, and high-stakes investing.
Latest News
Trump Administration Weakens Clean Air Act Enforcement
Black Hat Removes Hacker Linked To Epstein
Automattic Sought Royalty Fees From 10 Rivals, WP Engine Says
Logitech MX Vertical drops to $69.99 at Amazon
Apple Rolls Out Transfer To Android In iOS 26.3
Musk Outlines xAI Shakeup And Unveils Macrohard
Huawei Watch Adds Diabetes Risk Screening
Next Android Canary Build Nears Release
Pixel 10a Expected To Outlast Pixel 10 In Battery Life
Apple Releases iOS 26.3 With New Privacy Tools
PlayStation State Of Play Streams Live Worldwide
T-Mobile Rebuked Over In-Flight Wi-Fi Ad Claims
FindArticles
  • Contact Us
  • About Us
  • Write For Us
  • Privacy Policy
  • Terms of Service
  • Corrections Policy
  • Diversity & Inclusion Statement
  • Diversity in Our Team
  • Editorial Guidelines
  • Feedback & Editorial Contact Policy
FindArticles © 2025. All Rights Reserved.