
Astera Labs stakes claim in AI data-center interconnect after 115% revenue surge
Astera Labs reported $852.5M in 2025 revenue, up 115% year over year, marking a rapid commercial ramp in AI interconnect gear. Its combination of PCIe 6 fabric switching hardware and the COSMOS management stack is designed to reduce latency and increase bandwidth between GPUs, SSDs, and NICs inside dense racks.
The product set is explicitly aimed at tightening intra-rack and cross-rack data paths to support larger model training and inference. Astera positions software-defined fabrics to bridge generational hardware and to route around failing links, which supports resilient, high-throughput topologies for modern AI workloads. Management guidance indicates continued growth: the company targeted roughly $290M in near-term top-line, an increase the company described as substantial versus the prior year.
External forecasts back the demand thesis. Market research firms place the AI interconnect and high-performance switching opportunity in the multi‑billion-dollar range over the next decade, with divergent estimates reflecting scope differences between switching silicon and broader fabric ecosystems.
- 2025 Revenue: $852.5M (+115% YoY)
- Near-term Top-line Target: ~$290M (+83% YoY guidance)
- Analyst 2026 Growth Expectation: ~29% revenue growth
- Market Size Estimates: Mordor Intelligence ~15% CAGR to 2032 (> $40B); Dell'Oro Group estimate nearer $100B
Short-term market reactions have been volatile, reflecting the high beta typical of suppliers tightly linked to AI capex cycles. That price action does not negate the structural need for lower-latency, higher-bandwidth fabrics as model sizes and GPU clusters grow. Competitors include large integrated silicon vendors and network incumbents, but Astera’s emphasis on PCIe 6 fabrics and software-defined controls narrows a technical differentiation window.
For investors and technical buyers, the critical variables are adoption velocity at hyperscalers, interoperability with legacy racks, and the company’s ability to translate product wins into sustained margin expansion. Execution risks include supply-chain constraints, competitive pricing pressure, and the timing of large enterprise rollouts. If Astera sustains its deployment cadence, it could become a core supplier for scale-up AI architectures over the next several years.
Read Our Expert Analysis
Create an account or login for free to unlock our expert analysis and key takeaways for this development.
By continuing, you agree to receive marketing communications and our weekly newsletter. You can opt-out at any time.
Recommended for you

Nebius boosts GPU and data‑center spending to lock in AI capacity
Nebius sharply increased quarterly capital spending to buy AI processors and expand its global data‑center footprint, pushing secured electrical capacity above 2 GW and raising its year‑end target to more than 3 GW. The build‑out — including a planned 240 MW, GPU‑dense campus in Béthune, France — widens near‑term losses but is aimed at underpinning a multibillion‑dollar annualized revenue run‑rate by the end of 2026.
Arista’s move toward AMD accelerators nudges Nvidia lower and reshapes data-center dynamics
Arista said roughly one-fifth to one-quarter of recent deployments are built around AMD accelerators, prompting a modest market reaction that nudged Nvidia shares down and AMD shares up. The disclosure is an early, measurable sign of buyer diversification in AI infrastructure that will play out over procurement cycles, supply constraints and software-stack alignment.

