Work-Bench Snapshot: Augmenting Streaming and Batch Processing Workflows
The Work-Bench Snapshot Series explores the top people, blogs, videos, and more, shaping the enterprise on a particular topic we’re looking at from an investment standpoint.
This post was originally published in More Intelligent, my monthly newsletter that shares thematic research and the technology trends I’m digging into. I’m here to give you an investor’s view on the market and what’s next in enterprise software. Subscribe here!
Last week, I hosted a Data Center Roundtable with Zac Smith, co-founder of Packet, a bare-metal server provider that was acquired in 2020 by Equinix, the world’s largest global data center provider.
Recently, Zac and I have been exploring areas where venture scale opportunities could emerge in next-gen networking and infrastructure. So, we brought together a group of founders, data center developers, energy brokers, server providers, and Fortune 500 customers to discuss ideas and trends shaping the industry.
Below are themes from the discussion and investment areas where we’re actively exploring at Work-Bench. If you’re building in these areas and want to chat, please feel free to reach out!
Data centers have recently come in-vogue as AI research labs and hyperscalers look to invest in new-site construction and broader AI infrastructure to harness the power of LLMs and GPUs.
According to the WSJ, total spend for the global infrastructure market—including data centers, networking, and other hardware that supports the use of AI applications is expected to reach $423B by 2029, growing at a CAGR of ~44% over the next six years. On top of that, Citi forecasted that AI workloads will contribute to over 50% of data center IT load by 2030.
Ironically, despite this newfound demand, insiders to the data center world know that the industry is actually built around consistent, steady returns, which is why large private equity firms like Blackstone have spent the last 20+ years acquiring, investing in, and inking joint ventures across the data center landscape. As a result, the risks associated with scaled innovation have caused it to take a backseat to consistency.
Even with macro tailwinds propelling data centers into the limelight, many are skeptical that big-tech’s investment in AI infrastructure will pay off. Hyperscalers have spent an incremental $91B in capex along with legacy enterprises like Oracle are expecting to grow their data center footprint from 162 to 1000+ over the coming years. Still, there’s the ROI question. How sustainable is this over-allocation of compute and infrastructure?
This exploding demand for data center construction is causing a shortage of parts, property, equipment, and power - all integral components of data center makeup. The key hurdle to overcome as the AI supercycle ramps up is gaining access to efficient power and cooling solutions.
The AI boom of 2023 raised concerns over the amount of energy LLMs require when training on massive datasets and performing computations, because they operate on power hungry GPUs. By 2030, data centers are projected to consume a staggering 10% of the global power supply and consume more electricity than the entire nation of Brazil.
Given our nation’s energy constraints, the specific requirements for the new GPU-driven AI data centers are pushing builders to look for places where they can get lots of reliable electricity. Savvy operators are seeking alternatives to traditional natural gas, such as reliable renewables across hydrogen and nuclear energy.
Some examples:
In addition to harnessing energy, the lead time to get custom cooling systems is 5x longer than a few years ago and delivery times for backup generators have gone from as little as one month to as long as two years. This supply and demand imbalance has helped propel the stock of publicly traded companies like Vertiv, but could prove to be an opportunity for startups seeking to capitalize on the widespread shift.
Current data center construction is largely built to modern-day standards with general-purpose technology that can take multiple years to construct. Moving forward, we need to build interoperable data centers with modern software and hardware, so we can service new use cases, unique to each enterprise, and quickly update data centers with the best-in-class technology.
These two examples point to a future where enterprises will require bespoke data center systems to power their own unique use cases:
To make this a reality, we’ll need real collaboration across utility companies, real estate developers, and the OEMs that help procure parts and construct the data centers. Historically, these entities have acted as bottlenecks to each other, but need the other to get the job done. We could see a future where each of these players come together under one roof to create a new generation of infrastructure, or maybe even a world where hyperscalers and research labs bypass them altogether.
Lastly, data centers are an unregulated and siloed industry—there’s no public forum or central community to talk about the industry’s problems or spur innovation. In software, we have organizations like the Linux Foundation and the Apache Open Source Foundation that help developers manage and scale open technology projects and ecosystems.
If we are to move forward with new technology and thinking about how we can power and build data centers, more collaboration is certainly top of mind - who will be the OEMs that emerge ready to build next generation infrastructure?
Managing a data center is no easy task and obviously requires a strong understanding of hardware, infrastructure, networking, and software.
However, modern software development suggests a shift — future DevOps teams likely won’t look to manage their own infrastructure through the likes of AWS, GCP, or Azure.
Hyperscale clouds require tremendous infrastructure knowledge that are getting abstracted away as developers seek to work with vertically integrated cloud vendors to solve issues like latency, inflexibility in scaling resources, and a lack of vertical-specific tools, all of which can impact developer velocity.
In fact, different workloads will require a different vertically integrated stack to serve it. Today, engineers are getting more and more accustomed to serverless offerings from the likes of Cloudflare Workers and more, all which help abstract away the underlying knowledge of these platforms. As a result, there will be a new class of infrastructure providers that can capture the opportunity.
Right now, it feels like Netlify, Vercel, and Fly have the best chance of becoming the next true cloud platforms. These companies allow developers to focus on shipping code without having to think about the underlying infrastructure and provide greater efficiencies across a variety of customers. By bringing cloud-experiences to where developers work is unearthing tremendous time and cost savings, something that traditional cloud platforms overcomplicate and make challenging for developers.
Similarly, a new breed of AI cloud platforms has emerged, offering optimized compute resources, enhanced scalability, and developer-centric environments. Companies like Runpod and Together are making strides in building the AI cloud, where hardware, network, and software come together seamlessly to support next-gen applications. Their globally distributed GPU cloud platforms help engineers easily deploy AI when and where they want it. These specialized platforms are designed to handle the unique demands of AI workloads, from the high computational requirements of model training to the need for rapid scaling and efficient resource allocation as opposed to traditional tools like AWS Sagemaker, which is costly and inflexible.
Finally, to vertically integrate and increase the margin profile for the new cloud platforms, new points of presence (i.e., PoPs) and data center infrastructure will need to be built. One final question is how and where should these PoPs be built? Right now, Virginia is home to almost half of all domestic data centers and 35% of the worldwide population. As we look to the future, we’re seeing a brave new world for connectivity — companies like Axiom Space are putting data centers in outer-space to provide in-space cloud services without the need to connect back to terrestrial cloud infrastructure while Armada is bringing connectivity to remote environments.
We’re exploring several areas in the data center and next-gen networking space where there could be venture-scale outcomes across the stack.
We believe the best founder profiles for next-gen networking will be operators who truly understand the ins-and-outs of the software, networking, and hardware, trilogy. If that sounds like you and you’re exploring these areas, reach out – I’d love to meet! I’d love to meet you!
It’s clear, the world of data centers and cloud computing is rapidly evolving. As startups and established players alike navigate these shifts, the future of data centers holds tons of potential. The supply and demand imbalance for parts, property, and equipment, making it a particularly exciting time to be an OEM, cloud provider, or picks and shovels startup seeking to play a game of David and Goliath.