Table of Contents
- Executive Summary
- Lazard VGB AI Infra 40 List
- Report Scope and Methodology
- Key Market Trends
- Segmentation of AI Infrastructure
- Appendix I: Company Profiles
- Appendix II: References
This document has been prepared by Lazard & Co., Limited ("Lazard") solely for general information purposes and is based on publicly available information which has not been independently verified by Lazard. The information contained herein is preliminary and should not be relied upon for any purpose. No liability whatsoever is accepted, and neither Lazard nor any member of the Lazard Group (being Lazard Ltd and its direct and indirect subsidiary and associated undertakings) nor any of their respective directors, partners, officers, employees, representatives or other agents is, or will be, making any warranty, representation or undertaking (expressed or implied) concerning the accuracy or truthfulness of any of the information, ideas, forecasts, projections or of any of the views or opinions contained in this document or any other written or oral statement provided in connection herewith or for any errors, omissions or misstatements contained herein or for any reliance that any party may seek to place upon any such information. Nothing contained in this document constitutes, or should be relied upon as, (i) the giving of financial, investment or other advice by, or the issuance of research by, Lazard, or (ii) a promise or representation as to any matter whether as to the past or the future. Lazard undertakes no obligation to provide the recipient with access to any additional information or to update or correct any information contained herein. Interested parties should conduct their own investigation and analysis of the matters and companies referenced herein. Nothing contained in this document constitutes, or should be deemed to constitute, an offer or solicitation for the purchase or sale of any security. You undertake to keep this document confidential and to not distribute it to any third party, or excerpt from or reproduce this document (in whole or in part), without the prior written consent of Lazard. Lazard, which is a regulated financial adviser, only acts for those entities whom it has identified as its client in a signed engagement letter and no-one else and will not be responsible to anyone other than such client for providing the protections afforded to clients of Lazard nor for providing advice. Recipients are recommended to seek their own financial and other advice and should entirely rely solely on their own judgment, review and analysis of this document. Lazard or other members of the Lazard Group (i) may have acted in the past, or act currently or in the future as adviser to some of the companies referenced herein, (ii) may receive fees in connection with any such advisory engagements, (iii) may at any time be in contact with such companies in order to solicit them to enter into advisory engagements, and/or (iv) may from time to time have made, and may in the future make, investments in such companies. By accepting this document, recipients agree to be bound by the terms and conditions set out above.
Section One – Executive Summary
We are delighted to introduce the Lazard VGB AI Infra 40 list which showcases a selection of some of the most exciting growth-stage AI infrastructure companies we have identified through hundreds of discussions with CEOs and investors.
4 key market trends emerged through our research and company interviews:
Rethinking the Stack for the AI Age
The optimal architectures of hardware, software and associated tooling appear to be shifting significantly for the AI Era and technical efficiency. Architectures will likely become increasingly driven by AI use-case (e.g., specialized for edge compute) and flexible (agnostic to both underlying chips and foundational models).
Focus on AI Sustainability
Rapidly increasing energy requirements of AI training and inference is leading to a focus on architectural energy efficiency and is increasing the need for optimized energy and AI infrastructure integration.
Looming Data Constraints
Limited supply of high-quality language data is leading to opportunities in alternative and synthetic data sources.
The Dawn of Autonomous AI Agents
Co-pilot agents are just the start of harnessing the power of LLMs for agents. We see emerging opportunities in multi-agent and autonomous agent models as a possible key theme over the next 12 months.
This report looks at AI infrastructure as a space for potential investments at the growth stage. McKinsey1 estimates that while up to US$25.6 trillion p.a. in economic value could be added to the global economy by Generative Artificial Intelligence (GenAI) and its applications, the build-out of essential AI infrastructure is fundamental to realizing this potential: “infrastructure is destiny”, as one OpenAI executive recently put it2.
The current phase of AI infrastructure growth is however being hampered by a series of stark shortages: of hardware (GPUs and other AI chips), of sustainable energy sources, of high-quality training data and of talent. We believe that technologists from many companies including our VGB AI Infra 40 list are aiming to respond with innovative rethinking of hardware, software, data and tooling stacks, increasingly led by AI use cases, including the growing need for edge compute.
Section Two – Lazard VGB AI Infra 40 List
We are delighted to introduce the VGB AI Infra 40 list of selected companies arranged alphabetically by segment. Detailed profiles of the VGB AI Infra 40 companies are available in the Appendix.
Source: PitchBook Data, Inc.; Public Sources; Lazard VGB Insights
Section Three – Report Scope & Methodology
Since publishing our work on AI commercialization and go-to-market strategies in mid-2023, the Lazard VGB Insights team has spoken to 100+ CEOs of private AI companies, across the stack.
This report focuses on our selected AI infrastructure growth-stage companies and a complementary report on enterprise-ready Horizontal and Vertical AI Applications will follow at a later date.
Our screening methodology can be summarized as follows:
- We utilized extensive public information, industry reports and databases such as PitchBook to screen 2,000+ companies in our core markets of North America and Europe which we then identified for detailed research and interview.
- We only included companies categorized as “AI-centric”, meaning AI functionality was assessed to be core to the company’s native architecture or product offering. We excluded companies employing single-use features or enhancements.
- Companies considered by us to be incumbent or dominant players were not included, as well as foundational model builders, such as Anthropic or OpenAI, which we see as a maturing market segment.
- We limited our selection to companies with known valuations under US$1.5 billion.
Section Four – Key Market Trends
Introduction
The immense potential of the AI Era has been catalyzed at an astonishing pace in the 18 months since the launch of ChatGPT. Yet it can be challenging for observers to read signal versus noise amidst the use of both superlatives and skepticism. Indeed, a recent survey3 indicated that experts are split on whether AI-related public stocks are in a bubble: 40% said “Yes”, 45% said “No”.
It could be that both are correct. The investor Bill Janeway (who recently delivered a keynote at our VGB T500 Conference in London) argues that not all bubbles are alike, and that in “productive bubbles”, speculation attaches itself to transformational general-purpose technology but with the real potential to create new economies4.
However, the potential opportunity is too large to ignore. Generative AI has catalyzed a step-change and represents a reported5 US$1.3 trillion market opportunity by 2032, potentially quadrupling to as much as 12% of total global technology spend. Including the impact of new use cases and productivity gains, McKinsey estimates GenAI could add upwards of US$25.6 trillion of economic value per annum to the global economy6
- While there are still notable challenges to adoption, we are seeing significant enterprise-pull for use cases which is “mostly top down, coming from CEO/C-suite or a steering committee”, according to 71% of recently surveyed executives.8
- More than 80% of enterprises are expected to have used GenAI by 2026, up from less than 5% in 2023.9
- Almost all the companies in our AI Infra 40 list have paying marquee enterprise customers, typically on multi-year recurring contracts and deploying in production. We have, however, included a few pre-revenue companies by exception.
As outlined in our previous report on AI commercialization, we note a continued trend for VC-backed AI companies to aim to reach large audiences through monetized multi-year strategic partnerships, often backed by equity investment.
- Recent examples of such partnerships since our previous report include those between Amazon/ Anthropic, Snowflake/Mistral, and Google/HuggingFace.10
- We note, however, that the Big Tech partnership approach may have limited runway due to antitrust concerns. The US Justice Department and the FTC have reportedly agreed an approach to potential investigations into Nvidia, OpenAI and Microsoft (including the recent Inflection AI transaction).11
While the US received the vast majority of VC investment with 70% of the total in 202312, and 3x more than Europe13, we note emerging vibrant ecosystems in Canada (Toronto and Waterloo), UK and France.
- Although outside the detailed scope of this report, we note that moves towards “Sovereign AI”14 by some countries in the interests of cultural and linguistic preservation, economic growth, talent development, and cybersecurity may materially impact the funding environment in coming years.
- For more on the Geopolitics of AI, see this October 2023 report from Lazard Geopolitical Advisory.15
The energy and AI infrastructures to satisfy escalating potential demand is being built out today for the AI Era of tomorrow. While Nvidia has captured a dominant market position in the making of specialized chips for running generative AI models in the first phase of the AI Era, we believe that the broader AI infrastructure sector, or “picks and shovels”, could represent a significant investment opportunity, estimated to be valued at US$309 billion by 2031.16 The second phase of AI evolution will likely involve a broad range of companies building specialized AI-related infrastructure, across software and hardware, AIOps (automating IT systems using ML and big data), MLOps (standardizing the process of deploying ML systems), data infrastructure and more.
1. Rethinking the Stack for the AI Age
The AI infrastructure stack will be significantly different from historic datacenters, cloud, and software infrastructures due to the many unique characteristics and demands of AI workloads. Training currently accounts for the majority of computational requirements, but as market demand scales, efficient real-time inference and low latency may become equally critical.
While the transition from CPUs to AI chips including GPUs, ASICs and TPUs has significantly improved the efficiency of AI workloads, there is room for further optimization and innovation as well as rethinking hardware/software integration needs.
- The majority of GPUs are underutilized during peak times. Improving effective GPU deployment efficiency is set to become a key issue in 2024 through 2025.17
- A “staggering 74% of companies are dissatisfied with their current job scheduling tools and face resource allocation constraints regularly, while limited on-demand and self-serve access to GPU compute inhibits productivity”.18
Architecture innovations such as interconnect technologies and High Bandwidth Memory (HBM) have become key to the AI stack in order to optimize GPU usage:
- Ayar Labs’ optical I/O solution seeks to address data movement bottlenecks in AI systems, to result in higher bandwidth and lower latency with greater power efficiency. “We’re on the cusp of a new era in high performance computing as optical I/O becomes a ‘must have’ building block for meeting the exponentially growing, data-intensive demands of emerging technologies like generative AI”.19
- Meanwhile Celestial AI’s photonic fabric interconnect addresses the “Memory Wall”20 by aiming to enable bandwidth delivery directly to the point of compute within the chip.
- Cornelis Networks delivers end-to-end high-performance interconnect solutions with a proprietary scale-out architecture, incorporating telemetry-based adaptive routing, congestion control with low latency and enhanced support for messaging, memory models and AI optimization in large-scale hyperscaler, cloud AI and on-prem AI/HPC environments.
At the hardware/silicon level, there could also be significant opportunity for new and specialized players given market demand potential, the opportunity to diversify supply bases given the dominance of Nvidia’s estimated 80% market share21 as well as potential future supply chain constraints.22
- SiPearl seeks to provide a high-performance, low power microprocessor for high performance computing (HPC) and AI workloads which integrates Samsung’s HBM solution, to improve processing speed with reduced thermal resistance, rather than by simply adding more GPUs.23
Edge acceleration will require a range of AI-ready and energy efficient solution sets, whether for automotive, defense or enterprise for all of whom we are already seeing noticeable adoption.
- Edge computing market size was estimated at US$16 billion in 2023 and is expected to grow at 37% CAGR to US$156 billion by 2030.24
- Meanwhile, Gartner predicts that by the end of 2026, 100% of enterprise PC purchases will be an AI PC, with an integrated Neural Processing Unit (NPU).25
- SiMa’s embedded edge machine learning system-on-chip (MLSoC) aims to allow customers to run entire applications on a chip, while Axelera’s AI acceleration platform seeks to enable inference processing with YOLO (You Only Look Once, convolutional neural networks for real-time object detection) for edge AI computer vision applications.
Most of the emergent semiconductor/hardware category companies we spoke to are working on innovative technical approaches to workload management, recognizing that the computational requirements of data preparation, training and inference vary significantly. Integrated hardware / software solutions and use of computationally appropriate models will also vary significantly by use case.
- The use of industry or function specific GenAI models used by enterprise is expected to increase from approximately 1% in 2023 to 50% by 2027.26
2. Focus on AI Sustainability
The rapid growth in generative AI has sparked concerns about intensity of energy use. The International Energy Agency estimates that by 2026, datacenters could globally consume more than 1,000 terawatt-hours of electricity, more than double that of 2022 and roughly equal to Japan’s total energy usage.27 As energy consumption shifts from training towards inference as modern models are deployed at scale, energy efficiency throughout the model life cycle is increasingly under scrutiny.28 We identify two potential emerging trends resulting: Energy and AI infrastructure integration, and innovation in energy-efficient hardware and models.
Interdependence of energy supply chains and datacenter infrastructure is seeing “The Magnificent 7” (Mag7)29 integrating with energy infrastructure, increasingly co-locating datacenters with sustainable energy sources. Yet it is not clear that sufficient clean energy resources can meet demand.
- Microsoft and OpenAI are reportedly planning a massive US$100 billion, 5GW “Stargate” AI datacenter, potentially powered by alternative energy sources including nuclear, at an unspecified location.30 While Microsoft signed an agreement with nuclear power producer Constellation Energy in 2023,31 the analysis reports that this would not provide the scale of power required, and few existing global nuclear facilities could.32
- Microsoft also recently announced it is backing an estimated US$10 billion in renewable energy projects in a partnership with Brookfield AM.33
- Meanwhile in March 2024, Amazon acquired Talen Energy’s datacenter campus at a nuclear power plant in Pennsylvania for US$650 million.34
The capital intensity of the AI Era is reflected in the Mag7’s burgeoning Capex and R&D spend of US$374 billion in 2023. Amazon, Meta, Alphabet and Microsoft alone have already pledged to spend a combined US$200 billion on Capex in 2024, mostly on AI infrastructure, up 35% on 2023 figures. This represents as much as 21% of the total capex of the entire S&P500, up from 4% a decade ago.35
- Some analysts are beginning to question however how and when this “AI Arms Race” will generate commensurate returns.36 "Given the billions of dollars that Big Tech companies have been pouring into the AI boom, investors are cautious that this may ultimately result in infrastructure overbuild minus the promised future profits," according to one.37
- There is a question about the sustainability of such capital and energy intensive AI datacenters and potentially a complementary need for innovation solutions to address the issue.
While many hyperscalers have long claimed to be 100% renewable or carbon neutral, there are broader issues surrounding the displacement of renewable energy supplies from alternative uses such as EVs, which could lead to increasing scrutiny of the need for intensive AI energy use.38 We believe that innovations in chip technology, model architecture, orchestration, scheduling and energy efficient acceleration will likely play an important role and may present investment opportunities in the context of sustainability.
- Sustainability of energy resourcing is expected by analysts to result in 30% of GenAI implementations using energy-conserving computational methods by 2028.39
- We see the increasing use of more energy efficient Small Language Models such as Microsoft Phi3 as a trend for 2024, a trend which may present positive opportunities for AI developer platforms such as Anaconda, which supports the creation and flexible deployment of efficient models and a high-performance version of Python that maximizes efficiency of AI workloads, as well as Weights & Biases.40
- Nvidia’s acquisition of Run.ai could be indicative of the need for dynamic resource allocation and workload orchestration to minimize idle resource and optimize energy efficiency. Similarly, Anyscale aims to enable dynamic scaling, advanced scheduling and resource management tools as well as reduced data movement to improve energy efficiency across AI workloads. Octo AI’s strategy is to optimize models for energy efficient deployment on edge devices through advanced model compression and optimization of specific target edge devices.
- Cerebras’ wafer-scale chips and software stack seek to enable faster, more efficient processing of large AI models, reducing energy consumption. It is also designed to be air-cooled, addressing certain sustainability issues surrounding water-cooling.41 Lightmatter’s photonic AI accelerators and Untether AI’s approach to at-memory computation are also designed to be air-cooled.
3. Looming Data Constraints, Synthetic and Alternative Data
According to research from EpochAI, models are using increasing quantities of training datasets, growing at a rate of 2.8x per annum. There are limited resources of public human-generated text, with researchers estimating that high-quality data stocks for model training may become fully exploited by 2026-2032.42 This timing is uncertain as on the one hand, there may be advances in data efficiency, yet overtraining (using more data over reduced parameters with a view to optimizing compute for inference) could also exploit available public language data stocks even sooner.
Source: EpochAI, Lazard VGB Insights
There are a number of potential mitigating strategies and technologies to this bottleneck, including: speech recognition systems (e.g., OpenAI Whisper) enabling use of abundant audio data for training, Optical Character Recognition (OCR) enabling visual transformer models for paper academic documents (e.g., Meta’s Nougat), as well as the generation and use of synthetic data.
- By 2025, it is estimated that 75% of businesses will use generative AI to create synthetic customer data, up from less than 5% in 2023.43
We believe there could be a possibility of medium-term constraints on ML models transitioning from compute to data and synthetic data providers including those such as Gretel and Mostly.ai, as well as multi-modal data ingestion like Unstructured.io may present possible opportunities.
Our view is that there might also be a continuation of the trend for alternative data strategic partnerships such as those recently announced, including between both Google and OpenAI and Reddit and Stack Overflow.44
The scramble for training and fine-tuning data has begun to extend to proprietary data sources. OpenAI announced a data licensing deal with News Corp (Wall Street Journal, New York Post, Barron’s, The Times UK) in May 2024, “worth US$250 million in the next five years”, alongside existing agreements with The Associated Press, Financial Times, and Axel Springer (Politico).45
While there are further unused private data sources that might yet be exploited for training or fine-tuning, these are currently being hampered by privacy concerns and threat vectors. Accordingly, there may be opportunities for model and data security solutions such as Hidden Layer and Protect AI.46
4. The Dawn of Autonomous AI Agents
While GenAI has delivered significant opportunity over the last 18 months, we believe that the trifecta of transformers, internet-scale data and human feedback could be just the start of what is possible. While autonomous agents in the form of coding companions such as co-pilots have already started to emerge, multi-agent systems (MAS) for cooperative GenAI applications may also potentially transform the opportunity space.
- AI MAS are complex systems of multiple interacting intelligent agents, which learn from and collaborate with each other. Modern AI (GPT, BERT)47 may enable new opportunities for MAS, which have previously been constrained by complexity and coordination limitations.
- The market for autonomous agents is forecast to grow from US$5 billion in 2023 to US$29 billion by 2028 at a CAGR of 43%.48
- Established solutions such as Microsoft 365 Copilot and Github’s Copilot are already widely used by developers, albeit with some accuracy and security risks.49 Microsoft has further developed AutoGen as a framework to enable generation of multi-agent systems based on high-level specifications.
While still a relatively immature space for growth-stage investors, we believe that the use of multi-agent systems in AI infrastructure and the embedding of autonomous agents in applications could develop and become one of the important trends to watch over the next 12 months. Cognition Labs, the creator of autonomous coding agent Devin emerged from stealth only 6 months after formation raising US$175 million in April 2024 led by Founders Fund at a reported US$2 billion valuation.50
- The Open-Source project LangGraph seeks to facilitate efficient and complex information flows between multi-agent systems enabling them to collaborate and make decisions based on shared knowledge. The developer of LangGraph, LangChain recently (February 2024) announced a US$25 million Series A from Sequoia, Benchmark, and others.51
- Platforms for building and deploying AI applications using complex multi-agent approaches such as Abacus.ai could enable the creation of agents by chaining together user code, data transforms, ML models and LLM prompts which can access both LLMs and abacus APIs in one place.
- Agent monitoring-as-a-service will likely be necessary to develop multi-agent “artificial immune systems” with observability tooling including Fiddler and AccelData potentially seeing positive opportunities from the agent and trend towards MAS.
Section Five – Segmentation of AI Infrastructure
There are many ways to segment the AI infrastructure market, and our approach is designed to provide an overall market view while recognizing the distinct features of each sub-segment.
- This market map provides an overview of some companies we have identified as incumbent players and some selected emergent players, with our selected VGB AI Infra 40 companies highlighted in dotted boxes.
- Companies have been categorized by us based on what we have identified as their specific AI use case, using category abstractions. Certain companies, however, do not fit neatly into any single box but may have capabilities across multiple categories.
Source: Lazard VGB Insights
The following section provides our outline of what we perceive to be the features of each AI Infrastructure subsector as well as identifying some incumbents and some selected growth stage companies identified during our research.
Hardware / Silicon
Hyperscalers and Compute
Data Infrastructure
Source: PitchBook Data, Inc.; Public Sources; Lazard VGB Insights
Model Serving and Inference
Foundational Models
MLOps
Source: PitchBook Data, Inc.; Public Sources; Lazard VGB Insights
Market Map - Company Maturity
We looked at the maturity stages of companies across our AI infrastructure market map and from our analysis have observed the following:
- Significant capital seems to be being deployed to certain disruptors in the Hardware/Silicon and MLOps sector, possibly indicating that these segments are beginning to mature.
- The market landscape appears highly fragmented across all segments, in our view with no clearcut winners in each category, likely encouraging enterprises to adopt one of three strategies: embrace end-to-end solutions, construct their own bespoke systems, or select the "best of breed" companies.
- We might see category leaders continuing to attract through-the-cycle funding, and some mid-tier players may potentially be:
i. Forced into defensive mergers,
ii. Acquired by industry leaders, or
iii. Acquired by strategic investors
- Companies including Cerebras, Scale and Grafana that have raised >US$500 million, may possibly be looking towards the public markets within the next 12-18 months depending on IPO market sentiment.
Source: PitchBook Data, Inc.; Lazard VGB Insights
Growth Trajectories across AI Infrastructure Segments
Relying on the aforementioned market segmentation map and companies which we have selected, we used employee growth figures since 2021 relative to total funds raised as a hypothetical proxy for overall growth. Based entirely on that measure of growth and applied to our selected landscape of companies, our analysis revealed that:
- Many larger companies are continuing to experience rapid expansion (>50%+) even as they continue to scale.
- A concentrated group of companies seem to be experiencing rapid growth, particularly at the relatively early stages of funding (highlighted in the grey section of the chart below).Companies including Cerebras, Scale and Grafana that have raised >US$500 million, may possibly be looking towards the public markets within the next 12-18 months depending on IPO market sentiment.
Source: PitchBook Data, Inc.; Lazard VGB Insights
VI. Lazard VGB AI Infra 40 Profiles
- Axelera AI is a developer of a hardware and software platform for AI, designed to deliver exceptional performance within a power envelope of just a few watts while maintaining the flexibility to support multiple networks
- The platform combines a custom dataflow architecture with multicore in-memory computing. This enables clients to minimize power consumption and deliver edge applications for a sustainable future, promoting both efficiency and environmental responsibility
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- The company develops electronic-photonic chipsets designed for applications demanding high bandwidth, low latency, and power-efficient short-reach interconnects
- Utilizing industry-standard, cost-effective silicon processing techniques, the company creates optical-based interconnect chipsets and lasers to replace traditional electrical-based input-output systems. This technology allows companies to manage large volumes of data more effectively by miniaturizing fiber optic transceivers
- Celestial AI is a developer of an innovative data center and AI computing platform that aims to cater to deep learning and machine learning applications
- The company's technology combines the benefits of photonics, mixed-signal ASICs, and packaging to provide a substantial enhancement in computing performance. This enables clients to offer AI acceleration hardware and software alternatives, fostering advanced solutions for diverse applications.
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Cornami is a developer of reconfigurable computational fabric technology that aims to treat processor cores as scalable resources in the same way as memory, storage, and transistors
- The company's technology delivers scalability from thousands of cores on a single chip to millions across a system for real-time computing at the lowest cost, power, and latency available. It is focused on delivering real-time fully homomorphic encryption at market prices. This enables developers, large enterprises, and edge computing to deliver high performance anywhere and on any device with the lowest power consumption and latency, enhancing efficiency and productivity
- Cornelis Networks is a developer of purpose-built high-performance fabrics, designed for leading scientific, commercial, and government organizations
- The company offers high-performance fabrics that expedite computing, data analytics, and artificial intelligence workloads. This enables customers to effectively concentrate the computational power of multiple processing devices on a single problem, thereby enhancing both the result and accuracy simultaneously
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- d-Matrix is a developer of a computing platform tailored for GenAI and LLMS by offering an innovative in-memory computing technique for data centers
- The platform concentrates on addressing the physics of memory-compute integration using mixed-signal and digital signal processing techniques. This enables clients to benefit from enhanced computing efficiency, fostering improved performance and productivity
- Enfabrica, a developer of high-performance, ultra-scalable infrastructure silicon and software, aims to address the critical interconnect bottlenecks in next-generation computing workloads
- The company's foundational fabric technologies and products are designed to enable groundbreaking system efficiencies, topologies, and performance across various sectors, including hyper-scale cloud, edge, enterprise, fifth-generation/sixth-generation, and automotive infrastructure.
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Lightelligence is an operator of an optical computing platform that aims to accelerate information processing
- The platform employs artificial intelligence and cutting-edge technology to transmit data via photons, enabling users to convey information with considerably lower latency and higher throughput compared to conventional electronic circuits
- Lightmatter is leading the revolution in networking for AI.
- The company invented a leading 3D-stacked photonics engine, Passage™, capable of connecting thousands to millions of processors at the speed of light in extreme-scale data centers for the most advanced AI and HPC workloads
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- NextSilicon develops advanced computing architecture technology focused on enhancing future computer processing methods
- The company specializes in chip design and software development, using innovative software algorithms to speed up compute-intensive applications. This technology provides high-performance architecture for supercomputers, offering a new approach to chip technology that supports organizations in achieving greater computational efficiency
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- SiMa.ai facilitates the widespread adoption of high-performance machine learning inference at extremely low power in embedded edge applications
- It offers push-button performance, effortless deployment, and scaling at the embedded edge. This enables businesses to support traditional computing with high-performance, energy-efficient, safe, and secure machine learning inference
- Untether AI is a developer of AI chips designed to pioneer new frontiers in artificial intelligence applications
- The company's chips integrate near-memory design with digital processing to facilitate neural net inference that minimizes the distance data must travel. This enables clients to enhance inference efficiency while consuming fewer resources and energy, and requiring less supporting infrastructure, thereby promoting sustainable and efficient operations
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Lambda is a developer of a cloud computing platform tailored for large-scale artificial intelligence training and inference.
- The company’s product portfolio spans from on-prem GPU hardware to hosted GPUs in the cloud and includes Lambda’s proprietary software which enables deep learning / AI teams to access the tools they need via a singular interface regardless of the location of the compute resources (on-prem or cloud
- The solution is ideal for tasks such as natural language processing and drug discovery, enabling organizations to accelerate their AI initiatives with ease and precision
- NexGen Cloud operates as an Infrastructure-as-a-Service (IaaS) provider with a mission to bridge the gap between Web 2.0 and Web 3.0. Specializing in decentralized blockchain storage services, the company caters to the AI/ML, Meta, and Omniverse industries
- By offering robust and scalable solutions, it empowers large-scale projects to overcome challenges related to cost, transparency, and centralization, facilitating a seamless transition to the next generation of web technologies
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Together AI is an operator of a technology services platform dedicated to offering a decentralized cloud for AI
- The platform focuses on constructing extensive, open models that are user-friendly and open-source. This enables researchers, developers, and companies to harness and enhance artificial intelligence through a seamless integration of data, models, and computation platforms
- Anyscale aims to streamline distributed computing. The company's software specializes in an open-source framework that simplifies the creation of complex, compute-intensive applications by easing the underlying hardware decisions
- This enables software developers of all skill levels to build applications capable of running at any scale, from a laptop to a data center, promoting versatility and efficiency
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Modular’s is an integrated, composable suite of products that simplifies customers’ AI infrastructure so they can develop, deploy, and innovate faster
- Modular provides an engine that tries to improve the inferencing performance of AI models on CPUs and GPUs while delivering on cost savings
- Modular’s other flagship product, Mojo, is a programming language that aims to combine the usability of Python with features like caching, adaptive compilation techniques, and metaprogramming
- OctoAI is a developer of a technology is designed to assist engineering teams in swiftly deploying machine learning models on any hardware, cloud provider, or edge device
- The company's technology offers a managed service that utilizes machine learning to automate machine language code generation and optimization in multi-cloud environments. This ensures that the models continue to operate at peak efficiency, providing businesses with secure deployments of deep learning models, thereby enhancing productivity and security
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Abacus.AI is an autonomous AI platform that aims to assist organizations in creating large-scale, real-time customizable deep learning systems
- The platform offers an end-to-end autonomous AI service that trains machine and deep learning models for common enterprise AI use cases such as churn prediction, time-series forecasting, and deep-learning-based personalization. It also allows for the creation of custom, specific models with a state-of-the-art toolset. This enables clients to integrate cutting-edge deep learning models into their business processes or customer experiences, fostering innovation and improved efficiency
- Acceldata is a data and analytics platform designed to simplify data operations
- The platform provides information integration and data streaming services, enabling clients to stream, collect, and process data, construct data clusters, and gain actionable insights from the data. It also allows for optimization of workflow operations and capitalization on opportunities identified through predictive analytics. This empowers enterprises to proactively manage performance, security, data quality, and workflow, fostering improved efficiency and decision-making
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Anaconda provides an enterprise grade platform for open-source software development with a focus on AI and data science capabilities. Anaconda’s platform enables AI developers, data scientists and IT teams with secure access to curated open-source software artifacts, with security, policy and governance control, capabilities to develop and deploy on-prem LLMs, and a high-performance version of Python that optimizes workload performance for both numerical and general code.
- The platform helps AI developers and data scientists ensure complete reproducibility and reliability of their projects and in-production workloads, and supports IT teams with reobust governance and control over how developers leverage open-source software artifacts.
- Domino Data Lab is a developer of an enterprise data science management platform designed to assist companies in building and deploying ideas through collaborative, reusable, and reproducible analysis
- The platform expedites research, accelerates model deployment, and fosters collaboration for code-first data science teams at scale. This enables data scientists to contribute to various fields, such as developing medicines, increasing crop productivity, adapting risk models to major economic shifts, constructing cars, and enhancing customer support
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Edge Impulse is an ML development platform designed to bring about positive societal change through machine learning.
- The platform streamlines the process of building, deploying, and scaling embedded ML applications. This enables developers to create intelligent devices by simplifying the collection of real sensor data, live signal processing from raw data to neural networks, testing, and deployment to any target device, fostering innovation and efficiency in creating smart solutions
- Featurebase is a data virtualization software is designed to secure access to large, fragmented, and geographically dispersed datasets
- The company's software aids in the mass parallelization of large, high cardinality, ad hoc queries for managing massive, unbounded datasets and takes advantage of the efficiency, performance, and simplicity of bitmaps as a foundation for powering AI with real-time information. This enables clients to perform time-based sharing to capture streaming data and carry out segmentation based on historical data or time ranges, ultimately enhancing data accessibility and analysis capabilities
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Fiddler is an enterprise AI platform designed to create AI services that are transparent, explainable, and understandable
- The platform utilizes the AI engine to provide statistical metrics, performance monitoring, and security services through a common language, centralized controls, and actionable insights. This enables businesses to analyze, manage, and deploy their machine learning models at scale, enhancing efficiency and decision-making capabilities
- Gretel is a data categorization and identification platform, is designed to automatically generate an anonymized version of a dataset
- The platform leverages machine learning to categorize data across various customer identifiers such as names and addresses. It features automatic data labeling, power testing, and synthetics. This enables developers to safely and swiftly experiment, collaborate, and build with customer data, promoting innovation and data privacy
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- HiddenLayer’s AISec Platform is an AI/ML Protection Suite that ensures the integrity of customers’ models throughout the MLOps pipeline
- By ensuring the security of pretrained models, detecting malicious injections, and monitoring algorithm inputs and outputs for potential threats, The AISec Platform delivers an automated and scalable defense tailored for ML
- This enables proactive responses to attacks without necessitating access to private data or models
- Human Signal specializes in the development of advanced data labeling software engineered to centralize and streamline training data management
- The platform enhances operational efficiency by integrating robust management and annotator functionalities, which facilitate and optimize collaborative data labeling processes, quality assurance, and analytical evaluations. This enables enterprises to expedite dataset annotation and achieve precise, high-performance machine learning and artificial intelligence models at scale, thereby maintaining a competitive edge in the industry
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Latent AI is an inference platform designed to support edge computing workloads
- The company offers a quantization optimizer for edge AI devices to automate the exploration of low-bit-precision training deploying efficient neural networks for on-device intelligence and inference. This enables software developers to feasibly access, deploy, and manage AI for the edge, promoting innovation and efficiency in edge computing applications
- Lightning AI is a multi-cloud machine learning systems designed to aid in building simple research demos
- The platform offers monitoring, training management, single command cloud training, experiment analysis, engineering automation, and automated artifact backups. This enables engineers, data scientists, and AI researchers to save time and train machine learning models on the cloud directly from their laptops, enhancing efficiency and convenience
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Mad Street Den is a cloud-based AI platform is designed to build models of generalizable intelligence and create actionable ways to contextualize AI on a large scale
- The company's platform offers artificial intelligence and computer vision modules to facilitate various features, including object recognition, gaze tracking, emotion-expression detection, head and facial gestures, as well as 3D facial reconstruction. This enables clients to build models of generalizable intelligence on a grand scale, which can be deployed through meaningful applications across various industries, enhancing efficiency and innovation
- Mostly AI is a pioneer in GPU-powered technology designed to simulate synthetic customer data at scale. This cutting-edge technology enables the generation of an unlimited number of realistic and representative synthetic customers, closely mirroring the patterns and behaviors of actual customers with unprecedented accuracy
- By leveraging this advanced simulation capability, businesses can unlock a wealth of opportunities from previously inaccessible data, driving faster innovation while mitigating risks and reducing costs. This transformative approach empowers organizations to harness the full potential of their data assets, opening new avenues for growth and efficiency
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Protect AI is a cybersecurity platform is designed to concentrate on machine learning workflows and pipelines
- The company's platform offers innovative security products and performs security scans using machine learning models and artificial intelligence systems to access curated resources, learn best practices in machine learning security, listen to podcasts with thought leaders, and connect with a thriving community. This enables enterprises to build a safer, AI-powered world, fostering enhanced security and innovation
- Rescale, a developer of a cloud-based infrastructure platform, is designed to streamline scientific and engineering simulations
- The platform offers infinite scalability, customization tools, and the ability to make adjustments optimized for specific workloads, ultimately reducing turnaround times. This enables businesses to transform their information technology into unified, agile environments, enhancing overall outcomes and efficiency
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- RelationalAI is a relational knowledge graph system designed to address complex business challenges
- The company concentrates on the rich interdependencies and structures inherent in every business, complementing the modern data stack to expedite the development of intelligent data applications. This enables clients to implement intelligent applications with semantic layers on a data-centric foundation, lowering the barrier to codifying and utilizing knowledge, and ultimately enhancing business efficiency and decision-making
- Snorkel is an AI-powered programmatic data labeling tool is designed for extracting information from text documents such as scientific articles and electronic health records
- The company's tool leverages theoretically grounded techniques to perform data augmentation and slicing data into different critical subsets, and then identifies subsets of the data. This enables users to quickly leverage structured data resources available in domains such as bioinformatics, enhancing efficiency and productivity in data processing and analysis
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Unravel leverages AI, ML and analytics to offer actionable recommendations and automation, enabling businesses to understand and optimize their data-driven applications
- Unravel’s purpose-built AI data observability and FinOps for Databricks, Snowflake, BigQuery and other modern data stacks provides granular visibility for cost allocation, metadata correlation for data reliability, and AI-powered insights for data performance management
- Unstructured.io is an open-source data transformation platform designed to simplify the preprocessing of natural language data for downstream machine learning services
- The platform utilizes open-source libraries and application programming interfaces to construct custom preprocessing pipelines for labeling, training, or production machine learning pipelines. This enables clients to convert simple data into language data, fostering innovation and improved efficiency in processing natural language information
Source: Company Websites; Funding Press Releases; PitchBook Data, Inc.; Lazard VGB Insights
- Weights & Biases is a dataset optimization tool, is dedicated to creating high-quality software tools for deep learning practitioners
- The company's tool offers performance visualization for machine learning and assists teams in tracking their models, visualizing model performance, and effortlessly automating the training and enhancement of models. This enables companies to transform deep learning research projects into deployed software, fostering innovation and efficiency
Disclaimer
This document has been prepared by Lazard Frères & Co. LLC ("Lazard") solely for general information purposes and is based on publicly available information which has not been independently verified by Lazard. The information contained herein is preliminary and should not be relied upon for any purpose. No liability whatsoever is accepted, and neither Lazard nor any member of the Lazard Group (being Lazard Ltd and its direct and indirect subsidiary and associated undertakings) nor any of their respective directors, partners, officers, employees, representatives or other agents is, or will be, making any warranty, representation or undertaking (expressed or implied) concerning the accuracy or truthfulness of any of the information, ideas, forecasts, projections or of any of the views or opinions contained in this document or any other written or oral statement provided in connection herewith or for any errors, omissions or misstatements contained herein or for any reliance that any party may seek to place upon any such information. Lazard undertakes no obligation to provide the recipient with access to any additional information or to update or correct any information contained herein. Interested parties should conduct their own investigation and analysis of the companies and information referenced herein. Lazard only acts for those entities whom it has identified as its client in a signed engagement letter and no-one else and will not be responsible to anyone other than such client for providing the protections afforded to clients of Lazard nor for providing advice. Nothing contained in this document constitutes, or should be relied upon as, (i) the giving of financial, investment or other advice or recommendations by, or the issuance of research by, Lazard, or (ii) a promise or representation as to any matter whether as to the past or the future. Recipients are recommended to seek their own financial and other advice and should entirely rely solely on their own judgment, review and analysis of this document. Lazard or other members of the Lazard Group (A) may have acted in the past, or act currently or in the future as adviser to some of the companies referenced herein, (B) may receive fees in connection with any such advisory engagements, (C) may at any time be in contact with such companies in order to solicit them to enter into advisory engagements and (D) may from time to time have made, and may in the future make, investments in such companies. Nothing contained in this document constitutes, or should be deemed to constitute, an offer or solicitation for the purchase or sale of any security. You undertake to keep this document confidential and to not distribute it to any third party, or excerpt from or reproduce this document (in whole or in part), without the prior written consent of Lazard.
VII. References
1. McKinsey Digital, “The Economic Potential of Generative AI: The Next Productivity Frontier”, June 2023
2. Axios, “OpenAI’s Chris Lehane says AI is Critical Infrastructure”, April 2024, as quoted in Axios, “Behind the Curtain: AI’s Ominous Scarcity Crisis”, May 2024
3. Bank of America “Global Fund Manager Survey”, March 2024, as quoted in Morningstar, “Is the World in an AI Bubble? Money Managers are Split?”, March 2024
4. Bill Janeway, “Productive Bubbles”, in Noema Magazine, July 2021
5. Bloomberg Intelligence, “2023 Generative AI Growth Report”, June 2023, as quoted in Bloomberg Press Release, “Generative AI to Become a $1.3 Trillion Market by 2032, Research Finds”, June 2023
6. McKinsey Digital, “The Economic Potential of Generative AI: The Next Productivity Frontier”, June 2023
7. Additional quotations in text boxes from: Mark Benioff, “Salesforce+ Salesforce AI Day” video, cited in Forbes, “Shifting the AI Narrative: From Doomsday Fears to Pragmatic Solutions”, March 2024; GMO, “The Great Paradox of the U.S. Market”, March 2024; Jensen Huang, as cited in Reuters, “Chip Giant Nvidia Nears Trillion-Dollar Status on AI Bet”, May 2023; Bessemer Venture Partners, “Roadmap: AI Infrastructure”, June 2024
8. Gartner CEO Survey November 2023, quoted in Harvard Business Review, “5 Forces That Will Drive the Adoption of GenAI”, December 2023
9. Gartner Press Release, “More than 80% of Enterprises Will Have Used Generative AI APIs or Deployed Generative AI-Enabled Applications by 2026”, October 2023
10. See Press Releases: “Amazon and Anthropic Deepen Their Shared Commitment to Advancing Generative AI”, March 2024; “Snowflake Partners with Mistral AI to Bring Industry-Leading Language Models to Enterprises Through Snowflake Cortex”, March 2024; “Hugging Face and Google Partner for Open AI Collaboration”, January 2024
11. Reuters, “US sets Stage for Antitrust Probes into Microsoft, OpenAI and Nvidia”, June 2024; Fortune, “Why Microsoft’s Surprise Deal with $4 Billion Startup Inflection is the Most Important Non-Acquisition in AI”, March 2024; Reuters, “Microsoft Pays Inflection $650 Million in Licensing Deal While Poaching Top Talents”, March 2024
12. Stanford Institute for Human-Centered Artificial Intelligence, “Artificial Intelligence Index 2024”, Chapter 4: Economy, Figure 4.3.8, p.35
13. Dealroom.co, “State of AI Investing”, May 2024
14. See Anupam Chander, Haochen Sun (eds), “Data Sovereignty: From the Digital Silk Road to the Return of the State”, Oxford University Press 2023, Chapter 5: Andrew Keane Woods “Digital Sovereignty + Artificial Intelligence”; World Economic Forum, “Sovereign AI: What it is, and 6 Strategic Pillars for achieving it”, April 2024
15. Lazard Geopolitical Advisory, “The Geopolitics of Artificial Intelligence”, October 2023
16. Allied Market Research, “AI Infrastructure Market 2023”, September 2023
17. AI Infrastructure Alliance, “The State of Infrastructure at Scale 2024”, March 2024
18. AI Infrastructure Alliance, March 2024, ibid.
19. Ayar Labs CEO Charlie Wuischpard, quoted in Press Release, “Ayar Labs Showcases 4 Tbps Optically-Enabled Intel FPGA at Supercomputing 2023”, November 2023
20. The “Memory Wall” refers to a mismatch between the slow growth of on-chip memory capabilities and the dramatic expansion of data requirements for advanced AI. See Optics and Photonics News, “Celestial AI Cultivates a Photonic Fabric Ecosystem”, April 2024
21. The Wall Street Journal, “Nvidia’s Business is Booming. Here’s What Could Slow It Down”, May 2024
22. Quote from Sequoia, “Generative AI’s Act Two”, Sonya Huang and Pat Grady, September 2023
23. eeNews Embedded, “SiPearl Partners with Samsung for built-in HBM in Rhea”, May 2024
24. Grand View Research, “Edge Computing Market Size and Trends”, March 2024
25. Gartner, “Gartner Forecasts Worldwide AI Chips Revenue to Grow 33% in 2024”, May 2024
26. Gartner, “3 Bold and Actionable Predictions for the Future of GenAI”, April 2024
27. International Energy Agency (IEA), “Electricity 2024”, January 2024: IEA, Paris https://www.iea.org/reports/electricity-2024, Licence: CC BY 4.0
28. Luccioni, Sasha, Yacine Jernite, and Emma Strubell. "Power Hungry Processing: Watts Driving the Cost of AI Deployment?" In The 2024 ACM Conference on Fairness, Accountability, and Transparency, pp. 85-99. 2024, arXiv:2311.16863 [cs.LG]
29. The “Magnificent Seven” includes Meta, Apple, Tesla, Nvidia, Amazon, Microsoft, and Alphabet
30. Data Center Dynamics, “Microsoft and OpenAI Consider $100 Billion, 5GW ‘Stargate’ AI Data Center”, March 2024
31. Data Center Dynamics, “Microsoft Signs 24/7 Nuclear Power Deal with Constellation for Boydton Data Center”, June 2023
32. Data Center Dynamics, “Is Microsoft and OpenAI’s 5GW Stargate Supercomputer Feasible?”, April 2024
33. Financial Times, “Microsoft to Power Data Centers with Big Brookfield Renewables Deal”, May 2024
34. Data Center Dynamics, “AWS Acquires Talen’s Nuclear Data Center Campus in Pennsylvania”, March 2024. See also Talen Energy, “Mar-24 Business Update Presentation”, available at https://talenenergy.investorroom.com/financials-presentations
35. Capital Group, “Tech Giants Ratchet Up Spending in AI Race”, May 2024
36. See, for example, The Economist, “Big Tech’s Capex Splurge May be Irrationally Exuberant”, May 2024
37. Nicole Tanenbaum, Chequers Financial Management, quoted in “Big Tech’s AI Spending Spree Comes with a Catch”, May 2024
38. Google, “100% Renewable is Just the Beginning”, December 2016
39. Gartner, April 2024, ibid
40. Venture Beat, “Why Small Language Models are the Next Big Thing in AI”, April 2024
41. Each ChatGPT search uses an estimated gallon of water. Microsoft, Meta, and Alphabet have all set targets to become water positive by 2030. See Liontrust, “The New Investment Landscape for the Water Industry”, June 2024
42. Epoch AI, “Will We Run Out of Data? Limits of LLM Scaling Based on Human-Generated Data”, June 2024. See also full paper: Pablo Villalobos, Anson Ho, Jaime Sevilla, Tamay Besiroglu, Lennart Heim, and Marius Hobbhahn. “Will We Run Out of Data? Limits of LLM Scaling Based on Human-Generated Data”. ArXiv arXiv:2211.04325v2 [cs.LG], 2024. https://doi.org/10.48550/arXiv.2211.04325 Graph generated from selected model data in EpochAI’s Notable AI Models set, under Creative Commons Attribution license. Epoch AI, “Data on Notable AI Models”. Published online at epochai.org. Retrieved from ‘https://epochai.org/data/notable-ai-models’ [online resource]. Accessed 1 Jul 2024.
43. Gartner, April 2024, ibid
44. Company Press Releases, “Stack Overflow and Google Cloud Announce Strategic Partnership to Bring Generative AI to Millions of Developers”, February 2024; “Stack Overflow and OpenAI Partner to Strengthen the World’s Most Popular Large Language Models”, May 2024; “OpenAI and Reddit Partnership”, May 2024; Reddit, “Expanding our Partnership with Google”, February 2024
45. The Verge, “OpenAI’s News Corp Deal Licenses Content from WSJ, New York Post and More”, May 2024; OpenAI Announcement, “A Landmark Multi-Year Global Partnership with News Corp”, May 2024
46. For more on Security for AI, see Menlo Ventures, “Security for AI: The New Wave of Startups Racing to Secure the AI Stack”, February 2024
47. Bidirectional Encoder Representations from Transformers, LLM introduced by Google in October 2018
48. See for example, Markets and Markets, “Autonomous AI and Autonomous Agents Market”, June 2023
49. Axios, “When AI-Produced Code Goes Bad”, June 2024
50. Quartz, “An AI Startup That’s Not Even Six Months Old Says It’s Worth $2 Billion”, April 2024
51. Next Unicorn, “LangChain Secures $25 Million in Funding”, February 2024. See also LangChain Press Release, “Announcing the General Availability of LangSmith and Our Series A Led by Sequoia Capital”, February 2024