In the ongoing global artificial intelligence (AI) arms race, technologies like generative artificial intelligence (GenAI) have already demonstrated many transformative capabilities, enhancing analysis, speed, and scale across various industries—including cybersecurity.

Yet the pace of AI advancement shows no signs of slowing down, with industry leaders already exploring the next frontier: integrating AI into edge devices like personal computers (PCs) and mobile devices. In fact, worldwide shipments of AI personal computers (AI PCs) and GenAI smartphones are projected to reach 295 million units by the end of 2024—up from just 29 million in 2023. This anticipated surge underscores the critical role that AI compute security plays in safeguarding these emerging technologies from cyber threats, emphasizing the need to understand its development, significance, and benefits.


To date, AI—particularly specialized branches like GenAI—has predominantly relied on massive data centers that provide the extensive data storage, robust network infrastructure, and high computational power required for training and deploying large-language models. These data centers enable the handling of vast datasets and complex calculations that smaller, local systems typically cannot support. However, this approach poses significant challenges including limited computational power and capacity, high energy consumption, scalability and high cost challenges, and latency issues that can affect real-time processing and responsiveness. Latency refers to a delay in data transmission and processing, which can hinder the performance of AI applications. To mitigate these limitations and improve the overall user experience, many experts view edge computing as a solution.

Companies like Qualcomm, with its next-generation Snapdragon processor, and Google, with its Pixel 8 powered by Gemini Nano, are at the forefront of this transition, developing hardware and software solutions that bring AI capabilities closer to users. Processing data locally allows AI-enabled mobile devices to operate faster and more reliably, which is crucial for applications requiring real-time response. Industry leaders are also integrating AI into PCs to enhance their capabilities. For example, Microsoft’s CoPilot+  PCs aim to boost productivity by embedding AI capabilities directly into client PCs. Similarly, Intel’s AI PC Acceleration Program is expected to enable AI on more than 100 million PCs through 2025, improving various aspects of the user experience—from audio effects to video collaboration.

However, this shift also has the potential to amplify existing cybersecurity vulnerabilities and introduce new risks. As AI computations happen closer to the user and on more devices, the attack surface expands, necessitating robust security measures to combat potential misuse and abuse. Therefore, understanding the relationship between edge computing, AI compute security, and edge AI is critical to accurately assessing potential security threats and developing effective cybersecurity strategies. The following definitions and examples elucidate the relationship between these concepts.

Significance and Benefits of AI Compute Security

The integration of edge computing, AI compute security, and edge AI represents a paradigm shift in AI technology deployment and management. Edge computing provides the infrastructure for local data processing, which is essential for the real-time capabilities of edge AI. As noted previously, edge computing reduces latency, enhances performance, and improves user privacy by processing data locally on edge devices. AI compute security protects the data and models processed on these devices from cyber threats, thereby maintaining their integrity and privacy.

Together, these technologies enable the development of robust, efficient, and secure AI uses that leverage a hybrid approach, combining the real-time processing capabilities of edge computing with the advanced analytics and storage capacities of central data centers. This synergy is crucial to ensure the next generation of AI-enabled devices are not only powerful, but also secure and trustworthy. The combination of edge computing, edge AI, and AI compute security establishes a strong foundation for advanced AI technology, delivering significant benefits in security, efficiency, and performance. These benefits fall into three key categories.

1. Strengthens Data Integrity and Privacy Protections

AI compute security is critical in ensuring the integrity and security of data within edge devices. The architecture of edge computing substantially reduces the risk of unauthorized access and breaches by minimizing long-distance data transmission. A key element in this security approach is the secure boot process, which verifies the software integrity at startup and blocks any malicious code from executing. Such localized processing keeps personal data secure, directly enhancing user trust and ensuring data protection management. Alongside secure boot processes, hardware-based security and advanced encryption methods collectively strengthen the protection of sensitive information.

2. Enhances System Reliability and Resilience

Incorporating AI compute security enhances the resilience and reliability of edge devices, ensuring continuous operation even in the face of emerging and amplified cyber threats. By decentralizing AI computations, these devices can maintain functionality and service availability even if network connectivity is compromised. For example, AI PCs can process AI data locally without calling on cloud data centers. By securing the computations on the PC itself, these systems can continue to operate reliably during cyber incidents, mitigating the risks associated with network disruptions and cyberattacks.

3. Optimizes Resource Allocation and Availability

Implementing AI compute security measures optimizes the efficiency of security protocols, minimizing overhead while maintaining performance. This can include hardware-assisted security features, such as trusted platform modules, that provide isolated environments for sensitive computations. These features ensure that security tasks do not consume excessive computational resources, allowing the primary AI functions to operate without significant delays. Edge computing and edge AI contribute to system efficiency by enabling local data processing. Not only does this local processing speed up operations, but it also reduces the dependency on constant, reliable, high-speed internet connectivity, making the device more versatile and resilient in varied network conditions. Additionally, this efficiency translates to energy savings, as local processing is less resource-intensive than transmitting data to and from cloud servers.

Preparing for the Next AI Frontier

As AI technologies continue to advance, their integration into edge devices becomes increasingly critical. Yet to integrate these technologies safely, we must first develop comprehensive security solutions, which requires establishing robust security standards and guidance; promoting transparency and accountability; and fostering collaboration between industry, government, and academic stakeholders. Policymakers and industry leaders must prioritize the development of secure frameworks that address the unique security challenges posed by edge computing and edge AI while maximizing their opportunities and benefits. This knowledge is crucial to prepare for the next AI frontier and sets the stage for assessing security threats to AI compute security—the focus of our next article.

Get the latest artificial intelligence policy research and analysis in your inbox. Sign up for the R Street newsletter today.