Edge AI refers to the convergence of edge computing and artificial intelligence, where AI applications are deployed at the network edge on smart connected network endpoints.
This is different from the traditional IoT-AI application framework where the data generated by connected technologies is transmitted to a backend cloud system, processed by AI algorithms and the resulting control actions are transmitted across the network to the connected devices.
Instead of running AI models at the backend, they are configured onto processors or FPGA chips inside the connected devices operating at the network edge.
Examples of Edge AI include autonomous vehicles, smart traffic lights and the wider Internet of Vehicles (IoV) network where vehicles, traffic lights and emergency services can mediate between each other to coordinate emergency routes and diversions when necessary.
This coordination means a high level of processing efficiency and accelerated data-driven decision making in real-time. This is made possible in recent years for three key reasons:
(Learn more about edge computing types and IoT monitoring.)
With these drivers, the Intelligent Edge — that is, the infrastructure — meets Edge Intelligence, actual applications of intelligence on edge devices. The term edge AI is used synonymously with edge intelligence. These key elements derive the use of Intelligent Edge capabilities:
AI applications draw data from multiple devices on the network edge and embed intelligence into the automation functionality served by the connected devices.
AI models use data to infer a decision or control action using real-time data generated by these devices. This process takes place in real-time and is achieved using pre-trained AI models.
The networking systems and architecture are adapted to support Edge AI applications. The endpoints are not only sensors but a low-power computing system that can run AI models and are connected with an automation system that performs the required control actions intelligently.
The models can be programmed to train and adapt on new data streams generated by sensor endpoints. This approach requires model embedding onto dedicated FPGA devices or smart devices with onboard computing systems.
The network edge and smart devices involved in Edge AI applications are optimized for:
All of this sounds interesting, except that most networking and IoT devices at the network edge can only be used as a data source.
Machine learning models deployed on smart devices can work well if the model is small and the AI task is limited to solving a simple classification problem. As the model grows in complexity, Edge AI devices will undergo a steep accuracy-resource demand tradeoff. The number of parameters that need to be learned and configured on a computing chip can grow exponentially for AI applications designed to solve complex problems. This means that the device must be equipped with a capable processing chip which consumes high energy.
Ideally, an intelligent machine must be able to adapt and improve in its learning capacity. Edge AI applications can take advantage of distributed edge computing devices, each generating a variety of useful information to train the AI models. However, employing a federated learning approach presents its own set of challenges:
(Explore the ethics of artificial intelligence.)
While Edge AI is destined to play a crucial role in the future of AI adoption, it will also face the same adoption challenges as any conventional AI and digital technology.
Edge AI applications are developed for computing devices with limited computing capacity and hardware designed for specific tasks. The entire application will run between devices without interacting with a backend cloud network.
This means that data governance, security and user privacy capabilities must also be embedded into the AI system and the intelligent edge cannot truly be separated from a centralized computing environment.
See an error or have a suggestion? Please let us know by emailing ssg-blogs@splunk.com.
This posting does not necessarily represent Splunk's position, strategies or opinion.
The Splunk platform removes the barriers between data and action, empowering observability, IT and security teams to ensure their organizations are secure, resilient and innovative.
Founded in 2003, Splunk is a global company — with over 7,500 employees, Splunkers have received over 1,020 patents to date and availability in 21 regions around the world — and offers an open, extensible data platform that supports shared data across any environment so that all teams in an organization can get end-to-end visibility, with context, for every interaction and business process. Build a strong data foundation with Splunk.