Power efficient chipsets are set to be the main driver as artificial intelligence (AI) moves makes a significant shift from the cloud to the edge, according to ABI Research.
Mobile vendor Huawei is already introducing on-device AI training for battery power management in its P20 Pro handset (right), in partnership with Cambricon Technologies.
Chip vendors NVIDIA, Intel, and Qualcomm are also making a push to deliver the hardware that will enable automotive OEMs to experiment with on-device AI training to support their efforts in autonomous driving.
Training at the edge on-device is beginning to gain momentum in terms of R&D, but it could still take some take some time for it to be a realist approach in most segments.
“The massive growth in devices using AI is positive for all players in the ecosystem concerned, but critically those players enabling AI at the edge are going to see an increase in demand that the industry to date has overlooked,” said Jack Vernon, Industry Analyst of ABI Research.
“Vendors can no longer go on ignoring the potential of AI at the edge. As the market momentum continues to swing toward ultra-low latency and more robust analytics, end users must start to incorporate edge AI in their roadmap. They need to start thinking about new business models like end-to-end integration or chipset as a service,” he added.
According to the research firm, the shift from the cloud to the edge will happen initially in terms of inference (machine learning) and then by training.
This shift translates to a huge opportunity for chipset vendors with power efficient chipsets and other products that can meet the demand for edge AI computing.
Edge AI inference will grow from just six percent in 2017 to 43 percent in 2023.
“The shift to the edge for AI processing will be driven by cheaper edge hardware, mission-critical applications, a lack of reliable and cost-effective connectivity options, and a desire to avoid expensive cloud implementation. Consumer electronics, automotive, and machine vision vendors will play an initial critical role in driving the market for edge AI hardware. Scaling said hardware to a point where it becomes cost effective will enable a greater number of verticals to begin moving processing out of the cloud and on to the edge,” said Vernon.
ABI Research has identified 11 verticals ripe for the adoption of AI, including automotive, mobile devices, wearables, smart home, robotics, small unmanned aerial vehicles, smart manufacturing, smart retail, smart video, smart building, and oil and gas sectors and split across a further 58 use cases.
By 2023 the market will witness 1.2 billion shipment of devices capable of on-device AI inference – up from 79 million in 2017.
Cloud providers will still play a pivotal role, particularly when it comes to AI training. Out of three billion AI device shipments that will take place in 2023, more than 2.2 billion will rely on cloud service providers for AI training – this is still a real-term decline in the cloud providers market share for AI training, which currently stands around 99 percent, but will fall to 76 percent by 2023.
Hardware providers should not be too concerned about this shift away from the cloud, as AI training is likely to be supported by the same hardware, only at the edge, either on-premise servers or gateway systems.