The classic drivers of edge computing deployment are as much at work in the video surveillance industry as anywhere else.
Scalability
Rising AI adoption in increasingly data-intensive use cases, such as monitoring high definition video, requires more and more computing infrastructure, especially as the AI models themselves are growing very fast. The sheer number of edge devices (cameras and recorders) is an argument for moving from centralized to distributed processing. Edge AI is a significant factor in the distribution of inference workloads to the edge (cameras and devices), improving the efficiency of the video analytics solution.
Scalability 2
Rising AI adoption in increasingly data-intensive use cases, such as monitoring high definition video, requires more and more computing infrastructure, especially as the AI models themselves are growing very fast. The sheer number of edge devices (cameras and recorders) is an argument for moving from centralized to distributed processing. Edge AI is a significant factor in the distribution of inference workloads to the edge (cameras and devices), improving the efficiency of the video analytics solution.
Scalability 3
Rising AI adoption in increasingly data-intensive use cases, such as monitoring high definition video, requires more and more computing infrastructure, especially as the AI models themselves are growing very fast. The sheer number of edge devices (cameras and recorders) is an argument for moving from centralized to distributed processing. Edge AI is a significant factor in the distribution of inference workloads to the edge (cameras and devices), improving the efficiency of the video analytics solution.
Processing and Power Budgets
From an AI processor chip standpoint, the development breakthrough happened when two things were accomplished. One is to be able to process the video analytics at low power. Network cameras have a power budget. Typically, just three watts for the subsystem of SoC plus DRAM. The breakthrough on the technical side was a new generation of SoCs that enable analytics within this power budget.
Previously, it was only possible to do the same with GPUs, FPGAs, and CPUs, but the power would be prohibitively higher. This is perfect for a server, but it doesn’t fit an analytic camera. This advance was to create new architectures to be able to get the power budget low enough. In addition to this, the network camera also needs to work as a camera. It needs to be able to do image processing and video encoding, and sometimes up to 4K resolution. The second challenge from the chip design point of view, was to not only be able to do the analytics at low power, but to still keep the performance to do all the image processing, such as high dynamic range low light processing and H.265 video encoding, plus all of the video analytics, within a very low power budget.
AI Chipset Revenue by Device Processing and Power – Cameras Only
Resolution Increase and Bandwidth Management
There is an ongoing trend in the market toward higher resolution cameras. While resolution is just one aspect of video quality, the key benefit of higher-resolution cameras is sharper and more clearly defined pictures.
As with consumer video products, such as televisions, the typical resolution of video surveillance cameras continues to increase over time. However, the typical resolution of video surveillance cameras lags that of many consumer video products. For video surveillance, full HD resolution is often considered a sufficient compromise between image quality, bandwidth, storage, and cost for most indoor applications. Production of network cameras with resolution lower than full HD has reduced.
The trend will be from 2.1-megapixel (Full HD) resolution cameras accounting for the largest proportion of unit shipments now to 4–5-megapixel cameras accounting for the largest proportion in 2025. 4K (8.3 megapixel and higher) cameras could then account for the largest proportion in the longer term beyond 2030.
Increased camera resolution is a driver for more powerful edge AI. For example, if we think of a camera that is in an airport or a parking lot. If it is a 4K camera with a wide-angle lens, it can see people that are far off in the field of view. At 4K, it might have enough pixels on the target to be able to identify a face or a license plate, but at lower resolutions this will be more challenging.
Currently, a large proportion of server analytics algorithms don’t work on 4K input. Images are downscaled as the video is transferred to the server to allow bandwidth management, and then upscaled. This does not typically upscale to the full 4K image resolution.
By processing the image on the camera, the analysis can be performed on the raw 4K image, and then just the meta data will be transferred to the server for analysis. Due to this trend, Omdia estimates that the requirement for more powerful AI processing throughput will constantly increase. Consequently, camera chipsets capable of processing more than 5 TOPS will grow at a CAGR of 71% from 2020 to 2026.