Edge computing AI: an opportunity for data science
It is not new that edge computing AI is a big deal for data science and industrial application.
After all, Deloitte reports that in 2020 alone some 750 million edge AI chips would be sold, a market worth some $2.6 billion, and all capable of moving the capacity to run AI models onto an embedded device instead of relying on the cloud.
This investment coupled with enthusiasm for TinyML and the trend towards adopting edge AI for reasons ranging from lower computing costs to better environmental outcomes means that data scientists are hearing more and more about edge computing.
But taking your AI model to the edge is not always straightforward.
How can you help data science to think edge computing? How do you bring an AI model to the edge? How do you convert an existing Artificial intelligence model to something useful on a device? How should data from an embedded device be communicated to the cloud? And what type of hardware can you use for your edge computing, and why should it matter?
In this article, I’ll explain the answers to these questions and more to help demonstrate why taking artificial intelligence to the edge is a great step forward for data science.
What is the ‘edge’ in edge computing or edge AI?
Let’s start with the most basic question: just what is the ‘edge’ in ‘edge AI’ or ‘edge computing?
Put simply, the edge is shorthand for the processing of data closest to the end-user. In practical terms, this means that data might be processed on a connected device rather than pushed to the cloud for processing. It’s increasingly popular and with good reason. Audio and video processed closer to its capture can be processed and rendered faster than pushing it to the cloud or a local network, for example, and sensor data captured by an IoT device can be processed on that same device more efficiently, and with less chance of a data loss.
Before selecting the right Hardware platform, remember the reasons for using Edge computing AI?
Running an artificial intelligence model at the edge has four clear advantages for data science over-relying on the cloud:
- A reliable, always-up connection: When you can gather and process data on the same device instead of relying on a network connection to the cloud, you’ll avoid network connection issues.
- Goodbye to latency: When processing is local, you’ll avoid any and all issues around latency in communicating with the cloud.
- Fewer security and privacy issues: No need to communicate with the network or the cloud means reducing the risk that data will be intercepted, lost, stolen, or leaked.
- Lower bandwidth costs: Reduce the communication between devices and the cloud and the cost of your bandwidth is reduced, too.
With all this in favor of the edge computing you might get the impression that you don’t need the cloud at all. Well, not so fast.
Edge computing does not mean “no cloud”
Short answer: yes, you probably still need the cloud. While processing data on a remote device or elsewhere on the edge has its advantages, there are also limits. Training Artificial Intelligence models requires significant processing power that can usually only be found in the cloud.
With the latest research from IBM suggesting that AI models can be trained with more limited processing power, this might be changing. But this remains only a research topic for the moment and this new method of training networks is not yet available commercially.
That said, if your data science use case is anomaly detection and you have labels for your data, you do have one option: Nanoedge AI. Developed by Cartesiam, this framework allows the training of anomaly detection networks directly on the device. Outside of this very specific use case, though, the options are either to use an existing dataset to train the edge AI Model or send data to the cloud to start the creation of a new dataset.
Hardware & Software Selection
Which solution to send your data Edge to cloud?
There are really two ways to send your data to the cloud: buy or build.
On the buy-side, off-the-shelf solutions like Edge Impulse are built to connect edge devices to a cloud platform and securely transfer datasets between devices and the cloud. Developers can test their free pricing tier to learn how it works and experiments with the solution before stepping up to the enterprise level.
Alternatively, you could build a solution yourself based on a framework like MQTT or CoAP. MQTT is an open-source, industry-standard publish/subscribe messaging transport that is ideal for connecting remote devices with a small code footprint and with minimal network bandwidth. The Constrained Application Protocol, or CoAP, is an alternative to MQTT that is designed for the low-memory, low-power devices that constitute many IoT networks.