There's no cloud without a good edge, and vice versa, says expert panel

Ben Wodecki, Jr. Editor

July 1, 2021

2 Min Read

Deciding whether to deploy AI at the edge of the network or in a cloud data center should be judged on use cases, with data being a key consideration, speakers told the audience at the Industrial AI Summit earlier this week.

During a panel discussion on the risks and opportunities of deploying AI in production, Data Analysis Bureau CEO Eric Topham emphasized that every decision should revolve around the application, and what that company is trying to achieve with its data.

"It really depends on what you define as the 'edge,'” he said, adding that the quality of the data being sought after "is the be-all and end-all."

Matteo Dariol, innovation strategist at Bosch Rexroth, described AI deployment as "the art of asking the right question."

“If you’re asking the wrong question, you’re never going to get the right result. It’s crucial for any AI project to have a lot of different stakeholders sitting at the table. It’s a team effort.”

All you need is good data, a good cloud, and a good edge

Dariol said that while some end-users may see edge deployments as just another aspect of IT to manage, more are realizing the need to take care of the data that goes in and out of their AI models.

“People from around the company need to come together to make a decision on what use cases that firm is trying to sort – rather than placing the onus on one person,” he added.

Omdia analyst Alex West, who chaired the discussion, said that the application and data would be the factors determining whether it should be a cloud or edge deployment.

“For some, sensitive data, such as custom or ingredient data, will have to stay local. However, with other datasets, e.g. asset health, there's a greater comfort level with using the cloud.”

West also invoked issues related to latency requirements, as well as the cost and level of analytics, as being important considerations.

"The different edges can include anything from a sensor with [an] embedded processor (e.g. Raspberry Pi), through controllers and then on to on-prem servers," he added.

About the Author(s)

Ben Wodecki

Jr. Editor

Ben Wodecki is the Jr. Editor of AI Business, covering a wide range of AI content. Ben joined the team in March 2021 as assistant editor and was promoted to Jr. Editor. He has written for The New Statesman, Intellectual Property Magazine, and The Telegraph India, among others. He holds an MSc in Digital Journalism from Middlesex University.

Keep up with the ever-evolving AI landscape
Unlock exclusive AI content by subscribing to our newsletter!!

You May Also Like