Anzeige

Economies of Scale

Author: Jeff Bier, Founder Embedded Vision Alliance and President BDTI

The expansion of computer-vision-based systems and applications is enabled by many factors, including advances in processors, sensors and development tools. But, arguably, the single most important thing driving the proliferation of computer vision is deep learning.

Bild: Embedded Vision AllianceBild: Embedded Vision Alliance

Bild: Embedded Vision AllianceBild: Embedded Vision Alliance

Embedded Vision Summit 2020

Besides BDTI and the Embedded Vision Alliance Jeff Bier is also organizer of the yearly Embedded Vision Summit, the industry´s largest event for practical computer vision. The next event will take place in Santa Clara (California) from 18- 20 May 2020.

www.embedded-vision.com/summit www.embedded-vision.com

The fact that deep learning-based visual perception works extremely well - routinely achieving better results than older, hand-crafted algorithms - has been widely discussed. What is less widely understood, but equally important, is how the rise of deep learning is fundamentally changing the process and economics of developing solutions and building-block technologies for commercial computer vision applications. Prior to the widespread use of deep learning in commercial computer vision applications, developers created highly complex, unique algorithms for each application. These algorithms were usually highly tuned to the specifics of the application, including factors such as image sensor characteristics, camera position, and the nature of the background behind the objects of interest. Developing, testing and tuning these algorithms often consumed tens or even hundreds of person-years of work. Even if a company was fortunate enough to have enough people available with the right skills, the magnitude of the effort required meant that only a tiny fraction of potential computer vision applications could actually be addressed.

Less diverse algorithms

With deep learning, in contrast, we tend to re-use a relatively small handful of algorithms across a wide range of applications and imaging conditions. Instead of inventing new algorithms, we re-train existing, proven algorithms. As a consequence, the algorithms being deployed in commercial computer vision systems are becoming much less diverse. This has two important consequences:

- First, the economics of commercial computer vision applications and building-block technologies have fundamentally shifted. Take processors, for example. Five or ten years ago, developing a specialized processor to deliver significantly improved performance and efficiency on a wide range of computer vision tasks was nearly impossible, due to the extreme diversity of computer vision algorithms. Today, with the focus mainly on deep learning, it's much practical to create a specialized processor that accelerates vision workloads - and it's much easer for investors to see a path for such a processor to sell in large volumes, serving a wide range of applications.

- Second, the nature of computer vision algorithm development has changed. Instead of investing years of effort devising novel algorithms, increasingly these days we select among proven algorithms from the research literature, perhaps tweaking them a Bit for our needs. So, in commercial applications much less effort goes into designing algorithms. But deep learning algorithms require lots of data for training and validation. And not just any data. The data must be carefully curated for the algorithms to achieve high levels of accuracy. So, there's been a substantial shift in the focus of algorithm-related work in commercial computer vision applications, away from devising unique algorithms and towards obtaining the right quantities of the right types of training data.

Embedded Vision Alliance

Dieser Artikel erschien in inVISION 6 2019 - 07.11.19.
Für weitere Artikel besuchen Sie www.invision-news.de