Microprocessors and Microsystems 52 (2017) 2–22 Contents lists available at ScienceDirect Microprocessors and Microsystems journal homepage: www.elsevier.com/locate/micpro System-level design space identification for Many-Core Vision Processors Jones Yudi a,b, , Carlos Humberto Llanos b , Michael Huebner a a Embedded Systems for Information Technology, Ruhr-University Bochum, Germany b Automation and Control Group, University of Brasilia, Brazil a r t i c l e i n f o Article history: Received 22 January 2017 Revised 19 May 2017 Accepted 24 May 2017 Available online 26 May 2017 a b s t r a c t The current main trends in the embedded systems area, the Cyber-Physical Systems (CPS) and the Internet-of-Things (IoT), are leveraging the development of complex, distributed, low-power, and high- performance embedded systems. An important feature needed in this new Era is the embedded intelli- gence enabling to locally process data and actuate over the environment, without the need of a remote central processing server. In this context, emerged the Smart Cameras: devices able to acquire images and apply sophisticated algorithms for different Image Processing and Computer Vision (IP/CV) applications. Both the technology convergence and the evolution of embedded systems to multi/many-core architec- tures allow envisioning future cameras as many-core systems able to efficiently explore the natural IP/CV parallelism to meet embedded application’s constraints, e.g. real-time, power consumption, silicon area, temperature management, fault tolerance, among others. In this work, we show the development of a Many-Core Vision Processor architecture, suitable for future Smart Cameras. In our design methodology, we analyze several aspects involved, from high-level application analysis down to fine-grained operations and physical aspects (e.g. geometry and spatial distribution). The main analysis is performed using a Sys- temC/TLM2.0 simulator specially developed for this project. Silicon Area, Power Consumption and Timing estimations are also provided as results of an early Design-Space Exploration (DSE). Using these results we propose a first complete architecture, which is implemented in an FPGA. Details about the hardware implementation are provided, as well as synthesis results. In comparison to other works, from the litera- ture, the implemented architecture shows the potential of the project developed in this work. © 2017 Elsevier B.V. All rights reserved. 1. Introduction The evolution of the microelectronics technology has enabled the industry to integrate billions of transistors on a single chip, following the well-known Moore’s Law prediction. This evolu- tion enabled the development of new ideas, products and whole new markets based on the embedded processing systems. Follow- ing this trend, some new concepts emerged some years ago: the Cyber-Physical Systems (CPS), the Ubiquitous Computing (UC) and the Internet of Things (IoT), which are expected to change the in- teraction between humans and the surrounding environment. The computing capabilities embedded in different devices, daily-life objects, buildings, and so on, will be distributed, ubiquitous and transparent to the users [1]. Corresponding author. E-mail addresses: jones.morialvesdasilva@rub.de, jonesyudi@unb.br, jones.morialvesdasilva@ruhr-uni-bochum.de (J. Yudi), llanos@unb.br (C. Humberto Llanos), michael.huebner@rub.de (M. Huebner). Distributed embedded devices and intelligent sensors will col- lect and process huge amounts of data, enabling the pervasive- ness of the computational environment [2]. These sensors, also known as Smart Sensors, are devices able to perform not only pre-processing algorithms but also more complex applications with embedded and stand-alone intelligence. The Smart Cameras, a par- ticular type of Smart Sensor, will be essential as devices able to capture and interpret the human behavior in an environment, among other events/objects. These future cameras must be able to perform complex applications simultaneously, coping with real- time constraints. Also, other important requisites must be fulfilled: energy consumption, chip temperature control, reliability, Quality of Service (QoS), data security, privacy, power management, cost (silicon area), and so on [3]. Image Processing and Computer Vision (IP/CV) applications are computationally costly, mainly due to the huge amount of data to be processed. New technologies are enabling the use of in- creasing resolutions, and new applications are demanding high- performance with tight deadlines. In this scenario, there are cam- http://dx.doi.org/10.1016/j.micpro.2017.05.013 0141-9331/© 2017 Elsevier B.V. All rights reserved.