Export Publication

The publication can be exported in the following formats: APA (American Psychological Association) reference format, IEEE (Institute of Electrical and Electronics Engineers) reference format, BibTeX and RIS.

Export Reference (APA)
Rosario, V. M. Do, Breternitz, M. & Borin, E. (2021). Efficiency and scalability of Multi-Lane Capsule Networks (MLCN). Journal of Parallel and Distributed Computing. 155, 63-73
Export Reference (IEEE)
V. M. Rosario et al.,  "Efficiency and scalability of Multi-Lane Capsule Networks (MLCN)", in Journal of Parallel and Distributed Computing, vol. 155, pp. 63-73, 2021
Export BibTeX
@article{rosario2021_1716147107627,
	author = "Rosario, V. M. Do and Breternitz, M. and Borin, E.",
	title = "Efficiency and scalability of Multi-Lane Capsule Networks (MLCN)",
	journal = "Journal of Parallel and Distributed Computing",
	year = "2021",
	volume = "155",
	number = "",
	doi = "10.1016/j.jpdc.2021.04.010",
	pages = "63-73",
	url = "https://www.sciencedirect.com/journal/journal-of-parallel-and-distributed-computing"
}
Export RIS
TY  - JOUR
TI  - Efficiency and scalability of Multi-Lane Capsule Networks (MLCN)
T2  - Journal of Parallel and Distributed Computing
VL  - 155
AU  - Rosario, V. M. Do
AU  - Breternitz, M.
AU  - Borin, E.
PY  - 2021
SP  - 63-73
SN  - 0743-7315
DO  - 10.1016/j.jpdc.2021.04.010
UR  - https://www.sciencedirect.com/journal/journal-of-parallel-and-distributed-computing
AB  - Some Deep Neural Networks (DNN) have what we call lanes, or they can be reorganized as such. Lanes are paths in the network which are data-independent and typically learn different features or add resilience to the network. Given their data-independence, lanes are amenable for parallel processing. The Multi-lane CapsNet (MLCN) is a proposed reorganization of the Capsule Network which is shown to achieve better accuracy while bringing highly-parallel lanes. However, the efficiency and scalability of MLCN had not been systematically examined. In this work, we study the MLCN network with multiple GPUs finding that it is 2x more efficient than the original CapsNet when using model-parallelism. We introduce the load balancing problem of distributing heterogeneous lanes in homogeneous or heterogeneous accelerators and show that a simple greedy heuristic can be almost 50% faster than a naïve random approach. Further, we show that we can generate MLCN models with heterogeneous lanes with a balanced fit for a given set of devices. We describe a Neural Architectural Search generating MLCN models matching the device's memory that are load balanced. This search discovered models with 18.6% better accuracy for CIFAR-10.
ER  -