Skip to main content

ZNNi: Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUs

Author(s): Zlateski, A; Lee, K; Seung, H. Sebastian

Download
To refer to this page use: http://arks.princeton.edu/ark:/88435/pr1m24b
Full metadata record
DC FieldValueLanguage
dc.contributor.authorZlateski, A-
dc.contributor.authorLee, K-
dc.contributor.authorSeung, H. Sebastian-
dc.date.accessioned2021-10-08T19:44:11Z-
dc.date.available2021-10-08T19:44:11Z-
dc.date.issued2017-03-16en_US
dc.identifier.citationZlateski, A, Lee, K, Seung, HS. (2017). ZNNi: Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUs. 854 - 865. doi:10.1109/SC.2016.72en_US
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/pr1m24b-
dc.description.abstractSliding window convolutional networks (ConvNets) have become a popular approach to computer vision problems such as image segmentation and object detection and localization. Here we consider the parallelization of inference, i.e., the application of a previously trained ConvNet, with emphasis on 3D images. Our goal is to maximize throughput, defined as the number of output voxels computed per unit time. We propose CPU and GPU primitives for convolutional and pooling layers, which are combined to create CPU, GPU, and CPU-GPU inference algorithms. The primitives include convolution based on highly efficient padded and pruned FFTs. Our theoretical analyses and empirical tests reveal a number of interesting findings. For example, adding host RAM can be a more efficient way of increasing throughput than adding another GPU or more CPUs. Furthermore, our CPU-GPU algorithm can achieve greater throughput than the sum of CPU-only and GPU-only throughputs.en_US
dc.format.extent854 - 865en_US
dc.language.isoen_USen_US
dc.relation.ispartofInternational Conference for High Performance Computing, Networking, Storage and Analysis, SCen_US
dc.rightsAuthor's manuscripten_US
dc.titleZNNi: Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUsen_US
dc.typeConference Articleen_US
dc.identifier.doidoi:10.1109/SC.2016.72-
dc.date.eissued2017-03-16en_US
pu.type.symplectichttp://www.symplectic.co.uk/publications/atom-terms/1.0/journal-articleen_US

Files in This Item:
File Description SizeFormat 
ZNNi Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUs.pdf645.33 kBAdobe PDFView/Download


Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.