ZNNi: Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUs
Author(s): Zlateski, A; Lee, K; Seung, H. Sebastian
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr1m24b
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Zlateski, A | - |
dc.contributor.author | Lee, K | - |
dc.contributor.author | Seung, H. Sebastian | - |
dc.date.accessioned | 2021-10-08T19:44:11Z | - |
dc.date.available | 2021-10-08T19:44:11Z | - |
dc.date.issued | 2017-03-16 | en_US |
dc.identifier.citation | Zlateski, A, Lee, K, Seung, HS. (2017). ZNNi: Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUs. 854 - 865. doi:10.1109/SC.2016.72 | en_US |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/pr1m24b | - |
dc.description.abstract | Sliding window convolutional networks (ConvNets) have become a popular approach to computer vision problems such as image segmentation and object detection and localization. Here we consider the parallelization of inference, i.e., the application of a previously trained ConvNet, with emphasis on 3D images. Our goal is to maximize throughput, defined as the number of output voxels computed per unit time. We propose CPU and GPU primitives for convolutional and pooling layers, which are combined to create CPU, GPU, and CPU-GPU inference algorithms. The primitives include convolution based on highly efficient padded and pruned FFTs. Our theoretical analyses and empirical tests reveal a number of interesting findings. For example, adding host RAM can be a more efficient way of increasing throughput than adding another GPU or more CPUs. Furthermore, our CPU-GPU algorithm can achieve greater throughput than the sum of CPU-only and GPU-only throughputs. | en_US |
dc.format.extent | 854 - 865 | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartof | International Conference for High Performance Computing, Networking, Storage and Analysis, SC | en_US |
dc.rights | Author's manuscript | en_US |
dc.title | ZNNi: Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUs | en_US |
dc.type | Conference Article | en_US |
dc.identifier.doi | doi:10.1109/SC.2016.72 | - |
dc.date.eissued | 2017-03-16 | en_US |
pu.type.symplectic | http://www.symplectic.co.uk/publications/atom-terms/1.0/journal-article | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
ZNNi Maximizing the Inference Throughput of 3D Convolutional Networks on CPUs and GPUs.pdf | 645.33 kB | Adobe PDF | View/Download |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.