TUTOR: Training Neural Networks Using Decision Rules as Model Priors
Author(s): Hassantabar, Shayan; Terway, Prerit; Jha, Niraj K
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr1br8mg36
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Hassantabar, Shayan | - |
dc.contributor.author | Terway, Prerit | - |
dc.contributor.author | Jha, Niraj K | - |
dc.date.accessioned | 2023-12-24T15:21:45Z | - |
dc.date.available | 2023-12-24T15:21:45Z | - |
dc.date.issued | 2022-05-31 | en_US |
dc.identifier.citation | Hassantabar, Shayan, Terway, Prerit, Jha, Niraj K. (2023). TUTOR: Training Neural Networks Using Decision Rules as Model Priors. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 42 (2), 483 - 496. doi:10.1109/tcad.2022.3179245 | en_US |
dc.identifier.issn | 0278-0070 | - |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/pr1br8mg36 | - |
dc.description.abstract | The human brain has the ability to carry out new tasks with limited experience. It utilizes prior learning experiences to adapt the solution strategy to new domains. On the other hand, deep neural networks (DNNs) generally need large amounts of data and computational resources for training. However, this requirement is not met in many settings. To address these challenges, we propose the TUTOR DNN synthesis framework. TUTOR targets tabular datasets. It synthesizes accurate DNN models with limited available data and reduced memory/computational requirements. It consists of three sequential steps. The first step involves generation, verification, and labeling of synthetic data. The synthetic data generation module targets both the categorical and continuous features. TUTOR generates the synthetic data from the same probability distribution as the real data. It then verifies the integrity of the generated synthetic data using a semantic integrity classifier module. It labels the synthetic data based on a set of rules extracted from the real dataset. Next, TUTOR uses two training schemes that combine synthetic and training data to learn the parameters of the DNN model. These two schemes focus on two different ways in which synthetic data can be used to derive a prior on the model parameters and, hence, provide a better DNN initialization for training with real data. In the third step, TUTOR employs a grow-and-prune synthesis paradigm to learn both the weights and the architecture of the DNN to reduce model size while ensuring its accuracy. We evaluate the performance of TUTOR on nine datasets of various sizes. We show that in comparison to fully-connected DNNs, TUTOR, on an average, reduces the need for data by 5.9× (geometric mean), improves accuracy by 3.4%, and reduces the number of parameters (floating-point operations) by 4.7× (4.3×) (geometric mean). Thus, TUTOR enables a less data-hungry, more accurate, and more compact DNN synthesis. | en_US |
dc.format.extent | 483 - 496 | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartof | IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems | en_US |
dc.rights | Author's manuscript | en_US |
dc.title | TUTOR: Training Neural Networks Using Decision Rules as Model Priors | en_US |
dc.type | Journal Article | en_US |
dc.identifier.doi | doi:10.1109/tcad.2022.3179245 | - |
dc.identifier.eissn | 1937-4151 | - |
pu.type.symplectic | http://www.symplectic.co.uk/publications/atom-terms/1.0/journal-article | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
2010.05429.pdf | 10.18 MB | Adobe PDF | View/Download |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.