Please use this identifier to cite or link to this item: http://hdl.handle.net/10071/34358
Author(s): Bacellar, A.
Susskind, Z.
Breternitz Jr., M.
John, E.
John, L.
Lima, P.
França, F.
Editor: Salakhutdinov R., Kolter Z., Heller K., Weller A., Oliver N., Scarlett J., Berkenkamp F.
Date: 2024
Title: Differentiable weightless neural networks
Volume: 235
Book title/volume: Proceedings of the 41st International Conference on Machine Learning, PMLR
Pages: 2277 - 2295
Event title: Proceedings of Machine Learning Research
Reference: Bacellar, A., Susskind, Z., Breternitz Jr., M., John, E., John, L., Lima, P., & França, F. (2024). Differentiable weightless neural networks. In R. Salakhutdinov, Z. Kolter, K. Heller, A. Weller, N. Oliver, J. Scarlett, & F. Berkenkamp (Eds.), Proceedings of the 41st International Conference on Machine Learning, PMLR (pp. 2277-2295). ML Research Press. http://hdl.handle.net/10071/34358
ISSN: 2640-3498
Keywords: Machine learning
Differentiable networks
Weightless neural networks
Abstract: We introduce the Differentiable Weightless Neural Network (DWN), a model based on interconnected lookup tables. Training of DWNs is enabled by a novel Extended Finite Difference technique for approximate differentiation of binary values. We propose Learnable Mapping, Learnable Reduction, and Spectral Regularization to further improve the accuracy and efficiency of these models. We evaluate DWNs in three edge computing contexts: (1) an FPGA-based hardware accelerator, where they demonstrate superior latency, throughput, energy efficiency, and model area compared to state-of-the-art solutions, (2) a low-power microcontroller, where they achieve preferable accuracy to XGBoost while subject to stringent memory constraints, and (3) ultralow-cost chips, where they consistently outperform small models in both accuracy and projected hardware area. DWNs also compare favorably against leading approaches for tabular datasets, with higher average rank. Overall, our work positions DWNs as a pioneering solution for edge-compatible high-throughput neural networks.
Peerreviewed: yes
Access type: Open Access
Appears in Collections:ISTAR-CRI - Comunicações a conferências internacionais

Files in This Item:
File SizeFormat 
conferenceObject_105156.pdf1,66 MBAdobe PDFView/Open


FacebookTwitterDeliciousLinkedInDiggGoogle BookmarksMySpaceOrkut
Formato BibTex mendeley Endnote Logotipo do DeGóis Logotipo do Orcid 

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.