SciTePress - Publication Details
loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Authors: Kyle Rogers 1 ; Hao Yu 1 ; Seong-Eun Cho 2 ; Nancy Fulda 1 ; Jordan Yorgason 3 and Tyler Jarvis 2

Affiliations: 1 Department of Computer Science, Brigham Young University, Provo, Utah, U.S.A. ; 2 Department of Mathematics, Brigham Young University, Provo, Utah, U.S.A. ; 3 Cellular Biology and Physiology, Center for Neuroscience, Brigham Young University, Provo, Utah, U.S.A.

Keyword(s): Machine Learning, Matrix Abstraction, Biologically Inspired Learning Algorithm, Model Parallelization, Network Modularization, Backpropagation, Skip Connections, Neuromorphic.

Abstract: In this work we introduce a novel method for decoupling the backward pass of backpropagation using mathematical and biological abstractions to approximate the error gradient. Inspired by recent findings in neuroscience, our algorithm allows gradient information to skip groups of layers during the backward pass, such that weight updates at multiple depth levels can be calculated independently. We explore both gradient abstractions using the identity matrix as well as an abstraction that we derive mathematically for network regions that consist of piecewise-linear layers (including layers with ReLU and leaky ReLU activations). We validate the derived abstraction calculation method on a fully connected network with ReLU activations. We then test both the derived and identity methods on the transformer architecture and show the capabilities of each method on larger model architectures. We demonstrate empirically that a network trained using an appropriately chosen abstraction matrix can match the loss and test accuracy of an unmodified network, and we provide a roadmap for the application of this method toward depth-wise parallelized models and discuss the potential of network modularization by this method. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 8.209.245.224

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Rogers, K., Yu, H., Cho, S.-E., Fulda, N., Yorgason, J. and Jarvis, T. (2024). Decoupling the Backward Pass Using Abstracted Gradients. In Proceedings of the 16th International Conference on Agents and Artificial Intelligence - Volume 3: ICAART; ISBN 978-989-758-680-4; ISSN 2184-433X, SciTePress, pages 507-518. DOI: 10.5220/0012362800003636

@conference{icaart24,
author={Kyle Rogers and Hao Yu and Seong{-}Eun Cho and Nancy Fulda and Jordan Yorgason and Tyler Jarvis},
title={Decoupling the Backward Pass Using Abstracted Gradients},
booktitle={Proceedings of the 16th International Conference on Agents and Artificial Intelligence - Volume 3: ICAART},
year={2024},
pages={507-518},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0012362800003636},
isbn={978-989-758-680-4},
issn={2184-433X},
}

TY - CONF

JO - Proceedings of the 16th International Conference on Agents and Artificial Intelligence - Volume 3: ICAART
TI - Decoupling the Backward Pass Using Abstracted Gradients
SN - 978-989-758-680-4
IS - 2184-433X
AU - Rogers, K.
AU - Yu, H.
AU - Cho, S.
AU - Fulda, N.
AU - Yorgason, J.
AU - Jarvis, T.
PY - 2024
SP - 507
EP - 518
DO - 10.5220/0012362800003636
PB - SciTePress