TY - JOUR
T1 - Pseudoinvertible Neural Networks
AU - Bolluyt, Elijah D.
AU - Comaniciu, Cristina
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2024/2/1
Y1 - 2024/2/1
N2 - This work introduces novel architecture components and training procedures to create augmented neural networks with the ability to process data bidirectionally via an end-to-end approximate inverse. We develop pseudoinvertible neural network (PsI-NN) layers which function as drop-in replacements for corresponding convolutional and fully connected layers; by using these, existing architectures gain a pseudoinverse function which, with training, approximately reverses the forward function. For cases where learning both a task and its inverse are necessary or desirable, we show that PsI-NN enabled models match or exceed the quality of results generated by systems that use two separate models while drastically reducing system parameter count. We demonstrate this on two tasks: unpaired image translation and semisupervised classification. In both, PsI-NN greatly reduces parameter count without any loss of output quality; in semisupervised image classification, PsI-NN improves classification performance beyond the baseline autoencoder method. Our approach creates pseudoinvertible versions of existing architectures, circumventing the stringent constraints required to create a true inverse while allowing a single neural network to learn a task in two directions simultaneously.
AB - This work introduces novel architecture components and training procedures to create augmented neural networks with the ability to process data bidirectionally via an end-to-end approximate inverse. We develop pseudoinvertible neural network (PsI-NN) layers which function as drop-in replacements for corresponding convolutional and fully connected layers; by using these, existing architectures gain a pseudoinverse function which, with training, approximately reverses the forward function. For cases where learning both a task and its inverse are necessary or desirable, we show that PsI-NN enabled models match or exceed the quality of results generated by systems that use two separate models while drastically reducing system parameter count. We demonstrate this on two tasks: unpaired image translation and semisupervised classification. In both, PsI-NN greatly reduces parameter count without any loss of output quality; in semisupervised image classification, PsI-NN improves classification performance beyond the baseline autoencoder method. Our approach creates pseudoinvertible versions of existing architectures, circumventing the stringent constraints required to create a true inverse while allowing a single neural network to learn a task in two directions simultaneously.
KW - Semisupervised learning
KW - convolutional neural networks (CNN)
KW - unsupervised learning
UR - http://www.scopus.com/inward/record.url?scp=85165257668&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85165257668&partnerID=8YFLogxK
U2 - 10.1109/TAI.2023.3296573
DO - 10.1109/TAI.2023.3296573
M3 - Article
AN - SCOPUS:85165257668
VL - 5
SP - 602
EP - 612
JO - IEEE Transactions on Artificial Intelligence
JF - IEEE Transactions on Artificial Intelligence
IS - 2
ER -