Here is an animation of the three weight matrices of a neural network during training. We train a neural network to memorize random data, but as we train the network, we toggle L1 regularization on the weights so that the network is regularized for a number of generations, and after the regularization, the network trains without regularization. We repeat this process of turning on and off L1 regularization several times.
The L1 regularization hampers the ability of the network to memorize data, so we see an increase in the loss level of the network. The L1 regularization also increases the sparsity of the neural network, and this causes the entries in the square to appear more red,blue,green, or black.
The neural network is initialized to zero. The structure of the neural network is Chain(SkipConnection(Dense(mn,mn,atan) ,+),SkipConnection(Dense(mn,mn,atan),+),SkipConnection(Dense(mn,mn,atan),+)) where mn=50. This means that it has atan activation and a skip connection for each weight matrix. The skip connections allow us initialize the weight matrices to near zero, and the skip connections allow us to associate the inputs and the outputs of the weight matrices so that it makes sense to overlay these weight matrices in the animation.
The three weight matrices are colored red, green, and blue in the animation. We brighten the animation by taking the square root of the absolute values of all of the entries in these weight matrices.
We observe that after each period of regularization and non-regularization, the neural network reverts nearly to its initial state. This observation shows that L1 regularization (at least for the kind of network that I am training) cannot be used to easily change the structure of a neural network.
The notion of a neural network is not my own. I am simply making this animation to highlight how neural networks behave with respect to AI safety. It is good for AI safety that the neural network is able to recover its original state The neural network is only able to recover its original state under specialized conditions, so the state of a trained neural network is only somewhat special and it still contains a lot of randomness to it. On the other hand, since the network is able to recover its original state, we have a correspondence between neural networks trained with L1 regularization and networks trained without L1 regularization. Since L1 regularized networks are sparse, they are considered more interpretable than non-L1 regularized networks, but since L1 regularized networks correspond to non-L1 regularized networks, the non L1-regularized networks gain the interpretability from L1 regularized networks.
This animation shows that neural networks are able to recover their former states in some scenarios. I have made other visualizations that highlight instances in which neural networks are not able to recover their former states.
Unless otherwise stated, all algorithms featured on this channel are my own. You can go to
github.com/sponsors/jvanname to support my research on machine learning algorithms. I am also available to consult on the use of safe and interpretable AI for your business. I am designing machine learning algorithms for AI safety such as LSRDRs. In particular, my algorithms are designed to be more predictable and understandable to humans than other machine learning algorithms, and my algorithms can be used to interpret more complex AI systems such as neural networks. With more understandable AI, we can ensure that AI systems will be used responsibly and that we will avoid catastrophic AI scenarios. There is currently nobody else who is working on LSRDRs, so your support will ensure a unique approach to AI safety.
YouTube Comments - 0 Comments
Top Comments of this video!! :3