SkillAgentSearch skills...

Tsnet

Tensor Switching Networks

Install / Use

/learn @coxlab/Tsnet
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

TSNet

This repository provides code and scripts for reproducing the experimental results of Tensor Switching Networks by Chuan-Yung Tsai, Andrew Saxe, and David Cox.

TSNet is a novel neural network algorithm, which generalizes the Rectified Linear Unit (ReLU) nonlinearity to tensor-valued hidden units, and avoids the vanishing gradient problem by construction. Our experimental results show that the TSNet is not only more expressive, but also consistently learns faster than standard ReLU networks.

Requirements

Keras, Kerosene, Blessings, and libsvm-compact.

Usage

Use bash nips06/run.sh if you wish to run all of our experiments. Otherwise, use python tsnet_cli.py to run single models. For example, you can use the following commands to compare a simple single-hidden-layer ReLU network (first line) and its TS counterpart using the inverted backpropagation learning (second line):

python tsnet_cli.py -d mnist -n conv:0/16 relu:0 flat:0 sfmx:0/10 -e 10 -lrnparam 1e-3 1e-3 0.9 -v 1
python tsnet_cli.py -d mnist -n conv:2/16 relu:2 flat:0 sfmx:0/10 -e 10 -lrnparam 1e-3 1e-3 0.9 -v 1

By default, the Numpy backend supporting all learning algorithms is used, but you can also switch to the simpler and faster Keras backend by using -k. Please refer to cmp_mlp.py and cmp_cnn.py for more examples of how to define networks.

View on GitHub
GitHub Stars12
CategoryDevelopment
Updated2y ago
Forks9

Languages

Python

Security Score

60/100

Audited on Feb 5, 2024

No findings