Last updated:
0 purchases
pytorchspiking 0.1.0
PyTorchSpiking
PyTorchSpiking provides tools for training and running spiking neural networks
directly within the PyTorch framework. The main feature is
pytorch_spiking.SpikingActivation, which can be used to transform
any activation function into a spiking equivalent. For example, we can translate a
non-spiking model, such as
torch.nn.Sequential(
torch.nn.Linear(5, 10),
torch.nn.ReLU(),
)
into the spiking equivalent:
torch.nn.Sequential(
torch.nn.Linear(5, 10),
pytorch_spiking.SpikingActivation(torch.nn.ReLU()),
)
Models with SpikingActivation layers can be optimized and evaluated in the same way as
any other PyTorch model. They will automatically take advantage of PyTorchSpiking’s
“spiking aware training”: using the spiking activations on the forward pass and the
non-spiking (differentiable) activation function on the backwards pass.
PyTorchSpiking also includes various tools to assist in the training of spiking models,
such as filtering layers.
If you are interested in building and optimizing spiking neuron models, you may also
be interested in NengoDL. See
this page for a
comparison of the different use cases supported by these two packages.
Documentation
Check out the documentation for
Installation instructions
More detailed example introducing the features of PyTorchSpiking
API reference
Release history
0.1.0 (September 9, 2020)
Initial release
For personal and professional use. You cannot resell or redistribute these repositories in their original state.
There are no reviews.