Added sigmoid activation, and ability to do activation above linear layers.

This commit is contained in:
Alec Helbling
2023-01-25 17:36:44 -05:00
parent 4948c0ea4e
commit 301b230c73
7 changed files with 57 additions and 10 deletions

View File

@ -4,7 +4,6 @@ import random
import manim_ml.neural_network.activation_functions.relu as relu
class ActivationFunction(ABC, VGroup):
"""Abstract parent class for defining activation functions"""