Files

19 lines
479 B
Python

from manim import *
from manim_ml.neural_network.activation_functions.activation_function import (
ActivationFunction,
)
class ReLUFunction(ActivationFunction):
"""Rectified Linear Unit Activation Function"""
def __init__(self, function_name="ReLU", x_range=[-1, 1], y_range=[-1, 1]):
super().__init__(function_name, x_range, y_range)
def apply_function(self, x_val):
if x_val < 0:
return 0
else:
return x_val