Research Article

A Metaheuristic Approach to Map Driving Pattern for Analyzing Driver Behavior Using Big Data Analysis

Algorithm 5

Back-propagation algorithm.
Step 1: Initialize the ReLU derivation to create logical neuron
f(z) = max {0,z}: ReLU
Step 2: compute Activation function derivative z = 1 and the neuron operate in the active region z = 0
Step 3: Compute the Activation value is positive when z = 0.1 to increase input activation is positive
For negative it derivates 0
When z = 0 it chooses either 1 or 0.
Step 4: To Recalls gradient parameters of hidden layers are computed
Step 5: Derivation activate multiplicative factor Trais neurons
Instead of applying f(z) maxout unit divide Z into group of K value.
G(i) indicates of input for group i, {(i − 1) k + 1… ik}
Training freezes when z < 0
To overcome this issue, ReLU is proposed
f(z) = max {0,z} + α min {0,z}.
Value α different variant results are given below
α = −1 absolute value is rectification.
α = 0.01 small value is non-linearity called Leaky ReLU
α = left parameter during training
Step 6: The class ŷ ruled to predict the marginal weight fixed condition
ŷ = 
The parameters θ weight retains the softmax class at the defined active function
Step 7: To differentiate ReLU based cross-entropy with respect to repose dependencies last link layer form activation rule
ℓ(θ) = −)
H produces the relative output on logic condition with active input X