Neural Networks & Perceptron
Neural Networks & Perceptron
-
So we're creating truth table from this one.
-
For example if x1 and x2 is either 1 or 0, and both weights are 1/2. Suppose one equals zero, what are the value of the others to be at least equals the threshold?
-
We con do some equation, either x1 or x2 are zero. and create the point.
-
The result will be a plane that drawn to separate the y=0/y=1 threshold
-
We have specified the line separator for the plane. And if we plot x1,x2 like truth table. We ended up to must have both of them equals 1 to be in the green zone. And that's makes an "AND" table
-
Perceptron training is training our perceptron. It will be wastefull to just do perceptron by hand.
-
There's two options, perceptron rule or gradient descent.
-
Now what we do is some kind of other supervised learning, where we structure our data to have x data set and coressponding parameters, with the output as our data target.
-
Then we perform iteration in curly braces. Each weight will be increasing every iterations.
-
The delta weight will be learning rate . (y-h(x) . x data set
-
here, the H(x) is the activation unit, and the threshold got subtittute to the left side as bias unit.
-
If we cand find linear separable from the data, then the neural networks will find it.
-
Otherwise it goes to infinite loop.