# Neural Networks & Perceptron

- So we're creating truth table from this one.
- For example if x1 and x2 is either 1 or 0, and both weights are 1/2. Suppose one equals zero, what are the value of the others to be at least equals the threshold?
- We con do some equation, either x1 or x2 are zero. and create the point.
- The result will be a plane that drawn to separate the y=0/y=1 threshold

- We have specified the line separator for the plane. And if we plot x1,x2 like truth table. We ended up to must have both of them equals 1 to be in the green zone. And that's makes an "AND" table

- Perceptron training is training our perceptron. It will be wastefull to just do perceptron by hand.
- There's two options, perceptron rule or gradient descent.

- Now what we do is some kind of other supervised learning, where we structure our data to have x data set and coressponding parameters, with the output as our data target.
- Then we perform iteration in curly braces. Each weight will be increasing every iterations.
- The delta weight will be learning rate . (y-h(x) . x data set
- here, the H(x) is the activation unit, and the threshold got subtittute to the left side as bias unit.
- If we cand find linear separable from the data, then the neural networks will find it.
- Otherwise it goes to infinite loop.

Contents © 2017 Jonathan Hari Napitupulu. All code except stated otherwise is protected by MIT License - Powered by Nikola