Neural Networks & Perceptron

  |   Source
Neural Networks & Perceptron

  • So we're creating truth table from this one.
  • For example if x1 and x2 is either 1 or 0, and both weights are 1/2. Suppose one equals zero, what are the value of the others to be at least equals the threshold?
  • We con do some equation, either x1 or x2 are zero. and create the point.
  • The result will be a plane that drawn to separate the y=0/y=1 threshold

  • We have specified the line separator for the plane. And if we plot x1,x2 like truth table. We ended up to must have both of them equals 1 to be in the green zone. And that's makes an "AND" table

  • Perceptron training is training our perceptron. It will be wastefull to just do perceptron by hand.
  • There's two options, perceptron rule or gradient descent.

  • Now what we do is some kind of other supervised learning, where we structure our data to have x data set and coressponding parameters, with the output as our data target.
  • Then we perform iteration in curly braces. Each weight  will be increasing every iterations.
  • The delta weight will be learning rate . (y-h(x) . x data set
  • here, the H(x) is the activation unit, and the threshold got subtittute to the left side as bias unit.
  • If we cand find linear separable from the data, then the neural networks will find it.
  • Otherwise it goes to infinite loop.