top of page

Neural Networks

  • 작성자 사진: Shin Yoonah, Yoonah
    Shin Yoonah, Yoonah
  • 2022년 8월 4일
  • 2분 분량

최종 수정일: 2022년 8월 8일


Features and targets: Example

It is helpful to view the shape y as a decision function of x

--> non-linearly separable dataset

Use one dimension for simplicity


In the context of neural networks, its helpful to think of the classification problem as a decision function

We can represent the box, this box function is an example of a decision function


A neural network will approximate the function using learnable parameters

+ view the problem as trying to approximate the box function using logistic regression

anything in blue region y = one i.e dog

anything in red region y = zero i.e cat


If this was our cat-dog dataset, we cannot use a straight line to separate the data


Neural Networks

This line can be used to linearly separate same of the data, but some of the data is on the wrong side of the line

--> use the node to represent the line and the edges to represent the input x and output z


If we apply the logistic function, in the context of neural networks this is called the activation function


Represent the sigmoid function with the following node taking the input z from the linear function and producing an output

technically "A" is a function of z and x


Function "A" the activation function and the output of "A" is called the activation

this line looks like it can be use to separate the data


What happens when we apply the sigmoid function?

- after applying the sigmoid or activation function, we get an incorrect result for some of the samples


Consider the following sigmoid functions, we call them "A sub script one" and "A sub script two"

if we subtract the second, sigmoid function from the first sigmoid function we get something similar to the decision function

+ also apply the following operations with a linear function

i.e) just subtract the second activations from the first activation function

----> values will be learnable parameters


If we apply a threshold setting every value less than 0.5 to zero and greater than 0.5 to one

We get the exact function; we are trying to approximate

Then, classify the data, we obtain the parameters via gradient descent

--> use the graph to represent the process, we apply two linear functions to x and we get two outputs

--> to each linear function we apply a sigmoid


Then, we apply a second linear function to the outputs of the sigmoid

Usually, apply another function to the output of this linear function then apply a threshold


This diagram is used to represent a two-layer neural network, we have a hidden layer


Each linear function and activation is known as an artificial neuron

--> the hidden layer has two artificial neurons


The output layer has one artificial neuron, as it has two inputs; the input dimension for this neuron is two

The outputs of the activation function is a 2D plane that looks like this

--> it turns out that we can split the point using the following plane


In the same way, we can add more dimensions to the input, notice that there are a lot more weights between the input layer and hidden layer


These type of Neural Networks are called FeedForward Neural Networks or fully connected networks



More dimensions, more Neurals are required !!


Copyright Coursera All rights reserved

Comments


bottom of page