Skip to main content

Convolution Neural Network study material

 


Convolution Neural Network https://www.analyticsvidhya.com/blog/2021/05/convolutional-neural-networks-cnn/

Convolution Operation

In mathematics, convolution is a mathematical operation on two functions that produces a third function that expresses how the shape of one is modified by the other. The term convolution refers to both the result function and to the process of computing it. 

Max-pooling operation, questions based on these operations.

Benefits of Convolution operation

Convolutions in Neural Networks apply filters to extract features from actual data. A filter could be related to anything, for pictures of humans, one filter could be associated with seeing noses, another with eyes, and so on. Each feature extracted from input data will be in form of activation maps

Training process of convolution Neural Network.

These are the steps used to training the CNN (Convolutional Neural Network).

  1. Steps:
  2. Step 1: Upload Dataset.
  3. Step 2: The Input layer.
  4. Step 3: Convolutional layer.
  5. Step 4: Pooling layer.
  6. Step 5: Convolutional layer and Pooling Layer.
  7. Step 6: Dense layer.
  8. Step 7: Logit Layer.


Convolution Neural Net Architecture

AlexNet

VGG https://towardsdatascience.com/vgg-neural-networks-the-next-step-after-alexnet-3f91fa9ffe2c

VGG Neural Networks. While previous derivatives of AlexNet focused on smaller window sizes and strides in the first convolutional layer, VGG addresses another very important aspect of CNNs: depth. Let’s go over the architecture of VGG:

  • Input. VGG takes in a 224x224 pixel RGB image. For the ImageNet competition, the authors cropped out the center 224x224 patch in each image to keep the input image size consistent.
  • Convolutional Layers. The convolutional layers in VGG use a very small receptive field (3x3, the smallest possible size that still captures left/right and up/down). There are also 1x1 convolution filters which act as a linear transformation of the input, which is followed by a ReLU unit. The convolution stride is fixed to 1 pixel so that the spatial resolution is preserved after convolution.
  • Fully-Connected Layers. VGG has three fully-connected layers: the first two have 4096 channels each and the third has 1000 channels, 1 for each class.
  • Hidden Layers. All of VGG’s hidden layers use ReLU (a huge innovation from AlexNet that cut training time). VGG does not generally use Local Response Normalization (LRN), as LRN increases memory consumption and training time with no particular increase in accuracy.

ResNet.
Autoencoders- link

Architecture of Autoencoders,

Training of Autoencoders  

Different type  of Auto encoders -https://iq.openÍgenus.org/types-of-autoencoder/

  1. Under complete
  2. Sparse
  3. Denoising Autoencoders.

Principal Component Analysis -Í

Principal Component Analysis (PCA) is an unsupervised, non-parametric statistical technique primarily used for dimensionality reduction in machine learning.

Assumptions

PCA is based on the Pearson correlation coefficient framework and inherits similar assumptions.

  1. Sample size: Minimum of 150 observations and ideally a 5:1 ratio of observation to features (Pallant, 2010)
  2. Correlations: The feature set is correlated, so the reduced feature set effectively represents the original data space.
  3. Linearity: All variables exhibit a constant multivariate normal relationship, and principal components are a linear combination of the original features.
  4. Outliers: No significant outliers in the data as these can have a disproportionate influence on the results.

PCA and its comparison with autoencoders (Lecture 29 posted on November 24)Comparison

  1. PCA is essentially a linear transformation but Auto-encoders are capable of modelling complex non linear functions.
  2. PCA features are totally linearly uncorrelated with each other since features are projections onto the orthogonal basis. But autoencoded features might have correlations since they are just trained for accurate reconstruction.
  3. PCA is faster and computationally cheaper than autoencoders.
  4. A single layered autoencoder with a linear activation function is very similar to PCA.
  5. Autoencoder is prone to overfitting due to high number of parameters. (though regularization and careful design can avoid this)


Optimization in Deep Learning https://medium.com/@minions.k/optimization-techniques-popularly-used-in-deep-learning-3c219ec8e0cc

Comments

Popular posts from this blog

Create a socket for HTTP for web page upload and download

Create a socket for HTTP for web page upload and download. Aim: To write a java program for socket for HTTP for web page upload and download . Algorithm 1.Start the program. 2.Get the frame size from the user 3.To create the frame based on the user request. 4.To send frames to server from the client side. 5.If your frames reach the server it will send ACK signal to client otherwise it will send NACK signal to client. 6.Stop the program Program : Client import javax.swing.*; import java.net.*; import java.awt.image.*; import javax.imageio.*; import java.io.*; import java.awt.image.BufferedImage; import java.io.ByteArrayOutputStream; import java.io.File; import java.io.IOException; import javax.imageio.ImageIO; public class Client{ public static void main(String args[]) throws Exception{ Socket soc; BufferedImage img = null; soc=new Socket("localhost",4000); System.out.println("Client is running. ");  try { System.out.println("Reading image from disk. "); im...

Write a code simulating ARP /RARP protocols

   Write a code simulating ARP /RARP protocols . Aim:        To write a java program for simulating ARP/RARP protocols ALGORITHM: server 1. Create a server socket and bind it to port. 2. Listen for new connection and when a connection arrives, accept it. 3. Send server ‟ s date and time to the client. 4. Read client ‟ s IP address sent by the client. 5. Display the client details. 6. Repeat steps 2-5 until the server is terminated. 7. Close all streams. 8. Close the server socket. 9. Stop. Client 1. Create a client socket and connect it to the server ‟ s port number. 2. Retrieve its own IP address using built-in function. 3. Send its address to the server. 4. Display the date & time sent by the server. 5. Close the input and output streams. 6. Close the client socket. 7. Stop. Program Program for Address Resolutuion Protocol (ARP) using TCP Client: import java.io.*; import java.net.*; impor...

Write program to find ε – closure of all states of any given NFA with ε transition.

 Write program to find ε – closure of all states of any given NFA with ε transition. Agenda 1.Program 2.Input/Output 1.Program #include <stdio.h> #include <string.h> char  result[ 20 ][ 20 ], copy[ 3 ], states[ 20 ][ 20 ]; void  add_state( char  a[ 3 ],  int  i) {   strcpy(result[i], a); } void  display( int  n) {    int  k =  0 ;   printf( "nnn Epsilon closure of %s = { " , copy);    while  (k < n) {     printf( " %s" , result[k]);     k++;   }   printf( " } nnn" ); } int  main() {   FILE * INPUT;   INPUT = fopen( "input.dat" ,  "r" );    char  state[ 3 ];    int  end, i =  0 , n, k =  0 ;  ...