跳转至

神经网络

约 358 个字 预计阅读时间 1 分钟

Fundamentals: This includes understanding the structure of a neural network, such as layers, weights, biases, and activation functions (sigmoid, tanh, ReLU, etc.)

Training and Optimization: Familiarize yourself with backpropagation and different types of loss functions, like Mean Squared Error (MSE) and Cross-Entropy. Understand various optimization algorithms like Gradient Descent, Stochastic Gradient Descent, RMSprop, and Adam.

Overfitting: Understand the concept of overfitting (where a model performs well on training data but poorly on unseen data) and learn various regularization techniques (dropout, L1/L2 regularization, early stopping, data augmentation) to prevent it.

Implement a Multilayer Perceptron (MLP): Build an MLP, also known as a fully connected network, using PyTorch.

1.神经元模型

M-P神经元模型

激活函数

学习率

权重初始化:常数初始化,随机分布初始化,Xavier初始化,He初始化

2.感知机与多层网络(MLP)

多层前馈神经网络(Multi-Layer FFN)

前馈并不意味着网络信号不能向后传,而是指网络拓扑结构上不存在环或回路。

3.反向传播算法(BP)

LMS(Least Mean Square)算法(BP前身):将LMS推广到由非线性可微神经元组成的多层前馈网络,就得到BP算法。

梯度下降GD

随机梯度下降SGD

解决过拟合:早停,正则化(dropout,标签平滑,权重衰减)

4.全局最小与局部最小

跳出局部最小:启发式算法(模拟退火,遗传算法等)

5.其他神经网络

RBF网络

ART网络

SOM网络

级联相关网络

Elman网络

Boltzman机

深度学习(pre-training,fine-tuning)

深度信念网络DBN

权值共享CNN

6.深度学习