
目录
1.Natural Language Processing,NLP(自然语言处理)
7.Stochastic Gradient Descent, SGD (随机梯度下降)
8.Mini-batch Gradient Descent (小批量梯度下降)
Natural Language Processing (NLP) is the field of artificial intelligence that enables computers to understand, interpret, and generate human language. It bridges the gap between human communication and computer understanding, making it possible for machines to perform tasks like translation, sentiment analysis, and topic classification.
This is a subset of artificial intelligence that involves algorithms and statistical models that enable computers to perform specific tasks without using explicit instructions. Instead, they rely on patterns and inference derived from data. The goal of ML is to enable computers to learn from and make predictions or decisions based on data.
Inspired by the human brain, neural networks are a set of algorithms, modeled loosely after the human brain, that are designed to recognize patterns. They interpret sensory data through a kind of machine perception, labeling, or clustering of raw input. These networks can adapt to changing input, meaning they generate the best possible result without needing to redesign the output criteria.
Deep Learning is a subset of machine learning in artificial intelligence that structures algorithms in layers to create an "artificial neural network" that can learn and make intelligent decisions on its own. This technology powers advanced applications such as voice recognition and image analysis.
A Loss Function in machine learning measures the difference between the actual output and the predicted output of the model. It quantifies how well the prediction model performs by assigning a cost to prediction errors.
Gradient Descent is an optimization algorithm used to minimize some function by iteratively moving towards the minimum value of the function. It is commonly used in machine learning to find the best parameters for a model.
Stochastic Gradient Descent (SGD) is a variation of the gradient descent algorithm that updates the model's parameters using only a single sample or a small batch of samples, which makes the process faster and can help avoid local minima.
Mini-batch Gradient Descent is a balance between the full batch gradient descent and stochastic gradient descent. It updates the model's parameters using a subset of the training data, rather than the full dataset or individual samples, optimizing computational efficiency.
Backpropagation is a method used in artificial neural networks to calculate the gradient of the loss function with respect to each weight by the chain rule, effectively allowing for the optimization of weights to minimize loss.
Overfitting occurs when a machine learning model learns the detail and noise in the training data to the extent that it negatively impacts the model's performance on new data. This means the model is too complex, capturing noise as if it were a significant pattern, leading to poor generalization on unseen data.
以上
君子坐而论道,少年起而行之,共勉
