AI Sharing Circle

I didn't fill in the profile!
交叉验证(Cross-Validation)是什么,一文看懂

Cross-Validation (Cross-Validation) is what, an article to see and understand

Cross-Validation is a core method for assessing the generalization ability of a model in machine learning.The basic idea is to split the original data into a training set and a test set, and to obtain more reliable performance estimates by rotating the use of different data subsets for training and validation. This approach simulates ...
4dys ago
01.9K
随机森林(Random Forest)是什么,一文看懂

What is Random Forest (Random Forest), an article to read and understand

Random Forest (Random Forest) is an integrated learning algorithm that accomplishes machine learning tasks by constructing multiple decision trees and synthesizing their predictions. The algorithm is based on the Bootstrap aggregation idea, where multiple subsets of samples are randomly drawn from the original dataset with putback for each tree...
4dys ago
01.4K
损失函数(Loss Function)是什么,一文看懂

Loss Function (Loss Function) is what, an article to read and understand

Loss Function (Loss Function) is a core concept in Machine Learning, undertaking the important task of quantifying the prediction error of a model. This function mathematically measures the degree of difference between the model's predicted value and the true value, providing a clear directional guide for model optimization.
4dys ago
01.8K
超参数(Hyperparameter)是什么,一文看懂

Hyperparameter (Hyperparameter) is what, an article to see and understand

In machine learning, a hyperparameter is a configuration option that is preset manually before model training begins, rather than learned from data. The central role is to control the learning process itself, as if setting a set of operating rules for the algorithm. For example, the learning...
2wks ago
04K
决策树(Decision Tree)是什么,一文看懂

Decision Tree (Decision Tree) is what, an article to see and understand

Decision Tree (DT) is a tree-shaped predictive model that simulates the human decision-making process, classifying or predicting data through a series of rules. Each internal node represents a feature test, branches correspond to test results, and leaf nodes store the final decision. This algorithm uses a divide-and-conquer strategy...
2wks ago
04.6K
梯度下降(Gradient Descent)是什么,一文看懂

What is Gradient Descent (Gradient Descent), an article to read and understand

Gradient Descent is the core optimization algorithm for solving function minimization. The algorithm determines the direction of descent by calculating the gradient of the function (the vector consisting of the partial derivatives of each), and iteratively updating the parameters according to the rule θ = θ - η - ∇J(θ).
2wks ago
04K
逻辑回归(Logistic Regression)是什么,一文看懂

What is Logistic Regression (Logistic Regression), an article to read and understand

Logistic Regression is a statistical learning method used to solve binary classification problems. The central goal is to predict the probability that a sample belongs to a particular category based on input features. The model maps the linear output to between 0 and 1 by linearly combining the eigenvalues using an S-shaped function...
2wks ago
05.4K
正则化(Regularization)是什么,一文看懂

Regularization (Regularization) is what, an article to see and understand

Regularization is a core technique in machine learning and statistics to prevent model overfitting. Regularization controls the degree of fitting by adding a penalty term to the objective function that is related to the complexity of the model. Common forms include L1 and L2 regularization: the L1 produces sparse solutions and applies...
2wks ago
04.8K
生成对抗网络(Generative Adversarial Network)是什么,一文看懂

What is Generative Adversarial Network (GAN) in one article?

Generative Adversarial Network (GAN) is a deep learning model proposed by Ian Goodfellow et al. in 2014. The framework implements generative modeling by training two neural networks against each other...
2wks ago
04.8K
自注意力(Self-Attention)是什么,一文看懂

Self-Attention (Self-Attention) is what, an article to read and understand

Self-Attention is a key mechanism in deep learning, originally proposed and widely used in the Transformer architecture. The core idea is to allow the model to simultaneously attend to all positions in the input sequence, and compute each position by weighted aggregation of...
3wks ago
08K