AI Sharing Circle

I didn't fill in the profile!
生成对抗网络(Generative Adversarial Network)是什么,一文看懂

What is Generative Adversarial Network (GAN) in one article?

Generative Adversarial Network (GAN) is a deep learning model proposed by Ian Goodfellow et al. in 2014. The framework implements generative modeling by training two neural networks against each other...
3mos ago
022.8K
自注意力(Self-Attention)是什么,一文看懂

Self-Attention (Self-Attention) is what, an article to read and understand

Self-Attention is a key mechanism in deep learning, originally proposed and widely used in the Transformer architecture. The core idea is to allow the model to simultaneously attend to all positions in the input sequence, and compute each position by weighted aggregation of...
3mos ago
030.7K
多任务学习(Multi-Task Learning)是什么,一文看懂

What is Multi-Task Learning (MTL) in one article?

Multi-Task Learning (MTL) is not an isolated algorithm, but an intelligent machine learning paradigm.
3mos ago
022.9K
扩散模型(Diffusion Model)是什么,一文看懂

Diffusion Model (Diffusion Model) what is it, an article to read and understand

Diffusion Model (Diffusion Model) is a generative model specialized for creating new data samples such as images, audio or text. The core of the model is inspired by the process of diffusion in physics, which simulates the natural diffusion of particles from a region of high concentration to a region of low concentration. In the machine...
4mos ago
029.3K
模型微调(Fine-tuning)是什么,一文看懂

What is Fine-tuning, in one article?

Model fine-tuning (Fine-tuning) is a specific implementation of transfer learning in machine learning. The core process is based on pre-trained models, which utilize large-scale datasets to learn generic patterns and develop extensive feature extraction capabilities. The fine-tuning phase then introduces task-specific datasets to ...
4mos ago
023.9K
注意力机制(Attention Mechanism)是什么,一文看懂

Attention Mechanism (Attention Mechanism) is what, an article to read and understand

Attention Mechanism (Attention Mechanism) is a computational technique that mimics human cognitive processes, initially applied in the field of machine translation, and later becoming an important part of deep learning.
4mos ago
029.3K
Transformer 架构(Transformer Architecture)是什么,一文看懂

What is the Transformer Architecture in one article?

The Transformer architecture is a deep learning model designed for processing sequence-to-sequence tasks such as machine translation or text summarization. The core innovation is the complete reliance on self-attention mechanisms, eschewing traditional loops or convolutional structures. Allowing the model to process all elements of a sequence in parallel, large...
4mos ago
028.1K
预训练模型(Pre-trained Model)是什么,一文看懂

What is Pre-trained Model (Pre-trained Model), an article to read and understand

Pre-trained Model is a fundamental and powerful technique in the field of Artificial Intelligence, representing machine learning models that are pre-trained on large-scale datasets. Models form a broad knowledge base by processing massive amounts of information and learning generalized patterns and features from the data...
4mos ago
025.5K
大语言模型(Large Language Model)是什么,一文看懂

What is the Large Language Model (LLM) in one article?

Large Language Model (LLM) is a deep learning system trained on massive text data, with the Transformer architecture at its core. The self-attention mechanism of this architecture can effectively capture long-distance dependencies in language. The model's "large ...
4mos ago
026.9K
长短期记忆网络(Long Short-Term Memory)是什么,一文看懂

What is Long Short-Term Memory (LSTM) network, an article to read and understand

Long Short-Term Memory (LSTM) is a recurrent neural network variant specialized in processing sequence data. In the field of artificial intelligence, sequence data is widely used in tasks such as time series prediction, natural language processing and speech recognition.
4mos ago
022.8K