Command Palette
Search for a command to run...
Wiki
Machine Learning Glossary: Explore definitions and explanations of key AI and ML concepts
The Gated Recurrent Unit (GRU) is a variant of the Recurrent Neural Network (RNN) proposed by Cho et al. in 2014. The related paper is “Empirical Evaluation of Gate […]
AlexNet is a deep convolutional neural network (CNN) proposed by Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton in 2012 and used in the ImageNet image classification competition that year.
CART Decision Tree is a decision tree algorithm that can be used for classification and regression tasks.
Gradient Boosting is an ensemble learning algorithm that builds a strong prediction model by combining multiple weak prediction models (usually decision trees).
LeNet-5 is a pioneering work in the field of deep learning and convolutional neural networks, which laid the foundation for many key concepts in modern deep learning, such as convolutional layers, pooling layers, and fully connected layers.
Qualification questions focus on how to determine all the conditions or factors required for an action or event to be successfully performed in a changing environment.
ReAct proposes a general paradigm that combines advances in reasoning and action to enable language models to solve a variety of language reasoning and decision-making tasks.
Pre-training Once is a three-branch self-supervised training framework that introduces elastic student branches and randomly samples sub-networks for training in each pre-training step.
FlexAttention is a flexible attention mechanism designed to improve the efficiency of high-resolution vision-language models.
FlashAttention is an efficient and memory-friendly attention algorithm.
Causal Attention (CATT) is an innovative attention mechanism that improves the interpretability and performance of models by incorporating causal inference, especially in vision-language tasks. This mechanism was first proposed by researchers from Nanyang Technological University and Monash University in Australia in 20 […]
Thought Trees generalize the popular thought chaining approach to prompt language models and enable the exploration of coherent text units (thoughts) as intermediate steps in problem solving.
The MoMa architecture is a novel modality-aware mixture of experts (MoE) architecture designed for pre-training mixed-modality, early-fusion language models.
Multi-step Error Minimization (MEM) was published in 2024 by the Institute of Information Engineering of the Chinese Academy of Sciences, Nanyang Technological University, National University of Singapore, and Sun Yat-sen University in the paper “Multimodal Unlearnable E […]
The Geometric Langlands Conjecture is a geometric version of the Langlands program.
The Langlands Program is a highly influential research field in modern mathematics. It involves multiple branches of mathematics such as number theory, algebraic geometry and group representation theory, and attempts to reveal the profound connections between them.
An application-specific integrated circuit (ASIC) is an integrated circuit designed and manufactured according to specific user requirements and the needs of a specific electronic system.
Wall clock time is a term used to measure the running time of a program or process. It refers to the actual time taken from the start of program execution to the end, including all types of waiting and blocking time.
Pareto Front is a key concept in multi-objective optimization, which refers to a set of solutions that achieve the best trade-off between multiple objectives.
Stride is a term that is often used in image processing and convolutional neural networks (CNNs). In the context of image processing, stride refers to the number of steps that the operation window moves on the image when applying certain operations to the image, such as cropping, feature extraction, or filtering. For example, when cropping an image, […]
Dynamic Prompts is a prompting technique that allows prompts to be dynamically adjusted based on specific tasks or instances in natural language processing (NLP) and other artificial intelligence applications. This technique can significantly improve the performance and adaptability of models. Dyn […]
Simple Online and Realtime Tracking (SORT) is a practical multi-target tracking method that focuses on simple and efficient algorithms. It was presented by researchers from Queensland University of Technology and the University of Sydney at the 2016 IEEE International Conference on Image Processing. […]
Prioritized Experience Replay is a method for reinforcement learning that replays experiences at different frequencies based on their importance, thereby improving learning efficiency.
CoT technology decomposes complex problems into a series of step-by-step sub-problem answers, guiding the model to generate a detailed reasoning process, thereby improving the model's performance on complex tasks such as arithmetic reasoning, common sense reasoning, and symbolic reasoning.
The Gated Recurrent Unit (GRU) is a variant of the Recurrent Neural Network (RNN) proposed by Cho et al. in 2014. The related paper is “Empirical Evaluation of Gate […]
AlexNet is a deep convolutional neural network (CNN) proposed by Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton in 2012 and used in the ImageNet image classification competition that year.
CART Decision Tree is a decision tree algorithm that can be used for classification and regression tasks.
Gradient Boosting is an ensemble learning algorithm that builds a strong prediction model by combining multiple weak prediction models (usually decision trees).
LeNet-5 is a pioneering work in the field of deep learning and convolutional neural networks, which laid the foundation for many key concepts in modern deep learning, such as convolutional layers, pooling layers, and fully connected layers.
Qualification questions focus on how to determine all the conditions or factors required for an action or event to be successfully performed in a changing environment.
ReAct proposes a general paradigm that combines advances in reasoning and action to enable language models to solve a variety of language reasoning and decision-making tasks.
Pre-training Once is a three-branch self-supervised training framework that introduces elastic student branches and randomly samples sub-networks for training in each pre-training step.
FlexAttention is a flexible attention mechanism designed to improve the efficiency of high-resolution vision-language models.
FlashAttention is an efficient and memory-friendly attention algorithm.
Causal Attention (CATT) is an innovative attention mechanism that improves the interpretability and performance of models by incorporating causal inference, especially in vision-language tasks. This mechanism was first proposed by researchers from Nanyang Technological University and Monash University in Australia in 20 […]
Thought Trees generalize the popular thought chaining approach to prompt language models and enable the exploration of coherent text units (thoughts) as intermediate steps in problem solving.
The MoMa architecture is a novel modality-aware mixture of experts (MoE) architecture designed for pre-training mixed-modality, early-fusion language models.
Multi-step Error Minimization (MEM) was published in 2024 by the Institute of Information Engineering of the Chinese Academy of Sciences, Nanyang Technological University, National University of Singapore, and Sun Yat-sen University in the paper “Multimodal Unlearnable E […]
The Geometric Langlands Conjecture is a geometric version of the Langlands program.
The Langlands Program is a highly influential research field in modern mathematics. It involves multiple branches of mathematics such as number theory, algebraic geometry and group representation theory, and attempts to reveal the profound connections between them.
An application-specific integrated circuit (ASIC) is an integrated circuit designed and manufactured according to specific user requirements and the needs of a specific electronic system.
Wall clock time is a term used to measure the running time of a program or process. It refers to the actual time taken from the start of program execution to the end, including all types of waiting and blocking time.
Pareto Front is a key concept in multi-objective optimization, which refers to a set of solutions that achieve the best trade-off between multiple objectives.
Stride is a term that is often used in image processing and convolutional neural networks (CNNs). In the context of image processing, stride refers to the number of steps that the operation window moves on the image when applying certain operations to the image, such as cropping, feature extraction, or filtering. For example, when cropping an image, […]
Dynamic Prompts is a prompting technique that allows prompts to be dynamically adjusted based on specific tasks or instances in natural language processing (NLP) and other artificial intelligence applications. This technique can significantly improve the performance and adaptability of models. Dyn […]
Simple Online and Realtime Tracking (SORT) is a practical multi-target tracking method that focuses on simple and efficient algorithms. It was presented by researchers from Queensland University of Technology and the University of Sydney at the 2016 IEEE International Conference on Image Processing. […]
Prioritized Experience Replay is a method for reinforcement learning that replays experiences at different frequencies based on their importance, thereby improving learning efficiency.
CoT technology decomposes complex problems into a series of step-by-step sub-problem answers, guiding the model to generate a detailed reasoning process, thereby improving the model's performance on complex tasks such as arithmetic reasoning, common sense reasoning, and symbolic reasoning.