Wiki
We have compiled hundreds of related entries to help you understand "artificial intelligence"
FlexAttention is a flexible attention mechanism designed to improve the efficiency of high-resolution vision-language models.
FlashAttention is an efficient and memory-friendly attention algorithm.
Causal Attention (CATT) is an innovative attention mechanism that improves the interpretability and performance of models by incorporating causal inference, especially in vision-language tasks. This mechanism was first proposed by researchers from Nanyang Technological University and Monash University in Australia in 20 […]
Thought Trees generalize the popular thought chaining approach to prompt language models and enable the exploration of coherent text units (thoughts) as intermediate steps in problem solving.
The MoMa architecture is a novel modality-aware mixture of experts (MoE) architecture designed for pre-training mixed-modality, early-fusion language models.
Multi-step Error Minimization (MEM) was published in 2024 by the Institute of Information Engineering of the Chinese Academy of Sciences, Nanyang Technological University, National University of Singapore, and Sun Yat-sen University in the paper “Multimodal Unlearnable E […]
The Geometric Langlands Conjecture is a geometric version of the Langlands program.
The Langlands Program is a highly influential research field in modern mathematics. It involves multiple branches of mathematics such as number theory, algebraic geometry and group representation theory, and attempts to reveal the profound connections between them.
An application-specific integrated circuit (ASIC) is an integrated circuit designed and manufactured according to specific user requirements and the needs of a specific electronic system.
Wall clock time is a term used to measure the running time of a program or process. It refers to the actual time taken from the start of program execution to the end, including all types of waiting and blocking time.
Pareto Front is a key concept in multi-objective optimization, which refers to a set of solutions that achieve the best trade-off between multiple objectives.
Stride is a term that is often used in image processing and convolutional neural networks (CNNs). In the context of image processing, stride refers to the number of steps that the operation window moves on the image when applying certain operations to the image, such as cropping, feature extraction, or filtering. For example, when cropping an image, […]
Dynamic Prompts is a prompting technique that allows prompts to be dynamically adjusted based on specific tasks or instances in natural language processing (NLP) and other artificial intelligence applications. This technique can significantly improve the performance and adaptability of models. Dyn […]
Simple Online and Realtime Tracking (SORT) is a practical multi-target tracking method that focuses on simple and efficient algorithms. It was presented by researchers from Queensland University of Technology and the University of Sydney at the 2016 IEEE International Conference on Image Processing. […]
Prioritized Experience Replay is a method for reinforcement learning that replays experiences at different frequencies based on their importance, thereby improving learning efficiency.
CoT technology decomposes complex problems into a series of step-by-step sub-problem answers, guiding the model to generate a detailed reasoning process, thereby improving the model's performance on complex tasks such as arithmetic reasoning, common sense reasoning, and symbolic reasoning.
Parameter Efficient Fine-tuning (PERT) is a fine-tuning method for large pre-trained models that reduces computational and storage costs by fine-tuning only a small subset of model parameters while maintaining performance comparable to full-parameter fine-tuning.
In the field of artificial intelligence, a "world model" is a model that can characterize the state of the environment or the world and predict the transition between states. This model enables the agent to learn in a simulated environment and transfer the learned strategy to the real world, thereby improving learning efficiency and reducing risks. Jürgen S […]
Multimodal Contrastive Learning with Joint Example Selection (JEST) aims to address the high energy consumption problem during training of large language models such as ChatGPT.
Full Parameter Tuning is a model optimization technique in deep learning, especially used in the context of transfer learning or domain adaptation. It involves fine-tuning all parameters of a pre-trained model to adapt it to a specific task or dataset.
Occupancy grid network plays an important role in autonomous driving perception tasks. It is a network model that emphasizes geometry over semantics. It can assist autonomous driving systems in better perceiving free space and is a key technology for improving perception capabilities and forming a closed loop.
The core idea of realignment during decoding is to dynamically adjust the alignment of the model during the decoding process without retraining the model, thus saving computing resources and improving research efficiency.
3D Gaussian splatting is an advanced computer graphics technique that has important applications in point cloud rendering, volume data visualization, and volume reconstruction. This technique achieves higher quality rendering by converting discrete data points or voxels into continuous surface or volume representations.
Shadow mode testing is a testing method used in the field of autonomous driving. It is mainly used to verify and evaluate autonomous driving algorithms in real traffic environments while ensuring that it does not interfere with the driver and surrounding traffic.
FlexAttention is a flexible attention mechanism designed to improve the efficiency of high-resolution vision-language models.
FlashAttention is an efficient and memory-friendly attention algorithm.
Causal Attention (CATT) is an innovative attention mechanism that improves the interpretability and performance of models by incorporating causal inference, especially in vision-language tasks. This mechanism was first proposed by researchers from Nanyang Technological University and Monash University in Australia in 20 […]
Thought Trees generalize the popular thought chaining approach to prompt language models and enable the exploration of coherent text units (thoughts) as intermediate steps in problem solving.
The MoMa architecture is a novel modality-aware mixture of experts (MoE) architecture designed for pre-training mixed-modality, early-fusion language models.
Multi-step Error Minimization (MEM) was published in 2024 by the Institute of Information Engineering of the Chinese Academy of Sciences, Nanyang Technological University, National University of Singapore, and Sun Yat-sen University in the paper “Multimodal Unlearnable E […]
The Geometric Langlands Conjecture is a geometric version of the Langlands program.
The Langlands Program is a highly influential research field in modern mathematics. It involves multiple branches of mathematics such as number theory, algebraic geometry and group representation theory, and attempts to reveal the profound connections between them.
An application-specific integrated circuit (ASIC) is an integrated circuit designed and manufactured according to specific user requirements and the needs of a specific electronic system.
Wall clock time is a term used to measure the running time of a program or process. It refers to the actual time taken from the start of program execution to the end, including all types of waiting and blocking time.
Pareto Front is a key concept in multi-objective optimization, which refers to a set of solutions that achieve the best trade-off between multiple objectives.
Stride is a term that is often used in image processing and convolutional neural networks (CNNs). In the context of image processing, stride refers to the number of steps that the operation window moves on the image when applying certain operations to the image, such as cropping, feature extraction, or filtering. For example, when cropping an image, […]
Dynamic Prompts is a prompting technique that allows prompts to be dynamically adjusted based on specific tasks or instances in natural language processing (NLP) and other artificial intelligence applications. This technique can significantly improve the performance and adaptability of models. Dyn […]
Simple Online and Realtime Tracking (SORT) is a practical multi-target tracking method that focuses on simple and efficient algorithms. It was presented by researchers from Queensland University of Technology and the University of Sydney at the 2016 IEEE International Conference on Image Processing. […]
Prioritized Experience Replay is a method for reinforcement learning that replays experiences at different frequencies based on their importance, thereby improving learning efficiency.
CoT technology decomposes complex problems into a series of step-by-step sub-problem answers, guiding the model to generate a detailed reasoning process, thereby improving the model's performance on complex tasks such as arithmetic reasoning, common sense reasoning, and symbolic reasoning.
Parameter Efficient Fine-tuning (PERT) is a fine-tuning method for large pre-trained models that reduces computational and storage costs by fine-tuning only a small subset of model parameters while maintaining performance comparable to full-parameter fine-tuning.
In the field of artificial intelligence, a "world model" is a model that can characterize the state of the environment or the world and predict the transition between states. This model enables the agent to learn in a simulated environment and transfer the learned strategy to the real world, thereby improving learning efficiency and reducing risks. Jürgen S […]
Multimodal Contrastive Learning with Joint Example Selection (JEST) aims to address the high energy consumption problem during training of large language models such as ChatGPT.
Full Parameter Tuning is a model optimization technique in deep learning, especially used in the context of transfer learning or domain adaptation. It involves fine-tuning all parameters of a pre-trained model to adapt it to a specific task or dataset.
Occupancy grid network plays an important role in autonomous driving perception tasks. It is a network model that emphasizes geometry over semantics. It can assist autonomous driving systems in better perceiving free space and is a key technology for improving perception capabilities and forming a closed loop.
The core idea of realignment during decoding is to dynamically adjust the alignment of the model during the decoding process without retraining the model, thus saving computing resources and improving research efficiency.
3D Gaussian splatting is an advanced computer graphics technique that has important applications in point cloud rendering, volume data visualization, and volume reconstruction. This technique achieves higher quality rendering by converting discrete data points or voxels into continuous surface or volume representations.
Shadow mode testing is a testing method used in the field of autonomous driving. It is mainly used to verify and evaluate autonomous driving algorithms in real traffic environments while ensuring that it does not interfere with the driver and surrounding traffic.