Home > Technology peripherals > AI > Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

WBOY
Release: 2023-05-08 19:37:08
forward
1166 people have browsed it

Transformer, as an NLP pre-training model architecture, can effectively learn on large-scale unlabeled data. Research has proven that Transformer is the core architecture of NLP tasks since BERT.

Recent work has shown that the state space model (SSM) is a favorable competing architecture for modeling long-range sequences. SSM achieves state-of-the-art results on speech generation and Long Range Arena benchmarks, even outperforming the Transformer architecture. In addition to improving accuracy, the routing layer based on SSM will not exhibit quadratic complexity as the sequence length grows.

In this article, researchers from Cornell University, DeepMind and other institutions proposed Bidirectional Gated SSM (BiGS) for pre-training without attention. It mainly uses SSM routing is combined with an architecture based on multiplicative gating. The study found that SSM by itself performs poorly in pre-training for NLP, but when integrated into a multiplicative gated architecture, downstream accuracy improves.

Experiments show that BiGS is able to match the performance of BERT models when trained on the same data under controlled settings. With additional pretraining on longer instances, the model also maintains linear time when scaling the input sequence to 4096. Analysis shows that multiplicative gating is necessary and fixes some specific problems of SSM models on variable-length text inputs.

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

##Paper address: https://arxiv.org/pdf/2212.10544.pdf

Method Introduction

SSM connects the continuous input u (t) to the output y (t) through the following differential equation:

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

For discrete sequences, the SSM parameters are discretized, and the process can be approximated as:

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

##This equation Can be interpreted as a linear RNN, where x_k is a hidden state. y can also be calculated using convolutions:

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

An efficient way to use SSM in neural networks was shown by Gu et al., who developed the parameters A's approach, called HiPPO, resulted in a stable and efficient architecture called S4. This retains the ability of SSM to model long-term sequences while being more efficient than RNN training. Recently, researchers proposed a simplified diagonalized version of S4 that achieves similar results with a simpler approximation of the original parameters. At a high level, SSM-based routing provides an alternative to modeling sequences in neural networks without the attentional cost of secondary computations.

Pre-training model architecture

Can SSM replace attention in pre-training? To answer this question, the study considered two different architectures, the stacked architecture (STACK) and the multiplicative gated architecture (GATED) shown in Figure 1.

The stacked architecture with self-attention is equivalent to the BERT /transformer model, and the gated architecture is a bidirectional adaptation of the gated unit, which has also recently been used for unidirectional SSM. 2 sequence blocks (i.e., forward and backward SSM) with multiplicative gating are sandwiched in a feedforward layer. For a fair comparison, the size of the gated architecture is kept comparable to the stacked architecture.

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

Figure 1: Model variables. STACK is a standard transformer architecture, and GATED is based on gate control units. For the Routing component (dashed line), the study considers both bidirectional SSM (shown in the figure) and standard self-attention. Gated(X) represents element-wise multiplication.

Experimental results

Pre-training

Table 1 shows the main results of different pre-trained models on the GLUE benchmark. BiGS replicates BERT’s accuracy on token expansion. This result shows that SSM can replicate the accuracy of the pre-trained transformer model under such a computational budget. These results are significantly better than other non-attention-based pre-trained models. To achieve this accuracy, multiplicative gating is necessary. Without gating, the results of stacked SSM are significantly worse. To examine whether this advantage mainly comes from the use of gating, we trained an attention-based model using the GATE architecture; however, the results show that the model is actually less effective than BERT.

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

Table 1: GLUE results. (Top) Comparison of different architectures and routing under control settings. See Figure 2 for details. (Bottom) reported comparable results for other non-attention pretrained models based on CNN, LSTM and FNet.

Long-Form Task

Table 2 results show that SSM can be combined with Longformer EncoderDecoder (LED) was compared to BART, however, and the results showed that it performed just as well or even better in remote tasks. Compared to the other two methods, SSM has much less pre-training data. Even though SSM does not need to approximate at these lengths, the long form is still important.

Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.

Table 2: SCROLLS Encoder test results. The baseline models are both encoder-decoder models, one based on Longformer (LED) and the other based on BART. The input length is truncated.

#Please see the original paper for more information.

The above is the detailed content of Pre-training requires no attention, and scaling to 4096 tokens is no problem, which is comparable to BERT.. For more information, please follow other related articles on the PHP Chinese website!

Related labels:
source:51cto.com
Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
Popular Tutorials
More>
Latest Downloads
More>
Web Effects
Website Source Code
Website Materials
Front End Template