Home > Technology peripherals > AI > AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference

AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference

WBOY
Release: 2023-05-12 10:43:05
forward
1082 people have browsed it

It only takes 12 seconds to use Stable Diffusion to generate an image using only the computing power of the mobile phone.

And it’s the kind that has completed 20 iterations.

AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference

You must know that current diffusion models basically exceed 1 billion parameters. If you want to quickly generate a picture, you must either rely on cloud computing or the local hardware must be powerful enough. .

As large model applications gradually become more popular, running large models on personal computers and mobile phones is likely to be a new trend in the future.

As a result, Google researchers have brought this new result, called Speed ​​is all you need: Accelerate the inference speed of large-scale diffusion models on devices through GPU optimization .

AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference

Three-step optimization acceleration

This method is optimized for Stable Diffusion, but it can also be adapted to other diffusion models. The task is to generate images from text.

Specific optimization can be divided into three parts:

  • Design a special kernel
  • Improve the efficiency of the Attention model
  • Winograd Convolution acceleration

First look at the specially designed kernel, which includes group normalization and GELU activation functions.

Group normalization is implemented throughout the UNet architecture. The working principle of this normalization is to divide the channels of feature mapping into smaller groups and normalize each group independently, so that Group normalization is less dependent on batch size and can adapt to a wider range of batch sizes and network architectures.

The researchers designed a unique kernel in the form of a GPU shader that can execute all kernels in a single GPU command without any intermediate tensors.

The GELU activation function contains a large number of numerical calculations, such as penalties, Gaussian error functions, etc.

A dedicated shader is used to integrate these numerical calculations and the accompanying division and multiplication operations, so that these calculations can be placed in a simple draw call.

Draw call is an operation in which the CPU calls the image programming interface and instructs the GPU to render.

Next, when it comes to improving the efficiency of the Attention model, the paper introduces two optimization methods.

One is the partial fusion of the softmax function.

In order to avoid performing the entire softmax calculation on the large matrix A, this study designed a GPU shader to calculate the L and S vectors to reduce calculations, ultimately resulting in a tensor of size N×2. Then the softmax calculation and matrix multiplication of matrix V are merged.

This method significantly reduces the memory footprint and overall latency of the intermediate program.

AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference

It should be emphasized that the parallelism of the computational mapping from A to L and S is limited because the number of elements in the result tensor is smaller than the number of elements in the input tensor A Much more.

In order to increase parallelism and further reduce latency, this study organized the elements in A into blocks and divided the reduction operations into multiple parts.

The calculation is then performed on each block and then reduced to the final result.

Using carefully designed threading and memory cache management, lower latency can be achieved in multiple parts using a single GPU command.

Another optimization method is FlashAttention.

This is the IO-aware precise attention algorithm that became popular last year. There are two specific acceleration technologies: incremental calculation in blocks, that is, tiling, and recalculating attention in backward pass to operate all attention Integrated into CUDA kernel.

Compared with standard Attention, this method can reduce HBM (high bandwidth memory) access and improve overall efficiency.

However, the FlashAttention core is very register-intensive, so the team uses this optimization method selectively.

They use FlashAttention on Adreno GPU and Apple GPU with attention matrix d=40, and use partial fusion softmax function in other cases.

The third part is Winograd convolution acceleration.

Its principle is simply to use more addition calculations to reduce multiplication calculations, thereby reducing the amount of calculations.

But the disadvantages are also obvious, which will bring more video memory consumption and numerical errors, especially when the tile is relatively large.

The backbone of Stable Diffusion relies heavily on 3×3 convolutional layers, especially in the image decoder, where 90% of the layers are composed of 3×3 convolutional layers.

After analysis, researchers found that when using 4×4 tiles, it is the best balance point between model calculation efficiency and video memory utilization.

AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference

Experimental results

In order to evaluate the improvement effect, the researchers first conducted a benchmark test on a mobile phone.

AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference

#The results show that after using the acceleration algorithm, the speed of image generation on both phones has been significantly improved.

Among them, the delay on Samsung S23 Ultra was reduced by 52.2%, and the delay on iPhone 14 Pro Max was reduced by 32.9%.

Generate a 512×512 pixel image from text end-to-end on Samsung S23 Ultra, with 20 iterations and taking less than 12 seconds.

Paper address: //m.sbmmt.com/link/ba825ea8a40c385c33407ebe566fa1bc


The above is the detailed content of AI completes painting on mobile phone within 12 seconds! Google proposes new method to accelerate diffusion model inference. For more information, please follow other related articles on the PHP Chinese website!

Related labels:
source:51cto.com
Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
Popular Tutorials
More>
Latest Downloads
More>
Web Effects
Website Source Code
Website Materials
Front End Template