Home > Technology peripherals > AI > DIY digital content production using AI technology

DIY digital content production using AI technology

WBOY
Release: 2023-06-15 19:03:11
forward
1561 people have browsed it

DIY digital content production using AI technology

Background

The stunning performance of large models represented by chatgpt this year has completely ignited the field of AICG. Various gpt and AI mapping products are springing up like mushrooms after a rain. Behind every successful product are exquisite algorithms. This article will give you a detailed introduction to the process and code of how to use a mobile phone to take several photos of the same scene, then synthesize new perspectives and generate videos. The technology used in this article is NeRF (Neural Radiance Fields), which is a 3D reconstruction method based on deep learning that has emerged since 2020. It can generate high-quality images by learning the light transmission and radiation transfer of the scene. Scene rendering images and 3D models. Regarding its principles and literature, I have a reference list at the end for everyone to learn from. This article mainly introduces it from a new perspective of code usage and environment construction.

Environment setup

environment.yml modification

The hardware environment used in this article is GPU RTX3090, and the operating system is Windows 10. The software used is open source NeRF implementation (https://github.com/cjw531/nerf_tf2). Since RTX 3090 requires the support of CUDA 11.0 and above, and TensorFlow-gpu requires support of 2.4.0 and above, we did not choose the official https://github.com/bmild/nerf because the bmild environment uses tensorflow. -gpu==1.15, the version is too old. There will be the following problem when running https://github.com/bmild/nerf/issues/174#issue-1553410900. I also replied in this tt that I need to upgrade to 2.8. But even if you use https://github.com/cjw531/nerf_tf2, its environment is somewhat problematic. First of all, because it is connected to the foreign conda channel, the speed is very slow. Secondly, its environment uses tensorflow==2.8 and does not specify the version of tensorflow-gpu. for these two questions. We have modified environment.yml.

# To run: conda env create -f environment.ymlname: nerf_tf2channels:- https://mirrors.tuna.tsinghua.edu.cn/anaconda/cloud/conda-forge/- conda-forgedependencies:- python=3.7- pip- cudatoolkit=11.0- cudnn=8.0- numpy- matplotlib- imageio- imageio-ffmpeg- configargparse- ipywidgets- tqdm- pip:- tensorflow==2.8- tensorflow-gpu==2.8- protobuf==3.19.0- -i https://pypi.tuna.tsinghua.edu.cn/simple
Copy after login

Start the conda environment

Open cmd and enter the following command.

conda env create -f environment.yml
Copy after login

Add nerf_tf2 to jupyter, so that jupyter can easily view the running results of the system.

// 安装ipykernelconda install ipykernel
Copy after login
//是该conda环境在jupyter中显示python -m ipykernel install --user --name 环境名称 --python -m ipykernel install --user --name 环境名称 --display-name "jupyter中显示名称"display-name "jupyter中显示名称"
Copy after login
//切换到项目目录cd 到项目目录//激活conda环境activate nerf_tf2//在cmd启动jupyterjupyter notebook
Copy after login

Now the conda environment and jupyter are ready.

Data preparation

  1. Download and install colmap, my environment is windows (https://demuc.de/colmap/#download)
  2. Use https: //github.com/fyusion/llff provides imgs2poses.py to obtain the internal and external parameters of the pictures taken by your own camera. For example, we took 10 pictures, and the directory location where they are placed is very particular, D:/LanJing/AI /LLFF/data/images, which means it must be placed under the images subdirectory. The parameter you passed in is python imgs2poses.py D:/LanJing/AI/LLFF/data. Because the images_path in its code is written like this (https://github.com/Fyusion/LLFF/blob/master/llff/poses/colmap_wrapper.py#L28)

DIY digital content production using AI technology

Sample pictures taken by mobile phones

feature_extractor_args = ['colmap', 'feature_extractor','--database_path', os.path.join(basedir, 'database.db'),'--image_path', os.path.join(basedir, 'images'),'--ImageReader.single_camera', '1',# '--SiftExtraction.use_gpu', '0',]
Copy after login

DIY digital content production using AI technology

python imgs2poses.py

After running the imgs2poses.py file , generated the sparse directory, colmap_out.txt, database.db, poses_bounds.npy, then we created a new directory data/nerf_llff_data/ll under the nerf_tf2 project, and copied the above sparse directory and poses_bounds.npy to this directory. Finally, we configure a new file config_ll.txt. At this point our data preparation work is completed.

expname = ll_testbasedir = ./logsdatadir = ./data/nerf_llff_data/lldataset_type = llfffactor = 8llffhold = 8N_rand = 1024N_samples = 64N_importance = 64use_viewdirs = Trueraw_noise_std = 1e0
Copy after login

Training

Migrate open source software to the windows platform.

Since this open source software mainly supports mac and linux, it cannot run on windows and requires modification of load_llff.py.

DIY digital content production using AI technology

load_llff code migration

Run 300,000 times of batch training.

activate nerf_tf2python run_nerf.py --config config_ll.txt
Copy after login

Test

DIY digital content production using AI technology

##Render_demo running

Effect

Because we use a few input images, only 10 Zhang, so the effect is not very good, but the overall process is the same. Tips: Half of the official codes use 30 or even 100 pictures.

Our results


DIY digital content production using AI technology

Rendering from a new perspective

Official effect

DIY digital content production using AI technology

fernOfficial synthesis of new perspective effect

References

https://zhuanlan.zhihu.com/p/554093703.

https://arxiv.org/pdf/2003.08934.pdf.

https://zhuanlan.zhihu.com/p/593204605.

https://inst.eecs.berkeley.edu/~cs194-26/fa22/Lectures/nerf_lecture1.pdf.

The above is the detailed content of DIY digital content production using AI technology. For more information, please follow other related articles on the PHP Chinese website!

Related labels:
source:51cto.com
Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
Popular Tutorials
More>
Latest Downloads
More>
Web Effects
Website Source Code
Website Materials
Front End Template