NERFs (No, not that kind) - Computerphile

Computerphile
21 Dec 202313:35

Summary

TLDRThis video explores Neural Radiance Fields (NeRF), a cutting-edge AI technique for generating 3D reconstructions from 2D images. The host, with the help of a PhD student, explains how NeRF uses RGB images to create detailed 3D scenes, contrasting it with traditional methods like point clouds and meshes. They discuss the process of training a neural network to 'see' and render 3D objects, emphasizing the importance of multiple viewpoints for accurate reconstruction. The video also touches on the limitations, such as the need for a large dataset and the challenges of capturing dynamic scenes, while highlighting NeRF's potential in revolutionizing 3D rendering and asset creation.

Takeaways

  • ๐ŸŒ The script introduces Neural Radiance Fields (NeRF), a technology for generating new views of scenes from a series of RGB images.
  • ๐ŸŽ“ Lewis, a PhD student, explains NeRF, which he works with as part of his PhD, to the audience.
  • ๐ŸŒณ NeRF reconstructs 3D scenes with high detail from simple RGB images, unlike traditional 3D reconstruction methods like point clouds or meshes.
  • ๐Ÿ–ผ๏ธ Rendering with NeRF involves shooting rays into the environment and querying a neural network for color and density at various points along the ray.
  • ๐Ÿ” NeRF's 3D representation allows it to understand the environment, unlike other generative models like diffusion which lack a 3D context.
  • ๐Ÿ“ธ For effective training, NeRF requires a substantial number of images, ideally 250 or more, to achieve a good reconstruction.
  • ๐Ÿ“น The script discusses the challenges of capturing a dynamic scene like a Christmas tree with NeRF, including issues like motion blur and overfitting.
  • ๐Ÿšซ Changes in the scene or photobombing can lead to artifacts like 'floaters' where the model struggles to place certain pixels accurately.
  • ๐Ÿ“ˆ Despite its limitations, NeRF can be used to quickly generate 3D representations from images, which can then be converted into meshes for use in games or other applications.
  • ๐Ÿ”ฎ The future of 3D rendering may involve technologies like NeRF, but there are also emerging rivals like GAN splatting that show promise.

Q & A

  • What are Neural Radiance Fields (NeRF)?

    -Neural Radiance Fields, or NeRF, is a method used in AI literature for generating new views of scenes from a series of RGB images. It reconstructs 3D scenes using a neural network, unlike traditional 3D reconstruction methods such as point clouds, voxel grids, or meshes.

  • How does NeRF differ from traditional 3D reconstruction methods?

    -Traditional 3D reconstruction methods like point clouds, voxel grids, and meshes are discrete and can become complicated with many points or faces. NeRF, on the other hand, reconstructs scenes in high-quality detail using only RGB images, encoding the 3D scene within the parameters and weights of a neural network.

  • How does the rendering process in NeRF work?

    -In NeRF, rendering involves shooting rays into the environment and querying a series of points along these rays. The neural network then provides color and density for each point, which is different from traditional rendering methods that might use rasterization or ray tracing.

  • What is the significance of density in the context of NeRF?

    -Density in NeRF indicates whether a point along a ray is within an object or in empty space. A density of zero corresponds to empty space, while a density of one or more signifies that the point is inside an object, which is crucial for 3D representation.

  • Why is capturing multiple images from different angles important for NeRF?

    -Capturing multiple images from various angles is essential for NeRF to properly reconstruct a 3D scene. It allows the neural network to understand the environment from different perspectives, which is necessary for accurate rendering when the camera viewpoint changes.

  • What are the downsides of using NeRF for 3D reconstruction?

    -One downside of NeRF is the need for a large number of images for good reconstruction quality. Additionally, the method can struggle with scenes where objects change position or are not fully captured, leading to artifacts or noise in the rendered image.

  • How many images are typically needed for a good NeRF reconstruction?

    -For a good reconstruction, at least 250 images are recommended. However, the quality can vary depending on the technique used and the complexity of the scene.

  • What is the role of the camera positions in NeRF?

    -Camera positions are crucial in NeRF as they provide the spatial context for the images captured. They help the neural network understand where each image was taken from, which is necessary for reconstructing the 3D scene accurately.

  • Can NeRF be used to create 3D assets for games or other applications?

    -Yes, NeRF can be used to extract the 3D volume of objects, which can then be converted into meshes. This can speed up the creation of 3D assets for use in games or other applications, as it requires fewer manual design and modeling steps.

  • What is the current state of NeRF in terms of real-time rendering?

    -NeRF is not suitable for real-time rendering due to its computational intensity. It can take several seconds to render a single image, making it more suitable for pre-rendered scenes rather than real-time applications.

  • Are there any alternatives to NeRF for 3D reconstruction?

    -Yes, there are alternatives such as GAN splatting, which is a newer method also providing impressive results for 3D reconstruction and is worth exploring for potential advantages over NeRF.

Outlines

00:00

๐ŸŒ Introduction to Neural Radiance Fields

The script introduces Neural Radiance Fields (NeRFs), a technology that generates new views of scenes from RGB images using a neural network. Unlike traditional 3D reconstruction methods like point clouds, voxel grids, or meshes, NeRFs offer a continuous representation of a scene. The discussion involves a PhD student, Lewis, who explains that NeRFs use a series of RGB images to reconstruct a 3D scene, which is then used to render images from novel viewpoints. The process is akin to ray tracing but uses a neural network to determine the color and density at various points along a ray, encoding the 3D scene within the network's parameters.

05:01

๐ŸŽฅ Practical Application and Limitations of NeRFs

This section delves into the practical application of NeRFs, discussing the process of capturing images and training the neural network to reconstruct a 3D scene. It highlights the importance of multiple viewpoints for accurate reconstruction and the challenges of capturing a complete scene with limited camera movement. The script also touches on the downsides of NeRFs, such as the need for a substantial number of images for good reconstruction and the potential for artifacts when scenes change or are not fully captured. The conversation suggests that while NeRFs may not be as efficient as traditional rendering, they offer a powerful tool for 3D scene capture and reconstruction from real-world images.

10:01

๐Ÿ–ผ๏ธ Exploring NeRFs with a Christmas Tree Example

The script uses a Christmas tree as an example to demonstrate the capabilities and limitations of NeRFs. It discusses the process of capturing images with a phone, the challenges of dealing with motion blur and incomplete scene capture, and the subsequent training of the NeRF. The outcome is a 3D reconstruction that looks good from certain angles but degrades in quality when viewed from outside the captured image range. The discussion points out the need for a large dataset of images for high-quality reconstruction and the potential of NeRFs to be used for creating 3D assets for games or other applications. It also mentions the existence of rival technologies like GAN splatting and the potential future of 3D rendering.

Mindmap

Keywords

๐Ÿ’กNeural Radiance Fields (NeRF)

Neural Radiance Fields, or NeRF, is a technique in the field of computer vision and graphics that uses machine learning to generate high-quality 3D representations of a scene from a series of 2D images. It is a significant advancement in 3D reconstruction, as it can create detailed 3D models from simple RGB images. In the video, NeRF is the central theme, with the discussion focusing on how it reconstructs 3D scenes with high-quality details, unlike traditional methods like point clouds or meshes.

๐Ÿ’ก3D Reconstruction

3D Reconstruction refers to the process of creating a three-dimensional model of a real-world object or scene from 2D images or data. In the context of the video, 3D reconstruction is achieved through NeRF, which uses a neural network to infer the 3D structure of a scene. The script mentions traditional methods like point clouds and meshes, which are discrete and can become complex, contrasting with NeRF's ability to reconstruct scenes with high-quality detail from RGB images.

๐Ÿ’กRGB Images

RGB images are a common type of digital image that represents color information using three color channels: red, green, and blue. In the video script, RGB images are the input data for training the NeRF model. The script explains that NeRF can reconstruct 3D scenes from these simple images, which can be captured using everyday devices like smartphones, making the technology more accessible.

๐Ÿ’กRay Tracing

Ray tracing is a rendering technique used in computer graphics to simulate the path of light as it interacts with objects in a 3D environment. In the video, ray tracing is compared to the process NeRF uses to determine the color and density at various points along a ray path through a scene. The script explains that while traditional ray tracing might involve bouncing light off surfaces, NeRF uses a neural network to predict these properties.

๐Ÿ’กOverfitting

Overfitting in machine learning occurs when a model learns the training data too well, including its noise and outliers, to the point where it negatively impacts the model's performance on new, unseen data. In the video, overfitting is mentioned as a desired outcome for NeRF, as the goal is to accurately represent a specific scene from multiple angles, rather than generalize to other scenes.

๐Ÿ’กData Set

A data set in the context of the video refers to a collection of images and associated camera position data used to train the NeRF model. The script mentions that a good NeRF reconstruction requires a substantial number of images, ideally capturing the scene from various viewpoints. The term is used to describe the input materials that the neural network learns from to reconstruct the 3D scene.

๐Ÿ’กRendering

Rendering in computer graphics is the process of generating an image from a 3D model by means of computer programs. The video discusses how NeRF rendering differs from traditional rendering techniques. With NeRF, rendering involves sending out many rays and querying the neural network at various points along these rays to determine color and density, creating a detailed and accurate 3D representation.

๐Ÿ’กDensity

In the context of NeRF, density refers to a property that the neural network predicts for points along a ray path, indicating the likelihood of that point being occupied by an object in the 3D space. The script explains that density, along with color, is queried from the neural network for each point along a ray, contributing to the 3D understanding of the scene.

๐Ÿ’กCamera Position

Camera position data is crucial for NeRF as it provides the necessary information about where each input image was taken. This data is used during the training process to help the neural network understand the spatial relationships within the scene. The video script mentions the importance of having accurate camera position data for effective 3D reconstruction.

๐Ÿ’กReal-time Rendering

Real-time rendering refers to the ability of a system to generate 3D images or animations at the speed required for smooth, interactive viewing. The video script mentions that NeRF is not suitable for real-time rendering due to its computational complexity, which takes about 5 seconds to render a single image. This is in contrast to the expectations of real-time applications like video games or virtual reality.

Highlights

Introduction to Neural Radiance Fields (NeRF), a method for generating new views of scenes from RGB images.

NeRF's ability to reconstruct 3D scenes from simple RGB images, offering an alternative to traditional 3D reconstruction methods like point clouds and meshes.

Explanation of how NeRF works by training a neural network to reconstruct scenes in 3D from a series of RGB images.

The innovative aspect of NeRF that allows for rendering by querying a neural network for color and density at points along a ray.

Comparison of NeRF's rendering process to ray tracing, with a focus on the neural network's role in determining color and density.

The importance of capturing multiple images from different angles for effective NeRF training and reconstruction.

Discussion on the limitations of NeRF when there are changes or motion in the captured images, leading to artifacts like 'floaters'.

The practical challenge of capturing a sufficient dataset for NeRF, with a recommendation of at least 250 images for good reconstruction.

The trade-off between the quality of NeRF reconstruction and the density and diversity of the captured images.

Demonstration of how NeRF can be used to extract 3D information and potentially convert it into meshes for use in games or other 3D applications.

The current state of NeRF technology, including its limitations and the potential for future improvements.

Comparison of NeRF with other emerging techniques like GAN splatting, suggesting a competitive landscape in the field of 3D rendering.

The potential of NeRF and similar technologies to revolutionize 3D rendering and their implications for the industry.

Practical demonstration of capturing a scene with a phone and the subsequent NeRF training process.

Analysis of the quality of the NeRF reconstruction, highlighting areas of strength and weakness based on the captured data.

Discussion on the practical applications of NeRF, including its current capabilities and the potential for future development.

Transcripts

play00:00

yeah it's a different kind of ner we're

play00:01

looking at today so these are neural

play00:02

Radiance Fields right this is this is

play00:04

something that's been happening you know

play00:05

been going around the AI literature for

play00:07

a while a couple of years at least um

play00:09

really impressive ways of generating new

play00:12

views of

play00:15

scenes I'm familiar with how Nerfs work

play00:17

but I don't work with Nerfs every day so

play00:19

I brought in my tame PhD student here

play00:20

Lewis hello yeah and Lewis actually you

play00:23

know trains Nerfs he's working with

play00:25

Nerfs as part of his PhD and so you know

play00:27

you're going to explain to us how it

play00:28

works I hope so yeah what they do is

play00:30

they take a series of uh RGB images and

play00:34

from those RGB images of a scene they're

play00:36

able to reconstruct it in 3D using a

play00:39

neural network um traditional methods of

play00:42

3D reconstruction are things like Point

play00:44

clouds voxal grids meshes things like

play00:47

that but they're discret and for some

play00:49

scenarios that's all you need but for

play00:51

real life situations for example this

play00:53

room it can get very very complicated

play00:56

many many points many many faces on your

play00:59

mesh

play01:00

this NE Radiance field is is able to

play01:03

reconstruct it in very very high quality

play01:05

detail just only using simple RGB images

play01:08

that you can capture on your phone so I

play01:10

think the interesting thing about this

play01:11

for me is that we're actually not doing

play01:14

rendering in in a way you might expect

play01:15

so normally what you would do with

play01:16

rendering is you'd have some meshes but

play01:18

you then you know you you render them as

play01:21

pixels so you rasterize the image or you

play01:22

do ray tracing or something like this

play01:24

with a Nerf what we're actually doing is

play01:26

essentially a bit like ra tracing but

play01:29

we're actually using a newal Network to

play01:31

say okay this Ray is going to be this

play01:33

color and this Ray is going to be this

play01:34

color and so we basically building our

play01:36

3D scene into the kind of inside of a

play01:39

newal network so the parameters and

play01:41

weights of the newal network are

play01:42

actually encoding our Christmas tree or

play01:44

our car or whatever or our room or

play01:46

whatever else it is we're we're we're uh

play01:49

looking at I'm just going to draw here a

play01:51

trained Nerf for now so imagine we're

play01:54

looking at this from a side angle so we

play01:57

have something like this and and for the

play02:00

sake of this video as you will see later

play02:03

I'm going to draw you a nice Christmas

play02:04

tree um I hope you're better at drawing

play02:06

than me uh not a

play02:08

chance that's actually a lot better than

play02:10

me okay and you know I'll just shade it

play02:11

in green yeah there you go so and let's

play02:15

imagine that we have a camera up here

play02:19

now when I say camera an image has

play02:21

already been taken of of this Christmas

play02:23

tree in real life and what we're doing

play02:25

is simulating this this camera this

play02:27

camera in real life we're an image that

play02:31

looks like that makes sense looking at

play02:33

it from a above sort of view so how do

play02:36

we render this right so what we do we

play02:38

start shooting Rays into the environment

play02:41

like this this is what the neural

play02:43

network does so unlike things like

play02:46

diffusion which basically generate noise

play02:50

isn't it the stable diffusion and yeah

play02:52

yeah stuff like frogs on stilts and

play02:54

stuff like that yeah th this doesn't

play02:56

generate an image given a position in 3D

play03:00

and a and a view Direction it will give

play03:02

you a color and a density at that point

play03:05

in the environment that's important

play03:07

that's what is different from things

play03:09

like diffusion is that it has a 3D

play03:12

representation it understands the

play03:14

environment so now we shot this Ray

play03:16

through what we need to do is query a

play03:18

series of points along this Ray and ask

play03:21

the neural network okay I'm at this

play03:23

point I'm looking at it from this

play03:24

direction what should the color be what

play03:27

should the density be so let's start off

play03:29

by doing this so we're going to do this

play03:31

point

play03:32

here right what is the color and what is

play03:35

the density we go to neural network and

play03:36

it comes back and says the color is

play03:38

white doesn't matter because the density

play03:40

is zero why because we're in empty space

play03:44

right there's nothing there's nothing

play03:45

there yeah it's like air you can't

play03:47

render nothing so doesn't matter let's

play03:50

continue do it again nothing again

play03:54

nothing again nothing again

play03:57

nothing oops go thank you wow here we go

play04:02

we've created a point here and it's come

play04:04

back with a density of one AKA it's

play04:07

inside an object and it's come out of a

play04:09

color of green makes sense because we're

play04:12

now entering this Christmas tree and

play04:14

what we're going to do is do it again

play04:16

here

play04:18

here here here and you've queried all

play04:22

these points along the ray the neural

play04:24

networ come back with nothing nothing

play04:26

Well white white white but with let's

play04:28

say density 0 0

play04:30

green density one green density one

play04:32

green density one 0 0 perhaps the thing

play04:34

that when you first learn about this is

play04:35

hard to get your head around is normally

play04:37

in ra tracing what you would do is you

play04:38

would fire ARR into your scene usually

play04:40

based on your position of your pixels

play04:42

and say basically query what color is it

play04:44

and it might bounce around and do

play04:45

lighting or something like this but

play04:46

essentially it will come back and say

play04:47

yeah it's red and you paint that pixel

play04:49

red in this case we're actually doing

play04:52

multiple samples per Ray and we're

play04:55

saying what's here what's here what's

play04:56

here all the way along and what you'll

play04:57

do is sometimes you'll just shoot off

play04:59

and they there be nothing there

play05:00

sometimes you'll hit an object you'll

play05:01

intersect an object and for some time

play05:03

you'll see different colors and so your

play05:05

rendering process is going to be about

play05:06

sending out a lot of these rays and

play05:08

finding out where in 3D everything is

play05:11

right as opposed to just sending out a

play05:12

ray and going oh it's red yeah now that

play05:15

might seem really inefficient but

play05:16

actually this is the only way it trains

play05:17

right because if you trained a new

play05:20

network and just said yeah this is red

play05:22

this is red this is green this is red

play05:24

then it will work very nicely at just

play05:26

drawing that particular image yeah but

play05:28

you can't then move the camera over here

play05:30

and say okay what does it look like from

play05:32

above right where we haven't necessarily

play05:34

got any images so the idea would be that

play05:35

you train this with with sort of another

play05:38

camera my camera's worse than yours uh

play05:41

and another camera so you maybe have

play05:42

three cameras but now we can draw this

play05:44

one and we can draw this one and this

play05:45

one because we can shoot Rays out and we

play05:47

can do thing what's important here is

play05:50

that when you shoot the Rays out

play05:52

here they are intersecting let's say

play05:56

like that that's sort of bit iffy

play05:57

because it goes through there but that

play05:59

how you're able to finalize where that

play06:02

object is in 3D space because these

play06:03

points intersect on the Rays that you

play06:06

shoot which is why with Nerf one of the

play06:09

downsides maybe is that you need quite a

play06:12

few images for it to properly get a good

play06:15

reconstruction if you only have one

play06:17

image what will happen is it will look

play06:18

pretty great from anywhere near that

play06:20

image but if you move elsewhere and it

play06:22

will it will degenerate pretty quickly

play06:24

um so and another thing that's

play06:26

interesting about this is normally in

play06:28

any kind of AI or machine learning what

play06:30

you're trying to do is generalize your

play06:32

approach to some other data set so you

play06:33

say I want to train on this data set but

play06:35

ultimately I don't really care how well

play06:36

it works on that data set CU I already

play06:38

know all the answers what I care about

play06:40

is how it works on this data set and

play06:42

this data set but in Nerf we're actually

play06:44

learning this exact scene it's not going

play06:46

to reconstruct a different kind of tree

play06:48

cuz that's not in the training set we

play06:49

only care about producing images of this

play06:52

tree from different directions yeah

play06:53

overfitting to the max pretty much they

play06:55

tell you never to overfit but in this

play06:57

case overfit as much as as you can this

play07:00

is over fitting by Design right by so I

play07:02

figure we could just Trot off down the

play07:04

corridor and take some pictures and have

play07:06

a go now Christmas tree is actually a

play07:08

really hard example of pines and

play07:10

everything yeah I mean that's not an

play07:11

easy thing right and and and it's a good

play07:13

example in a way because we'll see some

play07:14

of the problems as well as the benefits

play07:15

of Nerf but also it is worth knowing

play07:18

that this is something that most

play07:19

reconstruction techniques are going to

play07:20

really struggle on right so this is a

play07:23

hard problem but it's also fun yes well

play07:25

you know when you say you've got to take

play07:26

quite a few pictures how many is quite a

play07:28

few for something like a if you want a

play07:30

good reconstruction at least 250 okay

play07:33

yeah and and those are the good images

play07:35

there are to be fair because Nur really

play07:36

popular there's lots of research going

play07:38

on and there are many many techniques

play07:40

that are trying to reduce the number

play07:42

they don't all work very well right so

play07:43

you know your mileage May VAR if you

play07:44

want really good reconstruction then a

play07:46

lot of images is what you're going to

play07:48

need and obviously if I'm obviously I'm

play07:50

a videographer so video any good video

play07:52

is good because video is a lot of I mean

play07:54

aside from maybe motion blur and things

play07:56

like that if you've got good quality

play07:57

frames that's just more and more shots

play07:59

so what have we got we've got a nice

play08:01

looking Christmas tree I didn't decorate

play08:03

this one and this is a very complicated

play08:05

scene right because we've got bushes at

play08:06

the side with huge numbers of leaves

play08:08

we've got sofas we've got whatever all

play08:11

this stuff is hanging off the the

play08:12

trellises it's a big big place as well

play08:15

one thing about normal Nerf that you see

play08:17

in the literature is that they run on

play08:18

fairly constrained scenes like you often

play08:20

you've used a robot or some other

play08:21

capture rig to capture very nice

play08:23

concentric circles of images all equally

play08:25

spaced everything's very constrained

play08:28

we're just going to kind of wave a

play08:29

around and see what happens it's more

play08:31

fun so all right so Lou will catch us

play08:33

some videos and then we'll uh I'll yeah

play08:36

I'll stand over here and not be in the

play08:37

shot what happens if you get things

play08:40

changing in these images it's not good

play08:43

that that's where you'll get a lot of I

play08:45

think they they call them floaters where

play08:47

it can't figure out where to put certain

play08:50

pixels in the scene cuz they're

play08:51

different throughout and then it would

play08:53

just create noise so you could be photob

play08:55

bombed and that would really cause a

play08:56

problem exactly you get kind of a

play08:58

ghostly mic

play09:00

and then fing back out again that's

play09:01

exactly it yeah so we've got the video

play09:03

now the next step is we need to get the

play09:06

camera positions and then we just got

play09:08

train the Nerf and uh that should

play09:11

take about an hour I've trained up this

play09:14

Nerf and currently we're we're viewing

play09:17

it for this I'm using Nerf Studio which

play09:19

is what a lot of people are using now

play09:21

cuz it's very userfriendly and it's very

play09:24

very good and this is the sort of thing

play09:26

that it looks like so you can see all of

play09:28

the images and where they are in 3D

play09:30

space and that was me going around that

play09:33

Christmas tree it takes a while for the

play09:35

for it the quality to increase because

play09:36

what this does it renders it very quick

play09:38

because you need to get an understanding

play09:40

of the environment but Nerfs are slow

play09:43

right for real-time rendering so it

play09:45

takes about 5 seconds for it to render a

play09:47

good image every time but if I move it

play09:49

around here and let me just get rid of

play09:51

the cameras here so you can just see the

play09:53

thing give it a few seconds and it

play09:56

should there we go now I want to talk

play09:59

talk about what's good and what's bad

play10:01

about this there are Nerf data sets out

play10:03

there that are fantastic and you'll be

play10:06

able to get really good high quality

play10:08

reconstructions from them this is is not

play10:10

so good because I took it on my phone

play10:13

and there's lots of things like motion

play10:14

blur different sort of issues I didn't

play10:16

capture the whole scene but things like

play10:18

the Christmas tree that looks pretty

play10:20

good to me you know you can see the bu

play10:23

balls on it and high quality you're not

play10:25

this tall right with the greatest

play10:26

respect if you come down yeah yeah will

play10:29

it be better if you're closer to where

play10:30

the original cameras were ex it should

play10:32

be oh hang on hang on it's a bit finicky

play10:36

mind you there we go so this is a player

play10:39

yeah it's a real time renderer and it

play10:41

looks good from this perspective because

play10:43

this was where the images were taken

play10:44

from cuz that's around my height where I

play10:46

was taking it from and this is probably

play10:48

very close to there you go it's very

play10:51

close to where these images were taken

play10:53

which is why you can see when it loads

play10:56

you can see the background somewhat you

play10:58

can see the tree Tre it makes sense the

play11:00

Christmas tree looks good as soon as you

play11:02

move out suddenly it looks pretty bad

play11:07

right and that is one of the things with

play11:08

Nerf is that it's very good where the

play11:10

images can see the scene there is going

play11:12

to be a lot of loss of quality when you

play11:14

go outside where we were capturing the

play11:15

images from would you call this a data

play11:17

set or a picture or an image or a scene

play11:19

what would you call it I would call this

play11:20

a data set data set so how big is a data

play11:22

set like that so this is around 300

play11:25

images so Nerf data sets have a series

play11:28

of Imes and then uh Json far with where

play11:31

all the camera positions are that's

play11:33

that's it but this is about 300 but

play11:35

they're all relatively if you see here

play11:38

they're all pretty close together right

play11:40

if you want a really good capture really

play11:43

good scene you want them spaced far

play11:45

apart capturing the entire thing I'm not

play11:48

Mr Fantastic with long arms I can only

play11:51

capture things around here right which

play11:53

is why when you go and look let's say

play11:56

over here in the background

play11:59

you'll see that the

play12:01

background here especially on the

play12:03

ceiling it's really not very good

play12:05

because not a lot of dat was captured in

play12:07

these images of the background so that's

play12:08

why it's not as good you see here this

play12:10

is just noise and the reason why that is

play12:13

just noise cuz I didn't actually capture

play12:14

any of that floor when I was capturing

play12:16

the video I just forgot which is why

play12:18

when when you look at it from here it

play12:20

looks awful but that's not really the

play12:22

fault of the Nerf per se it's more of a

play12:24

fault of me perhaps it's worth thinking

play12:26

about what we would actually use this

play12:27

for because ultimately people might be

play12:29

looking at this and going well it's not

play12:30

as good as a 3D render but but we

play12:32

actually only had to capture 300 images

play12:33

we didn't have to artistically design

play12:35

the tree in 3D we didn't have to paint

play12:37

all the all the meshes we didn't have to

play12:39

do any of that um and you know I

play12:42

couldn't do that anyway because you've

play12:43

seen my drawing abilities so but the

play12:45

other thing is that we also as well as a

play12:47

color we can also extract where the

play12:49

objects are which means that you can

play12:50

convert a lot of these objects into a

play12:52

mesh so you can use this multi view

play12:55

reconstruction to essentially obtain 3D

play12:57

Volume so you could then speed up your

play12:59

creation of an actual 3D asset that you

play13:01

could use in a game or something like

play13:02

this is this the future then of of 3D

play13:04

rendering is this how it's all going to

play13:05

work this month possibly yeah actually

play13:08

there's a new rival to Nerf called Gan

play13:11

splatting which is also providing

play13:12

incredibly impressive results so maybe

play13:14

that's video number

play13:18

two this network is maybe slightly

play13:20

better when it has a text estima the

play13:22

noise so you actually put in two images

play13:26

of dystopian abandoned futuristic I with

play13:29

overgown plants right and then I just

play13:31

put them in a four Loop and just produce

play13:33

200 of them so I can pick the nice ones

Rate This
โ˜…
โ˜…
โ˜…
โ˜…
โ˜…

5.0 / 5 (0 votes)

Related Tags
Neural Radiance Fields3D ReconstructionAI LiteratureRGB ImagesNeural NetworkRendering Techniques3D Scene EncodingOverfittingReal-time RenderingAI Research