0

I'm trying to do depth estimation with CNNs (this is my ultimate goal), but a problem that i found is: I just did image classifications with CNNs, using for example "CIFAR-10", "MNIST", "Cats vs Dogs", etc. To do depth estimation I need to output a new image (the NYUv2 dataset has the labeled images). So, I'll input an image like 256x256x3 and need to output another image with for example 228x228x3.

What I need to do? Can I just do the convolutions for a while and after that decrease the features maps and increase the dimension? Thanks

obs: I'm using Tensorflow 2.0

1

1 Answer 1

0

I suggest you use a type of UNet. This kind of architecture has downsampling layers, followed by up sampling layers to get back to the original spatial dimensions.

Sign up to request clarification or add additional context in comments.

Comments

Your Answer

By clicking “Post Your Answer”, you agree to our terms of service and acknowledge you have read our privacy policy.

Start asking to get answers

Find the answer to your question by asking.

Ask question

Explore related questions

See similar questions with these tags.