Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it possible to get more shots from various angles ? #19

Closed
altava-sgp opened this issue Oct 18, 2024 · 7 comments
Closed

Is it possible to get more shots from various angles ? #19

altava-sgp opened this issue Oct 18, 2024 · 7 comments

Comments

@altava-sgp
Copy link

At first thanks for your great job! 👍

I have a question.
Is it possible to get more shots from various angles ?
A only single shot is possible ?

Please excuse me if I am missing anything.

@haodong2000
Copy link
Collaborator

Hi @altava-sgp thanks so much for your interest! Sorry for my late response.

Do you mean to get multi views of depth and normal given a single RGB image?

For depth estimation: you can lift the depth maps into 3D point clouds using camera intrinsics and then render new depth maps or RGB from different views. As shown in this issue.

For normal, if you want mulri views of normal maps, you probably should do: normal -> depth -> 3D point clouds -> depth (new view) -> normal (new view). You can refer this repo for normal2depth and this repo for depth2normal.

Best,

@altava-sgp
Copy link
Author

@haodong2000 Thanks for your answer. I will try as you said. 👍

@altava-sgp
Copy link
Author

@haodong2000 I also hope Lotus can generate various normals like I said.
It will be much helpful for AI generation.
Please consider this seriously. 🙏

@altava-sgp
Copy link
Author

Like these.
normals_back_masked
normals_front_left_masked
normals_front_masked
normals_front_right_masked
normals_left_masked
normals_right_masked

@haodong2000
Copy link
Collaborator

Hi @altava-sgp , thanks for response.

I get your point. Currently, Lotus only supports generating relative normal maps only for the current view. That is the way we feed the GTs during training.

If you require normal maps that remain consistent across different views, I recommend leveraging camera extrinsics to transform the normal maps into world coordinates. It's training-free.

I am not sure how precise and consistent you need the normal maps to be or what kind of application you are aiming to build. This simple solution may struggle to achieve perfectly consistent multi-view normal maps. If you require absolute consistency, an alternative approach could be to first reconstruct the 3D model and then generate the normal maps through rendering.

@altava-sgp
Copy link
Author

@haodong2000 Thank you for your kind answer.
Actually, I am trying to generate a good 3D model from a single image.
I will try other methods, including the suggestions you gave. 😄

@haodong2000
Copy link
Collaborator

Hi, thanks for response!

Actually I have tried to reconstruct 3D mesh using purely relative normal and depth maps, please see Fig. 14 of the paper for more details.

It seems that the performance is good but not in very high-quality.

Thus, I believe that by aligning all views into same (world) coordinates, the performance will boost significantly.

Best,

@jingheya jingheya closed this as completed Dec 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants