Hello,
thank you so much for your amazing work ! 👍
I was wondering about a couple of things :
- I am really curious about the architecture of the network. In other GitHub issues, you said that 2D and 3D heatmap are independent and trained at the same time. How to feed 3D data into the network? Does 2D heatmap and offset were used in predict stage?
- How did you generate the coordinates of 3D data through unity. For the data collection stage , how does the action in front of the camera correspond to the action of avatar in the unity?
- What is the meaning of 2D and 3D offset?
I'll really appreciate your help.
Thank you
Hello,
thank you so much for your amazing work ! 👍
I was wondering about a couple of things :
I'll really appreciate your help.
Thank you