Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Regarding direct training deployment on real robot #1

Open
yolo01826 opened this issue May 18, 2024 · 2 comments
Open

Regarding direct training deployment on real robot #1

yolo01826 opened this issue May 18, 2024 · 2 comments

Comments

@yolo01826
Copy link

Dear Author,

I really like your work and have a few questions I'd like to ask you.

  1. Firstly, if the keypoint tracking method is trained on a single task with 100 human demonstrations and 10 robot teleoperations, can it yield a well-performing tracking model?
  2. Does the keypoint information passed into the policy consist of its pixel coordinates, or does it include some form of rotation?
  3. The paper mentions using 4 A100 GPUs for training. Is it possible to train a good single-task policy using a 4090 GPU instead?

Thank you.

@AlvinWen428
Copy link
Collaborator

Hi, thanks for your interest in our work.

  1. Yes. A single task makes it easier to learn a good track transformer.
  2. Only 2D coordinates are fed into the policy. Because there are multiple point tracks, the policy will infer the rotation by itself.
  3. Yes, but you need to tune the hyperparameters.

I hope my answers will be helpful.

@dblanm
Copy link

dblanm commented May 30, 2024

Hey! I want to follow on questions about training and deployment in real.

  1. Is there any example of the data required for providing real-world demonstrations with a robot?
  2. Will you provide code for the real-world experiments with the UR5?
  3. How are you sending the commands with the UR5, is this ROS, RTDE with movep and movej, or something different?
  4. Related to 3, Could you explain the pipeline: ATM tracker & policy + send commands to robots
  5. Could you also specify more in detail what is needed for doing this with real hardware: what inputs for doing this with real hardware, requirements in camera viewpoints, any required transformations of observations to the policy, calibrations, limits of robot/human demonstration velocities, GPUs required for doing inference, etc.

Thanks for the help!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants