Alexander Soare
|
92573486a8
|
Don't use async envs by default (#448)
|
2024-09-20 15:22:52 +02:00 |
|
Zhuoheng Li
|
a2592a5563
|
Provide more information to the user (#358)
Co-authored-by: Alexander Soare <alexander.soare159@gmail.com>
Co-authored-by: Remi <re.cadene@gmail.com>
|
2024-08-23 11:00:35 +01:00 |
|
Alexander Soare
|
f8a6574698
|
Add online training with TD-MPC as proof of concept (#338)
|
2024-07-25 11:16:38 +01:00 |
|
Remi
|
265b0ec44d
|
Refactor env to add key word arguments from config yaml (#223)
Co-authored-by: Simon Alibert <75076266+aliberts@users.noreply.github.com>
|
2024-05-30 13:45:22 +02:00 |
|
Alexander Soare
|
9d60dce6f3
|
Tidy up yaml configs (#121)
|
2024-04-30 16:08:59 +01:00 |
|
Remi
|
659c69a1c0
|
Refactor datasets into LeRobotDataset (#91)
Co-authored-by: Alexander Soare <alexander.soare159@gmail.com>
|
2024-04-25 12:23:12 +02:00 |
|
Remi
|
e760e4cd63
|
Move normalization to policy for act and diffusion (#90)
Co-authored-by: Alexander Soare <alexander.soare159@gmail.com>
|
2024-04-25 11:47:38 +02:00 |
|
Cadene
|
06573d7f67
|
online training works (loss goes down), remove repeat_action, eval_policy outputs episodes data, eval_policy uses max_episodes_rendered
|
2024-04-10 11:34:01 +00:00 |
|
Cadene
|
1149894e1d
|
rename handle -> task
|
2024-04-08 14:54:52 +00:00 |
|
Simon Alibert
|
3f6dfa4916
|
Add gym-aloha, rename simxarm -> xarm, refactor
|
2024-04-08 16:24:11 +02:00 |
|