Evaluate
Evaluate Trained Policy
You can use the record script from lerobot/record.py but with a policy checkpoint as input. For instance, run this command to record 10 evaluation episodes:
python -m lerobot.record  \
  --robot.type=giraffe_follower \
  --robot.port=/dev/ttyACM1 \
  --robot.cameras="{ up: {type: opencv, index_or_path: /dev/video10, width: 640, height: 480, fps: 30}, side: {type: intelrealsense, serial_number_or_name: 233522074606, width: 640, height: 480, fps: 30}}" \
  --robot.id=my_follower_arm \
  --display_data=false \
  --dataset.repo_id=$HF_USER/eval_giraffe \
  --dataset.single_task="Put lego brick into the transparent box" \
  --policy.path=${HF_USER}/my_policy
As you can see, it’s almost the same command as previously used to record your training dataset. Two things changed:
- There is an additional --control.policy.pathargument which indicates the path to your policy checkpoint with (e.g.outputs/train/eval_act_giraffe_test/checkpoints/last/pretrained_model). You can also use the model repository if you uploaded a model checkpoint to the hub (e.g.${HF_USER}/act_giraffe_test).
- The name of dataset begins by eval to reflect that you are running inference (e.g.${HF_USER}/eval_act_giraffe_test).