You're reading the documentation for a development version. For the latest released version, please have a look at master.
Procedure of Using DA Calibration Node
In order to perform calibration using this node, we need to capture camera images of the calibration tool from different views. The calibration tools that can be used in this node are grid circles and gird chessboards that examples of each are shown below.
Note that the captured images should contain a variety of images with different position and rotation angles to ensure a good calibration result. Once you have captured calibration images, you can add the calibration node into the flowchart to perform calibration. (Templates provided in the software) When adding this node to the flowchart, you will see the option for choosing the calibration mode. We provide three modes for the calibration node. First, the accumulation mode is used to load the captured images, poses and point clouds from the calibration board and detect the corner points (chessboard) or circle center points (circle board). When you choose the accumulation mode, you will be able to set different options and inputs.
As shown in the above image, you have to choose the hand-eye configuration. Also, the other inputs include the captured image, robot pose (base in gripper) for each robot position. In addition, you have to choose the used grid board type (calibration tool). Note that you can see the result of the accumulation node (detected points) by clicking on the node. In case the algorithm is not able to detect the corner/circle points, the image will be skipped and not used for final calibration.
After obtaining the corner/circle points, it is time to use the results and inputs from the accumulation mode to perform robot-camera calibration. This can be done by adding another calibration node in the flowchart and setting the mode to the final mode. The final mode will perform the main computation for the calibration.
In the final node, you have to specify the reference accumulation node, so, the results then can be used for calibration. Also, as the calibration algorithm will use an optimization based method to obtain the calibration results, you need to choose the used cost function. In case of using a 3D camera and 3D picking, it is a better option to use the RPDepth cost function as our experiments show better results for this cost function. However, note that when using RPDepth, you will need to set the option “using the point cloud” in the accumulation mode. By running the calibration node in the final mode, the calibration results will be obtained. More specifically, the main outputs of the final mode are the camera to gripper and base to board 3D transformations in the eye-in-hand scenario. This will be camera to base and gripper to board 3D transformations in the eye-to-hand scenario. These transformations then can be used in the picking flowchart for different robot picking operations.
Another existing mode for calibration mode is the load mode which simply loads the result file saved in the final mode. This can be used to visualize the result of the calibration without a need to run the previous nodes.