3DoF AR Tag Following Robot

Our team engineered a sophisticated 3 Degree of Freedom AR tag tracking robot, featuring a custom-made chassis crafted from 80/20 T-slot aluminum. Every aspect of the robot’s design and construction is meticulously tailored, showcasing a commitment to precision and innovation.

The primary objective of this project was to demonstrate the distinctions between position-based visual servoing and image-based visual servoing. The robot incorporated an ELP 8mp USB camera seamlessly interfaced with a custom Python program. Calibration of the camera was performed utilizing OpenCV’s integrated calibration function and a dataset comprising 20 sample images of a 9x6 checkerboard.

In the realm of position-based visual servoing, OpenCV transmitted the AR tag coordinates directly to the movement algorithm. The robot, responding with remarkable accuracy, would adjust its position by 30% of the distance towards the target position above the tag. While effective, this method demanded significant computational resources.

Conversely, in the domain of image-based visual servoing, OpenCV identified the four corners of the AR tag. These corner coordinates were then fed into the movement algorithm, leveraging a timestep of 1 and a Jacobian matrix to orchestrate the robot’s precise movement to the desired position. This approach, though faster, exhibited a tendency to overshoot and faced challenges when confronted with substantial changes in the AR tag’s position.

Alexander Besch
Alexander Besch
Electrical Controls Engineer

My research interests include artifically intelligent agents, intelligent robotics, and manufacturing.