Hi all,
I have a system that consists of N cameras on fixed positions. My goal is to find the relative poses between them assuming that the 1st camera's pose is (0,0,0),(0,0,0) (my world's coordinate system is 1st camera's system).
What I've done so far was to cv2.stereoCalibrate each pair of cameras and by using their relative poses I could get their pose with respect of my reference camera. This didn't seem adequate as I was propagating errors from the steroCalibration to other pairs of cameras.
To further optimise their poses I'm using the ceres bundle adjustment example, where I'm putting the poses that I've found before as initial guess, their intrinsics, some object points and the observations.
As object points' coordinates, I'm using apriltag detections coming from the 1st camera (as I'm using this camera's coordinate system as a reference).
While it seems that the solver does the job amazingly well and converges very soon, I'd like to raise this question. Is a single photo taken by all cameras enough, or should I consider using a moving target? The problem with the second approach however, would be that for each image, each detection should be considered as a new object point.
Any insight, suggestion would be more than welcome,
Thanks in advance