Do you have just a single camera or multiple ones? If it's just a single one you can simply set the pose to the identity and specify a large diagonal R-matrix (inverse of the covariance!) in camera_poses.yaml. See some of the single-camera examples in the tagslam_test repository, under "tests/".
If you have multiple cameras, then in camera_poses.yaml you specify just the pose of one camera, e.g. camera 0. You put up some tags (you must specify the world pose of at least one of them) and make a recording. TagSLAM can then figure out the poses of all the tags and it will also give you the extrinsics of the cameras that you did not specify in camera_poses.yaml. You can find these files in the ~/.ros/ subdirectory after you have completed the run. Note that you must also perform the service call ("dump") for tagslam to actually write those files. See documentation.
Once you have the extrinsic calibration tagslam will give you the poses of the tags and the camera poses (and thereby the robot pose). Only as long as it sees any tags of course, and the accuracy is limited if the tags are few and small. How the AMCL pose figures into that I don't know. I'm not familiar with that package.