Hi Edmund et al.
Any updates on this? I haven't observed the discrepancy myself with pretrained models downloaded from the model Zoo (e.g. human pose) and with the ones I happened to train myself. I have asked around and it seems like some people experience this issue, and some don't, so I would like to understand better what exactly causes this.
I have asked Mackenzie about this and she suggested this might be down to an effect of pre-scaling which might be enabled by default in deeplabcut (this is sometimes done for robustness of the network to changes in resolution from the original training set).
How exactly are you training the models? Are you using the DLC gui as is, or the python scripts?
Any accuracy issue will be the same regardless of whether you are playing videos offline or online, as the evaluation code is exactly the same and simply expects a sequence of frames as an input, regardless of where they came from.
This is not likely to be related to GPU computational power or hardware, as in that case it would affect DeepLabCut and Bonsai+DLC the same, as they both use TensorFlow under the hood, with very little overhead and no particular conversion.
From the examples Edmund shared with me the issue seemed much more a matter of scale and shift, as the proportions all seemed correct, which I don't believe you would expect if the tracking was independently failing to regress the locations of features due to poor training accuracy. I'm not sure if the failures reported by Blasiak and Konrad show the same pattern.
If anyone can share their training procedure and projects, that might help us to figure out what exactly is causing this discrepancy.