I will confine the answer to EMAN2.1, where you actually have a mechanism for monitoring "better" and "worse" (gold-standard). If you look at the report generated while EMAN2.1 runs, you will note that it shows you two different sets of FSC curves. It shows the gold standard resolution as it changes from iteration to iteration, and it shows the 'convergence plot' independently for the even and odd maps. By looking at both of these sets of curves, you can learn quite a lot about what's going on.
- if you observe the gold-standard resolution curve getting worse for one or two iterations, then getting better again, that is indicative of a bad initial model, and shows that both refinements have independently found their way to the correct structure. This is a very good sign.
- If you observe the gold-standard resolution starting at some point then getting worse, and eventually getting stuck at a worse point, this indicates that your initial target resolution was higher than the data could sustain, and you need to rerun the refinement with a lower (worse) target resolution.
- If you observe the gold standard curve getting stuck at some resolution, but the individual convergence plots continue to converge past this point, this is a clear indication of structural variability in the data, and means you really need to work with e2refinemulti.py to try and separate the data.
- If you are unable to achieve convergence & self-consistence in your map, there are two primary causes: 1) variability in the data or 2) a perversely bad initial model
- to exclude case 2 you need to rerun refinements with a couple of different starting models and observe what happens. If all of the starting models produce pretty similar results, then you know 2 isn't the issue.
- In case 1, running e2refinemulti, which you may need to run for a significantly larger number of iterations than e2refine_easy, is the best way to learn more. e2refine_easy should normally converge in 5-6 iterations, and should really never require more than 10 iterations for a homogeneous data set. With e2refinemulti, it can be worthwhile to run as many as 20 iterations in some unusual situations, though normally ~10 will give the results you are after.
Both Relion and EMAN can get stuck in local minima with heterogeneous particles, but there is a difference in behavior. In EMAN, if the data is very heterogeneous, and no self-consistent structure has been obtained, it will tend to 'bounce around' the local minima and continue to change at least subtly. In Relion, due to its use of ML, if it gets 'stuck' like this, it will tend to give very solid looking answers which won't change. It can be very difficult in this situation to tell whether the answers are 'real' or not.