I am actually interested in both.
From practical side of things I would benefit most from things that are more or less applicable in the paradigm of ceres-solver, even if it would require some modifications of ceres-solver itself (+ I could probably fix some known issues on my way).
From theoretical side of things - I would be thankful for insights and references on constrained non-linear least squares on manifolds with a boundary.
From practical side of things I also want to make local parameterizations and parameter block boundaries in Ceres more "compatible" with each other (it seems to me that applying constraints to parameter block with non-trivial local parameterization breaks invariant of local parameterization, when boundary conditions become utilized).
The particular problem I am solving now involves optimization over polynomials of a fixed degree, monotonically non-decreasing on the fixed interval [this is a convex subset of polynomials of a fixed degree d, and interior is a manifold with dimension d+1], and a bunch of SE3 and R^3 parameters.
So far I've tried:
* Sum-of-squares parameterization of polynomials (since derivative of monotonically increasing polynomial is positive, and every polynomial p positive on the interval can be rewritten in a certain way using squared polynomials), but it has some problems (zero gradients over parameters for p(x)=0, etc)
* Optimization directly over coefficients of monotonic polynomial, rejecting steps that result into "bad" roots of derivative; this seem to have some problems near boundary, but works better (in terms of number of cases when optimization succeeds)
Playing with hyper-parameters (degree of polynomial) for this problem makes me think that I might benefit from a more suitable optimization algorithm, that would work better on the boundary.
For example, when fitting a model to synthetic data, generated with polynomial p(x) of degree d, with roots of dp/dx near the ends of interval - typically it helps to fit polynomial of degree d+2, discard extra coefficients, and optimize again as polynomial of degree d.