Hi all,
I’ve recently started digging deeper into the StableHLO ecosystem, and I’m trying to better understand the role and maturity of the stablehlo_to_tf_saved_model utility.
I have a few questions regarding its intended use and real-world applicability:
What was the original motivation for this tool?
Since StableHLO is designed to be framework-independent and serve as a portable compiler IR, converting it back to a framework-specific representation (TensorFlow SavedModel) seems somewhat counterintuitive to me. I would have expected downstream usage to favor other MLIR dialects or compiler backends instead. Could someone clarify the primary use case this tool was designed for?
How mature and widely used is this path?
Has stablehlo_to_tf_saved_model been tested extensively beyond small or synthetic examples (I only found one test with a very simple model, 2 constants)? Is it used in production or large-scale internal pipelines anywhere in the ecosystem?
Scalability to large models (e.g., LLaMA 2 7B)
Is it realistic to use this utility with very large models (multi-billion parameter scale)?
In particular:
How are large constants/weights typically handled?
How are input_locations and state_dict expected to be configured for such models?
Is there tooling to automate this process, or is it currently a mostly manual setup?
If anyone has experience using this tool with large transformer models (or can point to relevant discussions or code), I would greatly appreciate any references or insights.
Thanks in advance!
Hamza