I feel this is technically correct but intentionally cheating. no one - including the model creators - expects that to be the interface; it undermines they entire value proposition of using an LLM in the first place if I need to engineer the inputs to ensure reproducability. I'd love to hear some real world scenarios that do this where it wouldn't be simpler to NOT use AI.
When should a model's output be deterministic?
When should a model's output be non-deterministic?
When many humans interact with the same model, then maybe the model should try different seed values, and make measurements.
When model interaction is limited to a single human, then maybe the model should try different seed values, and make measurements.