What if it's more than just limitations?

IMO, What if Anthropic announced a new safety classifier before announcing its new model just because of the risks that it's kind of very powerful