We don't use T5 exactly, we use a derivative that has a similar (but not identical) architecture and is also pre-trained differently. That model, combined with our factual generation dataset and clever prompt engineering, seems to be the secret sauce for reducing hallucination.
And thank you :) it's comments like this that really fire us up
And thank you :) it's comments like this that really fire us up