Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

We don't use T5 exactly, we use a derivative that has a similar (but not identical) architecture and is also pre-trained differently. That model, combined with our factual generation dataset and clever prompt engineering, seems to be the secret sauce for reducing hallucination.

And thank you :) it's comments like this that really fire us up



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: