Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
jjmarr
40 days ago
|
parent
|
context
|
favorite
| on:
“Erdos problem #728 was solved more or less autono...
Seeing a task-specific model be consistently better at
anything
is extremely surprising given rapid innovation in foundation models.
Have you tried Aristotle on other, non-Lean tasks? Is it better at logical reasoning in general?
runeblaze
39 days ago
[–]
Is it though? There is a reason gpt has codex variants. RL on a specific task raises the performance on that task
jjmarr
39 days ago
|
parent
[–]
Post-training doesn't transfer over when a new base model arrives so anyone who adopted a task-specific LLM gets burned when a new generational advance comes out.
runeblaze
38 days ago
|
root
|
parent
[–]
Resouce-affording, if you are chasing the frontier of some more niche task you redo your training regime on the new-gen LLMs
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search:
Have you tried Aristotle on other, non-Lean tasks? Is it better at logical reasoning in general?