Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
sota_pop
7 months ago
|
parent
|
context
|
favorite
| on:
Fine-tuning LLMs is a waste of time
Yes, that does sound very similar. To my knowledge, isn’t that (effectively) how the latest DeepSeek breakthroughs were made? (i.e. by leveraging chatgpt outputs to provide feedback for training the likes of R1)
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: