Hacker News new | past | comments | ask | show | jobs | submit login

You are exactly right! As I wanted to have a solution that works with many LLMs out of the box, I focused on chain of thoughts and few shot learnings.

Lots of paper show that fine-tuning only helps with steerability and form (https://arxiv.org/abs/2402.05119), therefore I thought it would be sufficient to provide just the right examples and it did work!

We do intend to create a decentralized dataset to further train models and have maybe a 2b or 7b model working well




What kind of problems are you seeing that you think can be improved with a fine tune?


Thank you for linking that paper!




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: