Hacker Newsnew | past | comments | ask | show | jobs | submit | bestcoder69's commentslogin


Llamafile is great but solves a slightly different problem very well: how do I easily download and run a single model without having any infrastructure in place first?

Ollama solves the problem of how I run many models without having to deal with many instances of infrastructure.


You don't need any infrastructure for llamafiles, you just download and run them (everywhere).


Yes, that's what I meant, sorry if it was confusing: The problem that Llamafiles solve is making it easy to set up one model without infrastructure.


It's actually more difficult to use on linux (compared to ollama) because of the weird binfmt contortions you have to go through.


What contortions? None of my machines needed more than `chmod +x` for llamafile to run.


permitted to force RTO without any data != ought to force RTO without any data


Why though?


I can already use multiple backends by writing different code. The value-add langchain would need to prove is whether i can get better results using their abstractions compared to me doing it manually. Every time I’ve looked at how langchain’s prompts are constructed, they went wayyy against LLM vendor guidance so I have doubts.

Also the downside of not being able to easily tweak prompts based on experiments (crucial!)

And not to mention the library doesn’t actually live up to this use case, and you immediately (IME) run into “you actually can’t use a _Chain with provider _ if you want to use their _ API”, so I ultimately did have to care about whats supposed to be abstracted over


Your comment gives better reasons than the article for not using LangChain.


They’ve been growth hacking the whole time pretty much, optimizing for virality. Eg integrating with every ai thing under the sun, so they could publish a seo-friendly “use gpt3 with someVecDb and lang chain” page, but for every permutation you can think. Easy for them to write since langchains abstractions are just unnecessary wrappers. They’ve also had meetups since very early on. The design seems to make langchain hard to remove since you’re no longer doing functional composition like you’d do in normal python - you’re combining Chains. You can’t insert your own log statements in between their calls so you have to onboard to langsmith for observability (their saas play). Now they have a DSL with their own binary operators :[

VC-backed, if you couldn’t guess already


openai cookbook! Instructor is a decent library that can help with the annoying parts without abstracting the whole api call - see it’s docs for RAG examples.


They released chat and non-chat (completion) versions of 3.5 at the same time so not really; the switch to chat model was orthogonal.

e: actually some of the pre-chatgpt models like code-davinci may have been considered part of the 3.5 series too


limewire teens were pioneers too


together.xyz, although they’re .ai now


Newt Gingrich proposed once that we make kids do this, but it was just gonna be for the poor kids.


Tesla could just negotiate instead of being antagonistic.


Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: