Theres tons of ai apps. They're all general use chatbots or coding agents. Manus, Cursor, ChatGPT. Almost every app that has a robust search uses a reranker llm. AI is everywhere.
As far as totally new products - I built one (Habit.am - wordless journaling for mental health) and new products require new habits, people trying new things, its not that easy to change people's behavior. It would be much easier for me to sell my little app if it was a literal plain old journal.
I don't want to put information into a black box of mystery that can then be used for other monetization purposes. I am still waiting for a realistic local solution.
I ran tests of 100 attempts with different prompt/scenario combinations. Each "attempt"/theory had 3 different system prompts wordings. Most of the prompts did not mention a colon, but it kept appearing. When I added negative instructions against using a colon, the quality went down (most of the tool calls were malformed, one common issue was markdown ticks in front)
It was only when my system prompt acted like colons were normal that I kept getting 100/100 perfect expected tool calls. I ranked my system prompts by which returned the most consistent commands.
I tried over 20 variations of different system prompts.
Once I changed my tool to expect the colon, it also felt like it was running/calling tools faster, but I need to do a larger test to be sure.
reply