LLama.cpp now has a web interface
I put together a simple web-chat that demonstrates how to use the SSE(ish) streaming in the server example. I also went ahead and served it from the root url, to make the server a bit more approach...
There is a discussion on Hacker News, but feel free to comment here as well.
Hey that's pretty cool, thanks for sharing! I'll give it a spin this evening. I think this will make the project much more accessible.