Skip Navigation

You're viewing a single thread.

40 comments
  • LLMs are quite impressive as chatbots all things considered. The conversations with them are way more realistic and almost as funny as the ones with the IRC markov chain my friend made as a freshman CS student.

    Of course, out bot's training data only included the IRC channel's logs of a few years and the Finnish Bible we later threw in for shits and giggles. A training set of approximately zero terabytes in total.

    LLMs are less a marvel of machine learning algorithms (though I admit they might play a part) and more one of data scraping. Based on their claims, they have already dug through the vast majority of publicly accessible world wide web, so where do you go from there? Sure, there are a lot of books that are not on the web, but feeding them in the machine is about as hard as getting them on the web to begin with.

40 comments