The Other NLP

Recently I've been playing with some deep learning software - OpenAI's GPT-2 and GPT-3, and EleutherAI's GPT-J-6G. These are NLP algorithms. No, not that discredited garbage Neuro-Linguistic Programming - in this case NLP stands for Natural Language Processing.

The basic idea of these recent efforts in deep learning is to take a piece of software that has been written to guess the next word in a sequence, and train it on a huge corpus of data. It turns out that the internet is a great source of natural language, and a lot of it is very easy to scrape and feed into one of these algorithms. So these pieces of code are trained on lots and lots of internet text.

This training is very processor intensive, needing thousands of hours of time on modern PCs using specialised AI chips on expensive graphics cards. However, once the algorithm has been trained, the dataset that has been created is just a few hundred megabytes in size, and can be quickly loaded into memory - the training only needs to be done once. At this point the software can be used to predict the next word in a sequence, and can keep doing this - creating whole sentences and paragraphs that actually make grammatical and logical sense. We will see below what these general NLP algorithms can do.

The same software can also be fine tuned by giving it a smaller set of data. Using its ability to put together coherent sentences, the software can emulate the data set it's been fine tuned on. So far I've been working on a couple of fun skeptical projects with this, although I have more ideas.

Anyway, it's probably easiest if I just give you some brief intros and show you the kinds of results I've been getting. Enjoy!

Mark Honeychurch

A couple of weeks ago I talked about the NZ Skeptics putting together a submission for the Ministry of Justice in response to their new Hate Speech proposals. You can read our submission on our ... (251 words)

Category: News

Mark Honeychurch

Using an older algorithm, GPT-2, with a fairly small data set that it had been pre-trained on, I fine tuned this particular piece of software on the entire back catalogue of QAnon posts. Below are 6 ... (295 words)

Category: News

Mark Honeychurch

This one surprised me a little. I fed GPT-2 about a dozen holy texts, with the intention of generating some text that was their distilled, combined wisdom. Instead, what I received each time I ran ... (770 words)

Category: News

Mark Honeychurch

One idea I've been given is to find a large set of Deepak Chopra quotes, and use that to train an algorithm to create nonsense quotes talking about quantum realities and the collapse of the wave ... (72 words)

Category: News