6

Twitter's open algorithm, Auto-GPT, LLMs as "calculators for words", S...

 1 year ago
source link: https://changelog.com/podcast/news-2023-04-03
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.
neoserver,ios ssh client

The Changelog

Twitter's open algorithm, Auto-GPT, LLMs as "calculators for words", SudoLang & stochastic parrots

Changelog News 2023-04-03

Featuring

All Episodes

Twitter publishes (some of) its recommendation algorithm, Toran Bruce Richards puts GPT-4 on autopilot, Simon Willison shares a good way for us to think about LLMs, Eric Elliot creates a powerful pseudocode programming language for LLMs & I define and demystify the term “stochastic parrot”.

Transcript

📝 Edit Transcript

Changelog

Click here to listen along while you enjoy the transcript. 🎧

What up, nerds?! I’m Jerod and this is Changelog News for the week of Monday, April 3rd 2023.

Let’s do it.

On Friday, Twitter Engineering wrote about and open sourced the recommendation algorithm. People who previously decided Elon Musk is very good thought this release was very good. People who previously decided Elon Musk is very bad thought this release was very bad.

Me? I thought it was just okay.

Band joke.

On one hand, it’s not the full algorithm. In fact, the architectural diagram released with the code shows this big cloud neural network in the middle called “Heavy Ranker”, and the training data for that model is not part of the release. On the other hand, there is a substantial amount of code in this release. We’re talking 235 thousand lines of Scala, 136 thousand lines of Java, 22 thousand lines of Python and 7 thousand lines of Rust in the main repo alone.

Lots of weird stuff has already been found, such as a metric tracking code that groups Twitter users into four types: democrat, republican, power user and… Elon.

And some useful stuff has been learned as well, such as how it weights tweet characteristics and engagement. Btw, likes are a 30x boost, retweets a 20x boost, and images/videos are a 2x boost, according to Rowan Cheung, who spent the weekend feeding the code into GPT-4, because of course he did.

Perhaps the best thing this open sourcing produced… a treasure trove of absolutely ridiculous new Issues posted on the project’s repo with titles like “there is bug”, “Twitter is pretty cool, but weird”, “my cat gave birth to 4 kittens”, and “Join the Fediverse”

Speaking of the fediverse… I’m happy to announce I’m open sourcing an approximation of Mastodon’s recommendation algorithm right here on Changelog News: Are you ready, write this down: select * from toots order by published_at desc;

Okay that was far too much non-AI content I am so sorry here comes some large language models. Like, really large ones.

Big fat guy, I’m talking like orca fat.

Oh my, I just had an idea. Maybe we can repurpose all those “yo momma so fat” jokes for large language models and they’ll still be funny without hurting anyboyd’s mommas feelings. Hey ChatGPT, write me 3 funny “yo LLM so large” jokes pronto:

  1. Yo LLM so fat, it needs a search engine just to find its own information.
  2. Yo LLM so fat, it can’t even fit in the cloud; it needs its own weather system.
  3. Yo LLM so fat, it’s the reason for global warming.

That’s just depressing.

Auto-GPT is an autonomous GPT-4 experiment. You provide a name, role, and up to 5 goals, and Auto-GPT will do actions to fulfill the goals — on its own. Not only is this a cool experiment in and of itself, it also has a breakthrough feature: it’s able to write its own Python code and execute it. That means it can recursively debug, develop and self-improve.

Speaking of LLMs, Simon Willison will be back on the podcast this Wednesday. We chatted with Simon for a full two hours and honestly I think this episode’s even better than the Stable Diffusion discussion we ad with him in September.

After our chat, Simon wrote up his thoughts on how we should be thinking about LLMs. He calls them “calculators for words”, saying “This is reflected in their name: a “language model” implies that they are tools for working with language. That’s what they’ve been trained to do, and it’s language manipulation where they truly excel.”

This makes LLMs very good at a specific set of things, including: summarization, question answering, fact extraction, rewrites, and suggesting titles.

Eric Elliot recently wrote about and released SudoLang: a powerful pseudocode programing language for LLMs. He didn’t write it himself, though, because that would be so passé.

I’m so 3008. You’re so 2000 and late.

SudoLang has commands, modifiers, and arguments. The spec is available on GitHub. All you have to do is paste it into ChatGPT with GPT-4+ running and you’ll be off to the races. About the project Eric says, “AI turns everyone into a conjurer, and we have only witnessed the tip of the iceberg.”

Have you heard people use the term “schostic parrot” when describing LLMs? I heard it a few times on our Practical AI podcast and didn’t really know what it meant, so I looked it up so you don’t have to. When people use this term in reference to a language model, they’re referring to the limitation of the model’s understanding of the words it is spitting out.

“stochastic” refers to the random element in generating responses, and “parrot” highlights the notion that these models merely repeat what they have seen during training, like parrots do when you teach them to speak words.

The term comes from a 2021 paper called: On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?

Dr. Jo on Quora summarizes it as: “a stochastic parrot has form but no underlying model. We ourselves create the illusion of an underlying mind.”

That is the news for now. We’ll be back in your ear holes on Wednesday with our conversation with Simon Willison, it’s a good one. Have a great week, share Changelog News with your friends if you dig it, and we’ll talk to you again next time.

Changelog

Our transcripts are open source on GitHub. Improvements are welcome. 💚


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK