0 • 0 Ratings
🗓️ 22 February 2024
⏱️ 11 minutes
🧾️ Download transcript
In this episode, we discuss the future of DIY AI with Stanford's ChatGPT clone, exploring how it empowers citizen data scientists and enthusiasts to experiment with AI-driven solutions and contribute to the advancement of machine learning research.
Invest in AI Box: https://Republic.com/ai-box
Get on the AI Box Waitlist: https://AIBox.ai/
Click on a timestamp to play from that location
0:00.0 | With companies like Microsoft investing over $10 billion into chat GPT, it kind of begs the question |
0:07.8 | how valuable is one of these large language learning models. |
0:12.4 | So today on the podcast, we're going to talk about something very interesting that recently |
0:16.4 | came out, which is the fact that some researchers at Stanford said that they have essentially been |
0:21.4 | able with some new technology that came out to clone chat GPT for less than $600. |
0:27.0 | Now, when we're looking at, you know, Microsoft, you know, in 2017 investing a billion dollars, |
0:32.3 | getting this kind of technology kicked off, and then just recently investing $10 billion into |
0:37.2 | the company to further |
0:39.4 | its efforts. It's pretty interesting that, you know, people are claiming they're able to duplicate |
0:43.3 | this for $600. So on the podcast today, we're going to dive into what exactly happened and what |
0:48.0 | this is going to mean for AI in the future. So first off, Stanford is calling it Alpaca AI. |
0:53.9 | And essentially, what they did was Facebook recently, or meta, they recently came out with what's called Lama model. So essentially it's Facebook's like open source language model. It's pretty much the smallest and cheapest of all the Lama models available. |
1:13.3 | And it's pre-trained on a trillion quote-unquote tokens, which tokens is not necessarily words. |
1:20.5 | It's kind of like bits of words or half. I don't know. It's just like how AI models measure what is being spit out and what it was trained on. So like one long |
1:29.2 | word might be like three tokens and like a short word might be one token. Anyways, not that important. |
1:34.5 | So pretty much this model has a certain amount of capacity baked into it, but it would leg |
1:40.0 | significantly behind chat GPT with most tasks. And the number one area is cost. |
1:46.3 | So the biggest competitive advantage in the GPT model comes largely just from its |
1:51.3 | the enormous amount of time and manpower that OpenA has put into the post training. |
1:56.3 | So it's one of the things, for example, it's just like one thing to have read a billion books, |
2:03.1 | but it's another to have, you know, chewed through large quantities of questions and answers, |
2:09.0 | conversion pairs that teach an AI what their actual job will be, right? |
... |
Please login to see the full transcript.
Disclaimer: The podcast and artwork embedded on this page are from In Machines we Trust, and are the property of its owner and not affiliated with or endorsed by Tapesearch.
Generated transcripts are the property of In Machines we Trust and are distributed freely under the Fair Use doctrine. Transcripts generated by Tapesearch are not guaranteed to be accurate.
Copyright © Tapesearch 2025.