Placeholder Image

Subtitles section Play video

  • (upbeat music)

  • - There have been a lot of news about ChatGPT lately

  • like people using ChatGPT to write essays,

  • ChatGPT hitting a hundred million users,

  • Google launching Bard to compete against ChatGPT

  • and Microsoft integrating ChatGPT

  • into all their products, and also the viral sensation

  • of CatGPT where it can answer all of your queries,

  • but as a cat, meow, meow, meow, meow, meow, meow.

  • ChatGPT, if you don't know already, it's a chat bot

  • by OpenAI where you can ask it many things.

  • For example, explaining complex topics

  • like explain why I'm a disappointment to my parents

  • or ask it more technical questions like,

  • how do I inherit more money than my brother from my parents?

  • A lot of people are using it to write essays, draft emails,

  • and even write code.

  • So I tried it myself, of course, as a YouTuber obviously,

  • my first question to it was, who is Joma Tech?

  • And it answered...

  • Are you fucking--

  • You know, ChatGPT has a lot of limitations,

  • like here we ask it to name colors

  • that don't have the letter E in them,

  • and this is what they gave us.

  • Orang, yllow, red, that's clearly wrong.

  • In all seriousness,

  • this is to demonstrate how ChatGPT works.

  • It's a pre-trained large language model,

  • meaning it was trained on text data

  • from the internet until the end of 2021.

  • So it won't know anything

  • about things that happened recently.

  • It doesn't have access to the internet.

  • It'll only predict the answer based

  • on what it has consumed already,

  • and the way it answers your question is

  • by predicting each word that comes next.

  • For example, if you ask GPT who Bard is,

  • it's not going to know.

  • You might ask Joma, didn't your channel launch in 2017

  • and ChatGPT was trained on internet data until 2021,

  • yet it doesn't know who you are?

  • Yeah, so there's actually a technical reason

  • and fuck you.

  • Recently ChatGPT hit a hundred million users.

  • It launched November 30th, 2022,

  • and this article came out February 3rd, 2023.

  • So it took two months to hit a hundred million users.

  • Who are these users and what are they doing with ChatGPT?

  • Well, it's pretty obvious, they're cheating with it.

  • Everybody's cheating such that

  • some school districts have banned access to ChatGPT.

  • If they can write essays, then they can pass exams.

  • ChatGPT was able to pass exams from law school,

  • business school, and medical school.

  • Three prestigious industries.

  • Now, this is why I went into coding

  • because I always thought that law school,

  • business school, and medical school,

  • it was too much about memorization

  • and you're bound to get replaced,

  • it just wasn't intellectual enough, you know?

  • All right, well,

  • I guess engineering is getting replaced, too.

  • ChatGPT passes Google coding interview,

  • which is known to be hard, but I guess not.

  • But note that it is for a L3 engineer,

  • which means it's a entry level, for those not in tech,

  • there's no L2 and L1, it starts at L3,

  • but this does raise questions about ChatGPT's ability

  • to change engineering jobs behind it,

  • and we're already seeing the change

  • as Amazon employees are already using ChatGPT

  • for coding even though that immediately after,

  • they told them to stop, warning them not

  • to share confidential information with ChatGPT.

  • What's happening is they're feeding ChatGPT

  • internal documents, which are confidential,

  • but OpenAI stores all that data.

  • You know, it reminds me of when I used to intern

  • at Microsoft and they didn't let us use Google

  • for searches because they think that they might spy on us.

  • I was like, relax, I'm an intern.

  • I'm not working on anything important.

  • In fact, I actually wasn't working at all.

  • You know, I was playing Overwatch all day,

  • but yeah, anyways, they forced us to use Bing for searches.

  • One thing that's being underreported

  • in mainstream media is the success of GitHub Copilot.

  • It's probably the most useful

  • and most well executed AI product currently out there.

  • Have I used it?

  • No, I haven't coded in forever.

  • Now, here's how it works.

  • The moment you write your code,

  • it's like auto complete on steroids, like this example,

  • it helps you write the whole drawScatterplot function

  • and it knows how to use a D3 library correctly.

  • Another example here, you can write a comment

  • explaining what you want your function to do

  • and it'll write the code for you.

  • Sometimes even the name

  • of the function will give it enough information

  • to write the rest of the code for you.

  • It's very powerful

  • because it's able to take your whole code base as context

  • and with that, make more accurate predictions that way.

  • For example, if you're building a trading bot

  • and you write the function get_tech_stock_prices,

  • it'll suggest, hey, I know you're going

  • through a rough time,

  • but building a trading bot is not going

  • to fix your insecurities and maybe you should just accept

  • that you'll be a disappointment for the rest of your life.

  • Okay.

  • How did all of this happen?

  • Why is AI so good suddenly?

  • The answer is the transformer model

  • which caused a paradigm shift

  • on how we build large language models, LLM.

  • By the way, this diagram means nothing to me.

  • It makes me look smart, so that's why I put it on there.

  • Before transformers,

  • the best natural language processing system used RNN,

  • and then it used LSTM,

  • but then Google Brain published a paper

  • in 2017 called "Attention is All You Need"

  • which is also my life's motto because I'm a narcissist.

  • The paper proposes a simple neural network model

  • they call transformer, which is based

  • on the self attention mechanism

  • which I don't fully understand, so I'll pretend

  • like I don't have time to explain it

  • but I also know that it allows for more parallelization

  • which means you can throw more hardware,

  • more GPUs to make your training go faster

  • and that's when things got crazy.

  • They kept adding more data and also added more parameters

  • and the model just got better.

  • So what did we do?

  • We made bigger models with more parameters

  • and shoved it a shit ton of data.

  • Sorry, I'm trying my best here to make the model bigger.

  • All right, fuck it.

  • Anyway, that gave us ready

  • to use pre-trained transformer models like Google's Bert,

  • and OpenAI's GPT, generative pre-trained transformers.

  • They crawled the whole web to get text data

  • from Wikipedia and Reddit.

  • This graph shows you how many parameters each model has.

  • So as you can see, we've been increasing the number

  • of parameters exponentially.

  • So OpenAI kept improving their GPT model

  • like how Goku kept becoming stronger each time

  • he reached a new Super Saiyan form.

  • While editing this,

  • I realized how unhelpful the "Dragon Ball" analogy was.

  • So I want to try again.

  • To recap, transformer was the model architecture,

  • a type of neural network.

  • Other types of models would be like RNN and LSTM.

  • Compared to RNN, transformers don't need

  • to process words one by one,

  • so it's way more efficient at training with lots of data.

  • OpenAI used the transformer model and pre-trained it

  • by feeding it a bunch of data from the internet

  • and they called that pre-trained model GPT-1.

  • Back then, NLP models would be trained from scratch

  • for a specific task like translation or summarization.

  • Both transformer, we get to pre-train the model first

  • then fine tune it for a specific task.

  • Then for GPT-2, they did the same thing, but more

  • and with a bigger model, hence with 1.5 billion parameters,

  • and then with GPT-3,

  • they went crazy and gave it 175 billion parameters.