The 10 Trillion Parameter AI Model With 300 IQ
The hosts consider what a world with ultra-intelligent models would look like and what potential unlocks could be made possible.
Transcript
If o one is this magical, what does it actually mean for founders and builders? One argument is it's bad for builders because maybe o one is just so powerful that OpenAI will just capture all the value. You mean they're gonna capture a light cone of all future value? Yeah. They're gonna capture a light cone of all present, past, and future value. Oh my god.
The alternative, more optimistic scenario is we see ourselves how much time the founder spend, especially during the batch, on getting prompts to work correctly, getting the outputs to be accurate. But if it becomes more deterministic and accurate, then they can just spend their time on bread and butter software things.
The winners will just be whoever builds the best, like, user experience and gets all these, like, nitty gritty details correct. Welcome back to another episode of the Light Cone. We are sort of in this moment where OpenAI has raised the largest venture round ever, $6,600,000,000 with a b. Here's what Sarah Fryer, the CFO of OpenAI said about how they're gonna use the money.
It's compute first, and it's not cheap. It's great talent, second. And then of course, it's all the normal operating expenses of a more traditional company. But I think there is no denying that you are we're on a a scaling law right now, where orders of magnitude matter. The next model is going to be an order of magnitude bigger, and the next one on and on.
And so that does make it very capital intensive. So it's really about orders of magnitude. Let's live in the future. There's 10,000,000,000,000 parameters out there. 10,000,000,000,000 parameter large language models. Two orders of magnitude out from the state of the art today. What happens?
Like, are people actually gonna be throwing queries and actually using these 10,000,000,000,000 parameter models? Seems like you'd be waiting, you know, ten minutes per token. Yeah.
For a bit of context right now, the frontier models, I mean, they're not public exactly how many param parameters they have, but they're roughly in the five hundreds of billions ish, like, LAMA three, four zero five billion. Anthropic is speculated to be 500,000,000,000. GPT four o, roughly about around that much. Getting to 10,000,000,000,000, that's a two order magnitude. Right?
I think the type of level of potential innovation could be the same leap we saw from GPT two, which was around 1,000,000,000 parameters that was released with the paper of a scaling loss, which was one of these seminal papers that people figure out, okay. This is transformer architecture that we figure What if we just throw a bunch of engineering and just do a lot of it?
Where does this scale in this logarithmic type of scaling? Then this was proved out when GPT 3. 5 or three got released. That was about a hundred billion parameters. So that's like that two order magnitude, and we saw what happened with that. That created this new flourishing era of AI companies, and we saw it.
We experienced this back in 2023 when we started seeing all these companies building on top of GBD 3 point 5 that was starting to work, and it created this giant wealth. So we could probably expect if this scaling law continues, the feeling will be similar to what we felt from that year of transition in 2022 to 2023. Yeah. That was the moment when everything changed.
So that would be pretty wild if that happens again. I think there's one interesting aspect to this, which is clearly the current generation state of the art models that are available, especially given o one chain of thoughts, they sort of basically rival normal intelligence. Like, you could make a strong case that AGI is basically already here.
The majority of the tasks that, you know, 98% of knowledge workers do day to day, it is now possible for a software engineer, probably sitting in front of Cursor, to write something that gets to, you know, 90 to 98% accuracy and actually do what a null a human knowledge worker with a 20 IQ would be doing all day. And that's sort of writ large.
Like, there are probably hundreds of companies that each of us have worked with over the past few years that are literally doing that day to day right now. You know, the weird interesting question is, like, at 10,000,000,000,000 parameters at, you know, 200 to 300 IQ, like, sort of ASI beyond what a normal human being normally could do, you know, what does that unlock?
There's an a great article in The Atlantic with Terence Tao, sort of famously this Taiwanese mathematician who is, like, literally north of 200 IQ, and how he uses ChatGPT right now, and it's sort of unlocking new capabilities for him. There are some examples of this happening, you know, quite a few times in human history. Like, you could argue that nuclear power was that fission.
You had to actually model theoretically that something like nuclear fission was possible before anyone, you know, experimentally tried to do it for your transforms. Yeah. Maybe the thing is if we think a lot of the capabilities right now are here, but it's not evenly distributed, if you go walk down the street and you talk to the random Joe, they don't feel the AI.
They're just living their normal life, and stuff is still just normal. It hasn't changed. But I think the counterexample is just sometimes these discovery take time for it to really pan out. This is example we're discussing. Fourier transform was this mathematical representation that Joseph Fourier discovered in the eighteen hundreds.
That was like a seminal thesis that he wrote about representing series of functions that were repeating in periods. And before Fourier transform, they were written as these long sums, series of sum that are very expensive to add them up and figure out how to really model the equation, basically.
But he found this very elegant way that instead of just doing sums of series, you could basically collapse all these math function into sines and cosines wave that only need two variables, basically, the amplitude and the period, and you could represent every periodic signal and function. I mean, it's not really cool math, which is like how some of this LLM and use case sounds like, okay. Cool.
They can do all this coding. But four year transform, it took another hundred and fifty years until the nineteen fifties. When people figure out what to do with this, it turned out that four year transform were super good at representing signals. And we need signals basically to represent everything in the analog world to be digital because bits are ones and zeros, and how do you compress that?
And one of the big applications as well is radio waves and made telecommunication a lot more efficient, image representation, encoding, information theory. It just unlocks so much of the modern world, like the Internet and cell towers work because of this theory. But it took a hundred and fifty years until the average Joe could feel the Fourier transform. Interesting. That's a really powerful idea.
I mean, that took a while then. I mean, apparently, in the nineteen fifties, that's the moment that color TV happened. So Unlocked by Fourier Transformers as well. That's right. If you apply it to the AI stuff that's happening today, though, it's like, one, where do you start the clock ticking from?
Like, it's not clear if you start it from the chat GBT three moment two years ago or from just like all of the research that's been going on for decades. Like, we might actually just be we've talked about this before, but we might actually be like decades into this now, it's starting to hit like the inflection moment potentially. Yeah. For sure.
I mean, if we run with Diana's example of fast Fourier transforms, like all the math that's underpinning all the this new AI stuff is linear algebra stuff that's like a hundred years old. Yeah. It just turns out how far you could push it. Probably have all the GPUs to compute it.
I guess that's one potential way that these 10,000,000,000,000 parameter models actually alter the face of what humans are capable of. Like, they sort of unlock something about the nature of reality and our ability to model it. And then somehow it leads to either nuclear weapons or the color TV.
The other big thing is just because this is all in software, like, compared to, like, Fourier transforms, like, a lot the applications we're seeing physical devices. Right? Like, record players or or or telephones, like you said. And so it takes a while for your technology to get adopted because you have to, like, buy your update device and all these things.
Now we have, like, Facebook and Google who have, like, you know, pretty decent percentages, like, the world using their software already, like, as soon as these things start rolling out. And I feel that's another thing that's starting to be noticed is Meta in particular coming out with the meta RayBands, the consumer, like, device.
Like, I think consumers, once this becomes something that's, like, visual in your, like, smart glasses plus, like, a voice app that you can talk to and it, like, is indistinguishable from a human being, like, that's gonna be a real change the world moment for me. They'll start feeling the AI once they can, like, talk to it all the time.
I mean, it seems like there's really a bifurcation in what we might expect when we have this capability. At the extreme end, you're going to have people like Terence Tau pushing the edge and boundary of our understanding of our, you know, modelable world.
And then, you know, maybe that's actually worth tens or hundreds of millions of dollars of inference to run these 10,000,000,000,000 parameter models. And then the more likely way this ends up being useful for the rest of us is actually in distillation.
So taking, you know, there's some evidence that, for instance, Meta's four four zero five b was mostly use useful to make their 70,000,000,000 parameter model much much better. And so and you actually see this today. There is sort of this moment there where we thought that, know, people might just go to GPT four and distill out all the weights.
And it seems like there's some evidence that certain governmental entities are doing that already. But GPT four itself and four you know, it became four o. OpenAI itself has now enabled distillation internal to its own API. So you can use o one, you can use even GPT four or four four o to distill it down into a much cheaper model that's internal to them, like g p four four o mini.
And that's sort of their, you know, lock in capability. Yeah. I don't think this is talked much about, but it is interesting that you have these giant models, the 400 or 500, whatever, billion parameter models that are basically the teacher models because they're the mega train with everything and took forever.
And they are the teacher model, master model that teaches a student model, which are these smaller ones that are faster and cheaper because doing inference for a four or five billion parameter model is very expensive. So we have evidence that all these distillation models are working. Companies in the batch, the they're building from the latest and greatest.
They're not going for the giant model with all of the parameters and give me the biggest thing to do so that it works the best. We have evidence that's not the case. People are not going for the big model, and we actually have stats in the batch. I mean, Harsh, we kinda talked about them. Yeah. Jared ran some numbers on this, and it's it's fascinating.
But I think the bigger meta point is even the fact that talking about the startups or the founders building this stuff are choosing, like, the smaller models versus the bigger models. I just have choice. And even, like, a year ago when this, like, entire industry started existing, like, everything was built on top of ChatGBT. Right? There was a was a % market share, the ChatGPT rapper meme.
And I feel like we've, especially over the last six months, seen people start talking about the other models, like Claude and Sonnet being sort of this word-of-mouth for almost being better at cogen than chat GBT, and people are just starting to use different models.
And so the numbers that Jared ran for the summer twenty four batch are fascinating because it seemed that that trend has just continued. Like, we have more diversification of LLMs and models that developers are building on top of.
And some of the stuff that really stood out is Claude has even just in six months from the winter batch to summer batch has gone from, like, 5% developer market share to, like, 25%. Of companies in the batch. Yes. Of companies in the batch, which is huge. That's like I've never seen a jump like that. Right. Larm has gone from 0% to 8%.
Like One thing that we know from running YC for a long time is that whatever the companies in the batch use is a very good predictor of what, like, the best companies in the world are using, and therefore, what products will be most successful.
A lot of YC's most successful companies, you could have basically predicted which ones they would be based on just looking, basically, just running a poll of what the companies in the batch use. If we just take a take OpenAI's latest fundraise off the table and the latest, like, the o one model off the table for a second, it would seem like amongst developers and builders, OpenAI was losing.
Like, they went from being the only game in town to just, like, bleeding market share to the other models at a pretty rapid rate. The interesting thing, though, is maybe they are coming back. Like, what's the stat that you pulled? It seems like or 15% of the batch are already using o one, even though it's not, like, fully available yet. Yeah. O one is only, like, two weeks old now. Yep. Yeah.
And we're seeing some interesting things with o one. We're actually hosting right now in person right now as we speak downstairs a hackathon to give y c companies early access to o one, and Sam himself was here. Did the kick kickoff. There's a bunch of open AI researchers and engineers working on it.
And it's only been about four hours of hacking, and we already heard of I already saw actually some demos as I was walking by to see some teams, and they already built things that were not possible before with any other model. Do you do you have some examples? One of the companies I'm working with is Freestyle.
They're building a cloud solution fully built with TypeScript with if you're familiar with durable objects with this really cool framework to that that makes front end and back end seamless to develop, and it's really cool to use. What was cool about them is they've just been working on it for a couple hours, and I saw a demo that was mind blown.
They basically got a version of Replit agent working with the product. All they had to prompt o one with was all their developer. Oh, sure. Some of their developer documentation and some of their code, and they could just prompt it, build me a web app that writes a to do list or this, and it would just boom. Just work.
And it was able to reason and inference with the documentation and took a lot longer, but it arrived and built the actual app. What's interesting for us to talk about is if o one is this magical, what does it actually mean for founders and builders?
And one argument is it's bad for builders because maybe o one is just so powerful that OpenAI will just capture all the value and everything that could be valuable and built on top of this stuff will just be owned by them. You mean they're gonna capture a light cone of all future value? Yeah. They're gonna capture a light cone of all present, past, and future value. Oh my god.
The alternative more optimistic scenario is we see ourselves how much time the founder spend, especially during the batch, on the tooling around getting the problem, like, getting prompts to work correctly, getting the outputs to be accurate, human in the loop. Like, all of this time spent just getting the core product workings is not deterministic.
But if it becomes more deterministic and accurate, then they can just spend their time on bread and butter software things, you know, like better UI, better customer experience, more sales, more relationships.
In which case, it's like, it may be a better time to start now than ever because you don't even have like, maybe all of the knowledge you learn around how to get, like, the prompts accurate and working was just temporary knowledge that's no longer relevant as these things get more powerful.
Actually, we had this conversation with Jake Heller from CaseText where getting the legal copilot to work to % was a huge unlock. And it was really hard. Yep. He, like, you know, we hated this whole talk about all the things he had to do to actually get the thing to be accurate enough. Yeah. Imagine if he didn't have to do any of that.
If if they just on day one, you could be guaranteed a % accuracy. They as though you're just building a web app on top of a database, the barrier to entry to build these things goes way down. There's gonna be more competition than ever, and then it will probably just become look more like a traditional winner takes all software market. Garrett has an example. So there's a company, Drymerge Yeah.
That you work with, and they went from 80% accuracy to pretty much a 99 or, for intents and purposes, a % using a one and unlocked a bunch of thing. You wanna talk about them? Yeah. Just by swapping out GPT four zero for for for o one.
I think there might be an even more bullish version, Harge, which is that there are use cases right now that people are not able to use LLMs for because even though they're trying to get the accuracy high enough, they just can't get it accurate enough Mhmm. For to actually be rolled out in production.
Like, especially if you think about, like, really mission critical jobs where the consequences of mistakes are dire, like, pretty hard to use LMs for that. As they keep getting more accurate, those applications will start to actually work. I guess there is a lot of evidence inside the YC greater portfolio. You know, I was meeting a company from 2017. I think I tweeted about them.
They were, you know, $50,000,000 annualized revenue at that point, but growing 50% a year. A year or two ago, they were not profitable.
They knew that they needed to raise more money, but in the year since, they automated about 60% of their customer support tickets, and they went from something that needed to raise another round imminently to something that was totally cash flow breakeven while still growing 50% year on year.
That's sort of like the dream scenario for building enterprise value because you're big enough that, you know, you're a going concern, and then you're literally compounding your growth with, like, no additional capital coming in.
So it's companies like that that actually end up becoming, like, half a billion, a billion dollars a year in revenue and, like, driving hundreds of millions of dollars in free cash flow. I mean, that's sort of the dream for founders at some level, and I think that that's one of the more dramatic examples that I've seen thus far. And I think it's sort of not an isolated case.
You know how we're sort of talking it's, you know, 2024 now, and we're still in this overhang moment where companies sort of on this path raised way too much money at, you know, 30 x or 40 x, you know, next twelve months multiple revenue, seemingly struggling, but, you know, also never going to raise another round.
Like, this is actually pretty good news for them because they actually can go from, like, not profitable to, you know, breakeven to then potentially very profitable. I think that narrative is not out there, and I think it's really, really good news for founders. I've already started to catch attention. Didn't know.
Klarner, CEO, got a lot of attention a few weeks ago for I mean, it's not unclear how much of it is real or not, but at least they're pitching that they're just, you know, replacing their internal systems of records for HR and sales with home built or LLM create apps, at least was like the insinuation. Yeah. What is it? They got rid of Workday. Yeah. That was it. Yeah. That's pretty wild, honestly.
I mean, so that's good. If you treat OpenAI as the Google of the next twenty years, you want to invest in Google and all the things that Google enabled like Airbnb. Google could do Airbnb. It probably won't. Yeah. Just from, like, I don't know, Coase's theorem of the firm, probably. It's just, like, too inefficient and too difficult. It requires too much domain expertise to actually pull that off.
So what are sort of the new Googles that are getting built out? There's these vertical agents. What are some examples that we'll have that we could talk about? I loved working with this company called TaxGPT from the last YC batch. They started off actually really literally a rapper and, like, you know, it's in the name, TaxGPT.
But my favorite example about them is, like, you know, it turned out that tax advice, you know, doing basic rag on you know, it's sort of like case text, actually. It was, you know, being able to do rag on existing case law and existing policy documents from the IRS or internationally.
That was just sort of the wedge that got them in front of, you know, tens of thousands of accountant accountants and accounting firms. And now what they're doing is building an enterprise business on document upload. So you sort of, you know, get them for cheap or free for the thing that people are googling for.
And then once they know about you and trust you, you get like this 10 or hundred thousand dollar a year ACV contract that then takes over real workflow that, you know, actually extinguishes tens to hundreds of hours of work per accountant. And another interesting thing about the o one model is we were just saying, originally, ChatGPT was the only thing you could build on top of.
OpenAI was the only game in town. Then there were all these models. I think the sort of alpha leak we have here, like, right now in this room is downstairs, people are building at the cutting edge of o one that even the public doesn't have access to. And what we're seeing is that this is a real major step forward.
Like, o one is going to be a big deal for any programmer engineer who is building an AI application. The interesting thing is, will this cycle repeat where it will give OpenAI a temporary lead, their market share will just, like, go, you know, back up towards a %. But then within six months, Lama will be updated. Claude will come up with its new release.
Gemini will keep getting better, and they'll just be like, you know, four different models that have equivalent levels of reasoning. Or will this be like the first time OpenAI has a a true breakthrough? And I was just defining true breakthrough as something that's actually defensible. Like, if no one else can replicate it, then that puts them in a really powerful position. But we don't know.
And I think that's what's interesting. It's like OpenAI seems like it is continually the one pushing the envelope, but they always seem to be the first ones to make major breakthroughs, but they have never been able to maintain the lead so far.
I think the only thing that's interesting about o one is that it makes a lot of the GPU needs even bigger because it's moving a lot of the computation needs a lot higher for inference because it's taking a lot more time to do a lot of the inference.
So I think it's gonna change also a lot of dynamics underneath for a lot of the companies building AI infrastructure as well, which is something food for thought. Seems like there are two different types of use cases. I believe they did just enable distillation from o one into four o.
And so it's conceivable that for relatively rote and repeating use cases, you could just sort of use o one for the difficult ones, and then you distill it out and then you pay four o or four o mini prices from there. And then there again, there's this other type of problem that is like very specific.
I mean, imagine most many code gen situations are a little bit more like that, where you need to pay like for the full o one experience because it's, you know, fairly detailed and specific. Depends who you're building for too. Right?
If you're a enterprise software and you can pass the cost on to your customer and they can tolerate at a higher latency and don't care as much about it being instant, then you can just use maybe o one a lot. If you're putting, like, consumer apps, probably not. But talking of consumer apps, mean, the other thing that was striking about OpenAI's release releases like this real time voice API.
Super cool. It's pretty remarkable. And I think the the most telling thing to me is that the ongoing usage based pricing is $9 per hour. And it sort of points to a sort of powerful thing.
Like, if I were a macro trader, I would be very very bearish on countries that have that rely very heavily on call centers right now because, you know, $9 an hour is sort of right there at what a call center would cost. This is another thing we're definitely seeing within the batch. Right? Like, it's clear that voice is a almost like a killer app.
Like, arguably, like, there's a company I just worked within this batch. I am they're in my group at least, Domu, just do sort of AI voice for debt collection, and their traction is just phenomenal. It's working incredibly well. A whole bunch of the voice apps in s twenty four were just, like, some of the fastest growing, like, just, like, explosive companies.
It's it's a clear trend for s twenty four. And I remember working with companies in the prior two batches that tried to do voice, and it just, like, wasn't working well enough. Yeah. Like, the latency was too high. Exactly. Like, it got confused if, like with interruptions and things like that. And it's, like, just turned the corner where it's, like, finally working.
There's another company I work with, Happy Robot, that landed on this idea that was doing a voice agent for coordinating all these phone calls for logistics. Think of a truck driver that needs to go from point a to point b. These are all, like, people just calling to check where you are. There's no, like, there's no, like, find my friends for it. And they started getting a lot of usage on this.
And I think we talked a bit about this before that at this point, AI has passed Turing tests and is solving all of these very menial problems over the phone. That's pretty wild.
I guess one thing that is maybe under discussed is to what degree the engineering teams that are in in these sort of incumbent industries, it feels like it's pretty binary either, you know, the vast majority of companies and organizations, especially the ones that were maybe founded four or more years ago, they actually don't take any of this seriously.
Like, have literally no initiatives on this stuff. And I sort of wonder how generational it is. Like, I'm realizing that eng managers and VP of eng, like, they're probably my age now. I'm 43 now. And, you know, if I wasn't here seeing exactly what was happening, I would be sort of tempted to say, like, this is just the same old thing, AI. Yeah. Yeah. Yeah.
But I I think it's the rate of improvement that people don't get if they're not as close to it as we are. I just think your average corporate enterprise person is certainly used to technology disrupting things, but over pretty long time frames. And if anything, they become cynical. So like, oh, the cloud. Like, cloud was such a buzzword for a long time.
It totally did change our enterprise software is build and delivered, but it took like a decade or so. And so I I suspect everyone's feeling that way about AI. It's just your natural default mode is to be cynical. Oh, yeah. Like, it's not gonna be ready for a while.
And then probably if you looked at this stuff even six months ago, like we were just talking about, if you looked at an AI voice app six months ago Oh, yeah. You're like, oh, this is this is years away from being anything that we need to take seriously. It's like, actually, like, three to four months later, like, it's like it's hit some real major inflection point.
And I think that's what takes even people within tech, it's surprising all of us how quickly this stuff is moving. It's the fastest any tech has ever improved, I think. Yep. Certainly faster than processors, certainly faster than the cloud. And it's kind of fun to actually watch. It's been remarkable to actually see another example of this in the batch.
So a lot of the technical founders, sometimes I sit with them and I just watch how they code, the before and the after before all of this wave of AI. Just standard. You have your IDE and things on the terminal. People ship fine. But the demos and products that we're seeing founders build during the batch is like a next level of polish.
And when you see them and see them code, it's like they're living in the future. They're really not just like at this point, GitHub Copilot is already kinda old news. They're using the latest, greatest coding assistant. A lot of them perhaps using something like continue or Cursor. Right? But this is something Jared pulled out as well when we asked the founders Oh, the IDs. Right? Yeah.
Yeah. We surveyed the summer '20 '4 founders, and half the batch is using Cursor compared to only 12% that's using GitHub Copilot. That was surprising to me. They're not even using the fully agentic coding agents like Replit. Haser's still sort of like Copilot phase stuff.
But even just going from, like, GitHub copilot to Cursor, which is, like, the next step up in terms of, like, how much the actual AI does, is this, like, incredible breakthrough. They ship very quickly. I mean, this is evident today in the hackathon. I was impressed with what they built. I was looking at their editors, Cursor. It's like, cool.
It's another sign where, like, the founders have the advantage. Right? Like, it feels to me again that when GitHub Copilot first came on the scene, it seemed it's GitHub plus Microsoft. It has all the developers, plus it has all the capital, plus it has all the access to, like, the inside track on OpenAI. How could any coding IED compete with them? It will just get subsumed.
And my cursor has come out of nowhere and is, like, you know, according to our numbers, like, five times the size of, like, GitHub Copilot within the batch. Which again, like you were saying earlier, was like, the the startup founders are actually usually the taste makers on this kind of thing.
I think there's certain types of businesses where it doesn't make sense to maybe go after startup founders as your early customers. But for developer tools, it definitely does. Like Stripe, AWS, both wanted to own YC batches in particular, and that worked out really well for them.
So it's probably a really good sign for Cursor, honestly, that they have, like, such good penetration within the YC batch. Yeah. I I would definitely say Cursor is pretty awesome, but AltaVista was awesome too. Yep. I remember using that as a search engine, and there was another version, and the next version was 10 times better. And so this is the way it's gonna go.
I mean, which the only people who win are actually developers because of all this competition. So I think, again, takes us to, like, the optimistic view of all of this stuff, which is as the models get more powerful, the winners will just be whoever builds the best, like, user experience and gets all these, like, nitty gritty details correct.
And so that's why Cursor can be GitHub Copilot that has all the advantages. Altavista is a great example. Like, there's still like, Google still came along and crushed them. Right? So there's still room for someone to keep doing to Cursor what Cursor has done to GitHub Copilot. So let's get back to 10,000,000,000,000 parameters.
What world do you think we will live in with this made real, with ASI or something approaching it? You know, what will humans actually do, and how much more awesome will it be?
Well, I'll give a steel man for a really bullish case, which is that the thing that is holding back the rate of scientific and technological progress is arguably the number of smart people who can actually analyze all the information, that we already know about the world. There's millions of scientific papers already out there, an incredible amount of data, but, like, try reading all of it.
It's far beyond the scale of any human's comprehension. And if we make the models smart enough that they can actually do original thinking and deep analysis with correct logic, and you could let loose an infinite a near infinite amount of intelligence on the near infinite infinite amount of data and knowledge that we have about the world.
You can just imagine it just coming out with just crazy scientific discoveries, room temperature fusion, room temperature superconductors, time travel, flying cars, all the stuff that humans haven't been able to invent yet. Like, with enough intelligence, maybe we'll finally invent it all. Sign me up for that future. Sounds great. I totally agree with you.
I think, you know, what this might be is not merely a bicycle for the mind. It might actually be a self driving car or even crazier, maybe a rocket to Mars. So with that, we'll see you guys next time.
✨ This content is provided for educational purposes. All rights reserved by the original authors. ✨
Related Videos
You might also be interested in these related videos