Professional Documents
Culture Documents
Better AI Models, Better Startups - 英语 (自动生成)
Better AI Models, Better Startups - 英语 (自动生成)
Better AI Models, Better Startups - 英语 (自动生成)
bunch of startups waiting with baited breath to see whether open AI is going
to kill their startup this is actually a really uh crazy moment for all startups
adding more types of modalities and more capabilities uh per model the the better
off every startup is you have to be on top of these announcements and be kind
of know what you're going to build in anticipation of them before someone else
does versus being worried about open AI or Google being the ones to build
the light cone I'm Gary this is Jared Harge and Diana and we're some of the
group Partners at YC who have funded companies that have gone on to be worth
interesting moment in the innovation of large language models and that we've
seen a lot of really new tech come out just in the last few weeks whether it's
GPT 40 it's uh Gemini 1.5 Harge how are you thinking about you
know what does it mean for these models to be so much better anytime I see a new
announcement from one of the big AI companies with the release of a new
model the first thing I think about is what does this mean for the startups and
in particular YC startups and when I was watching the open AI demos it was pretty
clear to me that they are really targeting consumer like all of the demos
were cool consumer use cases and applications which makes sense that's
kind of what chat gbt was was a consumer app that went really viral I just wonder
what it means for the consumer companies that we're funding and in particular
like how will they compete with open AI for these users what did you think like
even if we take it back like how do consumer products win from like first
principles like is it more about the product or the distribution and how do
you compete with open AI on either of those things yeah that's a great
question I mean I think ultimately it's both and then uh how I want it to be is
that the best product wins uh how it actually is is Whoever has the best
actually think we're at sort of uh in this moment where the better the model
becomes if you're already using four and suddenly four you know you can uh change
one line of code and suddenly be using 40 uh you basically just get smarter by
default every generation and that's really really powerful it means that you
I think we're entering this moment where the the IQ of these things is still you
know four is arguably around 85 it's not that high and then if the Next
Generation if CLA 3 really is at 100 or you know the next few models end up
being closer to you know 110 120 130 this is actually a really uh crazy
moment for all startups and uh the most interesting thing is like uh adding new
capabilities so having the same model be great at coding for instance uh that
means that you know you might have a breakthrough in reasoning not through
just the model reasoning itself but you could have the model actually write code
and have the code do better and even right now it seems like there's um a lot
of evidence that if instead of trying to prompt the model to do the work itself
you have it right code and you execute the code it can actually do things that
reasoning alone could not do so adding more types of modalities and more
capabilities uh per model the the better off every startup is I mean the cool
thing about uh 40 is that you can get better structure output in this
particular case they are better at getting Json which is getting signs of
getting large language models not just outputting English but more language for
computers so that you can build even better applications on top which is
signaling that this better model can be better for startups and make it easier
to integrate because one of the challenges for startups has been always
coing LMS to Output the right thing so you're actually process it in regular
business Logic the other thing I kind of thought about when I was looking at the
demos is as it relates to startups if only one of these companies has the most
powerful model by some distance then that is indeed bad for startups because
you have to depend on them being friendly and having like a nice API for
you're much safer off as a startup it was funny maybe coincidental maybe not
that like open AI announcement was like what two days before one day one day
before Google's right um what's the difference between the so under the hood
the way that GPT 40 works and then Gemini 1.5 works and do you have any
thing about 40 why was so interesting it was adding the speech modality and also
video processing on top of a text and the way they do that is still primarily
a text based Transformer model underneath basically GPT 4 and what they
done is bootstrap and added modules so that it has different caths to handle
this different type of data open AI famously also implemented and launched
recognition and probably that's what they're doing they took the architecture
of a whisper and then bolted it into GPT 4 and they also bolted dally and they
combined these and that became 40 so this is why in terms of the reasoning
capabilities 40 isn't better per se than four by any margin so it's how it works
it's kind of adding modules how they describe it on the white paper the
difference versus Gemini 1.5 which actually on the technical aspects and
merits I'm actually more excited by the Gemini one I know it's counterintuitive
because 40 and open AI has captured the zist of everyone and they're so good at
the demos Right singing Happy Birthday a bit off key that's like so humanid happy
birthday to you happy birthday to you happy birthday dear Jan happy birthday
in terms of demo but in terms of reading their white paper what's interesting
of expert and that is a technique that's new where they actually train from the
image audio and the whole network activates a specific path for these
different data types so instead of um the open AI model that has like kind of
modules this one truly is a one all model and what it does is different
parts of the network activate depending on the data input so it becomes very
energy efficient and I think the reason why uh Google was able to do it is
because they have the engineering Hammer they have tpus where they can really
afford to put a lot of data because it's very expensive to put not just all Texs
image and video and train this giant thing in a distributed cluster they have
TPU is like their I think it's their fifth generation now and it's pretty
cool what they done is that the first big model release that's using mure
experts I think they talked a bit about it on the previous BN but everyone was a
bit uh disolution after the demo of the duck was not real it is a duck yes but
this one was described better I mean the interesting thing is that I think this
time they learned their lesson and I think is actually working yeah and the
other cool thing about Gemini is uh it has a contact window of a million tokens
which is huge the GPD 40 is 128,000 so imagine what you can do with that
because that's about like five books of 500 words or more and the cool thing
about the Gemini 1.5 was their white paper has to saying that on Research
they proved it to work on a 10 million token window which brings a question for
all of you what does that mean for startups especially a lot of the
building tooling around rack which is a whole industrial right now maybe they
become obsolete what do you all think about that I feel like the people who
care a lot about data privacy and whe the data is stored are still going to
want some sort of rag system right like they want the data stored somewhere they
control it versus all in the context window it's not clear that that's going
to be the biggest part of the market like in general people who care this
much about any behind the scenes architectural thing tend to be like
early adopters but not like Mass Market consumer so my guess is people just want
like a massive context window because then you can start building the kinds of
consumer apps people are excited about right like the assistant that just has
of your like personal emails at it that's like a project that the hobbyists
on Reddit are doing a lot of is just try and get like your personal AI That's got
all the information on you but if you had like a infinite context window you
would need to do all of that I think you'd still need rag to be able to sort
of uh store everything and that's like sort of the long-term permanent memory
and then what you actually want is a separate workflow to pull out the
interesting things about uh that user and their intentions and then you
actually have a little like summary bullet point of things that you know
about the user you can actually kind of see some version of this even now in
chat GPT if you go into the settings under 40 it actually now has a memory
and so you can actually see a concrete version of this inside chat GPT I was
just using it to sort of generate some like where's Waldo uh images for my son
and uh it wasn't quite doing what I want wanted it kept using like making like
really deformed faces so I kept like prompting it back to back I was like no
no no I really want no deformed faces and then for a while it was like uh I
said I wanted a red robot in the corner and it kept making uh all of the
characters like various forms of red and I said no no no I really don't want you
to do it and I you know sort of repeated it four or five times and then I went
and looked at my settings and it was like Gary really doesn't want deformed
faces in his uh generous ated images we should also try not to use red it was
interesting to see that like literally from even like maybe 10 or 15 different
chat interactions um you know I was getting frustrated but it was definitely
the most interesting thing was that uh you could see what the machine had like
sort of pulled out from your interactions thus far and you could like
that the retrieval is actually accurate yeah and this is more I mean more
anecdotal in practice from what Founders have told us versus what the actual
exactly for what you said privacy and people wanting to fine-tune models on
their own data and not getting that licked out over the wire over the
internet and the other thing is um yeah maybe there still more accurate to do it
on your own when you really want that very precise information I think you
still need Rag and I think the analogy I like to think about this is sort of like
um processors back in the day in the 90s as uh when mlaw was actually Mo law
scaling it was not just a CPU processing speed getting faster but also memory
cache levels were also getting bigger and bigger but now more than 30 years
kinds of cashing for retrieving data out of like databases out of databases you
have maybe like a fast memory store with like Reddit for high availability and
then you still have things stored in your browser cach there still very much
lots of layers of how things will be cached and I think rag is going to be
this foundational thing that will stay and it'll be like how we work with
databases normally now just like lots of levels yeah yeah the tricky thing about
the context window I mean uh Gemini may have the team may have already fixed
this by now but certainly a lot of the Founders I talked to they said uh it's
sort of you know the million token context window sort of lacks uh
specificity literally uh if you ask for retrieval from its own context window
from you know or the prompt it actually sometimes just like can't seem to recall
it or can't seem to you know pick out the specific thing that you already fed
into it and uh the tricky thing there is like you'd rather have a 128k context
window that you knew was pretty Rock Solid rather than a system which where
you know it's still a bit of a black box you don't really know what's going on
and then for all you know it's just like sort of randomly picking up like half a
million of the tokens and that you know again like probably fixable you know I
can't imagine that that's like a permanent situation for you know uh a
million or 10 million uh token context window but something that we're seeing
from the field for now also in Enterprises like in business use cases
people care a lot about like what specific data is being retrieved who's
doing it like logging all of this stuff and permissioning around data so yeah
you can imagine having some kind of yeah a giant context window is not
necessarily what you want in Enterprise use case you actually probably want in
particular sensitive data stored somewhere else and retrieve like when
it's needed and know who's making the requests and filter it appropriately
exactly I think that will that will stay I was really encouraged what you said
actually about how the Google technology is maybe better than the open AI itself
it feels very googly actually it's like hey they've got the technology but they
just like don't know how to get like the Polish around it correct that means open
AI does not have this like Leap Forward unsalable Tech Advantage if Google has
should expect to see like meta come in and what we're seeing at the batch level
is just the models are pretty abstracted out right on a day-to-day basis like
Founders are already using different models to prototype versus like build
observability Ops software around this stuff just keeps progressing really
quickly so just funny my initial reaction whenever I hear like the model
they're all like we never talk about how Alliant they are on any one model I
worry if there's one model that's very very good and it'll be dominant and sort
of take over the world uh I'm less and less worried if there are many different
uh alternatives because then you have a market and a Marketplace equals uh you
know non- Monopoly pricing which means that uh you know 1,000 flowers can
actually Bloom like other startups can actually make choices and uh have gross
margin of their own and I'd much rather see you know thousands of companies make
a billion dollars a year each rather than you know one or two let alone seven
companies worth a trillion dollars and I think we have a dark horse that is yet
TBD we don't know when Lama 3 with 400 billion parameters comes out because
that's still being trained and that's like one that's like wow it could really
turn tables as well yeah the interesting thing about meta is I mean they have Pro
probably one of the largest clusters uh certainly I think I was reading um you
know in terms of who has paid Nvidia more money in the past year uh meta
apparently is number one by by um a a decent bit actually and the funny thing
is they have this giant cluster not because they necessarily have foreseen
couple years with large language models they acquire lots of gpus because they
needed to train their uh recommendation models right that use actually similar
architecture with de deep neural Nets to actually compete with Tik Tok because to
just like very classic Tech Innovation and disruption right like they're
basically worried about competing with Tik Tok they stop a bunch of gpus and
there turns out the gpus are just really valuable for this like completely
different use case that's going to change the world Jared like on that note
if you zoom out just like how does this cycle of hey like we're worried startups
are worried about the elephant in the room this case it's open AI maybe Google
competing and crushing them how does it play out to when we first moved out here
even like in that like era where Facebook was Rising Google was starting
to go from the search engine company to like the multi-product company do you
every time there's an open AI product release now it feels like there's a
bunch of startups waiting with baited breath to see whether open AI is going
to kill their startup and then there's all this internet commentary afterwards
about like which startups got killed by the latest open AI release and it
reminds me a lot of when we got to YC the the the three of us and the like
2005 to 2010 era there were all these companies who were innovating in the
same idea space as Google and Facebook building like related products and
services where the big question was always like what happens if Google does
this and when starts were pitching to investors that was like the main like a
big question that they'd always get from investors is like oh like but isn't
Google gonna do this the best response to that by the way was like well what if
Google gets into VC which it did that's a great VC on so a lot of the people who
are building AI apps now this is the first hype cycle they've ever been in
but we've all been through multiple hype cycles and so I think it's interesting
actually for the people who are in the middle of this hype cycle now where all
this is new to look back on the past hype cycles and see how the history of
what happened there can inform their decisions about what to work on if we
take Google as an example one thing that's interesting is if you look back
there was there was competing with Google in a very head-on way which was
hey we're going to build a better search engine um and YC definitely funded a lot
of companies trying that and I feel like the approach people would go after was
the vertical engine was say we're going to build a better Google for real estate
for example um some of those made it did they I which on well you could I mean
argue that um something like a red finin or Zillow clearly did have vertical
access to data and then or kayak for travel I guess or algolia for um company
Enterprise search search yeah that's true okay those yeah I hadn't thought of
yeah I hadn't thought of Zillow as a search engine but yeah it's essentially
that it's exactly that it's Vertical Search yeah but you have to monetize not
necessarily through the same way a search engine would you have to have
other services you have to become uh a broker you have to you know basically
make money in all these other ways CET different it doesn't look at all like
Google yeah and the data Integrations is very different like you have to really
poke and connect to MLS and a regular search engine would you wouldn't just
work with that like page rank wouldn't necessarily work with MLS yeah red Fin's
very interesting because I'm very addicted to red fin and it has actually
ultimately a great consumer is actually about buying just like a little bit of
someone's brain such that during the course of one's day I mean it doesn't
have to be every day but ideally it is you sort of think to use it and no one
of those companies would have said that they had better techn ol or they beat
Google on technology right like anyone who went up head head on against Google
for like the better general purpose search engine just got crushed and in
general most of the vertical search engs didn't work and certainly nothing that
looks anything like Google worked the the ones that I remember the most were
more ones that were in the vein of Google apps like when Google expanded
Beyond search and started launching Google Docs and sheets and slides and
maps and photos and all all these all these like like separate apps there were
a lot of companies that we funded y that were either going to be crushed or Not
by the next Google product yeah that's like the Santa Casa when you can bundle
software in I mean this like this is what Microsoft did to Netscape right
like once you can start bundling in software especially in the Enterprise
it's like people don't necessarily want to buy like 10 different solutions from
10 different vendors all the time if you can offer a good enough product across
several different use cases and ble them together Enterprises often want that I
mean famously uh Dropbox was in that R potential Rogue kale right because and
Drew because Drew actually talks about it when he comes back and give the
dinner talks about the fear when with Google Drive and Google hats his other
product Carousel thing right yeah in fact there is a time when um Dropbox had
launched this was after the batch and Google was working on Google Drive but
hadn't launched it was called G drive it was like the secret project inside of
Google and news of it leaked to the press and the whole world just decided
that like drop boxes Goose was cook like it was over Google was going to launch G
drive and because it was Google they had infinite money they were going to do the
same move that they're doing now who just throw like infinite money at the
product and give away like infinite storage for free how could have start to
possibly compete with Google you know spending billions of dollars to give
away infinite storage for free that was infinite tokens yeah and now it's
infinite tokens what are the big companies trying to do right now that
maybe you should avoid doing and uh the super obvious one is well uh open AI
simultaneously released the first version of the desktop app but that
version of the desktop app is merely uh sort of a skin on the web experience but
if you put two and two together surely it's going to look a lot more like her I
mean they've been really sh Scarlet voice they just pulled that right yeah
they're like oh shoot you know who knows are they getting sued who knows that's
that's what Twitter says today anyway but I think if you you look at the
details of that you know you can sort of sketch out what's going to happen with
um llms on the desktop and the desktop is sort of has access to all your files
basically the true personal assistant um that is directly consumer and then that
sounds like a whole category like you know we're going to interface with
computers and using potentially voice and certainly like ex we will have the
expectation of a lot of smarts and uh that you know that seems like where
that's where they're going and that's going to be one of the fights when I was
thinking back to like this first era of companies I guess one thought I had is
that it was fairly predictable actually what Google would build not 100%
predictable like Dropbox was like it was like unclear if Google would win that
space But like a lot of them are actually pretty obvious in hindsight um
like adtech for example like all of adtech just like never stuck around
because it was like too strategic to Google and Facebook and so they just had
to own all of it and like almost all of vertical search just didn't really
survive it's pretty easy to imagine what the next version of open AI like product
releases is going to be and if you can easily imagine that what you're building
is going to be in the next open AI release you know maybe it will be using
that framework it's like open AI really wants to capture just like the
imagination like the Sci-Fi imagination of everyone so it's like yeah like the
general purpose AI system that you just talk to and it figures out what you want
and does everything it seems hard to compete with them on that that's like
competing with Google on search right that's clearly going to be like the the
core because that was the early signs of why what chat gbt is being used for as
well just like a very very rudimentary right yeah which is the same thing with
would all use the same product anything that was like that was going to be
really tough as a startup yeah when I think of it for products I use like
perplexity norwi company but plexity is a product I use a lot because it's much
better for sort of research if I need to fix a toaster it's way easier for me to
type in like the model of the toaster into perplexity and get back like
specific links and YouTube videos just the whole workflow it was Diana who told
regular search yeah that's what I never I was trying to use perplexity for a
while and I couldn't get it and I was because I was trying to use it in the
same way I would use like the open AI the chat gbt app and I was like oh but
like chb is just so much better because I just like type in fuzzy things and it
figures it out and it comes back with smart things and perplexity just wasn't
as good for that use case but the specific hey I have this task that I
want like Source material back and links for it works much much much better it
doesn't capture the imagination right like open AI is not going to like
release a model that they demo the oh look like if you search it like gives
you the links back or it like shows you the YouTube videos that it's referring
to the demo is not as cool actually Gemini 1.5s has that feature and nobody
really talks about the demos from yeah from Google iio they're kind of like so
maybe one way to figure out how not to be roadkill is to like if you can build
the valuable but unsexy thing that open ey aren't going to demo on stage because
it doesn't like capture the Sci-Fi imagination you might survive yeah
that's definitely a whole line of thinking like Google was never going to
do instacart or door Dash's business so or ubber so all of that was fair game
and all of those turned out to be you know decacorn or you know potentially
you know even Airbnb like hundred billion dollar company the other thing
people always under estimate is just I think the size of new markets I remember
for a long time people didn't believe LinkedIn could be a big company because
like well like why CU like Facebook won social networking linkedin's just a
social network it's just going to be a like you have your work tab on your
Facebook profile like why would you need something else same thing with Twitter I
remember um when I first moved to San Francisco in 2007 some of the first
people I met were the early Facebook employees and they were like they saw
Twitter growing and they're like ah yeah we're going to like release stus updates
or something and just like Twitter's going to be done as just a feature but
yeah it turned out like Twitter was like a whole other thing instacart and door
Dash I think another great example of this again I remember iPhone comes out
Android becomes pervasive it's like oh there it's just going to be like Apple
and Google dominate mobile but there were all these things that they would
never build same in this AI World probably right there's all these things
that the big companies are never going to build and we probably have more
appetite for using multiple AI agent type apps than just like the one open AI
one and a huge like meta category that is basically almost anything that's B2B
like Google basically never built anything B2B they like basically only
Built Mass consumer software and so if you look at the YC unicorns like a ton
of them built you know some like B2B thing like you segment or something that
like Google was never going to build segment that's just like not interesting
to them I want because I think in B2B people really underestimate the human
being willing to go out and figure out who you sell to do the sales like Listen
to Someone Like give you all the things they're unhappy about and note them down
and take them back to your engineering team and say oh yeah we need to like
tweak this this and this and this and all these details right like and build
all these obscure edge cases like I think of one of our AI companies at YC
that's doing really well is called permit flow and they literally just
expedite the process for applying for construction permits and not just for
individuals but for like big construction companies now as well it's
like yeah like really hard to imagine that being the next open AI release
right like hey guys we built a a feature for for filing your construction permits
like can you yeah can you imagine turning up for your first day work as an
open AI engineer and they're like okay you're going to work on the construction
permit workflow feature they think it works that way well I guess if you join
those two ideas together something interesting happens though it seems sort
of inevitable sometime in the next two to five years you know assuming the open
AI her digital assistant comes out and then it's going to be on your desk top
it will actually know everything about you uh it'll know what you're doing and
know it'll know minute to minute what task you're trying to complete and then
it's conceivable you know if you masch that with sort of a launch that I think
they probably didn't invest enough into which was like the GPT store um you
could sort of imagine that might extend into B2B as well and then they would
sort of charge that Vig but I think the the thing that I don't think is going to
work for B2B actually is I think there's a lot of sensitivity ity around the
spaces with fintech and Healthcare I mean for good reasons they should be
very regulated and a lot of privacy data to Pro protect the consumers so I think
the other area that we've been having also success for AI B2B applications has
been in fintech we found that green light that's doing kyc using AI to
replace all the human behind that does a lot of the validation of consumer
identities or we also have a green board right right they both start with green
green board that was also doing a lot of the compliance things for banks as well
yeah Bronco is doing it in AR and there are a bunch of more companies doing
things in payments and just any of the boring dayto day that you know someone I
mean is sort of wrote doing it Um this can just basically supercharge that and
you know have one person do the work of 10 yeah we call this episode better
models better startups I think that is literally true for B2B companies where
it's like the underlying models like B2B software business models are so much
about how do I upsell like how do I make more money per customer next year than I
did this year and it just hey like every time the model gets better you can just
software and your end user doesn't care right like they just care about what the
function the software can do for them and so I think there's a world where the
models keep getting better you've got your choice of which one to use and the
additional functionality you just charge Ms your customers for and you make more
money yeah that's definitely what we're seeing at YC I mean last batch people
were making $6 million a year right at the beginning of the batch and it end up
being north of 30 million by the end of the batch so that's some really
outrageous Revenue growth in a very very short amount of time three or four
months and that's sort of on the back of what uh you know a few people working on
B2B software you know they can focus on a particular one that makes a lot of
money and then people are willing to Fork out a lot of cash if they see Roi
pretty much immediately there's not as many Founders working in this area as
there should be given the size of the opportunity like like to your to your
point har like people often underestimate how big these markets are
SAS like all the SAS combined right because like SAS is basically the tools
for the workers to do the jobs the AI like equivalent of SAS is like it it
just does the jobs tool plus the people yeah so like it should be just as large
someone's uh you know sort of you know cash flow statement right now but it'll
turn into software Revenue at 10x which will be interesting for market caps over
the next 10 20 years I was doing office hours with a startup this morning that
asked me this question about hey like you probably saw the GPT for for launch
like should we be worried about it um yeah my reply was you should be worried
about it but you should be worried about the other startups that are like
competing with you because ultimately it's all of the stuff we're talking
about it's whoever builds the best product on top of these models with all
the right nuances and details is going to win and that's going to be one of the
other startups in the space so I just think the meta thing as a startup now is
you're going to build in anticipation of them before someone else does versus
being worried about open AI or Google being the ones to build them let talk a
little bit about uh consumer because we did talk about what could be potentially
rill for Consumer startups if you're going against basically assistance some
sort of assistant type of thing opening eyes hinting well strongly direct and
companies what are those those things that they could flourish well here's an
edgy one anything that involves legal or PR risk is challenging for incumbents to
take on Microsoft giving money to open aai in the first place you could argue
was really about that I mean when image models image diffusion models first came
out at Google they were not allowed to generate the human form for uh PR and
legal LK risk reasons this is a large part of what created the opportunity for
open AI in the first place as Google was too scared to jeopardize their Golden
Goose by releasing this technology to the public the same thing could probably
be true now for startups things that are increasingly edgy are often the places
replica AI which was a AI NLP company working in this space for many years
even before llms were a thing still one of the top companies doing the AI
boyfriend or girlfriend and the wild thing about replica is that they've been
and earlier we were talking about you know a million token context window you
can imagine that virtual entity knowing everything about you like for many many
years like even your you know deepest darkest secrets and desires I mean
that's pretty wild stuff but um you it's going to look weird like that and um you
know people might not be paying attention I mean character AI has really
really deep retention and people are sort of spending hours per day sort of
using things like that so you know whatever happens in consumer it might be
nonobvious and and it might be very weird like that so there's a lot of kind
of more edgy stuff around uh deep fakes that are applied in different different
spaces so there's a company that you work with Jared Infinity AI right yeah
Infinity AI lets you turn any script into a movie and that movie can involve
famous characters and so like enables you to make famous people say whatever
is in your mind which is edgy which is part of what makes it like interesting
and cool Google would never launch that would never launch that and I think even
you know the the same move that open AI did to Google which is being willing to
release something that's really edgy well open AI is now the incumbent guys
they now can't release super edgy stuff like that anymore we're going to see a
interesting when you think about it like anything that's on the hey like I is
explicitly like a famous person this is explicitly using the likess of a famous
person for a profit is is going to get shut down on the other hand you have
like I If I make a meme with Will Smith and some like a caption like no one's
going to sue me for that and a lot of this content is like right in the middle
right it's like I'm not trying to build like a video that's literally I want
people to believe that it's like these people saying these things but what if
it's like a joke about a joke or a satire like where does that fit and yeah
you can't see you can't imagine Facebook or is going to roll this out on
Instagram anytime soon right like they want it they want to stay well clear of
that but you're already seeing this version of memes sort of 2.0 that are
basically deep fixed that are making the rounds and they're becoming viral tweets
right yeah hey why don't we clar out by going to a question that one of our
audience asked us on Twitter um so thank you sandip uh for this question question
is what specific update from open a Google meta excited each of you and why
I'll give one um the thing that really excited me about the open AI release was
the emotion in the generated voice and I didn't realize how much I was missing
this from the existing text to speech models until I heard the open AI voice
oh a bedtime story about robots and love I got you covered once upon a time in a
world not too different from ours there was a robot named bite it's amazingly
better compared to the incumbent Texas each model because it actually knows
what it's saying the existing ones by contrast sound so robotic they like
and the open AI one it felt like you were talking to a human my one was the
Brazilian her parents don't speak English and so I've been learning
Portuguese but it's coming along very slowly the like the idea of having just
communicate with anyone anywhere in the world is really exciting hey how's it
recently it's a massive idea I mean it could change the world you could go live
in a foreign country where you don't speak the language it it like it is huge
real is a pretty cool one I guess for me um what's funny about 40 is it sounds
like maybe it was actually just a reorg basically there was a reor get open Ai
and they realized uh they want all of the teams rowing in the same direction
and then uh what that means is probably really good things for both their
assistant desktop product uh but also eventually robotics which um might be a
really big deal down the road this Chinese company called unry announced a
$116,000 human biped robot though Twitter warns me that it's another
$50,000 if you actually want open API access previously they made a
$114,000 version of that robot um but I think unified models means more and more
likelihood that practical robotics is you know actually not that far away
famous last words of course we've been saying that uh pretty consistently for
many years in a row but this time it's different I think for me maybe a bit
more of a tech technical one I know it doesn't sound too too fancy but really
the half the cost is like a huge thing and if you extrapolate that what that
means is probably a lot of these models are hitting some kind of ASM totic
growth of how much better they can get which means also that they're becoming
more stable and it can open up the space for actual custom silicon to process all
of these and enable a lot more low power processing to enable Robotics and build
the device that you mentioned and actually have it in your packet and not
be Ted to the Internet so all these things that we could perhaps see uh
excitement of new tech product releases because I kind of missed those day when
every product Tech demo was like very exciting now it's just like kind of like
a feature true we could be excited about new things coming up well we're going to
be really excited to see what you guys all come up with that's it for this week
n [Music]