Channel
Interviewed Person
Mariano Cocirio

Better Stack
Interviewed: v0

My 50 cents
Interviewed: Malte Ubl

Software Engineering Daily
Interviewed: Tim Neutkens

This Dot Media
Interviewed: Guillermo Rauch
we like the well-known good parts of sess we are going broke with this because our requests are taking longer and longer and longer because the models are reasoning more and more and more think about some startups who are building I don't know AI to generate music to generate videos it takes a long time we are talking about minutes waiting and if you are not able to reutilize that then you are a bit Doom
fa hello everybody and welcome to the modern web podcast I'm your host Rob oel the VP of innovation at this. laabs joined today by my wonderful co-host Danny Thompson Danny is the director of Technology at this. Labs Danny how are you doing hey hey I'm great I'm excited to be here and to be honest really excited about today's topic that's right we have a really exciting topic because we are sitting down and talking with Mariano cerio who is the staff product manager at Vercel marianao how are you doing I'm doing fine thank you f
yeah and today we're going to be talking about fluid compute and I guess the first question is what is that so do you want to introduce us to what you guys just announced and just released yes flu compute is a new cloud computing model it's built on severals and it combines event driven execution and intelligent Resource Management it's basically would you like from servers would you like from serverless make it together you get fluid so as of the time of recording this this just came out like this is big news um
the announcements were fantastic and the blog article about it I've definitely been reading up on it um one point that it seems to be echoed is like serverless for AI is kind of broken and this is basically a solution in some way shape or form in order to handle that and so I guess if we kind of go into this what are like the main challenges with traditional serverless Computing why did we see this as being the solution for it especially like regards to Ai and what
even was the thought process to wanting to solve this and especially why is this the solution that we want to solve it with yeah I think you said right serus compute for AI is broken we need to accept that fact why is it because we have been optimizing sever lless to handle easy request like requests are going to be fast asking something to a database doing something and returning it meanwhile with AI we some sometimes s wanted to spend a little more time thinking reasoning agents talking to
each other and what's the problem with selfless you are paying for that idle time while you are not using the CPU you are going to have instances that are going to be just e waiting for response from I don't know OpenAI you call it Cloud whatever API you're using your own models hosted by your backhand and that time you're paying it and you don't want to be doing that meanwhile we when we
were in the past on servers what you would do is you would have your server you will hit just you know request on it and it will start queuing it and will start you know executing like one request executes if I have idle time then another request executes right after that one Etc so during the idle time of request one you can execute request two what fluid is bringing you is that have being being able to have
one single instance handling multiple request but at the same time those instances can scale down to zero and can scale up to infinite like they do on the sess model so you get the infinite scaling from serverless you get the scale down to Zero from serverless but you also get the reability of the instances and the efficiency of the servers it's the best of both worlds