Hi all 👋, makers Kyle + Erik + Sahil + Blake + Candice here! We’re excited for you to try out Banana.
Banana is an ML inference hosting platform on serverless GPUs.
Why did we build Banana?
We used to operate an ML consulting firm that helped companies build & host models in production. Through this work we realized how difficult it was to get models deployed, and how incredibly expensive it was. Customer models had to run on a fleet of always-on GPUs that would get maybe 10% utilization, which felt like a really big money pit and waste of compute.
During our time as consultants we built really efficient deployment infrastructure underneath us. Six months ago we made a pivot to focus solely on productizing our deployment infra into a hosting platform for ML teams to use that would remove the pain of deployment and reduce the cost of hosting models.
That brings us to today.
Banana provides inference hosting for ML models in three easy steps and a single line of code. We are 100% self serve, meaning users can sign up and deploy to production without ever talking to our 🍌 team. And thanks to being on serverless GPUs, customers see their compute costs reduced by upwards of 90%.
Try it out and let us know what you think!
@bp_banana Congratulations on the launch!
Banana could be addressing the need-gap: 'Democratisation of AI/ML/DL hardware', Posted on my problem validation forum - https://needgap.com/problems/50-... .
You're welcomed to explain how Banana addresses that problem there, So those who need it can find Banana easily.
Been using Banana in production for a good bit now, nothing but great things to share! Few notes:
- Product is insanely good. Specifically, we use it for indexing jobs requiring a good bit of GPU compute. These jobs are huge, sometimes involving up to 1M inferences of a large NL model. Banana is perfect for this use case, as we can burst up to 10+ GPUs, only pay for the compute we use, and quickly scale back down to near zero.
- Team is very strong, super responsive to questions and are experts at deploying & scaling ML models. We often get advice and recommendations from their team on how to best do something, and it's been really appreciated!
- Lastly, velocity / speed of iteration has been ridiculous. They're moving really quick, have an ambitious roadmap, and ship new features and improvements daily. It's been really cool to watch.
Would highly recommend anyone check them out!
@gallantlabs Thank you for your kind words & support, Morgan! Fantastic having you as a customer and inspiring seeing your progress as a team! Always a message away :)
This looks seriously impressive. I'll be trying it out soon. Seems like this will be a huge step up over Google Cloud Run in terms of speed. I see on your roadmap that you're planning on moving to beefier GPUs in the future. Which GPUs are you running now?
Also, from a technical perspective, I assume this works by moving models from CPU to GPU at inference time? Trying to wrap my head around how you're getting such fast cold starts.
@gregpriday Thanks for the support!
The roadmap is now! We used to only do T4 GPUs, but now we also support A100 GPUs which are yielding faster cold boots + inference + download speeds.
To understand how banana works it may be easier to think of us as a compiler company. When you send us a model we do stuff under-the-hood to make it run faster/cheaper. CPU/GPU memory hacks are definitely involved (how we load memory, where, when). A key point is none of our optimizations affect model outputs. This means we don't do weight quantization dynamic layer/node pruning which yield way smaller/faster models but does affect output.
Can't say enough good things about @erikdoingthings and this rockstar team. Having personally grappled with the problem of GPU provisioning and management for years (and having built an unsuccessful product around it), the way Banana has abstracted it is magical. Can see this becoming the DX standard for ML.
Hey, Congrats on your launch! 🚀
We will be launching soon too, (https://www.producthunt.com/upco...), I hope that you will support us as well,
Cheers 😊
@derekpankaew yes I'm sure this would have been solid for the Yolo models you were running at Next Fitness! With models that small too they would have scaled up in the matter of a couple seconds too :)
gpu coldstarts is a tough problem to solve! glad banana is taking on this challenge! definitely super cool product! especially for hobby projects that require gpus and you don't want to pay an arm and an leg to host a demo
Replies
Img2Prompt
Shoot YOUR Shot - AI Edition
Bone Health Tracker
Operand
Audioread (formerly Audiblogs)
Shoot YOUR Shot - AI Edition
Product Hunt
Shoot YOUR Shot - AI Edition
Audioread (formerly Audiblogs)
CloudPrint
Shoot YOUR Shot - AI Edition
CloudPrint
Shoot YOUR Shot - AI Edition
Img2Prompt
Shoot YOUR Shot - AI Edition
Img2Prompt
Shoot YOUR Shot - AI Edition
Snackeet
Shoot YOUR Shot - AI Edition
AppLogger by PLG Works
Img2Prompt
Private Wishlist
Img2Prompt
Img2Prompt
Kobra
Shoot YOUR Shot - AI Edition
Img2Prompt
Opus 0.2
Shoot YOUR Shot - AI Edition
Hampton
Img2Prompt
Listening.io
Shoot YOUR Shot - AI Edition
Ship That Shit
Img2Prompt