236: We Now Measures the Largest Chips Used to Generate an LLM - or a 21st century #$%& Measuring Contest

Episode 236 November 23, 2023 00:23:00
236: We Now Measures the Largest Chips Used to Generate an LLM - or a 21st century #$%& Measuring Contest
tcp.fm
236: We Now Measures the Largest Chips Used to Generate an LLM - or a 21st century #$%& Measuring Contest

Nov 23 2023 | 00:23:00

/

Show Notes

Welcome to episode 236 of the Cloud Pod Podcast, where the forecast is always cloudy! Are you wandering around every day wondering just who has the biggest one? Chips, we mean. Of course. Get your mind out of the gutter. Did you know Azure was winning that battle for like 8 whole minutes? Join us for episode 236 where we talk about chip size, LLM’s, updates to Bedrock, and Toxicity Detection - something you will never find applied to the podcast. Not on purpose, anyway.  Happy Thanksgiving!

Titles we almost went with this week:

A big thanks to this week’s sponsor:

Foghorn Consulting provides top-notch cloud and DevOps engineers to the world’s most innovative companies. Initiatives stalled because you have trouble hiring?  Foghorn can be burning down your DevOps and Cloud backlogs as soon as next week.

AI is Going Great! 

00:39  OpenAI’s New Weapon in Talent War With Google: $10 Million Pay Packages for Researchers (listeners note: paywall article) 01:30 Jonathan - “I guess it's quite a concern actually that since Google bought DeepMind they have pretty much two-thirds of the entire global AI talent at their own disposal. So I guess this is a desperate needs, call for desperate measures kind of thing.” 01:49 Nvidia Unveils New AI Chip, Upping Ante with AMD (listeners note: paywall article) 02:29 Matthew - “ I feel like we're seeing the speed curve of processors and now we're just watching the same things that happened in the 90s and 2000s happen with GPUs. It's like, it will double every 18 months. That's fine. Or here sooner.” 04:51 Report: Enterprise investment in generative AI shockingly low, while traditional AI is thriving    05:36 Ryan - “I don't see any way where this is going to not be a huge contributor to cloud spend in coming years. I'm actually more surprised that it's the traditional AI and machine learning is only 18%. But then you have to realize that, you know, that's we're also an industry that's still largely doing rented compute. So it makes sense.”

AWS

06:32 AWS Audit Manager now supports first third-party GRC integration 07:15 Justin - “Thank goodness, cause I'm, I was kind of thinking this was a walled garden that didn't make sense for a long time. So glad to see this one coming.” 07:42 Amazon Bedrock now provides access to Meta’s Llama 2 Chat 13B model  ADDITIONALLY -  Amazon Bedrock now provides access to Cohere Command Light and Cohere Embed English and multilingual models 09:12 New for Amazon Comprehend – Toxicity Detection 09:47 Ryan - “My very first thought when I read this is, you know, back in the day, I created a chatbot in IRC that would count swear words by user and you could run a command and it would just put that out. And so now I have an idea where plugging this into, you know, several team rooms in Slack or Teams and then giving a toxicity score would be pretty sweet. It would be pretty funny.” 10:28 Jonathan - “It's kind of interesting technology. I see use cases for it for sure for things like, you know, filtering reviews for online merchants, things that users post that end up on other people's websites. Makes a lot of sense. I guess I'm kind of concerned a little bit that this type of technology might end up in things like Teams or Zoom or any other kind of chat or Slack for that matter. And potentially like to report on a user's behavior or attitude or something else to kind of like their management in a way. Imagine that's quite a big brother-ish kind of technology, but I think the potential is there right now for this.” 12:59 Amazon Aurora MySQL zero-ETL integration with Amazon Redshift is now generally available 13:51 Justin - “The most interesting about this to me is this kind of breaks one of the main things about Amazon is that their services sort of all are independent of each other and they don't use the same storage subsystem. They don't do these things. And so now they've created a dependency where these things now have to work together. So that's kind of interesting, uh, paradigm shift. I love it. Uh, cause I hate running ETL jobs. Uh, and I can definitely see this being something I would use if I was on Aurora and I needed redshift. Um, so bravo, but also like, how does this work? I'm hoping maybe there's a re:Invent session that'll come up somewhere that details us a bit more. And I'll be keeping an eye out for that during re:Invent to see if I can learn more about how they're doing this magic in the backend.:

GCP

15:16  Introducing Cloud SQL in-place upgrade: move from Enterprise to Enterprise Plus with ease  16:32 Ryan - “I think what I liked most about this announcement is that they gave you a rollback procedure. You want to play out with the new enterprise and I've done that and then there's no way to turn it off. This is expensive and I don't want to pay for it and you have to kill the whole thing. So I like the fact that this can go both ways and you can see if you really need those advanced features or not.” 17:09  Google Cloud demonstrates the world’s largest distributed training job for large language models across 50000+ TPU v5e chips 

Azure

18:39 Azure sets a scale record in large language model training  19:16 Justin - “So sorry Azure, no record for you today.” 20:02 Matthew - “So I figured oracles business model out. They're just a layer on top of all the other hyperscalers, which breaks everything. It'll be fine.” 20:13 Justin - “It's really just a tech company on top of a bunch of lawyers.”

Closing

And that is the week in the cloud! We would like to thank our sponsors Foghorn Consulting. Check out our website, the home of the Cloud Pod where you can join our newsletter, slack team, send feedback or ask questions at theCloud Pod.net or tweet at us with hashtag #theCloud Pod

Other Episodes

Episode

September 14, 2019 57m47
Episode Cover

US East 1’s on Fire,  The Cloud Pod says let the !#*&@ burn – Ep 38

US-East-1 has a hiccup in a single AZ, Lambda fixes cold start launches inside a VPC, Google gets an AD service and Microsoft goes...

Listen

Episode

April 23, 2019 42m07s
Episode Cover

Episode 19: Announcing the new Cloud Pod Premium Tier

We are back to our normal show after our GCP Next recap.  This week the new AWS APAC region, Azure premium tiers and the...

Listen

Episode 133

September 10, 2021 00:41:03
Episode Cover

133: Google Cloud Serverless Functions now with Servers

On The Cloud Pod this week, AWS releases new features including Managed Grafana, GCP Serverless solves the cold start problem, and Wiz hacks into...

Listen