188
submitted 1 year ago by L4s@lemmy.world to c/technology@lemmy.world

Nvidia reveals new A.I. chip, says costs of running LLMs will ‘drop significantly’::Currently, Nvidia dominates the market for AI chips, with over 80% market share, according to some estimates.

all 28 comments
sorted by: hot top controversial new old
[-] BobKerman3999@feddit.it 77 points 1 year ago

So yeah Nvidia is mask off now and it is 100% an AI hardware company with a side business of graphics chips

[-] Even_Adder@lemmy.dbzer0.com 49 points 1 year ago

Nvidia has always been a tech company that also happens to make consumer graphics cards.

[-] BobKerman3999@feddit.it 9 points 1 year ago

Hmm, define "always" because at the time of riva and Tnt they definitely were a graphics card only company

[-] Even_Adder@lemmy.dbzer0.com 17 points 1 year ago

Yeah, I misspoke there, but for most of recent memory they've been doing big things besides consumer graphics cards. Nvidia launched its professional oriented graphics Quadro product line in 2000. They launched CUDA architecture in 2006 which opened up parallel processing capabilities of GPUs for use in science and research. They entered the data center and cloud computing market in the early 2010s, and in 2015 they launched the DRIVE product line.

[-] BobKerman3999@feddit.it -1 points 1 year ago

Unfortunately I know Nvidia since the big GC wars with matrix, 3dgx, ATI and the guys that built Kyro...

[-] Tatters@feddit.uk 34 points 1 year ago

And before that it was a bit mining company, with a side line of gaming graphics hardware.

[-] Haha@lemmy.world 15 points 1 year ago

It’s always been that way. Whether it’s been AI or something else. Nothing wrong with that.

[-] Not_mikey@lemmy.world 9 points 1 year ago

Not all bad, compared to crypto the vector transformations done for ml are relatively similar to those done by graphics processing. So any innovations on the ml front will probably yield improvements in graphics.

[-] synceDD@lemmy.world 7 points 1 year ago

Company focusing on their profits and not my 4k dlss witcher 3 ?!?!

[-] Zerfallen@lemmy.world 36 points 1 year ago

I'm sure the cost to the consumer will remain exactly the same, or somehow increase.

[-] GenderNeutralBro@lemmy.sdf.org 9 points 1 year ago

I'm not worried about that. There will be open competition, because most of this stuff is open-source. Cheaper hardware will open the door for anyone like you or me to set up our own services. Anyone can set up a server with their own hardware (or rent it from Amazon or wherever) and run their own chatbot (with blackjack! and hookers!) instead of using ChatGPT.

This is already possible on consumer hardware, just not with the biggest and best networks. Right now, if I wanted to run, say, BLOOM (an open-source LLM), I'd need to spend close to $100K on hardware. Obviously, that's out of reach for a hobbyist, so I'm limited to using smaller, less advanced networks like LLaMa or GPT-J. Cheaper hardware will help break the hold that the big players currently have over the industry.

[-] abhibeckert@lemmy.world 1 points 1 year ago* (last edited 1 year ago)

if I wanted to run, say, BLOOM (an open-source LLM), I’d need to spend close to $100K on hardware

Doesn't that dozens of notes with over a terabyte of RAM each? And state of the art networking?

Sounds closer to $100M than $100K.

[-] GenderNeutralBro@lemmy.sdf.org 5 points 1 year ago

If you want to train your own network like they did, you'd want something like that, yeah, but to run the trained network you "only" need ~360GB of memory.

For context, even if you wanted to run this in CPU, there are currently no A5 mobos (Ryzen 7000 series) that support more than 192GB of memory. You literally can't even run it on high-end consumer hardware.

this post was submitted on 10 Aug 2023
188 points (96.1% liked)

Technology

58150 readers
6948 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS