this post was submitted on 13 Aug 2023
1084 points (96.2% liked)

Technology

59314 readers
4948 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

As the AI market continues to balloon, experts are warning that its VC-driven rise is eerily similar to that of the dot com bubble.

you are viewing a single comment's thread
view the rest of the comments
[–] Not_Alec_Baldwin@lemmy.world 9 points 1 year ago (1 children)

I've started going down this rabbit hole. The takeaway is that if we define intelligence as "ability to solve problems", we've already created artificial intelligence. It's not flawless, but it's remarkable.

There's the concept of Artificial General Intelligence (AGI) or Artificial Consciousness which people are somewhat obsessed with, that we'll create an artificial mind that thinks like a human mind does.

But that's not really how we do things. Think about how we walk, and then look at a bicycle. A car. A train. A plane. The things we make look and work nothing like we do, and they do the things we do significantly better than we do them.

I expect AI to be a very similar monster.

If you're curious about this kind of conversation I'd highly recommend looking for books or podcasts by Joscha Bach, he did 3 amazing episodes with Lex.

[–] orphiebaby@lemmy.world -5 points 1 year ago* (last edited 1 year ago) (2 children)

Current "AI" doesn't solve problems. It doesn't understand context. It can't see fingers and say "those are fingers, make sure there's only five". It can't tell the difference between a truth and a lie. It can't say "well that can't be right!" It just regurgitates an amalgamation of things humans have showed it or said, with zero understanding. "Consciousness" and certainly "sapience" aren't really relevant factors here.

[–] magic_lobster_party@kbin.social 4 points 1 year ago (1 children)

You’re confusing AI with AGI. AGI is the ultimate goal of AI research. AI are all the steps along the way. Step by step, AI researchers figure out how to make computers replicate human capabilities. AGI is when we have an AI that has basically replicated all human capabilities. That’s when it’s no longer bounded by a particular problem.

You can use the more specific terms “weak AI” or “narrow AI” if you prefer.

Generative AI is just another step in the way. Just like how the emergence of deep learning was one step some years ago. It can clearly produce stuff that previously only humans could make, which in this case is convincing texts and pictures from arbitrary prompts. It’s accurate to call it AI (or weak AI).

[–] orphiebaby@lemmy.world -1 points 1 year ago

Yeah, well, "AGI" is not the end result of this generative crap. You're gonna have to start over with something different one way or another. This simply is not the way.

[–] minikieff@lemmy.world 0 points 1 year ago (1 children)
[–] orphiebaby@lemmy.world 0 points 1 year ago* (last edited 1 year ago)

No? There's a whole lot more to being human than being able to separate one object from another and identify it, recognize that object, and say "my database says that there should only be two of these in this context". Current "AI" can't even do that much-- especially not with art.

Do you know what "sapience" means, by the way?