this post was submitted on 20 Nov 2023
148 points (100.0% liked)

Technology

37717 readers
423 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

Article from The Atlantic, archive link: https://archive.ph/Vqjpr

Some important quotes:

The tensions boiled over at the top. As Altman and OpenAI President Greg Brockman encouraged more commercialization, the company’s chief scientist, Ilya Sutskever, grew more concerned about whether OpenAI was upholding the governing nonprofit’s mission to create beneficial AGI.

The release of GPT-4 also frustrated the alignment team, which was focused on further-upstream AI-safety challenges, such as developing various techniques to get the model to follow user instructions and prevent it from spewing toxic speech or “hallucinating”—confidently presenting misinformation as fact. Many members of the team, including a growing contingent fearful of the existential risk of more-advanced AI models, felt uncomfortable with how quickly GPT-4 had been launched and integrated widely into other products. They believed that the AI safety work they had done was insufficient.

Employees from an already small trust-and-safety staff were reassigned from other abuse areas to focus on this issue. Under the increasing strain, some employees struggled with mental-health issues. Communication was poor. Co-workers would find out that colleagues had been fired only after noticing them disappear on Slack.

Summary: Tech bros want money, tech bros want speed, tech bros want products.

Scientists want safety, researchers want to research...

you are viewing a single comment's thread
view the rest of the comments
[–] canis_majoris@lemmy.ca 4 points 1 year ago (1 children)

It's already easy to self host and we've optimized LLMs to run locally on not much serious hardware after we've trained them; I have GPT4ALL set up on my machine and it runs everything locally with my processor, no GPU or anything. Some of those datasets are uncensored, and I've seen what Stable Diffusion can do for image generation.

I tend to use the GPT-4 built into Edge with my O365 corporate plan, because it suits my needs better for day-to-day challenges. It can still audit code and summarize things, which is all I really need it to do here and there.

[–] cwagner@beehaw.org 5 points 1 year ago (1 children)

Nothing that runs on my GPU / CPU comes even close to GPT 3.5, GPT4 is not even in the same universe, and that’s with them running far more slowly.

[–] RandoCalrandian@kbin.social 1 points 11 months ago (1 children)

In my tests, the self hosted options that have access to a 30xx or 40xx graphics card return results far faster than gpt4

[–] cwagner@beehaw.org 1 points 11 months ago (1 children)

Which model are you talking about?

[–] RandoCalrandian@kbin.social 1 points 11 months ago (1 children)

Mistral for chatgpt, and i'm not saying it gives better answers, just that it's much faster than my web portal to gpt4

[–] cwagner@beehaw.org 1 points 11 months ago

Oh, faster is easy. GPT 3.5 is also far faster than GPT 4. Faster at quality replies is the issue.