this post was submitted on 19 Jun 2023
6 points (100.0% liked)

Technology

34838 readers
13 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
top 13 comments
sorted by: hot top controversial new old
[–] schnapsidee@feddit.de 5 points 1 year ago (2 children)

Decisions like this just prove how massive the market for a self-hostable alternative is. They're not banning it because it's a bad tool, they're banning it because they're concerned about what happens to the source code their engineers paste into it.

There are already a bunch of OSS attempts, and it likely won't take long until we have something of comparable quality to ChatGPT is available for companies to host on their own hardware.

[–] saplyng@kbin.social 2 points 1 year ago

Isn't Llama selfhostable?

[–] eight_byte@feddit.de 1 points 1 year ago* (last edited 1 year ago)

Companies are also banning ChatGPT because its unclear from where the code it spits out was stolen and how it’s licensed. Copy and pasting code from AI tools is an enormous legal risk for a software company.

[–] MargotRobbie@lemmy.world 1 points 1 year ago

Well of course, ChatGPT has already leaked Samsung Semiconductor's internal information earlier, and Apple is infamous for being secretive about their design.

[–] MentalEdge@sopuli.xyz -4 points 1 year ago* (last edited 1 year ago) (2 children)

How to neuter your own ability to compete: ban your workers from using the latest tool for boosting employee performance.

[–] SwingingKoala@discuss.tchncs.de 6 points 1 year ago (3 children)

Frankly, if not using chatgpt reduces your peformance significantly I wouldn't want to work with you. It would mean that you're not doing much more than copy and pasting random search results into the project and don't spend any time validating, vetting or testing them. Chatgpt is just a new interface to already existing data.

[–] PerogiBoi@lemmy.ca 1 points 1 year ago

I’m gonna vehemently disagree with you. As a knowledge worker, ChatGPT allows me to offload low level thinking and writing tasks so I can focus on bigger picture creative aspects.

GPT speeds up my quality work output by around half. Those who refuse to incorporate it into their work flow will find they fall behind compared to those who have successfully integrated it.

[–] animist@lemmy.one 0 points 1 year ago

Better stop using xerox machines to make copies and write everything out by hand

[–] MentalEdge@sopuli.xyz -1 points 1 year ago* (last edited 1 year ago) (1 children)

Then you don't have much faith for your co-workers competence in wielding any given tool to its greatest utility. Using an LLM like ChatGPT to access data hardly automatically means you're also a brain-dead search result copy-paster.

Yes, its a new interface for existing data, the same way digital files are to data on paper. Only ever using the latter is really inefficient, and stupid in a world where the digital files exist. Not that the hardcopies cant be to their own utility, or be used as corroborating data.

It's a really good interface, if you know how to use it. This is like banning search engines because you expect your workers to be expert at everything, so they shouldn't need support tools to sleuth for data.

This is like banning search engines because you expect your workers to be expert at everything,

More like banning your engineers from discussing their work with third parties. If you feed chatgpt the same queries that you would feed to a search engine you're probably not using it optimally.

[–] ulu_mulu@lemmy.world 4 points 1 year ago (1 children)

Leaking industry secrets is a much bigger concern that boosting productivity a little bit.

We're talking about very specialized engineering work, it's not something you can totally rely on a bot to do, though it might help sometimes, it's fully understandable for specialized companies to want to ban GPT internally, until there's a way for them to host a totally internal one.

[–] MentalEdge@sopuli.xyz 0 points 1 year ago (1 children)

On this I agree entirely. The potential for corporate espionage because of unwitting employees using an LLM through unofficial means is huge.

At the very least, the corporation itself would have to be the customer, so that watertight terms might be negotiated, not the employee.

[–] ulu_mulu@lemmy.world 1 points 1 year ago* (last edited 1 year ago)

I don't think being a customer would work either, language models are still on the training, noone knows exactly how users queries are used, that's a big no no for every company having to protect their secrets.

A self-hosted instance is a much better solution, if not the only "safe" one from that point of view, we'll get there.