this post was submitted on 19 Feb 2025
185 points (93.4% liked)

Technology

62853 readers
3422 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 29 comments
sorted by: hot top controversial new old
[–] vrighter@discuss.tchncs.de 16 points 22 hours ago (1 children)

not remove ic replace.

Also, stop calling releasing binary blobs of weights as open source

[–] brucethemoose@lemmy.world 1 points 12 hours ago

It's honestly not that big a deal, as it's not like knowing anything about how it was trained (beyond the config) would help you modify it. It's still highly modifiable. It's not like anyone can afford to replicate it.

It would be nice to publish the hyperparameters for research purposes, but... shrug.

I think a subset of the exact training data/hyperparameters would help with quantization-aware-training, maybe, but that's all I got.

[–] Korkki@lemmy.ml 42 points 1 day ago* (last edited 1 day ago) (1 children)

Now that it has been "freedomized" it instead says that Israel is actually 3000+ years old and Palestinians are invaders, and Israel has the right to defend itself like chatgpt does. Also that American style liberal democracy is the peak of human development and civilization. Don't kid yourself that this has anything to do with truth or making it more "accurate". Yeah just replace one set of official truths, half thrush with other set of official positions, half truths and outright lies to plug the gaps. Again, who fact checks the factcheckers? Even just out of spite I would not use any model that advertises itself as being trained to answer to the sensibilities of a western techbro liberals so that they can once more outsource their thinking to outside party, because at-least it isn't the CCP propaganda. Like it's so absolutely dreadful when deepseek copy-pastes official CPC party line word for word, on Chinese form of government instead of making up some Wikipedia tier wall of text where the word "authoritarian" is about as common of a word as "the".

I just don't get it man. What kind of non thinking cretin purposefully would use this kind of model? It's probably really the pettiness that gets to me so badly. Or maybe it's that while this is purposefully made to push a certain narrative, the deepseek's bias is probably just a result that it's been trainer on Chinese internet data.

[–] Squizzy@lemmy.world 3 points 2 hours ago

My chatgpt will list the questionable human righfs record of Israel ad agree with the conclusion that they are, under the same standards and logic, likely guilty of their accused atrocities as thenlikes of China or Russia.

[–] ivanafterall@lemmy.world 6 points 22 hours ago

Ctrl + F

Find: Chinese

Replace: God-damned Chinese

New model's ready!

[–] biofaust@lemmy.world 5 points 22 hours ago (1 children)

I run an uncensored version on my PC since weeks, there are multiple ones on HuggingFace.

[–] brucethemoose@lemmy.world 2 points 12 hours ago* (last edited 12 hours ago)

Not full R1, which is developed differently than any of the distillations.

[–] GissaMittJobb@lemmy.ml 13 points 1 day ago (1 children)

My god, could he have chosen a name more cringe than this? Fuck.

[–] MunkyNutts@lemmy.world 3 points 19 hours ago (2 children)

An attempt to appeal to U.S. conservatives maybe? You know all that 1776, 3%, ΜΟΛΩΝ ΛΑΒΕ etc. sticker suckers.

[–] GissaMittJobb@lemmy.ml 4 points 19 hours ago

From what I gather from the Perplexity CEO, he is just that type of Musk-tier cringelord, so yes, probably

[–] Beldarofremulak@discuss.online 2 points 18 hours ago
[–] thefluffiest@feddit.nl 23 points 1 day ago (2 children)

Great. Has it also removed American censorship and propaganda?

[–] DarkSpectrum@lemmy.world 21 points 1 day ago

I believe this is what was added

[–] FaceDeer@fedia.io -1 points 1 day ago (1 children)

Why would a Chinese-made AI have American censorship and propaganda in it?

[–] UnsavoryMollusk@lemmy.world 2 points 23 hours ago

They can add stuff too. At least it seems so, this model still give biased answers now but more in favor of the US... So who knows ?

[–] brucethemoose@lemmy.world 31 points 1 day ago* (last edited 1 day ago) (1 children)

That’s the great thing about open models. Censorship? Once identified, all it takes is one person and a bit of cash to get rid of it, though it seems Perplexity did a particularly good job (unlike some “abliterated” models that are pretty dumbed down).

[–] Even_Adder@lemmy.dbzer0.com 5 points 1 day ago (1 children)

Can't wait to try a distillation. The full model is huge.

[–] brucethemoose@lemmy.world 1 points 1 day ago (1 children)

In the 32B range? I think we have plenty of uncensored thinking models there, maybe try fusion 32B.

I'm not an expert though, as models trained from base Qwen have been sufficient for that, for me.

[–] Even_Adder@lemmy.dbzer0.com 1 points 1 day ago

I just want to mess with this one too. I had a hard time finding an abliterated one before that didn't fail the Tiananmen Square question regularly.

[–] FrankLaskey@lemmy.ml 18 points 1 day ago* (last edited 1 day ago) (4 children)

I think we can all agree that modifications to these models which remove censorship and propaganda on behalf of one particular country or party is valuable for the sake of accuracy and impartiality, but reading some of the example responses for the new model I honestly find myself wondering if they haven’t gone a bit further than that by replacing some of the old non-responses and positive portrayals of China and the CPC with a highly critical perspective typified by western governments which are hostile to China (in particular the US). Even the name of the model certainly doesn’t make it sound like neutrality and accuracy is their primary aim here.

[–] ZILtoid1991@lemmy.world 9 points 1 day ago

LLMs are gigantic bias reproduction machines, so it will never be perfect.

[–] Aatube@kbin.melroy.org 12 points 1 day ago

ehhhh, the only thing the model got quite wrong was the level of control on access to media, internet, and especially education. Other than that the article's example responses seem pretty on-point. (I only otherwise found a blemish where a few words needed further clarification; I found no other errors in my first reading.) Though I do also find the name of the model quite off-putting.

[–] brucethemoose@lemmy.world 1 points 1 day ago* (last edited 1 day ago)

Well you can merge it with the original model, to any degree, to get any sliding scale of “bias” you want.

Practically, though, I guess that’s not super practical, as very few have the hardware or cash to deploy a custom full R1 themselves.

[–] iopq@lemmy.world -4 points 1 day ago (1 children)

What part is highly critical of China? Facts can't be critical

[–] fruitycoder@sh.itjust.works 5 points 1 day ago (1 children)

Listen, I'm highly critical of the CCP, but LLMs aren't facts machines, they are make text like what they are trained on machines.

They have no grasp of truth, and we can only get some sense of truth of what the average collective text response of its dataset (at best!).

[–] iopq@lemmy.world 2 points 1 day ago

I'm talking about the example texts

[–] ZILtoid1991@lemmy.world 5 points 1 day ago (2 children)

IDK, but this seems like wankery to me. Just google it if you want to know about it, the AI isn't an "all knowing being" nor "the arbitrer of truth".

I have a feeling that a new logical fallacy will soon emerge (if it isn't already widespread on certain places of the internet), that will be "X is true because the LLM said so".

[–] fruitycoder@sh.itjust.works 3 points 1 day ago

It's really an extension of "Would some really do that? Just lie on the Internet?" But now "Would AI, which is built to create content like what people post on the Internet, really just lie?"

[–] KeenFlame@feddit.nu 2 points 1 day ago

Seems like almost everyone understands that it hallucinates.