this post was submitted on 21 Nov 2024
44 points (100.0% liked)
Fediverse
28480 readers
843 users here now
A community to talk about the Fediverse and all it's related services using ActivityPub (Mastodon, Lemmy, KBin, etc).
If you wanted to get help with moderating your own community then head over to !moderators@lemmy.world!
Rules
- Posts must be on topic.
- Be respectful of others.
- Cite the sources used for graphs and other statistics.
- Follow the general Lemmy.world rules.
Learn more at these websites: Join The Fediverse Wiki, Fediverse.info, Wikipedia Page, The Federation Info (Stats), FediDB (Stats), Sub Rehab (Reddit Migration), Search Lemmy
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Something I've thought about a bunch re: recommendation engines is the idea of a "sweet spot" that balances exploration and safety
Though actually I should start by saying that recommendation engines tend to aim to maximise engagement, which is why manosphere type content is so prevalent on places like YouTube if you go in with a fresh account — outrage generates engagement far more reliably than other content. I'm imagining a world where recommendation algorithms may be able to be individually tailored and trained, where I can let my goals shape the recommendations. I did some tinkering with a concept like this in the context of a personal music recommender, and I gave it an "exploration" slider, where at maximum, it'd suggest some really out-there stuff, but lower down might give me new songs from familiar artists. That project worked quite well, but it needs a lot of work to untangle before I can figure out how and why it worked so well.
That was a super individualistic program I made there, in that it was trained exclusively from data I gave it. One can get individual goals without having to rely on the data of just one person though - listenbrainz is very cool — its open source, and they are working on recommendation stuff (I've used listenbrainz as a user, but not yet as a contributor/developer)
Anyway, that exploration slider I mentioned is an aspect of the "sweet spot" I mentioned at the start. If we imagine a "benevolent" (aligned with the goals of its user) recommendation engine, and say that the goal you're after is you want to listen to more diverse music. For a random set of songs that are new to you, we could estimate how close they are to your current taste (getting this stuff into matrices is a big chunk of the work, ime). But maybe one of the songs is 10 arbitrary units away from the boundary of your "musical comfort zone". Maybe 10 units is too much too soon, too far away from your comfort zone. But maybe the song that's only 1 unit away is too similar to what you like already and doesn't feel stimulating and exciting in the way you expect the algorithm to feel. So maybe we could try what we think is a 4 or 5. Something novel enough to be exciting, but still feels safe.
Research has shown that recommendation algorithms can change affect our beliefs and our tastes [citation needed]. I got onto the music thing because I was thinking about the power in a recommendation algorithm, which is currently mostly used on keeping us consuming content like good cash cows. It's reasonable that so many people have developed an aversion to algorithmic recommendations, but I wish I could have a dash of algorithmic exploration, but with me in control (but not quite so in control as what you describe in your options 3). As someone who is decently well versed in machine learning (by scientist standards — I have never worked properly in software development or ML), I think it's definitely possible.