otl

joined 8 months ago
[–] otl@apubtest2.srcbeat.com 4 points 8 months ago

Huh wow didn't know Teams didn't have multiple screen sharing. Nice work.

Something does irk me about the whole approach, though. We have accomplished some amazing stuff in the software world. But Jesus almighty we have some serious interoperability problems, to the point where we literally just send pixels to one another because that's all we've got. Reminds me of that glib statement "the web has become just 4 websites full of screenshots of each other" (or however it goes).

[–] otl@apubtest2.srcbeat.com 2 points 8 months ago

No worries! Writing that down actually helped clarify some of my thoughts.

Something extra: distributed computing.

Let's say you have 3 processes that need to communicate with one another. There's heaps of tooling available in OSs to manage those processes. Logging, networking, filesystem access, privilege separation, resource allocation... all provided by the host OS without installing anything. But what if those 3 processes can't run on one "machine"? Which process should go where? What if it needs 8GB memory but there's only 6GB available on some of the machines? Who controls that?

Systems like Kubernetes, Nomad, Docker Swarm etc. offer a way to manage this. They let us say something like:

  • run this process (by specifying a container image),
  • give it at least these resources (xGB memory, xvCPUs)
  • let it communicate with these other processes (e.g. pods, overlay networks...)

These systems manage containers. If you want to do distributed computing and want to take advantage of those systems to manage it, stuff needs to be run in containers.

Containers are not the only way to do distributed computing - far from it! But over the past few years this particular approach has become popular in the, umm... "commercial software development industry".

Opinion. Are Linux containers something to look into as someone who doesn't work in the industry? Unless you're interested in how containers themselves work and/or distributed computing; frankly - no. Computers are still getting faster and cheaper. So why is all this stuff so popular in the commercial world? I'll end with some tongue-in-cheek.

Partly it's because the software development industry is made up of actual human beings who have their own emotions and desires. Distributed computing is a fun idea because tech people are faced with challenges tech people are interested in.

Boring: can we increase our real estate agency brand recognition by 200%? We could provide property listings as both a CSV and PDF to our partners! Our logo could go on the PDF! Wow! Who knows how popular our brand could be?

Fun: can we increase throughput in this part of the system by 200%? We might need to break that component out to run on a separate machine! Wow! Who knows who fast it could go?

[–] otl@apubtest2.srcbeat.com 2 points 8 months ago (1 children)

When you specify To: localtesting@aussie.zone how does the bridge know if you meant https://aussie.zone/c/localtesting or https://aussie.zone/u/localtesting instead?

Good question

The process of going from that email-like address to an ActivityPub Actor (https URL) is done by WebFinger. So whatever we get back from that lookup is who we send the message to. For example, when we look up localtesting@aussie.zone we get this response (truncated for readability):

{
  "rel": "self",
  "type": "application/activity+json",
  "href": "https://aussie.zone/c/localtesting",
  "properties": {
    "https://www.w3.org/ns/activitystreams#type": "Group"
  }
}

The value for type, application/activity+json tells us that we can send ActivityPub stuff to it.

You can play around with WebFinger in the browser.

Hope that makes sense?

[–] otl@apubtest2.srcbeat.com 1 points 8 months ago

Markdown was originally designed for exactly your use case. The Daring Fireball blog has been using Markdown for 20 years now.

I always forget how old Markdown is. And the syntax it's inspired from - plain-text email - has been in use for even longer!

[–] otl@apubtest2.srcbeat.com 7 points 8 months ago (5 children)

Containers are used for a whole bunch of reasons. I'll address just one: process isolation. I'll only do one because I've ran into times when containers were not helpful. And it may lead to some funny stories and interesting discussion from others!

A rule of thumb for me is that if the process is well-behaved, has its dependencies under control and doesn't keep uneccesary state, then it may not need the isolation provided by a container and all the tooling that comes with it.

On one extreme, should we run ls in a container? Probably not. It doesn't write to the filesystem and has only a handful of dependencies available on pretty much any Unix-like/Linux system.

But on the other extreme, what about that big bad internal Node.JS application which requires some weird outdated Python dependencies that has many hardcoded paths and package versions? The original developer is long gone. It dumps all sorts of shit to the filesystem. Nobody is really sure whether those files are used as a cache or they contain some critical state management. Who wants to spend the time and money to tidy that thing up? In this scenario containers can be used to hermetically seal a fragile thing. This can come back to bite you. Instead of actually improving the software to be portable and robust enough to work in varied execution environments (different operating systems, on a laptop, as a library...), you kick the can down the road.

[–] otl@apubtest2.srcbeat.com 4 points 8 months ago

Hey mate thanks for writing all that down. Gives me a few ideas - in particular how NNTP could fit in to the equation.

LKML can be accessed by a usenet client at nntp.lore.kernel.org. In theory, then, it should be possible to at least read both LKML and ActivityPub stuff hosted at the same NNTP server. To get that working means making sure all the conversion and mapping of concepts are as clean as possible. It's a good test!

Thanks again!

[–] otl@apubtest2.srcbeat.com 2 points 8 months ago

Don't tempt me :D

[–] otl@apubtest2.srcbeat.com 1 points 8 months ago (2 children)

I'd much rather have this put in front of LKML than the terrible interface they have right now.

Interesting. Could you go into this a bit more? Do you mean for example being able to use the Lemmy web UI to read LKML?

[–] otl@apubtest2.srcbeat.com 2 points 8 months ago

Oh wow awesome. Thanks so much!

It needs a bit more fleshing out, removing a bit of hardcoding too. And of course I want to publishit under an open source licence too. Thanks for the tip - I've written it down in the growing TODO file! :)

[–] otl@apubtest2.srcbeat.com 3 points 8 months ago

This is interesting, but have you considered porting to Usenet?

Yes that's on the list! Now that I have a whole bunch of RFC822 files, (in Maildirs) I can also serve them over read-only NNTP. This was the original goal actually - I like the idea of using the simpler protocol NNTP over IMAP to read stuff.

[–] otl@apubtest2.srcbeat.com 12 points 8 months ago (2 children)

The alternative is something like FediSeer where you can get sites guaranteed by others and block anything not given the all-clear, but that really harms the ability for new sites to appear.

If something like this were to gain lots of traction I'd hope it would be something not too difficult to implement by smaller new sites.

What really sucks is the situation with email now: it's really tricky to get stuff delivered if you're not Google/Microsoft. The barrier to entry is way, way too high :(

 

What have I done?! My abomination of an idea of bridging my email and ActivityPub progresses. If you see this message, something is working! Comments replies are welcome as it's a good test of this system :)

People keep saying ActivityPub is a lot like email. If it's so similar to email, could I use my email client to interact with the fediverse?

Previously I did this by writing a SMTP interface to the Mastodon HTTP API. That worked. But as we probably know, the fediverse is not Mastodon; it's really ActivityPub. The real deal would be working with ActivityPub directly, not the Mastodon HTTP API.

And that's now (mostly?) working! In shonky diagram form, sending looks like this:

laptop --SMTP--> my_server --ActivityPub--> fediverse

Replies look like this:

fediverse --ActivityPub--> my_server --SMTP--> mailbox <--IMAP-- laptop

my_server translates back and forth between ActivityPub messages and mail messages.

For example given the message:

Date: Wed, 6 Mar 2024 16:37:59 +1100
From: Oliver Lowe <otl@apubtest2.srcbeat.com>
To: localtesting@aussie.zone
Subject: test 2

test hello world!

The following ActivityPub message is created:

{
	"@context": "https://www.w3.org/ns/activitystreams",
	"id":"https://apubtest2.srcbeat.com/outbox/1709703480070628170",
	"type":"Note",
	"name":"test 2",
	"to": ["https://aussie.zone/c/localtesting","https://www.w3.org/ns/activitystreams#Public"],
	"cc": ["https://aussie.zone/c/localtesting"],
	"published":"2024-03-06T16:37:59+11:00",
	"attributedTo":"https://apubtest2.srcbeat.com/actor.json",
	"content":"test hello world!",
	"mediaType":"text/markdown"
}

There's still a lot of bugs (of course) and unimplemented bits (of course). I can't call this a proper fediverse service yet. I'm going to roll with this for a bit and see how it holds up.

view more: next ›