behohippy

joined 1 year ago
[–] behohippy@lemmy.world 2 points 1 year ago

The advancements in this space have moved so fast, it's hard to extract a predictive model on where we'll end up and how fast it'll get there.

Meta releasing LLaMA produced a ton of innovation from open source that showed you could run models that were nearly the same level as ChatGPT with less parameters, on smaller and smaller hardware. At the same time, almost every large company you can think of has prioritized integrating generative AI as a high strategic priority with blank cheque budgets. Whole industries (also deeply funded) are popping up around solving the context window memory deficiencies, prompt stuffing for better steerability, better summarization and embedding of your personal or corporate data.

We're going to see LLM tech everywhere in everything, even if it makes no sense and becomes annoying. After a few years, maybe it'll seem normal to have a conversation with your shoes?

[–] behohippy@lemmy.world 9 points 1 year ago

For the really old stuff, I used to do NetBSD. I'm sure their 32bit x86 support is still top notch.

[–] behohippy@lemmy.world 4 points 1 year ago

These are amazing. Dell, Lenovo and I think HP made these tiny things and they were so much easier to get than Pi's during the shortage. Plus they're incredibly fast in comparison.

[–] behohippy@lemmy.world 4 points 1 year ago

Subscribed. That last episode of AAA was heartbreaking.

[–] behohippy@lemmy.world 1 points 1 year ago

Bad article title. This is the "Textbooks are all you need" paper from a few days ago. It's programming focused and I think Python only. For general purpose LLM use, LLaMA is still better.

[–] behohippy@lemmy.world 0 points 1 year ago (1 children)

I hear good things about Traefik. Basically all I need is a reverse proxy that will handle re-writing URLs and websockets and slapping some ssl and auth on it. If something is easier for that, I'm all ears.

 

I host a ton of services running behind my nginx reverse proxy (basic auth + lets encrypt). On the whole it works really well with nearly everything I throw at it. Lately, there's been a lot of gradio/websocket/python stuff coming from the AI community like the local llama and stable diffusion stuff. Not sure what's causing it but there's always weird issues when I try to reverse proxy them.

Does anyone have some magic settings that "just work" with these weirdo web apps?

 

He's 5 today

[–] behohippy@lemmy.world 1 points 1 year ago

Yep, I'm using an RTX2070 for that right now. The LLMs are just executing on CPU.

[–] behohippy@lemmy.world 1 points 1 year ago (1 children)

Do you recommend this email provider? Lots of people looking to get off gmail lately.

[–] behohippy@lemmy.world 2 points 1 year ago (3 children)

Are you running your own mail server? I only ever integrated Spamassassin with postfix.

[–] behohippy@lemmy.world 2 points 1 year ago (4 children)

Stable Diffusion (Stability AI version), text-generation-webui (WizardLM), a text embedder service with Spacy, Bert and a bunch of sentence-transformer models, PiHole, Octoprint, Elasticsearch/Kibana for my IoT stuff, Jellyfin, Sonarr, FTB Minecraft (customized pack), a few personal apps I wrote myself (todo lists), SMB file shares, qBittorrent and Transmission (one dedicated to Sonarr)... Probably a ton of other stuff I'm forgetting.

[–] behohippy@lemmy.world 1 points 1 year ago

Yup, mostly running pretrained models for text embedding and some generative stuff. No real fine tuning.

[–] behohippy@lemmy.world 1 points 1 year ago

Yup, typically we get into it after upgrading an older PC or something and instead of selling the parts, just turn it into a server. You can also find all sorts of cheap/good stuff on ebay from office off-lease.

 

Ryzen 5900X, 64 gig DDR4-3200, 2tb ssd,10tb hdd and an RTX2070. Hosting Stable Diffusion, various llama.cpp instances with python bindings, jellyfin, sonarr, multiple modded minecraft servers, and a network file share.

 

She's mostly good. Mostly.

view more: next ›