Openstreetmap doesn’t seem to label bodies of water, but it should definitely add this label
Openstreetmap doesn’t seem to label bodies of water, but it should definitely add this label
What do you folks on hyprland/sway use for your shell / toolbars / launcher? I tried nwg and it was… OK but pretty clunky. No shade for the developers of the project, all the settings pages and system config stuff is a TON to put together…
I don’t need something as full-featured as KDE or Gnome Settings. I’d prefer a well-polished minimalist launcher and task manager / toolbar over something that does everything
Yeah yeah yeah… yeah
no fact checking and a ton of bots being added? Sounds like a fantastic place to spend time. /s
Posting pictures of riced-out tiling window manager desktops is the hot new game they’re looking for.
DNFTA
First a caveat/warning - you’ll need a beefy GPU to run larger models, there are some smaller models that perform pretty well.
Adding a medium amount of extra information for you or anyone else that might want to get into running models locally
If you look at https://ollama.com/library?sort=featured you can see models
Model size is measured by parameter count. Generally higher parameter models are better (more “smart”, more accurate) but it’s very challenging/slow to run anything over 25b parameters on consumer GPUs. I tend to find 8-13b parameter models are a sort of sweet spot, the 1-4b parameter models are meant more for really low power devices, they’ll give you OK results for simple requests and summarizing, but they’re not going to wow you.
If you look at the ‘tags’ for the models listed below, you’ll see things like 8b-instruct-q8_0
or 8b-instruct-q4_0
. The q part refers to quantization, or shrinking/compressing a model and the number after that is roughly how aggressively it was compressed. Note the size of each tag and how the size reduces as the quantization gets more aggressive (smaller numbers). You can roughly think of this size number as “how much video ram do I need to run this model”. For me, I try to aim for q8 models, fp16 if they can run in my GPU. I wouldn’t try to use anything below q4 quantization, there seems to be a lot of quality loss below q4. Models can run partially or even fully on a CPU but that’s much slower. Ollama doesn’t yet support these new NPUs found in new laptops/processors, but work is happening there.
It’s a good thing that real open source models are getting good enough to compete with or exceed OpenAI.
Really love arch and the AUR. I’ve been tempted to get nix set up for the rare cases when there’s no AUR package or the AUR package is unmaintained. I figure if there’s no package in the AUR or nixpkgs, it’s probably not worth running.
btop reports some gpu, network and disk information that I don’t think shows up in htop, feels a bit more comprehensive maybe? Both are fine, but I too use btop, it’s nice.
Random trivia: I think btop has been rewritten like 3-5 times now? It’s sort of an inside joke to the point that someone suggested another rewrite from C++ to Rust ( https://github.com/aristocratos/btop/issues/5 ). I guess the guy just likes writing system monitoring console apps.
FWIW they didn’t merge it, they closed the PR without merging, link to line that still exists on master.
The recent comments are from the announcement of the ladybird browser project which is forked from some browser code from Serenity OS, I guess people are digging into who wrote the code.
Not arguing that the new comments on the PR are good/bad or anything, just a bit of context.
Been 100% linux for like 6-9 months now, these stories make me thankful for finally making the switch.
I’ve tried to make the switch 3-4 times in the past and was stopped by 2 main things:
The experience was so much better this time and I really have no regrets. I don’t imagine I’ll ever run Windows again outside of a VM
And what about Dr Pepper Brisket?
Elon “Nick Cannon” Musk
Synology nas are nice. I will say there’s definitely a nice UI there and they generally work well. But there is a good bit of lock-in and there are some really reasonable roll-your-own hardware and software options these days.
If you want something that just works, doesn’t need to be super configurable and is easiest to set up and manage, get a synology. If you don’t mind putting in some work or if you need to really tweak some stuff, roll your own
Rip up the Reddit contract and don’t use that data to train the model. It’s the definition of a garbage in garbage out problem.
Fwiw they’re able to do the same thing by the sound of someone typing a password across the room. Not advocating for fingerprints or anything, just these exotic hacks are everywhere
I like that. If there was a site that did like The Razzies for movies but for technology enshitification, I would definitely watch, and probably follow a blog if it was done well
Also, reporting that Trump Tower (either) should be renamed Mexico Tower feels pretty good.