

don’t know the details but my general IT knowledge says that: single unicode character/glyph can take up to 4 bytes instead of 1 (ascii).


don’t know the details but my general IT knowledge says that: single unicode character/glyph can take up to 4 bytes instead of 1 (ascii).


Little bits and pieces but mostly bug fixes - I like my shit working but maintenance is not my strong suit, more of a traveling contributor or drive-thru fixer.
I believe I fixed calling in one electron messenger.com wrapper before - that was fun but these days I usually try to help the game BAR whenever I have extra time.
Edit: Keep forgetting but I am also maintaining few apps on AUR, nothing big except for maybe one helper tool/calculator for EVE online
Linux has kind of two forms of memory pages (entries in RAM), one is a file cache (page cache) and the other is “memory allocated by programs for work” (anonymous pages).
When you look at memory consumed by a process you are looking at RSS, page/file cache is part of kernel and for example in btop corresponds to Cached.
Page cache can never be moved into swap - that would be the same as duplicating the file from one place on a disk to another place on a (possibly different) disk.
If more memory is needed, page cache is evicted (written back into the respective file, if changed).
Only anonymous pages (not backed by anything permanent) can be moved into swap.
So what does “PostgreSQL heavily relies on the OSs disk cache” mean? The more free memory there is, the more files can be kept cached in RAM and the faster postgres can then retrieve these files.
When you add zram, you dedicate part of actual RAM to a compressed swap device which, as I said above, will never contain page cache.
In theory this still increases the total available memory but in reality that is only true if you configure the kernel to aggressively “swap” anonymous pages into the zram backed swap.
Notes: I tried to simplify this a bit so it might not be exact, also if you look at a process, the memory consumed by it is called RSS and it contains multiple different things not just memory directly allocated by the code of the program.


it’s not just Let’s Encrypt, it’s because of https://certificate.transparency.dev/


Flameshot: screenshotting tool with everything you would ever need for screenshots
have it on dietpi (pi 4) + tailscale at home to monitor my dedics


Unsure how well known it is, but flameshot (screenshot tool).
I prefer CLI usually, so: zoxide, the zsh git plugin for aliases (e.g.: gst is git status), fzf zsh plugin and the tldr command comes in handy sometimes.
Also, this might be useful just for me, but due to orientation of my living space, I have to fiddle with monitor brightness at least three times a day so I made myself a little Qt tray wrapper around ddcutil’s ddcquery which can change standard vesa monitors brightness/contrast (DDC/CI communication).
There is also ddcui/gddccontrol GUI that does the same thing.


For cli oriented folks, ncdu is a great cli alternative of QDirStat


Yeah Nick (the guy behind this) is one of a kind (in a good way) lol
Also recommend checking out this demo of NotCurses https://www.youtube.com/watch?v=cYhZ7myXyyg it’s great
There is also https://github.com/jokob-sk/NetAlertX


My instance is close to two years old now, and on average has had about 2 MAU, with no (local) communities.
Currently we have about 700 active federated communities (that had any federated activity within last month), out of 900.[1]
The on-disk size of both lemmy and pict-rs database[2]
postgres@postgres:~$ pwd
/var/lib/postgresql
postgres@postgres:~$ du -sh data/
31G data/
I use pict-rs with S3 provider and the bucket size is currently at 22.82 GB (read: external network storage, this is probably mostly just thumbnails[3]).
So in total there is almost 54GBs spent just for lemmy.
So assuming you have 100G remaining after system stuff and dedicate that box only to lemmy (and pict-rs media files) and use it mostly for yourself [4], you should be alright for about 3-4 years (assuming that I am gaining about 27GBs total per year and that you will federate with a similar amount of a similarly active communities).
If you offload media storage to a hosted S3 bucket[5] then you should be good for a lot longer as you will only need space for the postgres databases.
The rest is either dead (instance gone) or no one is subscribed to them anymore (as such my instance is not getting any new content from there: neither posts nor comments or votes) ↩︎
Postgres itself reports about 2G less, don’t really know why but I am guessing it has something to do with the filesystem being btrfs ↩︎
Edit: I currently do not use the “privacy” mode of pict-rs where it proxies all content (so that a bad guy can’t post an image link to his server and unmask users IPs), this would increase the S3 size and slightly postgres size. ↩︎
You should use Lemmy Subscriber Bot to automatically federate little bit of random communities so that public All feed is not exact copy (minus NSFW comms) of whatever you as the only user subscribe to. ↩︎
Though keep in mind that S3 buckets eventually cost some money too, for example Cloudflare R2 charges $0.015 per 1GB, above the first 10GBs. ↩︎


Someone did s/double/decimal/g (find ‘double’, replace by ‘decimal’) on the whole project.
Please decimal check
lol


Here is hoping this will at least make it easier to take back control of tuya devices


Also, proxy_buffering


Sounds like you need to instrument it yourself.
It could be as “easy” as calling the endpoints yourself and saving the sensor states in any kind of storage grafana supports, then making a dashboard on top of that data.
Maybe Zabbix could also work


Finally


Thanks, especially for that openwrt mesh bit, that might end up as the the best solution.


Looking into it, ty!


Good tip, thanks!
before getting a pocketbook I was using ReadEra and it worked alright (for basic reading)