Here you dropped this:
#define ifnt(x) if (!(x))
Here you dropped this:
#define ifnt(x) if (!(x))
How well does it work with super large repos (i.e. Linux, dpdk, etc)? In my experience git plugins (Vim fugitive, zsh git) tend to be a miss with anything larger than a personal project.
Also if the router blocks icmp for some reason you can always manually send an ARP request and check the response latency.
Is there a specific reason you’re looking at shadowsocks? The original developer has been MIA for years. People who used it in the past largely consider it insecure for its original stated purpose
trojan-gfw is a better modern replacement. However that requires a certificate in order to work. You can easily get one via lets encrypt.
At this point, let Shadowsocks, obfs, and kcp die a graceful death like GoAgent before it did.
I don’t think either of us is the target audience here. I can see a “cheaper” (questionable) Pro laptop being useful for students going into college with a limited budget. An undergrad CS/graphic design degree shouldn’t tax an 8gb machine too much, assuming students shut down everything else when doing their once-a-semester major rendering/compiling/model training. If people just want Macbook pro software with more ports, a “cheaper” machine is better than none. Personally, I would still get a used/refurbished machine though.
That being said, my current laptop workload tends to be emacs, qpdfview, Firefox, and tmux on EL9. For the remaining stuff, I usually just spin up a VM then ssh/xrdp into it. As for slack, teams, jabber, etc, I’m happy to report I’ve been out of industry/IT for 1+ years and don’t plan on going back anytime soon. For all I care, Apple can call their models unicorn edition. As long as it sells it’s not stupid.
Another thing you can look into is apptainer/singularity. Basically portable container binaries. Executing the binary automatically runs a program/drops you into a shell inside the container with your $HOME mounted inside. Stuff like cuda also work as long as your host system has appropriate drivers.
You can also port docker containers to apptainer directly via cli.
Oh wow. It supports Kobos as well. Gonna have to check this out. Thanks.
8gb RAM and 256 gb storage is perfectly fine for a pro-ish machine in 2023. What’s not fine is the price point they are offering it (but if idiots still buy that, that’s on them and not apple). I’ve been using a 8gb ram 256 gb storage Thinkpad for lecturing, small code demos, and light video editing (e.g. zoom recordings) this past year, it works perfectly fine. But as soon as I have to run my own research code, back to the 2022 Xeon I go.
Is it Apple’s fault people treat browser tabs as a bookmarking mechanism? No. Is it unethical for Apple to say that their 8GB model fits this weirdly common use case? Definitely.
Just in time to move to IPv6!
What someone does with their 16,777,215 private IPv4 addresses is none of our business…
Now just connect all of that with dumb L2 switches and watch those broadcasts fly!
No, the 2037 problem is fixing the Y2k38 problem in 2037.
Before that there’s no problem :)
Can’t comment much about the docker side since it’s not something I’m familiar with.
For the kernel part, assuming what you’re referring to as UUIDs is the pid namespace mechanism, I’m failing to see how that would add overhead with containers. The namespace lookups/permission checks are performed regardless of whether the process is in a container or not. There is no fast path for non-containerized processes. The worst overhead that this could add is probably one extra ptr chase in the namespace linked list.
Yep it’s Intel.
They said it up until their competitor started offering more than 4 cores as a standard.
A more recent example:
“Nobody needs more than 4 cores for personal use!”
Make sure to test your setup if you are using DAV. Large files can fail if your nextcloud setup is done incorrectly.
Source: idiot who misconfigured PHP that resulted in a DAV client stuck in a retry loop, then getting banned by my own firewall for DoS.
Or just any dpdk program, where any gdb caused slowdown causes the code to “behave as expected”
Isn’t the whole point of these things the “bloated” (CI/CD, issue tracker, merge requests, mirroring, etc) part? Otherwise we’d all be using bare git repos over ssh (which works great btw!)
It’s like complaining about IDE bloat while not using a text editor. Or complaining there’s too many knives in a knife set instead of buying just the chef knife.
CrowdSec has completely replaced fail2ban for me. It’s a bit harder to setup but it’s way more flexible with bans/statistics/etc. Also uses less ram.
It’s also fun to watch the ban counter go up for things that I would never think about configuring on fail2ban, such as nginx CVEs.
Edit: fixed url. Oops!
Simply changing the binary worked for me. Been more than 1 month and no migration issues.
It does still show gitea branding, however.