• 1 Post
  • 20 Comments
Joined 2 years ago
cake
Cake day: June 14th, 2023

help-circle
  • Intel NUC running Linux. Not the cheapest solution but can play anything and I have full control over it. At first I tried to find some kind of programmable remote but now we have a wireless keyboard with built-in touchpad.

    Biggest downside is that the hardware quality is kind of questionable and the first two broke after 3 years + a few months, so we’re on our third now.


  • This is my wireguard docker setup:

    version: "3.6"
    services:
      wireguard:
        image: linuxserver/wireguard
        container_name: wireguard
        cap_add:
          - NET_ADMIN
          - SYS_MODULE
        environment:
          - PUID=116
          - PGID=122
          - TZ=Europe/Stockholm
          - ALLOWEDIPS=192.168.1.0/24
        volumes:
          - /data/torrent/wireguard/config:/config
          - /lib/modules:/lib/modules
        ports:
          - 192.168.1.111:8122:8122  # Deluge webui
          - 192.168.1.111:9127:9127  # jackett webui
          - 192.168.1.111:9666:9666  # prowlarr webui
          - 51820:51820/udp           # wireguard
          - 192.168.1.111:58426:58426  # Deluge RPC
        sysctls:
          - net.ipv4.conf.all.src_valid_mark=1
          - net.ipv6.conf.all.disable_ipv6=1
          - net.ipv6.conf.default.disable_ipv6=1
        restart: unless-stopped
    

    Can reach the webuis from LAN, no other network configuration was necessary. 192.168.1.111 is the server’s LAN address. The other services are configured very similar to your qbittorrent, and don’t expose any ports. Can’t promise it’s 100% correct but it’s working for me.





  • For LLMs it entirely depends on what size models you want to use and how fast you want it to run. Since there’s diminishing returns to increasing model sizes, i.e. a 14B model isn’t twice as good as a 7B model, the best bang for the buck will be achieved with the smallest model you think has acceptable quality. And if you think generation speeds of around 1 token/second are acceptable, you’ll probably get more value for money using partial offloading.

    If your answer is “I don’t know what models I want to run” then a second-hand RTX3090 is probably your best bet. If you want to run larger models, building a rig with multiple (used) RTX3090 is probably still the cheapest way to do it.








  • Do you mean that you want to build the docker image on one computer, export it to a different computer where it’s going to run, and there shouldn’t be any traces of the build process on the first computer? Perhaps it’s possible with the –output option… Otherwise you could write a small script which combines the commands for docker build, export to file, delete local image, and clean up the system.