I’ve been doing this for at least a decade now and the drives are just as reliable as if you bought them normally. The only downside is having to block one of the pins on the SATA connector with kapton tape for it to work.
I’ve been doing this for at least a decade now and the drives are just as reliable as if you bought them normally. The only downside is having to block one of the pins on the SATA connector with kapton tape for it to work.
Damn I forgot no content existed online and could be profitable before YouTube came along and saved us all from the dark ages.
Exactly, then it could have just been a text list on a webpage and we’d all be better off.
I like the workflow of having a DNS record on my network for *.mydomain.com pointing to Nginx Proxy Manager, and just needing to plug in a subdomain, IP, and port whenever I spin up something new for super easy SSL. All you need is one let’s encrypt wildcard cert for your domain and you’re all set.
IIRC from running into this same issue, this won’t work the way you have the volume bind mounts set up because it will treat the movies and downloads directories as two separate file systems, which hardlinks don’t work across.
If you bind mounted /media/HDD1:/media/HDD1 it should work, but then the container will have access to the entire drive. You might be able to get around that by running the container as a different user and only giving that user access to those two directories, but docker is also really inconsistent about that in my experience.
If you want Proxmox to dynamically allocate resources you’ll need to use LXCs, not VMs. I don’t use VMs at all anymore for this exact reason.
I had done a few easier Linux installs on Raspberry Pis and VMs in the past, but when I decided I wanted to try using Linux as my daily driver on my desktop (dual-booted with Windows at the time) I decided to go with a manual Arch install using a guide and I would 100% recommend it if you’re trying to pick up Linux knowledge. It’s really not a difficult process to just follow step-by-step, but I looked up each command as they came up in the guide so I could try to understand what I was doing and why.
I don’t know what packages archinstall includes because I’ve never used it, but really the biggest thing for me learning was booting into a barebones Arch install. Looking into the different options for components and getting everything I needed setup and configured how I wanted was invaluable.
That being said, now that I know how, is that how I would choose to install it? Nah, I use the CachyOS installer now, but if I wanted stock Arch I’d probably use archinstall.
That I’m not sure of. My proxmox host is headless and none of my containers have a GUI so I haven’t tried.
You can also pass the GPU to multiple LXCs that will share it vs it being tied to a single VM. I use VMs as little as possible in Proxmox these days.
Glad to hear this is being worked on, thanks for sharing this. I assumed it was related to my config and was putting off looking into it further.
Visual discomfort because it looks like an slightly older app? What kind of issue is that???
You’ve met an iOS user.
Absolutely, if it was anything I needed or even really wanted to be sure was reliably available I’d never put it on a free VPS.
Now, something trivial like this that just requires installing wireguard and nginx, copying over some configs, and changing a DNS record? Hard to beat free.
I know everyone loves to shit on Oracle, but a free-tier Oracle VPS would solve this.
Or if you want something decent pay for a cheap VPS.
I mean that’s not inherently bad, what you do with that data could be though.
Hoarders Household power users
Then offer to host the database/servers on your own dime and make it free for everyone. It’s a completely reasonable price with the option to try it for free before paying.
I was skeptical about DeArrow but it is absolutely worth it. I don’t use the thumbnail replacer but removing clickbait titles makes YouTube so much more usable.
What you’re describing is the whole point of flatpaks. Just don’t use flatpaks then.
I’m sure the artists behind the music for the 20+ year old games this could be used for are really feeling the pain of their creative rights being abused from people trying to still enjoy their art after all this time, you wet blanket.
The issue is that the docker container will still be running as the LXC’s root user even if you specify another user to run as in the docker compose file or run command, and if root doesn’t have access to the dir the container will always fail.
The solution to this is to remap the unprivileged LXC’s root user to a user on the Proxmox host that has access to the dir using the LXC’s config file, mount the container’s filesystem using pct mount, and then chown everything in the container owned by the default root mapped user (100000).
These are the commands I use for this:
find /var/lib/lxc/xxx/rootfs -user 100000 -type f -exec chown username {} +; find /var/lib/lxc/xxx/rootfs -user 100000 -type d -exec chown username {} +; find /var/lib/lxc/xxx/rootfs -user 100000 -type l -exec chown -h username {} +; find /var/lib/lxc/xxx/rootfs -group 100000 -type f -exec chown :username {} +; find /var/lib/lxc/xxx/rootfs -group 100000 -type d -exec chown :username {} +; find /var/lib/lxc/xxx/rootfs -group 100000 -type l -exec chown -h :username {} +
(Replace xxx with the LXC number and username with the host user/UID)
If group permissions are involved you’ll also have to map those groups in the LXC config, create them in the LXC with the corresponding GIDs, add them as supplementary groups to the root user in the LXC, and then add them to the docker compose yaml using group_add.
It’s super confusing and annoying but this is the workflow I’m using now to avoid having to have any resources tied up in VMs unnecessarily.