• 1 Post
  • 155 Comments
Joined 3 years ago
cake
Cake day: June 15th, 2023

help-circle
  • tburkhol@lemmy.worldtoSelfhosted@lemmy.worldIPv6
    link
    fedilink
    English
    arrow-up
    5
    ·
    24 days ago

    Not familiar with opnSense, but on your PC, you can check the address it assigns - if it’s /128, it’s a single address.

    My ISP does not assign a prefix for delegation unless you specifically ask for it. I had to add “request_prefix 1” to my dhclient.conf file to get a /64 I assume opnSense has a friendly setting somewhere for that. For me, the key phrase was ‘prefix delegation.’ After I got that, I could search around and get my solution.



  • If you only need it to be accessible inside your home, then you just need to run your own DNS. Have your dhcp point at your DNS and your DNS declare itself the master for your domain.

    To get full functionality, you’ll probably want to have your registrar point to the public IP you get from your ISP as the domain’s authoritative name server.You should be able to script it to update the registrar when your ISP changes your IP, but that usually happens infrequently enough to do manually. Obviously can’t do that if you’re behind CGNAT.

    To get Lets Encrypt certificates, you can do the DNS challenge. If your ISP gives you a (even inconsistent) public IP, you can do fancy ‘views’ with your selfhosted DNS, where it responds with private IPs inside your network and your ISP-given IP outside your network. I have certbot set up to expose my DNS & web server just before it starts its renewal process, then close the firewall after. Once you have the certificate, you can move it to where ever it will actually be used.


  • To me, the nonstandard port is mostly nice for reducing log spam from scripts. The risk is that using a nonstandard port lulls one into a false sense of security and overlook good sshd practices. Good sshd practices will prevent the script-kiddies just as well as the non-standard port, while a non-standard port will not challenge a targeted attack. And, if you interact with multiple servers, it can be inconvenient to remember a different port for each one.



  • In the old days, university IT put essentially no access controls on their networks, so students’ dorm computers were completely exposed to the internet. Any service you started was immediately, globally accessible. Some big sites, including slashdot and facebook, got their start in some kid’s dorm room. I feel like access controls really got going in the early 00’s - first for residential, then for broader campus.

    Check with your IT people - they may have policy or conditions under which they will expose ports on your personal computer to the internet. Otherwise, your best bet is probably free-tier AWS or Oracle.

    Not free, but there are some ‘KVM VPS’ providers out there that will rent you a small, internet exposed computer pretty cheap. They can be a good platform for experimenting with self-hosting services, without exposing your personal equipment or home network. eg: 1CPU/1GB RAM/24GB SSD $12/year https://my.racknerd.com/cart.php?a=add&pid=903



  • Depending on the board in your mini-server, you may have enough SATA ports to plug in directly. I have a system similar to what you’re describing (N100 with 4x 2TB HDDs with 1.5TB data): 2 of those drives are set up in RAID1 (mirror), and once a month, I plug in one of the spares, rsync the array to it, and unplug it. Every 3 months or so, I swap the offline drive with an offsite drive. I used to use a USB dock for the offline drive, but I got a 3-bay hot-swap enclosure to make the whole process faster and easier.

    The server shares the array via NFS and SMB, and it is absolutely a NAS for all my other systems.

    If you expect to exceed 2TB data within 2 years, then you’ll need to replace all 4 of those 2TB drives in 2 years. You might, today, get a pair of 4 TB drives and one 2TB, use the 4TB as your main storage, the 2TBs as rotating backups, and wait until you actually outgrow 2TB to upgrade the backups.


  • I see you’re getting lots of advice just to use c/selfhosted as a free consultant. That’s good advice if you’re self-motivated and focused.

    If you want someone to be a coach through the process, to keep you focused and moving, that’s a) a slightly different skillset and b) worth putting in the description. I mention this only because I have a bunch of aspirational projects on my to–do list that have just sat there for literally years because of perfectionism, anxiety, and maybe some undiagnosed ADHD. I’ll also counter by noting that a lot of people, this time of year, buy a gym membership on the theory that spending the money will somehow force them actually to go to the gym, only to find that spent money is not actually a motivator.



  • If you want it to be an actual community service, then you want it to be something that outlives your residence, your tenure as event coordinator, and your interest in being the neighborhood IT guy. It’ll be much easier to transfer control of a VPS to your successor than to give them hardware that also hosts a bunch of your personal services.

    You can start with a very small, nearly free VPS while you recruit users & scale up as (if) anyone bites. Probably even get the HOA to pay for it.


  • tburkhol@lemmy.worldtoSelfhosted@lemmy.worldRaspberry Pi 4B
    link
    fedilink
    English
    arrow-up
    6
    ·
    5 months ago

    I got my Pi4 to be a media player - LibreElec or Kodi - for my old, not-smart TV. It plays my library of CDs&DVDs, frontend for OTA TV, and a variety of streaming services. Fanless, so it doesn’t distract from audio, low power, so I don’t mind leaving it on 24/7. You can configure it to listen to a USB IR receiver, but I control mine from phone via web. The actual media library/NAS and tvheaded run on an old desktop in another room.

    My favorite thing is all the sensors you can hook up. Adafruit & Sparkfun have a wide array of sensors with breakout boards for simplicity and well documented python libraries. I started just logging temperature, humidity, then air quality, CO2 to my own database and web page, but eventually expanded to full HomeAssisstant system.

    Pihole.



  • A lot depends on how many users you expect and how much media you expect. For one or two users with that stack, transcoding media is really the only CPU load. If most of your media is already in your desired format, then that’s not a big deal.

    My stack is pretty similar (no *arr, plus tvheadend, homeassistant and a kodi frontend) for two users and it sits near idle all day long. It runs on an N100 NAS system off Aliexpress with 16GB and will transcode 1080p to x264 at just about playback speed… System runs from a 100 GB nvme, with a couple half-full 4 TB WD Reds for data. 35-ish Watts, maybe an extra 5 when actively transcoding. Used to be ~150 USD,

    If you want a lot of 4k content, then I’d definitely go with the GTX 1660.