• 0 Posts
  • 45 Comments
Joined 1 year ago
cake
Cake day: June 12th, 2023

help-circle
  • I constantly hear this but I just want to be the counter argument here.

    Self hosting email is not the impossible tasks that people make it out to be. It is on the more advanced side of things though if you are hosting your primary email that you rely on.

    I’ve been hosting my own email forany years now and have had no issues whatsoever but I also have years of experience and know how email works better than many that have no interest in such.

    I would NOT recommended starting your self hosting journey with email but I will never discourage people from doing it.

    Take your time. Ask questions any time you don’t understand something. Be ready to learn a lot and design a solid plan for disaster recovery.







  • Just know that this is a step forward in the direction of making it technically possible to force people to usw the app.

    I disagree. There is nothing stopping that as it is. What this really does is remove one more level of control from the end user on their device.

    Ask yourself if you trust them to not try and profit from this.

    Im not sure what you are getting at here. Of course i don’t trust “them”. Nor do i trust any corp. It’s those reasons among others why i have completely removed google from my computing life and almost exclusively use open source software as well as self host functionally all network services.



  • In the scope of wireguard it’ll just be a matter of you building appropriate firewall rules.

    Since you want their internet traffic to go through you then i assime you’re effectively pushing a 0.0.0.0/0 route to your clients. You then need to add firewall rules on your server to block traffic to its local subnet and in the future allow traffic to only your jellyfin server.

    This is also pretty simple and nothing wrong with that setup.





  • themachine@lemmy.worldtoSelfhosted@lemmy.worldJourney into self-hosting
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    2 months ago

    You did not answer what VPN tech you are using.

    Without that knowledge i would recommend setting up tailscale and having your users use that. If you want to be fully self hosted you can also run Headscale as the control plane instead of relying on Tailscales own service.

    I recommend tailscale as it is very easy to grant a user privileges to ONLY use an endpoint as an exit node but also grant access to any other endpoints as needed (such as your future jellyfin server) via theor ACLs.




  • If the total data is 3tb and you want disk failure protection I would take your two 6tb disks and put them in a mirror. With the amount of data you have and the drive sizes at your disposal that makes the most sense. This leaves you with 3tb free for growth. If you wanted an additional backup I would recommend storing it in a different location entirely or pay a cloud provider like Backblaze.

    I would do this with ZFS but you can also do this via LVM or just straight md-raid/mdadm. I’m not sure what your issues are with zfs on popos but they should be resolvable as Ubuntu supports zfs fine to my knowledge.

    An alternative you could consider is using mergersfs to logically pool indivial filesystems on each of the disks and then use SnapRAID to provider some level of protection. You’ll have to look into that further if interests you as I don’t have to much info in my head related to that solution. Its not as safe as a mirror but its better than nothing.


  • Your title is about backups but your question seems mostly just about how to set up your storage for backups.

    You can go about pooling disks in a few ways but you first need to define what level of protection from failure you want. Before going further though, how much space do you project that you will need for backups?


  • If you want simple you’ll have to manually decrypt each time it needs doing.

    If you want it to be “automatic” then your best bet is something network based. A “simple” would be to just have a script ssh’s somewhere, pulls the decryption key, and then decrypts the disks. There’s plenty of flaws with this though as while a threat actor couldn’t swipe a single encrypted disk they could just log in as root, get your script, and pull the decryption key themselves.

    The optimal solution would be to also encrypt the root partition but now you need to do network based decryption at boot which adds further complexity. I’ve previously used Clevis and Tang to do this.

    I personally don’tencrypt my server root and only encrypt my data disks. Then ssh in on a reboot or power event and manually decrypt. It is the simplest and most secure option.



  • The primary reason a private track is private is to make it feasible to maintain a curated community. Many users are not good torrent citizens. Many users are not good netizens in the first place. More than a few will look to actively do harm. Keeping a mostly closed community allows the vetting of users and those who end up breaking the rules are dealt with swiftly.

    The extra barrier of entry also helps prevent bad actors from operating on the site. This is of course not a full proof thing but it is obviously much better than a public site.

    Additionally running a private tracker and site takes server resources that are not free. Limiting the total number of users is a way of maintaining uptime by staying within your operational limits.

    I’m sure there are other benefits for private trackers but these are at least a few.

    I am not going to explain why someone on the internet was mean to you. Given the tone of this post I wouldn’t be surprised if it was deserved.