So, I have a few services (Jellyfin, Home Assistant, etc) that I am running, and have been acessing via their IP’s and port numbers.

Recently, I started using NGINX so that I could setup entries in my Pi Hole, and access my services via some made up hostname (jellyfin.home, homeassistant.home, etc).

This is working great, but I also own a few domains, and thought of adding an SSL cert to them as well, which I have seen several tutorials on and it seems straight forward.

My questions:

  • Will there be any issues running SSL certs if all of my internal service are inward facing, with no WAN access? My understanding is that when I try to go to jellyfin.mydomainname.com, it will do the DNS lookup, which will point to a local address for NGINX on my network, which the requesting device will then point to and get the IP of the actual server.

  • Are there risks of anything being exposed externally if I use an actual CA for my cert? My main goal is to keep my home setup off of the internet.

  • phi@lemmy.world
    link
    fedilink
    English
    arrow-up
    11
    ·
    1 year ago

    i have a similar setup at home. the way i did it was using certbot and dns verification. i pointed my domain’s NSs to digitalocean’s NS and then i downloaded the certbot-digitalocean-dns plugin, created an API key for DO and stored it somewhere and then certbot took care of everything else. nothing is exposed to the internet

    • pacocascadero@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 year ago

      This is the way for services not exposed to the internet. Thera are multiple DNS providers supported (I use Cloudflare personally). At the other hand if the service is published to the internet HTTP validation is very simple to configure as well. I have stopped using Nginx as a reverse proxy and use Traefik for conteinerised services or Caddy for the rest. Both proxies support ACME protocol out of the box.

    • root@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Very nice! And you don’t have to worry about adding the cert to each device that wants to use the service, right? Since this isn’t a self hosted CA.

      • phi@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        exactly. that was the main thing i wanted to avoid. i also have nginx-proxy-manager in front of all my apps which also automates some things (like requesting new certs or renewing them when the time comes)

        • root@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Ooo, very nice! If I use that script, can I generate certificates for a made up domain within my network (eg *.homelab), or do I need to use a domain I actually own?

  • Aurailious@beehaw.org
    link
    fedilink
    English
    arrow-up
    5
    ·
    1 year ago

    If you use Let’s Encrypt, or any public CA, all of your domains and certificates will be public. You can use a wildcard to avoid revealing subdomains. There is a website that you can use to search what is available, but I don’t remember what it is.

    I suspect there aren’t any serious risks to having that information revealed. The only real reason would be privacy against which services you are using on that domain.

    • phi@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      yeah true but if the DNS records aren’t actually pointing anywhere then there’s no real threat no? because everything stays in the internal network

      • Aurailious@beehaw.org
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        You have to use a public DNS registrar, and that DNS record has to point to your public IP if you want to automate to a public CA. All of my subdomains are in my local DNS server though and I use a wilcard for them. So no one externally can go to jellyfin.mydomain.com, but they could go to www.mydomain.com to my IP, but that doesn’t forward on my router either.

        But also only automated scrappers are going to look for my domain too and they are going to be blocked in the same way automated scrappers for residential IPs are blocked. I could be wrong, but I don’t think there are ways to bypass security with knowing the domain name tied to an IP.

  • Omripresent@leddit.social
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    1 year ago

    Shouldn’t be any risk if it’s all local.

    For an internal domain you’ll need to set up your own internal CA to sign certs for your fqdns. The risk comes from any mishandling of that new CA since you’ll need to install it as a trusted root on all of your devices and if someone gets a hold of it nothing would stop them from creating a MITM attack for let’s say yourbank.com

    If you have the CA’s key under lock then you should be good.

    • pacocascadero@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 year ago

      Don’t use internal domain, use standard domain + split DNS instead. Much simpler to handle certificates for internal services with ACME protocol.

    • root@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Gotcha. Yeah I read about doing a self-hosted CA, but then I have to add the cert to every device that needs access to the service, which I don’t think the family would be thrilled about. I was going to use the cert generator in NGINX and use the key from my actual domain. This way I don’t need to add the certs manually.

      My only worry is exposing something accidentally, but if my firewall rules prevent any outside access from my services (Jellyfin, Nginx, Homelab, etc) and the only thing with internet is the device accessing it (a laptop or TV), then I think I should be ok…

      • Omripresent@leddit.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        If you have a domain you own that’s the way to go, I went by your .home naming assuming that’s what you’re using. Since .home can’t be registered similar to .local, LetsEncrypt wouldn’t be an option.

        I have a split DNS setup on my end so a service like jellyfin would resolve only internally since I want to limit it, but others would be both public and internal.

    • dan@upvote.au
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      For an internal domain you’ll need to set up your own internal CA

      No real need to run your own CA. As long as you have an actual domain name, you can use Let’s Encrypt with DNS challenges to get certificates for internal servers.

  • AyyLMAO@exploding-heads.com
    link
    fedilink
    English
    arrow-up
    3
    ·
    1 year ago

    Will there be any issues running SSL certs if all of my internal service are inward facing, with no WAN access?

    If you’re using a third party CA, periodically renewing certificates in my experience. The authority needs to be able to connect to the device it’s issuing a cert to, for it to handle a security challenge iirc.

    If you set up your own CA, none that I know of.

    My main goal is to keep my home setup off of the internet.

    Then I don’t understand the need for neither domain names nor third party signed certs. Can’t you use PiHole as a configurable DNS server, just make any domain name go to any of your local devices?

    • root@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      That was my concern too. NGINX would need access to the internet in order to renew the certs.

      Then I don’t understand the need for neither domain names nor third party signed certs. Can’t you use PiHole as a configurable DNS server, just make any domain name go to any of your local devices?

      Yes, that is how it is currently setup, and how I may end up leaving it. Right now, I can go to jellyfin.home, and that request gets routed to my pihole which has custom DNS entries, which then points to NGINX and NGINX forwards it to the correct IP/ port. All works as expected, except it is not https (which is not that big of a deal since all my stuff is restricted from the outside world). Just an OCD itch I’m trying to scratch.

      • AyyLMAO@exploding-heads.com
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        1 year ago

        Hey, I advocate https even for LAN only, most people don’t think about the Wifi attack vector. That’s why I use self signed certs on my LAN stuff, I just don’t care about that yellow padlock that disappears when I trust the website. I’ve only experienced a single app ever that didn’t accept self-signed (I’m looking at you wallabag app).

        I can understand how it would be different if family members suddenly starts asking if it’s true when their devices tell them the webside is potentially dangerous.

        Yes, it’s dangerous to surveillance capitalism

        People… watching money?

        Forget it, I’ll set it up with a regular cert and external access

    • root@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      I have heard of this, but I think if you self-host a CA, you have to add the cert to every device that wants access to the service right? For example, I’d have to add it to my TV if my TV connects to Jellyfin, to my laptop if my laptop needs access to Home Assistant, etc. I’m not sure my family would like that XD

  • ehleks@lemmy.javant.xyz
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    1 year ago

    You can use duckdns.org to create a subdomain, set the domain name to your local IP and then use let’s encrypt DNS challenge to issue a trusted certificate