this post was submitted on 05 Apr 2024
28 points (96.7% liked)

Selfhosted

40353 readers
409 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 1 year ago
MODERATORS
 

Hello everyone,

I would like to get started with selfhost with two projects.

Project A (for me): A NUC with Proxmox installed on it, two VMs including a Home Assistant and a NAS system that I haven't chosen yet.

The only question I have with this project is:

  • how to access the NAS and HA separately from the outside knowing that my access provider does not offer a static IP and that access to each VM must be differentiated from Proxmox.

~~Project B (for my uncle):~~ ~~A NUC (with Proxmox or not, I don't know yet, perhaps simpler for making backups), with HA but especially Frigate.~~ ~~The goal is to use Google Coral to do recognition on 3 video surveillance cameras.~~

~~My questions are:~~

  • ~~is Coral really useful with 3 cameras?~~
  • ~~do you need a Coral in USB or M.2 version?~~
  • ~~are there affordable NUCs with free M.2 slots?~~
  • ~~won't proxmox add a layer of complexity with Coral/Frigate/a Zigbee dongle?~~

Thank you in advance for your help and sorry if my post is long.

~~PS: if you have recommendations for cameras that work with Frigate and are self-powered with solar panels, I'll take them!~~

Edit : 8 april 2024

A little update. Thank you everyone for your super quick responses!

Regarding my uncle's project and after big discussions, he is going to buy Reolink cameras and that's it. This will be much simpler for maintenance than building a server.

Regarding my project: I chose a Beelink Mini S12 pro with an N100 processor (for its low consumption) with a 2.5 bay for an SSD for my Nextcloud.

I wondered if I wouldn't take the opportunity to add pihole and that's where new questions arise...

I see a lot of people installing Pihole on Docker, should I put it on Docker? Or create a VM?

Should Docker be installed on Proxmox or on a VM?

Is Proxmox really useful, shouldn't I better install HA/Nextcloud/Pihole under Docker directly?

Should I use LXC or Docker?

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 7 months ago (1 children)

I share bind-mounts currently between multiple LXC from the host Proxmox OS, configuration is pretty easy, and there are lots of tutorials online for getting started.

Now then:

Are you sharing SMB mounts? I have my HDDs passed through to OMV and have considered just trying to pass them through to other VMs, but never tried because I don't wanna break anything.

I have seen that you can share SMB to Proxmox and use them in Proxmox but don't know if you can use them in VMs too.

As it is I really struggled with mounting smb for a couple of weeks and then had an "aha" moment last weekend, and have it all figured out now.

The Tailnet idea was so I can just mount everything to the Tailnet and stop worrying about whether it's on this vlan or that. I was trying to set up an Openwrt container with VPN, which I could use for any container that needs a vpn, but then those containers couldn't see the main network properly...

I've given up on that now and have my SMB mounts all set up, but feel like pass-through would give better network speeds for moving things around.

[–] dai 3 points 7 months ago (1 children)

Yeah there is a workaround for using bind-mounts in Proxmox VMs: https://gist.github.com/Drallas/7e4a6f6f36610eeb0bbb5d011c8ca0be

If you wanted, and your drives are mounted to the Proxmox host (and not to a VM), try an LXC for the services you are running, if you require a VM then the above workaround would be recommended after backing up your data.

I've got my drives mounted in a container as shown here:

Basicboi config, but it's quick and gets the job done.

I'd originally gone down the same route as you had with VMs and shares, but it's was all too much after a while.

I'm almost rid of all my VMs, home assistant is currently the last package I've yet to migrate. Migrated my frigate to a docker container under nixos, tailscale exit node under nixos too while the vast majority of other packages are already in LXC.

[–] [email protected] 3 points 7 months ago (1 children)

This all sounds awesome. So eli5 I have all my drives mounted to Proxmox, then passed through to OMV in a VM.

I can just mount these same drives to containers no issues right now, and I can add them to VMs using your link?

I would like to get down to LXCs too, but I've found VMs so much easier to set up and use. I'll try your way

[–] dai 2 points 7 months ago* (last edited 7 months ago)

I've not tested the method linked but yeah it would seem like it's possible via this method.

My lone VM doesn't need a connection to those drives so I've not had a point to.

You could probably run OMV in an LXC and skip the overheads of a VM entirely. LXC are containers, you can just edit the config files for the containers on the host Proxmox and pass drives right through.

Your containers will need to be privileged, you can also clone a container and make it privileged if you have something setup already as unprivileged!