This project, https://neko.m1k1o.net/#/getting-started/examples , looks like a good base to try running regular GUI apps via docker & web.
edit: and here's the git with Dockerfiles, https://github.com/m1k1o/neko-apps
This project, https://neko.m1k1o.net/#/getting-started/examples , looks like a good base to try running regular GUI apps via docker & web.
edit: and here's the git with Dockerfiles, https://github.com/m1k1o/neko-apps
On proxmox you should be able to share any GPU (integrated or dedicated) to multiple LXCs while keeping it accessible to the host. I use intel integrated GPU in LXC for plex, jellyfin, and one with just ffmpeg I use to convert videos occasionally. I used these instructions as starting point/base when I set mine up on proxmox v7.x, https://forum.proxmox.com/threads/plex-hw-transcoding-lxc-and-jasper-lake-igpu-passthru.116163/
I had looked at instructions to assign the GPU to a specific VM but it looked like way too much work and people were saying it didn't always work for the 11th gen iGPUs. Thankfully I ran across the sharing method and it's been running stable since.
My info may be outdated as I last had G Fiber about a year ago but have moved out of their service area so stuck with AT&T fiber along with their horrible modem+router :(
When I first got the 2G down/1G up G Fiber service there was no bridge mode & had to use their provided device as modem+router+wifi. They updated it to add in a bridge mode option but I never tested it. I had dropped back down to 1G down & up before that option was available.
edit: forgot to mention I had read some people had luck using Unifi Dream Machine to plug in G Fiber's 2.5G SFP looking module but I wasn't willing to spend any more money on anything Unifi besides WiFi APs.
My last NAS & ESXi box were 12 years old when I retired them. I had thought about sticking with used enterprise gear but wanted a break to be a little lazy for a couple years. Storage is on Synology (DS1520+) and Proxmox runs on Asus PN63-S1 mini PC. Hyper Backup was primary reason I chose Synology (always been lazy about off-site backups) and docker feature has come in handy for things like secondary pihole & DNS. LXC with docker or podman have been able to cover majority of my needs in proxmox but still have Home Assistant & Unifi Network Controller on their own VMs. Home Assistant I have zero plans to move. Unifi I eventually plan to move over to docker but it works for now, albeit on an older version. Really need to up my documentation & diagram game, it's all a huge mess, lol.
Future plans would love to have closet full of used enterprise servers running proxmox with all flash ceph storage backend then can do whatever NAS distro I want as a VM. My budget is focused elsewhere for next year or two unfortunately so gonna be awhile unless something breaks.
Always like to hear about other setups as I am constantly re-thinking my own.
I have public wildcard DNS entry (*.REMOVEDDOMAIN.com) on Cloudflare on my primary domain that resolves to 192.168.10.120 (my Caddy host)
Caddyfile
{
email [email protected]
acme_dns cloudflare TOKENGOESHERE
}
portal.REMOVEDDOMAIN.com {
reverse_proxy 127.0.0.1:8081
}
speedtest.REMOVEDDOMAIN.com {
reverse_proxy 192.168.10.125:8181
}
You can self-host ACME server which lets you use certbot to do automatic renewals even for private, internal only certs. I don't know if it would work with NPM. I plan to test that out at some point in the future but my current setup works & I'm not ready to break it for a maybe yet :P
I use Caddy with the Cloudflare DNS plugin for Let's Encrypt DNS based challenges, should work for wildcard too but only have a couple subdomains so never tried to do that. My DNS entries are public but point at private IP ranges, e.g. nc.PRIVATEDOMAIN.COM resolves to 192.168.1.20 where Caddy sends the traffic to my Nextcloud docker
Yeah, Proxmox has a built in backup utility. I use it for nightly backup of all VMs and LXCs to cifs share on my NAS.
Free and centrally managed, not aware of any but definitely interested in something like that too.
My current setup has Proxmox backing up all LXC and VMs to Synology NAS then the Synology NAS backing up to Backblaze. Both run nightly. Using the built-in backup utility on Proxmox VE pointed at CIFS share on the Synology NAS.
Synology does have a software backup client available but I have never used it. For my desktops & laptops, they are easily reinstalled+reconfigured, I just make sure the data I care about is stored or synchronized to my NAS or the cloud. Nextcloud for files, Firefox sync for history+bookmarks, bitwarden client+vaultwarden for passwords, chezmoi for some dotfiles on some linux systems.
They changed quite a few things between DSM 6 & 7 and unfortunately one of them broke easy use of those USB sticks. I didn't want to mess with the internal config of the Synology NAS too much so used the VM approach with the HA image and mapped the USB stick to it.
I hope you like it :)
Another benefit to LXC is you can map devices, including GPU, to multiple LXC while keeping them accessible to the host. For my home setup I currently have 3 LXC with access to the iGPU, 1 for jellyfin+caddy via podman nested, 1 for moonfire-nvr via podman nested, and been trying to use 1 to figure out hardware transcoding with owncast through multiple install methods but no luck so far. I've also been playing with mapping rtl-sdr v3 devices, zigbee stick, zwave stick, and coral usb for a variety of projects lately.
edit: I forgot to answer the question and went straight to ranting, lol. LXC is like a bare-metal VM. You can install & run multiple things on them like a normal VM including podman or docker.