this post was submitted on 07 Dec 2023
17 points (100.0% liked)
technology
23218 readers
2 users here now
On the road to fully automated luxury gay space communism.
Spreading Linux propaganda since 2020
- Ways to run Microsoft/Adobe and more on Linux
- The Ultimate FOSS Guide For Android
- Great libre software on Windows
- Hey you, the lib still using Chrome. Read this post!
Rules:
- 1. Obviously abide by the sitewide code of conduct. Bigotry will be met with an immediate ban
- 2. This community is about technology. Offtopic is permitted as long as it is kept in the comment sections
- 3. Although this is not /c/libre, FOSS related posting is tolerated, and even welcome in the case of effort posts
- 4. We believe technology should be liberating. As such, avoid promoting proprietary and/or bourgeois technology
- 5. Explanatory posts to correct the potential mistakes a comrade made in a post of their own are allowed, as long as they remain respectful
- 6. No crypto (Bitcoin, NFT, etc.) speculation, unless it is purely informative and not too cringe
- 7. Absolutely no tech bro shit. If you have a good opinion of Silicon Valley billionaires please manifest yourself so we can ban you.
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Hi! Hopefully I can make some things make more sense.
Ah well I thought you were using TrueNAS Core which is based on FreeBSD rather than Linux. TrueNAS Scale is Linux though so nevermind!
These are just some lower-level tools that Proxmox is using to run VMs. qemu manages the virtual machines and cloud-init is a standardized configuration system for setting up VMs when they boot up. When you look at a VM's settings in proxmox, under the hood some of them are qemu settings and cloud-init settings. Number of cores is a qemu thing. Virtualized Ethernet card settings are a cloud-init thing. I'm only mentioning this because if proxmox things aren't making sense you might want to play around with these tools more directly until they make sense. Proxmox doesn't really make them easier to understand, just easier to discover and set in one interface.
SMB could work just fine! I just default to NFS when all I want is to share a directory between Linux systems. I just mean "share" when I say volume or dataset in terms of these two tools.
It'll be way easier to with with the lxc containers and VMs if they have static IPs, so definitely prioritize this! There are two perfectly valid ways to do this:
I use different software than dd-wrt (I use used low power enterprise stuff) so I don't know the exact way to do these in its settings, but it should be able to do either.
I really just mean subnet. CIDR is a way of describing IP ranges, doesn't really matter except I think proxmox might want you to use CIDR notation sometimes when doing some network configuration?
Basically your lxc containers / VMs have network settings somewhere on the proxmox web interface. Maybe under hardware? If you click to configure one of them, it'll pop up a dialog that will let you manually specify a static IP, gateway, and netmask. The static IP is an IP on your LAN. The gateway is the address of your router (like 192.168.0.1). The netmask is almost always 255.255.255.0. You might have to use CIDR notation for the static IP - I forget whether it's at this step or when installing proxmox itself. You'll know because it will raise a little input error warning that says your static IP setting is invalid. Slap a /24 on the end and it'll go away. Like 192.168.0.10/24.
It depends on how you installed and configured Jellyfin. You may be able to use the admin dashboard built into jellyfin.
Jellyfin stores a lot of settings and data in an SQLite database, which is just a file in the jellyfin config directory. You'll be able to avoid an entire category of ways that database can get corrupted if you don't put it on a network share.
iSCSI is a bit more challenging to use than smb. It's very good at what it does but if you can make smb work then I'd stick with that. But if you ever get an itch to try something different, including providing other kinds of shared storage for VMs, look into iSCSI.