Skip Navigation
InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)GA
gansheim @lemmy.world
Posts 1
Comments 9
Has anyone ever cloned a disk with the GPT partition scheme?
  • That's fair, I should have clarified that on most Enterprise Linux distros LVM is definitely the norm. I know Fedora switched to btrfs a few releases back and you may be right about Suse Tumbleweed but pretty sure Suse Leap uses LVM. CentOS, RHEL, Alma, etc. all still default to LVM, as the idea of keeping everything on a single partition is a bad idea and managing multiple partitions is significantly easier with LVM. More than likely that'll change when btrfs has a little more mileage on it and is trusted as "enterprise ready" but for now LVM is the way they go. MBR vs GPT and EFI vs non-EFI don't have a lot to do with it though, it's more about the ease of managing multiple partitions (or subvolumes if you're used to btrfs), as having a single partition for root, var, and home is bad idea jeans.

  • Has anyone ever cloned a disk with the GPT partition scheme?
  • That's fair, I should have clarified that on most Enterprise Linux distros LVM is definitely the norm. I know Fedora switched to btrfs a few releases back and you may be right about Suse Tumbleweed but pretty sure Suse Leap uses LVM. CentOS, RHEL, Alma, etc. all still default to LVM, as the idea of keeping everything on a single partition is a bad idea and managing multiple partitions is significantly easier with LVM. More than likely that'll change when btrfs has a little more mileage on it and is trusted as "enterprise ready" but for now LVM is the way they go. MBR vs GPT and EFI vs non-EFI don't have a lot to do with it though, it's more about the ease of managing multiple partitions (or subvolumes if you're used to btrfs), as having a single partition for root, var, and home is bad idea jeans.

  • Has anyone ever cloned a disk with the GPT partition scheme?
  • LVM is actually super common. Most Linux distros default to LVM unless you do custom partitioning. It's not just about the max number of partitions supported by the table. LVM provides a TON more flexibility and ease of management of partitions.

  • why does this community have no posts 🥹
  • At least in my experience, this isn't true for all pets. I ever only had one dog that ate all the food in her bowl no matter what. All the dogs and cats I've had besides that one, including the 2 dogs and cat I have now will eat until they're full, regardless if the bowl is empty, and then come back to it later.

  • Your lips don't touch when you say the word touch, but do touch when you say separate

    My 8 year old son told me this randomly after getting out of the shower this morning.

    33
    Help with Proxmox
  • Sounds like you're conflating terms a bit. The example storage volumes you gave sounds like ProxMox volumes, but then you switched to VM volumes in your example.

    The hypervisor (ProxMox) has storage volumes that are used for multiple things; these are sometimes called Datastores. One will be for the local ProxMox OS, another might be for sending logs to, and lastly you'll have one for VM storage. The VM Storage volume(s) are where your virtual machine files live, to include any virtual disks you provision.

    The storage type for the volume you create in ProxMox will dictate what you can and can't do with VM disks in that storage. They provide a list here: https://pve.proxmox.com/wiki/Storage

    Pay attention to things like if you can take a live VM snapshots and such.

    Getting back to your original issue, it sounds like you tried to create multiple VM disks using the same bus channel (i.e. SCSI0 or IDE0). You have to specify a different channel for each disk. First disk is channel 0, next is 1, etc.

    Assuming you configure a datastore for each storage class (slow spinny, fast ssd, super fast flash), you would then configure a virtual disk for the VM backed by each respective datastore depending on the storage class you want. Let's assume you're using you're the SCSI storage driver, disk 0 is your boot drive, and you want that on datastore0 which super fast NVMe, disk 1 is your general OS drive and you want that on datastore1 which is fast ssd, and disk 2 is your data drive and you want that on slow hdd. You'd configure your disks similar to the following: Disk0 SCSI0 Datastore0

    Disk1 SCSI1 Datastore1

    Disk2 SCSI2 Datastore2

    Lastly, if you're wondering what storage driver to use, it comes down to performance and compatibility of your VM. IDE is slowest but has good compatibility with pretty much everything, although I've heard rumors that newer systems may be dropping support for IDE

    SCSI is usually a good happy medium but has compatibility issues with older legacy systems

    VirtIO is generally supposed be fastest, but has more compatibility issues with a wider range of "older" systems

    Compatibility issues for each can/may be resolved with the right drivers, but then you might need to custom build your boot image to support them, especially in Windows.

    There's an older post in the forum that goes over the differences that should still be relevant here: https://forum.proxmox.com/threads/differences-between-ide-virtio-scsi.8356/

    I've only been using ProxMox for a couple years in my homelab, and mostly use VMware so I may have crossed the terms a bit, but hopefully this gives you the gist. I hope I understood your question/issue and that this helps.