I have been successfully using consumer SSD in my ESXi for few years. After migration to Proxmox, I have tried to use ZFS for virtual machines. First few weeks I was very happy about it. Then I noticed very fast increasing SSD wearout levels on my new SSDs. Proxmox support forums and reddit are full of users mentioning proxmox is destroying SSD, but I thought this is just rumor. Where is the true?
Five years ago before Christmas I bought my first SSD – Crucial MX300 750GB. I have put it into my linux server running some web hosting (running this blog for example), my emails. Second MX300 was put into Linux MD Raid1. Disk were running two years in this setup, then I have migrated into virtualized setup, running ESXi, booted from USB drive. SSD were re-used and running as hw Raid 1 using LSI-9211 card.
When installing Windows 2003 to Dell M600 blade system windows will not recognize any hard drives. You need to load additional drivers, by pressing F6 during setup. Drivers can be found at Dell FTP site, but Dell is not providing floppy image, just drivers.
LVM is a logical volume manager for the Linux kernel; it manages disk drives and similar mass-storage devices, in particular large ones. The term “volume” refers to a disk drive or partition thereof.
Every LVM operation makes backup of lvm metadata. Metadata archive stored in directory /etc/lvm/archive/ in easily readable text files. Each operation creates own metadata backup file, with information when it was created. For example:
With shrinking IT budgets you will at some point be tempted fill your RAID controllers with cheap desktop hard drives. You may also be tempted to replace your servers system drives with them as well. I’m going to give you some reasons why you should consider buying those pricey enterprise class drives instead.