Let me first say, that it does work for Linux guest. It doesn’t work on Windows guest because there is a know bug (/issue) with the default hardware layout – made of i440FX BIOS. VirtManager would not allow us to replace the settings, so we need to create the VM ourselves using XML. You can export your XML settings (of an existing VM) using the command
virsh dumpxml > /tmp/VM_NAME.xml
There are relevant fields there which you might want to save for later, like MAC addresses, network settings, and so on.
You can use this XML file to build your VM anew. Note that you will want to modify the network settings, the name and the UUID. Also – you will need a newer QEMU command (through the package qemu-system-x86), you can find in the Centos updates repository, . It has been providing me with /usr/bin/qemu-system-x86_64 command, which I am using, instead of the default qemu command used by default by VirtManager.
My Windows VM XML file (as a reference you can copy and use) is provided below. Major modifications are required to the hardware settings of the Windows VM – moving from PCI to PCIE, changing from IDE to SATA or VirtIO – and the provided XML gives a good reference of how this file should look like. This was taken from a machine tested to allow USB hot-add/remove via the method provided in my previous post.
It is an amazing news to me. I really love XenServer. I think that Citrix were able to make a good use of Linux mechanisms for the purposes of virtualization, without abusing the OS layer (like some of the other virtualization solutions did). The file locations are decent (for example – most parts are located…
The simple scripts in /etc/xen/scripts which manage networking are fine for most usages, however, when your server is using bonding together with VLAN tagging (802.11q) you should consider an alternative. A PDF document written by Mark Nielsen, GPS Senior Consultant, Red Hat, Inc (I lost the original link, sorry) named “BOND/VLAN/Xen Network Configuration” as a…
I am using autofs (or automout) to handle shares presented on an NFS server I am using. When I am not home, or when the server is inaccessible, I do not want to be delayed by autofs failure to mount. This failure could take a long while, causing Autofs delay mount. During that time –…
Assuming your storage is capable of replication, a bunch of VMs could be happily replicated to an alternate location, where you can start them on will (and on crisis, most likely). This procedure, in theory, is rather simple. I have discovered that it is less so, especially if your system goes into testing once a…
To burn Dual-Layer (or Double-Layer, of you stick to the official name for DVD+R) medias, I use growisofs. The syntax is as follow: growisofs -dvd-compat -use-the-force-luke=break:1913760 -Z /dev/scd0=file.iso Change the break blocks to match your own values, and replace the file.iso with the actual name of your ISO file. If you do not set layer…