From c730e973b7bf6e0a88bb7bd1bdf0840f6ef816b3 Mon Sep 17 00:00:00 2001 From: Fiona Ebner Date: Tue, 20 Dec 2022 09:35:58 +0100 Subject: [PATCH] tree-wide: switch to offical spelling of QEMU Signed-off-by: Fiona Ebner --- images/pve-software-stack.svg | 2 +- images/pve-software-stack.tt | 2 +- pct.adoc | 2 +- pve-admin-guide.adoc | 2 +- pve-faq.adoc | 2 +- pvesm.adoc | 4 +-- qm.adoc | 58 +++++++++++++++++------------------ qmeventd.adoc | 4 +-- vzdump.adoc | 6 ++-- 9 files changed, 41 insertions(+), 41 deletions(-) diff --git a/images/pve-software-stack.svg b/images/pve-software-stack.svg index 1ffd4f0..6f50e31 100644 --- a/images/pve-software-stack.svg +++ b/images/pve-software-stack.svg @@ -60,7 +60,7 @@ Guest OS -Qemu +QEMU diff --git a/images/pve-software-stack.tt b/images/pve-software-stack.tt index e07bf7c..019942f 100644 --- a/images/pve-software-stack.tt +++ b/images/pve-software-stack.tt @@ -59,7 +59,7 @@ graph pve_software_stack { - Qemu + QEMU diff --git a/pct.adoc b/pct.adoc index 90a6d04..fdeb6dd 100644 --- a/pct.adoc +++ b/pct.adoc @@ -59,7 +59,7 @@ VM, but without the additional overhead. This means that Proxmox Containers can be categorized as ``System Containers'', rather than ``Application Containers''. NOTE: If you want to run application containers, for example, 'Docker' images, it -is recommended that you run them inside a Proxmox Qemu VM. This will give you +is recommended that you run them inside a Proxmox QEMU VM. This will give you all the advantages of application containerization, while also providing the benefits that VMs offer, such as strong isolation from the host and the ability to live-migrate, which otherwise isn't possible with containers. diff --git a/pve-admin-guide.adoc b/pve-admin-guide.adoc index d0a1637..eaa2609 100644 --- a/pve-admin-guide.adoc +++ b/pve-admin-guide.adoc @@ -149,7 +149,7 @@ include::pvesh.1-synopsis.adoc[] :leveloffset: 0 -*qm* - Qemu/KVM Virtual Machine Manager +*qm* - QEMU/KVM Virtual Machine Manager ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :leveloffset: 1 diff --git a/pve-faq.adoc b/pve-faq.adoc index 8e203f4..5a41be7 100644 --- a/pve-faq.adoc +++ b/pve-faq.adoc @@ -161,4 +161,4 @@ Docker Engine command line interface. It is not recommended to run docker directly on your {pve} host. + NOTE: If you want to run application containers, for example, 'Docker' images, it -is best to run them inside a Proxmox Qemu VM. +is best to run them inside a Proxmox QEMU VM. diff --git a/pvesm.adoc b/pvesm.adoc index 76992dd..6ade1a4 100644 --- a/pvesm.adoc +++ b/pvesm.adoc @@ -93,7 +93,7 @@ That way you get a `shared` LVM storage. Thin Provisioning ~~~~~~~~~~~~~~~~~ -A number of storages, and the Qemu image format `qcow2`, support 'thin +A number of storages, and the QEMU image format `qcow2`, support 'thin provisioning'. With thin provisioning activated, only the blocks that the guest system actually use will be written to the storage. @@ -195,7 +195,7 @@ this property to select what this storage is used for. images::: -KVM-Qemu VM images. +QEMU/KVM VM images. rootdir::: diff --git a/qm.adoc b/qm.adoc index f923ccf..644ece5 100644 --- a/qm.adoc +++ b/qm.adoc @@ -7,7 +7,7 @@ qm(1) NAME ---- -qm - Qemu/KVM Virtual Machine Manager +qm - QEMU/KVM Virtual Machine Manager SYNOPSIS @@ -19,7 +19,7 @@ DESCRIPTION ----------- endif::manvolnum[] ifndef::manvolnum[] -Qemu/KVM Virtual Machines +QEMU/KVM Virtual Machines ========================= :pve-toplevel: endif::manvolnum[] @@ -29,52 +29,52 @@ endif::manvolnum[] // http://pve.proxmox.com/wiki/KVM // http://pve.proxmox.com/wiki/Qemu_Server -Qemu (short form for Quick Emulator) is an open source hypervisor that emulates a -physical computer. From the perspective of the host system where Qemu is -running, Qemu is a user program which has access to a number of local resources +QEMU (short form for Quick Emulator) is an open source hypervisor that emulates a +physical computer. From the perspective of the host system where QEMU is +running, QEMU is a user program which has access to a number of local resources like partitions, files, network cards which are then passed to an emulated computer which sees them as if they were real devices. A guest operating system running in the emulated computer accesses these devices, and runs as if it were running on real hardware. For instance, you can pass -an ISO image as a parameter to Qemu, and the OS running in the emulated computer +an ISO image as a parameter to QEMU, and the OS running in the emulated computer will see a real CD-ROM inserted into a CD drive. -Qemu can emulate a great variety of hardware from ARM to Sparc, but {pve} is +QEMU can emulate a great variety of hardware from ARM to Sparc, but {pve} is only concerned with 32 and 64 bits PC clone emulation, since it represents the overwhelming majority of server hardware. The emulation of PC clones is also one of the fastest due to the availability of processor extensions which greatly -speed up Qemu when the emulated architecture is the same as the host +speed up QEMU when the emulated architecture is the same as the host architecture. NOTE: You may sometimes encounter the term _KVM_ (Kernel-based Virtual Machine). -It means that Qemu is running with the support of the virtualization processor -extensions, via the Linux KVM module. In the context of {pve} _Qemu_ and -_KVM_ can be used interchangeably, as Qemu in {pve} will always try to load the KVM +It means that QEMU is running with the support of the virtualization processor +extensions, via the Linux KVM module. In the context of {pve} _QEMU_ and +_KVM_ can be used interchangeably, as QEMU in {pve} will always try to load the KVM module. -Qemu inside {pve} runs as a root process, since this is required to access block +QEMU inside {pve} runs as a root process, since this is required to access block and PCI devices. Emulated devices and paravirtualized devices -------------------------------------------- -The PC hardware emulated by Qemu includes a mainboard, network controllers, +The PC hardware emulated by QEMU includes a mainboard, network controllers, SCSI, IDE and SATA controllers, serial ports (the complete list can be seen in the `kvm(1)` man page) all of them emulated in software. All these devices are the exact software equivalent of existing hardware devices, and if the OS running in the guest has the proper drivers it will use the devices as if it -were running on real hardware. This allows Qemu to runs _unmodified_ operating +were running on real hardware. This allows QEMU to runs _unmodified_ operating systems. This however has a performance cost, as running in software what was meant to run in hardware involves a lot of extra work for the host CPU. To mitigate this, -Qemu can present to the guest operating system _paravirtualized devices_, where -the guest OS recognizes it is running inside Qemu and cooperates with the +QEMU can present to the guest operating system _paravirtualized devices_, where +the guest OS recognizes it is running inside QEMU and cooperates with the hypervisor. -Qemu relies on the virtio virtualization standard, and is thus able to present +QEMU relies on the virtio virtualization standard, and is thus able to present paravirtualized virtio devices, which includes a paravirtualized generic disk controller, a paravirtualized network card, a paravirtualized serial port, a paravirtualized SCSI controller, etc ... @@ -131,7 +131,7 @@ can specify which xref:qm_display[display type] you want to use. [thumbnail="screenshot/gui-create-vm-system.png"] Additionally, the xref:qm_hard_disk[SCSI controller] can be changed. If you plan to install the QEMU Guest Agent, or if your selected ISO image -already ships and installs it automatically, you may want to tick the 'Qemu +already ships and installs it automatically, you may want to tick the 'QEMU Agent' box, which lets {pve} know that it can use its features to show some more information, and complete some actions (for example, shutdown or snapshots) more intelligently. @@ -153,7 +153,7 @@ Hard Disk [[qm_hard_disk_bus]] Bus/Controller ^^^^^^^^^^^^^^ -Qemu can emulate a number of storage controllers: +QEMU can emulate a number of storage controllers: * the *IDE* controller, has a design which goes back to the 1984 PC/AT disk controller. Even if this controller has been superseded by recent designs, @@ -177,7 +177,7 @@ containing the drivers during the installation. // https://pve.proxmox.com/wiki/Paravirtualized_Block_Drivers_for_Windows#During_windows_installation. If you aim at maximum performance, you can select a SCSI controller of type _VirtIO SCSI single_ which will allow you to select the *IO Thread* option. -When selecting _VirtIO SCSI single_ Qemu will create a new controller for +When selecting _VirtIO SCSI single_ QEMU will create a new controller for each disk, instead of adding all disks to the same controller. * The *VirtIO Block* controller, often just called VirtIO or virtio-blk, @@ -252,7 +252,7 @@ IO Thread The option *IO Thread* can only be used when using a disk with the *VirtIO* controller, or with the *SCSI* controller, when the emulated controller type is *VirtIO SCSI single*. -With this enabled, Qemu creates one I/O thread per storage controller, +With this enabled, QEMU creates one I/O thread per storage controller, rather than a single thread for all I/O. This can increase performance when multiple disks are used and each disk has its own storage controller. @@ -274,7 +274,7 @@ allows you. Increasing the number of virtual CPUs (cores and sockets) will usually provide a performance improvement though that is heavily dependent on the use of the VM. Multi-threaded applications will of course benefit from a large number of -virtual CPUs, as for each virtual cpu you add, Qemu will create a new thread of +virtual CPUs, as for each virtual cpu you add, QEMU will create a new thread of execution on the host system. If you're not sure about the workload of your VM, it is usually a safe bet to set the number of *Total cores* to 2. @@ -299,7 +299,7 @@ the whole VM can use on the host. It is a floating point value representing CPU time in percent, so `1.0` is equal to `100%`, `2.5` to `250%` and so on. If a single process would fully use one single core it would have `100%` CPU Time usage. If a VM with four cores utilizes all its cores fully it would -theoretically use `400%`. In reality the usage may be even a bit higher as Qemu +theoretically use `400%`. In reality the usage may be even a bit higher as QEMU can have additional threads for VM peripherals besides the vCPU core ones. This setting can be useful if a VM should have multiple vCPUs, as it runs a few processes in parallel, but the VM as a whole should not be able to run all @@ -347,7 +347,7 @@ section for more examples. CPU Type ^^^^^^^^ -Qemu can emulate a number different of *CPU types* from 486 to the latest Xeon +QEMU can emulate a number different of *CPU types* from 486 to the latest Xeon processors. Each new processor generation adds new features, like hardware assisted 3d rendering, random number generation, memory protection, etc ... Usually you should select for your VM a processor type which closely matches the @@ -359,7 +359,7 @@ as your host system. This has a downside though. If you want to do a live migration of VMs between different hosts, your VM might end up on a new system with a different CPU type. If the CPU flags passed to the guest are missing, the qemu process will stop. To -remedy this Qemu has also its own CPU type *kvm64*, that {pve} uses by defaults. +remedy this QEMU has also its own CPU type *kvm64*, that {pve} uses by defaults. kvm64 is a Pentium 4 look a like CPU type, which has a reduced CPU flags set, but is guaranteed to work everywhere. @@ -627,7 +627,7 @@ _tap device_, ( a software loopback device simulating an Ethernet NIC ). This tap device is added to a bridge, by default vmbr0 in {pve}. In this mode, VMs have direct access to the Ethernet LAN on which the host is located. * in the alternative *NAT mode*, each virtual NIC will only communicate with -the Qemu user networking stack, where a built-in router and DHCP server can +the QEMU user networking stack, where a built-in router and DHCP server can provide network access. This built-in DHCP will serve addresses in the private 10.0.2.0/24 range. The NAT mode is much slower than the bridged mode, and should only be used for testing. This mode is only available via CLI or the API, @@ -1016,10 +1016,10 @@ see the section on xref:first_guest_boot_delay[Proxmox VE Node Management]. [[qm_qemu_agent]] -Qemu Guest Agent +QEMU Guest Agent ~~~~~~~~~~~~~~~~ -The Qemu Guest Agent is a service which runs inside the VM, providing a +The QEMU Guest Agent is a service which runs inside the VM, providing a communication channel between the host and the guest. It is used to exchange information and allows the host to issue commands to the guest. @@ -1477,7 +1477,7 @@ chosen, the first of: Managing Virtual Machines with `qm` ------------------------------------ -qm is the tool to manage Qemu/Kvm virtual machines on {pve}. You can +qm is the tool to manage QEMU/KVM virtual machines on {pve}. You can create and destroy virtual machines, and control execution (start/stop/suspend/resume). Besides that, you can use qm to set parameters in the associated config file. It is also possible to diff --git a/qmeventd.adoc b/qmeventd.adoc index 4e89e7d..31e7035 100644 --- a/qmeventd.adoc +++ b/qmeventd.adoc @@ -6,7 +6,7 @@ qmeventd(8) NAME ---- -qmeventd - PVE Qemu Eventd Daemon +qmeventd - PVE QEMU Eventd Daemon SYNOPSIS -------- @@ -18,7 +18,7 @@ DESCRIPTION endif::manvolnum[] ifndef::manvolnum[] -PVE Qemu Event Daemon +PVE QEMU Event Daemon ===================== :pve-toplevel: endif::manvolnum[] diff --git a/vzdump.adoc b/vzdump.adoc index 8658e38..ce46529 100644 --- a/vzdump.adoc +++ b/vzdump.adoc @@ -64,7 +64,7 @@ depending on the guest type. This mode provides the highest consistency of the backup, at the cost of a short downtime in the VM operation. It works by executing an -orderly shutdown of the VM, and then runs a background Qemu process to +orderly shutdown of the VM, and then runs a background QEMU process to backup the VM data. After the backup is started, the VM goes to full operation mode if it was previously running. Consistency is guaranteed by using the live backup feature. @@ -92,8 +92,8 @@ https://git.proxmox.com/?p=pve-qemu.git;a=blob_plain;f=backup.txt[here]. NOTE: {pve} live backup provides snapshot-like semantics on any storage type. It does not require that the underlying storage supports snapshots. Also please note that since the backups are done via -a background Qemu process, a stopped VM will appear as running for a -short amount of time while the VM disks are being read by Qemu. +a background QEMU process, a stopped VM will appear as running for a +short amount of time while the VM disks are being read by QEMU. However the VM itself is not booted, only its disk(s) are read. .Backup modes for Containers: -- 2.39.2