[pve-devel] applied: [PATCH pve-docs] fix typos/wording

Wolfgang Bumiller w.bumiller at proxmox.com
Thu Aug 23 14:35:17 CEST 2018


applied

On Thu, Aug 23, 2018 at 01:48:43PM +0200, David Limbeck wrote:
> Signed-off-by: David Limbeck <d.limbeck at proxmox.com>
> ---
>  ha-manager.adoc | 28 ++++++++++++++--------------
>  pveum.adoc      |  2 +-
>  qm.adoc         |  8 ++++----
>  vzdump.adoc     |  2 +-
>  4 files changed, 20 insertions(+), 20 deletions(-)
> 
> diff --git a/ha-manager.adoc b/ha-manager.adoc
> index 3edbc50..4d38583 100644
> --- a/ha-manager.adoc
> +++ b/ha-manager.adoc
> @@ -137,7 +137,7 @@ resource of type `vm` (virtual machine) with the ID 100.
>  
>  For now we have two important resources types - virtual machines and
>  containers. One basic idea here is that we can bundle related software
> -into such VM or container, so there is no need to compose one big
> +into such a VM or container, so there is no need to compose one big
>  service from other services, like it was done with `rgmanager`. In
>  general, a HA managed resource should not depend on other resources.
>  
> @@ -156,7 +156,7 @@ GUI, or simply use the command line tool, for example:
>  
>  The HA stack now tries to start the resources and keeps it
>  running. Please note that you can configure the ``requested''
> -resources state. For example you may want that the HA stack stops the
> +resources state. For example you may want the HA stack to stop the
>  resource:
>  
>  ----
> @@ -225,7 +225,7 @@ the following command:
>  
>  NOTE: This does not start or stop the resource.
>  
> -But all HA related task can be done on the GUI, so there is no need to
> +But all HA related tasks can be done in the GUI, so there is no need to
>  use the command line at all.
>  
>  
> @@ -253,7 +253,7 @@ handles node fencing.
>  .Locks in the LRM & CRM
>  [NOTE]
>  Locks are provided by our distributed configuration file system (pmxcfs).
> -They are used to guarantee that each LRM is active once and working. As a
> +They are used to guarantee that each LRM is active once and working. As an
>  LRM only executes actions when it holds its lock, we can mark a failed node
>  as fenced if we can acquire its lock. This lets us then recover any failed
>  HA services securely without any interference from the now unknown failed node.
> @@ -369,7 +369,7 @@ The LRM lost its lock, this means a failure happened and quorum was lost.
>  After the LRM gets in the active state it reads the manager status
>  file in `/etc/pve/ha/manager_status` and determines the commands it
>  has to execute for the services it owns.
> -For each command a worker gets started, this workers are running in
> +For each command a worker gets started, these workers are running in
>  parallel and are limited to at most 4 by default. This default setting
>  may be changed through the datacenter configuration key `max_worker`.
>  When finished the worker process gets collected and its result saved for
> @@ -381,19 +381,19 @@ The default value of at most 4 concurrent workers may be unsuited for
>  a specific setup. For example may 4 live migrations happen at the same
>  time, which can lead to network congestions with slower networks and/or
>  big (memory wise) services. Ensure that also in the worst case no congestion
> -happens and lower the `max_worker` value if needed. In the contrary, if you
> +happens and lower the `max_worker` value if needed. On the contrary, if you
>  have a particularly powerful high end setup you may also want to increase it.
>  
> -Each command requested by the CRM is uniquely identifiable by an UID, when
> -the worker finished its result will be processed and written in the LRM
> +Each command requested by the CRM is uniquely identifiable by a UID, when
> +the worker finishes its result will be processed and written in the LRM
>  status file `/etc/pve/nodes/<nodename>/lrm_status`. There the CRM may collect
>  it and let its state machine - respective the commands output - act on it.
>  
>  The actions on each service between CRM and LRM are normally always synced.
> -This means that the CRM requests a state uniquely marked by an UID, the LRM
> +This means that the CRM requests a state uniquely marked by a UID, the LRM
>  then executes this action *one time* and writes back the result, also
>  identifiable by the same UID. This is needed so that the LRM does not
> -executes an outdated command.
> +execute an outdated command.
>  With the exception of the `stop` and the `error` command,
>  those two do not depend on the result produced and are executed
>  always in the case of the stopped state and once in the case of
> @@ -430,11 +430,11 @@ lost agent lock::
>  
>  The CRM lost its lock, this means a failure happened and quorum was lost.
>  
> -It main task is to manage the services which are configured to be highly
> +Its main task is to manage the services which are configured to be highly
>  available and try to always enforce the requested state. For example, a
>  service with the requested state 'started' will be started if its not
>  already running. If it crashes it will be automatically started again.
> -Thus the CRM dictates the actions which the LRM needs to execute.
> +Thus the CRM dictates the actions the LRM needs to execute.
>  
>  When an node leaves the cluster quorum, its state changes to unknown.
>  If the current CRM then can secure the failed nodes lock, the services
> @@ -468,7 +468,7 @@ Resources
>  
>  The resource configuration file `/etc/pve/ha/resources.cfg` stores
>  the list of resources managed by `ha-manager`. A resource configuration
> -inside that list look like this:
> +inside that list looks like this:
>  
>  ----
>  <type>: <name>
> @@ -689,7 +689,7 @@ Start Failure Policy
>  ---------------------
>  
>  The start failure policy comes in effect if a service failed to start on a
> -node once ore more times. It can be used to configure how often a restart
> +node one or more times. It can be used to configure how often a restart
>  should be triggered on the same node and how often a service should be
>  relocated so that it gets a try to be started on another node.
>  The aim of this policy is to circumvent temporary unavailability of shared
> diff --git a/pveum.adoc b/pveum.adoc
> index c3eb870..b0bb72a 100644
> --- a/pveum.adoc
> +++ b/pveum.adoc
> @@ -223,7 +223,7 @@ of predefined roles which satisfies most needs.
>  
>  You can see the whole set of predefined roles on the GUI.
>  
> -Adding new roles can currently only be done from the command line, like
> +Adding new roles can be done via both GUI and the command line, like
>  this:
>  
>  [source,bash]
> diff --git a/qm.adoc b/qm.adoc
> index 28d2a38..1451f5d 100644
> --- a/qm.adoc
> +++ b/qm.adoc
> @@ -442,7 +442,7 @@ minimum amount you specified is always available to the VM, and if RAM usage on
>  the host is below 80%, will dynamically add memory to the guest up to the
>  maximum memory specified.
>  
> -When the host is becoming short on RAM, the VM will then release some memory
> +When the host is running low on RAM, the VM will then release some memory
>  back to the host, swapping running processes if needed and starting the oom
>  killer in last resort. The passing around of memory between host and guest is
>  done via a special `balloon` kernel driver running inside the guest, which will
> @@ -452,14 +452,14 @@ footnote:[A good explanation of the inner workings of the balloon driver can be
>  When multiple VMs use the autoallocate facility, it is possible to set a
>  *Shares* coefficient which indicates the relative amount of the free host memory
>  that each VM should take. Suppose for instance you have four VMs, three of them
> -running a HTTP server and the last one is a database server. To cache more
> +running an HTTP server and the last one is a database server. To cache more
>  database blocks in the database server RAM, you would like to prioritize the
>  database VM when spare RAM is available. For this you assign a Shares property
>  of 3000 to the database VM, leaving the other VMs to the Shares default setting
>  of 1000. The host server has 32GB of RAM, and is currently using 16GB, leaving 32
>  * 80/100 - 16 = 9GB RAM to be allocated to the VMs. The database VM will get 9 *
>  3000 / (3000 + 1000 + 1000 + 1000) = 4.5 GB extra RAM and each HTTP server will
> -get 1/5 GB.
> +get 1.5 GB.
>  
>  All Linux distributions released after 2010 have the balloon kernel driver
>  included. For Windows OSes, the balloon driver needs to be added manually and can
> @@ -516,7 +516,7 @@ of packets transferred.
>  
>  //http://blog.vmsplice.net/2011/09/qemu-internals-vhost-architecture.html
>  When using the VirtIO driver with {pve}, each NIC network queue is passed to the
> -host kernel, where the queue will be processed by a kernel thread spawn by the
> +host kernel, where the queue will be processed by a kernel thread spawned by the
>  vhost driver. With this option activated, it is possible to pass _multiple_
>  network queues to the host kernel for each NIC.
>  
> diff --git a/vzdump.adoc b/vzdump.adoc
> index 193e1cf..fb1ac3d 100644
> --- a/vzdump.adoc
> +++ b/vzdump.adoc
> @@ -25,7 +25,7 @@ Backup and Restore
>  :pve-toplevel:
>  endif::manvolnum[]
>  
> -Backups are a requirements for any sensible IT deployment, and {pve}
> +Backups are a requirement for any sensible IT deployment, and {pve}
>  provides a fully integrated solution, using the capabilities of each
>  storage and each guest system type. This allows the system
>  administrator to fine tune via the `mode` option between consistency
> -- 
> 2.11.0




More information about the pve-devel mailing list