Skip to content
Snippets Groups Projects
  1. Nov 04, 2024
  2. Oct 31, 2024
    • Matt Crees's avatar
      Prevent accidental downgrades of RabbitMQ · e6727bd4
      Matt Crees authored
      As version-check.yml is added to deploy.yml, we must make sure the
      tasks are only run when the rabbitmq container exists.
      
      Change-Id: Iaa31bae739110094affb5e402ed9ac40b153ac3d
      (cherry picked from commit f5ad7829)
      e6727bd4
  3. Aug 29, 2024
    • Matt Crees's avatar
      Add a precheck to catch RMQ SLURP upgrades · 4731e7f2
      Matt Crees authored
      Given we bump the RabbitMQ version each release, there is a manual
      upgrade to an intermediary RabbitMQ version needed before a skip-level
      upgrade can be performed.
      
      Change-Id: Id8a5ebe19a50ebdc59d12667889472c803b8d7c8
      4731e7f2
  4. Aug 12, 2024
    • Ivan Halomi's avatar
      Refactor of kolla_container_facts · 4ce47e22
      Ivan Halomi authored
      
      Refactor that prepares kolla_container_facts
      module for introducing more actions that will be moved
      from kolla_container module and kolla_container_volume_facts.
      
      This change is based on a discussion about adding a new action
      to kolla_container module that retrieves all names of the running
      containers. It was agreed that kolla-ansible should follow Ansible's
      direction of splitting modules between action modules and facts
      modules. Because of this, kolla_container_facts needs to be able
      to handle different requests for data about containers or volumes.
      
      Change-Id: Ieaec8f64922e4e5a2199db2d6983518b124cb4aa
      Signed-off-by: default avatarIvan Halomi <ivan.halomi@tietoevry.com>
      4ce47e22
  5. Jun 28, 2024
    • Roman Krček's avatar
      Performance: use filters for service dicts · fb3a8f5f
      Roman Krček authored
      Most roles are not leveraging the jinja filters available.
      According to [1] filtering the list of services makes the execution
      faster than skipping the tasks.
      
      This patchset also includes some cosmetic changes to genconfig.
      Individual services are now also using a jinja filter. This has
      no impact on performance, just makes the tasks look cleaner.
      
      Naming of some vars in genconfig was changed to "service" to make
      the tasks more uniform as some were previously using
      the service name and some were using "service".
      
      Three metrics from the deployment were taken and those were
      - overall deployment time [s]
      - time spent on the specific role [s]
      - CPU usage (measured with perf) [-]
      Overall genconfig time went down on avg. from 209s to 195s
      Time spent on the loadbalancer role went down on avg. from 27s to 23s
      Time spent on the neutron role went down on avg from 102s to 95s
      Time spent on the nova-cell role went down on avg. from 54s to 52s
      Also the average CPUs utilized reported by perf went down
      from 3.31 to 3.15.
      For details of how this was measured see the comments in gerrit.
      
      [1] - https://github.com/stackhpc/ansible-scaling/blob/master/doc/skip.md
      
      
      
      Change-Id: Ib0f00aadb6c7022de6e8b455ac4b9b8cd6be5b1b
      Signed-off-by: default avatarRoman Krček <roman.krcek@tietoevry.com>
      fb3a8f5f
  6. May 13, 2024
  7. Apr 25, 2024
  8. Apr 22, 2024
  9. Mar 12, 2024
  10. Dec 19, 2023
  11. Dec 13, 2023
  12. Nov 15, 2023
  13. Nov 14, 2023
    • Michal Nasiadka's avatar
      Introduce oneshot docker_restart_policy · cea076f3
      Michal Nasiadka authored
      docker_restart_policy: no causes systemd units to not get created
      and we use it in CI to disable restarts on services.
      
      Introducing oneshot policy to not create systemd unit for oneshot
      containers (those that are running bootstrap tasks, like db
      bootstrap and don't need a systemd unit), but still create systemd
      units for long lived containers but with Restart=No.
      
      Change-Id: I9e0d656f19143ec2fcad7d6d345b2c9387551604
      cea076f3
  14. Aug 25, 2023
  15. Jun 19, 2023
  16. Jun 17, 2023
  17. Apr 20, 2023
  18. Apr 19, 2023
    • Matt Crees's avatar
      Add precheck to fail if RabbitMQ HA needs configuring · a5331d32
      Matt Crees authored
      Currently, the process of enabling RabbitMQ HA with the variable
      ``om_enable_rabbitmq_high_availbility`` requires some manual steps to
      migrate from transient to mirrored queues. In preparation for setting
      this variable to ``True`` by default, this adds a precheck that will
      fail if a system is currently running non-mirrored queues and
      ``om_enable_rabbitmq_high_availbility`` is set to ``True``.
      
      Includes a helpful message informing the operator of their choice.
      Either follow the manual procedure to migrate the queues described in
      the docs, or set ``om_enable_rabbitmq_high_availbility`` to ``False``.
      
      The RabbitMQ HA section of the reference docs is updated to include
      these instructions.
      
      Change-Id: Ic5e64998bd01923162204f7bb289cc110187feec
      a5331d32
  19. Apr 13, 2023
    • Matt Crees's avatar
      Remove RabbitMQ ha-all policy when not required · c85b64d1
      Matt Crees authored
      With the addition of the variable
      `om_enable_rabbitmq_high_availability`, this feature in the upgrade
      task should be brought back. It is also now used in the deploy task. The
      `ha-all` policy is cleared only when
      `om_enable_rabbitmq_high_availability` is set to `false`.
      
      Change-Id: Ia056aa40e996b1f0fed43c0f672466c7e4a2f547
      c85b64d1
  20. Apr 12, 2023
  21. Jan 12, 2023
    • Mark Goddard's avatar
      Fix prechecks in check mode · 46aeb984
      Mark Goddard authored
      When running in check mode, some prechecks previously failed because
      they use the command module which is silently not run in check mode.
      Other prechecks were not running correctly in check mode due to e.g.
      looking for a string in empty command output or not querying which
      containers are running.
      
      This change fixes these issues.
      
      Closes-Bug: #2002657
      Change-Id: I5219cb42c48d5444943a2d48106dc338aa08fa7c
      46aeb984
  22. Jan 09, 2023
  23. Dec 21, 2022
    • Matt Crees's avatar
      Integrate oslo-config-validator · 6c2aace8
      Matt Crees authored
      Regularly, we experience issues in Kolla Ansible deployments because we
      use wrong options in OpenStack configuration files. This is because
      OpenStack services ignore unknown options. We also need to keep on top
      of deprecated options that may be removed in the future. Integrating
      oslo-config-validator into Kolla Ansible will greatly help.
      
      Adds a shared role to run oslo-config-validator on each service. Takes
      into account that services have multiple containers, and these may also
      use multiple config files. Service roles are extended to use this shared
      role. Executed with the new command ``kolla-ansible validate-config``.
      
      Change-Id: Ic10b410fc115646d96d2ce39d9618e7c46cb3fbc
      6c2aace8
  24. Nov 02, 2022
  25. Oct 28, 2022
  26. Aug 09, 2022
  27. Jul 27, 2022
  28. Jul 25, 2022
    • Michal Nasiadka's avatar
      Fix var-spacing · dcf5a8b6
      Michal Nasiadka authored
      ansible-lint introduced var-spacing - let's fix our code.
      
      Change-Id: I0d8aaf3c522a5a6a5495032f6dbed8a2be0251f0
      dcf5a8b6
  29. Mar 24, 2022
  30. Mar 18, 2022
  31. Feb 21, 2022
    • Doug Szumski's avatar
      Remove classic queue mirroring for internal RabbitMQ · 6bfe1927
      Doug Szumski authored
      When OpenStack is deployed with Kolla-Ansible, by default there
      are no durable queues or exchanges created by the OpenStack
      services in RabbitMQ. In Rabbit terminology, not being durable
      is referred to as `transient`, and this means that the queue
      is generally held in memory.
      
      Whether OpenStack services create durable or transient queues is
      traditionally controlled by the Oslo Notification config option:
      `amqp_durable_queues`. In Kolla-Ansible, this remains set to
      the default of `False` in all services. The only `durable`
      objects are the `amq*` exchanges which are internal to RabbitMQ.
      
      More recently, Oslo Notification has introduced support for
      Quorum queues [7]. These are a successor to durable classic
      queues, however it isn't yet clear if they are a good fit for
      OpenStack in general [8].
      
      For clustered RabbitMQ deployments, Kolla-Ansible configures all
      queues as `replicated` [1]. Replication occurs over all nodes
      in the cluster. RabbitMQ refers to this as 'mirroring of classic
      queues'.
      
      In summary, this means that a multi-node Kolla-Ansible deployment
      will end up with a large number of transient, mirrored queues
      and exchanges. However, the RabbitMQ documentation warns against
      this, stating that 'For replicated queues, the only reasonable
      option is to use durable queues: [2]`. This is discussed
      further in the following bug report: [3].
      
      Whilst we could try enabling the `amqp_durable_queues` option
      for each service (this is suggested in [4]), there are
      a number of complexities with this approach, not limited to:
      
      1) RabbitMQ is planning to remove classic queue mirroring in
         favor of 'Quorum queues' in a forthcoming release [5].
      2) Durable queues will be written to disk, which may cause
         performance problems at scale. Note that this includes
         Quorum queues which are always durable.
      3) Potential for race conditions and other complexity
         discussed recently on the mailing list under:
         `[ops] [kolla] RabbitMQ High Availability`
      
      The remaining option, proposed here, is to use classic
      non-mirrored queues everywhere, and rely on services to recover
      if the node hosting a queue or exchange they are using fails.
      There is some discussion of this approach in [6]. The downside
      of potential message loss needs to be weighed against the real
      upsides of increasing the performance of RabbitMQ, and moving
      to a configuration which is officially supported and hopefully
      more stable. In the future, we can then consider promoting
      specific queues to quorum queues, in cases where message loss
      can result in failure states which are hard to recover from.
      
      [1] https://www.rabbitmq.com/ha.html
      [2] https://www.rabbitmq.com/queues.html
      [3] https://github.com/rabbitmq/rabbitmq-server/issues/2045
      [4] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit
      [5] https://blog.rabbitmq.com/posts/2021/08/4.0-deprecation-announcements/
      [6] https://fuel-ccp.readthedocs.io/en/latest/design/ref_arch_1000_nodes.html#replication
      [7] https://bugs.launchpad.net/oslo.messaging/+bug/1942933
      [8] https://www.rabbitmq.com/quorum-queues.html#use-cases
      
      Partial-Bug: #1954925
      Change-Id: I91d0e23b22319cf3fdb7603f5401d24e3b76a56e
      6bfe1927
  32. Jan 09, 2022
    • LinPeiWen's avatar
      Support enable/disable rabbitmq prometheus plugins · 1f3dcce5
      LinPeiWen authored
      rabbitmq starting from 3.8.0, built-in Prometheus support,
      prometheus plugins are enabled by default, when the environment is
      "enable_prometheus is no", rabbitmq role will disable prometheus plugins
      
      Closes-Bug: #1885106
      
      Change-Id: I4d694d6224c813285d228d6bc7eece5731db1078
      1f3dcce5
  33. Aug 10, 2021
    • Radosław Piliszek's avatar
      Refactor and optimise image pulling · 9ff2ecb0
      Radosław Piliszek authored
      We get a nice optimisation by using a filtered loop instead
      of task skipping per service with 'when'.
      
      Partially-Implements: blueprint performance-improvements
      Change-Id: I8f68100870ab90cb2d6b68a66a4c97df9ea4ff52
      9ff2ecb0
  34. Jun 23, 2021
    • Mark Goddard's avatar
      Use ansible_facts to reference facts · ade5bfa3
      Mark Goddard authored
      By default, Ansible injects a variable for every fact, prefixed with
      ansible_. This can result in a large number of variables for each host,
      which at scale can incur a performance penalty. Ansible provides a
      configuration option [0] that can be set to False to prevent this
      injection of facts. In this case, facts should be referenced via
      ansible_facts.<fact>.
      
      This change updates all references to Ansible facts within Kolla Ansible
      from using individual fact variables to using the items in the
      ansible_facts dictionary. This allows users to disable fact variable
      injection in their Ansible configuration, which may provide some
      performance improvement.
      
      This change disables fact variable injection in the ansible
      configuration used in CI, to catch any attempts to use the injected
      variables.
      
      [0] https://docs.ansible.com/ansible/latest/reference_appendices/config.html#inject-facts-as-vars
      
      Change-Id: I7e9d5c9b8b9164d4aee3abb4e37c8f28d98ff5d1
      Partially-Implements: blueprint performance-improvements
      ade5bfa3
  35. Apr 14, 2021
  36. Dec 14, 2020
    • Mark Goddard's avatar
      Revert "Performance: Use import_tasks in the main plays" · db4fc85c
      Mark Goddard authored
      This reverts commit 9cae59be.
      
      Reason for revert: This patch was found to introduce issues with fluentd customisation. The underlying issue is not currently fully understood, but could be a sign of other obscure issues.
      
      Change-Id: Ia4859c23d85699621a3b734d6cedb70225576dfc
      Closes-Bug: #1906288
      db4fc85c
  37. Nov 19, 2020
  38. Oct 27, 2020
    • Radosław Piliszek's avatar
      Performance: Use import_tasks in the main plays · 9cae59be
      Radosław Piliszek authored
      Main plays are action-redirect-stubs, ideal for import_tasks.
      
      This avoids 'include' penalty and makes logs/ara look nicer.
      
      Fixes haproxy and rabbitmq not to check the host group as well.
      
      Change-Id: I46136fc40b815e341befff80b54a91ef431eabc0
      Partially-Implements: blueprint performance-improvements
      9cae59be
  39. Oct 12, 2020
    • Radosław Piliszek's avatar
      Performance: optimize genconfig · 3411b9e4
      Radosław Piliszek authored
      Config plays do not need to check containers. This avoids skipping
      tasks during the genconfig action.
      
      Ironic and Glance rolling upgrades are handled specially.
      
      Swift and Bifrost do not use the handlers at all.
      
      Partially-Implements: blueprint performance-improvements
      Change-Id: I140bf71d62e8f0932c96270d1f08940a5ba4542a
      3411b9e4
Loading