Skip to content
Snippets Groups Projects
  1. Feb 14, 2023
    • John Garbutt's avatar
      Improve RabbitMQ performance by reducing ha replicas · 6cf22b0c
      John Garbutt authored
      Currently we do not follow the RabbitMQ advice on replicas here:
      https://www.rabbitmq.com/ha.html#replication-factor
      
      Here we reduce the number of replicas to n // 2 + 1 as advised
      above. The hope it this helps speed up recovery from rabbit
      issues.
      
      Related-Bug: #1954925
      Change-Id: Ib6bcb26c499c9884faa4a0cd51abaec00cacb096
      6cf22b0c
    • Matt Crees's avatar
      Add flag to change RabbitMQ ha-mode definition · e13072a9
      Matt Crees authored
      Adds the flag `rabbitmq_ha_replica_count` to change how many different
      nodes a queue should be mirrored across. If the value is not set, then
      it defaults to "ha-mode":"all". This value is unset by default to avoid
      any unexpected changes to the RabbitMQ definitions.json file, as that
      would trigger an unexpected restart of RabbitMQ during the next deploy.
      
      Change-Id: Iee98cd937197a73a3b04aa8501fa325e8ecfff24
      e13072a9
  2. Feb 09, 2023
    • John Garbutt's avatar
      RabbitMQ: Support setting ha-promote-on-shutdown · 94f3ce0c
      John Garbutt authored
      By default ha-promote-on-shutdown=when-synced. However we are seeing
      issues with RabbitMQ automatically recovering when nodes are restarted.
      https://www.rabbitmq.com/ha.html#cluster-shutdown
      
      Rather than waiting for operator interventions, it is better we allow
      recovery to happen, even if that means we may loose some messages.
      A few failed and timed out operations is better than a totaly broken
      cloud. This is achieved using ha-promote-on-shutdown=always.
      
      Note, when a node failure is detected, this is already the default
      behaviour from 3.7.5 onwards:
      https://www.rabbitmq.com/ha.html#promoting-unsynchronised-mirrors
      
      This patch adds the option to change the ha-promote-on-shutdown
      definition, using the flag `rabbitmq_ha_promote_on_shutdown`. This
      value is unset by default to avoid any unexpected changes to the
      RabbitMQ definitions.json file, as that would trigger an unexpected
      restart of RabbitMQ during the next deploy.
      
      Related-Bug: #1954925
      
      Change-Id: I2146bda2c72ddac2c9923c6941b0596395fd9ab5
      94f3ce0c
  3. Feb 03, 2023
  4. Jan 29, 2023
  5. Jan 26, 2023
  6. Jan 23, 2023
    • Alex-Welsh's avatar
      Adding optional delay between l3 agent restarts · 391aa467
      Alex-Welsh authored
      This change serialises the neutron l3 agent restart process and adds a
      user configurable delay between restarts. This can prevent connectivity
      loss due to all agents being restarted at the same time.
      
      Routers increase the recovery time, making this issue more prevalent.
      
      Change-Id: I3be0ebfa12965e6ae32d1b5f13f8fd23c3f52b8c
      391aa467
  7. Jan 20, 2023
    • Stanislav Dmitriev's avatar
      Set scheduler.max_attempts for nova conductor · 0b62db7c
      Stanislav Dmitriev authored
      In order to honour configured max number of attempts
      it has to be presented in nova.conf inside of
      nova_conductor container, otherwise the default value
      of 3 will be used
      
      Closes-Bug: #2003587
      Change-Id: I928af332b8658223444594f96417830233057284
      0b62db7c
  8. Jan 19, 2023
  9. Jan 17, 2023
  10. Jan 16, 2023
  11. Jan 13, 2023
    • Matt Crees's avatar
      Add a flag to handle RabbitMQ high availability · 09df6fc1
      Matt Crees authored
      A combination of durable queues and classic queue mirroring can be used
      to provide high availability of RabbitMQ. However, these options should
      only be used together, otherwise the system will become unstable. Using
      the flag ``om_enable_rabbitmq_high_availability`` will either enable
      both options at once, or neither of them.
      
      There are some queues that should not be mirrored:
      * ``reply`` queues (these have a single consumer and TTL policy)
      * ``fanout`` queues (these have a TTL policy)
      * ``amq`` queues (these are auto-delete queues, with a single consumer)
      An exclusionary pattern is used in the classic mirroring policy. This
      pattern is ``^(?!(amq\\.)|(.*_fanout_)|(reply_)).*``
      
      Change-Id: I51c8023b260eb40b2eaa91bd276b46890c215c25
      09df6fc1
  12. Jan 12, 2023
    • Mark Goddard's avatar
      Fix prechecks in check mode · 46aeb984
      Mark Goddard authored
      When running in check mode, some prechecks previously failed because
      they use the command module which is silently not run in check mode.
      Other prechecks were not running correctly in check mode due to e.g.
      looking for a string in empty command output or not querying which
      containers are running.
      
      This change fixes these issues.
      
      Closes-Bug: #2002657
      Change-Id: I5219cb42c48d5444943a2d48106dc338aa08fa7c
      46aeb984
  13. Jan 11, 2023
  14. Jan 10, 2023
  15. Jan 05, 2023
  16. Dec 22, 2022
    • labedz's avatar
      ovn: add ovn-monitor-all variable · 20355edb
      labedz authored
      Setting ovn-monitor-all to 'true' will configure
      ovn-controller to monitor all OVS database records
      unconditionally. That will release some CPU resource
      from OVS Southbound DB but will increase number of events
      coming to ovn-controller.
      
      Default value is 'false'.
      
      Change-Id: I291e166013d8c88f00e84ceaf308251c352c9a79
      20355edb
    • Michal Nasiadka's avatar
      ovn: Change order of deployment · 3a94996b
      Michal Nasiadka authored
      ovn-controller should be deployed first according to OVN upgrade guide.
      Since we are getting newer OVN/OVS versions from RDO/Ubuntu in a cycle,
      let's apply that to deployment.
      
      Closes-Bug: #1979329
      
      Change-Id: I017aec611a057db1634cfc2634164b21cb210193
      3a94996b
  17. Dec 21, 2022
    • Matt Crees's avatar
      Integrate oslo-config-validator · 6c2aace8
      Matt Crees authored
      Regularly, we experience issues in Kolla Ansible deployments because we
      use wrong options in OpenStack configuration files. This is because
      OpenStack services ignore unknown options. We also need to keep on top
      of deprecated options that may be removed in the future. Integrating
      oslo-config-validator into Kolla Ansible will greatly help.
      
      Adds a shared role to run oslo-config-validator on each service. Takes
      into account that services have multiple containers, and these may also
      use multiple config files. Service roles are extended to use this shared
      role. Executed with the new command ``kolla-ansible validate-config``.
      
      Change-Id: Ic10b410fc115646d96d2ce39d9618e7c46cb3fbc
      6c2aace8
  18. Dec 20, 2022
  19. Dec 19, 2022
  20. Dec 13, 2022
  21. Dec 08, 2022
  22. Dec 01, 2022
  23. Nov 30, 2022
  24. Nov 28, 2022
  25. Nov 23, 2022
  26. Nov 18, 2022
  27. Nov 17, 2022
  28. Nov 15, 2022
  29. Nov 11, 2022
    • Doug Szumski's avatar
      Remove support for deploying OpenStack Monasca · adb8f89a
      Doug Szumski authored
      Kolla Ansible is switching to OpenSearch and is dropping support for
      deploying ElasticSearch. This is because the final OSS release of
      ElasticSearch has exceeded its end of life.
      
      Monasca is affected because it uses both Logstash and ElasticSearch.
      Whilst it may continue to work with OpenSearch, Logstash remains an
      issue.
      
      In the absence of any renewed interest in the project, we remove
      support for deploying it. This helps to reduce the complexity
      of log processing configuration in Kolla Ansible, freeing up
      development time.
      
      Change-Id: I6fc7842bcda18e417a3fd21c11e28979a470f1cf
      adb8f89a
Loading