Skip to content
Snippets Groups Projects
  1. Nov 26, 2024
  2. Aug 16, 2024
    • Michal Arbet's avatar
      Fix nova-libvirt's secrets - device or resource is busy · 82a642ed
      Michal Arbet authored
      This patch removes the nova_libvirt_secret container volume because
      it is a complete antipattern, and during testing, I found that
      it causes problems. When it was necessary to copy libvirt secrets
      from /etc/kolla/nova-libvirt/secrets, the container logs reported that
      the resource is busy - precisely because it was a mounted container
      volume. This, of course, is unnecessary because the secrets are copied
      to the kolla host in /etc/kolla/nova-libvirt/secrets.
      
      Closes-Bug: #2073678
      Change-Id: I715a6a95f9d32d62a8199727ddbaddd0dd7baa2d
      82a642ed
  3. Aug 12, 2024
    • Ivan Halomi's avatar
      Refactor of kolla_container_facts · 4ce47e22
      Ivan Halomi authored
      
      Refactor that prepares kolla_container_facts
      module for introducing more actions that will be moved
      from kolla_container module and kolla_container_volume_facts.
      
      This change is based on a discussion about adding a new action
      to kolla_container module that retrieves all names of the running
      containers. It was agreed that kolla-ansible should follow Ansible's
      direction of splitting modules between action modules and facts
      modules. Because of this, kolla_container_facts needs to be able
      to handle different requests for data about containers or volumes.
      
      Change-Id: Ieaec8f64922e4e5a2199db2d6983518b124cb4aa
      Signed-off-by: default avatarIvan Halomi <ivan.halomi@tietoevry.com>
      4ce47e22
  4. Jul 22, 2024
  5. Jun 28, 2024
    • Roman Krček's avatar
      Performance: use filters for service dicts · fb3a8f5f
      Roman Krček authored
      Most roles are not leveraging the jinja filters available.
      According to [1] filtering the list of services makes the execution
      faster than skipping the tasks.
      
      This patchset also includes some cosmetic changes to genconfig.
      Individual services are now also using a jinja filter. This has
      no impact on performance, just makes the tasks look cleaner.
      
      Naming of some vars in genconfig was changed to "service" to make
      the tasks more uniform as some were previously using
      the service name and some were using "service".
      
      Three metrics from the deployment were taken and those were
      - overall deployment time [s]
      - time spent on the specific role [s]
      - CPU usage (measured with perf) [-]
      Overall genconfig time went down on avg. from 209s to 195s
      Time spent on the loadbalancer role went down on avg. from 27s to 23s
      Time spent on the neutron role went down on avg from 102s to 95s
      Time spent on the nova-cell role went down on avg. from 54s to 52s
      Also the average CPUs utilized reported by perf went down
      from 3.31 to 3.15.
      For details of how this was measured see the comments in gerrit.
      
      [1] - https://github.com/stackhpc/ansible-scaling/blob/master/doc/skip.md
      
      
      
      Change-Id: Ib0f00aadb6c7022de6e8b455ac4b9b8cd6be5b1b
      Signed-off-by: default avatarRoman Krček <roman.krcek@tietoevry.com>
      fb3a8f5f
  6. Apr 22, 2024
  7. Apr 08, 2024
  8. Jan 30, 2024
  9. Nov 15, 2023
  10. Nov 14, 2023
    • Michal Nasiadka's avatar
      Introduce oneshot docker_restart_policy · cea076f3
      Michal Nasiadka authored
      docker_restart_policy: no causes systemd units to not get created
      and we use it in CI to disable restarts on services.
      
      Introducing oneshot policy to not create systemd unit for oneshot
      containers (those that are running bootstrap tasks, like db
      bootstrap and don't need a systemd unit), but still create systemd
      units for long lived containers but with Restart=No.
      
      Change-Id: I9e0d656f19143ec2fcad7d6d345b2c9387551604
      cea076f3
  11. Nov 09, 2023
  12. Oct 20, 2023
  13. Oct 04, 2023
    • Michal Nasiadka's avatar
      Generate hostnqn · 8b3e755b
      Michal Nasiadka authored
      hostnqn is generated using to_uuid filter
      
      Usually "nvme gen-hostnqn" command is used to generate
      hostnqn, and it has the format of:
      nqn.2014-08.org.nvmexpress:uuid:67dc8c8e-0262-4d81-ac51-ace7c25e4daa
      
      "nqn.2014-08.org.nvmexpress:uuid:" is always static
      
      Closes-Bug: #2035975
      
      Change-Id: I6ece4fe8c18c0167a2707c24693fbe39ed15cdba
      8b3e755b
  14. Sep 08, 2023
  15. Jun 28, 2023
  16. Jun 14, 2023
    • Michal Arbet's avatar
      Add support for multiple ceph files · fdf2385f
      Michal Arbet authored
      This patch is adding a feature for an option to copy different
      ceph configuration files and corresponding keyrings for cinder,
      glance, manila, gnocchi and nova services.
      
      This is especially useful when the deployment uses availability
      zones as below example.
      
        - Individual compute can read/write to individual ceph
          cluster in same AZ.
        - Cinder can write to several ceph clusters in several AZs.
        - Glance can use multistore and upload images to
          several ceph clusters in several AZs at once.
      
      Change-Id: Ie4d8ab5a3df748137835cae1c943b9180cd10eb1
      fdf2385f
  17. Feb 14, 2023
    • Mark Goddard's avatar
      Fix deploy/genconfig in check mode · 572ff2f8
      Mark Goddard authored
      Previously, when running one of the following commands:
      
        kolla-ansible deploy --check
        kolla-ansible genconfig --check
      
      deployment or configuration generation fails for various reasons.
      
      MariaDB fails to lookup the existing cluster.
      
      Keystone fails to generate cron config.
      
      Nova-cell fails to get the cell settings.
      
      Closes-Bug: #2002661
      Change-Id: I5e765f498ae86d213d0a4379ca5d473db1499962
      572ff2f8
  18. Jan 26, 2023
  19. Jan 12, 2023
    • Mark Goddard's avatar
      Fix prechecks in check mode · 46aeb984
      Mark Goddard authored
      When running in check mode, some prechecks previously failed because
      they use the command module which is silently not run in check mode.
      Other prechecks were not running correctly in check mode due to e.g.
      looking for a string in empty command output or not querying which
      containers are running.
      
      This change fixes these issues.
      
      Closes-Bug: #2002657
      Change-Id: I5219cb42c48d5444943a2d48106dc338aa08fa7c
      46aeb984
  20. Dec 21, 2022
    • Matt Crees's avatar
      Integrate oslo-config-validator · 6c2aace8
      Matt Crees authored
      Regularly, we experience issues in Kolla Ansible deployments because we
      use wrong options in OpenStack configuration files. This is because
      OpenStack services ignore unknown options. We also need to keep on top
      of deprecated options that may be removed in the future. Integrating
      oslo-config-validator into Kolla Ansible will greatly help.
      
      Adds a shared role to run oslo-config-validator on each service. Takes
      into account that services have multiple containers, and these may also
      use multiple config files. Service roles are extended to use this shared
      role. Executed with the new command ``kolla-ansible validate-config``.
      
      Change-Id: Ic10b410fc115646d96d2ce39d9618e7c46cb3fbc
      6c2aace8
  21. Nov 04, 2022
  22. Nov 02, 2022
  23. Oct 28, 2022
  24. Oct 07, 2022
  25. Sep 26, 2022
  26. Sep 21, 2022
  27. Aug 09, 2022
  28. Jul 25, 2022
    • Michal Nasiadka's avatar
      Fix var-spacing · dcf5a8b6
      Michal Nasiadka authored
      ansible-lint introduced var-spacing - let's fix our code.
      
      Change-Id: I0d8aaf3c522a5a6a5495032f6dbed8a2be0251f0
      dcf5a8b6
  29. Apr 22, 2022
    • Mark Goddard's avatar
      nova: use any_errors_fatal for once-per-cell tasks · 832989d0
      Mark Goddard authored
      We run some nova tasks once per cell, using a condition to match a
      single host in the cell. In other similar tasks, we use run_once, which
      will fail all hosts if the task fails. Typically these tasks are
      critical, and that is desirable. However, with the approach used in
      nova-cell to support multiple cells, if a once-per-cell task fails, then
      other hosts will continue to execute, which could lead to unexpected
      results.
      
      This change adds any_errors_fatal to the plays or blocks that run these
      tasks.
      
      Closes-Bug: #1948694
      
      Change-Id: I2a5871ccd4e8198171ef3239ce95f475f3e4b051
      832989d0
  30. Apr 05, 2022
    • Mark Goddard's avatar
      libvirt: Fix nova-libvirt-cleanup command · 188b3285
      Mark Goddard authored
      This change addresses an issue in the nova-libvirt-cleanup command,
      added in I46854ed7eaf1d5b5e3ccd8531c963427848bdc99.
      
      Check for rc=1 pgrep command, since a lack of matches is a pass.
      
      Also, use bash for set -o pipefail.
      
      Change-Id: Iffda0dfffce8768324ffec55e629134c70e2e996
      188b3285
  31. Mar 29, 2022
    • Mark Goddard's avatar
      nova: improve compute service registration failure handling · f1d3ff11
      Mark Goddard authored
      If any nova compute service fails to register itself, Kolla Ansible will
      fail the host that queries the Nova API. This is the first compute host
      in the inventory, and fails in the task:
      
          Waiting for nova-compute services to register themselves
      
      Other hosts continue, often leading to further errors later on. Clearly
      this is not idea.
      
      This change modifies the behaviour to query the compute service list
      until all expected hosts are present, but does not fail the querying
      host if they are not. A new task is added that executes for all hosts,
      and fails only those hosts that have not registered successfully.
      
      Alternatively, to fail all hosts in a cell when any compute service
      fails to register, set nova_compute_registration_fatal to true.
      
      Change-Id: I12c1928cf1f1fb9e28f1741e7fe4968004ea1816
      Closes-Bug: #1940119
      f1d3ff11
  32. Mar 21, 2022
    • Mark Goddard's avatar
      libvirt: add nova-libvirt-cleanup command · 80b311be
      Mark Goddard authored
      Change Ia1239069ccee39416b20959cbabad962c56693cf added support for
      running a libvirt daemon on the host, rather than using the nova_libvirt
      container. It did not cover migration of existing hosts from using a
      container to using a host daemon.
      
      This change adds a kolla-ansible nova-libvirt-cleanup command which may
      be used to clean up the nova_libvirt container, volumes and related
      items on hosts, once it has been disabled.
      
      The playbook assumes that compute hosts have been emptied of VMs before
      it runs. A future extension could support migration of existing VMs, but
      this is currently out of scope.
      
      Change-Id: I46854ed7eaf1d5b5e3ccd8531c963427848bdc99
      80b311be
    • Mark Goddard's avatar
      libvirt: make it possible to run libvirt on the host · 4e41acd8
      Mark Goddard authored
      In some cases it may be desirable to run the libvirt daemon on the host.
      For example, when mixing host and container OS distributions or
      versions.
      
      This change makes it possible to disable the nova_libvirt container, by
      setting enable_nova_libvirt_container to false. The default values of
      some Docker mounts and other paths have been updated to point to default
      host directories rather than Docker volumes when using a host libvirt
      daemon.
      
      This change does not handle migration of existing systems from using
      a nova_libvirt container to libvirt on the host.
      
      Depends-On: https://review.opendev.org/c/openstack/ansible-collection-kolla/+/830504
      
      Change-Id: Ia1239069ccee39416b20959cbabad962c56693cf
      4e41acd8
  33. Mar 18, 2022
  34. Mar 10, 2022
    • Mark Goddard's avatar
      libvirt: support SASL authentication · d2d4b53d
      Mark Goddard authored
      In Kolla Ansible OpenStack deployments, by default, libvirt is
      configured to allow read-write access via an unauthenticated,
      unencrypted TCP connection, using the internal API network.  This is to
      facilitate migration between hosts.
      
      By default, Kolla Ansible does not use encryption for services on the
      internal network (and did not support it until Ussuri). However, most
      other services on the internal network are at least authenticated
      (usually via passwords), ensuring that they cannot be used by anyone
      with access to the network, unless they have credentials.
      
      The main issue here is the lack of authentication. Any client with
      access to the internal network is able to connect to the libvirt TCP
      port and make arbitrary changes to the hypervisor. This could include
      starting a VM, modifying an existing VM, etc. Given the flexibility of
      the domain options, it could be seen as equivalent to having root access
      to the hypervisor.
      
      Kolla Ansible supports libvirt TLS [1] since the Train release, using
      client and server certificates for mutual authentication and encryption.
      However, this feature is not enabled by default, and requires
      certificates to be generated for each compute host.
      
      This change adds support for libvirt SASL authentication, and enables it
      by default. This provides base level of security. Deployments requiring
      further security should use libvirt TLS.
      
      [1] https://docs.openstack.org/kolla-ansible/latest/reference/compute/libvirt-guide.html#libvirt-tls
      
      Depends-On: https://review.opendev.org/c/openstack/kolla/+/833021
      Closes-Bug: #1964013
      Change-Id: Ia91ceeb609e4cdb144433122b443028c0278b71e
      d2d4b53d
  35. Jan 10, 2022
  36. Oct 27, 2021
  37. Oct 01, 2021
  38. Sep 28, 2021
    • Niklas Hagman's avatar
      Transition Keystone admin user to system scope · 2e933dce
      Niklas Hagman authored
      A system-scoped token implies the user has authorization to act on the
      deployment system. These tokens are useful for interacting with
      resources that affect the deployment as a whole, or exposes resources
      that may otherwise violate project or domain isolation.
      
      Since Queens, the keystone-manage bootstrap command assigns the admin
      role to the admin user with system scope, as well as in the admin
      project. This patch transitions the Keystone admin user from
      authenticating using project scoped tokens to system scoped tokens.
      This is a necessary step towards being able to enable the updated oslo
      policies in services that allow finer grained access to system-level
      resources and APIs.
      
      An etherpad with discussion about the transition to the new oslo
      service policies is:
      
      https://etherpad.opendev.org/p/enabling-system-scope-in-kolla-ansible
      
      
      
      Change-Id: Ib631e2211682862296cce9ea179f2661c90fa585
      Signed-off-by: default avatarNiklas Hagman <ubuntu@post.blinkiz.com>
      2e933dce
  39. Aug 12, 2021
    • Michal Arbet's avatar
      Trivial fix nova's healthchecks · 85879afc
      Michal Arbet authored
      Kolla-ansible upgrade task is calling different
      handlers as deploy task and these handlers are
      missing healthcheck key. This patch is fixing
      this.
      
      Closes-Bug: #1939679
      Change-Id: Id83d20bfd89c27ccf70a3a79938f428cdb5d40fc
      85879afc
Loading