Skip to content
Snippets Groups Projects
  1. Aug 15, 2019
    • Rafael Weingärtner's avatar
      Standardize the configuration of "oslo_messaging" section · 22a6223b
      Rafael Weingärtner authored
      After all of the discussions we had on
      "https://review.opendev.org/#/c/670626/2", I studied all projects that
      have an "oslo_messaging" section. Afterwards, I applied the same method
      that is already used in "oslo_messaging" section in Nova, Cinder, and
      others. This guarantees that we have a consistent method to
      enable/disable notifications across projects based on components (e.g.
      Ceilometer) being enabled or disabled. Here follows the list of
      components, and the respective changes I did.
      
      * Aodh:
      The section is declared, but it is not used. Therefore, it will
      be removed in an upcomming PR.
      
      * Congress:
      The section is declared, but it is not used. Therefore, it will
      be removed in an upcomming PR.
      
      * Cinder:
      It was already properly configured.
      
      * Octavia:
      The section is declared, but it is not used. Therefore, it will
      be removed in an upcomming PR.
      
      * Heat:
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Ceilometer:
      Ceilometer publishes some messages in the rabbitMQ. However, the
      default driver is "messagingv2", and not ''(empty) as defined in Oslo;
      these configurations are defined in ceilometer/publisher/messaging.py.
      Therefore, we do not need to do anything for the
      "oslo_messaging_notifications" section in Ceilometer
      
      * Tacker:
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Neutron:
      It was already properly configured.
      
      * Nova
      It was already properly configured. However, we found another issue
      with its configuration. Kolla-ansible does not configure nova
      notifications as it should. If 'searchlight' is not installed (enabled)
      the 'notification_format' should be 'unversioned'. The default is
      'both'; so nova will send a notification to the queue
      versioned_notifications; but that queue has no consumer when
      'searchlight' is disabled. In our case, the queue got 511k messages.
      The huge amount of "stuck" messages made the Rabbitmq cluster
      unstable.
      
      https://bugzilla.redhat.com/show_bug.cgi?id=1478274
      https://bugs.launchpad.net/ceilometer/+bug/1665449
      
      * Nova_hyperv:
      I added the same configurations as in Nova project.
      
      * Vitrage
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Searchlight
      I created a mechanism similar to what we have in AODH, Cinder, Nova,
      and others.
      
      * Ironic
      I created a mechanism similar to what we have in AODH, Cinder, Nova,
      and others.
      
      * Glance
      It was already properly configured.
      
      * Trove
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Blazar
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Sahara
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Watcher
      I created a mechanism similar to what we have in AODH, Cinder, Nova,
      and others.
      
      * Barbican
      I created a mechanism similar to what we have in Cinder, Nova,
      and others. I also added a configuration to 'keystone_notifications'
      section. Barbican needs its own queue to capture events from Keystone.
      Otherwise, it has an impact on Ceilometer and other systems that are
      connected to the "notifications" default queue.
      
      * Keystone
      Keystone is the system that triggered this work with the discussions
      that followed on https://review.opendev.org/#/c/670626/2
      
      . After a long
      discussion, we agreed to apply the same approach that we have in Nova,
      Cinder and other systems in Keystone. That is what we did. Moreover, we
      introduce a new topic "barbican_notifications" when barbican is
      enabled. We also removed the "variable" enable_cadf_notifications, as
      it is obsolete, and the default in Keystone is CADF.
      
      * Mistral:
      It was hardcoded "noop" as the driver. However, that does not seem a
      good practice. Instead, I applied the same standard of using the driver
      and pushing to "notifications" queue if Ceilometer is enabled.
      
      * Cyborg:
      I created a mechanism similar to what we have in AODH, Cinder, Nova,
      and others.
      
      * Murano
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Senlin
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Manila
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Zun
      The section is declared, but it is not used. Therefore, it will
      be removed in an upcomming PR.
      
      * Designate
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      * Magnum
      It was already using a similar scheme; I just modified it a little bit
      to be the same as we have in all other components
      
      Closes-Bug: #1838985
      
      Change-Id: I88bdb004814f37c81c9a9c4e5e491fac69f6f202
      Signed-off-by: default avatarRafael Weingärtner <rafael@apache.org>
      22a6223b
  2. Jul 23, 2019
  3. Jul 22, 2019
  4. Jul 18, 2019
    • Jason's avatar
      Do not recreate Blazar DB if using preconfigured · 7d284761
      Jason authored
      Most other services already gate the DB bootstrap operations with the
      'use_preconfigured_databases' variable; Blazar did not.
      
      Change-Id: I772b1cb92612c7e6936f052ed9947f93582f264c
      Unverified
      7d284761
    • Mark Goddard's avatar
      Fix glance bootstrap with file backend · 1abd15d4
      Mark Goddard authored
      Change https://review.opendev.org/#/c/670247/ attempted to fix glance
      deployment with the file backend. However, it added a new bug by being
      more strict about only generating configuration where the container will
      be deployed. This means that the current method of running the glance
      bootstrap container on any host in glance-api group could be broken,
      since it needs the container configuration.
      
      This change only runs the bootstrap container on hosts in the
      glance_api_hosts list, which in the case of the file backend typically
      only contains one host.
      
      This change also fixes up some logic during rolling upgrade, where we
      might not generate new configuration for the bootstrap host.
      
      Change-Id: I83547cd83b06ddefb3a9e1f39844537bdb32bd7f
      Related-Bug: #1836151
      1abd15d4
    • Raimund Hook's avatar
      Updated multi-region docs to include keepalived · 99463849
      Raimund Hook authored
      The keepalived_virtual_router_id should be changed from the default in
      the case of a multi-region deployment where the VIP of the different
      regions resides on the same subnet.
      
      This is not immediately clear - this change should make it more obvious.
      
      Change-Id: Ia4899ba407937d9f27832c9d123701729e89987a
      99463849
  5. Jul 17, 2019
  6. Jul 16, 2019
  7. Jul 15, 2019
  8. Jul 14, 2019
  9. Jul 12, 2019
    • Mark Goddard's avatar
      Fix ironic inspector iPXE boot with UEFI · 7b939756
      Mark Goddard authored
      The ironic inspector iPXE configuration includes the following kernel
      argument:
      
      initrd=agent.ramdisk
      
      However, the ramdisk is actually called ironic-agent.initramfs, so the
      argument should be:
      
      initrd=ironic-agent.initramfs
      
      In BIOS boot mode this does not cause a problem, but for compute nodes
      with UEFI enabled, it seems to be more strict about this, and fails to
      boot.
      
      Change-Id: Ic84f3b79fdd3cd1730ca2fb79c11c7a4e4d824de
      Closes-Bug: #1836375
      7b939756
    • Zuul's avatar
      83dbed6b
    • Raimund Hook's avatar
      Add Region and Multiples into default globals.yml · e72c49ed
      Raimund Hook authored
      Currently, the documentation around configuring regions directs
      you to make changes to openstack_region_name and multiple_regions_names
      in the globals.yml file.
      The defaults weren't represented in there which could potentially cause
      confusion. This change adds these defaults with a brief description.
      
      TrivialFix
      
      Change-Id: Ie0ff7e3dfb9a9355a9c9dbaf27151d90162806dd
      e72c49ed
    • Zuul's avatar
      ab3377d4
    • Raimund Hook's avatar
      Language tweaks in multi-region docs for clarity · fd07e3d9
      Raimund Hook authored
      Tweaked some of the language in doc/source/user/multi-regions.rst for
      clarity purposes.
      
      TrivialFix
      
      Change-Id: Icdd8da6886d0e39da5da80c37d14d2688431ba8f
      fd07e3d9
    • Mark Goddard's avatar
      During deploy, always sync DB · d5e5e885
      Mark Goddard authored
      A common class of problems goes like this:
      
      * kolla-ansible deploy
      * Hit a problem, often in ansible/roles/*/tasks/bootstrap.yml
      * Re-run kolla-ansible deploy
      * Service fails to start
      
      This happens because the DB is created during the first run, but for some
      reason we fail before performing the DB sync. This means that on the second run
      we don't include ansible/roles/*/tasks/bootstrap_service.yml because the DB
      already exists, and therefore still don't perform the DB sync. However this
      time, the command may complete without apparent error.
      
      We should be less careful about when we perform the DB sync, and do it whenever
      it is necessary. There is an argument for not doing the sync during a
      'reconfigure' command, although we will not change that here.
      
      This change only always performs the DB sync during 'deploy' and
      'reconfigure' commands.
      
      Change-Id: I82d30f3fcf325a3fdff3c59f19a1f88055b566cc
      Closes-Bug: #1823766
      Closes-Bug: #1797814
      d5e5e885
  10. Jul 11, 2019
  11. Jul 10, 2019
  12. Jul 09, 2019
Loading