- May 12, 2020
-
-
Zuul authored
-
Mark Goddard authored
Ansible 2.9.8 includes a regression on the fileglob plugin [1] that causes the HAProxy role to fail. This change blacklists Ansible 2.9.8 to work around the issue. [1] https://github.com/ansible/ansible/issues/69450 Change-Id: I12ca3b154fc7fed6a221880596e0acb5f6278bb7 Related-Bug: #1878192
-
- May 11, 2020
-
-
Zuul authored
-
Zuul authored
-
Zuul authored
-
Mark Goddard authored
The removal of Kolla Ceph deploy [1] broke gnocchi & external Ceph integration - the variable gnocchi_pool_name is referenced in the config template, but should now be ceph_gnocchi_pool_name. This change fixes the issue. Reported by Nick Wilson. [1] https://review.opendev.org/#/c/704309/12/ansible/roles/gnocchi/defaults/main.yml Change-Id: I7089781c0c4d7bce8a44cb8b1fca847dd0b7efd1 Closes-Bug: #1877974
-
Zuul authored
-
- May 07, 2020
-
-
Zuul authored
-
- May 06, 2020
-
-
Radosław Piliszek authored
Nova cells support introduced a slight regression that triggers odd behaviour when we tried switching to Apache (httpd) [1]. Bootstrap no longer applied permissions recursively to all log files, creating a discrepancy between normal and bootstrap runs and also Nova and other services such as Cinder (regarding bootstrap logging). This patch fixes it. Backport to Train. Not creating reno nor a bug record because it does not affect any current standard usage in any currently known way. Note this only really hides (standardizes?) the global issue that we don't control file permissions on newly created files too well. [1] https://review.opendev.org/724793 Change-Id: I35e9924ccede5edd2e1307043379aba944725143 Needed-By: https://review.opendev.org/724793
-
- May 05, 2020
-
-
Pierre Riteau authored
This note refers to configuration changes done in I626dc7afe9eabfbeb6c08137a3e6bbeebde2b332. Change-Id: I75a37b9d3b28964f353977baa3a9f49fc424d866 Closes-Bug: #1876985
-
generalfuzz authored
Switch URL composition from using VIP to FQDN to connect with Kibana and Elasticsearch services. Change-Id: I5d559ead1d6d5e928e76bb685e0f730868fd7b89 Closes-Bug: #1862419
-
Mark Goddard authored
This was addressed in I21689e22870c2f6206e37c60a3c33e19140f77ff but accidentally reverted in I4f74bfe07d4b7ca18953b11e767cf0bb94dfd67e. Change-Id: Id5fc458b0ca54bddfe9a43cb315dbcfeb2142395
-
- May 01, 2020
-
-
Radosław Piliszek authored
Fixes: - SB/NB DB address format (single host) for SB/NB DB daemon - SB/NB DB address format (all hosts) for Neutron / northd / ovn-ovs bootstrap - OVN tests Change-Id: I539773c48f89b731d068280c228ce11782bf5788 Closes-Bug: #1875222
-
Zuul authored
-
- Apr 30, 2020
-
-
Zuul authored
-
James Kirsch authored
This patch introduces an optional backend encryption for Horizon and Placement services. When used in conjunction with enabling TLS for service API endpoints, network communcation will be encrypted end to end, from client through HAProxy to the Horizon and Placement services. Change-Id: I9cb274141c95aea20e733baa623da071b30acf2d Partially-Implements: blueprint add-ssl-internal-network
-
James Kirsch authored
Add TLS support for Glance api using HAProxy to perform TLS termination. Change-Id: I77051baaeb5d3f7dd9002262534e7d35f3926809 Partially-Implements: blueprint add-ssl-internal-network
-
Hongbin Lu authored
Zun has a new component "zun-cni-daemon" which should be deployed in every compute nodes. It is basically an implementation of CNI (Container Network Interface) that performs the neutron port binding. If users is using the capsule (pod) API, the recommended deployment option is using "cri" as capsule driver. This is basically to use a CRI runtime (i.e. CRI plugin for containerd) for supporting capsules (pods). A CRI runtime needs a CNI plugin which is what the "zun-cni-daemon" provides. The configuration is based on the Zun installation guide [1]. It consits of the following steps: * Configure the containerd daemon in the host. The "zun-compute" container will use grpc to communicate with this service. * Install the "zun-cni" binary at host. The containerd process will invoke this binary to call the CNI plugin. * Run a "zun-cni-daemon" container. The "zun-cni" binary will communicate with this container via HTTP. Relevant patches: Blueprint: https://blueprints.launchpad.net/zun/+spec/add-support-cri-runtime Install guide: https://review.opendev.org/#/c/707948/ Devstack plugin: https://review.opendev.org/#/c/705338/ Kolla image: https://review.opendev.org/#/c/708273/ [1] https://docs.openstack.org/zun/latest/install/index.html Depends-On: https://review.opendev.org/#/c/721044/ Change-Id: I9c361a99b355af27907cf80f5c88d97191193495
-
- Apr 29, 2020
-
-
Zuul authored
-
Zuul authored
-
Zuul authored
-
Zuul authored
-
Noboru Iwamatsu authored
This patch updates the octavia controller deployment to use the latest octavia certificate configuration guide [1]. The dual CA changes were introduced in Train. [1] https://docs.openstack.org/octavia/latest/admin/guides/certificates.html Change-Id: If89ec0d631568db70690f1a69d00115c59abe678 Closes-Bug: #1862133
-
xiaojueguan authored
Change-Id: Ie3022d1721f43dc84e4228331d0d2f6f3a3c7ebd Closes-Bug: 1875613
-
Zuul authored
-
Zuul authored
-
- Apr 28, 2020
-
-
Marcin Juszkiewicz authored
Debian defaults to Python2 which is not complete in aarch64 images. This patch changes CI to always use Python3. We need to install several Python modules to have working ussuri jobs. "Failed to import the required Python library (setuptools) on primary's Python /usr/bin/python3." And then several Python2 ones for train->ussuri upgrade jobs: "Unable to find any of pip2 to use. pip needs to be installed." Change-Id: Ia0d3ff15d97d1cabbb0b8e7f32e8712ca3f94732
-
Marcin Juszkiewicz authored
Change-Id: I18f8855a758703968aba032add68add24b31f673 Closes-bug: #1875588
-
Xing Zhang authored
The octavia service communicates to the barbican service with public endpoint_type by default[1], it should use internal like other services. [1] https://github.com/openstack/octavia/blob/0056b5175f89070164849501ec6d960549b95e34/octavia/common/config.py#L533-L537 Closes-Bug: #1875618 Change-Id: I90d2b0aeac090a3e2366341e260232fc1f0d6492
-
- Apr 27, 2020
- Apr 26, 2020
-
-
Radosław Piliszek authored
Since haproxy is orchestrated via site.yml in a single play, it does not need flushing handlers as handlers run will happen at the end of this play. Change-Id: Ia3743575da707325be93c39b4a2bcae9211cacb2 Related-Bug: #1864810 Closes-Bug: #1875228
-
Radosław Piliszek authored
Follow-up on [1] "Avoid multiple haproxy restarts after reconfiguration". There is no need to duplicate handler name in listen. The issue was because we had two handlers with the same name in the same environment. This causes Ansible not to mark handler as already run. [1] https://review.opendev.org/708385 Change-Id: I5425a8037b6860ef71bce59becff8dfe5b601d4c Related-Bug: #1864810
-
Zuul authored
-
Zuul authored
-
Zuul authored
-