- Oct 12, 2021
-
-
Maksim Malchuk authored
This change updates documentation, examples and tests to support Ironic inspection through DHCP-relay. The dnsmasq service should be configured with more specific format set in the variable ``ironic_dnsmasq_dhcp_range``. See the dnsmasq manual page [1]. [1] https://thekelleys.org.uk/dnsmasq/docs/dnsmasq-man.html Change-Id: I9488a72db588e31289907668f1997596a8ccdec6 Signed-off-by:
Maksim Malchuk <maksim.malchuk@gmail.com>
-
- Oct 06, 2021
-
-
Piotr Parczewski authored
Curl would not work without -L on old openstack.org URLs. TrivialFix Change-Id: I7ed7bd336f4b874fc6027cb9ad20669b08e4a7e9
-
- Oct 04, 2021
-
-
Gaël THEROND (Fl1nt) authored
* Fix various typos and formatting. * Add documentation about custom collector backend. * Add documentation about custom storage backend. Change-Id: If937afc5ce2a2747f464fbaf38a5dcf2e57ba04f Closes-bug: #1940842
-
- Sep 30, 2021
-
-
Mark Goddard authored
* Register Swift-compatible endpoints in Keystone * Load balance across RadosGW API servers using HAProxy The support is exercised in the cephadm CI jobs, but since RGW is not currently enabled via cephadm, it is not yet tested. https://docs.ceph.com/en/latest/radosgw/keystone/ Implements: blueprint ceph-rgw Change-Id: I891c3ed4ed93512607afe65a42dd99596fd4dbf9
-
Michał Nasiadka authored
Change-Id: I51e2b62f563e66b6bb919621272662f3f8721eb2
-
- Sep 26, 2021
-
-
Michal Arbet authored
This patch adding option to control weight of haproxy backends per service via host variable. Example: [control] server1 haproxy_nova_api_weight=10 server2 haproxy_nova_api_weight=2 haproxy_keystone_internal_weight=10 server3 haproxy_keystone_admin_weight=50 If weight is not defined, everything is working as before. Change-Id: Ie8cc228198651c57f8ffe3eb060875e45d1f0700
-
- Sep 16, 2021
-
-
Radosław Piliszek authored
Docs adapted to match. Removed the unsupported-for-quay option to set up a pull-through cache. Closes-Bug: #1942134 Change-Id: If5a26b1ba4bf35bc29306c24f608396dbf5e3371
-
Michal Arbet authored
Haproxy was renamed in [1]. [1] https://review.opendev.org/c/openstack/kolla-ansible/+/770618 Change-Id: Ib2d7f0774fede570a8c4c315d83afd420c31da0b
-
- Aug 20, 2021
-
-
Radosław Piliszek authored
To follow best security practices and help fellow operators. More details inline and in the linked bug report. Closes-Bug: #1940547 Change-Id: Ide9e9009a6e272f20a43319f27d257efdf315f68
-
- Aug 06, 2021
-
-
Ilya Popov authored
Basically, there are three main installation scenario: Scenario 1: Ironic installation together with other openstack services including keystone. In this case variable enable_keystone is set to true and keystone service will be installed together with ironic installation. It is possible realise this scenario, no fix needed Scenario 2: Ironic installation with connection to already installed keystone. In this scenario we have to set enable_keystone to “No” to prevent from new keystone service installation during the ironic installation process. But in other hand, we need to have correct sections in ironic.conf to provide all information needed to connect to existing keystone. But all sections for keystone are added to ironic.conf only if enable_keystone var is set to “Yes”. It isn’t possible to realise this scenario. Proposed fix provide support for this scenario, where multiple regions share the same keystone service. Scenario 3: No keystone integration. Ironic don't connect to Keystone. It is possible realise this scenario, no fix needed Proposed solution also keep the default behaviour: if no enable_keystone_integration is manually defined by default it takes value of enable_keystone variable and all behaviour is the same. But if we don't want to install keystone and want to connect to existing one at the same time, it will be possible to set enable_keystone var to “No” (preventing keystone from installation) and at the same time set ironic_enable_keystone_integration to Yes to allow needed section appear in ironic.conf through templating. Change-Id: I0c7e9a28876a1d4278fb2ed8555c2b08472864b9
-
- Aug 05, 2021
-
-
Piotr Parczewski authored
Change-Id: I0d7c7f47e6653cf2903589a9c86798a8c6404af5
-
- Jul 28, 2021
-
-
Radosław Piliszek authored
As mentioned in the Iced014acee7e590c10848e73feca166f48b622dc commit message, in Ussuri+ we can use ``+sbwtdcpu none +sbwtdio none`` as well. This is due to relying on RMQ-provided erlang in version 23.x. This change adds the extra arguments by default. It should be backported down to Ussuri before we do a release with Iced014acee7e590c10848e73feca166f48b622dc. Change-Id: I32e247a6cb34d7f6763b544f247fd408dce2b3a2
-
Mark Goddard authored
In Ussuri, nova stopped using separate Ceph keys for the volumes and vms pools by default. Instead, we set ceph_nova_keyring to the value of ceph_cinder_keyring by default, which is ceph.client.cinder.keyring. This is in line with the Ceph OpenStack integration guide [1]. However, the user used by nova to access the vms pool (ceph_nova_user) defaults to nova, meaning that nova will still try to use a ceph.client.nova.keyring, which probably does not exist. We did not see this issue in CI, because we set ceph_nova_user to cinder. This change fixes the issue by setting ceph_nova_user to the value of ceph_cinder_user by default, which is cinder. Closes-Bug: #1934145 Related-Bug: #1928690 [1] https://docs.ceph.com/en/latest/rbd/rbd-openstack/ Change-Id: I6aa8db2214e07906f1f3e035411fc80ba911a274
-
- Jul 27, 2021
-
-
wu.chunyang authored
Nova always tries to create the rabbitmq user regardless of whether RabbitMQ is enabled or not. This ps also adds an external rabbitmq doc. Change-Id: Iec517226e4c82ea351889b55689a3efceaadcc76
-
- Jul 22, 2021
-
-
Mark Goddard authored
In the Xena release, Ironic removed the iSCSI driver [1]. The recommended driver is direct, which uses HTTP to transfer the disk image. This requires an HTTP server, and the simplest option is to use the one currently deployed when enable_ironic_ipxe is set to true. For this reason, this patch always enables the HTTP server running on the conductor. iPXE is still enabled separately, since it cannot currently be used at the same time as PXE. [1] https://review.opendev.org/c/openstack/ironic/+/789382 Change-Id: I30c2ad2bf2957ac544942aefae8898cdc8a61ec6
-
- Jul 21, 2021
-
-
Pierre Riteau authored
The variable octavia_amphora_flavor should be octavia_amp_flavor. The variable for customising network and subnet was only mentioned in the example. Change-Id: I3ba5a7ccc2c810fea12bc48584c064738e5aa35e
-
- Jul 02, 2021
-
-
Mark Goddard authored
Adds a new variable, 'disable_firewall', which defaults to true. If set to false, then the host firewall will not be disabled during kolla-ansible bootstrap-servers. Change-Id: Ie5131013012f89c8c3b91ca359ad17d9cb77efc8
-
- Jun 23, 2021
-
-
Michal Arbet authored
This patch is adding configuration option to manipulate with kernel option sysctl_net_ipv4_tcp_retries2. More informations about kernel option in [1][2] and RedHat suggestion [3] to set for DBs and HA. [1]: https://pracucci.com/linux-tcp-rto-min-max-and-tcp-retries2.html [2]: https://blog.cloudflare.com/when-tcp-sockets-refuse-to-die/ [3]: https://access.redhat.com/solutions/726753 Closes-Bug: #1917068 Change-Id: Ia0decbbfa4e33b1889b635f8bb1c9094567a2ce6
-
- Jun 07, 2021
-
-
John Garbutt authored
On machines with many cores, we were seeing excessive CPU load on systems that were not very busy. With the following Erlang VM argument we saw RabbitMQ CPU usage drop from about 150% to around 20%, on a system with 40 hyperthreads. +S 2:2 By default RabbitMQ starts N schedulers where N is the number of CPU cores, including hyper-threaded cores. This is fine when you assume all your CPUs are dedicated to RabbitMQ. Its not a good idea in a typical Kolla Ansible setup. Here we go for two scheduler threads. More details can be found here: https://www.rabbitmq.com/runtime.html#scheduling and here: https://erlang.org/doc/man/erl.html#emulator-flags +sbwt none This stops busy waiting of the scheduler, for more details see: https://www.rabbitmq.com/runtime.html#busy-waiting Newer versions of rabbit may need additional flags: "+sbwt none +sbwtdcpu none +sbwtdio none" But this patch should be back portable to older versions of RabbitMQ used in Train and Stein. Note that information on this tuning was found by looking at data from: rabbitmq-diagnostics runtime_thread_stats More details on that can be found here: https://www.rabbitmq.com/runtime.html#thread-stats Related-Bug: #1846467 Change-Id: Iced014acee7e590c10848e73feca166f48b622dc
-
- May 17, 2021
-
-
Michal Arbet authored
Change-Id: Id9a24c54db9a70c777e51ea10504159d5090f9a7
-
- May 11, 2021
-
-
Florian LEDUC authored
* Enables the Neutron packet logging framework for OVS (https://docs.openstack.org/neutron/latest/admin/config-logging.html). * Adds a toggle variable "enable_neutron_packet_logging" Change-Id: Ica3594cdac634b496949a06ed813dccd18090af4 Implements: blueprint neutron-log-service-plugin
-
- Apr 27, 2021
-
-
Radosław Piliszek authored
As agreed during IRC meeting. [1] [1] http://eavesdrop.openstack.org/meetings/kolla/2021/kolla.2021-04-07-15.00.html Change-Id: Ibffa019e2106784e97f5c482106bb5082f788752
-
Doug Szumski authored
In the Xena cycle it was decided to remove the Monasca Grafana fork due to lack of maintenance. This commit removes the service and provides a limited workaround using the Monasca Grafana datasource with vanilla Grafana. Depends-On: I9db7ec2df050fa20317d84f6cea40d1f5fd42e60 Change-Id: I4917ece1951084f6665722ba9a91d47764d3709a
-
- Apr 26, 2021
-
-
wuchunyang authored
Trivial Fix Change-Id: Ie08877e339455bed45ee467a87de9648678e88c5
-
- Apr 19, 2021
-
-
wuchunyang authored
Change-Id: I713f6fafe328e060a71dbb584e61603e547deaf6
-
Doug Szumski authored
The current behaviour is to support supplying a single folder of Grafana dashboards which can then be populated into a single folder in Grafana. Some users may wish to have sub-folders of Dashboards, and load these into separate dashboard folders in Grafana via a custom provisioning file. For example, a user may have a sub-folder of Ceph dashboards that they wish to keep separate from OpenStack dashboards. This patch supports sub-folders whilst not affecting the original mechanism. Trivial-Fix Change-Id: I9cd289a1ea79f00cee4d2ef30cbb508ac73f9767
-
- Apr 07, 2021
-
-
Doug Szumski authored
Minor corrections to doc and release note. Change-Id: I8a90cbac0b9a1eaa5f6c02271515f2357547f908
-
- Apr 06, 2021
-
-
Radosław Piliszek authored
Per [1]. [1] http://lists.openstack.org/pipermail/openstack-discuss/2021-February/020707.html Change-Id: Id6f3cd158bf5d01750971249b11364b6a8631789 Closes-Bug: #1885689
-
- Mar 26, 2021
-
-
Michal Nasiadka authored
Change-Id: I947c2940518c0f4872acaa977edeaca370dc9a96
-
- Mar 18, 2021
-
-
Bartosz Bezak authored
Change-Id: Ie888e84a3b6e27afc23f89f643fdaa58880aae6d
-
- Mar 16, 2021
-
-
Bartosz Bezak authored
Allow users to import custom grafana dashboards. Dashboards as JSON files should be placed into "{{ node_custom_config }}/grafana/dashboards/" folder. Change-Id: Id0f83b8d08541b3b74649f097b10c9450201b426
-
- Mar 07, 2021
-
-
Doug Szumski authored
Change-Id: Ief84e093829677c97c8df9a08aefca43b1e51aac
-
- Mar 04, 2021
-
-
Doug Szumski authored
This change allows a user to forward control plane logs directly to Elasticsearch from Fluentd, rather than via the Monasca Log API when Monasca is enabled. The Monasca Log API can continue to handle tenant logs. For many use cases this is simpler, reduces resource consumption and helps to decouple control plane logging services from tenant logging services. It may not always be desired, so is optional and off by default. Change-Id: I195e8e4b73ca8f573737355908eb30a3ef13b0d6
-
Doug Szumski authored
The Monasca alerting pipeline provides multi-tenancy alerts and notifications. It runs as an Apache Storm topology and generally places a significant memory and CPU burden on monitoring hosts, particularly when there are lot of metrics. This is fine if the alerting service is in use, but sometimes it is not. For example you may use Prometheus for monitoring the control plane, and wish to offer tenants a monitoring service via Monasca without alerting and notification functionality. In this case it makes sense to disable this part of the Monasca pipeline and this patch adds support for that. If the service is ever re-enabled, all alerts and notifications should spawn back automatically since they are persisted in the central mysql database cluster. Change-Id: I84aa04125c621712f805f41c8efbc92c8e156db9
-
- Mar 03, 2021
-
-
Doug Szumski authored
The Log Metrics service is an admin only service. We now have support in Fluentd via the Prometheus plugin to create metrics from logs. These metrics can be scraped into Monasca or Prometheus. It therefore makes sense to deprecate this service, starting by disabling it by default, and then removing it in the Xena release. This should improve the stability of the Monasca metrics pipeline by ensuring that all metrics pass via the Monasca API for validation, and ensure that metrics generated from logs are available to both Prometheus and Monasca users by default. Change-Id: I704feb4434c1eece3eb00c19dc5f934fd4bc27b4
-
Doug Szumski authored
Historically Monasca Log Transformer has been for log standardisation and processing. For example, logs from different sources may use slightly different error levels such as WARN, 5, or WARNING. Monasca Log Transformer is a place where these could be 'squashed' into a single error level to simplify log searches based on labels such as these. However, in Kolla Ansible, we do this processing in Fluentd so that the simpler Fluentd -> Elastic -> Kibana pipeline also benefits. This helps to avoid spreading out log parsing configuration over many services, with the Fluentd Monasca output plugin being yet another potential place for processing (which should be avoided). It therefore makes sense to remove this service entirely, and squash any existing configuration which can't be moved to Fluentd into the Log Perister service. I.e. by removing this pipeline, we don't loose any functionality, we encourage log processing to take place in Fluentd, or at least outside of Monasca, and we make significant gains in efficiency by removing a topic from Kafka which contains a copy of all logs in transit. Finally, users forwarding logs from outside the control plane, eg. from tenant instances, should be encouraged to process the logs at the point of sending using whichever framework they are forwarding them with. This makes sense, because all Logstash configuration in Monasca is only accessible by control plane admins. A user can't typically do any processing inside Monasca, with or without this change. Change-Id: I65c76d0d1cd488725e4233b7e75a11d03866095c
-
- Mar 02, 2021
-
-
Piotr Parczewski authored
Deprecates support for Prometheus v1.x. In Xena support for it will be removed from Kolla Ansible. Change-Id: I027b19621196c698e09f79af294ba1b5dbfc0516
-
- Feb 23, 2021
-
-
Piotr Parczewski authored
This commit adds possibility to deploy Neutron's DHCP agents in OVN networking scenario. Co-Authored-By:
Michał Nasiadka <mnasiadka@gmail.com> Change-Id: I073d04319b510182f5c1478e728c0c5bcc8799f1
-
- Feb 16, 2021
-
-
Doug Szumski authored
Update the Monasca docs to improve security considerations. Trivial-Fix Change-Id: I97eb8441466f8c6abdbd66068257765bdbe32d4d
-
- Feb 15, 2021
-
-
Pedro Henrique authored
This pull request adds support for the OpenID Connect authentication flow in Keystone and enables both ID and access token authentication flows. The ID token configuration is designed to allow users to authenticate via Horizon using an identity federation; whereas the Access token is used to allow users to authenticate in the OpenStack CLI using a federated user. Without this PR, if one wants to configure OpenStack to use identity federation, he/she needs to do a lot of configurations in the keystone, Horizon, and register quite a good number of different parameters using the CLI such as mappings, identity providers, federated protocols, and so on. Therefore, with this PR, we propose a method for operators to introduce/present the IdP's metadata to Kolla-ansible, and based on the presented metadata, Kolla-ansible takes care of all of the configurations to prepare OpenStack to work in a federated environment. Implements: blueprint add-openid-support Co-Authored-By:
Jason Anderson <jasonanderson@uchicago.edu> Change-Id: I0203a3470d7f8f2a54d5e126d947f540d93b8210
-