- Oct 12, 2021
-
-
Maksim Malchuk authored
This change adds the dnsmasq.log for the ironic-dnsmasq container and also enables more verbose logging when debug logging enabled. This can be triggered globbaly via 'openstack_logging_debug' or per service via 'ironic_logging_debug' or 'neutron_logging_debug'. Change-Id: I0e6b089beb88827effbcc365625eb2df902f5470 Signed-off-by:
Maksim Malchuk <maksim.malchuk@gmail.com>
-
- Sep 27, 2021
-
-
Radosław Piliszek authored
Closes-Bug: #1945070 Change-Id: I1b2a82b57cb9884b6c3c3ad07f6449ae29042a3d
-
- Aug 09, 2021
-
-
Doug Szumski authored
This patch adds support for integrating Prometheus with Fluentd. This can be used to extract useful information about the status of Fluentd, such as output buffer capacity and logging rate, and also to extract metrics from logs via custom Fluentd configuration. More information can be found here in [1]. [1] https://docs.fluentd.org/monitoring-fluentd/monitoring-prometheus Change-Id: I233d6dd744848ef1f1589a462dbf272ed0f3aaae
-
- Jun 23, 2021
-
-
Mark Goddard authored
By default, Ansible injects a variable for every fact, prefixed with ansible_. This can result in a large number of variables for each host, which at scale can incur a performance penalty. Ansible provides a configuration option [0] that can be set to False to prevent this injection of facts. In this case, facts should be referenced via ansible_facts.<fact>. This change updates all references to Ansible facts within Kolla Ansible from using individual fact variables to using the items in the ansible_facts dictionary. This allows users to disable fact variable injection in their Ansible configuration, which may provide some performance improvement. This change disables fact variable injection in the ansible configuration used in CI, to catch any attempts to use the injected variables. [0] https://docs.ansible.com/ansible/latest/reference_appendices/config.html#inject-facts-as-vars Change-Id: I7e9d5c9b8b9164d4aee3abb4e37c8f28d98ff5d1 Partially-Implements: blueprint performance-improvements
-
- Jun 11, 2021
-
-
Matthias Runge authored
the project is deprecated and in the process of being removed from OpenStack upstream. Change-Id: I9d5ebed293a5fb25f4cd7daa473df152440e8b50
-
- Jun 07, 2021
-
-
John Garbutt authored
We really want elasticsearch or monasca to catch all logs, to providd the required centrailsed logging. While these appears to make little material difference, it should make it harder for logs to not get caught by any of the outputs we have configured. TrivialFix Change-Id: I3bb74dcdc3cbe78cd1e1657f44e2a0af9d6508ef
-
- Jun 04, 2021
-
-
John Garbutt authored
When using elasticsearch 7 with fluentd, you seem to get a lot of warnings in the docker logs output that look like: [types removal] Specifying types in bulk requests is deprecated. The docs suggest adding suppress_type_name to stop these warnings, and that seems to work without affecting any functionality. Further info here: https://github.com/uken/fluent-plugin-elasticsearch/issues/785 Closes-Bug: #1930856 Change-Id: I45be67df3717f78d78bcdc7df69600ab8681922f
-
- Jun 03, 2021
-
-
John Garbutt authored
Currently the logs tagged with infra.mariadb.xinetd flow into elasticsearch with no hostname or programname attach, thus making navigating the logs very hard. The quick fix is renaming the tag to infra.mariadb-xinetd, which is just enought to ensure the logs are processed correctly with the existing filters. TrivialFix Change-Id: Icd72206de7c1f701bdf35c8fb3b128ef2dbe29a8
-
John Garbutt authored
Currently when elasticsearch log output is enabled there are lots of warnings going into elasticsearch about type being deprecated and needing to move to @type. This change stops those warnings. TrivialFix Change-Id: Ideac1925cb764ad0d7d8416f56d5e4a993c6d8b6
-
- May 10, 2021
-
-
John Garbutt authored
This is a follow up on the change with the following ID: I337f42e174393f68b43e876ef075a74c887a5314 TrivialFix Change-Id: Ibb67811d7b086ef9ef4c695ae589171af0c4d657
-
- Mar 25, 2021
-
-
Martin Chlumsky authored
This change also adds support for Octavia backend TLS. Closes-Bug: #1874228 Depends-On: https://review.opendev.org/c/openstack/kolla/+/779892 Change-Id: I5ff84aec4cdbc15f6a797391815243821dbdbd67
-
- Mar 04, 2021
-
-
Doug Szumski authored
This change allows a user to forward control plane logs directly to Elasticsearch from Fluentd, rather than via the Monasca Log API when Monasca is enabled. The Monasca Log API can continue to handle tenant logs. For many use cases this is simpler, reduces resource consumption and helps to decouple control plane logging services from tenant logging services. It may not always be desired, so is optional and off by default. Change-Id: I195e8e4b73ca8f573737355908eb30a3ef13b0d6
-
- Jan 27, 2021
-
-
Doug Szumski authored
There are a few issues fixed here: - The Barbican API service doesn't set a log file, so all the Barbican API service logs go to loadwsgi.py.log by default. - The logs in loadwsgi.py.log are not ingested properly by Fluentd. - uWSGI logs go to barbican-api.log. This would normally be used as the log file for the Barbican API service logs. This patch makes the following changes to address the above issues: - All uWSGI logs (from the Emperor and Vassals) go to barbican_api_uwsgi_access.log Although these logs aren't strictly all access logs, this follows the existing pattern for WSGI logs. - The Barbican API service logs are written to barbican-api.log instead of loadwsgi.py.log. This follows the pattern used by other OpenStack services. - Fluentd is configured to parse the Barbican API service logs as it would with other OpenStack Python services. Change-Id: I6d03fa8c81c52b6f061514a836bbd15bb6639aaf Closes-Bug: #1891343
-
- Dec 16, 2020
-
-
Ghanshyam Mann authored
Qinling project is retiring in Wallaby cycle[1]. This commit removes the ansible roles of Qinling project before its code is removed. Needed-By: https://review.opendev.org/c/openstack/qinling/+/764521 [1] http://lists.openstack.org/pipermail/openstack-discuss/2020-November/018638.html Change-Id: I6543bacff638b1649511f7e779807954c34ef570
-
- Oct 12, 2020
-
-
James Kirsch authored
Add TLS support for backend Neutron API Server communication using HAProxy to perform TLS termination. When used in conjunction with enabling TLS for service API endpoints, network communication will be encrypted end to end, from client through HAProxy to the Neutron service. Change-Id: Ib333a1f1bd12491df72a9e52d961161210e2d330 Partially-Implements: blueprint add-ssl-internal-network
-
- Sep 18, 2020
-
-
Michal Nasiadka authored
Change-Id: Iabc0115d3476a626df134cc70cb473bf6e72487e Closes-Bug: #1890439
-
- Aug 19, 2020
-
-
Rafael Weingärtner authored
The goal for this push request is to normalize the construction and use of internal, external, and admin URLs. While extending Kolla-ansible to enable a more flexible method to manage external URLs, we noticed that the same URL was constructed multiple times in different parts of the code. This can make it difficult for people that want to work with these URLs and create inconsistencies in a large code base with time. Therefore, we are proposing here the use of "single Kolla-ansible variable" per endpoint URL, which facilitates for people that are interested in overriding/extending these URLs. As an example, we extended Kolla-ansible to facilitate the "override" of public (external) URLs with the following standard "<component/serviceName>.<companyBaseUrl>". Therefore, the "NAT/redirect" in the SSL termination system (HAproxy, HTTPD or some other) is done via the service name, and not by the port. This allows operators to easily and automatically create more friendly URL names. To develop this feature, we first applied this patch that we are sending now to the community. We did that to reduce the surface of changes in Kolla-ansible. Another example is the integration of Kolla-ansible and Consul, which we also implemented internally, and also requires URLs changes. Therefore, this PR is essential to reduce code duplicity, and to facility users/developers to work/customize the services URLs. Change-Id: I73d483e01476e779a5155b2e18dd5ea25f514e93 Signed-off-by:
Rafael Weingärtner <rafael@apache.org>
-
- Aug 06, 2020
-
-
Kien Nguyen authored
Change-Id: Ib08544a265fe1e0d599a6243cb9d38ed9a7769e1
-
- Jul 21, 2020
-
-
Pierre Riteau authored
A "@type copy" statement is already present at the beginning of each match element, so extra "type copy" are not needed. They are causing the following warnings in fluentd logs: [warn]: parameter 'type' in <match syslog.local0.**> [warn]: parameter 'type' in <match syslog.local1.**> This commit also harmonizes indentation of the Monasca config block. Change-Id: I779c2b942d007acbdd43d999f2fc0cdc131d431f Related-Bug: #1885873
-
- Jul 02, 2020
-
-
Justinas Balciunas authored
Time format in Ruby Time.strptime is not accepting padding flags, therefore we need to remove them for the Fluentd to be able to parse MariaDB xinetd logs properly. Change-Id: Iabfa9afdcad505106a5580eb2d058273ee5f7c1f Closes-Bug: #1886002
-
Stig Telfer authored
In Fluentd v0.12, both the in memory and file buffer chunk size default to 8MB. In v1.0 the file buffer defaults to 256MB. This can exceed the Monasca Log or Unified API maximum chunk size which is set to 10MB. This can result in logs being rejected and filling the local buffer on disk. Change-Id: I9c495773db726a3c5cd94b819dff4141737a1d6e Closes-Bug: #1885885 Co-Authored-By:
Sebastian Luna Valero <sebastian.luna.valero@gmail.com>
-
- Jul 01, 2020
-
-
Stig Telfer authored
Resolve trivial syntax error in Fluentd output config for Monasca. Change-Id: I20b37bb83a76bfabb1126925a1b4f1f59767b7a3 Co-Authored-By:
Sebastian Luna Valero <sebastian.luna.valero@gmail.com> Closes-Bug: #1885873
-
- Jun 25, 2020
-
-
Mark Goddard authored
Currently there is no way to configure a CA certificate bundle file for fluentd to Elasticsearch communication. This change adds a new variable, 'fluentd_elasticsearch_cacert' with a default value set to the value of 'openstack_cacert. Closes-Bug: #1885109 Change-Id: I5bbf55a4dd4ccce9fa2635cee720139c088268e3
-
- Jun 20, 2020
-
-
gugug authored
more info: https://review.opendev.org/#/c/721733/ Depends-On: I561ead226f714d98c8e06e6027715a64c3a8e47e Depends-On: I21c9ab9820f78cf76adf11c5f0591c60f76372a8 Change-Id: Ic740d090211ee331b374a6dac69dfde466df7200 Co-Authored-By:
jacky06 <zhang.min@99cloud.net>
-
- May 23, 2020
-
-
Doug Szumski authored
The Monasca Log API has been removed and in this change we switch to using the unified API. If dedicated log APIs are required then this can be supported through configuration. Out of the box the Monasca API is used for both logs and metrics which is envisaged to work for most use cases. In order to use the unified API for logs, we need to disable the legacy Kafka client. We also rename the Monasca API config file to remove a warning about using the old style name. Depends-On: https://review.opendev.org/#/c/728638 Change-Id: I9b6bf5b6690f4b4b3445e7d15a40e45dd42d2e84
-
- May 20, 2020
-
-
Michal Nasiadka authored
Depends-On: https://review.opendev.org/710217/ Change-Id: I85652f23e487c40192106d23f2cdd45a3077deca
-
- May 15, 2020
-
-
Michal Nasiadka authored
Change-Id: I812665059783617d581d748e619b29426f89b353
-
- Apr 30, 2020
-
-
James Kirsch authored
Add TLS support for Glance api using HAProxy to perform TLS termination. Change-Id: I77051baaeb5d3f7dd9002262534e7d35f3926809 Partially-Implements: blueprint add-ssl-internal-network
-
Hongbin Lu authored
Zun has a new component "zun-cni-daemon" which should be deployed in every compute nodes. It is basically an implementation of CNI (Container Network Interface) that performs the neutron port binding. If users is using the capsule (pod) API, the recommended deployment option is using "cri" as capsule driver. This is basically to use a CRI runtime (i.e. CRI plugin for containerd) for supporting capsules (pods). A CRI runtime needs a CNI plugin which is what the "zun-cni-daemon" provides. The configuration is based on the Zun installation guide [1]. It consits of the following steps: * Configure the containerd daemon in the host. The "zun-compute" container will use grpc to communicate with this service. * Install the "zun-cni" binary at host. The containerd process will invoke this binary to call the CNI plugin. * Run a "zun-cni-daemon" container. The "zun-cni" binary will communicate with this container via HTTP. Relevant patches: Blueprint: https://blueprints.launchpad.net/zun/+spec/add-support-cri-runtime Install guide: https://review.opendev.org/#/c/707948/ Devstack plugin: https://review.opendev.org/#/c/705338/ Kolla image: https://review.opendev.org/#/c/708273/ [1] https://docs.openstack.org/zun/latest/install/index.html Depends-On: https://review.opendev.org/#/c/721044/ Change-Id: I9c361a99b355af27907cf80f5c88d97191193495
-
- Apr 17, 2020
-
-
Q.hongtao authored
Kolla Ansible was missing vitrage-persistor service required by Vitrage for data storage. Depends on fixing availability of Kolla image. Change-Id: I8158ba66b8b624f6bcb89da9c990a30a68b7187b Depends-On: Id5e143636f9a81e7294b775f3d8b9134bee58054 Closes-Bug: #1869319
-
- Mar 18, 2020
-
-
Radosław Piliszek authored
Fluentd cannot accept empty 'path' parameter. I refactored the service list following the general pattern we have. Change-Id: I83d820efcc7e86bac9f8bda26a8f8bece72159e6 Closes-bug: #1867953
-
- Feb 04, 2020
-
-
Doug Szumski authored
By default a retry limit of 17 exists. When the limit is reached buffered logs are discarded. To avoid this, we disable the retry limit. The risk of bringing down the host by filling the Fluent data docker volume is managed by the maximum buffer size which is 2GB by default. In summary, after this change, the net behaviour is that Fluentd should buffer up to a maximum of 2GB of logs locally, and attept to post them to the Monasca Log API at intervals not exceeding 30 minutes. Closes-Bug: #1855702 Change-Id: I0d5a3dab29635c00411f4f51e5a0721726df2abd
-
Doug Szumski authored
This enables buffering to file, rather than memory for Monasca logs. A dedicated docker volume is used for the file buffer. If a post to the Monasca Log API fails, retries will be made using an exponential backoff algorithm with a maximum retry interval of 30mins. The maximum interval is set relatively low to try and reduce the risk of large buffers accumulating, and therefore the risk of overloading the Monasca Log API. Closes-Bug: #1855700 Change-Id: Ib5286e9dbaf2bc92d2f4960b2131223ab5dbdbec
-
- Dec 09, 2019
-
-
Mark Goddard authored
WSGI log files use a different input configuration than OpenStack log files. Currently this depends on log files matching either *-access.log or *-error.log. Some services use *_access.log or *_error.log, so are not parsed correctly. This change modifies the fluentd configuration to accept an underscore or hyphen for WSGI log file names. Change-Id: I566d6cac0b6749054fd5422ec8f36f99dacb1db7 Closes-Bug: #1720371
-
Krzysztof Klimonda authored
Enable reconnect_on_error option so that ES plugin re-establishes a new session to the ES cluster on errors. Also, enable buffering to the file, so that the buffer survives container restarts. Co-Authored-By:
Michal Nasiadka <mnasiadka@gmail.com> Co-Authored-By:
Radosław Piliszek <radoslaw.piliszek@gmail.com> Co-Authored-By:
Doug Szumski <doug@stackhpc.com> Closes-Bug: #1830724 Change-Id: Ia40685b9d4fc02194e03c8791ddeb3d29d7f07f6
-
- Dec 01, 2019
-
-
Erik Olof Gunnar Andersson authored
Depends-On: https://review.opendev.org/692948/ Depends-On: https://review.opendev.org/692691/ Change-Id: I07827b896d36c3723697540fcff164224f6729af
-
- Nov 18, 2019
-
-
Michal Nasiadka authored
Opendaylight support has been deprecated in Train - time to remove it. Change-Id: I3a61bfbcbf366c327ea3e25d2424bc3fedca29f0
-
- Nov 05, 2019
-
-
lklimin authored
Change-Id: I49b24545501085d5a44f4de73f0c6dd21e06e2a0 Closes-Bug: #1835501
-
- Oct 29, 2019
-
-
Isaac Prior authored
MariaDB logs contain two different log message formats, one output from mysqld and one from mysqld_safe. This patch splits the message formats by tag and parses them separately. Change-Id: I58857be67ae387eeda7487811a6af85b0f95970c Closes-Bug: #1845629
-
- Oct 16, 2019
-
-
Radosław Piliszek authored
Introduce kolla_address filter. Introduce put_address_in_context filter. Add AF config to vars. Address contexts: - raw (default): <ADDR> - memcache: inet6:[<ADDR>] - url: [<ADDR>] Other changes: globals.yml - mention just IP in comment prechecks/port_checks (api_intf) - kolla_address handles validation 3x interface conditional (swift configs: replication/storage) 2x interface variable definition with hostname (haproxy listens; api intf) 1x interface variable definition with hostname with bifrost exclusion (baremetal pre-install /etc/hosts; api intf) neutron's ml2 'overlay_ip_version' set to 6 for IPv6 on tunnel network basic multinode source CI job for IPv6 prechecks for rabbitmq and qdrouterd use proper NSS database now MariaDB Galera Cluster WSREP SST mariabackup workaround (socat and IPv6) Ceph naming workaround in CI TODO: probably needs documenting RabbitMQ IPv6-only proto_dist Ceph ms switch to IPv6 mode Remove neutron-server ml2_type_vxlan/vxlan_group setting as it is not used (let's avoid any confusion) and could break setups without proper multicast routing if it started working (also IPv4-only) haproxy upgrade checks for slaves based on ipv6 addresses TODO: ovs-dpdk grabs ipv4 network address (w/ prefix len / submask) not supported, invalid by default because neutron_external has no address No idea whether ovs-dpdk works at all atm. ml2 for xenapi Xen is not supported too well. This would require working with XenAPI facts. rp_filter setting This would require meddling with ip6tables (there is no sysctl param). By default nothing is dropped. Unlikely we really need it. ironic dnsmasq is configured IPv4-only dnsmasq needs DHCPv6 options and testing in vivo. KNOWN ISSUES (beyond us): One cannot use IPv6 address to reference the image for docker like we currently do, see: https://github.com/moby/moby/issues/39033 (docker_registry; docker API 400 - invalid reference format) workaround: use hostname/FQDN RabbitMQ may fail to bind to IPv6 if hostname resolves also to IPv4. This is due to old RabbitMQ versions available in images. IPv4 is preferred by default and may fail in the IPv6-only scenario. This should be no problem in real life as IPv6-only is indeed IPv6-only. Also, when new RabbitMQ (3.7.16/3.8+) makes it into images, this will no longer be relevant as we supply all the necessary config. See: https://github.com/rabbitmq/rabbitmq-server/pull/1982 For reliable runs, at least Ansible 2.8 is required (2.8.5 confirmed to work well). Older Ansible versions are known to miss IPv6 addresses in interface facts. This may affect redeploys, reconfigures and upgrades which run after VIP address is assigned. See: https://github.com/ansible/ansible/issues/63227 Bifrost Train does not support IPv6 deployments. See: https://storyboard.openstack.org/#!/story/2006689 Change-Id: Ia34e6916ea4f99e9522cd2ddde03a0a4776f7e2c Implements: blueprint ipv6-control-plane Signed-off-by:
Radosław Piliszek <radoslaw.piliszek@gmail.com>
-