- Nov 11, 2022
-
-
Doug Szumski authored
Kolla Ansible is switching to OpenSearch and is dropping support for deploying ElasticSearch. This is because the final OSS release of ElasticSearch has exceeded its end of life. Monasca is affected because it uses both Logstash and ElasticSearch. Whilst it may continue to work with OpenSearch, Logstash remains an issue. In the absence of any renewed interest in the project, we remove support for deploying it. This helps to reduce the complexity of log processing configuration in Kolla Ansible, freeing up development time. Change-Id: I6fc7842bcda18e417a3fd21c11e28979a470f1cf
-
- Jun 20, 2022
-
-
Radosław Piliszek authored
Per comments on [1]. [1] https://review.opendev.org/c/openstack/kolla-ansible/+/843727 Change-Id: I60162b54bc06e158534d29311d4474b34750c64d
-
- May 02, 2022
-
-
Pierre Riteau authored
This is a follow up to I7e5c1e20c7b66b64cbd333f669ef8d8da60daaa8. Change-Id: I11a86f59c1fb9cddde3370b544ee7bf4e8ae4fb4
-
- Apr 23, 2022
-
-
Pierre Riteau authored
The ipmi-exporter code blocks had too much indentation. Change-Id: I2cf3ab4873d9df2dfe1375cf7e2e0e5dc6571120
-
- Apr 11, 2022
-
-
Marcin Juszkiewicz authored
Change-Id: Ia2f549ba119ac09c4d1e4279baf594a42480511f
-
- Feb 14, 2022
-
-
Will Szumski authored
Grafana requires the scrape interval to be set to be able to compute $__rate_interval. The default is 15s which does not match the kolla default of 60s. The symptom of not setting this is that you will see "no data" when zooming graphs that use rate queries. This occurs as the interval will be set to a period shorter than the scrape interval. The recommendation is that you use a common scrape interval for all jobs. See: - https://grafana.com/blog/2020/09/28/new-in-grafana-7.2-__rate_interval-for-prometheus-rate-queries-that-just-work/ - https://stackoverflow.com/questions/66369969/set-scrape-interval-in-provisioned-prometheus-data-source-in-grafana Change-Id: I7e5c1e20c7b66b64cbd333f669ef8d8da60daaa8
-
- Sep 16, 2021
-
-
Radosław Piliszek authored
Docs adapted to match. Removed the unsupported-for-quay option to set up a pull-through cache. Closes-Bug: #1942134 Change-Id: If5a26b1ba4bf35bc29306c24f608396dbf5e3371
-
- Aug 05, 2021
-
-
Piotr Parczewski authored
Change-Id: I0d7c7f47e6653cf2903589a9c86798a8c6404af5
-
- Apr 27, 2021
-
-
Doug Szumski authored
In the Xena cycle it was decided to remove the Monasca Grafana fork due to lack of maintenance. This commit removes the service and provides a limited workaround using the Monasca Grafana datasource with vanilla Grafana. Depends-On: I9db7ec2df050fa20317d84f6cea40d1f5fd42e60 Change-Id: I4917ece1951084f6665722ba9a91d47764d3709a
-
- Apr 19, 2021
-
-
Doug Szumski authored
The current behaviour is to support supplying a single folder of Grafana dashboards which can then be populated into a single folder in Grafana. Some users may wish to have sub-folders of Dashboards, and load these into separate dashboard folders in Grafana via a custom provisioning file. For example, a user may have a sub-folder of Ceph dashboards that they wish to keep separate from OpenStack dashboards. This patch supports sub-folders whilst not affecting the original mechanism. Trivial-Fix Change-Id: I9cd289a1ea79f00cee4d2ef30cbb508ac73f9767
-
- Apr 07, 2021
-
-
Doug Szumski authored
Minor corrections to doc and release note. Change-Id: I8a90cbac0b9a1eaa5f6c02271515f2357547f908
-
- Apr 06, 2021
-
-
Radosław Piliszek authored
Per [1]. [1] http://lists.openstack.org/pipermail/openstack-discuss/2021-February/020707.html Change-Id: Id6f3cd158bf5d01750971249b11364b6a8631789 Closes-Bug: #1885689
-
- Mar 18, 2021
-
-
Bartosz Bezak authored
Change-Id: Ie888e84a3b6e27afc23f89f643fdaa58880aae6d
-
- Mar 16, 2021
-
-
Bartosz Bezak authored
Allow users to import custom grafana dashboards. Dashboards as JSON files should be placed into "{{ node_custom_config }}/grafana/dashboards/" folder. Change-Id: Id0f83b8d08541b3b74649f097b10c9450201b426
-
- Mar 07, 2021
-
-
Doug Szumski authored
Change-Id: Ief84e093829677c97c8df9a08aefca43b1e51aac
-
- Mar 04, 2021
-
-
Doug Szumski authored
This change allows a user to forward control plane logs directly to Elasticsearch from Fluentd, rather than via the Monasca Log API when Monasca is enabled. The Monasca Log API can continue to handle tenant logs. For many use cases this is simpler, reduces resource consumption and helps to decouple control plane logging services from tenant logging services. It may not always be desired, so is optional and off by default. Change-Id: I195e8e4b73ca8f573737355908eb30a3ef13b0d6
-
Doug Szumski authored
The Monasca alerting pipeline provides multi-tenancy alerts and notifications. It runs as an Apache Storm topology and generally places a significant memory and CPU burden on monitoring hosts, particularly when there are lot of metrics. This is fine if the alerting service is in use, but sometimes it is not. For example you may use Prometheus for monitoring the control plane, and wish to offer tenants a monitoring service via Monasca without alerting and notification functionality. In this case it makes sense to disable this part of the Monasca pipeline and this patch adds support for that. If the service is ever re-enabled, all alerts and notifications should spawn back automatically since they are persisted in the central mysql database cluster. Change-Id: I84aa04125c621712f805f41c8efbc92c8e156db9
-
- Mar 03, 2021
-
-
Doug Szumski authored
The Log Metrics service is an admin only service. We now have support in Fluentd via the Prometheus plugin to create metrics from logs. These metrics can be scraped into Monasca or Prometheus. It therefore makes sense to deprecate this service, starting by disabling it by default, and then removing it in the Xena release. This should improve the stability of the Monasca metrics pipeline by ensuring that all metrics pass via the Monasca API for validation, and ensure that metrics generated from logs are available to both Prometheus and Monasca users by default. Change-Id: I704feb4434c1eece3eb00c19dc5f934fd4bc27b4
-
Doug Szumski authored
Historically Monasca Log Transformer has been for log standardisation and processing. For example, logs from different sources may use slightly different error levels such as WARN, 5, or WARNING. Monasca Log Transformer is a place where these could be 'squashed' into a single error level to simplify log searches based on labels such as these. However, in Kolla Ansible, we do this processing in Fluentd so that the simpler Fluentd -> Elastic -> Kibana pipeline also benefits. This helps to avoid spreading out log parsing configuration over many services, with the Fluentd Monasca output plugin being yet another potential place for processing (which should be avoided). It therefore makes sense to remove this service entirely, and squash any existing configuration which can't be moved to Fluentd into the Log Perister service. I.e. by removing this pipeline, we don't loose any functionality, we encourage log processing to take place in Fluentd, or at least outside of Monasca, and we make significant gains in efficiency by removing a topic from Kafka which contains a copy of all logs in transit. Finally, users forwarding logs from outside the control plane, eg. from tenant instances, should be encouraged to process the logs at the point of sending using whichever framework they are forwarding them with. This makes sense, because all Logstash configuration in Monasca is only accessible by control plane admins. A user can't typically do any processing inside Monasca, with or without this change. Change-Id: I65c76d0d1cd488725e4233b7e75a11d03866095c
-
- Mar 02, 2021
-
-
Piotr Parczewski authored
Deprecates support for Prometheus v1.x. In Xena support for it will be removed from Kolla Ansible. Change-Id: I027b19621196c698e09f79af294ba1b5dbfc0516
-
- Feb 16, 2021
-
-
Doug Szumski authored
Update the Monasca docs to improve security considerations. Trivial-Fix Change-Id: I97eb8441466f8c6abdbd66068257765bdbe32d4d
-
- Jan 27, 2021
-
-
Piotr Parczewski authored
There are inconsitencies across the documentation and the source code files when it comes to project's name (Kolla Ansible vs. Kolla-Ansible). This commit aims at unifying it so that the naming becomes consistent everywhere. Change-Id: I903b2e08f5458b1a1abc4af3abefe20b66c23a54
-
- Jan 18, 2021
-
-
Piotr Parczewski authored
Update for the example command line options. Change-Id: Ida6e882d1014cdc5e55383a5d5ba8fa0e824a693
-
- Jan 12, 2021
-
-
Piotr Parczewski authored
It is now possible to deploy either 1.x or 2.x version of Prometheus. The new 2.x version introduces breaking changes in terms of storage format and command line options. Change-Id: I80cc6f1947f3740ef04b29839bfa655b14fae146 Co-Authored-By:
Radosław Piliszek <radoslaw.piliszek@gmail.com>
-
- May 22, 2020
-
-
Raimund Hook authored
Kolla-ansible version 4.0.0 contained the steps to follow when logging in to Kibana for the first time. These got deleted when the process was seemingly automated, but the relevant machinery no longer works. See [1] as well. Backport to Ussuri, Train, Stein (possibly more). [1] https://review.opendev.org/726289 Change-Id: If65622dc78e7f8fd16e37ee31bc9f34eb9267549
-
- May 11, 2020
-
-
Will Szumski authored
This provides a generic mechanism to include extra files that you can reference in prometheus.yml, for example: scrape_targets: - job_name: ipmi params: module: default scrape_interval: 1m scrape_timeout: 30s metrics_path: /ipmi scheme: http file_sd_configs: - files: - /etc/prometheus/extras/file_sd/ipmi-exporter-targets.yml refresh_interval: 5m Change-Id: Ie2f085204b71725b901a179ee51541f1f383c6fa Related: blueprint custom-prometheus-targets
-
Will Szumski authored
This provides a mechanism to scrape targets defined outside of kolla-ansible. Depends-On: https://review.opendev.org/#/c/685671/ Change-Id: I0950341b147bb374b4128f09f807ef5a756f5dfa Related: blueprint custom-prometheus-targets
-
- Apr 23, 2020
-
-
Raimund Hook authored
Just making it slightly more readable - there was an extra 'an'. TrivialFix Change-Id: I488f702449e217335321988874b6c3ee3136f497 Signed-off-by:
Raimund Hook <openstack@sting-ray.za.net>
-
- Mar 30, 2020
-
-
Doug Szumski authored
Not everyone wants Kafka data stored on a Docker volume. This change allows a user to flexibly control where the data is stored. Change-Id: I2ba8c7a85c7bf2564f954a43c6e6dbb3257fe902
-
- Feb 19, 2020
-
-
Will Szumski authored
This allows you to tune the performance of InfluxDB by locating the volume on a drive that is separate to the default docker storage. Change-Id: Iea555a2702b225b30f5d7035b8a703d4f3376ee7
-
- Dec 09, 2019
-
-
Doug Szumski authored
This allows users to supply an Elasticsearch Curator actions file to manage log retention [1]. Curator then runs on a cron job, which defaults to every day. A default curator actions file is provided, which can be customised by the end user if required. [1] https://www.elastic.co/guide/en/elasticsearch/client/curator/current/actionfile.html Change-Id: Ide9baea9190ae849e61b9d8b6cff3305bdcdd534
-
- Sep 24, 2019
-
-
Dincer Celik authored
Change-Id: I8bb39eaf8a4239c37fcbf91b55ec8003542e2506
-
- Jun 20, 2019
-
-
Doug Szumski authored
This commit should help guide people migrating to Kolla Monasca through the murky depths of the migration process. Since Kolla did not support Monasca in Queens, some of these steps which could be automated are not. Change-Id: I79051cca27178c3cf1671f5c603e38baf929c55c
-
- Apr 08, 2019
-
-
Doug Szumski authored
The recent addition of this flag make the configuration of stand-alone Monasca slightly simpler. Change-Id: Ib4c03926daa3f0f3de0fa4412cd785d87ed5500c
-
- Mar 08, 2019
-
-
Doug Szumski authored
In some scenarios it may be useful to perform custom formatting of logs before forwarding them. For example, the JSON formatter plugin can be used to convert an event to JSON. Change-Id: I3dd9240c5910a9477456283b392edc9566882dcd
-
- Feb 25, 2019
-
-
Christian Berendt authored
Change-Id: Id8276448c6e779b2b4a0aafee45d953c4f009fc1
-
- Feb 14, 2019
-
-
Doug Szumski authored
Until the Monasca Kafka client fork is removed it is currently required to run Kafka in compatibility mode. It is also necessary to disable an optimisation in the Kafka brokers to clean up idle connections. This is because the optimisation was added after the Monasca Kafka client was forked, and the client hasn't been updated since. These settings are now applied automatically when Monasca is enabled. Change-Id: I6935f1fb29f4f731cf3c9a70a0adf4d5812ca55e
-
- Nov 23, 2018
-
-
Eduardo Gonzalez authored
Change index to ease identify what service want to look. Split docs into more specific folder such as networking and storage. Change-Id: Ic7ac12b3dd555fa5c018eeb897ccd4a5a2dfe8f3
-