This updates the mariadb chart to use the correct auth values for
the mariadb prometheus exporter. The correct credentials to use
are the credentials in the oslo_db endpoint
Change-Id: I2d325167d7ffdf911a56fe97b879cb13b0d4c195
Fix a naming issue with the cronjob's binary, and schedule the cron
job to run every 15 minutes for the gates. Additonally check to
to ensure we are only running on block devices. Also update the
script to work with ceph-volume created devices.
Change-Id: I8aedab0ac41c191ef39a08034fff3278027d7520
This adds xxx-job name prefixes to the Selenium jobs for consistency
This will also remove the "|| true" suffix that was added temporarily to
ensure the Kibana selenium job did not error. The fix for the issue
was merged so the quick fix is no longer needed and may prevent an
error when an issue actually occurs.
Change-Id: I16881974cbf618b31813964b17c090dbfe33fe51
This proposes moving the multinode job to a periodic job to
match the approach used in the openstack-helm repo.
This also adds the openstack-exporter to the aio monitoring job as
it was previously missing.
This also proposes moving the aio-logging and aio-monitoring jobs
to voting
Change-Id: Idcd4544e03facdcd2430683b66bd80c79e73a372
This helps verify Kibana is working properly by using
Selenium Webdriver to navigate to different index dashboards
and takes screenshot of each one. It also add the scripts to
the gates for single and multinode deployments.
Change-Id: Ic2c91734d1eaac0ea4e7985bf69082942166715d
This fixes the Nagios volume mount for the Elasticsearch query
file. Previously, the check for adding the volumemount to the
pod definition was incorrect. This fixes the conditional check,
and also adds the same conditional check to the configuration
secret
This adds a simple check to the monitoring and multinode jobs to
validate the resulting json gets mounted into the pod successfully
Change-Id: I2af289ccc4e1cff1669cb5e6e829514781b14dd3
This adds scripts using Selenium Webdriver to verify
the dashboards for Gafana, Nagios, and Prometheus are
reachable and functioning as expected. The scripts
create screenshots of each dashboard as well as
pages that can be navigated to.
It also adds the scripts to the gates for the single
and multinode deployments.
Change-Id: I1699e0ba8ff82ce8f59342cc71aad10cff7d2516
Change deployment script for rgw to not use the docker
bridge for public and cluster network overrides. Instead,
calculate network values in same way as other ceph multinodes
deployment steps
Change-Id: I2bacd1af1cc331d76a5d61f3b589ca6ef80b1b2e
Currently the cronjob is broken due to syntax and
permission issues.
Additionally move the cronjob from once a month to
every 15 minutes, and automatically disable the job
unless explicitly enabled.
Change-Id: Id72bdb286c805ccb0ea4e9fcf65fabca94a180dd
This updates the fluentd buffer output configurations to account
for the restraints of the jobs deploying fluentd. This also
renames the fluentd configuration key from td_agent to fluentd to
reflect the fact we're no longer deploying td-agent
This also updates the Elasticsearch default replicas and overrides
the replica counts in each Elasticsearch deployment to account for
resource constraints
Change-Id: I55dee410eced99c3e1645f7452e4306ad646e601
This organizes the single node gates for osh-infra by function.
This organization aims to improve the single node gates in the
following ways:
1. Reduce number of services deployed in single node jobs
2. Only deploy Ceph for logging job, as Elasticsearch requires
RGW for snapshot repositories.
3. Use NFS for storage for monitoring job, as Ceph is not a
requirement for any of the services here.
4. Remove duplicate services deployed to multiple single node jobs
5. Remove storage from openstack-support job, as the only service
requiring storage is rabbitmq. Rabbitmq is deployed with
storage enabled in the openstack-helm checks/gates.
This also removes the documentation for the single node deployments,
as those deployments do not make sense with this change. This should
be revisited as a follow-on once we have a clear path forward for
the larger gate refactoring work
Change-Id: I46951f76904fa2ab245a202d55f76019b7503362
This commit adds falco daemonset of the node for behavioral activity
monitor designed to detect anomalous activity.
Change-Id: I783a2acc03592471c81a8a54e1dc0df140b34a42
This adds inputs for kernel logs on the host, as well as dockerd
and kubelet logs via the systemd plugin. This also adds a filter
for adding the hostname to the kernel log events, for renaming the
fields for systemd logs as kibana can not visualize fields that
begin with an underscore, and adds elasticsearch indexes for both
kernel and systemd logs
Change-Id: I026470dd45a971047f1e5bd1cd49bd0889589d12
This ps adds the ability to use the ceph radosgw s3 api for
snapshot repositories. It removes the ability to use a RWM pvc, as
the radosgw solution provides a more robust approach for storing
index snapshots
Change-Id: Ie56ac41ccdc61bfadcac52b400cceb35403e9fae
This attempts to trim down the dev-deploy gates until further
gate refactoring is complete. This disables the elasticsearch and
fluentd exporters and removes the openstack exporter from the
single node deployment gates to ease the load on nodepool vms
Change-Id: If211511e8f52fe39d293966abbd7e62b45b65970
This moves the mariadb chart to openstack-helm-infra as part of
the effort to move charts to the appropriate repositories
Change-Id: Ife56e28de46c536108cebb4f4cdf6bad2a415289
Story: 2002204
Task: 21725
This adds the process exporter to both the developer and multinode
gates, along with adding the relevant deployment steps to the docs
Change-Id: I85d5c398fbbb62145c9bb4e3a885e9a774725e5a
This adds a ceph developer gate to openstack-helm-infra, which
depends on ceph moving to openstack-helm-infra. This also replaces
the NFS backed storage for the multinode gate with ceph instead
Change-Id: I11268463aa037a2e037217a2dbc89c7432c0d277
This deploys the ingress chart in the openstack-helm-infra dev
and multinode gates, which allows for enabling ingresses in the
charts where defined
Change-Id: I055c7b02d9af68f6e3c5eda33d69dd0b8b1b70ca
This adds ldap support to the grafana chart. This required updating
the version of Grafana to 5.0, as this version allows for using
configuration files to bootstrap the datasources and dashboards
instead of using the grafana http api. This was a necessary change
as using ldap for grafana presented issues trying to create the
datasource via the http api
This also adds a basic helm test for grafana. This test simply
verifies whether the prometheus datasource configured exists and
whether the number of dashboards reported by the admin api matches
the number of dashboards expected
Change-Id: I2e987cb425adba9f909722ffdb25b83f82710c4d
This adds required configuration for enabling LDAP through
the apache proxy in the elasticsearch and kibana charts by
default
Change-Id: Iaff8f328ff50944ddad94ec86b1134ca73750176
This moves the ldap chart to openstack-helm-infra from
openstack-helm, allowing for ldap to provide an authentication
mechanism for components of the LMA stack, and can still be used for
keystone in openstack-helm, as openstack-helm-infra is a required
project
Change-Id: I211bc47c7a3ae875614102c8f64daa1099f702e8
This PS stops pulling the charts default images, as the make file
target cannot target over-rides - resulting in longer gate runs with
twice as many images pulled than required in some cases.
Change-Id: I04c1d43d6ac9d2f509604709139583b085b406e3
See: https://review.openstack.org/#/c/566738/
This moves to using bash scripts for deploying the charts for the
multinode gates in osh-infra as opposed to using the ansible
chart builder
Change-Id: I2951ccf57250a5e6e272872f2d6f7a81bd11f184