coordination #112718
Updated by okurz about 2 years ago
## # Observation We received a lot of alerts over the weekend regarding failed minion jobs and others. Checking Grafana I can see that the problem started Saturday, 18th of June around 13:00 CET: https://stats.openqa-monitor.qa.suse.de/d/WebuiDb/webui-summary?orgId=1&from=1655549105000&to=now The amount of returned PostgreSQL rows looks very suspicious and is now five times as high as before: https://stats.openqa-monitor.qa.suse.de/d/WebuiDb/webui-summary?orgId=1&from=1655475539000&to=now&viewPanel=89 ## Suggestions * Load [OSD database dump](https://progress.opensuse.org/projects/openqav3/wiki/#Backup) from after the incident started and try to reproduce the problem * Research how to find out where heavy queries come from * Research what can cause rows returned to grow from <100k to 20-60M ## Rollback and cleanup steps * *DONE:* on osd `systemctl enable --now telegraf` * on osd `systemctl unmask --now salt-master` and ensure that /etc/telegraf/telegraf.d/telegraf-webui.conf is reverted * on osd `systemctl unmask --now openqa-scheduler` * *DONE:* Retrigger all incomplete jobs since 2022-06-18 with https://github.com/os-autoinst/scripts/blob/master/openqa-advanced-retrigger-jobs * Retrigger failed obs-sync trigger events: https://openqa.suse.de/admin/obs_rsync/-/pipelines * Retrigger failed qem-bot trigger events: https://gitlab.suse.de/qa-maintenance/bot-ng/-/pipelines * *DONE:* Retrigger failed openQA bot trigger events: https://gitlab.suse.de/qa-maintenance/openQABot/-/pipelines * Unmask and start on openqaworker10 and 13: `sudo systemctl unmask --now openqa-worker-cacheservice openqa-worker@{1..20}` * Remove /etc/openqa/templates/main/index.html.ep * Apply salt high state and check that files are back to maintained format, e.g. * /usr/share/openqa/script/openqa-gru