action #132947
closedBring back ada.qe.suse.de and fix it properly
0%
Description
Acceptance criteria¶
- AC1: VM owners of a VM on ada.qe.suse.de can use it as in before the AC outage
- AC2: We can remote control the machine
- AC3: https://racktables.nue.suse.com/index.php?page=object&tab=default&object_id=16675 is up-to-date
- AC4: https://gitlab.suse.de/search?project_id=419&search=ada.qe and corresponding for ada.qam and ada-mgmt is up-to-date
Suggestions¶
- Check power cables
- Add power cable connections to racktables
- Ensure mgmt ethernet is connected
- Add connection details to racktables
- Add corresponding entries for both mgmt and eth0 in https://gitlab.suse.de/OPS-Service/salt/ repo and delete old entries
Updated by okurz 10 months ago
- Copied from action #132500: NUE1-SRV2, .qa.suse.de, aarch64 workers offline due to heat-related SRV2 shutdown size:M added
Updated by nicksinger 10 months ago
machine connected to power again and running. IPMI can be reached via 10.161.228.239 with old QAM root password
Updated by nicksinger 10 months ago
AC2: https://gitlab.suse.de/openqa/salt-pillars-openqa/-/merge_requests/567
not sure about the racktables entry because it already contains "new" IP ranges - was it already changed as preparation for the move?
Updated by okurz 10 months ago
https://gitlab.suse.de/openqa/salt-pillars-openqa/-/merge_requests/567 merged
nicksinger wrote:
AC2: https://gitlab.suse.de/openqa/salt-pillars-openqa/-/merge_requests/567
not sure about the racktables entry because it already contains "new" IP ranges - was it already changed as preparation for the move?
hm. https://racktables.nue.suse.com/index.php?page=object&tab=default&object_id=16675 says "Migrate to new qe zone.", not "migrated". Maybe that was interrupted. See our related tickets below #116623 . Right now the management interface as you put into salt pillars is 10.161.228.239 which reverse-resolves to d239.qam.suse.de so a dynamic lease within the qam domain which is also managed in https://gitlab.suse.de/OPS-Service/salt/ . I suggest to just bring https://gitlab.suse.de/OPS-Service/salt/ up-to-date, e.g. delete old obsolete entries as we have found. Any more cleanup can be left until after we physically migrate the machine out of NUE1-SRV2
EDIT: By the way it is planned to move ada.qe.suse.de to PRG2e so this is when we should update all according DHCP/DNS/inventory-management entries, see https://netbox.suse.de/dcim/devices/7560/ and #132617 accordingly
Updated by okurz 10 months ago
- Related to action #132617: Move of selected LSG QE machines NUE1 to PRG2e size:M added
Updated by openqa_review 10 months ago
- Due date set to 2023-08-02
Setting due date based on mean cycle time of SUSE QE Tools
Updated by nicksinger 9 months ago
- Related to deleted (action #132617: Move of selected LSG QE machines NUE1 to PRG2e size:M)
Updated by nicksinger 9 months ago
- Blocked by action #132617: Move of selected LSG QE machines NUE1 to PRG2e size:M added
Updated by okurz 9 months ago
- Tags changed from infra, ada, qam, next-office-day, next-maxtorhof-visit to infra, ada, qam
- Due date deleted (
2023-08-02)
We won't be able to make it by 2023-08-02 considering the blocker. However @nicksinger what's the quick and dirty way to achieve "AC2: We can remote control the machine". I guess you might have already tried that with at least a manual IP entry?
Updated by okurz 9 months ago
- Blocked by deleted (action #132617: Move of selected LSG QE machines NUE1 to PRG2e size:M)
Updated by okurz 9 months ago
- Related to action #132617: Move of selected LSG QE machines NUE1 to PRG2e size:M added
Updated by okurz 9 months ago
Nevermind. I added an entry on https://racktables.nue.suse.com/index.php?page=object&tab=edit&object_id=16675 which covers AC2+3. That's good enough for now. The rest is to be done after #132617
@nicksinger please don't add any tickets in the "blocking" relation, just use the soft "related". As otherwise when we reconsider we always need to redo those steps to be able to resolve tickets.
Updated by okurz 9 months ago
- Status changed from Blocked to Resolved
See? Told you, I can't resolve the ticket in one step as described in #132947-13