summaryrefslogtreecommitdiffstats
path: root/playbooks/byo/openshift-cluster/upgrades/v3_4/upgrade.yml
Commit message (Collapse)AuthorAgeFilesLines
* Move excluder disablement into control plane and node upgrade playbooksScott Dodson2017-02-061-0/+4
| | | | | So that excluder is disabled and reset within the scope of each of those in addition to the overall playbook
* Cleaning repo cache earlierRussell Teague2017-01-191-2/+2
|
* Validate system restart policy during pre-upgrade.Devan Goodwin2017-01-181-0/+4
| | | | | | | | | | This was done far into the process potentially leaving the user in a difficult situation if they had now considered they were running the upgrade playbook on a host that would be restarted. Instead check configuration and what host we're running on in pre-upgrade and allow the user to abort before making any substantial changes. This is a step towards merging master upgrade into one serial process.
* Add master config hook for 3.4 upgrade and fix facts ordering for config ↵Andrew Butcher2016-12-161-0/+2
| | | | hook run.
* YAML LintingRussell Teague2016-12-121-1/+0
| | | | | * Added checks to make ci for yaml linting * Modified y(a)ml files to pass lint checks
* Fix rare failure to deploy new registry/router after upgrade.Devan Goodwin2016-11-211-2/+2
| | | | | | | | | | | | | | | | | | | | | | | Router/registry update and re-deploy was recently reordered to immediately follow control plane upgrade, right before we proceed to node upgrade. In some situations (small or single host clusters) it appears possible that the deployer pods are running when the node in question is evacuated for upgrade. When the deployer pod dies the deployment is failed and the router/registry continue running the old version, despite the deployment config being updated correctly. This change re-orderes the router/registry upgrade to follow node upgrade. However for separate control plane upgrade, the router/registry still occurs at the end. This is because router/registry seems like they should logically be included in a control plane upgrade, and presumably the user will not manually launch node upgrade so quickly as to trigger an evac on the node in question. Workaround for this problem when it does occur is simply to: oc deploy docker-registry --latest
* Add support for 3.4 upgrade.Devan Goodwin2016-10-251-0/+96
This is a direct copy of 3.3 upgrade playbooks, with 3.3 specific hooks removed and version numbers adjusted appropriately.