| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
| |
Move more checks outside of init/main.yml for
speeding up upgrades and other operational plays that
need to run.
|
|
|
|
|
| |
We set these variables using facts in init, no need
to duplicate the logic all around the codebase.
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
This commit refactors some upgrade code paths.
Touched areas are:
1) Reduces usage of 'oo_all_hosts' in various places,
especially when running upgrade_control_plane.
2) Reuses common code across the various upgrade*
playbooks.
3) Moves docker upgrade checks into container_runtime_role.
4) Combines smaller playbooks and plays to reduce file sprawl.
|
|
|
|
|
| |
Bring openshift_facts into scope for plays that utilize
openshift_service_type as it is defined there.
|
|\
| |
| | |
Remove openshift.common.service_type
|
| |
| |
| |
| |
| |
| |
| |
| | |
This commit removes openshift.common.service_type
in favor of openshift_service_type.
This commit also removes r_openshift_excluder_service_type
from plays in favor of using the role's defaults.
|
|\ \
| |/
|/|
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Include Deprecation: Convert to import_playbook
* Converts playbooks/ to use `import_playbook`.
* Updates remaining `include:` tasks to `include_tasks:`
Trello: https://trello.com/c/ZTyZu3UM/484-3-ansible-24-include-deprecation
|
| | |
|
|/
|
|
| |
Replace with `oc adm`
|
| |
|
|
|
|
|
|
| |
When a package install/update fails due to network blips or other spotty
availability, retry it. If the failure is a real failure (e.g. package
is really not there) it still fails after 3 tries (Ansible default).
|
|
|
|
| |
list the package with --installed.
|
| |
|
| |
|
|
|
|
|
|
|
|
|
| |
Most of this role's purpose was to set facts. The vast majority
of these facts were simply redefining user-supplied variables.
This commit also removes various artifacts leftover from
previous versions, as well as variables that seem to be
entirely unused.
|
| |
|
|
|
|
| |
Prevents playbooks from accidentally restarting the master service.
|
|\
| |
| | |
drain still pending in below files without fix :
|
| |
| |
| |
| |
| |
| |
| | |
playbooks/common/openshift-cluster/upgrades/docker/docker_upgrade.yml
playbooks/common/openshift-cluster/upgrades/upgrade_nodes.yml
Signed-off-by: jkaurredhat <jkaur@redhat.com>
|
| | |
|
| | |
|
| | |
|
|/ |
|
|
|
|
|
| |
The playbooks were crossing byo/common boundaries for task includes.
This moves all 'common' files/tasks into the 'common' folder.
|
|\
| |
| | |
Begin requiring Docker 1.12.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Building off the work done for Docker 1.10, we now require Docker 1.12
by default.
The upgrade process was already set to ensure you are running the latest
docker during upgrade, and the standalone docker upgrade playbook can
also be used if desired.
As before, you can override this Docker 1.12 requirement by setting a
docker_version=1.10.3 (or similar), and you can skip the default to
upgrade docker by setting docker_upgrade=False.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
The script nuke_images.sh was introduced in
0c31d72be3bf32f848eedad9859a81ba858f8c8f and seems that the shell
argument $1 was never used (the only $1 in the script back then refers
to a field in a awk script).
There was a reference to $2 that was always undefined/empty.
The script was then simplified in
b89c835e3235f2628b37de15713c311d1b5a4bad, removing any reference to $1
and $2.
This commit cleans up the only call site to the shell script.
|
| | |
|
| |
| |
| |
| |
| | |
* Added checks to make ci for yaml linting
* Modified y(a)ml files to pass lint checks
|
| | |
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
In 3.3 one of our services lays down a systemd drop-in for configuring
Docker networking to use lbr0. In 3.4, this has been changed but the
file must be cleaned up manually by us.
However, after removing the file docker requires a restart. This had big
implications particularly in containerized environments where upgrade is
a very fragile series of upgrading and service restarts.
To avoid double docker restarts, and thus double service restarts in
containerized environments, this change does the following:
- Skip restart during docker upgrade, if it is required. We will restart
on our own later.
- Skip containerized service restarts when we upgrade the services
themselves.
- Clean shutdown of all containerized services.
- Restart Docker. (always, previously this only happened if it needed an
upgrade)
- Ensure all containerized services are restarted.
- Restart rpm node services. (always)
- Mark node schedulable again.
At the end of this process, docker0 should be back on the system.
|
|/
|
|
|
| |
The Ansible package module will call the correct package manager for the
underlying OS.
|
| |
|
| |
|
| |
|
|\
| |
| | |
Allow overriding the Docker 1.10 requirement for upgrade.
|
| |
| |
| |
| | |
Respect an explicit docker_version, and the use of docker_upgrade=False.
|
|/ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
In a parallel step prior to real upgrade tasks, clear out all unused
Docker images on all hosts. This should be relatively safe to interrupt
as no real upgrade steps have taken place.
Once into actual upgrade, we again clear all images only this time with
force, and after stopping and removing all containers.
Both rmi commands use a new and hopefully less error prone command to do
the removal, this should avoid missed orphans as we were hitting before.
Added some logging around the current image count before and after this
step, most of them are only printed if we're crossing the 1.10 boundary
but one does not, just for additional information in your ansible log.
|
|
|
|
|
| |
This avoids the automatic image migration in 1.10, which can take a very
long time and potentially cause rpm db corruption.
|
| |
|
| |
|
| |
|
| |
|
|
We now handle the two pieces of upgrade that require a node evac in the
same play. (docker, and node itself)
|