| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
|
| |
Sometimes the node container is not started on a container-engine
restart. Use a weaker dependency on openvswitch that is causing this issue
Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1451192
Signed-off-by: Giuseppe Scrivano <gscrivan@redhat.com>
|
|
|
|
|
|
|
| |
- If using a system container: container-engine
- If using a package install: docker
Ref: https://bugzilla.redhat.com/show_bug.cgi?id=1448800
|
|
|
|
| |
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1427807
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
around across node restart
With the move to a CNI plugin, docker no longer handles IPAM, but CNI does through
openshift-sdn's usage of the 'host-local' CNI IPAM plugin. That plugin stores
IPAM allocations under /var/lib/cni/.
If the node container gets restarted, without presreving /var/lib/cni, the IPs
currently allocated to running pods get lost and on restart, openshift-sdn
may allocate those IPs to new pods causing duplicate allocations.
This never happened with docker because it has its own persistent IPAM store that
does not get removed when docker restarts. Also because (historically) when docker
restarted, all the containers died and the IP allocations were released by the
daemon.
Fix this by ensuring that IPAM allocations (which are tied to the life of the pod,
*not* the life of the openshift-node process) persist even if the openshift-node
process restarts.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1427789
|
| |
|
| |
|
|
|
|
| |
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1367937
|
| |
|
|
|
|
|
|
|
|
|
|
| |
* Conditionally bind mount /usr/bin/docker-current when it is present
* fix upgrade
* add atomic-openshift-node-dep files to uninstall playbook
* Fix variable expansion
|
|
|
|
|
| |
atomic-openshift-node.service file
This fixes https://bugzilla.redhat.com/show_bug.cgi?id=1336780
|
|
|
|
| |
Lifted from https://github.com/openshift/origin/pull/8317
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
There are points where the docker service is restarted during
the install. Sometimes the services that are PartOf docker.service
do not get restarted when docker is restarted.
https://bugzilla.redhat.com/show_bug.cgi?id=1318948
Systemd documentation recommends using "wants" as the perferred
means of creating startup linkages between units. This
patch makes the ose services wanted by the docker service rather
than multi-user.target.
This creates a downward link from the docker service to the ose
containerized services.
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
copied from https://github.com/eparis/kubernetes-ansible/blob/17f98edd7ff53e649b43e26822b8fbc0be42b233/roles/common/tasks/main.yml
|