| Commit message (Collapse) | Author | Age | Files | Lines |
|\
| |
| | |
Bug 1511357: Task "Reload generated facts" failed while scaling up containerized masters
|
| | |
|
|\ \
| | |
| | | |
Add new grafana playbook.
|
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
this patch introduce a new playbook and roles to deploy grafana automaticly
using openshift-ansible capabilities.
this patch will abstract the grafana deployment process specially for openshift.
|
|\ \ \
| | | |
| | | | |
openshift_checks: repair adhoc list-checks mode
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
fixes bug https://bugzilla.redhat.com/show_bug.cgi?id=1509157
docker_image_availability.__init__ was loading a variable that of course
was not defined for localhost, which is how this play runs. Fixed that.
Would be nice if there were a more generic way to solve that class of
problem but I can't think of anything for now that isn't worse.
Also noticed that this play gathers facts for localhost. That's not
necessary for what it's doing, so disabled that.
|
|\ \ \ \
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | | |
Automatic merge from submit-queue.
Check that rc != 0 for steps with 'failed_when'
Some parts don't check the result rc at all, some check that 'error' is present there.
Instead the safest way is to check for both
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1534538
|
| | | | | |
|
| | | | |
| | | | |
| | | | |
| | | | | |
This might mask some failures and continue instead of a failure
|
|\ \ \ \ \
| |_|_|_|/
|/| | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | | |
juanvallejo/jvallejo/fix-elasticsearch-diskspace-check
Automatic merge from submit-queue.
include elasticsearch container name
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1534988
Addresses cases where the elasticsearch pod with multiple containers does not have the `elasticsearch` container as its first one.
cc @sosiouxme
|
| | | | | |
|
|\ \ \ \ \
| | | | | |
| | | | | | |
Faciliate 3.7 to 3.8 to 3.9 upgrade
|
| | | | | | |
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
There are some obsoletes in 3.9 packages which lead yum to upgrade
directly from 3.7 to 3.9 when you ask for 3.8 packages while 3.9 repos
are enabled. Since we'd like to allow people to run one playbook to
upgrade from 3.7 to 3.8 to 3.9 we need to exclude those packages when
upgrading to 3.8
|
| | | | | | |
|
|\ \ \ \ \ \
| | | | | | |
| | | | | | | |
Don't use from ansible.module_utils.six as its no longer available in Ansible 2.4
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
Ansible 2.4
|
|\ \ \ \ \ \ \
| |_|_|_|/ / /
|/| | | | | | |
Add missing task that got dropped in a refactor
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
Removed in 28b254886bf17251e973bfdad4dd4c89e26187fc
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | | |
Update console config for API changes
|
| | |_|/ / / /
| |/| | | | | |
|
|\ \ \ \ \ \ \
| |/ / / / / /
|/| | | | | |
| | | | | | |
| | | | | | | |
Automatic merge from submit-queue.
Use master group ansible_ssh_user to pull bootstrap kubeconfig
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
pull bootstrap kubeconfig.
|
|\ \ \ \ \ \ \
| |_|_|_|/ / /
|/| | | | | | |
Web Console: use a different var for asset config
|
| | |_|_|_|/
| |/| | | |
| | | | | |
| | | | | |
| | | | | | |
This ensures mktemp from metrics role doesn't get reset when
update_asset_config is being included
|
|\ \ \ \ \ \
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
Automatic merge from submit-queue.
Add console RBAC template
Required for https://github.com/openshift/origin-web-console-server/pull/18
Origin changes: https://github.com/openshift/origin/pull/18075
/assign @sdodson
/cc @deads2k
|
| | | | | | | |
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | | |
Adding ability for ASB to interact with network resources.
|
| | | | | | | | |
|
| | | | | | | | |
|
|\ \ \ \ \ \ \ \
| |_|_|/ / / / /
|/| | | | | | | |
Remove tuned-profiles from list of master packages upgraded
|
| | |_|_|_|_|/
| |/| | | | | |
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
Automatic merge from submit-queue.
Update to AWS EC2 root vol size so that Health Check tasks pass
AWS EC2s under the ASG and LaunchConfig are deployed with the default 10gb root vol size. Heath Checks fail because / filesystem and /var directory is too small. Health Check wants a minimum of 40Gb. 10Gb is just too small for server operations in general. Let's increase to 100Gb. Cost increase for additional EBS storage will be minimal and give plenty of elbow room for normal server operations.
|
| | | | | | | | |
|
|\ \ \ \ \ \ \ \
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
Automatic merge from submit-queue.
Separate ELB & S3 from master node group provisioning
Moved s3 and elb provisioning out of `roles/openshift_aws/tasks/provisioning.yml` and into their own playbooks. These playbooks are now included in the provision playbook and in a WIP infrastructure playbook we intend to run up front in place of prerequisites.
@kwoodson what are your thoughts on something like this?
|
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
outside of the openshift_aws master provisioning tasks.
|
|\ \ \ \ \ \ \ \ \
| |_|_|/ / / / / /
|/| | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
Automatic merge from submit-queue.
Configure Kuryr CNI daemon
Kuryr CNI daemon is scalability improvement that moves watching K8s API
and VIF plugging into a separate entity called kuryr-daemon.
Kuryr-daemon will run in a container and serve requests from
kuryr-driver.
|
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
Kuryr CNI daemon is scalability improvement that moves watching K8s API
and VIF plugging into a separate entity called kuryr-daemon.
Kuryr-daemon will run in a container and serve requests from
kuryr-driver.
|
|\ \ \ \ \ \ \ \ \
| | | | | | | | | |
| | | | | | | | | | |
Clean up host-local IPAM data while nodes are drained
|
| | | | | | | | | | |
|
|\ \ \ \ \ \ \ \ \ \
| |_|_|_|_|_|_|_|_|/
|/| | | | | | | | | |
Add var for catalog controller to enable async bindings
|
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | | |
New variable is called
'openshift_service_catalog_async_bindings_enabled' and is set to false
by default.
Also, fixed a typo in the controller template for the DaemonSet update
strategy, which caused the strategy to be set to OnDelete rather than
the intended RollingUpdate.
|
|\ \ \ \ \ \ \ \ \ \
| | | | | | | | | | |
| | | | | | | | | | | |
Update latest image streams and templates
|
| | |_|_|_|_|_|_|_|/
| |/| | | | | | | | |
|
|\ \ \ \ \ \ \ \ \ \
| |_|_|_|_|_|_|_|/ /
|/| | | | | | | | | |
Move more plugins to lib_utils
|
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | |
| | | | | | | | | | |
This commit continues moving plugins into lib_utils.
This commit does not move any plugins for add-on roles
such as logging and metrics.
|
|\ \ \ \ \ \ \ \ \ \
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | | |
Automatic merge from submit-queue.
Fix Cinder Persistent Volume support
This documents how to use Cinder-backed persistent volumes with OpenStack.
It needed a change to the dynamic inventory because the "openstack" cloudprovider plugin does actually require internal name resolution -- and the `openshift_hostname` value must match the name of the Nova server.
In addition, we need to be able to specify the V2 of the Cinder API for now as described in: https://github.com/openshift/openshift-docs/issues/5730
|
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | | |
When deploying on OpenStack with internal DNS configured, this will set
`openshift_hostname` to the Nova server name instead of its IP address.
Without those two matching, the OpenStack cloud provider configuration
will fail and the OpenShift nodes will not start.
|
| | | | | | | | | | | |
|
|\ \ \ \ \ \ \ \ \ \ \
| |_|_|_|/ / / / / / /
|/| | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | |
| | | | | | | | | | | |
Automatic merge from submit-queue.
Update deployment and apiserver with new certs
Since new certificates are generated for every run, the apiservice
caBundle needs updating in order to have the on disk CA match what is in
Kubernetes.
Because the secrets are updated, the daemonset needs to do a rolling
update for the apiserver to pick up the new certs. Implemented here is
an added annotation to the api server such that the update occurs
automatically when the CA is changed.
---
There may be a better way to make the rolling update occur without adding an annotation, such as within ansible itself (I just didn't know how to do that). Also, I think that probably the controller needs to be updated too in order to ensure staying in sync with the api server.
Edit: I did not think that doing "oc apply" would cause the daemonset to pick up a change since the yaml may be exactly the same, but if it does then the annotation part can definitely be changed.
|