| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
|
|
|
| |
- Added installer_checkpoint role and callback plugin
- Added checkpoint 'Start' and 'End' plays to each installation phase
Additional items related to enabling proper checkpointing:
- Removed openshift_set_hostname and related task (related to 3.0)
- Added openshift-hosted entry point playbook
- Moved openshift metrics and logging out of openshift_hosted playbook
- Moved API Aggregation play to master install
|
|
|
|
|
|
|
|
|
|
|
|
| |
Currently, openshift_hosted role duplicates some logic
across separate task chains. This commit cleans up
the openshift_hosted role and converts it to be
primarily used with include_role to give better
logic to the playbooks that utilize this role.
This commit also refactors the playbook that calls
various openshift_hosted roles into individual playbooks.
This allows more granularity for advanced users.
|
|
|
|
| |
single location before they are no longer honored
|
|\
| |
| | |
Merged by openshift-bot
|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Most of this role's purpose was to set facts. The vast majority
of these facts were simply redefining user-supplied variables.
This commit also removes various artifacts leftover from
previous versions, as well as variables that seem to be
entirely unused.
|
|/
|
|
|
|
|
|
|
|
|
|
|
|
| |
A new role for installing prometheus on openshift.
Depends on `openshift_hosted_prometheus_deploy` flag
role creates:
- prometheus namespace
- prometheus clusterrolebinding and service account
- pvs for prometheus, alertmanager and alertbuffer for internal nfs
- prometheus pod with prometheus behind oauth-proxy, alertmanager and alert-buffer behind oauth-proxy
- prometheus and alertmanager configmaps
- prometheus and alerts services and direct routes
- prometheus, alertmanager and alert-buffer pvcs
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
| |
openshift_logging
|
|
|
|
| |
playbooks
|
| |
|
|\
| |
| | |
fixes 1417261. Points playbooks to the correct 3.5 roles for logging …
|
| | |
|
|\| |
|
| | |
|
| | |
|
|/
|
|
| |
openshift_hosted_logging playbook
|
| |
|
| |
|
| |
|
|
|
|
| |
and openshift_hosted.
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Problem was caused by facts not being set for that master. To fix this
patch cleans up the calculation of metricsPublicURL in general. Because
this value is used in openshift_master to template into the master
config file, we now define these facts more clearly in
openshift_master_facts, and add a dependency on this to
openshift_metrics.
The calculation of default sub-domain is also changed to remove it from
system facts (as neither of these are facts about the system) and
instead use plain variables.
|
|
|
|
| |
existing /etc/sysconfig/docker.
|
| |
|
| |
|
|\
| |
| | |
Install Registry by Default
|
| |
| |
| |
| |
| |
| | |
* Convert oc template calls to jsonpath.
* Wait for deployments to finish before restarting docker.
* Re-organize node ca configuration.
|
| |
| |
| |
| |
| |
| |
| | |
Instead of restricting cockpit-ui to Atomic Registry installations,
install it by default everywhere.
Fixes Bug 1371459
|
|/ |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
| |
Metrics deployer now checks for route activation. As such we need a router
before we install metrics.
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* [openshift_projects] Add openshift_projects role
* [openshift_hosted] hosted deployments use openshift_hosted_infra_selector if openshift_hosted_<component>_selector is not defined
* [openshift_hosted] move openshift_projects, openshift_serviceaccounts and openshift_metrics to dependencies of openshift_hosted
* [router] improve router deployment
- add router option to force subdomain
- add CA to router certificate options
* [registry] move registry config into openshift_hosted role
- additional registry fixes/tweaks
- add s3 storage support for registry
* [serviceaccount] fix up serviceaccount creation
|
|\
| |
| | |
wait metrics-deployer complete (need to configure nodes before hosted…
|
| | |
|
|/ |
|
|
* Move openshift_router to openshift_hosted role which will eventually
contain registry, metrics and logging.
* Adds option for specifying an openshift_hosted_router_certificate
cert and key pair.
* Removes dependency on node label variables and retrieves the node
list from the API s.t. this role can be applied to any cluster with
existing nodes. I've added an openshift_hosted playbook that occurs
after node install to account for this.
* Infrastructure nodes are selected using
openshift_hosted_router_selector which is based on deployment type
by default; openshift-enterprise -> "region=infra" and online ->
"type=infra".
|