| Commit message (Collapse) | Author | Age | Files | Lines |
|\
| |
| | |
Add support for Openstack based persistent volumes
|
| |
| |
| |
| | |
Signed-off-by: Sylvain Baubeau <sbaubeau@redhat.com>
|
| | |
|
|\ \
| | |
| | | |
port filter_plugins to ansible2
|
| |/ |
|
|/ |
|
|
|
|
|
|
|
| |
Previously we expected you to comma separate, but this wasn't really
clear. Insted we will use an explicit JSON list going forward.
The comma separated list will be supported for backwards compatability.
|
| |
|
|
|
|
| |
counting nodes.
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Move openshift_router to openshift_hosted role which will eventually
contain registry, metrics and logging.
* Adds option for specifying an openshift_hosted_router_certificate
cert and key pair.
* Removes dependency on node label variables and retrieves the node
list from the API s.t. this role can be applied to any cluster with
existing nodes. I've added an openshift_hosted playbook that occurs
after node install to account for this.
* Infrastructure nodes are selected using
openshift_hosted_router_selector which is based on deployment type
by default; openshift-enterprise -> "region=infra" and online ->
"type=infra".
|
| |
|
| |
|
| |
|
|
|
|
| |
default
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- openshift_master role update
- infra_nodes was previously being set to num_infra, which is an integer
value when using the cloud providers, added a new variable osm_infra_nodes
that is expected to be a list of hosts
- if openshift_infra_nodes is not already set, create it from the nodes that
have the region=infra label.
- Cloud provider config playbook updates
- override openshift_router_selector for cloud providers to avoid using the
default of 'region=infra' when deployment_type is not 'online'
- Set openshift_infra_nodes to g_infra_host for cloud providers
|
| |
|
|
|
|
|
| |
- Fix issue where None is passed to to_padded_yaml filter and invalid config
is generated.
|
| |
|
| |
|
| |
|
|\
| |
| | |
Docker on master aws
|
| | |
|
| | |
|
| | |
|
|/ |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
The list playbook listed the IPs of the VMs without logging their role like:
TASK: [debug ] ************************************************************
ok: [10.64.109.37] => {
"msg": "public:10.64.109.37 private:192.168.165.5"
}
ok: [10.64.109.47] => {
"msg": "public:10.64.109.47 private:192.168.165.6"
}
ok: [10.64.109.36] => {
"msg": "public:10.64.109.36 private:192.168.165.4"
}
ok: [10.64.109.215] => {
"msg": "public:10.64.109.215 private:192.168.165.2"
}
The list playbook now prints the information in a more structured way with
a list of masters, a list of nodes and the subtype of the nodes like:
TASK: [debug ] ************************************************************
ok: [localhost] => {
"msg": {
"lenaicnewlist": {
"master": [
{
"name": "10.64.109.215",
"private IP": "192.168.165.2",
"public IP": "10.64.109.215",
"subtype": "default"
}
],
"node": [
{
"name": "10.64.109.47",
"private IP": "192.168.165.6",
"public IP": "10.64.109.47",
"subtype": "compute"
},
{
"name": "10.64.109.37",
"private IP": "192.168.165.5",
"public IP": "10.64.109.37",
"subtype": "compute"
},
{
"name": "10.64.109.36",
"private IP": "192.168.165.4",
"public IP": "10.64.109.36",
"subtype": "infra"
}
]
}
}
}
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|\
| |
| | |
Make all the OpenStack resources be managed by a Heat Stack
|
| | |
|
|\ \
| |/
|/| |
External clustered etcd support
|
| | |
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
- Add support to bin/cluster for specifying etcd hosts
- defaults to 0, if no etcd hosts are selected, then configures embedded
etcd
- Updates for the byo inventory file for etcd and master as node by default
- Consolidation of cluster logic more centrally into common playbook
- Added etcd config support to playbooks
- Restructured byo playbooks to leverage the common openshift-cluster playbook
- Added support to common master playbook to generate and apply external etcd
client certs from the etcd ca
- start of refactor for better handling of master certs in a multi-master
environment.
- added the openshift_master_ca and openshift_master_certificates roles to
manage master certs instead of generating them in the openshift_master
role
- added etcd host groups to the cluster update playbooks
- aded better handling of host groups when they are either not present or are
empty.
- Update AWS readme
|