summaryrefslogtreecommitdiffstats
path: root/cts/scheduler/xml/stop-all-resources.xml
blob: 6ecd4d6d7342ecd83a68ea26faf71cc26ea0a144 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
<cib crm_feature_set="3.3.0" validate-with="pacemaker-3.3" epoch="1" num_updates="1" admin_epoch="1" cib-last-written="Tue May  5 12:04:36 2020" update-origin="cluster01" update-client="crmd" update-user="hacluster" have-quorum="1" dc-uuid="2">
  <configuration>
    <crm_config>
      <cluster_property_set id="cib-bootstrap-options">
        <nvpair id="cib-bootstrap-options-have-watchdog" name="have-watchdog" value="false"/>
        <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="2.0.4-1.e97f9675f.git.el7-e97f9675f"/>
        <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/>
        <nvpair id="cib-bootstrap-options-cluster-name" name="cluster-name" value="test-cluster"/>
        <nvpair id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="true"/>
        <nvpair id="cib-bootstrap-options-maintenance-mode" name="maintenance-mode" value="false"/>
        <nvpair id="cib-bootstrap-options-stop-all-resources" name="stop-all-resources" value="true"/>
      </cluster_property_set>
    </crm_config>
    <nodes>
      <node id="1" uname="cluster01"/>
      <node id="2" uname="cluster02"/>
    </nodes>
    <resources>
      <clone id="ping-clone">
        <primitive class="ocf" id="ping" provider="pacemaker" type="ping">
          <instance_attributes id="ping-instance_attributes">
            <nvpair id="ping-instance_attributes-dampen" name="dampen" value="5s"/>
            <nvpair id="ping-instance_attributes-host_list" name="host_list" value="192.168.122.1"/>
            <nvpair id="ping-instance_attributes-multiplier" name="multiplier" value="1000"/>
          </instance_attributes>
          <operations>
            <op id="ping-monitor-interval-10s" interval="10s" name="monitor" timeout="60s"/>
            <op id="ping-start-interval-0s" interval="0s" name="start" timeout="60s"/>
            <op id="ping-stop-interval-0s" interval="0s" name="stop" timeout="20s"/>
          </operations>
        </primitive>
      </clone>
      <primitive class="stonith" id="Fencing" type="fence_xvm">
        <instance_attributes id="Fencing-instance_attributes">
          <nvpair id="Fencing-instance_attributes-ip_family" name="ip_family" value="ipv4"/>
        </instance_attributes>
        <operations>
          <op id="Fencing-monitor-interval-60s" interval="60s" name="monitor"/>
        </operations>
      </primitive>
      <primitive class="ocf" id="dummy" provider="pacemaker" type="Dummy">
        <instance_attributes id="dummy-instance_attributes">
          <nvpair id="dummy-instance_attributes-op_sleep" name="op_sleep" value="6"/>
        </instance_attributes>
        <operations>
          <op id="dummy-migrate_from-interval-0s" interval="0s" name="migrate_from" timeout="20s"/>
          <op id="dummy-migrate_to-interval-0s" interval="0s" name="migrate_to" timeout="20s"/>
          <op id="dummy-monitor-interval-60s" interval="60s" name="monitor" on-fail="stop"/>
          <op id="dummy-reload-interval-0s" interval="0s" name="reload" timeout="20s"/>
          <op id="dummy-start-interval-0s" interval="0s" name="start" timeout="20s"/>
          <op id="dummy-stop-interval-0s" interval="0s" name="stop" timeout="20s"/>
        </operations>
      </primitive>
      <clone id="inactive-clone">
        <meta_attributes id="inactive-clone-meta_attributes">
          <nvpair id="inactive-clone-meta_attributes-target-role" name="target-role" value="stopped"/>
        </meta_attributes>
        <primitive id="inactive-dhcpd" class="lsb" type="dhcpd"/>
      </clone>
      <group id="inactive-group">
        <meta_attributes id="inactive-group-meta_attributes">
          <nvpair id="inactive-group-meta_attributes-target-role" name="target-role" value="stopped"/>
        </meta_attributes>
        <primitive class="ocf" id="inactive-dummy-1" provider="pacemaker" type="Dummy"/>
        <primitive class="ocf" id="inactive-dummy-2" provider="pacemaker" type="Dummy"/>
      </group>
      <bundle id="httpd-bundle">
        <docker image="pcmk:http" replicas="3"/>
        <network ip-range-start="192.168.122.131" host-netmask="24" host-interface="eth0">
          <port-mapping id="httpd-port" port="80"/>
        </network>
        <storage>
          <storage-mapping id="httpd-syslog" source-dir="/dev/log" target-dir="/dev/log" options="rw"/>
          <storage-mapping id="httpd-root" source-dir="/srv/html" target-dir="/var/www/html" options="rw"/>
          <storage-mapping id="httpd-logs" source-dir-root="/var/log/pacemaker/bundles" target-dir="/etc/httpd/logs" options="rw"/>
        </storage>
        <primitive class="ocf" id="httpd" provider="heartbeat" type="apache"/>
        <meta_attributes id="bundle-meta_attributes">
          <nvpair id="bundle-meta_attributes-target-role" name="target-role" value="Started"/>
        </meta_attributes>
      </bundle>
      <group id="exim-group">
        <primitive id="Public-IP" class="ocf" type="IPaddr" provider="heartbeat">
          <instance_attributes id="params-public-ip">
            <nvpair id="public-ip-addr" name="ip" value="192.168.1.1"/>
          </instance_attributes>
        </primitive>
        <primitive id="Email" class="lsb" type="exim"/>
      </group>
      <clone id="mysql-clone-group">
        <group id="mysql-group">
          <primitive id="mysql-proxy" class="lsb" type="mysql-proxy">
            <operations>
              <op name="monitor" interval="10s" id="mysql-proxy_mon" timeout="20s"/>
            </operations>
          </primitive>
        </group>
      </clone>
    </resources>
    <constraints/>
    <op_defaults/>
  </configuration>
  <status>
    <node_state id="2" uname="cluster02" in_ccm="true" crmd="online" crm-debug-origin="do_update_resource" join="member" expected="member"/>
    <node_state id="1" uname="cluster01" in_ccm="true" crmd="online" crm-debug-origin="do_update_resource" join="member" expected="member"/>
  </status>
</cib>