This project is mirrored from https://gitee.com/wangmingco/rook.git. Pull mirroring failed .
Repository mirroring has been paused due to too many failed attempts. It can be resumed by a project maintainer.
  1. 09 May, 2022 1 commit
  2. 01 Dec, 2020 10 commits
  3. 30 Nov, 2020 2 commits
  4. 27 Nov, 2020 3 commits
  5. 26 Nov, 2020 6 commits
  6. 25 Nov, 2020 6 commits
    • Satoru Takeuchi's avatar
      Merge pull request #6696 from leseb/fix-6685 · d3ebaedc
      Satoru Takeuchi authored
      ceph: fix metadata device passed by-id
      d3ebaedc
    • Sébastien Han's avatar
      Merge pull request #6693 from leseb/abort-reconcile-simple · 78f22fcf
      Sébastien Han authored
      ceph: ability to abort orchestration
      78f22fcf
    • Sébastien Han's avatar
      ceph: ability to abort orchestration · ad249904
      Sébastien Han authored
      We can now prioritize orchestrations on certain event. Today only two
      events will cancel on-going orchestrations (if any):
      
      * request for cluster deletion
      * request for cluster upgrade
      
      If one of the two are caught by the watcher we will cancel the on-going
      orchestration.
      For that we implemented a simple approach based on check points, where
      we will check for a cancellation request in certain part of the
      orchestration. Mainly before each mons/mgr/osds orchestration loops.
      
      This solution is not perfect, but we are waiting for the
      controller-runtime to release its 0.7 version which will embed context
      support. With that we will be able to cancel reconciles more precisely
      and rapidly.
      
      Operator log example:
      
      ```
      2020-11-24 13:54:59.499719 I | op-mon: parsing mon endpoints: a=10.109.126.120:6789
      2020-11-24 13:54:59.499719 I | op-mon: parsing mon endpoints: a=10.109.126.120:6789
      2020-11-25 12:59:12.986264 I | ceph-cluster-controller: done reconciling ceph cluster in namespace "rook-ceph"
      2020-11-25 13:07:33.776947 I | ceph-cluster-controller: CR has changed for "rook-ceph". diff=  v1.ClusterSpec{
              CephVersion: v1.CephVersionSpec{
                      Image:            "ceph/ceph:v15.2.5",
      -               AllowUnsupported: true,
      +               AllowUnsupported: false,
              },
              DriveGroups: nil,
              Storage:     {UseAllNodes: true, Selection: {UseAllDevices: &true}},
              ... // 20 identical fields
        }
      2020-11-25 13:07:33.777039 I | ceph-cluster-controller: reconciling ceph cluster in namespace "rook-ceph"
      2020-11-25 13:07:33.785088 I | op-mon: parsing mon endpoints: a=10.107.242.49:6789,b=10.109.71.30:6789,c=10.98.93.224:6789
      2020-11-25 13:07:33.788626 I | ceph-cluster-controller: detecting the ceph image version for image ceph/ceph:v15.2.5...
      2020-11-25 13:07:35.280789 I | ceph-cluster-controller: detected ceph image version: "15.2.5-0 octopus"
      2020-11-25 13:07:35.280806 I | ceph-cluster-controller: validating ceph version from provided image
      2020-11-25 13:07:35.285888 I | op-mon: parsing mon endpoints: a=10.107.242.49:6789,b=10.109.71.30:6789,c=10.98.93.224:6789
      2020-11-25 13:07:35.287828 I | cephclient: writing config file /var/lib/rook/rook-ceph/rook-ceph.config
      2020-11-25 13:07:35.288082 I | cephclient: generated admin config in /var/lib/rook/rook-ceph
      2020-11-25 13:07:35.621625 I | ceph-cluster-controller: cluster "rook-ceph": version "15.2.5-0 octopus" detected for image "ceph/ceph:v15.2.5"
      2020-11-25 13:07:35.642688 I | op-mon: start running mons
      2020-11-25 13:07:35.646323 I | op-mon: parsing mon endpoints: a=10.107.242.49:6789,b=10.109.71.30:6789,c=10.98.93.224:6789
      2020-11-25 13:07:35.654070 I | op-mon: saved mon endpoints to config map map[csi-cluster-config-json:[{"clusterID":"rook-ceph","monitors":["10.107.242.49:6789","10.109.71.30:6789","10.98.93.224:6789"]}] data:a=10.107.242.49:6789,b=10.109.71.30:6789,c=10.98.93.224:6789 mapping:{"node":{"a":{"Name":"minikube","Hostname":"minikube","Address":"192.168.39.3"},"b":{"Name":"minikube","Hostname":"minikube","Address":"192.168.39.3"},"c":{"Name":"minikube","Hostname":"minikube","Address":"192.168.39.3"}}} maxMonId:2]
      2020-11-25 13:07:35.868253 I | cephclient: writing config file /var/lib/rook/rook-ceph/rook-ceph.config
      2020-11-25 13:07:35.868573 I | cephclient: generated admin config in /var/lib/rook/rook-ceph
      2020-11-25 13:07:37.074353 I | op-mon: targeting the mon count 3
      2020-11-25 13:07:38.153435 I | op-mon: checking for basic quorum with existing mons
      2020-11-25 13:07:38.178029 I | op-mon: mon "a" endpoint is [v2:10.107.242.49:3300,v1:10.107.242.49:6789]
      2020-11-25 13:07:38.670191 I | op-mon: mon "b" endpoint is [v2:10.109.71.30:3300,v1:10.109.71.30:6789]
      2020-11-25 13:07:39.477820 I | op-mon: mon "c" endpoint is [v2:10.98.93.224:3300,v1:10.98.93.224:6789]
      2020-11-25 13:07:39.874094 I | op-mon: saved mon endpoints to config map map[csi-cluster-config-json:[{"clusterID":"rook-ceph","monitors":["10.107.242.49:6789","10.109.71.30:6789","10.98.93.224:6789"]}] data:a=10.107.242.49:6789,b=10.109.71.30:6789,c=10.98.93.224:6789 mapping:{"node":{"a":{"Name":"minikube","Hostname":"minikube","Address":"192.168.39.3"},"b":{"Name":"minikube","Hostname":"minikube","Address":"192.168.39.3"},"c":{"Name":"minikube","Hostname":"minikube","Address":"192.168.39.3"}}} maxMonId:2]
      2020-11-25 13:07:40.467999 I | cephclient: writing config file /var/lib/rook/rook-ceph/rook-ceph.config
      2020-11-25 13:07:40.469733 I | cephclient: generated admin config in /var/lib/rook/rook-ceph
      2020-11-25 13:07:41.071710 I | cephclient: writing config file /var/lib/rook/rook-ceph/rook-ceph.config
      2020-11-25 13:07:41.078903 I | cephclient: generated admin config in /var/lib/rook/rook-ceph
      2020-11-25 13:07:41.125233 I | op-mon: deployment for mon rook-ceph-mon-a already exists. updating if needed
      2020-11-25 13:07:41.327778 I | op-k8sutil: updating deployment "rook-ceph-mon-a" after verifying it is safe to stop
      2020-11-25 13:07:41.327895 I | op-mon: checking if we can stop the deployment rook-ceph-mon-a
      2020-11-25 13:07:44.045644 I | op-k8sutil: finished waiting for updated deployment "rook-ceph-mon-a"
      2020-11-25 13:07:44.045706 I | op-mon: checking if we can continue the deployment rook-ceph-mon-a
      2020-11-25 13:07:44.045740 I | op-mon: waiting for mon quorum with [a b c]
      2020-11-25 13:07:44.109159 I | op-mon: mons running: [a b c]
      2020-11-25 13:07:44.474596 I | op-mon: Monitors in quorum: [a b c]
      2020-11-25 13:07:44.478565 I | op-mon: deployment for mon rook-ceph-mon-b already exists. updating if needed
      2020-11-25 13:07:44.493374 I | op-k8sutil: updating deployment "rook-ceph-mon-b" after verifying it is safe to stop
      2020-11-25 13:07:44.493403 I | op-mon: checking if we can stop the deployment rook-ceph-mon-b
      2020-11-25 13:07:47.135524 I | op-k8sutil: finished waiting for updated deployment "rook-ceph-mon-b"
      2020-11-25 13:07:47.135542 I | op-mon: checking if we can continue the deployment rook-ceph-mon-b
      2020-11-25 13:07:47.135551 I | op-mon: waiting for mon quorum with [a b c]
      2020-11-25 13:07:47.148820 I | op-mon: mons running: [a b c]
      2020-11-25 13:07:47.445946 I | op-mon: Monitors in quorum: [a b c]
      2020-11-25 13:07:47.448991 I | op-mon: deployment for mon rook-ceph-mon-c already exists. updating if needed
      2020-11-25 13:07:47.462041 I | op-k8sutil: updating deployment "rook-ceph-mon-c" after verifying it is safe to stop
      2020-11-25 13:07:47.462060 I | op-mon: checking if we can stop the deployment rook-ceph-mon-c
      2020-11-25 13:07:48.853118 I | ceph-cluster-controller: CR has changed for "rook-ceph". diff=  v1.ClusterSpec{
              CephVersion: v1.CephVersionSpec{
      -               Image:            "ceph/ceph:v15.2.5",
      +               Image:            "ceph/ceph:v15.2.6",
                      AllowUnsupported: false,
              },
              DriveGroups: nil,
              Storage:     {UseAllNodes: true, Selection: {UseAllDevices: &true}},
              ... // 20 identical fields
        }
      2020-11-25 13:07:48.853140 I | ceph-cluster-controller: upgrade requested, cancelling any ongoing orchestration
      2020-11-25 13:07:50.119584 I | op-k8sutil: finished waiting for updated deployment "rook-ceph-mon-c"
      2020-11-25 13:07:50.119606 I | op-mon: checking if we can continue the deployment rook-ceph-mon-c
      2020-11-25 13:07:50.119619 I | op-mon: waiting for mon quorum with [a b c]
      2020-11-25 13:07:50.130860 I | op-mon: mons running: [a b c]
      2020-11-25 13:07:50.431341 I | op-mon: Monitors in quorum: [a b c]
      2020-11-25 13:07:50.431361 I | op-mon: mons created: 3
      2020-11-25 13:07:50.734156 I | op-mon: waiting for mon quorum with [a b c]
      2020-11-25 13:07:50.745763 I | op-mon: mons running: [a b c]
      2020-11-25 13:07:51.045108 I | op-mon: Monitors in quorum: [a b c]
      2020-11-25 13:07:51.054497 E | ceph-cluster-controller: failed to reconcile. failed to reconcile cluster "rook-ceph": failed to configure local ceph cluster: failed to create cluster: CANCELLING CURRENT ORCHESTATION
      2020-11-25 13:07:52.055208 I | ceph-cluster-controller: reconciling ceph cluster in namespace "rook-ceph"
      2020-11-25 13:07:52.070690 I | op-mon: parsing mon endpoints: a=10.107.242.49:6789,b=10.109.71.30:6789,c=10.98.93.224:6789
      2020-11-25 13:07:52.088979 I | ceph-cluster-controller: detecting the ceph image version for image ceph/ceph:v15.2.6...
      2020-11-25 13:07:53.904811 I | ceph-cluster-controller: detected ceph image version: "15.2.6-0 octopus"
      2020-11-25 13:07:53.904862 I | ceph-cluster-controller: validating ceph version from provided image
      ```
      
      Closes: https://github.com/rook/rook/issues/6587
      
      Signed-off-by: default avatarSébastien Han <seb@redhat.com>
      ad249904
    • Sébastien Han's avatar
      ceph: fix metadata device passed by-id · 5d9612c2
      Sébastien Han authored
      The code was assuming that devices were passed by the user as
      "/dev/sda", this is bad! We all know people should be using paths like
      /dev/disk/by-id so we must support them.
      
      Closes: https://github.com/rook/rook/issues/6685
      
      Signed-off-by: default avatarSébastien Han <seb@redhat.com>
      5d9612c2
    • Alexander Trost's avatar
      b04f8823
    • Sébastien Han's avatar
      Merge pull request #6417 from renan-campos/issue_6332 · 4d8af284
      Sébastien Han authored
      ceph: periodically prune crash entries older than user-provided days
      4d8af284
  7. 24 Nov, 2020 6 commits
  8. 19 Nov, 2020 6 commits