This project is mirrored from https://gitee.com/wangmingco/rook.git.
Pull mirroring failed .
Repository mirroring has been paused due to too many failed attempts. It can be resumed by a project maintainer.
Repository mirroring has been paused due to too many failed attempts. It can be resumed by a project maintainer.
- 23 Oct, 2019 13 commits
-
-
Travis Nielsen authored
Set the v1.1.3 release tag
-
Travis Nielsen authored
With the v1.1.3 patch release we need to set the new tag in the deployment manifests and documentation. Signed-off-by:
Travis Nielsen <tnielsen@redhat.com>
-
mergify[bot] authored
ceph: osd remove --conf flag (bp #4161)
-
mergify[bot] authored
fix operator reconcile to restart osd daemons (bp #3996)
-
mergify[bot] authored
Enable restoring a cluster after disaster recovery (bp #4021)
-
Sébastien Han authored
Even legacy osds were still running with this flag. In my previous patch I was trying to minimize the impact of the change and was afraid the config file was holding important information but that appears not to be true. The only option in the config could have the keyring path, but it's passed in the CLI startup line. Closes: https://github.com/rook/rook/issues/4063 Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit 9f053ca4)
-
mergify[bot] authored
Ceph: Added removeOSDsIfOutAndSafeToRemove to Cluster CR (bp #4116)
-
travisn authored
If mon deployments need to be created, the operator should go ahead and create all of them instead of waiting for quorum in between checking each mon. In particular, if a cluster is being restored in a disaster recovery situation, none of the mons deployments will exist. All of them should be created before checking for quorum. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit 53812c00)
-
travisn authored
Requests to enable mgr modules or call mgr modules should timeout rather than hang. For example, if creating a self signed cert request hangs, the mgr should continue with other actions in the orchestration Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit a72a9816)
-
travisn authored
Checking for clean PGs is a recurring event when reconciling nodes for upgrades and fencing. No need to log every request. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit 43979747)
-
Santosh Pillai authored
OSD on PVC does not upgrade when the user upgrades the ceph version on cluster-on-pvc yaml. The solution incudes: - upgrade osd prepare and daemon pods on upgrade - skip c-v prepare if filesystem is already present on the pvc device. - skip lvm release in case of upgrade. Signed-off-by:
Santosh Pillai <sapillai@redhat.com> (cherry picked from commit 8ea693a7)
-
rohan47 authored
OSDs can be removed automatically with the current mechanism if a new setting removeOSDsIfOutAndSafeToRemove is set to true. The default for all new or upgraded clusters should be false. Signed-off-by:
rohan47 <rohgupta@redhat.com> (cherry picked from commit 7f9611d4)
-
Travis Nielsen authored
ceph: rework csi keys and secrets (bp #4086)
-
- 22 Oct, 2019 4 commits
-
-
mergify[bot] authored
[OSD on PVC] add kubernetes version check. (bp #4009)
-
Sébastien Han authored
We know create 4 secrets containing 4 ceph keys where each of them have limited permissions to access the cluster. Closes: https://github.com/rook/rook/issues/4074 Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit fcc2e21f)
-
Sébastien Han authored
The command `fs new` already enable the application pool settings for the cephfs pools so we don't need to run that command again. Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit eafc2a47)
-
Sébastien Han authored
When we give up on waiting for the pod to be running which describe it to see what's wrong. Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit eccea9fa)
-
- 21 Oct, 2019 3 commits
-
-
Sébastien Han authored
Backport: ceph: expose prepare pod resource limits
-
Sébastien Han authored
The osd-prepare jobs should not inherit osd resources since this might cause scheduling issues. Instead, we can now use the new CRD property `prepareosd` in the resource section. Refer to the documentation for more details. Closes: https://github.com/rook/rook/issues/2502 Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit 1765ad65)
-
Santosh Pillai authored
- OSD on PVC provisioning fails if k8s version is less than 1.13.0 - updated document to reflect minimum k8s version required. - Added check to skip OSD on pvc provisoning if the minimum k8s version requirement is not met. Signed-off-by:
Santosh Pillai <sapillai@redhat.com> (cherry picked from commit edf0c7ae)
-
- 18 Oct, 2019 1 commit
-
-
mergify[bot] authored
Issue 4073 configure rgw non ec pool as metadata (bp #4087)
-
- 17 Oct, 2019 1 commit
-
-
Owen Tuz authored
This is used for S3 multipart uploads and should be configured in the same way as other metadata pools. Signed-off-by:
Owen Tuz <owen@segfault.re> (cherry picked from commit 15e3ecad)
-
- 16 Oct, 2019 16 commits
-
-
mergify[bot] authored
ceph: hide wrong error for clusterdisruption controller (bp #4094)
-
mergify[bot] authored
Multiple integration test fixes to improve CI stability (bp #4098)
-
mergify[bot] authored
ceph: detect mount fstype more accurately (bp #4109)
-
mergify[bot] authored
ceph: mgr do not override annotation (bp #4110)
-
Sébastien Han authored
Using `df --type ceph` has proven not to be reliable enough, so let's try with findmnt. Closes: https://github.com/rook/rook/issues/4107 Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit 9cfdc4f5)
-
Sébastien Han authored
The current implementation was overriding any previous annotations set on the object meta. Moving the logic to its own method as well as adding unit tests. Closes: https://github.com/rook/rook/issues/4106 Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit f6a5caa1)
-
travisn authored
The timeout of 80 seconds wasn't always sufficient for mon failover. The timeout is now increased to 150 seconds. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit 0cca2cde)
-
travisn authored
The two clusters were being created in parallel by the tests and sometimes causing the test to timeout. The operator handles the cluster serially so we don't gain anything by starting them asynchronously. With this change the tests will be started serially to match the behavior of the operator and avoid the issues with test timeouts. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit 378f209d)
-
travisn authored
The logs are only collected by default if a test failed. Ensure that we collect the logs if the test setup step failed. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit 98c21950)
-
travisn authored
The BlockCreateSuite and BlockMountUnmountSuite suites are almost identical for running block tests with different types of mounts. We can consolidate these into a single test suite and eliminate a couple of the tests that are already covered by other tests. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit a178cccd)
-
travisn authored
Testing if the create does any better in the tests than apply. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit 979b6fd2)
-
travisn authored
Adding more logging until we can track down the file-test pod cleanup issue Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit f5eb11d8)
-
Sébastien Han authored
The clusterdisruption starts up before the Ceph cluster is created and during the initilization, it'll display invalid error messages such as: 2019-10-14T04:22:21.243679454Z 2019-10-14 04:22:21.243668 E | clusterdisruption-controller: could not check cluster health: failed to get status: exit status 1 2019-10-14T04:22:22.524222494Z 2019-10-14 04:22:22.524179 I | exec: Running command: ceph status --connect-timeout=15 --cluster=openshift-storage --conf=/var/lib/rook/openshift-storage/openshift-storage.config --keyring=/var/lib/rook/openshift-storage/client.admin.keyring --format json --out-file /tmp/605071670 2019-10-14T04:22:22.617370007Z 2019-10-14 04:22:22.617321 I | exec: Error initializing cluster client: ObjectNotFound('error calling conf_read_file',) That error message is invalid since the cluster is not up and running YET. So let's simply log a message. Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit 05140d7a)
-
mergify[bot] authored
ceph: fix topologyAware support (bp #4099)
-
travisn authored
The unit tests in the backport require a change that was not required in master. Signed-off-by:
travisn <tnielsen@redhat.com>
-
Mateusz Los authored
add missing topologyAware and location env variables to OSD pods Signed-off-by:
Mateusz Los <los.mateusz@gmail.com> (cherry picked from commit cb52aea9)
-
- 15 Oct, 2019 1 commit
-
-
mergify[bot] authored
OwnerReference and test pod cleanup (bp #4090)
-
- 14 Oct, 2019 1 commit
-
-
travisn authored
The cluster finalizer in some scenarios was not being removed during cluster removal. Specifically, if the cluster CR had been modified, the operator would always fail to remove the finalizer. This could occur when the CR status is updated around the same time that the cluster is deleted. Therefore, we need to remove the finalizer from a freshly retrieved instance of the cluster. Signed-off-by:
travisn <tnielsen@redhat.com> (cherry picked from commit bc684400)
-