This project is mirrored from https://gitee.com/mirrors/nomad.git. Pull mirroring failed .
Repository mirroring has been paused due to too many failed attempts. It can be resumed by a project maintainer.
  1. 31 Jan, 2022 1 commit
  2. 28 Jan, 2022 9 commits
    • Tim Gross's avatar
      docs: add 1.2.5 to changelog · f0e1938d
      Tim Gross authored
      f0e1938d
    • Tim Gross's avatar
      docs: missing changelog for #11892 (#11959) · 57517b16
      Tim Gross authored
      57517b16
    • Tim Gross's avatar
      00df7dda
    • Tim Gross's avatar
      CSI: node unmount from the client before unpublish RPC (#11892) · 707b4b3e
      Tim Gross authored
      When an allocation stops, the `csi_hook` makes an unpublish RPC to the
      servers to unpublish via the CSI RPCs: first to the node plugins and
      then the controller plugins. The controller RPCs must happen after the
      node RPCs so that the node has had a chance to unmount the volume
      before the controller tries to detach the associated device.
      
      But the client has local access to the node plugins and can
      independently determine if it's safe to send unpublish RPC to those
      plugins. This will allow the server to treat the node plugin as
      abandoned if a client is disconnected and `stop_on_client_disconnect`
      is set. This will let the server try to send unpublish RPCs to the
      controller plugins, under the assumption that the client will be
      trying to unmount the volume on its end first.
      
      Note that the CSI `NodeUnpublishVolume`/`NodeUnstageVolume` RPCs can 
      return ignorable errors in the case where the volume has already been
      unmounted from the node. Handle all other errors by retrying until we
      get success so as to give operators the opportunity to reschedule a
      failed node plugin (ex. in the case where they accidentally drained a
      node without `-ignore-system`). Fan-out the work for each volume into
      its own goroutine so that we can release a subset of volumes if only
      one is stuck.
      707b4b3e
    • Tim Gross's avatar
      CSI: move terminal alloc handling into denormalization (#11931) · 593b8db9
      Tim Gross authored
      * The volume claim GC method and volumewatcher both have logic
      collecting terminal allocations that duplicates most of the logic
      that's now in the state store's `CSIVolumeDenormalize` method. Copy
      this logic into the state store so that all code paths have the same
      view of the past claims.
      * Remove logic in the volume claim GC that now lives in the state
      store's `CSIVolumeDenormalize` method.
      * Remove logic in the volumewatcher that now lives in the state
      store's `CSIVolumeDenormalize` method.
      * Remove logic in the node unpublish RPC that now lives in the state
      store's `CSIVolumeDenormalize` method.
      593b8db9
    • Tim Gross's avatar
      csi: ensure that PastClaims are populated with correct mode (#11932) · 7181e965
      Tim Gross authored
      In the client's `(*csiHook) Postrun()` method, we make an unpublish
      RPC that includes a claim in the `CSIVolumeClaimStateUnpublishing`
      state and using the mode from the client. But then in the
      `(*CSIVolume) Unpublish` RPC handler, we query the volume from the
      state store (because we only get an ID from the client). And when we
      make the client RPC for the node unpublish step, we use the _current
      volume's_ view of the mode. If the volume's mode has been changed
      before the old allocations can have their claims released, then we end
      up making a CSI RPC that will never succeed.
      
      Why does this code path get the mode from the volume and not the
      claim? Because the claim written by the GC job in `(*CoreScheduler)
      csiVolumeClaimGC` doesn't have a mode. Instead it just writes a claim
      in the unpublishing state to ensure the volumewatcher detects a "past
      claim" change and reaps all the claims on the volumes.
      
      Fix this by ensuring that...
      7181e965
    • Tim Gross's avatar
      CSI: resolve invalid claim states (#11890) · cdbb2bcf
      Tim Gross authored
      * csi: resolve invalid claim states on read
      
      It's currently possible for CSI volumes to be claimed by allocations
      that no longer exist. This changeset asserts a reasonable state at
      the state store level by registering these nil allocations as "past
      claims" on any read. This will cause any pass through the periodic GC
      or volumewatcher to trigger the unpublishing workflow for those claims.
      
      * csi: make feasibility check errors more understandable
      
      When the feasibility checker finds we have no free write claims, it
      checks to see if any of those claims are for the job we're currently
      scheduling (so that earlier versions of a job can't block claims for
      new versions) and reports a conflict if the volume can't be scheduled
      so that the user can fix their claims. But when the checker hits a
      claim that has a GCd allocation, the state is recoverable by the
      server once claim reaping completes and no user intervention is
      required; the blocked eval should complete. Differentiate the
      scheduler error produced by these two conditions.
      cdbb2bcf
    • Tim Gross's avatar
      csi: update leader's ACL in volumewatcher (#11891) · debffe24
      Tim Gross authored
      The volumewatcher that runs on the leader needs to make RPC calls
      rather than writing to raft (as we do in the deploymentwatcher)
      because the unpublish workflow needs to make RPC calls to the
      clients. This requires that the volumewatcher has access to the
      leader's ACL token.
      
      But when leadership transitions, the new leader creates a new leader
      ACL token. This ACL token needs to be passed into the volumewatcher
      when we enable it, otherwise the volumewatcher can find itself with a
      stale token.
      debffe24
    • Derek Strickland's avatar
  3. 19 Jan, 2022 1 commit
  4. 18 Jan, 2022 5 commits
  5. 17 Jan, 2022 5 commits
  6. 15 Jan, 2022 1 commit
  7. 14 Jan, 2022 15 commits
  8. 13 Jan, 2022 3 commits