mirror of https://github.com/helm/helm
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
571 lines
20 KiB
571 lines
20 KiB
/*
|
|
Copyright The Helm Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package action
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"fmt"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/pkg/errors"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/cli-runtime/pkg/resource"
|
|
|
|
"helm.sh/helm/v3/pkg/chart"
|
|
"helm.sh/helm/v3/pkg/chartutil"
|
|
"helm.sh/helm/v3/pkg/kube"
|
|
"helm.sh/helm/v3/pkg/postrender"
|
|
"helm.sh/helm/v3/pkg/release"
|
|
"helm.sh/helm/v3/pkg/releaseutil"
|
|
"helm.sh/helm/v3/pkg/storage/driver"
|
|
)
|
|
|
|
// Upgrade is the action for upgrading releases.
|
|
//
|
|
// It provides the implementation of 'helm upgrade'.
|
|
type Upgrade struct {
|
|
cfg *Configuration
|
|
|
|
ChartPathOptions
|
|
|
|
// Install is a purely informative flag that indicates whether this upgrade was done in "install" mode.
|
|
//
|
|
// Applications may use this to determine whether this Upgrade operation was done as part of a
|
|
// pure upgrade (Upgrade.Install == false) or as part of an install-or-upgrade operation
|
|
// (Upgrade.Install == true).
|
|
//
|
|
// Setting this to `true` will NOT cause `Upgrade` to perform an install if the release does not exist.
|
|
// That process must be handled by creating an Install action directly. See cmd/upgrade.go for an
|
|
// example of how this flag is used.
|
|
Install bool
|
|
// Devel indicates that the operation is done in devel mode.
|
|
Devel bool
|
|
// Namespace is the namespace in which this operation should be performed.
|
|
Namespace string
|
|
// SkipCRDs skips installing CRDs when install flag is enabled during upgrade
|
|
SkipCRDs bool
|
|
// Timeout is the timeout for this operation
|
|
Timeout time.Duration
|
|
// Wait determines whether the wait operation should be performed after the upgrade is requested.
|
|
Wait bool
|
|
// WaitForJobs determines whether the wait operation for the Jobs should be performed after the upgrade is requested.
|
|
WaitForJobs bool
|
|
// DisableHooks disables hook processing if set to true.
|
|
DisableHooks bool
|
|
// DryRun controls whether the operation is prepared, but not executed.
|
|
// If `true`, the upgrade is prepared but not performed.
|
|
DryRun bool
|
|
// Force will, if set to `true`, ignore certain warnings and perform the upgrade anyway.
|
|
//
|
|
// This should be used with caution.
|
|
Force bool
|
|
// ResetValues will reset the values to the chart's built-ins rather than merging with existing.
|
|
ResetValues bool
|
|
// ReuseValues will re-use the user's last supplied values.
|
|
ReuseValues bool
|
|
// Recreate will (if true) recreate pods after a rollback.
|
|
Recreate bool
|
|
// MaxHistory limits the maximum number of revisions saved per release
|
|
MaxHistory int
|
|
// Atomic, if true, will roll back on failure.
|
|
Atomic bool
|
|
// CleanupOnFail will, if true, cause the upgrade to delete newly-created resources on a failed update.
|
|
CleanupOnFail bool
|
|
// SubNotes determines whether sub-notes are rendered in the chart.
|
|
SubNotes bool
|
|
// Description is the description of this operation
|
|
Description string
|
|
// PostRender is an optional post-renderer
|
|
//
|
|
// If this is non-nil, then after templates are rendered, they will be sent to the
|
|
// post renderer before sending to the Kubernetes API server.
|
|
PostRenderer postrender.PostRenderer
|
|
// DisableOpenAPIValidation controls whether OpenAPI validation is enforced.
|
|
DisableOpenAPIValidation bool
|
|
// Get missing dependencies
|
|
DependencyUpdate bool
|
|
// Lock to control raceconditions when the process receives a SIGTERM
|
|
Lock sync.Mutex
|
|
}
|
|
|
|
type resultMessage struct {
|
|
r *release.Release
|
|
e error
|
|
}
|
|
|
|
// NewUpgrade creates a new Upgrade object with the given configuration.
|
|
func NewUpgrade(cfg *Configuration) *Upgrade {
|
|
return &Upgrade{
|
|
cfg: cfg,
|
|
}
|
|
}
|
|
|
|
// Run executes the upgrade on the given release.
|
|
func (u *Upgrade) Run(name string, chart *chart.Chart, vals map[string]interface{}) (*release.Release, error) {
|
|
ctx := context.Background()
|
|
return u.RunWithContext(ctx, name, chart, vals)
|
|
}
|
|
|
|
// RunWithContext executes the upgrade on the given release with context.
|
|
func (u *Upgrade) RunWithContext(ctx context.Context, name string, chart *chart.Chart, vals map[string]interface{}) (*release.Release, error) {
|
|
if err := u.cfg.KubeClient.IsReachable(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Make sure if Atomic is set, that wait is set as well. This makes it so
|
|
// the user doesn't have to specify both
|
|
u.Wait = u.Wait || u.Atomic
|
|
|
|
if err := chartutil.ValidateReleaseName(name); err != nil {
|
|
return nil, errors.Errorf("release name is invalid: %s", name)
|
|
}
|
|
u.cfg.Log("preparing upgrade for %s", name)
|
|
currentRelease, upgradedRelease, err := u.prepareUpgrade(name, chart, vals)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
u.cfg.Releases.MaxHistory = u.MaxHistory
|
|
|
|
u.cfg.Log("performing update for %s", name)
|
|
res, err := u.performUpgrade(ctx, currentRelease, upgradedRelease)
|
|
if err != nil {
|
|
return res, err
|
|
}
|
|
|
|
if !u.DryRun {
|
|
u.cfg.Log("updating status for upgraded release for %s", name)
|
|
if err := u.cfg.Releases.Update(upgradedRelease); err != nil {
|
|
return res, err
|
|
}
|
|
}
|
|
|
|
return res, nil
|
|
}
|
|
|
|
// prepareUpgrade builds an upgraded release for an upgrade operation.
|
|
func (u *Upgrade) prepareUpgrade(name string, chart *chart.Chart, vals map[string]interface{}) (*release.Release, *release.Release, error) {
|
|
if chart == nil {
|
|
return nil, nil, errMissingChart
|
|
}
|
|
|
|
// finds the last non-deleted release with the given name
|
|
lastRelease, err := u.cfg.Releases.Last(name)
|
|
if err != nil {
|
|
// to keep existing behavior of returning the "%q has no deployed releases" error when an existing release does not exist
|
|
if errors.Is(err, driver.ErrReleaseNotFound) {
|
|
return nil, nil, driver.NewErrNoDeployedReleases(name)
|
|
}
|
|
return nil, nil, err
|
|
}
|
|
|
|
// Concurrent `helm upgrade`s will either fail here with `errPending` or when creating the release with "already exists". This should act as a pessimistic lock.
|
|
if lastRelease.Info.Status.IsPending() {
|
|
return nil, nil, errPending
|
|
}
|
|
|
|
var currentRelease *release.Release
|
|
if lastRelease.Info.Status == release.StatusDeployed {
|
|
// no need to retrieve the last deployed release from storage as the last release is deployed
|
|
currentRelease = lastRelease
|
|
} else {
|
|
// finds the deployed release with the given name
|
|
currentRelease, err = u.cfg.Releases.Deployed(name)
|
|
if err != nil {
|
|
if errors.Is(err, driver.ErrNoDeployedReleases) &&
|
|
(lastRelease.Info.Status == release.StatusFailed || lastRelease.Info.Status == release.StatusSuperseded) {
|
|
currentRelease = lastRelease
|
|
} else {
|
|
return nil, nil, err
|
|
}
|
|
}
|
|
}
|
|
|
|
// determine if values will be reused
|
|
vals, err = u.reuseValues(chart, currentRelease, vals)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
if err := chartutil.ProcessDependencies(chart, vals); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
// Increment revision count. This is passed to templates, and also stored on
|
|
// the release object.
|
|
revision := lastRelease.Version + 1
|
|
|
|
options := chartutil.ReleaseOptions{
|
|
Name: name,
|
|
Namespace: currentRelease.Namespace,
|
|
Revision: revision,
|
|
IsUpgrade: true,
|
|
}
|
|
|
|
caps, err := u.cfg.getCapabilities()
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
valuesToRender, err := chartutil.ToRenderValues(chart, vals, options, caps)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
hooks, manifestDoc, notesTxt, err := u.cfg.renderResources(chart, valuesToRender, "", "", u.SubNotes, false, false, u.PostRenderer, u.DryRun)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
// Store an upgraded release.
|
|
upgradedRelease := &release.Release{
|
|
Name: name,
|
|
Namespace: currentRelease.Namespace,
|
|
Chart: chart,
|
|
Config: vals,
|
|
Info: &release.Info{
|
|
FirstDeployed: currentRelease.Info.FirstDeployed,
|
|
LastDeployed: Timestamper(),
|
|
Status: release.StatusPendingUpgrade,
|
|
Description: "Preparing upgrade", // This should be overwritten later.
|
|
},
|
|
Version: revision,
|
|
Manifest: manifestDoc.String(),
|
|
Hooks: hooks,
|
|
}
|
|
|
|
if len(notesTxt) > 0 {
|
|
upgradedRelease.Info.Notes = notesTxt
|
|
}
|
|
err = validateManifest(u.cfg.KubeClient, manifestDoc.Bytes(), !u.DisableOpenAPIValidation)
|
|
return currentRelease, upgradedRelease, err
|
|
}
|
|
|
|
func (u *Upgrade) performUpgrade(ctx context.Context, originalRelease, upgradedRelease *release.Release) (*release.Release, error) {
|
|
current, err := u.cfg.KubeClient.Build(bytes.NewBufferString(originalRelease.Manifest), false)
|
|
if err != nil {
|
|
// Checking for removed Kubernetes API error so can provide a more informative error message to the user
|
|
// Ref: https://github.com/helm/helm/issues/7219
|
|
if strings.Contains(err.Error(), "unable to recognize \"\": no matches for kind") {
|
|
return upgradedRelease, errors.Wrap(err, "current release manifest contains removed kubernetes api(s) for this "+
|
|
"kubernetes version and it is therefore unable to build the kubernetes "+
|
|
"objects for performing the diff. error from kubernetes")
|
|
}
|
|
return upgradedRelease, errors.Wrap(err, "unable to build kubernetes objects from current release manifest")
|
|
}
|
|
target, err := u.cfg.KubeClient.Build(bytes.NewBufferString(upgradedRelease.Manifest), !u.DisableOpenAPIValidation)
|
|
if err != nil {
|
|
return upgradedRelease, errors.Wrap(err, "unable to build kubernetes objects from new release manifest")
|
|
}
|
|
|
|
// It is safe to use force only on target because these are resources currently rendered by the chart.
|
|
err = target.Visit(setMetadataVisitor(upgradedRelease.Name, upgradedRelease.Namespace, true))
|
|
if err != nil {
|
|
return upgradedRelease, err
|
|
}
|
|
|
|
// Do a basic diff using gvk + name to figure out what new resources are being created so we can validate they don't already exist
|
|
existingResources := make(map[string]bool)
|
|
for _, r := range current {
|
|
existingResources[objectKey(r)] = true
|
|
}
|
|
|
|
var toBeCreated kube.ResourceList
|
|
for _, r := range target {
|
|
if !existingResources[objectKey(r)] {
|
|
toBeCreated = append(toBeCreated, r)
|
|
}
|
|
}
|
|
|
|
toBeUpdated, err := existingResourceConflict(toBeCreated, upgradedRelease.Name, upgradedRelease.Namespace)
|
|
if err != nil {
|
|
return nil, errors.Wrap(err, "rendered manifests contain a resource that already exists. Unable to continue with update")
|
|
}
|
|
|
|
toBeUpdated.Visit(func(r *resource.Info, err error) error {
|
|
if err != nil {
|
|
return err
|
|
}
|
|
current.Append(r)
|
|
return nil
|
|
})
|
|
|
|
if u.DryRun {
|
|
u.cfg.Log("dry run for %s", upgradedRelease.Name)
|
|
if len(u.Description) > 0 {
|
|
upgradedRelease.Info.Description = u.Description
|
|
} else {
|
|
upgradedRelease.Info.Description = "Dry run complete"
|
|
}
|
|
return upgradedRelease, nil
|
|
}
|
|
|
|
u.cfg.Log("creating upgraded release for %s", upgradedRelease.Name)
|
|
if err := u.cfg.Releases.Create(upgradedRelease); err != nil {
|
|
return nil, err
|
|
}
|
|
rChan := make(chan resultMessage)
|
|
ctxChan := make(chan resultMessage)
|
|
doneChan := make(chan interface{})
|
|
go u.releasingUpgrade(rChan, upgradedRelease, current, target, originalRelease)
|
|
go u.handleContext(ctx, doneChan, ctxChan, upgradedRelease)
|
|
select {
|
|
case result := <-rChan:
|
|
doneChan <- true
|
|
return result.r, result.e
|
|
case result := <-ctxChan:
|
|
return result.r, result.e
|
|
}
|
|
}
|
|
|
|
// Function used to lock the Mutex, this is important for the case when the atomic flag is set.
|
|
// In that case the upgrade will finish before the rollback is finished so it is necessary to wait for the rollback to finish.
|
|
// The rollback will be trigger by the function failRelease
|
|
func (u *Upgrade) reportToPerformUpgrade(c chan<- resultMessage, rel *release.Release, created kube.ResourceList, err error) {
|
|
u.Lock.Lock()
|
|
if err != nil {
|
|
rel, err = u.failRelease(rel, created, err)
|
|
}
|
|
c <- resultMessage{r: rel, e: err}
|
|
u.Lock.Unlock()
|
|
}
|
|
|
|
// Setup listener for SIGINT and SIGTERM
|
|
func (u *Upgrade) handleContext(ctx context.Context, done chan interface{}, c chan<- resultMessage, upgradedRelease *release.Release) {
|
|
go func() {
|
|
select {
|
|
case <-ctx.Done():
|
|
err := ctx.Err()
|
|
|
|
// when the atomic flag is set the ongoing release finish first and doesn't give time for the rollback happens.
|
|
u.reportToPerformUpgrade(c, upgradedRelease, kube.ResourceList{}, err)
|
|
case <-done:
|
|
return
|
|
}
|
|
}()
|
|
}
|
|
func (u *Upgrade) releasingUpgrade(c chan<- resultMessage, upgradedRelease *release.Release, current kube.ResourceList, target kube.ResourceList, originalRelease *release.Release) {
|
|
// pre-upgrade hooks
|
|
|
|
if !u.DisableHooks {
|
|
if err := u.cfg.execHook(upgradedRelease, release.HookPreUpgrade, u.Timeout); err != nil {
|
|
u.reportToPerformUpgrade(c, upgradedRelease, kube.ResourceList{}, fmt.Errorf("pre-upgrade hooks failed: %s", err))
|
|
return
|
|
}
|
|
} else {
|
|
u.cfg.Log("upgrade hooks disabled for %s", upgradedRelease.Name)
|
|
}
|
|
|
|
results, err := u.cfg.KubeClient.Update(current, target, u.Force)
|
|
if err != nil {
|
|
u.cfg.recordRelease(originalRelease)
|
|
u.reportToPerformUpgrade(c, upgradedRelease, results.Created, err)
|
|
return
|
|
}
|
|
|
|
if u.Recreate {
|
|
// NOTE: Because this is not critical for a release to succeed, we just
|
|
// log if an error occurs and continue onward. If we ever introduce log
|
|
// levels, we should make these error level logs so users are notified
|
|
// that they'll need to go do the cleanup on their own
|
|
if err := recreate(u.cfg, results.Updated); err != nil {
|
|
u.cfg.Log(err.Error())
|
|
}
|
|
}
|
|
|
|
if u.Wait {
|
|
if u.WaitForJobs {
|
|
if err := u.cfg.KubeClient.WaitWithJobs(target, u.Timeout); err != nil {
|
|
u.cfg.recordRelease(originalRelease)
|
|
u.reportToPerformUpgrade(c, upgradedRelease, results.Created, err)
|
|
return
|
|
}
|
|
} else {
|
|
if err := u.cfg.KubeClient.Wait(target, u.Timeout); err != nil {
|
|
u.cfg.recordRelease(originalRelease)
|
|
u.reportToPerformUpgrade(c, upgradedRelease, results.Created, err)
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// post-upgrade hooks
|
|
if !u.DisableHooks {
|
|
if err := u.cfg.execHook(upgradedRelease, release.HookPostUpgrade, u.Timeout); err != nil {
|
|
u.reportToPerformUpgrade(c, upgradedRelease, results.Created, fmt.Errorf("post-upgrade hooks failed: %s", err))
|
|
return
|
|
}
|
|
}
|
|
|
|
originalRelease.Info.Status = release.StatusSuperseded
|
|
u.cfg.recordRelease(originalRelease)
|
|
|
|
upgradedRelease.Info.Status = release.StatusDeployed
|
|
if len(u.Description) > 0 {
|
|
upgradedRelease.Info.Description = u.Description
|
|
} else {
|
|
upgradedRelease.Info.Description = "Upgrade complete"
|
|
}
|
|
u.reportToPerformUpgrade(c, upgradedRelease, nil, nil)
|
|
}
|
|
|
|
func (u *Upgrade) failRelease(rel *release.Release, created kube.ResourceList, err error) (*release.Release, error) {
|
|
msg := fmt.Sprintf("Upgrade %q failed: %s", rel.Name, err)
|
|
u.cfg.Log("warning: %s", msg)
|
|
|
|
rel.Info.Status = release.StatusFailed
|
|
rel.Info.Description = msg
|
|
u.cfg.recordRelease(rel)
|
|
if u.CleanupOnFail && len(created) > 0 {
|
|
u.cfg.Log("Cleanup on fail set, cleaning up %d resources", len(created))
|
|
_, errs := u.cfg.KubeClient.Delete(created)
|
|
if errs != nil {
|
|
var errorList []string
|
|
for _, e := range errs {
|
|
errorList = append(errorList, e.Error())
|
|
}
|
|
return rel, errors.Wrapf(fmt.Errorf("unable to cleanup resources: %s", strings.Join(errorList, ", ")), "an error occurred while cleaning up resources. original upgrade error: %s", err)
|
|
}
|
|
u.cfg.Log("Resource cleanup complete")
|
|
}
|
|
if u.Atomic {
|
|
u.cfg.Log("Upgrade failed and atomic is set, rolling back to last successful release")
|
|
|
|
// As a protection, get the last successful release before rollback.
|
|
// If there are no successful releases, bail out
|
|
hist := NewHistory(u.cfg)
|
|
fullHistory, herr := hist.Run(rel.Name)
|
|
if herr != nil {
|
|
return rel, errors.Wrapf(herr, "an error occurred while finding last successful release. original upgrade error: %s", err)
|
|
}
|
|
|
|
// There isn't a way to tell if a previous release was successful, but
|
|
// generally failed releases do not get superseded unless the next
|
|
// release is successful, so this should be relatively safe
|
|
filteredHistory := releaseutil.FilterFunc(func(r *release.Release) bool {
|
|
return r.Info.Status == release.StatusSuperseded || r.Info.Status == release.StatusDeployed
|
|
}).Filter(fullHistory)
|
|
if len(filteredHistory) == 0 {
|
|
return rel, errors.Wrap(err, "unable to find a previously successful release when attempting to rollback. original upgrade error")
|
|
}
|
|
|
|
releaseutil.Reverse(filteredHistory, releaseutil.SortByRevision)
|
|
|
|
rollin := NewRollback(u.cfg)
|
|
rollin.Version = filteredHistory[0].Version
|
|
rollin.Wait = true
|
|
rollin.WaitForJobs = u.WaitForJobs
|
|
rollin.DisableHooks = u.DisableHooks
|
|
rollin.Recreate = u.Recreate
|
|
rollin.Force = u.Force
|
|
rollin.Timeout = u.Timeout
|
|
if rollErr := rollin.Run(rel.Name); rollErr != nil {
|
|
return rel, errors.Wrapf(rollErr, "an error occurred while rolling back the release. original upgrade error: %s", err)
|
|
}
|
|
return rel, errors.Wrapf(err, "release %s failed, and has been rolled back due to atomic being set", rel.Name)
|
|
}
|
|
|
|
return rel, err
|
|
}
|
|
|
|
// reuseValues copies values from the current release to a new release if the
|
|
// new release does not have any values.
|
|
//
|
|
// If the request already has values, or if there are no values in the current
|
|
// release, this does nothing.
|
|
//
|
|
// This is skipped if the u.ResetValues flag is set, in which case the
|
|
// request values are not altered.
|
|
func (u *Upgrade) reuseValues(chart *chart.Chart, current *release.Release, newVals map[string]interface{}) (map[string]interface{}, error) {
|
|
if u.ResetValues {
|
|
// If ResetValues is set, we completely ignore current.Config.
|
|
u.cfg.Log("resetting values to the chart's original version")
|
|
return newVals, nil
|
|
}
|
|
|
|
// If the ReuseValues flag is set, we always copy the old values over the new config's values.
|
|
if u.ReuseValues {
|
|
u.cfg.Log("reusing the old release's values")
|
|
|
|
// We have to regenerate the old coalesced values:
|
|
oldVals, err := chartutil.CoalesceValues(current.Chart, current.Config)
|
|
if err != nil {
|
|
return nil, errors.Wrap(err, "failed to rebuild old values")
|
|
}
|
|
|
|
newVals = chartutil.CoalesceTables(newVals, current.Config)
|
|
|
|
chart.Values = oldVals
|
|
|
|
return newVals, nil
|
|
}
|
|
|
|
if len(newVals) == 0 && len(current.Config) > 0 {
|
|
u.cfg.Log("copying values from %s (v%d) to new release.", current.Name, current.Version)
|
|
newVals = current.Config
|
|
}
|
|
return newVals, nil
|
|
}
|
|
|
|
func validateManifest(c kube.Interface, manifest []byte, openAPIValidation bool) error {
|
|
_, err := c.Build(bytes.NewReader(manifest), openAPIValidation)
|
|
return err
|
|
}
|
|
|
|
// recreate captures all the logic for recreating pods for both upgrade and
|
|
// rollback. If we end up refactoring rollback to use upgrade, this can just be
|
|
// made an unexported method on the upgrade action.
|
|
func recreate(cfg *Configuration, resources kube.ResourceList) error {
|
|
for _, res := range resources {
|
|
versioned := kube.AsVersioned(res)
|
|
selector, err := kube.SelectorsForObject(versioned)
|
|
if err != nil {
|
|
// If no selector is returned, it means this object is
|
|
// definitely not a pod, so continue onward
|
|
continue
|
|
}
|
|
|
|
client, err := cfg.KubernetesClientSet()
|
|
if err != nil {
|
|
return errors.Wrapf(err, "unable to recreate pods for object %s/%s because an error occurred", res.Namespace, res.Name)
|
|
}
|
|
|
|
pods, err := client.CoreV1().Pods(res.Namespace).List(context.Background(), metav1.ListOptions{
|
|
LabelSelector: selector.String(),
|
|
})
|
|
if err != nil {
|
|
return errors.Wrapf(err, "unable to recreate pods for object %s/%s because an error occurred", res.Namespace, res.Name)
|
|
}
|
|
|
|
// Restart pods
|
|
for _, pod := range pods.Items {
|
|
// Delete each pod for get them restarted with changed spec.
|
|
if err := client.CoreV1().Pods(pod.Namespace).Delete(context.Background(), pod.Name, *metav1.NewPreconditionDeleteOptions(string(pod.UID))); err != nil {
|
|
return errors.Wrapf(err, "unable to recreate pods for object %s/%s because an error occurred", res.Namespace, res.Name)
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func objectKey(r *resource.Info) string {
|
|
gvk := r.Object.GetObjectKind().GroupVersionKind()
|
|
return fmt.Sprintf("%s/%s/%s/%s", gvk.GroupVersion().String(), gvk.Kind, r.Namespace, r.Name)
|
|
}
|