mirror of https://github.com/helm/helm
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
378 lines
13 KiB
378 lines
13 KiB
/*
|
|
Copyright The Helm Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package main
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
"time"
|
|
|
|
"github.com/spf13/cobra"
|
|
apierrors "k8s.io/apimachinery/pkg/api/errors"
|
|
"k8s.io/client-go/kubernetes"
|
|
|
|
"k8s.io/apimachinery/pkg/util/yaml"
|
|
"k8s.io/helm/cmd/helm/installer"
|
|
"k8s.io/helm/pkg/helm"
|
|
"k8s.io/helm/pkg/helm/helmpath"
|
|
"k8s.io/helm/pkg/helm/portforwarder"
|
|
"k8s.io/helm/pkg/version"
|
|
)
|
|
|
|
const initDesc = `
|
|
This command installs Tiller (the Helm server-side component) onto your
|
|
Kubernetes Cluster and sets up local configuration in $HELM_HOME (default ~/.helm/).
|
|
|
|
As with the rest of the Helm commands, 'helm init' discovers Kubernetes clusters
|
|
by reading $KUBECONFIG (default '~/.kube/config') and using the default context.
|
|
|
|
To set up just a local environment, use '--client-only'. That will configure
|
|
$HELM_HOME, but not attempt to connect to a Kubernetes cluster and install the Tiller
|
|
deployment.
|
|
|
|
When installing Tiller, 'helm init' will attempt to install the latest released
|
|
version. You can specify an alternative image with '--tiller-image'. For those
|
|
frequently working on the latest code, the flag '--canary-image' will install
|
|
the latest pre-release version of Tiller (e.g. the HEAD commit in the GitHub
|
|
repository on the master branch).
|
|
|
|
To dump a manifest containing the Tiller deployment YAML, combine the
|
|
'--dry-run' and '--debug' flags.
|
|
`
|
|
|
|
var (
|
|
stableRepositoryURL = "https://kubernetes-charts.storage.googleapis.com"
|
|
// This is the IPv4 loopback, not localhost, because we have to force IPv4
|
|
// for Dockerized Helm: https://github.com/kubernetes/helm/issues/1410
|
|
localRepositoryURL = "http://127.0.0.1:8879/charts"
|
|
tlsServerName string // overrides the server name used to verify the hostname on the returned certificates from the server.
|
|
tlsCaCertFile string // path to TLS CA certificate file
|
|
tlsCertFile string // path to TLS certificate file
|
|
tlsKeyFile string // path to TLS key file
|
|
tlsVerify bool // enable TLS and verify remote certificates
|
|
tlsEnable bool // enable TLS
|
|
)
|
|
|
|
type initCmd struct {
|
|
image string
|
|
clientOnly bool
|
|
canary bool
|
|
upgrade bool
|
|
namespace string
|
|
dryRun bool
|
|
forceUpgrade bool
|
|
skipRefresh bool
|
|
skipRepos bool
|
|
out io.Writer
|
|
client helm.Interface
|
|
home helmpath.Home
|
|
opts installer.Options
|
|
kubeClient kubernetes.Interface
|
|
serviceAccount string
|
|
maxHistory int
|
|
replicas int
|
|
wait bool
|
|
}
|
|
|
|
func newInitCmd(out io.Writer) *cobra.Command {
|
|
i := &initCmd{out: out}
|
|
|
|
cmd := &cobra.Command{
|
|
Use: "init",
|
|
Short: "Initialize Helm on both client and server",
|
|
Long: initDesc,
|
|
RunE: func(cmd *cobra.Command, args []string) error {
|
|
if len(args) != 0 {
|
|
return errors.New("This command does not accept arguments")
|
|
}
|
|
i.namespace = settings.TillerNamespace
|
|
i.home = settings.Home
|
|
i.client = ensureHelmClient(i.client)
|
|
|
|
return i.run()
|
|
},
|
|
}
|
|
|
|
f := cmd.Flags()
|
|
f.StringVarP(&i.image, "tiller-image", "i", "", "Override Tiller image")
|
|
f.BoolVar(&i.canary, "canary-image", false, "Use the canary Tiller image")
|
|
f.BoolVar(&i.upgrade, "upgrade", false, "Upgrade if Tiller is already installed")
|
|
f.BoolVar(&i.forceUpgrade, "force-upgrade", false, "Force upgrade of Tiller to the current helm version")
|
|
f.BoolVarP(&i.clientOnly, "client-only", "c", false, "If set does not install Tiller")
|
|
f.BoolVar(&i.dryRun, "dry-run", false, "Do not install local or remote")
|
|
f.BoolVar(&i.skipRefresh, "skip-refresh", false, "Do not refresh (download) the local repository cache")
|
|
f.BoolVar(&i.skipRepos, "skip-repos", false, "Skip adding the stable and local repositories")
|
|
f.BoolVar(&i.wait, "wait", false, "Block until Tiller is running and ready to receive requests")
|
|
|
|
// TODO: replace TLS flags with pkg/helm/environment.AddFlagsTLS() in Helm 3
|
|
//
|
|
// NOTE (bacongobbler): we can't do this in Helm 2 because the flag names differ, and `helm init --tls-ca-cert`
|
|
// doesn't conform with the rest of the TLS flag names (should be --tiller-tls-ca-cert in Helm 3)
|
|
f.BoolVar(&tlsEnable, "tiller-tls", false, "Install Tiller with TLS enabled")
|
|
f.BoolVar(&tlsVerify, "tiller-tls-verify", false, "Install Tiller with TLS enabled and to verify remote certificates")
|
|
f.StringVar(&tlsKeyFile, "tiller-tls-key", "", "Path to TLS key file to install with Tiller")
|
|
f.StringVar(&tlsCertFile, "tiller-tls-cert", "", "Path to TLS certificate file to install with Tiller")
|
|
f.StringVar(&tlsCaCertFile, "tls-ca-cert", "", "Path to CA root certificate")
|
|
f.StringVar(&tlsServerName, "tiller-tls-hostname", settings.TillerHost, "The server name used to verify the hostname on the returned certificates from Tiller")
|
|
|
|
f.StringVar(&stableRepositoryURL, "stable-repo-url", stableRepositoryURL, "URL for stable repository")
|
|
f.StringVar(&localRepositoryURL, "local-repo-url", localRepositoryURL, "URL for local repository")
|
|
|
|
f.BoolVar(&i.opts.EnableHostNetwork, "net-host", false, "Install Tiller with net=host")
|
|
f.StringVar(&i.serviceAccount, "service-account", "", "Name of service account")
|
|
f.IntVar(&i.maxHistory, "history-max", 0, "Limit the maximum number of revisions saved per release. Use 0 for no limit.")
|
|
f.IntVar(&i.replicas, "replicas", 1, "Amount of tiller instances to run on the cluster")
|
|
|
|
f.StringVar(&i.opts.NodeSelectors, "node-selectors", "", "Labels to specify the node on which Tiller is installed (app=tiller,helm=rocks)")
|
|
f.VarP(&i.opts.Output, "output", "o", "Skip installation and output Tiller's manifest in specified format (json or yaml)")
|
|
f.StringArrayVar(&i.opts.Values, "override", []string{}, "Override values for the Tiller Deployment manifest (can specify multiple or separate values with commas: key1=val1,key2=val2)")
|
|
f.BoolVar(&i.opts.AutoMountServiceAccountToken, "automount-service-account-token", true, "Auto-mount the given service account to tiller")
|
|
|
|
return cmd
|
|
}
|
|
|
|
// tlsOptions sanitizes the tls flags as well as checks for the existence of required
|
|
// tls files indicated by those flags, if any.
|
|
func (i *initCmd) tlsOptions() error {
|
|
i.opts.EnableTLS = tlsEnable || tlsVerify
|
|
i.opts.VerifyTLS = tlsVerify
|
|
|
|
if i.opts.EnableTLS {
|
|
missing := func(file string) bool {
|
|
_, err := os.Stat(file)
|
|
return os.IsNotExist(err)
|
|
}
|
|
if i.opts.TLSKeyFile = tlsKeyFile; i.opts.TLSKeyFile == "" || missing(i.opts.TLSKeyFile) {
|
|
return errors.New("missing required TLS key file")
|
|
}
|
|
if i.opts.TLSCertFile = tlsCertFile; i.opts.TLSCertFile == "" || missing(i.opts.TLSCertFile) {
|
|
return errors.New("missing required TLS certificate file")
|
|
}
|
|
if i.opts.VerifyTLS {
|
|
if i.opts.TLSCaCertFile = tlsCaCertFile; i.opts.TLSCaCertFile == "" || missing(i.opts.TLSCaCertFile) {
|
|
return errors.New("missing required TLS CA file")
|
|
}
|
|
}
|
|
|
|
// FIXME: remove once we use pkg/helm/environment.AddFlagsTLS() in Helm 3
|
|
settings.TLSEnable = tlsEnable
|
|
settings.TLSVerify = tlsVerify
|
|
settings.TLSServerName = tlsServerName
|
|
settings.TLSCaCertFile = tlsCaCertFile
|
|
settings.TLSCertFile = tlsCertFile
|
|
settings.TLSKeyFile = tlsKeyFile
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// run initializes local config and installs Tiller to Kubernetes cluster.
|
|
func (i *initCmd) run() error {
|
|
if err := i.tlsOptions(); err != nil {
|
|
return err
|
|
}
|
|
i.opts.Namespace = i.namespace
|
|
i.opts.UseCanary = i.canary
|
|
i.opts.ImageSpec = i.image
|
|
i.opts.ForceUpgrade = i.forceUpgrade
|
|
i.opts.ServiceAccount = i.serviceAccount
|
|
i.opts.MaxHistory = i.maxHistory
|
|
i.opts.Replicas = i.replicas
|
|
|
|
writeYAMLManifests := func(manifests []string) error {
|
|
w := i.out
|
|
for _, manifest := range manifests {
|
|
if _, err := fmt.Fprintln(w, "---"); err != nil {
|
|
return err
|
|
}
|
|
|
|
if _, err := fmt.Fprintln(w, manifest); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// YAML ending document boundary marker
|
|
_, err := fmt.Fprintln(w, "...")
|
|
return err
|
|
}
|
|
if len(i.opts.Output) > 0 {
|
|
var manifests []string
|
|
var err error
|
|
if manifests, err = installer.TillerManifests(&i.opts); err != nil {
|
|
return err
|
|
}
|
|
switch i.opts.Output.String() {
|
|
case "json":
|
|
for _, manifest := range manifests {
|
|
var out bytes.Buffer
|
|
jsonb, err := yaml.ToJSON([]byte(manifest))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
buf := bytes.NewBuffer(jsonb)
|
|
if err := json.Indent(&out, buf.Bytes(), "", " "); err != nil {
|
|
return err
|
|
}
|
|
if _, err = i.out.Write(out.Bytes()); err != nil {
|
|
return err
|
|
}
|
|
fmt.Fprint(i.out, "\n")
|
|
}
|
|
return nil
|
|
case "yaml":
|
|
return writeYAMLManifests(manifests)
|
|
default:
|
|
return fmt.Errorf("unknown output format: %q", i.opts.Output)
|
|
}
|
|
}
|
|
if settings.Debug {
|
|
var manifests []string
|
|
var err error
|
|
|
|
// write Tiller manifests
|
|
if manifests, err = installer.TillerManifests(&i.opts); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err = writeYAMLManifests(manifests); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
if i.dryRun {
|
|
return nil
|
|
}
|
|
|
|
if i.skipRepos {
|
|
if err := installer.InitializeWithoutRepos(i.home, i.out); err != nil {
|
|
return fmt.Errorf("error initializing: %s", err)
|
|
}
|
|
} else {
|
|
if err := installer.Initialize(i.home, i.out, i.skipRefresh, settings, stableRepositoryURL, localRepositoryURL); err != nil {
|
|
return fmt.Errorf("error initializing: %s", err)
|
|
}
|
|
}
|
|
fmt.Fprintf(i.out, "$HELM_HOME has been configured at %s.\n", settings.Home)
|
|
|
|
if !i.clientOnly {
|
|
if i.kubeClient == nil {
|
|
_, c, err := getKubeClient(settings.KubeContext, settings.KubeConfig)
|
|
if err != nil {
|
|
return fmt.Errorf("could not get kubernetes client: %s", err)
|
|
}
|
|
i.kubeClient = c
|
|
}
|
|
if err := installer.Install(i.kubeClient, &i.opts); err != nil {
|
|
if !apierrors.IsAlreadyExists(err) {
|
|
return fmt.Errorf("error installing: %s", err)
|
|
}
|
|
if i.upgrade {
|
|
if err := installer.Upgrade(i.kubeClient, &i.opts); err != nil {
|
|
return fmt.Errorf("error when upgrading: %s", err)
|
|
}
|
|
if err := i.ping(i.opts.SelectImage()); err != nil {
|
|
return err
|
|
}
|
|
fmt.Fprintln(i.out, "\nTiller (the Helm server-side component) has been updated to", i.opts.SelectImage(), ".")
|
|
} else {
|
|
debug("The error received while trying to init: %s", err)
|
|
fmt.Fprintln(i.out, "Warning: Tiller is already installed in the cluster.\n"+
|
|
"(Use --client-only to suppress this message, or --upgrade to upgrade Tiller to the current version.)")
|
|
}
|
|
} else {
|
|
fmt.Fprintln(i.out, "\nTiller (the Helm server-side component) has been installed into your Kubernetes Cluster.")
|
|
if !tlsVerify {
|
|
fmt.Fprintln(i.out, "\nPlease note: by default, Tiller is deployed with an insecure 'allow unauthenticated users' policy.\n"+
|
|
"To prevent this, run `helm init` with the --tiller-tls-verify flag.\n"+
|
|
"For more information on securing your installation see: https://v2.helm.sh/docs/securing_installation/")
|
|
}
|
|
}
|
|
if err := i.ping(i.opts.SelectImage()); err != nil {
|
|
return err
|
|
}
|
|
} else {
|
|
fmt.Fprintln(i.out, "Not installing Tiller due to 'client-only' flag having been set")
|
|
}
|
|
|
|
needsDefaultImage := !i.clientOnly && !i.opts.UseCanary && len(i.opts.ImageSpec) == 0 && version.BuildMetadata == "unreleased"
|
|
if needsDefaultImage {
|
|
fmt.Fprintf(i.out, "\nWarning: You appear to be using an unreleased version of Helm. Please either use the\n"+
|
|
"--canary-image flag, or specify your desired tiller version with --tiller-image.\n\n"+
|
|
"Ex:\n"+
|
|
"$ helm init --tiller-image gcr.io/kubernetes-helm/tiller:v2.8.2\n\n")
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (i *initCmd) ping(image string) error {
|
|
if i.wait {
|
|
_, kubeClient, err := getKubeClient(settings.KubeContext, settings.KubeConfig)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if !watchTillerUntilReady(settings.TillerNamespace, kubeClient, settings.TillerConnectionTimeout, image) {
|
|
return fmt.Errorf("tiller was not found. polling deadline exceeded")
|
|
}
|
|
|
|
// establish a connection to Tiller now that we've effectively guaranteed it's available
|
|
if err := setupConnection(); err != nil {
|
|
return err
|
|
}
|
|
i.client = newClient()
|
|
if err := i.client.PingTiller(); err != nil {
|
|
return fmt.Errorf("could not ping Tiller: %s", err)
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// watchTillerUntilReady waits for the tiller pod to become available. This is useful in situations where we
|
|
// want to wait before we call New().
|
|
//
|
|
// Returns true if it exists. If the timeout was reached and it could not find the pod, it returns false.
|
|
func watchTillerUntilReady(namespace string, client kubernetes.Interface, timeout int64, newImage string) bool {
|
|
deadlinePollingChan := time.NewTimer(time.Duration(timeout) * time.Second).C
|
|
checkTillerPodTicker := time.NewTicker(500 * time.Millisecond)
|
|
doneChan := make(chan bool)
|
|
|
|
defer checkTillerPodTicker.Stop()
|
|
|
|
go func() {
|
|
for range checkTillerPodTicker.C {
|
|
image, err := portforwarder.GetTillerPodImage(client.CoreV1(), namespace)
|
|
if err == nil && image == newImage {
|
|
doneChan <- true
|
|
break
|
|
}
|
|
}
|
|
}()
|
|
|
|
for {
|
|
select {
|
|
case <-deadlinePollingChan:
|
|
return false
|
|
case <-doneChan:
|
|
return true
|
|
}
|
|
}
|
|
}
|