refactor test

Signed-off-by: Austin Abro <AustinAbro321@gmail.com>
pull/13604/head
Austin Abro 9 months ago
parent 947425ee64
commit 807cc925f5
No known key found for this signature in database
GPG Key ID: 92EB5159E403F9D6

@ -140,7 +140,10 @@ func New(getter genericclioptions.RESTClientGetter, waiter Waiter) (*Client, err
if err != nil { if err != nil {
return nil, err return nil, err
} }
waiter = &kstatusWaiter{sw, nopLogger} waiter = &kstatusWaiter{
sw: sw,
log: nopLogger,
pausedAsReady: true}
} }
return &Client{ return &Client{
Factory: factory, Factory: factory,

@ -34,6 +34,7 @@ type kstatusWaiter struct {
// Add any necessary dependencies, e.g., Kubernetes API client. // Add any necessary dependencies, e.g., Kubernetes API client.
sw watcher.StatusWatcher sw watcher.StatusWatcher
log func(string, ...interface{}) log func(string, ...interface{})
pausedAsReady bool
} }
func (w *kstatusWaiter) Wait(resourceList ResourceList, timeout time.Duration) error { func (w *kstatusWaiter) Wait(resourceList ResourceList, timeout time.Duration) error {

@ -18,12 +18,12 @@ package kube // import "helm.sh/helm/v3/pkg/kube"
import ( import (
"errors" "errors"
"fmt"
"testing" "testing"
"time" "time"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
v1 "k8s.io/api/core/v1" v1 "k8s.io/api/core/v1"
"k8s.io/apimachinery/pkg/api/meta"
"k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured"
"k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/runtime"
"k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apimachinery/pkg/runtime/schema"
@ -34,7 +34,7 @@ import (
"sigs.k8s.io/cli-utils/pkg/testutil" "sigs.k8s.io/cli-utils/pkg/testutil"
) )
var podCurrentYaml = ` var podCurrent = `
apiVersion: v1 apiVersion: v1
kind: Pod kind: Pod
metadata: metadata:
@ -47,7 +47,7 @@ status:
phase: Running phase: Running
` `
var podYaml = ` var podNoStatus = `
apiVersion: v1 apiVersion: v1
kind: Pod kind: Pod
metadata: metadata:
@ -55,21 +55,62 @@ metadata:
namespace: ns namespace: ns
` `
func TestRunHealthChecks(t *testing.T) { var jobNoStatus = `
apiVersion: batch/v1
kind: Job
metadata:
name: test
namespace: qual
generation: 1
`
var jobComplete = `
apiVersion: batch/v1
kind: Job
metadata:
name: test
namespace: qual
generation: 1
status:
succeeded: 1
active: 0
conditions:
- type: Complete
status: "True"
`
func getGVR(t *testing.T, mapper meta.RESTMapper, obj *unstructured.Unstructured) schema.GroupVersionResource {
gvk := obj.GroupVersionKind()
mapping, err := mapper.RESTMapping(gvk.GroupKind(), gvk.Version)
require.NoError(t, err)
return mapping.Resource
}
func TestKWaitJob(t *testing.T) {
t.Parallel() t.Parallel()
tests := []struct { tests := []struct {
name string name string
podYamls []string objYamls []string
expectErrs []error expectErrs []error
}{ }{
{
name: "Job is complete",
objYamls: []string{jobComplete},
expectErrs: nil,
},
{
name: "Job is not complete",
objYamls: []string{jobNoStatus},
expectErrs: []error{errors.New("not all resources ready: context deadline exceeded: test: Job not ready, status: InProgress")},
},
{ {
name: "Pod is ready", name: "Pod is ready",
podYamls: []string{podCurrentYaml}, objYamls: []string{podCurrent},
expectErrs: nil, expectErrs: nil,
}, },
{ {
name: "one of the pods never becomes ready", name: "one of the pods never becomes ready",
podYamls: []string{podYaml, podCurrentYaml}, objYamls: []string{podNoStatus, podCurrent},
// TODO, make this better // TODO, make this better
expectErrs: []error{errors.New("not all resources ready: context deadline exceeded: in-progress-pod: Pod not ready, status: InProgress")}, expectErrs: []error{errors.New("not all resources ready: context deadline exceeded: in-progress-pod: Pod not ready, status: InProgress")},
}, },
@ -82,18 +123,22 @@ func TestRunHealthChecks(t *testing.T) {
fakeClient := dynamicfake.NewSimpleDynamicClient(scheme.Scheme) fakeClient := dynamicfake.NewSimpleDynamicClient(scheme.Scheme)
fakeMapper := testutil.NewFakeRESTMapper( fakeMapper := testutil.NewFakeRESTMapper(
v1.SchemeGroupVersion.WithKind("Pod"), v1.SchemeGroupVersion.WithKind("Pod"),
schema.GroupVersionKind{
Group: "batch",
Version: "v1",
Kind: "Job",
},
) )
pods := []runtime.Object{} objs := []runtime.Object{}
statusWatcher := watcher.NewDefaultStatusWatcher(fakeClient, fakeMapper) statusWatcher := watcher.NewDefaultStatusWatcher(fakeClient, fakeMapper)
for _, podYaml := range tt.podYamls { for _, podYaml := range tt.objYamls {
m := make(map[string]interface{}) m := make(map[string]interface{})
err := yaml.Unmarshal([]byte(podYaml), &m) err := yaml.Unmarshal([]byte(podYaml), &m)
require.NoError(t, err) require.NoError(t, err)
pod := &unstructured.Unstructured{Object: m} resource := &unstructured.Unstructured{Object: m}
pods = append(pods, pod) objs = append(objs, resource)
fmt.Println(pod.GetName()) gvr := getGVR(t, fakeMapper, resource)
podGVR := schema.GroupVersionResource{Group: "", Version: "v1", Resource: "pods"} err = fakeClient.Tracker().Create(gvr, resource, resource.GetNamespace())
err = fakeClient.Tracker().Create(podGVR, pod, pod.GetNamespace())
require.NoError(t, err) require.NoError(t, err)
} }
c.Waiter = &kstatusWaiter{ c.Waiter = &kstatusWaiter{
@ -102,16 +147,17 @@ func TestRunHealthChecks(t *testing.T) {
} }
resourceList := ResourceList{} resourceList := ResourceList{}
for _, pod := range pods { for _, obj := range objs {
list, err := c.Build(objBody(pod), false) list, err := c.Build(objBody(obj), false)
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
resourceList = append(resourceList, list...) resourceList = append(resourceList, list...)
} }
err := c.Wait(resourceList, time.Second*5) err := c.Wait(resourceList, time.Second*3)
if tt.expectErrs != nil { if tt.expectErrs != nil {
//TODO remove require
require.EqualError(t, err, errors.Join(tt.expectErrs...).Error()) require.EqualError(t, err, errors.Join(tt.expectErrs...).Error())
return return
} }

Loading…
Cancel
Save