split snapshot controller and update capability.
Signed-off-by: f10atin9 <f10atin9@kubesphere.io>
This commit is contained in:
@@ -26,9 +26,6 @@ import (
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
snapbeta1 "github.com/kubernetes-csi/external-snapshotter/client/v3/apis/volumesnapshot/v1beta1"
|
||||
snapfake "github.com/kubernetes-csi/external-snapshotter/client/v3/clientset/versioned/fake"
|
||||
snapinformers "github.com/kubernetes-csi/external-snapshotter/client/v3/informers/externalversions"
|
||||
storagev1 "k8s.io/api/storage/v1"
|
||||
v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
"k8s.io/apimachinery/pkg/runtime"
|
||||
@@ -40,7 +37,6 @@ import (
|
||||
"k8s.io/client-go/tools/cache"
|
||||
|
||||
ksfake "kubesphere.io/kubesphere/pkg/client/clientset/versioned/fake"
|
||||
ksinformers "kubesphere.io/kubesphere/pkg/client/informers/externalversions"
|
||||
)
|
||||
|
||||
var (
|
||||
@@ -51,17 +47,13 @@ type fixture struct {
|
||||
t *testing.T
|
||||
snapshotSupported bool
|
||||
// Clients
|
||||
k8sClient *k8sfake.Clientset
|
||||
snapshotClassClient *snapfake.Clientset
|
||||
ksClient *ksfake.Clientset
|
||||
k8sClient *k8sfake.Clientset
|
||||
ksClient *ksfake.Clientset
|
||||
// Objects from here preload into NewSimpleFake.
|
||||
storageObjects []runtime.Object // include StorageClass
|
||||
snapshotClassObjects []runtime.Object
|
||||
capabilityObjects []runtime.Object // include StorageClassCapability and ProvisionerCapability
|
||||
storageObjects []runtime.Object // include StorageClass
|
||||
// Objects to put in the store.
|
||||
storageClassLister []*storagev1.StorageClass
|
||||
snapshotClassLister []*snapbeta1.VolumeSnapshotClass
|
||||
csiDriverLister []*v1beta1.CSIDriver
|
||||
storageClassLister []*storagev1.StorageClass
|
||||
csiDriverLister []*v1beta1.CSIDriver
|
||||
// Actions expected to happen on the client.
|
||||
actions []core.Action
|
||||
}
|
||||
@@ -74,49 +66,34 @@ func newFixture(t *testing.T, snapshotSupported bool) *fixture {
|
||||
}
|
||||
|
||||
func (f *fixture) newController() (*StorageCapabilityController,
|
||||
k8sinformers.SharedInformerFactory,
|
||||
ksinformers.SharedInformerFactory,
|
||||
snapinformers.SharedInformerFactory) {
|
||||
k8sinformers.SharedInformerFactory) {
|
||||
|
||||
f.k8sClient = k8sfake.NewSimpleClientset(f.storageObjects...)
|
||||
f.ksClient = ksfake.NewSimpleClientset(f.capabilityObjects...)
|
||||
f.snapshotClassClient = snapfake.NewSimpleClientset(f.snapshotClassObjects...)
|
||||
|
||||
k8sInformers := k8sinformers.NewSharedInformerFactory(f.k8sClient, noReSyncPeriodFunc())
|
||||
ksInformers := ksinformers.NewSharedInformerFactory(f.ksClient, noReSyncPeriodFunc())
|
||||
snapshotInformers := snapinformers.NewSharedInformerFactory(f.snapshotClassClient, noReSyncPeriodFunc())
|
||||
|
||||
c := NewController(
|
||||
f.k8sClient.StorageV1().StorageClasses(),
|
||||
k8sInformers.Storage().V1().StorageClasses(),
|
||||
k8sInformers.Storage().V1beta1().CSIDrivers(),
|
||||
f.snapshotSupported,
|
||||
f.snapshotClassClient.SnapshotV1beta1().VolumeSnapshotClasses(),
|
||||
snapshotInformers.Snapshot().V1beta1().VolumeSnapshotClasses(),
|
||||
)
|
||||
|
||||
for _, storageClass := range f.storageClassLister {
|
||||
_ = k8sInformers.Storage().V1().StorageClasses().Informer().GetIndexer().Add(storageClass)
|
||||
}
|
||||
for _, snapshotClass := range f.snapshotClassLister {
|
||||
_ = snapshotInformers.Snapshot().V1beta1().VolumeSnapshotClasses().Informer().GetIndexer().Add(snapshotClass)
|
||||
}
|
||||
for _, csiDriver := range f.csiDriverLister {
|
||||
_ = k8sInformers.Storage().V1beta1().CSIDrivers().Informer().GetIndexer().Add(csiDriver)
|
||||
}
|
||||
|
||||
return c, k8sInformers, ksInformers, snapshotInformers
|
||||
return c, k8sInformers
|
||||
}
|
||||
|
||||
func (f *fixture) runController(scName string, startInformers bool, expectError bool) {
|
||||
c, k8sI, crdI, snapI := f.newController()
|
||||
c, k8sI := f.newController()
|
||||
|
||||
if startInformers {
|
||||
stopCh := make(chan struct{})
|
||||
defer close(stopCh)
|
||||
k8sI.Start(stopCh)
|
||||
crdI.Start(stopCh)
|
||||
snapI.Start(stopCh)
|
||||
}
|
||||
|
||||
err := c.syncHandler(scName)
|
||||
@@ -127,9 +104,8 @@ func (f *fixture) runController(scName string, startInformers bool, expectError
|
||||
}
|
||||
|
||||
var actions []core.Action
|
||||
actions = append(actions, f.snapshotClassClient.Actions()...)
|
||||
actions = append(actions, f.k8sClient.Actions()...)
|
||||
actions = append(actions, f.ksClient.Actions()...)
|
||||
//actions = append(actions, f.ksClient.Actions()...)
|
||||
filerActions := filterInformerActions(actions)
|
||||
if len(filerActions) != len(f.actions) {
|
||||
f.t.Errorf("count of actions: differ (-got, +want): %s", cmp.Diff(filerActions, f.actions))
|
||||
@@ -150,16 +126,6 @@ func (f *fixture) expectUpdateStorageClassAction(storageClass *storagev1.Storage
|
||||
schema.GroupVersionResource{Resource: "storageclasses"}, storageClass.Namespace, storageClass))
|
||||
}
|
||||
|
||||
func (f *fixture) expectCreateSnapshotClassAction(snapshotClass *snapbeta1.VolumeSnapshotClass) {
|
||||
f.actions = append(f.actions, core.NewCreateAction(
|
||||
schema.GroupVersionResource{Resource: "volumesnapshotclasses"}, snapshotClass.Namespace, snapshotClass))
|
||||
}
|
||||
|
||||
func (f *fixture) expectDeleteSnapshotClassAction(snapshotClass *snapbeta1.VolumeSnapshotClass) {
|
||||
f.actions = append(f.actions, core.NewDeleteAction(
|
||||
schema.GroupVersionResource{Resource: "volumesnapshotclasses"}, snapshotClass.Namespace, snapshotClass.Name))
|
||||
}
|
||||
|
||||
// filterInformerActions filters list and watch actions for testing resources.
|
||||
// Since list and watch don't change resource state we can filter it to lower
|
||||
// nose level in our tests.
|
||||
@@ -240,16 +206,6 @@ func newCSIDriver(name string) *v1beta1.CSIDriver {
|
||||
}
|
||||
}
|
||||
|
||||
func newSnapshotClass(storageClass *storagev1.StorageClass) *snapbeta1.VolumeSnapshotClass {
|
||||
return &snapbeta1.VolumeSnapshotClass{
|
||||
ObjectMeta: v1.ObjectMeta{
|
||||
Name: storageClass.Name,
|
||||
},
|
||||
Driver: storageClass.Provisioner,
|
||||
DeletionPolicy: snapbeta1.VolumeSnapshotContentDelete,
|
||||
}
|
||||
}
|
||||
|
||||
func getKey(sc *storagev1.StorageClass, t *testing.T) string {
|
||||
key, err := cache.DeletionHandlingMetaNamespaceKeyFunc(sc)
|
||||
if err != nil {
|
||||
@@ -263,46 +219,7 @@ func TestCreateStorageClass(t *testing.T) {
|
||||
fixture := newFixture(t, true)
|
||||
storageClass := newStorageClass("csi-example", "csi.example.com")
|
||||
storageClassUpdate := storageClass.DeepCopy()
|
||||
storageClassUpdate.Annotations = map[string]string{annotationSupportSnapshot: "true", annotationSupportClone: "true"}
|
||||
snapshotClass := newSnapshotClass(storageClass)
|
||||
csiDriver := newCSIDriver("csi.example.com")
|
||||
|
||||
// Objects exist
|
||||
fixture.storageObjects = append(fixture.storageObjects, storageClass)
|
||||
fixture.storageClassLister = append(fixture.storageClassLister, storageClass)
|
||||
fixture.csiDriverLister = append(fixture.csiDriverLister, csiDriver)
|
||||
|
||||
// Action expected
|
||||
fixture.expectCreateSnapshotClassAction(snapshotClass)
|
||||
fixture.expectUpdateStorageClassAction(storageClassUpdate)
|
||||
|
||||
// Run test
|
||||
fixture.run(getKey(storageClass, t))
|
||||
}
|
||||
|
||||
func TestDeleteStorageClass(t *testing.T) {
|
||||
storageClass := newStorageClass("csi-example", "csi.example.com")
|
||||
snapshotClass := newSnapshotClass(storageClass)
|
||||
|
||||
fixture := newFixture(t, true)
|
||||
// Object exist
|
||||
fixture.storageObjects = append(fixture.storageObjects, storageClass)
|
||||
fixture.snapshotClassObjects = append(fixture.snapshotClassObjects, snapshotClass)
|
||||
fixture.snapshotClassLister = append(fixture.snapshotClassLister, snapshotClass)
|
||||
|
||||
// Action expected
|
||||
fixture.expectDeleteSnapshotClassAction(snapshotClass)
|
||||
|
||||
// Run test
|
||||
fixture.run(getKey(storageClass, t))
|
||||
}
|
||||
|
||||
func TestCreateStorageClassNotSupportSnapshot(t *testing.T) {
|
||||
// K8S version < 1.17.0
|
||||
fixture := newFixture(t, false)
|
||||
storageClass := newStorageClass("csi-example", "csi.example.com")
|
||||
storageClassUpdate := storageClass.DeepCopy()
|
||||
storageClassUpdate.Annotations = map[string]string{annotationSupportSnapshot: "false", annotationSupportClone: "true"}
|
||||
storageClassUpdate.Annotations = map[string]string{annotationAllowSnapshot: "true", annotationAllowClone: "true"}
|
||||
csiDriver := newCSIDriver("csi.example.com")
|
||||
|
||||
// Objects exist
|
||||
@@ -320,10 +237,9 @@ func TestCreateStorageClassNotSupportSnapshot(t *testing.T) {
|
||||
func TestStorageClassHadAnnotation(t *testing.T) {
|
||||
fixture := newFixture(t, true)
|
||||
storageClass := newStorageClass("csi-example", "csi.example.com")
|
||||
storageClass.Annotations = map[string]string{annotationSupportSnapshot: "false", annotationSupportClone: "false"}
|
||||
storageClass.Annotations = map[string]string{annotationAllowSnapshot: "false", annotationAllowClone: "false"}
|
||||
storageClassUpdate := storageClass.DeepCopy()
|
||||
csiDriver := newCSIDriver("csi.example.com")
|
||||
snapshotClass := newSnapshotClass(storageClass)
|
||||
|
||||
//Object exist
|
||||
fixture.storageObjects = append(fixture.storageObjects, storageClass)
|
||||
@@ -331,7 +247,6 @@ func TestStorageClassHadAnnotation(t *testing.T) {
|
||||
fixture.csiDriverLister = append(fixture.csiDriverLister, csiDriver)
|
||||
|
||||
//Action expected
|
||||
fixture.expectCreateSnapshotClassAction(snapshotClass)
|
||||
fixture.expectUpdateStorageClassAction(storageClassUpdate)
|
||||
|
||||
//Run test
|
||||
@@ -341,36 +256,16 @@ func TestStorageClassHadAnnotation(t *testing.T) {
|
||||
func TestStorageClassHadOneAnnotation(t *testing.T) {
|
||||
fixture := newFixture(t, true)
|
||||
storageClass := newStorageClass("csi-example", "csi.example.com")
|
||||
storageClass.Annotations = map[string]string{annotationSupportSnapshot: "false"}
|
||||
storageClass.Annotations = map[string]string{annotationAllowSnapshot: "false"}
|
||||
storageClassUpdate := storageClass.DeepCopy()
|
||||
storageClassUpdate.Annotations[annotationSupportClone] = "true"
|
||||
storageClassUpdate.Annotations[annotationAllowClone] = "true"
|
||||
csiDriver := newCSIDriver("csi.example.com")
|
||||
snapshotClass := newSnapshotClass(storageClass)
|
||||
|
||||
//object exist
|
||||
//Object exist
|
||||
fixture.storageObjects = append(fixture.storageObjects, storageClass)
|
||||
fixture.storageClassLister = append(fixture.storageClassLister, storageClass)
|
||||
fixture.csiDriverLister = append(fixture.csiDriverLister, csiDriver)
|
||||
|
||||
//Action expected
|
||||
fixture.expectCreateSnapshotClassAction(snapshotClass)
|
||||
fixture.expectUpdateStorageClassAction(storageClassUpdate)
|
||||
|
||||
//Run test
|
||||
fixture.run(getKey(storageClass, t))
|
||||
}
|
||||
|
||||
func TestDeleteCSIDriver(t *testing.T) {
|
||||
fixture := newFixture(t, true)
|
||||
storageClass := newStorageClass("csi-example", "csi.example.com")
|
||||
storageClass.Annotations = map[string]string{annotationSupportSnapshot: "false", annotationSupportClone: "false"}
|
||||
storageClassUpdate := storageClass.DeepCopy()
|
||||
storageClassUpdate.Annotations = map[string]string{}
|
||||
|
||||
//object exist
|
||||
fixture.storageObjects = append(fixture.storageObjects, storageClass)
|
||||
fixture.storageClassLister = append(fixture.storageClassLister, storageClass)
|
||||
|
||||
//Action expected
|
||||
fixture.expectUpdateStorageClassAction(storageClassUpdate)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user