mirror of
https://github.com/edgelesssys/constellation.git
synced 2024-10-01 01:36:09 -04:00
d1ace13713
* Remove `--config` and `--master-secret` falgs * Add `--workspace` flag * In CLI, only work on files with paths created from `cli/internal/cmd` * Properly print values for GCP on IAM create when not directly updating the config --------- Signed-off-by: Daniel Weiße <dw@edgeless.systems>
227 lines
5.6 KiB
Go
227 lines
5.6 KiB
Go
/*
|
|
Copyright (c) Edgeless Systems GmbH
|
|
|
|
SPDX-License-Identifier: AGPL-3.0-only
|
|
*/
|
|
|
|
package terraform
|
|
|
|
import (
|
|
"io/fs"
|
|
"path"
|
|
"path/filepath"
|
|
"strings"
|
|
"testing"
|
|
|
|
"github.com/edgelesssys/constellation/v2/internal/cloud/cloudprovider"
|
|
"github.com/edgelesssys/constellation/v2/internal/file"
|
|
"github.com/spf13/afero"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestPrepareWorkspace(t *testing.T) {
|
|
testCases := map[string]struct {
|
|
pathBase string
|
|
provider cloudprovider.Provider
|
|
fileList []string
|
|
testAlreadyUnpacked bool
|
|
}{
|
|
"awsCluster": {
|
|
pathBase: "terraform",
|
|
provider: cloudprovider.AWS,
|
|
fileList: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
"modules",
|
|
},
|
|
},
|
|
"gcpCluster": {
|
|
pathBase: "terraform",
|
|
provider: cloudprovider.GCP,
|
|
fileList: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
"modules",
|
|
},
|
|
},
|
|
"qemuCluster": {
|
|
pathBase: "terraform",
|
|
provider: cloudprovider.QEMU,
|
|
fileList: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
"modules",
|
|
},
|
|
},
|
|
"gcpIAM": {
|
|
pathBase: path.Join("terraform", "iam"),
|
|
provider: cloudprovider.GCP,
|
|
fileList: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
},
|
|
},
|
|
"azureIAM": {
|
|
pathBase: path.Join("terraform", "iam"),
|
|
provider: cloudprovider.Azure,
|
|
fileList: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
},
|
|
},
|
|
"awsIAM": {
|
|
pathBase: path.Join("terraform", "iam"),
|
|
provider: cloudprovider.AWS,
|
|
fileList: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
},
|
|
},
|
|
"continue on (partially) unpacked": {
|
|
pathBase: "terraform",
|
|
provider: cloudprovider.AWS,
|
|
fileList: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
"modules",
|
|
},
|
|
testAlreadyUnpacked: true,
|
|
},
|
|
}
|
|
|
|
for name, tc := range testCases {
|
|
t.Run(name, func(t *testing.T) {
|
|
assert := assert.New(t)
|
|
require := require.New(t)
|
|
|
|
file := file.NewHandler(afero.NewMemMapFs())
|
|
testWorkspace := "unittest"
|
|
|
|
path := path.Join(tc.pathBase, strings.ToLower(tc.provider.String()))
|
|
err := prepareWorkspace(path, file, testWorkspace)
|
|
|
|
require.NoError(err)
|
|
checkFiles(t, file, func(err error) { assert.NoError(err) }, testWorkspace, tc.fileList)
|
|
|
|
if tc.testAlreadyUnpacked {
|
|
// Let's try the same again and check if we don't get a "file already exists" error.
|
|
require.NoError(file.Remove(filepath.Join(testWorkspace, "variables.tf")))
|
|
err := prepareWorkspace(path, file, testWorkspace)
|
|
assert.NoError(err)
|
|
checkFiles(t, file, func(err error) { assert.NoError(err) }, testWorkspace, tc.fileList)
|
|
}
|
|
|
|
err = cleanUpWorkspace(file, testWorkspace)
|
|
require.NoError(err)
|
|
|
|
checkFiles(t, file, func(err error) { assert.ErrorIs(err, fs.ErrNotExist) }, testWorkspace, tc.fileList)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestPrepareUpgradeWorkspace(t *testing.T) {
|
|
testCases := map[string]struct {
|
|
pathBase string
|
|
provider cloudprovider.Provider
|
|
oldWorkingDir string
|
|
newWorkingDir string
|
|
backupDir string
|
|
oldWorkspaceFiles []string
|
|
newWorkspaceFiles []string
|
|
expectedFiles []string
|
|
testAlreadyUnpacked bool
|
|
wantErr bool
|
|
}{
|
|
"works": {
|
|
pathBase: "terraform",
|
|
provider: cloudprovider.AWS,
|
|
oldWorkingDir: "old",
|
|
newWorkingDir: "new",
|
|
backupDir: "backup",
|
|
oldWorkspaceFiles: []string{"terraform.tfstate"},
|
|
expectedFiles: []string{
|
|
"main.tf",
|
|
"variables.tf",
|
|
"outputs.tf",
|
|
"modules",
|
|
"terraform.tfstate",
|
|
},
|
|
},
|
|
"state file does not exist": {
|
|
pathBase: "terraform",
|
|
provider: cloudprovider.AWS,
|
|
oldWorkingDir: "old",
|
|
newWorkingDir: "new",
|
|
backupDir: "backup",
|
|
oldWorkspaceFiles: []string{},
|
|
expectedFiles: []string{},
|
|
wantErr: true,
|
|
},
|
|
"terraform files already exist in new dir": {
|
|
pathBase: "terraform",
|
|
provider: cloudprovider.AWS,
|
|
oldWorkingDir: "old",
|
|
newWorkingDir: "new",
|
|
backupDir: "backup",
|
|
oldWorkspaceFiles: []string{"terraform.tfstate"},
|
|
newWorkspaceFiles: []string{"main.tf"},
|
|
wantErr: true,
|
|
},
|
|
}
|
|
|
|
for name, tc := range testCases {
|
|
t.Run(name, func(t *testing.T) {
|
|
require := require.New(t)
|
|
assert := assert.New(t)
|
|
|
|
file := file.NewHandler(afero.NewMemMapFs())
|
|
|
|
path := path.Join(tc.pathBase, strings.ToLower(tc.provider.String()))
|
|
|
|
createFiles(t, file, tc.oldWorkspaceFiles, tc.oldWorkingDir)
|
|
createFiles(t, file, tc.newWorkspaceFiles, tc.newWorkingDir)
|
|
|
|
err := prepareUpgradeWorkspace(path, file, tc.oldWorkingDir, tc.newWorkingDir, tc.backupDir)
|
|
|
|
if tc.wantErr {
|
|
require.Error(err)
|
|
} else {
|
|
require.NoError(err)
|
|
}
|
|
checkFiles(t, file, func(err error) { assert.NoError(err) }, tc.newWorkingDir, tc.expectedFiles)
|
|
checkFiles(t, file, func(err error) { assert.NoError(err) },
|
|
tc.backupDir,
|
|
tc.oldWorkspaceFiles,
|
|
)
|
|
})
|
|
}
|
|
}
|
|
|
|
func checkFiles(t *testing.T, fileHandler file.Handler, assertion func(error), dir string, files []string) {
|
|
t.Helper()
|
|
for _, f := range files {
|
|
path := filepath.Join(dir, f)
|
|
_, err := fileHandler.Stat(path)
|
|
assertion(err)
|
|
}
|
|
}
|
|
|
|
func createFiles(t *testing.T, fileHandler file.Handler, fileList []string, targetDir string) {
|
|
t.Helper()
|
|
require := require.New(t)
|
|
|
|
for _, f := range fileList {
|
|
path := filepath.Join(targetDir, f)
|
|
err := fileHandler.Write(path, []byte("1234"), file.OptOverwrite, file.OptMkdirAll)
|
|
require.NoError(err)
|
|
}
|
|
}
|