mirror of
https://github.com/edgelesssys/constellation.git
synced 2024-12-15 02:44:24 -05:00
ac1128d07f
* Add common backend for interacting with cryptsetup * Use common cryptsetup backend in bootstrapper * Use common cryptsetup backend in disk-mapper * Use common cryptsetup backend in csi lib --------- Signed-off-by: Daniel Weiße <dw@edgeless.systems>
290 lines
8.0 KiB
Go
290 lines
8.0 KiB
Go
/*
|
|
Copyright (c) Edgeless Systems GmbH
|
|
|
|
SPDX-License-Identifier: AGPL-3.0-only
|
|
*/
|
|
|
|
/*
|
|
Package setup handles setting up rejoinclient and recoveryserver for the disk-mapper.
|
|
|
|
On success of either of these services, the state disk is decrypted and the node is tainted as initialized by updating it's PCRs.
|
|
*/
|
|
package setup
|
|
|
|
import (
|
|
"context"
|
|
"crypto/rand"
|
|
"errors"
|
|
"fmt"
|
|
"io/fs"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"strconv"
|
|
"sync"
|
|
"syscall"
|
|
|
|
"github.com/edgelesssys/constellation/v2/disk-mapper/internal/systemd"
|
|
"github.com/edgelesssys/constellation/v2/internal/attestation"
|
|
"github.com/edgelesssys/constellation/v2/internal/attestation/initialize"
|
|
"github.com/edgelesssys/constellation/v2/internal/attestation/vtpm"
|
|
"github.com/edgelesssys/constellation/v2/internal/constants"
|
|
"github.com/edgelesssys/constellation/v2/internal/crypto"
|
|
"github.com/edgelesssys/constellation/v2/internal/file"
|
|
"github.com/edgelesssys/constellation/v2/internal/logger"
|
|
"github.com/edgelesssys/constellation/v2/internal/nodestate"
|
|
"github.com/spf13/afero"
|
|
"go.uber.org/zap"
|
|
)
|
|
|
|
const (
|
|
keyPath = "/run/cryptsetup-keys.d"
|
|
keyFile = "state.key"
|
|
stateDiskMappedName = "state"
|
|
stateDiskMountPath = "/var/run/state"
|
|
cryptsetupOptions = "cipher=aes-xts-plain64,integrity=hmac-sha256"
|
|
stateInfoPath = stateDiskMountPath + "/constellation/node_state.json"
|
|
msrdonly = 0x1 // same as syscall.MS_RDONLY
|
|
)
|
|
|
|
// Manager handles formatting, mapping, mounting and unmounting of state disks.
|
|
type Manager struct {
|
|
log *logger.Logger
|
|
csp string
|
|
diskPath string
|
|
fs afero.Afero
|
|
mapper DeviceMapper
|
|
mounter Mounter
|
|
config ConfigurationGenerator
|
|
openDevice vtpm.TPMOpenFunc
|
|
}
|
|
|
|
// New initializes a SetupManager with the given parameters.
|
|
func New(log *logger.Logger, csp string, diskPath string, fs afero.Afero,
|
|
mapper DeviceMapper, mounter Mounter, openDevice vtpm.TPMOpenFunc,
|
|
) *Manager {
|
|
return &Manager{
|
|
log: log,
|
|
csp: csp,
|
|
diskPath: diskPath,
|
|
fs: fs,
|
|
mapper: mapper,
|
|
mounter: mounter,
|
|
config: systemd.New(fs),
|
|
openDevice: openDevice,
|
|
}
|
|
}
|
|
|
|
// PrepareExistingDisk requests and waits for a decryption key to remap the encrypted state disk.
|
|
// Once the disk is mapped, the function taints the node as initialized by updating it's PCRs.
|
|
func (s *Manager) PrepareExistingDisk(recover RecoveryDoer) error {
|
|
uuid, err := s.mapper.DiskUUID()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
s.log.With(zap.String("uuid", uuid)).Infof("Preparing existing state disk")
|
|
endpoint := net.JoinHostPort("0.0.0.0", strconv.Itoa(constants.RecoveryPort))
|
|
|
|
passphrase, measurementSecret, err := recover.Do(uuid, endpoint)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to perform recovery: %w", err)
|
|
}
|
|
|
|
if err := s.mapper.MapDisk(stateDiskMappedName, string(passphrase)); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := s.mounter.MkdirAll(stateDiskMountPath, os.ModePerm); err != nil {
|
|
return err
|
|
}
|
|
|
|
// we do not care about cleaning up the mount point on error, since any errors returned here should cause a boot failure
|
|
if err := s.mounter.Mount(filepath.Join("/dev/mapper/", stateDiskMappedName), stateDiskMountPath, "ext4", msrdonly, ""); err != nil {
|
|
return err
|
|
}
|
|
|
|
measurementSalt, err := s.readMeasurementSalt(stateInfoPath)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
clusterID, err := attestation.DeriveClusterID(measurementSecret, measurementSalt)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// taint the node as initialized
|
|
if err := initialize.MarkNodeAsBootstrapped(s.openDevice, clusterID); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := s.saveConfiguration(passphrase); err != nil {
|
|
return err
|
|
}
|
|
|
|
return s.mounter.Unmount(stateDiskMountPath, 0)
|
|
}
|
|
|
|
// PrepareNewDisk prepares an instances state disk by formatting the disk as a LUKS device using a random passphrase.
|
|
func (s *Manager) PrepareNewDisk() error {
|
|
uuid, _ := s.mapper.DiskUUID()
|
|
s.log.With(zap.String("uuid", uuid)).Infof("Preparing new state disk")
|
|
|
|
// generate and save temporary passphrase
|
|
passphrase := make([]byte, crypto.RNGLengthDefault)
|
|
if _, err := rand.Read(passphrase); err != nil {
|
|
return err
|
|
}
|
|
if err := s.saveConfiguration(passphrase); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := s.mapper.FormatDisk(string(passphrase)); err != nil {
|
|
return err
|
|
}
|
|
|
|
return s.mapper.MapDisk(stateDiskMappedName, string(passphrase))
|
|
}
|
|
|
|
func (s *Manager) readMeasurementSalt(path string) ([]byte, error) {
|
|
handler := file.NewHandler(s.fs)
|
|
var state nodestate.NodeState
|
|
if err := handler.ReadJSON(path, &state); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if len(state.MeasurementSalt) != crypto.RNGLengthDefault {
|
|
return nil, errors.New("missing state information to retaint node")
|
|
}
|
|
|
|
return state.MeasurementSalt, nil
|
|
}
|
|
|
|
// saveConfiguration saves the given passphrase and cryptsetup mapping configuration to disk.
|
|
func (s *Manager) saveConfiguration(passphrase []byte) error {
|
|
// passphrase
|
|
if err := s.fs.MkdirAll(keyPath, os.ModePerm); err != nil {
|
|
return err
|
|
}
|
|
if err := s.fs.WriteFile(filepath.Join(keyPath, keyFile), passphrase, 0o400); err != nil {
|
|
return err
|
|
}
|
|
|
|
// systemd cryptsetup unit
|
|
return s.config.Generate(stateDiskMappedName, s.diskPath, filepath.Join(keyPath, keyFile), cryptsetupOptions)
|
|
}
|
|
|
|
// LogDevices logs all available block devices and partitions (lsblk like).
|
|
func (s *Manager) LogDevices() error {
|
|
var devices []fs.FileInfo
|
|
dirs, err := os.ReadDir("/sys/class/block")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
for _, file := range dirs {
|
|
if file.IsDir() {
|
|
continue
|
|
}
|
|
fileInfo, err := file.Info()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
devices = append(devices, fileInfo)
|
|
}
|
|
|
|
s.log.Infof("List of all available block devices and partitions:")
|
|
for _, device := range devices {
|
|
var stat syscall.Statfs_t
|
|
dev := "/dev/" + device.Name()
|
|
if err := syscall.Statfs(dev, &stat); err != nil {
|
|
s.log.With(zap.Error(err)).Errorf("failed to statfs %s", dev)
|
|
continue
|
|
}
|
|
|
|
// get the raw size, in bytes
|
|
size := stat.Blocks * uint64(stat.Bsize)
|
|
free := stat.Bfree * uint64(stat.Bsize)
|
|
avail := stat.Bavail * uint64(stat.Bsize)
|
|
|
|
s.log.Infof(
|
|
"Name: %-15s, Size: %-10d, Mode: %s, ModTime: %s, Size = %-10d, Free = %-10d, Available = %-10d\n",
|
|
dev,
|
|
device.Size(),
|
|
device.Mode(),
|
|
device.ModTime(),
|
|
size,
|
|
free,
|
|
avail)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// RecoveryServer interface serves a recovery server.
|
|
type RecoveryServer interface {
|
|
Serve(context.Context, net.Listener, string) (key, secret []byte, err error)
|
|
}
|
|
|
|
// RejoinClient interface starts a rejoin client.
|
|
type RejoinClient interface {
|
|
Start(context.Context, string) (key, secret []byte)
|
|
}
|
|
|
|
// NodeRecoverer bundles a RecoveryServer and RejoinClient.
|
|
type NodeRecoverer struct {
|
|
recoveryServer RecoveryServer
|
|
rejoinClient RejoinClient
|
|
}
|
|
|
|
// NewNodeRecoverer initializes a new nodeRecoverer.
|
|
func NewNodeRecoverer(recoveryServer RecoveryServer, rejoinClient RejoinClient) *NodeRecoverer {
|
|
return &NodeRecoverer{
|
|
recoveryServer: recoveryServer,
|
|
rejoinClient: rejoinClient,
|
|
}
|
|
}
|
|
|
|
// Do performs a recovery procedure on the given state disk.
|
|
// The method starts a gRPC server to allow manual recovery by a user.
|
|
// At the same time it tries to request a decryption key from all available Constellation control-plane nodes.
|
|
func (r *NodeRecoverer) Do(uuid, endpoint string) (passphrase, measurementSecret []byte, err error) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
|
|
lis, err := net.Listen("tcp", endpoint)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
defer lis.Close()
|
|
|
|
var once sync.Once
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
key, secret, serveErr := r.recoveryServer.Serve(ctx, lis, uuid)
|
|
once.Do(func() {
|
|
cancel()
|
|
passphrase = key
|
|
measurementSecret = secret
|
|
})
|
|
if serveErr != nil && !errors.Is(serveErr, context.Canceled) {
|
|
err = serveErr
|
|
}
|
|
}()
|
|
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
key, secret := r.rejoinClient.Start(ctx, uuid)
|
|
once.Do(func() {
|
|
cancel()
|
|
passphrase = key
|
|
measurementSecret = secret
|
|
})
|
|
}()
|
|
|
|
wg.Wait()
|
|
return passphrase, measurementSecret, err
|
|
}
|