kubekey/pkg/scope/instance.go
24sama b492f1c6d3 feat: add k3s provider e2e test
Signed-off-by: 24sama <jacksama@foxmail.com>
2022-11-01 16:56:10 +08:00

229 lines
7.4 KiB
Go

/*
Copyright 2022 The KubeSphere Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package scope
import (
"context"
"github.com/go-logr/logr"
"github.com/pkg/errors"
corev1 "k8s.io/api/core/v1"
"k8s.io/apimachinery/pkg/types"
"k8s.io/klog/v2/klogr"
clusterv1 "sigs.k8s.io/cluster-api/api/v1beta1"
bootstrapv1 "sigs.k8s.io/cluster-api/bootstrap/kubeadm/api/v1beta1"
"sigs.k8s.io/cluster-api/util/conditions"
"sigs.k8s.io/cluster-api/util/patch"
"sigs.k8s.io/controller-runtime/pkg/client"
infrav1 "github.com/kubesphere/kubekey/api/v1beta1"
"github.com/kubesphere/kubekey/pkg"
)
// InstanceScopeParams defines the input parameters used to create a new InstanceScope.
type InstanceScopeParams struct {
Client client.Client
Logger *logr.Logger
Cluster *clusterv1.Cluster
InfraCluster *ClusterScope
Machine *clusterv1.Machine
KKInstance *infrav1.KKInstance
}
// NewInstanceScope creates a new InstanceScope from the supplied parameters.
// This is meant to be called for each reconcile iteration.
func NewInstanceScope(params InstanceScopeParams) (*InstanceScope, error) {
if params.Client == nil {
return nil, errors.New("client is required when creating a InstanceScope")
}
if params.Cluster == nil {
return nil, errors.New("cluster is required when creating a InstanceScope")
}
if params.Machine == nil {
return nil, errors.New("machine is required when creating a InstanceScope")
}
if params.InfraCluster == nil {
return nil, errors.New("kk cluster is required when creating a InstanceScope")
}
if params.KKInstance == nil {
return nil, errors.New("kk instance is required when creating a InstanceScope")
}
if params.Logger == nil {
log := klogr.New().WithValues("cluster", params.Cluster.Name, "machine", params.Machine.Name,
"kkInstance", params.KKInstance.Name)
params.Logger = &log
}
helper, err := patch.NewHelper(params.KKInstance, params.Client)
if err != nil {
return nil, errors.Wrap(err, "failed to init patch helper")
}
return &InstanceScope{
Logger: *params.Logger,
client: params.Client,
patchHelper: helper,
Cluster: params.Cluster,
Machine: params.Machine,
InfraCluster: params.InfraCluster,
KKInstance: params.KKInstance,
}, nil
}
// InstanceScope defines a scope defined around a machine instance and its cluster.
type InstanceScope struct {
logr.Logger
client client.Client
patchHelper *patch.Helper
Cluster *clusterv1.Cluster
InfraCluster pkg.ClusterScoper
Machine *clusterv1.Machine
KKMachine *infrav1.KKMachine
KKInstance *infrav1.KKInstance
}
// Name returns the name of the KKInstance.
func (i *InstanceScope) Name() string {
return i.KKInstance.Name
}
// HostName returns the hostname of the KKInstance corresponding remote host.
func (i *InstanceScope) HostName() string {
return i.KKInstance.Spec.Name
}
// Namespace returns the namespace of the KKInstance.
func (i *InstanceScope) Namespace() string {
return i.KKInstance.Namespace
}
// InternalAddress returns the internal address of the KKInstance.
func (i *InstanceScope) InternalAddress() string {
return i.KKInstance.Spec.InternalAddress
}
// Arch returns the architecture of the KKInstance.
func (i *InstanceScope) Arch() string {
return i.KKInstance.Spec.Arch
}
// ContainerManager returns the ContainerManager struct of the KKInstance.
func (i *InstanceScope) ContainerManager() *infrav1.ContainerManager {
return &i.KKInstance.Spec.ContainerManager
}
// RepositoryEnabled returns whether the repository is enabled.
func (i *InstanceScope) RepositoryEnabled() bool {
return i.KKInstance.Spec.Repository != nil && i.KKInstance.Spec.Repository.ISO != ""
}
// RepositoryUseISO returns whether the repository uses ISO.
func (i *InstanceScope) RepositoryUseISO() bool {
return i.RepositoryEnabled() && i.KKInstance.Spec.Repository.ISO != infrav1.NONE
}
// Repository returns the repository of the KKInstance.
func (i *InstanceScope) Repository() *infrav1.Repository {
return i.KKInstance.Spec.Repository
}
// KubernetesVersion returns the Kubernetes version of the KKInstance.
func (i *InstanceScope) KubernetesVersion() string {
return *i.Machine.Spec.Version
}
// IsControlPlane returns whether the KKInstance is a control plane node.
func (i *InstanceScope) IsControlPlane() bool {
if _, ok := i.Machine.GetLabels()[clusterv1.MachineControlPlaneLabelName]; ok {
return true
}
return false
}
// GetRawBootstrapDataWithFormat returns the raw bootstrap data from the corresponding machine.spec.bootstrap.
func (i *InstanceScope) GetRawBootstrapDataWithFormat(ctx context.Context) ([]byte, bootstrapv1.Format, error) {
if i.Machine.Spec.Bootstrap.DataSecretName == nil {
return nil, "", errors.New("error retrieving bootstrap data: linked Machine's bootstrap.dataSecretName is nil")
}
secret := &corev1.Secret{}
key := types.NamespacedName{Namespace: i.Machine.Namespace, Name: *i.Machine.Spec.Bootstrap.DataSecretName}
if err := i.client.Get(ctx, key, secret); err != nil {
return nil, "", errors.Wrapf(err, "failed to retrieve bootstrap data secret for KKInstance %s/%s", i.Namespace(), i.Name())
}
value, ok := secret.Data["value"]
if !ok {
return nil, "", errors.New("error retrieving bootstrap data: secret value key is missing")
}
format := secret.Data["format"]
if string(format) == "" {
format = []byte(bootstrapv1.CloudConfig)
}
return value, bootstrapv1.Format(format), nil
}
// PatchObject persists the machine spec and status.
func (i *InstanceScope) PatchObject() error {
// Always update the readyCondition by summarizing the state of other conditions.
// A step counter is added to represent progress during the provisioning process (instead we are hiding during the deletion process).
applicableConditions := []clusterv1.ConditionType{
infrav1.KKInstanceBootstrappedCondition,
infrav1.KKInstanceBinariesReadyCondition,
infrav1.KKInstanceCRIReadyCondition,
infrav1.KKInstanceProvisionedCondition,
}
conditions.SetSummary(i.KKInstance,
conditions.WithConditions(applicableConditions...),
conditions.WithStepCounterIf(i.KKInstance.ObjectMeta.DeletionTimestamp.IsZero()),
conditions.WithStepCounter(),
)
return i.patchHelper.Patch(
context.TODO(),
i.KKInstance,
patch.WithOwnedConditions{Conditions: []clusterv1.ConditionType{
clusterv1.ReadyCondition,
infrav1.KKInstanceBootstrappedCondition,
infrav1.KKInstanceBinariesReadyCondition,
infrav1.KKInstanceCRIReadyCondition,
infrav1.KKInstanceProvisionedCondition,
infrav1.KKInstanceDeletingBootstrapCondition,
}})
}
// Close the InstanceScope by updating the instance spec, instance status.
func (i *InstanceScope) Close() error {
return i.PatchObject()
}
// HasFailed returns the failure state of the instance scope.
func (i *InstanceScope) HasFailed() bool {
return i.KKInstance.Status.FailureReason != nil || i.KKInstance.Status.FailureMessage != nil
}
// SetState sets the state of the KKInstance.
func (i *InstanceScope) SetState(state infrav1.InstanceState) {
i.KKInstance.Status.State = state
}