зеркало из https://github.com/Azure/aks-engine.git
303 строки
8.4 KiB
Go
303 строки
8.4 KiB
Go
// Copyright (c) Microsoft Corporation. All rights reserved.
|
|
// Licensed under the MIT license.
|
|
|
|
package main
|
|
|
|
import (
|
|
"fmt"
|
|
"log"
|
|
"os"
|
|
"os/signal"
|
|
"path/filepath"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/Azure/aks-engine/pkg/api/common"
|
|
"github.com/Azure/aks-engine/test/e2e/azure"
|
|
"github.com/Azure/aks-engine/test/e2e/config"
|
|
"github.com/Azure/aks-engine/test/e2e/engine"
|
|
"github.com/Azure/aks-engine/test/e2e/metrics"
|
|
"github.com/Azure/aks-engine/test/e2e/runner"
|
|
"github.com/Azure/go-autorest/autorest/to"
|
|
)
|
|
|
|
var (
|
|
cfg *config.Config
|
|
cccfg *config.CustomCloudConfig
|
|
acct *azure.Account
|
|
eng *engine.Engine
|
|
rgs []string
|
|
err error
|
|
pt *metrics.Point
|
|
cliProvisioner *runner.CLIProvisioner
|
|
)
|
|
|
|
func main() {
|
|
cwd, _ := os.Getwd()
|
|
cfg, err = config.ParseConfig()
|
|
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to parse configuration: %s\n", err)
|
|
}
|
|
cfg.CurrentWorkingDir = cwd
|
|
|
|
if cfg.IsCustomCloudProfile() {
|
|
cccfg, err = config.ParseCustomCloudConfig()
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to parse custom cloud configuration: %s\n", err)
|
|
}
|
|
if cfg.Name == "" {
|
|
err = cfg.UpdateCustomCloudClusterDefinition(cccfg)
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to update cluster definition: %s\n", cfg.ClusterDefinition)
|
|
}
|
|
}
|
|
cccfg.SetEnvironment()
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to set environment to azure account! %s\n", err)
|
|
}
|
|
}
|
|
|
|
acct, err = azure.NewAccount()
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to setup azure account: %s\n", err)
|
|
}
|
|
|
|
err := acct.Login()
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to login to azure account! %s\n", err)
|
|
}
|
|
|
|
err = acct.SetSubscription()
|
|
if err != nil {
|
|
log.Fatal("Error while trying to set azure subscription!")
|
|
}
|
|
pt = metrics.BuildPoint(cfg.Orchestrator, cfg.Location, cfg.ClusterDefinition, acct.SubscriptionID)
|
|
|
|
// If an interrupt/kill signal is sent we will run the clean up procedure
|
|
trap()
|
|
|
|
cliProvisioner, err = runner.BuildCLIProvisioner(cfg, acct, pt)
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to build CLI Provisioner:%s", err)
|
|
}
|
|
if cliProvisioner.Config.Name != "" {
|
|
// Store the hosts for future introspection
|
|
hosts, err := cliProvisioner.Account.GetHosts(cliProvisioner.Config.Name)
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to get hosts in resource group:%s", err)
|
|
}
|
|
var masters, agents []azure.VM
|
|
for _, host := range hosts {
|
|
if strings.Contains(host.Name, "master") {
|
|
masters = append(masters, host)
|
|
} else if strings.Contains(host.Name, "agent") {
|
|
agents = append(agents, host)
|
|
}
|
|
}
|
|
cliProvisioner.Masters = masters
|
|
cliProvisioner.Agents = agents
|
|
}
|
|
|
|
sa := acct.StorageAccount
|
|
|
|
// Soak test specific setup
|
|
if cfg.SoakClusterName != "" {
|
|
sa.Name = "acsesoaktests" + cfg.Location
|
|
sa.ResourceGroup.Name = "acse-test-infrastructure-storage"
|
|
sa.ResourceGroup.Location = cfg.Location
|
|
err = sa.CreateStorageAccount()
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to create storage account: %s\n", err)
|
|
}
|
|
err = sa.SetConnectionString()
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to set storage account connection string: %s\n", err)
|
|
}
|
|
provision := true
|
|
rgExists := true
|
|
rg := cfg.SoakClusterName
|
|
err = acct.SetResourceGroup(rg)
|
|
if err != nil {
|
|
rgExists = false
|
|
log.Printf("Error while trying to set RG:%s\n", err)
|
|
} else {
|
|
// set expiration time to 7 days = 168h for now
|
|
var d time.Duration
|
|
d, err = time.ParseDuration("168h")
|
|
if err != nil {
|
|
log.Fatalf("Unexpected error parsing duration: %s", err)
|
|
}
|
|
provision = acct.IsClusterExpired(d)
|
|
}
|
|
if provision || cfg.ForceDeploy {
|
|
log.Printf("Soak cluster %s does not exist or has expired\n", rg)
|
|
if rgExists {
|
|
log.Printf("Deleting Resource Group:%s\n", rg)
|
|
acct.DeleteGroup(rg, true)
|
|
}
|
|
log.Printf("Deleting Storage files:%s\n", rg)
|
|
sa.DeleteFiles(cfg.SoakClusterName)
|
|
cfg.Name = ""
|
|
} else {
|
|
log.Printf("Soak cluster %s exists, downloading output files from storage...\n", rg)
|
|
err = sa.DownloadFiles(cfg.SoakClusterName, "_output")
|
|
if err != nil {
|
|
log.Printf("Error while trying to download _output dir: %s, will provision a new cluster.\n", err)
|
|
log.Printf("Deleting Resource Group:%s\n", rg)
|
|
acct.DeleteGroup(rg, true)
|
|
log.Printf("Deleting Storage files:%s\n", rg)
|
|
sa.DeleteFiles(cfg.SoakClusterName)
|
|
cfg.Name = ""
|
|
} else {
|
|
cfg.SetSSHKeyPermissions()
|
|
}
|
|
}
|
|
}
|
|
// Only provision a cluster if there isn't a name present
|
|
if cfg.Name == "" {
|
|
err = cliProvisioner.Run()
|
|
rgs = cliProvisioner.ResourceGroups
|
|
eng = cliProvisioner.Engine
|
|
if err != nil {
|
|
if cfg.CleanUpIfFail {
|
|
teardown()
|
|
}
|
|
log.Fatalf("Error while trying to provision cluster:%s", err)
|
|
}
|
|
if cfg.SoakClusterName != "" {
|
|
err = sa.CreateFileShare(cfg.SoakClusterName)
|
|
if err != nil {
|
|
log.Printf("Error while trying to create file share:%s\n", err)
|
|
}
|
|
err = sa.UploadFiles(filepath.Join(cfg.CurrentWorkingDir, "_output"), cfg.SoakClusterName)
|
|
if err != nil {
|
|
log.Fatalf("Error while trying to upload _output dir:%s\n", err)
|
|
}
|
|
}
|
|
} else {
|
|
rgs = append(rgs, cliProvisioner.Config.Name)
|
|
engCfg, err := engine.ParseConfig(cfg.CurrentWorkingDir, cfg.ClusterDefinition, cfg.Name)
|
|
cfg.SetKubeConfig()
|
|
if err != nil {
|
|
if cfg.CleanUpIfFail {
|
|
teardown()
|
|
}
|
|
log.Fatalf("Error trying to parse Engine config:%s\n", err)
|
|
}
|
|
cs, err := engine.ParseInput(engCfg.ClusterDefinitionTemplate)
|
|
if err != nil {
|
|
if cfg.CleanUpIfFail {
|
|
teardown()
|
|
}
|
|
log.Fatalf("Error trying to parse engine template into memory:%s\n", err)
|
|
}
|
|
eng = &engine.Engine{
|
|
Config: engCfg,
|
|
ClusterDefinition: cs,
|
|
}
|
|
cliProvisioner.Engine = eng
|
|
}
|
|
|
|
if !cfg.SkipTest {
|
|
var resourceGroup string
|
|
if cliProvisioner.Account.ResourceGroup.Name != "" {
|
|
resourceGroup = cliProvisioner.Account.ResourceGroup.Name
|
|
} else {
|
|
resourceGroup = cfg.Name
|
|
}
|
|
|
|
if resourceGroup == "" {
|
|
if cfg.CleanUpIfFail {
|
|
teardown()
|
|
}
|
|
log.Fatalf("Resource Group is empty")
|
|
}
|
|
|
|
os.Setenv("RESOURCE_GROUP", resourceGroup)
|
|
|
|
g, err := runner.BuildGinkgoRunner(cfg, pt)
|
|
if err != nil {
|
|
if cfg.CleanUpIfFail {
|
|
teardown()
|
|
}
|
|
log.Fatalf("Error: Unable to parse ginkgo configuration!")
|
|
}
|
|
err = g.Run()
|
|
if err != nil {
|
|
if cfg.CleanUpIfFail {
|
|
teardown()
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
}
|
|
|
|
teardown()
|
|
os.Exit(0)
|
|
}
|
|
|
|
func trap() {
|
|
// If an interrupt/kill signal is sent we will run the clean up procedure
|
|
c := make(chan os.Signal, 1)
|
|
signal.Notify(c, os.Interrupt)
|
|
signal.Notify(c, os.Kill)
|
|
go func() {
|
|
for sig := range c {
|
|
log.Printf("Received Signal:%s ... Clean Up On Exit?:%v\n", sig.String(), cfg.CleanUpOnExit)
|
|
teardown()
|
|
os.Exit(1)
|
|
}
|
|
}()
|
|
}
|
|
|
|
func teardown() {
|
|
pt.RecordTotalTime()
|
|
pt.Write()
|
|
hostname := fmt.Sprintf("%s.%s.cloudapp.azure.com", cfg.Name, cfg.Location)
|
|
logsPath := filepath.Join(cfg.CurrentWorkingDir, "_logs", hostname)
|
|
err := os.MkdirAll(logsPath, 0755)
|
|
if err != nil {
|
|
log.Printf("cannot create directory for logs: %s", err)
|
|
}
|
|
|
|
if cfg.SoakClusterName == "" && !cfg.SkipLogsCollection {
|
|
err = cliProvisioner.FetchProvisioningMetrics(logsPath, cfg, acct)
|
|
if err != nil {
|
|
log.Printf("cliProvisioner.FetchProvisioningMetrics error: %s\n", err)
|
|
}
|
|
}
|
|
if !cfg.SkipLogsCollection {
|
|
if err := cliProvisioner.FetchActivityLog(acct, logsPath); err != nil {
|
|
log.Printf("cannot fetch the activity log: %v", err)
|
|
}
|
|
}
|
|
if !cfg.RetainSSH {
|
|
creds := filepath.Join(cfg.CurrentWorkingDir, "_output/", "*ssh*")
|
|
files, err := filepath.Glob(creds)
|
|
if err != nil {
|
|
log.Printf("failed to get ssh files using %s: %s\n", creds, err)
|
|
}
|
|
for _, file := range files {
|
|
err := os.Remove(file)
|
|
if err != nil {
|
|
log.Printf("failed to delete file %s: %s\n", file, err)
|
|
}
|
|
}
|
|
}
|
|
if cfg.CleanUpOnExit {
|
|
for _, rg := range rgs {
|
|
log.Printf("Deleting Group: %s\n", rg)
|
|
acct.DeleteGroup(rg, false)
|
|
}
|
|
// Delete once we reuse the cluster group for the connectedCluster resource
|
|
for _, addon := range eng.ClusterDefinition.Properties.OrchestratorProfile.KubernetesConfig.Addons {
|
|
if addon.Name == common.AzureArcOnboardingAddonName && to.Bool(addon.Enabled) &&
|
|
addon.Config["resourceGroup"] != "" &&
|
|
addon.Config["location"] != "" {
|
|
log.Printf("Deleting Arc Group: %s\n", fmt.Sprintf("%s-arc", cfg.Name))
|
|
acct.DeleteGroup(fmt.Sprintf("%s-arc", cfg.Name), false)
|
|
}
|
|
}
|
|
}
|
|
}
|