blob: 271ef99f24974f313e930920aab64bda922cea83 [file] [log] [blame]
package hst
import (
"bufio"
"errors"
"flag"
"fmt"
"io"
"log"
"net/http"
"net/http/httputil"
"os"
"os/exec"
"path/filepath"
"runtime"
"strings"
"time"
containerTypes "github.com/docker/docker/api/types/container"
"github.com/docker/docker/client"
"github.com/onsi/gomega/gmeasure"
"gopkg.in/yaml.v3"
. "github.com/onsi/ginkgo/v2"
. "github.com/onsi/gomega"
)
const (
DEFAULT_NETWORK_NUM int = 1
)
var IsPersistent = flag.Bool("persist", false, "persists topology config")
var IsVerbose = flag.Bool("verbose", false, "verbose test output")
var IsUnconfiguring = flag.Bool("unconfigure", false, "remove topology")
var IsVppDebug = flag.Bool("debug", false, "attach gdb to vpp")
var NConfiguredCpus = flag.Int("cpus", 1, "number of CPUs assigned to vpp")
var VppSourceFileDir = flag.String("vppsrc", "", "vpp source file directory")
var IsDebugBuild = flag.Bool("debug_build", false, "some paths are different with debug build")
var UseCpu0 = flag.Bool("cpu0", false, "use cpu0")
var IsLeakCheck = flag.Bool("leak_check", false, "run leak-check tests")
var ParallelTotal = flag.Lookup("ginkgo.parallel.total")
var NumaAwareCpuAlloc bool
var SuiteTimeout time.Duration
type HstSuite struct {
Containers map[string]*Container
StartedContainers []*Container
Volumes []string
NetConfigs []NetConfig
NetInterfaces map[string]*NetInterface
Ip4AddrAllocator *Ip4AddressAllocator
TestIds map[string]string
CpuAllocator *CpuAllocatorT
CpuContexts []*CpuContext
CpuCount int
Ppid string
ProcessIndex string
Logger *log.Logger
LogFile *os.File
Docker *client.Client
}
// used for colorful ReportEntry
type StringerStruct struct {
Label string
}
// ColorableString for ReportEntry to use
func (s StringerStruct) ColorableString() string {
return fmt.Sprintf("{{red}}%s{{/}}", s.Label)
}
// non-colorable String() is used by go's string formatting support but ignored by ReportEntry
func (s StringerStruct) String() string {
return s.Label
}
func getTestFilename() string {
_, filename, _, _ := runtime.Caller(2)
return filepath.Base(filename)
}
func (s *HstSuite) getLogDirPath() string {
testId := s.GetTestId()
testName := s.GetCurrentTestName()
logDirPath := logDir + testName + "/" + testId + "/"
cmd := exec.Command("mkdir", "-p", logDirPath)
if err := cmd.Run(); err != nil {
Fail("mkdir error: " + fmt.Sprint(err))
}
return logDirPath
}
func (s *HstSuite) newDockerClient() {
var err error
s.Docker, err = client.NewClientWithOpts(client.FromEnv, client.WithAPIVersionNegotiation())
s.AssertNil(err)
s.Log("docker client created")
}
func (s *HstSuite) SetupSuite() {
s.CreateLogger()
s.newDockerClient()
s.Log("Suite Setup")
RegisterFailHandler(func(message string, callerSkip ...int) {
s.HstFail()
Fail(message, callerSkip...)
})
var err error
s.Ppid = fmt.Sprint(os.Getppid())
// remove last number so we have space to prepend a process index (interfaces have a char limit)
s.Ppid = s.Ppid[:len(s.Ppid)-1]
s.ProcessIndex = fmt.Sprint(GinkgoParallelProcess())
s.CpuAllocator, err = CpuAllocator()
if err != nil {
Fail("failed to init cpu allocator: " + fmt.Sprint(err))
}
s.CpuCount = *NConfiguredCpus
}
func (s *HstSuite) AllocateCpus() []int {
cpuCtx, err := s.CpuAllocator.Allocate(len(s.StartedContainers), s.CpuCount)
// using Fail instead of AssertNil to make error message more readable
if err != nil {
Fail(fmt.Sprint(err))
}
s.AddCpuContext(cpuCtx)
return cpuCtx.cpus
}
func (s *HstSuite) AddCpuContext(cpuCtx *CpuContext) {
s.CpuContexts = append(s.CpuContexts, cpuCtx)
}
func (s *HstSuite) TearDownSuite() {
defer s.LogFile.Close()
defer s.Docker.Close()
s.Log("Suite Teardown")
s.UnconfigureNetworkTopology()
}
func (s *HstSuite) TearDownTest() {
s.Log("Test Teardown")
if *IsPersistent {
return
}
s.WaitForCoreDump()
s.ResetContainers()
if s.Ip4AddrAllocator != nil {
s.Ip4AddrAllocator.DeleteIpAddresses()
}
}
func (s *HstSuite) SkipIfUnconfiguring() {
if *IsUnconfiguring {
s.Skip("skipping to unconfigure")
}
}
func (s *HstSuite) SetupTest() {
s.Log("Test Setup")
s.StartedContainers = s.StartedContainers[:0]
s.SkipIfUnconfiguring()
s.SetupContainers()
}
func (s *HstSuite) SetupContainers() {
for _, container := range s.Containers {
if !container.IsOptional {
container.Run()
}
}
}
func (s *HstSuite) LogVppInstance(container *Container, maxLines int) {
if container.VppInstance == nil {
return
}
logSource := container.GetHostWorkDir() + defaultLogFilePath
file, err := os.Open(logSource)
if err != nil {
return
}
defer file.Close()
scanner := bufio.NewScanner(file)
var lines []string
var counter int
for scanner.Scan() {
lines = append(lines, scanner.Text())
counter++
if counter > maxLines {
lines = lines[1:]
counter--
}
}
s.Log("vvvvvvvvvvvvvvv " + container.Name + " [VPP instance]:")
for _, line := range lines {
s.Log(line)
}
s.Log("^^^^^^^^^^^^^^^\n\n")
}
func (s *HstSuite) HstFail() {
for _, container := range s.StartedContainers {
out, err := container.log(20)
if err != nil {
s.Log("An error occured while obtaining '" + container.Name + "' container logs: " + fmt.Sprint(err))
s.Log("The container might not be running - check logs in " + s.getLogDirPath())
continue
}
s.Log("\nvvvvvvvvvvvvvvv " +
container.Name + ":\n" +
out +
"^^^^^^^^^^^^^^^\n\n")
s.LogVppInstance(container, 20)
}
}
func (s *HstSuite) AssertNil(object interface{}, msgAndArgs ...interface{}) {
Expect(object).To(BeNil(), msgAndArgs...)
}
func (s *HstSuite) AssertNotNil(object interface{}, msgAndArgs ...interface{}) {
Expect(object).ToNot(BeNil(), msgAndArgs...)
}
func (s *HstSuite) AssertEqual(expected, actual interface{}, msgAndArgs ...interface{}) {
Expect(actual).To(Equal(expected), msgAndArgs...)
}
func (s *HstSuite) AssertNotEqual(expected, actual interface{}, msgAndArgs ...interface{}) {
Expect(actual).ToNot(Equal(expected), msgAndArgs...)
}
func (s *HstSuite) AssertContains(testString, contains interface{}, msgAndArgs ...interface{}) {
Expect(testString).To(ContainSubstring(fmt.Sprint(contains)), msgAndArgs...)
}
func (s *HstSuite) AssertNotContains(testString, contains interface{}, msgAndArgs ...interface{}) {
Expect(testString).ToNot(ContainSubstring(fmt.Sprint(contains)), msgAndArgs...)
}
func (s *HstSuite) AssertEmpty(object interface{}, msgAndArgs ...interface{}) {
Expect(object).To(BeEmpty(), msgAndArgs...)
}
func (s *HstSuite) AssertNotEmpty(object interface{}, msgAndArgs ...interface{}) {
Expect(object).ToNot(BeEmpty(), msgAndArgs...)
}
func (s *HstSuite) AssertMatchError(actual, expected error, msgAndArgs ...interface{}) {
Expect(actual).To(MatchError(expected))
}
func (s *HstSuite) CreateLogger() {
suiteName := s.GetCurrentSuiteName()
var err error
s.LogFile, err = os.Create("summary/" + suiteName + ".log")
if err != nil {
Fail("Unable to create log file.")
}
s.Logger = log.New(io.Writer(s.LogFile), "", log.LstdFlags)
}
// Logs to files by default, logs to stdout when VERBOSE=true with GinkgoWriter
// to keep console tidy
func (s *HstSuite) Log(arg any) {
logs := strings.Split(fmt.Sprint(arg), "\n")
for _, line := range logs {
s.Logger.Println(line)
}
if *IsVerbose {
GinkgoWriter.Println(arg)
}
}
func (s *HstSuite) Skip(args string) {
Skip(args)
}
func (s *HstSuite) SkipIfMultiWorker(args ...any) {
if *NConfiguredCpus > 1 {
s.Skip("test case not supported with multiple vpp workers")
}
}
func (s *HstSuite) SkipIfNotEnoughAvailableCpus() bool {
var MaxRequestedCpu int
if s.CpuAllocator.runningInCi {
MaxRequestedCpu = ((s.CpuAllocator.buildNumber + 1) * s.CpuAllocator.maxContainerCount * s.CpuCount)
} else {
MaxRequestedCpu = (GinkgoParallelProcess() * s.CpuAllocator.maxContainerCount * s.CpuCount)
}
if len(s.CpuAllocator.cpus)-1 < MaxRequestedCpu {
s.Skip(fmt.Sprintf("test case cannot allocate requested cpus (%d cpus * %d containers)", s.CpuCount, s.CpuAllocator.maxContainerCount))
}
return true
}
func (s *HstSuite) SkipUnlessExtendedTestsBuilt() {
imageName := "hs-test/nginx-http3"
cmd := exec.Command("docker", "images", imageName)
byteOutput, err := cmd.CombinedOutput()
if err != nil {
s.Log("error while searching for docker image")
return
}
if !strings.Contains(string(byteOutput), imageName) {
s.Skip("extended tests not built")
}
}
func (s *HstSuite) SkipUnlessLeakCheck() {
if !*IsLeakCheck {
s.Skip("leak-check tests excluded")
}
}
func (s *HstSuite) WaitForCoreDump() {
var filename string
var cmd *exec.Cmd
dir, err := os.Open(s.getLogDirPath())
if err != nil {
s.Log(err)
return
}
defer dir.Close()
files, err := dir.Readdirnames(0)
if err != nil {
s.Log(err)
return
}
for _, file := range files {
if strings.Contains(file, "core") {
filename = file
}
}
timeout := 60
waitTime := 5
if filename != "" {
filename := s.getLogDirPath() + filename
s.Log(fmt.Sprintf("WAITING FOR CORE DUMP (%s)", filename))
for i := waitTime; i <= timeout; i += waitTime {
fileInfo, err := os.Stat(filename)
if err != nil {
s.Log("Error while reading file info: " + fmt.Sprint(err))
return
}
currSize := fileInfo.Size()
s.Log(fmt.Sprintf("Waiting %ds/%ds...", i, timeout))
time.Sleep(time.Duration(waitTime) * time.Second)
fileInfo, _ = os.Stat(filename)
if currSize == fileInfo.Size() {
if *IsDebugBuild {
cmd = exec.Command("sudo", "gdb", "../../build-root/build-vpp_debug-native/vpp/bin/vpp",
"-c", filename, "-ex", "bt", "full", "-ex", "quit")
} else {
cmd = exec.Command("sudo", "gdb", "../../build-root/build-vpp-native/vpp/bin/vpp",
"-c", filename, "-ex", "bt", "full", "-ex", "quit")
}
s.Log(cmd)
output, _ := cmd.Output()
AddReportEntry("VPP Backtrace", StringerStruct{Label: string(output)})
os.WriteFile(s.getLogDirPath()+"backtrace.log", output, os.FileMode(0644))
if s.CpuAllocator.runningInCi {
err = os.Remove(filename)
if err == nil {
s.Log("removed " + filename)
} else {
s.Log(err)
}
}
return
}
}
}
}
func (s *HstSuite) ResetContainers() {
for _, container := range s.StartedContainers {
container.stop()
s.Log("Removing container " + container.Name)
if err := s.Docker.ContainerRemove(container.ctx, container.ID, containerTypes.RemoveOptions{RemoveVolumes: true}); err != nil {
s.Log(err)
}
}
}
func (s *HstSuite) GetNetNamespaceByName(name string) string {
return s.ProcessIndex + name + s.Ppid
}
func (s *HstSuite) GetInterfaceByName(name string) *NetInterface {
return s.NetInterfaces[s.ProcessIndex+name+s.Ppid]
}
func (s *HstSuite) GetContainerByName(name string) *Container {
return s.Containers[s.ProcessIndex+name+s.Ppid]
}
/*
* Create a copy and return its address, so that individial tests which call this
* are not able to modify the original container and affect other tests by doing that
*/
func (s *HstSuite) GetTransientContainerByName(name string) *Container {
containerCopy := *s.Containers[s.ProcessIndex+name+s.Ppid]
return &containerCopy
}
func (s *HstSuite) LoadContainerTopology(topologyName string) {
data, err := os.ReadFile(containerTopologyDir + topologyName + ".yaml")
if err != nil {
Fail("read error: " + fmt.Sprint(err))
}
var yamlTopo YamlTopology
err = yaml.Unmarshal(data, &yamlTopo)
if err != nil {
Fail("unmarshal error: " + fmt.Sprint(err))
}
for _, elem := range yamlTopo.Volumes {
volumeMap := elem["volume"].(VolumeConfig)
hostDir := volumeMap["host-dir"].(string)
workingVolumeDir := logDir + s.GetCurrentTestName() + volumeDir
volDirReplacer := strings.NewReplacer("$HST_VOLUME_DIR", workingVolumeDir)
hostDir = volDirReplacer.Replace(hostDir)
s.Volumes = append(s.Volumes, hostDir)
}
s.Containers = make(map[string]*Container)
for _, elem := range yamlTopo.Containers {
newContainer, err := newContainer(s, elem)
newContainer.Suite = s
newContainer.Name = newContainer.Suite.ProcessIndex + newContainer.Name + newContainer.Suite.Ppid
if err != nil {
Fail("container config error: " + fmt.Sprint(err))
}
s.Containers[newContainer.Name] = newContainer
}
}
func (s *HstSuite) LoadNetworkTopology(topologyName string) {
data, err := os.ReadFile(networkTopologyDir + topologyName + ".yaml")
if err != nil {
Fail("read error: " + fmt.Sprint(err))
}
var yamlTopo YamlTopology
err = yaml.Unmarshal(data, &yamlTopo)
if err != nil {
Fail("unmarshal error: " + fmt.Sprint(err))
}
s.Ip4AddrAllocator = NewIp4AddressAllocator()
s.NetInterfaces = make(map[string]*NetInterface)
for _, elem := range yamlTopo.Devices {
if _, ok := elem["name"]; ok {
elem["name"] = s.ProcessIndex + elem["name"].(string) + s.Ppid
}
if peer, ok := elem["peer"].(NetDevConfig); ok {
if peer["name"].(string) != "" {
peer["name"] = s.ProcessIndex + peer["name"].(string) + s.Ppid
}
if _, ok := peer["netns"]; ok {
peer["netns"] = s.ProcessIndex + peer["netns"].(string) + s.Ppid
}
}
if _, ok := elem["netns"]; ok {
elem["netns"] = s.ProcessIndex + elem["netns"].(string) + s.Ppid
}
if _, ok := elem["interfaces"]; ok {
interfaceCount := len(elem["interfaces"].([]interface{}))
for i := 0; i < interfaceCount; i++ {
elem["interfaces"].([]interface{})[i] = s.ProcessIndex + elem["interfaces"].([]interface{})[i].(string) + s.Ppid
}
}
switch elem["type"].(string) {
case NetNs:
{
if namespace, err := newNetNamespace(elem); err == nil {
s.NetConfigs = append(s.NetConfigs, &namespace)
} else {
Fail("network config error: " + fmt.Sprint(err))
}
}
case Veth, Tap:
{
if netIf, err := newNetworkInterface(elem, s.Ip4AddrAllocator); err == nil {
s.NetConfigs = append(s.NetConfigs, netIf)
s.NetInterfaces[netIf.Name()] = netIf
} else {
Fail("network config error: " + fmt.Sprint(err))
}
}
case Bridge:
{
if bridge, err := newBridge(elem); err == nil {
s.NetConfigs = append(s.NetConfigs, &bridge)
} else {
Fail("network config error: " + fmt.Sprint(err))
}
}
}
}
}
func (s *HstSuite) ConfigureNetworkTopology(topologyName string) {
s.LoadNetworkTopology(topologyName)
if *IsUnconfiguring {
return
}
for _, nc := range s.NetConfigs {
s.Log(nc.Name())
if err := nc.configure(); err != nil {
Fail("Network config error: " + fmt.Sprint(err))
}
}
}
func (s *HstSuite) UnconfigureNetworkTopology() {
if *IsPersistent {
return
}
for _, nc := range s.NetConfigs {
nc.unconfigure()
}
}
func (s *HstSuite) GetTestId() string {
testName := s.GetCurrentTestName()
if s.TestIds == nil {
s.TestIds = map[string]string{}
}
if _, ok := s.TestIds[testName]; !ok {
s.TestIds[testName] = time.Now().Format("2006-01-02_15-04-05")
}
return s.TestIds[testName]
}
func (s *HstSuite) GetCurrentTestName() string {
return strings.Split(CurrentSpecReport().LeafNodeText, "/")[1]
}
func (s *HstSuite) GetCurrentSuiteName() string {
return CurrentSpecReport().ContainerHierarchyTexts[0]
}
// Returns last 3 digits of PID + Ginkgo process index as the 4th digit
func (s *HstSuite) GetPortFromPpid() string {
port := s.Ppid
for len(port) < 3 {
port += "0"
}
return port[len(port)-3:] + s.ProcessIndex
}
func (s *HstSuite) StartServerApp(running chan error, done chan struct{}, env []string) {
cmd := exec.Command("iperf3", "-4", "-s", "-p", s.GetPortFromPpid())
if env != nil {
cmd.Env = env
}
s.Log(cmd)
err := cmd.Start()
if err != nil {
msg := fmt.Errorf("failed to start iperf server: %v", err)
running <- msg
return
}
running <- nil
<-done
cmd.Process.Kill()
}
func (s *HstSuite) StartClientApp(ipAddress string, env []string, clnCh chan error, clnRes chan string) {
defer func() {
clnCh <- nil
}()
nTries := 0
for {
cmd := exec.Command("iperf3", "-c", ipAddress, "-u", "-l", "1460", "-b", "10g", "-p", s.GetPortFromPpid())
if env != nil {
cmd.Env = env
}
s.Log(cmd)
o, err := cmd.CombinedOutput()
if err != nil {
if nTries > 5 {
clnCh <- fmt.Errorf("failed to start client app '%s'.\n%s", err, o)
return
}
time.Sleep(1 * time.Second)
nTries++
continue
} else {
clnRes <- fmt.Sprintf("Client output: %s", o)
}
break
}
}
func (s *HstSuite) StartHttpServer(running chan struct{}, done chan struct{}, addressPort, netNs string) {
cmd := newCommand([]string{"./http_server", addressPort, s.Ppid, s.ProcessIndex}, netNs)
err := cmd.Start()
s.Log(cmd)
if err != nil {
s.Log("Failed to start http server: " + fmt.Sprint(err))
return
}
running <- struct{}{}
<-done
cmd.Process.Kill()
}
func (s *HstSuite) StartWget(finished chan error, server_ip, port, query, netNs string) {
defer func() {
finished <- errors.New("wget error")
}()
cmd := newCommand([]string{"wget", "--timeout=10", "--no-proxy", "--tries=5", "-O", "/dev/null", server_ip + ":" + port + "/" + query},
netNs)
s.Log(cmd)
o, err := cmd.CombinedOutput()
if err != nil {
finished <- fmt.Errorf("wget error: '%v\n\n%s'", err, o)
return
} else if !strings.Contains(string(o), "200 OK") {
finished <- fmt.Errorf("wget error: response not 200 OK")
return
}
finished <- nil
}
/*
RunBenchmark creates Gomega's experiment with the passed-in name and samples the passed-in callback repeatedly (samplesNum times),
passing in suite context, experiment and your data.
You can also instruct runBenchmark to run with multiple concurrent workers.
Note that if running in parallel Gomega returns from Sample when spins up all samples and does not wait until all finished.
You can record multiple named measurements (float64 or duration) within passed-in callback.
runBenchmark then produces report to show statistical distribution of measurements.
*/
func (s *HstSuite) RunBenchmark(name string, samplesNum, parallelNum int, callback func(s *HstSuite, e *gmeasure.Experiment, data interface{}), data interface{}) {
experiment := gmeasure.NewExperiment(name)
experiment.Sample(func(idx int) {
defer GinkgoRecover()
callback(s, experiment, data)
}, gmeasure.SamplingConfig{N: samplesNum, NumParallel: parallelNum})
AddReportEntry(experiment.Name, experiment)
}
/*
LogHttpReq is Gomega's ghttp server handler which logs received HTTP request.
You should put it at the first place, so request is logged always.
*/
func (s *HstSuite) LogHttpReq(body bool) http.HandlerFunc {
return func(w http.ResponseWriter, req *http.Request) {
dump, err := httputil.DumpRequest(req, body)
if err == nil {
s.Log("\n> Received request (" + req.RemoteAddr + "):\n" +
string(dump) +
"\n------------------------------\n")
}
}
}