| package main |
| |
| import ( |
| "bufio" |
| "errors" |
| "flag" |
| "fmt" |
| "io" |
| "log" |
| "os" |
| "os/exec" |
| "strings" |
| "time" |
| |
| "github.com/edwarnicke/exechelper" |
| . "github.com/onsi/ginkgo/v2" |
| . "github.com/onsi/gomega" |
| "gopkg.in/yaml.v3" |
| ) |
| |
| const ( |
| DEFAULT_NETWORK_NUM int = 1 |
| ) |
| |
| var isPersistent = flag.Bool("persist", false, "persists topology config") |
| var isVerbose = flag.Bool("verbose", false, "verbose test output") |
| var isUnconfiguring = flag.Bool("unconfigure", false, "remove topology") |
| var isVppDebug = flag.Bool("debug", false, "attach gdb to vpp") |
| var nConfiguredCpus = flag.Int("cpus", 1, "number of CPUs assigned to vpp") |
| var vppSourceFileDir = flag.String("vppsrc", "", "vpp source file directory") |
| |
| type HstSuite struct { |
| containers map[string]*Container |
| vppContainerCount int |
| volumes []string |
| netConfigs []NetConfig |
| netInterfaces map[string]*NetInterface |
| ip4AddrAllocator *Ip4AddressAllocator |
| testIds map[string]string |
| cpuAllocator *CpuAllocatorT |
| cpuContexts []*CpuContext |
| cpuPerVpp int |
| pid string |
| logger *log.Logger |
| logFile *os.File |
| } |
| |
| func (s *HstSuite) SetupSuite() { |
| s.createLogger() |
| s.log("Suite Setup") |
| RegisterFailHandler(func(message string, callerSkip ...int) { |
| s.hstFail() |
| Fail(message, callerSkip...) |
| }) |
| var err error |
| s.pid = fmt.Sprint(os.Getpid()) |
| s.cpuAllocator, err = CpuAllocator() |
| if err != nil { |
| Fail("failed to init cpu allocator: " + fmt.Sprint(err)) |
| } |
| s.cpuPerVpp = *nConfiguredCpus |
| } |
| |
| func (s *HstSuite) AllocateCpus() []int { |
| cpuCtx, err := s.cpuAllocator.Allocate(s.vppContainerCount, s.cpuPerVpp) |
| s.assertNil(err) |
| s.AddCpuContext(cpuCtx) |
| return cpuCtx.cpus |
| } |
| |
| func (s *HstSuite) AddCpuContext(cpuCtx *CpuContext) { |
| s.cpuContexts = append(s.cpuContexts, cpuCtx) |
| } |
| |
| func (s *HstSuite) TearDownSuite() { |
| defer s.logFile.Close() |
| s.log("Suite Teardown") |
| s.unconfigureNetworkTopology() |
| } |
| |
| func (s *HstSuite) TearDownTest() { |
| s.log("Test Teardown") |
| if *isPersistent { |
| return |
| } |
| s.resetContainers() |
| s.removeVolumes() |
| s.ip4AddrAllocator.deleteIpAddresses() |
| } |
| |
| func (s *HstSuite) skipIfUnconfiguring() { |
| if *isUnconfiguring { |
| s.skip("skipping to unconfigure") |
| } |
| } |
| |
| func (s *HstSuite) SetupTest() { |
| s.log("Test Setup") |
| s.vppContainerCount = 0 |
| s.skipIfUnconfiguring() |
| s.setupVolumes() |
| s.setupContainers() |
| } |
| |
| func (s *HstSuite) setupVolumes() { |
| for _, volume := range s.volumes { |
| cmd := "docker volume create --name=" + volume |
| s.log(cmd) |
| exechelper.Run(cmd) |
| } |
| } |
| |
| func (s *HstSuite) setupContainers() { |
| for _, container := range s.containers { |
| if !container.isOptional { |
| container.run() |
| } |
| } |
| } |
| |
| func (s *HstSuite) logVppInstance(container *Container, maxLines int) { |
| if container.vppInstance == nil { |
| return |
| } |
| |
| logSource := container.getHostWorkDir() + defaultLogFilePath |
| file, err := os.Open(logSource) |
| |
| if err != nil { |
| return |
| } |
| defer file.Close() |
| |
| scanner := bufio.NewScanner(file) |
| var lines []string |
| var counter int |
| |
| for scanner.Scan() { |
| lines = append(lines, scanner.Text()) |
| counter++ |
| if counter > maxLines { |
| lines = lines[1:] |
| counter-- |
| } |
| } |
| |
| s.log("vvvvvvvvvvvvvvv " + container.name + " [VPP instance]:") |
| for _, line := range lines { |
| s.log(line) |
| } |
| s.log("^^^^^^^^^^^^^^^\n\n") |
| } |
| |
| func (s *HstSuite) hstFail() { |
| s.log("Containers: " + fmt.Sprint(s.containers)) |
| for _, container := range s.containers { |
| out, err := container.log(20) |
| if err != nil { |
| s.log("An error occured while obtaining '" + container.name + "' container logs: " + fmt.Sprint(err)) |
| continue |
| } |
| s.log("\nvvvvvvvvvvvvvvv " + |
| container.name + ":\n" + |
| out + |
| "^^^^^^^^^^^^^^^\n\n") |
| s.logVppInstance(container, 20) |
| } |
| } |
| |
| func (s *HstSuite) assertNil(object interface{}, msgAndArgs ...interface{}) { |
| Expect(object).To(BeNil(), msgAndArgs...) |
| } |
| |
| func (s *HstSuite) assertNotNil(object interface{}, msgAndArgs ...interface{}) { |
| Expect(object).ToNot(BeNil(), msgAndArgs...) |
| } |
| |
| func (s *HstSuite) assertEqual(expected, actual interface{}, msgAndArgs ...interface{}) { |
| Expect(actual).To(Equal(expected), msgAndArgs...) |
| } |
| |
| func (s *HstSuite) assertNotEqual(expected, actual interface{}, msgAndArgs ...interface{}) { |
| Expect(actual).ToNot(Equal(expected), msgAndArgs...) |
| } |
| |
| func (s *HstSuite) assertContains(testString, contains interface{}, msgAndArgs ...interface{}) { |
| Expect(testString).To(ContainSubstring(fmt.Sprint(contains)), msgAndArgs...) |
| } |
| |
| func (s *HstSuite) assertNotContains(testString, contains interface{}, msgAndArgs ...interface{}) { |
| Expect(testString).ToNot(ContainSubstring(fmt.Sprint(contains)), msgAndArgs...) |
| } |
| |
| func (s *HstSuite) assertNotEmpty(object interface{}, msgAndArgs ...interface{}) { |
| Expect(object).ToNot(BeEmpty(), msgAndArgs...) |
| } |
| |
| func (s *HstSuite) createLogger() { |
| suiteName := CurrentSpecReport().ContainerHierarchyTexts[0] |
| var err error |
| s.logFile, err = os.Create("summary/" + suiteName + ".log") |
| if err != nil { |
| Fail("Unable to create log file.") |
| } |
| s.logger = log.New(io.Writer(s.logFile), "", log.LstdFlags) |
| } |
| |
| // Logs to files by default, logs to stdout when VERBOSE=true with GinkgoWriter |
| // to keep console tidy |
| func (s *HstSuite) log(arg any) { |
| logs := strings.Split(fmt.Sprint(arg), "\n") |
| for _, line := range logs { |
| s.logger.Println(line) |
| } |
| if *isVerbose { |
| GinkgoWriter.Println(arg) |
| } |
| } |
| |
| func (s *HstSuite) skip(args string) { |
| Skip(args) |
| } |
| |
| func (s *HstSuite) SkipIfMultiWorker(args ...any) { |
| if *nConfiguredCpus > 1 { |
| s.skip("test case not supported with multiple vpp workers") |
| } |
| } |
| |
| func (s *HstSuite) SkipUnlessExtendedTestsBuilt() { |
| imageName := "hs-test/nginx-http3" |
| |
| cmd := exec.Command("docker", "images", imageName) |
| byteOutput, err := cmd.CombinedOutput() |
| if err != nil { |
| s.log("error while searching for docker image") |
| return |
| } |
| if !strings.Contains(string(byteOutput), imageName) { |
| s.skip("extended tests not built") |
| } |
| } |
| |
| func (s *HstSuite) resetContainers() { |
| for _, container := range s.containers { |
| container.stop() |
| } |
| } |
| |
| func (s *HstSuite) removeVolumes() { |
| for _, volumeName := range s.volumes { |
| cmd := "docker volume rm " + volumeName |
| exechelper.Run(cmd) |
| os.RemoveAll(volumeName) |
| } |
| } |
| |
| func (s *HstSuite) getNetNamespaceByName(name string) string { |
| return name + s.pid |
| } |
| |
| func (s *HstSuite) getInterfaceByName(name string) *NetInterface { |
| return s.netInterfaces[name+s.pid] |
| } |
| |
| func (s *HstSuite) getContainerByName(name string) *Container { |
| return s.containers[name+s.pid] |
| } |
| |
| /* |
| * Create a copy and return its address, so that individial tests which call this |
| * are not able to modify the original container and affect other tests by doing that |
| */ |
| func (s *HstSuite) getTransientContainerByName(name string) *Container { |
| containerCopy := *s.containers[name+s.pid] |
| return &containerCopy |
| } |
| |
| func (s *HstSuite) loadContainerTopology(topologyName string) { |
| data, err := os.ReadFile(containerTopologyDir + topologyName + ".yaml") |
| if err != nil { |
| Fail("read error: " + fmt.Sprint(err)) |
| } |
| var yamlTopo YamlTopology |
| err = yaml.Unmarshal(data, &yamlTopo) |
| if err != nil { |
| Fail("unmarshal error: " + fmt.Sprint(err)) |
| } |
| |
| for _, elem := range yamlTopo.Volumes { |
| volumeMap := elem["volume"].(VolumeConfig) |
| hostDir := volumeMap["host-dir"].(string) |
| workingVolumeDir := logDir + CurrentSpecReport().LeafNodeText + volumeDir |
| volDirReplacer := strings.NewReplacer("$HST_VOLUME_DIR", workingVolumeDir) |
| hostDir = volDirReplacer.Replace(hostDir) |
| s.volumes = append(s.volumes, hostDir) |
| } |
| |
| s.containers = make(map[string]*Container) |
| for _, elem := range yamlTopo.Containers { |
| newContainer, err := newContainer(s, elem) |
| newContainer.suite = s |
| newContainer.name += newContainer.suite.pid |
| if err != nil { |
| Fail("container config error: " + fmt.Sprint(err)) |
| } |
| s.containers[newContainer.name] = newContainer |
| } |
| } |
| |
| func (s *HstSuite) loadNetworkTopology(topologyName string) { |
| data, err := os.ReadFile(networkTopologyDir + topologyName + ".yaml") |
| if err != nil { |
| Fail("read error: " + fmt.Sprint(err)) |
| } |
| var yamlTopo YamlTopology |
| err = yaml.Unmarshal(data, &yamlTopo) |
| if err != nil { |
| Fail("unmarshal error: " + fmt.Sprint(err)) |
| } |
| |
| s.ip4AddrAllocator = NewIp4AddressAllocator() |
| s.netInterfaces = make(map[string]*NetInterface) |
| |
| for _, elem := range yamlTopo.Devices { |
| if _, ok := elem["name"]; ok { |
| elem["name"] = elem["name"].(string) + s.pid |
| } |
| |
| if peer, ok := elem["peer"].(NetDevConfig); ok { |
| if peer["name"].(string) != "" { |
| peer["name"] = peer["name"].(string) + s.pid |
| } |
| if _, ok := peer["netns"]; ok { |
| peer["netns"] = peer["netns"].(string) + s.pid |
| } |
| } |
| |
| if _, ok := elem["netns"]; ok { |
| elem["netns"] = elem["netns"].(string) + s.pid |
| } |
| |
| if _, ok := elem["interfaces"]; ok { |
| interfaceCount := len(elem["interfaces"].([]interface{})) |
| for i := 0; i < interfaceCount; i++ { |
| elem["interfaces"].([]interface{})[i] = elem["interfaces"].([]interface{})[i].(string) + s.pid |
| } |
| } |
| |
| switch elem["type"].(string) { |
| case NetNs: |
| { |
| if namespace, err := newNetNamespace(elem); err == nil { |
| s.netConfigs = append(s.netConfigs, &namespace) |
| } else { |
| Fail("network config error: " + fmt.Sprint(err)) |
| } |
| } |
| case Veth, Tap: |
| { |
| if netIf, err := newNetworkInterface(elem, s.ip4AddrAllocator); err == nil { |
| s.netConfigs = append(s.netConfigs, netIf) |
| s.netInterfaces[netIf.Name()] = netIf |
| } else { |
| Fail("network config error: " + fmt.Sprint(err)) |
| } |
| } |
| case Bridge: |
| { |
| if bridge, err := newBridge(elem); err == nil { |
| s.netConfigs = append(s.netConfigs, &bridge) |
| } else { |
| Fail("network config error: " + fmt.Sprint(err)) |
| } |
| } |
| } |
| } |
| } |
| |
| func (s *HstSuite) configureNetworkTopology(topologyName string) { |
| s.loadNetworkTopology(topologyName) |
| |
| if *isUnconfiguring { |
| return |
| } |
| |
| for _, nc := range s.netConfigs { |
| if err := nc.configure(); err != nil { |
| Fail("Network config error: " + fmt.Sprint(err)) |
| } |
| } |
| } |
| |
| func (s *HstSuite) unconfigureNetworkTopology() { |
| if *isPersistent { |
| return |
| } |
| for _, nc := range s.netConfigs { |
| nc.unconfigure() |
| } |
| } |
| |
| func (s *HstSuite) getTestId() string { |
| testName := CurrentSpecReport().LeafNodeText |
| |
| if s.testIds == nil { |
| s.testIds = map[string]string{} |
| } |
| |
| if _, ok := s.testIds[testName]; !ok { |
| s.testIds[testName] = time.Now().Format("2006-01-02_15-04-05") |
| } |
| |
| return s.testIds[testName] |
| } |
| |
| // Returns last 4 digits of PID |
| func (s *HstSuite) getPortFromPid() string { |
| port := s.pid |
| for len(port) < 4 { |
| port += "0" |
| } |
| return port[len(port)-4:] |
| } |
| |
| func (s *HstSuite) startServerApp(running chan error, done chan struct{}, env []string) { |
| cmd := exec.Command("iperf3", "-4", "-s", "-p", s.getPortFromPid()) |
| if env != nil { |
| cmd.Env = env |
| } |
| s.log(cmd) |
| err := cmd.Start() |
| if err != nil { |
| msg := fmt.Errorf("failed to start iperf server: %v", err) |
| running <- msg |
| return |
| } |
| running <- nil |
| <-done |
| cmd.Process.Kill() |
| } |
| |
| func (s *HstSuite) startClientApp(ipAddress string, env []string, clnCh chan error, clnRes chan string) { |
| defer func() { |
| clnCh <- nil |
| }() |
| |
| nTries := 0 |
| |
| for { |
| cmd := exec.Command("iperf3", "-c", ipAddress, "-u", "-l", "1460", "-b", "10g", "-p", s.getPortFromPid()) |
| if env != nil { |
| cmd.Env = env |
| } |
| s.log(cmd) |
| o, err := cmd.CombinedOutput() |
| if err != nil { |
| if nTries > 5 { |
| clnCh <- fmt.Errorf("failed to start client app '%s'.\n%s", err, o) |
| return |
| } |
| time.Sleep(1 * time.Second) |
| nTries++ |
| continue |
| } else { |
| clnRes <- fmt.Sprintf("Client output: %s", o) |
| } |
| break |
| } |
| } |
| |
| func (s *HstSuite) startHttpServer(running chan struct{}, done chan struct{}, addressPort, netNs string) { |
| cmd := newCommand([]string{"./http_server", addressPort, s.pid}, netNs) |
| err := cmd.Start() |
| s.log(cmd) |
| if err != nil { |
| s.log("Failed to start http server: " + fmt.Sprint(err)) |
| return |
| } |
| running <- struct{}{} |
| <-done |
| cmd.Process.Kill() |
| } |
| |
| func (s *HstSuite) startWget(finished chan error, server_ip, port, query, netNs string) { |
| defer func() { |
| finished <- errors.New("wget error") |
| }() |
| |
| cmd := newCommand([]string{"wget", "--timeout=10", "--no-proxy", "--tries=5", "-O", "/dev/null", server_ip + ":" + port + "/" + query}, |
| netNs) |
| s.log(cmd) |
| o, err := cmd.CombinedOutput() |
| if err != nil { |
| finished <- fmt.Errorf("wget error: '%v\n\n%s'", err, o) |
| return |
| } else if !strings.Contains(string(o), "200 OK") { |
| finished <- fmt.Errorf("wget error: response not 200 OK") |
| return |
| } |
| finished <- nil |
| } |