blob: dfb236b6725772bff218bd2c2e46b82cfa6e831f [file] [log] [blame]
package hst
import (
"context"
"encoding/json"
"fmt"
"go.fd.io/govpp/binapi/ethernet_types"
"io"
"net"
"os"
"os/exec"
"os/signal"
"strconv"
"strings"
"syscall"
"time"
"github.com/edwarnicke/exechelper"
. "github.com/onsi/ginkgo/v2"
"github.com/sirupsen/logrus"
"go.fd.io/govpp"
"go.fd.io/govpp/api"
"go.fd.io/govpp/binapi/af_packet"
interfaces "go.fd.io/govpp/binapi/interface"
"go.fd.io/govpp/binapi/interface_types"
"go.fd.io/govpp/binapi/session"
"go.fd.io/govpp/binapi/tapv2"
"go.fd.io/govpp/core"
)
const vppConfigTemplate = `unix {
nodaemon
log %[1]s%[4]s
full-coredump
cli-listen %[1]s%[2]s
runtime-dir %[1]s/var/run
gid vpp
}
api-trace {
on
}
api-segment {
gid vpp
}
socksvr {
socket-name %[1]s%[3]s
}
statseg {
socket-name %[1]s/var/run/vpp/stats.sock
}
plugins {
plugin default { disable }
plugin unittest_plugin.so { enable }
plugin quic_plugin.so { enable }
plugin af_packet_plugin.so { enable }
plugin hs_apps_plugin.so { enable }
plugin http_plugin.so { enable }
plugin http_static_plugin.so { enable }
plugin prom_plugin.so { enable }
plugin tlsopenssl_plugin.so { enable }
plugin ping_plugin.so { enable }
plugin nsim_plugin.so { enable }
plugin mactime_plugin.so { enable }
}
logging {
default-log-level debug
default-syslog-log-level debug
}
`
const (
defaultCliSocketFilePath = "/var/run/vpp/cli.sock"
defaultApiSocketFilePath = "/var/run/vpp/api.sock"
defaultLogFilePath = "/var/log/vpp/vpp.log"
)
type VppInstance struct {
Container *Container
AdditionalConfig []Stanza
Connection *core.Connection
ApiStream api.Stream
Cpus []int
CpuConfig VppCpuConfig
}
type VppCpuConfig struct {
PinMainCpu bool
PinWorkersCorelist bool
SkipCores int
}
type VppMemTrace struct {
Count int `json:"count"`
Size int `json:"bytes"`
Sample string `json:"sample"`
Traceback []string `json:"traceback"`
}
func (vpp *VppInstance) getSuite() *HstSuite {
return vpp.Container.Suite
}
func (vpp *VppInstance) getCliSocket() string {
return fmt.Sprintf("%s%s", vpp.Container.GetContainerWorkDir(), defaultCliSocketFilePath)
}
func (vpp *VppInstance) getRunDir() string {
return vpp.Container.GetContainerWorkDir() + "/var/run/vpp"
}
func (vpp *VppInstance) getLogDir() string {
return vpp.Container.GetContainerWorkDir() + "/var/log/vpp"
}
func (vpp *VppInstance) getEtcDir() string {
return vpp.Container.GetContainerWorkDir() + "/etc/vpp"
}
func (vpp *VppInstance) Start() error {
maxReconnectAttempts := 3
// Replace default logger in govpp with our own
govppLogger := logrus.New()
govppLogger.SetOutput(io.MultiWriter(vpp.getSuite().Logger.Writer(), GinkgoWriter))
core.SetLogger(govppLogger)
// Create folders
containerWorkDir := vpp.Container.GetContainerWorkDir()
vpp.Container.Exec("mkdir --mode=0700 -p " + vpp.getRunDir())
vpp.Container.Exec("mkdir --mode=0700 -p " + vpp.getLogDir())
vpp.Container.Exec("mkdir --mode=0700 -p " + vpp.getEtcDir())
// Create startup.conf inside the container
configContent := fmt.Sprintf(
vppConfigTemplate,
containerWorkDir,
defaultCliSocketFilePath,
defaultApiSocketFilePath,
defaultLogFilePath,
)
configContent += vpp.generateVPPCpuConfig()
for _, c := range vpp.AdditionalConfig {
configContent += c.ToString()
}
startupFileName := vpp.getEtcDir() + "/startup.conf"
vpp.Container.CreateFile(startupFileName, configContent)
// create wrapper script for vppctl with proper CLI socket path
cliContent := "#!/usr/bin/bash\nvppctl -s " + vpp.getRunDir() + "/cli.sock"
vppcliFileName := "/usr/bin/vppcli"
vpp.Container.CreateFile(vppcliFileName, cliContent)
vpp.Container.Exec("chmod 0755 " + vppcliFileName)
vpp.getSuite().Log("starting vpp")
if *IsVppDebug {
// default = 3; VPP will timeout while debugging if there are not enough attempts
maxReconnectAttempts = 5000
sig := make(chan os.Signal, 1)
signal.Notify(sig, syscall.SIGQUIT)
cont := make(chan bool, 1)
go func() {
<-sig
cont <- true
}()
vpp.Container.ExecServer("su -c \"vpp -c " + startupFileName + " &> /proc/1/fd/1\"")
fmt.Println("run following command in different terminal:")
fmt.Println("docker exec -it " + vpp.Container.Name + " gdb -ex \"attach $(docker exec " + vpp.Container.Name + " pidof vpp)\"")
fmt.Println("Afterwards press CTRL+\\ to continue")
<-cont
fmt.Println("continuing...")
} else {
// Start VPP
vpp.Container.ExecServer("su -c \"vpp -c " + startupFileName + " &> /proc/1/fd/1\"")
}
vpp.getSuite().Log("connecting to vpp")
// Connect to VPP and store the connection
sockAddress := vpp.Container.GetHostWorkDir() + defaultApiSocketFilePath
conn, connEv, err := govpp.AsyncConnect(
sockAddress,
maxReconnectAttempts,
core.DefaultReconnectInterval)
if err != nil {
vpp.getSuite().Log("async connect error: " + fmt.Sprint(err))
return err
}
vpp.Connection = conn
// ... wait for Connected event
e := <-connEv
if e.State != core.Connected {
vpp.getSuite().Log("connecting to VPP failed: " + fmt.Sprint(e.Error))
return e.Error
}
ch, err := conn.NewStream(
context.Background(),
core.WithRequestSize(50),
core.WithReplySize(50),
core.WithReplyTimeout(time.Second*5))
if err != nil {
vpp.getSuite().Log("creating stream failed: " + fmt.Sprint(err))
return err
}
vpp.ApiStream = ch
return nil
}
func (vpp *VppInstance) Vppctl(command string, arguments ...any) string {
vppCliCommand := fmt.Sprintf(command, arguments...)
containerExecCommand := fmt.Sprintf("docker exec --detach=false %[1]s vppctl -s %[2]s %[3]s",
vpp.Container.Name, vpp.getCliSocket(), vppCliCommand)
vpp.getSuite().Log(containerExecCommand)
output, err := exechelper.CombinedOutput(containerExecCommand)
vpp.getSuite().AssertNil(err)
return string(output)
}
func (vpp *VppInstance) GetSessionStat(stat string) int {
o := vpp.Vppctl("show session stats")
vpp.getSuite().Log(o)
for _, line := range strings.Split(o, "\n") {
if strings.Contains(line, stat) {
tokens := strings.Split(strings.TrimSpace(line), " ")
val, err := strconv.Atoi(tokens[0])
if err != nil {
Fail("failed to parse stat value %s" + fmt.Sprint(err))
return 0
}
return val
}
}
return 0
}
func (vpp *VppInstance) WaitForApp(appName string, timeout int) {
vpp.getSuite().Log("waiting for app " + appName)
for i := 0; i < timeout; i++ {
o := vpp.Vppctl("show app")
if strings.Contains(o, appName) {
return
}
time.Sleep(1 * time.Second)
}
vpp.getSuite().AssertNil(1, "Timeout while waiting for app '%s'", appName)
}
func (vpp *VppInstance) createAfPacket(
veth *NetInterface,
) (interface_types.InterfaceIndex, error) {
createReq := &af_packet.AfPacketCreateV3{
Mode: 1,
UseRandomHwAddr: true,
HostIfName: veth.Name(),
Flags: af_packet.AfPacketFlags(11),
}
if veth.HwAddress != (MacAddress{}) {
createReq.UseRandomHwAddr = false
createReq.HwAddr = veth.HwAddress
}
vpp.getSuite().Log("create af-packet interface " + veth.Name())
if err := vpp.ApiStream.SendMsg(createReq); err != nil {
vpp.getSuite().HstFail()
return 0, err
}
replymsg, err := vpp.ApiStream.RecvMsg()
if err != nil {
return 0, err
}
reply := replymsg.(*af_packet.AfPacketCreateV3Reply)
err = api.RetvalToVPPApiError(reply.Retval)
if err != nil {
return 0, err
}
veth.Index = reply.SwIfIndex
// Set to up
upReq := &interfaces.SwInterfaceSetFlags{
SwIfIndex: veth.Index,
Flags: interface_types.IF_STATUS_API_FLAG_ADMIN_UP,
}
vpp.getSuite().Log("set af-packet interface " + veth.Name() + " up")
if err := vpp.ApiStream.SendMsg(upReq); err != nil {
return 0, err
}
replymsg, err = vpp.ApiStream.RecvMsg()
if err != nil {
return 0, err
}
reply2 := replymsg.(*interfaces.SwInterfaceSetFlagsReply)
if err = api.RetvalToVPPApiError(reply2.Retval); err != nil {
return 0, err
}
// Add address
if veth.AddressWithPrefix() == (AddressWithPrefix{}) {
var err error
var ip4Address string
if ip4Address, err = veth.Ip4AddrAllocator.NewIp4InterfaceAddress(veth.Peer.NetworkNumber); err == nil {
veth.Ip4Address = ip4Address
} else {
return 0, err
}
}
addressReq := &interfaces.SwInterfaceAddDelAddress{
IsAdd: true,
SwIfIndex: veth.Index,
Prefix: veth.AddressWithPrefix(),
}
vpp.getSuite().Log("af-packet interface " + veth.Name() + " add address " + veth.Ip4Address)
if err := vpp.ApiStream.SendMsg(addressReq); err != nil {
return 0, err
}
replymsg, err = vpp.ApiStream.RecvMsg()
if err != nil {
return 0, err
}
reply3 := replymsg.(*interfaces.SwInterfaceAddDelAddressReply)
err = api.RetvalToVPPApiError(reply3.Retval)
if err != nil {
return 0, err
}
return veth.Index, nil
}
func (vpp *VppInstance) addAppNamespace(
secret uint64,
ifx interface_types.InterfaceIndex,
namespaceId string,
) error {
req := &session.AppNamespaceAddDelV4{
IsAdd: true,
Secret: secret,
SwIfIndex: ifx,
NamespaceID: namespaceId,
SockName: defaultApiSocketFilePath,
}
vpp.getSuite().Log("add app namespace " + namespaceId)
if err := vpp.ApiStream.SendMsg(req); err != nil {
return err
}
replymsg, err := vpp.ApiStream.RecvMsg()
if err != nil {
return err
}
reply := replymsg.(*session.AppNamespaceAddDelV4Reply)
if err = api.RetvalToVPPApiError(reply.Retval); err != nil {
return err
}
sessionReq := &session.SessionEnableDisable{
IsEnable: true,
}
vpp.getSuite().Log("enable app namespace " + namespaceId)
if err := vpp.ApiStream.SendMsg(sessionReq); err != nil {
return err
}
replymsg, err = vpp.ApiStream.RecvMsg()
if err != nil {
return err
}
reply2 := replymsg.(*session.SessionEnableDisableReply)
if err = api.RetvalToVPPApiError(reply2.Retval); err != nil {
return err
}
return nil
}
func (vpp *VppInstance) createTap(
tap *NetInterface,
tapId ...uint32,
) error {
var id uint32 = 1
if len(tapId) > 0 {
id = tapId[0]
}
createTapReq := &tapv2.TapCreateV3{
ID: id,
HostIfNameSet: true,
HostIfName: tap.Name(),
HostIP4PrefixSet: true,
HostIP4Prefix: tap.Ip4AddressWithPrefix(),
}
vpp.getSuite().Log("create tap interface " + tap.Name())
// Create tap interface
if err := vpp.ApiStream.SendMsg(createTapReq); err != nil {
return err
}
replymsg, err := vpp.ApiStream.RecvMsg()
if err != nil {
return err
}
reply := replymsg.(*tapv2.TapCreateV3Reply)
if err = api.RetvalToVPPApiError(reply.Retval); err != nil {
return err
}
tap.Peer.Index = reply.SwIfIndex
// Get name and mac
if err := vpp.ApiStream.SendMsg(&interfaces.SwInterfaceDump{
SwIfIndex: reply.SwIfIndex,
}); err != nil {
return err
}
replymsg, err = vpp.ApiStream.RecvMsg()
if err != nil {
return err
}
ifDetails := replymsg.(*interfaces.SwInterfaceDetails)
tap.Peer.name = ifDetails.InterfaceName
tap.Peer.HwAddress = ifDetails.L2Address
// Add address
addAddressReq := &interfaces.SwInterfaceAddDelAddress{
IsAdd: true,
SwIfIndex: reply.SwIfIndex,
Prefix: tap.Peer.AddressWithPrefix(),
}
vpp.getSuite().Log("tap interface " + tap.Name() + " add address " + tap.Peer.Ip4Address)
if err := vpp.ApiStream.SendMsg(addAddressReq); err != nil {
return err
}
replymsg, err = vpp.ApiStream.RecvMsg()
if err != nil {
return err
}
reply2 := replymsg.(*interfaces.SwInterfaceAddDelAddressReply)
if err = api.RetvalToVPPApiError(reply2.Retval); err != nil {
return err
}
// Set interface to up
upReq := &interfaces.SwInterfaceSetFlags{
SwIfIndex: reply.SwIfIndex,
Flags: interface_types.IF_STATUS_API_FLAG_ADMIN_UP,
}
vpp.getSuite().Log("set tap interface " + tap.Name() + " up")
if err := vpp.ApiStream.SendMsg(upReq); err != nil {
return err
}
replymsg, err = vpp.ApiStream.RecvMsg()
if err != nil {
return err
}
reply3 := replymsg.(*interfaces.SwInterfaceSetFlagsReply)
if err = api.RetvalToVPPApiError(reply3.Retval); err != nil {
return err
}
// Get host mac
netIntf, err := net.InterfaceByName(tap.Name())
if err == nil {
tap.HwAddress, _ = ethernet_types.ParseMacAddress(netIntf.HardwareAddr.String())
}
return nil
}
func (vpp *VppInstance) saveLogs() {
logTarget := vpp.Container.getLogDirPath() + "vppinstance-" + vpp.Container.Name + ".log"
logSource := vpp.Container.GetHostWorkDir() + defaultLogFilePath
cmd := exec.Command("cp", logSource, logTarget)
vpp.getSuite().Log(cmd.String())
cmd.Run()
}
func (vpp *VppInstance) Disconnect() {
vpp.Connection.Disconnect()
vpp.ApiStream.Close()
}
func (vpp *VppInstance) setDefaultCpuConfig() {
vpp.CpuConfig.PinMainCpu = true
vpp.CpuConfig.PinWorkersCorelist = true
vpp.CpuConfig.SkipCores = 0
}
func (vpp *VppInstance) generateVPPCpuConfig() string {
var c Stanza
var s string
startCpu := 0
if len(vpp.Cpus) < 1 {
return ""
}
c.NewStanza("cpu")
// If skip-cores is valid, use as start value to assign main/workers CPUs
if vpp.CpuConfig.SkipCores != 0 {
c.Append(fmt.Sprintf("skip-cores %d", vpp.CpuConfig.SkipCores))
vpp.getSuite().Log(fmt.Sprintf("skip-cores %d", vpp.CpuConfig.SkipCores))
}
if len(vpp.Cpus) > vpp.CpuConfig.SkipCores {
startCpu = vpp.CpuConfig.SkipCores
}
if vpp.CpuConfig.PinMainCpu {
c.Append(fmt.Sprintf("main-core %d", vpp.Cpus[startCpu]))
vpp.getSuite().Log(fmt.Sprintf("main-core %d", vpp.Cpus[startCpu]))
}
workers := vpp.Cpus[startCpu+1:]
if len(workers) > 0 {
if vpp.CpuConfig.PinWorkersCorelist {
for i := 0; i < len(workers); i++ {
if i != 0 {
s = s + ", "
}
s = s + fmt.Sprintf("%d", workers[i])
}
c.Append(fmt.Sprintf("corelist-workers %s", s))
vpp.getSuite().Log("corelist-workers " + s)
} else {
s = fmt.Sprintf("%d", len(workers))
c.Append(fmt.Sprintf("workers %s", s))
vpp.getSuite().Log("workers " + s)
}
}
return c.Close().ToString()
}
// EnableMemoryTrace enables memory traces of VPP main-heap
func (vpp *VppInstance) EnableMemoryTrace() {
vpp.getSuite().Log(vpp.Vppctl("memory-trace on main-heap"))
}
// GetMemoryTrace dumps memory traces for analysis
func (vpp *VppInstance) GetMemoryTrace() ([]VppMemTrace, error) {
var trace []VppMemTrace
vpp.getSuite().Log(vpp.Vppctl("save memory-trace trace.json"))
err := vpp.Container.GetFile("/tmp/trace.json", "/tmp/trace.json")
if err != nil {
return nil, err
}
fileBytes, err := os.ReadFile("/tmp/trace.json")
if err != nil {
return nil, err
}
err = json.Unmarshal(fileBytes, &trace)
if err != nil {
return nil, err
}
return trace, nil
}
// memTracesSuppressCli filter out CLI related samples
func memTracesSuppressCli(traces []VppMemTrace) []VppMemTrace {
var filtered []VppMemTrace
for i := 0; i < len(traces); i++ {
isCli := false
for j := 0; j < len(traces[i].Traceback); j++ {
if strings.Contains(traces[i].Traceback[j], "unix_cli") {
isCli = true
break
}
}
if !isCli {
filtered = append(filtered, traces[i])
}
}
return filtered
}
// MemLeakCheck compares memory traces at different point in time, analyzes if memory leaks happen and produces report
func (vpp *VppInstance) MemLeakCheck(first, second []VppMemTrace) {
totalBytes := 0
totalCounts := 0
trace1 := memTracesSuppressCli(first)
trace2 := memTracesSuppressCli(second)
report := ""
for i := 0; i < len(trace2); i++ {
match := false
for j := 0; j < len(trace1); j++ {
if trace1[j].Sample == trace2[i].Sample {
if trace2[i].Size > trace1[j].Size {
deltaBytes := trace2[i].Size - trace1[j].Size
deltaCounts := trace2[i].Count - trace1[j].Count
report += fmt.Sprintf("grow %d byte(s) in %d allocation(s) from:\n", deltaBytes, deltaCounts)
for j := 0; j < len(trace2[i].Traceback); j++ {
report += fmt.Sprintf("\t#%d %s\n", j, trace2[i].Traceback[j])
}
totalBytes += deltaBytes
totalCounts += deltaCounts
}
match = true
break
}
}
if !match {
report += fmt.Sprintf("\nleak of %d byte(s) in %d allocation(s) from:\n", trace2[i].Size, trace2[i].Count)
for j := 0; j < len(trace2[i].Traceback); j++ {
report += fmt.Sprintf("\t#%d %s\n", j, trace2[i].Traceback[j])
}
totalBytes += trace2[i].Size
totalCounts += trace2[i].Count
}
}
summary := fmt.Sprintf("\nSUMMARY: %d byte(s) leaked in %d allocation(s)\n", totalBytes, totalCounts)
AddReportEntry(summary, report)
}