| #!/usr/bin/env python3 |
| |
| import sys |
| import argparse |
| import keyword |
| import logging |
| import binascii |
| import os |
| from subprocess import Popen, PIPE |
| import ply.lex as lex |
| import ply.yacc as yacc |
| |
| assert sys.version_info >= (3, 5), "Not supported Python version: {}".format( |
| sys.version |
| ) |
| log = logging.getLogger("vppapigen") |
| |
| # Ensure we don't leave temporary files around |
| sys.dont_write_bytecode = True |
| |
| # |
| # VPP API language |
| # |
| |
| # Global dictionary of new types (including enums) |
| global_types = {} |
| |
| seen_imports = {} |
| |
| |
| def global_type_add(name, obj): |
| """Add new type to the dictionary of types""" |
| type_name = "vl_api_" + name + "_t" |
| if type_name in global_types: |
| raise KeyError("Attempted redefinition of {!r} with {!r}.".format(name, obj)) |
| global_types[type_name] = obj |
| |
| |
| # All your trace are belong to us! |
| def exception_handler(exception_type, exception, traceback): |
| print("%s: %s" % (exception_type.__name__, exception)) |
| |
| |
| # |
| # Lexer |
| # |
| class VPPAPILexer: |
| def __init__(self, filename): |
| self.filename = filename |
| |
| reserved = { |
| "service": "SERVICE", |
| "rpc": "RPC", |
| "returns": "RETURNS", |
| "null": "NULL", |
| "stream": "STREAM", |
| "events": "EVENTS", |
| "define": "DEFINE", |
| "typedef": "TYPEDEF", |
| "enum": "ENUM", |
| "enumflag": "ENUMFLAG", |
| "typeonly": "TYPEONLY", |
| "manual_print": "MANUAL_PRINT", |
| "manual_endian": "MANUAL_ENDIAN", |
| "dont_trace": "DONT_TRACE", |
| "autoreply": "AUTOREPLY", |
| "autoendian": "AUTOENDIAN", |
| "option": "OPTION", |
| "u8": "U8", |
| "u16": "U16", |
| "u32": "U32", |
| "u64": "U64", |
| "i8": "I8", |
| "i16": "I16", |
| "i32": "I32", |
| "i64": "I64", |
| "f64": "F64", |
| "bool": "BOOL", |
| "string": "STRING", |
| "import": "IMPORT", |
| "true": "TRUE", |
| "false": "FALSE", |
| "union": "UNION", |
| "counters": "COUNTERS", |
| "paths": "PATHS", |
| "units": "UNITS", |
| "severity": "SEVERITY", |
| "type": "TYPE", |
| "description": "DESCRIPTION", |
| } |
| |
| tokens = ["STRING_LITERAL", "COMMENT", "ID", "NUM"] + list(reserved.values()) |
| |
| t_ignore_LINE_COMMENT = "//.*" |
| |
| def t_FALSE(self, t): |
| r"false" |
| t.value = False |
| return t |
| |
| def t_TRUE(self, t): |
| r"false" |
| t.value = True |
| return t |
| |
| def t_NUM(self, t): |
| r"0[xX][0-9a-fA-F]+|-?\d+\.?\d*" |
| base = 16 if t.value.startswith("0x") else 10 |
| if "." in t.value: |
| t.value = float(t.value) |
| else: |
| t.value = int(t.value, base) |
| return t |
| |
| def t_ID(self, t): |
| r"[a-zA-Z_][a-zA-Z_0-9]*" |
| # Check for reserved words |
| t.type = VPPAPILexer.reserved.get(t.value, "ID") |
| return t |
| |
| # C string |
| def t_STRING_LITERAL(self, t): |
| r"\"([^\\\n]|(\\.))*?\" " |
| t.value = str(t.value).replace('"', "") |
| return t |
| |
| # C or C++ comment (ignore) |
| def t_COMMENT(self, t): |
| r"(/\*(.|\n)*?\*/)|(//.*)" |
| t.lexer.lineno += t.value.count("\n") |
| return t |
| |
| # Error handling rule |
| def t_error(self, t): |
| raise ParseError( |
| "Illegal character '{}' ({})" |
| "in {}: line {}".format( |
| t.value[0], hex(ord(t.value[0])), self.filename, t.lexer.lineno |
| ) |
| ) |
| |
| # Define a rule so we can track line numbers |
| def t_newline(self, t): |
| r"\n+" |
| t.lexer.lineno += len(t.value) |
| |
| literals = ":{}[];=.," |
| |
| # A string containing ignored characters (spaces and tabs) |
| t_ignore = " \t" |
| |
| |
| def vla_mark_length_field(block): |
| if isinstance(block[-1], Array): |
| lengthfield = block[-1].lengthfield |
| for b in block: |
| if b.fieldname == lengthfield: |
| b.is_lengthfield = True |
| |
| |
| def vla_is_last_check(name, block): |
| vla = False |
| for i, b in enumerate(block): |
| if isinstance(b, Array) and b.vla: |
| vla = True |
| if i + 1 < len(block): |
| raise ValueError( |
| 'VLA field "{}" must be the last field in message "{}"'.format( |
| b.fieldname, name |
| ) |
| ) |
| elif b.fieldtype.startswith("vl_api_"): |
| if global_types[b.fieldtype].vla: |
| vla = True |
| if i + 1 < len(block): |
| raise ValueError( |
| 'VLA field "{}" must be the last ' |
| 'field in message "{}"'.format(b.fieldname, name) |
| ) |
| elif b.fieldtype == "string" and b.length == 0: |
| vla = True |
| if i + 1 < len(block): |
| raise ValueError( |
| 'VLA field "{}" must be the last ' |
| 'field in message "{}"'.format(b.fieldname, name) |
| ) |
| return vla |
| |
| |
| class Processable: |
| type = "<Invalid>" |
| |
| def process(self, result): # -> Dict |
| result[self.type].append(self) |
| |
| |
| class Service(Processable): |
| type = "Service" |
| |
| def __init__(self, caller, reply, events=None, stream_message=None, stream=False): |
| self.caller = caller |
| self.reply = reply |
| self.stream = stream |
| self.stream_message = stream_message |
| self.events = [] if events is None else events |
| |
| |
| class Typedef(Processable): |
| type = "Typedef" |
| |
| def __init__(self, name, flags, block): |
| self.name = name |
| self.flags = flags |
| self.block = block |
| self.crc = str(block).encode() |
| self.manual_print = False |
| self.manual_endian = False |
| for f in flags: |
| if f == "manual_print": |
| self.manual_print = True |
| elif f == "manual_endian": |
| self.manual_endian = True |
| global_type_add(name, self) |
| |
| self.vla = vla_is_last_check(name, block) |
| vla_mark_length_field(self.block) |
| |
| def process(self, result): |
| result["types"].append(self) |
| |
| def __repr__(self): |
| return self.name + str(self.flags) + str(self.block) |
| |
| |
| class Using(Processable): |
| type = "Using" |
| |
| def __init__(self, name, flags, alias): |
| self.name = name |
| self.vla = False |
| self.block = [] |
| self.manual_print = True |
| self.manual_endian = True |
| |
| self.manual_print = False |
| self.manual_endian = False |
| for f in flags: |
| if f == "manual_print": |
| self.manual_print = True |
| elif f == "manual_endian": |
| self.manual_endian = True |
| |
| if isinstance(alias, Array): |
| a = {"type": alias.fieldtype, "length": alias.length} |
| else: |
| a = {"type": alias.fieldtype} |
| self.alias = a |
| self.using = alias |
| |
| # |
| # Should have been: |
| # self.crc = str(alias).encode() |
| # but to be backwards compatible use the block ([]) |
| # |
| self.crc = str(self.block).encode() |
| global_type_add(name, self) |
| |
| def process(self, result): # -> Dict |
| result["types"].append(self) |
| |
| def __repr__(self): |
| return self.name + str(self.alias) |
| |
| |
| class Union(Processable): |
| type = "Union" |
| |
| def __init__(self, name, flags, block): |
| self.manual_print = False |
| self.manual_endian = False |
| self.name = name |
| |
| for f in flags: |
| if f == "manual_print": |
| self.manual_print = True |
| elif f == "manual_endian": |
| self.manual_endian = True |
| |
| self.block = block |
| self.crc = str(block).encode() |
| self.vla = vla_is_last_check(name, block) |
| |
| global_type_add(name, self) |
| |
| def process(self, result): |
| result["types"].append(self) |
| |
| def __repr__(self): |
| return str(self.block) |
| |
| |
| class Define(Processable): |
| type = "Define" |
| |
| def __init__(self, name, flags, block, comment=None): |
| self.name = name |
| self.flags = flags |
| self.block = block |
| self.dont_trace = False |
| self.manual_print = False |
| self.manual_endian = False |
| self.autoreply = False |
| self.autoendian = 0 |
| self.options = {} |
| self.comment = comment |
| for f in flags: |
| if f == "dont_trace": |
| self.dont_trace = True |
| elif f == "manual_print": |
| self.manual_print = True |
| elif f == "manual_endian": |
| self.manual_endian = True |
| elif f == "autoreply": |
| self.autoreply = True |
| elif f == "autoendian": |
| self.autoendian = 1 |
| |
| remove = [] |
| for b in block: |
| if isinstance(b, Option): |
| self.options[b.option] = b.value |
| remove.append(b) |
| |
| block = [x for x in block if x not in remove] |
| self.block = block |
| self.vla = vla_is_last_check(name, block) |
| vla_mark_length_field(self.block) |
| |
| self.crc = str(block).encode() |
| |
| def autoreply_block(self, name, parent): |
| block = [Field("u32", "context"), Field("i32", "retval")] |
| # inherit the parent's options |
| for k, v in parent.options.items(): |
| block.append(Option(k, v)) |
| return Define(name + "_reply", [], block) |
| |
| def process(self, result): # -> Dict |
| tname = self.__class__.__name__ |
| result[tname].append(self) |
| if self.autoreply: |
| result[tname].append(self.autoreply_block(self.name, self)) |
| |
| def __repr__(self): |
| return self.name + str(self.flags) + str(self.block) |
| |
| |
| class Enum(Processable): |
| type = "Enum" |
| |
| def __init__(self, name, block, enumtype="u32"): |
| self.name = name |
| self.enumtype = enumtype |
| self.vla = False |
| self.manual_print = False |
| |
| count = -1 |
| block2 = [] |
| block3 = [] |
| bc_set = False |
| |
| for b in block: |
| if "value" in b: |
| count = b["value"] |
| else: |
| count += 1 |
| block2.append([b["id"], count]) |
| try: |
| if b["option"]["backwards_compatible"]: |
| pass |
| bc_set = True |
| except KeyError: |
| block3.append([b["id"], count]) |
| if bc_set: |
| raise ValueError( |
| "Backward compatible enum must " |
| "be last {!r} {!r}".format(name, b["id"]) |
| ) |
| self.block = block2 |
| self.crc = str(block3).encode() |
| global_type_add(name, self) |
| |
| def process(self, result): |
| result["types"].append(self) |
| |
| def __repr__(self): |
| return self.name + str(self.block) |
| |
| |
| class EnumFlag(Enum): |
| type = "EnumFlag" |
| |
| def __init__(self, name, block, enumtype="u32"): |
| super(EnumFlag, self).__init__(name, block, enumtype) |
| |
| for b in self.block: |
| if bin(b[1])[2:].count("1") > 1: |
| raise TypeError( |
| "%s is not a flag enum. No element in a " |
| "flag enum may have more than a " |
| "single bit set." % self.name |
| ) |
| |
| |
| class Import(Processable): |
| type = "Import" |
| _initialized = False |
| |
| def __new__(cls, *args, **kwargs): |
| if args[0] not in seen_imports: |
| instance = super().__new__(cls) |
| instance._initialized = False |
| seen_imports[args[0]] = instance |
| |
| return seen_imports[args[0]] |
| |
| def __init__(self, filename, revision): |
| if self._initialized: |
| return |
| self.filename = filename |
| # Deal with imports |
| parser = VPPAPI(filename=filename, revision=revision) |
| dirlist = dirlist_get() |
| f = filename |
| for dir in dirlist: |
| f = os.path.join(dir, filename) |
| if os.path.exists(f): |
| break |
| self.result = parser.parse_filename(f, None) |
| self._initialized = True |
| |
| def __repr__(self): |
| return self.filename |
| |
| |
| class Option(Processable): |
| type = "Option" |
| |
| def __init__(self, option, value=None): |
| self.option = option |
| self.value = value |
| self.crc = str(option).encode() |
| |
| def process(self, result): # -> Dict |
| result[self.type][self.option] = self.value |
| |
| def __repr__(self): |
| return str(self.option) |
| |
| def __getitem__(self, index): |
| return self.option[index] |
| |
| |
| class Array(Processable): |
| type = "Array" |
| |
| def __init__(self, fieldtype, name, length, modern_vla=False): |
| self.fieldtype = fieldtype |
| self.fieldname = name |
| self.modern_vla = modern_vla |
| if type(length) is str: |
| self.lengthfield = length |
| self.length = 0 |
| self.vla = True |
| else: |
| self.length = length |
| self.lengthfield = None |
| self.vla = False |
| |
| def __repr__(self): |
| return str([self.fieldtype, self.fieldname, self.length, self.lengthfield]) |
| |
| |
| class Field(Processable): |
| type = "Field" |
| |
| def __init__(self, fieldtype, name, limit=None): |
| # limit field has been expanded to an options dict. |
| |
| self.fieldtype = fieldtype |
| self.is_lengthfield = False |
| |
| if self.fieldtype == "string": |
| raise ValueError("The string type {!r} is an " "array type ".format(name)) |
| |
| if name in keyword.kwlist: |
| raise ValueError( |
| "Fieldname {!r} is a python keyword and is not " |
| "accessible via the python API. ".format(name) |
| ) |
| self.fieldname = name |
| self.limit = limit |
| |
| def __repr__(self): |
| return str([self.fieldtype, self.fieldname]) |
| |
| |
| class Counter(Processable): |
| type = "Counter" |
| |
| def __init__(self, path, counter): |
| self.name = path |
| self.block = counter |
| |
| def process(self, result): # -> Dict |
| result["Counters"].append(self) |
| |
| |
| class Paths(Processable): |
| type = "Paths" |
| |
| def __init__(self, pathset): |
| self.paths = pathset |
| |
| def __repr__(self): |
| return "%s(paths=%s)" % (self.__class__.__name__, self.paths) |
| |
| |
| class Coord: |
| """Coordinates of a syntactic element. Consists of: |
| - File name |
| - Line number |
| - (optional) column number, for the Lexer |
| """ |
| |
| __slots__ = ("file", "line", "column", "__weakref__") |
| |
| def __init__(self, file, line, column=None): |
| self.file = file |
| self.line = line |
| self.column = column |
| |
| def __str__(self): |
| str = "%s:%s" % (self.file, self.line) |
| if self.column: |
| str += ":%s" % self.column |
| return str |
| |
| |
| class ParseError(Exception): |
| pass |
| |
| |
| # |
| # Grammar rules |
| # |
| class VPPAPIParser: |
| tokens = VPPAPILexer.tokens |
| |
| def __init__(self, filename, logger, revision=None): |
| self.filename = filename |
| self.logger = logger |
| self.fields = [] |
| self.revision = revision |
| self.last_comment = None |
| |
| def _parse_error(self, msg, coord): |
| raise ParseError("%s: %s" % (coord, msg)) |
| |
| def _parse_warning(self, msg, coord): |
| if self.logger: |
| self.logger.warning("%s: %s" % (coord, msg)) |
| |
| def _coord(self, lineno, column=None): |
| return Coord(file=self.filename, line=lineno, column=column) |
| |
| def _token_coord(self, p, token_idx): |
| """Returns the coordinates for the YaccProduction object 'p' indexed |
| with 'token_idx'. The coordinate includes the 'lineno' and |
| 'column'. Both follow the lex semantic, starting from 1. |
| """ |
| last_cr = p.lexer.lexdata.rfind("\n", 0, p.lexpos(token_idx)) |
| if last_cr < 0: |
| last_cr = -1 |
| column = p.lexpos(token_idx) - (last_cr) |
| return self._coord(p.lineno(token_idx), column) |
| |
| def p_slist(self, p): |
| """slist : stmt |
| | slist stmt""" |
| if len(p) == 2: |
| p[0] = [p[1]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_stmt(self, p): |
| """stmt : define |
| | typedef |
| | option |
| | import |
| | enum |
| | enumflag |
| | union |
| | service |
| | paths |
| | comment |
| | counters""" |
| p[0] = p[1] |
| |
| def p_import(self, p): |
| """import : IMPORT STRING_LITERAL ';'""" |
| p[0] = Import(p[2], revision=self.revision) |
| |
| def p_path_elements(self, p): |
| """path_elements : path_element |
| | path_elements path_element""" |
| if len(p) == 2: |
| p[0] = p[1] |
| else: |
| if type(p[1]) is dict: |
| p[0] = [p[1], p[2]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_path_element(self, p): |
| """path_element : STRING_LITERAL STRING_LITERAL ';'""" |
| p[0] = {"path": p[1], "counter": p[2]} |
| |
| def p_paths(self, p): |
| """paths : PATHS '{' path_elements '}' ';'""" |
| p[0] = Paths(p[3]) |
| |
| def p_counters(self, p): |
| """counters : COUNTERS ID '{' counter_elements '}' ';'""" |
| p[0] = Counter(p[2], p[4]) |
| |
| def p_counter_elements(self, p): |
| """counter_elements : counter_element |
| | counter_elements counter_element""" |
| if len(p) == 2: |
| p[0] = [p[1]] |
| else: |
| if type(p[1]) is dict: |
| p[0] = [p[1], p[2]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_counter_element(self, p): |
| """counter_element : ID '{' counter_statements '}' ';'""" |
| p[0] = {**{"name": p[1]}, **p[3]} |
| |
| def p_counter_statements(self, p): |
| """counter_statements : counter_statement |
| | counter_statements counter_statement""" |
| if len(p) == 2: |
| p[0] = p[1] |
| else: |
| p[0] = {**p[1], **p[2]} |
| |
| def p_counter_statement(self, p): |
| """counter_statement : SEVERITY ID ';' |
| | UNITS STRING_LITERAL ';' |
| | DESCRIPTION STRING_LITERAL ';' |
| | TYPE ID ';'""" |
| p[0] = {p[1]: p[2]} |
| |
| def p_service(self, p): |
| """service : SERVICE '{' service_statements '}' ';'""" |
| p[0] = p[3] |
| |
| def p_service_statements(self, p): |
| """service_statements : service_statement |
| | service_statements service_statement""" |
| if len(p) == 2: |
| p[0] = [p[1]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_service_statement(self, p): |
| """service_statement : RPC ID RETURNS NULL ';' |
| | RPC ID RETURNS ID ';' |
| | RPC ID RETURNS STREAM ID ';' |
| | RPC ID RETURNS ID EVENTS event_list ';'""" |
| if p[2] == p[4]: |
| # Verify that caller and reply differ |
| self._parse_error( |
| "Reply ID ({}) should not be equal to Caller ID".format(p[2]), |
| self._token_coord(p, 1), |
| ) |
| if len(p) == 8: |
| p[0] = Service(p[2], p[4], p[6]) |
| elif len(p) == 7: |
| p[0] = Service(p[2], p[5], stream=True) |
| else: |
| p[0] = Service(p[2], p[4]) |
| |
| def p_service_statement2(self, p): |
| """service_statement : RPC ID RETURNS ID STREAM ID ';'""" |
| p[0] = Service(p[2], p[4], stream_message=p[6], stream=True) |
| |
| def p_event_list(self, p): |
| """event_list : events |
| | event_list events""" |
| if len(p) == 2: |
| p[0] = [p[1]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_event(self, p): |
| """events : ID |
| | ID ','""" |
| p[0] = p[1] |
| |
| def p_enum(self, p): |
| """enum : ENUM ID '{' enum_statements '}' ';'""" |
| p[0] = Enum(p[2], p[4]) |
| |
| def p_enum_type(self, p): |
| """enum : ENUM ID ':' enum_size '{' enum_statements '}' ';'""" |
| if len(p) == 9: |
| p[0] = Enum(p[2], p[6], enumtype=p[4]) |
| else: |
| p[0] = Enum(p[2], p[4]) |
| |
| def p_enumflag(self, p): |
| """enumflag : ENUMFLAG ID '{' enum_statements '}' ';'""" |
| p[0] = EnumFlag(p[2], p[4]) |
| |
| def p_enumflag_type(self, p): |
| """enumflag : ENUMFLAG ID ':' enumflag_size '{' enum_statements '}' ';'""" # noqa : E502 |
| if len(p) == 9: |
| p[0] = EnumFlag(p[2], p[6], enumtype=p[4]) |
| else: |
| p[0] = EnumFlag(p[2], p[4]) |
| |
| def p_enum_size(self, p): |
| """enum_size : U8 |
| | U16 |
| | U32 |
| | I8 |
| | I16 |
| | I32""" |
| p[0] = p[1] |
| |
| def p_enumflag_size(self, p): |
| """enumflag_size : U8 |
| | U16 |
| | U32""" |
| p[0] = p[1] |
| |
| def p_define(self, p): |
| """define : DEFINE ID '{' block_statements_opt '}' ';'""" |
| self.fields = [] |
| p[0] = Define(p[2], [], p[4], self.last_comment) |
| self.last_comment = None |
| |
| def p_define_flist(self, p): |
| """define : flist DEFINE ID '{' block_statements_opt '}' ';'""" |
| # Legacy typedef |
| if "typeonly" in p[1]: |
| self._parse_error( |
| "legacy typedef. use typedef: {} {}[{}];".format(p[1], p[2], p[4]), |
| self._token_coord(p, 1), |
| ) |
| else: |
| p[0] = Define(p[3], p[1], p[5], self.last_comment) |
| self.last_comment = None |
| |
| def p_flist(self, p): |
| """flist : flag |
| | flist flag""" |
| if len(p) == 2: |
| p[0] = [p[1]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_flag(self, p): |
| """flag : MANUAL_PRINT |
| | MANUAL_ENDIAN |
| | DONT_TRACE |
| | TYPEONLY |
| | AUTOENDIAN |
| | AUTOREPLY""" |
| if len(p) == 1: |
| return |
| p[0] = p[1] |
| |
| def p_typedef(self, p): |
| """typedef : TYPEDEF ID '{' block_statements_opt '}' ';'""" |
| p[0] = Typedef(p[2], [], p[4]) |
| |
| def p_typedef_flist(self, p): |
| """typedef : flist TYPEDEF ID '{' block_statements_opt '}' ';'""" |
| p[0] = Typedef(p[3], p[1], p[5]) |
| |
| def p_typedef_alias(self, p): |
| """typedef : TYPEDEF declaration""" |
| p[0] = Using(p[2].fieldname, [], p[2]) |
| |
| def p_typedef_alias_flist(self, p): |
| """typedef : flist TYPEDEF declaration""" |
| p[0] = Using(p[3].fieldname, p[1], p[3]) |
| |
| def p_block_statements_opt(self, p): |
| """block_statements_opt : block_statements""" |
| p[0] = p[1] |
| |
| def p_block_statements(self, p): |
| """block_statements : block_statement |
| | block_statements block_statement""" |
| if len(p) == 2: |
| p[0] = [p[1]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_block_statement(self, p): |
| """block_statement : declaration |
| | option""" |
| p[0] = p[1] |
| |
| def p_enum_statements(self, p): |
| """enum_statements : enum_statement |
| | enum_statements enum_statement""" |
| if len(p) == 2: |
| p[0] = [p[1]] |
| else: |
| p[0] = p[1] + [p[2]] |
| |
| def p_enum_statement(self, p): |
| """enum_statement : ID '=' NUM ',' |
| | ID ',' |
| | ID '[' field_options ']' ',' |
| | ID '=' NUM '[' field_options ']' ','""" |
| if len(p) == 3: |
| p[0] = {"id": p[1]} |
| elif len(p) == 5: |
| p[0] = {"id": p[1], "value": p[3]} |
| elif len(p) == 6: |
| p[0] = {"id": p[1], "option": p[3]} |
| elif len(p) == 8: |
| p[0] = {"id": p[1], "value": p[3], "option": p[5]} |
| else: |
| self._parse_error("ERROR", self._token_coord(p, 1)) |
| |
| def p_field_options(self, p): |
| """field_options : field_option |
| | field_options field_option""" |
| if len(p) == 2: |
| p[0] = p[1] |
| else: |
| p[0] = {**p[1], **p[2]} |
| |
| def p_field_option(self, p): |
| """field_option : ID |
| | ID '=' assignee ',' |
| | ID '=' assignee |
| |
| """ |
| if len(p) == 2: |
| p[0] = {p[1]: None} |
| else: |
| p[0] = {p[1]: p[3]} |
| |
| def p_variable_name(self, p): |
| """variable_name : ID |
| | TYPE |
| | SEVERITY |
| | DESCRIPTION |
| | COUNTERS |
| | PATHS |
| """ |
| p[0] = p[1] |
| |
| def p_comment(self, p): |
| """comment : COMMENT""" |
| self.last_comment = p[1] |
| p[0] = [] |
| |
| def p_declaration(self, p): |
| """declaration : type_specifier variable_name ';' |
| | type_specifier variable_name '[' field_options ']' ';' |
| """ |
| if len(p) == 7: |
| p[0] = Field(p[1], p[2], p[4]) |
| elif len(p) == 4: |
| p[0] = Field(p[1], p[2]) |
| else: |
| self._parse_error("ERROR", self._token_coord(p, 1)) |
| self.fields.append(p[2]) |
| |
| def p_declaration_array_vla(self, p): |
| """declaration : type_specifier variable_name '[' ']' ';'""" |
| p[0] = Array(p[1], p[2], 0, modern_vla=True) |
| |
| def p_declaration_array(self, p): |
| """declaration : type_specifier variable_name '[' NUM ']' ';' |
| | type_specifier variable_name '[' ID ']' ';'""" |
| |
| if len(p) != 7: |
| return self._parse_error( |
| "array: %s" % p.value, self._coord(lineno=p.lineno) |
| ) |
| |
| # Make this error later |
| if type(p[4]) is int and p[4] == 0: |
| # XXX: Line number is wrong |
| self._parse_warning( |
| "Old Style VLA: {} {}[{}];".format(p[1], p[2], p[4]), |
| self._token_coord(p, 1), |
| ) |
| |
| if type(p[4]) is str and p[4] not in self.fields: |
| # Verify that length field exists |
| self._parse_error( |
| "Missing length field: {} {}[{}];".format(p[1], p[2], p[4]), |
| self._token_coord(p, 1), |
| ) |
| p[0] = Array(p[1], p[2], p[4]) |
| |
| def p_option(self, p): |
| """option : OPTION ID '=' assignee ';' |
| | OPTION ID ';'""" |
| if len(p) == 4: |
| p[0] = Option(p[2]) |
| else: |
| p[0] = Option(p[2], p[4]) |
| |
| def p_assignee(self, p): |
| """assignee : NUM |
| | TRUE |
| | FALSE |
| | STRING_LITERAL""" |
| p[0] = p[1] |
| |
| def p_type_specifier(self, p): |
| """type_specifier : U8 |
| | U16 |
| | U32 |
| | U64 |
| | I8 |
| | I16 |
| | I32 |
| | I64 |
| | F64 |
| | BOOL |
| | STRING""" |
| p[0] = p[1] |
| |
| # Do a second pass later to verify that user defined types are defined |
| def p_typedef_specifier(self, p): |
| """type_specifier : ID""" |
| if p[1] not in global_types: |
| self._parse_error( |
| "Undefined type: {}".format(p[1]), self._token_coord(p, 1) |
| ) |
| p[0] = p[1] |
| |
| def p_union(self, p): |
| """union : UNION ID '{' block_statements_opt '}' ';'""" |
| p[0] = Union(p[2], [], p[4]) |
| |
| def p_union_flist(self, p): |
| """union : flist UNION ID '{' block_statements_opt '}' ';'""" |
| p[0] = Union(p[3], p[1], p[5]) |
| |
| # Error rule for syntax errors |
| def p_error(self, p): |
| if p: |
| if p.type == "COMMENT": |
| self.parser.errok() |
| return |
| self._parse_error("before: %s" % p.value, self._coord(lineno=p.lineno)) |
| else: |
| self._parse_error("At end of input", self.filename) |
| |
| def build(self, **kwargs): |
| self.parser = yacc.yacc(module=self, **kwargs) |
| |
| |
| class VPPAPI: |
| def __init__(self, debug=False, filename="", logger=None, revision=None): |
| self.lexer = lex.lex(module=VPPAPILexer(filename), debug=debug) |
| self.parser = VPPAPIParser(filename, logger, revision=revision) |
| self.parser.build(write_tables=False, debug=debug) |
| self.logger = logger |
| self.revision = revision |
| self.filename = filename |
| |
| def parse_string(self, code, debug=0, lineno=1): |
| self.lexer.lineno = lineno |
| return self.parser.parser.parse(code, lexer=self.lexer, debug=debug) |
| |
| def parse_fd(self, fd, debug=0): |
| data = fd.read() |
| return self.parse_string(data, debug=debug) |
| |
| def parse_filename(self, filename, debug=0): |
| if self.revision: |
| git_show = "git show {}:{}".format(self.revision, filename) |
| proc = Popen(git_show.split(), stdout=PIPE, encoding="utf-8") |
| try: |
| data, errs = proc.communicate() |
| if proc.returncode != 0: |
| print( |
| "File not found: {}:{}".format(self.revision, filename), |
| file=sys.stderr, |
| ) |
| sys.exit(2) |
| return self.parse_string(data, debug=debug) |
| except Exception: |
| sys.exit(3) |
| else: |
| try: |
| with open(filename, encoding="utf-8") as fd: |
| return self.parse_fd(fd, None) |
| except FileNotFoundError: |
| print("File not found: {}".format(filename), file=sys.stderr) |
| sys.exit(2) |
| |
| def process(self, objs): |
| s = {} |
| s["Option"] = {} |
| s["Define"] = [] |
| s["Service"] = [] |
| s["types"] = [] |
| s["Import"] = [] |
| s["Counters"] = [] |
| s["Paths"] = [] |
| crc = 0 |
| for o in objs: |
| try: |
| crc = binascii.crc32(o.crc, crc) & 0xFFFFFFFF |
| except AttributeError: |
| pass |
| |
| if type(o) is list: |
| for o2 in o: |
| if isinstance(o2, Service): |
| o2.process(s) |
| else: |
| o.process(s) |
| |
| msgs = {d.name: d for d in s["Define"]} |
| svcs = {s.caller: s for s in s["Service"]} |
| replies = {s.reply: s for s in s["Service"]} |
| seen_services = {} |
| |
| s["file_crc"] = crc |
| |
| for service in svcs: |
| if service not in msgs: |
| raise ValueError( |
| "Service definition refers to unknown message" |
| " definition: {}".format(service) |
| ) |
| if svcs[service].reply != "null" and svcs[service].reply not in msgs: |
| raise ValueError( |
| "Service definition refers to unknown message" |
| " definition in reply: {}".format(svcs[service].reply) |
| ) |
| if service in replies: |
| raise ValueError( |
| "Service definition refers to message" |
| " marked as reply: {}".format(service) |
| ) |
| for event in svcs[service].events: |
| if event not in msgs: |
| raise ValueError( |
| "Service definition refers to unknown " |
| "event: {} in message: {}".format(event, service) |
| ) |
| seen_services[event] = True |
| |
| # Create services implicitly |
| for d in msgs: |
| if d in seen_services: |
| continue |
| if d.endswith("_reply"): |
| if d[:-6] in svcs: |
| continue |
| if d[:-6] not in msgs: |
| raise ValueError("{} missing calling message".format(d)) |
| continue |
| if d.endswith("_dump"): |
| if d in svcs: |
| continue |
| if d[:-5] + "_details" in msgs: |
| s["Service"].append(Service(d, d[:-5] + "_details", stream=True)) |
| else: |
| raise ValueError("{} missing details message".format(d)) |
| continue |
| |
| if d.endswith("_details"): |
| if d[:-8] + "_get" in msgs: |
| if d[:-8] + "_get" in svcs: |
| continue |
| raise ValueError( |
| "{} should be in a stream service".format(d[:-8] + "_get") |
| ) |
| if d[:-8] + "_dump" in msgs: |
| continue |
| raise ValueError("{} missing dump or get message".format(d)) |
| |
| if d in svcs: |
| continue |
| if d + "_reply" in msgs: |
| s["Service"].append(Service(d, d + "_reply")) |
| else: |
| raise ValueError( |
| "{} missing reply message ({}) or service definition".format( |
| d, d + "_reply" |
| ) |
| ) |
| |
| return s |
| |
| def process_imports(self, objs, in_import, result): # -> List |
| for o in objs: |
| # Only allow the following object types from imported file |
| if in_import and not isinstance(o, (Enum, Import, Typedef, Union, Using)): |
| continue |
| if isinstance(o, Import): |
| result.append(o) |
| result = self.process_imports(o.result, True, result) |
| else: |
| result.append(o) |
| return result |
| |
| |
| # Add message ids to each message. |
| def add_msg_id(s): |
| for o in s: |
| o.block.insert(0, Field("u16", "_vl_msg_id")) |
| return s |
| |
| |
| dirlist = [] |
| |
| |
| def dirlist_add(dirs): |
| global dirlist |
| if dirs: |
| dirlist = dirlist + dirs |
| |
| |
| def dirlist_get(): |
| return dirlist |
| |
| |
| def foldup_blocks(block, crc): |
| for b in block: |
| # Look up CRC in user defined types |
| if b.fieldtype.startswith("vl_api_"): |
| # Recursively |
| t = global_types[b.fieldtype] |
| try: |
| crc = binascii.crc32(t.crc, crc) & 0xFFFFFFFF |
| crc = foldup_blocks(t.block, crc) |
| except AttributeError: |
| pass |
| return crc |
| |
| |
| def foldup_crcs(s): |
| for f in s: |
| f.crc = foldup_blocks(f.block, binascii.crc32(f.crc) & 0xFFFFFFFF) |
| |
| |
| def run_vppapigen( |
| input_file=None, |
| output=sys.stdout, |
| includedir=None, |
| debug=False, |
| show_name=None, |
| output_module="C", |
| outputdir=None, |
| pluginpath="", |
| git_revision=None, |
| ): |
| # reset globals |
| dirlist.clear() |
| global_types.clear() |
| seen_imports.clear() |
| |
| dirlist_add(includedir) |
| if not debug: |
| sys.excepthook = exception_handler |
| |
| # Filename |
| if show_name: |
| filename = show_name[0] |
| elif input_file: |
| filename = input_file |
| else: |
| filename = "" |
| |
| if debug: |
| logging.basicConfig(stream=sys.stdout, level=logging.WARNING) |
| else: |
| logging.basicConfig() |
| |
| # |
| # Generate representation |
| # |
| from importlib.machinery import SourceFileLoader |
| |
| # Default path |
| pluginpath = "" |
| if not pluginpath: |
| cand = [] |
| cand.append(os.path.dirname(os.path.realpath(__file__))) |
| cand.append(os.path.dirname(os.path.realpath(__file__)) + "/../share/vpp/") |
| for c in cand: |
| c += "/" |
| if os.path.isfile("{}vppapigen_{}.py".format(c, output_module.lower())): |
| pluginpath = c |
| break |
| else: |
| pluginpath = pluginpath + "/" |
| if pluginpath == "": |
| log.exception("Output plugin not found") |
| return 1 |
| module_path = "{}vppapigen_{}.py".format(pluginpath, output_module.lower()) |
| |
| try: |
| plugin = SourceFileLoader(output_module, module_path).load_module() |
| except Exception as err: |
| log.exception("Error importing output plugin: %s, %s", module_path, err) |
| return 1 |
| |
| parser = VPPAPI(debug=debug, filename=filename, logger=log, revision=git_revision) |
| |
| try: |
| if not input_file: |
| parsed_objects = parser.parse_fd(sys.stdin, log) |
| else: |
| parsed_objects = parser.parse_filename(input_file, log) |
| except ParseError as e: |
| print("Parse error: ", e, file=sys.stderr) |
| sys.exit(1) |
| |
| # Build a list of objects. Hash of lists. |
| result = [] |
| |
| # if the variable is not set in the plugin, assume it to be false. |
| try: |
| plugin.process_imports |
| except AttributeError: |
| plugin.process_imports = False |
| |
| if plugin.process_imports: |
| result = parser.process_imports(parsed_objects, False, result) |
| s = parser.process(result) |
| else: |
| s = parser.process(parsed_objects) |
| imports = parser.process_imports(parsed_objects, False, result) |
| s["imported"] = parser.process(imports) |
| |
| # Add msg_id field |
| s["Define"] = add_msg_id(s["Define"]) |
| |
| # Fold up CRCs |
| foldup_crcs(s["Define"]) |
| |
| # |
| # Debug |
| if debug: |
| import pprint |
| |
| pp = pprint.PrettyPrinter(indent=4, stream=sys.stderr) |
| for t in s["Define"]: |
| pp.pprint([t.name, t.flags, t.block]) |
| for t in s["types"]: |
| pp.pprint([t.name, t.block]) |
| |
| result = plugin.run(outputdir, filename, s) |
| if result: |
| if isinstance(output, str): |
| with open(output, "w", encoding="UTF-8") as f: |
| print(result, file=f) |
| else: |
| print(result, file=output) |
| else: |
| log.exception("Running plugin failed: %s %s", filename, result) |
| return 1 |
| return 0 |
| |
| |
| def run_kw_vppapigen(kwargs): |
| return run_vppapigen(**kwargs) |
| |
| |
| # |
| # Main |
| # |
| def main(): |
| if sys.version_info < ( |
| 3, |
| 5, |
| ): |
| log.exception( |
| "vppapigen requires a supported version of python. " |
| "Please use version 3.5 or greater. " |
| "Using %s", |
| sys.version, |
| ) |
| return 1 |
| |
| cliparser = argparse.ArgumentParser(description="VPP API generator") |
| cliparser.add_argument("--pluginpath", default="") |
| cliparser.add_argument("--includedir", action="append") |
| cliparser.add_argument("--outputdir", action="store") |
| cliparser.add_argument("--input") |
| cliparser.add_argument( |
| "--output", |
| nargs="?", |
| type=argparse.FileType("w", encoding="UTF-8"), |
| default=sys.stdout, |
| ) |
| |
| cliparser.add_argument("output_module", nargs="?", default="C") |
| cliparser.add_argument("--debug", action="store_true") |
| cliparser.add_argument("--show-name", nargs=1) |
| cliparser.add_argument( |
| "--git-revision", help="Git revision to use for opening files" |
| ) |
| args = cliparser.parse_args() |
| |
| return run_vppapigen( |
| includedir=args.includedir, |
| debug=args.debug, |
| outputdir=args.outputdir, |
| show_name=args.show_name, |
| input_file=args.input, |
| output_module=args.output_module, |
| pluginpath=args.pluginpath, |
| git_revision=args.git_revision, |
| output=args.output, |
| ) |
| |
| |
| if __name__ == "__main__": |
| sys.exit(main()) |