blob: 4ca9b954e671bed432f2b410151c554b514013a5 [file] [log] [blame]
#!/usr/bin/env python
from __future__ import print_function
import ply.lex as lex
import ply.yacc as yacc
import sys
import argparse
import logging
import binascii
import os
# Ensure we don't leave temporary files around
sys.dont_write_bytecode = True
#
# VPP API language
#
# Global dictionary of new types (including enums)
global_types = {}
global_crc = 0
def global_type_add(name):
'''Add new type to the dictionary of types '''
type_name = 'vl_api_' + name + '_t'
if type_name in global_types:
raise KeyError('Type is already defined: {}'.format(name))
global_types[type_name] = True
# All your trace are belong to us!
def exception_handler(exception_type, exception, traceback):
print("%s: %s" % (exception_type.__name__, exception))
#
# Lexer
#
class VPPAPILexer(object):
def __init__(self, filename):
self.filename = filename
reserved = {
'service': 'SERVICE',
'rpc': 'RPC',
'returns': 'RETURNS',
'null': 'NULL',
'stream': 'STREAM',
'events': 'EVENTS',
'define': 'DEFINE',
'typedef': 'TYPEDEF',
'enum': 'ENUM',
'typeonly': 'TYPEONLY',
'manual_print': 'MANUAL_PRINT',
'manual_endian': 'MANUAL_ENDIAN',
'dont_trace': 'DONT_TRACE',
'autoreply': 'AUTOREPLY',
'option': 'OPTION',
'u8': 'U8',
'u16': 'U16',
'u32': 'U32',
'u64': 'U64',
'i8': 'I8',
'i16': 'I16',
'i32': 'I32',
'i64': 'I64',
'f64': 'F64',
'bool': 'BOOL',
'string': 'STRING',
'import': 'IMPORT',
'true': 'TRUE',
'false': 'FALSE',
'union': 'UNION',
}
tokens = ['STRING_LITERAL',
'ID', 'NUM'] + list(reserved.values())
t_ignore_LINE_COMMENT = '//.*'
def t_NUM(self, t):
r'0[xX][0-9a-fA-F]+|\d+'
base = 16 if t.value.startswith('0x') else 10
t.value = int(t.value, base)
return t
def t_ID(self, t):
r'[a-zA-Z_][a-zA-Z_0-9]*'
# Check for reserved words
t.type = VPPAPILexer.reserved.get(t.value, 'ID')
return t
# C string
def t_STRING_LITERAL(self, t):
r'\"([^\\\n]|(\\.))*?\"'
t.value = str(t.value).replace("\"", "")
return t
# C or C++ comment (ignore)
def t_comment(self, t):
r'(/\*(.|\n)*?\*/)|(//.*)'
t.lexer.lineno += t.value.count('\n')
# Error handling rule
def t_error(self, t):
raise ParseError("Illegal character '{}' ({})"
"in {}: line {}".format(t.value[0],
hex(ord(t.value[0])),
self.filename,
t.lexer.lineno))
t.lexer.skip(1)
# Define a rule so we can track line numbers
def t_newline(self, t):
r'\n+'
t.lexer.lineno += len(t.value)
literals = ":{}[];=.,"
# A string containing ignored characters (spaces and tabs)
t_ignore = ' \t'
#
# Side-effect: Sets global_crc
#
def crc_block(block):
global global_crc
s = str(block).encode()
global_crc = binascii.crc32(s, global_crc)
return binascii.crc32(s) & 0xffffffff
class Service():
def __init__(self, caller, reply, events=[], stream=False):
self.caller = caller
self.reply = reply
self.stream = stream
self.events = events
class Typedef():
def __init__(self, name, flags, block):
self.name = name
self.flags = flags
self.block = block
self.crc = crc_block(block)
self.manual_print = False
self.manual_endian = False
for f in flags:
if f == 'manual_print':
self.manual_print = True
elif f == 'manual_endian':
self.manual_endian = True
global_type_add(name)
def __repr__(self):
return self.name + str(self.flags) + str(self.block)
class Using():
def __init__(self, name, alias):
global global_crc
self.name = name
if isinstance(alias, Array):
a = { 'type': alias.fieldtype,
'length': alias.length }
else:
a = { 'type': alias.fieldtype }
self.alias = a
self.crc = binascii.crc32(str(alias)) & 0xffffffff
global_crc = binascii.crc32(str(alias), global_crc)
global_type_add(name)
def __repr__(self):
return self.name + str(self.alias)
class Union():
def __init__(self, name, block):
self.type = 'Union'
self.manual_print = False
self.manual_endian = False
global global_crc
self.name = name
self.block = block
self.crc = crc_block(block)
global_type_add(name)
def __repr__(self):
return str(self.block)
class Define():
def __init__(self, name, flags, block):
self.name = name
self.flags = flags
self.block = block
self.crc = crc_block(block)
self.dont_trace = False
self.manual_print = False
self.manual_endian = False
self.autoreply = False
self.singular = False
for f in flags:
if f == 'dont_trace':
self.dont_trace = True
elif f == 'manual_print':
self.manual_print = True
elif f == 'manual_endian':
self.manual_endian = True
elif f == 'autoreply':
self.autoreply = True
for b in block:
if isinstance(b, Option):
if b[1] == 'singular' and b[2] == 'true':
self.singular = True
block.remove(b)
def __repr__(self):
return self.name + str(self.flags) + str(self.block)
class Enum():
def __init__(self, name, block, enumtype='u32'):
self.name = name
self.enumtype = enumtype
count = 0
for i, b in enumerate(block):
if type(b) is list:
count = b[1]
else:
count += 1
block[i] = [b, count]
self.block = block
self.crc = crc_block(block)
global_type_add(name)
def __repr__(self):
return self.name + str(self.block)
class Import():
def __init__(self, filename):
self.filename = filename
# Deal with imports
parser = VPPAPI(filename=filename)
dirlist = dirlist_get()
f = filename
for dir in dirlist:
f = os.path.join(dir, filename)
if os.path.exists(f):
break
if sys.version[0] == '2':
with open(f) as fd:
self.result = parser.parse_file(fd, None)
else:
with open(f, encoding='utf-8') as fd:
self.result = parser.parse_file(fd, None)
def __repr__(self):
return self.filename
class Option():
def __init__(self, option):
self.option = option
self.crc = crc_block(option)
def __repr__(self):
return str(self.option)
def __getitem__(self, index):
return self.option[index]
class Array():
def __init__(self, fieldtype, name, length):
self.type = 'Array'
self.fieldtype = fieldtype
self.fieldname = name
if type(length) is str:
self.lengthfield = length
self.length = 0
else:
self.length = length
self.lengthfield = None
def __repr__(self):
return str([self.fieldtype, self.fieldname, self.length,
self.lengthfield])
class Field():
def __init__(self, fieldtype, name):
self.type = 'Field'
self.fieldtype = fieldtype
self.fieldname = name
def __repr__(self):
return str([self.fieldtype, self.fieldname])
class Coord(object):
""" Coordinates of a syntactic element. Consists of:
- File name
- Line number
- (optional) column number, for the Lexer
"""
__slots__ = ('file', 'line', 'column', '__weakref__')
def __init__(self, file, line, column=None):
self.file = file
self.line = line
self.column = column
def __str__(self):
str = "%s:%s" % (self.file, self.line)
if self.column:
str += ":%s" % self.column
return str
class ParseError(Exception):
pass
#
# Grammar rules
#
class VPPAPIParser(object):
tokens = VPPAPILexer.tokens
def __init__(self, filename, logger):
self.filename = filename
self.logger = logger
self.fields = []
def _parse_error(self, msg, coord):
raise ParseError("%s: %s" % (coord, msg))
def _parse_warning(self, msg, coord):
if self.logger:
self.logger.warning("%s: %s" % (coord, msg))
def _coord(self, lineno, column=None):
return Coord(
file=self.filename,
line=lineno, column=column)
def _token_coord(self, p, token_idx):
""" Returns the coordinates for the YaccProduction object 'p' indexed
with 'token_idx'. The coordinate includes the 'lineno' and
'column'. Both follow the lex semantic, starting from 1.
"""
last_cr = p.lexer.lexdata.rfind('\n', 0, p.lexpos(token_idx))
if last_cr < 0:
last_cr = -1
column = (p.lexpos(token_idx) - (last_cr))
return self._coord(p.lineno(token_idx), column)
def p_slist(self, p):
'''slist : stmt
| slist stmt'''
if len(p) == 2:
p[0] = [p[1]]
else:
p[0] = p[1] + [p[2]]
def p_stmt(self, p):
'''stmt : define
| typedef
| option
| import
| enum
| union
| service'''
p[0] = p[1]
def p_import(self, p):
'''import : IMPORT STRING_LITERAL ';' '''
p[0] = Import(p[2])
def p_service(self, p):
'''service : SERVICE '{' service_statements '}' ';' '''
p[0] = p[3]
def p_service_statements(self, p):
'''service_statements : service_statement
| service_statements service_statement'''
if len(p) == 2:
p[0] = [p[1]]
else:
p[0] = p[1] + [p[2]]
def p_service_statement(self, p):
'''service_statement : RPC ID RETURNS NULL ';'
| RPC ID RETURNS ID ';'
| RPC ID RETURNS STREAM ID ';'
| RPC ID RETURNS ID EVENTS event_list ';' '''
if p[2] == p[4]:
# Verify that caller and reply differ
self._parse_error(
'Reply ID ({}) should not be equal to Caller ID'.format(p[2]),
self._token_coord(p, 1))
if len(p) == 8:
p[0] = Service(p[2], p[4], p[6])
elif len(p) == 7:
p[0] = Service(p[2], p[5], stream=True)
else:
p[0] = Service(p[2], p[4])
def p_event_list(self, p):
'''event_list : events
| event_list events '''
if len(p) == 2:
p[0] = [p[1]]
else:
p[0] = p[1] + [p[2]]
def p_event(self, p):
'''events : ID
| ID ',' '''
p[0] = p[1]
def p_enum(self, p):
'''enum : ENUM ID '{' enum_statements '}' ';' '''
p[0] = Enum(p[2], p[4])
def p_enum_type(self, p):
''' enum : ENUM ID ':' enum_size '{' enum_statements '}' ';' '''
if len(p) == 9:
p[0] = Enum(p[2], p[6], enumtype=p[4])
else:
p[0] = Enum(p[2], p[4])
def p_enum_size(self, p):
''' enum_size : U8
| U16
| U32 '''
p[0] = p[1]
def p_define(self, p):
'''define : DEFINE ID '{' block_statements_opt '}' ';' '''
self.fields = []
p[0] = Define(p[2], [], p[4])
def p_define_flist(self, p):
'''define : flist DEFINE ID '{' block_statements_opt '}' ';' '''
# Legacy typedef
if 'typeonly' in p[1]:
p[0] = Typedef(p[3], p[1], p[5])
else:
p[0] = Define(p[3], p[1], p[5])
def p_flist(self, p):
'''flist : flag
| flist flag'''
if len(p) == 2:
p[0] = [p[1]]
else:
p[0] = p[1] + [p[2]]
def p_flag(self, p):
'''flag : MANUAL_PRINT
| MANUAL_ENDIAN
| DONT_TRACE
| TYPEONLY
| AUTOREPLY'''
if len(p) == 1:
return
p[0] = p[1]
def p_typedef(self, p):
'''typedef : TYPEDEF ID '{' block_statements_opt '}' ';' '''
p[0] = Typedef(p[2], [], p[4])
def p_typedef_alias(self, p):
'''typedef : TYPEDEF declaration '''
p[0] = Using(p[2].fieldname, p[2])
def p_block_statements_opt(self, p):
'''block_statements_opt : block_statements '''
p[0] = p[1]
def p_block_statements(self, p):
'''block_statements : block_statement
| block_statements block_statement'''
if len(p) == 2:
p[0] = [p[1]]
else:
p[0] = p[1] + [p[2]]
def p_block_statement(self, p):
'''block_statement : declaration
| option '''
p[0] = p[1]
def p_enum_statements(self, p):
'''enum_statements : enum_statement
| enum_statements enum_statement'''
if len(p) == 2:
p[0] = [p[1]]
else:
p[0] = p[1] + [p[2]]
def p_enum_statement(self, p):
'''enum_statement : ID '=' NUM ','
| ID ',' '''
if len(p) == 5:
p[0] = [p[1], p[3]]
else:
p[0] = p[1]
def p_declaration(self, p):
'''declaration : type_specifier ID ';' '''
if len(p) != 4:
self._parse_error('ERROR')
self.fields.append(p[2])
p[0] = Field(p[1], p[2])
def p_declaration_array(self, p):
'''declaration : type_specifier ID '[' NUM ']' ';'
| type_specifier ID '[' ID ']' ';' '''
if len(p) != 7:
return self._parse_error(
'array: %s' % p.value,
self._coord(lineno=p.lineno))
# Make this error later
if type(p[4]) is int and p[4] == 0:
# XXX: Line number is wrong
self._parse_warning('Old Style VLA: {} {}[{}];'
.format(p[1], p[2], p[4]),
self._token_coord(p, 1))
if type(p[4]) is str and p[4] not in self.fields:
# Verify that length field exists
self._parse_error('Missing length field: {} {}[{}];'
.format(p[1], p[2], p[4]),
self._token_coord(p, 1))
p[0] = Array(p[1], p[2], p[4])
def p_option(self, p):
'''option : OPTION ID '=' assignee ';' '''
p[0] = Option([p[1], p[2], p[4]])
def p_assignee(self, p):
'''assignee : NUM
| TRUE
| FALSE
| STRING_LITERAL '''
p[0] = p[1]
def p_type_specifier(self, p):
'''type_specifier : U8
| U16
| U32
| U64
| I8
| I16
| I32
| I64
| F64
| BOOL
| STRING'''
p[0] = p[1]
# Do a second pass later to verify that user defined types are defined
def p_typedef_specifier(self, p):
'''type_specifier : ID '''
if p[1] not in global_types:
self._parse_error('Undefined type: {}'.format(p[1]),
self._token_coord(p, 1))
p[0] = p[1]
def p_union(self, p):
'''union : UNION ID '{' block_statements_opt '}' ';' '''
p[0] = Union(p[2], p[4])
# Error rule for syntax errors
def p_error(self, p):
if p:
self._parse_error(
'before: %s' % p.value,
self._coord(lineno=p.lineno))
else:
self._parse_error('At end of input', self.filename)
class VPPAPI(object):
def __init__(self, debug=False, filename='', logger=None):
self.lexer = lex.lex(module=VPPAPILexer(filename), debug=debug)
self.parser = yacc.yacc(module=VPPAPIParser(filename, logger),
write_tables=False, debug=debug)
self.logger = logger
def parse_string(self, code, debug=0, lineno=1):
self.lexer.lineno = lineno
return self.parser.parse(code, lexer=self.lexer, debug=debug)
def parse_file(self, fd, debug=0):
data = fd.read()
return self.parse_string(data, debug=debug)
def autoreply_block(self, name):
block = [Field('u32', 'context'),
Field('i32', 'retval')]
return Define(name + '_reply', [], block)
def process(self, objs):
s = {}
s['Option'] = {}
s['Define'] = []
s['Service'] = []
s['types'] = []
s['Import'] = []
s['Alias'] = {}
for o in objs:
tname = o.__class__.__name__
if isinstance(o, Define):
s[tname].append(o)
if o.autoreply:
s[tname].append(self.autoreply_block(o.name))
elif isinstance(o, Option):
s[tname][o[1]] = o[2]
elif type(o) is list:
for o2 in o:
if isinstance(o2, Service):
s['Service'].append(o2)
elif (isinstance(o, Enum) or
isinstance(o, Typedef) or
isinstance(o, Union)):
s['types'].append(o)
elif isinstance(o, Using):
s['Alias'][o.name] = o.alias
else:
if tname not in s:
raise ValueError('Unknown class type: {} {}'
.format(tname, o))
s[tname].append(o)
msgs = {d.name: d for d in s['Define']}
svcs = {s.caller: s for s in s['Service']}
replies = {s.reply: s for s in s['Service']}
seen_services = {}
for service in svcs:
if service not in msgs:
raise ValueError(
'Service definition refers to unknown message'
' definition: {}'.format(service))
if svcs[service].reply != 'null' and \
svcs[service].reply not in msgs:
raise ValueError('Service definition refers to unknown message'
' definition in reply: {}'
.format(svcs[service].reply))
if service in replies:
raise ValueError('Service definition refers to message'
' marked as reply: {}'.format(service))
for event in svcs[service].events:
if event not in msgs:
raise ValueError('Service definition refers to unknown '
'event: {} in message: {}'
.format(event, service))
seen_services[event] = True
# Create services implicitly
for d in msgs:
if d in seen_services:
continue
if msgs[d].singular is True:
continue
if d.endswith('_reply'):
if d[:-6] in svcs:
continue
if d[:-6] not in msgs:
raise ValueError('{} missing calling message'
.format(d))
continue
if d.endswith('_dump'):
if d in svcs:
continue
if d[:-5]+'_details' in msgs:
s['Service'].append(Service(d, d[:-5]+'_details',
stream=True))
else:
raise ValueError('{} missing details message'
.format(d))
continue
if d.endswith('_details'):
if d[:-8]+'_dump' not in msgs:
raise ValueError('{} missing dump message'
.format(d))
continue
if d in svcs:
continue
if d+'_reply' in msgs:
s['Service'].append(Service(d, d+'_reply'))
else:
raise ValueError(
'{} missing reply message ({}) or service definition'
.format(d, d+'_reply'))
return s
def process_imports(self, objs, in_import, result):
imported_objs = []
for o in objs:
# Only allow the following object types from imported file
if in_import and not (isinstance(o, Enum) or
isinstance(o, Union) or
isinstance(o, Typedef) or
isinstance(o, Import) or
isinstance(o, Using)):
continue
if isinstance(o, Import):
self.process_imports(o.result, True, result)
else:
result.append(o)
# Add message ids to each message.
def add_msg_id(s):
for o in s:
o.block.insert(0, Field('u16', '_vl_msg_id'))
return s
dirlist = []
def dirlist_add(dirs):
global dirlist
if dirs:
dirlist = dirlist + dirs
def dirlist_get():
return dirlist
#
# Main
#
def main():
cliparser = argparse.ArgumentParser(description='VPP API generator')
cliparser.add_argument('--pluginpath', default=""),
cliparser.add_argument('--includedir', action='append'),
if sys.version[0] == '2':
cliparser.add_argument('--input', type=argparse.FileType('r'),
default=sys.stdin)
else:
cliparser.add_argument('--input',
type=argparse.FileType('r', encoding='UTF-8'),
default=sys.stdin)
cliparser.add_argument('--output', nargs='?', type=argparse.FileType('w'),
default=sys.stdout)
cliparser.add_argument('output_module', nargs='?', default='C')
cliparser.add_argument('--debug', action='store_true')
cliparser.add_argument('--show-name', nargs=1)
args = cliparser.parse_args()
dirlist_add(args.includedir)
if not args.debug:
sys.excepthook = exception_handler
# Filename
if args.show_name:
filename = args.show_name[0]
elif args.input != sys.stdin:
filename = args.input.name
else:
filename = ''
if args.debug:
logging.basicConfig(stream=sys.stdout, level=logging.WARNING)
else:
logging.basicConfig()
log = logging.getLogger('vppapigen')
parser = VPPAPI(debug=args.debug, filename=filename, logger=log)
parsed_objects = parser.parse_file(args.input, log)
# Build a list of objects. Hash of lists.
result = []
parser.process_imports(parsed_objects, False, result)
s = parser.process(result)
# Add msg_id field
s['Define'] = add_msg_id(s['Define'])
file_crc = global_crc & 0xffffffff
#
# Debug
if args.debug:
import pprint
pp = pprint.PrettyPrinter(indent=4, stream=sys.stderr)
for t in s['Define']:
pp.pprint([t.name, t.flags, t.block])
for t in s['types']:
pp.pprint([t.name, t.block])
#
# Generate representation
#
import imp
# Default path
pluginpath = ''
if not args.pluginpath:
cand = []
cand.append(os.path.dirname(os.path.realpath(__file__)))
cand.append(os.path.dirname(os.path.realpath(__file__)) +
'/../share/vpp/')
for c in cand:
c += '/'
if os.path.isfile('{}vppapigen_{}.py'
.format(c, args.output_module.lower())):
pluginpath = c
break
else:
pluginpath = args.pluginpath + '/'
if pluginpath == '':
raise Exception('Output plugin not found')
module_path = '{}vppapigen_{}.py'.format(pluginpath,
args.output_module.lower())
try:
plugin = imp.load_source(args.output_module, module_path)
except Exception as err:
raise Exception('Error importing output plugin: {}, {}'
.format(module_path, err))
result = plugin.run(filename, s, file_crc)
if result:
print(result, file=args.output)
else:
raise Exception('Running plugin failed: {} {}'
.format(filename, result))
if __name__ == '__main__':
main()