Move index handling from expression parsing into value parsing

This commit is contained in:
Elara 2024-11-13 12:34:53 -08:00
parent ab9a7bb34e
commit 2278a16ca0
3 changed files with 31 additions and 19 deletions

View File

@ -35,7 +35,7 @@ class Interp:
def _eval_index(self, index: ast.Index) -> typing.Any:
val = self._convert_value(index.value)
if not hasattr(val, '__getitem__'):
raise ValueError(f'{index.value.pos}: value is not indexable ({type(val).__getitem__})')
raise ValueError(f'{index.value.pos}: value is not indexable ({type(val).__name__})')
index_val = self._convert_value(index.index)
if type(index_val) is int and hasattr(val, '__len__') and index_val >= len(val):
raise IndexError(f'{index.index.pos}: index out of range ({index_val} with length {len(val)})')

View File

@ -34,15 +34,14 @@ class Parser:
tok, pos, lit = self._scan()
if tok != lexer.Token.SQUARE or lit != ']':
raise ExpectedError(pos, 'closing square bracket', lit)
while self.lexer._peek(1) == '[':
self._scan()
index = self._parse_index(index)
return index
def _parse_expr(self) -> ast.Value:
left = self._parse_value()
tok, pos, lit = self._scan()
while tok == lexer.Token.SQUARE and lit == '[':
left = self._parse_index(left)
# Scan the next token for the next if statement
tok, pos, lit = self._scan()
if tok != lexer.Token.OPERATOR:
self._unscan(tok, pos, lit)
return left
@ -117,33 +116,35 @@ class Parser:
return ast.FunctionCall(pos=id_pos, name=id_lit, args=args)
def _parse_value(self) -> ast.Value:
out = None
tok, pos, lit = self._scan()
match tok:
case lexer.Token.INTEGER:
return ast.Integer(pos=pos, value=int(lit))
out = ast.Integer(pos=pos, value=int(lit))
case lexer.Token.FLOAT:
return ast.Float(pos=pos, value=float(lit))
out = ast.Float(pos=pos, value=float(lit))
case lexer.Token.BOOL:
return ast.Bool(pos=pos, value=(lit == 'true'))
out = ast.Bool(pos=pos, value=(lit == 'true'))
case lexer.Token.STRING:
return ast.String(pos=pos, value=pyast.literal_eval(lit))
out = ast.String(pos=pos, value=pyast.literal_eval(lit))
case lexer.Token.IDENT:
if self.lexer._peek(1) == '(':
self._unscan(tok, pos, lit)
return self._parse_func_call()
return ast.VariableRef(pos=pos, name=lit)
out = self._parse_func_call()
else:
out = ast.VariableRef(pos=pos, name=lit)
case lexer.Token.HEREDOC:
return ast.String(pos=pos, value=lit)
out = ast.String(pos=pos, value=lit)
case lexer.Token.OPERATOR:
return ast.UnaryExpression(pos=pos, op=ast.Operator(pos=pos, value=lit), value=self._parse_value())
out = ast.UnaryExpression(pos=pos, op=ast.Operator(pos=pos, value=lit), value=self._parse_value())
case lexer.Token.SQUARE:
if lit != '[':
raise ExpectedError(pos, repr('['), lit)
return self._parse_tuple(pos)
out = self._parse_tuple(pos)
case lexer.Token.CURLY:
if lit != '{':
raise ExpectedError(pos, repr('{'), lit)
return self._parse_object(pos)
out = self._parse_object(pos)
case lexer.Token.PAREN:
if lit != '(':
raise ExpectedError(pos, repr('('), lit)
@ -151,9 +152,15 @@ class Parser:
tok, pos, lit = self._scan()
if tok != lexer.Token.PAREN or lit != ')':
raise ExpectedError(pos, repr(')'), lit)
return expr
out = expr
case _:
raise ExpectedError(pos, 'value', lit)
raise ExpectedError(pos, 'value', lit)
if self.lexer._peek(1) == '[':
self._scan()
out = self._parse_index(out)
return out
def parse(self, until: tuple[lexer.Token, str] = (lexer.Token.EOF, '')) -> ast.AST:
tree = []

View File

@ -104,12 +104,17 @@ class TestRefs(unittest.TestCase):
self.assertEqual(cfg['x'], 456)
def test_index(self):
i = interp.Interp(io.StringIO("x = y[1]"), "TestRefs.test_index")
i = interp.Interp(io.StringIO('x = y[1]'), "TestRefs.test_index")
i['y'] = [123, 456, 789]
cfg = i.run()
self.assertIn('x', cfg)
self.assertEqual(cfg['x'], 456)
def test_multi_index(self):
cfg = interp.Interp(io.StringIO('x = ["123", "456", "789"][1][2]'), "TestRefs.test_multi_index").run()
self.assertIn('x', cfg)
self.assertEqual(cfg['x'], '6')
def test_func(self):
def y(a, b):
return a + b