mirror of
https://github.com/AleoHQ/leo.git
synced 2024-12-19 15:41:36 +03:00
fix some span and input parsing issues
This commit is contained in:
parent
b955b77ac2
commit
c8b44141eb
@ -154,7 +154,7 @@ pub trait ReconstructingReducer {
|
||||
value: Expression,
|
||||
) -> Result<DefinitionStatement> {
|
||||
Ok(DefinitionStatement {
|
||||
declaration_type: definition.declaration_type.clone(),
|
||||
declaration_type: definition.declaration_type,
|
||||
variable_names,
|
||||
type_,
|
||||
value,
|
||||
|
@ -114,7 +114,7 @@ impl<'a> ParserContext<'a> {
|
||||
Some(idx) => idx,
|
||||
};
|
||||
|
||||
looker(self.tokens.get(idx).unwrap_or_else(|| &self.dummy_eof))
|
||||
looker(self.tokens.get(idx).unwrap_or(&self.dummy_eof))
|
||||
}
|
||||
|
||||
/// Emit the error `err`.
|
||||
|
@ -100,7 +100,7 @@ impl ParserContext<'_> {
|
||||
}
|
||||
|
||||
self.expect(&Token::Colon)?;
|
||||
let type_ = self.parse_type()?.0;
|
||||
let type_ = self.parse_all_types()?.0;
|
||||
Ok(FunctionInput::Variable(FunctionInputVariable::new(
|
||||
name.clone(),
|
||||
mode,
|
||||
@ -124,7 +124,7 @@ impl ParserContext<'_> {
|
||||
|
||||
// Parse return type.
|
||||
let output = if self.eat(&Token::Arrow) {
|
||||
Some(self.parse_type()?.0)
|
||||
Some(self.parse_all_types()?.0)
|
||||
} else {
|
||||
None
|
||||
};
|
||||
|
@ -65,7 +65,7 @@ impl ParserContext<'_> {
|
||||
|
||||
let name = self.expect_ident()?;
|
||||
self.expect(&Token::Colon)?;
|
||||
let (type_, span) = self.parse_type()?;
|
||||
let (type_, span) = self.parse_non_ident_types()?;
|
||||
self.expect(&Token::Assign)?;
|
||||
let value = self.parse_primary_expression()?;
|
||||
self.expect(&Token::Semicolon)?;
|
||||
|
@ -252,8 +252,6 @@ impl ParserContext<'_> {
|
||||
Token::Const => Declare::Const,
|
||||
_ => unreachable!("parse_definition_statement_ shouldn't produce this"),
|
||||
};
|
||||
dbg!();
|
||||
|
||||
// Parse variable names.
|
||||
let variable_names = if self.peek_is_left_par() {
|
||||
let vars = self
|
||||
@ -272,7 +270,7 @@ impl ParserContext<'_> {
|
||||
// Parse an optional type ascription.
|
||||
let type_ = self
|
||||
.eat(&Token::Colon)
|
||||
.then(|| self.parse_type().map(|t| t.0))
|
||||
.then(|| self.parse_all_types().map(|t| t.0))
|
||||
.transpose()?;
|
||||
|
||||
self.expect(&Token::Assign)?;
|
||||
|
@ -57,23 +57,29 @@ impl ParserContext<'_> {
|
||||
|
||||
/// Returns a [`(Type, Span)`] tuple of AST nodes if the next token represents a type.
|
||||
/// Also returns the span of the parsed token.
|
||||
pub fn parse_type(&mut self) -> Result<(Type, Span)> {
|
||||
pub fn parse_non_ident_types(&mut self) -> Result<(Type, Span)> {
|
||||
let span = self.expect_any(TYPE_TOKENS)?;
|
||||
Ok((
|
||||
match &self.prev_token.token {
|
||||
Token::Field => Type::Field,
|
||||
Token::Group => Type::Group,
|
||||
Token::Address => Type::Address,
|
||||
Token::Bool => Type::Boolean,
|
||||
Token::Char => Type::Char,
|
||||
x => Type::IntegerType(Self::token_to_int_type(x).expect("invalid int type")),
|
||||
},
|
||||
span,
|
||||
))
|
||||
}
|
||||
|
||||
/// Returns a [`(Type, Span)`] tuple of AST nodes if the next token represents a type.
|
||||
/// Also returns the span of the parsed token.
|
||||
pub fn parse_all_types(&mut self) -> Result<(Type, Span)> {
|
||||
Ok(if let Some(ident) = self.eat_identifier() {
|
||||
let span = ident.span.clone();
|
||||
(Type::Identifier(ident), span)
|
||||
} else {
|
||||
let span = self.expect_any(TYPE_TOKENS)?;
|
||||
(
|
||||
match &self.prev_token.token {
|
||||
Token::Field => Type::Field,
|
||||
Token::Group => Type::Group,
|
||||
Token::Address => Type::Address,
|
||||
Token::Bool => Type::Boolean,
|
||||
Token::Char => Type::Char,
|
||||
x => Type::IntegerType(Self::token_to_int_type(x).expect("invalid int type")),
|
||||
},
|
||||
span,
|
||||
)
|
||||
self.parse_non_ident_types()?
|
||||
})
|
||||
}
|
||||
}
|
||||
|
@ -46,7 +46,7 @@ pub(crate) fn tokenize_iter<'a>(path: &'a str, input: &'a str) -> impl 'a + Iter
|
||||
iter::from_fn(move || {
|
||||
while input.len() > index {
|
||||
let token = match Token::eat(&input[index..]) {
|
||||
Err(e) => return Some(Err(e.into())),
|
||||
Err(e) => return Some(Err(e)),
|
||||
Ok(t) => t,
|
||||
};
|
||||
|
||||
@ -56,7 +56,7 @@ pub(crate) fn tokenize_iter<'a>(path: &'a str, input: &'a str) -> impl 'a + Iter
|
||||
if bytes[index] == 0x000D && matches!(bytes.get(index + 1), Some(0x000A)) {
|
||||
// Check carriage return followed by newline.
|
||||
line_no += 1;
|
||||
line_start = index + token_len;
|
||||
line_start = index + token_len + 1;
|
||||
index += token_len;
|
||||
} else if matches!(bytes[index], 0x000A | 0x000D) {
|
||||
// Check new-line or carriage-return
|
||||
|
Loading…
Reference in New Issue
Block a user