Added new ast nodes when new_id != -1
This commit is contained in:
@@ -38,16 +38,27 @@ std::pair<Token_Iterator, bool> Type_Rule(Token_Iterator iter, Token_Iterator en
|
||||
std::pair<Token_Iterator, bool> Or_Rule(Token_Iterator iter, Token_Iterator end, TokenPtr parent, bool keep, int new_id, Rule lhs, Rule rhs) {
|
||||
Token_Iterator new_iter;
|
||||
unsigned int prev_size = parent->children.size();
|
||||
TokenPtr prev_parent = parent;
|
||||
|
||||
if (new_id != -1) {
|
||||
parent = TokenPtr(new Token("", new_id, parent->filename));
|
||||
}
|
||||
|
||||
if (*iter != *end) {
|
||||
std::pair<Token_Iterator, bool> result = lhs(iter, end, parent);
|
||||
|
||||
if (result.second) {
|
||||
if (new_id != -1) {
|
||||
prev_parent->children.push_back(parent);
|
||||
}
|
||||
return std::pair<Token_Iterator, bool>(result.first, true);
|
||||
}
|
||||
else {
|
||||
result = rhs(iter, end, parent);
|
||||
if (result.second) {
|
||||
if (new_id != -1) {
|
||||
prev_parent->children.push_back(parent);
|
||||
}
|
||||
return std::pair<Token_Iterator, bool>(result.first, true);
|
||||
}
|
||||
}
|
||||
@@ -63,7 +74,14 @@ std::pair<Token_Iterator, bool> Or_Rule(Token_Iterator iter, Token_Iterator end,
|
||||
|
||||
std::pair<Token_Iterator, bool> And_Rule(Token_Iterator iter, Token_Iterator end, TokenPtr parent, bool keep, int new_id, Rule lhs, Rule rhs) {
|
||||
Token_Iterator lhs_iter, rhs_iter;
|
||||
unsigned int prev_size = parent->children.size();
|
||||
unsigned int prev_size;
|
||||
TokenPtr prev_parent = parent;
|
||||
|
||||
if (new_id != -1) {
|
||||
parent = TokenPtr(new Token("", new_id, parent->filename));
|
||||
}
|
||||
|
||||
prev_size = parent->children.size();
|
||||
|
||||
if (*iter != *end) {
|
||||
std::pair<Token_Iterator, bool> result = lhs(iter, end, parent);
|
||||
@@ -71,6 +89,9 @@ std::pair<Token_Iterator, bool> And_Rule(Token_Iterator iter, Token_Iterator end
|
||||
if (result.second) {
|
||||
result = rhs(result.first, end, parent);
|
||||
if (result.second) {
|
||||
if (new_id != -1) {
|
||||
prev_parent->children.push_back(parent);
|
||||
}
|
||||
return std::pair<Token_Iterator, bool>(result.first, true);
|
||||
}
|
||||
}
|
||||
|
@@ -49,7 +49,9 @@ struct Rule {
|
||||
}
|
||||
|
||||
Rule &operator=(const Rule &rule) {
|
||||
*impl = *(rule.impl);
|
||||
//*impl = *(rule.impl);
|
||||
impl->rule = rule.impl->rule;
|
||||
impl->keep = rule.impl->keep;
|
||||
|
||||
return *this;
|
||||
}
|
||||
|
@@ -11,7 +11,8 @@
|
||||
#include "langkit_parser.hpp"
|
||||
|
||||
class TokenType { public: enum Type { Whitespace, Identifier, Number, Operator, Parens_Open, Parens_Close,
|
||||
Square_Open, Square_Close, Curly_Open, Curly_Close, Comma, Quoted_String, Single_Quoted_String, Carriage_Return, Semicolon }; };
|
||||
Square_Open, Square_Close, Curly_Open, Curly_Close, Comma, Quoted_String, Single_Quoted_String, Carriage_Return, Semicolon,
|
||||
Function_Def}; };
|
||||
|
||||
void debug_print(TokenPtr token, std::string prepend) {
|
||||
std::cout << prepend << "Token: " << token->text << "(" << token->identifier << ") @ " << token->filename << ": (" << token->start.column
|
||||
@@ -49,14 +50,23 @@ std::string load_file(const char *filename) {
|
||||
|
||||
void parse(std::vector<TokenPtr> &tokens) {
|
||||
|
||||
/*
|
||||
Rule lhs;
|
||||
Rule rhs;
|
||||
Rule rule = lhs << rhs;
|
||||
lhs = Str("def", true);
|
||||
rhs = Id(TokenType::Identifier, true);
|
||||
*/
|
||||
|
||||
//Rule rule(TokenType::Function_Def);
|
||||
//rule = Str("def") | Str("int");
|
||||
|
||||
//Rule rule = Str("def", false) << Id(TokenType::Identifier, true);
|
||||
//Rule rule = Str("def", false) << Id(TokenType::Identifier);
|
||||
|
||||
Rule params;
|
||||
Rule rule(TokenType::Function_Def);
|
||||
rule = Ign(Str("def")) << Id(TokenType::Identifier) << Ign(Str("(")) << params << Ign(Str(")"));
|
||||
params = Id(TokenType::Identifier) << Ign(Str(",")) << Id(TokenType::Identifier);
|
||||
|
||||
Token_Iterator iter = tokens.begin(), end = tokens.end();
|
||||
TokenPtr parent(new Token("Root", 0, "test"));
|
||||
|
Reference in New Issue
Block a user