100 lines
3.1 KiB
C++
100 lines
3.1 KiB
C++
// This file is distributed under the BSD License.
|
|
// See LICENSE.TXT for details.
|
|
|
|
#include <boost/bind.hpp>
|
|
|
|
#include <iostream>
|
|
#include <map>
|
|
#include <fstream>
|
|
|
|
#include "langkit_lexer.hpp"
|
|
#include "langkit_parser.hpp"
|
|
|
|
class TokenType { public: enum Type { Whitespace, Identifier, Number, Operator, Parens_Open, Parens_Close,
|
|
Square_Open, Square_Close, Curly_Open, Curly_Close, Comma, Quoted_String, Single_Quoted_String, Carriage_Return, Semicolon }; };
|
|
|
|
void debug_print(Token &token) {
|
|
std::cout << "Token: " << token.text << "(" << token.identifier << ") @ (" << token.start.column
|
|
<< ", " << token.start.line << ") to (" << token.end.column << ", " << token.end.line << ") " << std::endl;
|
|
}
|
|
void debug_print(std::vector<Token> &tokens) {
|
|
for (unsigned int i = 0; i < tokens.size(); ++i) {
|
|
debug_print(tokens[i]);
|
|
}
|
|
}
|
|
|
|
std::string load_file(const char *filename) {
|
|
std::ifstream infile (filename, std::ios::in | std::ios::ate);
|
|
|
|
if (!infile.is_open()) {
|
|
std::cerr << "Can not open " << filename << std::endl;
|
|
exit(0);
|
|
}
|
|
|
|
std::streampos size = infile.tellg();
|
|
infile.seekg(0, std::ios::beg);
|
|
|
|
std::vector<char> v(size);
|
|
infile.read(&v[0], size);
|
|
|
|
std::string ret_val (v.empty() ? std::string() : std::string (v.begin(), v.end()).c_str());
|
|
|
|
return ret_val;
|
|
}
|
|
|
|
void parse(std::vector<Token> &tokens) {
|
|
Rule rule;
|
|
rule.rule = boost::bind(String_Rule, _1, _2, "def");
|
|
|
|
Token_Iterator iter = tokens.begin(), end = tokens.end();
|
|
|
|
|
|
std::pair<Token_Iterator, Token> results = rule(iter, end);
|
|
|
|
if (results.first != iter) {
|
|
std::cout << "Parse successful: " << std::endl;
|
|
debug_print(results.second);
|
|
}
|
|
|
|
}
|
|
|
|
|
|
int main(int argc, char *argv[]) {
|
|
std::string input;
|
|
|
|
Lexer lexer;
|
|
lexer.set_skip(Pattern("[ \\t]+", TokenType::Whitespace));
|
|
lexer.set_line_sep(Pattern("\\n|\\r\\n", TokenType::Carriage_Return));
|
|
lexer.set_command_sep(Pattern(";|\\r\\n|\\n", TokenType::Semicolon));
|
|
|
|
lexer << Pattern("[A-Za-z]+", TokenType::Identifier);
|
|
lexer << Pattern("[0-9]+(\\.[0-9]+)?", TokenType::Number);
|
|
lexer << Pattern("[!@#$%^&*\\-+=/<>]+", TokenType::Operator);
|
|
lexer << Pattern("\\(", TokenType::Parens_Open);
|
|
lexer << Pattern("\\)", TokenType::Parens_Close);
|
|
lexer << Pattern("\\[", TokenType::Square_Open);
|
|
lexer << Pattern("\\]", TokenType::Square_Close);
|
|
lexer << Pattern("\\{", TokenType::Curly_Open);
|
|
lexer << Pattern("\\}", TokenType::Curly_Close);
|
|
lexer << Pattern(",", TokenType::Comma);
|
|
lexer << Pattern("\"(?:[^\"\\\\]|\\\\.)*\"", TokenType::Quoted_String);
|
|
lexer << Pattern("'(?:[^'\\\\]|\\\\.)*'", TokenType::Single_Quoted_String);
|
|
|
|
if (argc < 2) {
|
|
std::cout << "Expression> ";
|
|
std::getline(std::cin, input);
|
|
while (input != "quit") {
|
|
std::vector<Token> tokens = lexer.lex(input);
|
|
debug_print(tokens);
|
|
parse(tokens);
|
|
|
|
std::cout << "Expression> ";
|
|
std::getline(std::cin, input);
|
|
}
|
|
}
|
|
else {
|
|
std::vector<Token> tokens = lexer.lex(load_file(argv[1]));
|
|
debug_print(tokens);
|
|
}
|
|
}
|