Remove useless code

This commit is contained in:
freestrings 2020-02-16 22:16:22 +09:00
parent 32eef3f7c7
commit 7a07bc7744
2 changed files with 167 additions and 144 deletions

View File

@ -13,8 +13,8 @@ mod utils {
use std::str::FromStr; use std::str::FromStr;
pub fn string_to_num<F, S: FromStr>(string: &str, msg_handler: F) -> Result<S, String> pub fn string_to_num<F, S: FromStr>(string: &str, msg_handler: F) -> Result<S, String>
where where
F: Fn() -> String, F: Fn() -> String,
{ {
match string.parse() { match string.parse() {
Ok(n) => Ok(n), Ok(n) => Ok(n),
@ -209,7 +209,8 @@ impl Parser {
fn array_keys(tokenizer: &mut TokenReader, first_key: String) -> ParseResult<Node> { fn array_keys(tokenizer: &mut TokenReader, first_key: String) -> ParseResult<Node> {
let mut keys = vec![first_key]; let mut keys = vec![first_key];
while tokenizer.peek_is(COMMA) {
while let Ok(Token::Comma(_)) = tokenizer.peek_token() {
Self::eat_token(tokenizer); Self::eat_token(tokenizer);
Self::eat_whitespace(tokenizer); Self::eat_whitespace(tokenizer);
@ -230,7 +231,7 @@ impl Parser {
debug!("#array_quote_value"); debug!("#array_quote_value");
match tokenizer.next_token() { match tokenizer.next_token() {
Ok(Token::SingleQuoted(_, val)) | Ok(Token::DoubleQuoted(_, val)) => { Ok(Token::SingleQuoted(_, val)) | Ok(Token::DoubleQuoted(_, val)) => {
if tokenizer.peek_is(COMMA) { if let Ok(Token::Comma(_)) = tokenizer.peek_token() {
Self::array_keys(tokenizer, val) Self::array_keys(tokenizer, val)
} else { } else {
Ok(Self::node(ParseToken::Key(val))) Ok(Self::node(ParseToken::Key(val)))
@ -529,41 +530,35 @@ impl Parser {
fn term(tokenizer: &mut TokenReader) -> ParseResult<Node> { fn term(tokenizer: &mut TokenReader) -> ParseResult<Node> {
debug!("#term"); debug!("#term");
if tokenizer.peek_is(AT) { return match tokenizer.peek_token() {
Self::eat_token(tokenizer); Ok(Token::At(_)) => {
let node = Self::node(ParseToken::Relative); Self::eat_token(tokenizer);
let node = Self::node(ParseToken::Relative);
return match tokenizer.peek_token() { match tokenizer.peek_token() {
Ok(Token::Whitespace(_, _)) => { Ok(Token::Whitespace(_, _)) => {
Self::eat_whitespace(tokenizer); Self::eat_whitespace(tokenizer);
Ok(node) Ok(node)
}
_ => Self::paths(node, tokenizer),
} }
_ => Self::paths(node, tokenizer), }
}; Ok(Token::Absolute(_)) => {
} Self::json_path(tokenizer)
},
if tokenizer.peek_is(ABSOLUTE) { Ok(Token::DoubleQuoted(_, _)) | Ok(Token::SingleQuoted(_, _)) => {
return Self::json_path(tokenizer); Self::array_quote_value(tokenizer)
} },
Ok(Token::Key(_, key)) => {
if tokenizer.peek_is(DOUBLE_QUOTE) || tokenizer.peek_is(SINGLE_QUOTE) { match key.as_bytes()[0] {
return Self::array_quote_value(tokenizer); b'-' | b'0'..=b'9' => Self::term_num(tokenizer),
} _ => Self::boolean(tokenizer),
}
if tokenizer.peek_is(KEY) { }
let key = if let Ok(Token::Key(_, k)) = tokenizer.peek_token() { _ => {
k.clone() Err(tokenizer.err_msg())
} else { }
unreachable!() };
};
return match key.as_bytes()[0] {
b'-' | b'0'..=b'9' => Self::term_num(tokenizer),
_ => Self::boolean(tokenizer),
};
}
Err(tokenizer.err_msg())
} }
fn op(prev: Node, tokenizer: &mut TokenReader) -> ParseResult<Node> { fn op(prev: Node, tokenizer: &mut TokenReader) -> ParseResult<Node> {
@ -610,7 +605,7 @@ impl Parser {
fn close_token(ret: Node, token: Token, tokenizer: &mut TokenReader) -> ParseResult<Node> { fn close_token(ret: Node, token: Token, tokenizer: &mut TokenReader) -> ParseResult<Node> {
debug!("#close_token"); debug!("#close_token");
match tokenizer.next_token() { match tokenizer.next_token() {
Ok(ref t) if t.partial_eq(token) => Ok(ret), Ok(ref t) if t.is_match_token_type(token) => Ok(ret),
_ => Err(tokenizer.err_msg()), _ => Err(tokenizer.err_msg()),
} }
} }

View File

@ -2,30 +2,6 @@ use std::result::Result;
use super::path_reader::{PathReader, ReaderError}; use super::path_reader::{PathReader, ReaderError};
pub const ABSOLUTE: &str = "$";
pub const DOT: &str = ".";
pub const AT: &str = "@";
pub const OPEN_ARRAY: &str = "[";
pub const CLOSE_ARRAY: &str = "]";
pub const ASTERISK: &str = "*";
pub const QUESTION: &str = "?";
pub const COMMA: &str = ",";
pub const SPLIT: &str = ":";
pub const OPEN_PARENTHESIS: &str = "(";
pub const CLOSE_PARENTHESIS: &str = ")";
pub const KEY: &str = "Key";
pub const DOUBLE_QUOTE: &str = "\"";
pub const SINGLE_QUOTE: &str = "'";
pub const EQUAL: &str = "==";
pub const GREATER_OR_EQUAL: &str = ">=";
pub const GREATER: &str = ">";
pub const LITTLE: &str = "<";
pub const LITTLE_OR_EQUAL: &str = "<=";
pub const NOT_EQUAL: &str = "!=";
pub const AND: &str = "&&";
pub const OR: &str = "||";
pub const WHITESPACE: &str = " ";
const CH_DOLLA: char = '$'; const CH_DOLLA: char = '$';
const CH_DOT: char = '.'; const CH_DOT: char = '.';
const CH_ASTERISK: char = '*'; const CH_ASTERISK: char = '*';
@ -86,61 +62,104 @@ pub enum Token {
} }
impl Token { impl Token {
pub fn partial_eq(&self, other: Token) -> bool { pub fn is_match_token_type(&self, other: Token) -> bool {
self.to_simple() == other.to_simple()
}
pub fn simple_eq(&self, str_token: &str) -> bool {
self.to_simple() == str_token
}
#[cfg_attr(tarpaulin, skip)]
fn to_simple(&self) -> &'static str {
match self { match self {
Token::Absolute(_) => ABSOLUTE, Token::Absolute(_) => match other {
Token::Dot(_) => DOT, Token::Absolute(_) => true,
Token::At(_) => AT, _ => false
Token::OpenArray(_) => OPEN_ARRAY, },
Token::CloseArray(_) => CLOSE_ARRAY, Token::Dot(_) => match other {
Token::Asterisk(_) => ASTERISK, Token::Dot(_) => true,
Token::Question(_) => QUESTION, _ => false
Token::Comma(_) => COMMA, },
Token::Split(_) => SPLIT, Token::At(_) => match other {
Token::OpenParenthesis(_) => OPEN_PARENTHESIS, Token::At(_) => true,
Token::CloseParenthesis(_) => CLOSE_PARENTHESIS, _ => false
Token::Key(_, _) => KEY, },
Token::DoubleQuoted(_, _) => DOUBLE_QUOTE, Token::OpenArray(_) => match other {
Token::SingleQuoted(_, _) => SINGLE_QUOTE, Token::OpenArray(_) => true,
Token::Equal(_) => EQUAL, _ => false
Token::GreaterOrEqual(_) => GREATER_OR_EQUAL, },
Token::Greater(_) => GREATER, Token::CloseArray(_) => match other {
Token::Little(_) => LITTLE, Token::CloseArray(_) => true,
Token::LittleOrEqual(_) => LITTLE_OR_EQUAL, _ => false
Token::NotEqual(_) => NOT_EQUAL, },
Token::And(_) => AND, Token::Asterisk(_) => match other {
Token::Or(_) => OR, Token::Asterisk(_) => true,
Token::Whitespace(_, _) => WHITESPACE, _ => false
},
Token::Question(_) => match other {
Token::Question(_) => true,
_ => false
},
Token::Comma(_) => match other {
Token::Comma(_) => true,
_ => false
},
Token::Split(_) => match other {
Token::Split(_) => true,
_ => false
},
Token::OpenParenthesis(_) => match other {
Token::OpenParenthesis(_) => true,
_ => false
},
Token::CloseParenthesis(_) => match other {
Token::CloseParenthesis(_) => true,
_ => false
},
Token::Key(_, _) => match other {
Token::Key(_, _) => true,
_ => false
},
Token::DoubleQuoted(_, _) => match other {
Token::DoubleQuoted(_, _) => true,
_ => false
},
Token::SingleQuoted(_, _) => match other {
Token::SingleQuoted(_, _) => true,
_ => false
},
Token::Equal(_) => match other {
Token::Equal(_) => true,
_ => false
},
Token::GreaterOrEqual(_) => match other {
Token::GreaterOrEqual(_) => true,
_ => false
},
Token::Greater(_) => match other {
Token::Greater(_) => true,
_ => false
},
Token::Little(_) => match other {
Token::Little(_) => true,
_ => false
},
Token::LittleOrEqual(_) => match other {
Token::LittleOrEqual(_) => true,
_ => false
},
Token::NotEqual(_) => match other {
Token::NotEqual(_) => true,
_ => false
},
Token::And(_) => match other {
Token::And(_) => true,
_ => false
},
Token::Or(_) => match other {
Token::Or(_) => true,
_ => false
},
Token::Whitespace(_, _) => match other {
Token::Whitespace(_, _) => true,
_ => false
},
} }
} }
} }
fn simple_matched_token(ch: char, pos: usize) -> Option<Token> {
match ch {
CH_DOLLA => Some(Token::Absolute(pos)),
CH_DOT => Some(Token::Dot(pos)),
CH_ASTERISK => Some(Token::Asterisk(pos)),
CH_LARRAY => Some(Token::OpenArray(pos)),
CH_RARRAY => Some(Token::CloseArray(pos)),
CH_LPAREN => Some(Token::OpenParenthesis(pos)),
CH_RPAREN => Some(Token::CloseParenthesis(pos)),
CH_AT => Some(Token::At(pos)),
CH_QUESTION => Some(Token::Question(pos)),
CH_COMMA => Some(Token::Comma(pos)),
CH_SEMICOLON => Some(Token::Split(pos)),
_ => None,
}
}
pub struct Tokenizer<'a> { pub struct Tokenizer<'a> {
input: PathReader<'a>, input: PathReader<'a>,
} }
@ -261,17 +280,25 @@ impl<'a> Tokenizer<'a> {
} }
fn other(&mut self, pos: usize, ch: char) -> Result<Token, TokenError> { fn other(&mut self, pos: usize, ch: char) -> Result<Token, TokenError> {
let fun = |c: &char| match simple_matched_token(*c, pos) { let fun = |c: &char| match c {
Some(_) => false, &CH_DOLLA
_ if c == &CH_LITTLE | &CH_DOT
|| c == &CH_GREATER | &CH_ASTERISK
|| c == &CH_EQUAL | &CH_LARRAY
|| c == &CH_AMPERSAND | &CH_RARRAY
|| c == &CH_PIPE | &CH_LPAREN
|| c == &CH_EXCLAMATION => | &CH_RPAREN
{ | &CH_AT
false | &CH_QUESTION
} | &CH_COMMA
| &CH_SEMICOLON
| &CH_LITTLE
| &CH_GREATER
| &CH_EQUAL
| &CH_AMPERSAND
| &CH_PIPE
| &CH_EXCLAMATION
=> false,
_ => !c.is_whitespace(), _ => !c.is_whitespace(),
}; };
let (_, mut vec) = self.input.take_while(fun).map_err(to_token_error)?; let (_, mut vec) = self.input.take_while(fun).map_err(to_token_error)?;
@ -281,20 +308,28 @@ impl<'a> Tokenizer<'a> {
pub fn next_token(&mut self) -> Result<Token, TokenError> { pub fn next_token(&mut self) -> Result<Token, TokenError> {
let (pos, ch) = self.input.next_char().map_err(to_token_error)?; let (pos, ch) = self.input.next_char().map_err(to_token_error)?;
match simple_matched_token(ch, pos) { match ch {
Some(t) => Ok(t), CH_DOLLA => Ok(Token::Absolute(pos)),
None => match ch { CH_DOT => Ok(Token::Dot(pos)),
CH_SINGLE_QUOTE => self.single_quote(pos, ch), CH_ASTERISK => Ok(Token::Asterisk(pos)),
CH_DOUBLE_QUOTE => self.double_quote(pos, ch), CH_LARRAY => Ok(Token::OpenArray(pos)),
CH_EQUAL => self.equal(pos, ch), CH_RARRAY => Ok(Token::CloseArray(pos)),
CH_GREATER => self.greater(pos, ch), CH_LPAREN => Ok(Token::OpenParenthesis(pos)),
CH_LITTLE => self.little(pos, ch), CH_RPAREN => Ok(Token::CloseParenthesis(pos)),
CH_AMPERSAND => self.and(pos, ch), CH_AT => Ok(Token::At(pos)),
CH_PIPE => self.or(pos, ch), CH_QUESTION => Ok(Token::Question(pos)),
CH_EXCLAMATION => self.not_equal(pos, ch), CH_COMMA => Ok(Token::Comma(pos)),
_ if ch.is_whitespace() => self.whitespace(pos, ch), CH_SEMICOLON => Ok(Token::Split(pos)),
_ => self.other(pos, ch), CH_SINGLE_QUOTE => self.single_quote(pos, ch),
}, CH_DOUBLE_QUOTE => self.double_quote(pos, ch),
CH_EQUAL => self.equal(pos, ch),
CH_GREATER => self.greater(pos, ch),
CH_LITTLE => self.little(pos, ch),
CH_AMPERSAND => self.and(pos, ch),
CH_PIPE => self.or(pos, ch),
CH_EXCLAMATION => self.not_equal(pos, ch),
_ if ch.is_whitespace() => self.whitespace(pos, ch),
_ => self.other(pos, ch),
} }
} }
@ -333,13 +368,6 @@ impl<'a> TokenReader<'a> {
} }
} }
pub fn peek_is(&self, simple_token: &str) -> bool {
match self.peek_token() {
Ok(t) => t.simple_eq(simple_token),
_ => false,
}
}
pub fn peek_token(&self) -> Result<&Token, TokenError> { pub fn peek_token(&self) -> Result<&Token, TokenError> {
match self.tokens.last() { match self.tokens.last() {
Some((_, t)) => { Some((_, t)) => {