Fix warnings
This commit is contained in:
parent
31a26b4a35
commit
760f935a87
3 changed files with 17 additions and 17 deletions
1
parser/.gitignore
vendored
1
parser/.gitignore
vendored
|
@ -1,2 +1,3 @@
|
||||||
target
|
target
|
||||||
Cargo.lock
|
Cargo.lock
|
||||||
|
*.swp
|
||||||
|
|
|
@ -85,7 +85,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
|
|
||||||
fn skip(&mut self) {
|
fn skip(&mut self) {
|
||||||
self.token = None;
|
self.token = None;
|
||||||
self.peek();
|
//self.peek();
|
||||||
}
|
}
|
||||||
fn pop_state(&mut self) {
|
fn pop_state(&mut self) {
|
||||||
self.state = self.states.pop().unwrap()
|
self.state = self.states.pop().unwrap()
|
||||||
|
@ -117,7 +117,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
if ev == Event::StreamEnd {
|
if ev == Event::StreamEnd {
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
self.load_document(&ev);
|
try!(self.load_document(&ev));
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -125,7 +125,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
assert_eq!(first_ev, &Event::DocumentStart);
|
assert_eq!(first_ev, &Event::DocumentStart);
|
||||||
|
|
||||||
let ev = try!(self.parse());
|
let ev = try!(self.parse());
|
||||||
let ev = try!(self.load_node(&ev));
|
try!(self.load_node(&ev));
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
@ -251,7 +251,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn _explict_document_start(&mut self) -> ParseResult {
|
fn _explict_document_start(&mut self) -> ParseResult {
|
||||||
let mut tok = try!(self.peek());
|
let tok = try!(self.peek());
|
||||||
if tok.1 != TokenType::DocumentStartToken {
|
if tok.1 != TokenType::DocumentStartToken {
|
||||||
return Err(ScanError::new(tok.0, "did not find expected <document start>"));
|
return Err(ScanError::new(tok.0, "did not find expected <document start>"));
|
||||||
}
|
}
|
||||||
|
@ -280,7 +280,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn parse_node(&mut self, block: bool, indentless_sequence: bool) -> ParseResult {
|
fn parse_node(&mut self, block: bool, indentless_sequence: bool) -> ParseResult {
|
||||||
let mut tok = try!(self.peek());
|
let tok = try!(self.peek());
|
||||||
match tok.1 {
|
match tok.1 {
|
||||||
TokenType::AliasToken => unimplemented!(),
|
TokenType::AliasToken => unimplemented!(),
|
||||||
TokenType::AnchorToken => unimplemented!(),
|
TokenType::AnchorToken => unimplemented!(),
|
||||||
|
@ -316,7 +316,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
fn block_mapping_key(&mut self, first: bool) -> ParseResult {
|
fn block_mapping_key(&mut self, first: bool) -> ParseResult {
|
||||||
// skip BlockMappingStartToken
|
// skip BlockMappingStartToken
|
||||||
if first {
|
if first {
|
||||||
let tok = try!(self.peek());
|
let _ = try!(self.peek());
|
||||||
//self.marks.push(tok.0);
|
//self.marks.push(tok.0);
|
||||||
self.skip();
|
self.skip();
|
||||||
}
|
}
|
||||||
|
@ -379,7 +379,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
fn flow_sequence_entry(&mut self, first: bool) -> ParseResult {
|
fn flow_sequence_entry(&mut self, first: bool) -> ParseResult {
|
||||||
// skip FlowMappingStartToken
|
// skip FlowMappingStartToken
|
||||||
if first {
|
if first {
|
||||||
let tok = try!(self.peek());
|
let _ = try!(self.peek());
|
||||||
//self.marks.push(tok.0);
|
//self.marks.push(tok.0);
|
||||||
self.skip();
|
self.skip();
|
||||||
}
|
}
|
||||||
|
@ -421,7 +421,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
fn block_sequence_entry(&mut self, first: bool) -> ParseResult {
|
fn block_sequence_entry(&mut self, first: bool) -> ParseResult {
|
||||||
// BLOCK-SEQUENCE-START
|
// BLOCK-SEQUENCE-START
|
||||||
if first {
|
if first {
|
||||||
let tok = try!(self.peek());
|
let _ = try!(self.peek());
|
||||||
//self.marks.push(tok.0);
|
//self.marks.push(tok.0);
|
||||||
self.skip();
|
self.skip();
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
use std::collections::VecDeque;
|
use std::collections::VecDeque;
|
||||||
use yaml::*;
|
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
||||||
pub enum TEncoding {
|
pub enum TEncoding {
|
||||||
|
@ -149,7 +148,7 @@ pub type ScanResult = Result<(), ScanError>;
|
||||||
impl<T: Iterator<Item=char>> Scanner<T> {
|
impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
/// Creates the YAML tokenizer.
|
/// Creates the YAML tokenizer.
|
||||||
pub fn new(rdr: T) -> Scanner<T> {
|
pub fn new(rdr: T) -> Scanner<T> {
|
||||||
let mut p = Scanner {
|
Scanner {
|
||||||
rdr: rdr,
|
rdr: rdr,
|
||||||
buffer: VecDeque::new(),
|
buffer: VecDeque::new(),
|
||||||
mark: Marker::new(0, 1, 0),
|
mark: Marker::new(0, 1, 0),
|
||||||
|
@ -164,15 +163,14 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
flow_level: 0,
|
flow_level: 0,
|
||||||
tokens_parsed: 0,
|
tokens_parsed: 0,
|
||||||
token_available: false,
|
token_available: false,
|
||||||
};
|
}
|
||||||
return p;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn lookhead(&mut self, count: usize) {
|
fn lookhead(&mut self, count: usize) {
|
||||||
if self.buffer.len() >= count {
|
if self.buffer.len() >= count {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
for i in 0..(count - self.buffer.len()) {
|
for _ in 0..(count - self.buffer.len()) {
|
||||||
self.buffer.push_back(self.rdr.next().unwrap_or('\0'));
|
self.buffer.push_back(self.rdr.next().unwrap_or('\0'));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -193,6 +191,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
fn ch_is(&self, c: char) -> bool {
|
fn ch_is(&self, c: char) -> bool {
|
||||||
self.buffer[0] == c
|
self.buffer[0] == c
|
||||||
}
|
}
|
||||||
|
#[allow(dead_code)]
|
||||||
fn eof(&self) -> bool {
|
fn eof(&self) -> bool {
|
||||||
self.ch_is('\0')
|
self.ch_is('\0')
|
||||||
}
|
}
|
||||||
|
@ -250,7 +249,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.lookhead(4);
|
self.lookhead(4);
|
||||||
|
|
||||||
if is_z(self.ch()) {
|
if is_z(self.ch()) {
|
||||||
self.fetch_stream_end();
|
try!(self.fetch_stream_end());
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -325,7 +324,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn fetch_more_tokens(&mut self) -> ScanResult {
|
pub fn fetch_more_tokens(&mut self) -> ScanResult {
|
||||||
let mut need_more = false;
|
let mut need_more;
|
||||||
loop {
|
loop {
|
||||||
need_more = false;
|
need_more = false;
|
||||||
if self.tokens.is_empty() {
|
if self.tokens.is_empty() {
|
||||||
|
@ -469,7 +468,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
// - * only allowed in block
|
// - * only allowed in block
|
||||||
unreachable!();
|
unreachable!();
|
||||||
}
|
}
|
||||||
self.remove_simple_key();
|
try!(self.remove_simple_key());
|
||||||
self.allow_simple_key();
|
self.allow_simple_key();
|
||||||
|
|
||||||
let start_mark = self.mark;
|
let start_mark = self.mark;
|
||||||
|
@ -719,7 +718,7 @@ a4:
|
||||||
".to_string();
|
".to_string();
|
||||||
let p = Scanner::new(s.chars());
|
let p = Scanner::new(s.chars());
|
||||||
for t in p {
|
for t in p {
|
||||||
//println!("{:?}", t);
|
// println!("{:?}", t);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue