Fix warnings
This commit is contained in:
parent
5c25f78c7f
commit
1014e35108
3 changed files with 17 additions and 17 deletions
1
saphyr/.gitignore
vendored
1
saphyr/.gitignore
vendored
|
@ -1,2 +1,3 @@
|
|||
target
|
||||
Cargo.lock
|
||||
*.swp
|
||||
|
|
|
@ -85,7 +85,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
|
||||
fn skip(&mut self) {
|
||||
self.token = None;
|
||||
self.peek();
|
||||
//self.peek();
|
||||
}
|
||||
fn pop_state(&mut self) {
|
||||
self.state = self.states.pop().unwrap()
|
||||
|
@ -117,7 +117,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
if ev == Event::StreamEnd {
|
||||
return Ok(());
|
||||
}
|
||||
self.load_document(&ev);
|
||||
try!(self.load_document(&ev));
|
||||
Ok(())
|
||||
}
|
||||
|
||||
|
@ -125,7 +125,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
assert_eq!(first_ev, &Event::DocumentStart);
|
||||
|
||||
let ev = try!(self.parse());
|
||||
let ev = try!(self.load_node(&ev));
|
||||
try!(self.load_node(&ev));
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
@ -251,7 +251,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
}
|
||||
|
||||
fn _explict_document_start(&mut self) -> ParseResult {
|
||||
let mut tok = try!(self.peek());
|
||||
let tok = try!(self.peek());
|
||||
if tok.1 != TokenType::DocumentStartToken {
|
||||
return Err(ScanError::new(tok.0, "did not find expected <document start>"));
|
||||
}
|
||||
|
@ -280,7 +280,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
}
|
||||
|
||||
fn parse_node(&mut self, block: bool, indentless_sequence: bool) -> ParseResult {
|
||||
let mut tok = try!(self.peek());
|
||||
let tok = try!(self.peek());
|
||||
match tok.1 {
|
||||
TokenType::AliasToken => unimplemented!(),
|
||||
TokenType::AnchorToken => unimplemented!(),
|
||||
|
@ -316,7 +316,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
fn block_mapping_key(&mut self, first: bool) -> ParseResult {
|
||||
// skip BlockMappingStartToken
|
||||
if first {
|
||||
let tok = try!(self.peek());
|
||||
let _ = try!(self.peek());
|
||||
//self.marks.push(tok.0);
|
||||
self.skip();
|
||||
}
|
||||
|
@ -379,7 +379,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
fn flow_sequence_entry(&mut self, first: bool) -> ParseResult {
|
||||
// skip FlowMappingStartToken
|
||||
if first {
|
||||
let tok = try!(self.peek());
|
||||
let _ = try!(self.peek());
|
||||
//self.marks.push(tok.0);
|
||||
self.skip();
|
||||
}
|
||||
|
@ -421,7 +421,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
|||
fn block_sequence_entry(&mut self, first: bool) -> ParseResult {
|
||||
// BLOCK-SEQUENCE-START
|
||||
if first {
|
||||
let tok = try!(self.peek());
|
||||
let _ = try!(self.peek());
|
||||
//self.marks.push(tok.0);
|
||||
self.skip();
|
||||
}
|
||||
|
|
|
@ -1,5 +1,4 @@
|
|||
use std::collections::VecDeque;
|
||||
use yaml::*;
|
||||
|
||||
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
||||
pub enum TEncoding {
|
||||
|
@ -149,7 +148,7 @@ pub type ScanResult = Result<(), ScanError>;
|
|||
impl<T: Iterator<Item=char>> Scanner<T> {
|
||||
/// Creates the YAML tokenizer.
|
||||
pub fn new(rdr: T) -> Scanner<T> {
|
||||
let mut p = Scanner {
|
||||
Scanner {
|
||||
rdr: rdr,
|
||||
buffer: VecDeque::new(),
|
||||
mark: Marker::new(0, 1, 0),
|
||||
|
@ -164,15 +163,14 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
|||
flow_level: 0,
|
||||
tokens_parsed: 0,
|
||||
token_available: false,
|
||||
};
|
||||
return p;
|
||||
}
|
||||
}
|
||||
|
||||
fn lookhead(&mut self, count: usize) {
|
||||
if self.buffer.len() >= count {
|
||||
return;
|
||||
}
|
||||
for i in 0..(count - self.buffer.len()) {
|
||||
for _ in 0..(count - self.buffer.len()) {
|
||||
self.buffer.push_back(self.rdr.next().unwrap_or('\0'));
|
||||
}
|
||||
}
|
||||
|
@ -193,6 +191,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
|||
fn ch_is(&self, c: char) -> bool {
|
||||
self.buffer[0] == c
|
||||
}
|
||||
#[allow(dead_code)]
|
||||
fn eof(&self) -> bool {
|
||||
self.ch_is('\0')
|
||||
}
|
||||
|
@ -250,7 +249,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
|||
self.lookhead(4);
|
||||
|
||||
if is_z(self.ch()) {
|
||||
self.fetch_stream_end();
|
||||
try!(self.fetch_stream_end());
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
|
@ -325,7 +324,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
|||
}
|
||||
|
||||
pub fn fetch_more_tokens(&mut self) -> ScanResult {
|
||||
let mut need_more = false;
|
||||
let mut need_more;
|
||||
loop {
|
||||
need_more = false;
|
||||
if self.tokens.is_empty() {
|
||||
|
@ -469,7 +468,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
|||
// - * only allowed in block
|
||||
unreachable!();
|
||||
}
|
||||
self.remove_simple_key();
|
||||
try!(self.remove_simple_key());
|
||||
self.allow_simple_key();
|
||||
|
||||
let start_mark = self.mark;
|
||||
|
@ -719,7 +718,7 @@ a4:
|
|||
".to_string();
|
||||
let p = Scanner::new(s.chars());
|
||||
for t in p {
|
||||
//println!("{:?}", t);
|
||||
// println!("{:?}", t);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue