Format with rustfmt 0.99.4
This commit is contained in:
parent
124d237be0
commit
5039af6862
8 changed files with 733 additions and 584 deletions
|
@ -17,14 +17,14 @@ fn dump_node(doc: &yaml::Yaml, indent: usize) {
|
||||||
for x in v {
|
for x in v {
|
||||||
dump_node(x, indent + 1);
|
dump_node(x, indent + 1);
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
yaml::Yaml::Hash(ref h) => {
|
yaml::Yaml::Hash(ref h) => {
|
||||||
for (k, v) in h {
|
for (k, v) in h {
|
||||||
print_indent(indent);
|
print_indent(indent);
|
||||||
println!("{:?}:", k);
|
println!("{:?}:", k);
|
||||||
dump_node(v, indent + 1);
|
dump_node(v, indent + 1);
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
print_indent(indent);
|
print_indent(indent);
|
||||||
println!("{:?}", doc);
|
println!("{:?}", doc);
|
||||||
|
|
|
@ -1,9 +1,8 @@
|
||||||
use std::fmt::{self, Display};
|
|
||||||
use std::convert::From;
|
use std::convert::From;
|
||||||
use std::error::Error;
|
use std::error::Error;
|
||||||
|
use std::fmt::{self, Display};
|
||||||
use yaml::{Hash, Yaml};
|
use yaml::{Hash, Yaml};
|
||||||
|
|
||||||
|
|
||||||
#[derive(Copy, Clone, Debug)]
|
#[derive(Copy, Clone, Debug)]
|
||||||
pub enum EmitError {
|
pub enum EmitError {
|
||||||
FmtError(fmt::Error),
|
FmtError(fmt::Error),
|
||||||
|
@ -91,7 +90,7 @@ fn escape_str(wr: &mut fmt::Write, v: &str) -> Result<(), fmt::Error> {
|
||||||
b'\x1e' => "\\u001e",
|
b'\x1e' => "\\u001e",
|
||||||
b'\x1f' => "\\u001f",
|
b'\x1f' => "\\u001f",
|
||||||
b'\x7f' => "\\u007f",
|
b'\x7f' => "\\u007f",
|
||||||
_ => { continue; }
|
_ => continue,
|
||||||
};
|
};
|
||||||
|
|
||||||
if start < i {
|
if start < i {
|
||||||
|
@ -118,7 +117,7 @@ impl<'a> YamlEmitter<'a> {
|
||||||
best_indent: 2,
|
best_indent: 2,
|
||||||
compact: true,
|
compact: true,
|
||||||
|
|
||||||
level: -1
|
level: -1,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -147,7 +146,9 @@ impl<'a> YamlEmitter<'a> {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn write_indent(&mut self) -> EmitResult {
|
fn write_indent(&mut self) -> EmitResult {
|
||||||
if self.level <= 0 { return Ok(()); }
|
if self.level <= 0 {
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
for _ in 0..self.level {
|
for _ in 0..self.level {
|
||||||
for _ in 0..self.best_indent {
|
for _ in 0..self.best_indent {
|
||||||
try!(write!(self.writer, " "));
|
try!(write!(self.writer, " "));
|
||||||
|
@ -163,12 +164,11 @@ impl<'a> YamlEmitter<'a> {
|
||||||
Yaml::String(ref v) => {
|
Yaml::String(ref v) => {
|
||||||
if need_quotes(v) {
|
if need_quotes(v) {
|
||||||
try!(escape_str(self.writer, v));
|
try!(escape_str(self.writer, v));
|
||||||
}
|
} else {
|
||||||
else {
|
|
||||||
try!(write!(self.writer, "{}", v));
|
try!(write!(self.writer, "{}", v));
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
}
|
||||||
Yaml::Boolean(v) => {
|
Yaml::Boolean(v) => {
|
||||||
if v {
|
if v {
|
||||||
try!(self.writer.write_str("true"));
|
try!(self.writer.write_str("true"));
|
||||||
|
@ -176,21 +176,21 @@ impl<'a> YamlEmitter<'a> {
|
||||||
try!(self.writer.write_str("false"));
|
try!(self.writer.write_str("false"));
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
}
|
||||||
Yaml::Integer(v) => {
|
Yaml::Integer(v) => {
|
||||||
try!(write!(self.writer, "{}", v));
|
try!(write!(self.writer, "{}", v));
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
}
|
||||||
Yaml::Real(ref v) => {
|
Yaml::Real(ref v) => {
|
||||||
try!(write!(self.writer, "{}", v));
|
try!(write!(self.writer, "{}", v));
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
}
|
||||||
Yaml::Null | Yaml::BadValue => {
|
Yaml::Null | Yaml::BadValue => {
|
||||||
try!(write!(self.writer, "~"));
|
try!(write!(self.writer, "~"));
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
}
|
||||||
// XXX(chenyh) Alias
|
// XXX(chenyh) Alias
|
||||||
_ => { Ok(()) }
|
_ => Ok(()),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -260,7 +260,7 @@ impl<'a> YamlEmitter<'a> {
|
||||||
self.level -= 1;
|
self.level -= 1;
|
||||||
}
|
}
|
||||||
self.emit_array(v)
|
self.emit_array(v)
|
||||||
},
|
}
|
||||||
Yaml::Hash(ref h) => {
|
Yaml::Hash(ref h) => {
|
||||||
if (inline && self.compact) || h.is_empty() {
|
if (inline && self.compact) || h.is_empty() {
|
||||||
try!(write!(self.writer, " "));
|
try!(write!(self.writer, " "));
|
||||||
|
@ -271,7 +271,7 @@ impl<'a> YamlEmitter<'a> {
|
||||||
self.level -= 1;
|
self.level -= 1;
|
||||||
}
|
}
|
||||||
self.emit_hash(h)
|
self.emit_hash(h)
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
try!(write!(self.writer, " "));
|
try!(write!(self.writer, " "));
|
||||||
self.emit_node(val)
|
self.emit_node(val)
|
||||||
|
@ -296,34 +296,45 @@ impl<'a> YamlEmitter<'a> {
|
||||||
/// * When the string looks like a date (e.g. 2014-12-31) (otherwise it would be automatically converted into a Unix timestamp).
|
/// * When the string looks like a date (e.g. 2014-12-31) (otherwise it would be automatically converted into a Unix timestamp).
|
||||||
fn need_quotes(string: &str) -> bool {
|
fn need_quotes(string: &str) -> bool {
|
||||||
fn need_quotes_spaces(string: &str) -> bool {
|
fn need_quotes_spaces(string: &str) -> bool {
|
||||||
string.starts_with(' ')
|
string.starts_with(' ') || string.ends_with(' ')
|
||||||
|| string.ends_with(' ')
|
|
||||||
}
|
}
|
||||||
|
|
||||||
string == ""
|
string == ""
|
||||||
|| need_quotes_spaces(string)
|
|| need_quotes_spaces(string)
|
||||||
|| string.starts_with(|character: char| {
|
|| string.starts_with(|character: char| match character {
|
||||||
match character {
|
|
||||||
':' | '&' | '*' | '?' | '|' | '-' | '<' | '>' | '=' | '!' | '%' | '@' => true,
|
':' | '&' | '*' | '?' | '|' | '-' | '<' | '>' | '=' | '!' | '%' | '@' => true,
|
||||||
_ => false,
|
_ => false,
|
||||||
}
|
|
||||||
})
|
})
|
||||||
|| string.contains(|character: char| {
|
|| string.contains(|character: char| match character {
|
||||||
match character {
|
'{'
|
||||||
'{' | '}' | '[' | ']' | ',' | '#' | '`' | '\"' | '\'' | '\\' | '\0' ... '\x06' | '\t' | '\n' | '\r' | '\x0e' ... '\x1a' | '\x1c' ... '\x1f' => true,
|
| '}'
|
||||||
|
| '['
|
||||||
|
| ']'
|
||||||
|
| ','
|
||||||
|
| '#'
|
||||||
|
| '`'
|
||||||
|
| '\"'
|
||||||
|
| '\''
|
||||||
|
| '\\'
|
||||||
|
| '\0'...'\x06'
|
||||||
|
| '\t'
|
||||||
|
| '\n'
|
||||||
|
| '\r'
|
||||||
|
| '\x0e'...'\x1a'
|
||||||
|
| '\x1c'...'\x1f' => true,
|
||||||
_ => false,
|
_ => false,
|
||||||
}
|
|
||||||
})
|
})
|
||||||
|| [// http://yaml.org/type/bool.html
|
|| [
|
||||||
|
// http://yaml.org/type/bool.html
|
||||||
// Note: 'y', 'Y', 'n', 'N', is not quoted deliberately, as in libyaml. PyYAML also parse
|
// Note: 'y', 'Y', 'n', 'N', is not quoted deliberately, as in libyaml. PyYAML also parse
|
||||||
// them as string, not booleans, although it is volating the YAML 1.1 specification.
|
// them as string, not booleans, although it is volating the YAML 1.1 specification.
|
||||||
// See https://github.com/dtolnay/serde-yaml/pull/83#discussion_r152628088.
|
// See https://github.com/dtolnay/serde-yaml/pull/83#discussion_r152628088.
|
||||||
"yes","Yes","YES","no","No","NO",
|
"yes", "Yes", "YES", "no", "No", "NO", "True", "TRUE", "true", "False", "FALSE",
|
||||||
"True", "TRUE", "true", "False", "FALSE", "false",
|
"false", "on", "On", "ON", "off", "Off", "OFF",
|
||||||
"on","On","ON","off","Off","OFF",
|
|
||||||
// http://yaml.org/type/null.html
|
// http://yaml.org/type/null.html
|
||||||
"null","Null","NULL", "~"
|
"null", "Null", "NULL", "~",
|
||||||
].contains(&string)
|
]
|
||||||
|
.contains(&string)
|
||||||
|| string.starts_with('.')
|
|| string.starts_with('.')
|
||||||
|| string.parse::<i64>().is_ok()
|
|| string.parse::<i64>().is_ok()
|
||||||
|| string.parse::<f64>().is_ok()
|
|| string.parse::<f64>().is_ok()
|
||||||
|
@ -349,7 +360,6 @@ a4:
|
||||||
- 2
|
- 2
|
||||||
";
|
";
|
||||||
|
|
||||||
|
|
||||||
let docs = YamlLoader::load_from_str(&s).unwrap();
|
let docs = YamlLoader::load_from_str(&s).unwrap();
|
||||||
let doc = &docs[0];
|
let doc = &docs[0];
|
||||||
let mut writer = String::new();
|
let mut writer = String::new();
|
||||||
|
@ -361,7 +371,7 @@ a4:
|
||||||
println!("emitted:\n{}", writer);
|
println!("emitted:\n{}", writer);
|
||||||
let docs_new = match YamlLoader::load_from_str(&writer) {
|
let docs_new = match YamlLoader::load_from_str(&writer) {
|
||||||
Ok(y) => y,
|
Ok(y) => y,
|
||||||
Err(e) => panic!(format!("{}", e))
|
Err(e) => panic!(format!("{}", e)),
|
||||||
};
|
};
|
||||||
let doc_new = &docs_new[0];
|
let doc_new = &docs_new[0];
|
||||||
|
|
||||||
|
@ -398,7 +408,7 @@ products:
|
||||||
}
|
}
|
||||||
let docs_new = match YamlLoader::load_from_str(&writer) {
|
let docs_new = match YamlLoader::load_from_str(&writer) {
|
||||||
Ok(y) => y,
|
Ok(y) => y,
|
||||||
Err(e) => panic!(format!("{}", e))
|
Err(e) => panic!(format!("{}", e)),
|
||||||
};
|
};
|
||||||
let doc_new = &docs_new[0];
|
let doc_new = &docs_new[0];
|
||||||
assert_eq!(doc, doc_new);
|
assert_eq!(doc, doc_new);
|
||||||
|
@ -506,7 +516,11 @@ bool1: false"#;
|
||||||
emitter.dump(doc).unwrap();
|
emitter.dump(doc).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
assert_eq!(expected, writer, "expected:\n{}\nactual:\n{}\n", expected, writer);
|
assert_eq!(
|
||||||
|
expected, writer,
|
||||||
|
"expected:\n{}\nactual:\n{}\n",
|
||||||
|
expected, writer
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
|
@ -520,7 +534,8 @@ bool1: false"#;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn test_empty_and_nested_flag(compact: bool) {
|
fn test_empty_and_nested_flag(compact: bool) {
|
||||||
let s = if compact { r#"---
|
let s = if compact {
|
||||||
|
r#"---
|
||||||
a:
|
a:
|
||||||
b:
|
b:
|
||||||
c: hello
|
c: hello
|
||||||
|
@ -528,7 +543,9 @@ a:
|
||||||
e:
|
e:
|
||||||
- f
|
- f
|
||||||
- g
|
- g
|
||||||
- h: []"# } else { r#"---
|
- h: []"#
|
||||||
|
} else {
|
||||||
|
r#"---
|
||||||
a:
|
a:
|
||||||
b:
|
b:
|
||||||
c: hello
|
c: hello
|
||||||
|
@ -537,7 +554,8 @@ e:
|
||||||
- f
|
- f
|
||||||
- g
|
- g
|
||||||
-
|
-
|
||||||
h: []"# };
|
h: []"#
|
||||||
|
};
|
||||||
|
|
||||||
let docs = YamlLoader::load_from_str(&s).unwrap();
|
let docs = YamlLoader::load_from_str(&s).unwrap();
|
||||||
let doc = &docs[0];
|
let doc = &docs[0];
|
||||||
|
|
|
@ -41,16 +41,16 @@
|
||||||
|
|
||||||
extern crate linked_hash_map;
|
extern crate linked_hash_map;
|
||||||
|
|
||||||
pub mod yaml;
|
|
||||||
pub mod scanner;
|
|
||||||
pub mod parser;
|
|
||||||
pub mod emitter;
|
pub mod emitter;
|
||||||
|
pub mod parser;
|
||||||
|
pub mod scanner;
|
||||||
|
pub mod yaml;
|
||||||
|
|
||||||
// reexport key APIs
|
// reexport key APIs
|
||||||
pub use scanner::ScanError;
|
pub use emitter::{EmitError, YamlEmitter};
|
||||||
pub use parser::Event;
|
pub use parser::Event;
|
||||||
|
pub use scanner::ScanError;
|
||||||
pub use yaml::{Yaml, YamlLoader};
|
pub use yaml::{Yaml, YamlLoader};
|
||||||
pub use emitter::{YamlEmitter, EmitError};
|
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
|
@ -58,8 +58,7 @@ mod tests {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_api() {
|
fn test_api() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
# from yaml-cpp example
|
# from yaml-cpp example
|
||||||
- name: Ogre
|
- name: Ogre
|
||||||
position: [0, 5, 0]
|
position: [0, 5, 0]
|
||||||
|
@ -104,8 +103,7 @@ mod tests {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_fail() {
|
fn test_fail() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
# syntax error
|
# syntax error
|
||||||
scalar
|
scalar
|
||||||
key: [1, 2]]
|
key: [1, 2]]
|
||||||
|
|
|
@ -26,7 +26,7 @@ enum State {
|
||||||
FlowMappingKey,
|
FlowMappingKey,
|
||||||
FlowMappingValue,
|
FlowMappingValue,
|
||||||
FlowMappingEmptyValue,
|
FlowMappingEmptyValue,
|
||||||
End
|
End,
|
||||||
}
|
}
|
||||||
|
|
||||||
/// `Event` is used with the low-level event base parsing API,
|
/// `Event` is used with the low-level event base parsing API,
|
||||||
|
@ -48,7 +48,7 @@ pub enum Event {
|
||||||
SequenceEnd,
|
SequenceEnd,
|
||||||
/// Anchor ID
|
/// Anchor ID
|
||||||
MappingStart(usize),
|
MappingStart(usize),
|
||||||
MappingEnd
|
MappingEnd,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Event {
|
impl Event {
|
||||||
|
@ -74,12 +74,10 @@ pub struct Parser<T> {
|
||||||
anchor_id: usize,
|
anchor_id: usize,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
pub trait EventReceiver {
|
pub trait EventReceiver {
|
||||||
fn on_event(&mut self, ev: Event);
|
fn on_event(&mut self, ev: Event);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
pub trait MarkedEventReceiver {
|
pub trait MarkedEventReceiver {
|
||||||
fn on_event(&mut self, ev: Event, _mark: Marker);
|
fn on_event(&mut self, ev: Event, _mark: Marker);
|
||||||
}
|
}
|
||||||
|
@ -121,9 +119,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
pub fn next(&mut self) -> ParseResult {
|
pub fn next(&mut self) -> ParseResult {
|
||||||
match self.current {
|
match self.current {
|
||||||
None => self.parse(),
|
None => self.parse(),
|
||||||
Some(_) => {
|
Some(_) => Ok(self.current.take().unwrap()),
|
||||||
Ok(self.current.take().unwrap())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -132,28 +128,28 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
None => {
|
None => {
|
||||||
self.token = Some(try!(self.scan_next_token()));
|
self.token = Some(try!(self.scan_next_token()));
|
||||||
Ok(self.token.as_ref().unwrap())
|
Ok(self.token.as_ref().unwrap())
|
||||||
},
|
}
|
||||||
Some(ref tok) => Ok(tok)
|
Some(ref tok) => Ok(tok),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn scan_next_token(&mut self) -> Result<Token, ScanError> {
|
fn scan_next_token(&mut self) -> Result<Token, ScanError> {
|
||||||
let token = self.scanner.next();
|
let token = self.scanner.next();
|
||||||
match token {
|
match token {
|
||||||
None =>
|
None => match self.scanner.get_error() {
|
||||||
match self.scanner.get_error() {
|
|
||||||
None => Err(ScanError::new(self.scanner.mark(), "unexpected eof")),
|
None => Err(ScanError::new(self.scanner.mark(), "unexpected eof")),
|
||||||
Some(e) => Err(e),
|
Some(e) => Err(e),
|
||||||
},
|
},
|
||||||
Some(tok) => Ok(tok)
|
Some(tok) => Ok(tok),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn fetch_token(&mut self) -> Token {
|
fn fetch_token(&mut self) -> Token {
|
||||||
self.token.take().expect("fetch_token needs to be preceded by peek_token")
|
self.token
|
||||||
|
.take()
|
||||||
|
.expect("fetch_token needs to be preceded by peek_token")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
fn skip(&mut self) {
|
fn skip(&mut self) {
|
||||||
self.token = None;
|
self.token = None;
|
||||||
//self.peek_token();
|
//self.peek_token();
|
||||||
|
@ -174,8 +170,11 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
Ok((ev, mark))
|
Ok((ev, mark))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn load<R: MarkedEventReceiver>(&mut self, recv: &mut R, multi: bool)
|
pub fn load<R: MarkedEventReceiver>(
|
||||||
-> Result<(), ScanError> {
|
&mut self,
|
||||||
|
recv: &mut R,
|
||||||
|
multi: bool,
|
||||||
|
) -> Result<(), ScanError> {
|
||||||
if !self.scanner.stream_started() {
|
if !self.scanner.stream_started() {
|
||||||
let (ev, mark) = try!(self.next());
|
let (ev, mark) = try!(self.next());
|
||||||
assert_eq!(ev, Event::StreamStart);
|
assert_eq!(ev, Event::StreamStart);
|
||||||
|
@ -203,8 +202,12 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn load_document<R: MarkedEventReceiver>(&mut self, first_ev: Event, mark: Marker, recv: &mut R)
|
fn load_document<R: MarkedEventReceiver>(
|
||||||
-> Result<(), ScanError> {
|
&mut self,
|
||||||
|
first_ev: Event,
|
||||||
|
mark: Marker,
|
||||||
|
recv: &mut R,
|
||||||
|
) -> Result<(), ScanError> {
|
||||||
assert_eq!(first_ev, Event::DocumentStart);
|
assert_eq!(first_ev, Event::DocumentStart);
|
||||||
recv.on_event(first_ev, mark);
|
recv.on_event(first_ev, mark);
|
||||||
|
|
||||||
|
@ -219,28 +222,33 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn load_node<R: MarkedEventReceiver>(&mut self, first_ev: Event, mark: Marker, recv: &mut R)
|
fn load_node<R: MarkedEventReceiver>(
|
||||||
-> Result<(), ScanError> {
|
&mut self,
|
||||||
|
first_ev: Event,
|
||||||
|
mark: Marker,
|
||||||
|
recv: &mut R,
|
||||||
|
) -> Result<(), ScanError> {
|
||||||
match first_ev {
|
match first_ev {
|
||||||
Event::Alias(..) | Event::Scalar(..) => {
|
Event::Alias(..) | Event::Scalar(..) => {
|
||||||
recv.on_event(first_ev, mark);
|
recv.on_event(first_ev, mark);
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
}
|
||||||
Event::SequenceStart(_) => {
|
Event::SequenceStart(_) => {
|
||||||
recv.on_event(first_ev, mark);
|
recv.on_event(first_ev, mark);
|
||||||
self.load_sequence(recv)
|
self.load_sequence(recv)
|
||||||
},
|
}
|
||||||
Event::MappingStart(_) => {
|
Event::MappingStart(_) => {
|
||||||
recv.on_event(first_ev, mark);
|
recv.on_event(first_ev, mark);
|
||||||
self.load_mapping(recv)
|
self.load_mapping(recv)
|
||||||
},
|
}
|
||||||
_ => { println!("UNREACHABLE EVENT: {:?}", first_ev);
|
_ => {
|
||||||
unreachable!(); }
|
println!("UNREACHABLE EVENT: {:?}", first_ev);
|
||||||
|
unreachable!();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn load_mapping<R: MarkedEventReceiver>(&mut self, recv: &mut R)
|
fn load_mapping<R: MarkedEventReceiver>(&mut self, recv: &mut R) -> Result<(), ScanError> {
|
||||||
-> Result<(), ScanError> {
|
|
||||||
let (mut key_ev, mut key_mark) = try!(self.next());
|
let (mut key_ev, mut key_mark) = try!(self.next());
|
||||||
while key_ev != Event::MappingEnd {
|
while key_ev != Event::MappingEnd {
|
||||||
// key
|
// key
|
||||||
|
@ -254,14 +262,12 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
let (ev, mark) = try!(self.next());
|
let (ev, mark) = try!(self.next());
|
||||||
key_ev = ev;
|
key_ev = ev;
|
||||||
key_mark = mark;
|
key_mark = mark;
|
||||||
|
|
||||||
}
|
}
|
||||||
recv.on_event(key_ev, key_mark);
|
recv.on_event(key_ev, key_mark);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn load_sequence<R: MarkedEventReceiver>(&mut self, recv: &mut R)
|
fn load_sequence<R: MarkedEventReceiver>(&mut self, recv: &mut R) -> Result<(), ScanError> {
|
||||||
-> Result<(), ScanError> {
|
|
||||||
let (mut ev, mut mark) = try!(self.next());
|
let (mut ev, mut mark) = try!(self.next());
|
||||||
while ev != Event::SequenceEnd {
|
while ev != Event::SequenceEnd {
|
||||||
try!(self.load_node(ev, mark, recv));
|
try!(self.load_node(ev, mark, recv));
|
||||||
|
@ -289,7 +295,6 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
State::BlockNode => self.parse_node(true, false),
|
State::BlockNode => self.parse_node(true, false),
|
||||||
// State::BlockNodeOrIndentlessSequence => self.parse_node(true, true),
|
// State::BlockNodeOrIndentlessSequence => self.parse_node(true, true),
|
||||||
// State::FlowNode => self.parse_node(false, false),
|
// State::FlowNode => self.parse_node(false, false),
|
||||||
|
|
||||||
State::BlockMappingFirstKey => self.block_mapping_key(true),
|
State::BlockMappingFirstKey => self.block_mapping_key(true),
|
||||||
State::BlockMappingKey => self.block_mapping_key(false),
|
State::BlockMappingKey => self.block_mapping_key(false),
|
||||||
State::BlockMappingValue => self.block_mapping_value(),
|
State::BlockMappingValue => self.block_mapping_value(),
|
||||||
|
@ -322,9 +327,8 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.state = State::ImplicitDocumentStart;
|
self.state = State::ImplicitDocumentStart;
|
||||||
self.skip();
|
self.skip();
|
||||||
Ok((Event::StreamStart, mark))
|
Ok((Event::StreamStart, mark))
|
||||||
},
|
}
|
||||||
Token(mark, _) => Err(ScanError::new(mark,
|
Token(mark, _) => Err(ScanError::new(mark, "did not find expected <stream-start>")),
|
||||||
"did not find expected <stream-start>")),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -340,19 +344,19 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.state = State::End;
|
self.state = State::End;
|
||||||
self.skip();
|
self.skip();
|
||||||
Ok((Event::StreamEnd, mark))
|
Ok((Event::StreamEnd, mark))
|
||||||
},
|
}
|
||||||
Token(_, TokenType::VersionDirective(..))
|
Token(_, TokenType::VersionDirective(..))
|
||||||
| Token(_, TokenType::TagDirective(..))
|
| Token(_, TokenType::TagDirective(..))
|
||||||
| Token(_, TokenType::DocumentStart) => {
|
| Token(_, TokenType::DocumentStart) => {
|
||||||
// explicit document
|
// explicit document
|
||||||
self._explict_document_start()
|
self._explict_document_start()
|
||||||
},
|
}
|
||||||
Token(mark, _) if implicit => {
|
Token(mark, _) if implicit => {
|
||||||
try!(self.parser_process_directives());
|
try!(self.parser_process_directives());
|
||||||
self.push_state(State::DocumentEnd);
|
self.push_state(State::DocumentEnd);
|
||||||
self.state = State::BlockNode;
|
self.state = State::BlockNode;
|
||||||
Ok((Event::DocumentStart, mark))
|
Ok((Event::DocumentStart, mark))
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
// explicit document
|
// explicit document
|
||||||
self._explict_document_start()
|
self._explict_document_start()
|
||||||
|
@ -369,11 +373,11 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
// return Err(ScanError::new(tok.0,
|
// return Err(ScanError::new(tok.0,
|
||||||
// "found incompatible YAML document"));
|
// "found incompatible YAML document"));
|
||||||
//}
|
//}
|
||||||
},
|
}
|
||||||
TokenType::TagDirective(..) => {
|
TokenType::TagDirective(..) => {
|
||||||
// TODO add tag directive
|
// TODO add tag directive
|
||||||
},
|
}
|
||||||
_ => break
|
_ => break,
|
||||||
}
|
}
|
||||||
self.skip();
|
self.skip();
|
||||||
}
|
}
|
||||||
|
@ -390,7 +394,10 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.skip();
|
self.skip();
|
||||||
Ok((Event::DocumentStart, mark))
|
Ok((Event::DocumentStart, mark))
|
||||||
}
|
}
|
||||||
Token(mark, _) => Err(ScanError::new(mark, "did not find expected <document start>"))
|
Token(mark, _) => Err(ScanError::new(
|
||||||
|
mark,
|
||||||
|
"did not find expected <document start>",
|
||||||
|
)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -404,10 +411,8 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
// empty scalar
|
// empty scalar
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
},
|
|
||||||
_ => {
|
|
||||||
self.parse_node(true, false)
|
|
||||||
}
|
}
|
||||||
|
_ => self.parse_node(true, false),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -418,8 +423,8 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.skip();
|
self.skip();
|
||||||
_implicit = false;
|
_implicit = false;
|
||||||
mark
|
mark
|
||||||
},
|
}
|
||||||
Token(mark, _) => mark
|
Token(mark, _) => mark,
|
||||||
};
|
};
|
||||||
|
|
||||||
// TODO tag handling
|
// TODO tag handling
|
||||||
|
@ -447,13 +452,18 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
if let Token(mark, TokenType::Alias(name)) = self.fetch_token() {
|
if let Token(mark, TokenType::Alias(name)) = self.fetch_token() {
|
||||||
match self.anchors.get(&name) {
|
match self.anchors.get(&name) {
|
||||||
None => return Err(ScanError::new(mark, "while parsing node, found unknown anchor")),
|
None => {
|
||||||
Some(id) => return Ok((Event::Alias(*id), mark))
|
return Err(ScanError::new(
|
||||||
|
mark,
|
||||||
|
"while parsing node, found unknown anchor",
|
||||||
|
))
|
||||||
|
}
|
||||||
|
Some(id) => return Ok((Event::Alias(*id), mark)),
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
unreachable!()
|
unreachable!()
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
Token(_, TokenType::Anchor(_)) => {
|
Token(_, TokenType::Anchor(_)) => {
|
||||||
if let Token(mark, TokenType::Anchor(name)) = self.fetch_token() {
|
if let Token(mark, TokenType::Anchor(name)) = self.fetch_token() {
|
||||||
anchor_id = try!(self.register_anchor(name, &mark));
|
anchor_id = try!(self.register_anchor(name, &mark));
|
||||||
|
@ -467,7 +477,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
} else {
|
} else {
|
||||||
unreachable!()
|
unreachable!()
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
Token(_, TokenType::Tag(..)) => {
|
Token(_, TokenType::Tag(..)) => {
|
||||||
if let tg @ TokenType::Tag(..) = self.fetch_token().1 {
|
if let tg @ TokenType::Tag(..) = self.fetch_token().1 {
|
||||||
tag = Some(tg);
|
tag = Some(tg);
|
||||||
|
@ -481,14 +491,14 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
} else {
|
} else {
|
||||||
unreachable!()
|
unreachable!()
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
_ => {}
|
_ => {}
|
||||||
}
|
}
|
||||||
match *try!(self.peek_token()) {
|
match *try!(self.peek_token()) {
|
||||||
Token(mark, TokenType::BlockEntry) if indentless_sequence => {
|
Token(mark, TokenType::BlockEntry) if indentless_sequence => {
|
||||||
self.state = State::IndentlessSequenceEntry;
|
self.state = State::IndentlessSequenceEntry;
|
||||||
Ok((Event::SequenceStart(anchor_id), mark))
|
Ok((Event::SequenceStart(anchor_id), mark))
|
||||||
},
|
}
|
||||||
Token(_, TokenType::Scalar(..)) => {
|
Token(_, TokenType::Scalar(..)) => {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
if let Token(mark, TokenType::Scalar(style, v)) = self.fetch_token() {
|
if let Token(mark, TokenType::Scalar(style, v)) = self.fetch_token() {
|
||||||
|
@ -496,29 +506,32 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
} else {
|
} else {
|
||||||
unreachable!()
|
unreachable!()
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
Token(mark, TokenType::FlowSequenceStart) => {
|
Token(mark, TokenType::FlowSequenceStart) => {
|
||||||
self.state = State::FlowSequenceFirstEntry;
|
self.state = State::FlowSequenceFirstEntry;
|
||||||
Ok((Event::SequenceStart(anchor_id), mark))
|
Ok((Event::SequenceStart(anchor_id), mark))
|
||||||
},
|
}
|
||||||
Token(mark, TokenType::FlowMappingStart) => {
|
Token(mark, TokenType::FlowMappingStart) => {
|
||||||
self.state = State::FlowMappingFirstKey;
|
self.state = State::FlowMappingFirstKey;
|
||||||
Ok((Event::MappingStart(anchor_id), mark))
|
Ok((Event::MappingStart(anchor_id), mark))
|
||||||
},
|
}
|
||||||
Token(mark, TokenType::BlockSequenceStart) if block => {
|
Token(mark, TokenType::BlockSequenceStart) if block => {
|
||||||
self.state = State::BlockSequenceFirstEntry;
|
self.state = State::BlockSequenceFirstEntry;
|
||||||
Ok((Event::SequenceStart(anchor_id), mark))
|
Ok((Event::SequenceStart(anchor_id), mark))
|
||||||
},
|
}
|
||||||
Token(mark, TokenType::BlockMappingStart) if block => {
|
Token(mark, TokenType::BlockMappingStart) if block => {
|
||||||
self.state = State::BlockMappingFirstKey;
|
self.state = State::BlockMappingFirstKey;
|
||||||
Ok((Event::MappingStart(anchor_id), mark))
|
Ok((Event::MappingStart(anchor_id), mark))
|
||||||
},
|
}
|
||||||
// ex 7.2, an empty scalar can follow a secondary tag
|
// ex 7.2, an empty scalar can follow a secondary tag
|
||||||
Token(mark, _) if tag.is_some() || anchor_id > 0 => {
|
Token(mark, _) if tag.is_some() || anchor_id > 0 => {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
Ok((Event::empty_scalar_with_anchor(anchor_id, tag), mark))
|
Ok((Event::empty_scalar_with_anchor(anchor_id, tag), mark))
|
||||||
},
|
}
|
||||||
Token(mark, _) => { Err(ScanError::new(mark, "while parsing a node, did not find expected node content")) }
|
Token(mark, _) => Err(ScanError::new(
|
||||||
|
mark,
|
||||||
|
"while parsing a node, did not find expected node content",
|
||||||
|
)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -545,20 +558,21 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.parse_node(true, true)
|
self.parse_node(true, true)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
// XXX(chenyh): libyaml failed to parse spec 1.2, ex8.18
|
// XXX(chenyh): libyaml failed to parse spec 1.2, ex8.18
|
||||||
Token(mark, TokenType::Value) => {
|
Token(mark, TokenType::Value) => {
|
||||||
self.state = State::BlockMappingValue;
|
self.state = State::BlockMappingValue;
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
},
|
}
|
||||||
Token(mark, TokenType::BlockEnd) => {
|
Token(mark, TokenType::BlockEnd) => {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
self.skip();
|
self.skip();
|
||||||
Ok((Event::MappingEnd, mark))
|
Ok((Event::MappingEnd, mark))
|
||||||
},
|
|
||||||
Token(mark, _) => {
|
|
||||||
Err(ScanError::new(mark, "while parsing a block mapping, did not find expected key"))
|
|
||||||
}
|
}
|
||||||
|
Token(mark, _) => Err(ScanError::new(
|
||||||
|
mark,
|
||||||
|
"while parsing a block mapping, did not find expected key",
|
||||||
|
)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -573,13 +587,13 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.state = State::BlockMappingKey;
|
self.state = State::BlockMappingKey;
|
||||||
// empty scalar
|
// empty scalar
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::BlockMappingKey);
|
self.push_state(State::BlockMappingKey);
|
||||||
self.parse_node(true, true)
|
self.parse_node(true, true)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
Token(mark, _) => {
|
Token(mark, _) => {
|
||||||
self.state = State::BlockMappingKey;
|
self.state = State::BlockMappingKey;
|
||||||
// empty scalar
|
// empty scalar
|
||||||
|
@ -593,7 +607,8 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
let _ = try!(self.peek_token());
|
let _ = try!(self.peek_token());
|
||||||
self.skip();
|
self.skip();
|
||||||
}
|
}
|
||||||
let marker: Marker = {
|
let marker: Marker =
|
||||||
|
{
|
||||||
match *try!(self.peek_token()) {
|
match *try!(self.peek_token()) {
|
||||||
Token(mark, TokenType::FlowMappingEnd) => mark,
|
Token(mark, TokenType::FlowMappingEnd) => mark,
|
||||||
Token(mark, _) => {
|
Token(mark, _) => {
|
||||||
|
@ -614,17 +629,17 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
| Token(mark, TokenType::FlowMappingEnd) => {
|
| Token(mark, TokenType::FlowMappingEnd) => {
|
||||||
self.state = State::FlowMappingValue;
|
self.state = State::FlowMappingValue;
|
||||||
return Ok((Event::empty_scalar(), mark));
|
return Ok((Event::empty_scalar(), mark));
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::FlowMappingValue);
|
self.push_state(State::FlowMappingValue);
|
||||||
return self.parse_node(false, false);
|
return self.parse_node(false, false);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
Token(marker, TokenType::Value) => {
|
Token(marker, TokenType::Value) => {
|
||||||
self.state = State::FlowMappingValue;
|
self.state = State::FlowMappingValue;
|
||||||
return Ok((Event::empty_scalar(), marker));
|
return Ok((Event::empty_scalar(), marker));
|
||||||
},
|
}
|
||||||
Token(_, TokenType::FlowMappingEnd) => (),
|
Token(_, TokenType::FlowMappingEnd) => (),
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::FlowMappingEmptyValue);
|
self.push_state(State::FlowMappingEmptyValue);
|
||||||
|
@ -653,16 +668,15 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
Token(marker, TokenType::Value) => {
|
Token(marker, TokenType::Value) => {
|
||||||
self.skip();
|
self.skip();
|
||||||
match try!(self.peek_token()).1 {
|
match try!(self.peek_token()).1 {
|
||||||
TokenType::FlowEntry
|
TokenType::FlowEntry | TokenType::FlowMappingEnd => {}
|
||||||
| TokenType::FlowMappingEnd => { },
|
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::FlowMappingKey);
|
self.push_state(State::FlowMappingKey);
|
||||||
return self.parse_node(false, false);
|
return self.parse_node(false, false);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
marker
|
marker
|
||||||
},
|
}
|
||||||
Token(marker, _) => marker
|
Token(marker, _) => marker,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -683,13 +697,15 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
self.skip();
|
self.skip();
|
||||||
return Ok((Event::SequenceEnd, mark));
|
return Ok((Event::SequenceEnd, mark));
|
||||||
},
|
}
|
||||||
Token(_, TokenType::FlowEntry) if !first => {
|
Token(_, TokenType::FlowEntry) if !first => {
|
||||||
self.skip();
|
self.skip();
|
||||||
},
|
}
|
||||||
Token(mark, _) if !first => {
|
Token(mark, _) if !first => {
|
||||||
return Err(ScanError::new(mark,
|
return Err(ScanError::new(
|
||||||
"while parsing a flow sequence, expectd ',' or ']'"));
|
mark,
|
||||||
|
"while parsing a flow sequence, expectd ',' or ']'",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
_ => { /* next */ }
|
_ => { /* next */ }
|
||||||
}
|
}
|
||||||
|
@ -698,7 +714,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
self.skip();
|
self.skip();
|
||||||
Ok((Event::SequenceEnd, mark))
|
Ok((Event::SequenceEnd, mark))
|
||||||
},
|
}
|
||||||
Token(mark, TokenType::Key) => {
|
Token(mark, TokenType::Key) => {
|
||||||
self.state = State::FlowSequenceEntryMappingKey;
|
self.state = State::FlowSequenceEntryMappingKey;
|
||||||
self.skip();
|
self.skip();
|
||||||
|
@ -727,7 +743,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
| Token(mark, TokenType::BlockEnd) => {
|
| Token(mark, TokenType::BlockEnd) => {
|
||||||
self.state = State::IndentlessSequenceEntry;
|
self.state = State::IndentlessSequenceEntry;
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::IndentlessSequenceEntry);
|
self.push_state(State::IndentlessSequenceEntry);
|
||||||
self.parse_node(true, false)
|
self.parse_node(true, false)
|
||||||
|
@ -747,25 +763,24 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.pop_state();
|
self.pop_state();
|
||||||
self.skip();
|
self.skip();
|
||||||
Ok((Event::SequenceEnd, mark))
|
Ok((Event::SequenceEnd, mark))
|
||||||
},
|
}
|
||||||
Token(_, TokenType::BlockEntry) => {
|
Token(_, TokenType::BlockEntry) => {
|
||||||
self.skip();
|
self.skip();
|
||||||
match *try!(self.peek_token()) {
|
match *try!(self.peek_token()) {
|
||||||
Token(mark, TokenType::BlockEntry)
|
Token(mark, TokenType::BlockEntry) | Token(mark, TokenType::BlockEnd) => {
|
||||||
| Token(mark, TokenType::BlockEnd) => {
|
|
||||||
self.state = State::BlockSequenceEntry;
|
self.state = State::BlockSequenceEntry;
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::BlockSequenceEntry);
|
self.push_state(State::BlockSequenceEntry);
|
||||||
self.parse_node(true, false)
|
self.parse_node(true, false)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
|
||||||
Token(mark, _) => {
|
|
||||||
Err(ScanError::new(mark,
|
|
||||||
"while parsing a block collection, did not find expected '-' indicator"))
|
|
||||||
}
|
}
|
||||||
|
Token(mark, _) => Err(ScanError::new(
|
||||||
|
mark,
|
||||||
|
"while parsing a block collection, did not find expected '-' indicator",
|
||||||
|
)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -777,7 +792,7 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.skip();
|
self.skip();
|
||||||
self.state = State::FlowSequenceEntryMappingValue;
|
self.state = State::FlowSequenceEntryMappingValue;
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::FlowSequenceEntryMappingValue);
|
self.push_state(State::FlowSequenceEntryMappingValue);
|
||||||
self.parse_node(false, false)
|
self.parse_node(false, false)
|
||||||
|
@ -791,17 +806,16 @@ impl<T: Iterator<Item=char>> Parser<T> {
|
||||||
self.skip();
|
self.skip();
|
||||||
self.state = State::FlowSequenceEntryMappingValue;
|
self.state = State::FlowSequenceEntryMappingValue;
|
||||||
match *try!(self.peek_token()) {
|
match *try!(self.peek_token()) {
|
||||||
Token(mark, TokenType::FlowEntry)
|
Token(mark, TokenType::FlowEntry) | Token(mark, TokenType::FlowSequenceEnd) => {
|
||||||
| Token(mark, TokenType::FlowSequenceEnd) => {
|
|
||||||
self.state = State::FlowSequenceEntryMappingEnd;
|
self.state = State::FlowSequenceEntryMappingEnd;
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
},
|
}
|
||||||
_ => {
|
_ => {
|
||||||
self.push_state(State::FlowSequenceEntryMappingEnd);
|
self.push_state(State::FlowSequenceEntryMappingEnd);
|
||||||
self.parse_node(false, false)
|
self.parse_node(false, false)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
Token(mark, _) => {
|
Token(mark, _) => {
|
||||||
self.state = State::FlowSequenceEntryMappingEnd;
|
self.state = State::FlowSequenceEntryMappingEnd;
|
||||||
Ok((Event::empty_scalar(), mark))
|
Ok((Event::empty_scalar(), mark))
|
||||||
|
|
|
@ -1,10 +1,10 @@
|
||||||
use std::collections::VecDeque;
|
use std::collections::VecDeque;
|
||||||
use std::{char, fmt};
|
|
||||||
use std::error::Error;
|
use std::error::Error;
|
||||||
|
use std::{char, fmt};
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
||||||
pub enum TEncoding {
|
pub enum TEncoding {
|
||||||
Utf8
|
Utf8,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
||||||
|
@ -15,7 +15,7 @@ pub enum TScalarStyle {
|
||||||
DoubleQuoted,
|
DoubleQuoted,
|
||||||
|
|
||||||
Literal,
|
Literal,
|
||||||
Foled
|
Foled,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
|
||||||
|
@ -30,7 +30,7 @@ impl Marker {
|
||||||
Marker {
|
Marker {
|
||||||
index: index,
|
index: index,
|
||||||
line: line,
|
line: line,
|
||||||
col: col
|
col: col,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -57,7 +57,7 @@ impl ScanError {
|
||||||
pub fn new(loc: Marker, info: &str) -> ScanError {
|
pub fn new(loc: Marker, info: &str) -> ScanError {
|
||||||
ScanError {
|
ScanError {
|
||||||
mark: loc,
|
mark: loc,
|
||||||
info: info.to_owned()
|
info: info.to_owned(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -79,8 +79,13 @@ impl Error for ScanError {
|
||||||
impl fmt::Display for ScanError {
|
impl fmt::Display for ScanError {
|
||||||
// col starts from 0
|
// col starts from 0
|
||||||
fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result {
|
fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result {
|
||||||
write!(formatter, "{} at line {} column {}", self.info,
|
write!(
|
||||||
self.mark.line, self.mark.col + 1)
|
formatter,
|
||||||
|
"{} at line {} column {}",
|
||||||
|
self.info,
|
||||||
|
self.mark.line,
|
||||||
|
self.mark.col + 1
|
||||||
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -110,7 +115,7 @@ pub enum TokenType {
|
||||||
Anchor(String),
|
Anchor(String),
|
||||||
/// handle, suffix
|
/// handle, suffix
|
||||||
Tag(String, String),
|
Tag(String, String),
|
||||||
Scalar(TScalarStyle, String)
|
Scalar(TScalarStyle, String),
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, PartialEq, Debug, Eq)]
|
#[derive(Clone, PartialEq, Debug, Eq)]
|
||||||
|
@ -199,14 +204,12 @@ fn is_alpha(c: char) -> bool {
|
||||||
match c {
|
match c {
|
||||||
'0'...'9' | 'a'...'z' | 'A'...'Z' => true,
|
'0'...'9' | 'a'...'z' | 'A'...'Z' => true,
|
||||||
'_' | '-' => true,
|
'_' | '-' => true,
|
||||||
_ => false
|
_ => false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#[inline]
|
#[inline]
|
||||||
fn is_hex(c: char) -> bool {
|
fn is_hex(c: char) -> bool {
|
||||||
(c >= '0' && c <= '9')
|
(c >= '0' && c <= '9') || (c >= 'a' && c <= 'f') || (c >= 'A' && c <= 'F')
|
||||||
|| (c >= 'a' && c <= 'f')
|
|
||||||
|| (c >= 'A' && c <= 'F')
|
|
||||||
}
|
}
|
||||||
#[inline]
|
#[inline]
|
||||||
fn as_hex(c: char) -> u32 {
|
fn as_hex(c: char) -> u32 {
|
||||||
|
@ -214,7 +217,7 @@ fn as_hex(c: char) -> u32 {
|
||||||
'0'...'9' => (c as u32) - ('0' as u32),
|
'0'...'9' => (c as u32) - ('0' as u32),
|
||||||
'a'...'f' => (c as u32) - ('a' as u32) + 10,
|
'a'...'f' => (c as u32) - ('a' as u32) + 10,
|
||||||
'A'...'F' => (c as u32) - ('A' as u32) + 10,
|
'A'...'F' => (c as u32) - ('A' as u32) + 10,
|
||||||
_ => unreachable!()
|
_ => unreachable!(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -363,7 +366,8 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
&& self.buffer[0] == '-'
|
&& self.buffer[0] == '-'
|
||||||
&& self.buffer[1] == '-'
|
&& self.buffer[1] == '-'
|
||||||
&& self.buffer[2] == '-'
|
&& self.buffer[2] == '-'
|
||||||
&& is_blankz(self.buffer[3]) {
|
&& is_blankz(self.buffer[3])
|
||||||
|
{
|
||||||
try!(self.fetch_document_indicator(TokenType::DocumentStart));
|
try!(self.fetch_document_indicator(TokenType::DocumentStart));
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
@ -372,7 +376,8 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
&& self.buffer[0] == '.'
|
&& self.buffer[0] == '.'
|
||||||
&& self.buffer[1] == '.'
|
&& self.buffer[1] == '.'
|
||||||
&& self.buffer[2] == '.'
|
&& self.buffer[2] == '.'
|
||||||
&& is_blankz(self.buffer[3]) {
|
&& is_blankz(self.buffer[3])
|
||||||
|
{
|
||||||
try!(self.fetch_document_indicator(TokenType::DocumentEnd));
|
try!(self.fetch_document_indicator(TokenType::DocumentEnd));
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
@ -402,8 +407,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
// plain scalar
|
// plain scalar
|
||||||
'-' if !is_blankz(nc) => self.fetch_plain_scalar(),
|
'-' if !is_blankz(nc) => self.fetch_plain_scalar(),
|
||||||
':' | '?' if !is_blankz(nc) && self.flow_level == 0 => self.fetch_plain_scalar(),
|
':' | '?' if !is_blankz(nc) && self.flow_level == 0 => self.fetch_plain_scalar(),
|
||||||
'%' | '@' | '`' => Err(ScanError::new(self.mark,
|
'%' | '@' | '`' => Err(ScanError::new(
|
||||||
&format!("unexpected character: `{}'", c))),
|
self.mark,
|
||||||
|
&format!("unexpected character: `{}'", c),
|
||||||
|
)),
|
||||||
_ => self.fetch_plain_scalar(),
|
_ => self.fetch_plain_scalar(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -442,7 +449,9 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if !need_more { break; }
|
if !need_more {
|
||||||
|
break;
|
||||||
|
}
|
||||||
try!(self.fetch_next_token());
|
try!(self.fetch_next_token());
|
||||||
}
|
}
|
||||||
self.token_available = true;
|
self.token_available = true;
|
||||||
|
@ -452,8 +461,9 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
|
|
||||||
fn stale_simple_keys(&mut self) -> ScanResult {
|
fn stale_simple_keys(&mut self) -> ScanResult {
|
||||||
for sk in &mut self.simple_keys {
|
for sk in &mut self.simple_keys {
|
||||||
if sk.possible && (sk.mark.line < self.mark.line
|
if sk.possible
|
||||||
|| sk.mark.index + 1024 < self.mark.index) {
|
&& (sk.mark.line < self.mark.line || sk.mark.index + 1024 < self.mark.index)
|
||||||
|
{
|
||||||
if sk.required {
|
if sk.required {
|
||||||
return Err(ScanError::new(self.mark, "simple key expect ':'"));
|
return Err(ScanError::new(self.mark, "simple key expect ':'"));
|
||||||
}
|
}
|
||||||
|
@ -476,9 +486,12 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
if self.flow_level == 0 {
|
if self.flow_level == 0 {
|
||||||
self.allow_simple_key();
|
self.allow_simple_key();
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
'#' => while !is_breakz(self.ch()) {
|
||||||
|
self.skip();
|
||||||
|
self.lookahead(1);
|
||||||
},
|
},
|
||||||
'#' => while !is_breakz(self.ch()) { self.skip(); self.lookahead(1); },
|
_ => break,
|
||||||
_ => break
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -488,7 +501,8 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.indent = -1;
|
self.indent = -1;
|
||||||
self.stream_start_produced = true;
|
self.stream_start_produced = true;
|
||||||
self.allow_simple_key();
|
self.allow_simple_key();
|
||||||
self.tokens.push_back(Token(mark, TokenType::StreamStart(TEncoding::Utf8)));
|
self.tokens
|
||||||
|
.push_back(Token(mark, TokenType::StreamStart(TEncoding::Utf8)));
|
||||||
self.simple_keys.push(SimpleKey::new(Marker::new(0, 0, 0)));
|
self.simple_keys.push(SimpleKey::new(Marker::new(0, 0, 0)));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -503,7 +517,8 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
try!(self.remove_simple_key());
|
try!(self.remove_simple_key());
|
||||||
self.disallow_simple_key();
|
self.disallow_simple_key();
|
||||||
|
|
||||||
self.tokens.push_back(Token(self.mark, TokenType::StreamEnd));
|
self.tokens
|
||||||
|
.push_back(Token(self.mark, TokenType::StreamEnd));
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -526,12 +541,8 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
|
|
||||||
let name = try!(self.scan_directive_name());
|
let name = try!(self.scan_directive_name());
|
||||||
let tok = match name.as_ref() {
|
let tok = match name.as_ref() {
|
||||||
"YAML" => {
|
"YAML" => try!(self.scan_version_directive_value(&start_mark)),
|
||||||
try!(self.scan_version_directive_value(&start_mark))
|
"TAG" => try!(self.scan_tag_directive_value(&start_mark)),
|
||||||
},
|
|
||||||
"TAG" => {
|
|
||||||
try!(self.scan_tag_directive_value(&start_mark))
|
|
||||||
},
|
|
||||||
// XXX This should be a warning instead of an error
|
// XXX This should be a warning instead of an error
|
||||||
_ => {
|
_ => {
|
||||||
// skip current line
|
// skip current line
|
||||||
|
@ -541,7 +552,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
}
|
}
|
||||||
// XXX return an empty TagDirective token
|
// XXX return an empty TagDirective token
|
||||||
Token(start_mark, TokenType::TagDirective(String::new(), String::new()))
|
Token(
|
||||||
|
start_mark,
|
||||||
|
TokenType::TagDirective(String::new(), String::new()),
|
||||||
|
)
|
||||||
// return Err(ScanError::new(start_mark,
|
// return Err(ScanError::new(start_mark,
|
||||||
// "while scanning a directive, found unknown directive name"))
|
// "while scanning a directive, found unknown directive name"))
|
||||||
}
|
}
|
||||||
|
@ -561,8 +575,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
if !is_breakz(self.ch()) {
|
if !is_breakz(self.ch()) {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a directive, did not find expected comment or line break"));
|
start_mark,
|
||||||
|
"while scanning a directive, did not find expected comment or line break",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
// Eat a line break
|
// Eat a line break
|
||||||
|
@ -585,8 +601,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
let major = try!(self.scan_version_directive_number(mark));
|
let major = try!(self.scan_version_directive_number(mark));
|
||||||
|
|
||||||
if self.ch() != '.' {
|
if self.ch() != '.' {
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a YAML directive, did not find expected digit or '.' character"));
|
*mark,
|
||||||
|
"while scanning a YAML directive, did not find expected digit or '.' character",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
self.skip();
|
self.skip();
|
||||||
|
@ -607,13 +625,17 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
if string.is_empty() {
|
if string.is_empty() {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a directive, could not find expected directive name"));
|
start_mark,
|
||||||
|
"while scanning a directive, could not find expected directive name",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
if !is_blankz(self.ch()) {
|
if !is_blankz(self.ch()) {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a directive, found unexpected non-alphabetical character"));
|
start_mark,
|
||||||
|
"while scanning a directive, found unexpected non-alphabetical character",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(string)
|
Ok(string)
|
||||||
|
@ -625,8 +647,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
while is_digit(self.ch()) {
|
while is_digit(self.ch()) {
|
||||||
if length + 1 > 9 {
|
if length + 1 > 9 {
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a YAML directive, found extremely long version number"));
|
*mark,
|
||||||
|
"while scanning a YAML directive, found extremely long version number",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
length += 1;
|
length += 1;
|
||||||
val = val * 10 + ((self.ch() as u32) - ('0' as u32));
|
val = val * 10 + ((self.ch() as u32) - ('0' as u32));
|
||||||
|
@ -635,8 +659,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
if length == 0 {
|
if length == 0 {
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a YAML directive, did not find expected version number"));
|
*mark,
|
||||||
|
"while scanning a YAML directive, did not find expected version number",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(val)
|
Ok(val)
|
||||||
|
@ -666,8 +692,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
if is_blankz(self.ch()) {
|
if is_blankz(self.ch()) {
|
||||||
Ok(Token(*mark, TokenType::TagDirective(handle, prefix)))
|
Ok(Token(*mark, TokenType::TagDirective(handle, prefix)))
|
||||||
} else {
|
} else {
|
||||||
Err(ScanError::new(*mark,
|
Err(ScanError::new(
|
||||||
"while scanning TAG, did not find expected whitespace or line break"))
|
*mark,
|
||||||
|
"while scanning TAG, did not find expected whitespace or line break",
|
||||||
|
))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -696,8 +724,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
suffix = try!(self.scan_tag_uri(false, false, &String::new(), &start_mark));
|
suffix = try!(self.scan_tag_uri(false, false, &String::new(), &start_mark));
|
||||||
|
|
||||||
if self.ch() != '>' {
|
if self.ch() != '>' {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a tag, did not find the expected '>'"));
|
start_mark,
|
||||||
|
"while scanning a tag, did not find the expected '>'",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
self.skip();
|
self.skip();
|
||||||
|
@ -727,8 +757,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
// XXX: ex 7.2, an empty scalar can follow a secondary tag
|
// XXX: ex 7.2, an empty scalar can follow a secondary tag
|
||||||
Ok(Token(start_mark, TokenType::Tag(handle, suffix)))
|
Ok(Token(start_mark, TokenType::Tag(handle, suffix)))
|
||||||
} else {
|
} else {
|
||||||
Err(ScanError::new(start_mark,
|
Err(ScanError::new(
|
||||||
"while scanning a tag, did not find expected whitespace or line break"))
|
start_mark,
|
||||||
|
"while scanning a tag, did not find expected whitespace or line break",
|
||||||
|
))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -736,8 +768,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
let mut string = String::new();
|
let mut string = String::new();
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
if self.ch() != '!' {
|
if self.ch() != '!' {
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a tag, did not find expected '!'"));
|
*mark,
|
||||||
|
"while scanning a tag, did not find expected '!'",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
string.push(self.ch());
|
string.push(self.ch());
|
||||||
|
@ -758,14 +792,21 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
// It's either the '!' tag or not really a tag handle. If it's a %TAG
|
// It's either the '!' tag or not really a tag handle. If it's a %TAG
|
||||||
// directive, it's an error. If it's a tag token, it must be a part of
|
// directive, it's an error. If it's a tag token, it must be a part of
|
||||||
// URI.
|
// URI.
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while parsing a tag directive, did not find expected '!'"));
|
*mark,
|
||||||
|
"while parsing a tag directive, did not find expected '!'",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
Ok(string)
|
Ok(string)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn scan_tag_uri(&mut self, directive: bool, _is_secondary: bool,
|
fn scan_tag_uri(
|
||||||
head: &str, mark: &Marker) -> Result<String, ScanError> {
|
&mut self,
|
||||||
|
directive: bool,
|
||||||
|
_is_secondary: bool,
|
||||||
|
head: &str,
|
||||||
|
mark: &Marker,
|
||||||
|
) -> Result<String, ScanError> {
|
||||||
let mut length = head.len();
|
let mut length = head.len();
|
||||||
let mut string = String::new();
|
let mut string = String::new();
|
||||||
|
|
||||||
|
@ -788,7 +829,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
'=' | '+' | '$' | ',' | '.' | '!' | '~' | '*' | '\'' | '(' | ')' | '[' | ']' => true,
|
'=' | '+' | '$' | ',' | '.' | '!' | '~' | '*' | '\'' | '(' | ')' | '[' | ']' => true,
|
||||||
'%' => true,
|
'%' => true,
|
||||||
c if is_alpha(c) => true,
|
c if is_alpha(c) => true,
|
||||||
_ => false
|
_ => false,
|
||||||
} {
|
} {
|
||||||
// Check if it is a URI-escape sequence.
|
// Check if it is a URI-escape sequence.
|
||||||
if self.ch() == '%' {
|
if self.ch() == '%' {
|
||||||
|
@ -803,25 +844,26 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
if length == 0 {
|
if length == 0 {
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while parsing a tag, did not find expected tag URI"));
|
*mark,
|
||||||
|
"while parsing a tag, did not find expected tag URI",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(string)
|
Ok(string)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn scan_uri_escapes(&mut self, _directive: bool, mark: &Marker)
|
fn scan_uri_escapes(&mut self, _directive: bool, mark: &Marker) -> Result<char, ScanError> {
|
||||||
-> Result<char, ScanError> {
|
|
||||||
let mut width = 0usize;
|
let mut width = 0usize;
|
||||||
let mut code = 0u32;
|
let mut code = 0u32;
|
||||||
loop {
|
loop {
|
||||||
self.lookahead(3);
|
self.lookahead(3);
|
||||||
|
|
||||||
if !(self.ch() == '%'
|
if !(self.ch() == '%' && is_hex(self.buffer[1]) && is_hex(self.buffer[2])) {
|
||||||
&& is_hex(self.buffer[1])
|
return Err(ScanError::new(
|
||||||
&& is_hex(self.buffer[2])) {
|
*mark,
|
||||||
return Err(ScanError::new(*mark,
|
"while parsing a tag, did not find URI escaped octet",
|
||||||
"while parsing a tag, did not find URI escaped octet"));
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
let octet = (as_hex(self.buffer[1]) << 4) + as_hex(self.buffer[2]);
|
let octet = (as_hex(self.buffer[1]) << 4) + as_hex(self.buffer[2]);
|
||||||
|
@ -832,15 +874,19 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
_ if octet & 0xF0 == 0xE0 => 3,
|
_ if octet & 0xF0 == 0xE0 => 3,
|
||||||
_ if octet & 0xF8 == 0xF0 => 4,
|
_ if octet & 0xF8 == 0xF0 => 4,
|
||||||
_ => {
|
_ => {
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while parsing a tag, found an incorrect leading UTF-8 octet"));
|
*mark,
|
||||||
|
"while parsing a tag, found an incorrect leading UTF-8 octet",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
code = octet;
|
code = octet;
|
||||||
} else {
|
} else {
|
||||||
if octet & 0xc0 != 0x80 {
|
if octet & 0xc0 != 0x80 {
|
||||||
return Err(ScanError::new(*mark,
|
return Err(ScanError::new(
|
||||||
"while parsing a tag, found an incorrect trailing UTF-8 octet"));
|
*mark,
|
||||||
|
"while parsing a tag, found an incorrect trailing UTF-8 octet",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
code = (code << 8) + octet;
|
code = (code << 8) + octet;
|
||||||
}
|
}
|
||||||
|
@ -857,8 +903,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
|
|
||||||
match char::from_u32(code) {
|
match char::from_u32(code) {
|
||||||
Some(ch) => Ok(ch),
|
Some(ch) => Ok(ch),
|
||||||
None => Err(ScanError::new(*mark,
|
None => Err(ScanError::new(
|
||||||
"while parsing a tag, found an invalid UTF-8 codepoint"))
|
*mark,
|
||||||
|
"while parsing a tag, found an invalid UTF-8 codepoint",
|
||||||
|
)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -873,8 +921,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn scan_anchor(&mut self, alias: bool)
|
fn scan_anchor(&mut self, alias: bool) -> Result<Token, ScanError> {
|
||||||
-> Result<Token, ScanError> {
|
|
||||||
let mut string = String::new();
|
let mut string = String::new();
|
||||||
let start_mark = self.mark;
|
let start_mark = self.mark;
|
||||||
|
|
||||||
|
@ -886,11 +933,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
}
|
}
|
||||||
|
|
||||||
if string.is_empty()
|
if string.is_empty() || match self.ch() {
|
||||||
|| match self.ch() {
|
|
||||||
c if is_blankz(c) => false,
|
c if is_blankz(c) => false,
|
||||||
'?' | ':' | ',' | ']' | '}' | '%' | '@' | '`' => false,
|
'?' | ':' | ',' | ']' | '}' | '%' | '@' | '`' => false,
|
||||||
_ => true
|
_ => true,
|
||||||
} {
|
} {
|
||||||
return Err(ScanError::new(start_mark, "while scanning an anchor or alias, did not find expected alphabetic or numeric character"));
|
return Err(ScanError::new(start_mark, "while scanning an anchor or alias, did not find expected alphabetic or numeric character"));
|
||||||
}
|
}
|
||||||
|
@ -937,13 +983,16 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
let start_mark = self.mark;
|
let start_mark = self.mark;
|
||||||
self.skip();
|
self.skip();
|
||||||
|
|
||||||
self.tokens.push_back(Token(start_mark, TokenType::FlowEntry));
|
self.tokens
|
||||||
|
.push_back(Token(start_mark, TokenType::FlowEntry));
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn increase_flow_level(&mut self) -> ScanResult {
|
fn increase_flow_level(&mut self) -> ScanResult {
|
||||||
self.simple_keys.push(SimpleKey::new(Marker::new(0, 0, 0)));
|
self.simple_keys.push(SimpleKey::new(Marker::new(0, 0, 0)));
|
||||||
self.flow_level = self.flow_level.checked_add(1)
|
self.flow_level = self
|
||||||
|
.flow_level
|
||||||
|
.checked_add(1)
|
||||||
.ok_or_else(|| ScanError::new(self.mark, "Recursion limit exceeded"))?;
|
.ok_or_else(|| ScanError::new(self.mark, "Recursion limit exceeded"))?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
@ -958,8 +1007,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
if self.flow_level == 0 {
|
if self.flow_level == 0 {
|
||||||
// Check if we are allowed to start a new entry.
|
// Check if we are allowed to start a new entry.
|
||||||
if !self.simple_key_allowed {
|
if !self.simple_key_allowed {
|
||||||
return Err(ScanError::new(self.mark,
|
return Err(ScanError::new(
|
||||||
"block sequence entries are not allowed in this context"));
|
self.mark,
|
||||||
|
"block sequence entries are not allowed in this context",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
let mark = self.mark;
|
let mark = self.mark;
|
||||||
|
@ -967,7 +1018,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.roll_indent(mark.col, None, TokenType::BlockSequenceStart, mark);
|
self.roll_indent(mark.col, None, TokenType::BlockSequenceStart, mark);
|
||||||
} else {
|
} else {
|
||||||
// - * only allowed in block
|
// - * only allowed in block
|
||||||
return Err(ScanError::new(self.mark, r#""-" is only valid inside a block"#))
|
return Err(ScanError::new(
|
||||||
|
self.mark,
|
||||||
|
r#""-" is only valid inside a block"#,
|
||||||
|
));
|
||||||
}
|
}
|
||||||
try!(self.remove_simple_key());
|
try!(self.remove_simple_key());
|
||||||
self.allow_simple_key();
|
self.allow_simple_key();
|
||||||
|
@ -975,7 +1029,8 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
let start_mark = self.mark;
|
let start_mark = self.mark;
|
||||||
self.skip();
|
self.skip();
|
||||||
|
|
||||||
self.tokens.push_back(Token(start_mark, TokenType::BlockEntry));
|
self.tokens
|
||||||
|
.push_back(Token(start_mark, TokenType::BlockEntry));
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1029,16 +1084,20 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
if is_digit(self.ch()) {
|
if is_digit(self.ch()) {
|
||||||
if self.ch() == '0' {
|
if self.ch() == '0' {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a block scalar, found an intendation indicator equal to 0"));
|
start_mark,
|
||||||
|
"while scanning a block scalar, found an intendation indicator equal to 0",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
increment = (self.ch() as usize) - ('0' as usize);
|
increment = (self.ch() as usize) - ('0' as usize);
|
||||||
self.skip();
|
self.skip();
|
||||||
}
|
}
|
||||||
} else if is_digit(self.ch()) {
|
} else if is_digit(self.ch()) {
|
||||||
if self.ch() == '0' {
|
if self.ch() == '0' {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a block scalar, found an intendation indicator equal to 0"));
|
start_mark,
|
||||||
|
"while scanning a block scalar, found an intendation indicator equal to 0",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
increment = (self.ch() as usize) - ('0' as usize);
|
increment = (self.ch() as usize) - ('0' as usize);
|
||||||
|
@ -1071,8 +1130,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
|
|
||||||
// Check if we are at the end of the line.
|
// Check if we are at the end of the line.
|
||||||
if !is_breakz(self.ch()) {
|
if !is_breakz(self.ch()) {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a block scalar, did not find expected comment or line break"));
|
start_mark,
|
||||||
|
"while scanning a block scalar, did not find expected comment or line break",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
if is_break(self.ch()) {
|
if is_break(self.ch()) {
|
||||||
|
@ -1081,7 +1142,11 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
if increment > 0 {
|
if increment > 0 {
|
||||||
indent = if self.indent >= 0 { (self.indent + increment as isize) as usize } else { increment }
|
indent = if self.indent >= 0 {
|
||||||
|
(self.indent + increment as isize) as usize
|
||||||
|
} else {
|
||||||
|
increment
|
||||||
|
}
|
||||||
}
|
}
|
||||||
// Scan the leading line breaks and determine the indentation level if needed.
|
// Scan the leading line breaks and determine the indentation level if needed.
|
||||||
try!(self.block_scalar_breaks(&mut indent, &mut trailing_breaks));
|
try!(self.block_scalar_breaks(&mut indent, &mut trailing_breaks));
|
||||||
|
@ -1093,8 +1158,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
while self.mark.col == indent && !is_z(self.ch()) {
|
while self.mark.col == indent && !is_z(self.ch()) {
|
||||||
// We are at the beginning of a non-empty line.
|
// We are at the beginning of a non-empty line.
|
||||||
trailing_blank = is_blank(self.ch());
|
trailing_blank = is_blank(self.ch());
|
||||||
if !literal && !leading_break.is_empty()
|
if !literal && !leading_break.is_empty() && !leading_blank && !trailing_blank {
|
||||||
&& !leading_blank && !trailing_blank {
|
|
||||||
if trailing_breaks.is_empty() {
|
if trailing_breaks.is_empty() {
|
||||||
string.push(' ');
|
string.push(' ');
|
||||||
}
|
}
|
||||||
|
@ -1115,7 +1179,9 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
}
|
}
|
||||||
// break on EOF
|
// break on EOF
|
||||||
if is_z(self.ch()) { break; }
|
if is_z(self.ch()) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
self.lookahead(2);
|
self.lookahead(2);
|
||||||
self.read_break(&mut leading_break);
|
self.read_break(&mut leading_break);
|
||||||
|
@ -1134,9 +1200,15 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
if literal {
|
if literal {
|
||||||
Ok(Token(start_mark, TokenType::Scalar(TScalarStyle::Literal, string)))
|
Ok(Token(
|
||||||
|
start_mark,
|
||||||
|
TokenType::Scalar(TScalarStyle::Literal, string),
|
||||||
|
))
|
||||||
} else {
|
} else {
|
||||||
Ok(Token(start_mark, TokenType::Scalar(TScalarStyle::Foled, string)))
|
Ok(Token(
|
||||||
|
start_mark,
|
||||||
|
TokenType::Scalar(TScalarStyle::Foled, string),
|
||||||
|
))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1144,8 +1216,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
let mut max_indent = 0;
|
let mut max_indent = 0;
|
||||||
loop {
|
loop {
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
while (*indent == 0 || self.mark.col < *indent)
|
while (*indent == 0 || self.mark.col < *indent) && self.buffer[0] == ' ' {
|
||||||
&& self.buffer[0] == ' ' {
|
|
||||||
self.skip();
|
self.skip();
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
}
|
}
|
||||||
|
@ -1155,8 +1226,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for a tab character messing the intendation.
|
// Check for a tab character messing the intendation.
|
||||||
if (*indent == 0 || self.mark.col < *indent)
|
if (*indent == 0 || self.mark.col < *indent) && self.buffer[0] == '\t' {
|
||||||
&& self.buffer[0] == '\t' {
|
|
||||||
return Err(ScanError::new(self.mark,
|
return Err(ScanError::new(self.mark,
|
||||||
"while scanning a block scalar, found a tab character where an intendation space is expected"));
|
"while scanning a block scalar, found a tab character where an intendation space is expected"));
|
||||||
}
|
}
|
||||||
|
@ -1208,21 +1278,24 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
/* Check for a document indicator. */
|
/* Check for a document indicator. */
|
||||||
self.lookahead(4);
|
self.lookahead(4);
|
||||||
|
|
||||||
if self.mark.col == 0 &&
|
if self.mark.col == 0
|
||||||
(((self.buffer[0] == '-') &&
|
&& (((self.buffer[0] == '-') && (self.buffer[1] == '-') && (self.buffer[2] == '-'))
|
||||||
(self.buffer[1] == '-') &&
|
|| ((self.buffer[0] == '.')
|
||||||
(self.buffer[2] == '-')) ||
|
&& (self.buffer[1] == '.')
|
||||||
((self.buffer[0] == '.') &&
|
&& (self.buffer[2] == '.')))
|
||||||
(self.buffer[1] == '.') &&
|
&& is_blankz(self.buffer[3])
|
||||||
(self.buffer[2] == '.'))) &&
|
{
|
||||||
is_blankz(self.buffer[3]) {
|
return Err(ScanError::new(
|
||||||
return Err(ScanError::new(start_mark,
|
start_mark,
|
||||||
"while scanning a quoted scalar, found unexpected document indicator"));
|
"while scanning a quoted scalar, found unexpected document indicator",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
if is_z(self.ch()) {
|
if is_z(self.ch()) {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"while scanning a quoted scalar, found unexpected end of stream"));
|
start_mark,
|
||||||
|
"while scanning a quoted scalar, found unexpected end of stream",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
self.lookahead(2);
|
self.lookahead(2);
|
||||||
|
@ -1237,10 +1310,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
string.push('\'');
|
string.push('\'');
|
||||||
self.skip();
|
self.skip();
|
||||||
self.skip();
|
self.skip();
|
||||||
},
|
}
|
||||||
// Check for the right quote.
|
// Check for the right quote.
|
||||||
'\'' if single => { break; },
|
'\'' if single => break,
|
||||||
'"' if !single => { break; },
|
'"' if !single => break,
|
||||||
// Check for an escaped line break.
|
// Check for an escaped line break.
|
||||||
'\\' if !single && is_break(self.buffer[1]) => {
|
'\\' if !single && is_break(self.buffer[1]) => {
|
||||||
self.lookahead(3);
|
self.lookahead(3);
|
||||||
|
@ -1277,8 +1350,12 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
'x' => code_length = 2,
|
'x' => code_length = 2,
|
||||||
'u' => code_length = 4,
|
'u' => code_length = 4,
|
||||||
'U' => code_length = 8,
|
'U' => code_length = 8,
|
||||||
_ => return Err(ScanError::new(start_mark,
|
_ => {
|
||||||
"while parsing a quoted scalar, found unknown escape character"))
|
return Err(ScanError::new(
|
||||||
|
start_mark,
|
||||||
|
"while parsing a quoted scalar, found unknown escape character",
|
||||||
|
))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
self.skip();
|
self.skip();
|
||||||
self.skip();
|
self.skip();
|
||||||
|
@ -1307,15 +1384,18 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.skip();
|
self.skip();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
c => { string.push(c); self.skip(); }
|
c => {
|
||||||
|
string.push(c);
|
||||||
|
self.skip();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
self.lookahead(2);
|
self.lookahead(2);
|
||||||
}
|
}
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
match self.ch() {
|
match self.ch() {
|
||||||
'\'' if single => { break; },
|
'\'' if single => break,
|
||||||
'"' if !single => { break; },
|
'"' if !single => break,
|
||||||
_ => {}
|
_ => {}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1368,9 +1448,15 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.skip();
|
self.skip();
|
||||||
|
|
||||||
if single {
|
if single {
|
||||||
Ok(Token(start_mark, TokenType::Scalar(TScalarStyle::SingleQuoted, string)))
|
Ok(Token(
|
||||||
|
start_mark,
|
||||||
|
TokenType::Scalar(TScalarStyle::SingleQuoted, string),
|
||||||
|
))
|
||||||
} else {
|
} else {
|
||||||
Ok(Token(start_mark, TokenType::Scalar(TScalarStyle::DoubleQuoted, string)))
|
Ok(Token(
|
||||||
|
start_mark,
|
||||||
|
TokenType::Scalar(TScalarStyle::DoubleQuoted, string),
|
||||||
|
))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1398,23 +1484,25 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
/* Check for a document indicator. */
|
/* Check for a document indicator. */
|
||||||
self.lookahead(4);
|
self.lookahead(4);
|
||||||
|
|
||||||
if self.mark.col == 0 &&
|
if self.mark.col == 0
|
||||||
(((self.buffer[0] == '-') &&
|
&& (((self.buffer[0] == '-') && (self.buffer[1] == '-') && (self.buffer[2] == '-'))
|
||||||
(self.buffer[1] == '-') &&
|
|| ((self.buffer[0] == '.')
|
||||||
(self.buffer[2] == '-')) ||
|
&& (self.buffer[1] == '.')
|
||||||
((self.buffer[0] == '.') &&
|
&& (self.buffer[2] == '.')))
|
||||||
(self.buffer[1] == '.') &&
|
&& is_blankz(self.buffer[3])
|
||||||
(self.buffer[2] == '.'))) &&
|
{
|
||||||
is_blankz(self.buffer[3]) {
|
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
if self.ch() == '#' { break; }
|
if self.ch() == '#' {
|
||||||
|
break;
|
||||||
|
}
|
||||||
while !is_blankz(self.ch()) {
|
while !is_blankz(self.ch()) {
|
||||||
if self.flow_level > 0 && self.ch() == ':'
|
if self.flow_level > 0 && self.ch() == ':' && is_blankz(self.ch()) {
|
||||||
&& is_blankz(self.ch()) {
|
return Err(ScanError::new(
|
||||||
return Err(ScanError::new(start_mark,
|
start_mark,
|
||||||
"while scanning a plain scalar, found unexpected ':'"));
|
"while scanning a plain scalar, found unexpected ':'",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
// indicators ends a plain scalar
|
// indicators ends a plain scalar
|
||||||
match self.ch() {
|
match self.ch() {
|
||||||
|
@ -1438,7 +1526,6 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
trailing_breaks.clear();
|
trailing_breaks.clear();
|
||||||
}
|
}
|
||||||
leading_break.clear();
|
leading_break.clear();
|
||||||
|
|
||||||
}
|
}
|
||||||
leading_blanks = false;
|
leading_blanks = false;
|
||||||
} else {
|
} else {
|
||||||
|
@ -1452,15 +1539,18 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.lookahead(2);
|
self.lookahead(2);
|
||||||
}
|
}
|
||||||
// is the end?
|
// is the end?
|
||||||
if !(is_blank(self.ch()) || is_break(self.ch())) { break; }
|
if !(is_blank(self.ch()) || is_break(self.ch())) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
self.lookahead(1);
|
self.lookahead(1);
|
||||||
|
|
||||||
while is_blank(self.ch()) || is_break(self.ch()) {
|
while is_blank(self.ch()) || is_break(self.ch()) {
|
||||||
if is_blank(self.ch()) {
|
if is_blank(self.ch()) {
|
||||||
if leading_blanks && (self.mark.col as isize) < indent
|
if leading_blanks && (self.mark.col as isize) < indent && self.ch() == '\t' {
|
||||||
&& self.ch() == '\t' {
|
return Err(ScanError::new(
|
||||||
return Err(ScanError::new(start_mark,
|
start_mark,
|
||||||
"while scanning a plain scalar, found a tab"));
|
"while scanning a plain scalar, found a tab",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
if leading_blanks {
|
if leading_blanks {
|
||||||
|
@ -1493,7 +1583,10 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.allow_simple_key();
|
self.allow_simple_key();
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(Token(start_mark, TokenType::Scalar(TScalarStyle::Plain, string)))
|
Ok(Token(
|
||||||
|
start_mark,
|
||||||
|
TokenType::Scalar(TScalarStyle::Plain, string),
|
||||||
|
))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn fetch_key(&mut self) -> ScanResult {
|
fn fetch_key(&mut self) -> ScanResult {
|
||||||
|
@ -1501,10 +1594,17 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
if self.flow_level == 0 {
|
if self.flow_level == 0 {
|
||||||
// Check if we are allowed to start a new key (not nessesary simple).
|
// Check if we are allowed to start a new key (not nessesary simple).
|
||||||
if !self.simple_key_allowed {
|
if !self.simple_key_allowed {
|
||||||
return Err(ScanError::new(self.mark, "mapping keys are not allowed in this context"));
|
return Err(ScanError::new(
|
||||||
|
self.mark,
|
||||||
|
"mapping keys are not allowed in this context",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
self.roll_indent(start_mark.col, None,
|
self.roll_indent(
|
||||||
TokenType::BlockMappingStart, start_mark);
|
start_mark.col,
|
||||||
|
None,
|
||||||
|
TokenType::BlockMappingStart,
|
||||||
|
start_mark,
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
try!(self.remove_simple_key());
|
try!(self.remove_simple_key());
|
||||||
|
@ -1530,8 +1630,12 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
self.insert_token(sk.token_number - tokens_parsed, tok);
|
self.insert_token(sk.token_number - tokens_parsed, tok);
|
||||||
|
|
||||||
// Add the BLOCK-MAPPING-START token if needed.
|
// Add the BLOCK-MAPPING-START token if needed.
|
||||||
self.roll_indent(sk.mark.col, Some(sk.token_number),
|
self.roll_indent(
|
||||||
TokenType::BlockMappingStart, start_mark);
|
sk.mark.col,
|
||||||
|
Some(sk.token_number),
|
||||||
|
TokenType::BlockMappingStart,
|
||||||
|
start_mark,
|
||||||
|
);
|
||||||
|
|
||||||
self.simple_keys.last_mut().unwrap().possible = false;
|
self.simple_keys.last_mut().unwrap().possible = false;
|
||||||
self.disallow_simple_key();
|
self.disallow_simple_key();
|
||||||
|
@ -1539,12 +1643,18 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
// The ':' indicator follows a complex key.
|
// The ':' indicator follows a complex key.
|
||||||
if self.flow_level == 0 {
|
if self.flow_level == 0 {
|
||||||
if !self.simple_key_allowed {
|
if !self.simple_key_allowed {
|
||||||
return Err(ScanError::new(start_mark,
|
return Err(ScanError::new(
|
||||||
"mapping values are not allowed in this context"));
|
start_mark,
|
||||||
|
"mapping values are not allowed in this context",
|
||||||
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
self.roll_indent(start_mark.col, None,
|
self.roll_indent(
|
||||||
TokenType::BlockMappingStart, start_mark);
|
start_mark.col,
|
||||||
|
None,
|
||||||
|
TokenType::BlockMappingStart,
|
||||||
|
start_mark,
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
if self.flow_level == 0 {
|
if self.flow_level == 0 {
|
||||||
|
@ -1559,8 +1669,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn roll_indent(&mut self, col: usize, number: Option<usize>,
|
fn roll_indent(&mut self, col: usize, number: Option<usize>, tok: TokenType, mark: Marker) {
|
||||||
tok: TokenType, mark: Marker) {
|
|
||||||
if self.flow_level > 0 {
|
if self.flow_level > 0 {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
@ -1571,7 +1680,7 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
let tokens_parsed = self.tokens_parsed;
|
let tokens_parsed = self.tokens_parsed;
|
||||||
match number {
|
match number {
|
||||||
Some(n) => self.insert_token(n - tokens_parsed, Token(mark, tok)),
|
Some(n) => self.insert_token(n - tokens_parsed, Token(mark, tok)),
|
||||||
None => self.tokens.push_back(Token(mark, tok))
|
None => self.tokens.push_back(Token(mark, tok)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1611,23 +1720,21 @@ impl<T: Iterator<Item=char>> Scanner<T> {
|
||||||
last.possible = false;
|
last.possible = false;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod test {
|
mod test {
|
||||||
use super::*;
|
|
||||||
use super::TokenType::*;
|
use super::TokenType::*;
|
||||||
|
use super::*;
|
||||||
|
|
||||||
macro_rules! next {
|
macro_rules! next {
|
||||||
($p:ident, $tk:pat) => {{
|
($p:ident, $tk:pat) => {{
|
||||||
let tok = $p.next().unwrap();
|
let tok = $p.next().unwrap();
|
||||||
match tok.1 {
|
match tok.1 {
|
||||||
$tk => {},
|
$tk => {}
|
||||||
_ => { panic!("unexpected token: {:?}",
|
_ => panic!("unexpected token: {:?}", tok),
|
||||||
tok) }
|
|
||||||
}
|
}
|
||||||
}}
|
}};
|
||||||
}
|
}
|
||||||
|
|
||||||
macro_rules! next_scalar {
|
macro_rules! next_scalar {
|
||||||
|
@ -1637,17 +1744,16 @@ macro_rules! next_scalar {
|
||||||
Scalar(style, ref v) => {
|
Scalar(style, ref v) => {
|
||||||
assert_eq!(style, $tk);
|
assert_eq!(style, $tk);
|
||||||
assert_eq!(v, $v);
|
assert_eq!(v, $v);
|
||||||
},
|
|
||||||
_ => { panic!("unexpected token: {:?}",
|
|
||||||
tok) }
|
|
||||||
}
|
}
|
||||||
}}
|
_ => panic!("unexpected token: {:?}", tok),
|
||||||
|
}
|
||||||
|
}};
|
||||||
}
|
}
|
||||||
|
|
||||||
macro_rules! end {
|
macro_rules! end {
|
||||||
($p:ident) => {{
|
($p:ident) => {{
|
||||||
assert_eq!($p.next(), None);
|
assert_eq!($p.next(), None);
|
||||||
}}
|
}};
|
||||||
}
|
}
|
||||||
/// test cases in libyaml scanner.c
|
/// test cases in libyaml scanner.c
|
||||||
#[test]
|
#[test]
|
||||||
|
@ -1671,8 +1777,7 @@ macro_rules! end {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_explicit_scalar() {
|
fn test_explicit_scalar() {
|
||||||
let s =
|
let s = "---
|
||||||
"---
|
|
||||||
'a scalar'
|
'a scalar'
|
||||||
...
|
...
|
||||||
";
|
";
|
||||||
|
@ -1687,8 +1792,7 @@ macro_rules! end {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_multiple_documents() {
|
fn test_multiple_documents() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
'a scalar'
|
'a scalar'
|
||||||
---
|
---
|
||||||
'a scalar'
|
'a scalar'
|
||||||
|
@ -1724,8 +1828,7 @@ macro_rules! end {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_a_flow_mapping() {
|
fn test_a_flow_mapping() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
{
|
{
|
||||||
a simple key: a value, # Note that the KEY token is produced.
|
a simple key: a value, # Note that the KEY token is produced.
|
||||||
? a complex key: another value,
|
? a complex key: another value,
|
||||||
|
@ -1751,8 +1854,7 @@ macro_rules! end {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_block_sequences() {
|
fn test_block_sequences() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
- item 1
|
- item 1
|
||||||
- item 2
|
- item 2
|
||||||
-
|
-
|
||||||
|
@ -1794,8 +1896,7 @@ macro_rules! end {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_block_mappings() {
|
fn test_block_mappings() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
a simple key: a value # The KEY token is produced here.
|
a simple key: a value # The KEY token is produced here.
|
||||||
? a complex key
|
? a complex key
|
||||||
: another value
|
: another value
|
||||||
|
@ -1842,13 +1943,11 @@ a sequence:
|
||||||
next!(p, BlockEnd);
|
next!(p, BlockEnd);
|
||||||
next!(p, StreamEnd);
|
next!(p, StreamEnd);
|
||||||
end!(p);
|
end!(p);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_no_block_sequence_start() {
|
fn test_no_block_sequence_start() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
key:
|
key:
|
||||||
- item 1
|
- item 1
|
||||||
- item 2
|
- item 2
|
||||||
|
@ -1870,8 +1969,7 @@ key:
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_collections_in_sequence() {
|
fn test_collections_in_sequence() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
- - item 1
|
- - item 1
|
||||||
- item 2
|
- item 2
|
||||||
- key 1: value 1
|
- key 1: value 1
|
||||||
|
@ -1914,8 +2012,7 @@ key:
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_collections_in_mapping() {
|
fn test_collections_in_mapping() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
? a sequence
|
? a sequence
|
||||||
: - item 1
|
: - item 1
|
||||||
- item 2
|
- item 2
|
||||||
|
@ -1955,8 +2052,7 @@ key:
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_spec_ex7_3() {
|
fn test_spec_ex7_3() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
{
|
{
|
||||||
? foo :,
|
? foo :,
|
||||||
: bar,
|
: bar,
|
||||||
|
|
|
@ -1,13 +1,13 @@
|
||||||
|
use linked_hash_map::LinkedHashMap;
|
||||||
|
use parser::*;
|
||||||
|
use scanner::{Marker, ScanError, TScalarStyle, TokenType};
|
||||||
use std::collections::BTreeMap;
|
use std::collections::BTreeMap;
|
||||||
|
use std::f64;
|
||||||
|
use std::i64;
|
||||||
|
use std::mem;
|
||||||
use std::ops::Index;
|
use std::ops::Index;
|
||||||
use std::string;
|
use std::string;
|
||||||
use std::i64;
|
|
||||||
use std::f64;
|
|
||||||
use std::mem;
|
|
||||||
use std::vec;
|
use std::vec;
|
||||||
use parser::*;
|
|
||||||
use scanner::{TScalarStyle, ScanError, TokenType, Marker};
|
|
||||||
use linked_hash_map::LinkedHashMap;
|
|
||||||
|
|
||||||
/// A YAML node is stored as this `Yaml` enumeration, which provides an easy way to
|
/// A YAML node is stored as this `Yaml` enumeration, which provides an easy way to
|
||||||
/// access your YAML document.
|
/// access your YAML document.
|
||||||
|
@ -62,7 +62,7 @@ fn parse_f64(v: &str) -> Option<f64> {
|
||||||
".inf" | ".Inf" | ".INF" | "+.inf" | "+.Inf" | "+.INF" => Some(f64::INFINITY),
|
".inf" | ".Inf" | ".INF" | "+.inf" | "+.Inf" | "+.INF" => Some(f64::INFINITY),
|
||||||
"-.inf" | "-.Inf" | "-.INF" => Some(f64::NEG_INFINITY),
|
"-.inf" | "-.Inf" | "-.INF" => Some(f64::NEG_INFINITY),
|
||||||
".nan" | "NaN" | ".NAN" => Some(f64::NAN),
|
".nan" | "NaN" | ".NAN" => Some(f64::NAN),
|
||||||
_ => v.parse::<f64>().ok()
|
_ => v.parse::<f64>().ok(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -81,31 +81,31 @@ impl MarkedEventReceiver for YamlLoader {
|
||||||
match ev {
|
match ev {
|
||||||
Event::DocumentStart => {
|
Event::DocumentStart => {
|
||||||
// do nothing
|
// do nothing
|
||||||
},
|
}
|
||||||
Event::DocumentEnd => {
|
Event::DocumentEnd => {
|
||||||
match self.doc_stack.len() {
|
match self.doc_stack.len() {
|
||||||
// empty document
|
// empty document
|
||||||
0 => self.docs.push(Yaml::BadValue),
|
0 => self.docs.push(Yaml::BadValue),
|
||||||
1 => self.docs.push(self.doc_stack.pop().unwrap().0),
|
1 => self.docs.push(self.doc_stack.pop().unwrap().0),
|
||||||
_ => unreachable!()
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
},
|
|
||||||
Event::SequenceStart(aid) => {
|
Event::SequenceStart(aid) => {
|
||||||
self.doc_stack.push((Yaml::Array(Vec::new()), aid));
|
self.doc_stack.push((Yaml::Array(Vec::new()), aid));
|
||||||
},
|
}
|
||||||
Event::SequenceEnd => {
|
Event::SequenceEnd => {
|
||||||
let node = self.doc_stack.pop().unwrap();
|
let node = self.doc_stack.pop().unwrap();
|
||||||
self.insert_new_node(node);
|
self.insert_new_node(node);
|
||||||
},
|
}
|
||||||
Event::MappingStart(aid) => {
|
Event::MappingStart(aid) => {
|
||||||
self.doc_stack.push((Yaml::Hash(Hash::new()), aid));
|
self.doc_stack.push((Yaml::Hash(Hash::new()), aid));
|
||||||
self.key_stack.push(Yaml::BadValue);
|
self.key_stack.push(Yaml::BadValue);
|
||||||
},
|
}
|
||||||
Event::MappingEnd => {
|
Event::MappingEnd => {
|
||||||
self.key_stack.pop().unwrap();
|
self.key_stack.pop().unwrap();
|
||||||
let node = self.doc_stack.pop().unwrap();
|
let node = self.doc_stack.pop().unwrap();
|
||||||
self.insert_new_node(node);
|
self.insert_new_node(node);
|
||||||
},
|
}
|
||||||
Event::Scalar(v, style, aid, tag) => {
|
Event::Scalar(v, style, aid, tag) => {
|
||||||
let node = if style != TScalarStyle::Plain {
|
let node = if style != TScalarStyle::Plain {
|
||||||
Yaml::String(v)
|
Yaml::String(v)
|
||||||
|
@ -117,27 +117,21 @@ impl MarkedEventReceiver for YamlLoader {
|
||||||
// "true" or "false"
|
// "true" or "false"
|
||||||
match v.parse::<bool>() {
|
match v.parse::<bool>() {
|
||||||
Err(_) => Yaml::BadValue,
|
Err(_) => Yaml::BadValue,
|
||||||
Ok(v) => Yaml::Boolean(v)
|
Ok(v) => Yaml::Boolean(v),
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
"int" => {
|
"int" => match v.parse::<i64>() {
|
||||||
match v.parse::<i64>() {
|
|
||||||
Err(_) => Yaml::BadValue,
|
Err(_) => Yaml::BadValue,
|
||||||
Ok(v) => Yaml::Integer(v)
|
Ok(v) => Yaml::Integer(v),
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"float" => {
|
"float" => match parse_f64(&v) {
|
||||||
match parse_f64(&v) {
|
|
||||||
Some(_) => Yaml::Real(v),
|
Some(_) => Yaml::Real(v),
|
||||||
None => Yaml::BadValue,
|
None => Yaml::BadValue,
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"null" => {
|
"null" => match v.as_ref() {
|
||||||
match v.as_ref() {
|
|
||||||
"~" | "null" => Yaml::Null,
|
"~" | "null" => Yaml::Null,
|
||||||
_ => Yaml::BadValue,
|
_ => Yaml::BadValue,
|
||||||
}
|
},
|
||||||
}
|
|
||||||
_ => Yaml::String(v),
|
_ => Yaml::String(v),
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
@ -149,7 +143,7 @@ impl MarkedEventReceiver for YamlLoader {
|
||||||
};
|
};
|
||||||
|
|
||||||
self.insert_new_node((node, aid));
|
self.insert_new_node((node, aid));
|
||||||
},
|
}
|
||||||
Event::Alias(id) => {
|
Event::Alias(id) => {
|
||||||
let n = match self.anchor_map.get(&id) {
|
let n = match self.anchor_map.get(&id) {
|
||||||
Some(v) => v.clone(),
|
Some(v) => v.clone(),
|
||||||
|
@ -186,7 +180,7 @@ impl YamlLoader {
|
||||||
mem::swap(&mut newkey, cur_key);
|
mem::swap(&mut newkey, cur_key);
|
||||||
h.insert(newkey, node.0);
|
h.insert(newkey, node.0);
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
_ => unreachable!(),
|
_ => unreachable!(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -255,35 +249,35 @@ impl Yaml {
|
||||||
pub fn is_null(&self) -> bool {
|
pub fn is_null(&self) -> bool {
|
||||||
match *self {
|
match *self {
|
||||||
Yaml::Null => true,
|
Yaml::Null => true,
|
||||||
_ => false
|
_ => false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn is_badvalue(&self) -> bool {
|
pub fn is_badvalue(&self) -> bool {
|
||||||
match *self {
|
match *self {
|
||||||
Yaml::BadValue => true,
|
Yaml::BadValue => true,
|
||||||
_ => false
|
_ => false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn is_array(&self) -> bool {
|
pub fn is_array(&self) -> bool {
|
||||||
match *self {
|
match *self {
|
||||||
Yaml::Array(_) => true,
|
Yaml::Array(_) => true,
|
||||||
_ => false
|
_ => false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn as_f64(&self) -> Option<f64> {
|
pub fn as_f64(&self) -> Option<f64> {
|
||||||
match *self {
|
match *self {
|
||||||
Yaml::Real(ref v) => parse_f64(v),
|
Yaml::Real(ref v) => parse_f64(v),
|
||||||
_ => None
|
_ => None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn into_f64(self) -> Option<f64> {
|
pub fn into_f64(self) -> Option<f64> {
|
||||||
match self {
|
match self {
|
||||||
Yaml::Real(ref v) => parse_f64(v),
|
Yaml::Real(ref v) => parse_f64(v),
|
||||||
_ => None
|
_ => None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -315,7 +309,7 @@ impl Yaml {
|
||||||
_ if v.parse::<i64>().is_ok() => Yaml::Integer(v.parse::<i64>().unwrap()),
|
_ if v.parse::<i64>().is_ok() => Yaml::Integer(v.parse::<i64>().unwrap()),
|
||||||
// try parsing as f64
|
// try parsing as f64
|
||||||
_ if parse_f64(v).is_some() => Yaml::Real(v.to_owned()),
|
_ if parse_f64(v).is_some() => Yaml::Real(v.to_owned()),
|
||||||
_ => Yaml::String(v.to_owned())
|
_ => Yaml::String(v.to_owned()),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -328,7 +322,7 @@ impl<'a> Index<&'a str> for Yaml {
|
||||||
let key = Yaml::String(idx.to_owned());
|
let key = Yaml::String(idx.to_owned());
|
||||||
match self.as_hash() {
|
match self.as_hash() {
|
||||||
Some(h) => h.get(&key).unwrap_or(&BAD_VALUE),
|
Some(h) => h.get(&key).unwrap_or(&BAD_VALUE),
|
||||||
None => &BAD_VALUE
|
None => &BAD_VALUE,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -354,8 +348,7 @@ impl IntoIterator for Yaml {
|
||||||
|
|
||||||
fn into_iter(self) -> Self::IntoIter {
|
fn into_iter(self) -> Self::IntoIter {
|
||||||
YamlIter {
|
YamlIter {
|
||||||
yaml: self.into_vec()
|
yaml: self.into_vec().unwrap_or_else(Vec::new).into_iter(),
|
||||||
.unwrap_or_else(Vec::new).into_iter()
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -374,8 +367,8 @@ impl Iterator for YamlIter {
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod test {
|
mod test {
|
||||||
use yaml::*;
|
|
||||||
use std::f64;
|
use std::f64;
|
||||||
|
use yaml::*;
|
||||||
#[test]
|
#[test]
|
||||||
fn test_coerce() {
|
fn test_coerce() {
|
||||||
let s = "---
|
let s = "---
|
||||||
|
@ -424,8 +417,7 @@ a7: 你好
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_multi_doc() {
|
fn test_multi_doc() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
'a scalar'
|
'a scalar'
|
||||||
---
|
---
|
||||||
'a scalar'
|
'a scalar'
|
||||||
|
@ -438,8 +430,7 @@ a7: 你好
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_anchor() {
|
fn test_anchor() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
a1: &DEFAULT
|
a1: &DEFAULT
|
||||||
b1: 4
|
b1: 4
|
||||||
b2: d
|
b2: d
|
||||||
|
@ -452,8 +443,7 @@ a2: *DEFAULT
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_bad_anchor() {
|
fn test_bad_anchor() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
a1: &DEFAULT
|
a1: &DEFAULT
|
||||||
b1: 4
|
b1: 4
|
||||||
b2: *DEFAULT
|
b2: *DEFAULT
|
||||||
|
@ -461,7 +451,6 @@ a1: &DEFAULT
|
||||||
let out = YamlLoader::load_from_str(&s).unwrap();
|
let out = YamlLoader::load_from_str(&s).unwrap();
|
||||||
let doc = &out[0];
|
let doc = &out[0];
|
||||||
assert_eq!(doc["a1"]["b2"], Yaml::BadValue);
|
assert_eq!(doc["a1"]["b2"], Yaml::BadValue);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
|
@ -475,8 +464,7 @@ a1: &DEFAULT
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_plain_datatype() {
|
fn test_plain_datatype() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
- 'string'
|
- 'string'
|
||||||
- \"string\"
|
- \"string\"
|
||||||
- string
|
- string
|
||||||
|
@ -555,15 +543,23 @@ a1: &DEFAULT
|
||||||
#[test]
|
#[test]
|
||||||
fn test_bad_docstart() {
|
fn test_bad_docstart() {
|
||||||
assert!(YamlLoader::load_from_str("---This used to cause an infinite loop").is_ok());
|
assert!(YamlLoader::load_from_str("---This used to cause an infinite loop").is_ok());
|
||||||
assert_eq!(YamlLoader::load_from_str("----"), Ok(vec![Yaml::String(String::from("----"))]));
|
assert_eq!(
|
||||||
assert_eq!(YamlLoader::load_from_str("--- #here goes a comment"), Ok(vec![Yaml::Null]));
|
YamlLoader::load_from_str("----"),
|
||||||
assert_eq!(YamlLoader::load_from_str("---- #here goes a comment"), Ok(vec![Yaml::String(String::from("----"))]));
|
Ok(vec![Yaml::String(String::from("----"))])
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
YamlLoader::load_from_str("--- #here goes a comment"),
|
||||||
|
Ok(vec![Yaml::Null])
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
YamlLoader::load_from_str("---- #here goes a comment"),
|
||||||
|
Ok(vec![Yaml::String(String::from("----"))])
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_plain_datatype_with_into_methods() {
|
fn test_plain_datatype_with_into_methods() {
|
||||||
let s =
|
let s = "
|
||||||
"
|
|
||||||
- 'string'
|
- 'string'
|
||||||
- \"string\"
|
- \"string\"
|
||||||
- string
|
- string
|
||||||
|
@ -620,9 +616,18 @@ c: ~
|
||||||
let out = YamlLoader::load_from_str(&s).unwrap();
|
let out = YamlLoader::load_from_str(&s).unwrap();
|
||||||
let first = out.into_iter().next().unwrap();
|
let first = out.into_iter().next().unwrap();
|
||||||
let mut iter = first.into_hash().unwrap().into_iter();
|
let mut iter = first.into_hash().unwrap().into_iter();
|
||||||
assert_eq!(Some((Yaml::String("b".to_owned()), Yaml::Null)), iter.next());
|
assert_eq!(
|
||||||
assert_eq!(Some((Yaml::String("a".to_owned()), Yaml::Null)), iter.next());
|
Some((Yaml::String("b".to_owned()), Yaml::Null)),
|
||||||
assert_eq!(Some((Yaml::String("c".to_owned()), Yaml::Null)), iter.next());
|
iter.next()
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
Some((Yaml::String("a".to_owned()), Yaml::Null)),
|
||||||
|
iter.next()
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
Some((Yaml::String("c".to_owned()), Yaml::Null)),
|
||||||
|
iter.next()
|
||||||
|
);
|
||||||
assert_eq!(None, iter.next());
|
assert_eq!(None, iter.next());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -641,30 +646,49 @@ c: ~
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_indentation_equality() {
|
fn test_indentation_equality() {
|
||||||
|
let four_spaces = YamlLoader::load_from_str(
|
||||||
let four_spaces = YamlLoader::load_from_str(r#"
|
r#"
|
||||||
hash:
|
hash:
|
||||||
with:
|
with:
|
||||||
indentations
|
indentations
|
||||||
"#).unwrap().into_iter().next().unwrap();
|
"#,
|
||||||
|
).unwrap()
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
let two_spaces = YamlLoader::load_from_str(r#"
|
let two_spaces = YamlLoader::load_from_str(
|
||||||
|
r#"
|
||||||
hash:
|
hash:
|
||||||
with:
|
with:
|
||||||
indentations
|
indentations
|
||||||
"#).unwrap().into_iter().next().unwrap();
|
"#,
|
||||||
|
).unwrap()
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
let one_space = YamlLoader::load_from_str(r#"
|
let one_space = YamlLoader::load_from_str(
|
||||||
|
r#"
|
||||||
hash:
|
hash:
|
||||||
with:
|
with:
|
||||||
indentations
|
indentations
|
||||||
"#).unwrap().into_iter().next().unwrap();
|
"#,
|
||||||
|
).unwrap()
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
let mixed_spaces = YamlLoader::load_from_str(r#"
|
let mixed_spaces = YamlLoader::load_from_str(
|
||||||
|
r#"
|
||||||
hash:
|
hash:
|
||||||
with:
|
with:
|
||||||
indentations
|
indentations
|
||||||
"#).unwrap().into_iter().next().unwrap();
|
"#,
|
||||||
|
).unwrap()
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
assert_eq!(four_spaces, two_spaces);
|
assert_eq!(four_spaces, two_spaces);
|
||||||
assert_eq!(two_spaces, one_space);
|
assert_eq!(two_spaces, one_space);
|
||||||
|
|
|
@ -3,8 +3,8 @@ extern crate yaml_rust;
|
||||||
extern crate quickcheck;
|
extern crate quickcheck;
|
||||||
|
|
||||||
use quickcheck::TestResult;
|
use quickcheck::TestResult;
|
||||||
use yaml_rust::{Yaml, YamlLoader, YamlEmitter};
|
|
||||||
use std::error::Error;
|
use std::error::Error;
|
||||||
|
use yaml_rust::{Yaml, YamlEmitter, YamlLoader};
|
||||||
|
|
||||||
quickcheck! {
|
quickcheck! {
|
||||||
fn test_check_weird_keys(xs: Vec<String>) -> TestResult {
|
fn test_check_weird_keys(xs: Vec<String>) -> TestResult {
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
#![allow(non_upper_case_globals)]
|
#![allow(non_upper_case_globals)]
|
||||||
extern crate yaml_rust;
|
extern crate yaml_rust;
|
||||||
|
|
||||||
use yaml_rust::parser::{Parser, EventReceiver, Event};
|
use yaml_rust::parser::{Event, EventReceiver, Parser};
|
||||||
use yaml_rust::scanner::TScalarStyle;
|
use yaml_rust::scanner::TScalarStyle;
|
||||||
|
|
||||||
// These names match the names used in the C++ test suite.
|
// These names match the names used in the C++ test suite.
|
||||||
|
@ -21,7 +21,7 @@ enum TestEvent {
|
||||||
}
|
}
|
||||||
|
|
||||||
struct YamlChecker {
|
struct YamlChecker {
|
||||||
pub evs: Vec<TestEvent>
|
pub evs: Vec<TestEvent>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl EventReceiver for YamlChecker {
|
impl EventReceiver for YamlChecker {
|
||||||
|
@ -39,30 +39,30 @@ impl EventReceiver for YamlChecker {
|
||||||
} else {
|
} else {
|
||||||
TestEvent::OnScalar
|
TestEvent::OnScalar
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
Event::Alias(_) => TestEvent::OnAlias,
|
Event::Alias(_) => TestEvent::OnAlias,
|
||||||
_ => { return } // ignore other events
|
_ => return, // ignore other events
|
||||||
};
|
};
|
||||||
self.evs.push(tev);
|
self.evs.push(tev);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn str_to_test_events(docs: &str) -> Vec<TestEvent> {
|
fn str_to_test_events(docs: &str) -> Vec<TestEvent> {
|
||||||
let mut p = YamlChecker {
|
let mut p = YamlChecker { evs: Vec::new() };
|
||||||
evs: Vec::new()
|
|
||||||
};
|
|
||||||
let mut parser = Parser::new(docs.chars());
|
let mut parser = Parser::new(docs.chars());
|
||||||
parser.load(&mut p, true).unwrap();
|
parser.load(&mut p, true).unwrap();
|
||||||
p.evs
|
p.evs
|
||||||
}
|
}
|
||||||
|
|
||||||
macro_rules! assert_next {
|
macro_rules! assert_next {
|
||||||
($v:expr, $p:pat) => (
|
($v:expr, $p:pat) => {
|
||||||
match $v.next().unwrap() {
|
match $v.next().unwrap() {
|
||||||
$p => {},
|
$p => {}
|
||||||
e => { panic!("unexpected event: {:?}", e); }
|
e => {
|
||||||
|
panic!("unexpected event: {:?}", e);
|
||||||
}
|
}
|
||||||
)
|
}
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
// auto generated from handler_spec_test.cpp
|
// auto generated from handler_spec_test.cpp
|
||||||
|
@ -77,7 +77,7 @@ include!("spec_test.rs.inc");
|
||||||
#[test]
|
#[test]
|
||||||
fn test_mapvec_legal() {
|
fn test_mapvec_legal() {
|
||||||
use yaml_rust::yaml::{Array, Hash, Yaml};
|
use yaml_rust::yaml::{Array, Hash, Yaml};
|
||||||
use yaml_rust::{YamlLoader, YamlEmitter};
|
use yaml_rust::{YamlEmitter, YamlLoader};
|
||||||
|
|
||||||
// Emitting a `map<map<seq<_>>, _>` should result in legal yaml that
|
// Emitting a `map<map<seq<_>>, _>` should result in legal yaml that
|
||||||
// we can parse.
|
// we can parse.
|
||||||
|
@ -138,4 +138,3 @@ fn test_mapvec_legal() {
|
||||||
|
|
||||||
YamlLoader::load_from_str(&out_str).unwrap();
|
YamlLoader::load_from_str(&out_str).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue