Doc clean up
This commit is contained in:
parent
d4dee92904
commit
f7cedb3536
2 changed files with 7 additions and 6 deletions
|
@ -1,6 +1,6 @@
|
|||
/// Character Sets
|
||||
///
|
||||
/// Sets of characters valid for making up tokens.
|
||||
/* charset.rs
|
||||
* Eryn Wells <eryn@erynwells.me>
|
||||
*/
|
||||
|
||||
use std::collections::HashSet;
|
||||
use std::iter::FromIterator;
|
||||
|
|
|
@ -1,4 +1,6 @@
|
|||
//! # Lexer
|
||||
/* lexer.rs
|
||||
* Eryn Wells <eryn@erynwells.me>
|
||||
*/
|
||||
|
||||
pub mod token;
|
||||
mod char;
|
||||
|
@ -58,8 +60,7 @@ impl Lexer {
|
|||
println!("> begin={}", self.begin);
|
||||
}
|
||||
|
||||
/// Get the substring between the two input indexes. This is the value to give to a new Token
|
||||
/// instance.
|
||||
/// Get the substring between the two input indexes. This is the value to give to a new Token instance.
|
||||
fn value(&self) -> String {
|
||||
self.input[self.begin .. self.forward].to_string()
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue