pub fn tokenize(
    input: &str,
    frontmatter_allowed: FrontmatterAllowed,
) -> impl Iterator<Item = Token>Expand description
Creates an iterator that produces tokens from the input string.
When parsing a full Rust document,
first strip_shebang and then allow frontmatters with FrontmatterAllowed::Yes.
When tokenizing a slice of a document, be sure to disallow frontmatters with FrontmatterAllowed::No