// pest. The Elegant Parser // Copyright (c) 2018 DragoČ™ Tiselice // // Licensed under the Apache License, Version 2.0 // or the MIT // license , at your // option. All files in the project carrying such notice may not be copied, // modified, or distributed except according to those terms. #![doc( html_root_url = "https://docs.rs/pest_derive", html_logo_url = "https://raw.githubusercontent.com/pest-parser/pest/master/pest-logo.svg", html_favicon_url = "https://raw.githubusercontent.com/pest-parser/pest/master/pest-logo.svg" )] #![warn(missing_docs, rust_2018_idioms, unused_qualifications)] #![recursion_limit = "256"] //! # pest generator //! //! This crate generates code from ASTs (which is used in the `pest_derive` crate). #[macro_use] extern crate quote; use std::env; use std::fs::File; use std::io::{self, Read}; use std::path::Path; use proc_macro2::TokenStream; use syn::{Attribute, DeriveInput, Expr, ExprLit, Generics, Ident, Lit, Meta}; #[macro_use] mod macros; mod docs; mod generator; use pest_meta::parser::{self, rename_meta_rule, Rule}; use pest_meta::{optimizer, unwrap_or_report, validator}; /// Processes the derive/proc macro input and generates the corresponding parser based /// on the parsed grammar. If `include_grammar` is set to true, it'll generate an explicit /// "include_str" statement (done in pest_derive, but turned off in the local bootstrap). pub fn derive_parser(input: TokenStream, include_grammar: bool) -> TokenStream { let ast: DeriveInput = syn::parse2(input).unwrap(); let (name, generics, contents) = parse_derive(ast); let mut data = String::new(); let mut paths = vec![]; for content in contents { let (_data, _path) = match content { GrammarSource::File(ref path) => { let root = env::var("CARGO_MANIFEST_DIR").unwrap_or_else(|_| ".".into()); // Check whether we can find a file at the path relative to the CARGO_MANIFEST_DIR // first. // // If we cannot find the expected file over there, fallback to the // `CARGO_MANIFEST_DIR/src`, which is the old default and kept for convenience // reasons. // TODO: This could be refactored once `std::path::absolute()` get's stabilized. // https://doc.rust-lang.org/std/path/fn.absolute.html let path = if Path::new(&root).join(path).exists() { Path::new(&root).join(path) } else { Path::new(&root).join("src/").join(path) }; let file_name = match path.file_name() { Some(file_name) => file_name, None => panic!("grammar attribute should point to a file"), }; let data = match read_file(&path) { Ok(data) => data, Err(error) => panic!("error opening {:?}: {}", file_name, error), }; (data, Some(path.clone())) } GrammarSource::Inline(content) => (content, None), }; data.push_str(&_data); if let Some(path) = _path { paths.push(path); } } let pairs = match parser::parse(Rule::grammar_rules, &data) { Ok(pairs) => pairs, Err(error) => panic!("error parsing \n{}", error.renamed_rules(rename_meta_rule)), }; let defaults = unwrap_or_report(validator::validate_pairs(pairs.clone())); let doc_comment = docs::consume(pairs.clone()); let ast = unwrap_or_report(parser::consume_rules(pairs)); let optimized = optimizer::optimize(ast); generator::generate( name, &generics, paths, optimized, defaults, &doc_comment, include_grammar, ) } fn read_file>(path: P) -> io::Result { let mut file = File::open(path.as_ref())?; let mut string = String::new(); file.read_to_string(&mut string)?; Ok(string) } #[derive(Debug, PartialEq)] enum GrammarSource { File(String), Inline(String), } fn parse_derive(ast: DeriveInput) -> (Ident, Generics, Vec) { let name = ast.ident; let generics = ast.generics; let grammar: Vec<&Attribute> = ast .attrs .iter() .filter(|attr| { let path = attr.meta.path(); path.is_ident("grammar") || path.is_ident("grammar_inline") }) .collect(); if grammar.is_empty() { panic!("a grammar file needs to be provided with the #[grammar = \"PATH\"] or #[grammar_inline = \"GRAMMAR CONTENTS\"] attribute"); } let mut grammar_sources = Vec::with_capacity(grammar.len()); for attr in grammar { grammar_sources.push(get_attribute(attr)) } (name, generics, grammar_sources) } fn get_attribute(attr: &Attribute) -> GrammarSource { match &attr.meta { Meta::NameValue(name_value) => match &name_value.value { Expr::Lit(ExprLit { lit: Lit::Str(string), .. }) => { if name_value.path.is_ident("grammar") { GrammarSource::File(string.value()) } else { GrammarSource::Inline(string.value()) } } _ => panic!("grammar attribute must be a string"), }, _ => panic!("grammar attribute must be of the form `grammar = \"...\"`"), } } #[cfg(test)] mod tests { use super::parse_derive; use super::GrammarSource; #[test] fn derive_inline_file() { let definition = " #[other_attr] #[grammar_inline = \"GRAMMAR\"] pub struct MyParser<'a, T>; "; let ast = syn::parse_str(definition).unwrap(); let (_, _, filenames) = parse_derive(ast); assert_eq!(filenames, [GrammarSource::Inline("GRAMMAR".to_string())]); } #[test] fn derive_ok() { let definition = " #[other_attr] #[grammar = \"myfile.pest\"] pub struct MyParser<'a, T>; "; let ast = syn::parse_str(definition).unwrap(); let (_, _, filenames) = parse_derive(ast); assert_eq!(filenames, [GrammarSource::File("myfile.pest".to_string())]); } #[test] fn derive_multiple_grammars() { let definition = " #[other_attr] #[grammar = \"myfile1.pest\"] #[grammar = \"myfile2.pest\"] pub struct MyParser<'a, T>; "; let ast = syn::parse_str(definition).unwrap(); let (_, _, filenames) = parse_derive(ast); assert_eq!( filenames, [ GrammarSource::File("myfile1.pest".to_string()), GrammarSource::File("myfile2.pest".to_string()) ] ); } #[test] #[should_panic(expected = "grammar attribute must be a string")] fn derive_wrong_arg() { let definition = " #[other_attr] #[grammar = 1] pub struct MyParser<'a, T>; "; let ast = syn::parse_str(definition).unwrap(); parse_derive(ast); } #[test] #[should_panic( expected = "a grammar file needs to be provided with the #[grammar = \"PATH\"] or #[grammar_inline = \"GRAMMAR CONTENTS\"] attribute" )] fn derive_no_grammar() { let definition = " #[other_attr] pub struct MyParser<'a, T>; "; let ast = syn::parse_str(definition).unwrap(); parse_derive(ast); } #[doc = "Matches dar\n\nMatch dar description\n"] #[test] fn test_generate_doc() { let input = quote! { #[derive(Parser)] #[grammar = "../tests/test.pest"] pub struct TestParser; }; let token = super::derive_parser(input, true); let expected = quote! { #[doc = "A parser for JSON file.\nAnd this is a example for JSON parser.\n\n indent-4-space\n"] #[allow(dead_code, non_camel_case_types, clippy::upper_case_acronyms)] #[derive(Clone, Copy, Debug, Eq, Hash, Ord, PartialEq, PartialOrd)] pub enum Rule { #[doc = "Matches foo str, e.g.: `foo`"] r#foo, #[doc = "Matches bar str\n\n Indent 2, e.g: `bar` or `foobar`"] r#bar, r#bar1, #[doc = "Matches dar\n\nMatch dar description\n"] r#dar } }; assert!( token.to_string().contains(expected.to_string().as_str()), "{}\n\nExpected to contains:\n{}", token, expected ); } }