mirror of
				https://github.com/python/cpython.git
				synced 2025-11-04 03:44:55 +00:00 
			
		
		
		
	Open and close files manually. It prevents from leaking files, preliminary creation of output files, and accidental closing of stdin and stdout.
		
			
				
	
	
		
			73 lines
		
	
	
	
		
			2.1 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			73 lines
		
	
	
	
		
			2.1 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
"""Generate Lib/keyword.py from the Grammar and Tokens files using pgen"""
 | 
						|
 | 
						|
import argparse
 | 
						|
 | 
						|
from .build import build_parser, generate_token_definitions
 | 
						|
from .c_generator import CParserGenerator
 | 
						|
 | 
						|
TEMPLATE = r'''
 | 
						|
"""Keywords (from "Grammar/python.gram")
 | 
						|
 | 
						|
This file is automatically generated; please don't muck it up!
 | 
						|
 | 
						|
To update the symbols in this file, 'cd' to the top directory of
 | 
						|
the python source tree and run:
 | 
						|
 | 
						|
    PYTHONPATH=Tools/peg_generator python3 -m pegen.keywordgen \
 | 
						|
        Grammar/python.gram \
 | 
						|
        Grammar/Tokens \
 | 
						|
        Lib/keyword.py
 | 
						|
 | 
						|
Alternatively, you can run 'make regen-keyword'.
 | 
						|
"""
 | 
						|
 | 
						|
__all__ = ["iskeyword", "issoftkeyword", "kwlist", "softkwlist"]
 | 
						|
 | 
						|
kwlist = [
 | 
						|
{keywords}
 | 
						|
]
 | 
						|
 | 
						|
softkwlist = [
 | 
						|
{soft_keywords}
 | 
						|
]
 | 
						|
 | 
						|
iskeyword = frozenset(kwlist).__contains__
 | 
						|
issoftkeyword = frozenset(softkwlist).__contains__
 | 
						|
'''.lstrip()
 | 
						|
 | 
						|
 | 
						|
def main() -> None:
 | 
						|
    parser = argparse.ArgumentParser(
 | 
						|
        description="Generate the Lib/keywords.py file from the grammar."
 | 
						|
    )
 | 
						|
    parser.add_argument(
 | 
						|
        "grammar", help="The file with the grammar definition in PEG format"
 | 
						|
    )
 | 
						|
    parser.add_argument(
 | 
						|
        "tokens_file", help="The file with the token definitions"
 | 
						|
    )
 | 
						|
    parser.add_argument(
 | 
						|
        "keyword_file",
 | 
						|
        help="The path to write the keyword definitions",
 | 
						|
    )
 | 
						|
    args = parser.parse_args()
 | 
						|
 | 
						|
    grammar, _, _ = build_parser(args.grammar)
 | 
						|
    with open(args.tokens_file) as tok_file:
 | 
						|
        all_tokens, exact_tok, non_exact_tok = generate_token_definitions(tok_file)
 | 
						|
    gen = CParserGenerator(grammar, all_tokens, exact_tok, non_exact_tok, file=None)
 | 
						|
    gen.collect_rules()
 | 
						|
 | 
						|
    with open(args.keyword_file, 'w') as thefile:
 | 
						|
        all_keywords = sorted(list(gen.keywords.keys()))
 | 
						|
        all_soft_keywords = sorted(gen.soft_keywords)
 | 
						|
 | 
						|
        keywords = "" if not all_keywords else "    " + ",\n    ".join(map(repr, all_keywords))
 | 
						|
        soft_keywords = (
 | 
						|
            "" if not all_soft_keywords else "    " + ",\n    ".join(map(repr, all_soft_keywords))
 | 
						|
        )
 | 
						|
        thefile.write(TEMPLATE.format(keywords=keywords, soft_keywords=soft_keywords))
 | 
						|
 | 
						|
 | 
						|
if __name__ == "__main__":
 | 
						|
    main()
 |