Spaces:
Runtime error
Runtime error
| from arguments import InitializationArguments | |
| from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser | |
| # Configuration | |
| parser = HfArgumentParser(InitializationArguments) | |
| args = parser.parse_args() | |
| # Load codeparrot tokenizer trained for Python code tokenization | |
| tokenizer = AutoTokenizer.from_pretrained(args.tokenizer_name) | |
| # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks | |
| config_kwargs = { | |
| "vocab_size": len(tokenizer), | |
| "scale_attn_by_inverse_layer_idx": True, | |
| "reorder_and_upcast_attn": True, | |
| } | |
| # Load model config (GPT-2 large in this case) | |
| config = AutoConfig.from_pretrained(args.config_name, **config_kwargs) | |
| # Initialize new model with config | |
| model = AutoModelForCausalLM.from_config(config) | |
| # Save model to the hub | |
| model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub) | |