from transformers import GPT2LMHeadModel, GPT2TokenizerFast # Load tokenizer and model model_path = "." # Path where you saved your model tokenizer = GPT2TokenizerFast.from_pretrained(model_path) model = GPT2LMHeadModel.from_pretrained(model_path) # Move model to GPU if available import torch device = torch.device("cuda" if torch.cuda.is_available() else "cpu") model.to(device) # Generate text from a prompt prompt = "Once upon a time" inputs = tokenizer(prompt, return_tensors="pt").to(device) outputs = model.generate( inputs.input_ids, max_length=1024, num_return_sequences=1, do_sample=True, top_k=50, top_p=0.95, temperature=0.8, pad_token_id=tokenizer.pad_token_id ) # Decode and print generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True) print(generated_text)