Spaces:
Sleeping
Sleeping
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline,EncoderDecoderConfig | |
tokenizer = AutoTokenizer.from_pretrained("codellama/CodeLlama-34b-Instruct-hf") | |
model = EncoderDecoderConfig.from_pretrained("codellama/CodeLlama-34b-Instruct-hf") | |
fix_pipeline = pipeline("text2text-generation", model=model, tokenizer=tokenizer) | |
def fix_code(code: str) -> str: | |
prompt = f"""The following Python code has some bugs. Fix the code:\n\n{code}\n\nCorrected version:""" | |
try: | |
result = fix_pipeline(prompt, max_length=256)[0]["generated_text"] | |
return result.strip() | |
except Exception as e: | |
return f"Error during fixing: {e}" |