Spaces:
Runtime error
Runtime error
| import os | |
| from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM | |
| from bark import SAMPLE_RATE, generate_audio, preload_models | |
| from scipy.io.wavfile import write as write_wav | |
| # Load local models for LLM and TTS | |
| preload_models() | |
| # Load the Hugging Face model locally | |
| tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neo-2.7B") | |
| model = AutoModelForCausalLM.from_pretrained("EleutherAI/gpt-neo-2.7B") | |
| generator = pipeline("text-generation", model=model, tokenizer=tokenizer) | |
| def generate_script(system_prompt: str, input_text: str, output_model=None): | |
| """Generate dialogue script using local GPT-Neo model.""" | |
| try: | |
| # Generate text using the local Hugging Face model | |
| response = generator(system_prompt + input_text, max_length=1000, do_sample=True) | |
| dialogue = response[0]["generated_text"] | |
| return dialogue | |
| except Exception as e: | |
| return f"Error generating script: {str(e)}" | |
| def generate_podcast_audio(text: str, output_file="podcast_audio.wav"): | |
| """Generate audio from text using Bark (local TTS).""" | |
| try: | |
| audio_array = generate_audio(text) | |
| write_wav(output_file, SAMPLE_RATE, audio_array) | |
| return output_file | |
| except Exception as e: | |
| return f"Error generating audio: {str(e)}" | |