import gradio as gr lpmc_client = gr.load("seungheondoh/LP-Music-Caps-demo", src="spaces") from gradio_client import Client client = Client("https://ysharma-explore-llamav2-with-tgi.hf.space/") def infer(audio_file): cap_result = lpmc_client( audio_file, # str (filepath or URL to file) in 'audio_path' Audio component api_name="/predict" ) print(cap_result) result = client.predict( cap_result, # str in 'Message' Textbox component api_name="/chat" ) print(result) return cap_result, result with gr.Blocks() as demo: with gr.Column(elem_id="col-container"): audio_input = gr.Audio(type="filepath", source="upload") infer_btn = gr.Button("Generate") lpmc_cap = gr.Textbox(label="Lp Music Caps caption") llama_trans_cap = gr.Textbox(label="Llama translation") img_result = gr.Video(label="Result") infer_btn.click(fn=infer, inputs=[audio_input], outputs=[lpmc_cap, llama_trans_cap]) demo.queue().launch()