@@ -593,6 +593,39 @@ export const keras_hub = (model: ModelData): string[] => {
593
593
return snippets ;
594
594
} ;
595
595
596
+ export const kimi_audio = ( model : ModelData ) : string [ ] => [
597
+ `# Example usage for KimiAudio
598
+ # pip install git+https://github.com/MoonshotAI/Kimi-Audio.git
599
+
600
+ from kimia_infer.api.kimia import KimiAudio
601
+
602
+ model = KimiAudio(model_path="${ model . id } ", load_detokenizer=True)
603
+
604
+ sampling_params = {
605
+ "audio_temperature": 0.8,
606
+ "audio_top_k": 10,
607
+ "text_temperature": 0.0,
608
+ "text_top_k": 5,
609
+ }
610
+
611
+ # For ASR
612
+ asr_audio = "asr_example.wav"
613
+ messages_asr = [
614
+ {"role": "user", "message_type": "text", "content": "Please transcribe the following audio:"},
615
+ {"role": "user", "message_type": "audio", "content": asr_audio}
616
+ ]
617
+ _, text = model.generate(messages_asr, **sampling_params, output_type="text")
618
+ print(text)
619
+
620
+ # For Q&A
621
+ qa_audio = "qa_example.wav"
622
+ messages_conv = [{"role": "user", "message_type": "audio", "content": qa_audio}]
623
+ wav, text = model.generate(messages_conv, **sampling_params, output_type="both")
624
+ sf.write("output_audio.wav", wav.cpu().view(-1).numpy(), 24000)
625
+ print(text)
626
+ ` ,
627
+ ] ;
628
+
596
629
export const lightning_ir = ( model : ModelData ) : string [ ] => {
597
630
if ( model . tags . includes ( "bi-encoder" ) ) {
598
631
return [
0 commit comments