|
top_md_1 = (""" |
|
<div align="center"> |
|
<div style="display:flex; gap: 0.25rem;" align="center"> |
|
FunClip: <a href='https://github.com/alibaba-damo-academy/FunClip'><img src='https://img.shields.io/badge/Github-Code-blue'></a> |
|
🌟支持我们: <a href='https://github.com/alibaba-damo-academy/FunClip/stargazers'><img src='https://img.shields.io/github/stars/alibaba-damo-academy/FunClip.svg?style=social'></a> |
|
</div> |
|
</div> |
|
|
|
基于阿里巴巴通义实验室自研并开源的[FunASR](https://github.com/alibaba-damo-academy/FunASR)工具包及Paraformer系列模型及语音识别、端点检测、标点预测、时间戳预测、说话人区分、热词定制化开源链路 |
|
|
|
准确识别,自由复制所需段落,或者设置说话人标识,一键裁剪、添加字幕 |
|
|
|
* Step1: 上传视频或音频文件(或使用下方的用例体验),点击 **<font color="#f7802b">识别</font>** 按钮 |
|
* Step2: 复制识别结果中所需的文字至右上方,或者右设置说话人标识,设置偏移与字幕配置(可选) |
|
* Step3: 点击 **<font color="#f7802b">裁剪</font>** 按钮或 **<font color="#f7802b">裁剪并添加字幕</font>** 按钮获得结果 |
|
|
|
🔥 FunClip现在集成了大语言模型智能剪辑功能,选择LLM模型进行体验吧~ |
|
""") |
|
|
|
top_md_2 = ("通过Modelscope创空间的访问可能由于多人同时使用的原因不稳定,推荐通过github项目中的代码自己部署。") |
|
|
|
top_md_3 = ("""访问FunASR项目与论文能够帮助您深入了解ParaClipper中所使用的语音处理相关模型: |
|
<div align="center"> |
|
<div style="display:flex; gap: 0.25rem;" align="center"> |
|
FunASR: <a href='https://github.com/alibaba-damo-academy/FunASR'><img src='https://img.shields.io/badge/Github-Code-blue'></a> |
|
FunASR Paper: <a href="https://arxiv.org/abs/2305.11013"><img src="https://img.shields.io/badge/Arxiv-2305.11013-orange"></a> |
|
🌟Star FunASR: <a href='https://github.com/alibaba-damo-academy/FunASR/stargazers'><img src='https://img.shields.io/github/stars/alibaba-damo-academy/FunASR.svg?style=social'></a> |
|
</div> |
|
</div> |
|
""") |
|
|
|
top_md_4 = ("""我们在「LLM智能裁剪」模块中提供三种LLM调用方式, |
|
1. 选择阿里云百炼平台通过api调用qwen系列模型,此时需要您准备百炼平台的apikey,请访问[阿里云百炼](https://bailian.console.aliyun.com/#/home); |
|
2. 选择GPT开头的模型即为调用openai官方api,此时需要您自备sk与网络环境; |
|
3. [gpt4free](https://github.com/xtekky/gpt4free?tab=readme-ov-file)项目也被集成进FunClip,可以通过它免费调用gpt模型; |
|
|
|
其中方式1与方式2需要在界面中传入相应的apikey |
|
方式3而可能非常不稳定,返回时间可能很长或者结果获取失败,可以多多尝试或者自己准备sk使用方式1 |
|
""") |
|
|
|
top_md_1_en = (""" |
|
Powered by Alibaba DAMO Academy's in-house developed and open-sourced [FunASR](https://github.com/alibaba-damo-academy/FunASR) toolkit and the Paraformer series models, offering open link to voice recognition, endpoint detection, punctuation prediction, timestamp prediction, speaker differentiation, and customization of hot words. |
|
Accurate recognition, freely copy the required paragraphs, or set speaker labels, one-click clipping, adding subtitles |
|
* Step1: Upload a video or audio file (or use the case below for a demo), then click **<font color="#f7802b">Recognize</font>** button |
|
* Step2: Copy the required text from the recognition results to the top right, or set speaking indicators on the right and configure offset and subtitle settings (optional) |
|
* Step3: Click the **<font color="#f7802b">Clip</font>** button or **<font color="#f7802b">Clip and Add Subtitles</font>** button to get the result |
|
|
|
🔥 FunClip now integrates the smart editing functionality with large language models, choose an LLM model for a try~ |
|
""") |
|
|
|
top_md_2_en = (""" |
|
Access to the HuggingFace space may be unstable due to simultaneous use by multiple users. It is recommended to deploy the code through the GitHub project yourself. |
|
<div align="center"> |
|
<div style="display:flex; gap: 0.25rem;" align="center"> |
|
FunClip: <a href='https://github.com/alibaba-damo-academy/FunClip'><img src='https://img.shields.io/badge/Github-Code-blue'></a> |
|
🌟 Support Us: <a href='https://github.com/alibaba-damo-academy/FunClip/stargazers'><img src='https://img.shields.io/github/stars/alibaba-damo-academy/FunClip.svg?style=social'></a> |
|
</div> |
|
</div> |
|
""") |
|
|
|
top_md_3_en = ("""Visiting the FunASR project and paper can help you gain an in-depth understanding of the speech processing models used in ParaClipper: |
|
<div align="center"> |
|
<div style="display:flex; gap: 0.25rem;" align="center"> |
|
FunASR: <a href='https://github.com/alibaba-damo-academy/FunASR'><img src='https://img.shields.io/badge/Github-Code-blue'></a> |
|
FunASR Paper: <a href="https://arxiv.org/abs/2305.11013"><img src="https://img.shields.io/badge/Arxiv-2305.11013-orange"></a> |
|
🌟Star FunASR: <a href='https://github.com/alibaba-damo-academy/FunASR/stargazers'><img src='https://img.shields.io/github/stars/alibaba-damo-academy/FunASR.svg?style=social'></a> |
|
</div> |
|
</div> |
|
""") |
|
|
|
top_md_4_en = ("""We provide three ways to use LLMs in the 'LLM Clipping' module: |
|
1. Choose Alibaba Cloud's BaiLian platform to call qwen series models via API, for which you need to prepare the BaiLian platform's apikey. Please visit [Alibaba Cloud BaiLian](https://bailian.console.aliyun.com/#/home); |
|
2. Choosing models that start with GPT denotes the use of the official OpenAI API, for which you need to provide your own sk and network environment; |
|
3. The [gpt4free](https://github.com/xtekky/gpt4free?tab=readme-ov-file) project is also integrated into FunClip, allowing free use of the gpt model; |
|
|
|
Both methods 1 and 2 require you to enter the respective apikey in the interface. |
|
Method 3 may be highly unstable, with potential long return times or failure to retrieve results. You may try multiple times or prepare your own sk to use method 1. |
|
""") |
|
|