|
|
3 nedēļas atpakaļ | |
|---|---|---|
| .idea | 3 nedēļas atpakaļ | |
| LivePortrait-main | 3 nedēļas atpakaļ | |
| assets | 3 nedēļas atpakaļ | |
| data | 3 nedēļas atpakaļ | |
| models | 3 nedēļas atpakaļ | |
| musetalk | 3 nedēļas atpakaļ | |
| ultralight | 3 nedēļas atpakaļ | |
| wav2lip | 3 nedēļas atpakaļ | |
| web | 3 nedēļas atpakaļ | |
| Dockerfile | 3 nedēļas atpakaļ | |
| LICENSE | 3 nedēļas atpakaļ | |
| README-EN.md | 3 nedēļas atpakaļ | |
| README.md | 3 nedēļas atpakaļ | |
| app.py | 3 nedēļas atpakaļ | |
| baseasr.py | 3 nedēļas atpakaļ | |
| basereal.py | 3 nedēļas atpakaļ | |
| cuda-13-0-0-download-archive | 3 nedēļas atpakaļ | |
| docker-compose.yml | 3 nedēļas atpakaļ | |
| hubertasr.py | 3 nedēļas atpakaļ | |
| interruption_flow.py | 3 nedēļas atpakaļ | |
| knowledge_base.py | 3 nedēļas atpakaļ | |
| knowledge_intro.py | 3 nedēļas atpakaļ | |
| knowledge_intro.py.backup | 3 nedēļas atpakaļ | |
| lightreal.py | 3 nedēļas atpakaļ | |
| lipasr.py | 3 nedēļas atpakaļ | |
| lipreal.py | 3 nedēļas atpakaļ | |
| livetalking.log | 3 nedēļas atpakaļ | |
| llm.py | 3 nedēļas atpakaļ | |
| logger.py | 3 nedēļas atpakaļ | |
| main.py | 3 nedēļas atpakaļ | |
| museasr.py | 3 nedēļas atpakaļ | |
| musereal.py | 3 nedēļas atpakaļ | |
| requirements.txt | 3 nedēļas atpakaļ | |
| run_train_and_clone.sh | 3 nedēļas atpakaļ | |
| tts_ceshi.py | 3 nedēļas atpakaļ | |
| ttsreal.py | 3 nedēļas atpakaļ | |
| webrtc.py | 3 nedēļas atpakaļ |
<a href="./LICENSE"><img src="https://img.shields.io/badge/license-Apache%202-dfd.svg"></a>
<a href="https://github.com/lipku/LiveTalking/releases"><img src="https://img.shields.io/github/v/release/lipku/LiveTalking?color=ffa"></a>
<a href=""><img src="https://img.shields.io/badge/python-3.10+-aff.svg"></a>
<a href=""><img src="https://img.shields.io/badge/os-linux%2C%20win%2C%20mac-pink.svg"></a>
<a href="https://github.com/lipku/LiveTalking/graphs/contributors"><img src="https://img.shields.io/github/contributors/lipku/LiveTalking?color=c4f042&style=flat-square"></a>
<a href="https://github.com/lipku/LiveTalking/network/members"><img src="https://img.shields.io/github/forks/lipku/LiveTalking?color=8ae8ff"></a>
<a href="https://github.com/lipku/LiveTalking/stargazers"><img src="https://img.shields.io/github/stars/lipku/LiveTalking?color=ccf"></a>
A real-time interactive streaming digital human system enabling synchronized audio-video conversation, which basically meets commercial application standards.
wav2lip Demo | ernerf Demo | musetalk Demo
Domestic Mirror Repository: https://gitee.com/lipku/LiveTalking
Tested on Ubuntu 24.04, Python 3.10, PyTorch 2.5.0, and CUDA 12.4.
conda create -n nerfstream python=3.10
conda activate nerfstream
# If your CUDA version is not 12.4 (check via "nvidia-smi"), install the corresponding PyTorch version from <https://pytorch.org/get-started/previous-versions/>
conda install pytorch==2.5.0 torchvision==0.20.0 torchaudio==2.5.0 pytorch-cuda=12.4 -c pytorch -c nvidia
pip install -r requirements.txt
For common installation issues, refer to the FAQ.
For CUDA environment setup on Linux, refer to this article: https://zhuanlan.zhihu.com/p/674972886
Troubleshooting for video connection issues: https://mp.weixin.qq.com/s/MVUkxxhV2cgMMHalphr2cg
wav2lip256.pth to the models directory of this project and rename it to wav2lip.pth.Extract the wav2lip256_avatar1.tar.gz archive and copy the entire extracted folder to data/avatars of this project.
Run the Project
Execute: python app.py --transport webrtc --model wav2lip --avatar_id wav2lip256_avatar1
The server must open the following ports: TCP: 8010; UDP: 1-65536
You can access the client in two ways:
(1) Open http://serverip:8010/webrtcapi.html in a browser. First click "start" to play the digital human video; then enter any text in the input box and submit it. The digital human will broadcast the text.
(2) Use the desktop client (download link: https://pan.quark.cn/s/d7192d8ac19b).
If you cannot access Hugging Face, run the following command before starting the project:
export HF_ENDPOINT=https://hf-mirror.com
For detailed usage instructions: https://livetalking-doc.readthedocs.io/
No prior installation is required; run directly with Docker:
docker run --gpus all -it --network=host --rm registry.cn-zhangjiakou.aliyuncs.com/codewithgpu3/lipku-livetalking:toza2irpHZ
The code is located in /root/livetalking. First run git pull to fetch the latest code, then execute commands as described in Sections 2 and 3.
The following images are available:
In the backend logs, inferfps refers to the GPU inference frame rate, and finalfps refers to the final streaming frame rate. Both need to be above 25 fps to achieve real-time performance. If inferfps is above 25 but finalfps is below 25, it indicates insufficient CPU performance.
Real-Time Inference Performance
| Model | GPU Model | FPS |
|---|---|---|
| wav2lip256 | RTX 3060 | 60 |
| wav2lip256 | RTX 3080Ti | 120 |
| musetalk | RTX 3080Ti | 42 |
| musetalk | RTX 3090 | 45 |
| musetalk | RTX 4090 | 72 |
A GPU of RTX 3060 or higher is sufficient for wav2lip256, while musetalk requires an RTX 3080Ti or higher.
The following extended features are available for users who are familiar with the open-source project and need to expand product capabilities:
For more details: https://livetalking-doc.readthedocs.io/en/latest/service.html
Videos developed based on this project and published on platforms such as Bilibili, WeChat Channels, and Douyin must include the LiveTalking watermark and logo.
If this project is helpful to you, please give it a "Star". Contributions from developers interested in improving this project are also welcome.
