diff --git a/.dockerignore b/.dockerignore new file mode 100644 index 0000000000000000000000000000000000000000..ccc1bdb3075385e2d70064dacf8d775fa7cca8e0 --- /dev/null +++ b/.dockerignore @@ -0,0 +1,6 @@ +tmp +cache +Dockerfile +docker-compose.yml +.dockerignore +.gitignore diff --git a/.gitignore b/.gitignore index 17675148b94be7facb981bef0d58c6052f204cb8..33b7875466356e997310d1e072d230cc674c171e 100644 --- a/.gitignore +++ b/.gitignore @@ -1,2 +1,3 @@ __pycache__ -tmp \ No newline at end of file +tmp +cache \ No newline at end of file diff --git a/Dockerfile b/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..774d406e86f5ef792a2cb10f7e06ade556443753 --- /dev/null +++ b/Dockerfile @@ -0,0 +1,13 @@ +FROM pytorch/pytorch:2.4.0-cuda12.1-cudnn9-devel + +ENV PYTHONUNBUFFERED 1 + +WORKDIR /usr/src/app + +# Install packages +RUN apt-get update && apt-get install -y git && rm -rf /var/lib/apt/lists/* + +COPY requirements.txt ./ +RUN pip install --no-cache-dir -r requirements.txt + +COPY . . diff --git a/README.md b/README.md index 822c765b84a455c4f2f3bf2175fe664f91c6ab3e..6fdc484615c27ca5bfc6fb30d4aabdaa6a514476 100644 --- a/README.md +++ b/README.md @@ -12,6 +12,7 @@ - [Modularity](#modularity) * [Setup](#setup) * [Usage](#usage) + - [Docker Server approach](#docker-server) - [Server/Client approach](#serverclient-approach) - [Local approach](#local-approach) * [Command-line usage](#command-line-usage) @@ -56,6 +57,15 @@ The pipeline can be run in two ways: - **Server/Client approach**: Models run on a server, and audio input/output are streamed from a client. - **Local approach**: Uses the same client/server method but with the loopback address. +### Docker Server + +#### Install the NVIDIA Container Toolkit + +https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html + +#### Start the docker container +```docker compose up``` + ### Server/Client Approach To run the pipeline on the server: diff --git a/docker-compose.yml b/docker-compose.yml new file mode 100644 index 0000000000000000000000000000000000000000..ed2c9d46d3227a0c78b835cfe30f8289d3cd8a2e --- /dev/null +++ b/docker-compose.yml @@ -0,0 +1,40 @@ +--- +version: "3.8" +services: + + pipeline: + build: + context: . + command: + - python3 + - s2s_pipeline.py + - --recv_host + - 0.0.0.0 + - --send_host + - 0.0.0.0 + - --lm_model_name + - microsoft/Phi-3-mini-4k-instruct + - --init_chat_role + - system + - --init_chat_prompt + - "You are a helpful assistant" + - --stt_compile_mode + - reduce-overhead + - --tts_compile_mode + - default + expose: + - 12345/tcp + - 12346/tcp + ports: + - 12345:12345/tcp + - 12346:12346/tcp + volumes: + - ./cache/:/root/.cache/ + - ./s2s_pipeline.py:/usr/src/app/s2s_pipeline.py + deploy: + resources: + reservations: + devices: + - driver: nvidia + device_ids: ['0'] + capabilities: [gpu] diff --git a/requirements.txt b/requirements.txt index e04ebea181f5323d7ce91c7a916c1c34c6d45ef8..55985f8b594a5af97b8bc8e9aa84376e56bb643d 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,4 +1,4 @@ -nltk==3.8.1 +nltk==3.9.1 parler_tts @ git+https://github.com/huggingface/parler-tts.git torch==2.4.0 sounddevice==0.5.0