diff --git a/.dockerignore b/.dockerignore
new file mode 100644
index 0000000000000000000000000000000000000000..ccc1bdb3075385e2d70064dacf8d775fa7cca8e0
--- /dev/null
+++ b/.dockerignore
@@ -0,0 +1,6 @@
+tmp
+cache
+Dockerfile
+docker-compose.yml
+.dockerignore
+.gitignore
diff --git a/.gitignore b/.gitignore
index 17675148b94be7facb981bef0d58c6052f204cb8..33b7875466356e997310d1e072d230cc674c171e 100644
--- a/.gitignore
+++ b/.gitignore
@@ -1,2 +1,3 @@
 __pycache__
-tmp
\ No newline at end of file
+tmp
+cache
\ No newline at end of file
diff --git a/Dockerfile b/Dockerfile
new file mode 100644
index 0000000000000000000000000000000000000000..774d406e86f5ef792a2cb10f7e06ade556443753
--- /dev/null
+++ b/Dockerfile
@@ -0,0 +1,13 @@
+FROM pytorch/pytorch:2.4.0-cuda12.1-cudnn9-devel
+
+ENV PYTHONUNBUFFERED 1
+
+WORKDIR /usr/src/app
+
+# Install packages
+RUN apt-get update && apt-get install -y git && rm -rf /var/lib/apt/lists/*
+
+COPY requirements.txt ./
+RUN pip install --no-cache-dir -r requirements.txt
+
+COPY . .
diff --git a/README.md b/README.md
index 822c765b84a455c4f2f3bf2175fe664f91c6ab3e..6fdc484615c27ca5bfc6fb30d4aabdaa6a514476 100644
--- a/README.md
+++ b/README.md
@@ -12,6 +12,7 @@
   - [Modularity](#modularity)
 * [Setup](#setup)
 * [Usage](#usage)
+  - [Docker Server approach](#docker-server)
   - [Server/Client approach](#serverclient-approach)
   - [Local approach](#local-approach)
 * [Command-line usage](#command-line-usage)
@@ -56,6 +57,15 @@ The pipeline can be run in two ways:
 - **Server/Client approach**: Models run on a server, and audio input/output are streamed from a client.
 - **Local approach**: Uses the same client/server method but with the loopback address.
 
+### Docker Server
+
+#### Install the NVIDIA Container Toolkit
+
+https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html
+
+#### Start the docker container
+```docker compose up```
+
 ### Server/Client Approach
 
 To run the pipeline on the server:
diff --git a/docker-compose.yml b/docker-compose.yml
new file mode 100644
index 0000000000000000000000000000000000000000..ed2c9d46d3227a0c78b835cfe30f8289d3cd8a2e
--- /dev/null
+++ b/docker-compose.yml
@@ -0,0 +1,40 @@
+---
+version: "3.8"
+services:
+
+  pipeline:
+    build:
+      context: .
+    command: 
+      - python3 
+      - s2s_pipeline.py 
+      - --recv_host 
+      - 0.0.0.0 
+      - --send_host 
+      - 0.0.0.0 
+      - --lm_model_name 
+      - microsoft/Phi-3-mini-4k-instruct 
+      - --init_chat_role 
+      - system
+      - --init_chat_prompt
+      - "You are a helpful assistant"
+      - --stt_compile_mode 
+      - reduce-overhead 
+      - --tts_compile_mode 
+      - default
+    expose:
+      - 12345/tcp
+      - 12346/tcp
+    ports:
+      - 12345:12345/tcp
+      - 12346:12346/tcp
+    volumes:
+      - ./cache/:/root/.cache/
+      - ./s2s_pipeline.py:/usr/src/app/s2s_pipeline.py
+    deploy:
+      resources:
+        reservations:
+          devices:
+            - driver: nvidia
+              device_ids: ['0']
+              capabilities: [gpu]
diff --git a/requirements.txt b/requirements.txt
index e04ebea181f5323d7ce91c7a916c1c34c6d45ef8..55985f8b594a5af97b8bc8e9aa84376e56bb643d 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -1,4 +1,4 @@
-nltk==3.8.1
+nltk==3.9.1
 parler_tts @ git+https://github.com/huggingface/parler-tts.git
 torch==2.4.0
 sounddevice==0.5.0