Skip to content

Commit 13dc664

Browse files
committed
Document Sync by Tina
1 parent 07ad203 commit 13dc664

File tree

5 files changed

+12
-12
lines changed

5 files changed

+12
-12
lines changed

docs/stable/getting_started/docker_quickstart.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -95,7 +95,7 @@ Open a new terminal, activate the `sllm` environment, and set the `LLM_SERVER_UR
9595

9696
```bash
9797
conda activate sllm
98-
export LLM_SERVER_URL=http://localhost:8343/
98+
export LLM_SERVER_URL=http://127.0.0.1:8343/
9999
```
100100

101101
Deploy a model to the ServerlessLLM server using the `sllm-cli`:
@@ -117,7 +117,7 @@ INFO 08-01 07:39:00 deploy.py:49] Model registered successfully.
117117

118118
Now, you can query the model by any OpenAI API client. For example, you can use the following Python code to query the model:
119119
```bash
120-
curl http://localhost:8343/v1/chat/completions \
120+
curl http://127.0.0.1:8343/v1/chat/completions \
121121
-H "Content-Type: application/json" \
122122
-d '{
123123
"model": "facebook/opt-1.3b",

docs/stable/getting_started/installation.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,13 +24,13 @@ cd ServerlessLLM
2424
conda create -n sllm python=3.10 -y
2525
conda activate sllm
2626
pip install -e .
27-
pip install -i https://test.pypi.org/simple/ --extra-index-url https://pypi.org/simple/ serverless_llm_store==0.0.1.dev4
27+
pip install -i https://test.pypi.org/simple/ --extra-index-url https://pypi.org/simple/ serverless_llm_store==0.0.1.dev5
2828
2929
# worker node
3030
conda create -n sllm-worker python=3.10 -y
3131
conda activate sllm-worker
3232
pip install -e ".[worker]"
33-
pip install -i https://test.pypi.org/simple/ --extra-index-url https://pypi.org/simple/ serverless_llm_store==0.0.1.dev4
33+
pip install -i https://test.pypi.org/simple/ --extra-index-url https://pypi.org/simple/ serverless_llm_store==0.0.1.dev5
3434
```
3535

3636
# vLLM Patch

docs/stable/getting_started/quickstart.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ ray start --head --port=6379 --num-cpus=4 --num-gpus=0 \
1919
In a new terminal, start the worker node:
2020
```bash
2121
conda activate sllm-worker
22-
ray start --address=localhost:6379 --num-cpus=4 --num-gpus=2 \
22+
ray start --address=0.0.0.0:6379 --num-cpus=4 --num-gpus=2 \
2323
--resources='{"worker_node": 1, "worker_id_0": 1}' --block
2424
```
2525

@@ -67,7 +67,7 @@ sllm-cli deploy --model facebook/opt-1.3b
6767

6868
Now, you can query the model by any OpenAI API client. For example, you can use the following Python code to query the model:
6969
```bash
70-
curl http://localhost:8343/v1/chat/completions \
70+
curl http://127.0.0.1:8343/v1/chat/completions \
7171
-H "Content-Type: application/json" \
7272
-d '{
7373
"model": "facebook/opt-1.3b",

docs/stable/serve/storage_aware_scheduling.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -136,7 +136,7 @@ echo '{
136136

137137
```bash
138138
conda activate sllm
139-
export LLM_SERVER_URL=http://localhost:8343/
139+
export LLM_SERVER_URL=http://127.0.0.1:8343/
140140

141141
sllm-cli deploy --config config-opt-2.7b.json
142142
sllm-cli deploy --config config-opt-1.3b.json
@@ -145,7 +145,7 @@ sllm-cli deploy --config config-opt-1.3b.json
145145
3. Verify the deployment.
146146

147147
```bash
148-
curl http://localhost:8343/v1/chat/completions \
148+
curl http://127.0.0.1:8343/v1/chat/completions \
149149
-H "Content-Type: application/json" \
150150
-d '{
151151
"model": "opt-2.7b",
@@ -155,7 +155,7 @@ curl http://localhost:8343/v1/chat/completions \
155155
]
156156
}'
157157

158-
curl http://localhost:8343/v1/chat/completions \
158+
curl http://127.0.0.1:8343/v1/chat/completions \
159159
-H "Content-Type: application/json" \
160160
-d '{
161161
"model": "opt-1.3b",

docs/stable/store/quickstart.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ conda activate sllm-store
2626

2727
### Install with pip
2828
```bash
29-
pip install -i https://test.pypi.org/simple/ --extra-index-url https://pypi.org/simple/ serverless_llm_store==0.0.1.dev4
29+
pip install -i https://test.pypi.org/simple/ --extra-index-url https://pypi.org/simple/ serverless_llm_store==0.0.1.dev5
3030
```
3131

3232
### Install from source
@@ -69,7 +69,7 @@ save_model(model, './models/facebook/opt-1.3b')
6969
2. Launch the checkpoint store server in a separate process:
7070
```bash
7171
# 'mem_pool_size' is the maximum size of the memory pool in GB. It should be larger than the model size.
72-
sllm-store-server --storage_path $PWD/models --mem_pool_size 32
72+
sllm-store-server --storage_path $PWD/models --mem_pool_size 4
7373
```
7474

7575
<!-- Running the server using a container:
@@ -224,7 +224,7 @@ After downloading the model, you can launch the checkpoint store server and load
224224
2. Launch the checkpoint store server in a separate process:
225225
```bash
226226
# 'mem_pool_size' is the maximum size of the memory pool in GB. It should be larger than the model size.
227-
sllm-store-server --storage_path $PWD/models --mem_pool_size 32
227+
sllm-store-server --storage_path $PWD/models --mem_pool_size 4
228228
```
229229

230230
3. Load the model in vLLM:

0 commit comments

Comments
 (0)