Skip to content

Commit

Permalink
Update endpoint in 01 tutorial
Browse files Browse the repository at this point in the history
Signed-off-by: Shaoting <[email protected]>
  • Loading branch information
Shaoting-Feng committed Mar 5, 2025
1 parent cb5ebb2 commit 27d8209
Showing 1 changed file with 17 additions and 13 deletions.
30 changes: 17 additions & 13 deletions tutorials/01-minimal-helm-installation.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,17 +6,21 @@ This tutorial guides you through a minimal setup of the vLLM Production Stack us

## Table of Contents

- [Introduction](#introduction)
- [Table of Contents](#table-of-contents)
- [Prerequisites](#prerequisites)
- [Steps](#steps)
- [1. Deploy vLLM Instance](#1-deploy-vllm-instance)
- [2. Validate Installation](#2-validate-installation)
- [3. Send a Query to the Stack](#3-send-a-query-to-the-stack)
- [3.1. Forward the Service Port](#31-forward-the-service-port)
- [3.2. Query the OpenAI-Compatible API to list the available models](#32-query-the-openai-compatible-api-to-list-the-available-models)
- [3.3. Query the OpenAI Completion Endpoint](#33-query-the-openai-completion-endpoint)
- [4. Uninstall](#4-uninstall)
- [Tutorial: Minimal Setup of the vLLM Production Stack](#tutorial-minimal-setup-of-the-vllm-production-stack)
- [Introduction](#introduction)
- [Table of Contents](#table-of-contents)
- [Prerequisites](#prerequisites)
- [Steps](#steps)
- [1. Deploy vLLM Instance](#1-deploy-vllm-instance)
- [1.1: Use Predefined Configuration](#11-use-predefined-configuration)
- [1.2: Deploy the Helm Chart](#12-deploy-the-helm-chart)
- [2. Validate Installation](#2-validate-installation)
- [2.1: Monitor Deployment Status](#21-monitor-deployment-status)
- [3. Send a Query to the Stack](#3-send-a-query-to-the-stack)
- [3.1: Forward the Service Port](#31-forward-the-service-port)
- [3.2: Query the OpenAI-Compatible API to list the available models](#32-query-the-openai-compatible-api-to-list-the-available-models)
- [3.3: Query the OpenAI Completion Endpoint](#33-query-the-openai-completion-endpoint)
- [4. Uninstall](#4-uninstall)

## Prerequisites

Expand Down Expand Up @@ -116,7 +120,7 @@ sudo kubectl port-forward svc/vllm-router-service 30080:80
Test the stack's OpenAI-compatible API by querying the available models:

```bash
curl -o- http://localhost:30080/models
curl -o- http://localhost:30080/v1/models
```

Expected output:
Expand All @@ -141,7 +145,7 @@ Expected output:
Send a query to the OpenAI `/completion` endpoint to generate a completion for a prompt:

```bash
curl -X POST http://localhost:30080/completions \
curl -X POST http://localhost:30080/v1/completions \
-H "Content-Type: application/json" \
-d '{
"model": "facebook/opt-125m",
Expand Down

0 comments on commit 27d8209

Please sign in to comment.