Skip to content

Commit 08f5115

Browse files
Refactor README to include available endpoints and environment variables
1 parent d30d1e4 commit 08f5115

File tree

1 file changed

+3
-0
lines changed

1 file changed

+3
-0
lines changed

README.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2,9 +2,11 @@
22
This is a simple fastapi based server mock that implements the OpenAI API.
33

44
Available endpoints:
5+
56
- /v1/chat/completion
67

78
Instead of running a LLM model to generate completions, it simply returns a response generated by surrogate models. Available surrogate models are:
9+
810
- "yes_no": returns random "Yes" or "No" response
911
- "ja_nein": returns random "Ja" or "Nein" response
1012
- "lorem_ipsum": returns random "lorem ipsum" text
@@ -15,6 +17,7 @@ docker pull ghcr.io/hummerichsander/openai_api_server_mock:v ... # replace ... w
1517
```
1618

1719
Environment variables:
20+
1821
- `CONTEXT_SIZE`: context size for the model (default: 4096)
1922
- `SLEEP_TIME`: sleep time in seconds before returning the response (default: 0)
2023
- `MAX_CONCURRENT_REQUESTS`: maximum number of concurrent requests (default: 10^9)

0 commit comments

Comments
 (0)