File tree Expand file tree Collapse file tree 1 file changed +3
-0
lines changed Expand file tree Collapse file tree 1 file changed +3
-0
lines changed Original file line number Diff line number Diff line change 22This is a simple fastapi based server mock that implements the OpenAI API.
33
44Available endpoints:
5+ 
56-  /v1/chat/completion
67
78Instead of running a LLM model to generate completions, it simply returns a response generated by surrogate models. Available surrogate models are:
9+ 
810-  "yes_no": returns random "Yes" or "No" response
911-  "ja_nein": returns random "Ja" or "Nein" response
1012-  "lorem_ipsum": returns random "lorem ipsum" text
@@ -15,6 +17,7 @@ docker pull ghcr.io/hummerichsander/openai_api_server_mock:v ... # replace ... w
1517``` 
1618
1719Environment variables:
20+ 
1821-  ` CONTEXT_SIZE ` : context size for the model (default: 4096)
1922-  ` SLEEP_TIME ` : sleep time in seconds before returning the response (default: 0)
2023-  ` MAX_CONCURRENT_REQUESTS ` : maximum number of concurrent requests (default: 10^9)
    
 
   
 
     
   
   
          
     
  
    
     
 
    
      
     
 
     
    You can’t perform that action at this time.
  
 
    
  
     
    
      
        
     
 
       
      
     
   
 
    
    
  
 
  
 
     
    
0 commit comments