|
| 1 | +name: 🐛 Bug report |
| 2 | +description: Raise an issue here if you find a bug. |
| 3 | +title: "[Bug]: " |
| 4 | +labels: ["bug"] |
| 5 | + |
| 6 | +body: |
| 7 | +- type: markdown |
| 8 | + attributes: |
| 9 | + value: > |
| 10 | + #### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm/issues?q=is%3Aissue+sort%3Acreated-desc+). |
| 11 | +- type: textarea |
| 12 | + attributes: |
| 13 | + label: Your current environment |
| 14 | + description: | |
| 15 | + Please run the following and paste the output below. |
| 16 | + ```sh |
| 17 | + wget https://raw.githubusercontent.com/vllm-project/vllm/main/collect_env.py |
| 18 | + # For security purposes, please feel free to check the contents of collect_env.py before running it. |
| 19 | + python collect_env.py |
| 20 | + ``` |
| 21 | + value: | |
| 22 | + ```text |
| 23 | + The output of `python collect_env.py` |
| 24 | + ``` |
| 25 | + validations: |
| 26 | + required: true |
| 27 | +- type: textarea |
| 28 | + attributes: |
| 29 | + label: 🐛 Describe the bug |
| 30 | + description: | |
| 31 | + Please provide a clear and concise description of what the bug is. |
| 32 | +
|
| 33 | + If relevant, add a minimal example so that we can reproduce the error by running the code. It is very important for the snippet to be as succinct (minimal) as possible, so please take time to trim down any irrelevant code to help us debug efficiently. We are going to copy-paste your code and we expect to get the same result as you did: avoid any external data, and include the relevant imports, etc. For example: |
| 34 | +
|
| 35 | + ```python |
| 36 | + from vllm import LLM, SamplingParams |
| 37 | +
|
| 38 | + prompts = [ |
| 39 | + "Hello, my name is", |
| 40 | + "The president of the United States is", |
| 41 | + "The capital of France is", |
| 42 | + "The future of AI is", |
| 43 | + ] |
| 44 | + sampling_params = SamplingParams(temperature=0.8, top_p=0.95) |
| 45 | +
|
| 46 | + llm = LLM(model="facebook/opt-125m") |
| 47 | +
|
| 48 | + outputs = llm.generate(prompts, sampling_params) |
| 49 | +
|
| 50 | + # Print the outputs. |
| 51 | + for output in outputs: |
| 52 | + prompt = output.prompt |
| 53 | + generated_text = output.outputs[0].text |
| 54 | + print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}") |
| 55 | + ``` |
| 56 | +
|
| 57 | + If the code is too long (hopefully, it isn't), feel free to put it in a public gist and link it in the issue: https://gist.github.com. |
| 58 | +
|
| 59 | + Please also paste or describe the results you observe instead of the expected results. If you observe an error, please paste the error message including the **full** traceback of the exception. It may be relevant to wrap error messages in ```` ```triple quotes blocks``` ````. |
| 60 | + placeholder: | |
| 61 | + A clear and concise description of what the bug is. |
| 62 | +
|
| 63 | + ```python |
| 64 | + # Sample code to reproduce the problem |
| 65 | + ``` |
| 66 | +
|
| 67 | + ``` |
| 68 | + The error message you got, with the full traceback. |
| 69 | + ``` |
| 70 | + validations: |
| 71 | + required: true |
| 72 | +- type: markdown |
| 73 | + attributes: |
| 74 | + value: > |
| 75 | + ⚠️ Please separate bugs of `transformers` implementation or usage from bugs of `vllm`. If you think anything is wrong with the models' output: |
| 76 | +
|
| 77 | + - Try the counterpart of `transformers` first. If the error appears, please go to [their issues](https://github.com/huggingface/transformers/issues?q=is%3Aissue+is%3Aopen+sort%3Aupdated-desc). |
| 78 | +
|
| 79 | + - If the error only appears in vllm, please provide the detailed script of how you run `transformers` and `vllm`, also highlight the difference and what you expect. |
| 80 | +
|
| 81 | + Thanks for contributing 🎉! |
0 commit comments