Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Token-level Probability Always 0.0 When Fine-tuning Llama2-7b Model on Single GPU #3979

Closed
MoOo2mini opened this issue Apr 2, 2024 · 1 comment
Labels
llm Large Language Model related

Comments

@MoOo2mini
Copy link

MoOo2mini commented Apr 2, 2024

Describe the bug
The token-level probabilities consistently appear as 0.0 when fine-tuning the Llama2-7b model using "Ludwig + DeepLearning.ai: Efficient Fine-Tuning for Llama2-7b on a Single GPU.ipynb".
https://colab.research.google.com/drive/1Ly01S--kUwkKQalE-75skalp-ftwl0fE?usp=sharing

below thing is my code that has a problem...
https://colab.research.google.com/drive/1OmbCKlPzlxm4__iThYqB9PSLUWZZVptz?usp=sharing

To Reproduce
Steps to reproduce the behavior:

  1. Fine-tune the Llama2-7b model using the provided notebook.
  2. Execute the model's predictions using the predict function with modified parameters, including setting skip_save_unprocessed_output to False and providing a specific output_directory.
  3. Despite modifications, the token-level probabilities remain 0.0.
ludwig.predict(
  dataset=None,
  data_format=None,
  split='full',
  batch_size=128,
  skip_save_unprocessed_output=True,
  skip_save_predictions=True,
  output_directory='results',
  return_type=<class 'pandas.core.frame.DataFrame'>,
  debug=False
)

Expected behavior
Token-level probabilities should reflect the model's confidence in predicting each token's output.

Screenshots
N/A

Environment:

  • OS: Ubuntu 20.04
  • Python version: 3.8.10
  • Ludwig version: 0.3.3

Additional context
The logger within the predict function does not seem to function as expected.

스크린샷 2024-04-02 오후 4 45 28
@alexsherstinsky
Copy link
Collaborator

alexsherstinsky commented Jul 26, 2024

Hello, @MoOo2mini -- thank you for using Ludwig's LLM fine-tuning capabilities and reporting your issue. We cannot reproduce your error, because we do not have access to your model:

FileNotFoundError: [Errno 2] No such file or directory: '/content/test/model_hyperparameters.json'

Could you please make your model available (e.g., on HuggingFace), and I will be happy to troubleshoot the problem.

Thank you very much.

@alexsherstinsky alexsherstinsky added the llm Large Language Model related label Jul 26, 2024
@mhabedank mhabedank closed this as not planned Won't fix, can't repro, duplicate, stale Oct 21, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
llm Large Language Model related
Projects
None yet
Development

No branches or pull requests

3 participants