From 45168e9e7a4a0ac27b4026249b0dc8c27dcc2423 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sat, 7 Jan 2023 16:54:49 -0300 Subject: [PATCH] Update the description --- convert-to-torch.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/convert-to-torch.py b/convert-to-torch.py index 07169b57..7dfe42fd 100644 --- a/convert-to-torch.py +++ b/convert-to-torch.py @@ -2,9 +2,9 @@ Converts a transformers model to .pt, which is faster to load. Example: -python convert.py models/opt-1.3b +python convert-to-torch.py models/opt-1.3b -Output will be written to torch-dumps/name-of-the-model.pt +The output will be written to torch-dumps/name-of-the-model.pt ''' from transformers import AutoModelForCausalLM, T5ForConditionalGeneration