From 5b817f916edee90ea2cc9fd682cd83aeacfbbca3 Mon Sep 17 00:00:00 2001 From: "xingjun.wang" Date: Sun, 22 Oct 2023 01:13:32 +0800 Subject: [PATCH] update --- modelscope/models/nlp/llama/text_generation.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/modelscope/models/nlp/llama/text_generation.py b/modelscope/models/nlp/llama/text_generation.py index 45b9d5f0..f42f62ef 100644 --- a/modelscope/models/nlp/llama/text_generation.py +++ b/modelscope/models/nlp/llama/text_generation.py @@ -108,6 +108,9 @@ class LlamaForTextGeneration(MsModelMixin, LlamaForCausalLM, TorchModel): max_length=gen_kwargs['max_length'], tokenizer=tokenizer) input_ids = prompt_ids.to(self.device) + + print(f'>>>input_ids in text_generation: {input_ids}') + generate_ids = self.generate(input_ids, **gen_kwargs) # remove input tokens generate_ids = generate_ids[:, input_ids.shape[1]:]