diff --git a/modelscope/models/nlp/llama/text_generation.py b/modelscope/models/nlp/llama/text_generation.py index 45b9d5f0..f42f62ef 100644 --- a/modelscope/models/nlp/llama/text_generation.py +++ b/modelscope/models/nlp/llama/text_generation.py @@ -108,6 +108,9 @@ class LlamaForTextGeneration(MsModelMixin, LlamaForCausalLM, TorchModel): max_length=gen_kwargs['max_length'], tokenizer=tokenizer) input_ids = prompt_ids.to(self.device) + + print(f'>>>input_ids in text_generation: {input_ids}') + generate_ids = self.generate(input_ids, **gen_kwargs) # remove input tokens generate_ids = generate_ids[:, input_ids.shape[1]:]