mirror of
https://github.com/modelscope/modelscope.git
synced 2026-02-24 20:19:51 +01:00
update
This commit is contained in:
@@ -1088,6 +1088,10 @@ class ChatGLM2ForConditionalGeneration(ChatGLMPreTrainedModel):
|
||||
return_dict=return_dict,
|
||||
)
|
||||
|
||||
print(
|
||||
f'\n>>transformer_outputs in ChatGLM2ForConditionalGeneration forward:\n {transformer_outputs}'
|
||||
)
|
||||
|
||||
hidden_states = transformer_outputs[0]
|
||||
if return_last_logit:
|
||||
hidden_states = hidden_states[-1:]
|
||||
@@ -1115,6 +1119,10 @@ class ChatGLM2ForConditionalGeneration(ChatGLMPreTrainedModel):
|
||||
output = (lm_logits, ) + transformer_outputs[1:]
|
||||
return ((loss, ) + output) if loss is not None else output
|
||||
|
||||
print(
|
||||
f'\n>>lm_logits in ChatGLM2ForConditionalGeneration forward:\n {lm_logits}'
|
||||
)
|
||||
|
||||
return CausalLMOutputWithPast(
|
||||
loss=loss,
|
||||
logits=lm_logits,
|
||||
|
||||
@@ -143,7 +143,7 @@ class LLMPipeline(Pipeline):
|
||||
|
||||
print(f'>>response in _process_single for llm_pipe: {response}')
|
||||
|
||||
print(f'\n>>self.model: {self.model}')
|
||||
print(f'\n>>self.model:\n {self.model}')
|
||||
|
||||
return response
|
||||
|
||||
|
||||
Reference in New Issue
Block a user