1.Add getting labels from dataset in "text_classificationfinetune_text_classification.py" to simplify user's operation in flex training. Parameters "--num_labels" and "--labels" were removed in "run_train.sh".
2.In "chatglm6b / finetune.py", building dataset from file is necessary to support flex training.
Link: https://code.alibaba-inc.com/Ali-MaaS/MaaS-lib/codereview/13382745
* support getting labels from dataset in sbert text classification and building dataset from file in chatglm-6b
* support getting labels from dataset in sbert text classification and building dataset from file in chatglm-6b
* remove repetitive labels in a concise manner of using set
* reserve parameter labels in finetune_text_classification
* Merge branch 'master' of http://gitlab.alibaba-inc.com/Ali-MaaS/MaaS-lib
reserve parameter labels in finetune_text_classification
* Merge branch 'support_text_cls_labels_chatglm_json'
reserve parameter labels in finetune_text_classification
Link: https://code.alibaba-inc.com/Ali-MaaS/MaaS-lib/codereview/13080086
* support lora for llama
* update baichuan
* remove work_dir
* fixbug: 1. change ConfigDict to list when hooks key not in config 2. ignore all bin files when preparing output folder
* 1. support device_map 2. remove the operation of to float when using lora
* add inference file
* add comment
* support device_map