diff --git a/BAAI_Aquila_Model_License.pdf b/BAAI_Aquila_Model_License.pdf old mode 100644 new mode 100755 diff --git a/examples/AltDiffusion-m18/generate.py b/examples/AltDiffusion-m18/generate.py old mode 100644 new mode 100755 index af7230bc..98ee65d9 --- a/examples/AltDiffusion-m18/generate.py +++ b/examples/AltDiffusion-m18/generate.py @@ -11,7 +11,7 @@ loader = AutoLoader(task_name="text2img", #contrastive learning model_name="AltDiffusion-m18", model_dir="./checkpoints", - use_fp16=False) + fp16=False) model = loader.get_model() model.eval() model.to(device) diff --git a/examples/AltDiffusion/generate.py b/examples/AltDiffusion/generate.py old mode 100644 new mode 100755 index 0c2b7bd4..bafe1ae1 --- a/examples/AltDiffusion/generate.py +++ b/examples/AltDiffusion/generate.py @@ -11,7 +11,7 @@ loader = AutoLoader(task_name="text2img", #contrastive learning model_name="AltDiffusion-m9", model_dir="./checkpoints", - use_fp16=False) + fp16=False) model = loader.get_model() model.eval() diff --git a/flagai/auto_model/auto_loader.py b/flagai/auto_model/auto_loader.py index 013eacec..bd27d9ed 100755 --- a/flagai/auto_model/auto_loader.py +++ b/flagai/auto_model/auto_loader.py @@ -220,8 +220,6 @@ def __init__(self, only_download_config=only_download_config, device=device, **kwargs) - if kwargs.get("use_fp16", None): - self.model.half() if model_type == "nlp": if brief_model_name in ["galactica",]: diff --git a/setup.py b/setup.py index 2920b487..4fbe8412 100755 --- a/setup.py +++ b/setup.py @@ -5,7 +5,7 @@ setup( name="flagai", - version="v1.7.2", + version="v1.7.3", description="FlagAI aims to help researchers and developers to freely train and test large-scale models for NLP/CV/VL tasks.", long_description=open("README.md", encoding="utf-8").read(), long_description_content_type="text/markdown", @@ -36,5 +36,7 @@ 'taming-transformers-rom1504 == 0.0.6', 'rouge-score == 0.1.2', 'sacrebleu == 2.3.1', + 'jsonlines == 3.1.0', + 'accelerate == 0.19.0' ] )