Update modeling_tinyllava_phi.py
Browse files- modeling_tinyllava_phi.py +11 -13
modeling_tinyllava_phi.py
CHANGED
@@ -1,12 +1,22 @@
|
|
1 |
# For licensing see accompanying LICENSE file.
|
2 |
-
# Copyright (C) 2024
|
|
|
3 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
import re
|
|
|
5 |
import torch
|
6 |
import torch.utils.checkpoint
|
7 |
from torch import nn
|
8 |
from torch.nn import functional as F
|
9 |
|
|
|
10 |
from transformers import PreTrainedModel
|
11 |
from transformers.modeling_outputs import CausalLMOutputWithPast
|
12 |
from transformers.generation.utils import GenerateOutput
|
@@ -16,19 +26,7 @@ from .configuration import TinyLlavaConfig, IGNORE_INDEX, IMAGE_TOKEN_INDEX, DEF
|
|
16 |
|
17 |
from transformers import AutoConfig, AutoModelForCausalLM, PhiForCausalLM
|
18 |
|
19 |
-
import time
|
20 |
-
from transformers.utils import logging
|
21 |
-
import dataclasses
|
22 |
-
from enum import auto, Enum
|
23 |
-
from typing import List, Tuple
|
24 |
|
25 |
-
# this import has to be relative, otherwise, when setting trust_remote_code=True
|
26 |
-
# huggingface transformers won't be able to load the module correctly
|
27 |
-
from typing import List, Optional, Union
|
28 |
-
import requests
|
29 |
-
from PIL import Image
|
30 |
-
from io import BytesIO
|
31 |
-
import base64
|
32 |
|
33 |
logger = logging.get_logger(__name__)
|
34 |
|
|
|
1 |
# For licensing see accompanying LICENSE file.
|
2 |
+
# Copyright (C) 2024 TinyLLaVA. All Rights Reserved.
|
3 |
+
import time
|
4 |
|
5 |
+
import dataclasses
|
6 |
+
from enum import auto, Enum
|
7 |
+
from typing import List, Tuple, Optional, Union
|
8 |
+
import requests
|
9 |
+
from PIL import Image
|
10 |
+
from io import BytesIO
|
11 |
+
import base64
|
12 |
import re
|
13 |
+
|
14 |
import torch
|
15 |
import torch.utils.checkpoint
|
16 |
from torch import nn
|
17 |
from torch.nn import functional as F
|
18 |
|
19 |
+
from transformers.utils import logging
|
20 |
from transformers import PreTrainedModel
|
21 |
from transformers.modeling_outputs import CausalLMOutputWithPast
|
22 |
from transformers.generation.utils import GenerateOutput
|
|
|
26 |
|
27 |
from transformers import AutoConfig, AutoModelForCausalLM, PhiForCausalLM
|
28 |
|
|
|
|
|
|
|
|
|
|
|
29 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
|
31 |
logger = logging.get_logger(__name__)
|
32 |
|