Skip to content
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.

Commit 3b99d4f

Browse files
authoredOct 30, 2024
rls2.5: fix t5, llava, whisper (#3347)
1 parent ead9765 commit 3b99d4f

File tree

12 files changed

+28
-18
lines changed

12 files changed

+28
-18
lines changed
 

‎docs/tutorials/features/fast_bert.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ Currently `ipex.fast_bert` API is only well optimized for training. For inferenc
99

1010
### Prerequisite
1111

12-
- Transformers 4.6.0 ~ 4.43.2
12+
- Transformers 4.6.0 ~ 4.45.0
1313

1414
### Usage Example
1515

‎examples/cpu/features/fast_bert/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
Currently `ipex.fast_bert` API is only well optimized for training. For inference, it ensures functionality, while to get peak perf, please use `ipex.optimize` API + torchscript.
66

77
# Prerequisite:
8-
Transformers 4.6.0 ~ 4.43.2
8+
Transformers 4.6.0 ~ 4.45.0
99

1010
# Usage Example:
1111
Training:

‎examples/cpu/features/fast_bert/fast_bert_inference_bf16.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
import torch
22
from transformers import BertModel
33

4-
model = BertModel.from_pretrained("bert-base-uncased")
4+
model = BertModel.from_pretrained("bert-base-uncased", attn_implementation="eager")
55
model.eval()
66

77
vocab_size = model.config.vocab_size

‎examples/cpu/features/fast_bert/fast_bert_training_bf16.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
from transformers import BertForSequenceClassification
33

44
model = BertForSequenceClassification.from_pretrained(
5-
"bert-base-uncased", return_dict=True
5+
"bert-base-uncased", return_dict=True, attn_implementation="eager"
66
)
77
model.train()
88
optimizer = torch.optim.Adam(model.parameters(), lr=1e-5)

‎examples/cpu/llm/fine-tuning/requirements.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,6 @@ black[jupyter]
66
datasets
77
fire
88
peft
9-
transformers==4.43.2
9+
transformers==4.45.0
1010
gradio
1111
sentencepiece

‎examples/cpu/llm/inference/distributed/run_accuracy_with_deepspeed.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -574,7 +574,10 @@ def _model_call(
574574
model_kwargs = {"attention_mask": attention_mask_batched}
575575
model_kwargs = (
576576
self.base_model._prepare_encoder_decoder_kwargs_for_generation(
577-
inputs, model_kwargs, "input_ids"
577+
inputs,
578+
model_kwargs,
579+
"input_ids",
580+
transformers.generation.configuration_utils.GenerationConfig(),
578581
)
579582
)
580583
(

‎examples/cpu/llm/inference/single_instance/run_accuracy.py

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -346,7 +346,10 @@ def _model_call(
346346
model_kwargs = {"attention_mask": attention_mask_batched}
347347
model_kwargs = (
348348
self.base_model._prepare_encoder_decoder_kwargs_for_generation(
349-
inputs, model_kwargs, "input_ids"
349+
inputs,
350+
model_kwargs,
351+
"input_ids",
352+
transformers.generation.configuration_utils.GenerationConfig(),
350353
)
351354
)
352355
(
@@ -872,7 +875,7 @@ def __init__(
872875
self._config = self._model.config
873876
self._config.torchscript = self._with_jit
874877
self._model.eval()
875-
if with_ipex and dtype not in ["int8", "int4", "nf4"]:
878+
if with_ipex:
876879
self._model = ipex.llm.optimize(
877880
self._model.eval(),
878881
dtype=infer_dtype,
@@ -1284,6 +1287,8 @@ def _collate(x):
12841287
"num_beams": gen_kwargs["num_beams"],
12851288
"max_new_tokens": gen_kwargs["max_new_tokens"],
12861289
}
1290+
if not self._with_ipex:
1291+
input_dict["use_cache"] = False
12871292
elif re.search("git", self.model.config.architectures[0], re.IGNORECASE):
12881293
input_ids = self._image_processor(
12891294
images=visuals, return_tensors="pt"

‎examples/cpu/llm/inference/single_instance/run_quantization.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,7 @@
88

99
import torch
1010
from torch.utils.data import DataLoader
11+
import transformers
1112
from transformers import AutoConfig
1213
from transformers import TextStreamer
1314
import intel_extension_for_pytorch as ipex
@@ -842,12 +843,11 @@ def collate_batch(self, batch):
842843
elif model.example_inputs_mode == EXAMPLE_INPUTS_MODE.KV_ENC:
843844
input_bs = int(args.batch_size * num_beams)
844845
model_kwargs = {}
845-
model_kwargs = (
846-
user_model._prepare_encoder_decoder_kwargs_for_generation(
847-
torch.vstack(input_ids_padded).unsqueeze(0),
848-
model_kwargs,
849-
"input_features",
850-
)
846+
model_kwargs = user_model._prepare_encoder_decoder_kwargs_for_generation(
847+
torch.vstack(input_ids_padded).unsqueeze(0),
848+
model_kwargs,
849+
"input_features",
850+
transformers.generation.configuration_utils.GenerationConfig(),
851851
)
852852
last_hidden_state = model_kwargs["encoder_outputs"][
853853
"last_hidden_state"

‎intel_extension_for_pytorch/cpu/tpp/fused_bert.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1256,7 +1256,7 @@ def fast_bert(model, dtype=torch.float, optimizer=None, unpad=False):
12561256
# tpp bert optimization depends on the transformers repo to implementate the related module
12571257
installed_pkg = {pkg.key for pkg in pkg_resources.working_set}
12581258
min_version = "4.6.0"
1259-
max_version = "4.43.2"
1259+
max_version = "4.45.0"
12601260
if "transformers" not in installed_pkg:
12611261
raise RuntimeError(
12621262
"Please installed the transformers with version: between {} and {}".format(

‎tests/cpu/test_deepspeed.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,8 +31,10 @@
3131
import transformers
3232
from transformers import AutoConfig
3333
except ImportError:
34+
import subprocess
35+
3436
subprocess.check_call(
35-
[sys.executable, "-m", "pip", "install", "transformers==4.43.2"]
37+
[sys.executable, "-m", "pip", "install", "transformers==4.45.0"]
3638
)
3739
import transformers
3840
from transformers import AutoConfig

‎tests/cpu/test_ipex_optimize_transformers.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717
from transformers import AutoConfig
1818
except ImportError:
1919
subprocess.check_call(
20-
[sys.executable, "-m", "pip", "install", "transformers==4.43.2"]
20+
[sys.executable, "-m", "pip", "install", "transformers==4.45.0"]
2121
)
2222
import transformers
2323
from transformers import AutoConfig

‎tests/cpu/test_tpp_ops.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
import subprocess
1313

1414
subprocess.check_call(
15-
[sys.executable, "-m", "pip", "install", "transformers==4.43.2"]
15+
[sys.executable, "-m", "pip", "install", "transformers==4.45.0"]
1616
)
1717
import transformers
1818
from common_utils import TestCase

0 commit comments

Comments
 (0)
Please sign in to comment.