8000 enable reorder_for_compute_comm_overlap · InternLM/lmdeploy@0462005 · GitHub
[go: up one dir, main page]

Skip to content

Commit 0462005

Browse files
committed
enable reorder_for_compute_comm_overlap
1 parent 66aa3ba commit 0462005

File tree

2 files changed

+4
-3
lines changed

2 files changed

+4
-3
lines changed

lmdeploy/pytorch/models/internvl.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -383,6 +383,7 @@ def forward(
383383
):
384384
if inputs_embeds is None and pixel_values is not None:
385385
# extract feature
386+
torch._inductor.config.reorder_for_compute_comm_overlap = True
386387
torch._dynamo.mark_dynamic(pixel_values, 0)
387388
vit_embeds = self.extract_feature(pixel_values)
388389
lang_embeds = self.language_model.get_input_embeddings()(input_ids)

requirements/runtime_cuda.txt

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,8 @@ safetensors
1616
sentencepiece
1717
shortuuid
1818
tiktoken
19-
torch<=2.5.1,>=2.0.0
20-
torchvision<=0.20.1,>=0.15.0
19+
torch<=2.6.0,>=2.0.0
20+
torchvision<=0.21.0,>=0.15.0
2121
transformers
22-
triton<=3.1.0,>=3.0.0; sys_platform == "linux"
22+
triton<=3.2.0,>=3.0.0; sys_platform == "linux"
2323
uvicorn

0 commit comments

Comments
 (0)
0