Skip to content

Commit e2db116

Browse files
[Model] Enable BLOOM on V1 (vllm-project#23488)
Signed-off-by: DarkLight1337 <[email protected]>
1 parent 416f059 commit e2db116

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

docs/models/supported_models.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -328,7 +328,7 @@ th {
328328
| `BaiChuanForCausalLM` | Baichuan2, Baichuan | `baichuan-inc/Baichuan2-13B-Chat`, `baichuan-inc/Baichuan-7B`, etc. | ✅︎ | ✅︎ | ✅︎ |
329329
| `BailingMoeForCausalLM` | Ling | `inclusionAI/Ling-lite-1.5`, `inclusionAI/Ling-plus`, etc. | ✅︎ | ✅︎ | ✅︎ |
330330
| `BambaForCausalLM` | Bamba | `ibm-ai-platform/Bamba-9B-fp8`, `ibm-ai-platform/Bamba-9B` | ✅︎ | ✅︎ | ✅︎ |
331-
| `BloomForCausalLM` | BLOOM, BLOOMZ, BLOOMChat | `bigscience/bloom`, `bigscience/bloomz`, etc. | | ✅︎ | |
331+
| `BloomForCausalLM` | BLOOM, BLOOMZ, BLOOMChat | `bigscience/bloom`, `bigscience/bloomz`, etc. | | ✅︎ | ✅︎ |
332332
| `BartForConditionalGeneration` | BART | `facebook/bart-base`, `facebook/bart-large-cnn`, etc. | | | |
333333
| `MBartForConditionalGeneration` | mBART | `facebook/mbart-large-en-ro`, `facebook/mbart-large-50`, etc. | | | |
334334
| `ChatGLMModel`, `ChatGLMForConditionalGeneration` | ChatGLM | `zai-org/chatglm2-6b`, `zai-org/chatglm3-6b`, `ShieldLM-6B-chatglm3`, etc. | ✅︎ | ✅︎ | ✅︎ |

vllm/model_executor/models/bloom.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@
4343
from vllm.model_executor.sampling_metadata import SamplingMetadata
4444
from vllm.sequence import IntermediateTensors
4545

46-
from .interfaces import SupportsPP, SupportsQuant, SupportsV0Only
46+
from .interfaces import SupportsPP, SupportsQuant
4747
from .utils import (AutoWeightsLoader, is_pp_missing_parameter,
4848
make_empty_intermediate_tensors_factory, make_layers,
4949
maybe_prefix)
@@ -313,7 +313,7 @@ def load_weights(self, weights: Iterable[tuple[str,
313313
return loaded_params
314314

315315

316-
class BloomForCausalLM(nn.Module, SupportsPP, SupportsV0Only, SupportsQuant):
316+
class BloomForCausalLM(nn.Module, SupportsPP, SupportsQuant):
317317

318318
def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
319319
super().__init__()

0 commit comments

Comments
 (0)