|
@@ -118,6 +118,15 @@ class Server:
|
|
|
use_auth_token=use_auth_token,
|
|
|
revision=revision,
|
|
|
)
|
|
|
+ elif converted_model_name_or_path == "facebook/opt-175b":
|
|
|
+ bloom_config = BloomConfig.from_pretrained(
|
|
|
+ "bigscience/bloom-petals"
|
|
|
+ )
|
|
|
+ bloom_config.hidden_size = 12288
|
|
|
+ bloom_config.n_layer = 96
|
|
|
+ bloom_config.n_head = 96
|
|
|
+ bloom_config.vocab_size = 50272
|
|
|
+ self.block_config = bloom_config
|
|
|
else:
|
|
|
def _patch_bloom_config(bloom_config: BloomConfig, opt_config: OPTConfig):
|
|
|
bloom_config.hidden_size = opt_config.hidden_size
|