|
@@ -30,6 +30,13 @@ pip install https://github.com/huggingface/transformers/archive/6589e510fa4e6c44
|
|
|
# run one bloom block for a few steps
|
|
|
python -m cli.inference_one_block --config cli/config.json # see other args
|
|
|
|
|
|
+# convert model from HF hub to a distributed format
|
|
|
+MY_WRITE_TOKEN=TODO_WRITE_TOKEN_FROM_https://huggingface.co/settings/token
|
|
|
+python -m cli.convert_model --model bigscience/bloom-6b3 \
|
|
|
+ --output_path ./converted_model --output_repo bigscience/test-bloomd-6b3 \
|
|
|
+ --use_auth_token $MY_WRITE_TOKEN # ^-- todo replace output repo with something you have access to
|
|
|
+
|
|
|
+
|
|
|
# minimalistic server with non-trained bloom blocks
|
|
|
python -m cli.run_server --prefix smol --block_config bigscience/bloom-6b3 --num_blocks 2 --identity_path ./server1.id --host_maddrs /ip4/127.0.0.1/tcp/31337
|
|
|
```
|