Framework | Domain | Model | Datasets | Tasks | Training | Inference | Reference |
---|---|---|---|---|---|---|---|
PopXL | NLP | Dolly 2.0 | N/A | Instruction Fine-tuned Text Generation | ❌ |
✅ |
Blog |
Dolly 2.0 is a 12B parameter language model trained and instruction fine-tuned by Databricks. By instruction fine-tuning the large language model (LLM), we obtain an LLM better suited for human interactivity. Crucially, Databricks released all code, model weights, and their fine-tuning dataset with an open-source license that permits commercial use. This makes Dolly 2.0 the world's first, truly open-source instruction-tuned LLM. This app requires a minimum of 4 IPUs to run and supports faster inference on POD16.
The best way to run this demo is on Paperspace Gradient's cloud IPUs because everything is already set up for you.
Before you can run this model for training or inference you will need to:
- Install and enable the Poplar SDK (see Poplar SDK setup)
- Install the Python requirements (see Environment setup)
To check if your Poplar SDK has already been enabled, run:
echo $POPLAR_SDK_ENABLED
If no path is provided, then follow these steps:
-
Navigate to your Poplar SDK root directory
-
Enable the Poplar SDK with:
cd poplar-<OS version>-<SDK version>-<hash>
. enable
More detailed instructions on setting up your Poplar environment are available in the [Poplar Quick Start guide] (https://docs.graphcore.ai/projects/poplar-quick-start/en/latest/)
To prepare your environment, follow these steps:
- Create and activate a Python3 virtual environment:
python3 -m venv <venv name>
source <venv path>/bin/activate
-
Navigate to this example's root directory
-
Install the Python requirements with:
pip3 install -r requirements.txt
It is highly recommended to run all inference scripts with the environmental variable
POPXL_CACHE_DIR
set to a directory of your choice. This will cache parts of the graph compilation, thus speeding up subsequent launches.
To run the benchmarking script, execute the following in the example root:
python inference.py --config <config_name>
where config_name
is a config found in config/inference.yml
.
To run an interactive inference script with trained weights, execute the following:
python run-inference.py --config <config-name>
Finally, this example includes a Jupyter notebook. Start a Jupyter server and
navigate to Dolly2-an-OSS-instruction-LLM.ipynb
to explore Dolly inference in
an interactive environment.
This example is licensed as per the LICENSE
file at the root of this
repository. This example, when executed, downloads and uses a checkpoint
provided by Databricks which is licensed under the Apache 2.0 License.