Created
August 28, 2024 16:42
-
-
Save JGalego/159ab9b3f989143f6b8bdcee384a1c2f to your computer and use it in GitHub Desktop.
Testing Llama Stack π¦
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| # 0a. Download, install and activate Miniconda | |
| # https://docs.anaconda.com/miniconda/ | |
| mkdir -p ~/miniconda3 | |
| wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh -O ~/miniconda3/miniconda.sh | |
| bash ~/miniconda3/miniconda.sh -b -u -p ~/miniconda3 | |
| rm ~/miniconda3/miniconda.sh | |
| ~/miniconda3/bin/conda init bash | |
| source ~/.bashrc | |
| # 0b. Install Llama Stack | |
| pip install llama-toolchain | |
| # 1. Install distribution | |
| llama distribution install --spec local --name local_llama_8b_instruct | |
| # 2. Download model | |
| llama download --source huggingface --model-id Meta-Llama3.1-8B-Instruct --hf-token $HF_TOKEN | |
| # 3. Start distribution | |
| llama distribution start --name local_llama_8b_instruct --port 5000 --disable-ipv6 | |
| # 4. Test it! | |
| conda activate conda activate local_llama_8b_instruct | |
| python -m llama_toolchain.inference.client localhost 5000 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment