mirror of
https://github.com/mudler/LocalAI.git
synced 2025-05-21 02:55:01 +00:00
feat: add flash-attn in nvidia and rocm envs (#1995)
Signed-off-by: Ludovic LEROUX <ludovic@inpher.io>
This commit is contained in:
parent
e152b07b74
commit
b4548ad72d
1 changed files with 12 additions and 6 deletions
|
@ -2,6 +2,7 @@
|
||||||
set -ex
|
set -ex
|
||||||
|
|
||||||
SKIP_CONDA=${SKIP_CONDA:-0}
|
SKIP_CONDA=${SKIP_CONDA:-0}
|
||||||
|
REQUIREMENTS_FILE=$1
|
||||||
|
|
||||||
# Check if environment exist
|
# Check if environment exist
|
||||||
conda_env_exists(){
|
conda_env_exists(){
|
||||||
|
@ -14,7 +15,7 @@ else
|
||||||
export PATH=$PATH:/opt/conda/bin
|
export PATH=$PATH:/opt/conda/bin
|
||||||
if conda_env_exists "transformers" ; then
|
if conda_env_exists "transformers" ; then
|
||||||
echo "Creating virtual environment..."
|
echo "Creating virtual environment..."
|
||||||
conda env create --name transformers --file $1
|
conda env create --name transformers --file $REQUIREMENTS_FILE
|
||||||
echo "Virtual environment created."
|
echo "Virtual environment created."
|
||||||
else
|
else
|
||||||
echo "Virtual environment already exists."
|
echo "Virtual environment already exists."
|
||||||
|
@ -28,11 +29,16 @@ if [ -d "/opt/intel" ]; then
|
||||||
pip install intel-extension-for-transformers datasets sentencepiece tiktoken neural_speed optimum[openvino]
|
pip install intel-extension-for-transformers datasets sentencepiece tiktoken neural_speed optimum[openvino]
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ "$PIP_CACHE_PURGE" = true ] ; then
|
# If we didn't skip conda, activate the environment
|
||||||
|
# to install FlashAttention
|
||||||
if [ $SKIP_CONDA -eq 0 ]; then
|
if [ $SKIP_CONDA -eq 0 ]; then
|
||||||
# Activate conda environment
|
|
||||||
source activate transformers
|
source activate transformers
|
||||||
fi
|
fi
|
||||||
|
if [[ $REQUIREMENTS_FILE =~ -nvidia.yml$ ]]; then
|
||||||
|
#TODO: FlashAttention is supported on nvidia and ROCm, but ROCm install can't be done this easily
|
||||||
|
pip install flash-attn --no-build-isolation
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ "$PIP_CACHE_PURGE" = true ] ; then
|
||||||
pip cache purge
|
pip cache purge
|
||||||
fi
|
fi
|
Loading…
Add table
Add a link
Reference in a new issue