Skip to content

Commit

Permalink
Add 0.1.82 builds to package index
Browse files Browse the repository at this point in the history
Restructure new release workflow
  • Loading branch information
jllllll committed Aug 28, 2023
1 parent a109e03 commit 504db3d
Show file tree
Hide file tree
Showing 54 changed files with 575 additions and 63 deletions.
4 changes: 2 additions & 2 deletions .github/workflows/build-wheels-full-release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.1.78'
default: 'v0.1.78'
description: 'Version tag of llama-cpp-python to build: v0.1.82'
default: 'v0.1.82'
required: true
type: string

Expand Down
110 changes: 51 additions & 59 deletions .github/workflows/build-wheels-prioritized-release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -8,20 +8,13 @@ on:
default: 'v0.1.82'
required: true
type: string
workflow_call:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.1.82'
default: 'v0.1.82'
required: true
type: string

permissions:
contents: write

jobs:
build_textgen_wheels_prio:
name: ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.avxver }}
name: Textgen Prio ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.avxver }}
runs-on: ${{ matrix.os }}
strategy:
matrix:
Expand Down Expand Up @@ -125,8 +118,7 @@ jobs:
path: ./dist/*.whl

build_wheels_main_prio:
name: ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.releasetag == 'wheels' && 'AVX2' || matrix.releasetag }}
needs: build_textgen_wheels_prio
name: Main Prio ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.releasetag == 'wheels' && 'AVX2' || matrix.releasetag }}
runs-on: ${{ matrix.os }}
strategy:
matrix:
Expand Down Expand Up @@ -208,17 +200,17 @@ jobs:
with:
name: ${{ matrix.releasetag == 'wheels' && 'AVX2' || matrix.releasetag }}
path: ./dist/*.whl

build_textgen_wheels:
name: ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.avxver }}
needs: build_wheels_main_prio
build_wheels_main:
name: Main ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.releasetag == 'wheels' && 'AVX2' || matrix.releasetag }}
needs: ["build_textgen_wheels_prio", "build_wheels_main_prio"]
runs-on: ${{ matrix.os }}
strategy:
matrix:
os: [ubuntu-20.04, windows-latest]
pyver: ["3.10", "3.8", "3.9", "3.11"]
cuda: ["11.7.1", "11.8.0", "12.0.1", "12.1.1", "12.2.0"]
avxver: ["AVX2","AVX","basic"]
pyver: ["3.10", "3.11", "3.9", "3.8", "3.7"]
cuda: ["11.7.1", "11.8.0", "12.0.1", "12.1.1", "12.2.0", "11.6.2"]
releasetag: ["wheels","AVX","basic","AVX512"]
exclude:
- pyver: "3.10"
cuda: "11.7.1"
Expand All @@ -227,7 +219,7 @@ jobs:
shell: pwsh
env:
CUDAVER: ${{ matrix.cuda }}
AVXVER: ${{ matrix.avxver }}
AVXVER: ${{ matrix.releasetag }}
PCKGVER: ${{ inputs.version }}

steps:
Expand Down Expand Up @@ -261,27 +253,6 @@ jobs:
mamba install -y 'cuda' $cudaChannels.TrimEnd().Split()
python -m pip install build wheel
- name: Change Package Name
run: |
$packageVersion = [version]$env:PCKGVER.TrimStart('v')
$setup = Get-Content 'setup.py' -raw
$pyproject = Get-Content 'pyproject.toml' -raw
$cmakelists = Get-Content 'CMakeLists.txt' -raw
$regexstr = '(?s)name="llama_cpp_python",(.+)(package_dir={"llama_cpp": "llama_cpp", "llama_cpp.server": "llama_cpp/server"},.+?packages=\["llama_cpp", "llama_cpp.server"],)'
if ($packageVersion -gt [version]'0.1.77') {$regexstr = '(?s)name="llama_cpp_python",(.+)(package_dir={"llama_cpp": "llama_cpp", "llama_cpp.server": "llama_cpp/server"},.+?package_data={"llama_cpp": \["py.typed"]},.+?packages=\["llama_cpp", "llama_cpp.server"],)'}
$regexmatch = [Regex]::Matches($setup,$regexstr)
if (!($regexmatch[0].Success)) {throw 'setup.py parsing failed'}
$newstr = 'name="llama_cpp_python_cuda",' + $regexmatch[0].Groups[1].Value + $regexmatch[0].Groups[2].Value.Replace('llama_cpp','llama_cpp_cuda')
$newsetup = $regexmatch[0].Result(('$`'+$newstr+'$'''))
$regexstr = '(?s)(?<=name = ")llama_cpp_python(".+?packages = \[{include = ")llama_cpp(".+)'
$regexmatch = [Regex]::Matches($pyproject,$regexstr)
if (!($regexmatch[0].Success)) {throw 'pyproject.toml parsing failed'}
$newpyproject = $regexmatch[0].Result(('$`'+'llama_cpp_python_cuda'+'$1llama_cpp_cuda$2'))
Copy-Item 'llama_cpp' 'llama_cpp_cuda' -recurse
New-Item 'setup.py' -itemType File -value $newsetup -force
New-Item 'pyproject.toml' -itemType File -value $newpyproject -force
New-Item 'CMakeLists.txt' -itemType File -value $cmakelists.Replace('llama_cpp','llama_cpp_cuda') -force
- name: Build Wheel
run: |
$packageVersion = [version]$env:PCKGVER.TrimStart('v')
Expand All @@ -296,37 +267,38 @@ jobs:
if ($packageVersion -gt [version]'0.1.68' -and $packageVersion -lt [version]'0.1.71' -and [version]$env:CUDAVER -ge [version]'12.0') {$env:CMAKE_ARGS = '-DLLAMA_CUBLAS=on -DCMAKE_CUDA_ARCHITECTURES=52;61-real;70-real;72-real;75-real;80-real;86-real;89-real;90'}
if ($packageVersion -gt [version]'0.1.68' -and $packageVersion -lt [version]'0.1.71' -and [version]$env:CUDAVER -lt [version]'11.8') {$env:CMAKE_ARGS = '-DLLAMA_CUBLAS=on -DCMAKE_CUDA_ARCHITECTURES=35-real;37-real;52;61-real;70-real;72-real;75-real;80-real;86'}
if ($packageVersion -lt [version]'0.1.66') {$env:CUDAFLAGS = '-arch=all'}
if ($env:AVXVER -eq 'AVX') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX2=off'; $buildtag = 'avx'}
if ($env:AVXVER -eq 'basic') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_FMA=off -DLLAMA_F16C=off'; $buildtag = 'basic'}
python -m build --wheel -C--build-option=egg_info "-C--build-option=--tag-build=+cu$cudaVersion$buildtag"
if ($env:AVXVER -eq 'AVX') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX2=off'}
if ($env:AVXVER -eq 'AVX512') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX512=on'}
if ($env:AVXVER -eq 'basic') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_FMA=off -DLLAMA_F16C=off'}
python -m build --wheel -C--build-option=egg_info "-C--build-option=--tag-build=+cu$cudaVersion"
- name: Upload files to a GitHub release
id: upload-release
uses: svenstaro/upload-release-action@2.6.1
continue-on-error: true
with:
file: ./dist/*.whl
tag: textgen-webui
tag: ${{ matrix.releasetag }}
file_glob: true
make_latest: false
overwrite: true

- uses: actions/upload-artifact@v3
if: steps.upload-release.outcome == 'failure'
with:
name: textgen-webui-wheels
name: ${{ matrix.releasetag == 'wheels' && 'AVX2' || matrix.releasetag }}
path: ./dist/*.whl
build_wheels_main:
name: ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.releasetag == 'wheels' && 'AVX2' || matrix.releasetag }}
needs: build_textgen_wheels

build_textgen_wheels:
name: Textgen ${{ matrix.os }} ${{ matrix.pyver }} ${{ matrix.cuda }} ${{ matrix.avxver }}
needs: build_wheels_main
runs-on: ${{ matrix.os }}
strategy:
matrix:
os: [ubuntu-20.04, windows-latest]
pyver: ["3.7", "3.8", "3.9", "3.10", "3.11"]
cuda: ["11.6.2", "11.7.1", "11.8.0", "12.0.1", "12.1.1", "12.2.0"]
releasetag: ["AVX","wheels","AVX512","basic"]
pyver: ["3.10", "3.11", "3.8", "3.9"]
cuda: ["11.7.1", "11.8.0", "12.0.1", "12.1.1", "12.2.0"]
avxver: ["AVX2","AVX","basic"]
exclude:
- pyver: "3.10"
cuda: "11.7.1"
Expand All @@ -335,7 +307,7 @@ jobs:
shell: pwsh
env:
CUDAVER: ${{ matrix.cuda }}
AVXVER: ${{ matrix.releasetag }}
AVXVER: ${{ matrix.avxver }}
PCKGVER: ${{ inputs.version }}

steps:
Expand Down Expand Up @@ -369,6 +341,27 @@ jobs:
mamba install -y 'cuda' $cudaChannels.TrimEnd().Split()
python -m pip install build wheel
- name: Change Package Name
run: |
$packageVersion = [version]$env:PCKGVER.TrimStart('v')
$setup = Get-Content 'setup.py' -raw
$pyproject = Get-Content 'pyproject.toml' -raw
$cmakelists = Get-Content 'CMakeLists.txt' -raw
$regexstr = '(?s)name="llama_cpp_python",(.+)(package_dir={"llama_cpp": "llama_cpp", "llama_cpp.server": "llama_cpp/server"},.+?packages=\["llama_cpp", "llama_cpp.server"],)'
if ($packageVersion -gt [version]'0.1.77') {$regexstr = '(?s)name="llama_cpp_python",(.+)(package_dir={"llama_cpp": "llama_cpp", "llama_cpp.server": "llama_cpp/server"},.+?package_data={"llama_cpp": \["py.typed"]},.+?packages=\["llama_cpp", "llama_cpp.server"],)'}
$regexmatch = [Regex]::Matches($setup,$regexstr)
if (!($regexmatch[0].Success)) {throw 'setup.py parsing failed'}
$newstr = 'name="llama_cpp_python_cuda",' + $regexmatch[0].Groups[1].Value + $regexmatch[0].Groups[2].Value.Replace('llama_cpp','llama_cpp_cuda')
$newsetup = $regexmatch[0].Result(('$`'+$newstr+'$'''))
$regexstr = '(?s)(?<=name = ")llama_cpp_python(".+?packages = \[{include = ")llama_cpp(".+)'
$regexmatch = [Regex]::Matches($pyproject,$regexstr)
if (!($regexmatch[0].Success)) {throw 'pyproject.toml parsing failed'}
$newpyproject = $regexmatch[0].Result(('$`'+'llama_cpp_python_cuda'+'$1llama_cpp_cuda$2'))
Copy-Item 'llama_cpp' 'llama_cpp_cuda' -recurse
New-Item 'setup.py' -itemType File -value $newsetup -force
New-Item 'pyproject.toml' -itemType File -value $newpyproject -force
New-Item 'CMakeLists.txt' -itemType File -value $cmakelists.Replace('llama_cpp','llama_cpp_cuda') -force
- name: Build Wheel
run: |
$packageVersion = [version]$env:PCKGVER.TrimStart('v')
Expand All @@ -383,31 +376,30 @@ jobs:
if ($packageVersion -gt [version]'0.1.68' -and $packageVersion -lt [version]'0.1.71' -and [version]$env:CUDAVER -ge [version]'12.0') {$env:CMAKE_ARGS = '-DLLAMA_CUBLAS=on -DCMAKE_CUDA_ARCHITECTURES=52;61-real;70-real;72-real;75-real;80-real;86-real;89-real;90'}
if ($packageVersion -gt [version]'0.1.68' -and $packageVersion -lt [version]'0.1.71' -and [version]$env:CUDAVER -lt [version]'11.8') {$env:CMAKE_ARGS = '-DLLAMA_CUBLAS=on -DCMAKE_CUDA_ARCHITECTURES=35-real;37-real;52;61-real;70-real;72-real;75-real;80-real;86'}
if ($packageVersion -lt [version]'0.1.66') {$env:CUDAFLAGS = '-arch=all'}
if ($env:AVXVER -eq 'AVX') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX2=off'}
if ($env:AVXVER -eq 'AVX512') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX512=on'}
if ($env:AVXVER -eq 'basic') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_FMA=off -DLLAMA_F16C=off'}
python -m build --wheel -C--build-option=egg_info "-C--build-option=--tag-build=+cu$cudaVersion"
if ($env:AVXVER -eq 'AVX') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX2=off'; $buildtag = 'avx'}
if ($env:AVXVER -eq 'basic') {$env:CMAKE_ARGS = $env:CMAKE_ARGS + ' -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_FMA=off -DLLAMA_F16C=off'; $buildtag = 'basic'}
python -m build --wheel -C--build-option=egg_info "-C--build-option=--tag-build=+cu$cudaVersion$buildtag"
- name: Upload files to a GitHub release
id: upload-release
uses: svenstaro/upload-release-action@2.6.1
continue-on-error: true
with:
file: ./dist/*.whl
tag: ${{ matrix.releasetag }}
tag: textgen-webui
file_glob: true
make_latest: false
overwrite: true

- uses: actions/upload-artifact@v3
if: steps.upload-release.outcome == 'failure'
with:
name: ${{ matrix.releasetag == 'wheels' && 'AVX2' || matrix.releasetag }}
name: textgen-webui-wheels
path: ./dist/*.whl

build_wheels_cpu:
name: Build CPU-only Wheels
needs: build_wheels_main
needs: build_textgen_wheels
uses: ./.github/workflows/build-wheels-cpu.yml
with:
version: ${{ inputs.version }}
Expand Down
2 changes: 1 addition & 1 deletion generate-html.ps1
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ Set-Location $PSScriptRoot
$destinationDir = if (Test-Path $(Join-Path $(Resolve-Path '.') 'index')) {Join-Path '.' 'index' -resolve} else {(New-Item 'index' -ItemType 'Directory').fullname}
$avxVersions = "AVX","AVX2","AVX512","basic"
$cudaVersions = "11.6","11.7","11.8","12.0","12.1","12.2","cpu"
$packageVersions = "0.1.62","0.1.66","0.1.67","0.1.68","0.1.69","0.1.70","0.1.71","0.1.72","0.1.73","0.1.74","0.1.76","0.1.77","0.1.78","0.1.79","0.1.80","0.1.81"
$packageVersions = "0.1.62","0.1.66","0.1.67","0.1.68","0.1.69","0.1.70","0.1.71","0.1.72","0.1.73","0.1.74","0.1.76","0.1.77","0.1.78","0.1.79","0.1.80","0.1.81","0.1.82"
$pythonVersions = "3.7","3.8","3.9","3.10","3.11"
$supportedSystems = 'linux_x86_64','win_amd64'
$wheelSource = 'https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download'
Expand Down
2 changes: 1 addition & 1 deletion generate-textgen-html.ps1
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ $destinationDir = if (Test-Path $(Join-Path $(Resolve-Path '.') 'index')) {Join-
$destinationDir = if (Test-Path $(Join-Path $destinationDir 'textgen')) {Join-Path $destinationDir 'textgen'} else {(New-Item $(Join-Path $destinationDir 'textgen') -ItemType 'Directory').fullname}
$avxVersions = "AVX","AVX2","basic"
$cudaVersions = "11.7","11.8","12.0","12.1","12.2"
$packageVersions = "0.1.73","0.1.74","0.1.76","0.1.77","0.1.78","0.1.79","0.1.80","0.1.81"
$packageVersions = "0.1.73","0.1.74","0.1.76","0.1.77","0.1.78","0.1.79","0.1.80","0.1.81","0.1.82"
$pythonVersions = "3.8","3.9","3.10","3.11"
$supportedSystems = 'linux_x86_64','win_amd64'
$wheelSource = 'https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download'
Expand Down
11 changes: 11 additions & 0 deletions index/AVX/cpu/llama-cpp-python/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -121,5 +121,16 @@
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.81+cpuavx-cp310-cp310-win_amd64.whl">llama_cpp_python-0.1.81+cpuavx-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.81+cpuavx-cp311-cp311-linux_x86_64.whl">llama_cpp_python-0.1.81+cpuavx-cp311-cp311-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.81+cpuavx-cp311-cp311-win_amd64.whl">llama_cpp_python-0.1.81+cpuavx-cp311-cp311-win_amd64.whl</a><br/>

<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp37-cp37m-linux_x86_64.whl">llama_cpp_python-0.1.82+cpuavx-cp37-cp37m-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp37-cp37m-win_amd64.whl">llama_cpp_python-0.1.82+cpuavx-cp37-cp37m-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp38-cp38-linux_x86_64.whl">llama_cpp_python-0.1.82+cpuavx-cp38-cp38-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp38-cp38-win_amd64.whl">llama_cpp_python-0.1.82+cpuavx-cp38-cp38-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp39-cp39-linux_x86_64.whl">llama_cpp_python-0.1.82+cpuavx-cp39-cp39-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp39-cp39-win_amd64.whl">llama_cpp_python-0.1.82+cpuavx-cp39-cp39-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp310-cp310-linux_x86_64.whl">llama_cpp_python-0.1.82+cpuavx-cp310-cp310-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp310-cp310-win_amd64.whl">llama_cpp_python-0.1.82+cpuavx-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp311-cp311-linux_x86_64.whl">llama_cpp_python-0.1.82+cpuavx-cp311-cp311-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.1.82+cpuavx-cp311-cp311-win_amd64.whl">llama_cpp_python-0.1.82+cpuavx-cp311-cp311-win_amd64.whl</a><br/>
</body>
</html>
11 changes: 11 additions & 0 deletions index/AVX/cu116/llama-cpp-python/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -176,5 +176,16 @@
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.81+cu116-cp310-cp310-win_amd64.whl">llama_cpp_python-0.1.81+cu116-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.81+cu116-cp311-cp311-linux_x86_64.whl">llama_cpp_python-0.1.81+cu116-cp311-cp311-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.81+cu116-cp311-cp311-win_amd64.whl">llama_cpp_python-0.1.81+cu116-cp311-cp311-win_amd64.whl</a><br/>

<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp37-cp37m-linux_x86_64.whl">llama_cpp_python-0.1.82+cu116-cp37-cp37m-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp37-cp37m-win_amd64.whl">llama_cpp_python-0.1.82+cu116-cp37-cp37m-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp38-cp38-linux_x86_64.whl">llama_cpp_python-0.1.82+cu116-cp38-cp38-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp38-cp38-win_amd64.whl">llama_cpp_python-0.1.82+cu116-cp38-cp38-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp39-cp39-linux_x86_64.whl">llama_cpp_python-0.1.82+cu116-cp39-cp39-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp39-cp39-win_amd64.whl">llama_cpp_python-0.1.82+cu116-cp39-cp39-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp310-cp310-linux_x86_64.whl">llama_cpp_python-0.1.82+cu116-cp310-cp310-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp310-cp310-win_amd64.whl">llama_cpp_python-0.1.82+cu116-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp311-cp311-linux_x86_64.whl">llama_cpp_python-0.1.82+cu116-cp311-cp311-linux_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.1.82+cu116-cp311-cp311-win_amd64.whl">llama_cpp_python-0.1.82+cu116-cp311-cp311-win_amd64.whl</a><br/>
</body>
</html>
Loading

0 comments on commit 504db3d

Please sign in to comment.