Skip to content

Commit

Permalink
Add 0.2.14 builds to package index
Browse files Browse the repository at this point in the history
  • Loading branch information
jllllll committed Nov 7, 2023
1 parent b6b39fe commit c180f07
Show file tree
Hide file tree
Showing 86 changed files with 650 additions and 28 deletions.
8 changes: 4 additions & 4 deletions .github/workflows/build-wheels-cpu.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand All @@ -21,8 +21,8 @@ on:
workflow_call:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand Down
8 changes: 4 additions & 4 deletions .github/workflows/build-wheels-macos.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand All @@ -21,8 +21,8 @@ on:
workflow_call:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand Down
8 changes: 4 additions & 4 deletions .github/workflows/build-wheels-oobabooga.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand All @@ -21,8 +21,8 @@ on:
workflow_call:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand Down
12 changes: 6 additions & 6 deletions .github/workflows/build-wheels-prioritized-release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.11'
default: 'v0.2.11'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string

Expand All @@ -18,15 +18,15 @@ jobs:
uses: ./.github/workflows/build-wheels-oobabooga.yml
with:
version: ${{ inputs.version }}
config: 'pyver:3.10;cuda:11.7.1'
config: 'pyver:3.10;cuda:12.1.1'

build_wheels_main_prio:
name: Main Prioritized
uses: ./.github/workflows/build-wheels.yml
with:
version: ${{ inputs.version }}
cpu: '0'
config: 'pyver:3.10;cuda:11.7.1'
config: 'pyver:3.10;cuda:12.1.1'

build_wheels_cpu_prio:
name: CPU-only Prioritized
Expand Down Expand Up @@ -57,15 +57,15 @@ jobs:
with:
version: ${{ inputs.version }}
cpu: '0'
exclude: 'pyver:3.10,cuda:11.7.1'
exclude: 'pyver:3.10,cuda:12.1.1'

build_textgen_wheels:
name: Textgen Wheels
needs: build_wheels_main
uses: ./.github/workflows/build-wheels-oobabooga.yml
with:
version: ${{ inputs.version }}
exclude: 'pyver:3.10,cuda:11.7.1'
exclude: 'pyver:3.10,cuda:12.1.1'

build_wheels_cpu:
name: CPU-only Wheels
Expand Down
8 changes: 4 additions & 4 deletions .github/workflows/build-wheels-rocm-full.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.11'
default: 'v0.2.11'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand All @@ -21,8 +21,8 @@ on:
workflow_call:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.11'
default: 'v0.2.11'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
config:
Expand Down
8 changes: 4 additions & 4 deletions .github/workflows/build-wheels.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
cpu:
Expand All @@ -26,8 +26,8 @@ on:
workflow_call:
inputs:
version:
description: 'Version tag of llama-cpp-python to build: v0.2.9'
default: 'v0.2.9'
description: 'Version tag of llama-cpp-python to build: v0.2.14'
default: 'v0.2.14'
required: true
type: string
cpu:
Expand Down
2 changes: 1 addition & 1 deletion generate-html.ps1
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ Set-Location $PSScriptRoot
$destinationDir = if (Test-Path $(Join-Path $(Resolve-Path '.') 'index')) {Join-Path '.' 'index' -resolve} else {(New-Item 'index' -ItemType 'Directory').fullname}
$avxVersions = "AVX","AVX2","AVX512","basic"
$cudaVersions = "11.6","11.7","11.8","12.0","12.1","12.2","rocm5.4.2","rocm5.5","rocm5.5.1","rocm5.6.1","cpu"
$packageVersions = (@(62)+66..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11).foreach({"$_".Insert(0,'0.2.')})
$packageVersions = (@(62)+66..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11+@(14)).foreach({"$_".Insert(0,'0.2.')})
$pythonVersions = "3.7","3.8","3.9","3.10","3.11"
$supportedSystems = 'linux_x86_64','win_amd64','macosx_11_0_x86_64','macosx_12_0_x86_64','macosx_13_0_x86_64','macosx_14_0_x86_64','macosx_11_0_arm64','macosx_12_0_arm64','macosx_13_0_arm64','macosx_14_0_arm64'
$wheelSource = 'https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download'
Expand Down
2 changes: 1 addition & 1 deletion generate-textgen-html.ps1
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ $destinationDir = if (Test-Path $(Join-Path $(Resolve-Path '.') 'index')) {Join-
$destinationDir = if (Test-Path $(Join-Path $destinationDir 'textgen')) {Join-Path $destinationDir 'textgen'} else {(New-Item $(Join-Path $destinationDir 'textgen') -ItemType 'Directory').fullname}
$avxVersions = "AVX","AVX2","basic"
$cudaVersions = "11.7","11.8","12.0","12.1","12.2","rocm5.4.2","rocm5.5","rocm5.5.1","rocm5.6.1"
$packageVersions = (73..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11).foreach({"$_".Insert(0,'0.2.')})
$packageVersions = (73..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11+@(14)).foreach({"$_".Insert(0,'0.2.')})
$pythonVersions = "3.8","3.9","3.10","3.11"
$supportedSystems = 'linux_x86_64','win_amd64'
$wheelSource = 'https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download'
Expand Down
9 changes: 9 additions & 0 deletions index/AVX/cpu/llama-cpp-python/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -273,5 +273,14 @@
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.11+cpuavx-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.11+cpuavx-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.11+cpuavx-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.11+cpuavx-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.11+cpuavx-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.11+cpuavx-cp311-cp311-win_amd64.whl</a><br/>

<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp38-cp38-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cpuavx-cp38-cp38-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp38-cp38-win_amd64.whl">llama_cpp_python-0.2.14+cpuavx-cp38-cp38-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp39-cp39-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cpuavx-cp39-cp39-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp39-cp39-win_amd64.whl">llama_cpp_python-0.2.14+cpuavx-cp39-cp39-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp310-cp310-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cpuavx-cp310-cp310-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.14+cpuavx-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cpuavx-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/cpu/llama_cpp_python-0.2.14+cpuavx-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.14+cpuavx-cp311-cp311-win_amd64.whl</a><br/>
</body>
</html>
9 changes: 9 additions & 0 deletions index/AVX/cu116/llama-cpp-python/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -328,5 +328,14 @@
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu116-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.11+cu116-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu116-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.11+cu116-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu116-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.11+cu116-cp311-cp311-win_amd64.whl</a><br/>

<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp38-cp38-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu116-cp38-cp38-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp38-cp38-win_amd64.whl">llama_cpp_python-0.2.14+cu116-cp38-cp38-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp39-cp39-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu116-cp39-cp39-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp39-cp39-win_amd64.whl">llama_cpp_python-0.2.14+cu116-cp39-cp39-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp310-cp310-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu116-cp310-cp310-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.14+cu116-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu116-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu116-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.14+cu116-cp311-cp311-win_amd64.whl</a><br/>
</body>
</html>
9 changes: 9 additions & 0 deletions index/AVX/cu117/llama-cpp-python/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -328,5 +328,14 @@
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu117-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.11+cu117-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu117-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.11+cu117-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu117-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.11+cu117-cp311-cp311-win_amd64.whl</a><br/>

<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp38-cp38-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu117-cp38-cp38-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp38-cp38-win_amd64.whl">llama_cpp_python-0.2.14+cu117-cp38-cp38-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp39-cp39-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu117-cp39-cp39-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp39-cp39-win_amd64.whl">llama_cpp_python-0.2.14+cu117-cp39-cp39-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp310-cp310-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu117-cp310-cp310-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.14+cu117-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu117-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu117-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.14+cu117-cp311-cp311-win_amd64.whl</a><br/>
</body>
</html>
9 changes: 9 additions & 0 deletions index/AVX/cu118/llama-cpp-python/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -328,5 +328,14 @@
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu118-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.11+cu118-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu118-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.11+cu118-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu118-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.11+cu118-cp311-cp311-win_amd64.whl</a><br/>

<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp38-cp38-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu118-cp38-cp38-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp38-cp38-win_amd64.whl">llama_cpp_python-0.2.14+cu118-cp38-cp38-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp39-cp39-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu118-cp39-cp39-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp39-cp39-win_amd64.whl">llama_cpp_python-0.2.14+cu118-cp39-cp39-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp310-cp310-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu118-cp310-cp310-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.14+cu118-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu118-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu118-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.14+cu118-cp311-cp311-win_amd64.whl</a><br/>
</body>
</html>
9 changes: 9 additions & 0 deletions index/AVX/cu120/llama-cpp-python/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -328,5 +328,14 @@
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu120-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.11+cu120-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu120-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.11+cu120-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.11+cu120-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.11+cu120-cp311-cp311-win_amd64.whl</a><br/>

<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp38-cp38-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu120-cp38-cp38-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp38-cp38-win_amd64.whl">llama_cpp_python-0.2.14+cu120-cp38-cp38-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp39-cp39-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu120-cp39-cp39-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp39-cp39-win_amd64.whl">llama_cpp_python-0.2.14+cu120-cp39-cp39-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp310-cp310-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu120-cp310-cp310-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp310-cp310-win_amd64.whl">llama_cpp_python-0.2.14+cu120-cp310-cp310-win_amd64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp311-cp311-manylinux_2_31_x86_64.whl">llama_cpp_python-0.2.14+cu120-cp311-cp311-manylinux_2_31_x86_64.whl</a><br/>
<a href="https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/AVX/llama_cpp_python-0.2.14+cu120-cp311-cp311-win_amd64.whl">llama_cpp_python-0.2.14+cu120-cp311-cp311-win_amd64.whl</a><br/>
</body>
</html>
Loading

0 comments on commit c180f07

Please sign in to comment.