From 97f2bad42a9732a5de23a3899617da6644a2b988 Mon Sep 17 00:00:00 2001
From: jllllll <3887729+jllllll@users.noreply.github.com>
Date: Thu, 14 Dec 2023 22:15:45 -0600
Subject: [PATCH] Add 0.2.23 builds to package index
---
README.md | 5 ++-
generate-html.ps1 | 4 +-
generate-textgen-html.ps1 | 3 +-
index/AVX/cpu/llama-cpp-python/index.html | 9 ++++
index/AVX/cu116/llama-cpp-python/index.html | 9 ++++
index/AVX/cu117/llama-cpp-python/index.html | 9 ++++
index/AVX/cu118/llama-cpp-python/index.html | 9 ++++
index/AVX/cu120/llama-cpp-python/index.html | 9 ++++
index/AVX/cu121/llama-cpp-python/index.html | 9 ++++
index/AVX/cu122/llama-cpp-python/index.html | 9 ++++
.../AVX/rocm5.4.2/llama-cpp-python/index.html | 5 +++
index/AVX/rocm5.5/llama-cpp-python/index.html | 5 +++
.../AVX/rocm5.6.1/llama-cpp-python/index.html | 5 +++
index/AVX2/cpu/llama-cpp-python/index.html | 9 ++++
index/AVX2/cu116/llama-cpp-python/index.html | 9 ++++
index/AVX2/cu117/llama-cpp-python/index.html | 9 ++++
index/AVX2/cu118/llama-cpp-python/index.html | 9 ++++
index/AVX2/cu120/llama-cpp-python/index.html | 9 ++++
index/AVX2/cu121/llama-cpp-python/index.html | 9 ++++
index/AVX2/cu122/llama-cpp-python/index.html | 9 ++++
.../rocm5.4.2/llama-cpp-python/index.html | 5 +++
.../AVX2/rocm5.5/llama-cpp-python/index.html | 5 +++
.../rocm5.6.1/llama-cpp-python/index.html | 5 +++
index/AVX512/cpu/llama-cpp-python/index.html | 9 ++++
.../AVX512/cu116/llama-cpp-python/index.html | 9 ++++
.../AVX512/cu117/llama-cpp-python/index.html | 9 ++++
.../AVX512/cu118/llama-cpp-python/index.html | 9 ++++
.../AVX512/cu120/llama-cpp-python/index.html | 9 ++++
.../AVX512/cu121/llama-cpp-python/index.html | 9 ++++
.../AVX512/cu122/llama-cpp-python/index.html | 9 ++++
index/basic/cpu/llama-cpp-python/index.html | 45 +++++++++++++++++++
index/basic/cu116/llama-cpp-python/index.html | 9 ++++
index/basic/cu117/llama-cpp-python/index.html | 9 ++++
index/basic/cu118/llama-cpp-python/index.html | 9 ++++
index/basic/cu120/llama-cpp-python/index.html | 9 ++++
index/basic/cu121/llama-cpp-python/index.html | 9 ++++
index/basic/cu122/llama-cpp-python/index.html | 9 ++++
.../rocm5.4.2/llama-cpp-python/index.html | 5 +++
.../basic/rocm5.5/llama-cpp-python/index.html | 5 +++
.../rocm5.6.1/llama-cpp-python/index.html | 5 +++
index/cpu.html | 9 ++++
index/cu116.html | 9 ++++
index/cu117.html | 9 ++++
index/cu118.html | 9 ++++
index/cu120.html | 9 ++++
index/cu121.html | 9 ++++
index/cu122.html | 9 ++++
index/rocm5.4.2.html | 5 +++
index/rocm5.5.html | 5 +++
index/rocm5.6.1.html | 5 +++
.../cu117/llama-cpp-python-cuda/index.html | 9 ++++
.../cu118/llama-cpp-python-cuda/index.html | 9 ++++
.../cu120/llama-cpp-python-cuda/index.html | 9 ++++
.../cu121/llama-cpp-python-cuda/index.html | 9 ++++
.../cu122/llama-cpp-python-cuda/index.html | 9 ++++
.../llama-cpp-python-cuda/index.html | 5 +++
.../rocm5.5/llama-cpp-python-cuda/index.html | 5 +++
.../llama-cpp-python-cuda/index.html | 5 +++
.../cu117/llama-cpp-python-cuda/index.html | 9 ++++
.../cu118/llama-cpp-python-cuda/index.html | 9 ++++
.../cu120/llama-cpp-python-cuda/index.html | 9 ++++
.../cu121/llama-cpp-python-cuda/index.html | 9 ++++
.../cu122/llama-cpp-python-cuda/index.html | 9 ++++
.../llama-cpp-python-cuda/index.html | 5 +++
.../rocm5.5/llama-cpp-python-cuda/index.html | 5 +++
.../llama-cpp-python-cuda/index.html | 5 +++
.../cu117/llama-cpp-python-cuda/index.html | 9 ++++
.../cu118/llama-cpp-python-cuda/index.html | 9 ++++
.../cu120/llama-cpp-python-cuda/index.html | 9 ++++
.../cu121/llama-cpp-python-cuda/index.html | 9 ++++
.../cu122/llama-cpp-python-cuda/index.html | 9 ++++
.../llama-cpp-python-cuda/index.html | 5 +++
.../rocm5.5/llama-cpp-python-cuda/index.html | 5 +++
.../llama-cpp-python-cuda/index.html | 5 +++
74 files changed, 599 insertions(+), 4 deletions(-)
diff --git a/README.md b/README.md
index 3e3f695a..3be64921 100644
--- a/README.md
+++ b/README.md
@@ -6,8 +6,9 @@ Requirements:
- CUDA 11.6 - 12.2
- CPython 3.8 - 3.11
-llama.cpp, and llama-cpp-python by extension, has migrated to using the new GGUF format and has dropped support for GGML.
-This applies to version 0.1.79+.
+> [!WARNING]
+> MacOS 11 and Windows ROCm wheels are unavailable for 0.2.22+.
+> This is due to build issues with llama.cpp that are not yet resolved.
ROCm builds for AMD GPUs: https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/tag/rocm
Metal builds for MacOS 11.0+: https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/tag/metal
diff --git a/generate-html.ps1 b/generate-html.ps1
index cca9470c..aeafc459 100644
--- a/generate-html.ps1
+++ b/generate-html.ps1
@@ -3,7 +3,7 @@ Set-Location $PSScriptRoot
$destinationDir = if (Test-Path $(Join-Path $(Resolve-Path '.') 'index')) {Join-Path '.' 'index' -resolve} else {(New-Item 'index' -ItemType 'Directory').fullname}
$avxVersions = "AVX","AVX2","AVX512","basic"
$cudaVersions = "11.6","11.7","11.8","12.0","12.1","12.2","rocm5.4.2","rocm5.5","rocm5.5.1","rocm5.6.1","cpu"
-$packageVersions = (@(62)+66..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11+14..20).foreach({"$_".Insert(0,'0.2.')})
+$packageVersions = (@(62)+66..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11+14..20+@(23)).foreach({"$_".Insert(0,'0.2.')})
$pythonVersions = "3.7","3.8","3.9","3.10","3.11"
$supportedSystems = 'linux_x86_64','win_amd64','macosx_11_0_x86_64','macosx_12_0_x86_64','macosx_13_0_x86_64','macosx_14_0_x86_64','macosx_11_0_arm64','macosx_12_0_arm64','macosx_13_0_arm64','macosx_14_0_arm64','macosx_11_0_aarch64','macosx_12_0_aarch64','macosx_13_0_aarch64','macosx_14_0_aarch64'
$wheelSource = 'https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download'
@@ -41,6 +41,8 @@ Foreach ($avxVersion in $avxVersions)
ForEach ($supportedSystem in $supportedSystems)
{
$doMacos = $avxVersion -eq 'basic' -and $cudaVersion -eq 'cpu' -and $supportedSystem.contains('macosx') -and (($packageVersion -eq '0.1.85' -and !$supportedSystem.contains('macosx_14_0')) -or [version]$packageVersion -gt [version]'0.2.4')
+ if ([version]$packageVersion -gt '0.2.20' -and $supportedSystem.contains('macosx_11_0')) {$doMacos = $false}
+ if ($cudaVersion.StartsWith('rocm') -and [version]$packageVersion -gt [version]"0.2.21" -and $supportedSystem -eq 'win_amd64') {continue}
if ($cudaVersion.StartsWith('rocm') -and $cudaVersion.Split('rocm')[-1] -ne '5.5.1' -and $supportedSystem -eq 'win_amd64') {continue}
if ($cudaVersion.StartsWith('rocm') -and $cudaVersion.Split('rocm')[-1] -eq '5.5.1' -and $supportedSystem -eq 'linux_x86_64') {continue}
if ([version]$packageVersion -gt [version]"0.1.85" -and $supportedSystem -eq 'linux_x86_64') {$supportedSystem = 'manylinux_2_31_x86_64'}
diff --git a/generate-textgen-html.ps1 b/generate-textgen-html.ps1
index 76cf517b..b4374f57 100644
--- a/generate-textgen-html.ps1
+++ b/generate-textgen-html.ps1
@@ -4,7 +4,7 @@ $destinationDir = if (Test-Path $(Join-Path $(Resolve-Path '.') 'index')) {Join-
$destinationDir = if (Test-Path $(Join-Path $destinationDir 'textgen')) {Join-Path $destinationDir 'textgen'} else {(New-Item $(Join-Path $destinationDir 'textgen') -ItemType 'Directory').fullname}
$avxVersions = "AVX","AVX2","basic"
$cudaVersions = "11.7","11.8","12.0","12.1","12.2","rocm5.4.2","rocm5.5","rocm5.5.1","rocm5.6.1"
-$packageVersions = (73..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11+14..20).foreach({"$_".Insert(0,'0.2.')})
+$packageVersions = (73..74+76..85).foreach({"$_".Insert(0,'0.1.')}) + (0..11+14..20+@(23)).foreach({"$_".Insert(0,'0.2.')})
$pythonVersions = "3.8","3.9","3.10","3.11"
$supportedSystems = 'linux_x86_64','win_amd64'
$wheelSource = 'https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download'
@@ -37,6 +37,7 @@ Foreach ($avxVersion in $avxVersions)
$pyVer = $pythonVersion.replace('.','')
ForEach ($supportedSystem in $supportedSystems)
{
+ if ($cudaVersion.StartsWith('rocm') -and [version]$packageVersion -gt [version]"0.2.21" -and $supportedSystem -eq 'win_amd64') {continue}
if ($cudaVersion.StartsWith('rocm') -and $cudaVersion.Split('rocm')[-1] -ne '5.5.1' -and $supportedSystem -eq 'win_amd64') {continue}
if ($cudaVersion.StartsWith('rocm') -and $cudaVersion.Split('rocm')[-1] -eq '5.5.1' -and $supportedSystem -eq 'linux_x86_64') {continue}
if ([version]$packageVersion -gt [version]"0.1.85" -and $supportedSystem -eq 'linux_x86_64') {$supportedSystem = 'manylinux_2_31_x86_64'}
diff --git a/index/AVX/cpu/llama-cpp-python/index.html b/index/AVX/cpu/llama-cpp-python/index.html
index 602129ff..fbf72294 100644
--- a/index/AVX/cpu/llama-cpp-python/index.html
+++ b/index/AVX/cpu/llama-cpp-python/index.html
@@ -336,5 +336,14 @@
llama_cpp_python-0.2.20+cpuavx-cp310-cp310-win_amd64.whl
llama_cpp_python-0.2.20+cpuavx-cp311-cp311-manylinux_2_31_x86_64.whl
llama_cpp_python-0.2.20+cpuavx-cp311-cp311-win_amd64.whl
+
+ llama_cpp_python-0.2.23+cpuavx-cp38-cp38-manylinux_2_31_x86_64.whl
+ llama_cpp_python-0.2.23+cpuavx-cp38-cp38-win_amd64.whl
+ llama_cpp_python-0.2.23+cpuavx-cp39-cp39-manylinux_2_31_x86_64.whl
+ llama_cpp_python-0.2.23+cpuavx-cp39-cp39-win_amd64.whl
+ llama_cpp_python-0.2.23+cpuavx-cp310-cp310-manylinux_2_31_x86_64.whl
+ llama_cpp_python-0.2.23+cpuavx-cp310-cp310-win_amd64.whl
+ llama_cpp_python-0.2.23+cpuavx-cp311-cp311-manylinux_2_31_x86_64.whl
+ llama_cpp_python-0.2.23+cpuavx-cp311-cp311-win_amd64.whl