Fixup sorting in determineBestBackend
This commit is contained in:
parent
3cd099ee87
commit
13806a3f06
@ -541,10 +541,7 @@ export default class llamacpp_extension extends AIEngine {
|
|||||||
const getBackendCategory = (backendString: string): string | undefined => {
|
const getBackendCategory = (backendString: string): string | undefined => {
|
||||||
if (backendString.includes('cu12.0')) return 'cuda-cu12.0'
|
if (backendString.includes('cu12.0')) return 'cuda-cu12.0'
|
||||||
if (backendString.includes('cu11.7')) return 'cuda-cu11.7'
|
if (backendString.includes('cu11.7')) return 'cuda-cu11.7'
|
||||||
if (backendString.includes('vulkan')) {
|
if (backendString.includes('vulkan')) return 'vulkan'
|
||||||
// Only return vulkan category if we have enough GPU memory
|
|
||||||
return hasEnoughGpuMemory ? 'vulkan' : undefined
|
|
||||||
}
|
|
||||||
if (backendString.includes('avx512')) return 'avx512'
|
if (backendString.includes('avx512')) return 'avx512'
|
||||||
if (backendString.includes('avx2')) return 'avx2'
|
if (backendString.includes('avx2')) return 'avx2'
|
||||||
if (
|
if (
|
||||||
@ -2011,7 +2008,7 @@ export default class llamacpp_extension extends AIEngine {
|
|||||||
const totalMemory = totalVRAM + totalRAM
|
const totalMemory = totalVRAM + totalRAM
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
`Total VRAM: ${totalVRAM} bytes, Total RAM: ${totalRAM} bytes, Free: ${usableRAM} bytes, Total Memory: ${totalMemory} bytes`
|
`Total VRAM: ${totalVRAM} bytes, Total RAM: ${totalRAM} bytes, Total Memory: ${totalMemory} bytes`
|
||||||
)
|
)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user