First upload version 0.0.1
This commit is contained in:
17
node_modules/node-llama-cpp/dist/gguf/insights/utils/resolveModelGpuLayersOption.d.ts
generated
vendored
Normal file
17
node_modules/node-llama-cpp/dist/gguf/insights/utils/resolveModelGpuLayersOption.d.ts
generated
vendored
Normal file
@@ -0,0 +1,17 @@
|
||||
import { LlamaModelOptions } from "../../../evaluator/LlamaModel/LlamaModel.js";
|
||||
import { BuildGpu } from "../../../bindings/types.js";
|
||||
import type { GgufInsights } from "../GgufInsights.js";
|
||||
export declare function resolveModelGpuLayersOption(gpuLayers: LlamaModelOptions["gpuLayers"], { ggufInsights, ignoreMemorySafetyChecks, getVramState, llamaVramPaddingSize, llamaGpu, llamaSupportsGpuOffloading, defaultContextFlashAttention, defaultContextSwaFullCache, useMmap }: {
|
||||
ggufInsights: GgufInsights;
|
||||
ignoreMemorySafetyChecks?: boolean;
|
||||
getVramState(): Promise<{
|
||||
total: number;
|
||||
free: number;
|
||||
}>;
|
||||
llamaVramPaddingSize: number;
|
||||
llamaGpu: BuildGpu;
|
||||
llamaSupportsGpuOffloading: boolean;
|
||||
defaultContextFlashAttention: boolean;
|
||||
defaultContextSwaFullCache: boolean;
|
||||
useMmap?: boolean;
|
||||
}): Promise<number>;
|
||||
Reference in New Issue
Block a user