Files
airllm-fork-nodejs/node_modules/node-llama-cpp/dist/bindings/utils/logBinaryUsageExampleToConsole.js
2026-02-05 15:27:49 +08:00

22 lines
1.1 KiB
JavaScript

import { removeUndefinedFields } from "../../utils/removeNullFields.js";
import { getExampleUsageCodeOfGetLlama } from "./getExampleUsageCodeOfGetLlama.js";
export function logBinaryUsageExampleToConsole(buildOptions, specifyGpuType, showLatestBuildUsageExample = true) {
console.log("To use the binary you've just built, use this code:");
const llamaOptions = removeUndefinedFields({
gpu: specifyGpuType
? buildOptions.gpu
: undefined,
cmakeOptions: buildOptions.customCmakeOptions.size === 0
? undefined
: Object.fromEntries([...buildOptions.customCmakeOptions.entries()].sort(([keyA], [keyB]) => keyA.localeCompare(keyB)))
});
console.log(getExampleUsageCodeOfGetLlama(Object.keys(llamaOptions).length === 0
? undefined
: llamaOptions));
if (showLatestBuildUsageExample) {
console.log();
console.log("To always use the latest binary you build using a CLI command, use this code:");
console.log(getExampleUsageCodeOfGetLlama("lastBuild"));
}
}
//# sourceMappingURL=logBinaryUsageExampleToConsole.js.map