From 9a18d3133ca219ace52457640526ee4993879948 Mon Sep 17 00:00:00 2001 From: hiro Date: Sun, 3 Dec 2023 14:35:22 +0700 Subject: [PATCH] chore: refactor openai file structure --- .../inference-openai-extension/package.json | 22 ++---------- .../inference-openai-extension/src/module.ts | 34 ------------------- 2 files changed, 3 insertions(+), 53 deletions(-) delete mode 100644 extensions/inference-openai-extension/src/module.ts diff --git a/extensions/inference-openai-extension/package.json b/extensions/inference-openai-extension/package.json index 5d5dac264..c32027ca8 100644 --- a/extensions/inference-openai-extension/package.json +++ b/extensions/inference-openai-extension/package.json @@ -1,21 +1,14 @@ { "name": "@janhq/inference-openai-extension", "version": "1.0.0", - "description": "Inference Engine for OpenAI Extension, powered by @janhq/nitro, bring a high-performance Llama model inference in pure C++.", + "description": "Inference Engine for OpenAI Extension that can be used with any OpenAI compatible API", "main": "dist/index.js", "module": "dist/module.js", "author": "Jan ", "license": "AGPL-3.0", "scripts": { "build": "tsc -b . && webpack --config webpack.config.js", - "downloadnitro:linux": "NITRO_VERSION=$(cat ./bin/version.txt) && download https://github.com/janhq/nitro/releases/download/v${NITRO_VERSION}/nitro-${NITRO_VERSION}-linux-amd64.tar.gz -e --strip 1 -o ./bin/linux-cpu && chmod +x ./bin/linux-cpu/nitro && chmod +x ./bin/linux-start.sh && download https://github.com/janhq/nitro/releases/download/v${NITRO_VERSION}/nitro-${NITRO_VERSION}-linux-amd64-cuda.tar.gz -e --strip 1 -o ./bin/linux-cuda && chmod +x ./bin/linux-cuda/nitro && chmod +x ./bin/linux-start.sh", - "downloadnitro:darwin": "NITRO_VERSION=$(cat ./bin/version.txt) && download https://github.com/janhq/nitro/releases/download/v${NITRO_VERSION}/nitro-${NITRO_VERSION}-mac-arm64.tar.gz -e --strip 1 -o ./bin/mac-arm64 && chmod +x ./bin/mac-arm64/nitro && download https://github.com/janhq/nitro/releases/download/v${NITRO_VERSION}/nitro-${NITRO_VERSION}-mac-amd64.tar.gz -e --strip 1 -o ./bin/mac-x64 && chmod +x ./bin/mac-x64/nitro", - "downloadnitro:win32": "download.bat", - "downloadnitro": "run-script-os", - "build:publish:darwin": "rimraf *.tgz --glob && npm run build && npm run downloadnitro && ../../.github/scripts/auto-sign.sh && cpx \"bin/**\" \"dist/bin\" && npm pack && cpx *.tgz ../../electron/pre-install", - "build:publish:win32": "rimraf *.tgz --glob && npm run build && npm run downloadnitro && cpx \"bin/**\" \"dist/bin\" && npm pack && cpx *.tgz ../../electron/pre-install", - "build:publish:linux": "rimraf *.tgz --glob && npm run build && npm run downloadnitro && cpx \"bin/**\" \"dist/bin\" && npm pack && cpx *.tgz ../../electron/pre-install", - "build:publish": "run-script-os" + "build:publish": "rimraf *.tgz --glob && npm run build && npm pack && cpx *.tgz ../../electron/pre-install" }, "exports": { ".": "./dist/index.js", @@ -24,19 +17,13 @@ "devDependencies": { "cpx": "^1.5.0", "rimraf": "^3.0.2", - "run-script-os": "^1.1.6", "webpack": "^5.88.2", "webpack-cli": "^5.1.4" }, "dependencies": { "@janhq/core": "file:../../core", - "download-cli": "^1.1.1", - "electron-log": "^5.0.1", "fetch-retry": "^5.0.6", - "kill-port": "^2.0.1", "path-browserify": "^1.0.1", - "rxjs": "^7.8.1", - "tcp-port-used": "^1.0.2", "ts-loader": "^9.5.0", "ulid": "^2.3.0" }, @@ -49,9 +36,6 @@ "README.md" ], "bundleDependencies": [ - "tcp-port-used", - "kill-port", - "fetch-retry", - "electron-log" + "fetch-retry" ] } diff --git a/extensions/inference-openai-extension/src/module.ts b/extensions/inference-openai-extension/src/module.ts deleted file mode 100644 index 305c2e804..000000000 --- a/extensions/inference-openai-extension/src/module.ts +++ /dev/null @@ -1,34 +0,0 @@ -const fetchRetry = require("fetch-retry")(global.fetch); - -const log = require("electron-log"); - -const OPENAI_BASE_URL = "https://api.openai.com/v1"; -const OPENAI_API_KEY = process.env.OPENAI_API_KEY; - -/** - * The response from the initModel function. - * @property error - An error message if the model fails to load. - */ -interface InitModelResponse { - error?: any; - modelFile?: string; -} -// /root/engine/nitro.json - -/** - * Initializes a Nitro subprocess to load a machine learning model. - * @param modelFile - The name of the machine learning model file. - * @returns A Promise that resolves when the model is loaded successfully, or rejects with an error message if the model is not found or fails to load. - */ -function initModel(wrapper: any): Promise { - const engine_settings = { - ...wrapper.settings, - }; - - return ( - ) -} - -module.exports = { - initModel, -};