chore: update lint to use type imports (#10970)

* chore: update lint to use type imports

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* revert

* chore: add workspaces and overrides to package.json

* refactor: clean up imports in lobe-web-browsing executor

---------

Co-authored-by: Claude <noreply@anthropic.com>
This commit is contained in:
Innei 2025-12-25 22:28:19 +08:00 committed by GitHub
parent 7f6bca71e7
commit e87bee6dd5
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
1419 changed files with 2685 additions and 2627 deletions

View file

@ -22,6 +22,14 @@ config.rules['unicorn/prefer-query-selector'] = 0;
config.rules['unicorn/no-array-callback-reference'] = 0;
// FIXME: Linting error in src/app/[variants]/(main)/chat/features/Migration/DBReader.ts, the fundamental solution should be upgrading typescript-eslint
config.rules['@typescript-eslint/no-useless-constructor'] = 0;
config.rules['@typescript-eslint/consistent-type-imports'] = [
'error',
{
disallowTypeAnnotations: false,
fixStyle: 'inline-type-imports',
prefer: 'type-imports',
},
];
config.overrides = [
{
@ -38,6 +46,26 @@ config.overrides = [
},
},
// Desktop app enables `emitDecoratorMetadata`, so some "type-only" references may be required at runtime
// for metadata reflection. Avoid auto-converting imports to `import type` in this scope.
{
files: ['apps/desktop/**/*.{ts,tsx,cts,mts}'],
parserOptions: {
project: ['./apps/desktop/tsconfig.json'],
tsconfigRootDir: __dirname,
},
rules: {
'@typescript-eslint/consistent-type-imports': [
'error',
{
disallowTypeAnnotations: false,
fixStyle: 'inline-type-imports',
prefer: 'no-type-imports',
},
],
},
},
{
files: ['src/store/image/**/*', 'src/types/generation/**/*'],
rules: {

View file

@ -14,6 +14,15 @@
"tts",
"stt"
],
"workspaces": [
"packages/*",
"packages/business/*",
"e2e",
"apps/desktop/src/main"
],
"overrides": {
"stylelint-config-clean-order": "7.0.0"
},
"homepage": "https://github.com/lobehub/lobe-chat",
"bugs": {
"url": "https://github.com/lobehub/lobe-chat/issues/new/choose"
@ -25,12 +34,6 @@
"license": "MIT",
"author": "LobeHub <i@lobehub.com>",
"sideEffects": false,
"workspaces": [
"packages/*",
"packages/business/*",
"e2e",
"apps/desktop/src/main"
],
"scripts": {
"prebuild": "tsx scripts/prebuild.mts && npm run lint",
"build": "cross-env NODE_OPTIONS=--max-old-space-size=6144 next build --webpack",
@ -131,9 +134,6 @@
"eslint --fix"
]
},
"overrides": {
"stylelint-config-clean-order": "7.0.0"
},
"dependencies": {
"@ant-design/icons": "^6.1.0",
"@ant-design/pro-components": "^2.8.10",
@ -450,14 +450,5 @@
"publishConfig": {
"access": "public",
"registry": "https://registry.npmjs.org"
},
"pnpm": {
"onlyBuiltDependencies": [
"@vercel/speed-insights"
],
"overrides": {
"jose": "^6.1.3",
"stylelint-config-clean-order": "7.0.0"
}
}
}
}

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref https://docs.ai21.com/reference/jamba-15-api-ref
const Ai21: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://302.ai/pricing/
const Ai302: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://ai.360.cn/platform/docs/overview
const Ai360: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const AiHubMix: ModelProviderCard = {
apiKeyUrl: 'https://lobe.li/9mZhb4T',

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const AkashChat: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Anthropic: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://learn.microsoft.com/azure/ai-services/openai/concepts/models
const Azure: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://learn.microsoft.com/azure/ai-services/openai/concepts/models
const Azure: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://platform.baichuan-ai.com/price
const Baichuan: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref :https://docs.aws.amazon.com/bedrock/latest/userguide/conversation-inference.html
// ref :https://us-east-1.console.aws.amazon.com/bedrock/home?region=us-east-1#/models

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
/**
* @see https://docs.bfl.ai/

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Cerebras: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref https://developers.cloudflare.com/workers-ai/models/#text-generation
// api https://developers.cloudflare.com/workers-ai/configuration/open-ai-compatibility

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Cohere: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://api.cometapi.com/pricing
const CometAPI: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
/**
* ComfyUI Provider Configuration

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const DeepSeek: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
/**
* @see https://fal.ai/models/fal-ai/flux/schnell

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://fireworks.ai/models?show=Serverless
// ref: https://fireworks.ai/pricing

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://ai.gitee.com/serverless-api/packages/1910
const GiteeAI: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://github.com/marketplace/models
const Github: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://ai.google.dev/gemini-api/docs/models/gemini
const Google: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref https://console.groq.com/docs/tool-use
const Groq: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Higress: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const HuggingFace: ModelProviderCard = {
apiKeyUrl: 'https://huggingface.co/settings/tokens',

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref https://cloud.tencent.com/document/product/1729/104753
const Hunyuan: ModelProviderCard = {

View file

@ -1,6 +1,6 @@
import { ENABLE_BUSINESS_FEATURES } from '@lobechat/business-const';
import { ChatModelCard, ModelProviderCard } from '@/types/llm';
import { type ChatModelCard, type ModelProviderCard } from '@/types/llm';
import Ai21Provider from './ai21';
import Ai302Provider from './ai302';

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// https://cloud.infini-ai.com/genstudio/model
// All models are currently free

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const InternLM: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Jina: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://ollama.com/library
const LMStudio: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const LobeHub: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://platform.minimaxi.com/document/Models
const Minimax: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://docs.mistral.ai/getting-started/models/
// ref: https://docs.mistral.ai/capabilities/function_calling/

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://modelscope.cn/docs/model-service/API-Inference/intro
const ModelScope: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://platform.moonshot.cn/docs/intro#模型列表
const Moonshot: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Nebius: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const NewAPI: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://novita.ai/model-api/product/llm-api
const Novita: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Nvidia: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Ollama: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const OllamaCloud: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://platform.openai.com/docs/deprecations
const OpenAI: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref :https://openrouter.ai/docs#models
const OpenRouter: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref :https://docs.perplexity.ai/docs/model-cards
const Perplexity: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const PPIO: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://developer.qiniu.com/aitokenapi
const Qiniu: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://help.aliyun.com/zh/model-studio/getting-started/models
const Qwen: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// Ref: https://replicate.com/docs
const Replicate: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const SambaNova: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Search1API: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref https://platform.sensenova.cn/pricing
// ref https://platform.sensenova.cn/release?path=/release-202409.md

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://siliconflow.cn/zh-cn/pricing
const SiliconCloud: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://www.xfyun.cn/doc/spark/HTTP%E8%B0%83%E7%94%A8%E6%96%87%E6%A1%A3.html#_3-%E8%AF%B7%E6%B1%82%E8%AF%B4%E6%98%8E
// ref: https://www.xfyun.cn/doc/spark/Web.html#_1-%E6%8E%A5%E5%8F%A3%E8%AF%B4%E6%98%8E

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://platform.stepfun.com/docs/llm/text
// 根据文档,阶级星辰大模型的上下文长度,其 k 的含义均为 1000

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref :https://ai-maas.wair.ac.cn/#/doc
const Taichu: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const TencentCloud: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://docs.together.ai/docs/chat-models
// ref: https://www.together.ai/pricing

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref :https://developers.upstage.ai/docs/getting-started/models
const Upstage: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const V0: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const VercelAIGateway: ModelProviderCard = {
apiKeyUrl: 'https://vercel.com/dashboard/ai-gateway',

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://ai.google.dev/gemini-api/docs/models/gemini
const VertexAI: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const VLLM: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref https://www.volcengine.com/docs/82379/1330310
const Doubao: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const BaiduWenxin: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://x.ai/about
const XAI: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const Xinference: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const ZenMux: ModelProviderCard = {
chatModels: [],

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
// ref: https://platform.lingyiwanwu.com/docs#%E6%A8%A1%E5%9E%8B%E4%B8%8E%E8%AE%A1%E8%B4%B9
const ZeroOne: ModelProviderCard = {

View file

@ -1,4 +1,4 @@
import { ModelProviderCard } from '@/types/llm';
import { type ModelProviderCard } from '@/types/llm';
const ZhiPu: ModelProviderCard = {
chatModels: [],

View file

@ -0,0 +1,26 @@
diff --git a/node_modules/.cache/logger/umi.log b/node_modules/.cache/logger/umi.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/parse/parsers/binary/index-node.cjs b/src/parse/parsers/binary/index-node.cjs
index a88ba4d8a0ddff6f4b545ed9368739d94f893eb2..b9b78a29e3c228e7d78e0861b38d734bdef2c854 100644
--- a/src/parse/parsers/binary/index-node.cjs
+++ b/src/parse/parsers/binary/index-node.cjs
@@ -3,7 +3,7 @@
var _interopRequireDefault = require("@babel/runtime-corejs3/helpers/interopRequireDefault").default;
exports.__esModule = true;
exports.default = void 0;
-var _buffer = require("#buffer");
+var _buffer = require("buffer");
var _apidomCore = require("@swagger-api/apidom-core");
var _ParserError = _interopRequireDefault(require("../../../errors/ParserError.cjs"));
var _Parser = _interopRequireDefault(require("../Parser.cjs"));
diff --git a/src/parse/parsers/binary/index-node.mjs b/src/parse/parsers/binary/index-node.mjs
index 3abce8c52fd5eb7b92f33b66fccf6896ccbc89fe..4a3f957cf127a71c7f2b4dfa56512271b6f785ab 100644
--- a/src/parse/parsers/binary/index-node.mjs
+++ b/src/parse/parsers/binary/index-node.mjs
@@ -1,4 +1,4 @@
-import { Buffer } from '#buffer'; // eslint-disable-line import/order
+import { Buffer } from 'buffer'; // eslint-disable-line import/order
import { ParseResultElement, StringElement } from '@swagger-api/apidom-core';
import ParserError from "../../../errors/ParserError.mjs";
import Parser from "../Parser.mjs";

View file

@ -3,3 +3,15 @@ packages:
- '.'
- 'e2e'
- 'apps/desktop/src/main'
overrides:
jose: ^6.1.3
stylelint-config-clean-order: 7.0.0
'@lobehub/chat-plugin-sdk>swagger-client': 3.36.0
'@swagger-api/apidom-reference': 1.1.0
patchedDependencies:
'@swagger-api/apidom-reference': patches/@swagger-api__apidom-reference.patch
onlyBuiltDependencies:
- '@vercel/speed-insights'

View file

@ -1,5 +1,5 @@
import { ChatErrorType } from '@lobechat/types';
import OpenAI, { ClientOptions } from 'openai';
import OpenAI, { type ClientOptions } from 'openai';
import urlJoin from 'url-join';
import { getLLMConfig } from '@/envs/llm';

View file

@ -1,5 +1,5 @@
import { ChatErrorType, ErrorType } from '@lobechat/types';
import OpenAI from 'openai';
import { ChatErrorType, type ErrorType } from '@lobechat/types';
import type OpenAI from 'openai';
import { getOpenAIAuthFromRequest } from '@/const/fetch';
import { createErrorResponse } from '@/utils/errorResponse';

View file

@ -1,5 +1,5 @@
import debug from 'debug';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { getServerDB } from '@/database/core/db-adaptor';
import { AiAgentService } from '@/server/services/aiAgent';

View file

@ -1,5 +1,5 @@
import debug from 'debug';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { getServerDB } from '@/database/core/db-adaptor';
import { AgentRuntimeCoordinator } from '@/server/modules/AgentRuntime';

View file

@ -1,6 +1,6 @@
import { createSSEHeaders, createSSEWriter } from '@lobechat/utils/server';
import debug from 'debug';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { StreamEventManager } from '@/server/modules/AgentRuntime';

View file

@ -1,5 +1,5 @@
import debug from 'debug';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { serverDBEnv } from '@/config/db';
import { serverDB } from '@/database/server';

View file

@ -1,5 +1,5 @@
import { and, eq } from 'drizzle-orm';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { account } from '@/database/schemas/betterAuth';
import { users } from '@/database/schemas/user';

View file

@ -1,5 +1,5 @@
import { eq } from 'drizzle-orm';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { users } from '@/database/schemas/user';
import { serverDB } from '@/database/server';

View file

@ -1,4 +1,4 @@
import { WebhookEvent } from '@clerk/nextjs/server';
import { type WebhookEvent } from '@clerk/nextjs/server';
import { headers } from 'next/headers';
import { Webhook } from 'svix';

View file

@ -4,7 +4,7 @@ import { chunk } from 'es-toolkit/compat';
import type { ListTopicsForMemoryExtractorCursor } from '@/database/models/topic';
import {
MemoryExtractionExecutor,
MemoryExtractionPayloadInput,
type MemoryExtractionPayloadInput,
MemoryExtractionWorkflowService,
buildWorkflowPayloadInput,
normalizeMemoryExtractionPayload,

View file

@ -3,7 +3,7 @@ import { chunk } from 'es-toolkit/compat';
import {
MemoryExtractionExecutor,
MemoryExtractionPayloadInput,
type MemoryExtractionPayloadInput,
MemoryExtractionWorkflowService,
buildWorkflowPayloadInput,
normalizeMemoryExtractionPayload,

View file

@ -1,5 +1,5 @@
import { MarketSDK } from '@lobehub/market-sdk';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
type RouteContext = {
params: Promise<{

View file

@ -1,5 +1,5 @@
import { MarketSDK } from '@lobehub/market-sdk';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
type RouteContext = {
params: Promise<{

View file

@ -1,5 +1,5 @@
import { MarketSDK } from '@lobehub/market-sdk';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
type RouteContext = {
params: Promise<{

View file

@ -1,5 +1,5 @@
import { MarketSDK } from '@lobehub/market-sdk';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
type RouteContext = {
params: Promise<{

View file

@ -1,5 +1,5 @@
import { MarketSDK } from '@lobehub/market-sdk';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
const MARKET_BASE_URL = process.env.NEXT_PUBLIC_MARKET_BASE_URL || 'https://market.lobehub.com';

View file

@ -1,12 +1,12 @@
import { AuthObject } from '@clerk/backend';
import { type AuthObject } from '@clerk/backend';
import {
AgentRuntimeError,
ChatCompletionErrorPayload,
ModelRuntime,
type ChatCompletionErrorPayload,
type ModelRuntime,
} from '@lobechat/model-runtime';
import { ChatErrorType, ClientSecretPayload } from '@lobechat/types';
import { ChatErrorType, type ClientSecretPayload } from '@lobechat/types';
import { getXorPayload } from '@lobechat/utils/server';
import { NextRequest } from 'next/server';
import { type NextRequest } from 'next/server';
import {
LOBE_CHAT_AUTH_HEADER,

View file

@ -1,5 +1,5 @@
import debug from 'debug';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { URL } from 'node:url';
import { oidcEnv } from '@/envs/oidc';

View file

@ -1,5 +1,5 @@
import debug from 'debug';
import { NextRequest, NextResponse, after } from 'next/server';
import { type NextRequest, NextResponse, after } from 'next/server';
import { OAuthHandoffModel } from '@/database/models/oauthHandoff';
import { serverDB } from '@/database/server';

View file

@ -1,6 +1,6 @@
import { correctOIDCUrl, getUserAuth } from '@lobechat/utils/server';
import debug from 'debug';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { OIDCService } from '@/server/services/oidc';

View file

@ -1,5 +1,5 @@
import debug from 'debug';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
import { OAuthHandoffModel } from '@/database/models/oauthHandoff';
import { serverDB } from '@/database/server';

View file

@ -1,13 +1,13 @@
import {
AGENT_RUNTIME_ERROR_SET,
ChatCompletionErrorPayload,
ModelRuntime,
type ChatCompletionErrorPayload,
type ModelRuntime,
} from '@lobechat/model-runtime';
import { ChatErrorType } from '@lobechat/types';
import { checkAuth } from '@/app/(backend)/middleware/auth';
import { createTraceOptions, initModelRuntimeWithUserPayload } from '@/server/modules/ModelRuntime';
import { ChatStreamPayload } from '@/types/openai/chat';
import { type ChatStreamPayload } from '@/types/openai/chat';
import { createErrorResponse } from '@/utils/errorResponse';
import { getTracePayload } from '@/utils/trace';

View file

@ -1,4 +1,4 @@
import { ChatCompletionErrorPayload, PullModelParams } from '@lobechat/model-runtime';
import { type ChatCompletionErrorPayload, type PullModelParams } from '@lobechat/model-runtime';
import { ChatErrorType } from '@lobechat/types';
import { checkAuth } from '@/app/(backend)/middleware/auth';

View file

@ -1,4 +1,4 @@
import { ChatCompletionErrorPayload } from '@lobechat/model-runtime';
import { type ChatCompletionErrorPayload } from '@lobechat/model-runtime';
import { ChatErrorType } from '@lobechat/types';
import { ModelProvider } from 'model-bank';
import { NextResponse } from 'next/server';

View file

@ -1,5 +1,5 @@
import { AgentRuntimeError } from '@lobechat/model-runtime';
import { ChatErrorType, ErrorType, TraceNameMap } from '@lobechat/types';
import { ChatErrorType, type ErrorType, TraceNameMap } from '@lobechat/types';
import { getXorPayload } from '@lobechat/utils/server';
import type { PluginRequestPayload } from '@lobehub/chat-plugin-sdk';
import { createGatewayOnEdgeRuntime } from '@lobehub/chat-plugins-gateway';

View file

@ -1,5 +1,5 @@
import { revalidateTag } from 'next/cache';
import { NextRequest, NextResponse } from 'next/server';
import { type NextRequest, NextResponse } from 'next/server';
export const GET = async (request: NextRequest) => {
if (!process.env.REVALIDATE_SECRET) {

Some files were not shown because too many files have changed in this diff Show more