mirror of
				https://github.com/langgenius/dify.git
				synced 2025-10-31 02:42:59 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			449 lines
		
	
	
		
			15 KiB
		
	
	
	
		
			TypeScript
		
	
	
	
	
	
			
		
		
	
	
			449 lines
		
	
	
		
			15 KiB
		
	
	
	
		
			TypeScript
		
	
	
	
	
	
| import { useCallback, useEffect, useRef, useState } from 'react'
 | |
| import produce from 'immer'
 | |
| import { EditionType, VarType } from '../../types'
 | |
| import type { Memory, PromptItem, ValueSelector, Var, Variable } from '../../types'
 | |
| import { useStore } from '../../store'
 | |
| import {
 | |
|   useIsChatMode,
 | |
|   useNodesReadOnly,
 | |
| } from '../../hooks'
 | |
| import useAvailableVarList from '../_base/hooks/use-available-var-list'
 | |
| import useConfigVision from '../../hooks/use-config-vision'
 | |
| import type { LLMNodeType, StructuredOutput } from './types'
 | |
| import { useModelList, useModelListAndDefaultModelAndCurrentProviderAndModel } from '@/app/components/header/account-setting/model-provider-page/hooks'
 | |
| import {
 | |
|   ModelFeatureEnum,
 | |
|   ModelTypeEnum,
 | |
| } from '@/app/components/header/account-setting/model-provider-page/declarations'
 | |
| import useNodeCrud from '@/app/components/workflow/nodes/_base/hooks/use-node-crud'
 | |
| import useOneStepRun from '@/app/components/workflow/nodes/_base/hooks/use-one-step-run'
 | |
| import { RETRIEVAL_OUTPUT_STRUCT } from '@/app/components/workflow/constants'
 | |
| import { checkHasContextBlock, checkHasHistoryBlock, checkHasQueryBlock } from '@/app/components/base/prompt-editor/constants'
 | |
| 
 | |
| const useConfig = (id: string, payload: LLMNodeType) => {
 | |
|   const { nodesReadOnly: readOnly } = useNodesReadOnly()
 | |
|   const isChatMode = useIsChatMode()
 | |
| 
 | |
|   const defaultConfig = useStore(s => s.nodesDefaultConfigs)[payload.type]
 | |
|   const [defaultRolePrefix, setDefaultRolePrefix] = useState<{ user: string; assistant: string }>({ user: '', assistant: '' })
 | |
|   const { inputs, setInputs: doSetInputs } = useNodeCrud<LLMNodeType>(id, payload)
 | |
|   const inputRef = useRef(inputs)
 | |
| 
 | |
|   const setInputs = useCallback((newInputs: LLMNodeType) => {
 | |
|     if (newInputs.memory && !newInputs.memory.role_prefix) {
 | |
|       const newPayload = produce(newInputs, (draft) => {
 | |
|         draft.memory!.role_prefix = defaultRolePrefix
 | |
|       })
 | |
|       doSetInputs(newPayload)
 | |
|       inputRef.current = newPayload
 | |
|       return
 | |
|     }
 | |
|     doSetInputs(newInputs)
 | |
|     inputRef.current = newInputs
 | |
|   }, [doSetInputs, defaultRolePrefix])
 | |
| 
 | |
|   // model
 | |
|   const model = inputs.model
 | |
|   const modelMode = inputs.model?.mode
 | |
|   const isChatModel = modelMode === 'chat'
 | |
| 
 | |
|   const isCompletionModel = !isChatModel
 | |
| 
 | |
|   const hasSetBlockStatus = (() => {
 | |
|     const promptTemplate = inputs.prompt_template
 | |
|     const hasSetContext = isChatModel ? (promptTemplate as PromptItem[]).some(item => checkHasContextBlock(item.text)) : checkHasContextBlock((promptTemplate as PromptItem).text)
 | |
|     if (!isChatMode) {
 | |
|       return {
 | |
|         history: false,
 | |
|         query: false,
 | |
|         context: hasSetContext,
 | |
|       }
 | |
|     }
 | |
|     if (isChatModel) {
 | |
|       return {
 | |
|         history: false,
 | |
|         query: (promptTemplate as PromptItem[]).some(item => checkHasQueryBlock(item.text)),
 | |
|         context: hasSetContext,
 | |
|       }
 | |
|     }
 | |
|     else {
 | |
|       return {
 | |
|         history: checkHasHistoryBlock((promptTemplate as PromptItem).text),
 | |
|         query: checkHasQueryBlock((promptTemplate as PromptItem).text),
 | |
|         context: hasSetContext,
 | |
|       }
 | |
|     }
 | |
|   })()
 | |
| 
 | |
|   const shouldShowContextTip = !hasSetBlockStatus.context && inputs.context.enabled
 | |
| 
 | |
|   const appendDefaultPromptConfig = useCallback((draft: LLMNodeType, defaultConfig: any, passInIsChatMode?: boolean) => {
 | |
|     const promptTemplates = defaultConfig.prompt_templates
 | |
|     if (passInIsChatMode === undefined ? isChatModel : passInIsChatMode) {
 | |
|       draft.prompt_template = promptTemplates.chat_model.prompts
 | |
|     }
 | |
|     else {
 | |
|       draft.prompt_template = promptTemplates.completion_model.prompt
 | |
| 
 | |
|       setDefaultRolePrefix({
 | |
|         user: promptTemplates.completion_model.conversation_histories_role.user_prefix,
 | |
|         assistant: promptTemplates.completion_model.conversation_histories_role.assistant_prefix,
 | |
|       })
 | |
|     }
 | |
|   }, [isChatModel])
 | |
|   useEffect(() => {
 | |
|     const isReady = defaultConfig && Object.keys(defaultConfig).length > 0
 | |
| 
 | |
|     if (isReady && !inputs.prompt_template) {
 | |
|       const newInputs = produce(inputs, (draft) => {
 | |
|         appendDefaultPromptConfig(draft, defaultConfig)
 | |
|       })
 | |
|       setInputs(newInputs)
 | |
|     }
 | |
|     // eslint-disable-next-line react-hooks/exhaustive-deps
 | |
|   }, [defaultConfig, isChatModel])
 | |
| 
 | |
|   const [modelChanged, setModelChanged] = useState(false)
 | |
|   const {
 | |
|     currentProvider,
 | |
|     currentModel,
 | |
|   } = useModelListAndDefaultModelAndCurrentProviderAndModel(ModelTypeEnum.textGeneration)
 | |
| 
 | |
|   const {
 | |
|     isVisionModel,
 | |
|     handleVisionResolutionEnabledChange,
 | |
|     handleVisionResolutionChange,
 | |
|     handleModelChanged: handleVisionConfigAfterModelChanged,
 | |
|   } = useConfigVision(model, {
 | |
|     payload: inputs.vision,
 | |
|     onChange: (newPayload) => {
 | |
|       const newInputs = produce(inputs, (draft) => {
 | |
|         draft.vision = newPayload
 | |
|       })
 | |
|       setInputs(newInputs)
 | |
|     },
 | |
|   })
 | |
| 
 | |
|   const handleModelChanged = useCallback((model: { provider: string; modelId: string; mode?: string }) => {
 | |
|     const newInputs = produce(inputRef.current, (draft) => {
 | |
|       draft.model.provider = model.provider
 | |
|       draft.model.name = model.modelId
 | |
|       draft.model.mode = model.mode!
 | |
|       const isModeChange = model.mode !== inputRef.current.model.mode
 | |
|       if (isModeChange && defaultConfig && Object.keys(defaultConfig).length > 0)
 | |
|         appendDefaultPromptConfig(draft, defaultConfig, model.mode === 'chat')
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|     setModelChanged(true)
 | |
|   }, [setInputs, defaultConfig, appendDefaultPromptConfig])
 | |
| 
 | |
|   useEffect(() => {
 | |
|     if (currentProvider?.provider && currentModel?.model && !model.provider) {
 | |
|       handleModelChanged({
 | |
|         provider: currentProvider?.provider,
 | |
|         modelId: currentModel?.model,
 | |
|         mode: currentModel?.model_properties?.mode as string,
 | |
|       })
 | |
|     }
 | |
|   }, [model.provider, currentProvider, currentModel, handleModelChanged])
 | |
| 
 | |
|   const handleCompletionParamsChange = useCallback((newParams: Record<string, any>) => {
 | |
|     const newInputs = produce(inputs, (draft) => {
 | |
|       draft.model.completion_params = newParams
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [inputs, setInputs])
 | |
| 
 | |
|   // change to vision model to set vision enabled, else disabled
 | |
|   useEffect(() => {
 | |
|     if (!modelChanged)
 | |
|       return
 | |
|     setModelChanged(false)
 | |
|     handleVisionConfigAfterModelChanged()
 | |
|     // eslint-disable-next-line react-hooks/exhaustive-deps
 | |
|   }, [isVisionModel, modelChanged])
 | |
| 
 | |
|   // variables
 | |
|   const isShowVars = (() => {
 | |
|     if (isChatModel)
 | |
|       return (inputs.prompt_template as PromptItem[]).some(item => item.edition_type === EditionType.jinja2)
 | |
| 
 | |
|     return (inputs.prompt_template as PromptItem).edition_type === EditionType.jinja2
 | |
|   })()
 | |
|   const handleAddEmptyVariable = useCallback(() => {
 | |
|     const newInputs = produce(inputRef.current, (draft) => {
 | |
|       if (!draft.prompt_config) {
 | |
|         draft.prompt_config = {
 | |
|           jinja2_variables: [],
 | |
|         }
 | |
|       }
 | |
|       if (!draft.prompt_config.jinja2_variables)
 | |
|         draft.prompt_config.jinja2_variables = []
 | |
| 
 | |
|       draft.prompt_config.jinja2_variables.push({
 | |
|         variable: '',
 | |
|         value_selector: [],
 | |
|       })
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [setInputs])
 | |
| 
 | |
|   const handleAddVariable = useCallback((payload: Variable) => {
 | |
|     const newInputs = produce(inputRef.current, (draft) => {
 | |
|       if (!draft.prompt_config) {
 | |
|         draft.prompt_config = {
 | |
|           jinja2_variables: [],
 | |
|         }
 | |
|       }
 | |
|       if (!draft.prompt_config.jinja2_variables)
 | |
|         draft.prompt_config.jinja2_variables = []
 | |
| 
 | |
|       draft.prompt_config.jinja2_variables.push(payload)
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [setInputs])
 | |
| 
 | |
|   const handleVarListChange = useCallback((newList: Variable[]) => {
 | |
|     const newInputs = produce(inputRef.current, (draft) => {
 | |
|       if (!draft.prompt_config) {
 | |
|         draft.prompt_config = {
 | |
|           jinja2_variables: [],
 | |
|         }
 | |
|       }
 | |
|       if (!draft.prompt_config.jinja2_variables)
 | |
|         draft.prompt_config.jinja2_variables = []
 | |
| 
 | |
|       draft.prompt_config.jinja2_variables = newList
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [setInputs])
 | |
| 
 | |
|   const handleVarNameChange = useCallback((oldName: string, newName: string) => {
 | |
|     const newInputs = produce(inputRef.current, (draft) => {
 | |
|       if (isChatModel) {
 | |
|         const promptTemplate = draft.prompt_template as PromptItem[]
 | |
|         promptTemplate.filter(item => item.edition_type === EditionType.jinja2).forEach((item) => {
 | |
|           item.jinja2_text = (item.jinja2_text || '').replaceAll(`{{ ${oldName} }}`, `{{ ${newName} }}`)
 | |
|         })
 | |
|       }
 | |
|       else {
 | |
|         if ((draft.prompt_template as PromptItem).edition_type !== EditionType.jinja2)
 | |
|           return
 | |
| 
 | |
|         const promptTemplate = draft.prompt_template as PromptItem
 | |
|         promptTemplate.jinja2_text = (promptTemplate.jinja2_text || '').replaceAll(`{{ ${oldName} }}`, `{{ ${newName} }}`)
 | |
|       }
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [isChatModel, setInputs])
 | |
| 
 | |
|   // context
 | |
|   const handleContextVarChange = useCallback((newVar: ValueSelector | string) => {
 | |
|     const newInputs = produce(inputs, (draft) => {
 | |
|       draft.context.variable_selector = newVar as ValueSelector || []
 | |
|       draft.context.enabled = !!(newVar && newVar.length > 0)
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [inputs, setInputs])
 | |
| 
 | |
|   const handlePromptChange = useCallback((newPrompt: PromptItem[] | PromptItem) => {
 | |
|     const newInputs = produce(inputRef.current, (draft) => {
 | |
|       draft.prompt_template = newPrompt
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [setInputs])
 | |
| 
 | |
|   const handleMemoryChange = useCallback((newMemory?: Memory) => {
 | |
|     const newInputs = produce(inputs, (draft) => {
 | |
|       draft.memory = newMemory
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [inputs, setInputs])
 | |
| 
 | |
|   const handleSyeQueryChange = useCallback((newQuery: string) => {
 | |
|     const newInputs = produce(inputs, (draft) => {
 | |
|       if (!draft.memory) {
 | |
|         draft.memory = {
 | |
|           window: {
 | |
|             enabled: false,
 | |
|             size: 10,
 | |
|           },
 | |
|           query_prompt_template: newQuery,
 | |
|         }
 | |
|       }
 | |
|       else {
 | |
|         draft.memory.query_prompt_template = newQuery
 | |
|       }
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [inputs, setInputs])
 | |
| 
 | |
|   // structure output
 | |
|   const { data: modelList } = useModelList(ModelTypeEnum.textGeneration)
 | |
|   const isModelSupportStructuredOutput = modelList
 | |
|     ?.find(provideItem => provideItem.provider === model?.provider)
 | |
|     ?.models.find(modelItem => modelItem.model === model?.name)
 | |
|     ?.features?.includes(ModelFeatureEnum.StructuredOutput)
 | |
| 
 | |
|   const [structuredOutputCollapsed, setStructuredOutputCollapsed] = useState(true)
 | |
|   const handleStructureOutputEnableChange = useCallback((enabled: boolean) => {
 | |
|     const newInputs = produce(inputs, (draft) => {
 | |
|       draft.structured_output_enabled = enabled
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|     if (enabled)
 | |
|       setStructuredOutputCollapsed(false)
 | |
|   }, [inputs, setInputs])
 | |
| 
 | |
|   const handleStructureOutputChange = useCallback((newOutput: StructuredOutput) => {
 | |
|     const newInputs = produce(inputs, (draft) => {
 | |
|       draft.structured_output = newOutput
 | |
|     })
 | |
|     setInputs(newInputs)
 | |
|   }, [inputs, setInputs])
 | |
| 
 | |
|   const filterInputVar = useCallback((varPayload: Var) => {
 | |
|     return [VarType.number, VarType.string, VarType.secret, VarType.arrayString, VarType.arrayNumber, VarType.file, VarType.arrayFile].includes(varPayload.type)
 | |
|   }, [])
 | |
| 
 | |
|   const filterJinjia2InputVar = useCallback((varPayload: Var) => {
 | |
|     return [VarType.number, VarType.string, VarType.secret, VarType.arrayString, VarType.arrayNumber].includes(varPayload.type)
 | |
|   }, [])
 | |
| 
 | |
|   const filterMemoryPromptVar = useCallback((varPayload: Var) => {
 | |
|     return [VarType.arrayObject, VarType.array, VarType.number, VarType.string, VarType.secret, VarType.arrayString, VarType.arrayNumber, VarType.file, VarType.arrayFile].includes(varPayload.type)
 | |
|   }, [])
 | |
| 
 | |
|   const {
 | |
|     availableVars,
 | |
|     availableNodesWithParent,
 | |
|   } = useAvailableVarList(id, {
 | |
|     onlyLeafNodeVar: false,
 | |
|     filterVar: filterMemoryPromptVar,
 | |
|   })
 | |
| 
 | |
|   // single run
 | |
|   const {
 | |
|     isShowSingleRun,
 | |
|     hideSingleRun,
 | |
|     getInputVars,
 | |
|     runningStatus,
 | |
|     handleRun,
 | |
|     handleStop,
 | |
|     runInputData,
 | |
|     runInputDataRef,
 | |
|     setRunInputData,
 | |
|     runResult,
 | |
|     toVarInputs,
 | |
|   } = useOneStepRun<LLMNodeType>({
 | |
|     id,
 | |
|     data: inputs,
 | |
|     defaultRunInputData: {
 | |
|       '#context#': [RETRIEVAL_OUTPUT_STRUCT],
 | |
|       '#files#': [],
 | |
|     },
 | |
|   })
 | |
| 
 | |
|   const inputVarValues = (() => {
 | |
|     const vars: Record<string, any> = {}
 | |
|     Object.keys(runInputData)
 | |
|       .filter(key => !['#context#', '#files#'].includes(key))
 | |
|       .forEach((key) => {
 | |
|         vars[key] = runInputData[key]
 | |
|       })
 | |
|     return vars
 | |
|   })()
 | |
| 
 | |
|   const setInputVarValues = useCallback((newPayload: Record<string, any>) => {
 | |
|     const newVars = {
 | |
|       ...newPayload,
 | |
|       '#context#': runInputDataRef.current['#context#'],
 | |
|       '#files#': runInputDataRef.current['#files#'],
 | |
|     }
 | |
|     setRunInputData(newVars)
 | |
|   }, [runInputDataRef, setRunInputData])
 | |
| 
 | |
|   const contexts = runInputData['#context#']
 | |
|   const setContexts = useCallback((newContexts: string[]) => {
 | |
|     setRunInputData({
 | |
|       ...runInputDataRef.current,
 | |
|       '#context#': newContexts,
 | |
|     })
 | |
|   }, [runInputDataRef, setRunInputData])
 | |
| 
 | |
|   const visionFiles = runInputData['#files#']
 | |
|   const setVisionFiles = useCallback((newFiles: any[]) => {
 | |
|     setRunInputData({
 | |
|       ...runInputDataRef.current,
 | |
|       '#files#': newFiles,
 | |
|     })
 | |
|   }, [runInputDataRef, setRunInputData])
 | |
| 
 | |
|   const allVarStrArr = (() => {
 | |
|     const arr = isChatModel ? (inputs.prompt_template as PromptItem[]).filter(item => item.edition_type !== EditionType.jinja2).map(item => item.text) : [(inputs.prompt_template as PromptItem).text]
 | |
|     if (isChatMode && isChatModel && !!inputs.memory) {
 | |
|       arr.push('{{#sys.query#}}')
 | |
|       arr.push(inputs.memory.query_prompt_template)
 | |
|     }
 | |
| 
 | |
|     return arr
 | |
|   })()
 | |
| 
 | |
|   const varInputs = (() => {
 | |
|     const vars = getInputVars(allVarStrArr)
 | |
|     if (isShowVars)
 | |
|       return [...vars, ...toVarInputs(inputs.prompt_config?.jinja2_variables || [])]
 | |
| 
 | |
|     return vars
 | |
|   })()
 | |
| 
 | |
|   return {
 | |
|     readOnly,
 | |
|     isChatMode,
 | |
|     inputs,
 | |
|     isChatModel,
 | |
|     isCompletionModel,
 | |
|     hasSetBlockStatus,
 | |
|     shouldShowContextTip,
 | |
|     isVisionModel,
 | |
|     handleModelChanged,
 | |
|     handleCompletionParamsChange,
 | |
|     isShowVars,
 | |
|     handleVarListChange,
 | |
|     handleVarNameChange,
 | |
|     handleAddVariable,
 | |
|     handleAddEmptyVariable,
 | |
|     handleContextVarChange,
 | |
|     filterInputVar,
 | |
|     filterVar: filterMemoryPromptVar,
 | |
|     availableVars,
 | |
|     availableNodesWithParent,
 | |
|     handlePromptChange,
 | |
|     handleMemoryChange,
 | |
|     handleSyeQueryChange,
 | |
|     handleVisionResolutionEnabledChange,
 | |
|     handleVisionResolutionChange,
 | |
|     isShowSingleRun,
 | |
|     hideSingleRun,
 | |
|     inputVarValues,
 | |
|     setInputVarValues,
 | |
|     visionFiles,
 | |
|     setVisionFiles,
 | |
|     contexts,
 | |
|     setContexts,
 | |
|     varInputs,
 | |
|     runningStatus,
 | |
|     isModelSupportStructuredOutput,
 | |
|     handleStructureOutputChange,
 | |
|     structuredOutputCollapsed,
 | |
|     setStructuredOutputCollapsed,
 | |
|     handleStructureOutputEnableChange,
 | |
|     handleRun,
 | |
|     handleStop,
 | |
|     runResult,
 | |
|     filterJinjia2InputVar,
 | |
|   }
 | |
| }
 | |
| 
 | |
| export default useConfig
 | 
