mirror of
https://github.com/langgenius/dify.git
synced 2024-11-16 11:42:29 +08:00
chore: massive update of the Gemini models based on latest documentation (#8822)
This commit is contained in:
parent
ecc951609d
commit
e5efd09ebb
|
@ -0,0 +1,48 @@
|
||||||
|
model: gemini-1.5-flash-001
|
||||||
|
label:
|
||||||
|
en_US: Gemini 1.5 Flash 001
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 1048576
|
||||||
|
parameter_rules:
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
- name: top_p
|
||||||
|
use_template: top_p
|
||||||
|
- name: top_k
|
||||||
|
label:
|
||||||
|
zh_Hans: 取样数量
|
||||||
|
en_US: Top k
|
||||||
|
type: int
|
||||||
|
help:
|
||||||
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
|
||||||
|
en_US: Only sample from the top K options for each subsequent token.
|
||||||
|
required: false
|
||||||
|
- name: max_tokens_to_sample
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 8192
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
- name: stream
|
||||||
|
label:
|
||||||
|
zh_Hans: 流式输出
|
||||||
|
en_US: Stream
|
||||||
|
type: boolean
|
||||||
|
help:
|
||||||
|
zh_Hans: 流式输出允许模型在生成文本的过程中逐步返回结果,而不是一次性生成全部结果后再返回。
|
||||||
|
en_US: Streaming output allows the model to return results incrementally as it generates text, rather than generating all the results at once.
|
||||||
|
default: false
|
||||||
|
pricing:
|
||||||
|
input: '0.00'
|
||||||
|
output: '0.00'
|
||||||
|
unit: '0.000001'
|
||||||
|
currency: USD
|
|
@ -0,0 +1,48 @@
|
||||||
|
model: gemini-1.5-flash-002
|
||||||
|
label:
|
||||||
|
en_US: Gemini 1.5 Flash 002
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 1048576
|
||||||
|
parameter_rules:
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
- name: top_p
|
||||||
|
use_template: top_p
|
||||||
|
- name: top_k
|
||||||
|
label:
|
||||||
|
zh_Hans: 取样数量
|
||||||
|
en_US: Top k
|
||||||
|
type: int
|
||||||
|
help:
|
||||||
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
|
||||||
|
en_US: Only sample from the top K options for each subsequent token.
|
||||||
|
required: false
|
||||||
|
- name: max_tokens_to_sample
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 8192
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
- name: stream
|
||||||
|
label:
|
||||||
|
zh_Hans: 流式输出
|
||||||
|
en_US: Stream
|
||||||
|
type: boolean
|
||||||
|
help:
|
||||||
|
zh_Hans: 流式输出允许模型在生成文本的过程中逐步返回结果,而不是一次性生成全部结果后再返回。
|
||||||
|
en_US: Streaming output allows the model to return results incrementally as it generates text, rather than generating all the results at once.
|
||||||
|
default: false
|
||||||
|
pricing:
|
||||||
|
input: '0.00'
|
||||||
|
output: '0.00'
|
||||||
|
unit: '0.000001'
|
||||||
|
currency: USD
|
|
@ -0,0 +1,48 @@
|
||||||
|
model: gemini-1.5-flash-8b-exp-0924
|
||||||
|
label:
|
||||||
|
en_US: Gemini 1.5 Flash 8B 0924
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 1048576
|
||||||
|
parameter_rules:
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
- name: top_p
|
||||||
|
use_template: top_p
|
||||||
|
- name: top_k
|
||||||
|
label:
|
||||||
|
zh_Hans: 取样数量
|
||||||
|
en_US: Top k
|
||||||
|
type: int
|
||||||
|
help:
|
||||||
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
|
||||||
|
en_US: Only sample from the top K options for each subsequent token.
|
||||||
|
required: false
|
||||||
|
- name: max_tokens_to_sample
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 8192
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
- name: stream
|
||||||
|
label:
|
||||||
|
zh_Hans: 流式输出
|
||||||
|
en_US: Stream
|
||||||
|
type: boolean
|
||||||
|
help:
|
||||||
|
zh_Hans: 流式输出允许模型在生成文本的过程中逐步返回结果,而不是一次性生成全部结果后再返回。
|
||||||
|
en_US: Streaming output allows the model to return results incrementally as it generates text, rather than generating all the results at once.
|
||||||
|
default: false
|
||||||
|
pricing:
|
||||||
|
input: '0.00'
|
||||||
|
output: '0.00'
|
||||||
|
unit: '0.000001'
|
||||||
|
currency: USD
|
|
@ -1,6 +1,6 @@
|
||||||
model: gemini-1.5-flash-latest
|
model: gemini-1.5-flash-latest
|
||||||
label:
|
label:
|
||||||
en_US: Gemini 1.5 Flash
|
en_US: Gemini 1.5 Flash Latest
|
||||||
model_type: llm
|
model_type: llm
|
||||||
features:
|
features:
|
||||||
- agent-thought
|
- agent-thought
|
||||||
|
|
|
@ -0,0 +1,48 @@
|
||||||
|
model: gemini-1.5-flash
|
||||||
|
label:
|
||||||
|
en_US: Gemini 1.5 Flash
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 1048576
|
||||||
|
parameter_rules:
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
- name: top_p
|
||||||
|
use_template: top_p
|
||||||
|
- name: top_k
|
||||||
|
label:
|
||||||
|
zh_Hans: 取样数量
|
||||||
|
en_US: Top k
|
||||||
|
type: int
|
||||||
|
help:
|
||||||
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
|
||||||
|
en_US: Only sample from the top K options for each subsequent token.
|
||||||
|
required: false
|
||||||
|
- name: max_tokens_to_sample
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 8192
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
- name: stream
|
||||||
|
label:
|
||||||
|
zh_Hans: 流式输出
|
||||||
|
en_US: Stream
|
||||||
|
type: boolean
|
||||||
|
help:
|
||||||
|
zh_Hans: 流式输出允许模型在生成文本的过程中逐步返回结果,而不是一次性生成全部结果后再返回。
|
||||||
|
en_US: Streaming output allows the model to return results incrementally as it generates text, rather than generating all the results at once.
|
||||||
|
default: false
|
||||||
|
pricing:
|
||||||
|
input: '0.00'
|
||||||
|
output: '0.00'
|
||||||
|
unit: '0.000001'
|
||||||
|
currency: USD
|
|
@ -0,0 +1,48 @@
|
||||||
|
model: gemini-1.5-pro-001
|
||||||
|
label:
|
||||||
|
en_US: Gemini 1.5 Pro 001
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 2097152
|
||||||
|
parameter_rules:
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
- name: top_p
|
||||||
|
use_template: top_p
|
||||||
|
- name: top_k
|
||||||
|
label:
|
||||||
|
zh_Hans: 取样数量
|
||||||
|
en_US: Top k
|
||||||
|
type: int
|
||||||
|
help:
|
||||||
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
|
||||||
|
en_US: Only sample from the top K options for each subsequent token.
|
||||||
|
required: false
|
||||||
|
- name: max_tokens_to_sample
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 8192
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
- name: stream
|
||||||
|
label:
|
||||||
|
zh_Hans: 流式输出
|
||||||
|
en_US: Stream
|
||||||
|
type: boolean
|
||||||
|
help:
|
||||||
|
zh_Hans: 流式输出允许模型在生成文本的过程中逐步返回结果,而不是一次性生成全部结果后再返回。
|
||||||
|
en_US: Streaming output allows the model to return results incrementally as it generates text, rather than generating all the results at once.
|
||||||
|
default: false
|
||||||
|
pricing:
|
||||||
|
input: '0.00'
|
||||||
|
output: '0.00'
|
||||||
|
unit: '0.000001'
|
||||||
|
currency: USD
|
|
@ -0,0 +1,48 @@
|
||||||
|
model: gemini-1.5-pro-002
|
||||||
|
label:
|
||||||
|
en_US: Gemini 1.5 Pro 002
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 2097152
|
||||||
|
parameter_rules:
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
- name: top_p
|
||||||
|
use_template: top_p
|
||||||
|
- name: top_k
|
||||||
|
label:
|
||||||
|
zh_Hans: 取样数量
|
||||||
|
en_US: Top k
|
||||||
|
type: int
|
||||||
|
help:
|
||||||
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
|
||||||
|
en_US: Only sample from the top K options for each subsequent token.
|
||||||
|
required: false
|
||||||
|
- name: max_tokens_to_sample
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 8192
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
- name: stream
|
||||||
|
label:
|
||||||
|
zh_Hans: 流式输出
|
||||||
|
en_US: Stream
|
||||||
|
type: boolean
|
||||||
|
help:
|
||||||
|
zh_Hans: 流式输出允许模型在生成文本的过程中逐步返回结果,而不是一次性生成全部结果后再返回。
|
||||||
|
en_US: Streaming output allows the model to return results incrementally as it generates text, rather than generating all the results at once.
|
||||||
|
default: false
|
||||||
|
pricing:
|
||||||
|
input: '0.00'
|
||||||
|
output: '0.00'
|
||||||
|
unit: '0.000001'
|
||||||
|
currency: USD
|
|
@ -1,6 +1,6 @@
|
||||||
model: gemini-1.5-pro-latest
|
model: gemini-1.5-pro-latest
|
||||||
label:
|
label:
|
||||||
en_US: Gemini 1.5 Pro
|
en_US: Gemini 1.5 Pro Latest
|
||||||
model_type: llm
|
model_type: llm
|
||||||
features:
|
features:
|
||||||
- agent-thought
|
- agent-thought
|
||||||
|
|
|
@ -0,0 +1,48 @@
|
||||||
|
model: gemini-1.5-pro
|
||||||
|
label:
|
||||||
|
en_US: Gemini 1.5 Pro
|
||||||
|
model_type: llm
|
||||||
|
features:
|
||||||
|
- agent-thought
|
||||||
|
- vision
|
||||||
|
- tool-call
|
||||||
|
- stream-tool-call
|
||||||
|
model_properties:
|
||||||
|
mode: chat
|
||||||
|
context_size: 2097152
|
||||||
|
parameter_rules:
|
||||||
|
- name: temperature
|
||||||
|
use_template: temperature
|
||||||
|
- name: top_p
|
||||||
|
use_template: top_p
|
||||||
|
- name: top_k
|
||||||
|
label:
|
||||||
|
zh_Hans: 取样数量
|
||||||
|
en_US: Top k
|
||||||
|
type: int
|
||||||
|
help:
|
||||||
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
|
||||||
|
en_US: Only sample from the top K options for each subsequent token.
|
||||||
|
required: false
|
||||||
|
- name: max_tokens_to_sample
|
||||||
|
use_template: max_tokens
|
||||||
|
required: true
|
||||||
|
default: 8192
|
||||||
|
min: 1
|
||||||
|
max: 8192
|
||||||
|
- name: response_format
|
||||||
|
use_template: response_format
|
||||||
|
- name: stream
|
||||||
|
label:
|
||||||
|
zh_Hans: 流式输出
|
||||||
|
en_US: Stream
|
||||||
|
type: boolean
|
||||||
|
help:
|
||||||
|
zh_Hans: 流式输出允许模型在生成文本的过程中逐步返回结果,而不是一次性生成全部结果后再返回。
|
||||||
|
en_US: Streaming output allows the model to return results incrementally as it generates text, rather than generating all the results at once.
|
||||||
|
default: false
|
||||||
|
pricing:
|
||||||
|
input: '0.00'
|
||||||
|
output: '0.00'
|
||||||
|
unit: '0.000001'
|
||||||
|
currency: USD
|
Loading…
Reference in New Issue
Block a user