mirror of
https://github.com/n8n-io/n8n.git
synced 2024-11-09 22:24:05 -08:00
Default max token to 128 + minor copy updates.
This commit is contained in:
parent
6500e985ee
commit
acc43c4ea1
|
@ -68,6 +68,7 @@ const completeOperations: INodeProperties[] = [
|
|||
sortable: true,
|
||||
multipleValues: true,
|
||||
},
|
||||
description: 'The prompt can be a list of messages or a single message, depending on what you want to achieve and the model you are using. <a href="https://platform.openai.com/docs/guides/chat">More info</a>.',
|
||||
displayOptions: {
|
||||
show: {
|
||||
resource: ['chat'],
|
||||
|
@ -89,14 +90,17 @@ const completeOperations: INodeProperties[] = [
|
|||
{
|
||||
name: 'Assistant',
|
||||
value: 'assistant',
|
||||
description: 'Store prior responses',
|
||||
},
|
||||
{
|
||||
name: 'System',
|
||||
value: 'system',
|
||||
description: 'Set the behavior of the assistant',
|
||||
},
|
||||
{
|
||||
name: 'User',
|
||||
value: 'user',
|
||||
description: 'Instruct the assistant',
|
||||
},
|
||||
],
|
||||
default: 'user',
|
||||
|
@ -218,7 +222,7 @@ const sharedOperations: INodeProperties[] = [
|
|||
{
|
||||
displayName: 'Maximum Number of Tokens',
|
||||
name: 'maxTokens',
|
||||
default: 16,
|
||||
default: 128,
|
||||
description:
|
||||
'The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 4096).',
|
||||
type: 'number',
|
||||
|
|
|
@ -59,7 +59,7 @@ const completeOperations: INodeProperties[] = [
|
|||
name: 'model',
|
||||
type: 'options',
|
||||
description:
|
||||
'The model which will generate the completion. <a href="https://beta.openai.com/docs/models/overview">Learn more</a>.',
|
||||
'The model which will generate the completion. <a href="https://beta.openai.com/docs/models/overview">More info</a>.',
|
||||
displayOptions: {
|
||||
show: {
|
||||
operation: ['complete'],
|
||||
|
@ -145,7 +145,7 @@ const editOperations: INodeProperties[] = [
|
|||
name: 'model',
|
||||
type: 'options',
|
||||
description:
|
||||
'The model which will generate the edited version. <a href="https://beta.openai.com/docs/models/overview">Learn more</a>.',
|
||||
'The model which will generate the edited version. <a href="https://beta.openai.com/docs/models/overview">More info</a>.',
|
||||
displayOptions: {
|
||||
show: {
|
||||
operation: ['edit'],
|
||||
|
@ -218,7 +218,7 @@ const moderateOperations: INodeProperties[] = [
|
|||
name: 'model',
|
||||
type: 'options',
|
||||
description:
|
||||
'The model which will classify the text. <a href="https://beta.openai.com/docs/models/overview">Learn more</a>.',
|
||||
'The model which will classify the text. <a href="https://beta.openai.com/docs/models/overview">More info</a>.',
|
||||
displayOptions: {
|
||||
show: {
|
||||
resource: ['text'],
|
||||
|
@ -397,7 +397,7 @@ const sharedOperations: INodeProperties[] = [
|
|||
{
|
||||
displayName: 'Maximum Number of Tokens',
|
||||
name: 'maxTokens',
|
||||
default: 16,
|
||||
default: 128,
|
||||
description:
|
||||
'The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 4096).',
|
||||
type: 'number',
|
||||
|
|
Loading…
Reference in a new issue