LLM Prompter

For each row in the input table, this node sends one prompt to the LLM and receives a corresponding response. Rows and the corresponding prompts are treated in isolation, i.e. the LLM cannot remember the contents of the previous rows or how it responded to them.

Note: If you use the Credentials Configuration node and do not select the "Save password in configuration (weakly encrypted)" option for passing the API key for the LLM connector node, the Credentials Configuration node will need to be reconfigured upon reopening the workflow, as the credentials flow variable was not saved and will therefore not be available to downstream nodes.

Options

Add system message

Specify whether a customizable system message is included in each prompt. This option is only available for chat models.

Available options:

  • None: No system message will precede the prompts.
  • Global: A specifiable system message will precede each prompt.
  • Column: Each prompt includes an individual system message specified in a column.
System message

The first message given to the model describing how it should behave.

Example: You are a helpful assistant that has to answer questions truthfully, and if you do not know an answer to a question, you should state that.

System message column

The column containing the system message for each prompt.

Prompt column

Column containing prompts for the LLM.

Response column name

Name for the column holding the LLM's responses.

Input Ports

Icon

A large language model or chat model.

Icon

A table containing a string column with prompts.

Output Ports

Icon

A table containing prompts and their respective answer.

Popular Predecessors

  • No recommendations found

Popular Successors

  • No recommendations found

Views

This node has no views

Workflows

  • No workflows found

Links

Developers

You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.