This node supplies an LLM agent with a set of tools and the conversation history, and prompts it with the user-provided query.
The conversation table is expected to have at least two string columns that define previous conversation. If this is the start of the conversation, the conversation table can be empty.
The agent always receives the full conversation table as context, which can lead to slower execution times for longer conversations, or execution failures if the context becomes too large for the LLM. If you experience such issues, you can truncate the conversation table by only keeping the last few messages, or use an LLM to summarize the conversation held so far.
If checked, prints the output of LangChain's debug mode to the console.
Select the column that specifies the alternating sender roles assigned to each message. Example values are 'Human' and 'AI'.
Select the column containing the messages composing the conversation history.
Specify the new message to be sent to the agent.
You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.
To use this node in KNIME, install the extension KNIME Python Extension Development (Labs) from the below update site following our NodePit Product and Node Installation Guide:
A zipped version of the software site can be downloaded here.
Deploy, schedule, execute, and monitor your KNIME workflows locally, in the cloud or on-premises – with our brand new NodePit Runner.
Try NodePit Runner!Do you have feedback, questions, comments about NodePit, want to support this platform, or want your own nodes or workflows listed here as well? Do you think, the search results could be improved or something is missing? Then please get in touch! Alternatively, you can send us an email to mail@nodepit.com.
Please note that this is only about NodePit. We do not provide general support for KNIME — please use the KNIME forums instead.