Connects to a Large Language Model (LLM) configured in the GenAI Gateway of the connected KNIME Hub using the authentication provided via the input port.
Use this node to generate text, answer questions, summarize content or perform other text-based tasks.
Select the ID of the LLM to use. If set via flow variable, the ID can be obtained from the KNIME Hub AI Model Lister node.
The maximum number of tokens to generate.
This value, plus the token count of your prompt, cannot exceed the model's context length.
Sampling temperature to use, between 0.0 and 2.0.
Higher values will lead to less deterministic answers.
Try 0.9 for more creative applications, and 0 for ones with a well-defined answer. It is generally recommended altering this, or Top-p, but not both.
Maximum number of concurrent requests a single node (e.g. LLM Prompter (Table)) can make to the GenAI Gateway. The more requests a node can make in parallel, the faster it executes. Too many requests might get rate-limited by some GenAI providers.
An alternative to sampling with temperature, where the model considers the results of the tokens (words) with top_p probability mass. Hence, 0.1 means only the tokens comprising the top 10% probability mass are considered.
You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.
To use this node in KNIME, install the extension KNIME Python Extension Development (Labs) from the below update site following our NodePit Product and Node Installation Guide:
A zipped version of the software site can be downloaded here.
Deploy, schedule, execute, and monitor your KNIME workflows locally, in the cloud or on-premises – with our brand new NodePit Runner.
Try NodePit Runner!