Calculates the entropy uncertainty score of a class probability distribution. Input are rows containing class probabilities P = p1, p2, ..., pn that must sum up to 1. Output will be the normalized Shannon entropy . This is defined by E(P) = H(P) / log(n) with H(P) = - sum(p_i*log(p_i) for each i in 1,...,n. The logarithm with base 2 is used. The normalization leads always to values between 0 and 1. A uniform probability distribution (i.e., most uncertain as all probabilities are equal to each other) has an entropy value of 1. If one of the class probabilities is 1 and the others 0, the highest certainty is given and the entropy value will be 0.
You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.
A zipped version of the software site can be downloaded here.
Deploy, schedule, execute, and monitor your KNIME workflows locally, in the cloud or on-premises – with our brand new NodePit Runner.Try NodePit Runner!
Do you have feedback, questions, comments about NodePit, want to support this platform, or want your own nodes or workflows listed here as well? Do you think, the search results could be improved or something is missing? Then please get in touch! Alternatively, you can send us an email to email@example.com, follow @NodePit on Twitter or botsin.space/@nodepit on Mastodon.
Please note that this is only about NodePit. We do not provide general support for KNIME — please use the KNIME forums instead.