Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

The Hellinger distance is a metric used to measure the similarity between two probability distributions. It is related to the Euclidean distance but applied in the space of probability distributions. The Hellinger distance ranges between 0 and 1, where 0 indicates that the two distributions are identical, and 1 indicates that they are completely dissimilar.

 

Let P and Q denote two probability measures on a measure space X with corresponding probability density functions p(x) and q(x). The Hellinger distance H(P, Q) between these two distributions is defined as follows:

 

For continuous probability distributions:

 

H(P, Q) = (1 / sqrt(2)) * [∫ (sqrt(p(x)) - sqrt(q(x)))^2 dx]^(1/2)

 

For discrete probability distributions:

 

H(P, Q) = (1 / sqrt(2)) * sqrt(∑ (sqrt(p_i) - sqrt(q_i))^2)

 

Alternatively, the discrete version can be written more compactly as:

 

H(P, Q) = (1 / sqrt(2)) * ||sqrt(P) - sqrt(Q)||_2

 

where || · ||_2 denotes the Euclidean norm.

 

Properties of the Hellinger Distance

 

Bounded Metric: The Hellinger distance satisfies the property 0 ≤ H(P, Q) ≤ 1.

Symmetry: It is symmetric, meaning H(P, Q) = H(Q, P).

Metric Properties: It satisfies the triangle inequality, making it a true metric.

Maximum Distance: The maximum value of 1 occurs when the two distributions have no overlap.

 

Relationship with Other Measures

 

The Hellinger distance is closely related to the Bhattacharyya Coefficient but differs in that the Hellinger distance is a proper metric (it satisfies the triangle inequality). This makes the Hellinger distance easier to work with in certain mathematical contexts.

 

Application

 

The Hellinger distance is used in various fields, including statistical inference, machine learning, and information theory, to quantify the difference between probability distributions. It is especially useful in scenarios where one needs to measure the divergence between two distributions that might represent different scenarios, populations, or experimental conditions.

References

1. Nikulin, M.S., 2001. Hellinger distance. In Encyclopedia of Mathematics. EMS Press. 2. Hellinger, E., 1909. Neue Begründung der Theorie quadratischer Formen von unendlichvielen Veränderlichen. Journal für die reine und angewandte Mathematik, 136, pp.210–271. 3. Jeffreys, H., 1946. An invariant form for the prior probability in estimation problems. Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences, 186(1007), pp.453–461. 4. Liese, F. and Miescke, K.-J., 2008. Statistical Decision Theory: Estimation, Testing, and Selection. Springer. 5. Pardo, L., 2006. Statistical Inference Based on Divergence Measures. Chapman and Hall/CRC.
catalogue Logos

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.