0
$\begingroup$

In "Interpretable Machine Learning: A Guide For Making Black Box Models Explainable", I found the following for Friedman's H-statistic: $$PD_{jk}(x_j, x_k) = PD_j (x_j) + PD_k (x_k),$$ where $PD_{jk}(x_j, x_k)$ is the two way partial dependence function of both features and $PD_j (x_j)$ and $PD_k (x_k)$ the partial dependence functions of the single features. Later, the H-statistic is calculated as follows: $$H_{jk}^2 = \frac{\sum_i [PD_{jk}(x_j^{(i)}, x_k^{(i)}) - PD_j(x_j^{(i)}) - PD_k(x_k^{(i)})]^2}{\sum_i PD_{jk}(x_j^{(i)}, x_k^{(i)})}$$

Wouldn't this equation be always zero, when combined with the upper equation?

Looking at the numerator, my thought process is the following: $$PD_{jk}(x_j^{(i)}, x_k^{(i)}) - PD_j(x_j^{(i)}) - PD_k(x_k^{(i)}) = PD_j (x_j^{(i)}) + PD_k (x_k^{(i)}) - PD_j(x_j^{(i)}) - PD_k(x_k^{(i)}) = 0.$$

The chapter can be found here: https://christophm.github.io/interpretable-ml-book/interaction.html

$\endgroup$

1 Answer 1

1
$\begingroup$

This only holds if the features don't interact as is stated in your reference:

"If two features do not interact, we can decompose the partial dependence function as follows (assuming the partial dependence functions are centered at zero): $$𝑃𝐷_{𝑗𝑘}(𝑥_𝑗,𝑥_𝑘)=𝑃𝐷_𝑗(𝑥_𝑗)+𝑃𝐷_𝑘(𝑥_𝑘)." $$

$\endgroup$

You must log in to answer this question.

Not the answer you're looking for? Browse other questions tagged .