question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Why is the HSIC not minimized but maximized?

See original GitHub issue

Thank you for such a great work! When reading the paper and code, I have the following questions.

According to the definition of HSIC, it measures the level of independence and HSIC(U, V) = 0 indicates that U and V are independent. A larger HSIC value indicates that U and V are dependent to some extent.

So, to debias the representation of network f by using a biased network g, shouldn’t we minimize HSIC(f, g)?

Besides, the for loop in line 62 seems redundant because the g_dim will be over-written by the last loop, right?

Looking forward to your reply. Thanks!

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:5

github_iconTop GitHub Comments

1reaction
SanghyukChuncommented, Feb 24, 2021

@Cogito2012 In theory, you have to alternatively update to correctly solve a minimax problem, which aims to find a saddle point:

If you jointly optimize min and max problems at the same time, it will not guarantee a correct saddle point. I even cannot sure that such optimization will be converged to a certain point.

Note that our conceptual objective function is

\min_f [L(f) + \max_g HSIC(f, g)]

which have to solve alternative updates for f and g, respectively.

Such minimax problem is popular in many machine learning methods such as

  • Generative adversarial networks (alternatively update generator – max and discriminator – min)
  • Adversarial training (alternatively update the worst case input --max and the model parameter – min)
1reaction
Cogito2012commented, Feb 24, 2021

Let me rephrase my concern. Can we simplify the alternative updating as minimizing the objective?

L(f) + L(g) + HSIC(f, \fixed{g}) - HSIC(\fixed{f}, g)

where \fixed{f} or \fixed{g} means the network parameters of f or g will not be updated by corresponding HSIC term.

If this is also working, then this kind of ReBias variant may be even easier to use.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Cannot click taskbar or window minimize, maximize, and close ...
Since the last week the computer has gone into a mode where the taskbar and window interaction icons cannot be clicked.
Read more >
Feature Selection via Dependence Maximization
Like the mutual in- formation, HSIC is a nonparametric dependence measure, which takes into account all modes of dependence between the variables (not...
Read more >
Eliminating accidental deviations to minimize ... - PLOS
Eliminating accidental deviations to minimize generalization error and maximize replicability: Applications in connectomics and genomics.
Read more >
Self-Supervised Learning with Kernel Dependence ... - arXiv
SSL-HSIC maximizes dependence between representations of transformations of an image and the image identity, while minimizing the kernelized ...
Read more >
Self-Supervised Learning with an Information Maximization ...
Correlative information maximization among alternative latent ... It's true in computer vision, but not true in speech or natural language ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found