Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Introduce our work on the Relaxed Indicator Matrix (RIM) Manifold #65

Open
wants to merge 2 commits into
base: master
Choose a base branch
from

Conversation

Yuan-Jinghui
Copy link

Dear Nicolas Boumal,

Hello! It’s a pleasure to leave a message here. I would like to introduce our work on the Relaxed Indicator Matrix (RIM) Manifold and contribute our open-source code to Manopt.

In clustering and classification, indicator matrices are fundamental concepts, but optimizing them is an NP-hard problem. To address this, machine learning has proposed various relaxation methods:

Relaxation on the Stiefel manifold: F \in { X \mid X^T X = I } , which leads to the well-known spectral clustering algorithm[1].
Relaxation on the single stochastic manifold: F \in { X \mid X 1_c = 1_n, X > 0 }, which derives the famous Fuzzy Kmeans algorithm[2].
Recent progress on the doubly stochastic manifold: ( F \in { X \mid X 1_c = 1_n, X^T 1_n = r, X > 0 }[3].
However, each approach has limitations:

Issue with (1): Slow computation and the need for post-processing (e.g., Kmeans after spectral clustering).
Issue with (2): Unconstrained column sums may lead to empty clusters or imbalanced classifications.
Issue with (3): Requires strong prior knowledge for distribution specification (non-robust) and solving high-dimensional equations (slow).
To address these, we propose the Relaxed Indicator Matrix (RIM) Manifold: F \in { X \mid X 1_c = 1_n, l < X^T 1_n < u, \ X > 0 } which only requires rough estimates of column distributions and generalizes both single and doubly stochastic manifolds. For optimization, RIM enables fast computation (orders of magnitude faster than doubly stochastic methods in experiments with millions of variables) by leveraging the Euclidean metric and Dykstra’s algorithm to efficiently compute geodesics for convergence.

We have open-sourced the code and we will submitt a preprint to arXiv soon (already sent to your email). We hope this work can benefit the open-source community (both machine learning and optimization).

Best regards
Jinghui
[1] Spectral embedded clustering: A framework for in-sample and out-of-sample spectral clustering
[2] Adaptive Fuzzy C-Means with Graph Embedding
[3] Graph Cuts with Arbitrary Size Constraints Through Optimal Transport

Dear Nicolas Boumal,

Hello! It’s a pleasure to leave a message here. I would like to introduce our work on the Relaxed Indicator Matrix (RIM) Manifold and contribute our open-source code to Manopt.

In clustering and classification, indicator matrices are fundamental concepts, but optimizing them is an NP-hard problem. To address this, machine learning has proposed various relaxation methods:

Relaxation on the Stiefel manifold: F \in { X \mid X^T X = I } , which leads to the well-known spectral clustering algorithm[1].
Relaxation on the single stochastic manifold: F \in { X \mid X 1_c = 1_n, X > 0 }, which derives the famous Fuzzy Kmeans algorithm[2].
Recent progress on the doubly stochastic manifold: ( F \in { X \mid X 1_c = 1_n, X^T 1_n = r, X > 0 }[3].
However, each approach has limitations:

Issue with (1): Slow computation and the need for post-processing (e.g., Kmeans after spectral clustering).
Issue with (2): Unconstrained column sums may lead to empty clusters or imbalanced classifications.
Issue with (3): Requires strong prior knowledge for distribution specification (non-robust) and solving high-dimensional equations (slow).

To address these, we propose the Relaxed Indicator Matrix (RIM) Manifold: F \in { X \mid X 1_c = 1_n, l < X^T 1_n < u, \ X > 0 } which only requires rough estimates of column distributions and generalizes both single and doubly stochastic manifolds. For optimization, RIM enables fast computation (orders of magnitude faster than doubly stochastic methods in experiments with millions of variables) by leveraging the Euclidean metric and Dykstra’s algorithm to efficiently compute geodesics for convergence.

We have open-sourced the code and we will submitt a preprint to arXiv soon (already sent to your email). We hope this work can benefit the open-source community (both machine learning and optimization).

Best regards
Jinghui
[1] Spectral embedded clustering: A framework for in-sample and out-of-sample spectral clustering 
[2] Adaptive Fuzzy C-Means with Graph Embedding
[3] Graph Cuts with Arbitrary Size Constraints Through Optimal Transport
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant