Title: | Pairwise Constraints Clustering |
---|---|
Description: | There are 4 main functions in this package: ckmeans(), lcvqe(), mpckm() and ccls(). They take an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output. |
Authors: | Tran Khanh Hiep, Nguyen Minh Duc |
Maintainer: | Tran Khanh Hiep <[email protected]> |
License: | GPL-3 |
Version: | 1.1 |
Built: | 2024-12-09 06:42:06 UTC |
Source: | CRAN |
There are 4 main functions in this package: ckmeans(), lcvqe(), mpckm() and ccls(). They take an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output.
The DESCRIPTION file:
Package: | conclust |
Type: | Package |
Title: | Pairwise Constraints Clustering |
Version: | 1.1 |
Date: | 2016-08-15 |
Author: | Tran Khanh Hiep, Nguyen Minh Duc |
Maintainer: | Tran Khanh Hiep <[email protected]> |
Description: | There are 4 main functions in this package: ckmeans(), lcvqe(), mpckm() and ccls(). They take an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output. |
License: | GPL-3 |
NeedsCompilation: | no |
Packaged: | 2016-08-15 08:45:24 UTC; Hiep |
Repository: | CRAN |
Date/Publication: | 2016-08-15 13:16:21 |
Index of help topics:
ccls Pairwise Constrained Clustering by Local Search ckmeans COP K-means algorithm conclust-package Pairwise Constraints Clustering lcvqe LCVQE algorithm mpckm MPC K-means algorithm
There are 4 main functions in this package: ckmeans(), lcvqe(), mpckm() and ccls(). They take an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output.
Tran Khanh Hiep, Nguyen Minh Duc
Maintainer: Tran Khanh Hiep <[email protected]>
Wagstaff, Cardie, Rogers, Schrodl (2001), Constrained K-means Clustering with Background Knowledge Bilenko, Basu, Mooney (2004), Integrating Constraints and Metric Learning in Semi-Supervised Clustering Dan Pelleg, Dorit Baras (2007), K-means with large and noisy constraint sets
Wagstaff, Cardie, Rogers, Schrodl (2001), Constrained K-means Clustering with Background Knowledge Bilenko, Basu, Mooney (2004), Integrating Constraints and Metric Learning in Semi-Supervised Clustering Dan Pelleg, Dorit Baras (2007), K-means with large and noisy constraint sets
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = ckmeans(data, k, mustLink, cantLink) pred pred = mpckm(data, k, mustLink, cantLink) pred pred = lcvqe(data, k, mustLink, cantLink) pred pred = ccls(data, k, mustLink, cantLink) pred
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = ckmeans(data, k, mustLink, cantLink) pred pred = mpckm(data, k, mustLink, cantLink) pred pred = lcvqe(data, k, mustLink, cantLink) pred pred = ccls(data, k, mustLink, cantLink) pred
This function takes an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output.
ccls(data, k = -1, mustLink, cantLink, maxIter = 1, tabuIter = 100, tabuLength = 20)
ccls(data, k = -1, mustLink, cantLink, maxIter = 1, tabuIter = 100, tabuLength = 20)
data |
The unlabeled dataset. |
k |
Number of clusters. |
mustLink |
A list of must-link constraints |
cantLink |
A list of cannot-link constraints |
maxIter |
Number of iteration |
tabuIter |
Number of iteration in Tabu search |
tabuLength |
The number of elements in the Tabu list |
This algorithm minimizes the clustering cost function using Tabu search.
A vector that represents the labels (clusters) of the data points
This is the first algorithm for pairwise constrained clustering by local search.
Tran Khanh Hiep Nguyen Minh Duc
Tran Khanh Hiep, Nguyen Minh Duc, Bui Quoc Trung (2016), Pairwise Constrained Clustering by Local Search.
Tran Khanh Hiep, Nguyen Minh Duc, Bui Quoc Trung (2016), Pairwise Constrained Clustering by Local Search.
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = ckmeans(data, k, mustLink, cantLink) pred
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = ckmeans(data, k, mustLink, cantLink) pred
This function takes an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output.
ckmeans(data, k, mustLink, cantLink, maxIter = 100)
ckmeans(data, k, mustLink, cantLink, maxIter = 100)
data |
The unlabeled dataset. |
k |
Number of clusters. |
mustLink |
A list of must-link constraints |
cantLink |
A list of cannot-link constraints |
maxIter |
Number of iteration |
This algorithm produces a clustering that satisfies all given constraints.
A vector that represents the labels (clusters) of the data points
The constraints should be consistent in order for the algorithm to work.
Tran Khanh Hiep Nguyen Minh Duc
Wagstaff, Cardie, Rogers, Schrodl (2001), Constrained K-means Clustering with Background Knowledge
Wagstaff, Cardie, Rogers, Schrodl (2001), Constrained K-means Clustering with Background Knowledge
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = ckmeans(data, k, mustLink, cantLink) pred
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = ckmeans(data, k, mustLink, cantLink) pred
This function takes an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output.
lcvqe(data, k, mustLink, cantLink, maxIter = 10)
lcvqe(data, k, mustLink, cantLink, maxIter = 10)
data |
The unlabeled dataset. |
k |
Number of clusters. |
mustLink |
A list of must-link constraints |
cantLink |
A list of cannot-link constraints |
maxIter |
Number of iteration |
This algorithm finds a clustering that satisfies as many constraints as possible
A vector that represents the labels (clusters) of the data points
This algorithm can handle noisy constraints.
Tran Khanh Hiep Nguyen Minh Duc
Dan Pelleg, Dorit Baras (2007), K-means with large and noisy constraint sets
Dan Pelleg, Dorit Baras (2007), K-means with large and noisy constraint sets
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = lcvqe(data, k, mustLink, cantLink) pred
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = lcvqe(data, k, mustLink, cantLink) pred
This function takes an unlabeled dataset and two lists of must-link and cannot-link constraints as input and produce a clustering as output.
mpckm(data, k, mustLink, cantLink, maxIter = 10)
mpckm(data, k, mustLink, cantLink, maxIter = 10)
data |
The unlabeled dataset. |
k |
Number of clusters. |
mustLink |
A list of must-link constraints |
cantLink |
A list of cannot-link constraints |
maxIter |
Number of iteration |
This algorithm finds a clustering that satisfies as many constraints as possible
A vector that represents the labels (clusters) of the data points
This is one of the best algorithm for clustering with constraints.
Tran Khanh Hiep Nguyen Minh Duc
Bilenko, Basu, Mooney (2004), Integrating Constraints and Metric Learning in Semi-Supervised Clustering
Bilenko, Basu, Mooney (2004), Integrating Constraints and Metric Learning in Semi-Supervised Clustering
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = mpckm(data, k, mustLink, cantLink) pred
data = matrix(c(0, 1, 1, 0, 0, 0, 1, 1), nrow = 4) mustLink = matrix(c(1, 2), nrow = 1) cantLink = matrix(c(1, 4), nrow = 1) k = 2 pred = mpckm(data, k, mustLink, cantLink) pred