Hierarchical Grouping to Optimize an Objective Function

@article{Ward1963HierarchicalGT,
  title={Hierarchical Grouping to Optimize an Objective Function},
  author={Joe H. Ward},
  journal={Journal of the American Statistical Association},
  year={1963},
  volume={58},
  pages={236-244}
}
  • J. H. Ward
  • Published 1 March 1963
  • Mathematics
  • Journal of the American Statistical Association
Abstract A procedure for forming hierarchical groups of mutually exclusive subsets, each of which has members that are maximally similar with respect to specified characteristics, is suggested for use in large-scale (n > 100) studies when a precise optimal solution for a specified number of groups is not practical. Given n sets, this procedure permits their reduction to n − 1 mutually exclusive sets by considering the union of all possible n(n − 1)/2 pairs and selecting a union having a maximal… 

Figures from this paper

CLUSTER: A hierarchical clustering program for large data sets (n>100)

TLDR
CLUSTER is basically an extension of Veldrnan's (1967) HGROUP program, and utilizes Ward's (1963) hierarchical procedure of successively combining n objects together to form n, n 1, n 2, "', 2 groups by minimizing an objective error function.

Integer Programming and the Theory of Grouping

Abstract This paper is written with three objectives in mind. First, to point out that the problem of grouping, where a larger number of elements n are combined into m mutually exclusive groups (m <

Cluster Analysis: An Application of Lagrangian Relaxation

TLDR
It is shown that the optimization algorithm is an effective solution technique for the homogeneous clustering problem, and also a good method for providing tight lower bounds for evaluating the quality of solutions generated by other procedures.

Clique optimization: A method to construct parsimonious ultrametric trees from similarity data

TLDR
A clustering procedure that reduces similarity data to a hierarchy of nested partitions modeled as an ordered sequence of clique sets, CLOPT constructs models that compare favorably with HCS and it does so using fewer parameters.

A biased random-key genetic algorithm for data clustering.

  • P. Festa
  • Computer Science
    Mathematical biosciences
  • 2013

Model selection for minimum-diameter partitioning.

  • M. BruscoD. Steinley
  • Computer Science
    The British journal of mathematical and statistical psychology
  • 2014
TLDR
The practical value of the proposed process for facilitating model selection for the minimum-diameter partitioning problem is shown, using five examples from the empirical literature to show the practical value.

A Repetitive Branch-and-Bound Procedure for Minimum Within-Cluster Sums of Squares Partitioning

TLDR
A new branch-and-bound algorithm for minimizing WCSS is presented that provides optimal solutions for problems with up to 240 objects and eight well-separated clusters and was successfully applied to three empirical data sets from the classification literature.

HSS: Compact Set of Partitions via Hybrid Selection

TLDR
This paper proposes a hybrid multiojective algorithm, HSS, that aims to return a reduced and yet diverse set of solutions that shows the effectiveness of HSS in selecting a reduced number of partitions.

Comparison of single and complete linkage clustering with the hierarchical factor classification of variables

TLDR
The Hierarchical Factor Classification method performed as good as, and in some cases better than, both single and complete linkage clustering in detecting the known group structures in simulated data, with the advantage that the groups of variables and the units can be viewed on principal planes where usual interpretations apply.

Using combinatorial optimization in model-based trimmed clustering with cardinality constraints

...

References

SHOWING 1-5 OF 5 REFERENCES

A HIERARCHICAL GROUPING PROCEDURE APPLIED TO A PROBLEM OF GROUPING PROFILES

Abstract : This report describes an application of a hierarchical grouping procedure to a problem of grouping test profiles to maximize the homogeneity of profiles within the same clusters, taking

AN ITERATIVE TECHNIQUE FOR CLUSTERING CRITERIA WHICH RETAINS OPTIMUM PREDICTIVE EFFICIENCY

TLDR
This paper describes an iterative technique, programmed for an electronic computer, that at each step r duces the number of criterion clusters and provides optimal weights for the tests and a cost table can be used to determine the appropriate balance between predictive efficiency and number of criteria clusters.

Note on Grouping

Abstract Suppose that it is required to condense observations of a variate into a small number of groups, the grouping intervals to be chosen to retain as much information as possible. One way of

Who belongs in the family?

I was sitting before my TV set, a while back, watching Captain Video and pondering the organizational problems of psychologists, psychometricians, psychodiagnosticians, psycho-somatists,

On Grouping for Maximum Homogeneity

Abstract Given a set of arbitrary numbers, what is a practical procedure for grouping them so that the variance within groups is minimized? An answer to this question, including a description of an