Quentin F. Stout

Computer Science and Engineering, University of Michigan

**Abstract:**
This paper gives an approach for computing isotonic regressions
at points in multidimensional space, with the ordering given by
domination (also known as component-wise ordering, dimensional ordering, majorization,
or matrix ordering).
That is, given points p=(p_{1},...,p_{d}) and
q=(q_{1},...,q_{d}) in d-dimensional space, q is said
to *dominate* p if p_{i} ≤ q_{i} for all
1 ≤i ≤ d.
Expressing this in standard dag (directed acyclic graph) terminology, or, equivalently, terminology of partial orderings,
q dominates p is the same as q *succeeds* p.

Given a set of n d-dimensional points
*V* = {v_{1}...v_{n}} where point v_{i}
has data value y_{i} and weight w_{i}, the
*L _{p} isotonic regression of the data* is a set of values
ý

max { w

Note that each dimension need only be a linearly ordered set. For example, one may be height, another shoe size, and a third S < M < L < XL shirt size, where y may correspond to weight. The isotonic requirement is based on the reasonable assumption that if any two of the dimensions are held constant then an increase in the third will increase the expected weight, or, at least, not decrease it.

The key to the results in this paper is efficiently computing an embedding of
V into a larger dag G=(V∪R, E).
G is bipartite, with every edge having one endpoint in V and one in R,
and for all
u, v ∈ V, v dominates u iff there is an r ∈ R such that
(u,r), (r,v) ∈ E.
We call G a *rendezvous graph*, and r is the unique rendezvous vertex for u and v.
R and E are only slightly larger than V, with both having size
Θ(n log^{d} n), and they can be constructed in time linear in their size.
Some of the algorithms exploit the fact that G is not much larger than
V, while others use its structure in a more central way.
By using the rendezvous vertices operations are performed over the
transitive closure more efficiently than they can be performed over the
transitive closure itself.
We also give a reduced rendezvous graph G^{'} with Θ(n log ^{d-1}n ) vertices and edges, where for all u, v ∈ V, v dominates u iff there is a path in G^{'} from u to v.
See the note below concerning the optimality of G^{'}.

Using G, algorithms are given for the L_{1} and L_{2}
metrics which are n/poly-log n faster than previous
algorithms, taking approximately Θ(n^{2}) and Θ(n^{3}) time, respectively.
A yet faster algorithm is given for L_{1} isotonic regression
with unweighted data.
L_{∞} is not unique, so algorithms are given for several
different variants that have been studied.
This includes the strict L_{∞} isotonic regression, which
is the limit, as p→∞, of L_{p} isotonic regression.
These too are n/poly-log n faster than previous algorithms, taking approximately Θ(n) time.

Note that another approach to minimizing the representation of the transitive closure of a dag is via its transitive reduction, i.e., the smallest subset of edges that has the same transitive closure as the original dag.
For some sets of multidimensional points this can be smaller than the rendezvous graph, e.g., if the points in 2-space are (1,1), (2,2), ... (n,n) then the transitive reduction has n-1 edges, not Θ(n log^{2} n).
However, the worst case of the transitive closure is Θ(n^{2}), e.g., if the data is (-5,-1), (-4, -2), (-3, -3), (-2, -4), (-1,-5), (1,4), (2,3), (3,2), (4,1) then the last 4 points dominate the first 5, and the transitive reduction has 20 edges.
This can be reduced by adding the point (0,0), which dominates the first 5 points and is dominated by the second 4, and the transitive reduction has 9 edges.
This is the essential concept of the rendezvous graph.

This embedding approach had first been used for
L_{∞} isotonic regression, posted on the web in 2008.
That paper was too long so I moved the approach to this paper,
extending it to other metrics and to the strict L_{∞}
isotonic regression.

Note: An open question is whether the size of the reduced rendezvous graph mentioned above is optimal.
That is, is it true that for any d there are sets S(n) of n d-dimensional points such that for any dag G containing S(n) among its vertices, where G has the property that for all u, v in S(n), v dominates u iff there is a path in G from u to v, then G has Ω(n log^{d-1} n) edges, where the implied constants depend upon d?
The paper shows that this is true for d=2, and it is trivially true for d=1, but for larger d appears to be an open question.
A related result (not mentioned in the paper) is that given a random set of n d-dimensional points, where in each dimension all of the coordinates are different and they are in random order independent of the other dimensions, then the expected number of edges in the transitive reduction is Θ(n log^{d-1} n), where the implied constants depend upon d.
This follows from a result of Bentley et al. (JACM 1978, pp. 536-543).
Modifying the constructions in the paper shows that the transitive reduction can be computed in Θ(R + n log^{d-1} n) time, where R is the size of the transitive reduction.
I haven't searched the literature, but the optimal time to find the transitive reduction in this setting may be an open question.

**Keywords:** weighted isotonic regression, isotonic regression algorithm,
monotonic, multidimensional, nonparametric, order-embedding, shape-constrained, transitive closure

**Complete paper**. This appears in *Algorithmica* 71 (2015), pp. 450-470.

Here are papers on
L_{∞} isotonic regression for arbitrary dags (also known as minimax regression),
L_{∞} isotonic regression for linear, multidimensional, and tree orders,
strict L_{∞} isotonic regression,
L_{1}isotonic regression (also known as median regression),
isotonic regression with a fixed number of level sets, and
unimodal regression.

Copyright © 2009-2016 Quentin F. Stout. |