Skip to content

Galactic Sorting Networks

April 24, 2014

An improved sorting network to appear at STOC 2014

Teaching page source

Michael Goodrich is a Chancellor’s Professor at the University of California at Irvine, and recently served a year as Chair of the UCI Computer Science Department. He has designed algorithms that cover all of the area spanned by axes of cleverness and high performance. This includes attention to algorithms that we have called (clever but) galactic for having enormous constant factors and/or high exponents that forestall practical implementation. An example is his paper last year with his colleague Pawel Pszona, “Cole’s Parametric Search Technique Made Practical,” which greatly tunes up an algorithm-improvement technique of Richard Cole. Now he has a new paper taking a notorious behemoth head-on.

Today Ken and I wish to talk about Michael’s recent work improving on the size of the famous AKS sorting network.

The initials AKS have become legion twice over in complexity. They are engraved on Wikipedia for the deterministic polynomial-time primality test of Manindra Agrawal, Neeraj Kayal, and Nitin Saxena proved in 2002. Two decades earlier, however, they designated the sorting network devised by Miklós Ajtai, János Komlós, and Endre Szemerédi. For sorting {n} elements these networks have asymptotically optimal size and depth, but as noted by Wikipedia,

…[they have] little or no practical application because of the large linear constants hidden by the Big-O notation.

Michael begs to differ, and will explain during his talk at the STOC conference being organized May 31—June 3 by Columbia University, with workshops there and the main conference on 34th Street. His paper is one of the few that is single authored, one of the few that is on sorting, and is titled “Zig-zag Sort: A Simple Deterministic Data-Oblivious Sorting Algorithm Running in {O(n \log n)} Time.”

Sorting Networks

A sorting network is one that sorts in an oblivious manner, meaning that the comparisons done at each step are the same for all inputs. The following diagram is used to represent a small such network: the comparisons are the vertical lines, the inputs enter on the left, and each vertical line forms a swap with the larger value moving up and the lower going down.


Usually “oblivious” is not a great thing to be. But for sorting algorithms it is quite a useful property: it makes the algorithm simpler to describe, it makes it easier to implement in hardware and even software, and makes it useful in many security applications. The latter follows since the operations at each step are the same independent of the input; thus, someone watching the comparison operations, not their values, can learn nothing from the algorithm’s execution. Of course there is an interesting cost to being oblivious: the algorithm runs in the same time even if the original input is already sorted. I find the latter an interesting issue, since in practice the input may be in sorted order or at least close to this. But that is another question.


The best sorting network for many years was due to Kenneth Batcher and has size {O(n (\log n)^{2}))} and depth {O((\log n)^{2})}, where {n} is the number of items to be sorted. The constant hidden behind the {O} notation is quite small. AKS brilliantly solved the longstanding open question of whether one needs only {O(n (\log n))} comparisons, getting depth {O(\log n)} to boot. But in 1998, Don Knuth noted that:

Batcher’s method is much better, unless {n} exceeds the total memory capacity of all computers on earth!

As we say here at GLL, the AKS sorting network is galactic: it needs that {n} be larger than {2^{78}} or so to finally be smaller than Batcher’s network for {n} items. Hence Knuth’s point—see here for some details. Also there is a whole book on sorting networks by Sherenaz Al-Haj Baddar and Batcher.

I must add that before 1983 the search for a better method than Batcher was a “hot” research problem. I worked on it trying to use random networks, but could not make any progress. Even in 1971, David Van Voorhis could beat Batcher when {n=16} by one comparison, and he parlayed this and other ideas into a method that improved Batcher by an additive term of form {-cn \log n} for a positive {c}. See this for the details.

The Amazing 0-1 Law

One of the coolest results in sorting theory is the 0-1 Law. This principle shows that a sorting network is valid if it can sort all {2^{n}} sequences of {0}s and {1}s. This not only drastically cuts down on the number of tests needed to determine the validity of a network, it is of great use in creating many constructions of sorting networks as well. The principle is found in Knuth’s book on sorting and is attributed to Willard Bouricius by him. Hoary it may be, but Michael uses it in several ways, most notably to simplify measures of distances of sub-arrays from being sorted.

It is interesting to note that the 0-1 law has been generalized to networks that sort most vectors, rather than all of them. This is contained in a 2005 paper by Sanguthevar Rajasekaran and Sandeep Sen. I thought also about this years ago, but never got a result that seemed interesting. Theirs does.

Again the 0-1 law is one of those few results that we have in theory that say that an algorithm works for all inputs from a large universe provided it works correctly for a small subset of the universe. It would be interesting to see if such theorems could be proved for other classes of algorithms, even if we have to restrict the type of algorithm greatly. Such theorems could be extremely useful, but they seem to be hard to find.

Zig-Zag Sorting

Here is Michael’s new algorithm in pseudocode:


There are some comments that I want to make, besides the usual invitation to see the paper for the details. One is that the code is quite short, and that while the code is short, the main issue is proving that the code works—we sketch some details in a moment. The fact that the proof of correctness is much harder than the algorithm is a recurrent theme in modern algorithms. The classic quicksort algorithm is even shorter:


This phenomenon suggests to me a possible conjecture:

Could there be very short algorithms for SAT or other hard problems, factoring for example, that are short as programs, but have long complex proofs?

A related conjecture is: can we somehow prove that there is no very short program for SAT that runs fast? I think one of my first statements on GLL was that we could not rule out a polynomial time algorithm for SAT that fits on a single page. Is there any way to make this into a serious conjecture? How would we formalize it?

The last comment is that Michael does have to sacrifice the {O(\log n)} depth attained by AKS. In fact, so far he does not even get {O(n)} depth—he says in a footnote that he gets only {O(n\log n)} parallel time.

Some Ingredients

The name “zig-zag” comes from places where the algorithm does a shellsort-type percolation going up rather than down, and more simply, flips upside down some sequences of elements that were nearly sorted from a previous step. This seems “counter-intuitive” as he says, but such steps are allowed also in Batcher’s bitonic sort, since flipping a bitonic sequence upside-down still leaves it bitonic. Such flips can ultimately be eliminated in Batcher’s network, as noted here, whereas they seem more integral to Michael’s networks.

The {\mathsf{Reduce}(A,B)} step needs only to be a “black box” that meets the following three partial-progress conditions on the {m}-element arrays {A} and {B}, for suitably chosen values of the parameters {\epsilon,\delta,\beta}. The first two were previously known, but the third one is new, and really helps “kick” the partial progress into completion:

  1. For any {k \leq m}, at most {\epsilon k} of the largest {k} elements in {A \cup B} wind up in {A}, and at most {\epsilon k} of the smallest {k} elements in {A \cup B} end up in {B}.

  2. For any {k \leq \frac{5}{6} m}, the above holds with {\epsilon} replaced by the somewhat smaller value {\beta}.

  3. For any {k \leq \frac{5}{6} m}, and any {k_1,k_2 \leq k}, whenever {k_1} is the number of the largest {k} elements in {A \cup B} that are still in {A}, at most {\delta k_1} of those {k} elements remain in {A} after the step, and whenever {k_2} is the count of elements in {B} that are among the {k} smallest in {A \cup B}, at most {\delta k_2} of the smallest {k} elements remain in {B} afterward.

In the third clause, it is not mandated that the {\delta k_1} elements be a subset of the original {k_1}—the analysis only cares about how the trace number of still-misplaced elements is “attenuated” by each application of {\mathsf{Reduce}}. The inductive description of the algorithm building on this progress is very pretty.

There remains the task of coming up with {\mathsf{Reduce}} gadgets for the needed {\epsilon,\delta,\beta} choices and for various sizes {m}. Here is where the dreaded fellow-traveler of galacticity, namely non-constructivity, enters in. The smallest gadget sizes are known only via non-constructive probabilistic/counting arguments. What the paper shows is that the “penalty” for making the gadgets constructive is appreciably smaller for his algorithm than for AKS, resulting in a relatively great savings. For non-constructive versions his algorithm is “competitive” while still being simpler.

Open Problems

Is there a really practical sorting network that is asymptotically optimal? Can it have {O(n)} or even {O(\log n)} depth?

13 Comments leave one →
  1. April 25, 2014 12:12 am

    If there is a polytime algorithm for SAT, there is also one with a very short code. It just has to run all possible algorithms of length at most C for n^C steps and try the solution given by them (it is easy to make an algorithm also give a solution if you have one that decides existence). If none of the solutions work, then return unsatisfiable. Hm, since this is also an algorithm for SAT, did I just prove that there is an algorithm whose size is at most O(log C) where C is the best possible constant of the running time? In fact we can make it even “shorter” by running algs for 2^C steps and so on. Couldn’t we just try all such algorithms and thus prove for any C that there is no alg for SAT running in n^C? I must be wrong somewhere…

    • April 25, 2014 1:39 am

      I see that I have treated the constant of the running time a little too liberally, so the D from Dn^C should not be ignored.

    • April 25, 2014 2:20 am

      In fact here is another approach that even works for sorting networks: Start to generate all possible sorting networks along with a proof that they run in time O(nlogn). If a valid proof is found, then run that algorithm. This is correct, as we know that AKS exists. Of course our hidden constant might be much worse but we can also include that in the algorithm, we only have to describe it (or a slightly bigger number), this would of course take very small space.

  2. Commenter permalink
    April 25, 2014 3:25 am

    While it’s a great paper, I don’t think it’s fair to give that pseudo code as evidence of the code being short when the pseudo code isn’t given for the “Reduce” subroutine! That seems to be where a lot of the action is.

  3. April 25, 2014 10:26 am

    yeah sorting networks seem to yet contain some deep mysteries. once wondered if other key algorithmic problems can be reduced to sorting somehow, it seems to be the case. also it seems you didnt cover this recent paper on optimal sorting networks by Bundala/Zavodny… right down your alley? any opinions? similar to the erdos discrepancy problem results (found via SAT solvers) which you did cover in much detail. more musings great moments in empirical/experimental math/TCS reseach

  4. April 27, 2014 3:47 am

    Are badpoor computer scientists common?

  5. Jim permalink
    April 29, 2014 9:50 am

    Sorting networks are certainly interesting things. The data-blind aspect always intrigued me as well and led me to a different consideration: is it possible to sort an integer array without doing any compares at all? I worked on this a while – basically I swapped out the compare between two integers with a few math operations that had the same effect. Thus, the network becomes truly blind – it is a series of operations which never depend on the data, but ends up sorting the data. Unfortunately I couldn’t figure out if I actually had a compare-less sort.

    In particular, let’s say I’m trying to sort the pair (a,b). Then the ordered pair is:

    [(a+b) – abs(a-b) , (a+b) + abs(a-b)]. This pair is ordered smallest to largest.

    However, how do I calculate abs(a-b) without using a compare at some point? I could calculate ((a-b)^2)^0.5 which would do the trick – but how do you actually do a square root without using a compare in the algorithm? I also tried bit-twiddling with the 2’s complement representation of integers, but couldn’t do that without a compare at some point.

    But, if you figure out that trick, then the sorting network will scale it up and allow compare-less sorting of arbitrary sized lists of integers. Neat!

  6. Richard Rankin permalink
    April 29, 2014 10:21 pm

    Are you sure it isn’t “snorting”?

    Why join the navy when you can be a pirate?

    – Steve Jobs

    On 4/24/14, 11:46 PM, “Gödels Lost Letter and P=NP” wrote: Pip posted: ” An improved sorting network to appear at STOC 2014 Teaching page source Michael Goodrich is a Chancellor’s Professor at the University of California at Irvine, and recently served a year as Chair of th”

Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s