Home

Awesome

Sorting Algorithms

sort.js

Various sorting algorithms implementations in JavaScript

sort.min.js

screenshot

Sorting Series, which is also a kind of discrete optimization problem (eg the permutation function perm of 0..N-1 which maximizes 0*a[perm[0]]+1*a[perm[1]]+..+(N-1)*a[perm[N-1]] is the permutation which sorts the array a in ascending order that is a[perm[0]] <= a[perm[1]] <= .. <= a[perm[N-1]]); lies at the center of Computer Science and Algorithms because of its many uses.

(Ref. Sorting Algorithm)

Furthermore Sorting, in one way or another, is integral part of many other important algorithms and applications (see eg. Knuth The Art Of Computer Programming)

For example Sorting is very closely associated to Searching, another topic of immense importance and applications.

Under certain sorting states, searching can be achieved in O(logN) time or even in O(1) time (constant) for almost every search term.

Sorting has 3 approaches:

(eg. NIST.gov maintains a dictionary of various algorithms at: http://xlinux.nist.gov/dads// )

Block vs. Online/Adaptive:

  1. In the Block case, the whole array is available at once for this case many algorithms are known (comparison-based => O(N^2), O(NlogN) complexities) and (number/count based => O(N) complexity) (see below)

  2. In the Adaptive/Online case, the input series is accesed one at a time (for example an time-input signal). In this case some of the previous algorithms can be transformed to work adaptively

    Apart from that, there are algorithms (like Dynamic Lists, Dynamic Heaps and Balanced Trees, Tries, eg AVL Trees) which keep an input sequence always in a 'sorted' state (with each new input) with relatively low complexity (eg O(logN)).

Comparison-Based vs. Arithmetic/Count-Based:

Is O(N) sorting possible for arbitrary random numbers??

Computing the value of a certain number n (in a fixed type of encoding, eg decimal-place) requires approximately O(logn) "primitive digit" operations. Since, statistically, the range of values of numbers in a list of given size is increasingly correlated to the size of the list as the size of the list increases (i.e lists of size N containing random numbers with values over the whole range 0..N have increasingly greater probability, as N increases, over all lists of same size N, containing random numbers whose values do not cover the whole range 0..N), one then has an overall complexity of O(NlogN) even for arithmetic-based sorting algorithms.

See for example "what is the true complexity of radix sort?".

Classical algorithms for integer sorting require assumptions about the size of the integers to be sorted, or else have a running time dependent on the size.

-- Sorting in Linear Time? Arne Andersson, Torben Hagerupt, Stefan Nilsson, Rajeev Ramam

However the catch here is that same holds for comparing arbitrary numbers, computationaly one has to compare primitive digit by primitive digit in sequence on average, hence an additional O(logn) complexity for comparison-based algorithms, over the O(NlogN) bound.

Is O(NlogN) complexity a kind of strict base line for this computational model??

According to Knuth's theoretical lower bound theorem for general (comparison) sorting algorithms (note O(logN!) = O(NlogN)): the O(NlogN) bound is asymptoticaly tight (see also information-theoretic lower bound for comparison sorts is Ω(NlogN) ).

A summary of various sorting and searching algorithms can be found in SORTING AND SEARCHING ALGORITHMS, Tom Niemann (pdf)

Included Algorithms


NOTE: The calculation of asymptotic complexity is done usually (using recursive relations) with the Master Theorem :

Refs. http://en.wikipedia.org/wiki/Master_theorem, http://en.wikipedia.org/wiki/Introduction_to_Algorithms

T(n) = aT(n/b) + f(n), a>=1, b>1

eg. for MergeSort => T(n) = 2T(n/2) + O(n) => T(n) = O(nlogn)


This package implements showcases of various (best known) sorting algorithms (and a couple of custom ones) for study, experimentation and use in applications In a concice library

Algorithms as a technology Suppose computers were infinitely fast and memory was free. Would you have any reason to study algorithms? The answer is yes, if for no other reason than that you would still like to demonstrate that your solution method terminates and does so with the correct answer. ...Of course, computers may be fast but not infinitely fast and memory may be cheap but not completely free. Computing time is therefore a bounded resource, and so is space in memory. These resources should be used wisely and algorithms that are efficient in terms of time and space will help you do so. This demostrates that algorithms, like computer hardware, are a technology . Total system performance depends on choosing efficient algorithms as much as choosing fast hardware. Just as rapid advances are being made in other computer technologies, they are being made in algorithms as well. (Introduction to algorithms, 2nd Ed. Cormen,Leiserson,Rivest,Stein)

Algorithms as a ecological technology Additionaly, every operation/instruction a computer performs has an energy consumption cost. Thus an efficient algorithm saves energy! An efficient algorithm performs a computation by trying to use the resources in the best possible manner, so effectively uses energy in the best possible manner. Where does energy come from? It comes from burning coal or gas (mainly). So there you have it; efficient code is ecological! Better start learning your complexity soon.

see also: