Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Selection sort
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{short description|Sorting algorithm}} {{No footnotes|date=May 2019}} {{Infobox Algorithm |name={{PAGENAMEBASE}}|class=[[Sorting algorithm]] |caption=Selection sort animation |data=[[Array data structure|Array]] |time=<math>O(n^2)</math> comparisons, <math>O(n)</math> swaps |best-time=<math>O(n^2)</math> comparisons, <math>O(1)</math> swap |average-time=<math>O(n^2)</math> comparisons, <math>O(n)</math> swaps |space=<math>O(1)</math> auxiliary |optimal=No |stable=No }} In [[computer science]], '''selection sort''' is an [[in-place algorithm|in-place]] [[comparison sort|comparison]] [[sorting algorithm]]. It has a [[Big O notation|O]](''n''<sup>2</sup>) [[time complexity]], which makes it inefficient on large lists, and generally performs worse than the similar [[insertion sort]]. Selection sort is noted for its simplicity and has performance advantages over more complicated algorithms in certain situations, particularly where [[auxiliary memory]] is limited. The algorithm divides the input list into two parts: a sorted sublist of items which is built up from left to right at the front (left) of the list and a sublist of the remaining unsorted items that occupy the rest of the list. Initially, the sorted sublist is empty and the unsorted sublist is the entire input list. The algorithm proceeds by finding the smallest (or largest, depending on sorting order) element in the unsorted sublist, exchanging (swapping) it with the leftmost unsorted element (putting it in sorted order), and moving the sublist boundaries one element to the right. The time efficiency of selection sort is quadratic, so there are a number of sorting techniques which have better time complexity than selection sort. == Example == Here is an example of this sort algorithm sorting five elements: {| class="wikitable" ! Sorted sublist ! Unsorted sublist ! Least element in unsorted list |- | () | style="text-align:right;" | (12, 25, 64, 11, 22) | 11 |- | (11) | style="text-align:right;" | (25, 64, 12, 22) | 12 |- | (11, 12) | style="text-align:right;" | (64, 25, 22) | 22 |- | (11, 12, 22) | style="text-align:right;" | (25, 64) | 25 |- | (11, 12, 22, 25) | style="text-align:right;" | (64) | 64 |- | (11, 12, 22, 25, 64) | style="text-align:right;" | () | |} [[Image:Selection-Sort-Animation.gif|right|thumb|Selection sort animation. Red is current min. Yellow is sorted list. Blue is current item.]] (Nothing appears changed on these last two lines because the last two numbers were already in order.) Selection sort can also be used on list structures that make add and remove efficient, such as a [[linked list]]. In this case it is more common to ''remove'' the minimum element from the remainder of the list, and then ''insert'' it at the end of the values sorted so far. For example: <pre style="overflow:auto; width:auto;"> arr[] = 64 25 12 22 11 // Find the minimum element in arr[0...4] // and place it at beginning 11 25 12 22 64 // Find the minimum element in arr[1...4] // and place it at beginning of arr[1...4] 11 12 25 22 64 // Find the minimum element in arr[2...4] // and place it at beginning of arr[2...4] 11 12 22 25 64 // Find the minimum element in arr[3...4] // and place it at beginning of arr[3...4] 11 12 22 25 64 </pre> == Implementations == Below is an implementation in [[C (programming language)|C]]. <syntaxhighlight lang="c" style="overflow:auto; width:auto;" line="1"> /* a[0] to a[aLength-1] is the array to sort */ int i,j; int aLength; // initialise to a's length /* advance the position through the entire array */ /* (could do i < aLength-1 because single element is also min element) */ for (i = 0; i < aLength-1; i++) { /* find the min element in the unsorted a[i .. aLength-1] */ /* assume the min is the first element */ int jMin = i; /* test against elements after i to find the smallest */ for (j = i+1; j < aLength; j++) { /* if this element is less, then it is the new minimum */ if (a[j] < a[jMin]) { /* found new minimum; remember its index */ jMin = j; } } if (jMin != i) { swap(&a[i], &a[jMin]); } } </syntaxhighlight> == Complexity == Selection sort is not difficult to analyze compared to other sorting algorithms, since none of the loops depend on the data in the array. Selecting the minimum requires scanning <math>n</math> elements (taking <math>n-1</math> comparisons) and then swapping it into the first position. Finding the next lowest element requires scanning the remaining <math>n-1</math> elements (taking <math>n-2</math> comparisons) and so on. Therefore, the total number of comparisons is : <math>(n-1)+(n-2)+\dots+1 = \sum_{i=1}^{n-1}i</math> By [[arithmetic progression]], : <math>\sum_{i=1}^{n-1}i= \frac{(n-1)+1}{2}(n-1)= \frac{1}{2}n(n-1)= \frac{1}{2}(n^2-n)</math> which is of complexity <math>O(n^2)</math> in terms of number of comparisons. == Comparison to other sorting algorithms == Among quadratic sorting algorithms (sorting algorithms with a simple average-case of [[Big O notation#Family of Bachmann–Landau notations|Θ(''n''<sup>2</sup>)]]), selection sort almost always outperforms [[bubble sort]] and [[gnome sort]]. [[Insertion sort]] is very similar in that after the ''k''th iteration, the first <math>k</math> elements in the array are in sorted order. Insertion sort's advantage is that it only scans as many elements as it needs in order to place the <math>k+1</math>st element, while selection sort must scan all remaining elements to find the <math>k+1</math>st element. Simple calculation shows that insertion sort will therefore usually perform about half as many comparisons as selection sort, although it can perform just as many or far fewer depending on the order the array was in prior to sorting. It can be seen as an advantage for some [[real-time computing|real-time]] applications that selection sort will perform identically regardless of the order of the array, while insertion sort's running time can vary considerably. However, this is more often an advantage for insertion sort in that it runs much more efficiently if the array is already sorted or "close to sorted." While selection sort is preferable to insertion sort in terms of number of writes (<math>n-1</math> swaps versus up to <math>n(n-1)/2</math> swaps, with each swap being two writes), this is roughly twice the theoretical minimum achieved by [[cycle sort]], which performs at most ''n'' writes. This can be important if writes are significantly more expensive than reads, such as with [[EEPROM]] or [[Flash memory]], where every write lessens the lifespan of the memory. Selection sort can be implemented without unpredictable branches for the benefit of CPU [[branch predictor]]s, by finding the location of the minimum with branch-free code and then performing the swap unconditionally. Finally, selection sort is greatly outperformed on larger arrays by <math>\Theta(n\log n)</math> [[divide and conquer algorithm|divide-and-conquer algorithms]] such as [[mergesort]]. However, insertion sort or selection sort are both typically faster for small arrays (i.e. fewer than 10–20 elements). A useful optimization in practice for the recursive algorithms is to switch to insertion sort or selection sort for "small enough" sublists. == Variants == [[Heapsort]] has been described as "nothing but an implementation of selection sort using the right [[data structure]]."<ref>{{cite book |first=Steven |last=Skiena |author-link=Steven Skiena |title=The Algorithm Design Manual |edition=3rd |publisher=Springer |year=2008 |page=116 |chapter=Searching and Sorting |isbn=978-3-030-54255-9 |quote=The name typically given to this algorithm, ''heapsort'', obscures the fact that the algorithm is nothing but an implementation of selection sort using the right data structure. |doi=10.1007/978-3-030-54256-6_4}}<!--DOI for chapter--></ref> It greatly improves the basic algorithm by using an [[implicit data structure|implicit]] [[heap (data structure)|heap]] data structure to find and remove each lowest element in <math>\Theta(\log n)</math> time, instead of normal selection sort's <math>\Theta(n)</math> inner loop, reducing the total running time to <math>\Theta(n\log n)</math>. A bidirectional variant of selection sort (called '''double selection sort''' or sometimes '''cocktail sort''' due to its similarity to [[cocktail shaker sort]]) finds ''both'' the minimum and maximum values in the list in every pass. This requires three comparisons per two items (a pair of elements is compared, then the greater is compared to the maximum and the lesser is compared to the minimum) rather than regular selection sort's one comparison per item, but requires only half as many passes, a net 25% savings. Selection sort can be implemented as a [[Sorting algorithm#Classification|stable sort]] if, rather than swapping in step 2, the minimum value is ''inserted'' into the first position and the intervening values shifted up. However, this modification either requires a data structure that supports efficient insertions or deletions, such as a linked list, or it leads to performing <math>\Theta(n^{2})</math> writes. In the '''bingo sort''' variant, items are sorted by repeatedly looking through the remaining items to find the greatest value and moving ''all'' items with that value to their final location.<ref>{{DADS|Bingo sort|bingosort}}</ref> Like [[counting sort]], this is an efficient variant if there are many duplicate values: selection sort does one pass through the remaining items for each ''item'' moved, while Bingo sort does one pass for each ''value''. After an initial pass to find the greatest value, subsequent passes move every item with that value to its final location while finding the next value as in the following [[pseudocode]] (arrays are zero-based and the for-loop includes both the top and bottom limits, as in [[Pascal (programming language)|Pascal]]): <syntaxhighlight lang="pascal"> bingo(array A) { This procedure sorts in ascending order by repeatedly moving maximal items to the end. } begin last := length(A) - 1; { The first iteration is written to look very similar to the subsequent ones, but without swaps. } nextMax := A[last]; for i := last - 1 downto 0 do if A[i] > nextMax then nextMax := A[i]; while (last > 0) and (A[last] = nextMax) do last := last - 1; while last > 0 do begin { Each main loop searches for the new nextMax while swapping items equal to prevMax into place. } prevMax := nextMax; nextMax := A[last]; for i := last - 1 downto 0 do if A[i] > nextMax then if A[i] <> prevMax then nextMax := A[i]; else begin swap(A[i], A[last]); last := last - 1; end while (last > 0) and (A[last] = nextMax) do last := last - 1; end; end; </syntaxhighlight> Thus, if on average there are more than two items with the same value, bingo sort can be expected to be faster because it executes the inner loop fewer times than selection sort. <!-- If you came here to write an implementation of selection sort, note that this page used to have implementations but they were moved to Wikibooks. Therefore, implementations should not be added here. --> == See also == * [[Selection algorithm]] == References == {{reflist}} {{refbegin}} * [[Donald Knuth]]. ''[[The Art of Computer Programming]]'', Volume 3: ''Sorting and Searching'', Third Edition. Addison–Wesley, 1997. {{ISBN|0-201-89685-0}}. Pages 138–141 of Section 5.2.3: Sorting by Selection. * Anany Levitin. ''Introduction to the Design & Analysis of Algorithms'', 2nd Edition. {{ISBN|0-321-35828-7}}. Section 3.1: Selection Sort, pp 98–100. * [[Robert Sedgewick (computer scientist)|Robert Sedgewick]]. ''Algorithms in C++, Parts 1–4: Fundamentals, Data Structure, Sorting, Searching: Fundamentals, Data Structures, Sorting, Searching Pts. 1–4'', Second Edition. Addison–Wesley Longman, 1998. {{ISBN|0-201-35088-2}}. Pages 273–274 {{refend}} == External links == {{wikibooks|Algorithm implementation|Sorting/Selection_sort|Selection sort}} * {{webarchive |url=https://web.archive.org/web/20150307110315/http://www.sorting-algorithms.com/selection-sort |date=7 March 2015 |title=Animated Sorting Algorithms: Selection Sort}} – graphical demonstration {{sorting}} [[Category:Comparison sorts]] [[Category:Articles with example C code]]
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)
Pages transcluded onto the current version of this page
(
help
)
:
Template:Cite book
(
edit
)
Template:DADS
(
edit
)
Template:ISBN
(
edit
)
Template:Infobox Algorithm
(
edit
)
Template:No footnotes
(
edit
)
Template:Refbegin
(
edit
)
Template:Refend
(
edit
)
Template:Reflist
(
edit
)
Template:Short description
(
edit
)
Template:Sister project
(
edit
)
Template:Sorting
(
edit
)
Template:Webarchive
(
edit
)
Template:Wikibooks
(
edit
)