Quick Sort Roberto Hibbler Dept

Total Page:16

File Type:pdf, Size:1020Kb

Quick Sort Roberto Hibbler Dept Quick Sort Roberto Hibbler Dept. of Computer Science Florida Institute of Technology Melbourne, FL 32901 [email protected] ABSTRACT 2. We show that theoretically Quick Sort performs no ͦ Given an array of elements, we want to arrange those elements worse than ͉ʚ͢ ʛ in the worst case, being equivalent into a sorted order. To sort those elements, we will need to make to current solutions. comparisons between the individual elements efficiently. Quick 3. We show that empirically Quick Sort is on average Sort uses a divide and conquer strategy to sort an array faster than Insertion Sort. efficiently while making the least number of comparisons between array elements. Our results show that for arrays with This paper will discuss in Section 2 comparison sort algorithms large numbers of array elements, Quick Sort is more efficient related to the problem, followed by the detailed approach of our than three other comparison sort algorithms, Bubble Sort, solution in Section 3, the evaluation of our results in Section 4, Insertion Sort, and Selection Sort. Our theoretical evaluation and our final conclusion in Section 5. shows that Quick Sort beats a quadratic time complexity, while our empirical evaluation shows that Quick Sort at times was 32 times faster than Insertion Sort, the current recognized most 2. RELATED WORK efficient comparison algorithm, with one real data set. The three algorithms that we will discuss are Bubble Sort, Selection Sort, and Insertion Sort. All three are comparison sort Keywords algorithms, just as Quick Sort. Quick Sort, sorting, comparisons, Selection Sort, time complexity The Bubble Sort algorithm works by continually swapping adjacent array elements if they are out of order until the array is 1. INTRODUCTION in sorted order. Every iteration through the array places at least The ability to arrange an array of elements into a defined order one element at its correct position. Although algorithmically is very important in Computer Science. Sorting is heavily used correct, Bubble Sort is inefficient for use with arrays with a with online stores, were the order that services or items were large number of array elements and has a ͉ʚͦ͢ʛ worst case time purchased determines what orders can be filled and who complexity. Knuth observed[1], also, that while Bubble Sort receives their order first. Sorting is also essential for the shares the worst-case time complexity with other prevalent database management systems used by banks and financial sorting algorithms, compared to them it makes far more element systems, such as the New York Stock Exchange, to track and swaps, resulting in poor interaction with modern CPU hardware. rank the billions of transactions that go on in one day. There are We intend to show that Quick Sort needs to make on average many algorithms which provide a solution to sorting arrays, fewer element swaps than Bubble Sort. including algorithms such as Bubble Sort[1], Insertion Sort[2], and Selection Sort[3]. While these algorithms are The Selection Sort algorithm arranges array elements in order by programmatically correct, they are not efficient for arrays with a first finding the minimum value in the array and swapping it large number of elements and exhibit quadratic time complexity. with the array element that is in its correct position depending on how the array is being arranged. The process is then repeated We are given an array of comparable values. We need to arrange with the second smallest value until the array is sorted. This these values into either an ascending or descending order. creates two distinctive regions within the array, the half that is sorted and the half that has not been sorted. Selection Sort shows We introduce the Quick Sort algorithm. The Quick Sort an improvement over Bubble Sort by not comparing all the algorithm is a divide-and-conquer algorithm. It takes an array elements in its unsorted half until it is time for that element to be and divides that array into sub arrays based on a pivot value. placed into its sorted position. This makes Selection Sort less Values less than the pivot value are shifted to within the array to affected by the input’s order. Though, it is still no less the left of the pivot while values greater are shifted to the right inefficient with arrays with a large number of array elements. of the pivot in a process called partitioning. These elements are Also, even with the improvements Selection Sort still shares the then divided into two sub arrays, one of the elements to the left same worst-case time complexity of ͉ʚͦ͢ʛ. We intend to show of the pivot and the other the right elements. The partitioning that Quick Sort will operate on average faster Selection Sort. process is done recursively until we have arrays of single elements. A single element is already sorted, and so the elements The Insertion Sort algorithm takes elements from the input array are combined back into sorted arrays two sub-arrays at a time, and places those elements in their correct place into a new array, until we are left with a final sorted array. We contribute the shifting existing array elements as needed. Insertion Sort following: improves over Selection Sort by only making as many comparisons as it needs to determine the correct position of the 1. We introduce the Quick Sort algorithm. current element, while Selection Sort makes comparisons against each element in the unsorted part of the array. In the As the pseudocode shows, after a new pivot element is randomly )v chosen (lines 3-4), the array is broken up into a left half and a average case, Insertion Sort’s time complexity is ͉ʚ ʛ, but its ͨ right half and sorted recursively (lines 5-6). In the partition ͦ worst case is ͉ʚ͢ ʛ, the same as Bubble Sort and Selection algorithm, the two sub arrays are compared to determine how Sort. The tradeoff of Insertion Sort is that on the average more they should be arranged (lines 12-15). In the following elements are swapped as array elements are shifted within the examples, using the given input, the partitioning of the array array with the addition of new elements. Even with its (Figure 3) and how the sub arrays are concatenated back into a limitations, Selection Sort is the current fastest comparison sorted array (Figure 4) are illustrated. based sorting algorithm since it equals Bubble Sort and Selection Sort in the worst case, but is exceptionally better in the Inputs – A: array of n elements average case. We intend to show that Quick Sort operates on (33 55 77 11 66 88 22 44) average faster than Insertion Sort. Output – array A in ascending order 3. APPROACH A large array with an arbitrary order needs to be arranged in an ascending or descending order, either lexicographically or numerically. Quick sort can solve this problem by using two key ideas. The first key idea of Quick sort is that a problem can be divided and conquered. The problem can be broken into smaller arrays, and those arrays can be solved. This is done by picking an element from the array, called the pivot, and reordering the array Figure 3: Shows the partitioning of the input array into so that all elements which are less than the pivot come before single element arrays. The red element is the pivot element the pivot and all elements greater than the pivot come after it. of the array. This process is called partitioning. Second, by dividing the array into two halves, elements less than the pivot and elements greater than the pivot, then partitioning those halves recursively into arrays of single elements, two sorted arrays are concatenated on either side of the pivot value into one array, as a single element array is already sorted. Refer to the following pseudocode: Input – A: array of n elements, left: left sub array from the pivot, right: right sub array from the pivot Output – array A sorted in ascending order 1. proc quicksort(A,left,right) 2. if right > left Figure 4: From the bottom up, shows the concatenating of 3. pivot = random(A) the single element arrays. The red element is the pivot 4. newPivot=partition(A,left,right,pivot) element. 5. quicksort(A, left, newPivot – 1) 6. quicksort(A,newPivot+1,right) 7. return A As the example shows, array A is broken in half based on the Figure 1: Pseudocode of the Quick Sort algorithm pivot until they are in arrays of only a single element, then those single elements are concatenated together until they form a single sorted array in ascending order. Input – A: array of n elements,left:array of m elements, right: array of k elements, pivot: int of pivot index 4. EVALUATION Output – array result sorted in ascending order 8. proc partition(A,left,right,pivot) 4.1 Theoretical Analysis 9. val = A[pivot] 4.1.1 Evaluation Criteria 10. swap(A[pivot],A[right]) 11. index = left All comparison based sorting algorithms count the comparisons 12. for I from left to right -1 of array elements as one of their key operations. The Quick Sort 13. if A[i] >= val algorithm can be evaluated by measuring the number of 14. swap(A[i],A[index]) comparisons between array elements. As the key operation, we 15. index=index+1 can measure the number of comparisons made to determine the 16. swap(A[i],A[right]) overall efficiency of the algorithm. We intend to show that 17. return index because the Quick Sort algorithm makes less comparisons over Figure 2: Pseudocode of the Partition step of Quick Sort the currently acknowledged most efficient algorithm, Insertion Sort[Sec 2], Quick Sort is the most efficient comparison sort algorithm.
Recommended publications
  • Lecture 11: Heapsort & Its Analysis
    Lecture 11: Heapsort & Its Analysis Agenda: • Heap recall: – Heap: definition, property – Max-Heapify – Build-Max-Heap • Heapsort algorithm • Running time analysis Reading: • Textbook pages 127 – 138 1 Lecture 11: Heapsort (Binary-)Heap data structure (recall): • An array A[1..n] of n comparable keys either ‘≥’ or ‘≤’ • An implicit binary tree, where – A[2j] is the left child of A[j] – A[2j + 1] is the right child of A[j] j – A[b2c] is the parent of A[j] j • Keys satisfy the max-heap property: A[b2c] ≥ A[j] • There are max-heap and min-heap. We use max-heap. • A[1] is the maximum among the n keys. • Viewing heap as a binary tree, height of the tree is h = blg nc. Call the height of the heap. [— the number of edges on the longest root-to-leaf path] • A heap of height k can hold 2k —— 2k+1 − 1 keys. Why ??? Since lg n − 1 < k ≤ lg n ⇐⇒ n < 2k+1 and 2k ≤ n ⇐⇒ 2k ≤ n < 2k+1 2 Lecture 11: Heapsort Max-Heapify (recall): • It makes an almost-heap into a heap. • Pseudocode: procedure Max-Heapify(A, i) **p 130 **turn almost-heap into a heap **pre-condition: tree rooted at A[i] is almost-heap **post-condition: tree rooted at A[i] is a heap lc ← leftchild(i) rc ← rightchild(i) if lc ≤ heapsize(A) and A[lc] > A[i] then largest ← lc else largest ← i if rc ≤ heapsize(A) and A[rc] > A[largest] then largest ← rc if largest 6= i then exchange A[i] ↔ A[largest] Max-Heapify(A, largest) • WC running time: lg n.
    [Show full text]
  • Quick Sort Algorithm Song Qin Dept
    Quick Sort Algorithm Song Qin Dept. of Computer Sciences Florida Institute of Technology Melbourne, FL 32901 ABSTRACT each iteration. Repeat this on the rest of the unsorted region Given an array with n elements, we want to rearrange them in without the first element. ascending order. In this paper, we introduce Quick Sort, a Bubble sort works as follows: keep passing through the list, divide-and-conquer algorithm to sort an N element array. We exchanging adjacent element, if the list is out of order; when no evaluate the O(NlogN) time complexity in best case and O(N2) exchanges are required on some pass, the list is sorted. in worst case theoretically. We also introduce a way to approach the best case. Merge sort [4] has a O(NlogN) time complexity. It divides the 1. INTRODUCTION array into two subarrays each with N/2 items. Conquer each Search engine relies on sorting algorithm very much. When you subarray by sorting it. Unless the array is sufficiently small(one search some key word online, the feedback information is element left), use recursion to do this. Combine the solutions to brought to you sorted by the importance of the web page. the subarrays by merging them into single sorted array. 2 Bubble, Selection and Insertion Sort, they all have an O(N2) time In Bubble sort, Selection sort and Insertion sort, the O(N ) time complexity that limits its usefulness to small number of element complexity limits the performance when N gets very big. no more than a few thousand data points.
    [Show full text]
  • Binary Search
    UNIT 5B Binary Search 15110 Principles of Computing, 1 Carnegie Mellon University - CORTINA Course Announcements • Sunday’s review sessions at 5‐7pm and 7‐9 pm moved to GHC 4307 • Sample exam available at the SCHEDULE & EXAMS page http://www.cs.cmu.edu/~15110‐f12/schedule.html 15110 Principles of Computing, 2 Carnegie Mellon University - CORTINA 1 This Lecture • A new search technique for arrays called binary search • Application of recursion to binary search • Logarithmic worst‐case complexity 15110 Principles of Computing, 3 Carnegie Mellon University - CORTINA Binary Search • Input: Array A of n unique elements. – The elements are sorted in increasing order. • Result: The index of a specific element called the key or nil if the key is not found. • Algorithm uses two variables lower and upper to indicate the range in the array where the search is being performed. – lower is always one less than the start of the range – upper is always one more than the end of the range 15110 Principles of Computing, 4 Carnegie Mellon University - CORTINA 2 Algorithm 1. Set lower = ‐1. 2. Set upper = the length of the array a 3. Return BinarySearch(list, key, lower, upper). BinSearch(list, key, lower, upper): 1. Return nil if the range is empty. 2. Set mid = the midpoint between lower and upper 3. Return mid if a[mid] is the key you’re looking for. 4. If the key is less than a[mid], return BinarySearch(list,key,lower,mid) Otherwise, return BinarySearch(list,key,mid,upper). 15110 Principles of Computing, 5 Carnegie Mellon University - CORTINA Example
    [Show full text]
  • COSC 311: ALGORITHMS HW1: SORTING Due Friday, September 22, 12Pm
    COSC 311: ALGORITHMS HW1: SORTING Due Friday, September 22, 12pm In this assignment you will implement several sorting algorithms and compare their relative per- formance. The sorting algorithms you will consider are: 1. Insertion sort 2. Selection sort 3. Heapsort 4. Mergesort 5. Quicksort We will discuss all of these algorithms in class. You should run your experiments on the department servers, remus/romulus (if you would like to write your code on another machine that is fine, but make sure you run the actual tim- ing experiments on remus/romulus). Instructions for how to access the servers can be found on the CS department web page under “Computing Resources.” If you are a Five College stu- dent who has previously taken an Amherst CS course or who enrolled during preregistration last spring, you should have an account already set up (you may need to change your password; go to https://www.amherst.edu/help/passwords). If you don’t already have an account, you can request one at https://sysaccount.amherst.edu/sysaccount/CoursePetition.asp. It will take a day to create the new account, so please do this right away. Please type up your responses to the questions below. I recommend using LATEX, which is a type- setting language that makes it easy to make math look good. If you’re not already familiar with it, I encourage you to practice! Your tasks: 1) Theoretical predictions. Rank the five sorting algorithms in order of how you expect their run- times to compare (fastest to slowest). Your ranking should be based on the asymptotic analysis of the algorithms.
    [Show full text]
  • Sorting Algorithms
    Sorting Algorithms Next to storing and retrieving data, sorting of data is one of the more common algorithmic tasks, with many different ways to perform it. Whenever we perform a web search and/or view statistics at some website, the presented data has most likely been sorted in some way. In this lecture and in the following lectures we will examine several different ways of sorting. The following are some reasons for investigating several of the different algorithms (as opposed to one or two, or the \best" algorithm). • There exist very simply understood algorithms which, although for large data sets behave poorly, perform well for small amounts of data, or when the range of the data is sufficiently small. • There exist sorting algorithms which have shown to be more efficient in practice. • There are still yet other algorithms which work better in specific situations; for example, when the data is mostly sorted, or unsorted data needs to be merged into a sorted list (for example, adding names to a phonebook). 1 Counting Sort Counting sort is primarily used on data that is sorted by integer values which fall into a relatively small range (compared to the amount of random access memory available on a computer). Without loss of generality, we can assume the range of integer values is [0 : m], for some m ≥ 0. Now given array a[0 : n − 1] the idea is to define an array of lists l[0 : m], scan a, and, for i = 0; 1; : : : ; n − 1 store element a[i] in list l[v(a[i])], where v is the function that computes an array element's sorting value.
    [Show full text]
  • Sorting Algorithms Properties Insertion Sort Binary Search
    Sorting Algorithms Properties Insertion Sort Binary Search CSE 3318 – Algorithms and Data Structures Alexandra Stefan University of Texas at Arlington 9/14/2021 1 Summary • Properties of sorting algorithms • Sorting algorithms – Insertion sort – Chapter 2 (CLRS) • Indirect sorting - (Sedgewick Ch. 6.8 ‘Index and Pointer Sorting’) • Binary Search – See the notation conventions (e.g. log2N = lg N) • Terminology and notation: – log2N = lg N – Use interchangeably: • Runtime and time complexity • Record and item 2 Sorting 3 Sorting • Sort an array, A, of items (numbers, strings, etc.). • Why sort it? – To use in binary search. – To compute rankings, statistics (min/max, top-10, top-100, median). – Check that there are no duplicates – Set intersection and union are easier to perform between 2 sorted sets – …. • We will study several sorting algorithms, – Pros/cons, behavior . • Insertion sort • If time permits, we will cover selection sort as well. 4 Properties of sorting • Stable: – It does not change the relative order of items whose keys are equal. • Adaptive: – The time complexity will depend on the input • E.g. if the input data is almost sorted, it will run significantly faster than if not sorted. • see later insertion sort vs selection sort. 5 Other aspects of sorting • Time complexity: worst/best/average • Number of data moves: copy/swap the DATA RECORDS – One data move = 1 copy operation of a complete data record – Data moves are NOT updates of variables independent of record size (e.g. loop counter ) • Space complexity: Extra Memory used – Do NOT count the space needed to hold the INPUT data, only extra space (e.g.
    [Show full text]
  • Binary Search Algorithm Anthony Lin¹* Et Al
    WikiJournal of Science, 2019, 2(1):5 doi: 10.15347/wjs/2019.005 Encyclopedic Review Article Binary search algorithm Anthony Lin¹* et al. Abstract In In computer science, binary search, also known as half-interval search,[1] logarithmic search,[2] or binary chop,[3] is a search algorithm that finds a position of a target value within a sorted array.[4] Binary search compares the target value to an element in the middle of the array. If they are not equal, the half in which the target cannot lie is eliminated and the search continues on the remaining half, again taking the middle element to compare to the target value, and repeating this until the target value is found. If the search ends with the remaining half being empty, the target is not in the array. Binary search runs in logarithmic time in the worst case, making 푂(log 푛) comparisons, where 푛 is the number of elements in the array, the 푂 is ‘Big O’ notation, and 푙표푔 is the logarithm.[5] Binary search is faster than linear search except for small arrays. However, the array must be sorted first to be able to apply binary search. There are spe- cialized data structures designed for fast searching, such as hash tables, that can be searched more efficiently than binary search. However, binary search can be used to solve a wider range of problems, such as finding the next- smallest or next-largest element in the array relative to the target even if it is absent from the array. There are numerous variations of binary search.
    [Show full text]
  • Sorting Algorithms
    Sorting Algorithms Chapter 12 Our first sort: Selection Sort • General Idea: min SORTED UNSORTED SORTED UNSORTED What is the invariant of this sort? Selection Sort • Let A be an array of n ints, and we wish to sort these keys in non-decreasing order. • Algorithm: for i = 0 to n-2 do find j, i < j < n-1, such that A[j] < A[k], k, i < k < n-1. swap A[j] with A[i] • This algorithm works in place, meaning it uses its own storage to perform the sort. Selection Sort Example 66 44 99 55 11 88 22 77 33 11 44 99 55 66 88 22 77 33 11 22 99 55 66 88 44 77 33 11 22 33 55 66 88 44 77 99 11 22 33 44 66 88 55 77 99 11 22 33 44 55 88 66 77 99 11 22 33 44 55 66 88 77 99 11 22 33 44 55 66 77 88 99 11 22 33 44 55 66 77 88 99 Selection Sort public static void sort (int[] data, int n) { int i,j,minLocation; for (i=0; i<=n-2; i++) { minLocation = i; for (j=i+1; j<=n-1; j++) if (data[j] < data[minLocation]) minLocation = j; swap(data, minLocation, i); } } Run time analysis • Worst Case: Search for 1st min: n-1 comparisons Search for 2nd min: n-2 comparisons ... Search for 2nd-to-last min: 1 comparison Total comparisons: (n-1) + (n-2) + ... + 1 = O(n2) • Average Case and Best Case: O(n2) also! (Why?) Selection Sort (another algorithm) public static void sort (int[] data, int n) { int i, j; for (i=0; i<=n-2; i++) for (j=i+1; j<=n-1; j++) if (data[j] < data[i]) swap(data, i, j); } Is this any better? Insertion Sort • General Idea: SORTED UNSORTED SORTED UNSORTED What is the invariant of this sort? Insertion Sort • Let A be an array of n ints, and we wish to sort these keys in non-decreasing order.
    [Show full text]
  • Challenge 1: Pair Insertion Sort
    VerifyThis – Competition, Uppsala 2017 Challenge 1: Pair Insertion Sort Although it is an algorithm with O(n²) complexity, this sorting algorithm is used in modern library implementations. When dealing with smaller numbers of elements, insertion sorts performs better than, e.g., quicksort due to a lower overhead. It can be implemented more efficiently if the array traversal (and rearrangement) is not repeated for every element individually. A Pair Insertion Sort in which two elements are handled at a time is used by Oracle's implementation of the Java Development Kit (JDK) for sorting primitive values. In the following code snippet a is the array to be sorted, and the integer variables left and right are valid indices into a that set the range to be sorted. for (int k = left; ++left <= right; k = ++left) { int a1 = a[k], a2 = a[left]; if (a1 < a2) { a2 = a1; a1 = a[left]; } while (a1 < a[--k]) { a[k + 2] = a[k]; } a[++k + 1] = a1; while (a2 < a[--k]) { a[k + 1] = a[k]; } a[k + 1] = a2; } int last = a[right]; while (last < a[--right]) { a[right + 1] = a[right]; } a[right + 1] = last; (in DualPivotQuicksort.java line 245ff, used for java.util.Arrays.sort(int[]) ) (This is an optimised version which uses the borders a[left] and a[right] as sentinels.) While the problem is proposed here as a Java implementation, the challenge does not use specific language features and can be formulated in other languages easily. VerifyThis – Competition, Uppsala 2017 A simplified variant of the algorithm in pseudo code for sorting an array A whose indices
    [Show full text]
  • Unit 5 Searching and Sorting Algorithms
    Sri vidya college of engineering and technology course material UNIT 5 SEARCHING AND SORTING ALGORITHMS INTRODUCTION TO SEARCHING ALGORITHMS Searching is an operation or a technique that helps finds the place of a given element or value in the list. Any search is said to be successful or unsuccessful depending upon whether the element that is being searched is found or not. Some of the standard searching technique that is being followed in data structure is listed below: 1. Linear Search 2. Binary Search LINEAR SEARCH Linear search is a very basic and simple search algorithm. In Linear search, we search an element or value in a given array by traversing the array from the starting, till the desired element or value is found. It compares the element to be searched with all the elements present in the array and when the element is matched successfully, it returns the index of the element in the array, else it return -1. Linear Search is applied on unsorted or unordered lists, when there are fewer elements in a list. For Example, Linear Search 10 14 19 26 27 31 33 35 42 44 = 33 Algorithm Linear Search ( Array A, Value x) Step 1: Set i to 1 Step 2: if i > n then go to step 7 Step 3: if A[i] = x then go to step 6 Step 4: Set i to i + 1 Step 5: Go to Step 2 EC 8393/Fundamentals of data structures in C unit 5 Step 6: Print Element x Found at index i and go to step 8 Step 7: Print element not found Step 8: Exit Pseudocode procedure linear_search (list, value) for each item in the list if match item == value return the item‟s location end if end for end procedure Features of Linear Search Algorithm 1.
    [Show full text]
  • Introduction to Algorithms, Third Edition
    Introduction to Algorithms Third Edition IFoundations Introduction This part will start you thinking about designing and analyzing algorithms. It is intended to be a gentle introduction to how we specify algorithms, some of the design strategies we will use throughout this book, and many of the fundamental ideas used in algorithm analysis. Later parts of this book will build upon this base. Chapter 1 provides an overview of algorithms and their place in modern com- puting systems. This chapter defines what an algorithm is and lists some examples. It also makes a case that we should consider algorithms as a technology, along- side technologies such as fast hardware, graphical user interfaces, object-oriented systems, and networks. In Chapter 2, we see our first algorithms, which solve the problem of sorting asequenceofn numbers. They are written in a pseudocode which, although not directly translatable to any conventional programming language, conveys the struc- ture of the algorithm clearly enough that you should be able to implement it in the language of your choice. The sorting algorithms we examine are insertion sort, which uses an incremental approach, and merge sort, which uses a recursive tech- nique known as “divide-and-conquer.” Although the time each requires increases with the value of n,therateofincreasediffersbetweenthetwoalgorithms.We determine these running times in Chapter 2, and we develop a useful notation to express them. Chapter 3 precisely defines this notation, which we call asymptotic notation. It starts by defining several asymptotic notations, which we use for bounding algo- rithm running times from above and/or below. The rest of Chapter 3 is primarily apresentationofmathematicalnotation,moretoensurethatyouruseofnotation matches that in this book than to teach you new mathematical concepts.
    [Show full text]
  • Algorithms ROBERT SEDGEWICK | KEVIN WAYNE
    Algorithms ROBERT SEDGEWICK | KEVIN WAYNE 2.2 MERGESORT ‣ mergesort ‣ bottom-up mergesort Algorithms ‣ sorting complexity FOURTH EDITION ‣ divide-and-conquer ROBERT SEDGEWICK | KEVIN WAYNE https://algs4.cs.princeton.edu Last updated on 9/26/19 4:36 AM Two classic sorting algorithms: mergesort and quicksort Critical components in the world’s computational infrastructure. Full scientific understanding of their properties has enabled us to develop them into practical system sorts. Quicksort honored as one of top 10 algorithms of 20th century in science and engineering. Mergesort. [this lecture] ... Quicksort. [next lecture] ... 2 2.2 MERGESORT ‣ mergesort ‣ bottom-up mergesort Algorithms ‣ sorting complexity ‣ divide-and-conquer ROBERT SEDGEWICK | KEVIN WAYNE https://algs4.cs.princeton.edu Mergesort Basic plan. Divide array into two halves. Recursively sort each half. Merge two halves. input M E R G E S O R T E X A M P L E sort left half E E G M O R R S T E X A M P L E sort right half E E G M O R R S A E E L M P T X merge results A E E E E G L M M O P R R S T X Mergesort overview 4 Abstract in-place merge demo Goal. Given two sorted subarrays a[lo] to a[mid] and a[mid+1] to a[hi], replace with sorted subarray a[lo] to a[hi]. lo mid mid+1 hi a[] E E G M R A C E R T sorted sorted 5 Mergesort: Transylvanian–Saxon folk dance http://www.youtube.com/watch?v=XaqR3G_NVoo 6 Merging: Java implementation private static void merge(Comparable[] a, Comparable[] aux, int lo, int mid, int hi) { for (int k = lo; k <= hi; k++) copy aux[k] = a[k]; int i = lo, j = mid+1; merge for (int k = lo; k <= hi; k++) { if (i > mid) a[k] = aux[j++]; else if (j > hi) a[k] = aux[i++]; else if (less(aux[j], aux[i])) a[k] = aux[j++]; else a[k] = aux[i++]; } } lo i mid j hi aux[] A G L O R H I M S T k a[] A G H I L M 7 Mergesort quiz 1 How many calls does merge() make to less() in order to merge two sorted subarrays, each of length n / 2, into a sorted array of length n? A.
    [Show full text]