Note: This assignment is used to assess the required outcomes for the course, as outlined in the course syllabus. These outcomes are:
These will be assessed using the following rubric:
In order to earn a course grade of C- or better, the assessment must result in Effective or Highly Effective for each outcome. |
Educational Objectives: After completing this assignment, the student should be able to accomplish the following:
Deliverables: One file
assign4.pdf
containing your analysis of List::Sort as answers to the first three questions and an analysis of efficiency of hash tables using HashTable::Analysis as answer to the fourth question.
By a scalability curve for an algorithm implementation we shall mean an equation whose form is determined by known asymptotic properties of the algorithm and whose coefficients are determined by a least squares fit to actual timing data for the algorithm as a function of input size. For example, the merge sort algorithm, implemented as a generic algorithm named g_merge_sort(), is known to have asymptotic runtime Θ(n log n), and we will use the following for its form:
R = A + B n log n
where R is the predicted run time on input of size n. To obtain the concrete scalability curve, we need to obtain actual timing data for the sort and use that data to find optimal values for the coefficients A and B. Note this curve will depend on the implementation all the way from source code to hardware, so it is important to keep the compiler and testing platform the same in order to compare efficiencies of different sorts using their concrete scalability curves.
The method for finding the coefficients A and B is the method of least squares. Assume that we have sample runtime data as follows:
Input size: n1 n2 ... nk Measured runtime: t1 t2 ... tk
and the scalability form is given by
f(n) = A + B g(n)
Define the total square error of the approximation to be the sum of squares of errors at each data point:
E = Σ [ti - f(ni)]2
where the sum is taken from i = 1 to i = k, k being the number of data points. The key observation in the method of least squares is that total square error E is minimized when the gadient of E is zero, that is, where all three partial derivatives DAE and DBE are zero. Calculating these partial derivatives gives:
DXE = 2 Σ [ti - f(ni)] DXf = 2 Σ [ti - (A + B g(ni))] DXf
(where X is A or B). This gives the partial derivatives of E in terms of those of f, which may be calculated to be:
DAf = 1
DBf = g(n)
(because n and g(n) are constant with respect to A and B.) Substituting these into the previous formula and setting the results equal to zero yields the following equations:
A Σ 1 + B Σ g(ni) = Σ ti
A Σ g(ni) + B Σ (g(ni))2 = Σ ti g(ni)
Rearranging and using Σ 1 = k yields:
k A + [Σ g(ni)] B = Σ ti [Σ g(ni)] A + [Σ (g(ni))2] B = Σ ti g(ni)
These are two linear equations in the unknowns A and B. With even a small amount of luck, they have a unique solution, and thus optimal values of A and B are determined. (Here is a link to a more detailed derivation in the quadratic case.)
Note that all of the coefficients in these equations may be calculated from the original data table and knowledge of the function g(n), in a spreadsheet or in a simple stand-alone program. The solution to the system of equations itself is probably easiest to find by hand by row-reducing the 2x3 matrix of coefficients to upper triangular form and then back-substitution.
Answer questions 1-4 in Assignment 4.