Skip to main content

Analyzing Algorithms or Data Structures

 Algorithm 

step-by-step procedure for performing some task in a finite amount of time.

Data structure

systematic way of organizing and accessing data.

The running time of an algorithm or data structure operation typically depends on a number of factors. Its running time can be studied by executing it on various test, inputs and recording the actual time spent in each execution. The running time of an algorithm or data structure method increases with the input size.

The running time is also affected by

Hardware environment 

  • processor
  • clock rate
  • memory
  • disk, etc.

Software environment 

  • operating system
  • programming language
  • compiler
  • interpreter, etc

Limitations - Experimenting running time 

  • Limited set of test inputs
  • Running times to be performed in the same hardware and software environments.
  • Algorithm to be implemented and executed

Experimentation alone is not sufficient, an analytic framework is required to

  • take into account of all possible inputs
  • evaluate the relative efficiency of any two algorithms
  • perform studying a high-level description of the algorithm without implementing & executing.

Components of methodology for algorithm analysis

  • a language to describe algorithms
  • a computational model
  • a metric for measuring running time
  • an approach to characterize the running times

Pseudo Code

  • mixture of natural language and high-level programming
  • more structured code intended for human reader, not for computer.
  • describe the main ideas behind a generic implementation of a data structure or algorithm.

pseudo-code mixes natural language with standard programming language constructs such as 

  • Expressions 
  • Method declaration
  • Decision structures
  • Condition statements
  • Array Indexing
  • Method calls
  • Method returns

Popular posts from this blog

Exercise 2 - Amdahl's Law

A programmer has parallelized 99% of a program, but there is no value in increasing the problem size, i.e., the program will always be run with the same problem size regardless of the number of processors or cores used. What is the expected speedup on 20 processors? Solution As per Amdahl's law, the speedup,  N - No of processors = 20 f - % of parallel operation = 99% = 1 / (1 - 0.99) + (0.99 / 20) = 1 / 0.01 + (0.99 / 20) = 16.807 The expected speedup on 20 processors is 16.807

Gaussian Elimination - Row reduction Algorithm

 Gaussian elimination is a method for solving matrix equations of the form, Ax=b.  This method is also known as the row reduction algorithm. Back  Substitution Solving the last equation for the variable and then work backward into the first equation to solve it.  The fundamental idea is to add multiples of one equation to the others in order to eliminate a variable and to continue this process until only one variable is left. Pivot row The row that is used to perform elimination of a variable from other rows is called the pivot row. Example: Solving a linear equation The augmented matrix for the above equation shall be The equation shall be solved using back substitution. The eliminating the first variable (x1) in the first row (Pivot row) by carrying out the row operation. As the second row become zero, the row will be shifted to bottom by carrying out partial pivoting. Now, the second variable (x2)  shall be eliminated by carrying out the row operation again. ...

Decision Tree Scalability Methods

 The scalable decision tree induction methods are RainForest and BOAT. RainForest method maintains an AVC set for each attribute at each node. AVC stands for Attribute Value Classlabel. BOAT , stands for Bootstrapped Optimistic Algorithm for Tree construction, uses a statistical technique known as bootstrapping., by which several smaller subsets are created. The several trees are created using the subsets and finally the full tree is generated using the trees created by smaller subsets. BOAT was found to be two to three times faster than RainForest.