In
computational complexity theory, the potential method is a method used to analyze the
amortized time and space complexity of a
data structure
In computer science, a data structure is a data organization, management, and storage format that is usually chosen for efficient access to data. More precisely, a data structure is a collection of data values, the relationships among them, a ...
, a measure of its performance over sequences of operations that smooths out the cost of infrequent but expensive operations.
[.]
Definition of amortized time
In the potential method, a function Φ is chosen that maps states of the data structure to non-negative numbers. If ''S'' is a state of the data structure, Φ(''S'') represents work that has been accounted for ("paid for") in the amortized analysis but not yet performed. Thus, Φ(''S'') may be thought of as calculating the amount of
potential energy
In physics, potential energy is the energy held by an object because of its position relative to other objects, stresses within itself, its electric charge, or other factors.
Common types of potential energy include the gravitational potentia ...
stored in that state.
The potential value prior to the operation of initializing a data structure is defined to be zero. Alternatively, Φ(''S'') may be thought of as representing the amount of disorder in state ''S'' or its distance from an ideal state.
Let ''o'' be any individual operation within a sequence of operations on some data structure, with ''S''
before denoting the state of the data structure prior to operation ''o'' and ''S''
after denoting its state after operation ''o'' has completed. Once Φ has been chosen, the amortized time for operation ''o'' is defined to be
:
where ''C'' is a non-negative constant of proportionality (in units of time) that must remain fixed throughout the analysis.
That is, the amortized time is defined to be the actual time taken by the operation plus ''C'' times the difference in potential caused by the operation.
When studying
asymptotic computational complexity In computational complexity theory, asymptotic computational complexity is the usage of asymptotic analysis for the estimation of computational complexity of algorithms and computational problems, commonly associated with the usage of the big O no ...
using
big O notation
Big ''O'' notation is a mathematical notation that describes the limiting behavior of a function when the argument tends towards a particular value or infinity. Big O is a member of a family of notations invented by Paul Bachmann, Edmund Lan ...
, constant factors are irrelevant and so the constant ''C'' is usually omitted.
Relation between amortized and actual time
Despite its artificial appearance, the total amortized time of a sequence of operations provides a valid
upper bound on the actual time for the same sequence of operations.
For any sequence of operations
, define:
* The total amortized time:
* The total actual time:
Then:
:
where the sequence of potential function values forms a
telescoping series in which all terms other than the initial and final potential function values cancel in pairs. Rearranging this, we obtain:
:
Since
and
,
, so the amortized time can be used to provide an accurate upper bound on the actual time of a sequence of operations, even though the amortized time for an individual operation may vary widely from its actual time.
Amortized analysis of worst-case inputs
Typically, amortized analysis is used in combination with a
worst case assumption about the input sequence. With this assumption, if ''X'' is a type of operation that may be performed by the data structure, and ''n'' is an integer defining the size of the given data structure (for instance, the number of items that it contains), then the amortized time for operations of type ''X'' is defined to be the maximum, among all possible sequences of operations on data structures of size ''n'' and all operations ''o
i'' of type ''X'' within the sequence, of the amortized time for operation ''o
i''.
With this definition, the time to perform a sequence of operations may be estimated by multiplying the amortized time for each type of operation in the sequence by the number of operations of that type.
Examples
Dynamic array
A
dynamic array
In computer science, a dynamic array, growable array, resizable array, dynamic table, mutable array, or array list is a random access, variable-size list data structure that allows elements to be added or removed. It is supplied with standard lib ...
is a data structure for maintaining an
array of items, allowing both
random access
Random access (more precisely and more generally called direct access) is the ability to access an arbitrary element of a sequence in equal time or any datum from a population of addressable elements roughly as easily and efficiently as any othe ...
to positions within the array and the ability to increase the array size by one. It is available in
Java as the "ArrayList" type and in
Python as the "list" type.
A dynamic array may be implemented by a data structure consisting of an array ''A'' of items, of some length ''N'', together with a number ''n'' ≤ ''N'' representing the positions within the array that have been used so far. With this structure, random accesses to the dynamic array may be implemented by accessing the same cell of the internal array ''A'', and when ''n'' < ''N'' an operation that increases the dynamic array size may be implemented simply by incrementing ''n''. However, when ''n'' = ''N'', it is necessary to resize ''A'', and a common strategy for doing so is to double its size, replacing ''A'' by a new array of length 2''n''.
This structure may be analyzed using the potential function:
:::Φ = 2''n'' − ''N''
Since the resizing strategy always causes ''A'' to be at least half-full, this potential function is always non-negative, as desired.
When an increase-size operation does not lead to a resize operation, Φ increases by 2, a constant. Therefore, the constant actual time of the operation and the constant increase in potential combine to give a constant amortized time for an operation of this type.
However, when an increase-size operation causes a resize, the potential value of Φ decreases to zero after the resize. Allocating a new internal array ''A'' and copying all of the values from the old internal array to the new one takes O(''n'') actual time, but (with an appropriate choice of the constant of proportionality ''C'') this is entirely cancelled by the decrease in the potential function, leaving again a constant total amortized time for the operation.
The other operations of the data structure (reading and writing array cells without changing the array size) do not cause the potential function to change and have the same constant amortized time as their actual time.
Therefore, with this choice of resizing strategy and potential function, the potential method shows that all dynamic array operations take constant amortized time. Combining this with the inequality relating amortized time and actual time over sequences of operations, this shows that any sequence of ''n'' dynamic array operations takes O(''n'') actual time in the worst case, despite the fact that some of the individual operations may themselves take a linear amount of time.
When the dynamic array includes operations that decrease the array size as well as increasing it, the potential function must be modified to prevent it from becoming negative. One way to do this is to replace the formula above for Φ by its
absolute value
In mathematics, the absolute value or modulus of a real number x, is the non-negative value without regard to its sign. Namely, , x, =x if is a positive number, and , x, =-x if x is negative (in which case negating x makes -x positive), an ...
.
Multi-Pop Stack
Consider a
stack
Stack may refer to:
Places
* Stack Island, an island game reserve in Bass Strait, south-eastern Australia, in Tasmania’s Hunter Island Group
* Blue Stack Mountains, in Co. Donegal, Ireland
People
* Stack (surname) (including a list of people ...
which supports the following operations:
* Initialize - create an empty stack.
* Push - add a single element on top of the stack, enlarging the stack by 1.
* Pop(''k'') - remove ''k'' elements from the top of the stack, where ''k'' is no more than the current stack size
Pop(''k'') requires O(''k'') time, but we wish to show that all operations take O(1) amortized time.
This structure may be analyzed using the potential function:
:::Φ = number-of-elements-in-stack
This number is always non-negative, as required.
A Push operation takes constant time and increases Φ by 1, so its amortized time is constant.
A Pop operation takes time O(''k'') but also reduces Φ by ''k'', so its amortized time is also constant.
This proves that any sequence of ''m'' operations takes O(''m'') actual time in the worst case.
Binary counter
Consider a
counter represented as a
binary number
A binary number is a number expressed in the base-2 numeral system or binary numeral system, a method of mathematical expression which uses only two symbols: typically "0" (zero) and "1" ( one).
The base-2 numeral system is a positional notatio ...
and supporting the following operations:
* Initialize: create a counter with value 0.
* Inc: add 1 to the counter.
* Read: return the current counter value.
For this example, we are ''not'' using the
transdichotomous machine model, but instead require one unit of time per bit operation in the increment. We wish to show that Inc takes O(1) amortized time.
This structure may be analyzed using the potential function:
:::Φ = number-of-bits-equal-to-1 =
hammingweight(counter)
This number is always non-negative and starts with 0, as required.
An Inc operation flips the
least significant bit
In computing, bit numbering is the convention used to identify the bit positions in a binary number.
Bit significance and indexing
In computing, the least significant bit (LSB) is the bit position in a binary integer representing the binary 1 ...
. Then, if the LSB were flipped from 1 to 0, then the next bit is also flipped. This goes on until finally a bit is flipped from 0 to 1, at which point the flipping stops. If the counter initially ends in ''k'' 1 bits, we flip a total of ''k''+1 bits, taking actual time ''k''+1 and reducing the potential by ''k''−1, so the amortized time is 2. Hence, the actual time for running ''m'' Inc operations is O(''m'').
Applications
The potential function method is commonly used to analyze
Fibonacci heaps, a form of
priority queue
In computer science, a priority queue is an abstract data-type similar to a regular queue or stack data structure in which each element additionally has a ''priority'' associated with it. In a priority queue, an element with high priority is se ...
in which removing an item takes logarithmic amortized time, and all other operations take constant amortized time. It may also be used to analyze
splay trees, a self-adjusting form of
binary search tree with logarithmic amortized time per operation.
[Goodrich and Tamassia, Section 3.4, "Splay Trees", pp. 185–194.]
References
{{DEFAULTSORT:Potential Method
Analysis of algorithms