Bias
The usual leftist tree is a ''height-biased'' leftist tree. However, other biases can exist, such as in the ''weight-biased'' leftist tree.S-value
S-values of a leftist tree The s-value (or rank) of a node is the distance from that node to the nearest empty position in the subtree rooted at that node. Put another way, the s-value of anull
child is implicitly zero. Other nodes have an s-value equal to one more the minimum of their children's s-values. Thus, in the example at right, all nodes with at least one missing child have an s-value of 1, while node 4 has an s-value of 2, since its right child (8) has an s-value of 1. (In some descriptions, the s-value of null children is assumed to be −1.)
Knowing the shortest path to the nearest missing leaf in the subtree rooted at ''x'' is exactly of ''s''(''x''), every node at depth ''s''(''x'')−1 or less has exactly 2 children since ''s''(''x'') would have been less if not. Meaning that the size of the tree rooted at ''x'' is at least . Thus, ''s''(''x'') is at most , ''m'' being the number of nodes of the subtree rooted at ''x''.
Operations on a height biased leftist tree
Most operations on a Height Biased Leftist Tree are done using the merge operation.Merging two Min HBLTs
The merge operation takes two Min HBLTs as input and returns a Min HBLT containing all the nodes in the original Min HBLTs put together. If either of A or B is empty, the merge returns the other one. In case of Min HBLTs, assume we have two trees rooted at A and B where A.key B.key. Otherwise we can swap A and B so that the condition above holds. The merge is done recursively by merging B with A's right subtree. This might change the S-value of A's right subtree. To maintain the leftist tree property, after each merge is done, we check if the S-value of right subtree became bigger than the S-value of left subtree during the recursive merge calls. If so, we swap the right and left subtrees (If one child is missing, it should be the right one). Since we assumed that A's root is greater than B's, the heap property is also maintained.Pseudocode for merging two min height biased leftist trees
MERGE(A, B) if A = null return B if B = null return A if A.key > B.key return MERGE(B, A) A.right := MERGE (A.right, B) ''// the result cannot be null since B is non-null'' if A.left = null then SWAP(A.left, A.right) A.s_value := 1 ''// since the right subtree is null, the shortest path to a descendant leaf from node A is 1'' return A if A.right.s_value > A.left.s_value then SWAP (A.right, A.left) A.s_value := A.right.s_value + 1 return AJava code for merging two min height biased leftist trees
Haskell code for merging two min height biased leftist trees
Example
An example of how the merge operation in a leftist tree works is depicted. The boxes represent each merge call.Insertion into a Min HBLT
Insertion is done using the merge operation. An insertion of a node into an already existing Min HBLT, creates a HBLT tree of size one with that node and merges it with the existing tree. INSERT (''A'', ''x'') ''B'' := CREATE_TREE(''x'') return MERGE(''A'', ''B'')Deletion of Min element from Min HBLT
The Min element in a Min HBLT is the root. Thus, in order to delete the Min, the root is deleted and its subtrees are merged to form the new Min HBLT. DELETE_MIN(''A'') ''x'' := ''A''.key ''A'' := MERGE (''A''.right, ''A''.left) return ''x''Initializing a height biased leftist tree
Initializing a height biased leftist tree is primarily done in one of two ways. The first is to merge each node one at a time into one HBLT. This process is inefficient and takes O(''nlogn'') time. The other approach is to use a queue to store each node and resulting tree. The first two items in the queue are removed, merged, and placed back into the queue. This can initialize a HBLT in O(''n'') time. This approach is detailed in the three diagrams supplied. A min height biased leftist tree is shown. To initialize a min HBLT, place each element to be added to the tree into a queue. In the example (see Part 1 to the left), the set of numbers , 8, 10, 9, 1, 3, 5, 6, 11are initialized. Each line of the diagram represents another cycle of the algorithm, depicting the contents of the queue. The first five steps are easy to follow. Notice that the freshly created HBLT is added to the end of the queue. In the fifth step, the first occurrence of an s-value greater than 1 occurs. The sixth step shows two trees merged with each other, with predictable results. In part 2 a slightly more complex merge happens. The tree with the lower value (tree x) has a right child, so merge must be called again on the subtree rooted by tree x's right child and the other tree. After the merge with the subtree, the resulting tree is put back into tree x. The s-value of the right child (s=2) is now greater than the s-value of the left child (s=1), so they must be swapped. The s-value of the root node 4 is also now 2. Part 3 is the most complex. Here, we recursively call merge twice (each time with the right child 's subtree that is not grayed out). This uses the same process described for part 2.Deletion of an arbitrary element from a Min HBLT
If we have a pointer to a node x in a Min HBLT, we can delete it as follows: Replace the node x with the result of merging its two subtrees and update the s-values of the nodes on the path from x to the root, swapping the right and left subtrees if necessary to maintain the leftist tree property. The upward traversal is continued until either we hit the root or the s-values does not change. Since we are deleting an element, the S-values on the path traversed cannot be increased. Every node that is already the right child of its parent and causes its parent's s-value to be decreased, will remain on the right. Every node that is its parent's left child and causes the parent's s-value to be decreased has to be swapped with its right sibling if the s-value becomes lower than the current s-value of the right child. Each node needs to have a pointer to its parent, so that we can traverse the path to the root updating the s-values. When the traversal ends at some node y, the nodes traversed all lie on the rightmost path rooted at node y. An example is shown below. It follows that the number of nodes traversed is at most log(m), m being the size of the subtree rooted at y. Thus, this operation also takes O(lg m) to perform.Merging two Min WBLTs
The merge operation in WBLTs can be done using a single top to bottom traversal since the number of nodes in the subtrees are known prior to recursive call to merge. Thus, we can swap left and right subtrees if the total number of nodes in the right subtree and the tree to be merged is bigger than the number of nodes in the left subtree. This allows the operations be completed in a single path and so improves the time complexity of the operations by a constant factor. The merge operation is depicted in the graph below.Other operations on WBLT
Insertions and deletion of the min element can be done in the same as for HBLTs using the merge operation. Although WBLTs outperform HBLTs in merge, insertion and deletion of the Min key by a constant factor, the ''O''(log ''n'') bound is not guaranteed when deleting an arbitrary element from WBLTs, since θ(''n'') nodes have to be traversed. If this was an HBLT, then deleting the leaf node with key 60 would take ''O''(1) time and updating the s-values is not needed since the length of rightmost path for all the nodes does not change. But in an WBLT tree, we have to update the weight of each node back to the root, which takes ''O''(''n'') worst case.Variants
Several variations on the basic leftist tree exist, which make only minor changes to the basic algorithm: * The choice of the left child as the taller one is arbitrary; a "rightist tree" would work just as well. * It is possible to avoid swapping children, but instead record ''which'' child is the tallest (in, for example, the least significant bit of the s-value) and use that in the merge operation. * The s-value used to decide which side to merge with could use a metric other than height. For example, weight (number of nodes) could be used.References
Further reading
* * {{cite book , author1=Dinesh P. Mehta , author2=Sartaj Sahni , title=Handbook of Data Structures and Applications , chapter=Chapter 5: Leftist trees , chapter-url=https://books.google.com/books?id=fQVZy1zcpJkC&pg=SA5-PA1&lpg=SA5-PA1 , date=28 October 2004 , publisher=CRC Press , isbn=978-1-4200-3517-9External links