diff options
Diffstat (limited to 'gson/src/main/java/com/google/gson/internal/LinkedHashTreeMap.java')
-rw-r--r-- | gson/src/main/java/com/google/gson/internal/LinkedHashTreeMap.java | 861 |
1 files changed, 861 insertions, 0 deletions
diff --git a/gson/src/main/java/com/google/gson/internal/LinkedHashTreeMap.java b/gson/src/main/java/com/google/gson/internal/LinkedHashTreeMap.java new file mode 100644 index 00000000..e251ec2f --- /dev/null +++ b/gson/src/main/java/com/google/gson/internal/LinkedHashTreeMap.java @@ -0,0 +1,861 @@ +/* + * Copyright (C) 2010 The Android Open Source Project + * Copyright (C) 2012 Google Inc. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package com.google.gson.internal; + +import java.io.ObjectStreamException; +import java.io.Serializable; +import java.util.AbstractMap; +import java.util.AbstractSet; +import java.util.Arrays; +import java.util.Comparator; +import java.util.ConcurrentModificationException; +import java.util.Iterator; +import java.util.LinkedHashMap; +import java.util.NoSuchElementException; +import java.util.Set; + +/** + * A map of comparable keys to values. Unlike {@code TreeMap}, this class uses + * insertion order for iteration order. Comparison order is only used as an + * optimization for efficient insertion and removal. + * + * <p>This implementation was derived from Android 4.1's TreeMap and + * LinkedHashMap classes. + */ +public final class LinkedHashTreeMap<K, V> extends AbstractMap<K, V> implements Serializable { + @SuppressWarnings({ "unchecked", "rawtypes" }) // to avoid Comparable<Comparable<Comparable<...>>> + private static final Comparator<Comparable> NATURAL_ORDER = new Comparator<Comparable>() { + public int compare(Comparable a, Comparable b) { + return a.compareTo(b); + } + }; + + Comparator<? super K> comparator; + Node<K, V>[] table; + final Node<K, V> header; + int size = 0; + int modCount = 0; + int threshold; + + /** + * Create a natural order, empty tree map whose keys must be mutually + * comparable and non-null. + */ + @SuppressWarnings("unchecked") // unsafe! this assumes K is comparable + public LinkedHashTreeMap() { + this((Comparator<? super K>) NATURAL_ORDER); + } + + /** + * Create a tree map ordered by {@code comparator}. This map's keys may only + * be null if {@code comparator} permits. + * + * @param comparator the comparator to order elements with, or {@code null} to + * use the natural ordering. + */ + @SuppressWarnings({ "unchecked", "rawtypes" }) // unsafe! if comparator is null, this assumes K is comparable + public LinkedHashTreeMap(Comparator<? super K> comparator) { + this.comparator = comparator != null + ? comparator + : (Comparator) NATURAL_ORDER; + this.header = new Node<K, V>(); + this.table = new Node[16]; // TODO: sizing/resizing policies + this.threshold = (table.length / 2) + (table.length / 4); // 3/4 capacity + } + + @Override public int size() { + return size; + } + + @Override public V get(Object key) { + Node<K, V> node = findByObject(key); + return node != null ? node.value : null; + } + + @Override public boolean containsKey(Object key) { + return findByObject(key) != null; + } + + @Override public V put(K key, V value) { + if (key == null) { + throw new NullPointerException("key == null"); + } + Node<K, V> created = find(key, true); + V result = created.value; + created.value = value; + return result; + } + + @Override public void clear() { + Arrays.fill(table, null); + size = 0; + modCount++; + + // Clear all links to help GC + Node<K, V> header = this.header; + for (Node<K, V> e = header.next; e != header; ) { + Node<K, V> next = e.next; + e.next = e.prev = null; + e = next; + } + + header.next = header.prev = header; + } + + @Override public V remove(Object key) { + Node<K, V> node = removeInternalByKey(key); + return node != null ? node.value : null; + } + + /** + * Returns the node at or adjacent to the given key, creating it if requested. + * + * @throws ClassCastException if {@code key} and the tree's keys aren't + * mutually comparable. + */ + Node<K, V> find(K key, boolean create) { + Comparator<? super K> comparator = this.comparator; + Node<K, V>[] table = this.table; + int hash = secondaryHash(key.hashCode()); + int index = hash & (table.length - 1); + Node<K, V> nearest = table[index]; + int comparison = 0; + + if (nearest != null) { + // Micro-optimization: avoid polymorphic calls to Comparator.compare(). + @SuppressWarnings("unchecked") // Throws a ClassCastException below if there's trouble. + Comparable<Object> comparableKey = (comparator == NATURAL_ORDER) + ? (Comparable<Object>) key + : null; + + while (true) { + comparison = (comparableKey != null) + ? comparableKey.compareTo(nearest.key) + : comparator.compare(key, nearest.key); + + // We found the requested key. + if (comparison == 0) { + return nearest; + } + + // If it exists, the key is in a subtree. Go deeper. + Node<K, V> child = (comparison < 0) ? nearest.left : nearest.right; + if (child == null) { + break; + } + + nearest = child; + } + } + + // The key doesn't exist in this tree. + if (!create) { + return null; + } + + // Create the node and add it to the tree or the table. + Node<K, V> header = this.header; + Node<K, V> created; + if (nearest == null) { + // Check that the value is comparable if we didn't do any comparisons. + if (comparator == NATURAL_ORDER && !(key instanceof Comparable)) { + throw new ClassCastException(key.getClass().getName() + " is not Comparable"); + } + created = new Node<K, V>(nearest, key, hash, header, header.prev); + table[index] = created; + } else { + created = new Node<K, V>(nearest, key, hash, header, header.prev); + if (comparison < 0) { // nearest.key is higher + nearest.left = created; + } else { // comparison > 0, nearest.key is lower + nearest.right = created; + } + rebalance(nearest, true); + } + + if (size++ > threshold) { + doubleCapacity(); + } + modCount++; + + return created; + } + + @SuppressWarnings("unchecked") + Node<K, V> findByObject(Object key) { + try { + return key != null ? find((K) key, false) : null; + } catch (ClassCastException e) { + return null; + } + } + + /** + * Returns this map's entry that has the same key and value as {@code + * entry}, or null if this map has no such entry. + * + * <p>This method uses the comparator for key equality rather than {@code + * equals}. If this map's comparator isn't consistent with equals (such as + * {@code String.CASE_INSENSITIVE_ORDER}), then {@code remove()} and {@code + * contains()} will violate the collections API. + */ + Node<K, V> findByEntry(Entry<?, ?> entry) { + Node<K, V> mine = findByObject(entry.getKey()); + boolean valuesEqual = mine != null && equal(mine.value, entry.getValue()); + return valuesEqual ? mine : null; + } + + private boolean equal(Object a, Object b) { + return a == b || (a != null && a.equals(b)); + } + + /** + * Applies a supplemental hash function to a given hashCode, which defends + * against poor quality hash functions. This is critical because HashMap + * uses power-of-two length hash tables, that otherwise encounter collisions + * for hashCodes that do not differ in lower or upper bits. + */ + private static int secondaryHash(int h) { + // Doug Lea's supplemental hash function + h ^= (h >>> 20) ^ (h >>> 12); + return h ^ (h >>> 7) ^ (h >>> 4); + } + + /** + * Removes {@code node} from this tree, rearranging the tree's structure as + * necessary. + * + * @param unlink true to also unlink this node from the iteration linked list. + */ + void removeInternal(Node<K, V> node, boolean unlink) { + if (unlink) { + node.prev.next = node.next; + node.next.prev = node.prev; + node.next = node.prev = null; // Help the GC (for performance) + } + + Node<K, V> left = node.left; + Node<K, V> right = node.right; + Node<K, V> originalParent = node.parent; + if (left != null && right != null) { + + /* + * To remove a node with both left and right subtrees, move an + * adjacent node from one of those subtrees into this node's place. + * + * Removing the adjacent node may change this node's subtrees. This + * node may no longer have two subtrees once the adjacent node is + * gone! + */ + + Node<K, V> adjacent = (left.height > right.height) ? left.last() : right.first(); + removeInternal(adjacent, false); // takes care of rebalance and size-- + + int leftHeight = 0; + left = node.left; + if (left != null) { + leftHeight = left.height; + adjacent.left = left; + left.parent = adjacent; + node.left = null; + } + int rightHeight = 0; + right = node.right; + if (right != null) { + rightHeight = right.height; + adjacent.right = right; + right.parent = adjacent; + node.right = null; + } + adjacent.height = Math.max(leftHeight, rightHeight) + 1; + replaceInParent(node, adjacent); + return; + } else if (left != null) { + replaceInParent(node, left); + node.left = null; + } else if (right != null) { + replaceInParent(node, right); + node.right = null; + } else { + replaceInParent(node, null); + } + + rebalance(originalParent, false); + size--; + modCount++; + } + + Node<K, V> removeInternalByKey(Object key) { + Node<K, V> node = findByObject(key); + if (node != null) { + removeInternal(node, true); + } + return node; + } + + private void replaceInParent(Node<K, V> node, Node<K, V> replacement) { + Node<K, V> parent = node.parent; + node.parent = null; + if (replacement != null) { + replacement.parent = parent; + } + + if (parent != null) { + if (parent.left == node) { + parent.left = replacement; + } else { + assert (parent.right == node); + parent.right = replacement; + } + } else { + int index = node.hash & (table.length - 1); + table[index] = replacement; + } + } + + /** + * Rebalances the tree by making any AVL rotations necessary between the + * newly-unbalanced node and the tree's root. + * + * @param insert true if the node was unbalanced by an insert; false if it + * was by a removal. + */ + private void rebalance(Node<K, V> unbalanced, boolean insert) { + for (Node<K, V> node = unbalanced; node != null; node = node.parent) { + Node<K, V> left = node.left; + Node<K, V> right = node.right; + int leftHeight = left != null ? left.height : 0; + int rightHeight = right != null ? right.height : 0; + + int delta = leftHeight - rightHeight; + if (delta == -2) { + Node<K, V> rightLeft = right.left; + Node<K, V> rightRight = right.right; + int rightRightHeight = rightRight != null ? rightRight.height : 0; + int rightLeftHeight = rightLeft != null ? rightLeft.height : 0; + + int rightDelta = rightLeftHeight - rightRightHeight; + if (rightDelta == -1 || (rightDelta == 0 && !insert)) { + rotateLeft(node); // AVL right right + } else { + assert (rightDelta == 1); + rotateRight(right); // AVL right left + rotateLeft(node); + } + if (insert) { + break; // no further rotations will be necessary + } + + } else if (delta == 2) { + Node<K, V> leftLeft = left.left; + Node<K, V> leftRight = left.right; + int leftRightHeight = leftRight != null ? leftRight.height : 0; + int leftLeftHeight = leftLeft != null ? leftLeft.height : 0; + + int leftDelta = leftLeftHeight - leftRightHeight; + if (leftDelta == 1 || (leftDelta == 0 && !insert)) { + rotateRight(node); // AVL left left + } else { + assert (leftDelta == -1); + rotateLeft(left); // AVL left right + rotateRight(node); + } + if (insert) { + break; // no further rotations will be necessary + } + + } else if (delta == 0) { + node.height = leftHeight + 1; // leftHeight == rightHeight + if (insert) { + break; // the insert caused balance, so rebalancing is done! + } + + } else { + assert (delta == -1 || delta == 1); + node.height = Math.max(leftHeight, rightHeight) + 1; + if (!insert) { + break; // the height hasn't changed, so rebalancing is done! + } + } + } + } + + /** + * Rotates the subtree so that its root's right child is the new root. + */ + private void rotateLeft(Node<K, V> root) { + Node<K, V> left = root.left; + Node<K, V> pivot = root.right; + Node<K, V> pivotLeft = pivot.left; + Node<K, V> pivotRight = pivot.right; + + // move the pivot's left child to the root's right + root.right = pivotLeft; + if (pivotLeft != null) { + pivotLeft.parent = root; + } + + replaceInParent(root, pivot); + + // move the root to the pivot's left + pivot.left = root; + root.parent = pivot; + + // fix heights + root.height = Math.max(left != null ? left.height : 0, + pivotLeft != null ? pivotLeft.height : 0) + 1; + pivot.height = Math.max(root.height, + pivotRight != null ? pivotRight.height : 0) + 1; + } + + /** + * Rotates the subtree so that its root's left child is the new root. + */ + private void rotateRight(Node<K, V> root) { + Node<K, V> pivot = root.left; + Node<K, V> right = root.right; + Node<K, V> pivotLeft = pivot.left; + Node<K, V> pivotRight = pivot.right; + + // move the pivot's right child to the root's left + root.left = pivotRight; + if (pivotRight != null) { + pivotRight.parent = root; + } + + replaceInParent(root, pivot); + + // move the root to the pivot's right + pivot.right = root; + root.parent = pivot; + + // fixup heights + root.height = Math.max(right != null ? right.height : 0, + pivotRight != null ? pivotRight.height : 0) + 1; + pivot.height = Math.max(root.height, + pivotLeft != null ? pivotLeft.height : 0) + 1; + } + + private EntrySet entrySet; + private KeySet keySet; + + @Override public Set<Entry<K, V>> entrySet() { + EntrySet result = entrySet; + return result != null ? result : (entrySet = new EntrySet()); + } + + @Override public Set<K> keySet() { + KeySet result = keySet; + return result != null ? result : (keySet = new KeySet()); + } + + static final class Node<K, V> implements Entry<K, V> { + Node<K, V> parent; + Node<K, V> left; + Node<K, V> right; + Node<K, V> next; + Node<K, V> prev; + final K key; + final int hash; + V value; + int height; + + /** Create the header entry */ + Node() { + key = null; + hash = -1; + next = prev = this; + } + + /** Create a regular entry */ + Node(Node<K, V> parent, K key, int hash, Node<K, V> next, Node<K, V> prev) { + this.parent = parent; + this.key = key; + this.hash = hash; + this.height = 1; + this.next = next; + this.prev = prev; + prev.next = this; + next.prev = this; + } + + public K getKey() { + return key; + } + + public V getValue() { + return value; + } + + public V setValue(V value) { + V oldValue = this.value; + this.value = value; + return oldValue; + } + + @SuppressWarnings("rawtypes") + @Override public boolean equals(Object o) { + if (o instanceof Entry) { + Entry other = (Entry) o; + return (key == null ? other.getKey() == null : key.equals(other.getKey())) + && (value == null ? other.getValue() == null : value.equals(other.getValue())); + } + return false; + } + + @Override public int hashCode() { + return (key == null ? 0 : key.hashCode()) + ^ (value == null ? 0 : value.hashCode()); + } + + @Override public String toString() { + return key + "=" + value; + } + + /** + * Returns the first node in this subtree. + */ + public Node<K, V> first() { + Node<K, V> node = this; + Node<K, V> child = node.left; + while (child != null) { + node = child; + child = node.left; + } + return node; + } + + /** + * Returns the last node in this subtree. + */ + public Node<K, V> last() { + Node<K, V> node = this; + Node<K, V> child = node.right; + while (child != null) { + node = child; + child = node.right; + } + return node; + } + } + + private void doubleCapacity() { + table = doubleCapacity(table); + threshold = (table.length / 2) + (table.length / 4); // 3/4 capacity + } + + /** + * Returns a new array containing the same nodes as {@code oldTable}, but with + * twice as many trees, each of (approximately) half the previous size. + */ + static <K, V> Node<K, V>[] doubleCapacity(Node<K, V>[] oldTable) { + // TODO: don't do anything if we're already at MAX_CAPACITY + int oldCapacity = oldTable.length; + @SuppressWarnings("unchecked") // Arrays and generics don't get along. + Node<K, V>[] newTable = new Node[oldCapacity * 2]; + AvlIterator<K, V> iterator = new AvlIterator<K, V>(); + AvlBuilder<K, V> leftBuilder = new AvlBuilder<K, V>(); + AvlBuilder<K, V> rightBuilder = new AvlBuilder<K, V>(); + + // Split each tree into two trees. + for (int i = 0; i < oldCapacity; i++) { + Node<K, V> root = oldTable[i]; + if (root == null) { + continue; + } + + // Compute the sizes of the left and right trees. + iterator.reset(root); + int leftSize = 0; + int rightSize = 0; + for (Node<K, V> node; (node = iterator.next()) != null; ) { + if ((node.hash & oldCapacity) == 0) { + leftSize++; + } else { + rightSize++; + } + } + + // Split the tree into two. + leftBuilder.reset(leftSize); + rightBuilder.reset(rightSize); + iterator.reset(root); + for (Node<K, V> node; (node = iterator.next()) != null; ) { + if ((node.hash & oldCapacity) == 0) { + leftBuilder.add(node); + } else { + rightBuilder.add(node); + } + } + + // Populate the enlarged array with these new roots. + newTable[i] = leftSize > 0 ? leftBuilder.root() : null; + newTable[i + oldCapacity] = rightSize > 0 ? rightBuilder.root() : null; + } + return newTable; + } + + /** + * Walks an AVL tree in iteration order. Once a node has been returned, its + * left, right and parent links are <strong>no longer used</strong>. For this + * reason it is safe to transform these links as you walk a tree. + * + * <p><strong>Warning:</strong> this iterator is destructive. It clears the + * parent node of all nodes in the tree. It is an error to make a partial + * iteration of a tree. + */ + static class AvlIterator<K, V> { + /** This stack is a singly linked list, linked by the 'parent' field. */ + private Node<K, V> stackTop; + + void reset(Node<K, V> root) { + Node<K, V> stackTop = null; + for (Node<K, V> n = root; n != null; n = n.left) { + n.parent = stackTop; + stackTop = n; // Stack push. + } + this.stackTop = stackTop; + } + + public Node<K, V> next() { + Node<K, V> stackTop = this.stackTop; + if (stackTop == null) { + return null; + } + Node<K, V> result = stackTop; + stackTop = result.parent; + result.parent = null; + for (Node<K, V> n = result.right; n != null; n = n.left) { + n.parent = stackTop; + stackTop = n; // Stack push. + } + this.stackTop = stackTop; + return result; + } + } + + /** + * Builds AVL trees of a predetermined size by accepting nodes of increasing + * value. To use: + * <ol> + * <li>Call {@link #reset} to initialize the target size <i>size</i>. + * <li>Call {@link #add} <i>size</i> times with increasing values. + * <li>Call {@link #root} to get the root of the balanced tree. + * </ol> + * + * <p>The returned tree will satisfy the AVL constraint: for every node + * <i>N</i>, the height of <i>N.left</i> and <i>N.right</i> is different by at + * most 1. It accomplishes this by omitting deepest-level leaf nodes when + * building trees whose size isn't a power of 2 minus 1. + * + * <p>Unlike rebuilding a tree from scratch, this approach requires no value + * comparisons. Using this class to create a tree of size <i>S</i> is + * {@code O(S)}. + */ + final static class AvlBuilder<K, V> { + /** This stack is a singly linked list, linked by the 'parent' field. */ + private Node<K, V> stack; + private int leavesToSkip; + private int leavesSkipped; + private int size; + + void reset(int targetSize) { + // compute the target tree size. This is a power of 2 minus one, like 15 or 31. + int treeCapacity = Integer.highestOneBit(targetSize) * 2 - 1; + leavesToSkip = treeCapacity - targetSize; + size = 0; + leavesSkipped = 0; + stack = null; + } + + void add(Node<K, V> node) { + node.left = node.parent = node.right = null; + node.height = 1; + + // Skip a leaf if necessary. + if (leavesToSkip > 0 && (size & 1) == 0) { + size++; + leavesToSkip--; + leavesSkipped++; + } + + node.parent = stack; + stack = node; // Stack push. + size++; + + // Skip a leaf if necessary. + if (leavesToSkip > 0 && (size & 1) == 0) { + size++; + leavesToSkip--; + leavesSkipped++; + } + + /* + * Combine 3 nodes into subtrees whenever the size is one less than a + * multiple of 4. For example we combine the nodes A, B, C into a + * 3-element tree with B as the root. + * + * Combine two subtrees and a spare single value whenever the size is one + * less than a multiple of 8. For example at 8 we may combine subtrees + * (A B C) and (E F G) with D as the root to form ((A B C) D (E F G)). + * + * Just as we combine single nodes when size nears a multiple of 4, and + * 3-element trees when size nears a multiple of 8, we combine subtrees of + * size (N-1) whenever the total size is 2N-1 whenever N is a power of 2. + */ + for (int scale = 4; (size & scale - 1) == scale - 1; scale *= 2) { + if (leavesSkipped == 0) { + // Pop right, center and left, then make center the top of the stack. + Node<K, V> right = stack; + Node<K, V> center = right.parent; + Node<K, V> left = center.parent; + center.parent = left.parent; + stack = center; + // Construct a tree. + center.left = left; + center.right = right; + center.height = right.height + 1; + left.parent = center; + right.parent = center; + } else if (leavesSkipped == 1) { + // Pop right and center, then make center the top of the stack. + Node<K, V> right = stack; + Node<K, V> center = right.parent; + stack = center; + // Construct a tree with no left child. + center.right = right; + center.height = right.height + 1; + right.parent = center; + leavesSkipped = 0; + } else if (leavesSkipped == 2) { + leavesSkipped = 0; + } + } + } + + Node<K, V> root() { + Node<K, V> stackTop = this.stack; + if (stackTop.parent != null) { + throw new IllegalStateException(); + } + return stackTop; + } + } + + private abstract class LinkedTreeMapIterator<T> implements Iterator<T> { + Node<K, V> next = header.next; + Node<K, V> lastReturned = null; + int expectedModCount = modCount; + + public final boolean hasNext() { + return next != header; + } + + final Node<K, V> nextNode() { + Node<K, V> e = next; + if (e == header) { + throw new NoSuchElementException(); + } + if (modCount != expectedModCount) { + throw new ConcurrentModificationException(); + } + next = e.next; + return lastReturned = e; + } + + public final void remove() { + if (lastReturned == null) { + throw new IllegalStateException(); + } + removeInternal(lastReturned, true); + lastReturned = null; + expectedModCount = modCount; + } + } + + final class EntrySet extends AbstractSet<Entry<K, V>> { + @Override public int size() { + return size; + } + + @Override public Iterator<Entry<K, V>> iterator() { + return new LinkedTreeMapIterator<Entry<K, V>>() { + public Entry<K, V> next() { + return nextNode(); + } + }; + } + + @Override public boolean contains(Object o) { + return o instanceof Entry && findByEntry((Entry<?, ?>) o) != null; + } + + @Override public boolean remove(Object o) { + if (!(o instanceof Entry)) { + return false; + } + + Node<K, V> node = findByEntry((Entry<?, ?>) o); + if (node == null) { + return false; + } + removeInternal(node, true); + return true; + } + + @Override public void clear() { + LinkedHashTreeMap.this.clear(); + } + } + + final class KeySet extends AbstractSet<K> { + @Override public int size() { + return size; + } + + @Override public Iterator<K> iterator() { + return new LinkedTreeMapIterator<K>() { + public K next() { + return nextNode().key; + } + }; + } + + @Override public boolean contains(Object o) { + return containsKey(o); + } + + @Override public boolean remove(Object key) { + return removeInternalByKey(key) != null; + } + + @Override public void clear() { + LinkedHashTreeMap.this.clear(); + } + } + + /** + * If somebody is unlucky enough to have to serialize one of these, serialize + * it as a LinkedHashMap so that they won't need Gson on the other side to + * deserialize it. Using serialization defeats our DoS defence, so most apps + * shouldn't use it. + */ + private Object writeReplace() throws ObjectStreamException { + return new LinkedHashMap<K, V>(this); + } +} |