Jump to content

Bucket queue

This is a good article. Click here for more information.
From Wikipedia, the free encyclopedia

This is the current revision of this page, as edited by David Eppstein (talk | contribs) at 00:35, 13 June 2024 (Mauricio Resende). The present address (URL) is a permanent link to this version.

(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

Bucket queue
An array of buckets, suitable for priorities in the range from 1 to 6. The minimum-priority element can be found in the leftmost non-empty bucket.
Typepriority queue
Invented1969
Invented byRobert Dial
Time complexity in big O notation
Operation Average Worst case
Insert O(1) O(1)
Find-min O(#priorities) O(#priorities)
Delete-min O(#priorities) O(#priorities)
Decrease-key O(1) O(1)
Space complexity

A bucket queue is a data structure that implements the priority queue abstract data type: it maintains a dynamic collection of elements with numerical priorities and allows quick access to the element with minimum (or maximum) priority. In the bucket queue, the priorities must be integers, and it is particularly suited to applications in which the priorities have a small range.[1] A bucket queue has the form of an array of buckets: an array data structure, indexed by the priorities, whose cells contain collections of items with the same priority as each other. With this data structure, insertion of elements and changes of their priority take constant time. Searching for and removing the minimum-priority element takes time proportional to the number of buckets or, by maintaining a pointer to the most recently found bucket, in time proportional to the difference in priorities between successive operations.

The bucket queue is the priority-queue analogue of pigeonhole sort (also called bucket sort), a sorting algorithm that places elements into buckets indexed by their priorities and then concatenates the buckets. Using a bucket queue as the priority queue in a selection sort gives a form of the pigeonhole sort algorithm.[2] Bucket queues are also called bucket priority queues[3] or bounded-height priority queues.[1] When used for quantized approximations to real number priorities, they are also called untidy priority queues[4] or pseudo priority queues.[5] They are closely related to the calendar queue, a structure that uses a similar array of buckets for exact prioritization by real numbers.

Applications of the bucket queue include computation of the degeneracy of a graph, fast algorithms for shortest paths and widest paths for graphs with weights that are small integers or are already sorted, and greedy approximation algorithms for the set cover problem. The quantized version of the structure has also been applied to scheduling[2] and to marching cubes in computer graphics.[4] The first use of the bucket queue[6] was in a shortest path algorithm by Dial (1969).[7]

Operation

[edit]

Basic data structure

[edit]

A bucket queue can handle elements with integer priorities in the range from 0 or 1 up to some known bound C, and operations that insert elements, change the priority of elements, or extract (find and remove) the element that has the minimum (or maximum) priority. It consists of an array A of container data structures; in most sources these containers are doubly linked lists but they could alternatively be dynamic arrays[3] or dynamic sets. The container in the pth array cell A[p] stores the collection of elements whose priority is p.

A bucket queue can handle the following operations:

  • To insert an element x with priority p, add x to the container at A[p].
  • To change the priority of an element, remove it from the container for its old priority and re-insert it into the container for its new priority.
  • To extract an element with the minimum or maximum priority, perform a sequential search in the array to find the first or last non-empty container, respectively, choose an arbitrary element from this container, and remove it from the container.

In this way, insertions and priority changes take constant time, and extracting the minimum or maximum priority element takes time O(C).[1][6][8]

Optimizations

[edit]

As an optimization, the data structure can start each sequential search for a non-empty bucket at the most recently-found non-empty bucket instead of at the start of the array. This can be done in either of two different ways, lazy (delaying these sequential searches until they are necessary) or eager (doing the searches ahead of time). The choice of when to do the search affects which of the data structure operations is slowed by these searches. Dial's original version of the structure used a lazy search. This can be done by maintaining an index L that is a lower bound on the minimum priority of any element currently in the queue. When inserting a new element, L should be updated to the minimum of its old value and the new element's priority. When searching for the minimum priority element, the search can start at L instead of at zero, and after the search L should be left equal to the priority that was found in the search.[7][9] Alternatively, the eager version of this optimization keeps L updated so that it always points to the first non-empty bucket. When inserting a new element with a priority smaller than L, the data structure sets L to the new priority, and when removing the last element from a bucket with priority L, it performs a sequential search through larger indexes until finding a non-empty bucket and setting L to the priority of the resulting bucket.[1]

In either of these two variations, each sequential search takes time proportional to the difference between the old and new values of L. This could be significantly faster than the O(C) time bound for the searches in the un-optimized version of the data structure. In many applications of priority queues such as Dijkstra's algorithm, the minimum priorities form a monotonic sequence, allowing a monotone priority queue to be used. In these applications, for both the lazy and eager variations of the optimized structure, the sequential searches for non-empty buckets cover disjoint ranges of buckets. Because each bucket is in at most one of these ranges, their numbers of steps add to at most C. Therefore, in these applications, the total time for a sequence of n operations is O(n + C), rather than the slower O(nC) time bound that would result without this optimization.[9] A corresponding optimization can be applied in applications where a bucket queue is used to find elements of maximum priority, but in this case it should maintain an index that upper-bounds the maximum priority, and the sequential search for a non-empty bucket should proceed downwards from this upper bound.[10]

Another optimization (already given by Dial 1969) can be used to save space when the priorities are monotonic and, throughout the course of an algorithm, always fall within a range of r values rather than extending over the whole range from 0 to C. In this case, one can index the array by the priorities modulo r rather than by their actual values. The search for the minimum priority element should always begin at the previous minimum, to avoid priorities that are higher than the minimum but have lower moduli. In particular, this idea can be applied in Dijkstra's algorithm on graphs whose edge lengths are integers in the range from 1 to r.[8]

Because creating a new bucket queue involves initializing an array of empty buckets, this initialization step takes time proportional to the number of priorities. A variation of the bucket queue described by Donald B. Johnson in 1981 instead stores only the non-empty buckets in a linked list, sorted by their priorities, and uses an auxiliary search tree to quickly find the position in this linked list for any new buckets. It takes time O(log log C) to initialize this variant structure, constant time to find an element with minimum or maximum priority, and time O(log log D) to insert or delete an element, where D is the difference between the nearest smaller and larger priorities to the priority of the inserted or deleted element.[11]

Example

[edit]

For example, consider a bucket queue with four priorities, the numbers 0, 1, 2, and 3. It consists of an array whose four cells each contain a collection of elements, initially empty. For the purposes of this example, can be written as a bracketed sequence of four sets: . Consider a sequence of operations in which we insert two elements and with the same priority 1, insert a third element with priority 3, change the priority of to 3, and then perform two extractions of the minimum-priority element.

  • After inserting with priority 1, .
  • After inserting with priority 1, .
  • After inserting z with priority 3, .
  • Changing the priority of x from 1 to three involves removing it from and adding it to , after which .
  • Extracting the minimum-priority element, in the basic version of the bucket queue, searches from the start of to find its first non-empty element: is empty but , a non-empty set. It chooses an arbitrary element of this set (the only element, ) as the minimum-priority element. Removing from the structure leaves .
  • The second extract operation, in the basic version of the bucket queue, searches again from the start of the array: , , , non-empty. In the improved variants of the bucket queue, this search starts instead at the last position that was found to be non-empty, . In either case, is found to be the first non-empty set. One of its elements is chosen arbitrarily as the minimum-priority element; for example, might be chosen. This element is removed, leaving .

Applications

[edit]

Graph degeneracy

[edit]

A bucket queue can be used to maintain the vertices of an undirected graph, prioritized by their degrees, and repeatedly find and remove the vertex of minimum degree.[1] This greedy algorithm can be used to calculate the degeneracy of a given graph, equal to the largest degree of any vertex at the time of its removal. The algorithm takes linear time, with or without the optimization that maintains a lower bound on the minimum priority, because each vertex is found in time proportional to its degree and the sum of all vertex degrees is linear in the number of edges of the graph.[12]

Dial's algorithm for shortest paths

[edit]

In Dijkstra's algorithm for shortest paths in directed graphs with edge weights that are positive integers, the priorities are monotone,[13] and a monotone bucket queue can be used to obtain a time bound of O(m + dc), where m is the number of edges, d is the diameter of the network, and c is the maximum (integer) link cost.[9][14] This variant of Dijkstra's algorithm is also known as Dial's algorithm,[9] after Robert B. Dial, who published it in 1969.[7] The same idea also works, using a quantized bucket queue, for graphs with positive real edge weights when the ratio of the maximum to minimum weight is at most c.[15] In this quantized version of the algorithm, the vertices are processed out of order, compared to the result with a non-quantized priority queue, but the correct shortest paths are still found.[5] In these algorithms, the priorities will only span a range of width c + 1, so the modular optimization can be used to reduce the space to O(n + c).[8][14]

A variant of the same algorithm can be used for the widest path problem. In combination with methods for quickly partitioning non-integer edge weights into subsets that can be assigned integer priorities, it leads to near-linear-time solutions to the single-source single-destination version of the widest path problem.[16]

Greedy set cover

[edit]

The set cover problem has as its input a family of sets. The output should be a subfamily of these sets, with the same union as the original family, including as few sets as possible. It is NP-hard, but has a greedy approximation algorithm that achieves a logarithmic approximation ratio, essentially the best possible unless P = NP.[17] This approximation algorithm selects its subfamily by repeatedly choosing a set that covers the maximum possible number of remaining uncovered elements.[18] A standard exercise in algorithm design asks for an implementation of this algorithm that takes linear time in the input size, which is the sum of sizes of all the input sets.[19]

This may be solved using a bucket queue of sets in the input family, prioritized by the number of remaining elements that they cover. Each time that the greedy algorithm chooses a set as part of its output, the newly covered set elements should be subtracted from the priorities of the other sets that cover them; over the course of the algorithm the number of these changes of priorities is just the sum of sizes of the input sets. The priorities are monotonically decreasing integers, upper-bounded by the number of elements to be covered. Each choice of the greedy algorithm involves finding the set with the maximum priority, which can be done by scanning downwards through the buckets of the bucket queue, starting from the most recent previous maximum value. The total time is linear in the input size.[10]

Scheduling

[edit]

Bucket queues can be used to schedule tasks with deadlines, for instance in packet forwarding for internet data with quality of service guarantees. For this application, the deadlines should be quantized into discrete intervals, and tasks whose deadlines fall into the same interval are considered to be of equivalent priority.[2]

A variation of the quantized bucket queue data structure, the calendar queue, has been applied to scheduling of discrete-event simulations, where the elements in the queue are future events prioritized by the time within the simulation that the events should happen. In this application, the ordering of events is critical, so the priorities cannot be approximated. Therefore, the calendar queue performs searches for the minimum-priority element in a different way than a bucket queue: in the bucket queue, any element of the first non-empty bucket may be returned, but instead the calendar queue searches all the elements in that bucket to determine which of them has the smallest non-quantized priority. To keep these searches fast, this variation attempts to keep the number of buckets proportional to the number of elements, by adjusting the scale of quantization and rebuilding the data structure when it gets out of balance. Calendar queues may be slower than bucket queues in the worst case (when many elements all land in the same smallest bucket) but are fast when elements are uniformly distributed among buckets causing the average bucket size to be constant.[20][21]

Fast marching

[edit]

In applied mathematics and numerical methods for the solution of differential equations, untidy priority queues have been used to prioritize the steps of the fast marching method for solving boundary value problems of the Eikonal equation, used to model wave propagation. This method finds the times at which a moving boundary crosses a set of discrete points (such as the points of an integer grid) using a prioritization method resembling a continuous version of Dijkstra's algorithm, and its running time is dominated by its priority queue of these points. It can be sped up to linear time by rounding the priorities used in this algorithm to integers, and using a bucket queue for these integers. As in Dijkstra's and Dial's algorithms, the priorities are monotone, so fast marching can use the monotone optimization of the bucket queue and its analysis. However, the discretization introduces some error into the resulting calculations.[4]

See also

[edit]
  • Soft heap, a different way of speeding up priority queues by using approximate priorities

References

[edit]
  1. ^ a b c d e Skiena, Steven S. (1998), The Algorithm Design Manual, Springer, p. 181, ISBN 9780387948607.
  2. ^ a b c Figueira, N. R. (1997), "A solution for the priority queue problem of deadline-ordered service disciplines", Proceedings of Sixth International Conference on Computer Communications and Networks, IEEE Computer Society Press, pp. 320–325, doi:10.1109/icccn.1997.623330, S2CID 5611516
  3. ^ a b Henzinger, Monika; Noe, Alexander; Schulz, Christian (2019), "Shared-memory exact minimum cuts", 2019 IEEE International Parallel and Distributed Processing Symposium, IPDPS 2019, Rio de Janeiro, Brazil, May 20-24, 2019, pp. 13–22, arXiv:1808.05458, doi:10.1109/IPDPS.2019.00013, S2CID 52019258
  4. ^ a b c Rasch, Christian; Satzger, Thomas (2009), "Remarks on the O(N) implementation of the fast marching method" (PDF), IMA Journal of Numerical Analysis, 29 (3): 806–813, doi:10.1093/imanum/drm028, MR 2520171
  5. ^ a b Robledo, Alicia; Guivant, José E. (2010), "Pseudo priority queues for real-time performance on dynamic programming processes applied to path planning" (PDF), in Wyeth, Gordon; Upcroft, Ben (eds.), Australasian Conference on Robotics and Automation
  6. ^ a b Edelkamp, Stefan; Schroedl, Stefan (2011), "3.1.1 Bucket Data Structures", Heuristic Search: Theory and Applications, Elsevier, pp. 90–92, ISBN 9780080919737. See also p. 157 for the history and naming of this structure.
  7. ^ a b c Dial, Robert B. (1969), "Algorithm 360: Shortest-path forest with topological ordering [H]", Communications of the ACM, 12 (11): 632–633, doi:10.1145/363269.363610, S2CID 6754003.
  8. ^ a b c Mehlhorn, Kurt; Sanders, Peter (2008), "10.5.1 Bucket Queues", Algorithms and Data Structures: The Basic Toolbox, Springer, p. 201, ISBN 9783540779773.
  9. ^ a b c d Bertsekas, Dimitri P. (1991), "Dial's algorithm", Linear Network Optimization: Algorithms And Codes, Cambridge, Massachusetts: MIT Press, pp. 72–75, ISBN 0-262-02334-2, MR 1201582
  10. ^ a b Lim, C. L.; Moffat, Alistair; Wirth, Anthony Ian (2014), "Lazy and eager approaches for the set cover problem", in Thomas, Bruce; Parry, Dave (eds.), Thirty-Seventh Australasian Computer Science Conference, ACSC 2014, Auckland, New Zealand, January 2014, CRPIT, vol. 147, Australian Computer Society, pp. 19–27. See in particular Section 2.4, "Priority Queue", p. 22.
  11. ^ Johnson, Donald B. (1981), "A priority queue in which initialization and queue operations take O(log log D) time", Mathematical Systems Theory, 15 (4): 295–309, doi:10.1007/BF01786986, MR 0683047, S2CID 35703411
  12. ^ Matula, David W.; Beck, L. L. (1983), "Smallest-last ordering and clustering and graph coloring algorithms", Journal of the ACM, 30 (3): 417–427, doi:10.1145/2402.322385, MR 0709826, S2CID 4417741.
  13. ^ Varghese, George (2005), Network Algorithmics: An Interdisciplinary Approach to Designing Fast Networked Devices, Morgan Kaufmann, pp. 78–80, ISBN 9780120884773.
  14. ^ a b Festa, Paola (2006), "Shortest path algorithms", in Resende, Mauricio G. C.; Pardalos, Panos M. (eds.), Handbook of Optimization in Telecommunications, Boston: Springer, pp. 185–210, doi:10.1007/978-0-387-30165-5_8; see in particular Section 8.3.3.6, "Dial's implementation", pp. 194–195.
  15. ^ Mehlhorn & Sanders (2008) (Exercise 10.11, p. 201) credit this idea to a 1978 paper of E. A. Dinic (Yefim Dinitz).
  16. ^ Gabow, Harold N.; Tarjan, Robert E. (1988), "Algorithms for two bottleneck optimization problems", Journal of Algorithms, 9 (3): 411–417, doi:10.1016/0196-6774(88)90031-4, MR 0955149
  17. ^ Dinur, Irit; Steurer, David (2014), "Analytical approach to parallel repetition", in Shmoys, David B. (ed.), Symposium on Theory of Computing, STOC 2014, New York, NY, USA, May 31 - June 03, 2014, ACM, pp. 624–633, arXiv:1305.1979, doi:10.1145/2591796.2591884, MR 3238990, S2CID 15252482
  18. ^ Johnson, David S. (1974), "Approximation algorithms for combinatorial problems", Journal of Computer and System Sciences, 9 (3): 256–278, doi:10.1016/S0022-0000(74)80044-9, MR 0449012
  19. ^ Cormen, Thomas H.; Leiserson, Charles E.; Rivest, Ronald L.; Stein, Clifford (2009) [1990], "Exercise 35.3-3", Introduction to Algorithms (3rd ed.), MIT Press and McGraw-Hill, p. 1122, ISBN 0-262-03384-4
  20. ^ Brown, R. (October 1988), "Calendar queues: a fast priority queue implementation for the simulation event set problem", Communications of the ACM, 31 (10): 1220–1227, doi:10.1145/63039.63045, S2CID 32086497
  21. ^ Erickson, K. Bruce; Ladner, Richard E.; LaMarca, Anthony (2000), "Optimizing static calendar queues", ACM Transactions on Modeling and Computer Simulation, 10 (3): 179–214, doi:10.1145/361026.361028