In
computational complexity theory
In theoretical computer science and mathematics, computational complexity theory focuses on classifying computational problems according to their resource usage, and relating these classes to each other. A computational problem is a task solved ...
, the element distinctness problem or element uniqueness problem is the problem of determining whether all the elements of a list are distinct.
It is a well studied problem in many different models of computation. The problem may be solved by
sorting the list and then checking if there are any consecutive equal elements; it may also be solved in linear expected time by a
randomized algorithm
A randomized algorithm is an algorithm that employs a degree of randomness as part of its logic or procedure. The algorithm typically uses uniformly random bits as an auxiliary input to guide its behavior, in the hope of achieving good performa ...
that inserts each item into a
hash table
In computing, a hash table, also known as hash map, is a data structure that implements an associative array or dictionary. It is an abstract data type that maps keys to values. A hash table uses a hash function to compute an ''index'', ...
and compares only those elements that are placed in the same hash table cell.
Several lower bounds in computational complexity are proved by reducing the element distinctness problem to the problem in question, i.e., by demonstrating that the solution of the element uniqueness problem may be quickly found after solving the problem in question.
Decision tree complexity
The number of comparisons needed to solve the problem of size
, in a comparison-based model of computation such as a
decision tree
A decision tree is a decision support tool that uses a tree-like model of decisions and their possible consequences, including chance event outcomes, resource costs, and utility. It is one way to display an algorithm that only contains co ...
or
algebraic decision tree, is
. Here,
invokes
big theta notation, meaning that the problem can be solved in a number of comparisons proportional to
(a
linearithmic function) and that all solutions require this many comparisons. In these models of computation, the input numbers may not be used to index the computer's memory (as in the hash table solution) but may only be accessed by computing and comparing simple algebraic functions of their values. For these models, an algorithm based on
comparison sort
A comparison sort is a type of sorting algorithm that only reads the list elements through a single abstract comparison operation (often a "less than or equal to" operator or a three-way comparison) that determines which of two elements should oc ...
solves the problem within a constant factor of the best possible number of comparisons. The same lower bound applies as well to the
expected number of comparisons in the
randomized
In common usage, randomness is the apparent or actual lack of pattern or predictability in events. A random sequence of events, symbols or steps often has no order and does not follow an intelligible pattern or combination. Individual rand ...
algebraic decision tree model.
Quantum complexity
Quantum algorithm
In quantum computing, a quantum algorithm is an algorithm which runs on a realistic model of quantum computation, the most commonly used model being the quantum circuit model of computation. A classical (or non-quantum) algorithm is a finite seq ...
s can solve this problem faster, in
queries. The optimal algorithm is by
Andris Ambainis
Andris Ambainis (born 18 January 1975) is a Latvian computer scientist active in the fields of quantum information theory and quantum computing.
Education and career
Ambainis has held past positions at the Institute for Advanced Study at Princet ...
.
Yaoyun Shi first proved a tight lower bound when the size of the range is sufficiently large. Ambainis and Kutin independently (and via different proofs) extended his work to obtain the lower bound for all functions.
Generalization: Finding repeated elements
Elements that occur more than
times in a multiset of size
may be found by a comparison-based algorithm, the
Misra–Gries heavy hitters algorithm Misra and Gries defined the ''heavy-hitters problem''
(though they did not introduce the term ''heavy-hitters'') and described the first algorithm
for it in the paper ''Finding repeated elements''. Their algorithm
extends the Boyer-Moore majority ...
, in time
. The element distinctness problem is a special case of this problem where
. This time is optimal under the
decision tree model of computation.
[.]
See also
*
Collision problem
References
{{reflist
Polynomial-time problems