Verhoeff Algorithm
The Verhoeff algorithm is a checksum formula for error detection developed by the Dutch mathematician Jacobus Verhoeff and was first published in 1969. It was the first decimal check digit algorithm which detects all singledigit errors, and all transposition errors involving two adjacent digits, which was at the time thought impossible with such a code. Goals Verhoeff had the goal of finding a decimal code—one where the check digit is a single decimal digit—which detected all singledigit errors and all transpositions of adjacent digits. At the time, supposed proofs of the nonexistence of these codes made base11 codes popular, for example in the ISBN check digit. His goals were also practical, and he based the evaluation of different codes on live data from the Dutch postal system, using a weighted points system for different kinds of error. The analysis broke the errors down into a number of categories: first, by how many digits are in error; for those with two digits in ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Checksum
A checksum is a smallsized block of data derived from another block of digital data for the purpose of detecting errors that may have been introduced during its transmission or storage. By themselves, checksums are often used to verify data integrity but are not relied upon to verify data authenticity. The procedure which generates this checksum is called a checksum function or checksum algorithm. Depending on its design goals, a good checksum algorithm usually outputs a significantly different value, even for small changes made to the input. This is especially true of cryptographic hash functions, which may be used to detect many data corruption errors and verify overall data integrity; if the computed checksum for the current data input matches the stored value of a previously computed checksum, there is a very high probability the data has not been accidentally altered or corrupted. Checksum functions are related to hash functions, fingerprints, randomization functi ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Error Detection
In information theory and coding theory with applications in computer science and telecommunication, error detection and correction (EDAC) or error control are techniques that enable reliable delivery of digital data over unreliable communication channels. Many communication channels are subject to channel noise, and thus errors may be introduced during transmission from the source to a receiver. Error detection techniques allow detecting such errors, while error correction enables reconstruction of the original data in many cases. Definitions ''Error detection'' is the detection of errors caused by noise or other impairments during transmission from the transmitter to the receiver. ''Error correction'' is the detection of errors and reconstruction of the original, errorfree data. History In classical antiquity, copyists of the Hebrew Bible were paid for their work according to the number of stichs (lines of verse). As the prose books of the Bible were hardly ever wr ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Jacobus Verhoeff
Jacobus "Koos" Verhoeff (20 February 1927 – 19 March 2018) was a Dutch mathematician, computer scientist, and artist. He is known for his work on error detection and correction, and on information retrieval. He has also held exhibitions of his mathematically inspired sculptures. He is best known for his checkdigit Verhoeff algorithm The Verhoeff algorithm is a checksum formula for error detection developed by the Dutch mathematician Jacobus Verhoeff and was first published in 1969. It was the first decimal check digit algorithm which detects all singledigit errors, and all ..., which is based on the dihedral group of order 10. His son, Tom Verhoeff, is a mathematician and computer scientist. Selected publications * * * * * * References 1927 births 2018 deaths 20thcentury Dutch mathematicians Dutch computer scientists Erasmus University Rotterdam faculty University of Amsterdam alumni Scientists from The Hague {{mathematicianstub ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Check Digit
A check digit is a form of redundancy check used for error detection on identification numbers, such as bank account numbers, which are used in an application where they will at least sometimes be input manually. It is analogous to a binary parity bit used to check for errors in computergenerated data. It consists of one or more digits (or letters) computed by an algorithm from the other digits (or letters) in the sequence input. With a check digit, one can detect simple errors in the input of a series of characters (usually digits) such as a single mistyped digit or some permutations of two successive digits. Design Check digit algorithms are generally designed to capture ''human'' transcription errors. In order of complexity, these include the following: * letter/digit errors, such as l → 1 or O → 0 * singledigit errors, such as 1 → 2 * transposition errors, such as 12 → 21 * twin errors, such as 11 → 22 * jump transpositions errors, such as 132 → 231 * jump t ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

International Standard Book Number
The International Standard Book Number (ISBN) is a numeric commercial book identifier that is intended to be unique. Publishers purchase ISBNs from an affiliate of the International ISBN Agency. An ISBN is assigned to each separate edition and variation (except reprintings) of a publication. For example, an ebook, a paperback and a hardcover edition of the same book will each have a different ISBN. The ISBN is ten digits long if assigned before 2007, and thirteen digits long if assigned on or after 1 January 2007. The method of assigning an ISBN is nationspecific and varies between countries, often depending on how large the publishing industry is within a country. The initial ISBN identification format was devised in 1967, based upon the 9digit Standard Book Numbering (SBN) created in 1966. The 10digit ISBN format was developed by the International Organization for Standardization (ISO) and was published in 1970 as international standard ISO 2108 (the 9digit SBN ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Dihedral Group
In mathematics, a dihedral group is the group of symmetries of a regular polygon, which includes rotations and reflections. Dihedral groups are among the simplest examples of finite groups, and they play an important role in group theory, geometry, and chemistry. The notation for the dihedral group differs in geometry and abstract algebra. In geometry, or refers to the symmetries of the gon, a group of order . In abstract algebra, refers to this same dihedral group. This article uses the geometric convention, . Definition Elements A regular polygon with n sides has 2n different symmetries: n rotational symmetries and n reflection symmetries. Usually, we take n \ge 3 here. The associated rotations and reflections make up the dihedral group \mathrm_n. If n is odd, each axis of symmetry connects the midpoint of one side to the opposite vertex. If n is even, there are n/2 axes of symmetry connecting the midpoints of opposite sides and n/2 axes of symmetry connecting ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Lookup Table
In computer science, a lookup table (LUT) is an array that replaces runtime computation with a simpler array indexing operation. The process is termed as "direct addressing" and LUTs differ from hash tables in a way that, to retrieve a value v with key k, a hash table would store the value v in the slot h(k) where h is a hash function i.e. k is used to compute the slot, while in the case of LUT, the value v is stored in slot k, thus directly addressable. The savings in processing time can be significant, because retrieving a value from memory is often faster than carrying out an "expensive" computation or input/output operation. The tables may be precalculated and stored in static program storage, calculated (or "prefetched") as part of a program's initialization phase ( memoization), or even stored in hardware in applicationspecific platforms. Lookup tables are also used extensively to validate input values by matching against a list of valid (or invalid) items in an array ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Damm Algorithm
In error detection, the Damm algorithm is a check digit algorithm that detects all singledigit errors and all adjacent transposition errors. It was presented by H. Michael Damm in 2004. Strengths and weaknesses Strengths The Damm algorithm is similar to the Verhoeff algorithm. It too will detect ''all'' occurrences of the two most frequently appearing types of transcription errors, namely altering one single digit, and transposing two adjacent digits (including the transposition of the trailing check digit and the preceding digit). But the Damm algorithm has the benefit that it makes do without the dedicatedly constructed permutations and its position specific powers being inherent in the Verhoeff scheme. Furthermore, a table of inverses can be dispensed with provided all main diagonal entries of the operation table are zero. The Damm algorithm does not suffer from exceeding the number of 10 possible values, resulting in the need for using a nondigit character (as the X ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Cayley Table
Named after the 19th century British mathematician Arthur Cayley, a Cayley table describes the structure of a finite group by arranging all the possible products of all the group's elements in a square table reminiscent of an addition or multiplication table. Many properties of a groupsuch as whether or not it is abelian, which elements are inverses of which elements, and the size and contents of the group's centercan be discovered from its Cayley table. A simple example of a Cayley table is the one for the group under ordinary multiplication: History Cayley tables were first presented in Cayley's 1854 paper, "On The Theory of Groups, as depending on the symbolic equation ''θ'' ''n'' = 1". In that paper they were referred to simply as tables, and were merely illustrativethey came to be known as Cayley tables later on, in honour of their creator. Structure and layout Because many Cayley tables describe groups that are not abelian, the product ''ab'' with respect to ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Commutative
In mathematics, a binary operation is commutative if changing the order of the operands does not change the result. It is a fundamental property of many binary operations, and many mathematical proofs depend on it. Most familiar as the name of the property that says something like or , the property can also be used in more advanced settings. The name is needed because there are operations, such as division and subtraction, that do not have it (for example, ); such operations are ''not'' commutative, and so are referred to as ''noncommutative operations''. The idea that simple operations, such as the multiplication and addition of numbers, are commutative was for many years implicitly assumed. Thus, this property was not named until the 19th century, when mathematics started to become formalized. A similar property exists for binary relations; a binary relation is said to be symmetric if the relation applies regardless of the order of its operands; for example, equality is sy ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Permutation
In mathematics, a permutation of a set is, loosely speaking, an arrangement of its members into a sequence or linear order, or if the set is already ordered, a rearrangement of its elements. The word "permutation" also refers to the act or process of changing the linear order of an ordered set. Permutations differ from combinations, which are selections of some members of a set regardless of order. For example, written as tuples, there are six permutations of the set , namely (1, 2, 3), (1, 3, 2), (2, 1, 3), (2, 3, 1), (3, 1, 2), and (3, 2, 1). These are all the possible orderings of this threeelement set. Anagrams of words whose letters are different are also permutations: the letters are already ordered in the original word, and the anagram is a reordering of the letters. The study of permutations of finite sets is an important topic in the fields of combinatorics and group theory. Permutations are used in almost every branch of mathematics, and in many other fields of s ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Modular Arithmetic
In mathematics, modular arithmetic is a system of arithmetic for integers, where numbers "wrap around" when reaching a certain value, called the modulus. The modern approach to modular arithmetic was developed by Carl Friedrich Gauss in his book '' Disquisitiones Arithmeticae'', published in 1801. A familiar use of modular arithmetic is in the 12hour clock, in which the day is divided into two 12hour periods. If the time is 7:00 now, then 8 hours later it will be 3:00. Simple addition would result in , but clocks "wrap around" every 12 hours. Because the hour number starts over at zero when it reaches 12, this is arithmetic ''modulo'' 12. In terms of the definition below, 15 is ''congruent'' to 3 modulo 12, so "15:00" on a 24hour clock is displayed "3:00" on a 12hour clock. Congruence Given an integer , called a modulus, two integers and are said to be congruent modulo , if is a divisor of their difference (that is, if there is an integer such that ). Congruence modu ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 