Vector space model or term vector model is an algebraic model for representing text documents (or more generally, items) as
vectors such that the distance between vectors represents the relevance between the documents. It is used in
information filtering,
information retrieval
Information retrieval (IR) in computing and information science is the task of identifying and retrieving information system resources that are relevant to an Information needs, information need. The information need can be specified in the form ...
,
index
Index (: indexes or indices) may refer to:
Arts, entertainment, and media Fictional entities
* Index (''A Certain Magical Index''), a character in the light novel series ''A Certain Magical Index''
* The Index, an item on the Halo Array in the ...
ing and relevancy rankings. Its first use was in the
SMART Information Retrieval System.
Definitions
In this section we consider a particular vector space model based on the
bag-of-words representation. Documents and queries are represented as vectors.
:
:
Each
dimension
In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coo ...
corresponds to a separate term. If a term occurs in the document, its value in the vector is non-zero. Several different ways of computing these values, also known as (term) weights, have been developed. One of the best known schemes is
tf-idf weighting (see the example below).
The definition of ''term'' depends on the application. Typically terms are single words,
keywords, or longer phrases. If words are chosen to be the terms, the dimensionality of the vector is the number of words in the vocabulary (the number of distinct words occurring in the
corpus
Corpus (plural ''corpora'') is Latin for "body". It may refer to:
Linguistics
* Text corpus, in linguistics, a large and structured set of texts
* Speech corpus, in linguistics, a large set of speech audio files
* Corpus linguistics, a branch of ...
).
Vector operations can be used to compare documents with queries.
Applications
Candidate documents from the corpus can be retrieved and ranked using a variety of methods.
Relevance
Relevance is the connection between topics that makes one useful for dealing with the other. Relevance is studied in many different fields, including cognitive science, logic, and library and information science. Epistemology studies it in gener ...
ranking
A ranking is a relationship between a set of items, often recorded in a list, such that, for any two items, the first is either "ranked higher than", "ranked lower than", or "ranked equal to" the second. In mathematics, this is known as a weak ...
s of documents in a keyword search can be calculated, using the assumptions of
document similarities theory, by comparing the deviation of angles between each document vector and the original query vector where the query is represented as a vector with same dimension as the vectors that represent the other documents.
In practice, it is easier to calculate the
cosine
In mathematics, sine and cosine are trigonometric functions of an angle. The sine and cosine of an acute angle are defined in the context of a right triangle: for the specified angle, its sine is the ratio of the length of the side opposite that ...
of the angle between the vectors, instead of the angle itself:
:
Where
is the intersection (i.e. the
dot product
In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
) of the document (d
2 in the figure to the right) and the query (q in the figure) vectors,
is the norm of vector d
2, and
is the norm of vector q. The
norm of a vector is calculated as such:
:
Using the cosine the similarity between document ''d
j'' and query ''q'' can be calculated as:
:
As all vectors under consideration by this model are element-wise nonnegative, a cosine value of zero means that the query and document vector are
orthogonal
In mathematics, orthogonality (mathematics), orthogonality is the generalization of the geometric notion of ''perpendicularity''. Although many authors use the two terms ''perpendicular'' and ''orthogonal'' interchangeably, the term ''perpendic ...
and have no match (i.e. the query term does not exist in the document being considered). See
cosine similarity for further information.
Term frequency-inverse document frequency weights
In the classic vector space model proposed by
Salton, Wong and Yang
G. Salton , A. Wong , C. S. Yang, A vector space model for automatic indexing
Communications of the ACM, v.18 n.11, p.613–620, Nov. 1975 the term-specific weights in the document vectors are products of local and global parameters. The model is known as term frequency-inverse document frequency model. The weight vector for document ''d'' is , where
:
and
* is term frequency of term ''t'' in document ''d'' (a local parameter)
* is inverse document frequency (a global parameter). is the total number of documents in the document set; is the number of documents containing the term ''t''.
Advantages
The vector space model has the following advantages over the Standard Boolean model:
#Allows ranking documents according to their possible relevance
#Allows retrieving items with a partial term overlap
Most of these advantages are a consequence of the difference in the density of the document collection representation between Boolean and term frequency-inverse document frequency approaches. When using Boolean weights, any document lies in a vertex in a n-dimensional hypercube
In geometry, a hypercube is an ''n''-dimensional analogue of a square ( ) and a cube ( ); the special case for is known as a ''tesseract''. It is a closed, compact, convex figure whose 1- skeleton consists of groups of opposite parallel l ...
. Therefore, the possible document representations are and the maximum Euclidean distance between pairs is . As documents are added to the document collection, the region defined by the hypercube's vertices become more populated and hence denser. Unlike Boolean, when a document is added using term frequency-inverse document frequency weights, the inverse document frequencies of the terms in the new document decrease while that of the remaining terms increase. In average, as documents are added, the region where documents lie expands regulating the density of the entire collection representation. This behavior models the original motivation of Salton and his colleagues that a document collection represented in a low density region could yield better retrieval results.
Limitations
The vector space model has the following limitations:
#Query terms are assumed to be independent, so phrases might not be represented well in the ranking
#Semantic sensitivity; documents with similar context but different term vocabulary won't be associated
Many of these difficulties can, however, be overcome by the integration of various tools, including mathematical techniques such as singular value decomposition
In linear algebra, the singular value decomposition (SVD) is a Matrix decomposition, factorization of a real number, real or complex number, complex matrix (mathematics), matrix into a rotation, followed by a rescaling followed by another rota ...
and lexical database
In digital lexicography, natural language processing, and digital humanities, a lexical resource is a language resource consisting of data regarding the lexemes of the lexicon of one or more languages e.g., in the form of a database.
Characteris ...
s such as WordNet
WordNet is a lexical database of semantic relations between words that links words into semantic relations including synonyms, hyponyms, and meronyms. The synonyms are grouped into ''synsets'' with short definitions and usage examples. It can thu ...
.
Models based on and extending the vector space model
Models based on and extending the vector space model include:
* Generalized vector space model
* Latent semantic analysis
Latent semantic analysis (LSA) is a technique in natural language processing, in particular distributional semantics, of analyzing relationships between a set of documents and the terms they contain by producing a set of concepts related to the d ...
* Term
* Rocchio Classification
* Random indexing
* Search Engine Optimization
Search engine optimization (SEO) is the process of improving the quality and quantity of Web traffic, website traffic to a website or a web page from web search engine, search engines. SEO targets unpaid search traffic (usually referred to as ...
Software that implements the vector space model
The following software packages may be of interest to those wishing to experiment with vector models and implement search services based upon them.
Free open source software
* Apache Lucene
Apache Lucene is a free and open-source search engine software library, originally written in Java by Doug Cutting. It is supported by the Apache Software Foundation and is released under the Apache Software License. Lucene is widely used as a ...
. Apache Lucene is a high-performance, open source, full-featured text search engine library written entirely in Java.
* OpenSearch (software) and Solr: the two most well-known search engine programs (many smaller exist) based on Lucene.
* Gensim is a Python+ NumPy framework for Vector Space modelling. It contains incremental (memory-efficient) algorithms for term frequency-inverse document frequency, latent semantic indexing, random projections and latent Dirichlet allocation
In natural language processing, latent Dirichlet allocation (LDA) is a Bayesian network (and, therefore, a generative statistical model) for modeling automatically extracted topics in textual corpora. The LDA is an example of a Bayesian topic ...
.
* Weka. Weka is a popular data mining package for Java including WordVectors and Bag Of Words models.
* Word2vec
Word2vec is a technique in natural language processing (NLP) for obtaining vector representations of words. These vectors capture information about the meaning of the word based on the surrounding words. The word2vec algorithm estimates these rep ...
. Word2vec uses vector spaces for word embeddings.
Further reading
* G. Salton (1962),
Some experiments in the generation of word and document associations
''Proceeding AFIPS '62 (Fall) Proceedings of the December 4–6, 1962, fall joint computer conference'', pages 234–250. ''(Early paper of Salton using the term-document matrix formalization)''
* G. Salton, A. Wong, and C. S. Yang (1975),
A Vector Space Model for Automatic Indexing
''Communications of the ACM'', vol. 18, nr. 11, pages 613–620. ''(Article in which a vector space model was presented)''
* David Dubin (2004)
The Most Influential Paper Gerard Salton Never Wrote
''(Explains the history of the Vector Space Model and the non-existence of a frequently cited publication)''
* ttp://nlp.stanford.edu/IR-book/html/htmledition/vector-space-classification-1.html Relationship of vector space search to the "k-Nearest Neighbor" search
See also
References
{{reflist