Expertise finding
   HOME

TheInfoList



OR:

Expertise finding is the use of tools for finding and assessing individual expertise. In the recruitment industry, expertise finding is the problem of searching for employable candidates with certain required
skill A skill is the learned ability to act with determined results with good execution often within a given amount of time, energy, or both. Skills can often be divided into domain-general and domain-specific skills. For example, in the domain of w ...
s set. In other words, it is the challenge of linking humans to expertise areas, and as such is a sub-problem of expertise retrieval (the other problem being expertise profiling).


Importance of expertise

It can be argued that human expertise is more valuable than capital, means of production or intellectual property. Contrary to expertise, all other aspects of capitalism are now relatively generic: access to capital is global, as is access to means of production for many areas of manufacturing.
Intellectual property Intellectual property (IP) is a category of property that includes intangible creations of the human intellect. There are many types of intellectual property, and some countries recognize more than others. The best-known types are patents, cop ...
can be similarly licensed. Furthermore, expertise finding is also a key aspect of
institutional memory Institutional memory is a collective set of facts, concepts, experiences and knowledge held by a group of people. Concept Institutional memory has been defined as "the stored knowledge within the organization." Within any organization, tools ...
, as without its experts an institution is effectively decapitated. However, finding and "licensing" expertise, the key to the effective use of these resources, remain much harder, starting with the very first step: finding expertise that you can trust. Until very recently, finding expertise required a mix of individual, social and collaborative practices, a haphazard process at best. Mostly, it involved contacting individuals one trusts and asking them for referrals, while hoping that one's judgment about those individuals is justified and that their answers are thoughtful. In the last fifteen years, a class of
knowledge management Knowledge management (KM) is the collection of methods relating to creating, sharing, using and managing the knowledge and information of an organization. It refers to a multidisciplinary approach to achieve organisational objectives by making ...
software has emerged to facilitate and improve the quality of expertise finding, termed "expertise locating systems". These software range from social networking systems to
knowledge base A knowledge base (KB) is a technology used to store complex structured and unstructured information used by a computer system. The initial use of the term was in connection with expert systems, which were the first knowledge-based systems. ...
s. Some software, like those in the social networking realm, rely on users to connect each other, thus using social filtering to act as "recommender systems". At the other end of the spectrum are specialized
knowledge base A knowledge base (KB) is a technology used to store complex structured and unstructured information used by a computer system. The initial use of the term was in connection with expert systems, which were the first knowledge-based systems. ...
s that rely on experts to populate a specialized type of
database In computing, a database is an organized collection of data stored and accessed electronically. Small databases can be stored on a file system, while large databases are hosted on computer clusters or cloud storage. The design of databases s ...
with their self-determined areas of expertise and contributions, and do not rely on user recommendations. Hybrids that feature expert-populated content in conjunction with user recommendations also exist, and are arguably more valuable for doing so. Still other expertise knowledge bases rely strictly on external manifestations of expertise, herein termed "gated objects", e.g.,
citation impact Citation impact is a measure of how many times an academic journal article or book or author is cited by other articles, books or authors. Citation counts are interpreted as measures of the impact or influence of academic work and have given ris ...
s for scientific papers or data mining approaches wherein many of the work products of an expert are collated. Such systems are more likely to be free of user-introduced biases (e.g.
ResearchScorecard
), though the use of computational methods can introduce other biases. There are also hybrid approaches which use user-generated data (e.g., member profiles), community-based signals (e.g., recommendations and skill endorsements), and personalized signals (e.g., social connection between searcher and results). Examples of the systems outlined above are listed in Table 1. Table 1: A classification of expertise location systems


Technical problems

A number of interesting problems follow from the use of expertise finding systems: * The matching of questions from non-expert to the database of existing expertise is inherently difficult, especially when the database does not store the requisite expertise. This problem grows even more acute with increasing ignorance on the part of the non-expert due to typical search problems involving use of keywords to search
unstructured data Unstructured data (or unstructured information) is information that either does not have a pre-defined data model or is not organized in a pre-defined manner. Unstructured information is typically text-heavy, but may contain data such as dates, num ...
that are not semantically normalized, as well as variability in how well an expert has set up their descriptive content pages. Improved question matching is one reason why third-party semantically normalized systems such as ResearchScorecard and BiomedExperts should be able to provide better answers to queries from non-expert users. * Avoiding expert-fatigue due to too many questions/requests from users of the system (ref. 1). * Finding ways to avoid "gaming" of the system to reap unjustified expertise credibility. * Infer expertise on implicit skills. Since users typically do not declare all of the skills they have, it is important to infer their implicit skills that are highly related their explicit ones. The inference step can significantly improve recall in expertise finding.


Expertise ranking

Means of classifying and ranking expertise (and therefore experts) become essential if the number of experts returned by a query is greater than a handful. This raises the following social problems associated with such systems: * How can expertise be assessed objectively? Is that even possible? * What are the consequences of relying on unstructured social assessments of expertise, such as user recommendations? * How does one distinguish ''authoritativeness'' as a proxy metric of expertise from simple ''popularity'', which is often a function of one's ability to express oneself coupled with a good social sense? * What are the potential consequences of the social or professional stigma associated with the use of an authority ranking, such as used in Technorati and ResearchScorecard)? * How to make expertise ranking personalized to each individual searcher? This is particularly important for recruiting purpose since given the same skills, recruiters from different companies, industries, locations might have different preferences for candidates and their varying areas of expertise.


Sources of data for assessing expertise

Many types of data sources have been used to infer expertise. They can be broadly categorized based on whether they measure "raw" contributions provided by the expert, or whether some sort of filter is applied to these contributions. Unfiltered data sources that have been used to assess expertise, in no particular ranking order: * self-reported expertise on networking platforms * expertise sharing through platforms * user recommendations * help desk tickets: what the problem was and who fixed it * e-mail traffic between users * documents, whether private or on the web, particularly publications * user-maintained web pages * reports (technical, marketing, etc.) Filtered data sources, that is, contributions that require approval by third parties (grant committees, referees, patent office, etc.) are particularly valuable for measuring expertise in a way that minimizes biases that follow from popularity or other social factors: *
patent A patent is a type of intellectual property that gives its owner the legal right to exclude others from making, using, or selling an invention for a limited period of time in exchange for publishing an enabling disclosure of the invention."A ...
s, particularly if issued * scientific publications * issued grants (failed grant proposals are rarely known beyond the authors) *
clinical trial Clinical trials are prospective biomedical or behavioral research studies on human participants designed to answer specific questions about biomedical or behavioral interventions, including new treatments (such as novel vaccines, drugs, diet ...
s * product launches * pharmaceutical drugs


Approaches for creating expertise content

* Manual, either by experts themselves (e.g., Skillhive) or by a curator (Expertise Finder) * Automated, e.g., using
software agent In computer science, a software agent or software AI is a computer program that acts for a user or other program in a relationship of agency, which derives from the Latin ''agere'' (to do): an agreement to act on one's behalf. Such "action on beha ...
s (e.g., MIT's ExpertFinder) or a combination of agents and human curation (e.g., ResearchScorecard ) * In industrial expertise search engines (e.g., LinkedIn), there are many signals coming into the ranking functions, such as, user-generated content (e.g., profiles), community-generated content (e.g., recommendations and skills endorsements) and personalized signals (e.g., social connections). Moreover, user queries might contain many other aspects rather required expertise, such as, locations, industries or companies. Thus, traditional information retrieval features like text matching are also important.
Learning to rank Learning to rank. Slides from Tie-Yan Liu's talk at WWW 2009 conference aravailable online or machine-learned ranking (MLR) is the application of machine learning, typically supervised, semi-supervised or reinforcement learning, in the construc ...
is typically used to combine all of these signals together into a ranking function


Collaborator discovery

In academia, a related problem is collaborator discovery, where the goal is to suggest suitable collaborators to a researcher. While expertise finding is an asynchronous problem (employer looking for employee), collaborator discovery can be distinguished from expertise finding by helping establishing more symmetric relationships (collaborations). Also, while in expertise finding the task often can be clearly characterized, this is not the case in academic research, where future goals are more fuzzy.


References

{{Reflist


Further reading

# Ackerman, Mark and McDonald, David (1998) "Just Talk to Me: A Field Study of Expertise Location" ''Proceedings of the 1998 ACM Conference on Computer Supported Cooperative Work''. # Hughes, Gareth and Crowder, Richard (2003) "Experiences in designing highly adaptable expertise finder systems" ''Proceedings of the DETC Conference 2003''. # Maybury, M., D'Amore, R., House, D. (2002). "Awareness of organizational expertise." ''International Journal of Human-Computer Interaction'' 14(2): 199-217. # Maybury, M., D'Amore, R., House, D. (2000). Automating Expert Finding. ''International Journal of Technology Research Management.'' 43(6): 12-15. # Maybury, M., D'Amore, R, and House, D. December (2001). Expert Finding for Collaborative Virtual Environments. ''Communications of the ACM 14''(12): 55-56. In Ragusa, J. and Bochenek, G. (eds). Special Section on Collaboration Virtual Design Environments. # Maybury, M., D'Amore, R. and House, D. (2002). Automated Discovery and Mapping of Expertise. In Ackerman, M., Cohen, A., Pipek, V. and Wulf, V. (eds.). ''Beyond Knowledge Management: Sharing Expertise.'' Cambridge: MIT Press. # Mattox, D., M. Maybury, ''et al.'' (1999). "Enterprise expert and knowledge discovery". ''Proceedings of the 8th International Conference on Human-Computer Interactions (HCI International 99)'', Munich, Germany. # Tang, J., Zhang J., Yao L., Li J., Zhang L. and Su Z.(2008) "ArnetMiner: extraction and mining of academic social networks" ''Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining''. # Viavacqua, A. (1999). "Agents for expertise location". ''Proceedings of the 1999 AAAI Spring Symposium on Intelligent Agents in Cyberspace'', Stanford, CA. Evaluation methods Metrics Analysis Impact assessment Knowledge sharing Library science Information retrieval genres Science studies