Scientometrics and Bibliometrics September 24, 2014 - News & events
What are Scientometrics and Bibliometrics?
Scientometrics and bibliometrics are methodological approaches in which the scientific literature itself becomes the subject of analysis. In a sense, they could be considered a science of science. Scientometrics researchers often attempt to measure the evolution of a scientific domain, the impact of scholarly publications, the patterns of authorship, and the process of scientific knowledge production. Scientometrics and bibliometrics often involve the monitoring of research, the assessment of the scientific contribution of authors, journals, or specific works, as well as the analysis of the dissemination process of scientific knowledge. Researchers in such approaches have developed methodological principles on ways to gather information produced by the activity of researchers’ communications, and have used specific methods such as citation analysis, social network analysis, co-word, and content analysis, as well as text-mining to achieve these goals. Many bibliometrics studies will focus on authorship or will measure the contribution of journal and research organizations, but it may also involve a content analysis of words in titles, abstracts, the full text of books, journal articles or conference proceedings, or keywords assigned to published articles by editors or librarians.
Example of studies using WordStat and QDA Miner
WordStat and QDA Miner have often been used for these kinds of research efforts (see the reference section below). For example, to celebrate the 25th anniversary of Human Communication Research, Stephen (1999) analyzed words in the titles of all 634 articles published during those years and published his results as the last article in the final issue of the 25th volume. Using co-word analysis and hierarchical clustering, he was able to identify relationships among concepts as well as changes in topics studied over time. West (2007) did a similar analysis of words in the titles of 345 papers published in the International Journal of Advertising.
Abstracts are also good indicators to understand and grasp the content of publications. They often provide a more detailed description of the scope of a paper, of its theoretical underpinning, its methodological approach, and of other contextual variables. Lonchamp (2012) used WordStat to analyze a corpus of 121 abstracts from the International Journal of Computer-Supported Collaborative Learning (IJCSCL) to investigate and map its content. Other researchers have analyzed much larger corpora consisting of tens of thousands of journal abstracts on gene expression in order to identify ambiguous designation of genes in the published articles (Coimbra, Vanderwall, and Oliveira (2010) or to reveal unexpected gene associations (Chaussabel & Sheer, 2002).
Keywords may also be good indicators of a paper’s content and have often been used in scientometrics studies (Fratesi, 2008; Lonchamp, 2012; Reinhold, Laesser, & Bazzi, 2014). For example, Fratesi (2008) used keywords and titles of about 175,000 articles published in 68 geology journals between 1945 and 2000 to track changes in the influence of subdisciplines over time.
Several researchers have also used the text-processing capabilities of WordStat to extract and analyze large corpora consisting of the full text of journal articles (Lonchamp, 2012; Anderson et al, 2007; Waismel-Manor, 2011). For example, Waismel-Manor (2011) retrieved 1,317 articles from the Campaigns & Elections journal using Lexis-Nexis, obtaining a text corpus of 1,736,042 words, and was able to demonstrate the gradual professionalization of campaign consultants over time. Anderson, Joly, and Fairhurst (2007) analyzed the full text of 149 articles published over a five-year period in four retail-trade publications to document how retailers used business intelligence and data-mining tools to implement customer relationship management (CRM) in retailing.
Bibliometrics studies have often focused on co-citation analysis, identifying the influence of authors and journals and the relationship among them. One good example is the study of Jacobsen, Punzalan, and Hedstrom (2013). They analyzed 165 articles on collective memory from four leading archival studies journals between 1980 and 2010 to identify which scholars and well-known works have been the most influential. Oleinik (2011) also used WordStat to study authorship and citation patterns across North American and Russian institutions.
Scientometrics studies may also focus on how concepts are being defined over time or in different domains. For example, Walterbush, Gräuler, and Teuteberg (2014) identify similarities and differences in definitions of the word “trust” through literature of research spanning over 50 years. The authors collected a set of 121 definitions from various domains and analyzed those with a word-stem frequency analysis in WordStat as well as with a qualitative data analysis using QDA Miner.
Software features for scientometrics and bibliometrics studies
Several features in QDA Miner and WordStat have proven to be useful for scientometrics and bibliometrics studies. This section presents a quick overview of some of those features.
- Data importation: The ability to import MS Word, RTF, HTML, as well as PDF files, and to associate metadata (such as dates, numerical and categorical data) with those articles, allowing one to easily create a corpus of full-text articles with relevant variables. The QDA Miner ability to import the Reference Information System (RIS) data files is also convenient for importing information from journal databases like ProQuest or bibliographic software such as EndNote or Reference Manager. The Document Conversion Wizard is especially useful to import data from other databases such as Lexis/Nexis, for splitting single documents into multiple ones, or for extracting variables from structured listings or reports.
- Editing, Tagging, and Annotating: Once imported into QDA Miner, documents may be edited, coded, and annotated manually, allowing one to perform content analysis with WordStat while ignoring irrelevant sections or focusing on specific ones. For example, manually tagging the reference sections of journal articles could allow one to perform co-citation analysis, while by manually coding the research method sections of journal articles, one could also describe the evolution of research methods over time or compare methods used in different journals.
- Text Pre-processing: WordStat’s ability to transform words into stems, to lemmatize and remove words of little semantic value (like prepositions, conjunctions, or pronouns) allows one to quickly focus on more relevant words and phrases.
- Words and Phrases Extraction: WordStat can process up to 300,000 words per second and quickly produce frequency counts of significant words, extract common phrases, and produce visual displays in the form of bar charts, word clouds, etc.
- Analysis of Co-Occurrence: The analysis of co-occurrences using statistical techniques such as hierarchical clustering, multidimensional scaling as well as visualization tools like the proximity plot, allows one to promptly identify topics and themes in a discipline. Such tools in WordStat have often been used for mapping scientific domains (Fratesi & Vacher, 2008; Friedman & Smiraglia, 2012; Lonchamp, 2012; Reinold, Lasser & Bazzi (2014).
- Comparative Analysis: The ability to compare frequencies of words, phrases or content categories across different sources (e.g., journals, countries) or to look for changes over time could be used to identify the evolution of a scientific discipline, the rise and fall of specific ideas or concepts, or to document the differentiation process of scientific publications or the geo-spatial distribution of scientific activities. One could compute simple statistical tests (like chi-square tests, F-tests, or correlations), create presentation-quality visualizations (such as bar charts, line charts, bubble charts, or heatmaps), and apply correspondence analysis.
- Application of content analysis dictionaries: The possibility in WordStat to build dictionaries of keywords, key phrases, and proximity rules allows one to focus on specific dimensions. For example, one may easily build a dictionary of authors or journals and perform co-citation pattern analysis. One may also create dictionaries to group together key terms into broader concepts, allow one to measure the prevalence of methodological traditions, theories, research topics, etc.
- Keyword-In-Context: The Keyword-in-Context (or KWIC) feature is an essential tool to test the validity of existing or user-built dictionaries by making sure words or phrases used to measure reference to specific topics are effectively capturing the intended meaning. When an item is found to be ambiguous, KWIC lists are also useful to identify proper disambiguation rules.