scholarly journals ETAT: Expository Text Analysis Tool

2002 ◽  
Vol 34 (1) ◽  
pp. 93-107 ◽  
Author(s):  
Eduardo Vidal-Abarca ◽  
Héctor Reyes ◽  
Ramiro Gilabert ◽  
Javier Calpe ◽  
Emilio Soria ◽  
...  
2020 ◽  
Author(s):  
Leah P. Macfadyen

Curriculum analysis is a core component of curriculum renewal. Traditional approaches to curriculum analysis are manual, slow and subjective, but some studies have suggested that text analysis might usefully be employed for exploration of curriculum. This concise paper outlines a pilot use case of content analytics to support curriculum review and analysis. I have co-opted Quantext – a relatively user-friendly text analysis tool designed to help educators explore student writing – for analysis of the text content of the 17 courses in our online master’s program. Quantext computed descriptive metrics and readability indices for each course and identified top keywords and ngrams per course. Compilation and comparison of these revealed frequent curricular topics and networks of thematic relationships between courses, in ways that both individual educators and curriculum committees can interpret and use for decision-making. Future Quantext features will allow even more sophisticated identification of curricular gaps and redundancies.


2016 ◽  
Vol 21 (1) ◽  
pp. 105-115 ◽  
Author(s):  
Michael Barlow

In this article, I provide a brief introduction to the operation and motivation behind the text analysis tool WordSkew. This program, currently available for Windows, is a variant of a typical concordance program. The distinguishing feature of the software is that it allows the user to specify the units of discourse and apposite ways of segmenting the discourse. The results of a search query are then given with respect to each segment. For example, sentences might be divided into ten segments (based on word counts) and the frequency of the search term is then provided for each segment. This process is repeated as required for other textual units.


2021 ◽  
Author(s):  
Russell J Jarvis ◽  
Patrick M. McGurrin ◽  
Rebecca Featherston ◽  
Marc Skov Madsen ◽  
Shivam Bansal ◽  
...  

Here we present a new text analysis tool that consists of a text analysis service and an author search service. These services were created by using or extending many existing Free and Open Source tools, including streamlit, requests, WordCloud, TextStat, and The Natural Language Tool Kit. The tool has the capability to retrieve journal hosting links and journal article content from APIs and journal hosting websites. Together, these services allow the user to review the complexity of a scientist’s published work relative to other online-based text repositories. Rather than providing feedback as to the complexity of a single text as previous tools have done, the tool presented here shows the relative complexity across many texts from the same author, while also comparing the readability of the author’s body of work to a variety of other scientific and lay text types. The goal of this work is to apply a more data-driven approach that provides established academic authors with statistical insights into their body of published peer reviewed work. By monitoring these readability metrics, scientists may be able to cater their writing to reach broader audiences, contributing to an improved global communication and understanding of complex topics.


2020 ◽  
Vol 46 (Supplement_1) ◽  
pp. S222-S223
Author(s):  
Tatiana Baxter ◽  
Hyeon-Seung Lee ◽  
Lénie Torregrossa ◽  
Seoyeon Kim ◽  
Sohee Park

Abstract Background Schizophrenia has been suggested to be a disorder of social communication, which depends on the way language is used to convey thoughts, beliefs, feelings, and intentions. Everyday language can also reveal personality, emotions, and social skills of the speaker. Extensive past research affirms the central importance of language and thought disorder as diagnostic features of schizophrenia, mostly focused on the neurocognitive aspects of language output collected during clinical interviews, and not on the social nature of language. In this study, we examined narratives written in response to viewing social scenes by individuals with schizophrenia (SZ) and matched controls (CO) using an automated computational linguistics and statistical-based text analysis tool that computes socially-relevant variables. Methods 23 individuals with schizophrenia (SZ) and 23 demographically matched controls (CO) were shown paintings of social situations, and were asked to write reflections describing what they thought and felt about these scenes. Two pictures were presented consecutively. There was no time limit. Resulting narratives were analyzed with the Linguistic Inquiry and Word Count program (LIWC; Pennebaker et al, 2015). LIWC computes basic linguistic variables such as the % of self-referring and non-self-referring pronouns, social and emotion words, and cognitive items. LIWC also generates 4 complex variables: formal and logical thinking patterns (“analytic”); social status or confidence (“clout”); authenticity, and emotional tone. Clinical symptoms in SZ were assessed using the Scale for the Assessment of Positive Symptoms (SAPS) and the Scale for the Assessment of Negative Symptoms (SANS). For all participants, the UCLA Loneliness Scale, the National Adult Reading Test (NART), and the Edinburgh Handedness Inventory (EHI) were administered. Results The two groups did not differ in NART or EHI. There was no group difference in the number of words produced. SZ produced greater number of pronouns than CO but this effect was driven by a higher % of self-referring pronouns in SZ than CO, with no group difference in non self-referring pronouns. For complex variables, CO scored significantly higher in authenticity than SZ but no group differences were observed in clout, analytics, or emotional tone. Loneliness was higher in SZ compared with CO. There were no other significant group differences. In SZ, pronoun use was correlated with positive symptoms (especially with ratings of hallucinations, bizarre behavior, delusions of mind-reading and thought broadcasting) and inversely correlated with negative symptoms (avolition, apathy and motor retardation). Social words were inversely correlated with SAPS Thought Disorder. Clout was inversely correlated with SANS Alogia and SAPS Thought Disorder. Authenticity was correlated with SANS Anhedonia and Asociality. In CO, loneliness was correlated with the % negative emotion words and NART was correlated with total number of emotion words. Discussion We used an automated linguistic analysis tool to extract information relevant to social communication from written narratives. We found group differences in the use of pronouns and authenticity. We also observed associations of clinical symptoms with certain social aspects of language use in schizophrenia. One advantage of automated text analysis tools is the minimization of implicit biases inherent in ratings of interviews. Limitations of this study include lack of direct social functioning measures and the sample size. Future work will incorporate linguistic text analysis within a social paradigm to directly examine the role of language use in social functioning.


2017 ◽  
Vol 27 (7) ◽  
pp. 756-771 ◽  
Author(s):  
Joanna K. Huxster ◽  
Matthew H. Slater ◽  
Jason Leddington ◽  
Victor LoPiccolo ◽  
Jeffrey Bergman ◽  
...  

This study examines the conflation of terms such as “knowledge” and “understanding” in peer-reviewed literature, and tests the hypothesis that little current research clearly distinguishes between importantly distinct epistemic states. Two sets of data are presented from papers published in the journal Public Understanding of Science. In the first set, the digital text analysis tool, Voyant, is used to analyze all papers published in 2014 for the use of epistemic success terms. In the second set of data, all papers published in Public Understanding of Science from 2010–2015 are systematically analyzed to identify instances in which epistemic states are empirically measured. The results indicate that epistemic success terms are inconsistently defined, and that measurement of understanding, in particular, is rarely achieved in public understanding of science studies. We suggest that more diligent attention to measuring understanding, as opposed to mere knowledge, will increase efficacy of scientific outreach and communication efforts.


2015 ◽  
Vol 16 (3) ◽  
pp. 252-262 ◽  
Author(s):  
Nisa Bakkalbasi ◽  
Melissa Goertzen

Purpose – Over the past decade, as the electronic book (e-book) collection continues to grow, Columbia University Libraries has been gathering information to develop policies related to e-book acquisition, discovery, and access. The purpose of this paper is to investigate users’ e-book search behavior and information needs across different disciplines. Design/methodology/approach – The research method utilizes text data from two sources: users’ e-book search queries that were entered into the libraries discovery tool called CLIO and e-book title words provided by the Counting Online Usage of Networked Electronic Resources (COUNTER) usage reports. The analysis involves identifying and quantifying certain words from users’ search queries with the purpose of examining the contexts within which these words were used. Findings – The prominence of topical words such as “history,” “social,” and “politics” in the list was an interesting reflection on the kinds of works users were looking for, as were the terms “handbook,” “guide,” and “manual.” The high frequency of these words imply that users were searching for broad topics, reference works, or other collections of instructions, all of which are intended to provide ready reference. Originality/value – Running search queries and e-book title words through a text analysis tool revealed new ideas related to what types of materials users search for and use. Text analysis of search terms and title words provided insight into the nature of e-book use, including broad topic (e.g. history), academic level of use (e.g. introductory), and genre/type (e.g. reference). While it is challenging to deduce reader intent from word frequency analysis, as text data remain widely open for interpretation, the methodology has significant strengths that drive us to continue to use in future studies.


2021 ◽  
pp. 1-25
Author(s):  
Estelle Joubert

This article offers a series of experiments exploring the potential for ‘distant reading’ in French music criticism. ‘Distant reading’, a term first coined by literary theorist Franco Moretti, refers to quantitative approaches that allow for new insights into a large corpus of texts by aggregating data. While the main corpus employed here is the Revue et gazette musicale de Paris (1831–1877), I also use secondary corpora of reviews of Félicien David's Herculanum in 1859, Berlioz's reviews of Gluck and Beethoven in the Journal des débats and reviews that mention Gabriel Fauré in the Library of Congress’ Chronicling America database. My experiments employ a text analysis tool named Voyant, built by Geoffrey Rockwell and Stéfan Sinclair, thereby also offering a basic introduction to the range of visualizations employed in distant reading. My experiments focus on areas in which quantitative methods are particularly well suited to generating new knowledge: corpus-wide visualizations and queries, moving beyond traditional text searching, investigations of music critics’ authorial styles and detecting sentiment in reviews, and finally, to geographies of music criticism.


Sign in / Sign up

Export Citation Format

Share Document