Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Series TitleSeries Title
-
Reading LevelReading Level
-
YearFrom:-To:
-
More FiltersMore FiltersContent TypeItem TypeIs Full-Text AvailableSubjectPublisherSourceDonorLanguagePlace of PublicationContributorsLocation
Done
Filters
Reset
70,783
result(s) for
"Text analysis"
Sort by:
Automated Text Analysis for Consumer Research
2018
The amount of digital text available for analysis by consumer researchers has risen dramatically. Consumer discussions on the internet, product reviews, and digital archives of news articles and press releases are just a few potential sources for insights about consumer attitudes, interaction, and culture. Drawing from linguistic theory and methods, this article presents an overview of automated text analysis, providing integration of linguistic theory with constructs commonly used in consumer research, guidance for choosing amongst methods, and advice for resolving sampling and statistical issues unique to text analysis. We argue that although automated text analysis cannot be used to study all phenomena, it is a useful tool for examining patterns in text that neither researchers nor consumers can detect unaided. Text analysis can be used to examine psychological and sociological constructs in consumer-produced digital text by enabling discovery or by providing ecological validity.
Journal Article
Handbook of research on opinion mining and text analytics on literary works and social media
\"This book uses artificial intelligence and big data analytics to conduct opinion mining and text analytics on literary works and social media, focusing on theories, method, applications and approaches of data analytic techniques that can be used to extract and analyze data from literary books and social media, in a meaningful pattern\"-- Provided by publisher.
No Longer Lost in Translation: Evidence that Google Translate Works for Comparative Bag-of-Words Text Applications
by
de Vries, Erik
,
Schoonvelde, Martijn
,
Schumacher, Gijs
in
Automatic text analysis
,
Automation
,
Comparative analysis
2018
Automated text analysis allows researchers to analyze large quantities of text. Yet, comparative researchers are presented with a big challenge: across countries people speak different languages. To address this issue, some analysts have suggested using Google Translate to convert all texts into English before starting the analysis (Lucas et al. 2015). But in doing so, do we get lost in translation? This paper evaluates the usefulness of machine translation for bag-of-words models—such as topic models. We use the europarl dataset and compare term-document matrices (TDMs) as well as topic model results from gold standard translated text and machine-translated text. We evaluate results at both the document and the corpus level. We first find TDMs for both text corpora to be highly similar, with minor differences across languages. What is more, we find considerable overlap in the set of features generated from human-translated and machine-translated texts. With regard to LDA topic models, we find topical prevalence and topical content to be highly similar with again only small differences across languages. We conclude that Google Translate is a useful tool for comparative researchers when using bag-of-words text models.
Journal Article
Computer-Assisted Keyword and Document Set Discovery from Unstructured Text
by
Roberts, Margaret E.
,
Lam, Patrick
,
King, Gary
in
AJPS WORKSHOP
,
Algorithms
,
Automatic text analysis
2017
The (unheralded) first step in many applications of automated text analysis involves selecting keywords to choose documents from a large text corpus for further study. Although all substantive results depend on this choice, researchers usually pick keywords in ad hoc ways that are far from optimal and usually biased. Most seem to think that keyword selection is easy, since they do Google searches every day, but we demonstrate that humans perform exceedingly poorly at this basic task. We offer a better approach, one that also can help with following conversations where participants rapidly innovate language to evade authorities, seek political advantage, or express creativity; generic web searching; eDiscovery; look-alike modeling; industry and intelligence analysis; and sentiment and topic analysis. We develop a computer-assisted (as opposed to fully automated or human-only) statistical approach that suggests keywords from available text without needing structured data as inputs. This framing poses the statistical problem in a new way, which leads to a widely applicable algorithm. Our specific approach is based on training classifiers, extracting information from (rather than correcting) their mistakes, and summarizing results with easy-to-understand Boolean search strings. We illustrate how the technique works with analyses of English texts about the Boston Marathon bombings, Chinese social media posts designed to evade censorship, and others.
Journal Article
A multidimensional approach for detecting irony in Twitter
by
Reyes, Antonio
,
Veale, Tony
,
Rosso, Paolo
in
Assessments
,
Automatic text analysis
,
Automotive engineering
2013
Irony is a pervasive aspect of many online texts, one made all the more difficult by the absence of face-to-face contact and vocal intonation. As our media increasingly become more social, the problem of irony detection will become even more pressing. We describe here a set of textual features for recognizing irony at a linguistic level, especially in short texts created via social media such as Twitter postings or \"tweets\". Our experiments concern four freely available data sets that were retrieved from Twitter using content words (e.g. \"Toyota\") and user-generated tags (e.g. \"#irony\"). We construct a new model of irony detection that is assessed along two dimensions: representativeness and relevance. Initial results are largely positive, and provide valuable insights into the figurative issues facing tasks such as sentiment analysis, assessment of online reputations, or decision making.
Journal Article
Simultaneously Discovering and Quantifying Risk Types from Textual Risk Disclosures
2014
Managers and researchers alike have long recognized the importance of corporate textual risk disclosures. Yet it is a nontrivial task to discover and quantify variables of interest from unstructured text. In this paper, we develop a variation of the latent Dirichlet allocation topic model and its learning algorithm for simultaneously discovering and quantifying risk types from textual risk disclosures. We conduct comprehensive evaluations in terms of both conventional statistical fit and substantive fit with respect to the quality of discovered information. Experimental results show that our proposed method outperforms all competing methods, and could find more meaningful topics (risk types). By taking advantage of our proposed method for measuring risk types from textual data, we study how risk disclosures in 10-K forms affect the risk perceptions of investors. Different from prior studies, our results provide support for all three competing arguments regarding whether and how risk disclosures affect the risk perceptions of investors, depending on the specific risk types disclosed. We find that around two-thirds of risk types lack informativeness and have no significant influence. Moreover, we find that the informative risk types do not necessarily increase the risk perceptions of investors-the disclosure of three types of systematic and liquidity risks will increase the risk perceptions of investors, whereas the other five types of unsystematic risks will decrease them.
Data, as supplemental material, are available at
http://dx.doi.org/10.1287/mnsc.2014.1930
.
This paper was accepted by Alok Gupta, special issue on business analytics
.
Journal Article
Issues and Best Practices in Content Analysis
by
Watson, Brendan R.
,
Lovejoy, Jennette
,
Lacy, Stephen
in
Algorithms
,
Authorship
,
Best practice
2015
This article discusses three issues concerning content analysis method and ends with a list of best practices in conducting and reporting content analysis projects. Issues addressed include the use of search and databases for sampling, the differences between content analysis and algorithmic text analysis, and which reliability coefficients should be calculated and reported. The “Best Practices” section provides steps to produce reliable and valid content analysis data and the appropriate reporting of those steps so the project can be properly evaluated and replicated.
Journal Article
The Methodological Divide of Sociology: Evidence from Two Decades of Journal Publications
by
Schwemmer, Carsten
,
Wieczorek, Oliver
in
Academic disciplines
,
Automatic text analysis
,
Prestige
2020
Past research indicates that Sociology is a low-consensus discipline, where different schools of thought have distinct expectations about suitable scientific practices. This division of Sociology into different subfields is to a large extent related to methodology and choices between qualitative or quantitative research methods. Relying on theoretical constructs of the academic prestige economy, boundary demarcation and taste for research, we examine the methodological divide in generalist Sociology journals. Using automated text analysis for 8737 abstracts of articles published between 1995 and 2017, we discover evidence of this divide, but also of an entanglement between methodological choices and different research topics. Moreover, our results suggest a marginally increasing time trend for the publication of quantitative research in generalist journals. We discuss how this consolidation of methodological practices could enforce the entrenchment of different schools of thought, which ultimately reduces the potential for innovative and effective sociological research.
Journal Article