Hausarbeiten logo
Shop
Shop
Tutorials
De En
Shop
Tutorials
  • How to find your topic
  • How to research effectively
  • How to structure an academic paper
  • How to cite correctly
  • How to format in Word
Trends
FAQ
Go to shop › Speech Science / Linguistics

(Mean) type/token ratios

Computing the type/token ratio

Title: (Mean) type/token ratios

Term Paper , 2008 , 16 Pages

Autor:in: Jörn Piontek (Author)

Speech Science / Linguistics

Excerpt & Details   Look inside the ebook
Summary Excerpt Details

In this paper I will show how it is possible to compute the type/token ratio of a text by using the programming language Perl. First of all an overview about the topic will be given, including the definition of both the terms type and token and how they are used in the context of this program. Then I will explain how the program works and give further rationale for its shortcomings. Although the program is rather simple, some knowledge of the programming language Perl will be needed for the respective parts in this paper. Then I will proceed to do a short analysis of different texts and their respective type/token ratios. These texts were taken from the British National Corpus and Project Gutenberg. The results will show the need for a different measure of lexical density. One example of such a measure is the mean type/token ratio which I will go into shortly. In the Conclusion there will be a short critique of the expressiveness of type/token ratios as well as a short overview about current research on this topic.

Excerpt


Table of Contents

1 Introduction

2 Type/token ratios

2.1 Types and tokens

2.2 Type/token ratio

3 The Program

3.1 Computing the type/token ratio

3.2 Demonstration

4 Mean type/token ratios

5 Conclusion

6 References

7 Appendix

Objectives and Topics

This paper examines the computational calculation of type/token ratios, exploring the limitations of basic measurements and proposing mean type/token ratios as a more robust alternative for analyzing lexical diversity across texts of varying lengths.

  • Definition and linguistic distinction between types and tokens
  • Implementation of TTR-calculating software using the Perl programming language
  • Impact of text length and sampling on lexical diversity metrics
  • Challenges in automated tokenization and lemmatization
  • Introduction and evaluation of mean type/token ratios to mitigate skewing

Excerpt from the Book

2.1 Types and tokens

As the usage of the terms type and token is a lot more intuitive and clearly defined in the context of the program as opposed to the linguistic or even philosophical definition, this will be used as the starting point of this section.

Both the sentences (1a) and (1b) contain 10 words, in a certain sense. One could argue, however, that sentence (1b) only contains 9 words because “right” appears twice. These two different point of view represent the distinction between types and tokens as it is made by the program.

(1) a. Tom used his left hand to open the right door.

b. Tom used his right hand to open the right door.

All the words in sentence (1a) differ from each other. Thus, the number of tokens and the number of types are equal. In sentence (1b) there are 10 tokens as well. The word “right” appears twice and is therefore only counted as one type. No matter how often a certain word form appears in one text, it is only one type.

Summary of Chapters

1 Introduction: This chapter outlines the paper's intent to demonstrate the calculation of type/token ratios using Perl and introduces the need for mean type/token ratios.

2 Type/token ratios: This chapter defines the core concepts of types and tokens and explains how the type/token ratio serves as an indicator of lexical diversity.

3 The Program: This chapter describes the technical implementation of the Perl program for calculating ratios and provides a practical demonstration using texts from the British National Corpus.

4 Mean type/token ratios: This chapter introduces the mean type/token ratio as a standardized measure designed to eliminate the skewing effects of varying sample sizes.

5 Conclusion: This chapter summarizes the limitations of standard type/token ratios and reflects on the necessity of more advanced measures in computational text analysis.

6 References: This chapter lists the academic literature and sources cited throughout the paper.

7 Appendix: This chapter provides the complete source code for the programs used to calculate type/token and mean type/token ratios.

Keywords

Type/token ratio, TTR, computational linguistics, Perl, lexical diversity, tokenization, mean type/token ratio, corpus linguistics, text analysis, vocabulary diversity, sample size, lemmatization, language acquisition, statistical linguistics.

Frequently Asked Questions

What is the primary focus of this paper?

The paper focuses on the computational method of calculating type/token ratios to measure lexical diversity and addresses the technical and theoretical limitations of this metric.

What are the central thematic fields discussed?

The core themes include linguistic definitions of types and tokens, computer-aided text analysis, the influence of text length on statistical measures, and software implementation.

What is the main objective or research question?

The objective is to show how to compute these ratios using Perl and to argue for the use of "mean" type/token ratios to compensate for the skewing effects of different text lengths.

Which scientific method is employed?

The author uses a computational approach, implementing custom algorithms in the Perl programming language to analyze text data sourced from the British National Corpus and Project Gutenberg.

What is covered in the main section?

The main section covers the definitions of types and tokens, the logic behind the Perl program implementation, a demonstration using actual corpus data, and an evaluation of mean type/token ratios.

Which keywords characterize this work?

Key terms include Type/token ratio, lexical diversity, computational linguistics, Perl, corpus analysis, and sample size distortion.

Why does the author mention that standard TTR is "flawed"?

The author explains that standard TTR is highly dependent on the number of words in a sample; therefore, comparing texts of different lengths produces distorted results.

How does the program handle word forms like "going" and "went"?

The current program does not perform lemmatization and therefore treats different inflected forms as separate types, which the author identifies as a limitation.

What is the purpose of segmenting the text?

Segmenting the text into smaller, equal-sized parts and calculating the average ratio (mean TTR) is intended to neutralize the influence of total text length on the final metric.

Excerpt out of 16 pages  - scroll top

Details

Title
(Mean) type/token ratios
Subtitle
Computing the type/token ratio
College
University of Münster  (Englisches Seminar)
Course
Computational Text Analysis
Author
Jörn Piontek (Author)
Publication Year
2008
Pages
16
Catalog Number
V168529
ISBN (eBook)
9783640867271
ISBN (Book)
9783640867813
Language
English
Tags
computing
Product Safety
GRIN Publishing GmbH
Quote paper
Jörn Piontek (Author), 2008, (Mean) type/token ratios, Munich, GRIN Verlag, https://www.hausarbeiten.de/document/168529
Look inside the ebook
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
Excerpt from  16  pages
Hausarbeiten logo
  • Facebook
  • Instagram
  • TikTok
  • Shop
  • Tutorials
  • FAQ
  • Payment & Shipping
  • About us
  • Contact
  • Privacy
  • Terms
  • Imprint