46 Text Overview
46.1 Text Overview
Sometimes you have fields that are free long-form text. This cannot be dealt with using the methods we saw in Chapter 14, as the natural text will have enough variation that counting the full-text fields is not going to be fruitful. Imagine this fictitious short book reviews
- This book is excellent
- Best book I have read all year
- This book is absolutely excellent
- This book is far from excellent
- Not worth your time
On a character level, these are all uniquely different. But to us English speakers, the first 3 are positive, and the last 2 are negative. We will in the upcoming chapters look at all the methods and their details, to see how we can modify and extract as much information out of this as we can.
We will adhere to the #BenderRule1 and make it clear that while the examples you will see in this book are going to be in English, it doesnβt mean that the results you see can be replicated in other languages. All languages have unique things about them that make these types of calculations harder or easier to use.
In these chapters, we will be studying methods to work with text. Notably, text and language are different things, with many languages starting as spoken words, and a smaller number of these develop a writing system. This is important to remember, as the written text will almost always have less information than the language it is transcribing. You have seen examples of this when you send a text message that wasnβt received correctly, because of the missing non-verbal information.
Text data, like all other types of data, isnβt guaranteed to contain information that helps your modeling data at hand. This is especially true with shorter text fields but can be true with just about anything.
The types of operations you do to get text data into numeric data are various, but they can usually be split into several different tasks. You donβt have to do everything in this list to get good results, and some off-the-shelf methods will sometimes combine 2 or more of these steps into one. This is another reason why it is important to read the documentation of the implementation you are using.
There is also the possibility that you know exactly what type of information is important in your text field. In that case, you will be better served by manually extracting the information with regular expression or related text processing tools, we look at examples of that in Chapter 47.
The coding portions of the following chapters will be focused on using text as features in addition to other features. This is unlike other modeling tasks where text is the whole input, such as in translation or text2image tasks. Despite this, they would still be a worthwhile read for practitioners working on those types of tasks.
The chapters in this section read quite differently than the other chapters, as many of the chapters assume that preview chapters are used. This is because some of the chapters describe part of the text handling rather than a specific method. See the diagram below for possible workflows.
create diagram of how the chapters work together in a flow chart kind of way
46.2 Text cleaning
In Chapter 48, we will look over the ways we take raw text and get it ready for later tasks. This work deals with encoding issues, standardization, and cases and sometimes you need to get rid of a lot of unwanted chunks.
46.3 Tokenization
Once the text is cleaned, we need to split it into a smaller unit of information such that we can count it, this is called tokenization and we will visit that in Chapter 49.
46.4 Modifying tokens
Once you have the data as tokens, one of the things you might want to do is modify them in various ways. This could be things like changing the endings to words or changing the words entirely. We see examples of this in Chapter 50 and Chapter 51.
46.5 Filtering tokens
The tokens you create might not all be of the same quality. Depending on your choice of tokenizer, there will be reasons for you to remove some of the tokens you have created. We see examples of this in Chapter 52 and Chapter 53.
46.6 Counting tokens
We have gotten to the end of the line and we are ready to turn the tokens into numeric variables we can use. There are many different ways we look at them in Chapter 54, Chapter 55, Chapter 56, Chapter 57, and Chapter 58.
46.7 Embeddings
Another way to use text is to work with embeddings, this is another powerful tool that can give you good performance. We look at some of them in Chapter 59 and Chapter 60.
https://thegradient.pub/the-benderrule-on-naming-the-languages-we-study-and-why-it-matters/β©οΈ