|
The Coronavirus Corpus is designed to be the
definitive record of the social, cultural, and economic impact of the coronavirus (COVID-19) in 2020 and beyond.
The corpus (which was first released in May 2020) is currently about
1559 million words in size,
and it continues to grow by 3-4 million words each day. You can also look at "collocates" (nearby words) to see what is being said about a certain topic, such as (verbs near) virus, or any word near ban (v), stockpile, disinfect*, or remotely. And you can even see the collocates of a word in each 10-day period since Jan 2020 (e.g. stockpile). The corpus also allows you to see the patterns in which a word occurs, as with stay-at-home, social, economic, or hoard*. You can also compare between different time periods, to see how our view of things have changed over time. (And you can even compare between the 20 countries in the corpus). Interesting comparisons over time might include phrases like social * or economic * that were more common in Jan/Feb than in Apr/May, or words near BAN or OBEY that were more common in Apr-May than in Jan-Feb. Click on any of the links in the search form on the search page for context-sensitive help, and to see the range of queries that the corpus offers (LIST discusses the search syntax). You might pay special attention to the virtual corpora, which allow you to create personalized collections of texts related to a particular area of interest. Finally, the corpus is related to many other corpora of English that we have created. These corpora were formerly known as the "BYU Corpora"), and they offer unparalleled insight into variation in English.
LIST display
Find single words like
stay-at-home,
all forms of a word like FLATTEN,
words matching patterns like
*open*,
phrases like more * than or
social NOUN.
You can also search by synonyms (e.g.
scary), and
customized wordlists like
clothes. In each case, you see each individual
matching string.
You can use parts of speech as part of your query. For example, ADJ eyes would find a two word string, composed of an adjective followed by the word eyes. Some other examples are: rough NOUN, NAME Smith, VERB * money, TALK ADV, NUM people, LET PRON VERB. An easy way to use part of speech tags is by selecting them from the drop-down list (click on [PoS] to show it). You can also type the part of speech tags directly into the search form. Click here for a list of these part of speech tags. Previously, you had to use the part of speech tag (from the link above) inside of brackets, e.g. [j*]. But that's a bit cumbersome for mobile phones, and there are now different ways of specifying the part of speech -- all of which work equally as well. For example, all of the following would find the same strings: ADJ eyes, [j*] eyes, J eyes, _j eyes.
If you are using Type 1 or Type 4 above, you can use wildcards for the part of speech tag. For example, [nn2*] = plural nouns, [n*] = all nouns, [*n*] = nouns (including ambiguous noun/adj tags), etc. If you are using Type 2 or Type 3, it needs to be upper case: short NOUN. You can also add a part of speech tag to the end of any word, but you need to use either Type 1 or Type 4 above. For example, end would find end with any part of speech, but end.[n*] or end_n would limit it to end as a noun, and end_v or end_v would limit it to end as a verb. Make sure that you separate the word and the part of speech with a period / full stop and bracket (Type 1) or an underscore (Type 4), and that there is no space. Remember also that you can combine these with lemma searches to find all forms of a word with a given part of speech, e.g. END_v or END.[v*]. If you don't know what the part of speech tag is for a given word (or the words in a phrase), just select [OPTIONS] and then [GROUP BY] = [NONE] (SHOW POS). For example, see the PoS tags for light, back, front, or in light of If you capitalize an entire word, it will find all forms of that word. For example, FLATTEN (as in flatten the curve) would find all forms of flatten (flatten, flattens, flattening, flattened), whereas flatten would just find the single form flatten. Another example: =BAN_v would find all of the forms of all synonyms of ban (banning, forbids, preventing, disallowed, etc), whereas =ban would just find prevent, disallow, etc. (Notice that we have also added the part of speech _v to the end, to limit these to verbs.) You can search by all of the synonyms of a given word, which provides powerful "semantically-based" searches of the corpus. For example, you can find the synonynms of scary, worried, or ban (v). Of course you can use the synonyms as part of phrases as well. For example, =BAN * NOUN, =clever =man, or =strong ARGUMENT. As the last example shows, synonyms can be very useful when you are a non-native speaker, and you want to know which related words are used in a particular context. As =clean * NOUN shows, not every token will actually be a synonym of a given word in every case. For example, scour may be a synonym of clean in scour the sink, but not in scour the library for good books. Note the it is often useful to limit the synonyms to those with a particular part of speech, as in clean_v. It is often also useful to find all forms of the synonyms, by capitalizing the word: CLEAN. And of course you can combine these as well, for example all forms of all synonyms of clean as a verb. Finally, note that you can click on the [S] to find synonyms for each word in the results set. This allows you to follow a "synonym chain" from one word to another to another...
The Hansard and EEBO corpora have been "semantically tagged", and you can use these tags as part of your search. A few examples are given below.
"User lists" or "customized lists" are word lists that you create -- related to a certain topic (e.g. sports, clothing, or emotions), words that are grammatically related (e.g. a certain subset of adverbs or pronouns), or any other listing that you might want. For example, click here to run a query based on two sample word lists that we created -- one with a list of colors, and the other with a short list of parts of clothing. There
are two ways to create a customized list: You can later view the lists that you have created, and modify the wordlist (add or delete words), or delete a list entirely. Once created, you can re-use a wordlist in queries at any time in the future -- they remain stored in the database on the server. The easiest way to include a list in the main search window is to just select it in the wordlist window. If desired, you can also type it into the search form directly. The format is: @listName 1.Select [SAVE LISTS] = 'YES' in the search form 2. Run a query, such as (finding synonyms of beautiful) 3. Select and de-select words from the list by clicking in the checkbox to the left of each word. Only the words that you select will be saved to the list. You can use the checkbox to the left of the [CONTEXT] button to select or de-select the entire list. 4. Enter the name you want to give to the list (in this case, maybe beautiful-syn). 5. Make sure you really have selected some words (step 2 above), and then click [Submit] to save your list. 6. If you want, select the list that you've saved in the customized wordlists interface. You can add to the list, modify entries (click M), or delete words from the list. 7. Finally, you can then re-use this list as part of subsequent queries. For example, if [mark_davies@byu.edu] has created and stored the list [beautiful-syn] then he could find cases of was ADV followed by one of these adjectives.
Many of the examples shown in the other sections are for individual words. But you can combine the different types of searches to create fairly complex phrases. For example:
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
CHART display
If you are interested in a set of words or a grammatical construction, then the LIST option shows the frequency of each matching form (end up being, ended up saying, etc), while the CHART option shows the total frequency in each section. (in the Coronavirus Corpus, 10-day periods since early 2020). It is also possible to see the frequency of words and phrases by day, e.g. hoarding, Italy, For example: coronavirus, COVID-19, social distanc*, FLATTEN the curve, WORK * home, SHELTER in place, Zoom, pandemic, hoard*, toilet paper, quarantin*, curbside, pandemic, reopen, defy, (COVID|coronavirus) fatigue; (political) budget deficit, red/blue states, Boris Johnson, President Trump; (places) Wuhan, China, Italy, Spain, Sweden, and New York City.
COLLOCATES display
See what words occur near other words, which provides great
insight into meaning and usage.
For example, nouns after
dangerous
or
reopen,
verbs before
virus,
or any word near
ban (v),
stockpile,
disinfect*, or
remotely.
The collocates search finds words near another word (i.e. within a "cloud" of nearby words), whereas the
LIST search finds an exact string of words.
For both the WORD and COLLOCATES field, you can include the full range of searches, including words, lemmas, substrings, parts of speech, and synonyms. For example, the following are searches for collocates of gap (n): any word, nouns, adjective, the word fill, synonyms of large.
Select the "span" (number of words to the left and the right) for the
collocates. Use + to search more than four words to the left or right, and 0 to
exclude the words to the left or right. If you don't select a span, it will
default to 4 words left and 4 words right.
You can use collocates to do "variable length" searches, where there might be 0-4 (or more) words between two other sets of words or phrases.
For example, you could find all of the following with one simple search.
COMPARE WORDS display
Compare the collocates of two words, to see how they differ in meaning and usage. (Note: the "compare words" function may not be as useful in this corpus as in the other corpora from English-Corpora.org).
For example,
utter and sheer (note the negative collocates with utter),
warm and hot,
small and little, or adjectives near
boy and girl.
KWIC (Keyword in Context) display
See the patterns in which a word occurs, by sorting the words to the left and/or right. For example:
stay-at-home,
social,
economic,
hoard*,
ban (n),
ban (v).
Select the words that you want to sort with. Select L for 1, 2, and 3 words to the left. Select R for 1, 2, and 3 words to the right. You could also, for example, sort by one word to the left, then one and two words to the right. Click * to clear the entries and start over. Use the dropdown list to the left (POS or _pos) to input tags for "parts of speech" (PoS, e.g. nouns or verbs) into your search string. By default, it will add the PoS as a "full word", as in the searches strong NOUN or ADJ eyes. You can also have the PoS added as a "tag" on the end of a word, to limit the word to that PoS, as in the searches strike_n or and FIND_v. To make it insert PoS tags after words, click on _pos. To change it back to PoS as a separate "word", click on POS. You can find a wealth of information for the top 60,000 words in the corpus. As the following examples with bread show, you can see:
You can find a wealth of information for the top 40,000 words in the corpus, including:
SECTIONS
SHOW Determines whether the frequency is shown for each "section" of the corpus
(in the case of the Coronavirus Corpus, the 10-day periods since Jan 2020).
For example, the
synonyms of beautiful in
each section and
overall.
Select a time period (e.g. 2015 or 2015-2016) or country (e.g. US, GB, IN(dia) )
(Optional) Select a second (set of) section(s) against which to compare the sections chosen above
Two final examples show how the corpus can be used to find changes over just a couple of weeks. The first search compares all nouns in Mar 21-31 to Mar 1-10. Although there
are some strange entries there (just ignore them), we can see that in late March, people were talking about hydroxychloroquine, (closing) campgrounds, curfews,
takeout (food), ventilators, and (closing of) salons much more than 10-20 days earlier. The second search shows that between Mar 21-31 and Apr 21-30, people were
"pushing back" on restrictions, and that there was increasing mention of protesting, reopening (closed stores), and easing and loosening restrictions.
OTHER OPTIONS
# HITS is the number of results. # KWIC is the number of results for a KWIC (concordances) search. GROUP BY determines whether words are grouped by word form (e.g. decide and decided separately), lemma (e.g. all forms of decide together), and whether you see the part of speech for word (e.g. beat as a noun and verb displayed separately). CASE SENSITIVE determines whether She thought and she thought would be two different searches, or The Office, the Office, and the office. DISPLAY shows raw frequency, occurrences per million words, or a combination of these. SAVE LISTS allows you to create a wordlist from the results and then re-use it later in your searches.
SORT / LIMIT
Sort by raw frequency (e.g. hard * ) or by "relevance" ( hard *). Relevance uses the Mutual Information score. It is often useful to specify the minumim frequency when you are sorting by "relevance", to eliminate very low frequency strings. For example, collocates of green where minimum frequency = 1 (strange once-off strings) and where minimum frequency = 20. Note also that when you do a collocates search and you don't specify anything for the collocates field, it will automatically set MINIMUM to MUT INFO = 3 (Mutual Information score). It does this to remove high frequency noise words like the, to, with, etc. If you want to see more of these words, lower the MI score; to see less, increase it.
VIRTUAL CORPORA
Create a "virtual corpus" -- essentially your own personalized corpus within CORONA. You can create the corpus either by keywords in the texts (e.g. texts with the words hoarding, lockdown, or Italy), or information about the texts (e.g. date, title, or source), or a combination of keyword and text information. You can then edit your virtual corpora, search within a particular virtual corpus, compare the frequency of a word, phrase or grammatical construction in your different virtual corpora, and also create "keyword lists" based on the texts in your virtual corpus.
Click on any of the links above for more information. To create a virtual corpus by keywords, enter a word or phrase to the left, and then set TEXTS/VIRTUAL to FIND TEXTS (do it / undo; must be logged in first). You might also want to set SORT/LIMIT to RELEVANCE and MINIMUM FREQUENCY to something like 5 (the minimum number of times you want the word to occur in a text) (do it / undo). After clicking SUBMIT, you will see a list of matching texts from the corpus. For example, see matching texts for hoard*, contact tracing, or social distanc*. On the "results" page, choose how many texts you want in your virtual corpus, and then click SAVE LIST. After the virtual corpus is created, you might want to click on FIND KEYWORDS to see whether the corpus is providing the focus that you want. You can create a virtual corpus by selecting texts that match certain criteria -- such as title of the source (e.g. New York Times) or the title of the article, the topic, the date, and so on. Click on CREATE CORPUS to the left to see the interface to select the texts. As an example, this list was created by searching for articles dealing with contact tracing. Note that in that search form, you can also make sure that the texts have certain words in them. If you want more control in finding texts with certain words, you might want to search by keywords. See list that was created by searching CORONA for articles dealing with contact tracing.
See sample editing page (from Wikipedia corpus, but similar to CORONA). Explanation: You can add to or delete texts from your virtual corpus, or move texts from one virtual corpus to another. You can also rename and delete corpora, temporarily "ignore" corpora (for example, when you're comparing corpora. Finally, you can arrange virtual corpora into user-defined categories (science, religion, sports, etc). You can see what words occur much more in a particular virtual corpus than in the corpus overall. For example, see the keywords from the virtual corpus that is composed of articles dealing with contact tracing. Once you have created a virtual corpus (by keyword or by text metadata), then you can search that set of texts as though it were its own corpus. You can search for matching strings, collocates (nearby words), and retrieve re-sortable concordance (KWIC) lines. To search one of the corpora, just select it from your list of virtual corpora, and then fill out the rest of the search form as you normally would. For example, you can search for the word smartphone in the contact tracing virtual corpus. (Click on the word in the results list, and you will see that all of the occurrences are from your virtual corpus.) If you have created multiple virtual corpora, then you can compare the frequency of a word, phrase, or grammatical construction in these different corpora. Just enter the word or phrase in the search form (as you would do normally), and then select MY CORPORA (try it; must be logged in first -). |