Skip to main content

Advertisement

Evaluation of cutoff policies for term extraction

Article metrics

  • 1388 Accesses

  • 3 Citations

Abstract

Background

This paper presents a policy to choose cutoff points to identify potentially relevant terms in a given domain. Term extraction methods usually generate term lists ordered according to a relevance criteria, and the literature is abundant to offer different relevance indices. However, very few studies turn their attention to how many terms should be kept, i.e., to a cutoff policy.

Methods

Our proposed policy provides an estimation of the portion of this list which preserves a good balance between recall and precision, adopting a refined term extraction and tf-dcf relevance index.

Results

A practical study was conducted based on terms extracted from a Brazilian Portuguese corpus, and the results were quantitatively analyzed according to a previously defined reference list.

Conclusions

Even thou different extraction procedures and different relevance indices could brought a different outcome, our policy seems to deliver a good balance for the method adopted in our experiments and it is likely to be able to be generalized to other methods.

Background

Automatic identification of relevant terms for a given domain is an extremely important task for a myriad of natural language processing applications. For instance, any ontology learning effort is doomed to fail if the concept identification step has a poor performance. In fact, any other steps to automatically build an ontology rely on the concept identification [14].

Also, text categorization applications can be much more effective if a good relevant term identification is available.

An important part in the process of identifying relevant terms is the extraction of terms and the computation of their frequencies of use as term relevance index. For the term extraction itself, many software tools are available [5, 6]. These tools usually offer high-quality extraction, regardless being implemented on the basis of linguistic or statistical approaches. As for relevance indices, many theoretical formulations are available [711] and it is safe to assume that a reliable relevance-based rank of extracted terms is not difficult to obtain.

Unfortunately, even assuming a nearly perfect ranked list of terms, it is still difficult to decide how many relevant terms are sufficiently relevant to cover the domain essential terms. Even the high-quality extraction procedures deliver some terms that maybe present in a corpus, but are not really relevant to the domain. The computation of a relevance index, and the ranking according to this index, is likely to push such irrelevant terms to the last positions of the ranked list.

Once the extracted terms are ranked according to the relevance, one needs to decide how many extracted terms are relevant enough. Such decision is made by adopting a cutoff point, i.e., choosing a point in the ranked list to discard all extracted terms below this point and to keep the terms from this point up. A cutoff point is present in virtually all term extraction endeavors, but almost all referred scientific works [4, 1222] apply cutoff points based solely on guess and trial.

That being said, there are three traditional approaches to guess a cutoff point: absolute, threshold-based, and relative cutoff points. However, none of the referred work try to establish a generic policy to actually choose a point to cutoff the ranked term list. The major contribution of our paper is, therefore, to propose a policy to choose an actual cutoff point to a given term list, i.e., to offer researchers and practitioners an algorithmic solution to choose the discard point, instead of leaving this choice to a guess.

It is important to call the reader’s attention about the difference between an approach to choose a cutoff point and a cutoff point policy. An approach to choose a cutoff point is a method to determine how to cut off a term list, e.g., to keep the n top ranked terms of a list. A cutoff point policy is a more ambitious task, since it consists in the definition of a way to determine a cutoff point and the actual point where the list should be cut off, e.g., to keep the n top ranked terms of a list and a formula to estimate the value of n.

The application scope of the term selection policy proposed in this paper implies the following assumptions:

  • Given a domain corpus, a list of terms is extracted, and, for the purpose of this paper, we will focus on the relevant terms identification on a list of terms extracted from a domain corpus, even thou most of the definitions made here can be generalized to lists of terms built from virtually any other sources than domain corpora;

  • A relevance index is computed (usually based on some sort of frequency in the corpus) assigning to each term a numeric value proportional to its domain relevance; hence, the term list can be ranked according to this index;

  • Some of the extracted terms are sufficiently relevant and should be considered representative of the domain, those terms will be referred as valid terms;

  • Some of the extracted terms, on the contrary, are not sufficiently relevant and should be discarded; those terms will be referred as invalid terms;

  • Each extracted term is either valid or invalid.

In consequence, our policy will represent a contribution to practical cases where these assumptions are true. However, our proposed policy can still be beneficial even to practical cases where not all these assumptions can be verified. For a start, the definition of valid and invalid terms is frequently arguable. While some terms are clearly classified as valid or invalid, others terms remain in a grey area and they need to be arbitrarily classified into one group or another.

Another issue is the accuracy of the relevance index, since not necessarily all valid terms will be ranked in higher positions than all invalid terms. That being said, giving the efficacy of the relevance indices [23], it is only reasonable to assume that the relevance index-based ranking is strongly correlated to the terms’ actual relevance. Note that this assumption is probably as secure as to assume that terms that are not present in the list of extracted terms can never be considered relevant.

As a final consequence of these assumptions, the identification of the relevant terms of a domain consists in choosing a cutoff point to the ranked list aiming to have the valid terms above this point and the invalid terms below. In other words, it is necessary to find out the point that separates the valid and invalid terms.

By applying precision and recall metrics, which are common in information retrieval domain [24], it becomes clear that to maximize the precision, one only needs to consider very few or possibly just the most relevant terms and to discard all others. Similarly, it is easy to maximize the recall by considering all extracted terms, i.e., not discarding any. Besides those two trivial approaches, the challenge is to balance precision and recall keeping a set of extracted terms with most, ideally all, relevant terms and very few, or ideally none, irrelevant terms.

In this context, the challenge is to find out a trade-off between maximizing the precision (cutting off many terms) and maximizing the recall (cutting off few terms). Therefore, this paper’s goal is to propose a term selection policy as the definition of a cutoff point to the list of extracted terms in order to balance precision and recall with respect to a reference list indicating valid terms, i.e., terms judge as relevant to the domain. To fulfill this objective, a practical study is developed through the application of three traditional cutoff point approaches (absolute, threshold, and relative) to a term list extracted from a corpus.

The result of these approaches for several parameters is compared to a reference list (gold standard) assumed to be the full set of valid terms for the corpus. Such experiments allow the quantitative comparison of traditional cutoff policies. Consequently, a new and more effective policy is proposed.

It is important to stress that in all approaches, it is possible to achieve better and worse precision-recall balances, but our claim is that using traditional approaches, there is no simple way to determine whereas a given threshold would result in a reasonably balanced cutoff point. Therefore, it is not our goal to show that our policy offers a better reduction of extracted term list sizes than traditional approaches. Instead, we want to propose a way to estimate a cutoff point that takes into account more than any traditional approach alone.

This section presents the corpora employed in this paper’s experiments, the reference list corresponding to the relevant terms of the corpora, a brief description of the extraction tool and relevance index computation and, finally, the formulation of traditional information retrieval metrics. Next, the three traditional cutoff policy approaches: “Absolute”, “Threshold-based” and “Relative” and their results over a practical examples of different arbitrarily chosen cutoff points are presented.

The corpora and reference lists

This paper’s experiments were conducted over five different Portuguese domain corpora. The first one is the Pediatrics corpus (Ped) [25] composed of 281 texts extracted from the Jornal de Pediatria, a Brazilian scholar journal on Pediatrics.

The more important particularity of this corpus is the availability of a reference list of relevant bigrams and trigrams. This reference list was manually built in the context of TEXTCC research project, conducted by Maria José Finatto group at the Federal University of Rio Grande do Sul (UFRGS).

This list is composed of 1,534 bigrams and 2,660 trigrams chosen among the extracted terms that were found at least four times in the corpus. Terminology students and domain specialists (pediatricians) issued their opinion whereas a term was representative of the Pediatrics domain, and only terms with all participant agreement were kept. This list is publicly available at Finatto’s group web page: http://www6.ufrgs.br/textecc/.

The other four domain corpora employed in this paper’s experiments were developed in Lucelene Lopes Ph.D. thesis [3]. These corpora cover the topics of stochastic modeling (SM), data mining (DM), parallel processing (PP), and geology (Geo). All these corpora are composed not only by scientific articles but also by larger documents as Ph.D. thesis and M.Sc. dissertations. Table 1 describes numerical characteristics for the five mentioned corpora.

Table 1 Corpora characteristics

An important information about these corpora is the fact that they were carefully revised manually to prevent frequent textual problems such as char encoding and inclusion of non-phrasal information. Unfortunately, the four additional corpora (SM, DM, PP, and Geo) do not have previously established reference lists. Therefore, no cutoff numerical comparison with a gold standard was possible to these corpora. However, they were needed as contrastive corpora to compute the relevance index for the pediatrics corpus extracted terms (see the “The relevance index” section). Additionally, we also apply the traditional approaches and our proposed policy over the terms extracted from these contrastive corpora in order to observe the sheer impact on the term list size.

The reader interested in further information about these corpora construction and its characteristics can find detailed information in a previous publication [26]. As for the corpora data itself, the reader may freely download all these corpora from the PUCRS Natural Language Processing group at http://www.inf.pucrs.br/peg/lucelenelopes/ll_crp.html; therefore, all experiments conducted here could be replicated or compared with other approaches.

The extraction tool

All experiments with cutoff points conducted in this paper were based on the list of extracted terms from the five aforementioned corpora. Each of these corpora were submitted to the E χ ATO lp software tool [6], a term extractor based on linguistic knowledge [20].

The full extraction procedure includes a previous annotation of part-of-speech (POS) and grammatical information performed by the PALAVRAS parser [27]. The annotated texts are submitted to E χ ATO lp extraction software tool that basically identifies all noun phrases and submits these noun phrases to a set of linguistic heuristics to either avoid unappropriated terms [28].

As result, a rather large set of terms is extracted from each corpora. Figure 1 depicts how many terms classified according to their number of words (unigrams, bigrams, etc. —the terms indicated as M-grams have 10 or more words) were extracted from each corpora. In this picture, we present the number of noun phrases identified originally by PALAVRAS (left hand side bars—indicated with an asterisk) and the number of actually extracted terms after applying E χ ATO lp heuristics (right hand side bars).

Fig. 1
figure1

Total number of extracted terms

The use of E χ ATO lp heuristics provides a considerable re-arrangement of extracted terms, including a large number of unigrams and a significant reduction of terms with more than one word (bigrams, trigrams, and so on). As the reader may verify in a specific publication about the employed extraction [28], such re-arrangement is highly beneficial to quality of the extracted terms.

Table 2 presents numerically the number of extracted terms produced by E χ ATO lp for each corpora. Note that this number of terms reflects the number of terms after applying all E χ ATO lp linguistic heuristics, i.e., it corresponds to the number depicted in the right hand side columns of Fig. 1.

Table 2 Total number of extracted terms

While the total number of extracted terms represented in Table 2 stands for every term extracted from the corpora, a more important number of terms is expressed in Table 3. In this table, the number of terms express how many distinct terms were extracted. As expected, there are much more repetition in the smaller n-grams.

Table 3 Number of distinct extracted terms

The relevance index

Among the myriad of options to use as relevance index, the experiments conducted in this paper consider the term frequency, disjoint corpora frequency (tf-dcf) index [23] as estimation of each term relevance. This index is based on the use of contrasting corpora, i.e., it considers that each term relevance is directly proportional to the number of occurrences of the term in the target corpus and inversely proportional to the number of occurrences in other corpora.

Numerically, the tf-dcf index of a term t belonging to a corpus c, considering a set of contrasting corpora $\mathcal {G}$ is expressed as:

$$ \textit{tf-dcf}_{t}^{\,(c)} = \frac{\textit{tf}_{t}^{\;(c)}}{\prod\limits_{\forall g \in \mathcal{G}} 1 + \log \left(1 + \textit{tf}_{t}^{\,(g)} \right)} $$
((1))

where $\textit {tf}_{t}^{\;(c)}$ express the absolute frequency, i.e., the number of occurrences, of term t in corpus c.

The main characteristic of the tf-dcf index is to express the relevance of a term as its absolute frequency in the corpus, if this term does not appear in the contrasting corpora. The terms of the target corpus that also appears in the contrasting corpora are penalized geometrically according to its absolute frequency in each contrasting corpora. Consequently, a term that appears in many contrasting corpora tends to suffer a great penalization.

In fact, compared to other relevance indices [911], tf-dcf is often more effective because it penalizes terms according not only to the number of occurrences in contrasting corpora but also to the number of contrasting corpora in which it appears.

Nevertheless, the tf-dcf index mathematical structure keeps a semantic associated to the absolute frequency, since the numeric value of tf-dcf of a given term is bounded by its absolute frequency. As a consequence, the numeric value of the tf-dcf remains quite intuitive.

Information retrieval metrics

All comparisons of term lists made in this paper compute three information theory metrics that are frequently employed in a myriad of scientific works dealing with information retrieval. The metrics are precision (P), recall (R), and F-measure (F) [24].

These metrics consider the existence of two sets of different origin, e.g., two term lists. One of these sets is considered reliable, e.g., a reference list, denoted $\mathcal {RL}$ . The other set is the one to be compared to the reliable one, e.g., a extracted term list, denoted $\mathcal {EL}$ .

The precision of the set $\mathcal {EL}$ with respect to the reference set $\mathcal {RL}$ is expressed by the ratio between the number of identified terms that are also present in the reference list, i.e., the cardinality of the intersection between $\mathcal {EL}$ and $\mathcal {RL}$ and the number of identified terms, i.e., the cardinality of $\mathcal {EL}$ . Identified terms, in this context, mean terms that were extracted from the corpus and considered relevant to the target domain. Tuus, precision expresses the percentage of terms correctly considered as relevant to the domain, i.e., the percentage of true positives with respect to the all positive terms. Numerically:

$$ P = \frac{\mid \mathcal{RL} \cap \mathcal{EL} \mid}{\mid \mathcal{EL} \mid} $$
((2))

Recall index expresses the ratio between the number of identified terms that are also present in the reference list and the number of terms in the reference list, i.e., the cardinality of $\mathcal {RL}$ . In such way, recall expresses the percentage of reference list terms that were identified, i.e., the percentage of true positives with respect to all true terms. Numerically:

$$ R = \frac{\mid \mathcal{RL} \cap \mathcal{EL} \mid}{\mid \mathcal{RL} \mid} $$
((3))

As mentioned, considering as relevant very few extracted terms tends to improve the precision at the expense of recall reduction. On the contrary, considering as relevant a large number of extracted terms tends to increase the recall and reduce the precision. Consequently, it is necessary to find a way to evaluate the balance between precision and recall. F-measure numerically express such balance as the harmonic average between precision and recall, i.e., F-measure value is always a value between P and R, and as the difference between P and R grows, the value of F tends to be closer to the smaller one.

$$ F = \frac{2 \times P \times R}{P + R} $$
((4))

The use of those metrics is largely spread in several domains [2931]. The Natural Language Processing area and, specially in term extraction, many works numerically support their contribution using such metrics [7, 3235].

Absolute approach

The simplest and more popular way to apply cutoff points is to choose an arbitrary number of terms to consider as relevant. This approach is called absolute cutoff point, and its difficulty resides in estimating how many terms should be considered, since there is not a reasonable way to estimate which number will be adequate.

Despite all that, several works in the literature use this approach and arbitrarily choosing cutoff points with empirical basis [1216, 1822]. It is worth noting that often this approach is accompanied by the manipulation of terms in separated lists according to the number of words, i.e., lists of unigrams, lists of bigrams, and so on.

Such kind of manipulation seems to make sense since the relevance indices tend to behave quite differently according to the number of words in each term. For instance, the absolute frequency of unigrams tends to be much higher than the absolute frequency of bigrams. In consequence, a bigram that appears, for example, 20 times, tends to be much more relevant than a unigram that also appears 20 times. For instance, considering the Geology corpus previously mentioned, the more frequent unigram was bacia (“basin” in English) with 2390 occurrences, while the more frequent bigram was matéria orgânica (“organic matter” in English) with only 430 occurrences.

Experiments with absolute cutoff points to terms extracted from pediatrics corpus

By applying the absolute cutoff points method to the bigrams and trigrams of the previously mentioned pediatrics corpus, we obtain precision, recall, and F-measure as shown in Fig. 2 for the considered reference lists mentioned in the “The corpora and reference lists” section (corresponding numeric values are given in Table 4). For these experiments, we arbitrarily choose absolute cutoff points from 100 to 3500 terms with an increment of 100.

Fig. 2
figure2

Precision (P), recall (R), and F-measure (F) using absolute cutoff points

Table 4 Precision (P), recall (R), F-measure (F), and list sizes using absolute cutoff points

The first observation from Fig. 2 is the fact that the precision and recall curves do cross as expected, since the precision drops and recall increases as the absolute cutoff point increases, i.e., the considered extracted terms lists becomes less restrictive. Such crossing occurs nearly at 1500 bigrams and 2700 trigrams, which is not a surprise since the reference lists have, respectively, 1534 bigrams and 2660 trigrams. Obviously, lists with less terms than the reference list will never reach a 100 % recall.

The numeric values presented in Fig. 2 indicate that the best balanced choice was found for 2000 bigrams, with 81 %, and 3300 trigrams, with 84 %. It is interesting to notice that the best F-measure value are not found when the precision and recall curves cross (1500 bigrams and 2700 trigrams). The optimal values of F-measure are found in the situation where the precision and recall values present an inflection point. This is particularly clear observing Fig. 2 bigrams curve for recall and its flexion point at 2000.

It is observable that as the list sizes grow, the precision values drop slower than the recall values increase. In fact, the optimal F-measure values seem to be associated to the near stagnation of the recall values. For bigrams, it occurs with a list of 2000 terms, when the recall is around 93 %. This is a curious behavior, since it means that nearly 500 bigrams more than the reference list must be considered. This represents that approximately 25 % of the considered bigrams are invalid (out of reference list). Hence, the precision value stays around 72 %.

For trigrams, an analogous behavior is observed, since approximately 600 additional trigrams must be considered (3300 trigrams for 2660 in the reference list) in order to deliver the higher F-measure value. However, both precision and recall values are slightly higher (76 and 94 %) than for bigrams; thus, the F-measure is a little higher for trigrams.

Threshold-based approach

Another popular approach to establish cutoff points is to choose a threshold for the relevance index. Once again, it is an arbitrary choice of which index value is considered high enough to represent a valid term.

A reasonable number of works in the area adopt the threshold method using the simplest relevance index, the absolute frequency. This is the case of the Bourigault and Lame work [13] that assumes that terms occurring at least 10 times in the corpus are relevant enough. Obviously, such choice is greatly dependent on the size of corpus since occurring nine times maybe not relevant to Borrigault and Lame experiment, but it is quite significative for a small corpus 100,000 words.

Additionally, as mentioned before, the number of occurrences also varies considerably according to the number of words. Consequently, a same threshold for unigrams and for bigrams may represent very different levels of restriction.

Some authors [15] claim that cutoff points by threshold values of absolute frequency have a direct relation with the corpus size (number of words). Such intuitive assumption may be useful for many practical cases, but it is a known fact that the number of occurrences of terms in a corpus decreases exponentially [36].

The exact format of exponential decrease may vary according to the extraction method. For instance, terms extracted according to a pure statistical method tend to follow a Zipf law. According to Zipf law formulation, the frequency of a word in a corpus is inversely proportional to its rank. Therefore, the second more frequent word tends to appear the half of times than the more frequent one, and the third more frequent word tends to occur one third of times as the more frequent one [37].

However, linguistic-based extraction process tend to behave differently. For instance, the four more frequent unigrams in the pediatrics corpus are criança (“child”) with 2055 occurrences, estudo (“study”) with 1529 occurrences, paciente (“patient”) with 1505 occurrences, and doença (“disease”) with 784 occurrences.

Therefore, it is impractical to consider a generic formula to estimate a reasonable threshold considering solely the corpus size. Because of that, our experiments were conducted considering some arbitrary values of threshold.

Experimenting threshold-based cutoff points to terms extracted from pediatrics corpus

To conduct the experiments using the threshold-based cutoff points over a list os ranked terms, it is necessary to choose threshold values in accordance with the relevance index. It is important to recall that the tf-dcf index has a similar semantic to the absolute frequency; therefore, we can choose threshold values having in mind the same semantics, i.e., for threshold purposes, the tf-dcf index can be understood as the plain number of occurrences.

Figure 3 shows precision, recall, and F-measure for different threshold values from 0 to 15 for bigrams and trigrams of the pediatrics corpus (numeric values given in Table 5). In these experiments, threshold 0 means to consider all extracted terms, i.e., do not discard extracted terms. As the threshold value increases, the list becomes more restrictive.

Fig. 3
figure3

Precision (P), recall (R), and F-measure (F) using threshold-based cutoff points

Table 5 Precision (P), recall (R), F-measure (F), and list sizes using threshold-based cutoff points

Analogously to the analysis of the results for absolute cutoff points, Fig. 3 starts showing that once again, the precision and recall curves cross each other. However, the threshold values seem insufficient in terms of granularity to locate the inflection points as were observed in Fig. 2.

The optimal F-measure values were found for threshold 3 (bigrams) and for threshold 2 (trigrams) with values as good as the best absolute cutoff points (2000 bigrams and 3300 trigrams). The adjacent threshold values show significant drops in the F-measure value, illustrating the poor granularity of the threshold-based cutoff points.

In such way, the use of a threshold-based cutoff point alone seems to be too gross to find a balance between precision and recall. Even though, the definition of an arbitrary threshold is often found in many term extraction works [13, 15].

Relative approach

A rarely found alternative [17] to the absolute and threshold approaches is the use of relative cutoff points. In this approach, the number of terms to be considered is a fixed proportion of the extracted terms. As in the other approaches, it is still necessary to arbitrarily choose the percentage of terms to consider, but the biggest advantage of the relative cutoff approach is to offer an option that is independent of the extraction process or corpora size, even thought it is still dependent on the size of the extracted term list.

Specifically, this approach consists in the selection of a percentage of the top ranked extracted terms. Consequently, this approach only needs to choose a numeric value between 0 and 100 %, being those close to 0 % very restrictive cutoff points (delivering high precision) and those close to 100 % (delivering high recall).

Experimenting relative cutoff points to terms extracted from pediatrics corpus

As for the other traditional approaches, we conduct our experiments adopting different relative values for cutoff points. Specifically, we investigate the metrics for cutoff points considering from 1 % up to 30 % of bigrams and trigrams extracted from the pediatrics corpus.

Figure 4 depicts the precision, recall, and F-measure values obtained for these relative cutoff points (numeric values at Table 6). The first observation from the metrics in Fig. 4 is that the granularity of results is, like for the absolute cutoff point results, good enough to observe the inflections of precision and recall curves.

Fig. 4
figure4

Precision (P), recall (R), and F-measure (F) using relative cutoff points

Table 6 Precision (P), Recall (R), F-measure (F) and List Sizes using relative cutoff points

The more balanced values were obtained for 13 % of the bigrams and 18 % of the trigrams. In fact, we notice that for relative cutoff points from 8 to 15 % (for bigrams) and from 14 to 22 % (for trigrams), a F-measure value of at least 75 % is achieved. Thus, it is possible to consider that a relative cutoff point around 14 and 15 % was a good trade-off for bigrams and trigrams.

Evidently, these results may not generalize to other extracted term lists, or even other reference lists. Nevertheless, we see no reason to consider that these precision-recall balanced region would not appear in other experiments. It is important to notice that unlike the other traditional approaches, relative cutoff points seem to deliver a larger reasonable balanced region. For instance, absolute and threshold-based cutoff points do not have cutoff values where both bigrams and trigrams have F-measure above 75. Observing Table 4, bigrams cutoff points deliver F-measure above 75 % for absolute values between 1300 and 2300, while trigrams F-measure above 75 % is found for values from 2500 through 3500. Observing Table 5, bigrams cutoff points deliver F-measure above 75 % for values 3 and 4, while for trigrams only threshold 2 delivers F-measure above 75 %.

Pushing the analysis a little bit further than just the numerical analysis of the metrics over bigrams and trigrams in comparison to the reference list, we observe that larger terms, i.e., terms with four or more words are likely to have low values of the relevance index. Specifically, we observe that even the 15 % top ranked 4-grams contain terms that were found one single time in the corpus. Such rare terms are even more likely to be found in the top ranked positions for lists as the number of words of the terms increase. For instance, among the 1567 extracted 8-grams, only 38 terms were found more than once in the corpus, and for the 1093 extracted 9-grams, the number of terms found at least twice in the corpus was only 30.

Consequently, it seems that even relative cutoff points cannot be considered alone as an effective way to estimate a precision-recall balanced point for all extracted term lists, despite the finding of reasonable values of F-measure (above 75 %) for values of 14 and 15 % in bigrams and trigrams lists.

Ultimately, these experiments were our motivation to look for an alternative way to estimate cutoff points without being limited to the traditional approaches of absolute, threshold-based, and relative values. Hence, we suggest the adoption of a policy to choose a cutoff point for virtually any extracted term list.

Methods

The results for the traditional approaches presented in the previous section have shown some cutoff point values delivering balanced percentages of precision and recall, i.e., higher F-measures values. However, the goal of this paper is more than finding out good parameters for the bigrams and trigrams of the pediatrics corpus. Our goal is to propose a policy that has potential to be applied to any extracted term lists delivering satisfactory results.

Finding a single policy, regardless of the approach, does not seem possible since even for the two lists submitted to numerical experiments, it was difficult to define one single situation delivering optimal F-measure values to bigrams and trigrams. Therefore, we propose a hybrid method to estimate a reasonable cutoff point, and considering the experiments conducted in this paper, we will assume that an F-measure above or equal to 75 % is sufficiently balanced. Specifically, we propose a policy combining threshold-based and relative approaches.

It is important to clarify that the use of absolute cutoff points is not feasible, since in a generalization context, it is not possible to suggest any arbitrary fixed number of terms to consider. The approaches based on relative and threshold cutoff points are naturally flexible to the problem, since the size of the extracted list is sufficiently dependent on the corpus size.

Following the same reasoning, it is not enough to consider a same threshold to all extracted term lists; however, it seems reasonable to adopt a small threshold to avoid considering terms that would drop significantly the precision values. Therefore, we propose to adopt as the basic step of the proposed policy to consider only terms that have a tf-dcf value equal or superior to 2.

This choice of a threshold 2 for tf-dcf is conservative, since for bigrams of the pediatrics corpus, a threshold 3 would be a better choice. It is worthy to remember that for the trigrams of pediatrics corpus, the choice of threshold 3 would discard an important number of valid trigrams.

A more reliable part of the proposed policy is the adoption of a relative cutoff point of 15 % of the extracted terms. As seen before, a 15 % relative cutoff point would be a good trade-off for bigrams and trigrams of the pediatrics corpus. It is important to remind that the purpose of our proposed policy is to offer a practical way to estimate a good cutoff point to a given extracted term list. Such estimation was not possible using only traditional approaches as absolute or threshold-based approaches.

Even for relative cutoff points, where the experiments with values 14 and 15 % delivered reasonably high F-measure values (above 75 %) for bigrams and trigrams, we had observe that terms with more words (4-grams and up), therefore likely to be less relevant, were encountered in the top ranked 15 % extracted terms. Consequently, we suggest the policy considering, initially, the use of a 15 % relative cutoff point, followed by a restriction accepting only terms with a relevance index above or equal to 2, i.e., a composition with a threshold-based cutoff point.

Summarizing our cutoff policy, we propose to:

  • consider the extracted terms as a set of lists with terms with the same number of words, duly ranked by tf-dcf relevance index;

  • discard 85 % of the extracted terms of each list, keeping the top 15 % terms with the higher tf-dcf values;

  • discard among the remaining 15 % of each list, the terms with a tf-dcf value inferior to 2.

Results and discussion

Applying this policy to the extracted terms of the five corpora described in the “The corpora and reference lists” section, we obtain the number of terms presented in Table 7. In order to have a graphic impression of the proposed cutoff point application over the five corpora, Fig. 5 depicts the reduction in terms of the number of remaining terms. The left hand side bars depict the number of terms as outputted by the extraction (Table 3), while right hand side bars (identified with an asterisk) depict the number of terms considered as relevant to each domain (Table 7).

Fig. 5
figure5

Comparative analysis of the proposed policy application

Table 7 Number of identified relevant terms

As mentioned before, due to a lack of reference lists, we can only observe the reduction of the sheer number of remaining terms for all the extracted lists. It is also noticeable that lists of large terms were subject to reductions below 15 %, since such lists have a large number of terms found only once in their corpora. Nevertheless, for the bigrams and trigrams of the pediatrics corpus, it is possible to compare the remaining terms with the reference lists. Such comparison delivers the following values of precision, recall, and F-measure:

$$P = 62\, \% \;\; R = 94\,\% \;\; F = 75\,\% \;\;\; \textrm{for bigrams} $$
$$P = 77\,\% \;\; R = 79\,\% \;\; F = 78\,\% \;\;\; \textrm{for trigrams} $$

Conclusions

The main purpose of this paper was to propose a policy to estimate a reasonably balanced cutoff point that may be applicable to virtually any list of extracted terms. The first difficulty in such kind of work was to find closely similar work, since much of the available literature dealing with the choice of cutoff points do not focus on alternatives to estimate a fairly balanced cutoff point. In fact, at the authors’ best knowledge, most of authors limit themselves to briefly justify the adoption of a specific value for a traditional cutoff approach. Therefore, we believe that our work is an original contribution by focusing on the alternatives to estimate cutoff points and proposing a policy to be employed by practitioners and researchers that have to face the choice of a cutoff point.

In this paper, several experiments with cutoff points were conducted. For all these experiments, we started with a qualified list of extracted terms, duly ranked with an effective relevance index. Unfortunately, among all corpora employed, only the pediatrics corpus had a reliable and independently developed reference list to be used to the numerical analysis.

In such context, we were able to propose a hybrid method to identify the relevant terms to domain. The results were informally analyzed, and it seems that the considered extracted terms are rather relevant and specific to each domain. However, a formal analysis of these lists is an interesting future work. Such work was not yet carried out, due to the lack of domain specialists. Nevertheless, the practical application of the proposed policy to reduce the extracted terms was empirically approved by reducing with quality the term extracted in applied linguistic projects conducted by TEXTCC research group leaded by Maria José Finatto at the Federal University of Rio Grande do Sul (UFRGS).

Another interesting future work to consider is to look for more sophisticated analysis of the terms, adopting, besides the tf-dcf index, other indications of term relevance. We could gather term information regarding its grammatical role, for instance, terms that are subject of sentence may be more inclined to be relevant. Such kind of initiative would correspond to a multi-valued analysis that can be carried out even using machine learning tools [38]. Among such theoretical tools, it is natural to employ classification methods [3941] to improve the quality of the relevant term selection.

Despite of such ambitious future work, the current state of the policy proposed in this paper is already being employed in several research initiatives for Brazilian Portuguese. However, new controlled experiments would improve the confidence in our policy. Other suggestions of future work include the use of other extraction tools, e.g., [5, 13, 20], or even other relevance indices, e.g., [10, 11, 42]. The main obstacle to these works is the availability of corpora accompanied with reference lists.

It is also important to mention that the experience with a larger set of corpora and reference lists may lead to a revision of the proposed policy. After all, the choice of a good cutoff point is dependent of many factors, as the quality of the extraction procedure, and the quality of relevance ranking of terms. Consequently, the results presented in our paper may not generalize to other extraction efforts, including similar extraction procedures applied to other languages.

At the authors’ best knowledge, there are very few similar works to evaluate the options of cutoff points in general and none for term extraction from Portuguese corpora. Therefore, this paper offers some help to researchers and practitioners to define cutoff points to their own experiments.

References

  1. 1

    Maedche A, Staab S (2001) Ontology learning for the semantic web. IEEE Intell Syst 16(2): 72–79. doi:10.1109/5254.920602.

  2. 2

    Cimiano P (2006) Ontology learning and population from text: algorithms, evaluation and applications. Springer.

  3. 3

    Lopes L (2012) Extração automática de conceitos a partir de textos em língua portuguesa. PhD thesis, PUCRS University - Computer Science Department, Porto Alegre, Brazil.

  4. 4

    Conrado M, Pardo T, Rezende S (2013) A machine learning approach to automatic term extraction using a rich feature set In: Proceedings of the 2013 NAACL HLT Student Research Workshop, 16–23.. Association for Computational Linguistics. http://aclweb.org/anthology/N13-2003.

  5. 5

    Banerjee S, Pedersen T (2003) The design, implementation and use of the Ngram statistics package In: CICLing’03 Proceedings of the 4th international conference on Computational linguistics and intelligent text processing, 370–381.. Springer-Verlag, Berlin, Heildelberg.

  6. 6

    Lopes L, Fernandes P, Vieira R, Fedrizzi G (2009) ExATO lp—an automatic tool for term extraction from Portuguese language corpora In: Proceedings of the 4th Language & Technology Conference (LTC ’09), 427–431.. Faculty of Mathematics and Computer Science of Adam Mickiewicz University.

  7. 7

    Manning CD, Schütze H (1999) Foundations of statistical natural language processing. MIT Press, Cambridge, USA.

  8. 8

    Chung TM (2003) A corpus comparison approach for terminology extraction. Terminology 9(2): 221–246.

  9. 9

    Kit C, Liu X (2008) Measuring mono-word termhood by rank difference via corpus comparison. Terminology 14(2): 204–229.

  10. 10

    Park Y, Patwardhan S, Visweswariah K, Gates SC (2008) An empirical analysis of word error rate and keyword error rate In: INTERSPEECH, 2070–2073.

  11. 11

    Kim SN, Baldwin T, Kan MY (2009) Extracting domain-specific words—a statistical approach. In: Pizzato L Schwitter R (eds)Proceedings of the 2009 Australasian Language Technology Association Workshop, 94–98.. Australasian Language Technology Association, Sydney, Australia.

  12. 12

    Pao ML (1978) Automatic text analysis based on transition phenomena of word occurrences. J Am Soc Inform Sci 29(3): 121–124. doi:10.1002/asi.4630290303.

  13. 13

    Bourigault D, Lame G (2002) Analyse distributionnelle et structuration de terminologie. application a la construction d’une ontologie documentaire du droit. Traitement automatique des langues 43(1): 129–150.

  14. 14

    Milios E, Zhang Y, He B, Dong L (2003) Automatic term extraction and document similarity in special text corpora In: 6th Conference of the Pacific Association for Computational Linguistics, 275–284, Halifax, Nova Scotia, Canada. http://users.cs.dal.ca/~eem/res/pubs/pubs/pacling2003.pdf.

  15. 15

    Wermter J, Hahn U (2005) Paradigmatic modifiability statistics for the extraction of complex multi-word terms In: Proc. of the Conf. on Human Language Technology. HLT ’05, 843–850.. Assoc. for Comput. Ling., Stroudsburg, PA, USA. doi:10.3115/1220575.1220681.

  16. 16

    Yang H, Callan J (2008) Ontology generation for large email collections In: Proceedings of the 2008 International Conference on Digital Government Research. dg.o ’08, 254–261.. Digital Government Society of North America. http://dl.acm.org/citation.cfm?id=1367832.1367875.

  17. 17

    Maynard D, Li Y, Peters W (2008) NLP techniques for term extraction and ontology population In: Proceedings of the 2008 Conference on Ontology Learning and Population, 107–127.. IOS Press, Amsterdam, The Netherlands. http://dl.acm.org/citation.cfm?id=1563823.1563834.

  18. 18

    Lopes L, Vieira R, Finatto MJ, Zanette A, Martins D, Ribeiro Jr LC (2009) Automatic extraction of composite terms for construction of ontologies: an experiment in the health care area. RECIIS 3(1): 72–84.

  19. 19

    Evert S (2010) Google web 1T 5-grams made easy (but not for the computer) In: Proceedings of the NAACL HLT 2010 Sixth Web as Corpus Workshop. WAC-6 ’10, 32–40.. Association for Computational Linguistics, Stroudsburg, PA, USA. http://dl.acm.org/citation.cfm?id=1868765.1868770.

  20. 20

    Lopes L, Oliveira LH, Vieira R (2010) Portuguese term extraction methods: comparing linguistic and statistical approaches In: PROPOR 2010 – International Conference on Computational Processing of Portuguese Language.

  21. 21

    Awawdeh R, Anderson T (2010) Improving search in tag-based systems with automatically extracted keywords. In: Bi Y Williams M. -A. (eds)Knowledge Science, Engineering and Management. Lecture Notes in Computer Science, 378–387.. Springer. doi:10.1007/978-3-642-15280-1_35.

  22. 22

    Ding J, Zhou S, Guan J (2011) miRFam: an effective automatic miRNA classification method based on n-grams and a multiclass SVM. BMC Bioinformatics 12(1): 216. doi:10.1186/1471-2105-12-216.

  23. 23

    Lopes L, Fernandes P, Vieira R (2012) Domain term relevance through tf-dcf In: Proceedings of the 2012 International Conference on Artificial Intelligence (ICAI 2012), 1001–1007.. CSREA Press, Las Vegas, USA.

  24. 24

    van Rijsbergen CJ (1975) Information retrieval. Butterworths, London, UK.

  25. 25

    Coulthard RJ (2005) The application of corpus methodology to translation: the JPED parallel corpus and the Pediatrics comparable corpus. PhD thesis, UFSC, Florianópolis, Brazil.

  26. 26

    Lopes L, Vieira R (2013) Building domain specific parsed corpora in portuguese language In: Proceedings of the X National Meeting on Artificial and Computational Intelligence (ENIAC), 1–12.

  27. 27

    Bick E (2000) The parsing system PALAVRAS: automatic grammatical analysis of portuguese in constraint grammar framework. PhD thesis, Arhus University.

  28. 28

    Lopes L, Vieira R (2012) Improving quality of portuguese term extraction In: PROPOR 2012 – International Conference on Computational Processing of Portuguese Language.

  29. 29

    Boreczky JS, Rowe LA (1996) Comparison of video shot boundary detection techniques. J Electron Imaging 5(2): 122–128.

  30. 30

    Thomas J, Milward D, Ouzounis C, Pulman S, Carroll M (2000) Automatic extraction of protein interactions from scientific abstracts In: Pacific Symposium on Biocomputing, 538–549.

  31. 31

    Fernandes P, Lopes L, Ruiz DDA (2010) The impact of random samples in ensemble classifiers In: SAC’10: Proceedings of the 2010 ACM Symposium on Applied Computing, 1002–1009.. ACM, New York, USA. doi:10.1145/1774088.1774300.

  32. 32

    Witten IH, Moffat A, Bell TC (1999) Managing gigabytes: compressing and indexing documents and images. Morgan Kaufmann, San Francisco.

  33. 33

    Hulth A (2004) Enhancing linguistically oriented automatic keyword extraction In: Proceedings of HLT-NAACL 2004: Short Papers. HLT/NAACL, 17–20.. ACM, New York, USA.

  34. 34

    Lopes L, Vieira R, Finatto MJ, Martins D (2010) Extracting compound terms from domain corpora. J Braz Comput Soc 16: 247–259. doi:10.1007/s13173-010-0020-4.

  35. 35

    da Silva Conrado M, Felippo A, Salgueiro Pardo T, Rezende S (2014) A survey of automatic term extraction for brazilian portuguese. J Braz Comput Soc 20(1): 12. doi:10.1186/1678-4804-20-12.

  36. 36

    Spärck-Jones K (1972) A statistical interpretation of term specificity and its application in retrieval. J Doc 28(1): 11–21. doi:10.1108/eb026526.

  37. 37

    Zipf GK (1935) The psycho-biology of language—an introduction to dynamic philology. Houghton-Mifflin Company, Boston, USA.

  38. 38

    Mitchell T (1997) Machine learning. McGraw-Hill.

  39. 39

    Bauer E, Kohavi R (1999) An empirical comparison of voting classification algorithms: bagging, boosting and variants. Mach Learn 36(1–2): 105–139.

  40. 40

    Lopes L, Scalabrin EE, Fernandes P (2008) An empirical study of combined classifiers for knowledge discovery on medical data bases In: APweb 2008 Workshops (LNCS 4977), 110–121.

  41. 41

    Witten IH, Frank E, Hall MA (2011) Data mining: practical machine learning tools and techniques. 3rd edn. Morgan Kaufmann.

  42. 42

    Wu HC, Luk RWP, Wong KF, Kwok KL (2008) Interpreting TF-IDF term weights as making relevance decisions. ACM Trans Inform Syst 26: 13–11337. doi:10.1145/1361684.1361686.

Download references

Acknowledgements

Lucelene Lopes is funded by CAPES and FAPERGS DOCFIX grant. Renata Vieira is partially funded by CNPq.

Author information

Correspondence to Lucelene Lopes.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

This work was part of LL PhD dissertation and RV was her advisor. LL carried out most of the investigation, implementation, experiments, and writing. RV gave important ideas to the project and also contributed in the writing and revision of this manuscript. Both authors read and approved the final manuscript.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Keywords

  • Cutoff Point
  • Absolute Frequency
  • Reference List
  • Noun Phrase
  • Relevant Term