Brought to you by:
Paper The following article is Free article

Classification of Astrophysics Journal Articles with Machine Learning to Identify Data for NED

, , , , , , , , , and

Published 2022 January 11 © 2022. The Astronomical Society of the Pacific. All rights reserved.
, , Citation Tracy X. Chen et al 2022 PASP 134 014501 DOI 10.1088/1538-3873/ac3c36

1538-3873/134/1031/014501

Abstract

The NASA/IPAC Extragalactic Database (NED) is a comprehensive online service that combines fundamental multi-wavelength information for known objects beyond the Milky Way and provides value-added, derived quantities and tools to search and access the data. The contents and relationships between measurements in the database are continuously augmented and revised to stay current with astrophysics literature and new sky surveys. The conventional process of distilling and extracting data from the literature involves human experts to review the journal articles and determine if an article is of extragalactic nature, and if so, what types of data it contains. This is both labor intensive and unsustainable, especially given the ever-increasing number of publications each year. We present here a machine learning (ML) approach developed and integrated into the NED production pipeline to help automate the classification of journal article topics and their data content for inclusion into NED. We show that this ML application can successfully reproduce the classifications of a human expert to an accuracy of over 90% in a fraction of the time it takes a human, allowing us to focus human expertise on tasks that are more difficult to automate.

Export citation and abstract BibTeX RIS

1. Introduction

The NASA/IPAC Extragalactic Database (NED 1 ) is a comprehensive database of multi-wavelength data for extragalactic objects that provides a systematic, ongoing fusion of information integrated from hundreds of large sky surveys and tens of thousands of research publications. The database content includes both published information for distinct astrophysical objects such as nomenclature, positions, redshifts, photometry, diameters, images and spectra, and derived quantities such as estimates of Galactic extinction, reference-frame velocity corrections to redshifts, sizes and luminosities (Helou & Madore 1988; Mazzarella 2020).

In order to better support scientists, educators, and observatories in planning new observations, performing data analysis and modeling, making new discoveries, and preparing publications of new results, it is essential to keep the database content as current as possible. To achieve this, NED routinely reviews, extracts, and integrates data from high impact journals including Astronomy & Astrophysics (A&A 2 ), the Astronomical Journal (AJ 3 ), the Astrophysical Journal (ApJ 4 ), the Astrophysical Journal Letters (ApJL 5 ), the Astrophysical Journal Supplement (ApJS 6 ), the Monthly Notices of the Royal Astronomical Society (MNRAS 7 ) and Letters 8 , and Nature 9 . Until recently, a human expert would first scan the articles and determine if the topic of an article is appropriate to be included in the database, i.e., if the article contains data for extragalactic objects. When the topic is considered relevant, the expert would further review the content of the article to determine which of the different types of data should be extracted and integrated into the database. This is a very labor-intensive process. In 2018 alone, NED reviewed a total of 14,123 journal articles, which translated to an average of ∼270 articles per week. Given the increasing volume of publications and the growing volume and complexity of data therein, it is neither efficient nor sustainable to rely only on human effort to continue this work. We therefore explored applying machine learning (ML) technique to classify literature with the goals of automating and accelerating the process of selecting journal articles and data for inclusion into NED while maintaining a high level of consistency with human classifications.

This paper is organized as follows. Section 2 describes the method and software we selected for our ML application, and how we implemented it in our use case. Section 3 discusses the sample curation and training for selecting journal articles that are of interest to NED, the performance of our best model validated against the testing sets, and its deployment into the NED production pipeline. In Section 4, we extend the approach to further identify the different data content in the articles, e.g., astrometry, photometry and redshift data. We summarize our study and the results in Section 5.

2. Method

2.1. Algorithm

ML algorithms are generally divided into two groups: supervised and unsupervised. The classification of NED relevant articles and data calls for a supervised learning algorithm as we want to learn the mapping of features in journal articles to the decisions of human experts over the last 10–20 years, and use the learned relationship to predict the decisions on new, previously-unseen articles. We expected to have a high dimensionality in feature space (∼1 M features) as the input consists of fairly large documents (∼0.1–1 M bytes per article), and a fairly large number of them (∼ several thousand articles). Hence, we chose to focus on leveraging existing open-source software that is well-reputed to build models over text data. Additionally, the ML solution and feature extractor need to operate in a pipeline/background mode without graphical interfaces, remote connections, or human interaction. After careful study, we chose the Stanford Classifier (SC 10 ) developed by the Stanford Natural Language Processing Group. The SC implements a Maximum Entropy classifier with Conjugate-Gradient Descent optimization (MaxECG), which is described in detail in Manning & Klein (2003). The program includes an easy-to-use command-line interface with several tunable algorithms, as well as a feature extraction and mapping/encoding component. The software is well supported with a strong user base, has an open source licensing under the GNU General Public License 11 , and has also been updated and improved over the years. For the work described here, we used version 3.9.2, which was released in 2018. It is worth pointing out that our goal was not to perform a computer science experiment in search of a "best" solution, but to apply a well-tested package that has been successfully used in other application domains. For example, the Max Entropy (MaxEnt) SC is one of four leading supervised text classification algorithms that were compared for sentiment analysis of movie recommendations in Twitter feeds (Armentano et al. 2015).

2.2. Overview of ML Processing for NED

The ML process involves two major cycles: the learning (or training) cycle and the prediction cycle. Figure 1 illustrates the general article classification process in NED.

Figure 1.

Figure 1. Machine learning journal article classification process in NED.

Standard image High-resolution image

In the learning cycle, models are constructed and evaluated using training sets and test sets that are independent subsets of the input data. Our input data are the HTML files of articles published in the astronomical journals (see §1), acquired from the journal websites via their ADS links, and the labels placed by human experts which we adopt as ground truth. The Stanford Classifier is used to perform feature characterization, identification and extraction on the training sets. Features can be words, parts of words, stems of words , characters, groups of characters (letter-n-grams 12 ), and groups of words (word-n-grams). We experimented with different feature models and found the "Chris4" word-shape 13 , with leading or trailing 2–5 letter-n-grams, as well as whole-words, performed the best. We also learned that feature extraction is the most expensive computational activity, about 10 times the demand of the ML training itself. We therefore extract features from each article in a training set, combine them into an intermediate catalog of feature sets, and reuse them in subsequent training exercises. Positive or negative weights of the features for each class are learned in the training process based on the priors. Table 1 presents some examples of the features with the weights learned in the training process. It is noticeable that the weights of the features are sensitive to the letter cases. In the examples here, the words "galaxy", "Galaxy", "GALAXY" and "GALaxy" are shown to have very different weights.

Table 1. Example Features and Weights Extracted from a Training Sample

FeatureWeightNote
galaxy0.2875Whole word
galax0.3084Leading 5 letter-n-grams
xies0.3237Trailing 4 letter-n-grams
GALAXY0.0957Whole word
Galaxy0.0370Whole word
GALaxy0.0001Whole word

Download table as:  ASCIITypeset image

Multiple candidate models are trained and examined by varying the choice of the training and test sets, from which we select the best-performing model. The primary evaluation metrics we use in accessing the performance of the models are the classification accuracy and recall. Accuracy is defined as the fraction of the correct predictions (i.e., consistent with the decisions by the human experts) to the total number of predictions made:

Equation (1)

Here Tp represents the true positives where the model correctly predicts the positive class, Tn represents the true negatives where the model correctly predicts the negative class, Fp represents the false positives where the model incorrectly predicts the positive class, and Fn represents the false negatives where the model incorrectly predicts the negative class. Recall is defined as the fraction of actual positives that are identified correctly:

Equation (2)

The initial requirement for our ML application was to minimize the human effort and accelerate the rate of classification while maintaining 90% consistency with human classifications. By looking at the instances where the ML approach differ from the human decisions, we found that classifications by our human expert, though very accurate, typically have an error rate of ∼1%. This led us to set our goal as maximizing the efficiency of classification at a high accuracy rate (∼90%) and an acceptable recall rate (∼80%–90%).

In the prediction cycle, we extract features from new, previously unseen articles and feed them into the selected model. The model then returns a predicted label for the new article with a probability indicating the likelihood of the prediction. In Section 3 and Section 4, respectively, we describe specific applications of this methodology to successfully identify articles appropriate for NED and to classify the types of data contained withing the papers (object positions, photometry, redshifts, etc.), helping expedite the data ingestion from the literature into the database.

We note that, because the classifier is trained on past publications and used to predict future publications (domain shift), it is better to accept a slightly higher rate of false positives because all the positives will be further inspected by human experts in order to clean, reformat, and validate the data for ingestion into the database. During that process, the false positives will be effectively removed and not contaminate the end result. This still saves substantial time that used to be spent examining the true negatives, allowing human resources to work on the more difficult-to-automate data validation and extraction tasks. We are aware that NED will inevitably suffer from some incompleteness due to the false negatives. However, the main goal for this ML approach is to maximize efficiency while keeping as current as possible with the increasing flow of data from the literature. Some omissions are acceptable with an appreciable gain of processing more data into NED more quickly overall, which will be a greater benefit to a larger number of users. In addition, authors/users are generally quick to inform us if there is an important omission, and the missed data can always be added afterwards.

3. NED-Appropriate Topics Classification

Our first implementation of this ML approach is to identify if the topic of an article is relevant to NED or not, i.e., NED-Appropriate Papers (NAPs) or Not NED-Appropriate papers (NNAs). Historic data show that among the journals we reviewed, about a quarter to a third of the articles of articles published each year contain extragalactic objects and therefore need to be selected and further categorized for inclusion into the NED database.

3.1. Training and Testing

We selected training and testing sets from 16 yr (2003–2018) of AJ articles. There were 6987 articles in total, with 1984 NAPs and 5003 NNAs based on human classifications. We acquired the HTML files of these articles through the ADS API. 14 We then extracted the features from the raw HTMLs, i.e., no stripping of the HTML formatting, as this saves considerable time in processing. For each article we extracted on the order of a million features, and it took a few days of CPU time in total to generate all the feature sets for the 6987 articles.

From the 16 yr of AJ articles, we created 100 training/testing iteration samples, each containing the 1984 NAPs and 1984 randomly selected articles from the 5003 NNAs. Equal number of positives and negatives were chosen to avoid an imbalanced data set, which may result in better performance on the larger class and worse performance on the smaller class, and may also affect the use of regular accuracy as a evaluation metric (Baron 2019). For each iteration, we processed 10 folds of the SC-MaxECG algorithm on a randomly selected 90% of the articles for training and 10% for testing. Essentially, this created and tested 1000 models with different choices of training and test sets. The resulting classifier models typically selected 720,000 features of significance and scored an accuracy ranging from 89% to 97%, and a recall ranging from 92% to 96% when evaluated on the test sets. The 100 rounds with 10 folds of ML computations used 22 CPU hours in total.

Table 2 provides an example output generated by one of the ML models. For each article (represented by its ADS Bibcode), the classifier outputs a label for the article, and a probability of this article belonging to this class. For example, the classifier decided that article 2012A&A...537A..29R is 90.8% likely to be an NAP and article 2017ApJ...841L..16V is 99.8% likely to be an NNA. We adopted the default probability threshold of 0.5 in SC as our decision cutoff, i.e., an article would be labeled NAP if p(NAP) > 0.5 and NNA if p(NNA) > 0.5. We found that this threshold of p = 0.5 corresponds closely to the optimum point where the difference between the true positive rate and the false positive rate is maximized for the NAP/NNA classifier. This is illustrated in Figure 2 using the Receiver Operating Characteristic (ROC) curve (Fawcett 2006; Tharwat 2021) of the best trained model, where the true positive rate is plotted against the false positive rate and the optimum point is shown to occur at a threshold of p = 0.5082. Additionally, the predictions made with this cutoff threshold do meet our goal for accuracy and recall.

Figure 2.

Figure 2. Receiver Operating Characteristic (ROC) curve for the NAP/NNA classifier implemented in NED. Different probability points are labeled as well as the optimum point (OP) where the classifier discriminates the maximum true positive rate for the least false positive rate. Here the OP occurs at a true positive rate of 0.9319 and a false positive rate of 0.06473, which corresponds to a probability threshold of p = 0.5082, very close to the default p = 0.5 that we adopted.

Standard image High-resolution image

Table 2. Example Output from a NAP/NNA Classifier Model

BibcodeLabelProbability a
2012A&A...537A..29RNAP0.908
2014MNRAS.444.1862SNAP0.992
2018MNRAS.475.5513VNAP1.000
2018ApJ...868L..15SNNA0.998
2015ApJ...799..237UNNA0.882
2016ApJ...821L..12DNAP1.000
2013ApJ...763...19RNAP1.000
2016MNRAS.459.3432MNNA0.559
2018MNRAS.473.5237KNNA0.702
2011MNRAS.411.2426GNNA0.999
2017ApJ...841L..16VNNA0.998

Note.

a This is the likelihood of an article to be classified as one type or the other. We require a probability value to be >0.5 in order to label an article to be NED-Appropriate Papers (NAP) or Not NED-Appropriate papers (NNA).

Download table as:  ASCIITypeset image

During the testing, we found that 455 articles were classified opposite to NEDs historical choice (i.e., false positives and false negatives) in at least one out of the 100 iterations, and 137 articles classified opposite to NEDs choice in all 100 iterations. This can mean there are mistakes in the prior classification, but as the ML is "learning" from the priors, it more likely indicates that there is an inconsistency (in the vocabulary, syntax, or both), or simply not enough information (distinguishing terms/features).

When applying the best trained model on the entire corpus of AJ articles, we were able to reproduce the human classifications to an accuracy of 97% and a recall of 94%. We further validated the model on other journals from the same period of time (2003-2018), and were able to reproduce the priors to an accuracy of 93% for ApJ and ApJL, 91% for ApJS, and 90% for MNRAS and MNRAS Letters, individually, with a recall of >91% for all these journals.

We also experimented, but not elaborately, with the alternative Naive Bayes (NBy) classifier available in the SC package. NBy classifiers are known to compete favorably with MaxEnt classifiers given large training sets, i.e., ≫10,000 items in the corpus; MaxEnt classifiers perform best at general text classification with small corpora, i.e., ≲10,000 (Potts 2011; Yuret 2010; Pang et al. 2002). We evaluated the performance using computational speed, accuracy and recall. We found that MaxEnt does become slightly more expensive in computation as the number of training items increases. However, at the scale of the NED literature set (1000's), this is easily overcome by computing facilities at hand. The NBy method is problematic for small training sets, as accuracy and recall fall off quickly. In the few experiments conducted, the accuracy of the NBy classifier ranged from 40% to 70%, with no noticeable improvement in speed.

Other areas that were examined include: (1) Using a simpler word-shape, and/or not using letter-n-grams. This reduces the number of features substantially (factor of ∼2/3) and therefore computation. However, accuracy and recall suffered immensely, falling by more than 30%; (2) Using only defined keywords, known by experience to be related to the selection. We confirmed that the classifiers generally did select features that include forms of the NED keywords historically used by humans in skimming articles for significant information. It was in this process where we found that letter-case was significant in weighting such features.

3.2. Deployment to Production

The best performing model, dubbed SC-classifier-AJ-PF1984-064 (trained on 1984 positive priors and like number of negative priors of AJ articles), was selected for deployment in the NED production pipeline and has been used to automate the identification of NED-appropriate papers among new articles published in the astronomical journals since early 2020.

Figure 3 illustrates the acquisition, evaluation, and loading process in the NED production, with the ML classifier highlighted. The tool takes a list of files containing the pre-downloaded article texts (in HTML obtained from publishers' on-line archives), constructs the input needed to inform the feature extractor, invokes the feature extractor and the classifier with the configured model on the extracted features, and outputs the report and log files. The former records only the positive classifications that will be used as input to NED's existing loader program. The latter records all the classifications and the computed likelihood values, which are reserved for future testing and tuning.

Figure 3.

Figure 3. The acquisition, evaluation and loading process in the NED production pipeline, with the ML classifier highlighted.

Standard image High-resolution image

Since the performance of the selected model may degrade due to the changing of input data (e.g., shifts in research topics or journal formats can introduce new features), periodic model retraining is necessary to stay current with the astronomical literature. We set a periodic training schedule of once per year. The new training sets are from human classifications on one or two volumes per journal every 6 months, and the collected false positives detected in the data preparation stage. We deployed our model into NED production in early 2020 and collected our first batch of new human classifications in the summer of 2020. We compared the labels assigned by human experts with the predictions from the ML classifier, and present our results in Table 3. It is shown that although the model was trained on pre-2019 AJ articles, the performance is still strong: the overall accuracy and recall are both at 92% for articles published in 2020. We plan to incorporate these new training sets in 2021/2022. Our short-term focus, however, is to expand the scope of the ML application to cover classifying types of data in the NAP articles.

Table 3. Performance of the Topic Classifier on 2020 Articles

 ApJS 249A&A 638A&A 639MNRAS 494AJ 160ApJ 895Overall
     Issue 1, 2  
Number of articles36178160476962001,146
Accuracy97%94%96%89%100%87%92%
Recall89%100%98%96%100%71%92%

Download table as:  ASCIITypeset image

4. Data Content Classification

The success of the ML application to automate and accelerate the selection of NED relevant articles from astronomical journals encouraged us to further extend the approach to include the data content classifications, i.e., what types of data an article has to offer. The general methodology was the same as in our previous implementation, but was now applied to NED "data-frames" for astrometry (positions), kinematics (redshift), photometry, and in-literature acknowledgement of NED, in addition to the appropriateness for an article to be included in NED (NAPs).

The types of data NED routinely integrates into the database contain more than the selected few above. These categories were chosen mostly due to the fact that they have the largest sample size in the database. Additionally, positions, photometry and redshift are among the most fundamental properties NED keeps for an object, and acknowledgement is an important metric used as a basis for productivity and impact by data archives such as NED.

4.1. Training and Testing

We initially assembled a sample containing articles from all the NED core journals (A&A, AJ, ApJ, ApJL, ApJS, MNRAS and MNRAS Letters) between the year 2004 and 2019. The NED database management system (DBMS) was used to randomly select training and test sets, based on desired criteria (NAPs or NNAs; contains specific data types or not). However, during the training/testing process, we found that articles published before 2011 were more prone to classification error. Upon examination, we found that two issues were involved. One was a significant format change in some journals publication around 2010 (e.g., A&A and MNRAS), which created some variations in features that compromised the performance of the model. The other was due to the fact that there were less data in NED for the years before 2011, compared to the years after. Since our goal is to have the model predict future instances accurately, it is therefore important to make sure the data the model is trained on are similar to what is expected in the near future. Hence, we revised our sample selection to include only the articles from 2011 through 2019. Table 4 shows the total size of the final sample, together with the number of positive priors in each of the data categories to be modeled. It is worth pointing out that in the sample selection the NAP and acknowledgement classifications are independent, i.e., an article can acknowledge the use of NED without providing any data on objects, hence is not a NAP. The classifiers on different data types (astrometry, photometry, and kinematics) are only trained on samples that are NAP-condition-positive; that is, for example, the astrometry training set (both positives and negatives) is taken from the NAP subset.

Table 4. The Sample

CategoryNumber of articles
Total83,755
NAP21,820
Astrometry6660
Photometry9040
Kinematics7248
Acknowledgement5501

Note. This table gives the total number of articles in A&A, AJ, ApJ, ApJL, ApJS, MNRAS and MNRAS Letters from 2011 to 2019, and the number of positive priors in each of the data categories.

Download table as:  ASCIITypeset image

For each of the 5 categories (NAP, acknowledgement, astrometry, photometry, and kinematics), we processed 20 iterations of the SC-MaxECG on randomly selected subsets from the entire sample. Each subset contains 90%–95% of the positives and an equal number of negatives in each category for training, and 5%–10% of the positives and negatives for testing. All together, we trained 100 classifier models, and it took 144 CPU hours (not including feature extraction time). Since feature extraction of individual articles and the concatenation of different feature sets from all the articles in the sample are very time-consuming, on the order of days in CPU time, we chose to do it once at the beginning and used the same combined feature file in the training of all five classifiers.

The resulting models from the 20 iterations have an accuracy range of 91%–93%, 96%–98%, 87%–93%, 88%–92% and 88%–92% for NAP, acknowledgement, astrometry, kinematics and photometry classifiers, separately. Table 5 presents, for the best-performed model of each classifier, the total number of articles and the number of positive priors in the subset, together with the accuracy and recall when evaluated against the respective testing set. The accuracy rates are at >92% for all the classifiers, and the recall rates are at >93% except for the acknowledgement category. In the case of acknowledgements, there was a shortage of positive priors due to a shift in the definition of positives for this category. "Acknowledgement of NED" was initially defined as any mention of NED in the article text using ADS keyword extraction, or classification by ADS as having cited NED, or classification by NED team through visual inspection. The training sample was selected using the initial definition. However, we later revised the definition to only consider the NED team confirmed acknowledgements or citation, which resulted in a smaller training set and subsequently a somewhat limited yet still sufficient accuracy and recall in testing.

Table 5. Input and Performance of the Best Classifier in Each Data Type Category

ClassifierNAPAcknowledgementAstrometryRedshiftPhotometry
Sample Size a 41,45810,45013,65413,75617,176
Positives20,7048866,3176,8918,578
Accuracy93%98%93%92%92%
Recall93%80%96%93%94%

Note.

a This is the size of the subset used to train the classifier, containing 90%–95% of the positive priors and an equal number of negative priors randomly selected from the total sample in each category for training.

Download table as:  ASCIITypeset image

4.2. Deployment to Production

We integrated the above-mentioned best-performing models to production at the beginning of 2021. Our initial approach was to remove the articles identified as NNA from the input of the data content classifiers, so that the data type classification would only occur on articles that were already found to be NAP. However, we found that it was considerably faster to run the entire input set (including both the NAPs and NNAs) through all classifiers, and generate recommendations/classifications based on the combined output. We therefore augmented the program to run all five classifiers simultaneously on each article, extract results and output them in the required format of the NED loader program. During the initial deployment, we randomly selected a few volumes from different journals and had our human experts label the articles in order to assess the performance of the models. Table 6 provides a few examples of the results of classification accuracy based on post-classification human verification. The total number of articles as well as the number of NAPs in each volume are listed here to caution that these are all small samples, hence the accuracy and recall rates are very sensitive to the mis-classification of even a few articles. For example, ApJ 914 contains only four articles with NED acknowledgement in the entire volume. The ML approach mis-classified three, resulting in a low recall of 25%. Nonetheless, the classifiers are shown to perform well outside of their training domain (year 2011–2019) on these articles published in 2021. As of September 2021, this ML classifier has been in full operations for 8 months, and has reduced ∼80% of the labor for the content classification task. This translated to a team gain of 0.4 Full-Time Equivalent (FTE) which was then focused on tasks that cannot be automated.

Table 6. Performance of the New Data-type Classifier on Some 2020 and 2021 Journal Articles

Journal NAPAcknowledgementAstrometryRedshiftPhotometry
A&A 645Accuracy96%99%91%93%91%
(160:38)Recall89%88%77%91%82%
A&A 651Accuracy93%96%97%94%93%
(138:36)Recall83%62%87%67%77%
ApJ 904Accuracy96%99%94%94%91%
(233:54)Recall91%75%100%92%83%
ApJ 905Accuracy96%97%93%93%93%
(215:60)Recall92%60%96%96%95%
ApJ 906Accuracy95%99%95%92%90%
(151:49)Recall90%80%100%100%94%
ApJ 914Accuracy97%98%95%93%94%
(185:59)Recall92%25%94%88%100%
ApJS 252Accuracy97%97%94%97%81%
(32:11)Recall91%75%88%89%100%
ApJS 255Accuracy97%93%87%93%90%
(30:12)Recall100%71%83%80%100%
MNRAS 499Accuracy91%98%90%93%90%
(451:151)Recall97%79%83%87%91%

Note. The total number of articles and the number of NAPs in each volume are given in the parenthesis under the journal volume name.

Download table as:  ASCIITypeset image

As has been discussed previously in Section 3.2, these classifiers are trained on past publications and are being applied to publications that are new, with information/features that sometimes have not been seen before. Therefore, it is important to monitor the trend of accuracy and recall and be prepared to refresh the models with more current publications. Currently, we are keeping the periodic retraining schedule of once per year. The schedule may be changed based on the trend of the performance, which we are assessing every 6 months using human classified labels on randomly selected journal articles.

5. Conclusion

We explored the use of ML techniques to process articles in the astronomical literature, selecting those that are appropriate for inclusion into NED, and categorizing the different types of data contained in the articles for further extraction and ingestion. We showed that this automated process can reproduce the historic classifications by human experts of articles relevant to NED and selected data types (astrometry, photometry, kinematics and acknowledgement) to an accuracy of 92% or greater. It is worth noting that this ML approach does not completely remove human involvement in the process. Human expertise and learning are needed for marginal cases that are not resolved by existing capabilities. Further, the trained model is applied to a domain-shifted data set for predictions. Newly-emerged data content, variations in formats and representations can all contribute to deteriorating performance over time. Hence, there needs to be a continuing education program for retraining and updating the classifier models with new literature, which will require new labels identified by human experts periodically.

Overall, we have demonstrated the feasibility and value of using ML to reduce the human effort in this historically labor-intensive task while maintaining a high level of accuracy and consistency. This ML application on both subject and content classifications has been in operations since January 2021, and has gained the team about 0.4 FTE, i.e., the equivalent of 4 person-years of labor over a decade of NED operations, that we have refocused on tasks that have little or no hope of being handled with artificial intelligence or machine learning anytime soon, such as analyzing and understanding the data content well enough to clean, reformat, apply or validate metadata to prepare input to the database.

We want to acknowledge the past and continuing work of the Natural Language Processing Group at Stanford University, without which this application would have been much more difficult. We are also grateful to the American Astronomical Society Journals and IOP Publishing, the Oxford University Press, and EDP Sciences for their support in providing extensive and on-going access to their publications. We thank the anonymous referee for the critical review and comments which resulted in substantial improvements to the article. This work was funded by the National Aeronautics and Space Administration through a cooperative agreement with the California Institute of Technology.

Facilities: ADS - , NED. -

Software: Stanford Classifier v3.9.2, The Stanford Natural Language Processing Group, https://nlp.stanford.edu/software/classifier.shtml

Footnotes

Please wait… references are loading.