453
Views
0
CrossRef citations to date
0
Altmetric
Computer Programming, System and Service

Semi-automatic staging area for high-quality structured data extraction from scientific literature

ORCID Icon, ORCID Icon, ORCID Icon, ORCID Icon, , , ORCID Icon, ORCID Icon, ORCID Icon & ORCID Icon show all
Article: 2286219 | Received 08 Sep 2023, Accepted 16 Nov 2023, Published online: 14 Dec 2023

References

  • Foppiano L, Castro PB, Suarez PO, et al. Automatic extraction of materials and properties from superconductors scientific literature. Sci Technol Adv Mater. 2023;3(1):2153633. doi: 10.1080/27660400.2022.2153633
  • Oliveira ON, Oliveira MJ. Materials discovery with machine learning and knowledge discovery. Front Chem. 2022;10:10. doi: 10.3389/fchem.2022.930369
  • Jain A, Ong SP, Hautier G, et al. Commentary: the materials project: a materials genome approach to accelerating materials innovation. APL Mater. 2013;1(1):011002. doi: 10.1063/1.4812323
  • Curtarolo S, Setyawan W, Hart GL, et al. Aflow: an automatic framework for high-throughput materials discovery. Comput Mater Sci. 2012;58:218–12. https://www.sciencedirect.com/science/article/pii/S0927025612000717
  • Draxl C, Scheffler M. The nomad laboratory: from data sharing to artificial intelligence. J Phys Mater. 2019;2(3):036001. doi: 10.1088/2515-7639/ab13bb
  • Pratheepan T. Global publication productivity in materials science research: a scientometric analysis. Indian J Inf Sources Serv. 2019 Feb;9(1):111–116. https://ojs.trp.org.in/index.php/ijiss/article/view/583
  • Blokhin E, Villars P. The PAULING FILE project and materials platform for data science: from big data toward materials genome. Cham: Springer International Publishing; 2018. p. 1–26. doi: 10.1007/978-3-319-42913-7_62-1
  • Ishii M, Sakamoto K. Structuring superconductor data with ontology: reproducing historical datasets as knowledge bases. Sci Technol Adv Mater. 2023;3(1):2223051. doi: 10.1080/27660400.2023.2223051
  • Roter B, Dordevic S. Predicting new superconductors and their critical temperatures using machine learning. Phys C. 2020;575:1353689. doi: 10.1016/j.physc.2020.1353689
  • Stanev V, Oses C, Kusne A, et al. Machine learning modeling of superconducting critical temperature. Npj Comput Mater. 2017;4(1):4. doi: 10.1038/s41524-018-0085-8
  • Tran H, Vu TN. Machine-learning approach for discovery of conventional superconductors. arXiv preprint arXiv:221103265. 2022.
  • Konno T, Kurokawa H, Nabeshima F, et al. Deep learning model for finding new superconductors. Phys Rev B. 2021;103(1):014509. doi: 10.1103/PhysRevB.103.014509
  • Klie JC, Bugert M, Boullosa B, et al. The INCEpTION platform: machine-assisted and knowledge-oriented interactive annotation. In: Proceedings of the 27th International Conference on Computational Linguistics: System Demonstrations; Santa Fe, New Mexico; 2018. p. 5–9. https://www.aclweb.org/anthology/C18-2002
  • Nakayama H, Kubo T, Kamura J, et al. Doccano: text annotation tool for human; 2018. Software. https://github.com/doccano/doccano
  • Ong SP, Richards WD, Jain A, et al. Python materials genomics pymatgen: a robust open-source python library for materials analysis. Comput Mater Sci. 2013 2;68:314–319. doi: 10.1016/j.commatsci.2012.10.028
  • Kononova O, Huo H, He T, et al. Text-mined dataset of inorganic materials synthesis recipes. Sci Data. 2019 Oct;6(1):203. doi: https://doi.org/10.1038/s41597-019-0224-1
  • Tkachenko M, Malyuk M, Holmanyuk A, et al. Label studio: data labeling software; 2020-2022. Open source software https://github.com/heartexlabs/label-studio
  • Foppiano L, Dieb S, Suzuki A, et al. Supermat: construction of a linked annotated dataset from superconductors-related publications. Sci Technol Adv Mater: Methods. 2021;1(1):34–44. doi: 10.1080/27660400.2021.1918396
  • Beltagy I, Lo K, Cohan A. SciBERT: a pretrained language model for scientific text. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP); Nov; Hong Kong, China: Association for Computational Linguistics; 2019. p. 3615–3620. https://aclanthology.org/D19-1371
  • DeLFT contributors. Delft https://github.com/kermitt2/delft. 2018–2023.
  • Kirkpatrick J, Pascanu R, Rabinowitz NC, et al. Overcoming catastrophic forgetting in neural networks. CoRr.2016;abs/1612.00796. http://arxiv.org/abs/1612.00796
  • Contributors G. Grobid https://github.com/kermitt2/grobid 2008 — 2023.