Using Crawlers for Targeted Data Extraction from a Local Multi-File Database

Authors

  • Alexandru ANGHELESCU West University of Timișoara
  • Ciprian-Viorel STUPINEAN Romanian Society of Medical Informatics, 300222 Timișoara, Romania; Department of Computer Science, Faculty of Mathematics and Computer Science, Babeş-Bolyai University, 400084 Cluj-Napoca, Romania https://orcid.org/0009-0005-8268-0074
  • Ariana-Anamaria CORDOȘ Romanian Society of Medical Informatics, 300222 Timișoara, Romania; Department of Public Health, Faculty of Political, Administrative and Communication Sciences, Babeş-Bolyai University, 400084 Cluj-Napoca, Romania https://orcid.org/0000-0003-2853-4058

Keywords:

Crawler, C#, Automated Data Extraction, Email, Data

Abstract

Background: A crawler is a software program used to extract data in an automated manner. This study aimed to demonstrate how a crawler can extract specific data from multiple diverse .pdf files. Methods and Materials: To achieve this objective, a C# .NET9 application was developed, capable of processing a folder (local database) containing specific .pdf files. The application sequentially read each file and extracted relevant information. The ability of the crawler to extract the e-mail addresses of corresponding authors from academic papers was evaluated as a .pdf file may have contained multiple articles. In addition to email addresses, names of corresponding authors were extracted where possible. The PdfPig library was used to access the data since the input data were .pdf files. The output consisted of a CSV file containing all extracted email addresses. The input dataset included 19 books of abstracts and 180 articles. Results: During testing, the application managed to extract 929 email addresses and 77 names. However, due to pattern inconsistencies, name extraction was possible only for articles, not for books of abstracts. Further, evaluation on precision and accuracy was performed. While there was only 1 line extracted that did not contain emails out of the 880 lines, 34.55% of them needed corrective actions. In 213 instances text was attached to the e-mails (e.g. country names: Spain, Israel etc. or other words like keyword or abstract), country prefixes were attached in 156 cases and in 2 lines there additional full stops at the beginning or end of the e-mail. Discussion: Crawlers can be effective in extracting specific data from big databases of files simultaneously. In medical research, this ability can have an impact on productivity when dealing with data collection for research purposes. On the other hand, it poses a risk when personal information, e-mails in this case, become accessible for malicious purposes. Future work should explore compliance with data protection regulations, such as GDPR, and methods to ensure responsible data use. Conclusion: Besides the usefulness of crawlers in extracting email addresses, they prove their efficiency while dealing with the data gathering part of the research. By using a crawler, the researcher may be able to save some time, just by not dealing with the data extraction part of the study, while dealing with a large database of studies to cite.

Downloads

Published

02.05.2025

How to Cite

1.
ANGHELESCU A, STUPINEAN C-V, CORDOȘ A-A. Using Crawlers for Targeted Data Extraction from a Local Multi-File Database. Appl Med Inform [Internet]. 2025 May 2 [cited 2025 May 17];47(Suppl. 1):S52. Available from: https://ami.info.umfcluj.ro/index.php/AMI/article/view/1116

Issue

Section

Special Issue - RoMedINF