2020 |
From Publications to Knowledge Graphs (Paper in Conference Proceedings) Constantopoulos, Panos; Pertsas, Vayianos Flouris, Giorgos; Laurent, Dominique; Plexousakis, Dimitris; Spyratos, Nicolas; Tanaka, Yuzuru (Ed.): Information Search, Integration, and Personalization. Revised Selected Papers from 13th International Workshop, ISIP 2019, Heraklion, Greece, May 9–10, 2019., Pages: 18-33, Springer, 2020. @inproceedings{Constantopoulos2020, title = {From Publications to Knowledge Graphs}, author = {Panos Constantopoulos and Vayianos Pertsas}, editor = {Giorgos Flouris and Dominique Laurent and Dimitris Plexousakis and Nicolas Spyratos and Yuzuru Tanaka}, year = {2020}, date = {2020-03-27}, booktitle = {Information Search, Integration, and Personalization. Revised Selected Papers from 13th International Workshop, ISIP 2019, Heraklion, Greece, May 9–10, 2019.}, pages = {18-33}, publisher = {Springer}, abstract = {We address the task of compiling structured documentation of research processes in the form of knowledge graphs by automatically extracting information from publications and associating it with information from other sources. This challenge has not been previously addressed at the level described here. We have developed a process and a system that leverages existing information from DBpedia, retrieves articles from repositories, extracts and interrelates various kinds of named and non-named entities by exploiting article metadata, the structure of text as well as syntactic, lexical and semantic constraints, and populates a knowledge base in the form of RDF triples. An ontology designed to represent scholarly practices is driving the whole process. Rule -based and machine learning - based methods that account for the nature of scientific texts and a wide variety of writing styles have been developed for the task. Evaluation on datasets from three disciplines, Digital Humanities, Bioinformatics, and Medicine, shows very promising performance.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } We address the task of compiling structured documentation of research processes in the form of knowledge graphs by automatically extracting information from publications and associating it with information from other sources. This challenge has not been previously addressed at the level described here. We have developed a process and a system that leverages existing information from DBpedia, retrieves articles from repositories, extracts and interrelates various kinds of named and non-named entities by exploiting article metadata, the structure of text as well as syntactic, lexical and semantic constraints, and populates a knowledge base in the form of RDF triples. An ontology designed to represent scholarly practices is driving the whole process. Rule -based and machine learning - based methods that account for the nature of scientific texts and a wide variety of writing styles have been developed for the task. Evaluation on datasets from three disciplines, Digital Humanities, Bioinformatics, and Medicine, shows very promising performance. |
2018 |
Ontology-Driven Extraction of Research Processes (Paper in Conference Proceedings) Pertsas, Vayianos; Constantopoulos, Panos; Androutsopoulos, Ion Vrandečić, Denny; Bontcheva, Kalina; Suárez-Figueroa, Mari Carmen; Presutti, Valentina; Celino, Irene; Sabou, Marta; Kaffee, Lucie-Aimée; Simperl, Elena (Ed.): The Semantic Web – ISWC 2018. 17th International Semantic Web Conference, Monterey, CA, USA, October 8–12, 2018, Proceedings, Volume: 11136 of the series Lecture Notes in Computer Science Pages: 162-178, Springer, 2018, ISBN: 978-3-030-00670-9. @inproceedings{Pertsas2018b, title = {Ontology-Driven Extraction of Research Processes}, author = {Vayianos Pertsas and Panos Constantopoulos and Ion Androutsopoulos}, editor = {Denny Vrandečić and Kalina Bontcheva and Mari Carmen Suárez-Figueroa and Valentina Presutti and Irene Celino and Marta Sabou and Lucie-Aimée Kaffee and Elena Simperl}, doi = {10.1007/978-3-030-00671-6_10}, isbn = {978-3-030-00670-9}, year = {2018}, date = {2018-09-18}, booktitle = {The Semantic Web – ISWC 2018. 17th International Semantic Web Conference, Monterey, CA, USA, October 8–12, 2018, Proceedings}, volume = {11136}, pages = {162-178}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, abstract = {We address the automatic extraction from publications of two key concepts for representing research processes: the concept of research activity and the sequence relation between successive activities. These representations are driven by the Scholarly Ontology, specifically conceived for documenting research processes. Unlike usual named entity recognition and relation extrac- tion tasks, we are facing textual descriptions of activities of widely variable length, while pairs of successive activities often span multiple sentences. We developed and experimented with several sliding window classifiers using Logistic Regression, SVMs, and Random Forests, as well as a two-stage pipeline classifier. Our classifiers employ task-specific features, as well as word, part-of-speech and dependency embeddings, engineered to exploit distinctive traits of research publications written in English. The extracted activities and sequences are associated with other relevant information from publication metadata and stored as RDF triples in a knowledge base. Evaluation on datasets from three disciplines, Digital Humanities, Bioinformatics, and Medicine, shows very promising performance.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } We address the automatic extraction from publications of two key concepts for representing research processes: the concept of research activity and the sequence relation between successive activities. These representations are driven by the Scholarly Ontology, specifically conceived for documenting research processes. Unlike usual named entity recognition and relation extrac- tion tasks, we are facing textual descriptions of activities of widely variable length, while pairs of successive activities often span multiple sentences. We developed and experimented with several sliding window classifiers using Logistic Regression, SVMs, and Random Forests, as well as a two-stage pipeline classifier. Our classifiers employ task-specific features, as well as word, part-of-speech and dependency embeddings, engineered to exploit distinctive traits of research publications written in English. The extracted activities and sequences are associated with other relevant information from publication metadata and stored as RDF triples in a knowledge base. Evaluation on datasets from three disciplines, Digital Humanities, Bioinformatics, and Medicine, shows very promising performance. |
Ontology-Driven Information Extraction from Research Publications (Paper in Conference Proceedings) Pertsas, Vayianos; Constantopoulos, Panos Méndez, Eva; Crestani, Fabio; Ribeiro, Cristina; David, Gabriel; Lopes, João Correia (Ed.): Digital Libraries for Open Knowledge. TPDL 2018., Volume: 11057 of the series Lecture Notes in Computer Science Pages: 241-253, Springer, 2018, ISBN: 978-3-030-00065-3. @inproceedings{Pertsas2018, title = {Ontology-Driven Information Extraction from Research Publications}, author = {Vayianos Pertsas and Panos Constantopoulos}, editor = {Eva Méndez and Fabio Crestani and Cristina Ribeiro and Gabriel David and João Correia Lopes}, doi = {10.1007/978-3-030-00066-0_21}, isbn = {978-3-030-00065-3}, year = {2018}, date = {2018-09-05}, booktitle = {Digital Libraries for Open Knowledge. TPDL 2018.}, volume = {11057}, pages = {241-253}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, abstract = {Extraction of information from a research article, association with other sources and inference of new knowledge is a challenging task that has not yet been entirely addressed. We present Research Spotlight, a system that leverages existing information from DBpedia, retrieves articles from repositories, extracts and interrelates various kinds of named and non-named entities by exploiting article metadata, the structure of text as well as syntactic, lexical and semantic constraints, and populates a knowledge base in the form of RDF triples. An ontology designed to represent scholarly practices is driving the whole process. The system is evaluated through two experiments that measure the overall accuracy in terms of token- and entity- based precision, recall and F1 scores, as well as entity boundary detection, with promising results.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Extraction of information from a research article, association with other sources and inference of new knowledge is a challenging task that has not yet been entirely addressed. We present Research Spotlight, a system that leverages existing information from DBpedia, retrieves articles from repositories, extracts and interrelates various kinds of named and non-named entities by exploiting article metadata, the structure of text as well as syntactic, lexical and semantic constraints, and populates a knowledge base in the form of RDF triples. An ontology designed to represent scholarly practices is driving the whole process. The system is evaluated through two experiments that measure the overall accuracy in terms of token- and entity- based precision, recall and F1 scores, as well as entity boundary detection, with promising results. |
2016 |
MORe : A micro-service oriented aggregator (Paper in Conference Proceedings) Gavrilis, Dimitris; Nomikos, Vagelis; Kravvaritis, Konstantinos; Angelis, Stavros; Papatheodorou, Christos; Constantopoulos, Panos Volume: 672 of the series Communications in Computer and Information Science Springer International Publishing, 2016, ISBN: 978-3-319-49157-8. @inproceedings{Gavrilis2016, title = {MORe : A micro-service oriented aggregator}, author = {Dimitris Gavrilis and Vagelis Nomikos and Konstantinos Kravvaritis and Stavros Angelis and Christos Papatheodorou and Panos Constantopoulos }, url = {http://www.springer.com/gp/book/9783319491561}, doi = {10.1007/978-3-319-49157-8}, isbn = {978-3-319-49157-8}, year = {2016}, date = {2016-11-22}, issuetitle = {Metadata and Semantics Research 10th International Conference, Proceedings}, volume = {672}, publisher = {Springer International Publishing}, series = {Communications in Computer and Information Science}, abstract = {Metadata aggregation is a task increasingly encountered in many projects involving data repositories. The small number of specialized software for this task indicates that in most cases customized software is used to perform aggregation, which in turn relates to the highly complex tasks and architectures involved. In this paper, the metadata and object repository aggregator (MORe) is presented, which has been effectively used in numerous projects and provides an easy and flexible way of aggregating metadata from multiple sources and in multiple formats. Its flexible and scalable architecture exploits cloud technologies and allows storing content into different storage systems, defining workflows dynamically and extending the system with external services. One of the most important aspects of MORe is its curation/enrichment services which allow curation managers to automatically apply and execute enrichment plans employing enrichment micro-services in order to aggregated data.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Metadata aggregation is a task increasingly encountered in many projects involving data repositories. The small number of specialized software for this task indicates that in most cases customized software is used to perform aggregation, which in turn relates to the highly complex tasks and architectures involved. In this paper, the metadata and object repository aggregator (MORe) is presented, which has been effectively used in numerous projects and provides an easy and flexible way of aggregating metadata from multiple sources and in multiple formats. Its flexible and scalable architecture exploits cloud technologies and allows storing content into different storage systems, defining workflows dynamically and extending the system with external services. One of the most important aspects of MORe is its curation/enrichment services which allow curation managers to automatically apply and execute enrichment plans employing enrichment micro-services in order to aggregated data. |
Integrating data for archaeology (Paper in Conference Proceedings) Gavrilis, Dimitris; Fihn, Johan; Sebastian,; Olsson, Olof; Afiontzi, Eleni; Felicetti, Achille; Niccolucci, Franco 22nd Annual Meeting of EAA , Pages: 339-340, Saulius Jokuzys Publishing-Printing House, 2016. @inproceedings{Gavrilis2016c, title = {Integrating data for archaeology}, author = {Dimitris Gavrilis and Johan Fihn and Sebastian and Olof Olsson and Eleni Afiontzi and Achille Felicetti and Franco Niccolucci}, url = {http://eaavilnius2016.lt/wp-content/uploads/2016/08/Book-Abstract-A4_08-23_net.pdf}, year = {2016}, date = {2016-07-30}, booktitle = {22nd Annual Meeting of EAA }, pages = {339-340}, publisher = {Saulius Jokuzys Publishing-Printing House}, abstract = {In the past years, infrastructure projects in the Archaeology domain have focused on data aggregation in order to bring to the end users the vast amount of information gathered from various organizations and stakeholders. The typical processes found in a data aggregation infrastructure include: ingestion, normalization, transformation and validation processes that mainly focus on the homogenization and cleaning of heterogenous data. A portal is usually employed to present this information to the end users and is met with limited success due to the vast information contained. In order to increase the quality of services that are provided to end users, the European funded project Ariadne (http://www.ariadne-infrastructure.eu/) aims at integrating this data by modelling the underlying domain and providing the technical framework for automatic integration of heterogeneous resources. The heart of the infrastructure lies in the underlying domain model: Ariadne Catalog Data Model (ACDM), a DCAT derived model which models a large number of entities such as Agents, Language resources, datasets, collections, reports, services, databases, etc. With the help of a of micro-service oriented architecture and a set of powerful enrichment micro-services all aggregated data are transformed into XML and RDF, annotated over subject, space and time with the help of AAT, Geonames and Perio.do thesauri (thus establishing a common reference) and interlinked with each other based on their structural or logical relationiships. The data integration services can mine for links among resources, link them together and against language resources such as vocabularies. Complex records can be split into their individual components, represented, enriched and stored separately while maintaining their identity using semantic linking. Each integrated resource is assigned a URI and published to: a) Virtuoso RDF Store in RDF which provides a SPARQL interface b) to Elastic Search in JSON which provides a powerful indexing mechanism that can help present and associate resources accurately in real-time. This approach can provide developers and creative industries with the means to create innovative applications and mine information from the RDF store. End users ranging from simple visitors to domain researchers can access this data through the infrastructure’s portal which is capable of hiding the complexity of this plethora of data, filter the results using a plethora of filters and present connected resources in a way that can help guide the user instead of confusing him/her. The technical infrastructure has been developed using various programming languages such as Java, PHP, Javascript, it is distributed spanning multiple virtual machines and brings together different established technologies and components. Both the technical infrastructure and the portal will be presented and demonstrated. }, howpublished = {In 22nd Annual Meeting of the European Association of Archaeologists: Conference Abstracts}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } In the past years, infrastructure projects in the Archaeology domain have focused on data aggregation in order to bring to the end users the vast amount of information gathered from various organizations and stakeholders. The typical processes found in a data aggregation infrastructure include: ingestion, normalization, transformation and validation processes that mainly focus on the homogenization and cleaning of heterogenous data. A portal is usually employed to present this information to the end users and is met with limited success due to the vast information contained. In order to increase the quality of services that are provided to end users, the European funded project Ariadne (http://www.ariadne-infrastructure.eu/) aims at integrating this data by modelling the underlying domain and providing the technical framework for automatic integration of heterogeneous resources. The heart of the infrastructure lies in the underlying domain model: Ariadne Catalog Data Model (ACDM), a DCAT derived model which models a large number of entities such as Agents, Language resources, datasets, collections, reports, services, databases, etc. With the help of a of micro-service oriented architecture and a set of powerful enrichment micro-services all aggregated data are transformed into XML and RDF, annotated over subject, space and time with the help of AAT, Geonames and Perio.do thesauri (thus establishing a common reference) and interlinked with each other based on their structural or logical relationiships. The data integration services can mine for links among resources, link them together and against language resources such as vocabularies. Complex records can be split into their individual components, represented, enriched and stored separately while maintaining their identity using semantic linking. Each integrated resource is assigned a URI and published to: a) Virtuoso RDF Store in RDF which provides a SPARQL interface b) to Elastic Search in JSON which provides a powerful indexing mechanism that can help present and associate resources accurately in real-time. This approach can provide developers and creative industries with the means to create innovative applications and mine information from the RDF store. End users ranging from simple visitors to domain researchers can access this data through the infrastructure’s portal which is capable of hiding the complexity of this plethora of data, filter the results using a plethora of filters and present connected resources in a way that can help guide the user instead of confusing him/her. The technical infrastructure has been developed using various programming languages such as Java, PHP, Javascript, it is distributed spanning multiple virtual machines and brings together different established technologies and components. Both the technical infrastructure and the portal will be presented and demonstrated. |
2015 |
A one-class approach to cardiotocogram assessment (Paper in Conference Proceedings) Gavrilis, Dimitris; Nikolakopoulos, George; Georgoulas, George 37th Annual Conference of the IEEE Engineering in Medicine and Biology Society, Pages: 518-521, IEEE, 2015. @inproceedings{Gavrilis2015, title = {A one-class approach to cardiotocogram assessment}, author = {Dimitris Gavrilis and George Nikolakopoulos and George Georgoulas}, url = {http://emb.citengine.com/event/embc-2015/paper-details?pdID=4250 http://www.dcu.gr/wp-content/uploads/2016/10/A-one-class-approach-to-cardiotocogram-assessment-EMBC2015v2.pdf}, year = {2015}, date = {2015-01-01}, booktitle = {37th Annual Conference of the IEEE Engineering in Medicine and Biology Society}, pages = {518-521}, address = {IEEE}, crossref = {DBLP:conf/embc/2015}, abstract = {Cardiotocogram (CTG) is the most widely means for the assessment of fetal condition. CTG consists of two traces one depicting the Fetal Heart Rate (FHR), and the other the Uterine Contractions (UC) activity. Many automatic methods have been proposed for the interpretation of the CTG. Most of them rely either on a binary classification approach or on a multiclass approach to come up with a decision about the class that the tracing belongs to. This work investigates the use of a one-class approach to the assessment of cardiotocograms building a model for the healthy data. The preliminary results are promising indicating that normal traces could be used as part of an automatic system.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Cardiotocogram (CTG) is the most widely means for the assessment of fetal condition. CTG consists of two traces one depicting the Fetal Heart Rate (FHR), and the other the Uterine Contractions (UC) activity. Many automatic methods have been proposed for the interpretation of the CTG. Most of them rely either on a binary classification approach or on a multiclass approach to come up with a decision about the class that the tracing belongs to. This work investigates the use of a one-class approach to the assessment of cardiotocograms building a model for the healthy data. The preliminary results are promising indicating that normal traces could be used as part of an automatic system. |
Measuring Quality in Metadata Repositories (Paper in Conference Proceedings) Gavrilis, Dimitris; Makri, Dimitra Nefeli; Papachristopoulos, Leonidas; Angelis, Stavros; Kravvaritis, Konstantinos; Papatheodorou, Christos; Constantopoulos, Panos Kapidakis, Sarantos; Mazurek, Cezary; Werla, Marcin (Ed.): Research and Advanced Technology for Digital Libraries, 19th International Conference on Theory and Practice of Digital Libraries, TPDL 2015, Poznań, Poland, September 14-18, 2015, Proceedings, Volume: 9316 of the series Lecture Notes in Computer Science Pages: 56-67, 2015. @inproceedings{Gavrilis2015, title = {Measuring Quality in Metadata Repositories}, author = {Dimitris Gavrilis and Dimitra Nefeli Makri and Leonidas Papachristopoulos and Stavros Angelis and Konstantinos Kravvaritis and Christos Papatheodorou and Panos Constantopoulos}, editor = {Sarantos Kapidakis and Cezary Mazurek and Marcin Werla}, url = {http://link.springer.com/chapter/10.1007%2F978-3-319-24592-8_5#page-1}, doi = {10.1007/978-3-319-24592-8_5}, year = {2015}, date = {2015-01-01}, booktitle = {Research and Advanced Technology for Digital Libraries, 19th International Conference on Theory and Practice of Digital Libraries, TPDL 2015, Poznań, Poland, September 14-18, 2015, Proceedings}, volume = {9316}, pages = {56-67}, series = {Lecture Notes in Computer Science}, abstract = {The need for good quality metadata records becomes a necessity given the large quantities of digital content that is available through digital repositories and the increasing number of web services that use this content. The context in which metadata are generated and used affects the problem in question and therefore a flexible metadata quality evaluation model that can be easily and widely used has yet to be presented. This paper proposes a robust multidimensional metadata quality evaluation model that measures metadata quality based on five metrics and by taking into account contextual parameters concerning metadata generation and use. An implementation of this metadata quality evaluation model is presented and tested against a large number of real metadata records from the humanities domain and for different applications.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The need for good quality metadata records becomes a necessity given the large quantities of digital content that is available through digital repositories and the increasing number of web services that use this content. The context in which metadata are generated and used affects the problem in question and therefore a flexible metadata quality evaluation model that can be easily and widely used has yet to be presented. This paper proposes a robust multidimensional metadata quality evaluation model that measures metadata quality based on five metrics and by taking into account contextual parameters concerning metadata generation and use. An implementation of this metadata quality evaluation model is presented and tested against a large number of real metadata records from the humanities domain and for different applications. |
Discovering the Topical Evolution of the Digital Library Evaluation Community (Paper in Conference Proceedings) Papachristopoulos, Leonidas; Kleidis, Nikos; Sfakakis, Michalis; Tsakonas, Giannis; Papatheodorou, Christos Metadata and Semantics Research - 9th Research Conference, 9th Research Conference, MTSR 2015, Manchester, UK, September 9-11, 2015, Proceedings, Volume: 544 of the series Communications in Computer and Information Science Pages: 101-112, 2015. @inproceedings{Papachristopoulos2015, title = {Discovering the Topical Evolution of the Digital Library Evaluation Community}, author = {Leonidas Papachristopoulos and Nikos Kleidis and Michalis Sfakakis and Giannis Tsakonas and Christos Papatheodorou}, url = {http://link.springer.com/chapter/10.1007%2F978-3-319-24129-6_9}, year = {2015}, date = {2015-01-01}, booktitle = {Metadata and Semantics Research - 9th Research Conference, 9th Research Conference, MTSR 2015, Manchester, UK, September 9-11, 2015, Proceedings}, issuetitle = {Communications in Computer and Information Science}, volume = {544}, pages = {101-112}, chapter = {Metadata and Semantics Research}, series = { Communications in Computer and Information Science}, abstract = {The successful management of textual information is a rising challenge for all the researchers’ communities, in order firstly to assess its current and previous statuses and secondly to enrich the level of their metadata description. The huge amount of unstructured data that is produced has consequently populated text mining techniques for its interpretation, selection and metadata enrichment opportunities that provides. Scientific production regarding Digital Libraries (DLs) evaluation has been grown in size and has broaden the scope of coverage as it consists a complex and multidimensional field. The current study proposes a probabilistic topic modeling implemented on a domain corpus from the JCDL, ECDL/TDPL and ICADL conferences proceedings in the period 2001-2013, aiming at the unveiling of its topics and subject temporal analysis, for exploiting and extracting semantic metadata from large corpora in an automatic way.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The successful management of textual information is a rising challenge for all the researchers’ communities, in order firstly to assess its current and previous statuses and secondly to enrich the level of their metadata description. The huge amount of unstructured data that is produced has consequently populated text mining techniques for its interpretation, selection and metadata enrichment opportunities that provides. Scientific production regarding Digital Libraries (DLs) evaluation has been grown in size and has broaden the scope of coverage as it consists a complex and multidimensional field. The current study proposes a probabilistic topic modeling implemented on a domain corpus from the JCDL, ECDL/TDPL and ICADL conferences proceedings in the period 2001-2013, aiming at the unveiling of its topics and subject temporal analysis, for exploiting and extracting semantic metadata from large corpora in an automatic way. |
Cultural heritage content re-use: An aggregator (Paper in Conference Proceedings) Gavrilis, Dimitris; Ioannides, Marinos; Theofanous, Eirini ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Proceedings of the 25th International CIPA Symposium 2015, Volume: II-5/W3 Pages: 83-87, Copernicus Publications, 2015. @inproceedings{Gavrilis2015, title = {Cultural heritage content re-use: An aggregator}, author = {Dimitris Gavrilis and Marinos Ioannides and Eirini Theofanous}, url = {http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/II-5-W3/83/2015/isprsannals-II-5-W3-83-2015.pdf}, doi = {:10.5194/isprsannals-II-5-W3-83-2015 }, year = {2015}, date = {2015-00-00}, booktitle = {ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Proceedings of the 25th International CIPA Symposium 2015}, volume = {II-5/W3}, pages = {83-87}, publisher = {Copernicus Publications}, abstract = {This paper introduces a use case of re-using aggregated and enriched metadata for the tourism creative industry. The MORe aggregation and enrichment framework is presented along with an example for enriching cultural heritage objects harvested from a number of Omeka repositories. The enriched content is then published both to the EU Digital Library Europeana (www.europeana.eu) and to an Elastic Search component that feeds a portal aimed at providing tourists with interesting information. }, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } This paper introduces a use case of re-using aggregated and enriched metadata for the tourism creative industry. The MORe aggregation and enrichment framework is presented along with an example for enriching cultural heritage objects harvested from a number of Omeka repositories. The enriched content is then published both to the EU Digital Library Europeana (www.europeana.eu) and to an Elastic Search component that feeds a portal aimed at providing tourists with interesting information. |
2014 |
Integrating library and cultural heritage data models: the BIBFRAME - EDM case (Paper in Conference Proceedings) Zapounidou, Sophia; Sfakakis, Michalis; Papatheodorou, Christos Katsikas, Sokratis; Hatzopoulos, Michael; Apostolopoulos, Theodoros; Anagnostopoulos, Dimosthenis; Carayiannis, Elias; Varvarigou, Theodora; Nikolaidou, Mara (Ed.): PCI '14 Proceedings of the 18th Panhellenic Conference on Informatics, Pages: 1-6, ACM, 2014, ISBN: 978-1-4503-2897-5. @inproceedings{Zapounidou2014, title = {Integrating library and cultural heritage data models: the BIBFRAME - EDM case}, author = {Sophia Zapounidou and Michalis Sfakakis and Christos Papatheodorou}, editor = {Sokratis K. Katsikas and Michael Hatzopoulos and Theodoros Apostolopoulos and Dimosthenis Anagnostopoulos and Elias Carayiannis and Theodora A. Varvarigou and Mara Nikolaidou }, url = {http://doi.acm.org/10.1145/2645791.2645805}, isbn = {978-1-4503-2897-5}, year = {2014}, date = {2014-10-02}, booktitle = {PCI '14 Proceedings of the 18th Panhellenic Conference on Informatics}, pages = {1-6}, publisher = {ACM}, abstract = {Libraries create and preserve bibliographic data using the MARC family of standards to encode and interchange them. Aggregation and exposure of these data into the Semantic Web universe is a key issue in libraries and is approached on the basis of library data conceptual models. Examining the way that data are represented in each data model, as well as possible mappings between different data models is an important step towards interoperability. This paper aims to contribute to the desired interoperability by attempting to map core classes and properties between two well known conceptual models, namely BIBFRAME and EDM. BIBFRAME aims to transform the widely used MARC data structure in libraries to the Linked Data context and EDM is the model developed and used in the Europeana Cultural Heritage aggregation portal.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Libraries create and preserve bibliographic data using the MARC family of standards to encode and interchange them. Aggregation and exposure of these data into the Semantic Web universe is a key issue in libraries and is approached on the basis of library data conceptual models. Examining the way that data are represented in each data model, as well as possible mappings between different data models is an important step towards interoperability. This paper aims to contribute to the desired interoperability by attempting to map core classes and properties between two well known conceptual models, namely BIBFRAME and EDM. BIBFRAME aims to transform the widely used MARC data structure in libraries to the Linked Data context and EDM is the model developed and used in the Europeana Cultural Heritage aggregation portal. |
From Europeana Cloud to Europeana Research: The challenges of a community-driven platform exploiting Europeana content (Paper in Conference Proceedings) Benardou, Agiatis; Dallas, Costis; Dunning, Alastair Ioannides, Marinos; Thalmann, Nadia Magnenat; Fink, Eleanor; Zarnic, Roko; Yen, Alex Yianing; Quak, Ewald (Ed.): Digital Heritage. Progress in Cultural Heritaage: Documentation, Preservation, and Protection - 5th International Conference, EuroMed 2014 , Volume: 8740 of the series Lecture Notes in Computer Science Pages: 802-210, Springer, 2014. @inproceedings{Benardou2014, title = {From Europeana Cloud to Europeana Research: The challenges of a community-driven platform exploiting Europeana content}, author = {Agiatis Benardou and Costis Dallas and Alastair Dunning}, editor = {Marinos Ioannides and Nadia Magnenat - Thalmann and Eleanor E. Fink and Roko Zarnic and Alex - Yianing Yen and Ewald Quak }, url = {http://link.springer.com/chapter/10.1007%2F978-3-319-13695-0_82}, year = {2014}, date = {2014-01-01}, booktitle = {Digital Heritage. Progress in Cultural Heritaage: Documentation, Preservation, and Protection - 5th International Conference, EuroMed 2014 }, issuetitle = {Digital Heritage. Progress in Cultural Heritage: Documentation, Preservation, and Protection}, volume = {8740}, pages = {802-210}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, abstract = {This paper presents Europeana Cloud, a Best Practice Network coordinated by the Europeana Foundation, which aims at setting the foundations and building Europeana Research, a platform allowing third parties to develop tools and services based on Europeana content. Through a collaborative, user-centred and a mixed methods approach, we have tried to identify the needs of researchers in the fields of the Humanities and the Social Sciences, while attempting to actively engage various sub-disciplines in the course of our work. }, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } This paper presents Europeana Cloud, a Best Practice Network coordinated by the Europeana Foundation, which aims at setting the foundations and building Europeana Research, a platform allowing third parties to develop tools and services based on Europeana content. Through a collaborative, user-centred and a mixed methods approach, we have tried to identify the needs of researchers in the fields of the Humanities and the Social Sciences, while attempting to actively engage various sub-disciplines in the course of our work. |
Where and how knowledge on digital library evaluation spreads: a case study on conference literature (Paper in Conference Proceedings) Papachristopoulos, Leonidas; Mitrelis, Angelos; Tsakonas, Giannis; Papatheodorou, Christos Proceedings of the Libraries in the Digital Age (LIDA) Conference 2014, Volume: 13 2014. @inproceedings{Papachristopoulos2014, title = {Where and how knowledge on digital library evaluation spreads: a case study on conference literature}, author = {Leonidas Papachristopoulos and Angelos Mitrelis and Giannis Tsakonas and Christos Papatheodorou}, url = {http://www.dcu.gr/wp-content/uploads/2016/10/01.pdf}, year = {2014}, date = {2014-01-01}, booktitle = {Proceedings of the Libraries in the Digital Age (LIDA) Conference 2014}, volume = {13}, abstract = {Scholarly communication has not remained unaffected by the advance of the social networking culture. The traditional bibliometric paradigm is strongly questioned as a tool that accurately portrays the impact of research outcomes. New metrics, such as download or view rates and shares, have been proposed as alternative ways for measuring the impact of digital content published in the form of articles, datasets, etc. Mendeley's Readership Statistics are one of these metrics, based on the assumption that there is a linkage between a paper in a collection and the interests of the collection owner. The current study explores the ‘altmetric’ aspects of the literature of the digital libraries evaluation domain, as it is expressed in two major conferences of the field, namely JCDL and ECDL. Our corpus consists of 224 papers, for which we extract readership data from Mendeley and examine in how many 1 Corresponding author. collections these papers belong to. Our goal is to investigate whether readership statistics can help us to understand where and to whom DL evaluation research has impact. Therefore the data are analyzed statistically to produce indicators of geographical and topical distribution of Mendeley readers as well as to explore and classify their profession. Finally it derived that there is a loose correlation between the number of Google Scholar citations and the number of Mendeley readers.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Scholarly communication has not remained unaffected by the advance of the social networking culture. The traditional bibliometric paradigm is strongly questioned as a tool that accurately portrays the impact of research outcomes. New metrics, such as download or view rates and shares, have been proposed as alternative ways for measuring the impact of digital content published in the form of articles, datasets, etc. Mendeley's Readership Statistics are one of these metrics, based on the assumption that there is a linkage between a paper in a collection and the interests of the collection owner. The current study explores the ‘altmetric’ aspects of the literature of the digital libraries evaluation domain, as it is expressed in two major conferences of the field, namely JCDL and ECDL. Our corpus consists of 224 papers, for which we extract readership data from Mendeley and examine in how many 1 Corresponding author. collections these papers belong to. Our goal is to investigate whether readership statistics can help us to understand where and to whom DL evaluation research has impact. Therefore the data are analyzed statistically to produce indicators of geographical and topical distribution of Mendeley readers as well as to explore and classify their profession. Finally it derived that there is a loose correlation between the number of Google Scholar citations and the number of Mendeley readers. |
Describing Research Data: A Case Study for Archaeology (Paper in Conference Proceedings) Aloia, Nicola; Papatheodorou, Christos; Gavrilis, Dimitris; Debole, Franca; Meghini, Carlo Meersman, Robert; Panetto, Herve; Dillon, Tharam; Missikoff, Michele; Liu, Lin; Pastor, Oscar; Cuzzocrea, Alfredo; Sellis, Timos (Ed.): On the Move to Meaningful Internet Systems: {OTM} 2014 Conferences - Confederated International Conferences: CoopIS, and ODBASE 2014 , Volume: 8841 of the series Lecture Notes in Computer Science Pages: 768-775, Springer, 2014. @inproceedings{Aloia2014, title = {Describing Research Data: A Case Study for Archaeology}, author = {Nicola Aloia and Christos Papatheodorou and Dimitris Gavrilis and Franca Debole and Carlo Meghini}, editor = {Robert Meersman and Herve Panetto and Tharam S. Dillon and Michele Missikoff and Lin Liu and Oscar Pastor and Alfredo Cuzzocrea and Timos K. Sellis }, url = {http://link.springer.com/chapter/10.1007%2F978-3-662-45563-0_48}, year = {2014}, date = {2014-01-01}, booktitle = {On the Move to Meaningful Internet Systems: {OTM} 2014 Conferences - Confederated International Conferences: CoopIS, and ODBASE 2014 }, issuetitle = {On the Move to Meaningful Internet Systems: OTM 2014 Conferences}, volume = {8841}, pages = {768-775}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, abstract = {The growth of the digital resources produced by the research activities demand the development of e-Infrastructures in which researchers can access remote facilities, select and re-use huge volumes of data and services, run complex experimental processes and share results. Data registries aim to describe uniformly the data of e-Infrastructures contributing to the re-usability and interoperability of big scientific data. However the current situation requires the development of powerful resource integration mechanisms that step beyond the principles guaranteed by the data registries standards. This paper proposes a conceptual model for describing data resources and services and extends the existing specifications for the development of data registries. The model has been implemented in the context of the ARIADNE project, a EU funded project that focuses on the integration of Archaeological digital resources all over the Europe.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The growth of the digital resources produced by the research activities demand the development of e-Infrastructures in which researchers can access remote facilities, select and re-use huge volumes of data and services, run complex experimental processes and share results. Data registries aim to describe uniformly the data of e-Infrastructures contributing to the re-usability and interoperability of big scientific data. However the current situation requires the development of powerful resource integration mechanisms that step beyond the principles guaranteed by the data registries standards. This paper proposes a conceptual model for describing data resources and services and extends the existing specifications for the development of data registries. The model has been implemented in the context of the ARIADNE project, a EU funded project that focuses on the integration of Archaeological digital resources all over the Europe. |
Library Data Integration: Towards BIBFRAME Mapping to EDM (Paper in Conference Proceedings) Zapounidou, Sophia; Sfakakis, Michalis; Papatheodorou, Christos Closs, Sissi; Studer, Rudi; Garoufallou, Emmanouel; Sicilia, Miguel Angel (Ed.): Metadata and Semantics Research. International Conference on Theory and Practice of Digital Libraries, TPDL 2013, Valletta, Malta, September 22-26, 2013. Proceedings, Volume: 478 of the series Communications in Computer and Information Science Pages: 262-273, Springer, 2014. @inproceedings{Zapounidou2014, title = {Library Data Integration: Towards BIBFRAME Mapping to EDM}, author = {Sophia Zapounidou and Michalis Sfakakis and Christos Papatheodorou}, editor = {Sissi Closs and Rudi Studer and Emmanouel Garoufallou and Miguel - Angel Sicilia }, url = {http://dx.doi.org/10.1007/978-3-319-13674-5_25}, year = {2014}, date = {2014-01-01}, booktitle = {Metadata and Semantics Research. International Conference on Theory and Practice of Digital Libraries, TPDL 2013, Valletta, Malta, September 22-26, 2013. Proceedings}, volume = {478}, pages = {262-273}, publisher = {Springer}, series = {Communications in Computer and Information Science}, abstract = {Libraries create and preserve bibliographic data using the MARC family of standards to encode and interchange them. Aggregation and exposure of these data into the Semantic Web universe is a key issue in libraries and is approached on the basis of library data conceptual models. Examining the way that data are represented in each data model, as well as possible mappings between different data models is an important step towards interoperability. This paper aims to contribute to the desired interoperability by attempting to map core classes and properties between two well known conceptual models, namely BIBFRAME and EDM. BIBFRAME aims to transform the widely used MARC data structure in libraries to the Linked Data context and EDM is the model developed and used in the Europeana Cultural Heritage aggregation portal. }, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Libraries create and preserve bibliographic data using the MARC family of standards to encode and interchange them. Aggregation and exposure of these data into the Semantic Web universe is a key issue in libraries and is approached on the basis of library data conceptual models. Examining the way that data are represented in each data model, as well as possible mappings between different data models is an important step towards interoperability. This paper aims to contribute to the desired interoperability by attempting to map core classes and properties between two well known conceptual models, namely BIBFRAME and EDM. BIBFRAME aims to transform the widely used MARC data structure in libraries to the Linked Data context and EDM is the model developed and used in the Europeana Cultural Heritage aggregation portal. |
2013 |
A Curation-Oriented Thematic Aggregator (Paper in Conference Proceedings) Gavrilis, Dimitris; Angelis, Stavros; Dallas, Costis Research and Advanced Technology for Digital Libraries. International Conference on Theory and Practice of Digital Libraries, TPDL 2013, Valletta, Malta, September 22-26, 2013. Proceedings , of the series Lecture Notes in Computer Science Pages: 132–137, 2013. @inproceedings{Gavrilis2013, title = {A Curation-Oriented Thematic Aggregator}, author = {Dimitris Gavrilis and Stavros Angelis and Costis Dallas}, url = {http://link.springer.com/chapter/10.1007%2F978-3-642-40501-3_13}, year = {2013}, date = {2013-09-02}, booktitle = {Research and Advanced Technology for Digital Libraries. International Conference on Theory and Practice of Digital Libraries, TPDL 2013, Valletta, Malta, September 22-26, 2013. Proceedings }, journal = {Proceedings of the 17th International Conference on Theory and Practice of Digital Libraries, TPDL 2013}, pages = {132--137}, series = {Lecture Notes in Computer Science}, abstract = {The emergence of the European Digital Library (Europeana) presents the need for aggregating content using a more intelligent and effective approach, taking into account the need to support potential changes in target metadata schemas and new services. This paper presents the concept, architecture and services provided by a curation-oriented, OAIS-compliant thematic metadata aggregator, developed and used in the CARARE project, that addresses these challenges.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The emergence of the European Digital Library (Europeana) presents the need for aggregating content using a more intelligent and effective approach, taking into account the need to support potential changes in target metadata schemas and new services. This paper presents the concept, architecture and services provided by a curation-oriented, OAIS-compliant thematic metadata aggregator, developed and used in the CARARE project, that addresses these challenges. |
Preservation Aspects of a Curation-Oriented Thematic Aggregator (Paper in Conference Proceedings) Gavrilis, Dimitris; Angelis, Stavros; Papatheodorou, Christos; CostisDallas,; Constantopoulos, Panos Proceedings of the 10th International Conference on Preservation of Digital Objects, Pages: 246-251, 2013. @inproceedings{Gavrilis2013, title = {Preservation Aspects of a Curation-Oriented Thematic Aggregator}, author = {Dimitris Gavrilis and Stavros Angelis and Christos Papatheodorou and CostisDallas and Panos Constantopoulos}, url = {http://purl.pt/24107/1/iPres2013_PDF/Preservation%20Aspects%20of%20a%20Curation-Oriented%20Thematic%20Aggregator.pdf}, year = {2013}, date = {2013-09-01}, booktitle = {Proceedings of the 10th International Conference on Preservation of Digital Objects}, pages = {246-251}, abstract = {The emergence of the European Digital Library (Europeana) foregrounds the need for aggregating content using smarter and more efficient ways taking into account its context and production circumstances. This paper presents the main functionalities of MoRe, a curation oriented aggregator that addresses digital preservation issues. MoRe combines aggregation, digital curation and preservation capabilities in a package that shields content providers from changes, and that ensures efficient, high volume metadata processing. It aggregates data from a wide community of archaeological content providers and integrates them to a common metadata schema. The system provides added-value digital curation services for metadata quality monitoring and enrichment so that to ensure metadata reliability. Furthermore it provides preservation workflows which guarantee effective record keeping of all transactions and the current status of the repository. }, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The emergence of the European Digital Library (Europeana) foregrounds the need for aggregating content using smarter and more efficient ways taking into account its context and production circumstances. This paper presents the main functionalities of MoRe, a curation oriented aggregator that addresses digital preservation issues. MoRe combines aggregation, digital curation and preservation capabilities in a package that shields content providers from changes, and that ensures efficient, high volume metadata processing. It aggregates data from a wide community of archaeological content providers and integrates them to a common metadata schema. The system provides added-value digital curation services for metadata quality monitoring and enrichment so that to ensure metadata reliability. Furthermore it provides preservation workflows which guarantee effective record keeping of all transactions and the current status of the repository. |
Achieving Interoperability between the CARARE schema for Monuments and Sites and the Europeana Data Model (Paper in Conference Proceedings) Isaac, Antoine; Charles, Valentine; Fernie, Kate; Dallas, Costis; Gavrilis, Dimitris; Angelis, Stavros Proceedings of the International Conference on Dublin Core and Metadata Applications - DC-2013, Pages: 115-125, 2013. @inproceedings{Isaac2013, title = {Achieving Interoperability between the CARARE schema for Monuments and Sites and the Europeana Data Model}, author = {Antoine Isaac and Valentine Charles and Kate Fernie and Costis Dallas and Dimitris Gavrilis and Stavros Angelis}, url = {DCMI International Conference on Dublin Core and Metadata Applications}, year = {2013}, date = {2013-09-01}, booktitle = {Proceedings of the International Conference on Dublin Core and Metadata Applications - DC-2013}, pages = {115-125}, abstract = {Mapping between different data models in a data aggregation context always presents significant interoperability challenges. In this paper, we describe the challenges faced and solutions developed when mapping the CARARE schema designed for archaeological and architectural monuments and sites to the Europeana Data Model (EDM), a model based on Linked Data principles, for the purpose of integrating more than two million metadata records from national monument collections and databases across Europe into the Europeana digital library.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Mapping between different data models in a data aggregation context always presents significant interoperability challenges. In this paper, we describe the challenges faced and solutions developed when mapping the CARARE schema designed for archaeological and architectural monuments and sites to the Europeana Data Model (EDM), a model based on Linked Data principles, for the purpose of integrating more than two million metadata records from national monument collections and databases across Europe into the Europeana digital library. |
Charting the digital library evaluation domain with a semantically enhanced mining methodology (Paper in Conference Proceedings) Afiontzi, Eleni; Kazadeis, Giannis; Papachristopoulos, Leonidas; Sfakakis, Michalis; Tsakonas, Giannis; Papatheodorou, Christos Proceedings of the 13th ACM/IEE-CS joint conference on Digital libraries, JCDL , Pages: 125-134 , 2013. @inproceedings{Afiontzi2013, title = {Charting the digital library evaluation domain with a semantically enhanced mining methodology}, author = {Eleni Afiontzi and Giannis Kazadeis and Leonidas Papachristopoulos and Michalis Sfakakis and Giannis Tsakonas and Christos Papatheodorou}, url = {http://dl.acm.org/citation.cfm?id=2467713}, doi = {10.1145/2467696.2467713}, year = {2013}, date = {2013-07-01}, booktitle = {Proceedings of the 13th ACM/IEE-CS joint conference on Digital libraries, JCDL }, pages = {125-134 }, abstract = {The digital library evaluation field has an evolving nature and it is characterized by a noteworthy proclivity to enfold various methodological orientations. Given the fact that the scientific literature in the specific domain is vast, researchers require tools that will exhibit either commonly acceptable practices, or areas for further investigation. In this paper, a data mining methodology is proposed to identify prominent patterns in the evaluation of digital libraries. Using Machine Learning techniques, all papers presented in the ECDL and JCDL conferences between the years 2001 and 2011 were categorized as relevant or non-relevant to the DL evaluation domain. Then, the relevant papers were semantically annotated according to the Digital Library Evaluation Ontology (DiLEO) vocabulary. The produced set of annotations was clustered to evaluation patterns for the most frequently used tools, methods and goals of the domain. Our findings highlight the expressive nature of DiLEO, place emphasis on semantic annotation as a necessary step in handling domain-centric corpora and underline the potential of the proposed methodology in the profiling of evaluation activities.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The digital library evaluation field has an evolving nature and it is characterized by a noteworthy proclivity to enfold various methodological orientations. Given the fact that the scientific literature in the specific domain is vast, researchers require tools that will exhibit either commonly acceptable practices, or areas for further investigation. In this paper, a data mining methodology is proposed to identify prominent patterns in the evaluation of digital libraries. Using Machine Learning techniques, all papers presented in the ECDL and JCDL conferences between the years 2001 and 2011 were categorized as relevant or non-relevant to the DL evaluation domain. Then, the relevant papers were semantically annotated according to the Digital Library Evaluation Ontology (DiLEO) vocabulary. The produced set of annotations was clustered to evaluation patterns for the most frequently used tools, methods and goals of the domain. Our findings highlight the expressive nature of DiLEO, place emphasis on semantic annotation as a necessary step in handling domain-centric corpora and underline the potential of the proposed methodology in the profiling of evaluation activities. |
2012 |
Defining User Requirements for Holocaust Research Infrastractures and Services in the EHRI Project (Paper in Conference Proceedings) Benardou, Agiatis; Dallas, Costis Proceedings of the 2012 iConference, Pages: 644-645, 2012. @inproceedings{Benardou2012, title = {Defining User Requirements for Holocaust Research Infrastractures and Services in the EHRI Project}, author = {Agiatis Benardou and Costis Dallas}, url = {http://dl.acm.org/citation.cfm?doid=2132176.2132322}, year = {2012}, date = {2012-01-02}, booktitle = {Proceedings of the 2012 iConference}, pages = {644-645}, abstract = {The poster presents the background, conceptual framework, methodology and initial results of a mixed research project, investigating information practice and user requirements of historians, humanities scholars and social scientists working on the Holocaust. The results of the study will be a foundation for the specification of functionalities of the European digital infrastructure planned as part of the EU-funded European Holocaust Research Infrastructure project, and consisting of the EHRI search/portal and the EHRI Virtual Research Environment. Particular issues to be dealt with are the summarization of qualitative evidence from semi-open interviews by means of a conceptualization of relevant descriptive codes of research activities, resource types and tools/services; identification of specific user requirements and of different researcher profiles through statistical analysis of an online questionnaire defined on the basis of initial qualitative research; and, identification and theorization of special needs of Holocaust research, seen as a highly multi- and inter-disciplinary field of inquiry.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The poster presents the background, conceptual framework, methodology and initial results of a mixed research project, investigating information practice and user requirements of historians, humanities scholars and social scientists working on the Holocaust. The results of the study will be a foundation for the specification of functionalities of the European digital infrastructure planned as part of the EU-funded European Holocaust Research Infrastructure project, and consisting of the EHRI search/portal and the EHRI Virtual Research Environment. Particular issues to be dealt with are the summarization of qualitative evidence from semi-open interviews by means of a conceptualization of relevant descriptive codes of research activities, resource types and tools/services; identification of specific user requirements and of different researcher profiles through statistical analysis of an online questionnaire defined on the basis of initial qualitative research; and, identification and theorization of special needs of Holocaust research, seen as a highly multi- and inter-disciplinary field of inquiry. |
2011 |
A New Architecture and Approach to Asset Representation for Europeana Aggregation: The CARARE Way (Paper in Conference Proceedings) Papatheodorou, Christos; Dallas, Costis; Ertmann-Christiansen, Christian; Fernie, Kate; Gavrilis, Dimitris; Masci, Maria Emilia; Constantopoulos, Panos; Angelis, Stavros Metadata and Semantic Research, Proceedings of the 5th International Conference, Communications in Computer and Information Science, Volume: 240 of the series Communications in Computer and Information Science Pages: 412-423, 2011. @inproceedings{Papatheodorou2011, title = {A New Architecture and Approach to Asset Representation for Europeana Aggregation: The CARARE Way}, author = {Christos Papatheodorou and Costis Dallas and Christian Ertmann-Christiansen and Kate Fernie and Dimitris Gavrilis and Maria Emilia Masci and Panos Constantopoulos and Stavros Angelis}, url = {http://www.springerlink.com/content/qh565142q69x0763/}, year = {2011}, date = {2011-10-03}, booktitle = {Metadata and Semantic Research, Proceedings of the 5th International Conference, Communications in Computer and Information Science}, volume = {240}, pages = {412-423}, chapter = {Metadata and Semantic Research}, series = {Communications in Computer and Information Science}, abstract = {This paper presents a new metadata aggregation approach based on a mediating repository that intends to ensure the integrity, authenticity and semantic enrichment of metadata provided to Europeana by heterogeneous collections. Primary metadata are mapped to CARARE schema, a schema suitable for describing archaeological and architectural heritage assets, digital resources, collections, as well as events associated with them. The paper specifies the proposed schema and discusses the overall architecture of the proposed approach.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } This paper presents a new metadata aggregation approach based on a mediating repository that intends to ensure the integrity, authenticity and semantic enrichment of metadata provided to Europeana by heterogeneous collections. Primary metadata are mapped to CARARE schema, a schema suitable for describing archaeological and architectural heritage assets, digital resources, collections, as well as events associated with them. The paper specifies the proposed schema and discusses the overall architecture of the proposed approach. |