当前位置: X-MOL 学术ACM J. Comput. Cult. Herit. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Annotations as Knowledge Practices in Image Archives: Application of Linked Open Usable Data and Machine Learning
ACM Journal on Computing and Cultural Heritage ( IF 2.4 ) Pub Date : 2023-11-16 , DOI: 10.1145/3625301
Murielle Cornut 1 , Julien Antoine Raemy 2 , Florian Spiess 3
Affiliation  

We reflect on some of the preliminary findings of the Participatory Knowledge Practices in Analogue and Digital Image Archives (PIA) research project around annotations of photographic archives from the Swiss Society for Folklore Studies (SSFS) as knowledge practices, the underlying technological decisions, and their impact. The aim is not only to seek more information but to find new approaches of understanding the way in which people’s memory relate to the collective, public form of archival memory and ultimately how users figure in and shape the digital archive.

We provide a proof-of-concept workflow based on automatically generated annotations comprising 53,481 photos that were subjected to object detection using Faster R-CNN Inception ResNet V2. Of the detected objects, 184,609 have a detection score greater than 0.5, 123,529 have a score greater than 0.75, and 88,442 have a score greater than 0.9. A threshold of 0.75 was set for the dissemination of our annotations, compatible with the W3C Web Annotation Data Model (WADM) and embedded in our IIIF Manifests.

In the near future, the workflow will be upgraded to allow for the co-existence of various, and occasionally conflicting, assertions made by both human and machine users. We believe that Linked Open Usable Data (LOUD) standards should be used to improve the sustainability of such an ecosystem and to foster collaboration between actors in cultural heritage.



中文翻译:

注释作为图像档案中的知识实践:链接的开放可用数据和机器学习的应用

我们反思了模拟和数字图像档案参与式知识实践 (PIA) 研究项目的一些初步结果,这些研究项目围绕瑞士民俗研究学会 (SSFS) 的摄影档案注释作为知识实践、潜在的技术决策及其影响。其目的不仅是寻求更多信息,而且是找到新的方法来理解人们的记忆与档案记忆的集体、公共形式的关系,以及最终用户如何理解和塑造数字档案。

我们提供了一个基于自动生成注释的概念验证工作流程,其中包含 53,481 张照片,这些照片使用 Faster R-CNN Inception ResNet V2 进行了对象检测。在检测到的对象中,184,609 个检测分数大于 0.5,123,529 个分数大于 0.75,88,442 个分数大于 0.9。我们为注释的传播设置了 0.75 的阈值,与 W3C Web 注释数据模型 (WADM) 兼容并嵌入到我们的 IIIF 清单中。

在不久的将来,工作流程将升级,以允许人类和机器用户做出的各种(有时甚至是冲突的)断言共存。我们认为,应使用关联开放可用数据(LOUD)标准来提高此类生态系统的可持续性,并促进文化遗产参与者之间的合作。

更新日期:2023-11-16
down
wechat
bug