Calculating an Air Quality Index (AQI) typically uses data streams from air quality sensors deployed at fixed locations and the calculation is a real time process. If one or a number of sensors are broken or offline, then the real time AQI value cannot be computed. Estimating AQI values for some point in the future is a predictive process and uses historical AQI values to train and build models. In this work we focus on gap filling in air quality data where the task is to predict the AQI at 1, 5 and 7 days into the future. The scenario is where one or a number of air, weather and traffic sensors are offline and explores prediction accuracy under such situations. The work is part of the MediaEval'2022 Urban Air: Urban Life and Air Pollution task submitted by the DCU-Insight-AQ team and uses multimodal and crossmodal data consisting of AQI, weather and CCTV traffic images for air pollution prediction.
translated by 谷歌翻译
As part of the MediaEval 2022 Predicting Video Memorability task we explore the relationship between visual memorability, the visual representation that characterises it, and the underlying concept portrayed by that visual representation. We achieve state-of-the-art memorability prediction performance with a model trained and tested exclusively on surrogate dream images, elevating concepts to the status of a cornerstone memorability feature, and finding strong evidence to suggest that the intrinsic memorability of visual content can be distilled to its underlying concept or meaning irrespective of its specific visual representational.
translated by 谷歌翻译
This paper describes the 5th edition of the Predicting Video Memorability Task as part of MediaEval2022. This year we have reorganised and simplified the task in order to lubricate a greater depth of inquiry. Similar to last year, two datasets are provided in order to facilitate generalisation, however, this year we have replaced the TRECVid2019 Video-to-Text dataset with the VideoMem dataset in order to remedy underlying data quality issues, and to prioritise short-term memorability prediction by elevating the Memento10k dataset as the primary dataset. Additionally, a fully fledged electroencephalography (EEG)-based prediction sub-task is introduced. In this paper, we outline the core facets of the task and its constituent sub-tasks; describing the datasets, evaluation metrics, and requirements for participant submissions.
translated by 谷歌翻译
Recent developments in in-situ monitoring and process control in Additive Manufacturing (AM), also known as 3D-printing, allows the collection of large amounts of emission data during the build process of the parts being manufactured. This data can be used as input into 3D and 2D representations of the 3D-printed parts. However the analysis and use, as well as the characterization of this data still remains a manual process. The aim of this paper is to propose an adaptive human-in-the-loop approach using Machine Learning techniques that automatically inspect and annotate the emissions data generated during the AM process. More specifically, this paper will look at two scenarios: firstly, using convolutional neural networks (CNNs) to automatically inspect and classify emission data collected by in-situ monitoring and secondly, applying Active Learning techniques to the developed classification model to construct a human-in-the-loop mechanism in order to accelerate the labeling process of the emission data. The CNN-based approach relies on transfer learning and fine-tuning, which makes the approach applicable to other industrial image patterns. The adaptive nature of the approach is enabled by uncertainty sampling strategy to automatic selection of samples to be presented to human experts for annotation.
translated by 谷歌翻译
The Predicting Media Memorability task in the MediaEval evaluation campaign has been running annually since 2018 and several different tasks and data sets have been used in this time. This has allowed us to compare the performance of many memorability prediction techniques on the same data and in a reproducible way and to refine and improve on those techniques. The resources created to compute media memorability are now being used by researchers well beyond the actual evaluation campaign. In this paper we present a summary of the task, including the collective lessons we have learned for the research community.
translated by 谷歌翻译
我们通过在预测视频记忆力的任务中对视觉变压器进行了微调,调查了流行的犯罪剧电视连续剧CSI的5季跨度的记忆。通过使用详细的注释语料库结合视频记忆性分数来调查犯罪戏剧电视的流行类型,我们展示了如何从视频拍摄中产生的记忆性分数中推断出含义。我们执行定量分析,将视频拍摄的记忆与节目的各个方面相关联。我们在本文中提供的见解说明了视频记忆力在应用程序中使用多媒体在教育,市场营销,索引以及在这里的情况下的重要性的重要性,即电视和电影制作。
translated by 谷歌翻译
尽管现在使用自我监督方法构建的计算机视觉模型现在很普遍,但仍然存在一些重要问题。自我监督的模型是否学习高度冗余的频道功能?如果一个自我监督的网络可以动态选择重要的渠道并摆脱不必要的渠道怎么办?目前,与计算机视觉中的有监督的对手相比,通过自我训练预先训练的Convnet在下游任务上获得了可比的性能。但是,有一些自我监督模型的缺点,包括大量参数,计算昂贵的培训策略以及对下游任务更快推断的明确需求。在这项工作中,我们的目标是通过研究如何将用于监督学习的标准渠道选择方法应用于经过自学训练的网络。我们验证我们在一系列目标预算上验证我们的发现$ t_ {d} $,用于跨不同数据集的图像分类任务的频道计算,特别是CIFAR-10,CIFAR-100和IMAGENET-100,获得了与原始网络的可比性性能when selecting all channels but at a significant reduction in computation reported in terms of FLOPs.
translated by 谷歌翻译
可以通过串联协作来启用第二语言学习,在该协作中,学生在呼叫中学习其他学生的母语时,将学生分组为视频电话会议。这使学生处于在线环境中,更外向的人可以积极贡献和进行对话,而那些更害羞和不确定其第二语言技能的人可以通过电话坐下来坐下来。我们已经构建并部署了L2L系统,该系统记录了所有参与者在呼叫中的对话说话的时间。我们生成可视化的,包括每个呼叫中​​每个学生的参与率和时间表,并在仪表板上呈现。我们最近制定了一种称为个人对话波动率的措施,以表明每个学生在每个呼叫中​​对对话的贡献如何。我们介绍了来自大学学习Frenchm的19个讲英语的学生的样本的对话波动率措施的分析,在一个教学学期的86个串联电信呼叫中。我们的分析表明,有必要研究互动的本质,看看分配给他们的讨论主题的选择是否太难了,这可能会以某种方式影响他们的参与。
translated by 谷歌翻译
本文介绍了我们对Mediaeval 2021中预测媒体难忘任务的方法,该任务旨在通过设置自动预测视频难忘的任务来解决媒体难忘的问题。今年,我们从比较角度解决任务,寻求深入了解三个探索的模式,并使用去年提交的结果(2020年)作为参考点。我们在TRECVID2019 DataSet上测试的最佳性交短期难忘模型(0.132)就像去年一样 - 是基于帧的CNN,没有接受任何TRECVID数据培训,以及我们测试的最佳短期难忘性模型(0.524)在Memento10k DataSet上,是一个贝叶斯乘坐回归,符合Densenet121视觉功能。
translated by 谷歌翻译
本文介绍了预测媒体难忘性的Mediaeval 2021,这是今年第4版的任务,因为短期和长期视频难忘性的预测仍然是一个具有挑战性的任务。在2021年,使用两个视频数据集:第一,TRECVID 2019视频到文本数据集的子集;其次,Memento10K数据集是为了提供探索交叉数据集泛化的机会。另外,介绍了基于脑电图(EEG)的预测导频子任务。在本文中,我们概述了任务的主要方面,并描述了参与者提交的数据集,评估指标和要求。
translated by 谷歌翻译