众包技术依靠人群输入可能对决策至关重要的信息。这项工作检查了报告技术的混淆。我们表明,报告平台的广泛使用具有独特的安全性和隐私影响,并引入了威胁模型和相应的分类法,以概述该领域中众多攻击向量中的一些。然后,我们对有争议的现实世界报告热线的呼叫日志数据集进行了经验分析,并确定旨在阻碍平台合法性的协调混淆策略。我们提出了各种统计措施,以量化这种混淆策略在我们数据集中报告攻击的结构和语义特征方面的强度。
translated by 谷歌翻译
如今,算法在控制或影响我们生活的各个方面的许多技术系统中起着关键作用。结果,提供解释以满足用户和组织的需求,越来越多地受到法律法规,行为准则和公众的期望。但是,由于法律和法规没有规定如何满足这种期望,因此通常会留下组织来设计自己的解释性方法,不可避免地增加合规性和良好的治理成本。因此,我们提出了“通过设计的解释性”,这是一种以主动措施为特征的整体方法,包括在决策系统设计中的解释能力。本文介绍了软件工程工作流程中解释性方法的技术步骤,以实现域专家针对特定应用程序上下文提出的要求的解释能力。解释性逐设计方法的输出是一组配置,允许可重复使用的服务(称为解释助手)利用应用程序提供的日志并创建可以查询以提取相关数据点的出处痕迹,而这又可以是用于解释计划,以构建向消费者个性化的解释。遵循这些步骤,组织将能够设计其决策系统,以产生满足指定要求的解释,无论是根据法律,法规或业务需求而设计的。我们将方法应用于两个应用程序,从而部署了解释助理,展示了解释功能。最后,测量了相关的开发成本,表明构建解释的方法在开发时间方面是可以探讨的,每个解释句子可能低至两个小时。
translated by 谷歌翻译
随着自动决策解决方案越来越多地应用于日常生活的各个方面,因此为各种利益相关者(即决策者,决策者,审计师,监管机构...)产生有意义的解释能力变得至关重要。在本文中,我们提出了一种解释的分类法,该分类是作为该项目目的的整体“解释性划分”方法的一部分。该分类法的建立是为了为在组织层面设定的各种监管框架或政策所引起的广泛要求提供解释,以转化高级合规性要求或满足业务需求。分类法包括九个维度。它被用作被认为是侦探控制的解释的独立分类器,以帮助支持性自动化的合规策略。通过一系列示例证明了分类法的可机械性格式,并以轻度本体的形式提供了使用这种分类法的解释性的好处。
translated by 谷歌翻译
我们介绍了泰国抑郁症的第一个公开的有用的语料库。我们的语料库由几个在线博客中的抑郁症的专家验证案例编制。我们试验两种不同的基于LSTM的模型和两种不同的基于伯特模型。我们在检测抑郁症时达到77.53 \%的准确性。这为同一语料库的未来研究人员建立了一个很好的基准。此外,我们确定需要在比维基百科更多种多样的语料库培训的泰国嵌入。我们的语料库,代码和培训的型号在Zenodo上公开发布。
translated by 谷歌翻译
Variational inference uses optimization, rather than integration, to approximate the marginal likelihood, and thereby the posterior, in a Bayesian model. Thanks to advances in computational scalability made in the last decade, variational inference is now the preferred choice for many high-dimensional models and large datasets. This tutorial introduces variational inference from the parametric perspective that dominates these recent developments, in contrast to the mean-field perspective commonly found in other introductory texts.
translated by 谷歌翻译
Knowledge graphs (KG) have served as the key component of various natural language processing applications. Commonsense knowledge graphs (CKG) are a special type of KG, where entities and relations are composed of free-form text. However, previous works in KG completion and CKG completion suffer from long-tail relations and newly-added relations which do not have many know triples for training. In light of this, few-shot KG completion (FKGC), which requires the strengths of graph representation learning and few-shot learning, has been proposed to challenge the problem of limited annotated data. In this paper, we comprehensively survey previous attempts on such tasks in the form of a series of methods and applications. Specifically, we first introduce FKGC challenges, commonly used KGs, and CKGs. Then we systematically categorize and summarize existing works in terms of the type of KGs and the methods. Finally, we present applications of FKGC models on prediction tasks in different areas and share our thoughts on future research directions of FKGC.
translated by 谷歌翻译
Few Shot Instance Segmentation (FSIS) requires models to detect and segment novel classes with limited several support examples. In this work, we explore a simple yet unified solution for FSIS as well as its incremental variants, and introduce a new framework named Reference Twice (RefT) to fully explore the relationship between support/query features based on a Transformer-like framework. Our key insights are two folds: Firstly, with the aid of support masks, we can generate dynamic class centers more appropriately to re-weight query features. Secondly, we find that support object queries have already encoded key factors after base training. In this way, the query features can be enhanced twice from two aspects, i.e., feature-level and instance-level. In particular, we firstly design a mask-based dynamic weighting module to enhance support features and then propose to link object queries for better calibration via cross-attention. After the above steps, the novel classes can be improved significantly over our strong baseline. Additionally, our new framework can be easily extended to incremental FSIS with minor modification. When benchmarking results on the COCO dataset for FSIS, gFSIS, and iFSIS settings, our method achieves a competitive performance compared to existing approaches across different shots, e.g., we boost nAP by noticeable +8.2/+9.4 over the current state-of-the-art FSIS method for 10/30-shot. We further demonstrate the superiority of our approach on Few Shot Object Detection. Code and model will be available.
translated by 谷歌翻译
Unsupervised domain adaptation (UDA) for semantic segmentation is a promising task freeing people from heavy annotation work. However, domain discrepancies in low-level image statistics and high-level contexts compromise the segmentation performance over the target domain. A key idea to tackle this problem is to perform both image-level and feature-level adaptation jointly. Unfortunately, there is a lack of such unified approaches for UDA tasks in the existing literature. This paper proposes a novel UDA pipeline for semantic segmentation that unifies image-level and feature-level adaptation. Concretely, for image-level domain shifts, we propose a global photometric alignment module and a global texture alignment module that align images in the source and target domains in terms of image-level properties. For feature-level domain shifts, we perform global manifold alignment by projecting pixel features from both domains onto the feature manifold of the source domain; and we further regularize category centers in the source domain through a category-oriented triplet loss and perform target domain consistency regularization over augmented target domain images. Experimental results demonstrate that our pipeline significantly outperforms previous methods. In the commonly tested GTA5$\rightarrow$Cityscapes task, our proposed method using Deeplab V3+ as the backbone surpasses previous SOTA by 8%, achieving 58.2% in mIoU.
translated by 谷歌翻译
The development of social media user stance detection and bot detection methods rely heavily on large-scale and high-quality benchmarks. However, in addition to low annotation quality, existing benchmarks generally have incomplete user relationships, suppressing graph-based account detection research. To address these issues, we propose a Multi-Relational Graph-Based Twitter Account Detection Benchmark (MGTAB), the first standardized graph-based benchmark for account detection. To our knowledge, MGTAB was built based on the largest original data in the field, with over 1.55 million users and 130 million tweets. MGTAB contains 10,199 expert-annotated users and 7 types of relationships, ensuring high-quality annotation and diversified relations. In MGTAB, we extracted the 20 user property features with the greatest information gain and user tweet features as the user features. In addition, we performed a thorough evaluation of MGTAB and other public datasets. Our experiments found that graph-based approaches are generally more effective than feature-based approaches and perform better when introducing multiple relations. By analyzing experiment results, we identify effective approaches for account detection and provide potential future research directions in this field. Our benchmark and standardized evaluation procedures are freely available at: https://github.com/GraphDetec/MGTAB.
translated by 谷歌翻译
The performance of inertial navigation systems is largely dependent on the stable flow of external measurements and information to guarantee continuous filter updates and bind the inertial solution drift. Platforms in different operational environments may be prevented at some point from receiving external measurements, thus exposing their navigation solution to drift. Over the years, a wide variety of works have been proposed to overcome this shortcoming, by exploiting knowledge of the system current conditions and turning it into an applicable source of information to update the navigation filter. This paper aims to provide an extensive survey of information aided navigation, broadly classified into direct, indirect, and model aiding. Each approach is described by the notable works that implemented its concept, use cases, relevant state updates, and their corresponding measurement models. By matching the appropriate constraint to a given scenario, one will be able to improve the navigation solution accuracy, compensate for the lost information, and uncover certain internal states, that would otherwise remain unobservable.
translated by 谷歌翻译