异步事件序列广泛分布在自然界和人类活动中,例如地震记录,社交媒体中的用户活动等。如何蒸馏来自这些看似混乱的数据是研究人员专注的持久主题。最有用的模型之一是点过程模型,在此基础上,研究人员获得了许多明显的结果。此外,近年来,提出了神经网络基础的点过程模型,特别是复发性神经网络(RNN),并与传统模型进行比较,其性能大大提高。变压器模型的启发,可以有效地学习序列数据而无需反复和卷积结构,变压器鹰过程出现,并实现了最先进的性能。然而,有一些研究证明,转换变压器中的递归计算可以进一步提高变压器性能。因此,我们出现了一种新型的变压器鹰过程模型,通用变压器鹰过程(UTHP),其中包含递归机制和自我关注机制,并提高了模型的局部感知能力,我们还介绍了卷积神经网络(CNN)在位置方向前馈部分。我们对几个数据集进行实验,以验证UTHP的有效性,并在引入递归机制后探索变化。这些关于多个数据集的实验表明,与以前的最先进模型相比,我们提出的新模式的性能具有一定的改进。
translated by 谷歌翻译
近年来,霍克斯进程的异步序列的知识是一个值得关注的主题,基于神经网络的鹰过程逐渐成为最热门研究的领域,特别是基于复发神经网络(RNN)。然而,这些模型仍然包含RNN的一些固有缺点,例如消失和爆炸梯度和长期依赖性问题。同时,基于自我关注的变压器在文本处理和语音识别等顺序建模中取得了巨大成功。虽然变压器鹰过程(THP)已经获得了巨大的性能改进,但是THP不会有效地利用异步事件中的时间信息,因为这些异步序列,事件发生时刻与事件的类型一样重要,而传统的THPS只是转换时间信息进入位置编码并将其添加为变压器的输入。考虑到这一点,我们提出了一种新型的基于变压器的霍克斯工艺模型,暂时关注增强变压器鹰过程(TAA-THP),我们修改了传统的DOT产品注意力结构,并介绍了关注结构的时间编码。我们对多种合成和现实生活数据集进行多项实验,以验证我们提出的TAA-THP模型的性能,与现有的基线模型相比,在不同测量上实现的显着改进,包括在测试数据集上的日志可能性,并预测事件类型的准确性和发生时间。此外,通过烧蚀研究,我们通过比较模型的性能和没有时间关注的模型的性能,生动地证明了引入额外的时间关注的优点。
translated by 谷歌翻译
抽象的。我们遇到的大多数真实世界数据都是异步事件序列,因此过去几十年的特点是在社交网络,电子医疗记录和金融交易领域实施各种点进程。在开始时,霍克斯过程及其变体可以同时模拟复杂序列中不同事件之间的自触发和相互触发模式,以清晰和定量的方式更受欢迎.Later On,随着神经网络的进步,神经网络的进程陆续提出,逐渐成为一个研究热点。变压器鹰过程(THP)的提议取得了巨大的性能改进,因此掀起了基于变压器的神经鹰过程的新升级。但是,THP不会充分利用异步事件序列中发生的发生时间和事件类型的信息。它只是添加了事件类型转换的编码和将时间转换的位置编码到源编码。与此同时,从单个变压器构建的学习者将导致偏差不可避免。为了缓解这些问题,我们提出了一个三变形率霍克斯进程(TRI-THP)模型,其中将事件和时间信息作为辅助信息添加到DOT-Product Idition中,以形成新的多回力。 TRI-THP的有效性由一系列关于现实世界和合成数据的一系列精心设计的实验证明。
translated by 谷歌翻译
The development of social media user stance detection and bot detection methods rely heavily on large-scale and high-quality benchmarks. However, in addition to low annotation quality, existing benchmarks generally have incomplete user relationships, suppressing graph-based account detection research. To address these issues, we propose a Multi-Relational Graph-Based Twitter Account Detection Benchmark (MGTAB), the first standardized graph-based benchmark for account detection. To our knowledge, MGTAB was built based on the largest original data in the field, with over 1.55 million users and 130 million tweets. MGTAB contains 10,199 expert-annotated users and 7 types of relationships, ensuring high-quality annotation and diversified relations. In MGTAB, we extracted the 20 user property features with the greatest information gain and user tweet features as the user features. In addition, we performed a thorough evaluation of MGTAB and other public datasets. Our experiments found that graph-based approaches are generally more effective than feature-based approaches and perform better when introducing multiple relations. By analyzing experiment results, we identify effective approaches for account detection and provide potential future research directions in this field. Our benchmark and standardized evaluation procedures are freely available at: https://github.com/GraphDetec/MGTAB.
translated by 谷歌翻译
Interview has been regarded as one of the most crucial step for recruitment. To fully prepare for the interview with the recruiters, job seekers usually practice with mock interviews between each other. However, such a mock interview with peers is generally far away from the real interview experience: the mock interviewers are not guaranteed to be professional and are not likely to behave like a real interviewer. Due to the rapid growth of online recruitment in recent years, recruiters tend to have online interviews, which makes it possible to collect real interview data from real interviewers. In this paper, we propose a novel application named EZInterviewer, which aims to learn from the online interview data and provides mock interview services to the job seekers. The task is challenging in two ways: (1) the interview data are now available but still of low-resource; (2) to generate meaningful and relevant interview dialogs requires thorough understanding of both resumes and job descriptions. To address the low-resource challenge, EZInterviewer is trained on a very small set of interview dialogs. The key idea is to reduce the number of parameters that rely on interview dialogs by disentangling the knowledge selector and dialog generator so that most parameters can be trained with ungrounded dialogs as well as the resume data that are not low-resource. Evaluation results on a real-world job interview dialog dataset indicate that we achieve promising results to generate mock interviews. With the help of EZInterviewer, we hope to make mock interview practice become easier for job seekers.
translated by 谷歌翻译
Dynamic treatment regimes assign personalized treatments to patients sequentially over time based on their baseline information and time-varying covariates. In mobile health applications, these covariates are typically collected at different frequencies over a long time horizon. In this paper, we propose a deep spectral Q-learning algorithm, which integrates principal component analysis (PCA) with deep Q-learning to handle the mixed frequency data. In theory, we prove that the mean return under the estimated optimal policy converges to that under the optimal one and establish its rate of convergence. The usefulness of our proposal is further illustrated via simulations and an application to a diabetes dataset.
translated by 谷歌翻译
Temporal sentence grounding (TSG) aims to identify the temporal boundary of a specific segment from an untrimmed video by a sentence query. All existing works first utilize a sparse sampling strategy to extract a fixed number of video frames and then conduct multi-modal interactions with query sentence for reasoning. However, we argue that these methods have overlooked two indispensable issues: 1) Boundary-bias: The annotated target segment generally refers to two specific frames as corresponding start and end timestamps. The video downsampling process may lose these two frames and take the adjacent irrelevant frames as new boundaries. 2) Reasoning-bias: Such incorrect new boundary frames also lead to the reasoning bias during frame-query interaction, reducing the generalization ability of model. To alleviate above limitations, in this paper, we propose a novel Siamese Sampling and Reasoning Network (SSRN) for TSG, which introduces a siamese sampling mechanism to generate additional contextual frames to enrich and refine the new boundaries. Specifically, a reasoning strategy is developed to learn the inter-relationship among these frames and generate soft labels on boundaries for more accurate frame-query reasoning. Such mechanism is also able to supplement the absent consecutive visual semantics to the sampled sparse frames for fine-grained activity understanding. Extensive experiments demonstrate the effectiveness of SSRN on three challenging datasets.
translated by 谷歌翻译
Human parsing aims to partition humans in image or video into multiple pixel-level semantic parts. In the last decade, it has gained significantly increased interest in the computer vision community and has been utilized in a broad range of practical applications, from security monitoring, to social media, to visual special effects, just to name a few. Although deep learning-based human parsing solutions have made remarkable achievements, many important concepts, existing challenges, and potential research directions are still confusing. In this survey, we comprehensively review three core sub-tasks: single human parsing, multiple human parsing, and video human parsing, by introducing their respective task settings, background concepts, relevant problems and applications, representative literature, and datasets. We also present quantitative performance comparisons of the reviewed methods on benchmark datasets. Additionally, to promote sustainable development of the community, we put forward a transformer-based human parsing framework, providing a high-performance baseline for follow-up research through universal, concise, and extensible solutions. Finally, we point out a set of under-investigated open issues in this field and suggest new directions for future study. We also provide a regularly updated project page, to continuously track recent developments in this fast-advancing field: https://github.com/soeaver/awesome-human-parsing.
translated by 谷歌翻译
A storyboard is a roadmap for video creation which consists of shot-by-shot images to visualize key plots in a text synopsis. Creating video storyboards however remains challenging which not only requires association between high-level texts and images, but also demands for long-term reasoning to make transitions smooth across shots. In this paper, we propose a new task called Text synopsis to Video Storyboard (TeViS) which aims to retrieve an ordered sequence of images to visualize the text synopsis. We construct a MovieNet-TeViS benchmark based on the public MovieNet dataset. It contains 10K text synopses each paired with keyframes that are manually selected from corresponding movies by considering both relevance and cinematic coherence. We also present an encoder-decoder baseline for the task. The model uses a pretrained vision-and-language model to improve high-level text-image matching. To improve coherence in long-term shots, we further propose to pre-train the decoder on large-scale movie frames without text. Experimental results demonstrate that our proposed model significantly outperforms other models to create text-relevant and coherent storyboards. Nevertheless, there is still a large gap compared to human performance suggesting room for promising future work.
translated by 谷歌翻译
In the new era of personalization, learning the heterogeneous treatment effect (HTE) becomes an inevitable trend with numerous applications. Yet, most existing HTE estimation methods focus on independently and identically distributed observations and cannot handle the non-stationarity and temporal dependency in the common panel data setting. The treatment evaluators developed for panel data, on the other hand, typically ignore the individualized information. To fill the gap, in this paper, we initialize the study of HTE estimation in panel data. Under different assumptions for HTE identifiability, we propose the corresponding heterogeneous one-side and two-side synthetic learner, namely H1SL and H2SL, by leveraging the state-of-the-art HTE estimator for non-panel data and generalizing the synthetic control method that allows flexible data generating process. We establish the convergence rates of the proposed estimators. The superior performance of the proposed methods over existing ones is demonstrated by extensive numerical studies.
translated by 谷歌翻译