Transformer models have achieved great success across many NLP problems. However, previous studies in automated ICD coding concluded that these models fail to outperform some of the earlier solutions such as CNN-based models. In this paper we challenge this conclusion. We present a simple and scalable method to process long text with the existing transformer models such as BERT. We show that this method significantly improves the previous results reported for transformer models in ICD coding, and is able to outperform one of the prominent CNN-based methods.
translated by 谷歌翻译
最近的文献中的许多作品介绍了使用CNNS(卷积神经网络)来识别图像中语义特性的语义映射方法。通常预定义并限制在特定任务中通常是预定义的,属性(例如:室内大小,地方类别和对象)及其课程(例如:厨房和厨房和浴室,用于地方类别)。因此,在映射的构建期间获取和处理的所有视觉数据都丢失,并且仅在地图上仍然存在识别的语义属性。相比之下,这项工作介绍了一种拓扑语义映射方法,它使用由CNN(Googlenet)提取的深度视觉特征,从在机器人操作的环境中的多个视图中捕获的2D图像,以通过平均值,统一的视觉表示每个拓扑节点覆盖的区域中获取的功能。这些表示允许灵活地识别区域的语义属性,并在其他视觉任务中使用。具有现实世界室内数据集的实验表明,该方法能够整合区域的视觉特征,并使用它们识别对象和将类别识别为语义属性,并指示图像的拓扑位置,具有非常有前途的结果。
translated by 谷歌翻译