长期护理(LTC)居民的一半营养不良的住院治疗,死亡率,发病率较低。当前的跟踪方法是主观和耗时的。本文介绍了专为LTC设计的自动食品成像和营养进气跟踪(AFINI-T)技术。我们提出了一种用于食品分类的新型卷积Automencoder,在我们的模拟LTC食物摄入数据集上培训了用于食品分类,并在我们的模拟LTC食物摄入数据集上进行测试(每种餐路;每次最多15级;前1个分类准确度:88.9%;意味着进气错误: - 0.4 ml $ \ PM $ 36.7毫升)。营养摄入量的估计与质量的营养估计与质量($ ^ 2 $ 0.92至0.99)之间的营养估计与方法之间的良好符合($ \ sigma $ = -2.7至-0.01;零在协议的每一个限制中, 。 AFINI-T方法是深度学习的动力计算营养传感系统,可以提供更准确地和客观地跟踪LTC驻留食物摄入量的新颖手段,以支持和防止营养不良跟踪策略。
translated by 谷歌翻译
在免赠款稀疏代码多访问(GF-SCMA)系统中,主动用户检测(AUD)是一个主要的性能瓶颈,因为它涉及复杂的组合问题,这使用户和接收器的争夺资源的联合设计是至关重要的,但是一个具有挑战性的问题。为此,我们建议对编码器侧的两个序列生成网络(PGN)和解码器端的数据辅助AUD进行基于自动编码器(AE)的关节优化。提出的AE的核心体系结构是解码器中新型的用户活动提取网络(UAEN),该网络从SCMA CodeWord数据中提取先验用户活动信息,以获取数据辅助AUD。对拟议的AE进行的端到端培训可以使争夺资源的联合优化,即序列序列,每个序列,每个序列与其中一本代码书关联,并从序言和基于SCMA的数据传输中提取用户活动信息。此外,我们在端到端培训之前为UAEN提出了一个自制的预训练计划,以确保AE网络内部深处的UAEN的收敛性。仿真结果表明,与基于最先进的DL的AUD方案相比。
translated by 谷歌翻译
Covid-19流行病仍然有一个毁灭性的全球影响,并对世界各地努力努力的医疗系统带来了巨大的负担。鉴于资源有限,准确的患者三环和护理规划在对抗Covid-19的斗争中至关重要,并且在护理计划中的一个重要任务是确定患者是否应录取医院的重症监护单位(ICU)。通过对透明和值得信赖的ICU入学临床决策支持的推动,我们基于患者临床数据引入Covid-Net Clinical ICU,是ICU入学预测的神经网络。由透明信任的以信赖的方法驱动,拟议的Covid-Net临床ICU是使用来自医院Sirio-Libanes的临床数据集,包括1,925个Covid-19患者记录,并且能够预测Covid-19阳性患者要求ICU入场,准确性为96.9%,以便在持续流行下,为医院提供更好的护理计划。我们使用定量说明策略进行了系统级洞察发现,以研究不同临床特征的决策影响,并获得可操作的洞察,以提高预测性能。我们进一步利用了一套信任量化指标,以获得对Covid-Net临床ICU的可信度的更深入的见解。通过深入挖掘临床预测模型的时间和为何进行某些决策,我们可以发现决策中的关键因素,以获得关键的临床决策支持任务,如ICU准入预测,并确定可以信任临床预测模型的情况以获得更高的问责制。
translated by 谷歌翻译
Many recent works on understanding deep learning try to quantify how much individual data instances influence the optimization and generalization of a model, either by analyzing the behavior of the model during training or by measuring the performance gap of the model when the instance is removed from the dataset. Such approaches reveal characteristics and importance of individual instances, which may provide useful information in diagnosing and improving deep learning. However, most of the existing works on data valuation require actual training of a model, which often demands high-computational cost. In this paper, we provide a training-free data valuation score, called complexity-gap score, which is a data-centric score to quantify the influence of individual instances in generalization of two-layer overparameterized neural networks. The proposed score can quantify irregularity of the instances and measure how much each data instance contributes in the total movement of the network parameters during training. We theoretically analyze and empirically demonstrate the effectiveness of the complexity-gap score in finding 'irregular or mislabeled' data instances, and also provide applications of the score in analyzing datasets and diagnosing training dynamics.
translated by 谷歌翻译
While the capabilities of autonomous systems have been steadily improving in recent years, these systems still struggle to rapidly explore previously unknown environments without the aid of GPS-assisted navigation. The DARPA Subterranean (SubT) Challenge aimed to fast track the development of autonomous exploration systems by evaluating their performance in real-world underground search-and-rescue scenarios. Subterranean environments present a plethora of challenges for robotic systems, such as limited communications, complex topology, visually-degraded sensing, and harsh terrain. The presented solution enables long-term autonomy with minimal human supervision by combining a powerful and independent single-agent autonomy stack, with higher level mission management operating over a flexible mesh network. The autonomy suite deployed on quadruped and wheeled robots was fully independent, freeing the human supervision to loosely supervise the mission and make high-impact strategic decisions. We also discuss lessons learned from fielding our system at the SubT Final Event, relating to vehicle versatility, system adaptability, and re-configurable communications.
translated by 谷歌翻译
Attention mechanisms form a core component of several successful deep learning architectures, and are based on one key idea: ''The output depends only on a small (but unknown) segment of the input.'' In several practical applications like image captioning and language translation, this is mostly true. In trained models with an attention mechanism, the outputs of an intermediate module that encodes the segment of input responsible for the output is often used as a way to peek into the `reasoning` of the network. We make such a notion more precise for a variant of the classification problem that we term selective dependence classification (SDC) when used with attention model architectures. Under such a setting, we demonstrate various error modes where an attention model can be accurate but fail to be interpretable, and show that such models do occur as a result of training. We illustrate various situations that can accentuate and mitigate this behaviour. Finally, we use our objective definition of interpretability for SDC tasks to evaluate a few attention model learning algorithms designed to encourage sparsity and demonstrate that these algorithms help improve interpretability.
translated by 谷歌翻译
In robotics and computer vision communities, extensive studies have been widely conducted regarding surveillance tasks, including human detection, tracking, and motion recognition with a camera. Additionally, deep learning algorithms are widely utilized in the aforementioned tasks as in other computer vision tasks. Existing public datasets are insufficient to develop learning-based methods that handle various surveillance for outdoor and extreme situations such as harsh weather and low illuminance conditions. Therefore, we introduce a new large-scale outdoor surveillance dataset named eXtremely large-scale Multi-modAl Sensor dataset (X-MAS) containing more than 500,000 image pairs and the first-person view data annotated by well-trained annotators. Moreover, a single pair contains multi-modal data (e.g. an IR image, an RGB image, a thermal image, a depth image, and a LiDAR scan). This is the first large-scale first-person view outdoor multi-modal dataset focusing on surveillance tasks to the best of our knowledge. We present an overview of the proposed dataset with statistics and present methods of exploiting our dataset with deep learning-based algorithms. The latest information on the dataset and our study are available at https://github.com/lge-robot-navi, and the dataset will be available for download through a server.
translated by 谷歌翻译
Crowdsourcing has emerged as an effective platform to label a large volume of data in a cost- and time-efficient manner. Most previous works have focused on designing an efficient algorithm to recover only the ground-truth labels of the data. In this paper, we consider multi-choice crowdsourced labeling with the goal of recovering not only the ground truth but also the most confusing answer and the confusion probability. The most confusing answer provides useful information about the task by revealing the most plausible answer other than the ground truth and how plausible it is. To theoretically analyze such scenarios, we propose a model where there are top-two plausible answers for each task, distinguished from the rest of choices. Task difficulty is quantified by the confusion probability between the top two, and worker reliability is quantified by the probability of giving an answer among the top two. Under this model, we propose a two-stage inference algorithm to infer the top-two answers as well as the confusion probability. We show that our algorithm achieves the minimax optimal convergence rate. We conduct both synthetic and real-data experiments and demonstrate that our algorithm outperforms other recent algorithms. We also show the applicability of our algorithms in inferring the difficulty of tasks and training neural networks with the soft labels composed of the top-two most plausible classes.
translated by 谷歌翻译
Recent advances in deep learning have enabled us to address the curse of dimensionality (COD) by solving problems in higher dimensions. A subset of such approaches of addressing the COD has led us to solving high-dimensional PDEs. This has resulted in opening doors to solving a variety of real-world problems ranging from mathematical finance to stochastic control for industrial applications. Although feasible, these deep learning methods are still constrained by training time and memory. Tackling these shortcomings, Tensor Neural Networks (TNN) demonstrate that they can provide significant parameter savings while attaining the same accuracy as compared to the classical Dense Neural Network (DNN). In addition, we also show how TNN can be trained faster than DNN for the same accuracy. Besides TNN, we also introduce Tensor Network Initializer (TNN Init), a weight initialization scheme that leads to faster convergence with smaller variance for an equivalent parameter count as compared to a DNN. We benchmark TNN and TNN Init by applying them to solve the parabolic PDE associated with the Heston model, which is widely used in financial pricing theory.
translated by 谷歌翻译
Artificial neural networks can learn complex, salient data features to achieve a given task. On the opposite end of the spectrum, mathematically grounded methods such as topological data analysis allow users to design analysis pipelines fully aware of data constraints and symmetries. We introduce a class of persistence-based neural network layers. Persistence-based layers allow the users to easily inject knowledge about symmetries (equivariance) respected by the data, are equipped with learnable weights, and can be composed with state-of-the-art neural architectures.
translated by 谷歌翻译