This is a continuation of our recent paper in which we developed the theory of sequential parametrized motion planning. A sequential parametrized motion planning algorithm produced a motion of the system which is required to visit a prescribed sequence of states, in a certain order, at specified moments of time. In the previous publication we analysed the sequential parametrized topological complexity of the Fadell - Neuwirth fibration which in relevant to the problem of moving multiple robots avoiding collisions with other robots and with obstacles in the Euclidean space. Besides, in the preceeding paper we found the sequential parametrised topological complexity of the Fadell - Neuwirth bundle for the case of the Euclidean space $\Bbb R^d$ of odd dimension as well as the case $d=2$. In the present paper we give the complete answer for an arbitrary $d\ge 2$ even. Moreover, we present an explicit motion planning algorithm for controlling multiple robots in $\Bbb R^d$ having the minimal possible topological complexity; this algorithm is applicable to any number $n$ of robots and any number $m\ge 2$ of obstacles.
translated by 谷歌翻译
The number of international benchmarking competitions is steadily increasing in various fields of machine learning (ML) research and practice. So far, however, little is known about the common practice as well as bottlenecks faced by the community in tackling the research questions posed. To shed light on the status quo of algorithm development in the specific field of biomedical imaging analysis, we designed an international survey that was issued to all participants of challenges conducted in conjunction with the IEEE ISBI 2021 and MICCAI 2021 conferences (80 competitions in total). The survey covered participants' expertise and working environments, their chosen strategies, as well as algorithm characteristics. A median of 72% challenge participants took part in the survey. According to our results, knowledge exchange was the primary incentive (70%) for participation, while the reception of prize money played only a minor role (16%). While a median of 80 working hours was spent on method development, a large portion of participants stated that they did not have enough time for method development (32%). 25% perceived the infrastructure to be a bottleneck. Overall, 94% of all solutions were deep learning-based. Of these, 84% were based on standard architectures. 43% of the respondents reported that the data samples (e.g., images) were too large to be processed at once. This was most commonly addressed by patch-based training (69%), downsampling (37%), and solving 3D analysis tasks as a series of 2D tasks. K-fold cross-validation on the training set was performed by only 37% of the participants and only 50% of the participants performed ensembling based on multiple identical models (61%) or heterogeneous models (39%). 48% of the respondents applied postprocessing steps.
translated by 谷歌翻译
State-of-the-art object detectors are fast and accurate, but they require a large amount of well annotated training data to obtain good performance. However, obtaining a large amount of training annotations specific to a particular task, i.e., fine-grained annotations, is costly in practice. In contrast, obtaining common-sense relationships from text, e.g., "a table-lamp is a lamp that sits on top of a table", is much easier. Additionally, common-sense relationships like "on-top-of" are easy to annotate in a task-agnostic fashion. In this paper, we propose a probabilistic model that uses such relational knowledge to transform an off-the-shelf detector of coarse object categories (e.g., "table", "lamp") into a detector of fine-grained categories (e.g., "table-lamp"). We demonstrate that our method, RelDetect, achieves performance competitive to finetuning based state-of-the-art object detector baselines when an extremely low amount of fine-grained annotations is available ($0.2\%$ of entire dataset). We also demonstrate that RelDetect is able to utilize the inherent transferability of relationship information to obtain a better performance ($+5$ mAP points) than the above baselines on an unseen dataset (zero-shot transfer). In summary, we demonstrate the power of using relationships for object detection on datasets where fine-grained object categories can be linked to coarse-grained categories via suitable relationships.
translated by 谷歌翻译
Learning classifiers using skewed or imbalanced datasets can occasionally lead to classification issues; this is a serious issue. In some cases, one class contains the majority of examples while the other, which is frequently the more important class, is nevertheless represented by a smaller proportion of examples. Using this kind of data could make many carefully designed machine-learning systems ineffective. High training fidelity was a term used to describe biases vs. all other instances of the class. The best approach to all possible remedies to this issue is typically to gain from the minority class. The article examines the most widely used methods for addressing the problem of learning with a class imbalance, including data-level, algorithm-level, hybrid, cost-sensitive learning, and deep learning, etc. including their advantages and limitations. The efficiency and performance of the classifier are assessed using a myriad of evaluation metrics.
translated by 谷歌翻译
Large language models (LLMs) have been shown to be able to perform new tasks based on a few demonstrations or natural language instructions. While these capabilities have led to widespread adoption, most LLMs are developed by resource-rich organizations and are frequently kept from the public. As a step towards democratizing this powerful technology, we present BLOOM, a 176B-parameter open-access language model designed and built thanks to a collaboration of hundreds of researchers. BLOOM is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total). We find that BLOOM achieves competitive performance on a wide variety of benchmarks, with stronger results after undergoing multitask prompted finetuning. To facilitate future research and applications using LLMs, we publicly release our models and code under the Responsible AI License.
translated by 谷歌翻译
Spiking Neural Networks (SNNs) are bio-plausible models that hold great potential for realizing energy-efficient implementations of sequential tasks on resource-constrained edge devices. However, commercial edge platforms based on standard GPUs are not optimized to deploy SNNs, resulting in high energy and latency. While analog In-Memory Computing (IMC) platforms can serve as energy-efficient inference engines, they are accursed by the immense energy, latency, and area requirements of high-precision ADCs (HP-ADC), overshadowing the benefits of in-memory computations. We propose a hardware/software co-design methodology to deploy SNNs into an ADC-Less IMC architecture using sense-amplifiers as 1-bit ADCs replacing conventional HP-ADCs and alleviating the above issues. Our proposed framework incurs minimal accuracy degradation by performing hardware-aware training and is able to scale beyond simple image classification tasks to more complex sequential regression tasks. Experiments on complex tasks of optical flow estimation and gesture recognition show that progressively increasing the hardware awareness during SNN training allows the model to adapt and learn the errors due to the non-idealities associated with ADC-Less IMC. Also, the proposed ADC-Less IMC offers significant energy and latency improvements, $2-7\times$ and $8.9-24.6\times$, respectively, depending on the SNN model and the workload, compared to HP-ADC IMC.
translated by 谷歌翻译
主动脉(COA)患者特异性计算流体动力学(CFD)研究的目的 - 在资源约束设置中的研究受到可用成像方式和速度数据采集的可用成像方式的限制。多普勒超声心动图被视为合适的速度获取方式,因为其可用性和安全性较高。这项研究旨在调查经典机器学习(ML)方法的应用,以创建一种适当且可靠的方法,用于从多普勒超声心动图图像中获得边界条件(BCS),用于使用CFD进行血液动力学建模。方法 - 我们提出的方法结合了ML和CFD,以模拟感兴趣区域内的血流动力学流动。该方法的关键特征是使用ML模型来校准CFD模型的入口和出口边界条件(BCS)。 ML模型的关键输入变量是患者心率,因为这是研究中测得的血管的时间变化的参数。在研究的CFD组件中使用ANSYS Fluent,而Scikit-Learn Python库则用于ML分量。结果 - 我们在干预前对严重COA的真实临床案例进行了验证。将我们的模拟的最大缩回速度与从研究中使用的几何形状获得的患者获得的测量最大骨质速度进行了比较。在用于获得BCS的5 mL模型中,顶部模型在测得的最大骨质速度的5 \%之内。结论 - 该框架表明,它能够考虑在测量之间考虑患者心率的变化。因此,当在每个血管上缩放心率时,可以在生理上逼真的BC计算,同时提供合理准确的溶液。
translated by 谷歌翻译
将用户搜索查询与广告商实时竞标相关的关键字匹配是赞助搜索中的一个至关重要问题。在文献中,已经探索了两种广泛的方法来解决此问题:(i)在共享空间中学习查询和出价关键字的密集检索(DR),以及(ii)自然语言生成(NLG) - 学会直接生成给定查询的投标关键字。在这项工作中,我们首先对这两种方法进行了实证研究,并表明它们提供了添加剂的补充优势。特别是,从NLG检索到的很大一部分的关键字尚未由DR和反之亦然。然后,我们证明有可能将这两种方法的优势有效地结合在一个模型中。具体而言,我们提出了心脏:一种新型的多任务融合框架,在该框架中,我们共同优化共享编码器以同时执行DR和非自动性NLG。通过对30多个跨越20多种语言的搜索查询进行的广泛实验,我们表明,与使用相同的GPU计算的基线方法相比,心脏检索高质量的出价关键字40.3%。我们还证明,在单个心脏模型上推断与在两种不同的DR和NLG基线模型上推断为2倍计算一样好。此外,我们表明,接受心脏目标训练的DR模型要比接受标准对比度损失功能的训练的模型要好得多。最后,我们表明我们的心目标可以用于除赞助搜索并实现显着绩效提高以外的短文本检索任务。
translated by 谷歌翻译
联合学习(FL)是最近开发的机器学习领域,其中大量分布式客户端的私人数据用于在中央服务器协调下开发全球模型,而无需明确暴露数据。标准的FL策略具有许多重要的瓶颈,包括庞大的沟通要求和对客户资源的高影响。文献中已经描述了一些试图解决这些问题的策略。在本文中,提出了一个基于“模型生长”概念的新方案。最初,服务器部署了低复杂性的小型模型,该模型经过训练,可以在最初的回合中捕获数据复杂性。当这种模型的性能饱和时,服务器会借助保留功能的转换切换到较大的模型。随着客户处理更多数据的处理,模型的复杂性会增加,并且整体过程一直持续到达到所需的性能为止。因此,最复杂的模型仅在我们的方法的最后阶段进行广播,从而大大降低了通信成本和客户计算要求。该方法对三个标准基准进行了广泛的测试,并证明可以大大降低通信和客户计算,同时与当前最有效的策略相比实现了可比的准确性。
translated by 谷歌翻译
在其表示中,已经发现接受过文本数据训练的神经网络模型编码不希望的语言或敏感属性。删除此类属性是不平凡的,因为属性,文本输入和学习的表示之间存在复杂的关系。最近的工作提出了事后和对抗方法,以从模型的表示中删除此类不需要的属性。通过广泛的理论和经验分析,我们表明这些方法可以适得其反:它们无法完全删除属性,在最坏的情况下,最终可能会破坏所有与任务相关的功能。原因是方法对探测分类器的依赖作为属性的代理。即使在最有利的条件下,当属性在表示空间中的特征可以提供100%的学习探测分类器时,我们证明事后或对抗方法将无法正确删除属性。这些理论含义通过经验实验在合成,多NLI和Twitter数据集的模型上证实。对于敏感的属性去除(例如公平性),我们建议您谨慎使用这些方法,并提出伪造度量,以评估最终分类器的质量。
translated by 谷歌翻译