实现接近真实机器人的高度准确的运动学或模拟器模型可以促进基于模型的控制(例如,模型预测性控制或线性质量调节器),基于模型的轨迹计划(例如,轨迹优化),并减少增强学习方法所需的学习时间。因此,这项工作的目的是学习运动学和/或模拟器模型与真实机器人之间的残余误差。这是使用自动调节和神经网络实现的,其中使用自动调整方法更新神经网络的参数,该方法应用了从无味的Kalman滤波器(UKF)公式进行方程式。使用此方法,我们仅使用少量数据对这些残差错误进行建模 - 当我们直接从硬件操作中学习改善模拟器/运动学模型时,这是必要的。我们演示了关于机器人硬件(例如操纵器组)的方法,并表明,通过学习的残差错误,我们可以进一步缩小运动学模型,模拟和真实机器人之间的现实差距。
translated by 谷歌翻译
具有单个刚体模型的凸模型预测控制(MPC)在真实的腿部机器人上表现出强烈的性能。但是,凸MPC受其假设的限制,例如旋转角度和预定义的步态,从而限制了潜在溶液的丰富性。我们删除了这些假设,并使用单个刚体模型解决了完整的混合企业非凸编程。我们首先离线收集预处理问题的数据集,然后学习问题解决方案图以快速解决MPC的优化。如果可以找到温暖的启动,则可以接近全球最优性解决离线问题。通过根据初始条件产生各种步态和行为来测试所提出的控制器。硬件测试根据传感器反馈演示了在线步态生成和适应性超过50 Hz。
translated by 谷歌翻译
尽管腿部机器人的运动计划表现出了巨大的成功,但具有灵活的多指抓握的腿部机器人的运动计划尚未成熟。我们提出了一个有效的运动计划框架,用于同时解决运动(例如,质心动力学),抓地力(例如,贴片接触)和触点(例如步态)问题。为了加速计划过程,我们建议基于乘数的交替方向方法(ADMM)提出分布式优化框架,以求解原始的大型混合构成非整数非线性编程(MINLP)。最终的框架使用混合构成二次编程(MIQP)来求解联系人和非线性编程(NLP)来求解非线性动力学,这些动力学在计算方面更可行,对参数较不敏感。此外,我们通过微蜘蛛抓手从极限表面明确执行补丁接触约束。我们在硬件实验中演示了我们提出的框架,这表明多限制机器人能够实现各种动作,包括在斜坡角度45 {\ deg}的情况下进行较短的计划时间。
translated by 谷歌翻译
本文介绍了Scalucs,这是一种四足动物,该机器人在地上,悬垂和天花板上爬上攀爬,并在地面上爬行。 Scaleer是最早的自由度四束机器人之一,可以在地球的重力下自由攀爬,也是地面上最有效的四足动物之一。在其他最先进的登山者专门攀登自己的地方,Scaleer承诺使用有效载荷\ Textit {和}地面运动实践自由攀爬,这实现了真正的多功能移动性。新的攀登步态滑冰步态通过利用缩放器的身体连锁机制来增加有效载荷。 Scaleer在地面上达到了最大归一化的运动速度,即$ 1.87 $ /s,$ 0.56 $ m /s,$ 1.2 $ /min,或$ 0.42 $ m /min /min的岩石墙攀爬。有效载荷能力达到地面上缩放器重量的233美元,垂直墙上的$ 35 $%。我们的山羊抓手是一种机械适应的两指抓手,成功地抓住了凸凸和非凸的对象,并支持缩放器。
translated by 谷歌翻译
我们展示了一个具有自动调整的入口控制器,该控制器可用于单个和多点接触机器人(例如,带有点脚或多指握把的腿部机器人)。控制器的目标是跟踪每个接触点的扳手轮廓,同时考虑旋转摩擦引起的额外扭矩。我们的接收控制器在在线操作期间具有自适应性,该方法通过自动调整方法调整了控制器的收益,同时遵循几个培训目标,以促进控制器稳定性,例如尽可能接近跟踪扳手配置文件,以确保控制输出在实力之内限制最小化滑移并避免运动学奇异性。我们使用多限制的攀登机器人来证明控制器在硬件上的鲁棒性,用于操纵和运动任务。
translated by 谷歌翻译
In the past years, deep learning has seen an increase of usage in the domain of histopathological applications. However, while these approaches have shown great potential, in high-risk environments deep learning models need to be able to judge their own uncertainty and be able to reject inputs when there is a significant chance of misclassification. In this work, we conduct a rigorous evaluation of the most commonly used uncertainty and robustness methods for the classification of Whole-Slide-Images under domain shift using the H\&E stained Camelyon17 breast cancer dataset. Although it is known that histopathological data can be subject to strong domain shift and label noise, to our knowledge this is the first work that compares the most common methods for uncertainty estimation under these aspects. In our experiments, we compare Stochastic Variational Inference, Monte-Carlo Dropout, Deep Ensembles, Test-Time Data Augmentation as well as combinations thereof. We observe that ensembles of methods generally lead to higher accuracies and better calibration and that Test-Time Data Augmentation can be a promising alternative when choosing an appropriate set of augmentations. Across methods, a rejection of the most uncertain tiles leads to a significant increase in classification accuracy on both in-distribution as well as out-of-distribution data. Furthermore, we conduct experiments comparing these methods under varying conditions of label noise. We observe that the border regions of the Camelyon17 dataset are subject to label noise and evaluate the robustness of the included methods against different noise levels. Lastly, we publish our code framework to facilitate further research on uncertainty estimation on histopathological data.
translated by 谷歌翻译
Charisma is considered as one's ability to attract and potentially also influence others. Clearly, there can be considerable interest from an artificial intelligence's (AI) perspective to provide it with such skill. Beyond, a plethora of use cases opens up for computational measurement of human charisma, such as for tutoring humans in the acquisition of charisma, mediating human-to-human conversation, or identifying charismatic individuals in big social data. A number of models exist that base charisma on various dimensions, often following the idea that charisma is given if someone could and would help others. Examples include influence (could help) and affability (would help) in scientific studies or power (could help), presence, and warmth (both would help) as a popular concept. Modelling high levels in these dimensions for humanoid robots or virtual agents, seems accomplishable. Beyond, also automatic measurement appears quite feasible with the recent advances in the related fields of Affective Computing and Social Signal Processing. Here, we, thereforem present a blueprint for building machines that can appear charismatic, but also analyse the charisma of others. To this end, we first provide the psychological perspective including different models of charisma and behavioural cues of it. We then switch to conversational charisma in spoken language as an exemplary modality that is essential for human-human and human-computer conversations. The computational perspective then deals with the recognition and generation of charismatic behaviour by AI. This includes an overview of the state of play in the field and the aforementioned blueprint. We then name exemplary use cases of computational charismatic skills before switching to ethical aspects and concluding this overview and perspective on building charisma-enabled AI.
translated by 谷歌翻译
Deep learning-based 3D human pose estimation performs best when trained on large amounts of labeled data, making combined learning from many datasets an important research direction. One obstacle to this endeavor are the different skeleton formats provided by different datasets, i.e., they do not label the same set of anatomical landmarks. There is little prior research on how to best supervise one model with such discrepant labels. We show that simply using separate output heads for different skeletons results in inconsistent depth estimates and insufficient information sharing across skeletons. As a remedy, we propose a novel affine-combining autoencoder (ACAE) method to perform dimensionality reduction on the number of landmarks. The discovered latent 3D points capture the redundancy among skeletons, enabling enhanced information sharing when used for consistency regularization. Our approach scales to an extreme multi-dataset regime, where we use 28 3D human pose datasets to supervise one model, which outperforms prior work on a range of benchmarks, including the challenging 3D Poses in the Wild (3DPW) dataset. Our code and models are available for research purposes.
translated by 谷歌翻译
This article concerns Bayesian inference using deep linear networks with output dimension one. In the interpolating (zero noise) regime we show that with Gaussian weight priors and MSE negative log-likelihood loss both the predictive posterior and the Bayesian model evidence can be written in closed form in terms of a class of meromorphic special functions called Meijer-G functions. These results are non-asymptotic and hold for any training dataset, network depth, and hidden layer widths, giving exact solutions to Bayesian interpolation using a deep Gaussian process with a Euclidean covariance at each layer. Through novel asymptotic expansions of Meijer-G functions, a rich new picture of the role of depth emerges. Specifically, we find that the posteriors in deep linear networks with data-independent priors are the same as in shallow networks with evidence maximizing data-dependent priors. In this sense, deep linear networks make provably optimal predictions. We also prove that, starting from data-agnostic priors, Bayesian model evidence in wide networks is only maximized at infinite depth. This gives a principled reason to prefer deeper networks (at least in the linear case). Finally, our results show that with data-agnostic priors a novel notion of effective depth given by \[\#\text{hidden layers}\times\frac{\#\text{training data}}{\text{network width}}\] determines the Bayesian posterior in wide linear networks, giving rigorous new scaling laws for generalization error.
translated by 谷歌翻译
In this paper we study the smooth strongly convex minimization problem $\min_{x}\min_y f(x,y)$. The existing optimal first-order methods require $\mathcal{O}(\sqrt{\max\{\kappa_x,\kappa_y\}} \log 1/\epsilon)$ of computations of both $\nabla_x f(x,y)$ and $\nabla_y f(x,y)$, where $\kappa_x$ and $\kappa_y$ are condition numbers with respect to variable blocks $x$ and $y$. We propose a new algorithm that only requires $\mathcal{O}(\sqrt{\kappa_x} \log 1/\epsilon)$ of computations of $\nabla_x f(x,y)$ and $\mathcal{O}(\sqrt{\kappa_y} \log 1/\epsilon)$ computations of $\nabla_y f(x,y)$. In some applications $\kappa_x \gg \kappa_y$, and computation of $\nabla_y f(x,y)$ is significantly cheaper than computation of $\nabla_x f(x,y)$. In this case, our algorithm substantially outperforms the existing state-of-the-art methods.
translated by 谷歌翻译