We are announcing the release of our state-of-the-art off-policy model-free reinforcement learning algorithm, soft actor-critic (SAC). This algorithm has been developed jointly at UC Berkeley and Google Brain, and we have been using it internally for our robotics experiment. Soft actor-critic is, to our knowledge, one of the most efficient model-free algorithms available today, making it especially well-suited for real-world robotic learning. In this post, we will benchmark SAC against state-of-the-art model-free RL algorithms and showcase a spectrum of real-world robot examples, ranging from manipulation to locomotion. We also release our implementation of SAC, which is particularly designed for real-world robotic systems.
Document worth reading: “Small Sample Learning in Big Data Era”
As a promising area in artificial intelligence, a new learning paradigm, called Small Sample Learning (SSL), has been attracting prominent research attention in the recent years. In this paper, we aim to present a survey to comprehensively introduce the current techniques proposed on this topic. Specifically, current SSL techniques can be mainly divided into two categories. The first category of SSL approaches can be called ‘concept learning’, which emphasizes learning new concepts from only few related observations. The purpose is mainly to simulate human learning behaviors like recognition, generation, imagination, synthesis and analysis. The second category is called ‘experience learning’, which usually co-exists with the large sample learning manner of conventional machine learning. This category mainly focuses on learning with insufficient samples, and can also be called small data learning in some literatures. More extensive surveys on both categories of SSL techniques are introduced and some neuroscience evidences are provided to clarify the rationality of the entire SSL regime, and the relationship with human learning process. Some discussions on the main challenges and possible future research directions along this line are also presented. Small Sample Learning in Big Data Era
LoyaltyOne: Manager, CPG [Westborough, MA]
At: LoyaltyOne Location: Westborough, MAWeb: www.loyalty.comPosition: Manager, CPG
NLP Breakthrough Imagenet Moment has arrived
By Sebastian Ruder, Insight Centre.
If you did not already know
Algebraic Machine Learning Machine learning algorithms use error function minimization to fit a large set of parameters in a preexisting model. However, error minimization eventually leads to a memorization of the training dataset, losing the ability to generalize to other datasets. To achieve generalization something else is needed, for example a regularization method or stopping the training when error in a validation dataset is minimal. Here we propose a different approach to learning and generalization that is parameter-free, fully discrete and that does not use function minimization. We use the training data to find an algebraic representation with minimal size and maximal freedom, explicitly expressed as a product of irreducible components. This algebraic representation is shown to directly generalize, giving high accuracy in test data, more so the smaller the representation. We prove that the number of generalizing representations can be very large and the algebra only needs to find one. We also derive and test a relationship between compression and error rate. We give results for a simple problem solved step by step, hand-written character recognition, and the Queens Completion problem as an example of unsupervised learning. As an alternative to statistical learning, \enquote{algebraic learning} may offer advantages in combining bottom-up and top-down information, formal concept derivation from data and large-scale parallelization. …
Top Stories of 2018: 9 Must-have skills you need to become a Data Scientist, updated; Python eats away at R: Top Software for Analytics, Data Science, Machine Learning in 2018
Spark + AI Summit: learn best practices in ML and DL, latest frameworks, and more – special KDnuggets offer
For 2019, we added a new technical track called Databricks Tech Talks. Learn how data professionals are applying the Unified Analytics Platform in real-life use cases. Sessions will be technical and include demos. This is your Spark moment. Seize the biggest discount offered:$700 off for the year end.
Pdftools 2.0: powerful pdf text extraction tools
rOpenSci - open tools for open science
发表于
Whats new on arXiv
Text Data Augmentation Made Simple By Leveraging NLP Cloud APIs
Implementing ResNet with MXNET Gluon and Comet.ml for Image Classification
By Cecelia Shao, Comet.ml