Largely preprogrammed for their tasks, and most commercial robot applications still have a very limited ability to interact and physically engage with humans.In other words, something as simple as cutting bread is still challenging for robots. And there are still no robots that can wash the dishes.
provide a set of helpful services to accelerate software development. Intelligent code completion is one of the most useful features in IDEs, which suggests next probable code tokens, such as method calls or object fields, based on existing code in the context. Traditionally, code completion relies heavily on compile-time type information to predict next tokens.But there has also been some success for dynamically-typed languages, where researchers have treated these languages as natural languages, and trained code completion systems with the help of large codebases (e.g. GitHub):
In particular, neural language models such as Recurrent Neural Networks (RNNs) can capture sequential distributions and deep semantics.Here the authors propose their own system, based on improvements on existing techniques. A system which turns out to be quite effective.
Genetic programming is an effective technique for inductive synthesis of programs from tests, i.e. training examples of desired input-output behavior. (But) Programs synthesized in this way are not guaranteed to generalize beyond the training set...The authors then describe an improvement over such a (too) ''simplistic'' Genetic Algorithm approach. An improvement, that synthesizes correct programs fast, using few examples.
This work may pave the way for effective hybridization of heuristic search methods like GP with spec-based synthesis.For more about the context for this work, see also the Sygus Competition.
In both deep learning (DL) and deep reinforcement learning (DRL), training results in a function f that has a fixed structure (given by a deep neural network).
...
In both DL and DRL, the most common algorithm for minimizing the error function is stochastic gradient descent where the parameter vector θ is modified incrementally by taking steps in the direction of the gradient.
Solvers take a convenient description of a particular model instance and automatically compute its solution. For a classical planner, the inputs are classical planning problems and the output is a plan that solves the problem.I.e.
Learners require training, which is often slow, but then are fast; solvers can deal with new problems with no training but after some deliberation. Solvers are thus general, domain independent as they are called...Psychologist Daniel Kahneman refers to this as ''slow and fast thinking''.
... provided a suitable representation of the problems; learners need experience on related problems.
A key restriction of learners relying on neural networks is that the size of their inputs x is fixed. This implies that learners cannot emulate solvers even over specific domains.Still, the speed of the learners is obviously a good thing. Just as explanation and accountability through models (solvers) is a good thing.
The solutions to some of the bottlenecks faced by each approach, that often have to do with representational issues, may lie in the integration that is required to resemble the integration of Systems 1 and 2 in the human mind.An awesome talk!
A similarly fundamental importance as Physics or Biology. To have a name for talking about it, we used ''IPsI-science'' (or ''Intellectics'' suggested by the author more than three decades ago). While Physics is concerned with matter and Biology with life, IPsI-science deals with IPsI-stuff (ie. informational, psychological and intellectual stuff).Of course ...!
The long term goal is to build an intelligent Angry Birds playing agent that can play new levels better than the best human players.
This is a very difficult problem as it requires agents to predict the outcome of physical actions without having complete knowledge of the world, and then to select a good action out of infinitely many possible actions. This is an essential capability of future AI systems that interact with the physical world.
The Angry Birds AI competition provides a simplified and controlled environment for developing and testing these capabilities.
Deep neural networks are now better than humans at tasks such as face recognition and object recognition. They've mastered the ancient game of Go.According to Tegmark, the answer is that the Universe is governed by a tiny subset of all possible functions/laws.
Actually, a bit puzzling, as there are orders of magnitude more mathematical functions than possible networks to approximate them. And yet deep neural networks somehow get the right answer...
Meaning that deep neural networks don't have to approximate any possible mathematical function, only a tiny subset of them.For more, see ''Link Between Deep Neural Networks and the Nature of the Universe'' and ''Why does deep and cheap learning work so well?''.
To build a conversational system with moderate intelligence is challenging, and requires abundant dialogue data and interdisciplinary techniques...Generally speaking, we have two types of conversational AI. One is focussed on helping the user with a particular task. Buying a ticket, reserving a seat etc. Other chatbots deal with conversations in (more) open domains, where the hope is that they will entertain us or give us emotional companionship.
Ideally, the intelligent conversational AI should be able to output grammatical, coherent responses that are diverse and interesting.Which is a difficult task, even for humans...
Nelson Mandela had died aged 95 in Johannesburg.It might be easy for us to see that these two sentences, and events, are related. But clearly not so easy for an AI program to see:
The world has lost a great man.
Event coreference resolution consists of grouping together the text expressions that refer to real-world events (also called event mentions) into a set of clusters such that all the mentions from the same cluster correspond to a unique event.The tricky thing is that:
Event mentions are more diverse syntactic objects, including not only noun phrases, but also verb phrases, sentences and arguably whole paragraphs...Making this a very difficult problem. But, rather amazingly, some results have been achieved.
Physically embodied agents offer a way to learn to understand natural language in the context of the world to which it refers.Connecting symbols (linguistic tokens and learning) to real-world percepts and actions is at the core of giving meaning to these symbols.
...
Human language does not exist in isolation; it is learned, understood, and applied in the physical world in which people exist.
Language-using robots must learn how words are grounded in the noisy, perceptual world in which a robot operates... Still, natural language systems can benefit from the rich contextual information provided by sensor data about the world.A super interesting talk!
Following the order that a human gives -- seems like a good property for a robot to have. But, we humans are not perfect and we may give orders that are not best aligned to our preferences. We show that when a human is not perfectly rational then a robot that tries to infer and act according to the human's underlying preferences can always perform better.Self-driving cars that always ''play-nice'' might never be able to switch lanes on a motorway...
Both the first industrial revolution, from 1760 up to 1840, and the second, from 1840 up to 1914, brought a significant change in working hours...This tendency of decreasing work hours from 1870 led people, like Keynes, to predict a four-hour working day (i.e. a 20-hour working week). But that, of course, never materialized.
First, going from 180 days per year and 11 hours per day, to almost full year working and a 69-hour week and after the first revolution.
Starting from 1870 data shows a consistent drop of working hours. A trend that lasted until 1980 (when we observe a rise in working hours again, especially in the new world).
Given our arguments above, we think the logical outcome of improved automation is an increase in working hours, a continuation of an existing trend.
A computational approach to learning whereby an agent tries to maximize the total amount of reward it receives when interacting with a complex, uncertain environment.Some of the key ideas can be found in Sutton and Barto's RL book.
Klopf recognized that essential aspects of adaptive behaviour were being lost as learning researchers came to focus almost exclusely on supervised learning. What was missing, according to Klopf, were the hedonic aspects of behaviour, the drive to achieve some result from the environment, to control the environment towards desired ends and away from undesired ends. This is the esential idea of trial-and-error learning.And reinforcement learning is therefore obviously (also) relevant in disciplines like machine learning, neuroscience, behaviorist psychology etc.:
Satisfaction to the animal will, other things being equal, be more firmly connected with the situation, so that, when it recurs, they will be more likely to recur; those which are accompanied or closely followed by discomfort to the animal will, other things being equal, have their connections with that situation weakened.Which brings us to ''learning with a critic'':
Where the ''critic'' returns the reinforcement, which evaluates the quality of the action (in the environment) taken by the agent depending on its actual state.Some reinforcement learning milestones:
Aamas 2014 | Areadne 2014
Nasslli 2012 | WCE 2013 | CogSci 2013 | CogSci 2014
About www.simonlaub.net | Site Index | Post Index | NeuroSky | Connections | Future Minds | Mind Design | Contact Info
© August 2018 Simon Laub - www.simonlaub.dk - www.simonlaub.net - simonlaub.com
Original page design - August 10th 2018. Simon Laub - Aarhus, Denmark, Europe.