Learning by playing
Getting children (and adults) to tidy up after themselves can be a challenge, but we face an even greater challenge trying to get our AI agents to do the same. Success depends on the mastery of several core visuo-motor skills: approaching an object, grasping and lifting it, opening a box and putting things inside of it. To make matters more complicated, these skills must be applied in the right sequence.Control tasks, like tidying up a table or stacking objects, require an agent to determine how, when and where to coordinate the nine joints of its simulated arms and fingers to move correctly and achieve its objective. The sheer number of possible combinations of movements at any given time, along with the need to carry out a long sequence of correct actions constitute a serious exploration problemmaking this a particularly interesting area for reinforcement learning research.Techniques like reward shaping, apprenticeship learning or learning from demonstrations can help with the exploration problem. However, these methods rely on a considerable amount of knowledge about the taskthe problem of learning complex control problems from scratch with minimal prior knowledge is still an open challenge.Our new paper proposes a new learning paradigm called Scheduled Auxiliary Control (SAC-X) which seeks to overcome this exploration issue.Read More
Related Google News:
- Scaling deep retrieval with TensorFlow Recommenders and Vertex AI Matching Engine May 1, 2023
- Seeing the World: Vertex AI Vision Developer Toolkit May 1, 2023
- Google Cloud and Equinix: Building Excellence in ML Operations (MLOps) May 1, 2023
- Framing up FinOps: All about Google Cloud billing tools May 1, 2023
- Google at ICLR 2023 May 1, 2023
- Google Workspace Updates Weekly Recap - April 28, 2023 April 28, 2023
- Full HD in Google Meet live streams April 28, 2023
- New Alert Center notifications for Apple push certificates April 28, 2023