Latest news about Bitcoin and all cryptocurrencies. Your daily crypto news habit.
As the playoffs are in full swing, everyone is littering their speech with baseball apologies, âWe are only at the bottom of the first inning for all these technologies,â declared Pano Anthos at last weekâs RoboLab forum. When illustrating his point, Anthos deconstructed one of XRC Labsâ portfolio companies which utilizes computer vision to identify store merchandise. âRight now the technology is not very smart, it requires low paid workers in Asia to label the images in the database to work,â explains Anthos. Ken Pilot later reiterated this sentiment in sharing his experiences with leading brands that are still trying to comprehend the impact of mobile, much less unmanned, systems. âThe store of the future,â describes Pilot, âwill be as convenient as Amazon is for online consumers, enabling anyone to find what they want, when they want, without asking.â While that is the goal, both luminaries agreed there will be plenty of missteps along the way with companies deploying technology just for the sake of it. Ending the evening, almost on cue, the two speakers jokingly pointed at a robot store associate as an example of the current state of silliness.
The sun maybe setting on what David Letterman would call, âStupid Robot Tricks,â as intelligent machines are beginning to surpass humans in a wide variety of manual and intellectual pursuits. In March 2016, Googleâs Deep Mind software program AlphaGo defeated the reining Go champion Lee Sedol. Go, a Chinese game that originated more than 3,000 years ago, is said to be googol times more complex than chess. Lee was previously considered the greatest player in the past decade with 18 world titles. Today, AlphaGo holds the ranking title.
Deconstructing how the Deep Mind team was able to cross a once-impossible threshold for computer scientists could provide a primer on the tools available to roboticists. According to the AlphaGo website, âTraditional AI methods, which construct a search tree over all possible positions, donât have a chance in Go. This is because of the sheer number of possible moves and the difficulty of evaluating the strength of each possible board position.â Instead the researchers combined the traditional search tree approach with a deep learning system. âOne neural network, the âpolicy networkâ, selects the next move to play. The other neural network, the âvalue networkâ, predicts the winner of the game.â However, the key of AlphaGo was having the AI go through a rigorous approach of âreinforcement learning,â where it plays itself thousands of times from the database of games. âWe showed AlphaGo a large number of strong amateur games to help it develop its own understanding of what reasonable human play looks like. Then we had it play against different versions of itself thousands of times, each time learning from its mistakes and incrementally improving until it became immensely strong.â
By October 2017, the AI became so powerful it bypassed the reinforcement learning process that contained human input of professional and amateur games to only play earlier versions of itself. The new program, AlphaGo Zero, beat the previous one that defeated Lee Sedol months earlier by 100 games to 0, making it the greatest Go player in history. Deep Mind is now looking to apply this logic to âA wide set of structured problems that share similar properties to a game like Go, such as planning tasks or problems where a series of actions have to be taken in the correct sequence. Examples could include protein folding, reducing energy consumption or searching for revolutionary new materials.â
Reinforcement learning techniques are not limited to games of strategy, last week researchers at the University of Californiaâs Berkeley Artificial Intelligence Research (BAIR) Lab presented a paper using YouTube videos to train humanoids in mimicking movements. Utilizing a similar methodology as AlphaGo, the BAIR team developed a deep learning neural network that approximates the motion of actors seen online into programming steps for robots. In their post on Robohub.org the team expounded, âA staggering 300 hours of videos are uploaded to YouTube every minute. Unfortunately, it is still very challenging for our machines to learn skills from this vast volume of visual data.â In order to access this treasure trove of training data, programmers today are forced to purchasing and ferrying around bulky motion capture (mocap) equipment to create their own demonstration videos. âMocap systems also tend to be restricted to indoor environments with minimal occlusion, which can limit the types of skills that can be recorded,â says the authors, Xue Bin (Jason) Peng and Angjoo Kanazawa. Tackling this challenge, Peng and Kanazawa set out to create a seamless AI platform for unmanned systems to learn skills by unpacking hours of online video clips.
The paper states: âIn this work, we present a framework for learning skills from videos (SFV). By combining state-of-the-art techniques in computer vision and reinforcement learning, our system enables simulated characters to learn a diverse repertoire of skills from video clips. Given a single monocular video of an actor performing some skill, such as a cartwheel or a backflip, our characters are able to learn policies that reproduce that skill in a physics simulation, without requiring any manual pose annotations.â The video is fed through an agent that breaks down the movements into three stages: âpose estimation, motion reconstruction, and motion imitation.â The first stage predicts the frames following a subject initial pose. Then the âmotion reconstructionâ reorganizes these predictions into âreference motion.â The final process simulates the data with animated characters that continue to train via reinforcement learning. The SFV platform is actually an update to Peng and Kanazawaâs earlier system, DeepMimic, for using motion capture video. To date, the results have been staggering with 20 different skills acquired just from ordinary online videos, as shown below:
Peng and Kanazawa are hopeful that such simulations could be leveraged in the future to enable machines to navigate new environments: âEven though the environments are quite different from those in the original videos, the learning algorithm still develops fairly plausible strategies for handling these new environments.â The team is also optimistic about its contribution to furthering the development of mobile unmanned systems, âAll in all, our framework is really just taking the most obvious approach that anyone can think of when tackling the problem of video imitation. The key is in decomposing the problem into more manageable components, picking the right methods for those components, and integrating them together effectively.â
Humbly, the Berkley lab admits that most YouTube videos are still too complicated for their AI to imitate. Whimsically, Peng and Kanazawa single out dancing âGangnam styleâ as once of these hurdles. âWe still have all of our work ahead of us,â declares the researchers, âand we hope that this work will help inspire future techniques that will enable agents to take advantage of the massive volume of publicly available video data to acquire a truly staggering array of skills.â After all, itâs still the first inning. Batter up!
YouTube Is Making Machines Really Smart was originally published in Hacker Noon on Medium, where people are continuing the conversation by highlighting and responding to this story.
Disclaimer
The views and opinions expressed in this article are solely those of the authors and do not reflect the views of Bitcoin Insider. Every investment and trading move involves risk - this is especially true for cryptocurrencies given their volatility. We strongly advise our readers to conduct their own research when making a decision.