Researchers from Massachusetts Institute of Technology (MIT) and Google have created separate algorithms which have the ability to use sight, sound, and text to communicate with humans and interact in different environments more effectively. This would give artificial intelligence another progression towards freedom in a new development.
In two public papers, the researchers at Google said: “Deep learning yields great results across many fields, from speech recognition, image classification, to translation. We present a single model that yields good results on a number of problems spanning multiple domains.”
The researchers believe that with the algorithms’ new-found abilities, they will be potentially able to teach other AI algorithms communication skills using deep learning, in future research studies in the field with little human interference.
An MIT research paper noted: “The goal is to create representations that are robust in another way: we learn representations that are aligned across modality. We believe aligned cross-modal representations will have a large impact in computer vision because they are fundamental components for machine perception to understand relationships between modalities.”
The reason for the tests was to check whether the algorithms could identify and impart utilizing human senses.
The trials were led in basic stages with the goal that the algorithms knew unambiguously what was being told while recognizing different senses and how they could react to them properly in a genuine situation.
A typical case was the AI’s capacity of recognizing and deconstructing different types of data it was given with the goal that it could respond accordingly. These included sounds and pictures of vehicles, people, and individuals with portrayals of their appearances and activities being conferred.
The two separate studies adopted comparative yet different approaches to the research. Google focused on AI interpretation between dialects, while MIT examined how the AI could develop sentences.
Google researchers explained: “We believe that this treads a path towards interesting future work on more deep learning architectures, especially since our model shows transfer learning from tasks with a large amount of available data to ones where the data is limited.”
Recently AI researchers at Facebook by chance found that algorithms were able to communicate in a machine-type language until it was amended back to human. Currently, AI does not have the ability to perform more than one of the five human senses in practical tests. But it is hoped that the experiments conducted by MIT and Google have driven forward these next stages of AI development.