[ad_1]
<div _ngcontent-c16 = "" innerhtml = "
Discovery researchers Listen – the program looking for signs of intelligent life in the universe – detected 72 new radio bursts from distant galaxies using AI.
In 2017, the UC Berkeley listening team looking for Center for Research on Extraterrestrial Intelligence (SETI), detected 21 fast radio bursts (FRB) by analyzing more than 400 terabytes of data. Gerry Zhang and colleagues recently found 72 new bursts using convolutional neural networks to review the same dataset.
In this statement, lead researcher Andrew Siemion revealed his enthusiasm for the "potential of machine learning algorithms to detect signals missed by conventional algorithms."
"That the FRBs themselves turn out to be signatures of extraterrestrial technologies, Breakthrough Listen helps push the boundaries of a new and rapidly growing field of our understanding of the universe around us. "
Why is it important
Neural convolution networks (CNN), the technology behind object detection, are at the heart of many recent advances. To allow autonomous cars to recognize cars and pedestrians, to estimate obesity rates thanks to Google Maps images. This technology has proven to be essential in helping humans find patterns in images they would otherwise have discovered.
This technology has the potential to reconstruct old data sets of films or images useful to researchers. As Zhang and his colleagues have shown, CNNs can be used to browse older data sets to find patterns that traditional machine learning methods may lack.
How it works
Convolutional neural networks learn to represent parts of an image at different layers of abstraction. These systems are usually composed of a sequence of computing layers, each operating on the output of the previous one. For example, one can imagine a 4-layer neural network where layer 1 learns to detect contours and lines, layer 2 learns to detect basic geometric shapes and layer 3 learns to detect patterns such as wheels or outlines. The last layer learns to use the output of Layer 3 to choose the most likely category from an options list.
In this example, Zhang's team created an image dataset where an image "had at least one FRB pulse" and another image only of the noise. The team then simulated impulses and placed them on some of the noisy images. Simulation is a key technique deployed by researchers where they may not have the data they want.
Their 17-layer CNN is formed on 200,000 actual images and 200,000 false images have detected pulses with 88% accuracy and real pulses with an accuracy of 98%.
Their journal and the project website have more details about their project, including their dataset.
">
Discovery researchers Listen – the program looking for signs of intelligent life in the universe – detected 72 new radio bursts from distant galaxies using AI.
In 2017, the UC Berkeley listening team looking for Center for Research on Extraterrestrial Intelligence (SETI), detected 21 fast radio bursts (FRB) by analyzing more than 400 terabytes of data. Gerry Zhang and colleagues recently found 72 new bursts using convolutional neural networks to review the same dataset.
In this statement, lead researcher Andrew Siemion revealed his enthusiasm for the "potential of machine learning algorithms to detect signals missed by conventional algorithms."
"That the FRBs themselves turn out to be signatures of extraterrestrial technologies, Breakthrough Listen helps push the boundaries of a new and rapidly growing field of our understanding of the universe around us. "
Why is it important
Neural convolution networks (CNN), the technology behind object detection, are at the heart of many recent advances. To allow autonomous cars to recognize cars and pedestrians, to estimate obesity rates thanks to Google Maps images. This technology has proven to be essential in helping humans find patterns in images they would otherwise have discovered.
This technology has the potential to reconstruct old data sets of films or images useful to researchers. As Zhang and his colleagues have shown, CNNs can be used to browse older data sets to find patterns that traditional machine learning methods may lack.
How it works
Convolutional neural networks learn to represent parts of an image at different layers of abstraction. These systems are usually composed of a sequence of computing layers, each operating on the output of the previous one. For example, one can imagine a 4-layer neural network where layer 1 learns to detect contours and lines, layer 2 learns to detect basic geometric shapes and layer 3 learns to detect patterns such as wheels or outlines. The last layer learns to use the output of Layer 3 to choose the most likely category from an options list.
In this example, Zhang's team created an image dataset where an image "had at least one FRB pulse" and another image only of the noise. The team then simulated impulses and placed them on some of the noisy images. Simulation is a key technique deployed by researchers where they may not have the data they want.
Their 17-layer CNN is formed on 200,000 actual images and 200,000 false images have detected pulses with 88% accuracy and real pulses with an accuracy of 98%.
Their journal and the project website have more details about their project, including their dataset.