Reading as a god
Chapter 252 Shallow Learning
Chapter 252 Shallow Learning
"Nature"~
One of the earliest scientific journals in the world, it is also one of the most authoritative and prestigious academic journals in the world.
首版於1869年11月4日,到现在(2015年)传承都快150年了~
And while most scientific journals today are dedicated to a specific field, Nature is one of the few (other similar journals include Science and Proceedings of the National Academy of Sciences) that still publishes primary research papers from many scientific fields periodicals.
In many fields of scientific research, the most important and cutting-edge research results are published in the form of short articles in "Nature" every year.
Although the evaluation of the impact factor is not completely objective, the impact factor of 40+ shows its influence~
Although I have thought a lot in my mind, Zhang Shan still can't fully understand the concept of posting on this~
Just when Zhang Shan was inflated, the system poured cold water:
"The host has a 0% pass rate when submitting level 100 papers, and the current pass rate is 1% when the host submits level 99 papers~"
"When the level of the paper submitted by the host is n level, the pass rate will decrease by n% for every n level increase compared to level 0~"
Zhang Shan:. . .
Yes, white excited for a long time~
According to the system, there will only be a 9% pass rate when submitting grade 19 papers in the future.
But having said that, the current level 1 papers in the system are published in the rhythm of NATURE.
Where will the grade 9 papers be published in the future?
Now speaking, it is not a 100% pass rate to submit to nature.
It's a 99% pass rate~
Although it sounds plausible~
But Zhang Shan always has a dark face, and the 99.99 winning rate sometimes overturns~
What will be the specific result now, how can I say well~
For the article that contributed to nature, Zhang Shan is not interested at all, but is more interested in the two 0-level articles~
Although these two papers are still the heights that can only be achieved by outstanding human beings.
But with Zhang Shan's wisdom, he quickly figured out the ins and outs of the paper "Further application of the Generative Adversarial Networks":
Deep learning training a model requires a lot of manually labeled data.
In image recognition, millions of manually labeled data may be required; in speech recognition, tens of thousands of hours of manually labeled data may be required; and machine translation requires tens of millions of double-sentence pairs for training. , These are the embodiment of big data.
However, it is often very expensive to find experts to label data, and for some applications, it is difficult to find large-scale labeled data, such as some intractable diseases, or some relatively rare application scenarios.
The cost of labeling data is extremely high.
For example, for machine translation, if a human is hired to translate, the cost of a word is about 5-10 cents, and the average length of a sentence is about 30 words. If Zhang Shan needs to label 1000 million double sentence pairs , That is to say, Zhang Shan needs to find experts to translate 1000 million sentences, and the cost of this annotation is almost 2200 million US dollars.
The cost of data labeling is very, very high. It is difficult or not feasible for a start-up company or some companies just getting involved in artificial intelligence to spend such a large sum of money to label data.
Therefore, one of the current frontiers of deep learning is how to learn from unlabeled data.
The generative adversarial network described in Zhang Shan's article plays such a role.
The main purpose of a generative adversarial network is to learn a generative model so that the generative adversarial network can generate many images that look like real natural images.
The idea of generating confrontation networks to solve this problem is different from the previous methods. Generative confrontation networks learn two neural networks at the same time: one neural network generates images, and the other neural network classifies images to distinguish real images from generated images. Image.
In the generative adversarial network, the first neural network is the generative neural network. The purpose of the generative adversarial network is to hope that the generated image is very similar to the real image in nature. In this case, the second network behind it is That classifier cannot distinguish between real-world images and generated images; and the purpose of the second neural network, that is, the classifier, the generative adversarial network is to correctly distinguish the generated images, that is, fake images, from real ones. Nature images can be distinguished.
The purpose of these two neural networks is actually different. They can be trained together to get a good generative neural network.
When the generative confrontation network was first proposed, it was mainly for the generation of images.
What Zhang Shan proposed in his thesis is obviously to apply this method to various problems.
However, the focus of the paper is still Zhang Shan on how to learn from unlabeled data!
In the article, he proposed a new idea called dual learning.
The idea of dual learning is very different from the previous generative confrontation learning.
Zhang Shan found that many artificial intelligence tasks have dual properties in structure.
In machine translation, Zhang Shan translates Chinese into English. This is a task, but Zhang Shan also needs to translate English into Chinese. This is a dual task.
Between the original task and the dual task, their input and output are just reversed.
Originally, Zhang Shan felt guilty because the thesis was made systematically, but now it seems that the thesis was written completely according to his ideas.
Because of the previous excellence in multiple languages.
If Zhang Shan wrote this thesis himself, the most likely application involving the dual property would be translation work.
But the dual work doesn't stop there.
In speech processing, speech recognition is to convert speech into text, and speech synthesis is to convert text into speech, which are also two tasks that are dual to each other.
In image understanding, looking at pictures means generating a descriptive sentence for a picture. The dual task of the generative adversarial network is to generate a picture for a sentence. One of these two tasks is from image to text, and the other is One is from text to image.In the dialogue system, answering questions and generating questions are also dual issues. The former is to generate answers for given questions, and the latter is to generate questions for given answers.
In search engines, returning relevant documents for a given search term and returning keywords for a given document or advertisement are also dual problems: the main task of a search engine is to match some documents with the search terms submitted by users and return the most relevant documents ; When an advertiser submits an advertisement, the advertising platform needs to recommend some keywords to him so that his advertisement can be displayed and clicked by the user when the user searches for these words.
Dual learning attempts to apply the dual properties of this structure to machine learning.
The basic idea is relatively simple, and Zhang Shan uses machine translation as an example to illustrate.
When we want to translate a Chinese sentence into English, we can first use a Chinese-to-English translation model to translate this sentence into an English sentence. Because there is no English label, we don’t know whether the English translation is good or bad. How good and how bad.Zhang Shan then uses the English-to-Chinese translation model to translate the English sentence into a Chinese sentence. In this way, Zhang Shan obtains a new Chinese sentence.
The whole process includes two steps of forward translation and reverse translation which are dual to each other.
Then Zhang Shan compares the original Chinese sentence with the Chinese sentence obtained later. If the two translation models are good, the two Chinese sentences should be similar. If the two models are not good or one model is not good, get The two Chinese sentences of are not similar.Therefore, Zhang Shan can obtain feedback information from unlabeled data through this dual process, know whether Zhang Shan’s model is working well or not, and then train and update the forward and reverse models based on these feedback information, so as to achieve unlabeled data purpose of learning.
Zhang Shan did some experiments in machine translation and found that through the process of dual learning, Zhang Shan only needs to use 10% of the labeled data (about 100 million English and French bilingual sentence pairs), plus a lot of unlabeled data, reaching 100 % Accuracy of the model trained on labeled data (1200 million English-French bilingual sentence pairs).
The cost of labeling 1000 million training corpora is about 2200 million US dollars. If Zhang Shan can reduce the labor cost of labeling from 2200 million US dollars to 200 million US dollars, this will be a very good result, which can greatly reduce the company’s operating costs and improve operating efficiency. .
Many problems used to be due to the limitation of unlabeled data, and there was no way to use deep learning technology.
Now that Zhang Shan can learn from unlabeled data, deep learning technology can be applied to many applications and problems.
The thesis is already very good here!
But is it just that~
Zhang Shan continued to read, and was soon shocked!
Because the paper seems to propose a new concept.
Although deep learning is very popular now, in the final analysis, deep learning mainly learns from big data, that is, through a lot of labeled data, some models are learned by using deep learning algorithms.
Although called the name of artificial intelligence.
But this way of learning is very different from human intelligence.
Humans learn from small samples.
People classify images, and only a few samples are needed to achieve accurate classification.
When a two- or three-year-old child starts to know the world, if he wants to know what kind of animal is a dog, we show him a few pictures of dogs, and tell him the characteristics of a dog, and how it differs from other animals like cats or sheep If it is, the child can quickly and accurately identify the dog.
But like a deep residual neural network, generally speaking, a category requires thousands of pictures to be fully trained and get more accurate results.
Another example is car driving. Generally speaking, most people can drive on the road after training in a driving school, that is, dozens of hours of study and hundreds of kilometers of practice.
However, unmanned vehicles like today may have traveled millions of kilometers, and they still cannot reach the level of fully automatic driving.
The reason is that after limited training, humans can cope with various complex road conditions combined with rules and knowledge, but the current AI does not have the ability to think logically, associate and reason, and must rely on big data to cover various possible road conditions. The possible road conditions are almost endless.
With the improvement of each ability, Zhang Shan now has a deep understanding of people.
Human intelligence includes many aspects. The most basic stage is cognitive intelligence, which is the cognition of the whole world.
Although now for image recognition and speech recognition, AI has almost reached the human level, of course, it may be able to reach the human level under certain specific constraints.
But in fact, this kind of cognitive task is very simple for human beings. Now the things that AI can do or the level that can be achieved are actually very easy for humans to do.
It's just that AI may be faster in speed, and the cost will be lower after scaling up, and it doesn't need to rest for 24 hours.A more challenging question is whether artificial intelligence can do things that humans cannot or are difficult to do well.
The reason why AI does well in cognitive tasks such as image recognition and speech recognition is that these tasks are static. The so-called static is a given input, and the prediction result will not change over time.
However, decision-making problems often have complex interactions with the environment. In some scenarios, how to make optimal decisions, these optimal decisions are often dynamic and will change over time.
Some people are now trying to apply AI to the financial market, such as how to use AI technology to analyze stocks, predict stock rises and falls, give advice on stock trading, and even replace people in stock trading. This type of problem is a dynamic decision-making problem.
The second difficulty in decision-making problems lies in the mutual influence of various factors, which can affect the whole body.
The ups and downs of one stock will affect other stocks, and one person's investment decision, especially the investment decision of a large institution, may have an impact on the entire market, which is different from static cognitive tasks.
On static cognitive tasks our predictions do not have any influence on the question (e.g. other images or speech).
But in the stock market, any decision, especially the investment strategy of a large institution, will have an impact on the entire market, other investors, and the future.
At present, deep learning has achieved great success in static tasks. How to extend and extend this success to such complex dynamic decision-making problems is also one of the current challenges of deep learning.
Zhang Shan believes that one possible idea is game machine learning.
In game machine learning, by observing the environment and the behavior of other individuals, and constructing a different personalized behavior model for each individual, AI can think twice before acting.
Choose an optimal policy that will adapt to changes in the environment and changes in the behavior of other individuals.
……
In this paper, Zhang Shan proposed a kind of machine learning that is almost completely anti-deep learning - shallow learning.
Emphasize the importance of enhancing game machine learning, emphasizing the logic and speculative nature of AI, and greatly reducing the amount of "machine learning" tasks.
There is no doubt that this is a whole new way of machine learning!
At the very least, the performance of this new model in processing dynamic information will be revolutionary.
(End of this chapter)
"Nature"~
One of the earliest scientific journals in the world, it is also one of the most authoritative and prestigious academic journals in the world.
首版於1869年11月4日,到现在(2015年)传承都快150年了~
And while most scientific journals today are dedicated to a specific field, Nature is one of the few (other similar journals include Science and Proceedings of the National Academy of Sciences) that still publishes primary research papers from many scientific fields periodicals.
In many fields of scientific research, the most important and cutting-edge research results are published in the form of short articles in "Nature" every year.
Although the evaluation of the impact factor is not completely objective, the impact factor of 40+ shows its influence~
Although I have thought a lot in my mind, Zhang Shan still can't fully understand the concept of posting on this~
Just when Zhang Shan was inflated, the system poured cold water:
"The host has a 0% pass rate when submitting level 100 papers, and the current pass rate is 1% when the host submits level 99 papers~"
"When the level of the paper submitted by the host is n level, the pass rate will decrease by n% for every n level increase compared to level 0~"
Zhang Shan:. . .
Yes, white excited for a long time~
According to the system, there will only be a 9% pass rate when submitting grade 19 papers in the future.
But having said that, the current level 1 papers in the system are published in the rhythm of NATURE.
Where will the grade 9 papers be published in the future?
Now speaking, it is not a 100% pass rate to submit to nature.
It's a 99% pass rate~
Although it sounds plausible~
But Zhang Shan always has a dark face, and the 99.99 winning rate sometimes overturns~
What will be the specific result now, how can I say well~
For the article that contributed to nature, Zhang Shan is not interested at all, but is more interested in the two 0-level articles~
Although these two papers are still the heights that can only be achieved by outstanding human beings.
But with Zhang Shan's wisdom, he quickly figured out the ins and outs of the paper "Further application of the Generative Adversarial Networks":
Deep learning training a model requires a lot of manually labeled data.
In image recognition, millions of manually labeled data may be required; in speech recognition, tens of thousands of hours of manually labeled data may be required; and machine translation requires tens of millions of double-sentence pairs for training. , These are the embodiment of big data.
However, it is often very expensive to find experts to label data, and for some applications, it is difficult to find large-scale labeled data, such as some intractable diseases, or some relatively rare application scenarios.
The cost of labeling data is extremely high.
For example, for machine translation, if a human is hired to translate, the cost of a word is about 5-10 cents, and the average length of a sentence is about 30 words. If Zhang Shan needs to label 1000 million double sentence pairs , That is to say, Zhang Shan needs to find experts to translate 1000 million sentences, and the cost of this annotation is almost 2200 million US dollars.
The cost of data labeling is very, very high. It is difficult or not feasible for a start-up company or some companies just getting involved in artificial intelligence to spend such a large sum of money to label data.
Therefore, one of the current frontiers of deep learning is how to learn from unlabeled data.
The generative adversarial network described in Zhang Shan's article plays such a role.
The main purpose of a generative adversarial network is to learn a generative model so that the generative adversarial network can generate many images that look like real natural images.
The idea of generating confrontation networks to solve this problem is different from the previous methods. Generative confrontation networks learn two neural networks at the same time: one neural network generates images, and the other neural network classifies images to distinguish real images from generated images. Image.
In the generative adversarial network, the first neural network is the generative neural network. The purpose of the generative adversarial network is to hope that the generated image is very similar to the real image in nature. In this case, the second network behind it is That classifier cannot distinguish between real-world images and generated images; and the purpose of the second neural network, that is, the classifier, the generative adversarial network is to correctly distinguish the generated images, that is, fake images, from real ones. Nature images can be distinguished.
The purpose of these two neural networks is actually different. They can be trained together to get a good generative neural network.
When the generative confrontation network was first proposed, it was mainly for the generation of images.
What Zhang Shan proposed in his thesis is obviously to apply this method to various problems.
However, the focus of the paper is still Zhang Shan on how to learn from unlabeled data!
In the article, he proposed a new idea called dual learning.
The idea of dual learning is very different from the previous generative confrontation learning.
Zhang Shan found that many artificial intelligence tasks have dual properties in structure.
In machine translation, Zhang Shan translates Chinese into English. This is a task, but Zhang Shan also needs to translate English into Chinese. This is a dual task.
Between the original task and the dual task, their input and output are just reversed.
Originally, Zhang Shan felt guilty because the thesis was made systematically, but now it seems that the thesis was written completely according to his ideas.
Because of the previous excellence in multiple languages.
If Zhang Shan wrote this thesis himself, the most likely application involving the dual property would be translation work.
But the dual work doesn't stop there.
In speech processing, speech recognition is to convert speech into text, and speech synthesis is to convert text into speech, which are also two tasks that are dual to each other.
In image understanding, looking at pictures means generating a descriptive sentence for a picture. The dual task of the generative adversarial network is to generate a picture for a sentence. One of these two tasks is from image to text, and the other is One is from text to image.In the dialogue system, answering questions and generating questions are also dual issues. The former is to generate answers for given questions, and the latter is to generate questions for given answers.
In search engines, returning relevant documents for a given search term and returning keywords for a given document or advertisement are also dual problems: the main task of a search engine is to match some documents with the search terms submitted by users and return the most relevant documents ; When an advertiser submits an advertisement, the advertising platform needs to recommend some keywords to him so that his advertisement can be displayed and clicked by the user when the user searches for these words.
Dual learning attempts to apply the dual properties of this structure to machine learning.
The basic idea is relatively simple, and Zhang Shan uses machine translation as an example to illustrate.
When we want to translate a Chinese sentence into English, we can first use a Chinese-to-English translation model to translate this sentence into an English sentence. Because there is no English label, we don’t know whether the English translation is good or bad. How good and how bad.Zhang Shan then uses the English-to-Chinese translation model to translate the English sentence into a Chinese sentence. In this way, Zhang Shan obtains a new Chinese sentence.
The whole process includes two steps of forward translation and reverse translation which are dual to each other.
Then Zhang Shan compares the original Chinese sentence with the Chinese sentence obtained later. If the two translation models are good, the two Chinese sentences should be similar. If the two models are not good or one model is not good, get The two Chinese sentences of are not similar.Therefore, Zhang Shan can obtain feedback information from unlabeled data through this dual process, know whether Zhang Shan’s model is working well or not, and then train and update the forward and reverse models based on these feedback information, so as to achieve unlabeled data purpose of learning.
Zhang Shan did some experiments in machine translation and found that through the process of dual learning, Zhang Shan only needs to use 10% of the labeled data (about 100 million English and French bilingual sentence pairs), plus a lot of unlabeled data, reaching 100 % Accuracy of the model trained on labeled data (1200 million English-French bilingual sentence pairs).
The cost of labeling 1000 million training corpora is about 2200 million US dollars. If Zhang Shan can reduce the labor cost of labeling from 2200 million US dollars to 200 million US dollars, this will be a very good result, which can greatly reduce the company’s operating costs and improve operating efficiency. .
Many problems used to be due to the limitation of unlabeled data, and there was no way to use deep learning technology.
Now that Zhang Shan can learn from unlabeled data, deep learning technology can be applied to many applications and problems.
The thesis is already very good here!
But is it just that~
Zhang Shan continued to read, and was soon shocked!
Because the paper seems to propose a new concept.
Although deep learning is very popular now, in the final analysis, deep learning mainly learns from big data, that is, through a lot of labeled data, some models are learned by using deep learning algorithms.
Although called the name of artificial intelligence.
But this way of learning is very different from human intelligence.
Humans learn from small samples.
People classify images, and only a few samples are needed to achieve accurate classification.
When a two- or three-year-old child starts to know the world, if he wants to know what kind of animal is a dog, we show him a few pictures of dogs, and tell him the characteristics of a dog, and how it differs from other animals like cats or sheep If it is, the child can quickly and accurately identify the dog.
But like a deep residual neural network, generally speaking, a category requires thousands of pictures to be fully trained and get more accurate results.
Another example is car driving. Generally speaking, most people can drive on the road after training in a driving school, that is, dozens of hours of study and hundreds of kilometers of practice.
However, unmanned vehicles like today may have traveled millions of kilometers, and they still cannot reach the level of fully automatic driving.
The reason is that after limited training, humans can cope with various complex road conditions combined with rules and knowledge, but the current AI does not have the ability to think logically, associate and reason, and must rely on big data to cover various possible road conditions. The possible road conditions are almost endless.
With the improvement of each ability, Zhang Shan now has a deep understanding of people.
Human intelligence includes many aspects. The most basic stage is cognitive intelligence, which is the cognition of the whole world.
Although now for image recognition and speech recognition, AI has almost reached the human level, of course, it may be able to reach the human level under certain specific constraints.
But in fact, this kind of cognitive task is very simple for human beings. Now the things that AI can do or the level that can be achieved are actually very easy for humans to do.
It's just that AI may be faster in speed, and the cost will be lower after scaling up, and it doesn't need to rest for 24 hours.A more challenging question is whether artificial intelligence can do things that humans cannot or are difficult to do well.
The reason why AI does well in cognitive tasks such as image recognition and speech recognition is that these tasks are static. The so-called static is a given input, and the prediction result will not change over time.
However, decision-making problems often have complex interactions with the environment. In some scenarios, how to make optimal decisions, these optimal decisions are often dynamic and will change over time.
Some people are now trying to apply AI to the financial market, such as how to use AI technology to analyze stocks, predict stock rises and falls, give advice on stock trading, and even replace people in stock trading. This type of problem is a dynamic decision-making problem.
The second difficulty in decision-making problems lies in the mutual influence of various factors, which can affect the whole body.
The ups and downs of one stock will affect other stocks, and one person's investment decision, especially the investment decision of a large institution, may have an impact on the entire market, which is different from static cognitive tasks.
On static cognitive tasks our predictions do not have any influence on the question (e.g. other images or speech).
But in the stock market, any decision, especially the investment strategy of a large institution, will have an impact on the entire market, other investors, and the future.
At present, deep learning has achieved great success in static tasks. How to extend and extend this success to such complex dynamic decision-making problems is also one of the current challenges of deep learning.
Zhang Shan believes that one possible idea is game machine learning.
In game machine learning, by observing the environment and the behavior of other individuals, and constructing a different personalized behavior model for each individual, AI can think twice before acting.
Choose an optimal policy that will adapt to changes in the environment and changes in the behavior of other individuals.
……
In this paper, Zhang Shan proposed a kind of machine learning that is almost completely anti-deep learning - shallow learning.
Emphasize the importance of enhancing game machine learning, emphasizing the logic and speculative nature of AI, and greatly reducing the amount of "machine learning" tasks.
There is no doubt that this is a whole new way of machine learning!
At the very least, the performance of this new model in processing dynamic information will be revolutionary.
(End of this chapter)
You'll Also Like
-
Dragon Ball: Reborn as a Saiyan, They Covet Me
Chapter 120 1 hours ago -
Bleach: The Invincible Slacker from Rukongai
Chapter 201 3 hours ago -
Elves: Check out Ash's top ten legendary companions!
Chapter 194 3 hours ago -
I, Nurarihyon, the Lord of All Monsters
Chapter 121 3 hours ago -
Hogwarts: Past life exposed, I am the Sorcerer Supreme
Chapter 43 3 hours ago -
Shura White Tiger in Sailing
Chapter 186 3 hours ago -
National Secret Realm: Dig out Queen Medusa at the beginning
Chapter 223 3 hours ago -
Dissecting corpses in the morning, blind dating with Boss Yang in the afternoon
Chapter 209 3 hours ago -
Life in Tokyo starting as a child star
Chapter 184 3 hours ago -
Simultaneous time travel: Starting from the inherently evil Uchiha
Chapter 306 3 hours ago