Comments by "Mikko Rantalainen" (@MikkoRantalainen) on "Sabine Hossenfelder"
channel.
-
129
-
55
-
47
-
46
-
22
-
14
-
10
-
8
-
6
-
5
-
4
-
4
-
4
-
4
-
4
-
4
-
3
-
3
-
3
-
3
-
3
-
3
-
3
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
2
-
1
-
1
-
1
-
1
-
1
-
1
-
1
-
IMO, I think this can be battled the same way AI should be trained even today: start with a small amount of trusted data (hand selected articles and books from various fields verified by experts). Then for every new piece of content, make the AI estimate the quality of the data (can it be deduced from the trusted data?) and skip training if the content is not (yet?) considered high enough quality. Note that the quality of the data is about it being well supported by the existing knowledge. Do this multiple times to create chains of high quality data (that is, the potential content was not deemed trusted earlier but now that the AI has learned more, it will estimate the quality of the same data differently).
Keep track of the estimated quality of a given piece of data and recompute the estimated quality again for all documents every now and then. If the quality estimate of the AI is good enough, it should increase the estimated quality of the content over time (because more chains allows accepting more new data) and cases where previously trusted content turns on untrusted later would point out problems in the system.
Also run the estimation against known good high quality data not included in the training set every now and then. These should be considered high quality data but the AI may fail to identify the data correctly, which would demonstrate lack of general understanding by the AI.
Once you demonstrate that the estimated quality matches well enough with the expert evaluations of the same content, you can start to train the AI to understand misunderstandings of humans, too. Train low quality content as examples of humans failing to think / research correctly.
In the end, you should have an AI that can successfully estimate quality of any new content and automatically use it to either extend its knowledge (chains of known good content) or to automatically learn it as an example of low quality content that the AI should avoid but the AI should be made aware of.
If the AI doesn't have negative feedback from failed human content, it cannot understand failures in tasks given to said AI.
1
-
1
-
1
-
1