| Literature DB >> 35052032 |
Diederik Aerts1, Lester Beltran1.
Abstract
In previous research, we showed that 'texts that tell a story' exhibit a statistical structure that is not Maxwell-Boltzmann but Bose-Einstein. Our explanation is that this is due to the presence of 'indistinguishability' in human language as a result of the same words in different parts of the story being indistinguishable from one another, in much the same way that 'indistinguishability' occurs in quantum mechanics, also there leading to the presence of Bose-Einstein rather than Maxwell-Boltzmann as a statistical structure. In the current article, we set out to provide an explanation for this Bose-Einstein statistics in human language. We show that it is the presence of 'meaning' in 'texts that tell a story' that gives rise to the lack of independence characteristic of Bose-Einstein, and provides conclusive evidence that 'words can be considered the quanta of human language', structurally similar to how 'photons are the quanta of electromagnetic radiation'. Using several studies on entanglement from our Brussels research group, we also show, by introducing the von Neumann entropy for human language, that it is also the presence of 'meaning' in texts that makes the entropy of a total text smaller relative to the entropy of the words composing it. We explain how the new insights in this article fit in with the research domain called 'quantum cognition', where quantum probability models and quantum vector spaces are used in human cognition, and are also relevant to the use of quantum structures in information retrieval and natural language processing, and how they introduce 'quantization' and 'Bose-Einstein statistics' as relevant quantum effects there. Inspired by the conceptuality interpretation of quantum mechanics, and relying on the new insights, we put forward hypotheses about the nature of physical reality. In doing so, we note how this new type of decrease in entropy, and its explanation, may be important for the development of quantum thermodynamics. We likewise note how it can also give rise to an original explanatory picture of the nature of physical reality on the surface of planet Earth, in which human culture emerges as a reinforcing continuation of life.Entities:
Keywords: Bose–Einstein statistics; electromagnetic radiation; entanglement; human culture; human language; indistinguishability; life; quantum cognition; thermodynamics; von Neumann entropy
Year: 2021 PMID: 35052032 PMCID: PMC8775064 DOI: 10.3390/e24010006
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
An energy scale representation of the words of the Winnie the Pooh story ‘In Which Piglet Meets a Haffalump’ by A. A. Milne as published in [32]. The words are in the column ‘Words concepts cognitons’ and the energy levels are in the column ‘Energy levels ’, and are attributed according to the ‘numbers of appearances’ in the column ‘Appearance numbers ’, such that lower energy levels correspond to a higher order of appearances, and the value of the energy levels is determined according to (1). The ‘amounts of energies radiated by the words of energy level ’ are in the column ‘Energies from data ’. In the columns ‘Bose–Einstein modeling’, ‘Maxwell–Boltzmann modeling’, ‘Energies Bose–Einstein’, and ‘Energies Maxwell–Boltzmann’ are respectively the predicted values of the Bose–Einstein and the Maxwell–Boltzmann model of the ‘numbers of appearances’, and of the ‘radiated energies’. In the graphs of Figure 2, we can see that a maximum is reached for the energy level , corresponding to the word First, which appears seven times in the Winnie the Pooh story. If we use the analogy with light, we can say that the radiation spectrum of the story ‘In Which Piglet Meets a Haffalump’ has a maximum at First, which would hence be, again in analogy with light, the dominant color of the story. We have indicated this radiation peak by underlining the word First, its energy level , and the amount of energy 522.79 the story radiates, following the Bose-Einstein model. The omitted words are those between the word Put at number 78, and the word Wishing at 538, and this set of omitted words are indicated by a blank space in the Table.
| Words Concepts Cognitons | Energy Levels | Appearance Numbers | Bose–Einstein Modeling | Maxwell–Boltzmann Modeling | Energies from Data | Energies Bose–Einstein | Energies Maxwell–Boltzmann |
|---|---|---|---|---|---|---|---|
|
| 0 | 133 | 129.05 | 28.29 | 0 | 0 | 0 |
|
| 1 | 111 | 105.84 | 28.00 | 111 | 105.84 | 28.00 |
|
| 2 | 91 | 89.68 | 27.69 | 182 | 179.36 | 55.38 |
|
| 3 | 85 | 77.79 | 27.40 | 255 | 233.36 | 82.19 |
|
| 4 | 70 | 68.66 | 27.11 | 280 | 274.65 | 108.43 |
|
| 5 | 69 | 61.45 | 26.82 | 345 | 307.23 | 234.09 |
|
| 6 | 61 | 55.59 | 26.53 | 366 | 333.55 | 159.20 |
|
| 7 | 59 | 50.75 | 26.25 | 413 | 355.24 | 183.76 |
|
| 8 | 47 | 46.68 | 25.97 | 376 | 373.40 | 207.78 |
|
| 9 | 46 | 43.20 | 25.70 | 414 | 388.82 | 231.27 |
|
| 10 | 41 | 40.21 | 25.42 | 410 | 402.05 | 254.24 |
|
| 11 | 40 | 37.59 | 25.15 | 440 | 413.52 | 276.69 |
|
| 12 | 39 | 35.30 | 24.89 | 468 | 423.55 | 298.64 |
|
| 13 | 28 | 33.26 | 24.62 | 364 | 432.38 | 320.09 |
|
| 14 | 26 | 31.44 | 24.36 | 364 | 440.21 | 341.05 |
|
| 15 | 25 | 29.81 | 24.10 | 375 | 447.19 | 361.53 |
|
| 16 | 25 | 28.34 | 23.86 | 400 | 453.44 | 381.53 |
|
| 17 | 23 | 27.00 | 23.59 | 391 | 459.07 | 401.07 |
|
| 18 | 23 | 25.79 | 23.34 | 414 | 464.15 | 420.15 |
|
| 19 | 23 | 24.67 | 23.09 | 437 | 468.77 | 438.78 |
|
| 20 | 23 | 23.65 | 22.85 | 460 | 472.96 | 456.97 |
|
| 21 | 23 | 22.70 | 22.61 | 483 | 476.79 | 474.72 |
|
| 22 | 21 | 21.83 | 22.37 | 462 | 480.30 | 492.05 |
|
| 23 | 20 | 21.02 | 22.13 | 460 | 483.51 | 508.95 |
|
| 24 | 20 | 20.27 | 21.89 | 480 | 486.47 | 525.43 |
|
| 25 | 20 | 19.57 | 21.66 | 500 | 489.19 | 541.51 |
|
| 26 | 19 | 18.91 | 21.43 | 494 | 491.71 | 557.19 |
|
| 27 | 18 | 18.30 | 21.20 | 486 | 494.03 | 572.47 |
|
| 28 | 18 | 17.72 | 20.98 | 504 | 496.18 | 587.37 |
|
| 29 | 18 | 17.18 | 20.75 | 522 | 498.18 | 601.89 |
|
| 30 | 17 | 16.67 | 20.53 | 510 | 500.03 | 616.03 |
|
| 31 | 15 | 16.19 | 20.32 | 465 | 501.75 | 629.80 |
|
| 32 | 15 | 15.73 | 20.10 | 480 | 503.34 | 643.21 |
|
| 33 | 15 | 15.30 | 19.89 | 495 | 504.83 | 656.26 |
|
| 34 | 15 | 14.89 | 19.68 | 510 | 506.22 | 668.97 |
|
| 35 | 14 | 14.50 | 19.47 | 490 | 507.51 | 681.33 |
|
| 36 | 14 | 14.13 | 19.26 | 504 | 508.71 | 693.35 |
|
| 37 | 14 | 13.78 | 19.06 | 518 | 509.83 | 705.03 |
|
| 38 | 13 | 13.44 | 18.85 | 494 | 510.88 | 716.40 |
|
| 39 | 13 | 13.12 | 18.65 | 507 | 511.86 | 727.44 |
|
| 40 | 13 | 12.82 | 18.45 | 520 | 512.77 | 738.17 |
|
| 41 | 12 | 12.53 | 18.26 | 492 | 513.62 | 748.59 |
|
| 42 | 12 | 12.25 | 18.06 | 504 | 514.41 | 758.70 |
|
| 43 | 12 | 11.98 | 17.87 | 516 | 515.15 | 768.51 |
|
| 44 | 12 | 11.72 | 17.68 | 528 | 515.84 | 778.03 |
|
| 45 | 12 | 11.48 | 17.49 | 540 | 516.48 | 778.26 |
|
| 46 | 11 | 11.24 | 17.31 | 506 | 517.08 | 796.20 |
|
| 47 | 11 | 11.01 | 17.12 | 517 | 517.64 | 804.87 |
|
| 48 | 11 | 10.79 | 16.94 | 528 | 518.15 | 813.26 |
|
| 49 | 10 | 10.58 | 16.76 | 490 | 518.63 | 821.39 |
|
| 50 | 10 | 10.38 | 16.58 | 500 | 519.08 | 829.25 |
|
| 51 | 10 | 10.19 | 16.41 | 510 | 519.49 | 836.85 |
|
| 52 | 10 | 10.00 | 16.23 | 520 | 519.87 | 844.19 |
|
| 53 | 10 | 9.82 | 16.06 | 530 | 520.22 | 851.29 |
|
| 54 | 10 | 9.64 | 15.89 | 540 | 520.54 | 858.13 |
|
| 55 | 9 | 9.47 | 15.72 | 495 | 520.83 | 864.74 |
|
| 56 | 9 | 9.31 | 15.56 | 504 | 521.10 | 871.11 |
|
| 57 | 9 | 9.15 | 15.39 | 513 | 521.35 | 877.25 |
|
| 58 | 9 | 8.99 | 15.23 | 522 | 521.57 | 883.15 |
|
| 59 | 9 | 8.84 | 15.07 | 531 | 521.77 | 888.84 |
|
| 60 | 8 | 8.70 | 14.91 | 480 | 521.95 | 894.30 |
|
| 61 | 8 | 8.56 | 14.75 | 488 | 522.11 | 899.55 |
|
| 62 | 8 | 8.43 | 14.59 | 496 | 522.25 | 904.58 |
|
| 63 | 8 | 8.29 | 14.44 | 504 | 522.37 | 909.41 |
|
| 64 | 8 | 8.16 | 14.28 | 512 | 522.48 | 914.03 |
|
| 65 | 8 | 8.04 | 14.13 | 520 | 522.57 | 918.45 |
|
| 66 | 8 | 7.92 | 13.98 | 528 | 522.64 | 922.67 |
|
| 67 | 7 | 7.80 | 13.83 | 469 | 522.70 | 926.70 |
|
| 68 | 7 | 7.69 | 13.68 | 476 | 522.74 | 930.54 |
|
| 69 | 7 | 7.58 | 13.54 | 483 | 522.77 | 934.20 |
|
| 70 | 7 | 7.47 | 13.40 | 490 | 522.78 | 937.67 |
|
|
| 7 | 7.36 | 13.25 | 497 |
| 940.96 |
|
| 72 | 7 | 7.26 | 13.11 | 504 | 522.78 | 944.08 |
|
| 73 | 7 | 7.16 | 12.97 | 511 | 522.76 | 947.02 |
|
| 74 | 7 | 7.06 | 12.84 | 518 | 522.72 | 949.79 |
|
| 75 | 7 | 6.97 | 12.70 | 525 | 522.68 | 952.40 |
|
| 76 | 7 | 6.88 | 12.56 | 532 | 522.63 | 954.85 |
|
| 77 | 7 | 6.79 | 12.43 | 539 | 522.56 | 957.13 |
|
| 78 | 7 | 6.70 | 12.30 | 546 | 522.49 | 959.27 |
| … | … | … | … | … | … | … | … |
| … | … | … | … | … | … | … | … |
|
| 538 | 1 | 0.67 | 0.09 | 538 | 359.92 | 48.65 |
|
| 539 | 1 | 0.67 | 0.09 | 539 | 359.58 | 48.22 |
|
| 540 | 1 | 0.67 | 0.09 | 540 | 359.24 | 47.80 |
|
| 541 | 1 | 0.66 | 0.09 | 541 | 358.90 | 47.38 |
|
| 542 | 1 | 0.66 | 0.09 | 542 | 358.55 | 46.96 |
| 2655 | 2655.00 | 2654.96 | 242,891 | 242,891.01 | 242,889.76 |
Figure 1In (a) we represent the ‘number of appearances’ of words in the Winnie the Pooh story ‘In Which Piglet Meets a Haffalump’ [32], ranked from lowest energy level, corresponding to the most often appearing word, to highest energy level, corresponding to the least often appearing word as listed in Table 1. The blue graph (Series 1) represents the data, i.e., the collected numbers of appearances from the story (column ‘Appearance numbers ’ of Table 1), the red graph (Series 2) is a Bose–Einstein distribution model for these numbers of appearances (column ‘Bose–Einstein modeling’ of Table 1), and the green graph (Series 3) is a Maxwell–Boltzmann distribution model (column ‘Maxwell–Boltzmann modeling’ of Table 1). In (b) we represent the log/log graphs of the ‘numbers of appearances’ and their Bose–Einstein and Maxwell–Boltzmann models. The red and blue graphs coincide almost completely in both (a,b) whereas the green graph does not coincide at all with the blue graph of the data. This shows that the Bose–Einstein distribution is a good model for the numbers of appearances, while the Maxwell–Boltzmann distribution is not.
Figure 2A representation of the ‘energy distribution’ of the Winnie the Pooh story ‘In Which Piglet Meets a Haffalump’ [32] as listed in Table 1. The blue graph (Series 1) represents the energy radiated by the story per energy level (column ‘Energies from data ’ of Table 1), the red graph (Series 2) represents the energy radiated by the Bose–Einstein model of the story per energy level (column ‘Energies Bose–Einstein’ of Table 1), and the green graph (Series 3) represents the energy radiated by the Maxwell–Boltzmann model of the story per energy level (column ‘Energies Maxwell–Boltzmann’ of Table 1).
The data from the five different studies of ‘entanglement’ performed in our group and calculation of von Neumann entropy and purity.
| Experiments | Probabilities | Entropy | Purity | |||
|---|---|---|---|---|---|---|
| The 2011 Cognitive Experiment Study | ||||||
|
|
|
|
|
|
| |
|
|
|
|
|
|
|
|
| The Google Books Natural Language Processing Study | ||||||
|
|
|
|
|
|
| |
|
| 0 |
|
| 0 |
|
|
| The COCA Natural Language Processing Study | ||||||
|
|
|
|
|
|
| |
|
| 0 |
|
| 0 |
|
|
| The Google Images Study | ||||||
|
|
|
|
|
|
| |
|
|
|
|
|
|
|
|
| The 2021 Cognitive Experiment Study | ||||||
|
|
|
|
|
|
| |
|
|
|
|
|
|
|
|