{\huge Predictions We Didn't Make}
Happy New Year 2018 \vspace{.5in}
Muhammad Afzal Upal is Chair of the Computing and Information Science Department at Mercyhurst University. He works in machine learning and cognitive science, most specifically making inferences from textual data. In joint papers he has refined a quantitative approach to the idea of postdiction originally suggested by Walter Kintsch in the 1980s.
Today we review some postdictions from 2017 and wish everyone a Happy New Year 2018.
In a 2007 paper, postdictability $latex {\text{post}(C)}&fg=000000$ of a concept $latex {C}&fg=000000$ is defined as ``the ease with which a concept’s inclusion in the text can be justified after the textual unit containing that concept has been read.'' This contrasts with $latex {\text{pre}(C) =}&fg=000000$ ``the ease with which the occurrence of the concept can be predicted prior to the concept having been read.'' The main equation defines the extent $latex {M(C)}&fg=000000$ to which the concept---or event, I may add---is memorable by
$latex \displaystyle M(C) = K\cdot L(C)\cdot(\text{post}(C) - \text{pre}(C)), &fg=000000$
where $latex {L(C)}&fg=000000$ is the prior likelihood of $latex {C}&fg=000000$ emerging and $latex {K}&fg=000000$ is a constant. It says that $latex {C}&fg=000000$ is most memorable if you couldn't have predicted it but after you see it you slap your forehead and say, ``Ah!'' It relates to what makes ideas ``stick.''Mercyhurst is in Erie, Pennsylvania. Erie had lots of snow this past week. Record-breaking snow. More than Buffalo usually gets. We had several relatives and friends who had to drive through it on their way to Michigan and Pittsburgh and points further south. Was that karma? coincidence with this post? memorable in a way that fits the framework?
And how about the Buffalo Bills making the playoffs after a miracle TD by the Cincinnati Bengals on fourth-and-12 from midfield in the final minute knocked out the Baltimore Ravens? In designing a two-week unit on data science for my department's new team-taught Freshman Seminar on ``The Internet,'' I had the foresight to use years-since-a-team's-last-playoff-win (not playoff appearance) as the definition of ``playoff drought'' in activity examples. Hence---unless the Bills upset the Jacksonville Jaguars next Sunday---the local ``nudge'' of my materials will work equally well for next fall's offering. Can one quantify my prescience as prediction? postdiction? Let's consider some more-germane examples.
Some Standing Predictions
Last January we did not do a predictions or year-in-review post as we had done in all seven previous years. We were caught up in questions over Laszló Babai's graph isomorphism theorem and other matters. Several predictions were recurring, so let's suppose we made them also for 2017:
Postdictions From 2017
Since some of our perennial questions have entered a steady state, it is time to find new categories. A week-plus ago we noticed that Altmetric publishes a top-100 list based on their ``Altmetric Attention Score'' every November 15. So it is natural to postulate:
The answer with regard to the 2017 list is ``yes'' but the reason is unfortunate---it is Norbert Blum's incorrect $latex {\mathsf{P \neq NP}}&fg=000000$ paper coming in at \#38. Blum gave a formal retraction and subsequent explanation, which we added as updates to our own item on the claim. The only other paper labeled ``Information and Computer Sciences'' is the AlphaGo Zero paper at \#74. Actually, Blum's paper was tagged ``Research and Reproducibility.''
AlphaGo Zero and most recently AlphaZero spring to mind. With much swallowing of pride from my starting out as a chessplayer, I'm not sure that games of perfect information should ultimately be regarded as ``human-centric.'' Based on my current understanding of the AlphaZero paper and comments by Gwern Branwen in our previous post, what strikes me as the most stunning fact is the following:
Chess can be encoded into a `kernel' $latex {K}&fg=000000$ of well under 1GB such that $latex {K}&fg=000000$ + small search comprehensively outperforms an almost 1,000x larger search.
More on the human-centric side, however---and allowing supervision---the most surprise and attention seems to have gone to the 2017 Stanford study adapting off-the-shelf facial-analysis software to distinguish sexual orientation from photographs with accuracy upwards of 90%, compared to human subjects at 52% from a balanced sample, which is barely better than random. For utility we would nominate LipNet, which achieves over 95% accuracy in lip-reading from video data, but the paper dates to December 2016.
The lip-reading success may be the more predictable. The extent to which it and the ``gaydar'' application are postdictable appears to be the same as our having a common understanding of what deep neural nets are capable of---which does not require being able to explain how they work. Setting up grounds beforehand for the justification by which Upal and his co-authors define postdiction might be a fair way of ``giving credit for a postdiction.''
As Lance Fortnow notes in his own 2017 review, the complexity result of the year is split between two papers claiming to prove full dichotomy for nonuniform CSPs---where dichotomy means that they are either in P or NP-complete. Meanwhile we have devoted numerous posts to Jin-Yi Cai's work on dichotomy between P and \#P-completeness, including recently. So can we get some credit for prediction? or for postdiction? Anyway, we make it a prime prediction for 2018 that there will be notable further progress in this line.
We specify quantum supremacy to mean mean building a physical device that achieves a useful algorithmic task that cannot be performed in equal time by classical devices using commensurate resources. The words ``useful'' and ``commensurate'' are subjective but the former rules out having ``simulating natural quantum systems'' as the task and conveys John Preskill's emphasis in his original 2012 supremacy paper that the quantum device must be controllable. The latter rules out using whole server farms to match what a refrigerator-sized quantum device can do. The notion involves concrete rather than asymptotic complexity, so we are not positing anything about the hardness of factoring, and intensional tasks like Simon's Problem don't count---not to mention our doubts on the fairness of its classical-quantum comparison. Aram Harrow and Ashley Montanaro said more about supremacy requirements in this paper.
Our ``postdiction'' gets a yes for 2017 from the claims in this Google-led paper that 50-qubit devices would suffice to achieve supremacy and are nearly at hand, versus this IBM-led rebuttal showing that classical computers can emulate a representative set of 50-qubit computations. The notion of emulation allows polling for state information of the quantum circuit computation being emulated, so this is not even confronting the question of solving the task by other means---or proving that classical resources of some concrete size $latex {X}&fg=000000$ cannot solve all length-$latex {n}&fg=000000$ cases of the task at all. Recent views on the controversy are expressed in this November 14 column in Nature, which links this October 22 post by Scott Aaronson (see also his paper with Lijie Chen, ``Complexity-Theoretic Foundations of Quantum Supremacy Experiments'') and this December 4 paper by Cristian and Elena Calude which evokes the Google-IBM case.
That is, the notions of algorithm and protocol will fuse into greater structures with multiple objectives besides solving a task. In 2016 we noted Noam Nisan's 2012 Gödel Prize-winning paper with Amir Ronen titled ``Algorithm Mechanism Design.'' Noam's 2016 Knuth Prize citation stated, ``A mechanism is an algorithm or protocol that is explicitly designed so that rational participants, motivated purely by their self-interest, will achieve the designer’s goals.'' In November we covered mechanisms for algorithmic fairness. There is a nicely accessible survey titled ``Algorithms versus mechanisms: how to cope with strategic input?'' by Rad Niazadeh who works in this area. It alloys techniques from many veins of theory and has a practical gold mine of applications. What we are watching for is the emergence of single powerful new ideas from this pursuit.
Open Problems
What concepts $latex {C}&fg=000000$ do you think will have the highest $latex {M(C)}&fg=000000$ in 2018?