Definitions
- Let \(L\) be a language of propositional logic, and \(S\) be sentences in \(L\).
- A valuation is \(\mathbb V:S\to [0,1]\).
- A pricing is \(\Pj: S\to \Q \cap [0,1]\).
- A belief state is a pricing with finite support. (Syntactically. Semantically, think of it as probabilities.)
- Valuation sequence \(\mathbb N\to (S\to [0,1])\).
- A market is a computable sequence of pricings \(\Pj_i:S\to \Q\cap [0,1]\). (i.e., \(\N \to (S\to \Q\cap [0,1])\)).
- A computable belief sequence is a market with each \(\Pj_i\) having finite support. (Syntactically. Semantically, think of it as probabilities.)
- A deductive process \(\ol D:\N^+\to \text{Fin}(S)\) (finite subsets of \(S\)) is a computable nested sequence \(D_1\subeq D_2\subeq\cdots\) of sentences. Let \(D_\iy:=\bigcup_n D_n\).
- A world is a truth assignment \(\mathbb W: S\to \mathbb B\). True/false in \(\mathbb W\) means \(\mathbb W(\phi)=0,1\).
- Efficiently computable \((x_n)_{n=1}^{\iy}\) means \(x_n\) computable in time \(\poly(n)\).
- A valuation feature \[
\al:\ub{[0,1]^{S\times \N^+}}{\text{valuation sequences}}\to \R
\] is a continuous function such that \(\al(\ol{\mathbb V})\) depends only on \(\mathbb V_{\le n}\) for some \(n\in \N^+\) called the rank.
- Price feature (cf. evaluation functional) \[\psi^{*n}(\ol{\mathbb V}):=\mathbb V_n(\phi)\]
- Expressible feature is built up from price features, \(\Q, +, \times, \max, \max(\cdot, 1)^{-1}\). \(\mathcal{EF}\) is the set of expressible features, \(\mathcal{EF}_n\) of rank \(\le n\) (commutative ring).
- A trading strategy is an affine combination \[T = \pa{-\sum_i \xi_i \phi_i^{*n}} + \sum_i \xi_i \phi_i^{*n}\] where \(\phi_i\in S\), \(\xi_i\) are expressible features of rank \(\le n\). (I.e., it is an element of \(\ker(\ph)\) where \[\ph:\mathcal{EF}_n^{\opl S} \opl \mathcal{EF_n}
= \mathcal{EF}_n^{\opl S\cup \{1\}} \mapsto \mathcal{EF_n}\] given by \(\ph:x\opl y\mapsto x^{*n}+y\).) Let \(T[\phi]\) be the coefficient of \(\phi\), \(T[1]\) be the constant term.
- A trader \(\ol T\) is a sequence \((T_1,T_2,\ldots)\) where each \(T_n\) is a trading strategy for day \(n\).
- Think of \(\phi-\phi^{*n}\) as meaning “buy share of \(\phi_i\) at prevailing price.”
- Example of trading strategy: Arbitrage \(\psi\) against \(\neg \neg \psi\). \[
[(\neg \neg \phi)^{*n} - \phi^{*n}](\phi-\phi^{*n}) + [\phi^{*n}-(\neg\neg\phi)^{*5}](\neg \neg \phi - (\neg\neg \phi)^{*5}).
\]
- Define evaluation of a value function on a \(\mathcal F\)-combination, \((S\to [0,1]) \times (\mathcal F^{\opl(S\cup \{1\})}) \to \mathcal F\) in the natural way, sending \(\phi\) to \(\mathbb V(\phi)\) and extending by linearity. (\(\mathbb V(1)=1\).)
- \(\ol T\) exploits \(\mathbb V\) relative to a deductive process \(\ol D\) if \[
\set{\mathbb W\pa{\sumz in \mathbb V_i(T_i)}}{n\in \N^+, \mathbb W\in PC(D_n)}
\] is bounded below, but not bounded above. This set is the set of plausible assessments of net worth. “The trader can make unbounded returns with bounded risk.”
- Ex. if PA proves \(\phi\vee \psi\), then a trader who buys \(\phi,\psi\) at combined price \(<1\) exploits the market.
- \(\ol{\Pj}\) satisfies the logical induction criterion relative to deductive process \(\ol D\) if there is no efficiently computable trader \(\ol T\) that exploits \(\ol\Pj\) relative to \(\ol D\). \(\ol \Pj\) is a logical inductor.
- A logical inductor over a \(\Ga\)-complete deductive process \(\ol D\) is a logical inductor over \(\Ga\).
- \(\ol\Pj\) assigns \(\ol p\) to \(\ol\phi\) in a timely manner if \(\Pj_n(\phi_n) \simeq_n p_n\).
Theorem
For any deductive process \(\ol D\), there exists a computable belief sequence \(\ol \Pj\) satisfying the logical induction criterion relative to \(\ol D\).
For any recursively axiomatizable \(\Ga\), there exists a computable belief sequence that is a logical inductor over \(\Ga\).
Intuition: Consider any polynomial-time method for efficiently identifying patterns in logic. If the market prices don’t learn to reflect that pattern, a clever trader can use it to exploit the market. For example, if there is a polynomial-time method for identifying theorems that are always underpriced by the market, a clever trader could use that pattern to buy those theorems low, exploiting the market. To avoid exploitation, logical inductors must learn to identify many different types of patterns in logical statements.
(Note: I expected that the logical inductor would be the trader, but no, it’s the market! The traders are people trying to take advantage of the market. But I’m thinking of the AI against nature, and nature throwing catastrophes in the sense of trying to find things the AI would reason badly about.)
Properties
Limit properties
- Convergence: \(\Pj_\iy(\phi) :=\limn \Pj_n(\phi)\) exists.
- Limit coherence: \(\Pj_\iy\) defines a probability measure on the set of worlds consistent with \(\Ga\). \[
\Pj(\mathbb W(\phi)=1):=\Pj_\iy(\phi).
\]
- Occam bounds: There exists \(C>0\) such that, letting \(\ka(\phi)\) be prefix complexity of \(\phi\),
\[\begin{align}
\Pj_\iy(\phi) &\ge C 2^{-\ka(\phi)}\\
\Pj_\iy(\phi) &\le 1-C2^{-\ka(\phi)}.
\end{align}\]
- Domination of universal semimeasure \(M\): Let \(\ol{\si}_n\) be the statement: the first \(n\) bits of the observed sequence is \(\si_{\le n}\). There is universal \(C\) such that \(\Pj_\iy(\ol \si_{\le n}) \ge CM(\ol\si_{\le n})\).
Pattern recognition
Definitions
- Write \(x_n\simeq_n y_n\) for \(\limn x_n-y_n=0\). \(\gtrsim\) and \(\lesssim\) for \(\liminf\ge 0\), \(\limsup\le 0\).
- A divergent weighting is \(\ol w\in [0,1]^{\N^+}\) with \(\su w_n=\iy\). \(\ol q\) is generable from \(\ol\Pj\) if there exists e.c. \(\mathcal{EF}\)-progression \(\ol{q^{\dagger}}(\ol\Pj) =q_n\). (\(q_n^{\dagger}(\ol\Pj) = \ol\Pj_n(q_n^{\dagger})\).)
- “Divergent weightings generable from \(\ol\Pj\) are the pattern detectors that logical inductors can use.”
- Elements of \(\R^{\opl (S\cup \{1\})}\) are constraints.
- Bounded combination sequences \(BCS(\ol \Pj)\): \(\ol\Pj\)-generable \(\ol A\) (\(A_n\in \R^{\opl (S\cup \{1\})}\)) that are bounded (\(\exists b, \forall n,|A_n[1]|\le b\)).
Properties
- Provability induction: Let \(\ol{\phi}\) be e.c. sequence of theorems. Then \[
\Pj_n(\phi_n) \simeq_n 1
\] For disprovable sentences, \(\simeq_n0\). (Analogy: Ramanujan and Hardy)
- Preemptive learning: For e.c. sequence \(\ol\phi\), \(\liminf_{n\to \iy}\Pj_n(\phi_n) = \liminf_{n\to \iy} \sup_{m\ge n} \Pj_m(\phi_n).\) and similarly for inf/sup switched.
- if \(P\) always eventually pushes \(\phi_n\) up to a probability at least \(p\), then it will learn to assign each \(\phi_n\) a probability at least \(p\) in a timely manner (and similarly for least upper bounds). (p.26 in main paper)
- Learning pseudorandom frequencies. \(\ol{\phi}\) of decidable sentences is pseudorandom with frequency p wrt set of divergent weightings \(S\) if for all \(\ol w\in S\), \(\limn \fc{\sumo in w_i \one_{\phi_i \text{ is theorem in }\Ga}}{\sumo in w_i}=p\). For \(\ol\phi\) e.c., if \(\ol\phi\) is pseudorandom over all \(\ol\Pj\)-generable divergent weightings, \(\Pj_n(\phi_n) \simeq_n p\).
- Think of \(\ol{\Pj}\)-generable as meaning: trading strategies you could come up with looking at past history of beliefs. This still doesn’t fit my picture of traders as adversaries, rather than unkind nature.
- Affine coherence: Let \(\ol A\in BCS(\ol\Pj)\). \[
\liminf_{n\to \iy} \inf_{\mathbb W\in C(\Ga)} \mathbb W(A_n) \le \liminf_{n\to \iy} \Pj_\iy (A_n) \le \lim_{n\to \iy} \Pj_n (A_n).
\] Reverse inequalities for sup.
- “Tie ground truth on \(\ol A\) to value of \(\ol A\) on main diagonal.”
- “learns in a timely manner to respect all linear inequalities that actually hold between sentences, so long as those relationships can be enumerated in polynomial time.” Ex. if exactly one of \(A_n,B_n,C_n\) is true, will get that.
- What is each inequality saying?
I’m confused: Consider a sequence \(y_n=f(x_n)\), \(f:B^{2n}\to B^n\) which is hard-to-invert. Theorem \(n\) is \(y_n\in \im f\). How can we hope to assign probabilities \(\to 1\)?
Can we add true randomness? Randomness in \(\Ga\) is not allowed because axioms are recursively computable. (Footnote 3 in paper says we can add randomness to market.) How about allowing traders to be random?
Self-knowledge
Definitions
- Logically uncertain variable (LUV): formula \(X\) free in one variable that defines a unique value via \(\Ga\): (\(\exists!\)) \[
\Ga \perp \exists x: \forall x': X(x')\to x'=x.
\] Value of \(X\) in \(\mathbb W\in C(\Ga)\) is \[
\mathbb W(X):= \sup \set{x\in [0,1]}{\mathbb W("X\ge x")=1}.
\] Because it may not be clear that \(X\) has a unique value, “\(X\ge p\)” is shorthand for \(\forall x: X(x)\to x<p\).
- Ex. “\(\nu\) is 1/0 if Goldbach’s conjecture is true/false.”
- Problem with a defining \(\E\) the normal way: \(\ol\Pj\) may not have figured out that \(X\) takes a unique value!
- Indicator LUV \(\one(\phi) := "(\phi \wedge (\nu = 1))\vee (\neg \psi \wedge (\nu=0))"\).
- Approximate expectation operator with precision \(k\): For \(X\) a \([0,1]\)-LUV, \[
\E_k^{\mathbb V}(X):=\sumz i{k-1}\rc k \mathbb V("X>i/k").
\] Let \(\E_n:=\E_n^{\Pj_n}\).
- \(f:\N^+\to \N^+\) is deferral function if \(f(n)>n\) for all \(n\), \(f(n)\) can be computed in time \(\poly(f(n))\). Say \(f\) defers \(n\) to \(f(n)\).
- Continuous threshold indicator \(\Ind_\de(x>y)\) interpolates between 0 and 1 on \([y,y+\de]\).
Results
You can use the logic to encode the computation of the market prices, i.e. \(\ul{\Pj}_{\ul{n}}(\ul{\phi_n})\). (I’m omitting underlines in the following theorem statements.)
- Introspection: Let \(\ol \phi\) be e.c. sequence of sentences, \(\ol a,\ol b\) be e.c. sequences of probabilities expressible from \(\ol \Pj\). For any e.c. sequence of \(\Q^+\) with \(\ol \de\to 0\), there exists e.c. sequence of \(\Q^+\) with \(\ol\ep\to 0\), s.t. for all \(n\),
\[\begin{align}
\Pj_n(\phi_n)&\in (a_n+\de_n,b_n-\de_n)&\implies \Pj_n("a_n<\Pj_n(\phi_n)<b_n")&>1-\ep_n\\
\Pj_n(\phi_n)&\nin (a_n-\de_n,b_n+\de_n)&\implies \Pj_n("a_n<\Pj_n(\phi_n)<b_n")&<\ep_n
\end{align}\]
- “If there is e.c. pattern of the form your probabilities on \(\ol{\phi}\) will be in \((a,b)\) then \(\ol{\Pj}\) learns to believe that pattern iff it is true subject to finite precision.”
- Paradox resistance: Define paradoxical sentences \(\ol{\chi^p}\) \[
\Ga \vdash \ul{\chi_n^p} \lra (\Pj_n(\chi_n^p)<p).
\] Then \(\lim_{n\to \iy} \Pj_n(\chi_n^p) = p\).
- Expectations
- Convergence: \(\E_\iy\) exists.
- Linearity of expectation. \(\ol \al,\ol\be\) e.c. sequences of \(\Q\) expressible from \(\ol\Pj\), \(\ol X, \ol Y, \ol Z\) e.c. of \([0,1]\)-LUV. \[
\forall n, \Ga\vdash Z_n = \al_nX_n +\be_nY_n\implies \al_n\E_n(X_n) + \be_n\E_n(Y_n) \simeq_n \E_n(Z_n).
\]
- Expectation coherence. $B\(BLCS(\)$) set of bounded LUV-combination sequences. \[
\liminf_{n\to \iy} \inf_{\mathbb W\in C(\Ga)} \mathbb W(B_n) \le \liminf_{n\to \iy} \E_\iy (B_n) \le \lim_{n\to \iy} \E_n (B_n)
\] and reverse for sup.
- No expected net update: \(f\) deferral, \(\ol\phi\) e.c. sequence of sentences. \[\Pj(\phi_n) \simeq_n \E_n("\Pj_{f(n)} (\phi_n)").\]
- If \(\ol\Pj\) on day \(n\) believes on day \(f(n)\) it will believe \(\phi_n\) whp, then it already believes \(\phi_n\) whp today. “In other words, logical inductors learn to adopt their predicted future beliefs as their current beliefs in a timely manner.”
- Self-trust: \(f\) deferral, \(\ol\phi\) e.c., \(\ol\de\in (\Q^+)^n\) e.c., \(\ol p\) e.c. sequence of rational probs expressible from \(\ol\Pj\). Then \[
\E_n ("\one(\phi_n)\Ind_{\de_n} (\Pj_{f(n)}(\phi_n)>p_n)") \gtrsim_n p_n \E_n ("\Ind_{\de_n}(\Pj_{f(n)}(\phi_n)>p_n)").
\]
- “Trust that if beliefs change, they must have changed for good reason”
- Roughly: If we ask \(\ol\Pj\) what it believes about \(\phi\) now if it learned it was going to believe \(\phi\) wp \(\ge p\) in the future, it will answer with probability \(\ge p\). (Some subtlety with continuous indicators, paradoxical sentences.)
Extended paper: Intro
- 13 Gaifman inductivity: Given a \(\Pi_1\) statement \(\phi = \forall x, \psi\), as the set of examples approaches all examples, belief in \(\phi\) approaches 1.
- 18 Use of old evidence: When a bounded reasoner comes up with a new theory that neatly describes anomalies in the old theory, the old evidence should count in favor of the new theory. (Why isn’t this Bayesian? Bayesian keeps track of all hypotheses at all times.)
- 1, 2, 13 cannot be satisfied simultaneously (why?). 1, 6, 13, weak 2 cannot be satisfied simultaneously.
- Algorithm doesn’t satisfy 13, 14, 15. 16, 17 are unclear.
- 1 and 12 are key.
Properties
- Convergence and Coherence: In the limit, the prices of a logical inductor describe a belief state which is fully logically consistent, and represents a probability distribution over all consistent worlds.
- Timely Learning: For any efficiently computable sequence of theorems, a logical inductor learns to assign them high probability in a timely manner, regardless of how difficult they are to prove. (And similarly for assigning low probabilities to refutable statements.)
- Learning Statistical Patterns: If a sequence of sentences appears pseudorandom to all reasoners with the same runtime as the logical inductor, it learns the appropriate statistical summary (assigning, e.g., 10% probability to the claim “the nth digit of \(\pi\) is a 7” for large n, if digits of \(\pi\) are actually hard to predict).
- Calibration and Unbiasedness: Logical inductors are well-calibrated and, given good feedback, unbiased.
- Learning Logical Relationships: Logical inductors inductively learn to respect logical constraints that hold between different types of claims, such as by ensuring that mutually exclusive sentences have probabilities summing to at most 1.
- Non-Dogmatism: The probability that a logical inductor assigns to an independent sentence \(\phi\) is bounded away from 0 and 1 in the limit, by an amount dependent on the complexity of \(\phi\). In fact, logical inductors strictly dominate the universal semimeasure in the limit. This means that we can condition logical inductors on independent sentences, and when we do, they perform empirical induction.
- Conditionals: Given a logical inductor P, the market given by the conditional probabilities \(P(- | \psi)\) is a logical inductor over \(\ol D\) extended to include \(\psi\). Thus, when we condition logical inductors on new axioms, they continue to perform logical induction.
- Expectations: Logical inductors give rise to a well-behaved notion of the expected value of a logically uncertain variable.
- Trust in Consistency: If the theory \(\Ga\) underlying a logical inductor’s deductive process is expressive enough to talk about itself, then the logical inductor learns inductively to trust \(\Ga\).
- Reasoning about Halting: If there’s an efficient method for generating programs that halt, a logical inductor will learn in a timely manner that those programs halt (often long before having the resources to evaluate them). If there’s an efficient method for generating programs that don’t halt, a logical inductor will at least learn not to expect them to halt for a very long time.
- Introspection: Logical inductors “know what they know”, in that their beliefs about their current probabilities and expectations are accurate.
- Self-Trust: Logical inductors trust their future beliefs.
Others not covered before
- 4.2.3 persistence of knowledge: If \(\ol\Pj\) assigns \(\ol p\) to \(\ol\phi\) in the limit, then \(\ol\Pj\) assigns probability near \(p_n\) to \(\phi_n\) at times \(m\ge n\).
- 4.3 Calibration and unbiasedness is hard to check.
- Check conditional rather than marginal probabilities.
- restrict our consideration to sequences where the average frequency of truth converges; or
- look at subsequences of \(\phi\) where P has “good feedback” about the truth values of previous elements of the subsequence, in a manner defined below.
- 4.3.3 Recurring calibration: In any sequence, consider the theorems that were assigned probabilities in \((a,b)\). (\(a<\Pj_i(\phi_i)<b\).) This sequence has a limit point in \([a,b]\).
- If the frequency of truth diverges (like \(1-2+4-8...\)) then it’s still well-calibrated infinitely often.
- A trader can cheat; we want unbiasedness: no efficient method for detecting bias in beliefs.
- 4.3.6 Recurring unbiasedness: For any decidable ec \(\ol\phi\), any generable divergent weighting, weighted average of \(\Pj_i(\phi_i) - Thm_\Ga(\phi)\) has limit point 0.
- Bias converges when weighting is sparse enough so that \(\ol\Pj\) can gather sufficient feedback between guesses.
- 4.3.8 Unbiasedness from feedback: deferral \(f\). for \(Supp \ol w\subeq \im f\), \(Thm_\Ga(\phi_{f(n)})\) computable in \(O(f(n+1))\) time. (Figure out how previous elements turned out before forced to predict the next one.)
- 4.4 Learning statistical patterns
- Weaken pseudorandom.
- \(\ol w\) is \(f\)-patient if weight it places between \(n\), \(f(n)\) is bounded.
- 4.4.4. \(\ol p\)-varied pseudorandom sequence: Relative to set \(S\) of \(f\)-patient divergent weightings, \(\E_w(p_i-Thm_\Ga(\phi_i))\simeq 0_n\).
- 4.4.5. If there is some \(f\) such that \(\ol\phi\) is \(\ol p\)-varied pseudorandom, \(\Pj_n(\phi_n)\simeq_np_n\).
- (? where is assumption of \(n\) being decided before \(n+1\) presented?)
- 4.5 Logical relationships
- 4.5.1 Exclusive-exhaustive
- 4.5.5 Affine coherence: learns in a timely manner to respect all linear inequalities that actually hold between sentences, so long as those relationships can be enumerated in polynomial time.
- (is BCS bound for all \(\Pj\)?)
- 4.5.6 Persistence of affine knowledge.
- 4.5.7. Affine preemptive
- … (skip)
- 4.6 Non-dogmatism
- 4.6.1 Closure under finite perturbations
- Non-dogmatism doesn’t guarantee reasonable beliefs about theories
- Uniform non-dogmatism: For any computably enumerable sequence such that \(\Ga\cup \ol\phi\) is consistent, there is \(\ep>0\), \(\forall n\), \(\Pj_\iy(\phi_n)\ge \ep\).
- Domination, strict domination of universal semimeasure. (Logical inductors assign positive probabilities to set of completions of theories, universal semimeasures do not.)
- 4.7 Conditionals
- Define \(\mathbb V(\phi|\psi)\). \(=1\) if \(\mathbb V(\phi\wedge \psi)\ge \mathbb V(\psi)\).
- 4.7.2 Closure under conditioning. \(\Pj_n(-|\bigwedge_{i\le n}\psi_i)\) is logical inductor over \(\Ga\cup \Psi\).
- 4.8 Expectations
- Idea of expectation definition: integration by parts naturally puts in terms of \(\mathbb V\).
- 4.8.6 Expectation of indicators.
- 4.9 Trust in consistency
- \(Con(\Ga')(\nu)\) is “There is no proof of \(\perp\) from \(\ol{\Ga'}\) with \(\le \nu\) symbols.”
- Belief in finitistic consistency: for \(f\) computable, \(\Pj_n(Con(\Ga)(``f(n)"))\simeq_n1\).
- Learns to trust PA inductively.
- Also true if replace \(\Ga\) with \(\Ga'\) any recursively axiomatizable consistent theory.
- 4.9.4 Disbelief in (sequence of) inconsistent theories.
- 4.10 Halting
- 4.10.1/2. Learn (ec) halting patterns “\(m_n\) halts on input \(x_n\)” and nonhalting patterns.
- Only apply to cases where \(\Ga\) can prove machines halt or don’t. Else,
- 4.10.3 learning not to anticipate halting: in \(f(n)\) steps.
- It is impossible to tell whether or not a Turing machine halts in full generality, but for large classes of well-behaved computer programs (such as e.c. sequences of halting programs and provably non-halting programs) it’s quite possible to develop reasonable and accurate beliefs
- 4.11 Introspection
- Expectations of probabilities, iterated expectations
- 4.12 Self-trust
- 4.12.1 Expected future expectations: current expectation on day \(n\) already equal to expected value in \(f(n)\) days.
- 4.12.2/3: No expected net update under conditionals.
[Q: can SoS give reasonable estimates of probabilities to a given statement given a web of propositional statements?]
[Don’t want traders to do more computation beyond propositional correctness… they can still utilize eventual theoremhood by not keeping too much open.]
Construction
Given deductive process \(\ol D\), construct computable belief sequence \(\ol{LIA}\).
- Belief history \(\Pj_\le n\): sequence of belief states.
- \(n\)-strategy history \(T_{\le n}\): finite list of trading strategies, \(T_i\) is \(i\)-strategy.
MarketMaker
Sets market prices anticipating what a single trader is about to do.
Fixed point lemma Let \(T_n\) be \(n\)-strategy, \(\Pj_{\le n-1}\) belief history. There exists \(\mathbb V\), \(\Supp(\mathbb V)\subeq \Supp(T_n):=S'\), s.t. \[
\forall \mathbb\in \mathcal W, \mathbb W(T_n(\Pj_{\le n-1}, \mathbb V))\le 0.
\]
Proof. Define \(V' = [0,1]^{S'}\), \[
f(\mathbb V)(\phi):= \text{clamp}_{[0,1]} [\mathbb V(\phi) + T(\Pj_{\le n-1}, \mathbb V)[\phi]].
\] Idea: if the trader buys under \(\mathbb V\), then increase the price. (Note the scaling of \(T\) doesn’t really matter.) If the trader doesn’t buy/sell \(\phi\), there is no change.
By Brouwer on the simply connected compact \([0,1]^{S'}\), there is a fixed point \(\mathbb V^{\text{fix}}\). \(T_n\) only buys shares when \(\mathbb V^{\text{fix}}=1\), and sell when \(=0\): no profit!
MarketMaker: on input \(T_n, \Pj_{\le n-1}\), return \(\Pj\subeq [0,1]^{S'}\), \[
\forall \mathbb\in \mathcal W, \mathbb W(T_n(\Pj_{\le n-1}, \mathbb V))\le 2^{-n}.
\]
(Note we haven’t restricted to consistent worlds at all.)
Idea: brute force approximate search.
Let \(\mathbb W' = \mathbb W\one_{S'}\). Consider \[g:\mathbb V \mapsto \max_{\mathbb W'\in \mathcal W'} \mathbb W'(T_n(\Pj_{\le n-1}, \mathbb V)).\] We find a rational belief state \(\Pj\in (g')^{-1}((-\iy, 2^{-n}))\) because we can compute \(g\).
Lemma. MarketMaker \(\Pj_n=MarketMaker_n(T_n,\Pj_{\le n-1})\) is not exploitable by \(\ol T\) relative to any \(\ol D\).
(No assumptions on \(D\).)
Proof. Sum \(\le 1\).
Budgeter
Alter trader to stay within budget \(b\).
If there is some PC world where the trader could have lost $b or more, do nothing. Else, scale down to stay within budget in worst possible world.
Define Budgeter\({}_n^{\ol D}(b, T_{\le n}, \Pj_{\le n-1})\) by:
- if \(\mathbb W(\sum_{i\le m}T_i(\Pj_{\le i}))\le -b\) for some \(m<n\), \(\mathbb W \in PC(D_m)\), then 0.
- else: \(T_n \inf_{\mathbb W\in PC(D_n)}\ba{\max\pa{1,\fc{-\mathbb W(T_n)}{b+ \mathbb W(\sum_{i\le n-1} T_i(\Pj_{\le i}))}}^{-2}}\). (can write as \(\min\pa{1, \fc{b+ \mathbb W(...)}{-\mathbb W(T_n)}}\).)
Lemma (Properties):
- B doesn’t change \(T\) if for all past times \(m\le n\), in all PC worlds (\(\mathbb W\in PC(D_m)\)) the value is \(>-b\) (\(\mathbb W(\sum_{i\le m}T_i(\ol \Pj))>-b\)).
- (Stays within budget) \(\mathbb W(\sum_{i\le n}B_i^b(\ol\Pj)) \ge -b\).
- If \(\ol T\) exploits \(\ol\Pj\) relative to \(\ol D\)< then so does \(\ol B^b\) for some \(b\in \N^+\). (Proof. choose \(b\) to be the lower bound, then \(T\) doesn’t change.)
TradingFirm
Uses budgeter to combine infinite (enumerable) sequence of e.c. traders into a single trader that exploits a given market if any e.c. trader exploits the market.
(5.3.1) there exists a computable sequence of e.c. traders containing every e.c. trader.
Idea of construction of TradingFirm
- At finite step \(n\), incorporate the first \(n\) traders in the list \(\ol T^k\).
- We don’t know the right \(b\), so define a converging sum over \(b\).
In math:
- \(S_n^k = \one_{n\ge k} T^k_n\).
- TradingFirm\({}_n^{\ol D} (\Pj_{\le n-1}) = \sumo k{\iy}\sumo b{\iy}2^{-k-b} \text{Budgeter}_n^{\ol D} (b, S_{\le n}^k, \Pj_{\le n-1})\).
Note this is a computable finite sum because we can compute a lower bound on \(\mathbb W(\sum_{i\le m} S_i^k (\Pj_{\le m}))\), \(-C_n:=-\sum_{i\le n}\ve{S_i^k (\ol{\mathbb V})}_1\). When the budget is more than \(C_n\), the trading strategy doesn’t change.
Lemma (Trading firm dominance): If there exists any e.c. trader \(\ol T\) that exploits \(\ol\Pj\) relative to \(\ol D\), then \((\text{TradingFirm}_n^{\ol D}(\Pj_{\le n-1}))_{n\in \N^+}\) also exploits \(\ol\Pj\) relative to \(\ol D\).
Proof. Use repeatedly: If \(A_n\) exploits \(\ol\Pj\) and \(\mathbb W(\sum_{i\le n}A_n(\Pj)) \ge c_1 \mathbb W(\sum_{i\le n}T_n^k(\Pj)) + c_2\) for \(c_1>0\), then \(B_n\) exploits \(\ol \Pj\).
LIA (Logical Induction Algorithm)
Uses MarketMaker to make market not exploitable by TradingFirm.
Define \[
LIA_n := \text{MarketMaker}_n(\text{TradingFirm}_n^{\ol D}(LIA_{\le n-1}), LIA_{\le n-1}).
\]
LIA satisfies the logical induction criterion.
Proof. If any e.c. trader exploits LIA, then so does the trading firm, contradiction.
Runtime and convergence
Tradeoff between runtime of \(\Pj_n\) as function of \(n\) and how quickly \(\Pj_n(\phi) \to \Pj_\iy(\phi)\).
Selected proofs
Idea: if a market doesn’t satisfy nice properties, then there are ec traders taking advantage of this.
- Convergence: \(\Pj_\iy(\phi):=\limn \Pj_n(\phi)\) exists.
- Proof sketch: If \(\ol\Pj\) never makes up its mind about \(\phi\) then it can be exploited by a trader that buys at \(<p-\ep\) and sells at \(>p+\ep\). (cf. upcrossing inequality) Technicalities
- Use a continuous indicator function.
- Need to track net \(\phi\)-shares bought (holdings \(H_n\)). Don’t buy too many, so maintain bounded risk. (E.g. Cap at 1 at all times.)
- Limit coherence:
- Show that 3 properties are satisfied:
- \(\Ga \vdash \phi\implies \Pj_\iy(\phi)=1\).
- \(\Ga \vdash \neg \phi\implies \Pj_\iy(\phi)=0\).
- \(\Ga \vdash \neg(\phi\wedge \psi)\implies \Pj_\iy(\phi\vee \psi) = \Pj_\iy(\phi)+\Pj_\iy(\psi)\).
- Otherwise, we can construct a trader taking advantage of this “inconsistency”.
- Gaifman showed that these properties imply that \(\Pj\) extends to a probability measure. (Kolmogorov extension over \(\{0,1\}^S\)?)
- It’s important here that the trader considers PC worlds—over PC worlds, these trading strategies are bounded below. (More generally, if you allow other computable transformations, I think what happens is that \(\Pj_\iy\) will be guaranteed to respect those transformations.)
- Non-dogmatism
- Proof that if \(\Ga\not\vdash \neg \phi\), then \(\Pj_\iy(\phi)>0\). Otherwise, for every \(k\), make sure the trader buys one share of \(\phi\) for price \(\le 2^{-k}\).
- Idea: we can have \(\sum a_k 2^{-k}\) bounded (spend a bounded amount of money in any world) but \(\sum a_k (1-2^{-k})\) unbounded (potentially unbounded returns in the world where \(\phi\) is true).
- Pseudorandom frequencies
- Buy \(\phi_n\) shares whenever price goes below \(p-\ep\). \(p\) proportion pays out (by pseudorandomness—since there is by definition no way you can get something other than \(p\) by a poly trader looking at \(\Pj\)). Make trades continuous and budget.
- (Q: this seems probabilistic—how do you prevent having bad luck for arbitrarily long bounded time?)
- This seems more complicated than the rest; read this proof.
- Provability induction
- Use pseudorandomness with answer 1.
- Confusion: what if deduction process doesn’t hit all theorems?
Discussion
Compare with Solomonoff induction.
This (uncomputable) algorithm is impractical, but has nevertheless been of theoretical use: its basic idiom-consult a series of experts, reward accurate predictions, and penalize complexity-is commonplace in statistics, predictive analytics, and machine learning.
experts consulted by logical inductors don’t make predictions about what is going to happen next; instead, they observe the aggregated advice of all the experts (including themselves) and attempt to exploit inefficiencies in that aggregate model.
consider the task of designing an AI system that reasons about the behavior of computer programs, or that reasons about its own beliefs and its own effects on the world. While practical algorithms for achieving these feats are sure to make use of heuristics and approximations, we believe scientists will have an easier time designing robust and reliable systems if they have some way to relate those approximations to theoretical algorithms that are known to behave well in principle
3 takeaways.
- Make predictions by combining advice from ensemble of experts (Solomonoff + Gaifman). Note LIA many only see sequence of sets from a theorem prover. (Ex. only see “#1 is true” “#2 is false or #3 is true”, etc.)
- No meta-cognition (what to think about)
- Not practical—bad bounds
- Keep experts small. Experts do not predict the world. They identify inconsistencies even if they don’t know what’s actually happening.
- Make trading functions continuous.
Showing traders the current market prices is not trivial, because the market prices on day n depend on which trades are made on day n, creating a circular dependency. Our framework breaks this cycle by requiring that the traders use continuous betting strategies, guaranteeing that stable beliefs can be found… something like continuity is strictly necessary, if the market is to have accurate beliefs about itself.
This breaks paradoxes. Ex. \(\chi = "\Pj_n(\chi)<0.5"\).
Generality:
- not tied to any specific notion of efficiency.
- definition of trader flexible.
- not specific to domain of logic (all that is necessary is a set of atomic events that can be “true” or “false”, a language for talking about Boolean combinations of those atoms, and a deductive process that asserts things about those atoms over time)
Open problems
- Decision rationality: target specific, decision-relevant claims; reason as efficiently as possible about those claims. (Ex. reason about one sentence in particular.)
- Answers counterpossible questions.
- Why? Need to reason what would happen if agent had output a vs. b.
- Use of old evidence.
a strong solution to the problem of old evidence isn’t just about finding new ways to use old data every so often; it’s about giving a satisfactory account of how to algorithmically generate new scientific theories.
- Efficiency
Imagine we pick some limited domain of reasoning, and a collection of constant- and linear-time traders. Imagine we use standard approximation methods (such as gradient descent) to find approximately-stable market prices that aggregate knowledge from those traders.
Misc
“No dutch book” in expected utility theory, Bayesian probability theory
Given a sequence of bits which follow a polytime generable pattern, will it learn it in the limit?
Limitation (5.5.1): If \(\ol\Pj\) is a logical inductor over \(\Ga\) representing computable functions, \(\forall \phi, \Ga\vdash \phi, \forall n>f(\phi,\ep), \Pj_n(\phi) >1-\ep\), then \(f\) is uncomputable.
Otherwise, you can use \(f\) to design an algorithm that tells whether \(\Ga\vdash \phi\). (Falsify \(\phi\) by finding \(n>f(\phi,\rc b)\) and \(\Pj_n(\phi)\le 1-\rc b\).
If \(D\) does not eventually spew out a theorem \(\phi\), then there’s no guarantee of convergence for \(\phi\)?
Extended paper
- In what sense is self-trust impossible in formal logic? p. 8
- Where do you use the fact, in proving the consequences, that it was against polytime continuous traders with coefficients in \(\mathcal EF\)?