Abstract
In their 2010 paper The Search for a Search: Measuring the Information Cost of Higher Level Search, the authors William A. Dembski and Robert J. Marks II present as one of two results their so-called Horizontal No Free Lunch Theorem. One of the consequences of this theorem is their remark: If no information about a search exists, so that the underlying measure is uniform, then, on average, any other assumed measure will result in negative active information, thereby rendering the search performance worse than random search. This is quite surprising, as one would expect in the tradition of the No Free Lunch theorem that the performances are equally good (or bad). Using only very basic elements of probability theory, this essay shows that their remark is wrong - as is their theorem.Definitions
The situation is quite simple: There is a set Ω1 - the search space - with N elements, so w.l.o.g. Ω={1,…,N}. Let T1⊂Ω1 denote the target. At first assume that T1 consists of a single element. To find this element T1 you have K guesses and you aren't allowed to guess the same element twice. Such a sequence of K different guesses can be described as an ordered K-tuple (n1,n2,…,nK) of K pairwise different numbers and is called a search. The N!(N−K)! different searches build the space ΩK. A search strategy is a probability measure Q on ΩK. Such a measure can be given by assigning a probability Q((n1,n2,…,nK))=q(n1,n2,…,nK) for each of the (pure) strategies (n1,n2,…,nK) in ΩK: q(n1,n2,…,nK) is the probability to choose the strategy (n1,n2,…,nK) from ΩK. So obviously, q(n1,n2,…,nK)≥0∀(n1,n2,…,nK)∈ΩK and ∑(n1,n2,…,nK)∈ΩKq(n1,n2,…,nK)=1. Random search (i.e., the random search strategy) correspond to the uniform distribution U on ΩK, i.e. U((n1,n2,…,nK))=(N−K)!N!. Any other search strategy is called an assisted search. This shows the small scope of the concept of an assisted search in the papers of Dembski and Marks: it is not possible to encode the influence of an oracle or a fitness function into the measure Q, so most examples presented by the authors (like partioned search or easter egg hunt are not covered by this model. The performance of a search strategy P(Q,δT1) for finding an element T1∈Ω1 is defined as the probability to name T1 in one of the K guesses of the search.Some calculations
Using the characteristic function χ{n1,n2,…,nK}(T1)={1ifT1∈{n1,n2,…,nK}0otherwise this probability can be written as: P(Q,δT1)=∑(n1,n2,…,nK)∈ΩKq(n1,n2,…,nK)⋅χ{n1,n2,…,nK}(T1) The characteristic function is independent of the order of the elements n1,n2,…nK. So the expression can be streamlined using instead of the space of ordered n-tuples ΩK the space PK(Ω1) of subsets {n1,n2,…,nK} with K elements of Q1. This space has (NK) elements. Q on ΩK gives instantly rise to a measure on PK(Ω1) - again called Q - by setting: Q({n1,n2,…,nK})=∑σ∈SKq(σ1(n1),σ2(n2),…,σK(nK)). Here, σ=(σ1,σ2,…,σK) are the K! elements of the group of permutations of K elements, SK. Thus, the probability to search successfully for T1 is P(Q,δT1)=∑{n1,n2,…,nK}∈PK(Ω1)Q({n1,n2,…,nK})⋅χ{n1,n2,…,nK}(T1). But how does this search perform on average - assuming that the underlying measure is uniform? The underlying measure gives the probability with which a target element is chosen: if there is a uniform underlying measure UΩ, each element 1,2,…,N is a the target with a probability of 1N. The average performance is given by: P(Q,UΩ)=∑Nl=11N∑{n1,n2,…,nK}∈PK(Ω1)Q({n1,n2,…,nK})⋅χ{n1,n2,…,nK}(l) =1N∑{n1,n2,…,nK}∈PK(Ω1)Q({n1,n2,…,nK})⋅(∑Nl=1χ{n1,n2,…,nK}(l)) =1N∑{n1,n2,…,nK}∈PK(Ω1)Q({n1,n2,…,nK})⋅K =KN∑{n1,n2,…,nK}∈PK(Ω1)Q({n1,n2,…,nK}) =KN. This holds for every measure Q on the space of the searches, especially for the uniform measure, leading to theSpecial Remark: If no information about a search for a single element exists, so that the underlying measure is uniform, then, on average, any other assumed measure will result in the same search performance as random search.
But perhaps this result depends on the special condition of an elementary target? What happens when Tm={t1,t2,…,tm}∈Pm(Ω1)? Then a search is seen as a success if it identifies at least one element of the target Tm, thus P(Q,δT1)=∑{n1,n2,…,nK}∈PK(Ω1)q{n1,n2,…,nK}⋅maxt∈Tm{χ{n1,n2,…,nK}(t)}. Calculating the average over all elements of Pm(Ω1) results in: P(Q,UPm(Ω1))=1(Nm)∑T∈Pm(Ω1)∑{n1,n2,…,nK}∈PK(Ω1)q{n1,n2,…,nK}⋅maxt∈Tm{χ{n1,n2,…,nK}(t)}. =1(Nm)∑{n1,n2,…,nK}∈PK(Ω1)q{n1,n2,…,nK}∑T∈Pm(Ω1)maxt∈Tm{χ{n1,n2,…,nK}(t)} =1(Nm)∑{n1,n2,…,nK}∈PK(Ω1)q{n1,n2,…,nK}⋅((Nm)−(N−Km)) =(Nm)−(N−Km)(Nm)∑{n1,n2,…,nK}∈PK(Ω1)q{n1,n2,…,nK} =(Nm)−(N−Km)(Nm)=1−(N−K)(N−K−1)…(N−K−m+1)N(N−1)…(N−m+1). Again the probability of a successful search - and by definition the performance - is on average independent of the measure used on the space of the searches! (Note that the expression becomes KN for m=1 and 1 for N−K<m). The result of this calculation is the
Less Special Remark: If no information about a search for a target exists other than the number of elements the target exists from, so that the underlying measure is uniform, then, on average, any other assumed measure will result in the same search performance as random search.
As the performance of the random search and any other search doesn't differ on any subset of targets of fixed cardinality, the next remark is elementary:
General Remark: If no information about a search exists, so that the underlying measure is uniform, then, on average, any other assumed measure will result in the same search performance as random search.
Obviously this General Remark negates the
Remark of Dembski and Marks: If no information about a search exists, so that the underlying measure is uniform, then, on average, any other assumed measure will result in negative active information, thereby rendering the search performance worse than random search.
No comments:
Post a Comment