822 research outputs found
De-Quantising the Solution of Deutsch's Problem
Probably the simplest and most frequently used way to illustrate the power of
quantum computing is to solve the so-called {\it Deutsch's problem}. Consider a
Boolean function and suppose that we have a
(classical) black box to compute it. The problem asks whether is constant
(that is, ) or balanced (). Classically, to solve
the problem seems to require the computation of and , and then
the comparison of results. Is it possible to solve the problem with {\em only
one} query on ? In a famous paper published in 1985, Deutsch posed the
problem and obtained a ``quantum'' {\em partial affirmative answer}. In 1998 a
complete, probability-one solution was presented by Cleve, Ekert, Macchiavello,
and Mosca. Here we will show that the quantum solution can be {\it
de-quantised} to a deterministic simpler solution which is as efficient as the
quantum one. The use of ``superposition'', a key ingredient of quantum
algorithm, is--in this specific case--classically available.Comment: 8 page
Spurious, Emergent Laws in Number Worlds
We study some aspects of the emergence of logos from chaos on a basal model
of the universe using methods and techniques from algorithmic information and
Ramsey theories. Thereby an intrinsic and unusual mixture of meaningful and
spurious, emerging laws surfaces. The spurious, emergent laws abound, they can
be found almost everywhere. In accord with the ancient Greek theogony one could
say that logos, the Gods and the laws of the universe, originate from "the
void," or from chaos, a picture which supports the unresolvable/irreducible
lawless hypothesis. The analysis presented in this paper suggests that the
"laws" discovered in science correspond merely to syntactical correlations, are
local and not universal.Comment: 24 pages, invited contribution to "Contemporary Natural Philosophy
and Philosophies - Part 2" - Special Issue of the journal Philosophie
Is Feasibility in Physics Limited by Fantasy Alone?
Although various limits on the predicability of physical phenomena as well as
on physical knowables are commonly established and accepted, we challenge their
ultimate validity. More precisely, we claim that fundamental limits arise only
from our limited imagination and fantasy. To illustrate this thesis we give
evidence that the well-known Turing incomputability barrier can be trespassed
via quantum indeterminacy. From this algorithmic viewpoint, the "fine tuning"
of physical phenomena amounts to a "(re)programming" of the universe.Comment: 10 pages, contribution to "A Computable Universe," ed. by Hector
Zenil (World Scientific, Singapore, 2012), pp. 539-54
A priori Knowledge and the Kochen-Specker Theorem
We introduce and formalize a notion of "a priori knowledge" about a quantum
system, and show some properties about this form of knowledge. Finally, we show
that the Kochen-Specker theorem follows directly from this study. This version
is a draft version, the bibliography in particular is extremely scarce.
Comments welcome
Computing A Glimpse of Randomness
A Chaitin Omega number is the halting probability of a universal Chaitin
(self-delimiting Turing) machine. Every Omega number is both computably
enumerable (the limit of a computable, increasing, converging sequence of
rationals) and random (its binary expansion is an algorithmic random sequence).
In particular, every Omega number is strongly non-computable. The aim of this
paper is to describe a procedure, which combines Java programming and
mathematical proofs, for computing the exact values of the first 64 bits of a
Chaitin Omega:
0000001000000100000110001000011010001111110010111011101000010000. Full
description of programs and proofs will be given elsewhere.Comment: 16 pages; Experimental Mathematics (accepted
A Non-Probabilistic Model of Relativised Predictability in Physics
Little effort has been devoted to studying generalised notions or models of
(un)predictability, yet is an important concept throughout physics and plays a
central role in quantum information theory, where key results rely on the
supposed inherent unpredictability of measurement outcomes. In this paper we
continue the programme started in [1] developing a general, non-probabilistic
model of (un)predictability in physics. We present a more refined model that is
capable of studying different degrees of "relativised" unpredictability. This
model is based on the ability for an agent, acting via uniform, effective
means, to predict correctly and reproducibly the outcome of an experiment using
finite information extracted from the environment. We use this model to study
further the degree of unpredictability certified by different quantum
phenomena, showing that quantum complementarity guarantees a form of
relativised unpredictability that is weaker than that guaranteed by
Kochen-Specker-type value indefiniteness. We exemplify further the difference
between certification by complementarity and value indefiniteness by showing
that, unlike value indefiniteness, complementarity is compatible with the
production of computable sequences of bits.Comment: 10 page
- …
