www.nature.com/scientificreports/ |
www.nature.co /scientificr ports |
suai.ru/our-contacts |
quantum machine learning |
for the trial, and the stimulus information (coherence, direction, beep times). Everything was presented and recorded in Matlab using Psychtoolbox and a joystick mouse emulator19.
Procedure. Participants volunteered for the experiment by signing up through the laboratory on-line experiment recruitment system, which included mainly Michigan State students and the East Lansing community. Upon entering the lab, they completed informed consent and were briefed on the intent and procedures of the study. e rst experimental session included extensive training on using the scale and joystick, including approximately 60 practice trials on making accurate responses to speci c numbers, single responses to the stimulus, making two accurate responses to numbers in a row, and making two responses to the stimulus (as in the full trials).
Subsequent experimental sessions started with 30–40 “warm-up’ ‘ trials that were not recorded. A er training or warm-up, participants completed 22 ( rst session) or 28 (subsequent sessions) blocks of 12 trials, evenly split between con dence timings and stimulus coherence levels. e timing and coherence manipulations were random within-block, so each block of 12 trials included every combination of coherence (4 levels) and con dence timing. A er every block of trials, they completed 3 test trials where they were asked to hit a particular number on the con dence scale rather than respond based on the stimulus. is was included to get a handle on how accurate and precise the participants could be when using the joystick and understand how much motor error was likely factoring into their responses. Ultimately, motor error was controlled by grouping responses into the three main con dence levels - motor error was far less than the distance between con dence categories on the physical scale.
At the conclusion of the experiment, participants were debriefed on its intent and paid $10 plus up to $5 according to their performance. Performance was assessed using a strictly proper scoring rule20 so that the optimal response was to give a con dence response that re ected their expected accuracy. Participants received updates on the number of points they received at the end of each block of the experiment, including at the end of the study.
Mathematical Models. ere are di erent types of Markov processes that have been used for evidence accumulation. One type is a discrete state and time Markov chain21, another type is a discrete state and continuous time random walk process22, another type is a continuous state and discrete time random walk1,23, and fourth type is a continuous state and continuous time di usion process24. However, the discrete state models converge to make the same predictions as the di usion process when there are a large number of states and the step size approaches zero25.
Like the Markov models, there are di erent types of quantum processes. One type is a discrete state continuous time version10,11 and another type is a continuous state and time version8.
To facilitate the model comparison, we tried to make parallel assumptions for the two models. e use of a discrete state and continuous time version for both the Markov and quantum models serves this purpose very well. Additionally, a large number of states were used to closely approximate the predictions of continuous state and time processes.
For both models, we used an approximately continuous set of mental belief states. e set consisted of N = 99
states j {1, …, 99}, where 1 corresponds to a belief that the dots are certainly not moving to the right (i.e., a belief that they are certainly moving to the le ), 50 corresponds to completely uncertain belief state, and 99 corresponds to a belief that the dots are certainly moving to the right. We used 1–99 states instead of 0–100 states because we categorized the states into three categories and 99 can be equally divided into three sets. For a Markov
model, the use of N = 99 belief states produces a very closely approximation to a di usion process.
For the Markov model, we de ne ϕj(t) as the probability that an individual is located at a belief state j at time t for a single trial, which is a positive real number between 0 and 1, and ∑ϕj(t) = 1. ese 99 state probabilities
form a N ×1 column matrix denoted as ϕ(t). For the quantum model, we de ne ψj as the amplitude that an individual assigns to the belief state located a evidence level j on a single trial (the probability of selecting that belief
state equals |ψj|2). e amplitudes are complex numbers with modulus less than or equal to one, and ∑|ψ|2 = 1.
e 99 amplitudes form a N ×1 column matrix denoted as ψ(t). Both models assumed a narrow, approximately normally distributed (mean zero, standard deviation =5 steps in the 99 states), initial probability distribution at the start (t =0) of each trial of the task.
e probability distribution for the Markov process evolves from to time τ + t according to the Kolmogorov transition law ϕ(t + τ) = T(t) ϕ(τ), where T(t) is a transition matrix de ned by the matrix exponential function T(t) = exp(t K). Transition matrix element Tij is the probability to transit from the state in column j to the state in row i. e intensity matrix K is a N ×N matrix de ned by matrix elements Kij = α > 0 for i = j − 1, Kij = β > 0 for i = j + 1, Kii = −α − β, and zero otherwise. e amplitude distribution for the quantum process evolves from to time τ + t according to the Schrödinger unitary law ψ(t + τ) = U(t) ψ(τ), where U(t) is
a unitary matrix de ned by the matrix exponential function U(t) = exp(−i t H). Unitary matrix element Uij is the amplitude to transit from the state in column j to the state in row i. e Hamiltonian matrix H is a N ×N
Hermitian matrix de ned by matrix elements Hij = σ for i = j + 1, Hij = σ for i = j − 1, Hii = μ Ni , and zero otherwise.
For both models, we mapped the 99 belief states to 3 categories using the following three orthogonal projection matrices ML, MM, and MH. Define 1 as a vector of 33 ones, and define 0 as a vector of 33 zeros. Then
ML = diag[1, 0, 0], MM = diag[0, 1, 0] MH = diag[0, 0, 1]. Finally, de ne 
X 
1 as the sum of all the elements in
the vector X, and X 2 as the sum of the squared magnitude of the elements in the vector X. For the Markov model, the probabililty of reporting rating l at time t2 equals