Abstract
Many natural processes rely on optimizing the success ratio of a search process. We use an experimental setup consisting of a simple online game in which players have to find a target hidden on a board, to investigate how the rounds are influenced by the detection of cues. We focus on the search duration and the statistics of the trajectories traced on the board. The experimental data are explained by a family of randomwalkbased models and probabilistic analytical approximations. If no initial information is given to the players, the search is optimized for cues that cover an intermediate spatial scale. In addition, initial information about the extension of the cues results, in general, in faster searches. Finally, strategies used by informed players turn into nonstationary processes in which the length of e ach displacement evolves to show a welldefined characteristic scale that is not found in noninformed searches.
Introduction
The problem of searching for targets whose location is unknown arises in many fields and at different scales^{1,2,3}. Numerous examples appear in the natural sciences including in ecology^{4,5,6,7,8}, biochemistry^{9,10,11} and chemistry^{12}. In addition, many human activities involve situations where a target has to be found. Some instances are the location of a lost object, rescue operations, or fugitive prosecutions^{13}. More recently, the development of eyetracking technology has allowed the study of visual searches on screens^{14,15,16}. In order to understand the social, biological and physical mechanisms behind these processes, it is essential to have empirical evidence of the performance of different strategies and how they are affected by environmental cues, regardless of whether they are employed by humans, animals or bacteria^{17}. Such data are also required to verify the mathematical models that have been proposed^{18,19,20,21,22,23,24,25}, and to develop improved protocols.
Situations in which a target has to be located appear in a large variety of scenarios, which allows the design of multiple strategies to find a successful solution. Such strategies can be classified in many different ways, according to one or more of their properties^{1}. For instance, stochastic or systematic processes are distinguished depending on the type of search rule^{2} and the amount of directional information available determines the existence of bias towards preferred regions^{26,27}. Finally, differences may also be attributable to the movement pattern, such as cruising versus ambush^{28} and to the frequency of the reorientation events, such as intensive (frequent) versus extensive (infrequent)^{29,30}. The effectiveness of a particular choice within each category is determined by the properties and the state of the searcher, the target, and the environment where the task has to be accomplished. For instance, searchers with memory that navigate relatively predictable environments do not employ purely random strategies but combine a stochastic component with knowledge acquired through previous experience. There is therefore a learning process that plays an important role in the emergence of new rules^{31,32}. In other scenarios, individuals who live in groups may incorporate information gathered by conspecifics with their own in order to improve foraging efficiency. It has been recently showed that intermediate combinations between both types of cues result in more efficient searches regardless of the nature of the mobility pattern^{33} and the spatial distribution of the targets^{34,35}. However, the precise optimal balance between social and individual information is determined by each specific setup.
In all of these scenarios, interactions with the environment provide the searcher with information that may alter the effectiveness of a given strategy over the course of the search. Therefore, in the most general case, search strategies must be interpreted as dynamical processes consisting of several components rather than fixed procedures. For instance, many predators respond to the detection of cues indicating the proximity of prey by increasing their turning angles and reducing their speed in order to scan the local environment more carefully^{36,37}, which leads to concentration of the search activity in areas of high prey density^{38}. This behavior has been reported in several species of insects^{39}, seabirds^{40,41} and also in human searchers looking for hidden resources in open environments^{42}. Other phenomena that trigger sudden changes in individual movement behavior are changes between habitats^{43} and changes in the amount and quality of information gathered by the searcher^{7}.
In this work we propose the use of computer games as a new experimental approach with which it is possible to address these and related questions in humans. This is particularly intriguing since, due to their cognitive abilities, individuals might show a large diversity of complex responses to the same stimulus. Despite substantial efforts aimed at understanding the theoretical concepts behind many search processes, a reliable and unifying empirical framework in which these ideas may be tested is still lacking. The family of games presented here is a good candidate to fill this gap, as they can be accessed online by a large number of players. This results in the generation of large and clean datasets. In addition, the rules and setup of the game can be experimentally manipulated so that different mechanisms or strategies can be rigorously tested. First, we address several questions related to search efficiency and investigate how the strategies change due to the amount and the quality of information acquired by the player at different stages of the game. In a second step, the main features of these patterns are extracted from the data and used to develop a family of random walk models that can be applied to predict human search behavior in other configurations of the game. The variety of experiments shown in this work reinforces the flexibility of our approach and aims to open a new route for the study of search problems.
In the following section, after presenting the characteristics of the game, we show the empirical results obtained from two different setups. In the first case, players have no information about the configuration of the board, whereas in the second study they are provided with partial information about it. Then, we formulate a family of models that capture the main mechanisms behind the experimental results and derive analytical approximations to show the robustness of the results. Finally, all the previous steps are combined to develop a comprehensive framework in which it is possible to predict the optimal configuration of the landscape that yields faster searches. The paper finishes with a discussion of the results and opportunities for new lines of research.
Results
Experimental setup
We consider a simple game in which a single target has to be found. It slightly resembles the classic minesweeper, although the objective is to find a unique target (mine) instead of avoiding a collection of them. The interface consists of N × N squares that can be explored by the player through successive clicks with the mouse. There are three classes of cells depending on their color after being clicked (unclicked cells are always blue): (i) black cells are typically far from the target, (ii) yellow cells indicate that the target may be one of the neighboring cells and (iii) the single red cell is the target. The target is randomly located within a patch of yellow cells. Therefore, it provides partial information about the configuration of the board. Two different geometries for this set of yellow cells are explored here. First, in the next two sections they form a N_{y} × N_{y} neighborhood square region (Fig. 1a). Second, in the last section of the Results they will outline a random patch whose size will be measured in terms of the number of yellow cells. Further details about the implementation of these random neighborhoods will be provided in that section. The discovery of a yellow cell indicates that the player is in the neighborhood of the target and thus reduces the area that needs to be scanned. For simplicity we fixed N = 20 in all the experiments and then manipulated N_{y}.
To generate the dataset players access anonymously the game online and are asked to find the target using as few clicks (jumps on the board) as possible. Since players are not identified separately, we cannot identify the number of rounds played by each user. The rounds are all independent (different configurations of the board) and each one is represented by the trajectory traced by the player on the board. Finally, the experimental setup also includes a timer. In order to study the limit in which searches are more stochastic, players are requested to find the target as quickly as possible. This constraint also mimics many real situations both for humans and nonhumans in which time is a limitation for the search. Some instances are human rescue operations or animal foraging while avoiding predators.
In the following sections we investigate i) how the duration of the search, represented by the number of mouse clicks, changes with the size of the target’s neighborhood (also called yellow region); and ii) the statistical properties of the searching patterns as defined by the distance between clicks d_{i} (jump length) and the turn angles θ_{i}. By definition, we consider turns to the left to be between 0^{°} and 180^{°} and turns to the right to be between 180^{°} and 360^{°} (see Fig. 1b for a definition of both quantities). We consider two classes of experiments: a) blind searches, where the player is given no a priori knowledge of the size of the neighborhood, and b) searches with initial information, where the value of N_{y} is given to the player at the beginning of the round. The objective of performing both classes of experiments is twofold: on the one hand to investigate whether players adapt their searching strategies when they have better information about the landscape and, on the other hand, to examine how search efficiency changes when the reliability of the information provided by the yellow cells increases.
Experiments with blind searchers and square neighborhoods
For this first series of experiments neither the exact size, the position of the yellow region, nor a range of possible dimensions was given to the searchers. Before starting the round, each player only knew that a target (red square) was hidden in the board and it might be randomly placed inside a square vicinity of yellow cells of unknown size. The uncertainty in the size of the neighborhood reduces the reliability of the information acquired by the player when a yellow cell is clicked and favors the efficiency of random strategies^{1}. Our dataset consists of 500 rounds with N_{y} ranging from N_{y} = 1, which means that the target does not have a neighborhood, to N_{y} = 13. A distribution of the number of rounds for each value of N_{y} is shown in the Supplementary Table I. We first measure the mean number of clicks needed to find the target as a function of the lateral length of its yellow neighborhood (black squares in Fig. 2a).
Due to the design of the experiments, there is a tradeoff between finding the yellow region and finding the target inside it. Larger neighborhoods are easier to locate but make the final detection of the target inside them harder. Smaller neighborhoods, however, need on average more steps to be found but make the target within them easier to locate (Fig. 2b). According to our results, this tradeoff is balanced at intermediate sizes of the neighborhood, . This resembles the foraging dynamics of animals that exchange information about food location with their conspecifics, so that both spreading information over distances that are either too large or too short may slow down the search^{35}. Following this analogy, we refer to the the size of the yellow area that minimizes the number of clicks needed to find the target as the optimal interaction range. The standard deviation of the number of jumps is also minimal at the optimal range, which means a narrowing in the distribution of clicks used to detect the target and therefore a reduction in the stochasticity of the search. In the limit of zero information (i.e. no yellow cells or Ny = 1, or the whole board is yellow, Ny = N), the probability of finding the target on the first click is given by the inverse of the number of available cells, 1/N^{2}. In any subsequent movement, m, this probability is given by
where the first term yields the probability of not having found the target in the previous m − 1 clicks and the second term yields the probability of hitting the target once m−1 squares have been visited. Eq. (1) reduces to 1/N^{2} regardless of the value of m. Therefore, the probability of detection in the limit N_{y} = 1 (and N_{y} = N) follows a uniform distribution of mean N^{2}/2 and standard deviation , which is in good agreement with data (black squares in Fig. 2a).
Next, we analyze all the trajectories traced by the players in every round. To facilitate this, the experimental setup saves the sequence of clicks in each round, from which we calculate the length of each displacement and the angle of each turn. We identify extensive and intensive searching modes that depend on whether the player has detected a yellow cell or not respectively (Fig. 3a). In both situations the jump lengths can be fitted using exponential distributions, with the intensive phase showing a lower mean value 1/λ_{int} = 2.04 and 1/λ_{ext} = 3.70 (1/λ_{int}, 1/λ_{ext} are the mean length of the displacements in the intensive and the extensive phase respectively). Therefore, the typical size of the jumps is reduced once the player finds the yellow area as the detection of the cue (represented by a yellow cell) triggers an arearestricted search^{36,37}. Although the player does not know how big the neighborhood is and therefore how reliable the information is, the trajectories recorded after the discovery of the yellow region still show shorter distances between turns, suggesting that players switch to an intensive search mode once they find the yellow region^{1}. It is important to remark that, although alternation between extensive (motion phase) an intensive modes (scanning phase) is also characteristic of intermittent searches, the player is not performing an intermittent search as it has been defined in the literature before^{2}. The differences lie in two points. First, in our study the switch between reorientation modes is triggered by the external cue instead of taking place at random. Second, the detection of the target may take place in both phases instead of being limited to the extensive one. Regarding to the type of motion, we study, however, a spatially intermittent search since the player performs a saltatory trajectory in which the target can be found only if the searcher lands on it. This differs from the case of a cruise forager who looks for targets while moving and that would constitute a completely different study.
Regarding the turn angles, both the extensive (before the first encounter with a yellow cell) and the intensive phases (after detecting the first yellow cell) show correlations between subsequent turn angles (Fig. 3b,c, respectively). This could indicate that the strategies are not completely random but contain some systematic features. In fact, a frequent strategy consists of tracing a series of short jumps in the same direction. To reduce searching times, players show a tendency to scan a direction doing several consecutive clicks. This behavior is also seen in the distributions of jump lengths, since they show a large deviation from the exponential for onecell length jumps, which are overrepresented in the dataset (Fig. 3a). The higher frequency of turning angles closer to zero is linked to the higher presence of jumps of length one. The explanation for this persistence in the direction of movement shown in Fig. 3b,c is probably a combination between the attempt of some players to design purely systematic strategies and the intrinsic tendency of humans to keep visually scanning in the same direction^{16}.
As an exception, movements done immediately after a yellowtoblack transition show a strong tendency to reverse the direction, as this sequence in the colors of the cells indicates that the player is moving away from the target (Fig. 3d).
Experiments with initial information and square neighborhoods. The case of N_{y} = 5 as compared to the blind case
In this second series of experiments the players know the size of the yellow region, which is fixed at the optimal interaction range N_{y} = 5. This increases the quality of the information obtained when one of the yellow cells is found as the player can limit the search area. The position of this area, as well as the location of the target inside it, is random, changes from round to round and is unknown to the player.
Data from 230 rounds were collected. As a general result, a priori information accelerates the search and reduces its stochasticity. Blind searchers need on average 31.30 clicks to find the target when N_{y} = 5 (subset of 65 rounds from the 500 trajectories analyzed for blind experiments, see Supplementary Table I), while informed players use 25.5 clicks. The twotailed P value on the difference of these mean values obtained using an unpaired ttest, 3 × 10^{−4}, is highly statistically significant. The standard deviation also decreases, indicating a narrowing in the distribution of the number of displacements and therefore in the randomness of the process: σ_{b} = 14.10 for blind searchers and σ_{i} = 10.50 for the informed ones. To find out what stage of the search is more strongly affected by the initial information, we analyze the number of clicks done in each phase of the search. We repeat this factorization for the blind and the informed cases and compare both of them (Fig. 4a). From Fig. 4b and c, we observe that all of the reduction in the number of jumps accumulates in the intensive phase, while the extensive stage remains unaltered by the initial information. More interestingly, if the number of displacements that take place between two yellow cells is subtracted from the total number of jumps of the intensive phase (Fig. 4d), we observe that this quantity remains almost unchanged. There is, however, an important reduction in the number of displacements that correspond to the rest of the combinations of cells (black to yellow, yellow to black and black to black jumps; Fig. 4e). In fact, the percentage of yellowtoyellow movements that take place during the intensive phase increases from 54% to a 75% in the informed searches. This result indicates that having information about the size of the yellow zone allows a faster detection of its limits and therefore reduces the number of movements spent to find the target.
Regarding to the statistical analysis of the trajectories, initial information about N_{y} also yields some differences in the distributions of the lengths of the jumps and the turning angles. Informed players adapt their displacements during the extensive phase, concentrating the length of their movements around the size of the yellow neighborhood, N_{y} = 5 (Fig. 5a). If we analyze the whole set of informed rounds, we observe a strong dominance of movements of length N_{y} = 5 (green squares in Fig. 5a). This is due to the presence of approximately 50 rounds in the dataset where players performed optimally designed systematic strategies that consist of moving in jumps of fixed length N_{y} during the extensive phase. We will come back to this in the description of a random walk based model for this process. For the purposes of this section we will remove these systematic rounds and focus on the subset of stochastic strategies formed by the other 180 rounds. The distribution of the length of the displacements is still dominated by jumps that cover a distance of the order of N_{y} (red circles in Fig. 5a). For the subset of blind searchers with N_{y} = 5, however, this distribution does not show a well defined typical scale and instead, players explore several scales as they look for the yellow region (Fig. 5b). For the intensive phase, informed searches also show a higher abundance of onecell displacements than the distribution of the blind searches (inset of Fig. 5a and b respectively). This result is independent of whether or not the systematic deterministic strategies are included within the analyzed dataset and is due to the fact that knowing the neighborhood size reduces exploration during this phase. Finally, giving the size of the yellow neighborhood to the players in advance also has an effect on the distribution of turns made by the searcher immediately after a yellowtoblack displacement. This distribution is shown in Fig. 5c for informed strategies and in Fig. 5d for blind searches. Although in both cases the movement shows a strong bias backwards, informed searches result in distributions with a stronger peak around θ = 180^{°}. This is due to the fact that players do not have to find out the size of the neighborhood of the target and consistent with the factorization of the number of clicks shown in Fig. 4e.
We conclude this section with an analysis of the trajectories during the extensive phase, in order to find the mechanism by which a characteristic length scale appears in the jump length distribution. We find the existence of a feedback between the searcher and the environment that makes the extensive phase nonstationary (the mean value of the distribution changes with time). This feedback allows a progressive narrowing of the jump length distribution around N_{y} as the extensive phase evolves and the searcher gathers and accumulates information from the landscape. Since the player has perfect memory about his trajectory (visited cells remain open), trajectories that start with large displacements tend to create landscapes that are fragmented in patches of length N_{y} in which long movements are inefficient. To show the existence of this feedback we split the data of the extensive phase in four subdivisions: (i) from jump 1 to 5, (ii) 6 to 10, (iii) 11 to 15, and (iv) 16 to the end. The distributions for each of these pieces are shown in Fig. 6a,b,c and d respectively, and they can be fitted by a family of gamma distributions (dashed lines in each panel) of decreasing mean, mode and variance (See Table 1 for numerical values of these parameters and details of the distributions). Then, the total distribution of Fig. 5a can be approximated by a gamma function defined in terms of the parameters of the distributions of the pieces (dashed line in Fig. 5a). This approach shows an excellent agreement with a direct fitting of the whole extensive phase (full line in Fig. 5a).
At this point, we have shown the existence of an optimal size for the neighborhood of the target, as well as an improvement in the search efficiency when the value of N_{y} is revealed at the beginning of the round. In addition, these informed strategies evolve through information gathering during the extensive phase towards a dominant jumping distance equal to the lateral length of the neighborhood of the target, N_{y}. In the following sections we develop a theoretical framework and a family of models based on random walks to study the basic principles behind these results and how they can be transfered to more general scenarios, with irregular shapes for the information region.
Model for blind searchers: numerical simulations and analytical approximation
We develop a minimalistic searching model based on random walks to explain previous experimental results on the basis of simple dynamical rules. The model has the three main ingredients obtained from the data analysis: (i) two modes of movement defined by the mean length of the displacements: 1/λ_{int} and 1/λ_{ext}; (ii) in the absence of any information (no yellow cell clicked) the direction is completely random (uniform distribution in the turning angles); and (iii) when cues are obtained (a yellow cell has been detected), the searcher has a bias towards unvisited cells surrounding a yellow one. The choice of a uniform distribution for the turning angles is a consequence of using a purely exponential distribution for the length of the displacements (see Fig. 3a). The high persistence shown by the experimental turning angle distribution, which can be approximated by a uniform distribution except for that peak at θ = 0 (see Fig. 3b,c), comes from the high presence of jumps of length one. Disregarding the high frequency of unitylength movements also implies disregarding the higher abundance of turning angles close to zero and therefore using a uniform distribution for θ. The third assumption aims to capture the influence of the information provided to the searcher when a yellow cell is found, as well as the strong tendency to go back to yellow cells exhibited by the distribution of turning angles in Fig. 3d.
The results of the simulations (green curve in Fig. 2) show an excellent agreement with the experimental data (black curve) both in the mean average number of jumps and in its standard deviation. Simulations reproduce at least the two first moments of the number of clicks distribution.
Except in the limits N_{y} = 1 (no yellow cells) and N_{y} = N (yellow cells occupy the whole board), it is hard to obtain exact analytical expressions for the average total number of jumps needed to find the target. However, it is possible to obtain the distribution for the length of the extensive phase
where P_{i}(N_{y}) is the probability of having an extensive phase of i jumps when the neighborhood of the target has a lateral length N_{y} and is the probability of finding a yellow cell in the i − th mouse click. In words, the probability of having an extensive phase with i jumps is given by the probability of not finding a yellow cell in all the previous movements multiplied by the probability of finding one in the i − th movement. Given Eq. (2), the mean length of the extensive phase is
For the length of the intensive phase however we can only give and upper and a lower limit, assuming that after the detection of the first yellow cell all the movements are to neighboring cells. Therefore, the target is found on average after jumps in the intensive phase when the neighborhood of the target is large and after (N_{y} + 2)^{2}/2 movements when the neighborhood is small. These two limits account for the decreasing probability of visiting cells outside the neighborhood when increasing its size. For small values of N_{y} it is very likely to reach the border of the neighborhood before detecting the target and thus to return to the black region. Combining these two results for the intensive phase with the length of the extensive phase obtained in Eq. (3), we obtain two theoretical approximations to the total number of clicks
The combination of these two expressions gives an approximated range for the length of the search (magenta region in Fig. 2) that shows an excellent agreement with empirical data and numerical simulations of the model.
Model for searches with initial information. The design of optimal strategies.
Knowing the size of the yellow region at the beginning of the game changes the nature of the search as the information gathered by the player with each movement may be used to design the next displacement. This reinforces the nonMarkovian nature of the informed search process as the player uses all the previous steps to discard cells that have not been visited yet and results in selfadaptive strategies that evolve towards displacements of length N_{y}. Also, as the value of N_{y} is known, the number of exits from the neighborhood of the target diminishes (Fig. 4e). In a first approach to model this effect, we modify the model used for blind searches using the new experimental distribution of the length of the displacements in both the extensive and the intensive modes(Fig. 5a). Therefore, instead of using the exponential distributions of Fig. 3 we sample the histograms of Fig. 5a (red circles) and its inset, that are obtained from experimental searches with initial information. This approach overestimates both the length of the extensive and the intensive phases, which results in a clearly higher average number of movements; 33.50 jumps, σ = 19.00 for the model and 25.50 jumps, σ = 10.50 in the data (MB green bars and DI gray bars in Fig. 7 respectively). This is due to the fact that the model does not integrate the information about the size of the target to a priori discard some of the cells during the intensive and the extensive phase.
In a first approach to remove this discrepancy, we hypothesize that the most important differences arise in the modeling of the intensive phase. During this stage, given a certain number of yellow cells and some of their neighboring black squares, our experimental results suggest that human players are able to discriminate the real border of the neighborhood of the target and thus reduce the number of erroneous displacements. The model that we developed for blind searches lacks this ingredient, which increases the duration of the intensive phase since more black cells are open. To correct this, we modify the model and include the effect that previous movements, together with knowing the size of the neighborhood of the target, have on the intensive phase (See Methods for a detailed description). In this new approach, once the first yellow cell has been detected and based on all the previous movements, only those cells that can possibly be part of a 5 × 5 yellow square have a nonzero probability of being visited by the searcher. This mechanism reduces the number of times that black cells are visited once a yellow cell has been found as the model is able to discriminate all the possible borders of the neighborhood of the target. With this new ingredient the efficiency of the model increases (MI blue bar in Fig. 7a) and the number of jumps in the intensive phase shows excellent agreement with the experimental data (DI gray and MI blue bars in Fig. 7c). However, despite this substantial improvement as compared to the blind model, significant differences still remain between empirical data and numerical results. The source of this disagreement arises from the extensive phase (DI gray and MI blue bars in Fig. 7b). To correct this, we next modify the extensive phase of the model.
During the extensive phase, players are able to discriminate regions where the target cannot be placed as a 5 × 5 square would not fit. To incorporate this in the model, we first compute the probability of jumping to each of the nonvisited cells of the board according to the histogram in Fig. 5a. Then, for each cell we obtain all the possible squares of lateral length 5 to which it could belong and set the probability of jumping to that cell to zero if all these squares contain at least one open black cell (See Methods for more details). With this mechanism the extensive phase becomes more efficient and the agreement of the model with the experimental data is excellent. More importantly, this comes from a precise fitting of both the intensive and the extensive phase individually (DI gray and MII black bars in Fig. 7a,b and c).
Optimal strategy
However, both actual player strategies and random walk models are much less efficient than entirely systematic protocols. Knowing a typical size of the target in advance allows the design of optimized strategies that minimize the number of incorrect steps. Particularly important is to shorten the extensive phase, as within the neighborhood of the target all the cells are equivalent and it is equally likely to find the target in any position. In fact, during the experimental rounds with initial information, one of the players developed one of these searching methods by repeatedly playing with the same size of N_{y} = 5.
This strategy optimizes the extensive phase and only allows for two yellowtoblack transitions during the intensive phase (Fig. 8a). Given a value for N_{y}, the search rule is given the following steps:
 1
Divide the board in theoretical squares of size N_{y} × N_{y} (see Fig. 8a)
 2
Click in the upper right corner of each subdivision. Start with those squares whose upper right cell has more neighbors and continue with those in the borders. This reduces the length of the extensive phase on average as corners that are farther from the border are more likely to contain a yellow cell.
 3
Once a yellow cell is found, visit consecutive squares in a given direction (horizontal in Fig. 8a for N_{y} = 5) until finding a black position. Then, if the number of yellow cells in the row is lower than N_{y}, complete it.
 4
Repeat the same operation in the other direction starting from one already known yellow cell.
 5
Once the neighborhood of the target has been delimited, move inside it until finding the target.
In the particular case of N_{y} = 5, the average number of movements before target detection following this strategy is 19.03 (10^{4} realizations) and it is always lower than 42. In addition the extensive phase has a duration of 5.90 clicks on average, which is about 50% lower than the experimental result. This improvement is much higher than the one observed for the intensive phase, which can be optimized by players once they are provided with initial informaition about the landscape (see gray and red bars in Fig. 7b,c for a comparison). In real human scenarios, this result suggests that efforts put into optimizing the extensive phase may pay off more than equivalent efforts to optimize the intensive phase.
Applying this optimal strategy to many sizes of the yellow region (Fig. 8b) we observe that the tradeoff between finding the neighborhood of the target (yellow diamonds in Fig. 8b) and finding the target inside it (blue circles in Fig. 8b) balances at intermediate values of N_{y}. Following theoretical results for blind experiments, analytical expressions can be obtained for the mean number of movements during both phases and therefore for the optimal interaction range. The mean number of clicks during the intensive phase is as the target can be in any cell with the same probability (green dashed line in Fig. 8b) (we only consider the lower bound obtained for blind experiments since this optimal protocol minimizes the number of erroneous movements). To obtain the mean number of movements in the extensive stage, we assume that the upper right corner of each subdivision of the board (Fig. 8a)is equally likely to have a yellow cell. Therefore, the number of steps is given by . This is not completely true, as cells close to the border have a lower probability of being yellow, but it is a good approximation (black dashed line in Fig. 8b fitting yellow diamonds). At the optimal interaction range both functions intersect, which gives for our experimental setup with N = 20. This result is in excellent agreement with the value obtained from the experiments (Fig. 8c) and suggests, together with the theoretical approximation, that the optimal interaction range is independent of the searching strategy. This result suggests the possibility of using this theoretical framework to predict the optimal size of the neighborhood of the target in more general scenarios.
Anticipating the optimal range of interaction for random neighborhoods
In this section we allow the target to adopt different sizes and random shapes across rounds. In order to facilitate the formulation of theoretical predictions, the neighborhood is built starting from a triangle of varying base b_{y} (see Methods for a detailed description and Fig. 9) where the target is embedded. Then, the region is randomized by turning 30% of the cells black. In this way, we implement random neighborhoods that vary in form and size from round to round but with an underlying fixed pattern. Before starting the game, players know that the neighborhood has now a varying form and size (Fig. 9), but they are given no information about the way it is constructed.
The optimal interaction range can be evaluated from an independent estimation of the number of movements needed in the extensive and the intensive phases. The length of the extensive phase is obtained following the same steps used for square neighborhoods; the probability of finding a yellow cell in the i−th movement is given by Eq. (2) from where the mean length of the extensive phase is obtained using Eq. (3). This quantity is shown by the magenta circles in (Fig. 10b). To approximate the number of movements used in the intensive phase, which will give us the optimal interaction range we used the underlying triangle shape of the neighborhood of the target. This calculation provides lower and upper bounds for the average duration of the intensive phase. The lower bound is obtained assuming that all the cells from the original target have the same probability of being visited but all the cells that do not belong to it will never be clicked. The total number of cells that form this original triangle is (b_{y}/2 + 0.5)^{2} and since all the cells can be visited with the same probability, the lower limit for the length of the intensive phase is given by . The upper limit is obtained assuming that the first cells that do not belong to the triangle in each direction also have a nonzero probability of being visited. This results in an upper bound for the length of the intensive phase given by . Both limits are shown by the magenta circles in Fig. 10c. Finally, the total number of movements, i.e., the sum of the extensive and the intensive phase, is shown by the magenta circles in Fig. 10a, with an estimated optimal neighborhood size in between 18 and 25 yellow cells. It is important to note the difference between the optimal interaction range for random and square neighborhoods, which shows the non triviality of predicting optimal interaction ranges for different geometries.
We tested these a priori predictions with a series of experiments using an experimental setup with neighborhoods that consist of 5, 16, 33, 55 and 69 cells (plus the target red cell, see Supplementary Table II for a distribution of the number of rounds with each size). 301 rounds were analyzed and the observed mean number of clicks is shown in Fig. 10a. We also split each round into the extensive and intensive phases and the results are shown in panels b and c of Fig. 10. The good agreement between the predicted values and the results obtained with the experiments shows the robustness of the theoretical approach developed in simpler scenarios.
Discussion
We have developed a novel approach to study human search problems by building a simple game that can be accessed online. This approach facilitates the collection of large and clean experimental datasets. By combining data analysis with probabilistic calculations and numerical simulations of existing and new models, it is possible to obtain a deeper understanding of how humans approach simple search tasks and how their strategies differ from optimal patterns.
A comprehensive analysis of the trajectories on the board of the game (length of the displacements and turning angles) shows that players follow strategies consisting of two modes. The detection of cues about the location of the target triggers an arearestricted search mainly characterized by shorter movements on average^{36,37}. In the context of existing studies, these processes are usually modeled by composite random walks that consist of an extensive phase and an intensive one. In the particular instance of animal foraging, the latter is triggered by encountering a food item and is characterized by shorter steps and larger turning angles (relative to the extensive mode)^{1,44,45,46}. Our findings show that the duration of the search is minimal when the cues extend over intermediate spatial scales as compared to the system size. The tradeoff between locating a cue and finding the target among the cues is balanced, which results in faster searches. Although this result seems to be robust against changes in the total system size, considering larger landscapes could offer a richer phenomenology in the analysis of the trajectories on the board as well as in the features of both phases.
In the simplest scenario studied here, in which no information is given about the size of the neighborhood of the target, developing a systematic searching rule as opposed to following a stochastic trajectory does not provide a significant advantage. A systematic scan of the environment usually provides higher efficiencies by minimizing the probability of revisiting a certain region. In this setup, however, cells remain open once they are visited, providing players with a perfect memory about the history of their movements. As a consequence, neither random nor systematic players click more than once on a cell, regions are not revisited, and both protocols offer equivalent results. This scenario however changes when some information about the nature of the target is provided to the players. In that case an optimal systematic strategy can be constructed based on this information. Interestingly, our data show that one of these optimal strategies was developed by a particular player who repeatedly played several rounds in the same landscape. This result opens the door to explore a broad range of questions at the interface between landscape variability, the searcher’s memory, and learning abilities, which has recently become an important topic in movement ecology^{32}. Most animals do not follow completely random strategies, but combine this stochastic component with spatial memory and learning^{31,47,48}. To investigate the importance of cognitive skills such as learning or memory in the development of optimal strategies, our approach could easily be extended to allow landscapes where the position of the target exhibits a certain degree of persistence across rounds of the game. In addition, in order to compare how more complex decisionmaking processes come into play, it would be particularly interesting to compare the results presented here with the outcome of a new round of experiments in which players are not requested to find the target in the shortest possible time.
In fact, we have shown that, when they have some knowledge about the landscape (size of the neighborhood of the target), players use the additional information obtained in each movement step to increase search efficiency. In this scenario, the effect of the information gathered during the whole process has to be included in theoretical models to reproduce experimental results. Introducing a more realistic finite memory by allowing clicked cells to revert back to the unclicked state after some time arises as a future line of research.
More importantly, however, the excellent agreement between our experimental data and simple theoretical models suggest that this onlinegame based methodology could be applicable to address more complex scenarios. Energy budget related questions can be addressed by introducing a metabolic cost that penalizes longer movements and evolutionary aspects of search problems may be addressed by allowing pairs of players to compete and selecting those using more efficient strategies. This would mimic environments where different individuals compete for limited resources and could shed some light on the driving forces behind the evolution of optimal searching. The effect of cooperative interactions among players on search efficiency could also be addressed. Many species forage in groups as opposed to individually. The methodology that has been presented here would facilitate, given a certain landscape, exploration of the level of confidence that players place on movements performed by previous participants. Before every movement of the new player, the choice of previous searchers at that same moment can be shown to the new player to investigate whether and how much the current player trusts on previous participants. In addition, if the neighborhood of the target is changed, or multiple targets are included, it would be possible to explore the relationships between use of social information versus personal experience for tasks of increasing difficulty. Finally, in this study we have focused on the case of saltatory searches, in which the target can be detected only if the searcher lands on it. A next step should consider the more general scenario of cruising searches, in which the target can be detected at any point of the displacements^{1}. Such setup would provide a higher flexibility in constructing more complex landscapes with different gradients of information that could allow the study of taxisdriven searches.
In summary, and in view of the large and exciting range of possibilities for future exploration, we expect that this general framework will complement purely theoretical efforts to unveil the fundamental mechanisms that drive a wide variety of search scenarios.
Methods
Ethics statement
The anonymity of all the participants was maintained during the whole experimental protocol. Participants accessed the game remotely through internet and non of their personal data was stored. No ethical concerns are involved other than preserving the anonymity of participants. Informed consent was obtained from all subjects. The procedure was checked and approved by the Committee of Ethics in Research of the University of the Balearic Islands, since the game was hosted in the web domain of one of its research institutes, the Institute for CrossDisciplinary Physics and Complex Systems (IFISC) The experiments were subsequently carried out in accordance with the approved guidelines.
Fitting of the partial distributions of displacement lengths to gamma distributions in informed searches
We showed that, for informed searches, the length of the displacements when players are given a priori information about the landscape follow a series of gamma distributions whose probability density function is given by
where α and β are real positive parameters. For known values of α and β, the mean value of the distribution can be obtained as αβ, the variance as αβ^{2} and the mode (the value that appears most often in the distribution) as β(α − 1). All the parameters shown in Table 1 were obtained using the maximum likelihood estimation. Results shown in the last row of Table 1 correspond to a distribution that is a mixture of all four component distributions. Given the mean value and variance of these distributions, we can assume that they all have the same weight in the composition since all the subsets of the trajectory have the same length. The mixed distribution can be obtained as:
Implementation of the random walk model for blind searches
We have developed a minimalistic model based on composite random walks to understand the basic features of the search strategies used by the players. We initialize the model from a random configuration of the board in which the target is placed in a random position inside a smaller square of lateral length N_{y}. To mimic the experimental setup, we fix the size of the board so it has 20 cells on each side and explore N_{y} varying between 1 and 20. The searcher is placed in a random position of the board and the dynamics starts. The algorithm consists of the following steps:
 1
Obtain the probability of jumping from the current position, i to the rest of the cells in the board j. This is given by the experimental jump length distributions, so P_{ij} = exp(−λ_{γ}r_{ij})/λ_{γ}, where γ ≡ {in, ext} and r_{ij} is the distance between two cells. The two values of λ are obtained from the experimental data and define the extensive and the intensive phase: 1/λ_{int} = 2.05 and 1/λ_{ext} = 3.70.
 2
As in the game the player has perfect memory of previous moves, so the probability of jumping to already visited cells is set to zero.
 3
If any of the visited cells belongs to the neighborhood of the target (yellow cell), then we multiply the probability of jumping to each of its unvisited neighbors by a bias factor η = 10^{3} whose effect is to keep the searcher around the cues and avoid unrealistic escapes from them. The existence of such a bias is suggested by the distribution of turn angles shown in Fig. 3d that shows a high probability of returning to the yellow region when it is left. Our results are, however, independent of the numerical value of this bias provided that it is strong enough to trap the searcher close to the yellow cells.
 4
Renormalize all the jumping probabilities so .
 5
Sort a uniform random number u between 0 and 1 and move to a cell k when .
These steps are repeated until the target is found, then the number of movements is saved and the system restarted for a new realization.
Implementation of the random walk model for searches with initial information
To introduce the effect of having initial information about the configuration of the landscape (size of the yellow region) we modify the randomwalk model presented for blind searches. Simulations are set as in the first model, starting from a 20 × 20 cells board where the target is randomly placed inside a square region of lateral length N_{y} = 5. The position of this region is also random in the board and changes across realizations. The searcher is placed at an initial random position and the dynamics starts. The algorithm has two well differentiated parts for the intensive and the extensive phase:
• Extensive phase:
 1
Obtain the distance from the current position of the searcher, i, to every other cell in the board, j, and assign a jumping probability, P_{ij}, by taking a random sample from the histogram in Fig. 5a.
 2
As in the game the player has perfect memory of previous moves, so the probability of jumping to already visited cells is set to zero.
 3
For every cell j in the board obtain all the possible 5 × 5 squares to which it can belong. If all of them have any open black cell, then set the probability of jumping to j to zero. This step is skipped in the intermediate model where only the intensive phase is improved.
 4
Renormalize all the jumping probabilities so they sum one.
 5
Sort a uniform random number u between 0 and 1 and move to a cell k when .
• Intensive phase, after the first yellow cell is hit:
 1
Obtain all the possible neighborhoods of the target to which the first detected yellow cell can belong.
 2
Count the number of open cells of both classes (black and yellow) in each of those possible neighborhoods of the target.
 3
Pick those 5 × 5 squares that include all the open yellow cells and none of the black ones.
 4
Set the probability of jumping to all other of the rest of the cells of the board to zero.
 5
From the histogram in the inset of Fig. 5a, obtain the probability P_{ij} of jumping to the cells that belong to the chosen 5 × 5 squares.
 6
Renormalize all the jumping probabilities so they sum one.
 7
Sort a uniform random number u between 0 and 1 and move to a cell k when .
Additional Information
How to cite this article: MartínezGarcía, R. et al. Online games: a novel approach to explore how partial information influences human random searches. Sci. Rep. 7, 40029; doi: 10.1038/srep40029 (2017).
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
Méndez, V., Campos, D. & Bartumeus, F. Random search strategies. In Stochastic Foundations in Movement Ecology,]' Springer Series in Synergetics, 177–205 (Springer Berlin: Heidelberg, 2014).
Bénichou, O., Loverdo, C., Moreau, M. & Voituriez, R. Intermittent search strategies. Reviews of Modern Physics 83, 81–129 (2011).
Kagan, E. & BenGal, I. Search and Foraging: Individual Motion and Swarm Dynamics (CRC Press, 2015).
Viswanathan, G., da Luz, M. G. E., Raposo, E. P. & Stanley, H. E. The physics of foraging: an introduction to random searches and biological encounters (Cambridge University Press, 2011), 1 edn.
Viswanathan, G. M. et al. Optimizing the success of random searches. Nature 401, 911–4 (1999).
Méndez, V., Campos, D. & Bartumeus, F. Biological searches and random animal motility. In Stochastic Foundations in Movement Ecology, Springer Series in Synergetics, 267–288 (Springer Berlin: Heidelberg, 2014).
Bartumeus, F., da Luz, M. G. E., Viswanathan, G. M. & Catalan, J. Animal search strategies: a quantitative randomwalk analysis. Ecology 86, 3078–3087 (2005).
Edwards, A. M. et al. Revisiting lévy flight search patterns of wandering albatrosses, bumblebees and deer. Nature 449, 1044–1048 (2007).
Gorman, J. & Greene, E. C. Visualizing onedimensional diffusion of proteins along dna. Nature structural & molecular biology 15, 768–774 (2008).
Kantsler, V., Dunkel, J., Blayney, M. & Goldstein, R. E. Rheotaxis facilitates upstream navigation of mammalian sperm cells. eLife 3 (2014).
Bonnet, I. et al. Sliding and jumping of single ecorv restriction enzymes on noncognate dna. Nucleic acids research 36, 4118–4127 (2008).
Hänggi, P., Talkner, P. & Borkovec, M. Reactionrate theory: fifty years after kramers. Rev. Mod. Phys. 62, 251–341 (1990).
Frost, J. & Stone, L. D. Review of search theory: advances and applications to search and rescue decision support. Tech. Rep., DTIC Document (2001).
Najemnik, J. & Geisler, W. S. Optimal eye movement strategies in visual search. Nature 434, 387–391 (2005).
Credidio, H. F., Teixeira, E. N., Reis, S. D., Moreira, A. A. & Andrade Jr, J. S. Statistical patterns of visual search for hidden objects. Scientific reports 2 (2012).
Amor, T. A., Reis, S. D., Campos, D., Herrmann, H. J. & Andrade Jr, J. S. Persistence in eye movement during visual search. Scientific reports 6 (2016).
Levin, S. A. The problem of pattern and scale in ecology: the robert h. macarthur award lecture. Ecology 73, 1943–1967 (1992).
Bartumeus, F., Catalan, J., Fulco, U., Lyra, M. & Viswanathan, G. Optimizing the encounter rate in biological interactions: Lévy versus brownian strategies. Physical Review Letters 88, 097901 (2002).
Hein, A. M. & McKinley, S. A. Sensing and decisionmaking in random search. Proceedings of the National Academy of Sciences of the United States of America 109, 12070–12074 (2012).
Bénichou, O., Coppey, M., Moreau, M., Suet, P.H. & Voituriez, R. Optimal search strategies for hidden targets. Phys. Rev. Lett. 94, 198101 (2005).
Chupeau, M., Bénichou, O. & Voituriez, R. Cover times of random searches. Nature Physics 11, 844–847 (2015).
Vergassola, M., Villermaux, E. & Shraiman, B. I. ‘Infotaxis’ as a strategy for searching without gradients. Nature 445, 406–9 (2007).
Campos, D., Abad, E., Méndez, V., Yuste, S. & Lindenberg, K. Optimal search strategies of spacetime coupled random walkers with finite lifetimes. Physical Review E 91, 052115 (2015).
Abe, M. S. & Shimada, M. Lévy walks suboptimal under predation risk. PLoS Comput Biol 11, e1004601 (2015).
Roberts, W. M. et al. A stochastic neuronal model predicts random search behaviors at multiple spatial scales in C. elegans. eLife;10.7554/eLife.12572 (2016).
Patlak, C. S. Random walk with persistence and external bias. The Bulletin of mathematical biophysics 15, 311–338 (1953).
Codling, E. A., Plank, M. J. & Benhamou, S. Random walk models in biology. Journal of the Royal Society Interface 5, 813–834 (2008).
O’brien, W. J., Browman, H. I. & Evans, B. I. Search strategies of foraging animals. American Scientist 78, 152–160 (1990).
Jonsen, I. D., Flemming, J. M. & Myers, R. A. Robust statespace modeling of animal movement data. Ecology 86, 2874–2880 (2005).
McClintock, B. T. et al. A general discretetime modeling framework for animal movement using multistate random walks. Ecological Monographs 82, 335–349 (2012).
Merkle, J., Fortin, D. & Morales, J. A memorybased foraging tactic reveals an adaptive mechanism for restricted space use. Ecology letters 17, 924–931 (2014).
Fagan, W. F. et al. Spatial memory and animal movement. Ecology letters 16, 1316–1329 (2013).
MartínezGarcía, R., Calabrese, J. M. & López, C. Optimal search in interacting populations: Gaussian jumps versus Lévy flights. Physical Review E 89, 032718 (2014).
Bhattacharya, K. & Vicsek, T. Collective foraging in heterogeneous landscapes. Journal of The Royal Society Interface 11, 20140674 (2014).
MartínezGarcía, R., Calabrese, J. M., Mueller, T., Olson, K. A. & López, C. Optimizing the Search for Resources by Sharing Information: Mongolian Gazelles as a Case Study. Physical Review Letters 110, 248106 (2013).
Hassell, M. P. The dynamics of arthropod predatorprey systems (Princeton University Press, 1978).
Curio, E. The ethology of predation, vol. 7 (Springer Science & Business Media, 2012).
Kareiva, P. & Odell, G. Swarms of predators exhibit “preytaxis” if individual predators use arearestricted search. The American Naturalist 233–270 (1987).
Kareiva, P. Patchiness, dispersal, and species interactions: consequences for communities of herbivorous insects. Community ecology 192–206 (1986).
Weimerskirch, H., Pinaud, D., Pawlowski, F. & Bost, C.A. Does prey capture induce arearestricted search ? a finescale study using gps in a marine predator, the wandering albatross. The American Naturalist 170, 734–743 (2007).
Fauchald, P. & Tveraa, T. Using firstpassage time in the analysis of arearestricted search and habitat selection. Ecology 84, 282–288 (2003).
Hills, T. T., Kalff, C. & Wiener, J. M. Adaptive lévy processes and arearestricted search in human foraging. PLoS One 8, e60488 (2013).
Ovaskainen, O. Habitatspecific movement parameters estimated using markrecapture data and a diffusion model. Ecology 85, 242–257 (2004).
AugerMéthé, M., Derocher, A. E., Plank, M. J., Codling, E. A. & Lewis, M. A. Differentiating the lévy walk from a composite correlated random walk. Methods in Ecology and Evolution 6, 1179–1189 (2015).
Benhamou, S. Efficiency of areaconcentrated searching behaviour in a continuous patchy environment. Journal of Theoretical Biology 159, 67–81 (1992).
Morales, J., Haydon, D., Frair, J., Holsinger, K. & Fryxell, J. Extracting more out of relocation data: building movement models as mixtures of random walks. Ecology 85, 2436–2445 (2004).
Boyer, D. & SolisSalas, C. Random walks with preferential relocations to places visited in the past and their application to biology. Physical Review Letters 112, 240601 (2014).
Polansky, L., Kilian, W. & Wittemyer, G. Elucidating the significance of spatial memory on movement decisions by african savannah elephants using statespace models. Proceedings of the Royal Society of London B: Biological Sciences 282, 20143042 (2015).
Acknowledgements
We acknowledge Antònia Tugores, Rubén Tolosa and Iharob al Asimi Espina for advice in the development of the experimental setup. We are also grateful to George W. Constable for useful discussions and to Frederic Bartumeus for useful discussions and a critical reading of the manuscript. This work is funded by the Gordon and Betty Moore Foundation through Grant GBMF2550.06 to RMG, Universitat de les Illes Balears through a 2015 Young Visiting Scholar grant to RMG, the US National Science Foundation through grant ABI 1458748 to JMC and Ministerio de Economía y Competitividad and Fondo Europeo de Desarrollo Regional through project CTM201566407P (MINECO/FEDER) to CL.
Author information
Authors and Affiliations
Contributions
R.MG conceived the study, implemented the experimental setup, and did the numerical simulations. All the authors designed the experiments, analyzed and discussed the results and contributed to the writing and revision of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing financial interests.
Supplementary information
Rights and permissions
This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/
About this article
Cite this article
MartínezGarcía, R., Calabrese, J. & López, C. Online games: a novel approach to explore how partial information influences human random searches. Sci Rep 7, 40029 (2017). https://doi.org/10.1038/srep40029
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/srep40029
Further reading

Efficient Lévy walks in virtual human foraging
Scientific Reports (2021)

Spatial ecoevolutionary feedbacks mediate coexistence in preypredator systems
Scientific Reports (2019)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.