Abstract
Many cellular patterns exhibit a reaction-diffusion component, suggesting that Turing instability may contribute to pattern formation. However, biological gene-regulatory pathways are more complex than simple Turing activator-inhibitor models and generally do not require fine-tuning of parameters as dictated by the Turing conditions. To address these issues, we employ random matrix theory to analyze the Jacobian matrices of larger networks with robust statistical properties. Our analysis reveals that Turing patterns are more likely to occur by chance than previously thought and that the most robust Turing networks have an optimal size, consisting of only a handful of molecular species, thus significantly increasing their identifiability in biological systems. Broadly speaking, this optimal size emerges from a trade-off between the highest stability in small networks and the greatest instability with diffusion in large networks. Furthermore, we find that with multiple immobile nodes, differential diffusion ceases to be important for Turing patterns. Our findings may inform future synthetic biology approaches and provide insights into bridging the gap to complex developmental pathways.
Similar content being viewed by others
Introduction
Spatial patterns and structures are common in biological systems, ranging from microbial communities and biofilms to developmental biology and ecological systems1,2,3. In a seminal work, Alan Turing proposed a self-organizing, emergent mechanism for pattern formation based on a diffusion-driven instability4, later formalized by Gierer and Meinhardt in terms of a slowly diffusing activator and a fast-diffusing inhibitor molecule5. According to Turing’s definition, the chemical reactions are stable without diffusion, leading to a homogeneous steady state, but become unstable with diffusion, forming a periodic pattern at certain wave numbers. The mathematical analysis is generally based on reaction-diffusion models, implemented using partial differential equations, and linear stability analysis to investigate the effect of perturbations on a steady state6. Nonetheless, employing Turing models to understand biological patterns is still a topic of debate.
The Turing mechanism clearly has biological relevance in cell and developmental biology, likely explaining aspects of digit formation in mice, scale-patterns in zebrafish, fingerprint formation, and cortical folds of the fetal brain7,8,9,10. However, this mechanism has two main drawbacks: extreme simplicity and a need for fine-tuning. The latter requires parameters to be carefully chosen to produce patterns. In contrast to simple activator-inhibitor Turing models, biological gene regulatory networks in embryonic development generally consist of hundreds of molecular species and are notable for their immense complexity, hierarchical structure, and tolerance to noise11. In contrast, the Turing conditions lead to a lack of structural robustness, which is at odds with the noise tolerance and evolutionary adaptability required for such patterning solutions to occur12,13,14. However, there are also bottom-up approaches to further elucidate the issues with Turing models.
As developmental biology is complex, synthetic biology provides an alternative route, following Feynman’s mantra: “What I cannot create, I do not understand”15. By implementing the Turing mechanism from scratch in cells that communicate via chemicals, our understanding of how to generate stable pattern can be systematically improved. Two previous synthetic biology attempts failed due to the lack of differential diffusion, leading to irregular patterns. First, stochastic Turing patterns were engineered in E. coli cells. The implemented circuit contained self-activation and lateral inhibition, with two diffusible quorum-sensing signals16. Second, solitary patterns were engineered in HEK cells using the Nodal-Lefty system17. The issue is that small molecules have roughly the same diffusion constant, although extra nonspecific binding can help slow down diffusion. In a recent work, a more robust three-node network was implemented using synthetic circuits of six genes with small diffusible quorum-sensing molecules and extra control cassettes18. This showed regular patterns in growing bacterial colonies, but the variability is immense, and robustness is still limited. Methods to increase the reproducibility of these patterns, such as solving the inverse problem given an experimental pattern, have also been recently developed19. Hence, while there is progress in the rational design of circuits with specific properties, the lack of control over robustness represents a significant downside in tissue engineering, patterned biomaterial deposition, and bridging developmental programs20,21,22.
The need to develop theoretical frameworks beyond the two-equation Turing model has been recognized previously. An exhaustive exploration of all two- and three-node topologies showed that three nodes are, on average, more robust than two nodes, pointing toward complexity increasing robustness23. Note that “nodes” refers here to genes and links to regulatory interactions. These networks should not be confused with Turing models on graphs, which deal with discrete topologies, not continuous space24,25,26,27. More specifically, approximately 60% of all topologies produced patterns for some parameter combinations, but the parameter space producing Turing patterns was overall minuscule, around 0.1%. This finding is supported by another study that explored networks with up to four nodes28. Motivated by Robert May’s work in theoretical ecology29 (based on earlier work by Eugene Wigner30) a random matrix approach for \(N \le 6\) showed that larger networks add robustness by reducing the “diffusive threshold”, i.e., softening the requirement of differential diffusion31. This raises the question: why not investigate even larger matrices? While the diagonalization of Jacobian matrices is a slow \(O(N^3)\) process, it is still efficient for significantly larger networks with modern computers. The advantages of exploring random Jacobian matrices are apparent; such an approach produces excellent statistics, and biological realism can be introduced through the distributions and topologies.
Here, we go beyond small Turing networks and use large random Jacobian matrices to systematically explore how network size affects robustness of patterning continuous space. We begin by motivating distributions for Jacobian matrix elements from explicit small Hill-function-based models for gene regulation, as relevant to synthetic biology18. We then continue by exploring large random networks with up to \(N=100\) nodes using corresponding random Jacobian matrices with two diffusers and variable sparsity. We identify an optimal network size \(N_\text {opt}\sim 5-8\), arising from a tradeoff between stability without diffusion and instability with diffusion, each with opposite N-dependencies. This increased robustness relieves the constraints on parameters, known as Turing conditions, including differential diffusion for the two diffusing species. Our results are expected to renew the quest for robust Turing patterns in synthetic biology and to identify Turing modules in larger networks of developmental biology.
Network graph representations of pre-defined Turing topologies. (Ai) 2-node Gierer-Meinhardt network. (Bi) 3-node network based on18,23. (C) Extended 4-node network. (D) Large \(N>\!\!>1\) network. Image was produced with Python library NetworkX v3.1 for graph generation. (E) Number of different Turing-network topologies for N nodes, exemplified by all 2 and 3-node networks in (Aii) and (Bii), respectively. Dark blue (red) arrows indicate activation (inhibition), dashed lines denote missing links, and wiggly lines indicate diffusion.
Results
We are interested in the robustness of large Turing networks, that is reaction-diffusion models with many molecular species with a large parameter space to support diffusion-driven instabilities. Fig. 1 shows example networks, ranging from small (\(N=2\) nodes) to large (\(N>\!\!>1\)). We restrict ourselves to two diffusing species similar to Turing’s original paper. In order to understand how to model large networks we begin by investigating examples of smaller networks to gain intuition. For a network of N nodes and hence N species with time and space dependent concentrations \(\varvec{X}(\varvec{r},t)=\{x_1(\varvec{r}, t), x_2(\varvec{r}, t), \dots , x_N(\varvec{r},t)\}\), the general dynamics are given by the following partial differential equations
where \(\varvec{D}=diag(D_1, D_2, 0, \dots , 0)\) is the diagonal diffusion matrix, and Laplacian \(\nabla ^2=\sum _{i=1}^d\partial ^2/\partial r_i^2\) in d dimensions, given by the sum of second spatial derivatives. For simplicity, we assume an infinite domain, allowing us to neglect boundary effects and to use a continuous wave number. The nonlinearities are given by the interactions between the nodes. As shown in Fig. 1Aii and Bii, we consider all possibilities for each connection, depicted as black for activation, red for inhibition and dotted for no interaction. We will refer to activation and inhibition nodes as positive and negative edges respectively. Curved lines denote diffusion, which as mentioned will only be considered for the first two nodes. As our motivation is to learn how to build these with synthetic gene circuits, we model the activating and inhibiting interactions with Hill functions, together with basal expression and degradation terms:
with \(\varvec{\theta } =\{\varvec{b}, \varvec{V}, \varvec{K}, \varvec{n}, \varvec{\mu }\}\) respectively the sets of basal and maximal expression rates, concentration thresholds, Hill coefficients, and degradation rates. Furthermore, \(S_i^{+ (-)}\) denotes the set of positive (negative) edges ending in node i, where the different regulators act (and saturate) independently (non-competitively) of each other. Following23, we sample parameters from a wide range of allowed values in arbitrary units.
To understand whether such parameter combinations produce Turing patterns, we use linear stability analysis following Turing’s approach (see Methods for additional details). Briefly, first the model needs to have a stable homogeneous steady state \(\varvec{X}^*\) without diffusion, defined by
which we solve with the Newton-Raphson method using different initial conditions. Second, we linearize the dynamic equations assuming small perturbations around steady state, using \(x_i(\varvec{r},t)=x_i^*+\delta x(\varvec{r},t)\), leading to
with Jacobian matrix \(\varvec{J}\) of first derivative with matrix elements \(J_{ij}=\partial f_i/\partial x_j\).
Dispersion relations illustrating different Turing instabilities. (A) Turing type I instability with a peak and negative real part for \(k>\!\!>1\). (B) Turing type II instability which stays positive (note this is however not possible when all species diffuse). The y axis represents the maximum real part of the eigenvalue and the x axis shows wave number k. The plots are generated from random matrices with network size \(N = 10\) and variance \(\sigma ^2 = 0.25\).
To remove the second derivatives, we Fourier transform in the space domain, or more simply apply a wave-like perturbation \(\delta x(\varvec{r},t)=\delta \tilde{x}(t)\exp {(i\varvec{k}\varvec{r})}\) with \(\varvec{k}=\{k_1, k_2, \dots , k_d)\) and \(\varvec{k}\varvec{r}\) the scalar product between the \(\varvec{k}\) and \(\varvec{r}\) vectors. Plugged into Eq. (4), this leads to a new Jacobian matrix with modified diagonal matrix elements and extra parameter \(k=|\varvec{k}|\), given by
indicating that for a rotationally invariant system, the dependence is only on the modulus of \(\varvec{k}\), rendering linear stability analysis an effective one-dimensional problem. As a result, diagonalization of the modified Jacobian matrix leads to k-dependent eigenvalues, called dispersion relations, where the one with the largest real part, termed \(Re(\lambda _\text {max}(k))\), determines the stability. For a Turing instability, we require the homogeneous steady state without diffusion to be stable (corresponding to k=0), i.e. with a negative real part given by \(Re(\lambda _\text {max}(k(0)))<0\). Furthermore, with diffusion, we require instability for a wave number (\(k>0\)) and thus a positive real part given by \(Re(\lambda _\text {max}(k))>0\). For a classic Turing instability leading to a well-defined wave pattern, there is a \(k_\text {max}>0\), for which the dispersion has a clear maximum, and for \(k\rightarrow \infty\), the system becomes stable again (negative real part of eigenvalue). This type of instability is called Turing I (Fig. 2A). However, there are other possibilities, such as the instability for \(k\rightarrow \infty\) (Turing II in Fig. 2B). In the latter case, there is no well defined wavelength for pattern formation, and hence Turing II cases are generally not considered in the following. (Note that this is a simplified classification scheme - in other schemes our Turing I is given by Turing Ia, and our Turing II is given by Turing Ib, IIa, and IIb with subtle differences23.) Yet, another type of instability is the Turing-Hopf, for which the imaginary part can additionally lead to oscillations (not shown).
Focusing initially on small networks (\(N = 2-4\)), we employed Latin hypercube sampling of \(10^7\) parameter sets and filtered them for Turing I instabilities. Fig. 3 shows the fits of the empirical histograms of the Jacobian matrix elements (j) to the beta distribution \(B(j; \alpha ,\beta )\) (see Methods for details on parameter searches and fitting procedure). The beta distribution depends on two parameters, \(\alpha\) and \(\beta\), allowing a wide range of distributions from Gaussian-like symmetric to non-Gaussian shapes with various levels of skewness and kurtosis. We generally observe that the Turing-I matrix elements are more narrowly distributed in line with Turing I having to fulfill Turing conditions, leading to a subset of parameters and hence Jacobian matrix elements. This can be clearly seen in Fig. 3A (see Fig. S1 in the Supplementary Figures for additional histograms). To fulfill the Turing conditions, the (1,1) Jacobian matrix elements need to be positive for activation, which is the case for the histogram and fitted line (blue), but not for the orange line, representing the non-Turing case. Note, while Node 1 is an activator, degradation contributes to the diagonal matrix elements, potentially leading to negative values. Furthermore, the (2,2) matrix elements need to be negative for inhibition, which is here already fulfilled for non-Turing due to model constraints (degradation). In contrast, the (1,2) and (2,1) matrix elements can, respectively, be either negative and positive (here the case even for non-Turing due to model constraints), or the other way around (however not for our 2-equation model)6. We generally did not observe multimodal distributions, which indicates a certain level of simplicity in the distributions.
The difference between Turing and non-Turing Jacobian matrix elements is further visible in Fig. 4, where panel A shows beehive plots based on the empirical data from Fig. 3. Clearly, the variances of the off-diagonal matrix elements of the non-Turing cases are much broader than the Turing-I cases. This is further illustrated in panel B, showing the sum of the Jacobian matrix elements (equivalent plots can be made for the mean of those random variables). For larger matrices, we expect to observe the central limit theorem, with the sums calculated from \(N(N-1)\) off-diagonal Jacobian matrix elements. For increasing matrix size, this trend is already evident for \(N=4\), although there is still a significant cusp at zero: The mean is approximately zero, the variance approaches a finite value (as variances are positive numbers), and the skewness and kurtosis vanish. Hence, not surprisingly, on average a Gaussian distribution with zero mean and finite variance describes the sum of Jacobian matrix elements. Note that this trend would be even stronger, if we average over all possible network topologies with N nodes, which scales as \(N_G = 3^N 9^{\left( {\begin{array}{c}N\\ 2\end{array}}\right) }\) and hence is superexponentially increasing with N (see Fig. 1E). In the following, we exploit these insights and use random Jacobian matrices directly.
Empirical distributions of Jacobian matrix elements. (A) Histograms for parameters leading to Turing I instabilities (light blue) and fits to the beta distribution for Turing I (blue line) and non-Turing (orange line). Empirical distributions are computed based on parameter sampling for the pre-defined 2–4-node network topologies in Fig. 1Ai, Bi, and C for 2-node (A), 3-node (B), and 4-node (C) networks. Note some matrix elements are zero due to missing links in the network (or entries in the adjacency matrix).
Jacobian matrix elements are different for Turing I and non-Turing. (A) Beehive plots of the mean (top) and variance (bottom) of the empirical distributions of the non-zero off-diagonal Jacobian matrix elements from Fig. 3 (based on the pre-defined 2-4-node network topologies in Fig. 1Ai, Bi, and C). A statistical Brown-Forsythe test was utilized to verify whether the difference in variance between each pair of Turing and non-Turing distributions was statistically significant. For all pairs of the Jacobian distributions except one, we found that the test showed a statistically significant difference between the variances. Furthermore, 12 out of 16 p-values were under 0.001, showing a strong statistical significance. (B) Sum of off-diagnal matrix elements, showing a clear visual difference between Turing I (light blue) and non-Turing (light orange) for the 2-4 node topologies. The distributions become increasingly symmetric, and ultimately Gaussian, for large numbers of nodes.
Inspired by Robert May‘s statistical treatment of non-equilibrium ecological communities that studied asymmetric random Jacobians to represent linearized population dynamics, contrasting Wigner’s symmetric (Hermitian) random matrices29,32, we analyzed Turing instabilities by sampling asymmetric random Jacobian matrices that represent the linearized reaction dynamics. Similar to May’s “neutral interaction” model, our Jacobian matrix without diffusion is given by
where \(\varvec{G}\) is a random matrix with elements \(g_{ij}\) randomly assigned from a Gaussian distribution that has mean value zero and variance Var\((g_{ij}) = \sigma ^2\), but with diagonal terms \(g_{ii} = 0\). For simplicity (and convenience), the latter are modeled by the identity matrix, \(\varvec{I}\), responsible for providing stability. In such a Turing system, a positive value of \(J_{ij}\) (or \(g_{ii}\)) implies activation while a negative value implies inhibition, similarly to previous matrix representations of Turing systems23,33. The identity matrix \(\varvec{I}\) represents the degradation of the species. Further self-interactions are neglected, leading to the curious result that there are no Turing instabilities for \(N=2\) (see proof in Methods). While the Gaussian distribution is motivated by Fig. 4, there is no reason to expect the reaction kinetics in a reaction-diffusion system to be independent as there was no reason for the population dynamics in the work of May29 to be independent. However, detailed understanding of realistic models and their parameters, or Jacobian matrices of linearized models is missing. Hence, the statistical random-matrix approach has been powerful in understanding the principles of stability in population dynamics34,35 and small Turing systems (\(N\le 6\))31.
Next, we investigate the eigenvalue distributions of such random matrices (see Fig. 5A). In the absence of diffusion (i.e. at \(k = 0\)), the eigenvalues are distributed in a circle with radius \(\gamma = \sigma \sqrt{N}\) according to May’s “circular law”, derived from random ecological communities29,32 and exact in the asymptotic limit. In the presence of diffusion, the eigenvalue distribution can be described as having “in-bulk” and “outlier” distributions (see Fig. 5B). The outlier distribution has extreme negative real parts, following approximately \(-N-k^2\sum _{i=1}^ND_i\) (see Methods), which tends to minus infinity for increasing k and network size N. As we are only interested in the maximal real part to determine the linear stability, it is safe to ignore the “outliers” and only focus on the “in-bulk” eigenvalue distribution, containing the maximal real eigenvalues.
Eigenvalue distribution of a large random network. (A) Circular scatter plot and corresponding kernel density estimations of eigenvalues in the complex plain (green at the top and blue on the right) for network size \(N = 100\), variance \(\sigma ^2 = 0.01\), and diffusion constants \(D_1 = 1\), \(D_2 = 10\). k-dependent movement of the eigenvalues is visible, and finite-N effects in the kernel density estimates as well, such as suppression of the density at the boundary of the circle. While the system without diffusion has all the eigenvalues distribute around \(\gamma =1\), the same system with diffusion has most but not all of its eigenvalues distribute in a circle with radius \(\gamma =1\). (B) Full eigenvalue distribution at a large scale with focus on the outliers (the eigenvalue circle now appears like a vertical line.) Green, red and blue points correspond to wave numbers \(k<1\), \(k = 1\) and \(k>1\), respectively.
We subsequently sampled thousands of random matrices to analyse how network size affects the stability of a reaction-diffusion system. We observed three relationships between network size and stability. Firstly, in the absence of diffusion, a random reaction-diffusion system with fixed radius \(\gamma\) is likely more stable for smaller network size (Fig. 6A). Asymptotically, the circular law predicts a step function for \(N\rightarrow \infty\), i.e. stability for radii \(\gamma <1\) (as circles are fully to the left of the imaginary axis) and instability for \(\gamma >1\) (as circles cross the imaginary axis)30. Secondly, a previously stable system without diffusion is more likely to turn unstable (i.e. having a diffusion-driven instability) when diffusion is present for larger network size (Fig. 6B). Therefore, thirdly, out of all random matrices, Turing instabilities are maximized (i.e. found the most) at an optimal intermediate network size (Fig. 6C).
As the Turing-Hopf instability is competing with the Turing instability, and Turing I is of importance in producing spatial patterns, we analyzed the effects of network size on Turing I and Turing-Hopf instabilities. Out of stable systems without diffusion, the systems are likely to become Turing-Hopf unstable in the presence of diffusion with larger network size (Fig. 6B). This trend for the Turing-Hopf instability is similar to the trend for Turing I. However, particularly at a very large network size, the Turing-Hopf instability becomes more likely to occur and ultimately out-competes Turing I, as can be seen for \(N = 50\) in the figure. A similar trend is observed for Turing I relative to the overall Turing instabilities of all types: Turing I is more likely for larger network size being approximately half the occurrence of all Turing instabilities. Among all Turing instability types, we found that Turing II is the most common to arise (4.2% maximum occurrence of all random reaction-diffusion systems), followed by Turing I (3.6%).
Relationships between network size N and radius \(\gamma =\sigma \sqrt{N}\). \(10^3\) random matrices are sampled for each N and \(\gamma\) pair. (A) Percentage of stable random matrices without diffusion. (B) Percentage of previously stable random matrices turning unstable with diffusion. ’Hopf’ indicates Turing-Hopf. (C) Percentage of random matrices leading to Turing instabilities (Turing I, II, and Turing-Hopf). (D) Percentage of random matrices leading to Turing I instabilities. The key point is \(\gamma =1\), while for larger \(\gamma\) values the statistics become unreliable.
Without having an experimental understanding of the variability of interaction strength between nodes, an optimal network size should produce Turing I instabilities for a wide range of variances. Such a network size would be most robust to changes in interaction strength, which is a desirable outcome from a biological perspective. To find the exact network size that is the most optimal, we plotted the occurrence of Turing instabilities for different network sizes and variances from \(N=2\) to 100 (Fig. 7A). Based on the heat map, the majority of the Turing instabilities arose just above \(\gamma =1\), as we should expect since \(\gamma>\!\!>1\) would likely result in an unstable system without diffusion, but \(\gamma<\!\!<1\) would make it harder for a stable system to become unstable with diffusion. In fact, the occurrence of Turing is “pinned” to radius \(\gamma =1\) for large N, and the percentage density then follows closely the circular law \(\sigma ^2\sim 1/N\).
Is there an optimal network size for highest robustness of the occurance of Turing patterns? We expect such an \(N_\text {opt}\) as there are no Turing patterns for \(N=2\) and for \(N\rightarrow \infty\). Turing patterns disappear as the eigenvalue distribution becomes a step function, allowing no instability with diffusion. To confirm our intuition, we then calculated the percentage of Turing I for each network size from the heat map (Fig. 7B), demonstrating that \(N_\text {opt}=5\) is the most optimal network size based on its highest percentage of Turing patterns. Thus, for a network with two diffusible species, having additional three immobile nodes gives the optimal network size for Turing instabilities. Compared to \(N = 3\) with one non-diffusible species (4.73%), Turing instabilities are three times more likely to occur for \(N+\text {opt} = 5\) (13.86%).
The peak is however not at the first allowed N value, i.e. 3, but at a higher value, pointing towards another principle for this maximum in Turing frequency. For finite-\(N\) random matrices and \(\gamma \gtrsim 1\), the scaling of the eigenvalue density, projected onto the real axis, near \(x = 0\) can be approximated as
(see the Supplementary Text for details). For small \(N\), the density of marginal stable matrices i.e. near \(x=0\), is suppressed. This is due to eigenvalues being correlated and biased toward the center of the eigenvalue circle due to finite N effects. As a result, there are less Turing cases for small N than expected based on the asymptotic form of Robert May’s circular law. This ultimately leads to the value \(N_\text {opt}=5\).
Optimal network size for highest Turing robustness. (A) Heat map of percentage occurrence of Turing I in random matrices of different network sizes N and variances \(\sigma ^2\) for \(\gamma =1\) and zero off-diagonal element sparsity (or connectivity \(C = 1\)). (B) Corresponding percentage shares of each network size N. For our parameters, the optimal network size is \(N = 5\). The inset provides a more detailed view of network sizes from 2 to 8 for finer resolution..
While our random matrices are fully connected, it was suggested that biological networks tend to be sparsely connected due to its evolutionary advantage in preserving robustness36. Let us assume that the probability of observing k connections is binomially distributed \(\left( {\begin{array}{c}n\\ k\end{array}}\right) p^k(1-p)^{n-k}\) where n the number of matrix elements to consider (’trials’) such as \(n=N(N-1)\) for the number of off-diagonal matrix elements, and p the probability for having a connection (’success’). As the variance \(np(1-p)\) is largest for \(p=0.5\), corresponding to the maximum entropy distribution, we would expect on average 50% connections. This corresponds to the maximum number of possible network structures found by evolution. However, does sparsity also translate into higher Turing robustness? We incorporated sparsity but found that its effect is minor, without qualitatively changing our results (see Figs. S2 and S3 in the Supplementary Figures). The peak value of the Turing I percentage slightly reduces with sparsity (\(\approx 14\), 12, and 10% for respective sparsities 0, 25, and 50%), while the corresponding optimal network sizes increase slightly (\(N_\text {opt}=5\), 7, and 8 for above sparsities). Hence, overall the robustness of Turing is reduced with sparsity, pointing to feedback being important for pattern formation. Note the circular law also applies to sparsity, but needs modification. The radius is just rescaled to \(\gamma =\sigma \sqrt{NC}\) where connectivity \(C=p\) is the probability for having an off-diagonal matrix element30. This points to our results being broadly representative for a large range of random Turing networks.
Additionally, we tested full Gaussian matrices with Gaussian-distributed diagonal elements (allowing self-interactions), both with and without self-degradation (\(-1\)), as well as anti-symmetric matrices with diagonal elements fixed at -1, as in the original setup. The anti-symmetry is designed to capture correlations arising from potential mass-conserved reactions. The full Gaussian matrices with fixed degradation produce an optimal network size (see Fig. S4–S5), whereas the anti-symmetric matrices with \(-1\) on the diagonals do not produce Turing patterns at any network size or variance, in line with known theorems37. This is because the eigenvalues of real anti-symmetric matrices are purely imaginary or zero, and adding −1 to the diagonal shifts them along the real axis, leaving no positive real parts to enable Turing instabilities. For the full Gaussian matrices, Turing patterns are possible for \(N=2\) when self-interactions are present, as degradation is no longer fixed at \(-1\) for all nodes. However, when there is no fixed degradation of \(-1\), no clear peak in network size emerges, as the eigenvalue distribution becomes more diffuse without a dominant eigenvalue near \(-1\) to stabilize the system.
A major constraint on Turing patterns is that the inhibitor has to diffuse much faster than the activator. How does having immobile nodes change this constraint? Previous work on small networks hinted towards a softening of this constraint23,28,31. By varying the diffusion of the two diffusible species modeled in our random reaction-diffusion systems, we sampled systems of \(N = 2\), 3, 5, and 50 nodes, with \(N=2\) again not supporting Turing for our random matrices with equal degradation of \(-1\). Importantly, we found that differential diffusivity is important for Turing I instabilities to arise for very few immobile nodes, but not for systems with many immobile nodes: Based on Fig. 8, a network of \(N = 3\) with one immobile node has nearly zero percent Turing instabilities arising at equal diffusivity (i.e. at the identity line \(D_1/D_2=1\)). As the diffusion ratio deviates from 1, Turing instabilities become more likely for \(N = 3\), creating a distinct gap around the identity line. Notably, the plot is symmetric with respect to the identity line for all N since the assignment to activator or inhibitor is arbitrary. In contrast, in the presence of more immobile nodes, Turing instabilities can arise at equal diffusivities: As network size increases, the gap at the identity line becomes more narrow, indicating Turing instabilities occurring at a wider combination of the diffusion parameters (see also Fig. S6 in the Supplementary Figures). For the large network size of \(N=50\), we observed a similar occurrence of Turing instability for the majority of the diffusion parameter combinations, except for when the diffusion parameters are both extremely low, approximating the case of no diffusion. Since the Turing instability is a diffusion-driven instability, it is reasonable to have no Turing instability arising here. Additionally, there is no ’dark line’ (indicating zero or low percentage) at the very bottom or very left of all surface plots, indicating that a Turing instability can arise in systems with only one diffusible species. The absence of differential diffusion is even stronger when consider all Turing cases (Turing I, II, and Turing-Hopf) - see Fig. S7 of the Supplementary Figures. In summary, having numerous immobile nodes significantly relaxes the constraint of differential diffusivity from the classic Turing conditions.
Effect of diffusion constants on Turing pattern formation. Heat map of percentage occurrence of Turing I in random matrices for different diffusion parameters, \(D_1\) and \(D_2\): (A) \(N = 2\), \(\sigma ^2 = 0.5\); (B) \(N = 3\), \(\sigma ^2 = 0.33\); (C) \(N = 5\), \(\sigma ^2 = 0.2\) (D) \(N = 50\), \(\sigma ^2 = 0.02\). For increasing N, the constraint of differential diffusivity begins to vanish.
Discussion
Turing networks have traditionally been small in size, making a direct mapping to complex developmental pathways with many unknown molecular species and parameter values difficult. Another drawback of using Turing models in biology has been the requirement to fine-tune parameters to fulfill Turing’s conditions on stability without diffusion and instability for certain wave numbers with diffusion, in contrast to biology’s robustness and evolvability38. Here, we circumvented these issues by using large random matrices to describe linearized Jacobian systems, greatly extending previous studies on \(N \le 6\)31. By extensively sampling matrix elements, we obtained excellent statistics for networks with up to 100 molecular species.
Applying Robert May’s circular law29 to cases without and with diffusion of two designated species, in line with Turing’s original model4, we identified an optimal network size for maximal robustness, reaching 13.86% of sampled matrices. This value is orders of magnitude higher than previous estimates for small networks23. Our results are robust even when sparsity is introduced, leading only to minor changes in the observed Turing percentage or optimal network size. Importantly, an optimum in Turing robustness emerges in our model due to the tradeoff between maximizing stability without diffusion, which favors small networks with eigenvalues having negative real parts, and maximizing instability with diffusion, which favors large networks with eigenvalues having positive real parts. According to the circular law, small networks have small radii for the distribution of their eigenvalues, while large networks have large radii. For increasingly large Turing networks, we further found that differential diffusivity is not required, in line with studies on Turing topologies28. Having many immobile nodes effectively allows for delays and hence different diffusion constants. We hope these results shed new light on Turing mechanisms in developmental systems and help make Turing mechanisms more relevant to quantitative biology.
While we focus on the sweet spot of Turing network size for maximal robustness, a recent study of very large networks supports our findings, that complex random networks with \(N>\!\!>1\) do not tend to spontaneously form Turing patterns39. When not constraining \(\sigma ^2\), this is because eigenvalue circles cross over to the positive real axis, leading to instability even without diffusion. Even when constraining the radius to \(\gamma =1\), this is still the case, as we found, since the eigenvalue distribution becomes a step function, eliminating the important marginal stable cases. Remedies for overcoming above mentioned issues are identified in39, such as allowing for non-diffusive interactions including cross-diffusion, nonlocal interactions, or advection (chemotaxis)40.
To make our model interpretable, we used a number of simplifying assumptions, resulting in limitations of our predictions. A major simplification is our diagonal matrix elements, which are set to the same value, representing equal degradation rates for all species. This led to the convenient behavior that all eigenvalues are centered around a single point on the negative real axis, following Robert May’s circular law. Without that assumption, the eigenvalue distributions would still be restricted by circles, but now follow the less restrictive Gerschgorin theorem. Based on the latter, the radius scales as \(\sim N\), and not \(\sim \sqrt{N}\) as in May’s circular law. This leads not only to larger circles but also to unions of overlapping circles, restricting the real parts less efficiently41. Using the same degradation rates also does not allow for flexibility in the levels of self-activation and inhibition often observed in biological pathways. Choosing degradation rates from a distribution introduces correlations among corresponding off-diagonal matrix elements \(J_{ij}\) and \(J_{ji}\) and hence different shapes of eigenvalue distributions30.
Due to the optimal Turing network size for relatively small N, we did not need to investigate very large matrices. However, to speed up the investigation of large random Jacobians, the concept of reactivity can be included42. Reactivity, describing the transient growth potential of local perturbations in absence of diffusion, could provide a preliminary filter that avoids the full diagonalization of the system’s Jacobian. This is because the reactivity is determined by the largest eigenvalue of the Hermitian part of the Jacobian matrix, which is computationally cheaper to compute than the eigenvalues of the full Jacobian matrix.
Our work opens new avenues for exciting research in systems biology. In future work, more specific small networks could be investigated similar to the ones shown in Fig. 1, to better understand the distributions of Jacobian matrix elements with pattern-forming capability from sampled parameter values. In addition, examining random networks would allow one to obtain joint distributions of matrix elements and hence correlations among these, leading to renewed investigations into the role of network topology33. Of course, more diffusing species could be included, and the effects of boundary conditions investigated. The latter effect was removed in our study by considering systems of infinite domain size, in line with previous studies23.
Another interesting future direction is to study random Turing networks on graphs, moving from continuous to discrete topologies, with diffusion replaced by hopping along the graph’s edges24,25,26,27. Due to more versatile interactions, robustness may be increased. While classical Turing patterns are associated with regular spatial structures, on graphs, the lack of a physical notion of space makes patterns inherently dependent on the network’s topology. Patterns may appear irregular on random graphs or when nodes are visualized arbitrarily, but they reflect the underlying eigenmodes of the graph’s Laplacian and the topology’s influence on reaction-diffusion dynamics. The challenge lies in interpreting these patterns in the context of the system being modeled, as their biological or functional meaning may not rely on spatial regularity but rather on network-specific features. However, applications are vast, ranging from biochemical reactors and neuronal networks to tissues of living cells, synthetic systems, and patchy predator-prey ecosystems.
In conclusion, we find that Turing patterns do not only occur more frequently by chance than previously thought, but also that there is a surprising sweet spot of network size for most robust Turing patterns. This optimal network size occurred via a stability and instability tradeoff, which did not change significantly when varying the sparsity of the network. This finding supports the idea that the Turing mechanism, if employed by evolution in developmental pathways, is represented by relatively small and hence likely identifiable modules. Understanding their embedding and hierarchical organization in large developmental networks will be worthwhile endeavors.
Methods
Parameter searches including Latin hypercube sampling
Parameter sampling was done with Latin hypercube sampling, which is a statistical method for generating a sample of parameter values from a multidimensional space, ensuring that each parameter is evenly sampled across its range by dividing the space into equal probability intervals18,43. Its strength lies in efficiently covering the input space with fewer samples compared to simple random sampling, reducing variance while maintaining representative coverage. Specifically, for Fig. 3 we sampled \(10^7\) parameter combinations for each network topology using the lhs function from pyDOE library, using a loguniform distribution with the following bounds: \(V_i:\ (0.1, 10)\), \(K_{ij}:\ (0.01, 1)\), \(b_i:\ (0.001, 0.1)\), \(\mu _i:\ (0.01, 1)\) (see Eq. 2). The diffusion constants for Species 1 and 2 were fixed to 1 and 10, and the Hill coefficient was set to \(n_{ij}=2\) throughout. For each parameter set we calculated the unique steady states using 10 different initial guess using the Newton-Raphson method (function root with ’hybrid’ setting). In Figs. 6 and 7 we used \(10^3\) random matrices for each N and \(\gamma\) pair, and in Fig. 8 we used \(10^3\) random matrices for each \(D_1\) and \(D_2\) pair.
Linear stability analysis
For each steady state of a parameter set, and there can be more than one due to multistability, we conducted linear stability analysis by calculating and diagonalizing the Jacobian matrices for wave numbers from 0 to \(k_{max}=10\) using stead size \(\Delta k=0.01\) and with linalg.eigenvals function from numpy library. Subsequently, we analyzed for Turing I, II etc using a simplified version of the classification scheme in23.
Numerical implementation and code availability
For Fig. 3, the high-dimensional parameter space created by nested loops was flattened, and ran in parallel by dividing up the linear parameter array equally on different cores using the multiprocessing library. The code used for producing Figs. 2, 3, 4, 5, 6, 7 and 8 can be accessed on
\(github.com/Endresgroup/Random\_Turing\_networks\).
Fitting of beta distribution
Fitting a beta distribution to the data of the Jacobian matrix elements in Fig. 3 involves estimating the distribution’s parameters, typically using the method of moments or maximum likelihood estimation (MLE). The method of moments involves calculating the sample mean and variance to derive initial estimates for the shape parameters \(\alpha\) and \(\beta\). MLE, on the other hand, maximizes the log-likelihood function, which measures how well the beta distribution with specific parameters explains the observed data, to find the best-fitting parameters. Both methods aim to provide a beta distribution that accurately represents the underlying data distribution. Here, we used the method of moments to obtain an initial guess for the parameter values of \(\alpha\) and \(\beta\), which we then used as an initial guess for the maximization of the log-likelihood function.
Mathematical proofs
Proof that the outliers of the Jacobian matrices have extreme negative real parts, following approximately \(-N-k^2\sum _{i=1}^ND_i\). This is because the trace is given by \(Tr(\varvec{J})=Tr(\varvec{J_0})-k^2Tr(\varvec{D})=-Tr(\varvec{1})-k^2Tr(\varvec{D}) = -N-k^2\sum _{i=1}^ND_i = \sum _{i=1}^N \lambda _i\), which is invariant under similarity transformations during diagonalization and hence equals the sum of all the eigenvalues \(\lambda _i\) (\(i=1,\dots , N\)). Or in other words, the mean of the eigenvalues stays the same under such a transformation. Hence, at least one eigenvalue, if real, or a complex conjugate pair, if with an imaginary part, becomes an outlier and tends to minus infinity for increasing wave number k at a given network size N. This is relevant for Fig. 5A.
Proof that there are no Turing instabilities for random matrices with \(N=2\): Based on Eq. (5), the k-dependent Jacobian matrix is given by
Hence, the trace is always negative for all k, i.e.
fulfilling one of Turing’s conditions. Furthermore, the determinant without diffusion (\(k=0\)) is required to be positive, and hence must be
leading to the required stability in absence of diffusion. However, with diffusion (\(k>0\)), the determinant cannot become negative as needed for a saddle node and hence instability. Instead, we have
Hence, the last of Turing’s conditions cannot be fulfilled for two nodes, which is relevant for Figs. 7A and 8A. Note, Jacobian matrix Eq. (7) is different from Turing’s original linear model, in which the diagonal matrix element of the activator is positive without diffusion4. In our case, all diagonal matrix elements are \(-1\), leading to increased stability.
Data availability
The datasets generated and analyzed in Figs. 3 and 4 are available in the Zenodo repository https://zenodo.org/records/13142658
References
Maini P. Bones, feathers, teeth and coat marking: a unified model. Scientific Progress. (1997). https://ora.ox.ac.uk/objects/uuid:f31e4aad-7188-44de-8143-6c8839385077.
Urban, M.C., Strauss, S.Y., Pelletier, F., Palkovacs, E.P., Leibold, M.A. & Hendry A.P, et al. Evolutionary origins for ecological patterns in space. In: Proceedings of the National Academy of Sciences. 2020 Jul;117(30):17482-90. Publisher: Proceedings of the National Academy of Sciences. https://www.pnas.org/doi/abs/10.1073/pnas.1918960117.
Martinez-Garcia, R., Tarnita, C. E. & Bonachela, J. A. Spatial patterns in ecological systems: from microbial colonies to landscapes. Emerging Topics Life Sci. 6(3), 245–58. https://doi.org/10.1042/ETLS20210282 (2022).
Turing, A. M. The chemical basis of morphogenesis. Philos. Trans. R. Soc. Lond. B Biol. Sci. 237(641), 37–72. https://doi.org/10.1098/rstb.1952.0012 (1952).
Gierer, A. & Meinhardt, H. A theory of biological pattern formation. Kybernetik 12(1), 30–9. https://doi.org/10.1007/BF00289234 (1972).
Murray, J. D. (ed.) Mathematical Biology: I. An introduction of interdisciplinary applied mathematics 17 (Springer, 2002). https://doi.org/10.1007/b98868.
Lefèvre, J. & Mangin, J. F. A reaction-diffusion model of human brain development. PLOS Comput. Biol. 6(4), e1000749 (2010).
Sheth, R. et al. Hox Genes Regulate Digit Patterning by Controlling the Wavelength of a Turing-Type Mechanism. Science 338(6113), 1476–80. https://doi.org/10.1126/science.1226804 (2012).
Watanabe, M. & Kondo, S. Is pigment patterning in fish skin determined by the Turing mechanism? Trends in Genetics 88–96 (Elsevier, 2015).
Glover, J. D. et al. The developmental basis of fingerprint pattern formation and variation. Cell 186(5), 940–56.e20 (2023).
Samarasinghe, S. & Minh-Thai, T. N. A Comprehensive conceptual and computational dynamics framework for autonomous regeneration of form and function in biological organisms. PNAS Nexus. 2(2), pgac308 (2023).
Maini, P. K., Woolley, T. E., Baker, R. E., Gaffney, E. A. & Lee, S. S. Turing’s model for biological pattern formation and the robustness problem. Interface Focus. 2(4), 487–96. https://doi.org/10.1098/rsfs.2011.0113 (2012).
Vittadello, S. T., Leyshon, T. & Schnoerr, D. Stumpf MPH. Turing pattern design principles and their robustness. Philosophical Trans. Royal Soc. A Mathe. Phys. Eng. Sci. 379(2213), 20200272. https://doi.org/10.1098/rsta.2020.0272 (2021).
Krause, A. L., Gaffney, E. A., Jewell, T. J., Klika, V. & Walker, B. J. Turing Instabilities are Not Enough to Ensure Pattern Formation. Bull. Math. Biol. 86(2), 21. https://doi.org/10.1007/s11538-023-01250-4 (2024).
Way, M. What I cannot create, I do not understand. J. Cell Sci. 130(18), 2941–2. https://doi.org/10.1242/jcs.209791 (2017).
Karig, D., Martini, K. M., Lu, T., DeLateur, N. A., Goldenfeld, N. & Weiss, R. Stochastic Turing patterns in a synthetic bacterial population. In: Proceedings of the National Academy of Sciences. (2018) 115(26):6572-7. Publisher: Proceedings of the National Academy of Sciences. https://www.pnas.org/doi/abs/10.1073/pnas.1720770115.
Sekine, R., Shibata, T. & Ebisuya, M. Synthetic mammalian pattern formation driven by differential diffusivity of Nodal and Lefty. Nat. Commun. 9(1), 5456 (2018).
Tica, J. et al. A three-node Turing gene circuit forms periodic spatial patterns in bacteria. Cell Syst. 15, 1–10 (2024).
Matas-Gil, A. & Endres, R. Unraveling biochemical spatial patterns: machine learning approaches to the inverse problem of stationary Turing patterns. iScience 04(27), 109822–2 (2024).
Cao, Y. et al. Programmable assembly of pressure sensors using pattern-forming bacteria. Nat. Biotechnol. 35(11), 1087–93 (2017).
Din, M. O., Martin, A., Razinkov, I., Csicsery, N. & Hasty, J. Interfacing gene circuits with microelectronics through engineered population dynamics. Sci. Adv. 6(21), 8344. https://doi.org/10.1126/sciadv.aaz8344 (2020).
Davies, J. Using synthetic biology to explore principles of development. Development 144(7), 1146–58. https://doi.org/10.1242/dev.144196 (2017).
Scholes, N. S., Schnoerr, D., Isalan, M. & Stumpf, M. P. H. A comprehensive network atlas reveals that turing patterns are common but not robust. Cell Syst. 9(3), 243–57.e4 (2019).
Othmer, H. G. & Scriven, L. E. Instability and dynamic pattern in cellular networks. J. Theoretical Biol. 32(3), 507–37 (1971).
Nakao, H. & Mikhailov, A. S. Turing patterns in network-organized activator-inhibitor systems. Nat. Phys. 6(7), 544–50 (2010).
Kouvaris, N. E., Hata, S. & Guilera, A. D. Pattern formation in multiplex networks. Sci. Rep. 5(1), 10840 (2015).
Muolo, R., Giambagli, L., Nakao, H., Fanelli, D. & Carletti, T. Turing patterns on discrete topologies: from networks to higher-order structures. arXiv; 2024. ArXiv:2407.07663. http://arxiv.org/abs/2407.07663.
Marcon, L., Diego, X., Sharpe, J. & Müller, P. High-throughput mathematical analysis identifies Turing networks for patterning with equally diffusing signals. eLife 5, e14022. https://doi.org/10.7554/eLife.14022 (2016).
May, R. M. Will a Large Complex System be Stable? Nature. Aug;238(5364):413-4. Number: 5364 Publisher: Nature Publishing Group. (1972) https://www.nature.com/articles/238413a0.
Allesina, S. & Tang, S. The stability-complexity relationship at age 40: a random matrix perspective. Population Ecol. 57(1), 63–75. https://doi.org/10.1007/s10144-014-0471-0 (2015).
Haas, P. A. & Goldstein, R. E. Turing’s diffusive threshold in random reaction-diffusion systems. Phys. Rev. Lett. 126(23), 238101. https://doi.org/10.1103/PhysRevLett.126.238101 (2021).
Stone, L. The feasibility and stability of large complex biological networks: a random matrix approach. Sci. Rep. 8(1), 8246 (2018).
Diego, X., Marcon, L., Müller, P. & Sharpe, J. Key features of turing systems are determined purely by network topology. Phys. Rev. X 8(2), 021071. https://doi.org/10.1103/PhysRevX.8.021071 (2018).
Allesina, S. & Tang, S. Stability criteria for complex ecosystems. Nature 483(7388), 205–8 (2012).
Serván, C., Capitán, J., Grilli, J., Morrison, K. & Allesina, S. Coexistence of many species in random ecosystems. Nat. Ecol. Evol. 2(8), 1237–42 (2018).
Pavlopoulos, G. A. et al. Using graph theory to analyze biological networks. BioData Mining. 4(1), 10. https://doi.org/10.1186/1756-0381-4-10 (2011).
Petersen, K. B., Pedersen, M. S. The Matrix Cookbook; (2012) http://www2.imm.dtu.dk/pubdb/views/edoc_download.php/3274/pdf/imm3274.pdf.
Wagner, A. Robustness and evolvability: a paradox resolved. In: Proceedings of the Royal Society B: Biological Sciences. 275(1630):91-100. Publisher: Royal Society. (2007) https://doi.org/10.1098/rspb.2007.1137.
Piskovsky, V., Maini, P. K. Will a large complex system form Turing patterns?. bioRxiv; 2024.10.30.621030 Section: New Results. (2024) https://www.biorxiv.org/content/10.1101/2024.10.30.621030v1.
Zhao, H., Košmrlj, A. & Datta, S. S. Chemotactic motility-induced phase separation. Phys. Rev. Lett. 131(11), 118301. https://doi.org/10.1103/PhysRevLett.131.118301 (2023).
Pazuki, R. & Endres, R. Robustness of Turing models and gene regulatory networks with a sweet spot. Phys. Rev. E 109(6), 064305. https://doi.org/10.1103/PhysRevE.109.064305 (2024).
Neubert, M. G., Caswell, H. & Murray, J. D. Transient dynamics and pattern formation: reactivity is necessary for Turing instabilities. Math. Biosci. 175(1), 1–11 (2002).
Mckay, M. D., Beckman, R. J. & Conover, W. J. A comparison of three methods for selecting values of input variables in the analysis of output from a computer code. Technometrics[SPACE]https://doi.org/10.1080/00401706.2000.10485979 (2000).
Acknowledgements
We thank Martina Oliver Huidobro, Mark Isalan, David Lacoste, and Roozbeh Pazuki for helpful discussions, and funding through a studentship from the Department of Life Sciences at Imperial College London (to AMG) and the AI-4-EB Consortium for Bioengineered Cells and Systems (BBSRC award BB/W013770/1) (to RGE).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Shaberi, H.S.A., Kappassov, A., Matas-Gil, A. et al. Optimal network sizes for most robust Turing patterns. Sci Rep 15, 2948 (2025). https://doi.org/10.1038/s41598-025-86854-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-025-86854-7