As mentioned in the introduction, we will start with spatial models without price. We will add price dynamics later.
We use a 2-dimensional grid with periodic boundary conditions. Sites are numbered . Each site belongs to a cluster, denoted by . Initially, each site belongs to ``itself'', that is, , and thus cluster numbers also go from to .
The dynamics is such that in each time step we randomly pick a cluster, delete it, and the corresponding sites are taken over by neighboring clusters. Since the details, in particular with respect to the time scaling, make a difference, we give a more technical version of the model. In each time step, we first select a cluster for deletion by randomly picking a number between and . All sites belonging to the cluster (i.e. ) are marked as ``dead''. We then let adjoining clusters grow into the ``dead'' area. Because of the interpretation later in the paper, in our model the ``dead'' sites play the active role. In parallel, they all pick randomly one of their four nearest neighbors. If that neighbor is not dead (i.e. belongs to a cluster), then the previously dead site will join that cluster. This step is repeated over and over, until no dead sites are left. Only then, time is advanced and the next cluster is selected for deletion.
In physics this is called a domain coarsening scheme (e.g. [11]): Clusters are selected and deleted, and their area is taken over by the neighbors. This happens with a total separation of time scales, that is, we do not pick another cluster for deletion before the distribution of the last deleted cluster has finished. Fig. 1 shows an example. We will call a cluster of size larger than zero ``active''.
[width=0.4]basic-1-gz.eps
[height=0.4]basic-2-gz.eps
|
Note that it is possible to pick a cluster that has already been deleted. In that case, nothing happens except that the clock advances by one. This implies that there are two reasonable definitions of time:
At any particular time step, there is a typical cluster size. In
fact, in cluster time, since there are
clusters, the average cluster size as a
function of cluster time is
.
However, if one averages over all time steps, we find a scaling
law. In cluster time, it is numerically close to
where is the cluster size, is the number of clusters of
size , and is the number of clusters with size larger than
. In natural time, the large clusters have more weight since time
moves more slowly near the end of the coarsening process. The result
is again a scaling law (Fig. 2 (left)), but with
exponents increased by one:
|
In view of evolution, for example in economics or in biology, it is realistic to inject new small clusters. A possibility is to inject them at random positions. So in each time step, before the cluster deletion described above, in addition with probability we pick one random site and inject a cluster of size one at . That is, we set to . This is followed by the usual cluster deletion. It will be explained in more detail below what this means in terms of system-wide injection and deletion rates.
This algorithm maintains the total separation of time scales between the cluster deletion (slow time scale) and cluster growth (fast time scale). That is, no other cluster will be deleted as long as there are still ``dead'' sites in the system. Note that the definition of time in this section corresponds to natural time.
The probability that the injected cluster is really new is reduced by
the probability to select a cluster that is already active. The
probability of selecting an already active cluster is , where
is again the number of active clusters. In consequence, the
effective injection rate is
The cluster size distribution for the model of this section is numerically close to a log-normal distribution, see Fig. 2 (right). Indeed, the position of the distribution moves with (not shown). In contrast to Sec. 3.1, this is now a steady state result.
It is maybe intuitively clear that the injection mechanism of the
model described in Sec. 3.2 destroys the scaling law from
the basic model without injection (Sec. 3.1), since
injection at random positions introduces a typical spatial scale. One
injection process that actually generates steady-state scaling is
injection along a 1-d line. Instead of the random injection of
Sec. 3.2, we now permanently set
In this case, we numerically find a stationary cluster size
distribution (Fig. 3 (right)) with
[width=0.4]snap-line-gz.eps
[width=0.55]line-scaling-gpl.eps
|
One could ask what would happen without space. A possible translation of our model into ``no space'' is: Do in parallel: Instead of picking one of your four nearest neighbors, you pick an arbitrary other agent (random neighbor approximation). If that agent is not dead, copy its cluster number. Do this over and over again in parallel, until all agents are part of a cluster again. A cluster is now no longer a spatially connected structure, but just a set of agents. In that case, we obtain again power laws for the size distribution, but this time with slopes that depend on the injection rate (Fig. 4); see Sec. 4.4 for details.
[width=0.49]nsptl-p10-gpl.eps
[width=0.49]nsptl-p01-gpl.eps
|
Fig. 5 shows actual retail company size distributions from the 1992 U.S. economic census [12], using annual sales as a proxy for company size. We use the retail sector because we think that it is closest to our modelling assumptions -- this is discussed at the end of Sec. 6. We show two curves: establishment size, and firm size. It is clear that in order to be comparable with our model assumptions, we need to look at establishment size rather than at company size.
Census data comes in unequally spaced bins; the procedure to convert it into useable data is described in the appendix. Also, the last four data points for firm size (not for the establishment size, however) were obtained via a different method than the other data points; for details, again see the appendix.
From both plots, one can see that there is a typical establishment size around $400000 annual sales; and the typical firm size is a similar number. This number intuitively makes sense: With, say, income of 10% of sales, smaller establishments will not provide a reasonable income.
One can also see from the plots that the region around that typical
size can be fitted by a log-normal. We also see, however, that for
larger numbers of annual sales, such a fit is impossible since the
tail is much fatter. A scaling law with
This is, however, at odds with investigations in the literature. For example, Ref. [13] find a log-normal, and by using a Zipf plot they show that for large companies the tail is less fat than a log-normal. However, there is a huge difference between our and their data: They only use publicely traded companies, while our data refers to all companies in the census. Indeed, one finds that their plot has its maximum at annual sales of , which is already in the tail of our distribution. This implies that the small scale part of their distribution comes from the fact that small companies are typically not publicely traded. In consequence, it reflects the dynamics of companies entering and exiting from the stock market, not entry and exit of the company itself.
We conclude that from available data, company size distributions are between a log-normal and a power law with or . Further investigation of this goes beyond the scope of this paper.
[width=0.49]sales-loglog-gpl.eps
[width=0.49]sales-accum-gpl.eps
|