Our mission is to help computational modelers at all levels engage in the establishment and adoption of community standards and good practices for developing and sharing computational models. Model authors can freely publish their model source code in the Computational Model Library alongside narrative documentation, open science metadata, and other emerging open science norms that facilitate software citation, reproducibility, interoperability, and reuse. Model authors can also request peer review of their computational models to receive a DOI.
All users of models published in the library must cite model authors when they use and benefit from their code.
Please check out our model publishing tutorial and contact us if you have any questions or concerns about publishing your model(s) in the Computational Model Library.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with additional detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 10 of 318 results for "Tim Dorscheidt" clear search
An ABM, derived from a case study and a series of surveys with greenhouse growers in the Westland, Netherlands. Experiments using this model showshow that the greenhouse horticulture industry displays diversity, adaptive complexity and an uneven distribution, which all suggest that the industry is an evolving system.
Using chains of replicas of Atwood’s Machine, this model explores implications of the Maximum Power Principle. It is one of a series of models exploring the dynamics of sustainable economics – PSoup, ModEco, EiLab, OamLab, MppLab, TpLab, EiLab.
In CmLab we explore the implications of the phenomenon of Conservation of Money in a modern economy. This is one of a series of models exploring the dynamics of sustainable economics – PSoup, ModEco, EiLab, OamLab, MppLab, TpLab, CmLab.
The largely dominant meritocratic paradigm of highly competitive Western cultures is rooted on the belief that success is due mainly, if not exclusively, to personal qualities such as talent, intelligence, skills, smartness, efforts, willfulness, hard work or risk taking. Sometimes, we are willing to admit that a certain degree of luck could also play a role in achieving significant material success. But, as a matter of fact, it is rather common to underestimate the importance of external forces in individual successful stories. It is very well known that intelligence (or, more in general, talent and personal qualities) exhibits a Gaussian distribution among the population, whereas the distribution of wealth - often considered a proxy of success - follows typically a power law (Pareto law), with a large majority of poor people and a very small number of billionaires. Such a discrepancy between a Normal distribution of inputs, with a typical scale (the average talent or intelligence), and the scale invariant distribution of outputs, suggests that some hidden ingredient is at work behind the scenes. In a recent paper, with the help of this very simple agent-based model realized with NetLogo, we suggest that such an ingredient is just randomness. In particular, we show that, if it is true that some degree of talent is necessary to be successful in life, almost never the most talented people reach the highest peaks of success, being overtaken by mediocre but sensibly luckier individuals. As to our knowledge, this counterintuitive result - although implicitly suggested between the lines in a vast literature - is quantified here for the first time. It sheds new light on the effectiveness of assessing merit on the basis of the reached level of success and underlines the risks of distributing excessive honors or resources to people who, at the end of the day, could have been simply luckier than others. With the help of this model, several policy hypotheses are also addressed and compared to show the most efficient strategies for public funding of research in order to improve meritocracy, diversity and innovation.
EMMIT is an end-user developed agent-based simulation of malaria transmission. The simulation’s development is a case study demonstrating an approach for non-technical investigators to easily develop useful simulations of complex public health problems. We focused on malaria transmission, a major global public health problem, and insecticide resistance (IR), a major problem affecting malaria control. Insecticides are used to reduce transmission of malaria caused by the Plasmodium parasite that is spread by the Anopheles mosquito. However, the emergence and spread of IR in a mosquito population can diminish the insecticide’s effectiveness. IR results from mutations that produce behavioral changes or biochemical changes (such as detoxification enhancement, target site alterations) in the mosquito population that provide resistance to the insecticide. Evolutionary selection for the IR traits reduces the effectiveness of an insecticide favoring the resistant mosquito population. It has been suggested that biopesticides, and specifically those that are Late Life Acting (LLA), could address this problem. LLA insecticides exploit Plasmodium’s approximate 10-day extrinsic incubation period in the mosquito vector, a delay that limits malaria transmission to older infected mosquitoes. Since the proposed LLA insecticide delays mosquito death until after the exposed mosquito has a chance to produce several broods of offspring, reducing the selective pressure for resistance, it delays IR development and gives the insecticide longer effectivity. Such insecticides are designed to slow the evolution of IR thus maintaining their effectiveness for malaria control. For the IR problem, EMMIT shows that an LLA insecticide could work as intended, but its operational characteristics are critical, primarily the mean-time-to-death after exposure and the associated standard deviation. We also demonstrate the simulation’s extensibility to other malaria control measures, including larval source control and policies to mitigate the spread of IR. The simulation was developed using NetLogo as a case study of a simple but useful approach to public health research.
Digital social networks facilitate the opinion dynamics and idea flow and also provide reliable data to understand these dynamics. Public opinion and cooperation behavior are the key factors to determine the capacity of a successful and effective public policy. In particular, during the crises, such as the Corona virus pandemic, it is necessary to understand the people’s opinion toward a policy and the performance of the governance institutions. The problem of the mathematical explanation of the human behaviors is to simplify and bypass some of the essential process. To tackle this problem, we adopted a data-driven strategy to extract opinion and behavioral patterns from social media content to reflect the dynamics of society’s average beliefs toward different topics. We extracted important subtopics from social media contents and analyze the sentiments of users at each subtopic. Subsequently, we structured a Bayesian belief network to demonstrate the macro patters of the beliefs, opinions, information and emotions which trigger the response toward a prospective policy. We aim to understand the factors and latent factors which influence the opinion formation in the society. Our goal is to enhance the reality of the simulations. To capture the dynamics of opinions at an artificial society we apply agent-based opinion dynamics modeling. We intended to investigate practical implementation scenarios of this framework for policy analysis during Corona Virus Pandemic Crisis. The implemented modular modeling approach could be used as a flexible data-driven policy making tools to investigate public opinion in social media. The core idea is to put the opinion dynamics in the wider contexts of the collective decision-making, data-driven policy-modeling and digital democracy. We intended to use data-driven agent-based modeling as a comprehensive analysis tools to understand the collective opinion dynamics and decision making process on the social networks and uses this knowledge to utilize network-enabled policy modeling and collective intelligence platforms.
Criminal organizations operate in complex changing environments. Being flexible and dynamic allows criminal networks not only to exploit new illicit opportunities but also to react to law enforcement attempts at disruption, enhancing the persistence of these networks over time. Most studies investigating network disruption have examined organizational structures before and after the arrests of some actors but have disregarded groups’ adaptation strategies.
MADTOR simulates drug trafficking and dealing activities by organized criminal groups and their reactions to law enforcement attempts at disruption. The simulation relied on information retrieved from a detailed court order against a large-scale Italian drug trafficking organization (DTO) and from the literature.
The results showed that the higher the proportion of members arrested, the greater the challenges for DTOs, with higher rates of disrupted organizations and long-term consequences for surviving DTOs. Second, targeting members performing specific tasks had different impacts on DTO resilience: targeting traffickers resulted in the highest rates of DTO disruption, while targeting actors in charge of more redundant tasks (e.g., retailers) had smaller but significant impacts. Third, the model examined the resistance and resilience of DTOs adopting different strategies in the security/efficiency trade-off. Efficient DTOs were more resilient, outperforming secure DTOs in terms of reactions to a single, equal attempt at disruption. Conversely, secure DTOs were more resistant, displaying higher survival rates than efficient DTOs when considering the differentiated frequency and effectiveness of law enforcement interventions on DTOs having different focuses in the security/efficiency trade-off.
Overall, the model demonstrated that law enforcement interventions are often critical events for DTOs, with high rates of both first intention (i.e., DTOs directly disrupted by the intervention) and second intention (i.e., DTOs terminating their activities due to the unsustainability of the intervention’s short-term consequences) culminating in dismantlement. However, surviving DTOs always displayed a high level of resilience, with effective strategies in place to react to threatening events and to continue drug trafficking and dealing.
Identifying how organisms respond to environmental stressors remains of central importance as human impacts continue to shift the environmental conditions for countless species. Some mammals are able to mitigate these environmental stressors at the cellular level, but the mechanisms by which cells are able to do this and how these strategies vary among species is not well understood. At the cellular level, it is difficult to identify the temporal dynamics of the system through empirical data because fine-grained time course samples are both incomplete and limited by available resources. To help identify the mechanisms by which animal cells mitigate extreme environmental conditions, we propose an agent-based model to capture the dynamics of the system. In the model, agents are regulatory elements and genes, and are able to impact the behaviors of each other. Rather than imposing rules for these interactions among agents, we will begin with randomized sets of rules and calibrate the model based on empirical data of cellular responses to stress. We will apply a common-garden framework to cultured cells from 16 mammalian species, which will yield genomic data and measures of cell morphology and physiology when exposed to different levels of temperature, glucose, and oxygen. These species include humans, dolphins, bats, and camels, among others, which vary in how they respond to environmental stressors, offering a comparative approach for identifying mechanistic rules whereby cells achieve robustness to environmental stressors. For calibration of the model, we will iteratively select for rules that best lead to the emergent outcomes observed in the cellular assays. Our model is generalized for any species, any cell type, and any environmental stressor, offering many applications of the model beyond our study. This study will increase our understanding of how organisms mitigate environmental stressors at the cellular level such that we can better address how organisms are impacted by and respond to extreme environmental conditions.
This generic individual-based model of a bird colony shows how the influence neighbour’s stress levels synchronize the laying date of neighbours and also of large colonies. The model has been used to demonstrate how this form of simulation model can be recognised as being ‘event-driven’, retaining a history in the patterns produced via simulated events and interactions.
This is a model of organizational behavior in the hierarchy in which personnel decisions are made.
The idea of the model is that the hierarchy, busy with operations, is described by such characteristics as structure (number and interrelation of positions) and material, filling these positions (persons with their individual performance). A particular hierarchy is under certain external pressure (performance level requirement) and is characterized by the internal state of the material (the distribution of the perceptions of others over the ensemble of persons).
The World of the model is a four-level hierarchical structure, consisting of shuff positions of the top manager (zero level of the hierarchy), first-level managers who are subordinate to the top manager, second-level managers (subordinate to the first-level managers) and positions of employees (the third level of the hierarchy). ) subordinated to the second-level managers. Such a hierarchy is a tree, i.e. each position, with the exception of the position of top manager, has a single boss.
Agents in the model are persons occupying the specified positions, the number of persons is set by the slider (HumansQty). Personas have some operational performance (harisma, an unfortunate attribute name left over from the first edition of the model)) and a sense of other personas’ own perceptions. Performance values are distributed over the ensemble of persons according to the normal law with some mean value and variance.
The value of perception by agents of each other is positive or negative (implemented in the model as numerical values equal to +1 and -1). The distribution of perceptions over an ensemble of persons is implemented as a random variable specified by the probability of negative perception, the value of which is set by the control elements of the model interface. The numerical value of the probability equal to 0 corresponds to the case in which all persons positively perceive each other (the numerical value of the random variable is equal to 1, which corresponds to the positive perception of the other person by the individual).
The hierarchy is occupied with operational activity, the degree of intensity of which is set by the external parameter Difficulty. The level of productivity of each manager OAIndex is equal to the level of productivity of the department he leads and is the ratio of the sum of productivity of employees subordinate to the head to the level of complexity of the work Difficulty. An increase in the numerical value of Difficulty leads to a decrease in the OAIndex for all subdivisions of the hierarchy. The managerial meaning of the OAIndex indicator is the percentage of completion of the load specified for the hierarchy as a whole, i.e. the ratio of the actual performance of the structural subdivisions of the hierarchy to the required performance, the level of which is specified by the value of the Difficulty parameter.
…
Displaying 10 of 318 results for "Tim Dorscheidt" clear search