Our mission is to help computational modelers at all levels engage in the establishment and adoption of community standards and good practices for developing and sharing computational models. Model authors can freely publish their model source code in the Computational Model Library alongside narrative documentation, open science metadata, and other emerging open science norms that facilitate software citation, reproducibility, interoperability, and reuse. Model authors can also request peer review of their computational models to receive a DOI.
All users of models published in the library must cite model authors when they use and benefit from their code.
Please check out our model publishing tutorial and contact us if you have any questions or concerns about publishing your model(s) in the Computational Model Library.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with additional detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 10 of 22 results for "Richard Taylor" clear search
While the world’s total urban population continues to grow, not all cities are witnessing such growth, some are actually shrinking. This shrinkage causes several problems to emerge including population loss, economic depression, vacant properties and the contraction of housing markets. Such problems challenge efforts to make cities sustainable. While there is a growing body of work on study shrinking cities, few explore such a phenomenon from the bottom up using dynamic computational models. To overcome this issue this paper presents an spatially explicit agent-based model stylized on the Detroit Tri-county area, an area witnessing shrinkage. Specifically, the model demonstrates how through the buying and selling of houses can lead to urban shrinkage from the bottom up. The model results indicate that along with the lower level housing transactions being captured, the aggregated level market conditions relating to urban shrinkage are also captured (i.e., the contraction of housing markets). As such, the paper demonstrates the potential of simulation to explore urban shrinkage and potentially offers a means to test polices to achieve urban sustainability.
The Non-Deterministic model of affordable housing Negotiations (NoD-Neg) is designed for generating hypotheses about the possible outcomes of negotiating affordable housing obligations in new developments in England. By outcomes we mean, the probabilities of failing the negotiation and/or the different possibilities of agreement.
The model focuses on two negotiations which are key in the provision of affordable housing. The first is between a developer (DEV) who is submitting a planning application for approval and the relevant Local Planning Authority (LPA) who is responsible for reviewing the application and enforcing the affordable housing obligations. The second negotiation is between the developer and a Registered Social Landlord (RSL) who buys the affordable units from the developer and rents them out. They can negotiate the price of selling the affordable units to the RSL.
The model runs the two negotiations on the same development project several times to enable agents representing stakeholders to apply different negotiation tactics (different agendas and concession-making tactics), hence, explore the different possibilities of outcomes.
The model produces three types of outputs: (i) histograms showing the distribution of the negotiation outcomes in all the simulation runs and the probability of each outcome; (ii) a data file with the exact values shown in the histograms; and (iii) a conversation log detailing the exchange of messages between agents in each simulation run.
MELBIS-V1 is a spatially explicit agent-based model that allows the geospatial simulation of the decision-making process of newcomers arriving in the bilingual cities and boroughs of the island of Montreal, Quebec in CANADA, and the resulting urban segregation spatial patterns. The model was implemented in NetLogo, using geospatial raster datasets of 120m spatial resolution.
MELBIS-V2 enhances MELBIS-V1 to implement and simulate the decision-making processes of incoming immigrants, and to analyze the resulting spatial patterns of segregation as immigrants arrive and settle in various cities in Canada. The arrival and segregation of immigrants is modeled with MELBIS-V2 and compared for three major Canadian immigration gateways, including the City of Toronto, Metro Vancouver, and the City of Calgary.
The Garbage Can Model of Organizational Choice (GCM) is a fundamental model of organizational decision-making originally propossed by J.D. Cohen, J.G. March and J.P. Olsen in 1972. In their model, decisions are made out of random meetings of decision-makers, opportunities, solutions and problems within an organization.
With this model, these very same agents are supposed to meet in society at large where they make decisions according to GCM rules. Furthermore, under certain additional conditions decision-makers, opportunities, solutions and problems form stable organizations. In this artificial ecology organizations are born, grow and eventually vanish with time.
We develop an IBM that predicts how interactions between elephants, poachers, and law enforcement affect poaching levels within a virtual protected area. The model is theoretical at this stage and is not meant to provide a realistic depiction of poaching, but instead to demonstrate how IBMs can expand upon the existing modelling work done in this field, and to provide a framework for future research. The model could be further developed into a useful management support tool to predict the outcomes of various poaching mitigation strategies at real-world locations. The model was implemented in NetLogo version 6.1.0.
We first compared a scenario in which poachers have prescribed, non-adaptive decision-making and move randomly across the landscape, to one in which poachers adaptively respond to their memories of elephant locations and where other poachers have been caught by law enforcement. We then compare a situation in which ranger effort is distributed unevenly across the protected area to one in which rangers patrol by adaptively following elephant matriarchal herds.
This model extends the original Artifical Anasazi (AA) model to include individual agents, who vary in age and sex, and are aggregated into households. This allows more realistic simulations of population dynamics within the Long House Valley of Arizona from AD 800 to 1350 than are possible in the original model. The parts of this model that are directly derived from the AA model are based on Janssen’s 1999 Netlogo implementation of the model; the code for all extensions and adaptations in the model described here (the Artificial Long House Valley (ALHV) model) have been written by the authors. The AA model included only ideal and homogeneous “individuals” who do not participate in the population processes (e.g., birth and death)–these processes were assumed to act on entire households only. The ALHV model incorporates actual individual agents and all demographic processes affect these individuals. Individuals are aggregated into households that participate in annual agricultural and demographic cycles. Thus, the ALHV model is a combination of individual processes (birth and death) and household-level processes (e.g., finding suitable agriculture plots).
As is the case for the AA model, the ALHV model makes use of detailed archaeological and paleoenvironmental data from the Long House Valley and the adjacent areas in Arizona. It also uses the same methods as the original model (from Janssen’s Netlogo implementation) to estimate annual maize productivity of various agricultural zones within the valley. These estimates are used to determine suitable locations for households and farms during each year of the simulation.
This model is an extension of the Artificial Long House Valley (ALHV) model developed by the authors (Swedlund et al. 2016; Warren and Sattenspiel 2020). The ALHV model simulates the population dynamics of individuals within the Long House Valley of Arizona from AD 800 to 1350. Individuals are aggregated into households that participate in annual agricultural and demographic cycles. The present version of the model incorporates features of the ALHV model including realistic age-specific fertility and mortality and, in addition, it adds the Black Mesa environment and population, as well as additional methods to allow migration between the two regions.
As is the case for previous versions of the ALHV model as well as the Artificial Anasazi (AA) model from which the ALHV model was derived (Axtell et al. 2002; Janssen 2009), this version makes use of detailed archaeological and paleoenvironmental data from the Long House Valley and the adjacent areas in Arizona. It also uses the same methods as the original AA model to estimate annual maize productivity of various agricultural zones within the Long House Valley. A new environment and associated methods have been developed for Black Mesa. Productivity estimates from both regions are used to determine suitable locations for households and farms during each year of the simulation.
The model aims to illustrate how Earned Value Management (EVM) provides an approach to measure a project’s performance by comparing its actual progress against the planned one, allowing it to evaluate trends to formulate forecasts. The instance performs a project execution and calculates the EVM performance indexes according to a Performance Measurement Baseline (PMB), which integrates the description of the work to do (scope), the deadlines for its execution (schedule), and the calculation of its costs and the resources required for its implementation (cost).
Specifically, we are addressing the following questions: How does the risk of execution delay or advance impact cost and schedule performance? How do the players’ number or individual work capacity impact cost and schedule estimations to finish? Regardless of why workers cause delays or produce overruns in their assignments, does EVM assess delivery performance and help make objective decisions?
To consider our model realistic enough for its purpose, we use the following patterns: The model addresses classic problems of Project Management (PM). It plays the typical task board where workers are assigned to complete a task backlog in project performance. Workers could delay or advance in the task execution, and we calculate the performance using the PMI-recommended Earned Value.
Patagonia PSMED is an agent-based model designed to study a simple case of Evolution of Ethnic Differentiation. It replicates how can hunter-gatherer societies evolve and built cultural identities as a consequence of the way they interacted.
This model is based on the Narragansett Bay, RI recreational fishery. The two types of agents are piscivorous fish and fishers (shore and boat fishers are separate “breeds”). Each time step represents one week. Open season is weeks 1-26, assuming fishing occurs during half the year. At each weekly time step, fish agents grow, reproduce, and die. Fisher agents decide whether or not to fish based on their current satisfaction level, and those that do go fishing attempt to catch a fish. If they are successful, they decide whether to keep or release the fish. In our publication, this model was linked to an Ecopath with Ecosim food web model where the commercial harvest of forage fish affected the biomass of piscivorous fish - which then became the starting number of piscivorous fish for this ABM. The number of fish caught in a season of this ABM was converted to a fishing pressure and input back into the food web model.
Displaying 10 of 22 results for "Richard Taylor" clear search