Research Article
Coordinator-Master-Worker Model For Efficient Large Scale Network Simulation
@INPROCEEDINGS{10.4108/icst.simutools.2013.251739, author={Bilel Ben romdhanne and Navid Nikaein and Christian Bonnet}, title={Coordinator-Master-Worker Model For Efficient Large Scale Network Simulation}, proceedings={Sixth International Conference on Simulation Tools and Techniques}, publisher={ICST}, proceedings_a={SIMUTOOLS}, year={2013}, month={7}, keywords={pdes simualtion architecture hybrid archtecture gpu gpgpu}, doi={10.4108/icst.simutools.2013.251739} }
- Bilel Ben romdhanne
Navid Nikaein
Christian Bonnet
Year: 2013
Coordinator-Master-Worker Model For Efficient Large Scale Network Simulation
SIMUTOOLS
ACM
DOI: 10.4108/icst.simutools.2013.251739
Abstract
In this work, we propose a coordinator-master-worker (CMW) model for medium to extra-large scale network simulation. The model supports distributed and parallel simulation for a heterogeneous computing node architecture with both multi-core CPUs and GPUs. The model aims at maximizing the hardware usage rate while reducing the overall management overhead. In the CMW model, the coordinator is the top-level simulation CPU process that performs an initial partitioning of the simulation into multiple instances and is responsible for load balancing and synchronization services among all the active masters. The master is also a CPU process and provides event scheduling, synchronization, and communication services to the workers. It manages workers operating potentially on different computing resources within the same shared memory context and communicates with the coordinator and others masters through the messages passing interface. The worker is the elementary actor of CMW model that performs the simulation routines and interacts with the input and output data, and can be a CPU or a GPU thread.
Compared to existing master-worker models, the CMW is natively parallel and GPU compliant, and can be extended to support additional computing resources. The performance gain of the model is evaluated through different benchmarking scenarios using low-cost publicly available GPU platforms. The results have been shown that the speedup up to 3000 times can be achieved.