Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5366789
Xinyu Xing, Shivakant Mishra
Nowadays, tariffs and packages on offer for broadband worldwide gives a general story of increased speed and reduced prices. As a complementary to fixed-line broadband access, WiFi networks are undoubtedly taking off in many American households. Due to the combination of wireless technology and fixed-line broadband access, prior available bandwidth measurement tools may not be an appropriate solution for a common broadband subscriber. In this paper, we utilize abget in combination with Multi-Router Traffic Grapher (MRTG) to target the tight link of the Internet. Based on the observation that the tight link of the Internet in the context of an in-home wireless broadband access is usually on the edge of the Internet, we then introduce ABODE, a single-end, light-weight tool for estimating available bandwidth in the context of an in-home wireless bandwidth environment. ABODE harnesses ICMP echo request messages to generate stable, rate-controlled traffic flows. Based on the timestamp of ICMP echo reply messages, ABODE performs available bandwidth estimation by calculating the drift of the time centroid over a traffic flow. To verify the accuracy of ABODE, we use MRTG data to compare with the estimation results of ABODE. The measurement shows that ABODE is capable of efficiently estimating available bandwidth in the context of an in-home wireless broadband environment.
{"title":"Where is the tight link in a home wireless broadband environment?","authors":"Xinyu Xing, Shivakant Mishra","doi":"10.1109/MASCOT.2009.5366789","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5366789","url":null,"abstract":"Nowadays, tariffs and packages on offer for broadband worldwide gives a general story of increased speed and reduced prices. As a complementary to fixed-line broadband access, WiFi networks are undoubtedly taking off in many American households. Due to the combination of wireless technology and fixed-line broadband access, prior available bandwidth measurement tools may not be an appropriate solution for a common broadband subscriber. In this paper, we utilize abget in combination with Multi-Router Traffic Grapher (MRTG) to target the tight link of the Internet. Based on the observation that the tight link of the Internet in the context of an in-home wireless broadband access is usually on the edge of the Internet, we then introduce ABODE, a single-end, light-weight tool for estimating available bandwidth in the context of an in-home wireless bandwidth environment. ABODE harnesses ICMP echo request messages to generate stable, rate-controlled traffic flows. Based on the timestamp of ICMP echo reply messages, ABODE performs available bandwidth estimation by calculating the drift of the time centroid over a traffic flow. To verify the accuracy of ABODE, we use MRTG data to compare with the estimation results of ABODE. The measurement shows that ABODE is capable of efficiently estimating available bandwidth in the context of an in-home wireless broadband environment.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116920071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5363146
B. Haverkort
Since the 1970's, the scientific field of model-based performance and dependability evaluation has been flourishing. Starting with breakthroughs in the area of closed queueing networks in the 1970's, the 1980's brought new results on state-based methods, such as those for stochastic Petri nets and matrix-geometric methods, whereas the 1990's introduced process algebra-type models. Since the turn of the century, techniques for stochastic model checking are being introduced, to name just a few major developments. The applicability of all these techniques has been boosted enormously through Moore's law; these days, stochastic models with tens of millions of states can easily be dealt with on a standard desktop or laptop computer. A dozen or so dedicated conferences serve the scientific field, as well as a number of scientific journals. However, for the field as a whole to make progress, it is important to step back, and to consider how all these as-such important developments have really changed the way computer and communication systems are being designed and operated. The answer to this question is most probable rather disappointing. I do observe a rather strong discrepancy between what is being published in top conferences and journals, and what is being used in real practice. Blaming industry for this would be too easy a way out. Currently, we do not see model-based performance and dependability evaluation as key step in the design process for new computer and communication systems. Moreover, in the exceptional cases that we do see performance and dependability evaluation being part of a design practice, the employed techniques are not the ones referred to above, but instead, depending on the application area, techniques like discrete-event simulation on the basis of hand-crafted simulation programs (communication protocols), or techniques based on (non-stochastic) timed-automata or timeless behavioral models (embedded systems). In all these cases, however, the scalability of the employed methods, also for discrete-event simulation, forms a limiting factor. Still, industry is serving the world with ever better, faster and more impressive computing machinery and software! What went wrong? When and why did ”our field” land on a side track? In this presentation I will argue that it is probably time for a change, for a change toward a new way of looking at performance and dependability models and evaluation of computer and communication systems, a way that is, if you like, closer to the way physicists deal with very large scale systems, by applying different type of abstractions. In particular, I will argue that computer scientist should “stop counting things”. Instead, a more fluid way of thinking about system behavior is deemed to be necessary to be able to evaluate the performance and dependability of the next generation of very large scale omnipresent systems. First successes of such new approaches have recently been reported. Will be witness a
{"title":"Time — It's time for a change","authors":"B. Haverkort","doi":"10.1109/MASCOT.2009.5363146","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5363146","url":null,"abstract":"Since the 1970's, the scientific field of model-based performance and dependability evaluation has been flourishing. Starting with breakthroughs in the area of closed queueing networks in the 1970's, the 1980's brought new results on state-based methods, such as those for stochastic Petri nets and matrix-geometric methods, whereas the 1990's introduced process algebra-type models. Since the turn of the century, techniques for stochastic model checking are being introduced, to name just a few major developments. The applicability of all these techniques has been boosted enormously through Moore's law; these days, stochastic models with tens of millions of states can easily be dealt with on a standard desktop or laptop computer. A dozen or so dedicated conferences serve the scientific field, as well as a number of scientific journals. However, for the field as a whole to make progress, it is important to step back, and to consider how all these as-such important developments have really changed the way computer and communication systems are being designed and operated. The answer to this question is most probable rather disappointing. I do observe a rather strong discrepancy between what is being published in top conferences and journals, and what is being used in real practice. Blaming industry for this would be too easy a way out. Currently, we do not see model-based performance and dependability evaluation as key step in the design process for new computer and communication systems. Moreover, in the exceptional cases that we do see performance and dependability evaluation being part of a design practice, the employed techniques are not the ones referred to above, but instead, depending on the application area, techniques like discrete-event simulation on the basis of hand-crafted simulation programs (communication protocols), or techniques based on (non-stochastic) timed-automata or timeless behavioral models (embedded systems). In all these cases, however, the scalability of the employed methods, also for discrete-event simulation, forms a limiting factor. Still, industry is serving the world with ever better, faster and more impressive computing machinery and software! What went wrong? When and why did ”our field” land on a side track? In this presentation I will argue that it is probably time for a change, for a change toward a new way of looking at performance and dependability models and evaluation of computer and communication systems, a way that is, if you like, closer to the way physicists deal with very large scale systems, by applying different type of abstractions. In particular, I will argue that computer scientist should “stop counting things”. Instead, a more fluid way of thinking about system behavior is deemed to be necessary to be able to evaluate the performance and dependability of the next generation of very large scale omnipresent systems. First successes of such new approaches have recently been reported. Will be witness a ","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128877329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5366135
Christos A. Papageorgiou, Konstantinos Birkos, T. Dagiuklas, S. Kotsopoulos
In this work we present an obstacle-aware human mobility model for ad hoc networks. Typical examples where the nodes of mobile ad hoc networks are human-operated are natural or man-made disasters, military activities or healthcare services. In these scenarios, obstacles are an integral part of the areas where such networks are deployed in order to facilitate communication among the firemen, policemen, medics, soldiers, etc. In the proposed mobility model, the nodes of the network move around the obstacles in a natural and realistic way. A recursive procedure is followed by each node according to which every time an obstacle is encountered between the node's current position and the final destination point, the node moves to the obstacle's vertex that is closest to the destination. This process is repeated until the destination is reached. The obstacles are also taken into account in modeling the signal propagation. When a packet is transmitted through an obstacle, the power at which it is received is attenuated by a certain value representing the physical layer phenomena suffered by the signal. The model is implemented as an add-on module in Network Simulator ns-2. A thorough simulation study conducted highlights the differences of the proposed model with other mobility models, by investigating the properties of the resulting network topologies and their impact on network performance.
{"title":"An obstacle-aware human mobility model for ad hoc networks","authors":"Christos A. Papageorgiou, Konstantinos Birkos, T. Dagiuklas, S. Kotsopoulos","doi":"10.1109/MASCOT.2009.5366135","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5366135","url":null,"abstract":"In this work we present an obstacle-aware human mobility model for ad hoc networks. Typical examples where the nodes of mobile ad hoc networks are human-operated are natural or man-made disasters, military activities or healthcare services. In these scenarios, obstacles are an integral part of the areas where such networks are deployed in order to facilitate communication among the firemen, policemen, medics, soldiers, etc. In the proposed mobility model, the nodes of the network move around the obstacles in a natural and realistic way. A recursive procedure is followed by each node according to which every time an obstacle is encountered between the node's current position and the final destination point, the node moves to the obstacle's vertex that is closest to the destination. This process is repeated until the destination is reached. The obstacles are also taken into account in modeling the signal propagation. When a packet is transmitted through an obstacle, the power at which it is received is attenuated by a certain value representing the physical layer phenomena suffered by the signal. The model is implemented as an add-on module in Network Simulator ns-2. A thorough simulation study conducted highlights the differences of the proposed model with other mobility models, by investigating the properties of the resulting network topologies and their impact on network performance.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129984842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5366714
M. Asif, S. Majumdar
Web services are getting popular in the domain of business to business electronic commerce and in automating information exchange between business processes because of the interoperability they provide in a distributed heterogeneous environment. Most existing systems use web services that are hosted in fixed infrastructures. Hosting web services on wireless mobile devices is challenging because of their limited resources. The few efforts made by researchers in the past are effective only for hosting of simple web service (WS) applications. However, the existing techniques are not adequate for hosting of WS applications that involve large and complex business processes or invoked by multiple concurrent WS clients. This research proposes to use an application partitioning approach so that the execution of some parts (partitions) of a WS application can be offloaded to a powerful computing node. A critical analysis of existing partitioning approaches for traditional and mobile applications is presented and a graph theory based algorithm for WS application partitioning is proposed. A performance analysis of the algorithm is made through prototyping and measurements.
{"title":"A graph-based algorithm for partitioning of mobile web services","authors":"M. Asif, S. Majumdar","doi":"10.1109/MASCOT.2009.5366714","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5366714","url":null,"abstract":"Web services are getting popular in the domain of business to business electronic commerce and in automating information exchange between business processes because of the interoperability they provide in a distributed heterogeneous environment. Most existing systems use web services that are hosted in fixed infrastructures. Hosting web services on wireless mobile devices is challenging because of their limited resources. The few efforts made by researchers in the past are effective only for hosting of simple web service (WS) applications. However, the existing techniques are not adequate for hosting of WS applications that involve large and complex business processes or invoked by multiple concurrent WS clients. This research proposes to use an application partitioning approach so that the execution of some parts (partitions) of a WS application can be offloaded to a powerful computing node. A critical analysis of existing partitioning approaches for traditional and mobile applications is presented and a graph theory based algorithm for WS application partitioning is proposed. A performance analysis of the algorithm is made through prototyping and measurements.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134368768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5366170
M. L. Santamaría, S. Galmés, R. Puigjaner
Time-driven sensor networks are devoted to the continuous reporting of ambient data to the base station. In many cases, these data are provided by nodes that have been deployed in a structured manner, either by selecting strategic locations or by adopting some regular sampling pattern. In either case, the resulting inter-node distances may not be small, and thus additional supporting nodes may be necessary. This suggests that the problem could be better addressed from a network planning perspective. In this sense, a particular approach is proposed and, as part of it, the paper focuses on optimizing the lifetime that can be predicted from the network topology, which is assumed to be a static data gathering tree. It is shown that this problem requires the exploration of all possible spanning trees, since the energy consumed by a node depends on its workload, which in turn depends on how this node is connected to its neighborhood. Because this is an NP-hard problem, the use of a heuristic approach is required. Then, an algorithm based on simulated annealing is proposed, which converges asymptotically to the global optimum. This algorithm is tested on different scenarios and its computational complexity is proved to be linearly dependent on the number of nodes.
{"title":"Simulated annealing approach to optimizing the lifetime of sparse time-driven sensor networks","authors":"M. L. Santamaría, S. Galmés, R. Puigjaner","doi":"10.1109/MASCOT.2009.5366170","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5366170","url":null,"abstract":"Time-driven sensor networks are devoted to the continuous reporting of ambient data to the base station. In many cases, these data are provided by nodes that have been deployed in a structured manner, either by selecting strategic locations or by adopting some regular sampling pattern. In either case, the resulting inter-node distances may not be small, and thus additional supporting nodes may be necessary. This suggests that the problem could be better addressed from a network planning perspective. In this sense, a particular approach is proposed and, as part of it, the paper focuses on optimizing the lifetime that can be predicted from the network topology, which is assumed to be a static data gathering tree. It is shown that this problem requires the exploration of all possible spanning trees, since the energy consumed by a node depends on its workload, which in turn depends on how this node is connected to its neighborhood. Because this is an NP-hard problem, the use of a heuristic approach is required. Then, an algorithm based on simulated annealing is proposed, which converges asymptotically to the global optimum. This algorithm is tested on different scenarios and its computational complexity is proved to be linearly dependent on the number of nodes.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133138661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5367053
Johannes Semmler, K. Wolter, P. Reinecke
Ensuring Quality of Service (QoS) in wireless networks poses an open problem in many application domains. We propose an automatic on-line QoS monitoring and management infrastructure that can be incorporated into existing network setups. Based on model-based assessment of current and future QoS conditions, our solution will control traffic in the network through a combination of admission control, enforced handover, traffic shaping and transmission parameter adjustments. Correctness of the model is evaluated through experimental evaluation and simulations. We implement a prototype of the proposed system using open-source components.
{"title":"On-line monitoring for model-based QoS management in IEEE 802.11 wireless networks","authors":"Johannes Semmler, K. Wolter, P. Reinecke","doi":"10.1109/MASCOT.2009.5367053","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5367053","url":null,"abstract":"Ensuring Quality of Service (QoS) in wireless networks poses an open problem in many application domains. We propose an automatic on-line QoS monitoring and management infrastructure that can be incorporated into existing network setups. Based on model-based assessment of current and future QoS conditions, our solution will control traffic in the network through a combination of admission control, enforced handover, traffic shaping and transmission parameter adjustments. Correctness of the model is evaluated through experimental evaluation and simulations. We implement a prototype of the proposed system using open-source components.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"136 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133665973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5366701
Dong Zhang, R. Bunt, N. Osgood
A wireless transmitter can use a channel to transmit bits over distance per second (in general, either more bits over shorter distance or fewer bits over longer distance). We call this the Channel Transport Capacity. This paper analyzes how the one-hop channel transport capacity limits multi-hop network capacity in cellular wireless mesh networks. The main results of this paper are analytic expressions for capacity bounds for multi-radio, multi-channel, multi-rate, multi-hop and carrier sense based cells in cellular wireless mesh networks. We validated our analytic results with simulations.
{"title":"Capacity bounds for cellular wireless mesh networks","authors":"Dong Zhang, R. Bunt, N. Osgood","doi":"10.1109/MASCOT.2009.5366701","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5366701","url":null,"abstract":"A wireless transmitter can use a channel to transmit bits over distance per second (in general, either more bits over shorter distance or fewer bits over longer distance). We call this the Channel Transport Capacity. This paper analyzes how the one-hop channel transport capacity limits multi-hop network capacity in cellular wireless mesh networks. The main results of this paper are analytic expressions for capacity bounds for multi-radio, multi-channel, multi-rate, multi-hop and carrier sense based cells in cellular wireless mesh networks. We validated our analytic results with simulations.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"303 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123451770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5366158
Eric Anderson, Christopher Hoover, Xiaozhou Li, Joseph A. Tucek
Distributed systems are notoriously difficult to implement and debug. One important tool for understanding the behavior of distributed systems is tracing. Unfortunately, effective tracing for modern distributed systems faces several challenges. First, many interesting behaviors in distributed systems only occur rarely, or at full production scale. Hence we need tracing mechanisms which impose minimal overhead, in order to allow always-on tracing of production instances. Second, for high-speed systems, messages can be delivered in significantly less time than the error of traditional time synchronization techniques such as network time protocol (NTP), necessitating time adjustment techniques with much higher precision. Third, distributed systems today may generate millions of events per second systemwide, resulting in traces consisting of billions of events. Such large traces can overwhelm existing trace analysis tools. These challenges make effective tracing difficult. We present techniques that address these three challenges. Our contributions include 1) a low-overhead tracing mechanism, which allows tracing of large systems without impacting their behavior or performance (0.14 μs/event), 2) a post hoc technique for producing highly accurate time synchronization across hosts (within 10 /ts, compared to between 100 μs to 2 ms for NTP), and 3) incremental data processing techniques which facilitate analyzing traces containing billions of trace points on desktop systems. We have successfully applied these techniques to two distributed systems, a cooperative caching system and a distributed storage system, and from our experience, we believe our techniques are applicable to other distributed systems.
{"title":"Efficient tracing and performance analysis for large distributed systems","authors":"Eric Anderson, Christopher Hoover, Xiaozhou Li, Joseph A. Tucek","doi":"10.1109/MASCOT.2009.5366158","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5366158","url":null,"abstract":"Distributed systems are notoriously difficult to implement and debug. One important tool for understanding the behavior of distributed systems is tracing. Unfortunately, effective tracing for modern distributed systems faces several challenges. First, many interesting behaviors in distributed systems only occur rarely, or at full production scale. Hence we need tracing mechanisms which impose minimal overhead, in order to allow always-on tracing of production instances. Second, for high-speed systems, messages can be delivered in significantly less time than the error of traditional time synchronization techniques such as network time protocol (NTP), necessitating time adjustment techniques with much higher precision. Third, distributed systems today may generate millions of events per second systemwide, resulting in traces consisting of billions of events. Such large traces can overwhelm existing trace analysis tools. These challenges make effective tracing difficult. We present techniques that address these three challenges. Our contributions include 1) a low-overhead tracing mechanism, which allows tracing of large systems without impacting their behavior or performance (0.14 μs/event), 2) a post hoc technique for producing highly accurate time synchronization across hosts (within 10 /ts, compared to between 100 μs to 2 ms for NTP), and 3) incremental data processing techniques which facilitate analyzing traces containing billions of trace points on desktop systems. We have successfully applied these techniques to two distributed systems, a cooperative caching system and a distributed storage system, and from our experience, we believe our techniques are applicable to other distributed systems.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124872223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5366277
Livathinos S. Nikolaos
One of the most widely used simulation environments for mobile wireless networks is the Network Simulator 2 (NS-2). However NS-2 stores its outcome in a text file, so there is a need for a visualization tool to animate the simulation of the wireless network. The purpose of this tool is to help the researcher examine in detail how the wireless protocol works both on a network and a node basis. It is clear that much of this information is protocol dependent and cannot be depicted properly by a general purpose animation process. Existing animation tools do not provide this level of information nor permit the specific protocol to control the animation at all. EXAMS is an NS-2 visualization tool for mobile simulations which makes possible the portrayal of NS-2's internal information like transmission properties and node's data structures. This is mainly possible due to EXAMS extensible architecture which separates the animation process into a general and a protocol specific part. The latter can be developed independently by the protocol designer and loaded on demand. These and other useful characteristics of the EXAMS tool can be an invaluable help for a researcher in order to investigate and debug a mobile networking protocol.
{"title":"EXtensible animator for mobile simulations: EXAMS","authors":"Livathinos S. Nikolaos","doi":"10.1109/MASCOT.2009.5366277","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5366277","url":null,"abstract":"One of the most widely used simulation environments for mobile wireless networks is the Network Simulator 2 (NS-2). However NS-2 stores its outcome in a text file, so there is a need for a visualization tool to animate the simulation of the wireless network. The purpose of this tool is to help the researcher examine in detail how the wireless protocol works both on a network and a node basis. It is clear that much of this information is protocol dependent and cannot be depicted properly by a general purpose animation process. Existing animation tools do not provide this level of information nor permit the specific protocol to control the animation at all. EXAMS is an NS-2 visualization tool for mobile simulations which makes possible the portrayal of NS-2's internal information like transmission properties and node's data structures. This is mainly possible due to EXAMS extensible architecture which separates the animation process into a general and a protocol specific part. The latter can be developed independently by the protocol designer and loaded on demand. These and other useful characteristics of the EXAMS tool can be an invaluable help for a researcher in order to investigate and debug a mobile networking protocol.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124890502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2009-12-28DOI: 10.1109/MASCOT.2009.5367039
Stefan Podlipnig
Ordinal time series analysis is a simple approach to the investigation of experimental data. The basic idea is to consider the order relations between the values of a time series and not the values themselves. This results in fast and robust algorithms that extract the basic intrinsic structure of the given data series. This paper gives a short overview of this approach and describes the application of simple ordinal time series methods like rank autocorrelation, local rank autocorrelation and permutation entropy to workload traces from large grid computing systems. We show how these methods can be used to extract important correlation information from experimental traces and how these methods outperform traditional methods.
{"title":"Applying ordinal time series methods to grid workload traces","authors":"Stefan Podlipnig","doi":"10.1109/MASCOT.2009.5367039","DOIUrl":"https://doi.org/10.1109/MASCOT.2009.5367039","url":null,"abstract":"Ordinal time series analysis is a simple approach to the investigation of experimental data. The basic idea is to consider the order relations between the values of a time series and not the values themselves. This results in fast and robust algorithms that extract the basic intrinsic structure of the given data series. This paper gives a short overview of this approach and describes the application of simple ordinal time series methods like rank autocorrelation, local rank autocorrelation and permutation entropy to workload traces from large grid computing systems. We show how these methods can be used to extract important correlation information from experimental traces and how these methods outperform traditional methods.","PeriodicalId":275737,"journal":{"name":"2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems","volume":"161 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125891112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}