Generation of Distributed System Test-Beds from High-Level Software Architecture Descriptions
Proceedings of the 16th IEEE international conference on Automated software engineering
Model-Based Performance Prediction in Software Development: A Survey
IEEE Transactions on Software Engineering
Performance Model Interchange Format (PMIF 2.0): XML Definition and Implementation
QEST '04 Proceedings of the The Quantitative Evaluation of Systems, First International Conference
Stability of Feature Selection Algorithms
ICDM '05 Proceedings of the Fifth IEEE International Conference on Data Mining
Performance modeling for service oriented architectures
Companion of the 30th international conference on Software engineering
Performance evaluation of component-based software systems: A survey
Performance Evaluation
Model-driven web engineering performance prediction with layered queue networks
ICWE'10 Proceedings of the 10th international conference on Current trends in web engineering
Ginpex: deriving performance-relevant infrastructure properties through goal-oriented experiments
Proceedings of the joint ACM SIGSOFT conference -- QoSA and ACM SIGSOFT symposium -- ISARCS on Quality of software architectures -- QoSA and architecting critical systems -- ISARCS
Proceedings of the 2013 ACM SIGPLAN international conference on Object oriented programming systems languages & applications
Hi-index | 0.00 |
Designing complex multi-tier applications that must meet strict performance requirements is a challenging software engineering problem. Ideally, the application architect could derive accurate performance predictions early in the project life-cycle, leveraging initial application design-level models and a description of the target software and hardware platforms. To this end, we have developed a capacity planning tool suite for component-based applications, called Revel8tor. The tool adheres to the model driven development paradigm and supports benchmarking and performance prediction for J2EE, .Net and Web services platforms. The suite is composed of three different tools: MDAPerf, MDABench and DSLBench. MDAPerf allows annotation of design diagrams and derives performance analysis models. MDABench allows a customized benchmark application to be modeled in the UML 2.0 Testing Profile and automatically generates a deployable application, with measurement automatically conducted. DSLBench allows the same benchmark modeling and generation to be conducted using a simple performance engineering Domain Specific Language (DSL) in Microsoft Visual Studio. DSLBench integrates with Visual Studio and reuses its load testing infrastructure. Together, the tool suite can assist capacity planning across platforms in an automated fashion.