Research Article Open Access

Feature Subset Selection for Hot Method Prediction using Genetic Algorithm wrapped with Support Vector Machines

Sandra Johnson1 and Valli Shanmugam1
  • 1 , Afganistan
Journal of Computer Science
Volume 7 No. 5, 2011, 707-714

DOI: https://doi.org/10.3844/jcssp.2011.707.714

Submitted On: 27 December 2010 Published On: 7 May 2011

How to Cite: Johnson, S. & Shanmugam, V. (2011). Feature Subset Selection for Hot Method Prediction using Genetic Algorithm wrapped with Support Vector Machines. Journal of Computer Science, 7(5), 707-714. https://doi.org/10.3844/jcssp.2011.707.714

Abstract

Problem statement: All compilers have simple profiling-based heuristics to identify and predict program hot methods and also to make optimization decisions. The major challenge in the profile-based optimization is addressing the problem of overhead. The aim of this work is to perform feature subset selection using Genetic Algorithms (GA) to improve and refine the machine learnt static hot method predictive technique and to compare the performance of the new models against the simple heuristics. Approach: The relevant features for training the predictive models are extracted from an initial set of randomly selected ninety static program features, with the help of the GA wrapped with the predictive model using the Support Vector Machine (SVM), a Machine Learning (ML) algorithm. Results: The GA-generated feature subsets containing thirty and twenty nine features respectively for the two predictive models when tested on MiBench predict Long Running Hot Methods (LRHM) and frequently called hot methods (FCHM) with the respective accuracies of 71% and 80% achieving an increase of 19% and 22%. Further, inlining of the predicted LRHM and FCHM improve the program performance by 3% and 5% as against 4% and 6% with Low Level Virtual Machines (LLVM) default heuristics. When intra-procedural optimizations (IPO) are performed on the predicted hot methods, this system offers a performance improvement of 5% and 4% as against 0% and 3% by LLVM default heuristics on LRHM and FCHM respectively. However, we observe an improvement of 36% in certain individual programs. Conclusion: Overall, the results indicate that the GA wrapped with SVM derived feature reduction improves the hot method prediction accuracy and that the technique of hot method prediction based optimization is potentially useful in selective optimization.

  • 1,039 Views
  • 1,804 Downloads
  • 1 Citations

Download

Keywords

  • Feature selection
  • genetic algorithm
  • hot method prediction
  • support vector machine
  • virtual machine
  • Frequently Called Hot Methods (FCHM)
  • Machine Learning (ML)
  • Intra-Procedural Optimizations (IPO)
  • Total Prediction Accuracy (TPA)
  • Hot Method Prediction Accuracy (HMPA)
  • Low Level Virtual Machine (LLVM)