Skip to Main Content
We investigate the sample weighting effect on genetic parallel programming (GPP). GPP evolves parallel programs to solve the training samples in a training set. Usually, the samples are captured directly from a real-world system. The distribution of samples in a training set can be extremely biased. Standard GPP assigns equal weights to all samples. It slows down evolution because crowded regions of samples dominate the fitness evaluation causing premature convergence. We present 4 sample weighting (SW) methods, i.e. equal SW, class-equal SW, static SW (SSW) and dynamic SW (DSW). We evaluate the 4 methods on 7 training sets (3 Boolean functions and 4 UCI medical data classification databases). Experimental results show that DSW is superior in performance on all tested problems. In the 5-input symmetry Boolean function experiment, SSW and DSW boost the evolutionary performance by 465 and 745 times respectively. Due to the simplicity and effectiveness of SSW and DSW, they can also be applied to different population-based evolutionary algorithms.