As multi-core processors with tens or hundreds of cores begin to grow, system optimization issues once faced only by the High-Performance Computing (HPC). To satisfy the requirement, one can leverage multi-core architectures to parallelize traffic monitoring so as to progress information processing capabilities over traditional uni-processor architectures. In this paper an effective scheduling framework for multi-core processors that strike a balance between control over the system and an effective network traffic control mechanism for high-performance computing is proposed. In the proposed Cache Fair Thread Scheduling (CFTS), information supplied by the user to guide threads scheduling and also, where necessary, gives the programmer fine control over thread placement. Cloud computing has recently received considerable attention, as a promising approach for delivering network traffic services by improving the utilization of data centre resources. The primary goal of scheduling framework is to improve application throughput and overall system utilization in cloud applications. The resultant aim of the framework is to improve fairness so that each thread continues to make good forward progress. The experimental results show that the parallel CFTS could not only increase the processing rate, but also keep a well performance on stability which is important for cloud computing. This makes, it an effective network traffic control mechanism for cloud computing.