Journal of South China University of Technology (Natural Science Edition) ›› 2015, Vol. 43 ›› Issue (11): 35-46,53.doi: 10.3969/j.issn.1000-565X.2015.11.006

• Computer Science & Technology • Previous Articles     Next Articles

A Transition-Based Word Segmentation Model on Microblog with Text Normalization

Huang Min1 Ding Ping1,2 Luo Hai-biao2   

  1. 1. School of Software Engineering,South China University of Technology,Guangzhou 510006, Guangdong,China;2.Research Center of Parallel Software Research Center,Institute of Software Application Technology,Guangzhou & CAS,Guangzhou 511458,Guangdong,China
  • Received:2015-03-10 Revised:2015-06-07 Online:2015-11-25 Published:2015-10-01
  • Contact: 黄敏( 1976-) ,女,博士,副教授,主要从事并行计算和移动云计算研究 E-mail:minh@scut.edu.cn
  • About author:黄敏( 1976-) ,女,博士,副教授,主要从事并行计算和移动云计算研究
  • Supported by:
    广东省公益研究与能力建设专项(2014A040401018);广东省促进科技服务业发展计划项目(2013B040404009);
    广东省新媒体与品牌传播创新应用重点实验室资助项目(2013WSYS0002)

Abstract: In order to harness the strong horsepower of multi-core processors and meet the demand of high parallelism,a new parallel conjugate gradient algorithm is proposed,which focuses on solving the linear equations of large-scale sparse matrices. For the GPU coprocessors,the memory hierarchy of GPU is effectively utilized,optimization methods,such as thread and matrix mappings,data merging and data multiplexing,are adopted,and an effective thread scheduling is conducted to hide the high latency of accessing the global memory of GPU. For Xeon Phi processors,the computation of high parallelism is effectively utilized to optimize data communication and transmission,data dependence reduction,vectorization and asynchronous computation,and effective thread scheduling is also conducted to hide the high latency of accessing global memory of GPU. Finally,the proposed algorithm is proved to be feasible and correct by tests on GPU and Xeon Phi,and its parallel efficiencies in two different ways are compared. It is found that the proposed algorithm on GPU has a better acceleration effect than itself on Xeon Phi.

Key words: conjugate gradient method, graphics processing unit, Xeon Phi, parallel optimization, sparse matrix-vector multiplication