Skip to content
2000
Volume 11, Issue 2
  • ISSN: 2352-0965
  • E-ISSN: 2352-0973

Abstract

Background: Many applications in voice processing have high inherent parallelism. Field programmable gate array (FPGA) has shown very high performance in spite of its low operational frequency by fully extracting the parallelism. Nevertheless, recent CPU and graphic processing unit (GPU) have also an inherent for high performance. Methods: In fact, it becomes possible to utilize the parallelism using multi-cores, which support improved single instruction multiple data (SIMD) instruction. Recent GPUs support a large number of cores, and have a potential for high performance in many applications. Our goals are at first to compare GPU and FPGA implementation of the linear prediction coding (LPC) algorithm, in order to understand the trade-off between the flexibility but relatively low speed of an FPGA and the high speed and fixed architecture of the GPU. Secondly, we try to apply various levels optimization from overlapping data transfers to fine-tuning operation sequences. Results: The experimental results highlight the relative strengths and limitations of the two systems. Conclusion: Our experiments show that, for several samples corresponding to several speeches coding, GPU manages speedups of up to 3x compared to the FPGA and around 35x compared to a sequential execution.

Loading

Article metrics loading...

/content/journals/raeeng/10.2174/2352096511666180116154942
2018-06-01
2025-06-19
Loading full text...

Full text loading...

/content/journals/raeeng/10.2174/2352096511666180116154942
Loading
This is a required field
Please enter a valid email address
Approval was a Success
Invalid data
An Error Occurred
Approval was partially successful, following selected items could not be processed due to error
Please enter a valid_number test