Abstract
Due to the reduction in computational complexity, FFTs are usually performed to enable a faster convolution implementation in seismic computations. However, on current parallel computation platforms, such as multi‐core processors, Graphic Processing Units (GPUs), and Field Programmable Gate Arrays (FPGAs), the performance is not only determined by the computational complexity but also relates to other factors, such as the parallelism and the memory access pattern of the algorithm. In our work, we investigate optimized designs of convolutions and FFTs on different parallel computation platforms with different problem and stencil sizes. Experiment results show that, for many stencil sizes used in practical seismic applications, the direct convolution approach demonstrates a better performance than the FFT‐based approach on parallel platforms. For 1D cases, the parallel performance of the FFT‐based approach is limited by the data dependency of the FFT algorithm. 3D FFT‐based approaches use cache poorly. Only 2D FFTs scales well with the parallel computation capacity of modern architectures. The technological trends indicate that these findings will continue.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.