![Loading...](https://link.springer.com/static/c4a417b97a76cc2980e3c25e2271af3129e08bbe/images/pdf-preview/spacer.gif)
-
Chapter and Conference Paper
The Optimization and Parallelization of Two-Dimensional Zigzag Scanning on the Matrix
With the expansion of applications, such as image processing, scientific computing, numerical simulation, biomedicine, social network and so on, enormous quantities of data need to be crunched in order to get ...
-
Chapter and Conference Paper
Rgs-SpMM: Accelerate Sparse Matrix-Matrix Multiplication by Row Group Splitting Strategy on the GPU
The Sparse Matrix-Matrix Multiplication (SpMM) operation is widely used in different fields, especially the recently popular GNN framework. Researchers have designed many kernels on the GPU to accelerate the S...
-
Chapter and Conference Paper
The Parallelization and Optimization of K-means Algorithm Based on MGPUSim
Although the k-means algorithm has been parallelized into different platforms, it has not yet been explored on multi-GPU architecture thoroughly. This paper presents a study of parallelizing k-means on a novel...
-
Chapter and Conference Paper
Optimization and Analysis of Parallel Back Propagation Neural Network on GPU Using CUDA
Graphic Processing Unit (GPU) can achieve remarkable performance for dataset-oriented application such as Back Propagation Network (BPN) under reasonable task decomposition and memory optimization. However, ad...
-
Chapter and Conference Paper
Parallelizing Block Cryptography Algorithms on Speculative Multicores
Although block cryptography algorithms have been parallelized into different platforms, they have not yet been explored on speculative multicore architecture thoroughly, especially under CBC, CFB and OFB modes...
-
Chapter and Conference Paper
Dynamic Resource Tuning for Flexible Core Chip Multiprocessors
Technology evolving has forced the coming of chip multiprocessors (CMP) era, and enabled architects to place an increasing number of cores on single chip. For the abundance of computing resources, a fundamenta...
-
Chapter and Conference Paper
Balancing Thread Partition for Efficiently Exploiting Speculative Thread-Level Parallelism
General-purpose computing is taking an irreversible step toward on-chip parallel architectures. One way to enhance the performance of chip multiprocessors is the use of thread-level speculation (TLS). Identify...