scispace - formally typeset
Search or ask a question
Topic

Degree of parallelism

About: Degree of parallelism is a research topic. Over the lifetime, 1515 publications have been published within this topic receiving 25546 citations.


Papers
More filters
Journal ArticleDOI
TL;DR: Parallel processing has been an important concern of the computer engineer for many years and is employed to achieve high performance by replication or a unique function by interconnecting different processes.
Abstract: Parallel processing has been an important concern of the computer engineer for many years. It is employed to achieve high performance by replication or a unique function by interconnecting different processes. In each case the engineering problem, of providing the structure to interconnect the processors, is a combination of issues concerning not only the implementation technology but also the needs of the application.

4 citations

Book ChapterDOI
14 Sep 2020
TL;DR: This paper addressed the problem of the optimization of UDFs in data-intensive workflows and presented the approach to construct a cost model to determine the degree of parallelism for parallelizable UDF’s.
Abstract: Optimizing Data Processing Pipelines (DPPs) is challenging in the context of both, data warehouse architectures and data science architectures. Few approaches to this problem have been proposed so far. The most challenging issue is to build a cost model of the whole DPP, especially if user defined functions (UDFs) are used. In this paper we addressed the problem of the optimization of UDFs in data-intensive workflows and presented our approach to construct a cost model to determine the degree of parallelism for parallelizable UDFs .

4 citations

Journal ArticleDOI
TL;DR: A new concept, based on the use of random signals, is proposed for a computing instrument for simulating the reliability aspects of a system, using a simulator designed and constructed using large-scale integrated circuits.

4 citations

Book ChapterDOI
14 Jan 2019
TL;DR: This work presents an iterative, greedy algorithm for automatically partitioning spreadsheets into load-balanced, acyclic groups of cells that can be scheduled to run on shared-memory multicore processors.
Abstract: Spreadsheets are popular tools for end-user development and complex modelling but can suffer from poor performance. While end-users are usually domain experts they are seldom IT professionals that can leverage today’s abundant multicore architectures to offset such poor performance. We present an iterative, greedy algorithm for automatically partitioning spreadsheets into load-balanced, acyclic groups of cells that can be scheduled to run on shared-memory multicore processors. A big-step cost semantics for the spreadsheet formula language is used to estimate work and guide partitioning. The algorithm does not require end-users to modify the spreadsheet in any way. We implement three extensions to the algorithm for further accelerating computation; two of which recognise common cell structures known as cell arrays that naturally express a degree of parallelism. To the best of our knowledge, no such automatic algorithm has previously been proposed for partitioning spreadsheets. We report a maximum 24-fold speed-up on 48 logical cores.

4 citations

Journal ArticleDOI
TL;DR: In this paper, a more effective parallel router that combines static and dynamic load balance in parallel routing for FPGAs is proposed, where hierarchical region partitioning is explored to assign routing tasks to different cores for static load balance.
Abstract: Routing is a very complex process in the field programmable gate array (FPGA) CAD flow. The increase of both FPGA size and design complexity leads to a long routing time hindering the productivity. In this article, we propose a more effective parallel router that combines static and dynamic load balance in parallel routing for FPGAs. First, we explore hierarchical region partitioning to assign routing tasks to different cores for static load balance. Then, we coordinate message propagation and task migration at runtime so that load balance between cores can be dynamically maintained in parallel routing. Finally, we combine static and dynamic load balance in the parallel routing for a higher degree of parallelism. Our parallel router performs on the multicore distributed-memory systems and the communication between cores is through message passing interface messages. We demonstrate the effectiveness of our parallel router using large-scale Titan designs. On average, our parallel router can scale up to 32 cores to achieve about $17\times $ speedup with slight loss of quality, compared with the latest VTR 8 router.

4 citations


Network Information
Related Topics (5)
Server
79.5K papers, 1.4M citations
85% related
Scheduling (computing)
78.6K papers, 1.3M citations
83% related
Network packet
159.7K papers, 2.2M citations
80% related
Web service
57.6K papers, 989K citations
80% related
Quality of service
77.1K papers, 996.6K citations
79% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20221
202147
202048
201952
201870
201775