scispace - formally typeset
Search or ask a question

Answers from top 8 papers

More filters
Papers (8)Insight
GPUswap ensures that all data is permanently accessible to the GPU and thus allows applications to submit commands to the GPU directly at any time, without the need for software scheduling.
Our experiments show that ScaleGPU can run a GPU application on any GPU memory size and also improves performance significantly.
Open accessProceedings ArticleDOI
Jianting Zhang, Simin You, Le Gruenwald 
29 Jun 2015
7 Citations
The tiny GPU cluster is equipped with standard gigabyte Ethernet network while has much less computing power and energy footprint when compared with a regular GPU cluster and represents a new platform with more balanced compute to communication ratio.
This clearly shows that direct GPU-to-GPU transfers are the key factor in obtaining good performance on multi-GPU systems.
To the best of our knowledge, this is the first paper to provide a comprehensive solution for MPI two-sided and one-sided GPU-to-GPU communication within a node, using CUDA IPC.
ScaleGPU also achieves high performance by minimizing the amount of CPU-GPU data transfers and by utilizing the GPU memory's high bandwidth.
In particular, we demonstrate the advantage of direct GPU-to-GPU transfers over data exchanges via host main memory.

Related Questions

What is the impact of GPU boosting computational fluid dynamics?5 answersThe use of GPUs for computational fluid dynamics (CFD) has had a significant impact on performance. GPU acceleration has been applied to various CFD methods, such as the Kinetic Theory of Gases, MARS (Multi-interfaces Advection and Reconstruction Solver), Piecewise-Parabolic Method (PPM), and finite volume method for solving Euler and Navier-Stokes equations. These studies have shown substantial speedup compared to CPU implementations, ranging from 4 times fasterto 20-50 times faster. The use of GPUs allows for faster computation of multi-phase flows, interface tracking, and fluid flow simulations, enabling interactive visualization and analysis. However, it is important to note that double-precision computation on GPUs requires high precision. Overall, the application of GPUs in CFD has provided a promising perspective for designing high-performance and cost-effective software frameworks for fluid dynamics simulations.
What are the advantages and disadvantages of using GPUs for AI?3 answersGPUs offer several advantages for AI applications. They can process a considerable amount of data, improve task management, and enhance learning personal experiences. Additionally, GPUs provide cost-effective hardware acceleration, making supercomputing power more accessible and affordable for desktop applications. However, there are also some disadvantages to using GPUs for AI. Some concerns include the possible loss of traditional educational jobs, the costs of implementing AI systems, errors in programming and error-processing, and the potential lack of human relationships in the classroom. Furthermore, there are differing opinions on the ability of AI to control students' behavior and direct learning, improve the efficiency of the educational system, provide notes and reviews, reduce dependency on teachers, and enhance social interaction.
Do you need a graphics card for a Plex server?8 answers
Do you need a video card for Plex server?9 answers
How can I access my NAS without Internet?10 answers
Is Windows Server good for Nas?7 answers

See what other people are reading

What is the relationship between memory bandwidth and the efficiency of data processing in modern computing?
5 answers
The relationship between memory bandwidth and the efficiency of data processing in modern computing systems is critically interdependent, as highlighted by a body of research. Modern computing systems, designed predominantly to move data towards computation, face significant performance, scalability, and energy bottlenecks due to this architecture. These challenges are exacerbated by the increasing data-intensiveness of applications, where data access from memory emerges as a key bottleneck because memory bandwidth and energy do not scale well with demand. The dichotomy between computation and data storage/movement, where a substantial portion of system energy is consumed and performance is lost in moving data, underscores the inefficiency of traditional computing paradigms. To address these inefficiencies, research has explored processing-in-memory (PIM) technologies, which aim to reduce or eliminate data movement by placing computation mechanisms closer to or inside memory storage. This approach leverages the analog operational properties of DRAM and 3D-stacked memory technology to perform massively parallel operations in memory, thereby potentially alleviating the memory bandwidth bottleneck. Furthermore, the importance of memory bandwidth is not limited to traditional computing systems but extends to specialized computing platforms like FPGAs. Efficient use of memory bandwidth is essential for fully utilizing the processing capabilities of these platforms, as demonstrated by FPGA-based implementations of two-dimensional fast Fourier transform (2D-FFT), which address the memory bandwidth bottleneck through algorithm and data path design. In addition, strategies for increasing global cache reuse and optimizing data access patterns have been proposed to mitigate bandwidth limitations and improve data processing efficiency. These strategies include fusing computations on the same data and grouping data used by the same computation to enable contiguous memory access. In summary, the relationship between memory bandwidth and data processing efficiency is a central concern in modern computing, driving innovations in memory system architecture, algorithm design, and processing paradigms to overcome inherent limitations.
What is pMTOR?
5 answers
pMTOR stands for "previously merging for masking repeats," a method proposed for masking repeats in DNA fragment assembly. The Pebble Bed Modular Reactor (PBMR) technology, used for electricity production and process heat applications, is a significant development in the nuclear energy sector. Additionally, a resource-oriented computing model called COR has been introduced to integrate user-level multithreading and communication in symmetrical multiprocessor clusters, enhancing distributed application design. Moreover, a new free-form profilometer, the MPR-700, has been designed for precise 3D metrological measurements of optics, offering high accuracy and speed in data acquisition. South Africa's PBMR programme has seen advancements, including changes in turbine orientation and core design, with a power uprate to 400 MWt, making it a prominent next-generation reactor development effort.
Is bwa mem recommended for long reads?
5 answers
Yes, BWA-MEM is recommended for long reads due to its efficiency and performance improvements. Studies have shown that BWA-MEM, especially when optimized with specific parameters like -k 19 -r 2.5, can provide mapping efficiency comparable to other methods like BWA-aln but with faster run times, making it suitable for processing long reads. Additionally, efforts have been made to accelerate BWA-MEM through efficient architecture-aware implementations, resulting in significant speedups on single-thread and single-socket processors, making it a favorable choice for handling large volumes of data from high-throughput sequencing technologies. Furthermore, implementations of BWA-MEM on GPUs have shown promising results, achieving notable speedups over CPU-based versions, further enhancing its suitability for processing long reads efficiently.
What are the key factors that influence the effectiveness of particle transport modeling using OpenFOAM?
5 answers
The effectiveness of particle transport modeling using OpenFOAM is influenced by several key factors. These include the resolution of Computational Fluid Dynamics (CFD) simulations, the efficiency of the particle-tracking solver, the accuracy of turbulence modeling, and the impact of particle velocity and temperature on deposition efficiency. Additionally, the choice of coupling regime, particle boundary conditions, and dispersion models play crucial roles in accurately predicting particle dispersion. Optimization of memory usage and computational time through specific classes for radiation sources data and hierarchical parallelization further enhance the modeling capabilities of OpenFOAM for complex radiation transport problems. OpenFOAM's versatility allows for customization in modeling particulates settling under flow fields and molecular flow, making it a solid foundation for contamination modeling in various scenarios.
What is the historical and cultural significance of shamanism in architecture?
5 answers
Shamanism holds historical and cultural significance in architecture due to its role in shaping ritual spaces and landscapes. The practice of shamanism, characterized by its recurrent features worldwide, has been proposed as one of the earliest specialized professions, influencing the design and use of ritual architecture. Shamanic beliefs have influenced the interpretation of rock art and contributed to the cultural identity of indigenous peoples, showcasing the integration of spiritual practices with architectural elements. Additionally, shamanism's ability to enhance social group cohesion distinguishes it from other magico-religious practices, emphasizing its impact on community structures and potentially influencing architectural layouts to facilitate communal rituals. Overall, shamanism's presence in architecture reflects its deep-rooted connection to cultural practices and beliefs across various societies.
What is the meaning of the word utilization?
5 answers
Utilization refers to the effective use or application of something, such as information systems, computer programs, or materials, to achieve a specific purpose or goal. It involves making the most out of available resources or features. In the context of information systems, utilization is crucial for enhancing performance. For computer programs, increasing utilization involves analyzing user behavior to identify unused or underutilized features and then guiding users towards utilizing them effectively. Designers also focus on utilization by creating products that encourage users to explore and make the most of the available features and materials. In the realm of technology, utilization can involve optimizing memory systems to efficiently access and utilize data for computational tasks.
How to solve a slope geographical problem?
5 answers
To solve a slope geographical problem efficiently, various approaches can be employed. One method involves utilizing high-performance computational systems like "TianHe 1-A" to implement new parallel algorithms for slope analysis, optimizing geographical computations. Another strategy includes creating hazard zonation maps using geographic information systems (GIS) and statistical weighting techniques to assess instability potential in different regions. Additionally, integrating research and application in physical geography can provide valuable insights, such as studying mass movement patterns and dynamic equilibrium principles using quantitative DEM data. Furthermore, for accurate slope computations in Geographic Information Systems, an iterative aggregation strategy can be applied, scaling logarithmically in window size and enabling efficient processing of high-resolution Digital Elevation Model data. These diverse methods showcase the multidimensional approaches available for addressing slope geographical problems effectively.
What is the definition of mind?
5 answers
The mind can be defined as a cognitive structure that organizes various mental capacities along dimensions such as bodily sensation, cognition, and sometimes emotion, as indicated by research across diverse cultures. In the realm of advanced parallel computer architecture, MIND (Memory, Intelligence, and Network Device) stands out as a Processor-in-Memory (PIM) architecture that integrates DRAM bit cells and CMOS logic devices on the same silicon die, supporting global shared memory and efficient parallel execution models. Furthermore, the human mind is described as a functional capability of the brain, responsible for subjective interpretations based on neurocognitive processes, cognitive judgments, and emotional experiences, shaping individual actions and responses based on detected external events and changes over time and space. Integrating Western and Indian perspectives, the concept of mind is pivotal for understanding psychological processes, cognition, consciousness, self, and paranormal phenomena, offering complementary insights for a comprehensive understanding of human functioning and development.
What is impulsive adoption?
5 answers
Impulsive adoption refers to the spontaneous and immediate acceptance and utilization of a technology or application without much premeditation. In the context of mobile shopping apps in India, the Unified Theory of Acceptance and Use of Technology 2 (UTAUT2) model was extended to include impulsiveness as a construct in understanding the adoption of such apps. This study found that impulsiveness, along with other factors like performance expectancy, effort expectancy, hedonic motivation, habit, and behavioral intention, significantly influences the use behavior of mobile shopping apps. The addition of impulsiveness to the model highlights the impulsive nature of some users in quickly embracing and engaging with mobile shopping applications, showcasing the diverse factors influencing technology adoption in different contexts.
Because is hot?
5 answers
Fever is a physiological response that results from increased prostaglandin synthesis in the hypothalamus due to pyrogenic cytokines, signaling immune system activation. Fever can be caused by various factors such as infections, immune disorders, malignancy, and drug side effects, among others. It is a common reason for seeking pediatric care, with most cases being due to viral illnesses that require supportive care and parental education. Despite fever being a beneficial response aiding in clearing infections, concerns about fever, termed fever phobia, persist among parents and healthcare professionals, leading to unnecessary treatments and anxiety. The management of fever in children involves educating parents, preventing complications, and providing supportive care, emphasizing the importance of understanding the underlying causes and appropriate management strategies.
How many citations watson?
5 answers
Watson and Crick's 1953 paper on the double helix structure of DNA has been cited extensively. Despite being cited over 2,000 times since 1961, Watson and Crick themselves have only referenced their own paper twice. The reasons for citing their work vary, with a significant proportion of citations being historical rather than actively using the findings. Additionally, a parallel version of Galton-Watson processes has been studied for generating tree-shaped structures, with over 49,000 citations attesting to the usefulness of random trees. Furthermore, the WATSON COST Action, initiated in 2020, aims to integrate scientific knowledge on stable isotopes in water to understand water dynamics in the Earth's Critical Zone, involving around 230 members from 37 European countries.