An Efficient Stack Data Management for Scratchpad Memory based Multi-core Processors

Research output: Patent

Abstract

As technology scales, the number of cores in processors will rapidly increase in order to avoid the power wall. In processors that have hundreds and thousands of cores, traditional memory architectures in which a coherent memory interface is provided to all the cores in hardware, will not be feasible. The memory management functionality is expected to migrate from hardware (in existing systems) to the software domain. As a result, existing programs that have been written assuming that processor hardware will provide memory management support will not execute correctly on these modern many-core architectures. To make existing programs run correctly on such modern many-core architectures, programmers will have to manually change the program by inserting memory management functionality in the program. However, this is extremely difficult, and even if possible, manual memory management is tedious and a highly error-prone process. Researchers at Arizona State University have developed a compiler that can automatically compile a given application to execute on a scratchpad memory based multi-core architecture. A novel stack data management technique comprises a modified compiler infrastructure and optimized stack data management runtime library. The compiler can manage limitless stack data, solve invalid pointers, perform stack management with smaller footprint on local memory, and scale with the number of cores. Benchmark results ran on an IBM Cell processor show overhead of only 3.23% in performance. Potential Applications Chip Manufacturers reduce programming overhead of implementing scratchpad memory based multi-core processors. Application Developers create motivation to efficiently utilize the multitude of multi-core processors available in the market. Benefits and Advantages Automated compiler can automatically compile applications for scratchpad memory on multi-core architecture. Unlimited Stack Data in Local Memory the compiler can request dynamic memory allocation in the global memory and support unlimited stack data in the local memory. Better Stack Pointer Management pointer problems caused by pointing to other stack frames in global memory will no longer become an issue. Reduced Number of Direct Memory Access (DMA) stack frames are managed at a coarser granularity. Improved Performance stack data is managed in linear fashion, which reduces the management overhead on maintaining memory status if memory is fragmented by circular fashion. Therefore, our management overhead is only 3% on average. Download Original PDF For more information about the inventor(s) and their research, please see Dr. Aviral Shrivastava's directory webpage
Original languageEnglish (US)
StatePublished - Jan 28 2013

Fingerprint

Information management
Data storage equipment
Computer hardware
Storage allocation (computer)
Memory architecture
Computer programming
Interfaces (computer)
Computer systems

Cite this

@misc{f43a0e168b3540cd89ce5445a7b666cd,
title = "An Efficient Stack Data Management for Scratchpad Memory based Multi-core Processors",
abstract = "As technology scales, the number of cores in processors will rapidly increase in order to avoid the power wall. In processors that have hundreds and thousands of cores, traditional memory architectures in which a coherent memory interface is provided to all the cores in hardware, will not be feasible. The memory management functionality is expected to migrate from hardware (in existing systems) to the software domain. As a result, existing programs that have been written assuming that processor hardware will provide memory management support will not execute correctly on these modern many-core architectures. To make existing programs run correctly on such modern many-core architectures, programmers will have to manually change the program by inserting memory management functionality in the program. However, this is extremely difficult, and even if possible, manual memory management is tedious and a highly error-prone process. Researchers at Arizona State University have developed a compiler that can automatically compile a given application to execute on a scratchpad memory based multi-core architecture. A novel stack data management technique comprises a modified compiler infrastructure and optimized stack data management runtime library. The compiler can manage limitless stack data, solve invalid pointers, perform stack management with smaller footprint on local memory, and scale with the number of cores. Benchmark results ran on an IBM Cell processor show overhead of only 3.23{\%} in performance. Potential Applications Chip Manufacturers reduce programming overhead of implementing scratchpad memory based multi-core processors. Application Developers create motivation to efficiently utilize the multitude of multi-core processors available in the market. Benefits and Advantages Automated compiler can automatically compile applications for scratchpad memory on multi-core architecture. Unlimited Stack Data in Local Memory the compiler can request dynamic memory allocation in the global memory and support unlimited stack data in the local memory. Better Stack Pointer Management pointer problems caused by pointing to other stack frames in global memory will no longer become an issue. Reduced Number of Direct Memory Access (DMA) stack frames are managed at a coarser granularity. Improved Performance stack data is managed in linear fashion, which reduces the management overhead on maintaining memory status if memory is fragmented by circular fashion. Therefore, our management overhead is only 3{\%} on average. Download Original PDF For more information about the inventor(s) and their research, please see Dr. Aviral Shrivastava's directory webpage",
author = "Aviral Shrivastava",
year = "2013",
month = "1",
day = "28",
language = "English (US)",
type = "Patent",

}

TY - PAT

T1 - An Efficient Stack Data Management for Scratchpad Memory based Multi-core Processors

AU - Shrivastava, Aviral

PY - 2013/1/28

Y1 - 2013/1/28

N2 - As technology scales, the number of cores in processors will rapidly increase in order to avoid the power wall. In processors that have hundreds and thousands of cores, traditional memory architectures in which a coherent memory interface is provided to all the cores in hardware, will not be feasible. The memory management functionality is expected to migrate from hardware (in existing systems) to the software domain. As a result, existing programs that have been written assuming that processor hardware will provide memory management support will not execute correctly on these modern many-core architectures. To make existing programs run correctly on such modern many-core architectures, programmers will have to manually change the program by inserting memory management functionality in the program. However, this is extremely difficult, and even if possible, manual memory management is tedious and a highly error-prone process. Researchers at Arizona State University have developed a compiler that can automatically compile a given application to execute on a scratchpad memory based multi-core architecture. A novel stack data management technique comprises a modified compiler infrastructure and optimized stack data management runtime library. The compiler can manage limitless stack data, solve invalid pointers, perform stack management with smaller footprint on local memory, and scale with the number of cores. Benchmark results ran on an IBM Cell processor show overhead of only 3.23% in performance. Potential Applications Chip Manufacturers reduce programming overhead of implementing scratchpad memory based multi-core processors. Application Developers create motivation to efficiently utilize the multitude of multi-core processors available in the market. Benefits and Advantages Automated compiler can automatically compile applications for scratchpad memory on multi-core architecture. Unlimited Stack Data in Local Memory the compiler can request dynamic memory allocation in the global memory and support unlimited stack data in the local memory. Better Stack Pointer Management pointer problems caused by pointing to other stack frames in global memory will no longer become an issue. Reduced Number of Direct Memory Access (DMA) stack frames are managed at a coarser granularity. Improved Performance stack data is managed in linear fashion, which reduces the management overhead on maintaining memory status if memory is fragmented by circular fashion. Therefore, our management overhead is only 3% on average. Download Original PDF For more information about the inventor(s) and their research, please see Dr. Aviral Shrivastava's directory webpage

AB - As technology scales, the number of cores in processors will rapidly increase in order to avoid the power wall. In processors that have hundreds and thousands of cores, traditional memory architectures in which a coherent memory interface is provided to all the cores in hardware, will not be feasible. The memory management functionality is expected to migrate from hardware (in existing systems) to the software domain. As a result, existing programs that have been written assuming that processor hardware will provide memory management support will not execute correctly on these modern many-core architectures. To make existing programs run correctly on such modern many-core architectures, programmers will have to manually change the program by inserting memory management functionality in the program. However, this is extremely difficult, and even if possible, manual memory management is tedious and a highly error-prone process. Researchers at Arizona State University have developed a compiler that can automatically compile a given application to execute on a scratchpad memory based multi-core architecture. A novel stack data management technique comprises a modified compiler infrastructure and optimized stack data management runtime library. The compiler can manage limitless stack data, solve invalid pointers, perform stack management with smaller footprint on local memory, and scale with the number of cores. Benchmark results ran on an IBM Cell processor show overhead of only 3.23% in performance. Potential Applications Chip Manufacturers reduce programming overhead of implementing scratchpad memory based multi-core processors. Application Developers create motivation to efficiently utilize the multitude of multi-core processors available in the market. Benefits and Advantages Automated compiler can automatically compile applications for scratchpad memory on multi-core architecture. Unlimited Stack Data in Local Memory the compiler can request dynamic memory allocation in the global memory and support unlimited stack data in the local memory. Better Stack Pointer Management pointer problems caused by pointing to other stack frames in global memory will no longer become an issue. Reduced Number of Direct Memory Access (DMA) stack frames are managed at a coarser granularity. Improved Performance stack data is managed in linear fashion, which reduces the management overhead on maintaining memory status if memory is fragmented by circular fashion. Therefore, our management overhead is only 3% on average. Download Original PDF For more information about the inventor(s) and their research, please see Dr. Aviral Shrivastava's directory webpage

M3 - Patent

ER -