Academic literature on the topic 'Memory hierarchy'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Memory hierarchy.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Memory hierarchy"

1

Balasubramonian, R., D. H. Albonesi, A. Buyuktosunoglu, and S. Dwarkadas. "A dynamically tunable memory hierarchy." IEEE Transactions on Computers 52, no. 10 (October 2003): 1243–58. http://dx.doi.org/10.1109/tc.2003.1234523.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Tabak, Daniel. "Cache and Memory Hierarchy Design." ACM SIGARCH Computer Architecture News 23, no. 3 (June 1995): 28. http://dx.doi.org/10.1145/203618.564957.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Bauer, Michael, John Clark, Eric Schkufza, and Alex Aiken. "Programming the memory hierarchy revisited." ACM SIGPLAN Notices 46, no. 8 (September 7, 2011): 13–24. http://dx.doi.org/10.1145/2038037.1941558.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Graefe, Goetz. "Sorting in a Memory Hierarchy with Flash Memory." Datenbank-Spektrum 11, no. 2 (July 15, 2011): 83–90. http://dx.doi.org/10.1007/s13222-011-0062-6.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Lopes, Alba Sandyra Bezerra, Ivan Saraiva Silva, and Luciano Volcan Agostini. "A Memory Hierarchy Model Based on Data Reuse for Full-Search Motion Estimation on High-Definition Digital Videos." International Journal of Reconfigurable Computing 2012 (2012): 1–10. http://dx.doi.org/10.1155/2012/473725.

Full text
Abstract:
The motion estimation is the most complex module in a video encoder requiring a high processing throughput and high memory bandwidth, mainly when the focus is high-definition videos. The throughput problem can be solved increasing the parallelism in the internal operations. The external memory bandwidth may be reduced using a memory hierarchy. This work presents a memory hierarchy model for a full-search motion estimation core. The proposed memory hierarchy model is based on a data reuse scheme considering the full search algorithm features. The proposed memory hierarchy expressively reduces the external memory bandwidth required for the motion estimation process, and it provides a very high data throughput for the ME core. This throughput is necessary to achieve real time when processing high-definition videos. When considering the worst bandwidth scenario, this memory hierarchy is able to reduce the external memory bandwidth in 578 times. A case study for the proposed hierarchy, using32×32search window and8×8block size, was implemented and prototyped on a Virtex 4 FPGA. The results show that it is possible to reach 38 frames per second when processing full HD frames (1920×1080pixels) using nearly 299 Mbytes per second of external memory bandwidth.
APA, Harvard, Vancouver, ISO, and other styles
6

Isabel, Maria. "Stress amplifies memory for social hierarchy." Frontiers in Neuroscience 1, no. 1 (November 1, 2007): 175–84. http://dx.doi.org/10.3389/neuro.01.1.1.013.2007.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Qiming Hou, Xin Sun, Kun Zhou, C. Lauterbach, and D. Manocha. "Memory-Scalable GPU Spatial Hierarchy Construction." IEEE Transactions on Visualization and Computer Graphics 17, no. 4 (April 2011): 466–74. http://dx.doi.org/10.1109/tvcg.2010.88.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Yotov, Kamen, Keshav Pingali, and Paul Stodghill. "Automatic measurement of memory hierarchy parameters." ACM SIGMETRICS Performance Evaluation Review 33, no. 1 (June 6, 2005): 181–92. http://dx.doi.org/10.1145/1071690.1064233.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Mei, Xinxin, and Xiaowen Chu. "Dissecting GPU Memory Hierarchy Through Microbenchmarking." IEEE Transactions on Parallel and Distributed Systems 28, no. 1 (January 1, 2017): 72–86. http://dx.doi.org/10.1109/tpds.2016.2549523.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

DEVILLERS, OLIVIER. "THE DELAUNAY HIERARCHY." International Journal of Foundations of Computer Science 13, no. 02 (April 2002): 163–80. http://dx.doi.org/10.1142/s0129054102001035.

Full text
Abstract:
We propose a new data structure to compute the Delaunay triangulation of a set of points in the plane. It combines good worst case complexity, fast behavior on real data, small memory occupation and the possibility of fully dynamic insertions and deletions. The location structure is organized into several levels. The lowest level just consists of the triangulation, then each level contains the triangulation of a small sample of the level below. Point location is done by walking in a triangulation to determine the nearest neighbor of the query at that level, then the walk restarts from the neighbor at the level below. Using a small subset (3%) to sample a level allows a small memory occupation; the walk and the use of the nearest neighbor to change levels quickly locate the query.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Memory hierarchy"

1

Gan, Yee Ling. "Redesigning the memory hierarchy for memory-safe programming languages." Thesis, Massachusetts Institute of Technology, 2018. http://hdl.handle.net/1721.1/119765.

Full text
Abstract:
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2018.
This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 69-75).
We present Hotpads, a new memory hierarchy designed from the ground up for modern, memory-safe languages like Java, Go, and Rust. Memory-safe languages hide the memory layout from the programmer. This prevents memory corruption bugs, improves programmability, and enables automatic memory management. Hotpads extends the same insight to the memory hierarchy: it hides the memory layout from software and enables hardware to take control over it, dispensing with the conventional flat address space abstraction. This avoids the need for associative caches and virtual memory. Instead, Hotpads moves objects across a hierarchy of directly-addressed memories. It rewrites pointers to avoid most associative lookups, provides hardware support for memory allocation, and unifies hierarchical garbage collection and data placement. As a result, Hotpads improves memory performance and efficiency substantially, and unlocks many new optimizations. This thesis contributes important optimizations for Hotpads and a comprehensive evaluation of Hotpads against prior work.
by Yee Ling Gan.
M. Eng.
APA, Harvard, Vancouver, ISO, and other styles
2

Dublish, Saumay Kumar. "Managing the memory hierarchy in GPUs." Thesis, University of Edinburgh, 2018. http://hdl.handle.net/1842/31205.

Full text
Abstract:
Pervasive use of GPUs across multiple disciplines is a result of continuous adaptation of the GPU architectures to address the needs of upcoming application domains. One such vital improvement is the introduction of the on-chip cache hierarchy, used primarily to filter the high bandwidth demand to the off-chip memory. However, in contrast to traditional CPUs, the cache hierarchy in GPUs is presented with significantly different challenges such as cache thrashing and bandwidth bottlenecks, arising due to small caches and high levels of memory traffic. These challenges lead to severe congestion across the memory hierarchy, resulting in high memory access latencies. In memory-intensive applications, such high memory access latencies often get exposed and can no longer be hidden through multithreading, and therefore adversely impact system performance. In this thesis, we address the inefficiencies across the memory hierarchy in GPUs that lead to such high levels of congestion. We identify three major factors contributing to poor memory system performance: first, disproportionate and insufficient bandwidth resources in the cache hierarchy; second, poor cache management policies; and third, high levels of multithreading. In order to revitalize the memory hierarchy by addressing the above limitations, we propose a three-pronged approach. First, we characterize the bandwidth bottlenecks present across the memory hierarchy in GPUs and identify the architectural parameters that are most critical in alleviating congestion. Subsequently, we explore the architectural design space to mitigate the bandwidth bottlenecks in a cost-effective manner. Second, we identify significant inter-core reuse in GPUs, presenting an opportunity to reuse data among the L1s. We exploit this reuse by connecting the L1 caches with a lightweight ring network to facilitate inter-core communication of shared data. We show that this technique reduces traffic to the L2 cache, freeing up the bandwidth for other accesses. Third, we present Poise, a machine learning approach to mitigate cache thrashing and bandwidth bottlenecks by altering the levels of multi-threading. Poise comprises a supervised learning model that is trained offline on a set of profiled kernels to make good warp scheduling decisions. Subsequently, a hardware inference engine is used to predict good warp scheduling decisions at runtime using the model learned during training. In summary, we address the problem of bandwidth bottlenecks across the memory hierarchy in GPUs by exploring how to best scale, supplement and utilize the existing bandwidth resources. These techniques provide an effective and comprehensive methodology to mitigate the bandwidth bottlenecks in the GPU memory hierarchy.
APA, Harvard, Vancouver, ISO, and other styles
3

Taylor, Nathan Bryan. "Cachekata : memory hierarchy optimization via dynamic binary translation." Thesis, University of British Columbia, 2013. http://hdl.handle.net/2429/44335.

Full text
Abstract:
As hardware parallelism continues to increase, CPU caches can no longer be considered a transparent, hardware-level performance optimization. Adverse cache impact on performance is entirely workload-dependent and may depend on runtime factors. The operating system must begin to treat CPU caches like any other shared hardware resource to effectively support workloads on parallel hardware. We present a binary translation system called Cachekata that provides a byte-granular memory remapping facility within the OS in an efficient manner. Cachekata is incorporated into a larger system, Plastic, which diagnoses and corrects instances of false sharing occurring within running applications. Our implementation is able to achieve a 3-6x performance improvement on known examples of false sharing in parallel benchmarks.
APA, Harvard, Vancouver, ISO, and other styles
4

Kim, Jinwoo. "Memory hierarchy management through off-line computational learning." Diss., Georgia Institute of Technology, 2003. http://hdl.handle.net/1853/8194.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Greenwald, Benjamin Eliot 1975. "A technique for compilation to exposed memory hierarchy." Thesis, Massachusetts Institute of Technology, 1999. http://hdl.handle.net/1721.1/87163.

Full text
Abstract:
Thesis (S.M.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, June 2000.
"September 1999."
Includes bibliographical references (p. 55-56).
by Benjamin Eliot Greenwald.
S.M.
APA, Harvard, Vancouver, ISO, and other styles
6

Dimić, Vladimir. "Runtime-assisted optimizations in the on-chip memory hierarchy." Doctoral thesis, Universitat Politècnica de Catalunya, 2020. http://hdl.handle.net/10803/670363.

Full text
Abstract:
Following Moore's Law, the number of transistors on chip has been increasing exponentially, which has led to the increasing complexity of modern processors. As a result, the efficient programming of such systems has become more difficult. Many programming models have been developed to answer this issue. Of particular interest are task-based programming models that employ simple annotations to define parallel work in an application. The information available at the level of the runtime systems associated with these programming models offers great potential for improving hardware design. Moreover, due to technological limitations, Moore's Law is predicted to eventually come to an end, so novel paradigms are necessary to maintain the current performance improvement trends. The main goal of this thesis is to exploit the knowledge about a parallel application available at the runtime system level to improve the design of the on-chip memory hierarchy. The coupling of the runtime system and the microprocessor enables a better hardware design without hurting the programmability. The first contribution is a set of insertion policies for shared last-level caches that exploit information about tasks and task data dependencies. The intuition behind this proposal revolves around the observation that parallel threads exhibit different memory access patterns. Even within the same thread, accesses to different variables often follow distinct patterns. The proposed policies insert cache lines into different logical positions depending on the dependency type and task type to which the corresponding memory request belongs. The second proposal optimizes the execution of reductions, defined as a programming pattern that combines input data to form the resulting reduction variable. This is achieved with a runtime-assisted technique for performing reductions in the processor's cache hierarchy. The proposal's goal is to be a universally applicable solution regardless of the reduction variable type, size and access pattern. On the software level, the programming model is extended to let a programmer specify the reduction variables for tasks, as well as the desired cache level where a certain reduction will be performed. The source-to-source compiler and the runtime system are extended to translate and forward this information to the underlying hardware. On the hardware level, private and shared caches are equipped with functional units and the accompanying logic to perform reductions at the cache level. This design avoids unnecessary data movements to the core and back as the data is operated at the place where it resides. The third contribution is a runtime-assisted prioritization scheme for memory requests inside the on-chip memory hierarchy. The proposal is based on the notion of a critical path in the context of parallel codes and a known fact that accelerating critical tasks reduces the execution time of the whole application. In the context of this work, task criticality is observed at a level of a task type as it enables simple annotation by the programmer. The acceleration of critical tasks is achieved by the prioritization of corresponding memory requests in the microprocessor.
Siguiendo la ley de Moore, el número de transistores en los chips ha crecido exponencialmente, lo que ha comportado una mayor complejidad en los procesadores modernos y, como resultado, de la dificultad de la programación eficiente de estos sistemas. Se han desarrollado muchos modelos de programación para resolver este problema; un ejemplo particular son los modelos de programación basados en tareas, que emplean anotaciones sencillas para definir los Trabajos paralelos de una aplicación. La información de que disponen los sistemas en tiempo de ejecución (runtime systems) asociada con estos modelos de programación ofrece un enorme potencial para la mejora del diseño del hardware. Por otro lado, las limitaciones tecnológicas hacen que la ley de Moore pueda dejar de cumplirse próximamente, por lo que se necesitan paradigmas nuevos para mantener las tendencias actuales de mejora de rendimiento. El objetivo principal de esta tesis es aprovechar el conocimiento de las aplicaciones paral·leles de que dispone el runtime system para mejorar el diseño de la jerarquía de memoria del chip. El acoplamiento del runtime system junto con el microprocesador permite realizar mejores diseños hardware sin afectar Negativamente en la programabilidad de dichos sistemas. La primera contribución de esta tesis consiste en un conjunto de políticas de inserción para las memorias caché compartidas de último nivel que aprovecha la información de las tareas y las dependencias de datos entre estas. La intuición tras esta propuesta se basa en la observación de que los hilos de ejecución paralelos muestran distintos patrones de acceso a memoria e, incluso dentro del mismo hilo, los accesos a diferentes variables a menudo siguen patrones distintos. Las políticas que se proponen insertan líneas de caché en posiciones lógicas diferentes en función de los tipos de dependencia y tarea a los que corresponde la petición de memoria. La segunda propuesta optimiza la ejecución de las reducciones, que se definen como un patrón de programación que combina datos de entrada para conseguir la variable de reducción como resultado. Esto se consigue mediante una técnica asistida por el runtime system para la realización de reducciones en la jerarquía de la caché del procesador, con el objetivo de ser una solución aplicable de forma universal sin depender del tipo de la variable de la reducción, su tamaño o el patrón de acceso. A nivel de software, el modelo de programación se extiende para que el programador especifique las variables de reducción de las tareas, así como el nivel de caché escogido para que se realice una determinada reducción. El compilador fuente a Fuente (compilador source-to-source) y el runtime ssytem se modifican para que traduzcan y pasen esta información al hardware subyacente, evitando así movimientos de datos innecesarios hacia y desde el núcleo del procesador, al realizarse la operación donde se encuentran los datos de la misma. La tercera contribución proporciona un esquema de priorización asistido por el runtime system para peticiones de memoria dentro de la jerarquía de memoria del chip. La propuesta se basa en la noción de camino crítico en el contexto de los códigos paralelos y en el hecho conocido de que acelerar tareas críticas reduce el tiempo de ejecución de la aplicación completa. En el contexto de este trabajo, la criticidad de las tareas se considera a nivel del tipo de tarea ya que permite que el programador las indique mediante anotaciones sencillas. La aceleración de las tareas críticas se consigue priorizando las correspondientes peticiones de memoria en el microprocesador.
Seguint la llei de Moore, el nombre de transistors que contenen els xips ha patit un creixement exponencial, fet que ha provocat un augment de la complexitat dels processadors moderns i, per tant, de la dificultat de la programació eficient d’aquests sistemes. Per intentar solucionar-ho, s’han desenvolupat diversos models de programació; un exemple particular en són els models basats en tasques, que fan servir anotacions senzilles per definir treballs paral·lels dins d’una aplicació. La informació que hi ha al nivell dels sistemes en temps d’execució (runtime systems) associada amb aquests models de programació ofereix un gran potencial a l’hora de millorar el disseny del maquinari. D’altra banda, les limitacions tecnològiques fan que la llei de Moore pugui deixar de complir-se properament, per la qual cosa calen nous paradigmes per mantenir les tendències actuals en la millora de rendiment. L’objectiu principal d’aquesta tesi és aprofitar els coneixements que el runtime System té d’una aplicació paral·lela per millorar el disseny de la jerarquia de memòria dins el xip. L’acoblament del runtime system i el microprocessador permet millorar el disseny del maquinari sense malmetre la programabilitat d’aquests sistemes. La primera contribució d’aquesta tesi consisteix en un conjunt de polítiques d’inserció a les memòries cau (cache memories) compartides d’últim nivell que aprofita informació sobre tasques i les dependències de dades entre aquestes. La intuïció que hi ha al darrere d’aquesta proposta es basa en el fet que els fils d’execució paral·lels mostren diferents patrons d’accés a la memòria; fins i tot dins el mateix fil, els accessos a variables diferents sovint segueixen patrons diferents. Les polítiques que s’hi proposen insereixen línies de la memòria cau a diferents ubicacions lògiques en funció dels tipus de dependència i de tasca als quals correspon la petició de memòria. La segona proposta optimitza l’execució de les reduccions, que es defineixen com un patró de programació que combina dades d’entrada per aconseguir la variable de reducció com a resultat. Això s’aconsegueix mitjançant una tècnica assistida pel runtime system per dur a terme reduccions en la jerarquia de la memòria cau del processador, amb l’objectiu que la proposta sigui aplicable de manera universal, sense dependre del tipus de la variable a la qual es realitza la reducció, la seva mida o el patró d’accés. A nivell de programari, es realitza una extensió del model de programació per facilitar que el programador especifiqui les variables de les reduccions que usaran les tasques, així com el nivell de memòria cau desitjat on s’hauria de realitzar una certa reducció. El compilador font a font (compilador source-to-source) i el runtime system s’amplien per traduir i passar aquesta informació al maquinari subjacent. A nivell de maquinari, les memòries cau privades i compartides s’equipen amb unitats funcionals i la lògica corresponent per poder dur a terme les reduccions a la pròpia memòria cau, evitant així moviments de dades innecessaris entre el nucli del processador i la jerarquia de memòria. La tercera contribució proporciona un esquema de priorització assistit pel runtime System per peticions de memòria dins de la jerarquia de memòria del xip. La proposta es basa en la noció de camí crític en el context dels codis paral·lels i en el fet conegut que l’acceleració de les tasques que formen part del camí crític redueix el temps d’execució de l’aplicació sencera. En el context d’aquest treball, la criticitat de les tasques s’observa al nivell del seu tipus ja que permet que el programador les indiqui mitjançant anotacions senzilles. L’acceleració de les tasques crítiques s’aconsegueix prioritzant les corresponents peticions de memòria dins el microprocessador.
APA, Harvard, Vancouver, ISO, and other styles
7

Low, Douglas Wai Kok. "Network processor memory hierarchy designs for IP packet classification /." Thesis, Connect to this title online; UW restricted, 2005. http://hdl.handle.net/1773/6973.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Vitkovskiy, Arseniy <1979&gt. "Memory hierarchy and data communication in heterogeneous reconfigurable SoCs." Doctoral thesis, Alma Mater Studiorum - Università di Bologna, 2008. http://amsdottorato.unibo.it/1127/1/Tesi_Vitkovskiy_Arseniy.pdf.

Full text
Abstract:
The miniaturization race in the hardware industry aiming at continuous increasing of transistor density on a die does not bring respective application performance improvements any more. One of the most promising alternatives is to exploit a heterogeneous nature of common applications in hardware. Supported by reconfigurable computation, which has already proved its efficiency in accelerating data intensive applications, this concept promises a breakthrough in contemporary technology development. Memory organization in such heterogeneous reconfigurable architectures becomes very critical. Two primary aspects introduce a sophisticated trade-off. On the one hand, a memory subsystem should provide well organized distributed data structure and guarantee the required data bandwidth. On the other hand, it should hide the heterogeneous hardware structure from the end-user, in order to support feasible high-level programmability of the system. This thesis work explores the heterogeneous reconfigurable hardware architectures and presents possible solutions to cope the problem of memory organization and data structure. By the example of the MORPHEUS heterogeneous platform, the discussion follows the complete design cycle, starting from decision making and justification, until hardware realization. Particular emphasis is made on the methods to support high system performance, meet application requirements, and provide a user-friendly programmer interface. As a result, the research introduces a complete heterogeneous platform enhanced with a hierarchical memory organization, which copes with its task by means of separating computation from communication, providing reconfigurable engines with computation and configuration data, and unification of heterogeneous computational devices using local storage buffers. It is distinguished from the related solutions by distributed data-flow organization, specifically engineered mechanisms to operate with data on local domains, particular communication infrastructure based on Network-on-Chip, and thorough methods to prevent computation and communication stalls. In addition, a novel advanced technique to accelerate memory access was developed and implemented.
APA, Harvard, Vancouver, ISO, and other styles
9

Vitkovskiy, Arseniy <1979&gt. "Memory hierarchy and data communication in heterogeneous reconfigurable SoCs." Doctoral thesis, Alma Mater Studiorum - Università di Bologna, 2008. http://amsdottorato.unibo.it/1127/.

Full text
Abstract:
The miniaturization race in the hardware industry aiming at continuous increasing of transistor density on a die does not bring respective application performance improvements any more. One of the most promising alternatives is to exploit a heterogeneous nature of common applications in hardware. Supported by reconfigurable computation, which has already proved its efficiency in accelerating data intensive applications, this concept promises a breakthrough in contemporary technology development. Memory organization in such heterogeneous reconfigurable architectures becomes very critical. Two primary aspects introduce a sophisticated trade-off. On the one hand, a memory subsystem should provide well organized distributed data structure and guarantee the required data bandwidth. On the other hand, it should hide the heterogeneous hardware structure from the end-user, in order to support feasible high-level programmability of the system. This thesis work explores the heterogeneous reconfigurable hardware architectures and presents possible solutions to cope the problem of memory organization and data structure. By the example of the MORPHEUS heterogeneous platform, the discussion follows the complete design cycle, starting from decision making and justification, until hardware realization. Particular emphasis is made on the methods to support high system performance, meet application requirements, and provide a user-friendly programmer interface. As a result, the research introduces a complete heterogeneous platform enhanced with a hierarchical memory organization, which copes with its task by means of separating computation from communication, providing reconfigurable engines with computation and configuration data, and unification of heterogeneous computational devices using local storage buffers. It is distinguished from the related solutions by distributed data-flow organization, specifically engineered mechanisms to operate with data on local domains, particular communication infrastructure based on Network-on-Chip, and thorough methods to prevent computation and communication stalls. In addition, a novel advanced technique to accelerate memory access was developed and implemented.
APA, Harvard, Vancouver, ISO, and other styles
10

Senni, Sophiane. "Exploration of non-volatile magnetic memory for processor architecture." Thesis, Montpellier, 2015. http://www.theses.fr/2015MONTS264/document.

Full text
Abstract:
De par la réduction continuelle des dimensions du transistor CMOS, concevoir des systèmes sur puce (SoC) à la fois très denses et énergétiquement efficients devient un réel défi. Concernant la densité, réduire la dimension du transistor CMOS est sujet à de fortes contraintes de fabrication tandis que le coût ne cesse d'augmenter. Concernant l'aspect énergétique, une augmentation importante de la puissance dissipée par unité de surface frêne l'évolution en performance. Ceci est essentiellement dû à l'augmentation du courant de fuite dans les transistors CMOS, entraînant une montée de la consommation d'énergie statique. En observant les SoCs actuels, les mémoires embarquées volatiles tels que la SRAM et la DRAM occupent de plus en plus de surface silicium. C'est la raison pour laquelle une partie significative de la puissance totale consommée provient des composants mémoires. Ces deux dernières décennies, de nouvelles mémoires non volatiles sont apparues possédant des caractéristiques pouvant aider à résoudre les problèmes des SoCs actuels. Parmi elles, la MRAM est une candidate à fort potentiel car elle permet à la fois une forte densité d'intégration et une consommation d'énergie statique quasi nulle, tout en montrant des performances comparables à la SRAM et à la DRAM. De plus, la MRAM a la capacité d'être non volatile. Ceci est particulièrement intéressant pour l'ajout de nouvelles fonctionnalités afin d'améliorer l'efficacité énergétique ainsi que la fiabilité. Ce travail de thèse a permis de mener une exploration en surface, performance et consommation énergétique de l'intégration de la MRAM au sein de la hiérarchie mémoire d'un processeur. Une première exploration fine a été réalisée au niveau mémoire cache pour des architectures multicoeurs. Une seconde étude a permis d'évaluer la possibilité d'intégrer la MRAM au niveau registre pour la conception d'un processeur non volatile. Dans le cadre d'applications des objets connectés, de nouvelles fonctionnalités ainsi que les intérêts apportés par la non volatilité ont été étudiés et évalués
With the downscaling of the complementary metal-oxide semiconductor (CMOS) technology,designing dense and energy-efficient systems-on-chip (SoC) is becoming a realchallenge. Concerning the density, reducing the CMOS transistor size faces up to manufacturingconstraints while the cost increases exponentially. Regarding the energy, a significantincrease of the power density and dissipation obstructs further improvement inperformance. This issue is mainly due to the growth of the leakage current of the CMOStransistors, which leads to an increase of the static energy consumption. Observing currentSoCs, more and more area is occupied by embedded volatile memories, such as staticrandom access memory (SRAM) and dynamic random access memory (DRAM). As a result,a significant proportion of total power is spent into memory systems. In the past twodecades, alternative memory technologies have emerged with attractive characteristics tomitigate the aforementioned issues. Among these technologies, magnetic random accessmemory (MRAM) is a promising candidate as it combines simultaneously high densityand very low static power consumption while its performance is competitive comparedto SRAM and DRAM. Moreover, MRAM is non-volatile. This capability, if present inembedded memories, has the potential to add new features to SoCs to enhance energyefficiency and reliability. In this thesis, an area, performance and energy exploration ofembedding the MRAM technology in the memory hierarchy of a processor architectureis investigated. A first fine-grain exploration was made at cache level for multi-core architectures.A second study evaluated the possibility to design a non-volatile processorintegrating MRAM at register level. Within the context of internet of things, new featuresand the benefits brought by the non-volatility were investigated
APA, Harvard, Vancouver, ISO, and other styles
More sources

Books on the topic "Memory hierarchy"

1

Sun, Guangyu. Exploring Memory Hierarchy Design with Emerging Memory Technologies. Cham: Springer International Publishing, 2014. http://dx.doi.org/10.1007/978-3-319-00681-9.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Hexsel, Roberto A. The performance of SCI memory hierarchies. Edinburgh: University of Edinburgh, Dept. of Computer Science, 1994.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Przybylski, Steven A. Cache and memory hierarchy design: A performance-directed approach. San Mateo, Calif: Morgan Kaufmann Publishers, 1990.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

Sardashti, Somayeh, Angelos Arelakis, Per Stenström, and David A. Wood. A Primer on Compression in the Memory Hierarchy. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-031-01751-3.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

The fractal structure of data reference: Applications to the memory hierarchy. Boston, MA: Kluwer Academic, 2000.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

T, Chronopoulos A. Implementation of preconditioned S-step conjugate gradient methods on a multiprocessor system with memory hierarchy. Urbana, IL (1304 W. Springfield Ave., Urbana 61801): Dept. of Computer Science, University of Illinois at Urbana-Champaign, 1987.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
7

T, Chronopoulos A. Implementation of s-step methods on parallel vector architectures. Urbana, IL (1304 W. Springfield Ave., Urbana 61801): Dept. of Computer Science, University of Illinois at Urbana-Champaign, 1987.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
8

Henning, Grant. A study of the effects of variation of short-term memory load, reading response length, and processing hierarchy on TOEFL listening comprehension item performance. Princeton, NJ: Educational Testing Service, 1991.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
9

Cache and Memory Hierarchy Design. Elsevier, 1990. http://dx.doi.org/10.1016/c2009-0-27582-9.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Sun, Guangyu. Exploring Memory Hierarchy Design with Emerging Memory Technologies. Springer, 2013.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Book chapters on the topic "Memory hierarchy"

1

Manegold, Stefan. "Memory Hierarchy." In Encyclopedia of Database Systems, 1–8. New York, NY: Springer New York, 2016. http://dx.doi.org/10.1007/978-1-4899-7993-3_657-2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Rochange, Christine, Sascha Uhrig, and Pascal Sainrat. "Memory Hierarchy." In Time-Predictable Architectures, 69–104. Hoboken, NJ, USA: John Wiley & Sons, Inc., 2014. http://dx.doi.org/10.1002/9781118790229.ch4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Manegold, Stefan. "Memory Hierarchy." In Encyclopedia of Database Systems, 1707–13. Boston, MA: Springer US, 2009. http://dx.doi.org/10.1007/978-0-387-39940-9_657.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Manegold, Stefan. "Memory Hierarchy." In Encyclopedia of Database Systems, 2222–29. New York, NY: Springer New York, 2018. http://dx.doi.org/10.1007/978-1-4614-8265-9_657.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Mohammad, Baker. "Embedded Memory Hierarchy." In Analog Circuits and Signal Processing, 29–35. New York, NY: Springer New York, 2013. http://dx.doi.org/10.1007/978-1-4614-8881-1_3.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Waidyasooriya, Hasitha Muthumala, Masanori Hariyama, and Kunio Uchiyama. "Exploiting the Memory Hierarchy." In Design of FPGA-Based Computing Systems with OpenCL, 75–91. Cham: Springer International Publishing, 2017. http://dx.doi.org/10.1007/978-3-319-68161-0_5.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Sardashti, Somayeh, Angelos Arelakis, Per Stenström, and David A. Wood. "Memory Compression." In A Primer on Compression in the Memory Hierarchy, 33–43. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-031-01751-3_4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Nieplocha, Jarek, Robert Harrison, and Ian Foster. "Explicit Management of Memory Hierarchy." In Advances in High Performance Computing, 185–99. Dordrecht: Springer Netherlands, 1997. http://dx.doi.org/10.1007/978-94-011-5514-4_11.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Sun, Guangyu. "Replacing Different Levels of the Memory Hierarchy with NVMs." In Exploring Memory Hierarchy Design with Emerging Memory Technologies, 13–67. Cham: Springer International Publishing, 2013. http://dx.doi.org/10.1007/978-3-319-00681-9_2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Sun, Guangyu. "Introduction." In Exploring Memory Hierarchy Design with Emerging Memory Technologies, 1–11. Cham: Springer International Publishing, 2013. http://dx.doi.org/10.1007/978-3-319-00681-9_1.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Memory hierarchy"

1

CAIANIELLO, E. R. "STRUCTURE AND HIERARCHY." In Scientific Highlights in Memory of Léon Van Hove. WORLD SCIENTIFIC, 1993. http://dx.doi.org/10.1142/9789812795977_0001.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Bauer, Michael, John Clark, Eric Schkufza, and Alex Aiken. "Programming the memory hierarchy revisited." In the 16th ACM symposium. New York, New York, USA: ACM Press, 2011. http://dx.doi.org/10.1145/1941553.1941558.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Ayers, Grant, Jung Ho Ahn, Christos Kozyrakis, and Parthasarathy Ranganathan. "Memory Hierarchy for Web Search." In 2018 IEEE International Symposium on High Performance Computer Architecture (HPCA). IEEE, 2018. http://dx.doi.org/10.1109/hpca.2018.00061.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Fatahalian, Kayvon, Timothy Knight, Mike Houston, Mattan Erez, Daniel Horn, Larkhoon Leem, Ji Park, et al. "Sequoia: Programming the Memory Hierarchy." In ACM/IEEE SC 2006 Conference (SC'06). IEEE, 2006. http://dx.doi.org/10.1109/sc.2006.55.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Van Laer, Anouk, William Wang, and Chris Emmons. "Inefficiencies in the Cache Hierarchy." In MEMSYS '15: International Symposium on Memory Systems. New York, NY, USA: ACM, 2015. http://dx.doi.org/10.1145/2818950.2818980.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Liu, Lei, Hao Yang, Yong Li, Mengyao Xie, Lian Li, and Chenggang Wu. "Memos: A full hierarchy hybrid memory management framework." In 2016 IEEE 34th International Conference on Computer Design (ICCD). IEEE, 2016. http://dx.doi.org/10.1109/iccd.2016.7753305.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Li, Pengcheng, Hao Luo, and Chen Ding. "Rethinking a heap hierarchy as a cache hierarchy: a higher-order theory of memory demand (HOTM)." In ISMM '16: International Symposium on Memory Management. New York, NY, USA: ACM, 2016. http://dx.doi.org/10.1145/2926697.2926708.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Lioris, Theodoros, Grigoris Dimitroulakos, and Kostas Masselos. "XMSIM: EXtensible Memory SIMulator for Early Memory Hierarchy Evaluation." In 2010 IEEE Computer Society Annual Symposium on VLSI (ISVLSI). IEEE, 2010. http://dx.doi.org/10.1109/isvlsi.2010.106.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Shoushtari, Majid, Amir M. Rahmani, and Nikil Dutt. "Quality-configurable memory hierarchy through approximation." In ESWEEK'17: THIRTEENTH EMBEDDED SYSTEM WEEK. New York, NY, USA: ACM, 2017. http://dx.doi.org/10.1145/3125501.3125525.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Yotov, Kamen, Keshav Pingali, and Paul Stodghill. "Automatic measurement of memory hierarchy parameters." In the 2005 ACM SIGMETRICS international conference. New York, New York, USA: ACM Press, 2005. http://dx.doi.org/10.1145/1064212.1064233.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Memory hierarchy"

1

Соловйов, Володимир Миколайович, V. Saptsin, and D. Chabanenko. Markov chains applications to the financial-economic time series predictions. Transport and Telecommunication Institute, 2011. http://dx.doi.org/10.31812/0564/1189.

Full text
Abstract:
In this research the technology of complex Markov chains is applied to predict financial time series. The main distinction of complex or high-order Markov Chains and simple first-order ones is the existing of after-effect or memory. The technology proposes prediction with the hierarchy of time discretization intervals and splicing procedure for the prediction results at the different frequency levels to the single prediction output time series. The hierarchy of time discretizations gives a possibility to use fractal properties of the given time series to make prediction on the different frequencies of the series. The prediction results for world’s stock market indices are presented.
APA, Harvard, Vancouver, ISO, and other styles
2

Соловйов, Володимир Миколайович, Vladimir Saptsin, and Dmitry Chabanenko. Prediction of financial time series with the technology of high-order Markov chains. AGSOE, March 2009. http://dx.doi.org/10.31812/0564/1131.

Full text
Abstract:
In this research the technology of complex Markov chains, i.e. Markov chains with a memory is applied to forecast the financial time-series. The high-order Markov chains can be simplified to first-order ones by generalizing the states in Markov chains. Considering the *generalized state* as the sequence of states makes a possibility to model high-order Markov chains like first-order ones. The adaptive method of defining the states is proposed, it is concerned with the statistic properties of price returns. The algorithm of prediction includes the next steps: (1) Generate the hierarchical set of time discretizations; (2) Reducing the discretiza- tion of initial data and doing prediction at the every time-level (3) Recurrent conjunction of prediction series of different discretizations in a single time-series. The hierarchy of time discretizations gives a possibility to review long-memory properties of the series without increasing the order of the Markov chains, to make prediction on the different frequencies of the series. The technology is tested on several time-series, including: EUR/USD Forex course, the World’s indices, including Dow Jones, S&P 500, RTS, PFTS and other.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography