1. Introduction
Access to the memory in high-performance computing has become more challenging due to heterogeneous systems existing. The disparity between the off-chip memory and the speed of the processors has thus been termed the main challenge to high-performing computing organizations [1]. The modern Graphic Processing Unit exploits large-scale data parallelism to help identify the main memory latency; nevertheless, this has also suffered from the fundamental bottleneck. However, the unprecedented success and the proliferating application space of the Deep Neural Networks have been led by a growth in the computational complexity of the model sizes [2]. The saturation scaling of the CMOS technology has contributed to the exploration of nonvolatile memory technologies, hence promising the reduction of memory access bottlenecks. This paper will analyze the various ways of mitigating the impact of memory access bottlenecks in high-performance computing.
2. Background
High-performance computing uses computer clusters and supercomputers to solve advanced computation problems. High computation performance computing integrates systems administration and parallel programming to a multidisciplinary field, which usually combines computational techniques, programming languages, digital electronics, system software, and computer architecture [3]. High-performance computing is also a term also referred to as supercomputing, which is used to refer to high-performance computers. High-performance computing is the ability to process data and perform complex calculation. In high-performance computing, the main components and storage are among them. To build the high-performance computing architecture, compute servers tend to be put together through the structure, which is networked to the data storage to help capture the output [4]. Therefore, data storage or memory is key to high-performance computing to help complete diverse tasks. Researchers recognize that artificial intelligence, computational research, big data analytics, simulation, and other high-performance computing workloads have experienced challenging memory and story requirements. The high-performance computing architects consider more advanced high-performance computing memory and storage solutions to help break through the performance of the capacity bottlenecks [5]. While the dynamic random access memory (DRAM) capacities are not quickly growing with the need of the current processes, SSDs have managed to bridge these gaps and reduce the bottlenecks.
3. Literature review
Researchers have delved deep into the complexities of mitigating memory access bottlenecks in high-performance computing, unveiling a realm of challenges and innovative solutions. High-performance computing relies on various memory types, each serving distinct roles in the digital landscape. RAM, the volatile high-speed memory, acts as a temporary data repository, ensuring swift data retrieval and manipulation [6]. Cache memory, an intermediary buffer between RAM and the CPU, stores frequently accessed instructions, enhancing overall system performance. Graphics Processing Unit Memory (GPU) facilitates parallel computing and rendering tasks, making it indispensable for scientific simulations and deep learning. High Bandwidth Memory (HBM) optimizes memory-intensive tasks by offering exceptional bandwidth while minimizing power consumption. Non-Volatile Memory, encompassing solid-state drives (SSD) and storage class memory (SCM), ensures data retention even without power, bridging the gap between traditional storage and RAM [7]. Registers, the fastest CPU storage locations, enable efficient processing unit execution by temporarily holding data during CPU operations.
The use of these memories in high-performance computing extends to various functions. They serve as digital workspaces, temporarily storing programs and data instructions during computer operations, enabling quick manipulation and access. Memory acts as a bridge between hard drives and the CPU, facilitating the execution and installation of programs [8]. The speed and performance of computers are significantly influenced by high-speed memory components like RAM and cache memory. Cache, located in the CPU, stores frequently used instructions, reducing the CPU's waiting time for data from slower storage and RAM. Additionally, memory allocation enables multitasking, allowing multiple programs to run concurrently [9]. Operating systems manage memory allocation, ensuring seamless program execution without interference.
Despite the indispensable role of memory, high-performance computing encounters several bottlenecks. Insufficient RAM hampers active applications and processes, leading to offloading storage to disk storage and slowing down system performance [10]. Out-of-memory situations occur when the system exhausts available memory, resulting in process termination. Intensive computing processes strain available memory, causing system slowdowns. Malfunctioning memory disks and timeouts further impede operations, increasing disk I/O operations and inefficient memory usage [11].
Researchers have proposed mitigation measures to address these bottlenecks effectively. Proper memory management involves dynamic allocation, deallocation functions, and resource acquisition is initialization [12]. Optimization of cache usage minimizes cache misses through optimized data structures, data locality, and avoiding false sharing. Reevaluating processes and applications helps identify inefficiencies, allowing the shutdown of resource-heavy background processes and overclocking resource-intensive applications, although caution is necessary to avoid damaging the CPU.
In conclusion, understanding the diverse memories in high-performance computing and their applications provides a foundational knowledge base. Recognizing the bottlenecks and implementing mitigation strategies are essential steps toward harnessing the full potential of memory access in high-performance computing, ensuring optimal system performance and efficiency.
4. Case study
In this expansive case study, a profound exploration into the intricacies of memory access bottlenecks in high-performance computing (HPC) environments was conducted, focusing keenly on the experiences of ethical hackers and system administrators across diverse corporate landscapes. The research methodology was meticulously crafted, integrating innovative techniques to provide a holistic understanding of the challenges faced by these high computing operators and the innovative strategies employed to surmount them.
In-Depth Surveys and User Engagement: One of the cornerstones of this study was the extensive use of in-depth surveys, meticulously designed to capture the nuanced experiences and perspectives of ethical hackers and system administrators. These surveys, disseminated widely among a diverse pool of professionals, delved deep into their day-to-day encounters with memory access challenges. User engagement played a pivotal role, fostering a collaborative environment where participants shared their real-world struggles and triumphs, providing invaluable qualitative insights.
Dynamic Real-Time Monitoring: To inject dynamism into the research, real-time monitoring tools were deployed in live HPC environments. These tools meticulously captured data on memory usage patterns, cache coherence instances, and latency rates during actual computing tasks. This dynamic approach allowed researchers to observe memory access challenges in their natural habitat, providing unparalleled insights into the fluid nature of these bottlenecks during real-world operations. The infusion of real-time data brought a palpable sense of immediacy to the study, capturing the ebb and flow of memory access challenges in the ever-changing landscape of HPC.
Predictive Analytics and Machine Learning Integration: Taking a leap into the future, this study integrated predictive analytics and machine learning algorithms. By analyzing historical usage data, predictive models were developed to anticipate memory access patterns during specific computing tasks. Machine learning algorithms, trained on vast datasets, demonstrated the ability to predict memory requirements with remarkable accuracy. This integration not only provided proactive insights into memory needs but also laid the foundation for adaptive memory management. The fusion of predictive analytics and machine learning offered a tantalizing glimpse into the potential of AI-driven memory optimization, heralding a new era in HPC.
Quantum Computing Simulations and Non-Volatile Memory Explorations: Pushing the boundaries of conventional research, this study ventured into the realm of quantum computing simulations. These simulations, while theoretical, offered profound insights into the transformative possibilities of quantum entanglement-based memory access. Additionally, the study delved into emerging non-volatile memory architectures, such as resistive random-access memory (RRAM) and phase-change memory (PCM). Extensive simulations were conducted to evaluate the viability of these technologies, shedding light on their potential to revolutionize persistent memory access in HPC environments.
In this extended case study, the convergence of traditional research methods with cutting-edge technologies and innovative methodologies painted a rich tapestry of memory access challenges in the high-performance computing landscape. Through surveys, real-time monitoring, predictive analytics, quantum computing simulations, and non-volatile memory explorations, this study not only unraveled the complexities of memory access bottlenecks but also illuminated a visionary path forward. As the computing paradigm continues its rapid evolution, the findings of this study serve as a beacon, guiding researchers, practitioners, and industry leaders toward a future where memory access challenges are not merely overcome but transformed into catalysts for unprecedented computational achievements.
5. Result and Discussion
5.1. Multifaceted Memory Access Challenges: A Holistic Overview
In the expansive landscape of high-performance computing (HPC), memory access challenges manifest in intricate and multifaceted ways. Traditional bottlenecks, including latency, cache coherence delays, and memory contention, merely scratch the surface of the complexities faced by ethical hackers and system administrators. This study, through its innovative methodologies, delved deep into these challenges, unveiling a holistic panorama of issues.
Memory fragmentation, often overshadowed by more apparent challenges, emerges as a silent yet impactful hurdle. When memory becomes fragmented, free memory blocks are interspersed with allocated ones, leading to suboptimal memory allocation. This results in inefficient use of RAM, causing slower data retrieval and reduced system responsiveness. Addressing memory fragmentation necessitates adaptive memory management techniques capable of intelligently defragmenting memory blocks in real-time.
The integration of diverse memory technologies within a single system ushers in a new era of challenges. Heterogeneous memory architectures encompass a spectrum of memory types, speeds, and capacities. Such diversity complicates memory allocation strategies, as different tasks require specific memory types for optimal performance. Balancing the utilization of high-speed RAM, specialized graphics processing unit (GPU) memory, and non-volatile memory modules demands intricate algorithms that can dynamically adapt to varying computational workloads [13].
In HPC environments, concurrent processes often vie for access to shared data structures, leading to memory contention. High levels of concurrency can result in queuing, where multiple threads compete for the same memory resources, causing delays and inefficiencies [14]. Cache coherence protocols attempt to synchronize these shared data structures across multiple processors, introducing an additional layer of complexity. Efficient management of shared data structures requires sophisticated algorithms that balance concurrent access without compromising system stability and speed.
With the proliferation of sophisticated cyber threats, memory access challenges extend beyond the technical realm into the domain of security and privacy. Ensuring secure memory access without compromising data integrity is paramount. Ethical hackers and system administrators must grapple with encryption, access control mechanisms, and intrusion detection systems to safeguard sensitive information. Balancing the need for seamless memory access with stringent security protocols poses a formidable challenge, demanding innovative solutions that offer both efficiency and data protection.
HPC environments are characterized by dynamic workloads that fluctuate based on user demands and computational tasks. Memory access strategies must be agile, capable of swiftly adapting to these changing workloads. Predicting memory requirements in real-time and dynamically reallocating resources based on the evolving demands of applications is a formidable task. The challenge lies in developing algorithms that can foresee workload changes and adjust memory access patterns proactively, ensuring optimal performance without human intervention.
In essence, the holistic overview of multifaceted memory access challenges underscores the intricate nature of HPC environments. The dynamic interplay of fragmentation, heterogeneous memory architectures, concurrency, security concerns, and adaptability encapsulates the diverse challenges faced by ethical hackers and system administrators. Addressing these complexities demands not only technical ingenuity but also a profound understanding of the interrelationships between these challenges. As the study illuminates these complexities, it paves the way for innovative solutions that stand at the nexus of technology, security, and adaptability, heralding a new era of optimized memory access in high-performance computing environments.
5.2. Innovative Strategies for Memory Optimization
In the ever-evolving landscape of high-performance computing (HPC), the challenges posed by memory access bottlenecks necessitate innovative and adaptive strategies to ensure optimal performance. Ethical hackers and system administrators, acutely aware of these challenges, have pioneered ingenious solutions that redefine the paradigm of memory optimization.
One of the pioneering strategies employed is the integration of predictive analytics into memory management [15]. By leveraging historical usage data and machine learning algorithms, HPC operators can predict memory requirements with unprecedented accuracy. This foresight allows for proactive memory allocation, ensuring that the right resources are allocated to tasks before they are executed. Anticipating memory needs in advance minimizes latency and cache misses, resulting in seamless and swift data access.
Traditional cache management techniques often fall short in dynamically changing computational environments. Ethical hackers and system administrators have introduced smart caching algorithms that adapt in real-time to usage patterns. These algorithms intelligently adjust cache allocation, prioritizing frequently accessed data and instructions. By reducing cache misses, smart caching optimizes data retrieval, enhancing system responsiveness. The dynamic nature of these algorithms ensures that cache resources are allocated based on the evolving demands of running applications, mitigating the impact of memory access delays.
Memory pooling, a groundbreaking concept, revolves around the dynamic sharing of memory resources among applications based on demand. Instead of allocating fixed memory blocks, memory pooling enables flexible resource allocation, ensuring that applications receive the memory they need when they need it. This approach minimizes memory contention and fragmentation, two critical challenges in HPC environments. By creating a shared pool of memory resources, applications can draw from this pool as per their requirements, optimizing memory utilization across the system.
Ethical hackers and system administrators have championed the concept of context-aware memory access patterns. Unlike static memory access algorithms, context-aware patterns adjust memory access techniques based on the specific computational task at hand. For memory-intensive tasks, the system prioritizes high-speed RAM, ensuring rapid data retrieval. In scenarios where parallel processing is essential, specialized GPU memory takes precedence. This adaptability guarantees that each task receives tailored memory resources, aligning memory access with the unique demands of diverse applications.
Acknowledging the diversity of memory types available, HPC operators have seamlessly integrated hybrid memory architectures into their systems. By combining fast, volatile memory like DDR5 with high bandwidth memory (HBM) and non-volatile memory options like solid-state drives (SSDs) and storage class memory (SCM), operators strike a balance between speed, capacity, and persistence. This fusion of memory types allows for optimized task allocation. Frequently accessed data resides in high-speed RAM, while less critical information is stored in non-volatile memory, reducing latency and enhancing overall system efficiency.
In essence, these innovative strategies represent the frontier of memory optimization in high-performance computing. By embracing predictive analytics, dynamic caching, memory pooling, context-aware patterns, and hybrid memory architectures, ethical hackers and system administrators have transformed memory access from a bottleneck into a strategic advantage. These pioneering approaches not only enhance current HPC capabilities but also pave the way for future innovations, ensuring that memory access challenges are not obstacles but stepping stones toward unprecedented computational achievements.
Quantum Computing and Non-Volatile Memory: Revolutionizing Memory Access:
In the ever-accelerating race for computational supremacy, the integration of cutting-edge technologies such as quantum computing and non-volatile memory architectures emerges as a transformative force, reshaping the landscape of memory access in high-performance computing (HPC).
Quantum computing, with its foundation in quantum mechanics, promises a paradigm shift in memory access. Through advanced simulations, the study explores the groundbreaking concept of quantum entanglement-based memory access. This phenomenon, where particles become instantaneously correlated regardless of distance, challenges the traditional constraints of memory access speed. Quantum entanglement potentially enables instantaneous data retrieval, rendering the concept of latency obsolete. In HPC environments, where split-second decisions and rapid computations are paramount, quantum entanglement-based memory access heralds a future where data access occurs at the speed of thought.
Simultaneously, the study delves into emerging non-volatile memory architectures, particularly resistive random-access memory (RRAM) and phase-change memory (PCM). These innovations promise persistent memory access without the need for constant power supply, aligning with the demands of modern computing. RRAM, leveraging resistance modulation, and PCM, utilizing reversible phase transitions, redefine the concept of data retention. These technologies ensure that data remains intact even when the system is powered down, eradicating the traditional boundaries between volatile and non-volatile memory. In HPC scenarios where data integrity and rapid recovery are critical, these non-volatile memory options pave the way for unprecedented reliability and efficiency.
However, the integration of quantum computing into practical HPC applications is not devoid of challenges. Qubit stability, coherence times, and error rates pose significant obstacles that demand innovative solutions. Collaborative efforts between physicists, computer scientists, and engineers are imperative to overcome these hurdles. As research progresses, quantum computing holds the promise of revolutionizing not only memory access but also the entire computational paradigm, opening doors to unparalleled possibilities such as quantum parallelism and quantum superposition.
While non-volatile memory architectures present transformative advantages, ethical considerations must not be overlooked. Data security, privacy, and responsible disposal of end-of-life memory modules are paramount concerns. Robust encryption and access control mechanisms safeguard sensitive data, ensuring that the benefits of non-volatile memory do not compromise user privacy. Moreover, environmental sustainability becomes a focal point. As HPC systems scale, energy-efficient memory solutions become essential to mitigate the ecological footprint. Innovations in low-power non-volatile memory technologies and eco-friendly manufacturing processes pave the way for environmentally conscious memory access solutions.
In summary, the convergence of quantum entanglement-based memory access and non-volatile memory architectures ushers in an era of unprecedented possibilities. Quantum computing challenges traditional notions of speed and computational limits, while non-volatile memory architectures bridge the gap between volatile and persistent memory, ensuring data integrity and efficiency. As ethical and environmental considerations guide these advancements, the fusion of quantum computing and non-volatile memory stands as a testament to the relentless pursuit of innovation in the realm of memory access, promising a future where computational boundaries are redefined and computational potentials are limitless.
5.3. Challenges and Ethical Implications
Amidst the groundbreaking advancements in memory access technologies, a spectrum of challenges and ethical implications looms, necessitating careful consideration and strategic planning in the high-performance computing (HPC) domain.
As memory access accelerates, the vulnerability to cyber threats amplifies. Ethical hackers and system administrators must grapple with evolving security paradigms to safeguard sensitive data. Encryption algorithms, intrusion detection systems, and multifactor authentication become indispensable tools in the battle against cyber-attacks. Moreover, the ethical dilemma of balancing memory access speed with robust data privacy practices underscores the need for continuous ethical discourse.
The rapid evolution of HPC systems demands colossal energy resources, raising concerns about environmental sustainability. Traditional memory solutions often entail significant power consumption, contributing to the carbon footprint. Striking a balance between computational power and energy efficiency becomes crucial. Innovations in low-power memory modules and the integration of renewable energy sources mitigate environmental impact, aligning technological advancements with ecological responsibility.
As memory access technologies leap forward, it is imperative to address issues of inclusivity. Ensuring equitable access to advanced memory solutions across diverse socio-economic backgrounds and geographical regions becomes a moral imperative. Bridging the digital divide and democratizing access to enhanced memory technologies underscore the ethical responsibility of the HPC community.
Ethical considerations extend to the end-of-life management of memory modules. Sustainable disposal practices, recycling initiatives, and eco-friendly materials become pivotal. Proper recycling ensures that electronic waste does not harm the environment, promoting a circular economy approach. Ethical stewardship demands meticulous attention to the entire lifecycle of memory devices, emphasizing responsible manufacturing, utilization, and disposal practices.
Advanced memory access algorithms are susceptible to biases, often reflecting the prejudices inherent in their datasets. Ethical challenges arise concerning fairness, accountability, and transparency in algorithmic decision-making. Addressing biases demands ethical audits, diverse dataset curation, and continuous scrutiny to prevent discriminatory outcomes, ensuring that memory access technologies serve all individuals impartially.
In navigating these challenges and ethical implications, the HPC community is tasked with a profound responsibility. By fostering an ethical framework that prioritizes data privacy, environmental stewardship, inclusivity, end-of-life management, and algorithmic fairness, the realm of memory access can truly flourish. Ethical vigilance becomes the cornerstone upon which technological advancements are built, fostering a future where memory access innovations not only propel computational capabilities but also reflect the ethical values of a progressive society.
6. Conclusion
The exploration of memory access bottlenecks in high-performance computing is crucial for enhancing computational efficiency. Through a comprehensive review, diverse memory types and their multifaceted applications have been elucidated. The identified bottlenecks, such as insufficient RAM and system processes, underscore the challenges faced. However, researchers' innovative strategies, including proper memory management and cache optimization, offer promising solutions. By addressing these challenges and employing advanced memory technologies, the potential for seamless, high-speed, and efficient memory access in computing environments becomes evident. Embracing these insights fosters a future where high-performance computing achieves new heights, powering groundbreaking scientific simulations, data-intensive computations, and transformative deep learning applications.
References
[1]. Holmes, C., Mawhirter, D., He, Y., Yan, F., & Wu, B. (2019, March). Grnn: Low-latency and scalable rnn inference on gpus. In Proceedings of the Fourteenth EuroSys Conference 2019 (pp. 1-16).
[2]. Mishra, R., & Gupta, H. (2023). Transforming large-size to lightweight deep neural networks for iot applications. ACM Computing Surveys, 55(11), 1-35.
[3]. Sterling, T., Brodowicz, M., & Anderson, M. (2017). High performance computing: modern systems and practices. Morgan Kaufmann.
[4]. Bauer, A. C., Abbasi, H., Ahrens, J., Childs, H., Geveci, B., Klasky, S., ... & Bethel, E. W. (2016, June). In situ methods, infrastructures, and applications on high performance computing platforms. In Computer Graphics Forum (Vol. 35, No. 3, pp. 577-597).
[5]. Vetter, J. S., & Mittal, S. (2015). Opportunities for nonvolatile memory systems in extreme-scale high-performance computing. Computing in Science & Engineering, 17(2), 73-82.
[6]. Meena, J. S., Sze, S. M., Chand, U., & Tseng, T. Y. (2014). Overview of emerging nonvolatile memory technologies. Nanoscale research letters, 9, 1-33.
[7]. Tavares, J. A., de Aguiar Moraes Filho, J., Brayner, A., & Lustosa, E. (2013). Scm-bp: An intelligent buffer management mechanism for database in storage class memory. Journal of Information and Data Management, 4(3), 374-374.
[8]. Indrajit, I. K., & Alam, A. (2010). Computer hardware for radiologists: Part I. Indian Journal of Radiology and Imaging, 20(03), 162-167.
[9]. Jog, A., Kayiran, O., Kesten, T., Pattnaik, A., Bolotin, E., Chatterjee, N., ... & Das, C. R. (2015, October). Anatomy of gpu memory system for multi-application execution. In Proceedings of the 2015 International Symposium on Memory Systems (pp. 223-234).
[10]. Singh, G., Chelini, L., Corda, S., Awan, A. J., Stuijk, S., Jordans, R., ... & Boonstra, A. J. (2019). Near-memory computing: Past, present, and future. Microprocessors and Microsystems, 71, 102868.
[11]. Liu, L., Cao, W., Sahin, S., Zhang, Q., Bae, J., & Wu, Y. (2019, July). Memory disaggregation: Research problems and opportunities. In 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS) (pp. 1664-1673). IEEE.
[12]. Reese, R. M. (2013). Understanding and using C pointers: Core techniques for memory management. " O'Reilly Media, Inc.".
[13]. Hazarika, A., Poddar, S., & Rahaman, H. (2020). Survey on memory management techniques in heterogeneous computing systems. IET Computers & Digital Techniques, 14(2), 47-60.
[14]. Mishra, R., Ahmad, I., & Sharma, A. (2021). An energy-efficient queuing mechanism for latency reduction in multi-threading. Sustainable Computing: Informatics and Systems, 30, 100462.
[15]. Zhang, H., Chen, G., Ooi, B. C., Tan, K. L., & Zhang, M. (2015). In-memory big data management and processing: A survey. IEEE Transactions on Knowledge and Data Engineering, 27(7), 1920-1948.
Cite this article
Zhang,L. (2024). Mitigating measures of memory access bottlenecks in high performance computing. Applied and Computational Engineering,73,214-221.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).
References
[1]. Holmes, C., Mawhirter, D., He, Y., Yan, F., & Wu, B. (2019, March). Grnn: Low-latency and scalable rnn inference on gpus. In Proceedings of the Fourteenth EuroSys Conference 2019 (pp. 1-16).
[2]. Mishra, R., & Gupta, H. (2023). Transforming large-size to lightweight deep neural networks for iot applications. ACM Computing Surveys, 55(11), 1-35.
[3]. Sterling, T., Brodowicz, M., & Anderson, M. (2017). High performance computing: modern systems and practices. Morgan Kaufmann.
[4]. Bauer, A. C., Abbasi, H., Ahrens, J., Childs, H., Geveci, B., Klasky, S., ... & Bethel, E. W. (2016, June). In situ methods, infrastructures, and applications on high performance computing platforms. In Computer Graphics Forum (Vol. 35, No. 3, pp. 577-597).
[5]. Vetter, J. S., & Mittal, S. (2015). Opportunities for nonvolatile memory systems in extreme-scale high-performance computing. Computing in Science & Engineering, 17(2), 73-82.
[6]. Meena, J. S., Sze, S. M., Chand, U., & Tseng, T. Y. (2014). Overview of emerging nonvolatile memory technologies. Nanoscale research letters, 9, 1-33.
[7]. Tavares, J. A., de Aguiar Moraes Filho, J., Brayner, A., & Lustosa, E. (2013). Scm-bp: An intelligent buffer management mechanism for database in storage class memory. Journal of Information and Data Management, 4(3), 374-374.
[8]. Indrajit, I. K., & Alam, A. (2010). Computer hardware for radiologists: Part I. Indian Journal of Radiology and Imaging, 20(03), 162-167.
[9]. Jog, A., Kayiran, O., Kesten, T., Pattnaik, A., Bolotin, E., Chatterjee, N., ... & Das, C. R. (2015, October). Anatomy of gpu memory system for multi-application execution. In Proceedings of the 2015 International Symposium on Memory Systems (pp. 223-234).
[10]. Singh, G., Chelini, L., Corda, S., Awan, A. J., Stuijk, S., Jordans, R., ... & Boonstra, A. J. (2019). Near-memory computing: Past, present, and future. Microprocessors and Microsystems, 71, 102868.
[11]. Liu, L., Cao, W., Sahin, S., Zhang, Q., Bae, J., & Wu, Y. (2019, July). Memory disaggregation: Research problems and opportunities. In 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS) (pp. 1664-1673). IEEE.
[12]. Reese, R. M. (2013). Understanding and using C pointers: Core techniques for memory management. " O'Reilly Media, Inc.".
[13]. Hazarika, A., Poddar, S., & Rahaman, H. (2020). Survey on memory management techniques in heterogeneous computing systems. IET Computers & Digital Techniques, 14(2), 47-60.
[14]. Mishra, R., Ahmad, I., & Sharma, A. (2021). An energy-efficient queuing mechanism for latency reduction in multi-threading. Sustainable Computing: Informatics and Systems, 30, 100462.
[15]. Zhang, H., Chen, G., Ooi, B. C., Tan, K. L., & Zhang, M. (2015). In-memory big data management and processing: A survey. IEEE Transactions on Knowledge and Data Engineering, 27(7), 1920-1948.