PLDI 2022 Co-Located Events
43nd ACM SIGPLAN International Conference on Programming Language Design and Implementation (PLDI 2022)
Powered by
Conference Publishing Consulting

2022 ACM SIGPLAN International Symposium on Memory Management (ISMM 2022), June 14, 2022, San Diego, CA, USA

ISMM 2022 – Proceedings

Contents - Abstracts - Authors

2022 ACM SIGPLAN International Symposium on Memory Management (ISMM 2022)


Title Page

Welcome from the Chairs
With great pleasure, we welcome you to the ACM SIGPLAN International Symposium on Memory Management (ISMM) 2022. This is the 21st event in the ISMM series. As COVID restrictions ease, this event is in a hybrid format for the first time. We are pleased to see that our colleagues and friends in the memory management community remain active and productive in this challenging year.


Reconsidering OS Memory Optimizations in the Presence of Disaggregated Memory
Shai Bergman, Priyank Faldu, Boris GrotORCID logo, Lluís Vilanova ORCID logo, and Mark Silberstein ORCID logo
(Technion, Israel; ARM, UK; University of Edinburgh, UK; Imperial College London, UK)
Tiered memory systems introduce an additional memory level with higher-than-local-DRAM access latency and require sophisticated memory management mechanisms to achieve cost-efficiency and high performance. Recent works focus on byte-addressable tiered memory architectures which offer better performance than pure swap-based systems. We observe that adding disaggregation to a byte-addressable tiered memory architecture requires important design changes that deviate from the common techniques that target lower-latency non-volatile memory systems. Our comprehensive analysis of real workloads shows that the high access latency to disaggregated memory undermines the utility of well-established memory management optimizations Based on these insights, we develop HotBox – a disaggregated memory management subsystem for Linux that strives to maximize the local memory hit rate with low memory management overhead. HotBox introduces only minor changes to the Linux kernel while outperforming state-of-the-art systems on memory-intensive benchmarks by up to 2.25×.

Publisher's Version
MemSweeper: Virtualizing Cluster Memory Management for High Memory Utilization and Isolation
AmirHossein Seyri ORCID logo, Abhisek Pan, and Balajee Vamanan ORCID logo
(University of Illinois at Chicago, USA; Microsoft, USA)
Memory caches are critical components of modern web services that improve response times and reduce the load on backend databases. In multi-tenant clouds, several instances of caches compete for memory. The current state-of-the-art is to statically allocate memory for cache instances (e.g., based on cost-tier) but such allocation tends to be sub-optimal as memory demands of instances often vary with time and not known apriori. We propose MemSweeper, which dynamically manages memory between cache instances. MemSweeper uses a novel, score-based metric and an associated algorithm to identify cache instances whose working sets fit well within their allocated memory and thus can relinquish a portion of the memory without suffering appreciable loss in their hit rates. Using a combination of synthetic and production traces on a real implementation, we show that MemSweeper achieves 74% improvement (on average) in the miss rate of critical tenants without degrading the performance of other tenants.

Publisher's Version
Concurrent and Parallel Garbage Collection for Lightweight Threads on Multicore Processors
Katsuhiro Ueno ORCID logo and Atsushi Ohori
(Niigata University, Japan; Tohoku University, Japan)
This paper develops a concurrent and parallel garbage collection (GC) method that works with a lightweight thread library realizing the standard M:N threading model. The GC algorithm is organized as a set of procedures that are called from a user-level thread at various occasions and executed in the context of the OS-level threads owning the user-level thread. The procedures realize an on-the-fly collection that does not stop any thread. All OS-level threads cooperatively perform the collection in parallel. This construction achieves the same degree of parallelism as underlying lightweight thread scheduling. We have implemented the algorithm in a Standard ML compiler and have evaluated the performance with sequential and parallel benchmark programs. Our implementation shows good parallel scalability comparable to C programs directly using the lightweight threads library.

Publisher's Version
Replication-Based Object Persistence by Reachability
Kotaro Matsumoto ORCID logo, Tomoharu UgawaORCID logo, and Hideya Iwasaki ORCID logo
(Kochi University of Technology, Japan; University of Tokyo, Japan; University of Electro-Communications, Japan)
The emergence of non-volatile memory (NVM) presents opportunities for making in-memory data of application programs persistent at a small cost. An adequate abstraction is required for programming languages to be able to utilize NVM. Here, persistence by reachability is a suitable abstraction for managed languages. In this abstraction, all objects are volatile when they are created and become persistent later depending on their reachability from the predefined roots. The state-of-the-art in the implementations of persistence by reachability creates objects in DRAM and moves them to NVM when they become persistent. This implementation has two inefficiencies. One is the read barriers to get the current location of objects; the other is to read values of persistent objects from NVM, which is slower than DRAM.
This paper proposes a new algorithm to realize persistence by reachability. The proposed algorithm does not move objects to NVM to make them persistent. Rather, it creates replicas of the objects in NVM. After replication, the original copy in DRAM is kept synchronized with the replica. The program can obtain the contents of a persistent object by reading from DRAM without read barriers. We preliminarily implemented it in the HotSpot VM of OpenJDK and evaluated its overhead. The results showed that the overhead of making objects persistent was 2.7% on average. The overhead of writing to persistent objects varied from 0.4% to 335.5%, depending on write frequency. The overhead imposed on programs that do not make any object persistent was similar to the previous work.

Publisher's Version

proc time: 1.38