[libos] Fix munmap conflict chunk range and vma range
This commit is contained in:
		
							parent
							
								
									61b9639081
								
							
						
					
					
						commit
						171faccea7
					
				| @ -54,6 +54,7 @@ impl Eq for Chunk {} | ||||
| 
 | ||||
| impl Debug for Chunk { | ||||
|     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { | ||||
|         write!(f, "range = {:?}, ", self.range); | ||||
|         match self.internal() { | ||||
|             ChunkType::SingleVMA(vma) => write!(f, "Single VMA chunk: {:?}", vma), | ||||
|             ChunkType::MultiVMA(internal_manager) => write!(f, "default chunk: {:?}", self.range()), | ||||
|  | ||||
| @ -240,8 +240,8 @@ impl VMManager { | ||||
|             align_up(size, PAGE_SIZE) | ||||
|         }; | ||||
|         let munmap_range = { VMRange::new_with_size(addr, size) }?; | ||||
|         let chunk = { | ||||
|         let current = current!(); | ||||
|         let chunk = { | ||||
|             let process_mem_chunks = current.vm().mem_chunks().read().unwrap(); | ||||
|             let chunk = process_mem_chunks | ||||
|                 .iter() | ||||
| @ -260,8 +260,10 @@ impl VMManager { | ||||
|         // Case 1: the overlapping chunk IS NOT a super set of munmap range
 | ||||
|         if !chunk.range().is_superset_of(&munmap_range) { | ||||
|             // munmap range spans multiple chunks
 | ||||
| 
 | ||||
|             // Must lock the internal manager first here in case the chunk's range and vma are conflict when other threads are operating the VM
 | ||||
|             let mut internal_manager = self.internal(); | ||||
|             let overlapping_chunks = { | ||||
|                 let current = current!(); | ||||
|                 let process_mem_chunks = current.vm().mem_chunks().read().unwrap(); | ||||
|                 process_mem_chunks | ||||
|                     .iter() | ||||
| @ -273,8 +275,7 @@ impl VMManager { | ||||
|             for chunk in overlapping_chunks.iter() { | ||||
|                 match chunk.internal() { | ||||
|                     ChunkType::SingleVMA(_) => { | ||||
|                         let mut internl_manager = self.internal(); | ||||
|                         internl_manager.munmap_chunk(chunk, Some(&munmap_range))? | ||||
|                         internal_manager.munmap_chunk(chunk, Some(&munmap_range))? | ||||
|                     } | ||||
|                     ChunkType::MultiVMA(manager) => manager | ||||
|                         .lock() | ||||
| @ -297,8 +298,22 @@ impl VMManager { | ||||
|                     .munmap_range(munmap_range); | ||||
|             } | ||||
|             ChunkType::SingleVMA(_) => { | ||||
|                 // Single VMA Chunk could be updated during the release of internal manager lock. Get overlapping chunk again.
 | ||||
|                 // This is done here because we don't want to acquire the big internal manager lock as soon as entering this function.
 | ||||
|                 let mut internal_manager = self.internal(); | ||||
|                 return internal_manager.munmap_chunk(&chunk, Some(&munmap_range)); | ||||
|                 let overlapping_chunk = { | ||||
|                     let process_mem_chunks = current.vm().mem_chunks().read().unwrap(); | ||||
|                     process_mem_chunks | ||||
|                         .iter() | ||||
|                         .find(|&chunk| chunk.range().intersect(&munmap_range).is_some()) | ||||
|                         .map(|chunk| chunk.clone()) | ||||
|                 }; | ||||
|                 if let Some(overlapping_chunk) = overlapping_chunk { | ||||
|                     return internal_manager.munmap_chunk(&overlapping_chunk, Some(&munmap_range)); | ||||
|                 } else { | ||||
|                     warn!("no overlapping chunks anymore"); | ||||
|                     return Ok(()); | ||||
|                 } | ||||
|             } | ||||
|         } | ||||
|     } | ||||
|  | ||||
| @ -1,5 +1,5 @@ | ||||
| include ../test_common.mk | ||||
| 
 | ||||
| EXTRA_C_FLAGS := -Wno-return-stack-address | ||||
| EXTRA_LINK_FLAGS := | ||||
| EXTRA_LINK_FLAGS := -lpthread | ||||
| BIN_ARGS := | ||||
|  | ||||
| @ -10,6 +10,7 @@ | ||||
| #include <string.h> | ||||
| #include <fcntl.h> | ||||
| #include <sys/syscall.h> | ||||
| #include <pthread.h> | ||||
| #include "test_fs.h" | ||||
| 
 | ||||
| // ============================================================================
 | ||||
| @ -677,6 +678,78 @@ int test_munmap_whose_range_intersects_with_multiple_mmap_regions() { | ||||
|     return 0; | ||||
| } | ||||
| 
 | ||||
| 
 | ||||
| static void *child_routine(void *_arg) { | ||||
|     size_t len = 1 * MB; | ||||
|     void *old_addr = _arg; | ||||
|     void *new_addr = old_addr + 3 * len; | ||||
| 
 | ||||
|     void *addr = mremap(old_addr, len, 2 * len, MREMAP_MAYMOVE | MREMAP_FIXED, | ||||
|                         new_addr); | ||||
|     if (addr != new_addr) { | ||||
|         printf("mremap fail in client routine\n"); | ||||
|         return (void *) -1; | ||||
|     } | ||||
| 
 | ||||
|     int ret = munmap(addr, 2 * len); | ||||
|     if (ret < 0) { | ||||
|         printf("munmap failed"); | ||||
|         return (void *) -1; | ||||
|     } | ||||
| 
 | ||||
|     return NULL; | ||||
| } | ||||
| 
 | ||||
| int test_mremap_concurrent() { | ||||
|     int prot = PROT_READ | PROT_WRITE; | ||||
|     int flags = MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED; | ||||
|     size_t len = 1 * MB; | ||||
|     void *child_ret; | ||||
| 
 | ||||
|     // Allocate continuous single vma chunks at the end of the default chunk
 | ||||
|     size_t hint_1 = HINT_BEGIN + DEFAULT_CHUNK_SIZE; | ||||
|     void *addr = mmap((void *)hint_1, len, prot, flags, -1, 0); | ||||
|     if ((size_t)addr != hint_1) { | ||||
|         THROW_ERROR("fixed mmap with good hint failed"); | ||||
|     } | ||||
| 
 | ||||
|     size_t hint_2 = hint_1 + len; | ||||
|     addr = mmap((void *)hint_2, len, prot, flags, -1, 0); | ||||
|     if ((size_t)addr != hint_2) { | ||||
|         THROW_ERROR("fixed mmap spans over two chunks failed"); | ||||
|     } | ||||
| 
 | ||||
|     // create a child thread to mremap chunk from hint_2 to address after the chunk which starts with hint_3
 | ||||
|     pthread_t child_thread = 0; | ||||
|     int ret = pthread_create(&child_thread, NULL, child_routine, (void *)hint_2); | ||||
|     if (ret < 0) { | ||||
|         THROW_ERROR("pthread create failed"); | ||||
|     } | ||||
| 
 | ||||
|     // mremap chunk from hint_1 to hint_3
 | ||||
|     size_t hint_3 = hint_2 + len; | ||||
|     void *ret_addr = mremap((void *)hint_1, len, len * 2, MREMAP_MAYMOVE | MREMAP_FIXED, | ||||
|                             (void *)hint_3); | ||||
|     if (ret_addr != (void *)hint_3) { | ||||
|         THROW_ERROR("mremap failed"); | ||||
|     } | ||||
| 
 | ||||
|     ret = pthread_join(child_thread, &child_ret); | ||||
|     if (ret < 0 || child_ret != NULL) { | ||||
|         THROW_ERROR("pthread_join failed"); | ||||
|     } | ||||
| 
 | ||||
|     if (munmap((void *)hint_3, len * 2) < 0) { | ||||
|         THROW_ERROR("munmap failed"); | ||||
|     } | ||||
| 
 | ||||
|     if (check_buf_is_munmapped((void *)hint_1, len * 5) < 0) { | ||||
|         THROW_ERROR("munmap does not really free the memory"); | ||||
|     } | ||||
| 
 | ||||
|     return 0; | ||||
| } | ||||
| 
 | ||||
| int test_munmap_whose_range_intersects_with_several_chunks() { | ||||
|     int prot = PROT_READ | PROT_WRITE; | ||||
|     int flags = MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED; | ||||
| @ -1390,6 +1463,7 @@ static test_case_t test_cases[] = { | ||||
|     TEST_CASE(test_mprotect_with_non_page_aligned_size), | ||||
|     TEST_CASE(test_mprotect_multiple_vmas), | ||||
|     TEST_CASE(test_mprotect_grow_down), | ||||
|     TEST_CASE(test_mremap_concurrent), | ||||
| }; | ||||
| 
 | ||||
| int main() { | ||||
|  | ||||
		Loading…
	
		Reference in New Issue
	
	Block a user