mirror of
				https://github.com/RGBCube/serenity
				synced 2025-10-31 04:32:44 +00:00 
			
		
		
		
	LibJS: Check the ASAN fake stack for heap pointers when ASAN is enabled
This is a similar strategy to what v8 does. Use the ASAN API function __asan_addr_is_in_fake_stack to check any fake stack frames associated with each stack address we scan. This fully allows running test-js -g with the option detect_stack_use_after_return turned on.
This commit is contained in:
		
							parent
							
								
									7d71acf1bb
								
							
						
					
					
						commit
						31440687a3
					
				
					 2 changed files with 49 additions and 20 deletions
				
			
		|  | @ -24,6 +24,10 @@ | ||||||
| #    include <serenity.h> | #    include <serenity.h> | ||||||
| #endif | #endif | ||||||
| 
 | 
 | ||||||
|  | #ifdef HAS_ADDRESS_SANITIZER | ||||||
|  | #    include <sanitizer/asan_interface.h> | ||||||
|  | #endif | ||||||
|  | 
 | ||||||
| namespace JS { | namespace JS { | ||||||
| 
 | 
 | ||||||
| #ifdef AK_OS_SERENITY | #ifdef AK_OS_SERENITY | ||||||
|  | @ -132,6 +136,46 @@ void Heap::gather_roots(HashTable<Cell*>& roots) | ||||||
|     } |     } | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | static void add_possible_value(HashTable<FlatPtr>& possible_pointers, FlatPtr data) | ||||||
|  | { | ||||||
|  |     if constexpr (sizeof(FlatPtr*) == sizeof(Value)) { | ||||||
|  |         // Because Value stores pointers in non-canonical form we have to check if the top bytes
 | ||||||
|  |         // match any pointer-backed tag, in that case we have to extract the pointer to its
 | ||||||
|  |         // canonical form and add that as a possible pointer.
 | ||||||
|  |         if ((data & SHIFTED_IS_CELL_PATTERN) == SHIFTED_IS_CELL_PATTERN) | ||||||
|  |             possible_pointers.set(Value::extract_pointer_bits(data)); | ||||||
|  |         else | ||||||
|  |             possible_pointers.set(data); | ||||||
|  |     } else { | ||||||
|  |         static_assert((sizeof(Value) % sizeof(FlatPtr*)) == 0); | ||||||
|  |         // In the 32-bit case we will look at the top and bottom part of Value separately we just
 | ||||||
|  |         // add both the upper and lower bytes as possible pointers.
 | ||||||
|  |         possible_pointers.set(data); | ||||||
|  |     } | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | #ifdef HAS_ADDRESS_SANITIZER | ||||||
|  | __attribute__((no_sanitize("address"))) void Heap::gather_asan_fake_stack_roots(HashTable<FlatPtr>& possible_pointers, FlatPtr addr) | ||||||
|  | { | ||||||
|  |     void* begin = nullptr; | ||||||
|  |     void* end = nullptr; | ||||||
|  |     void* real_stack = __asan_addr_is_in_fake_stack(__asan_get_current_fake_stack(), reinterpret_cast<void*>(addr), &begin, &end); | ||||||
|  | 
 | ||||||
|  |     if (real_stack != nullptr) { | ||||||
|  |         for (auto* real_stack_addr = reinterpret_cast<void const* const*>(begin); real_stack_addr < end; ++real_stack_addr) { | ||||||
|  |             void const* real_address = *real_stack_addr; | ||||||
|  |             if (real_address == nullptr) | ||||||
|  |                 continue; | ||||||
|  |             add_possible_value(possible_pointers, reinterpret_cast<FlatPtr>(real_address)); | ||||||
|  |         } | ||||||
|  |     } | ||||||
|  | } | ||||||
|  | #else | ||||||
|  | void Heap::gather_asan_fake_stack_roots(HashTable<FlatPtr>&, FlatPtr) | ||||||
|  | { | ||||||
|  | } | ||||||
|  | #endif | ||||||
|  | 
 | ||||||
| __attribute__((no_sanitize("address"))) void Heap::gather_conservative_roots(HashTable<Cell*>& roots) | __attribute__((no_sanitize("address"))) void Heap::gather_conservative_roots(HashTable<Cell*>& roots) | ||||||
| { | { | ||||||
|     FlatPtr dummy; |     FlatPtr dummy; | ||||||
|  | @ -145,32 +189,16 @@ __attribute__((no_sanitize("address"))) void Heap::gather_conservative_roots(Has | ||||||
| 
 | 
 | ||||||
|     auto* raw_jmp_buf = reinterpret_cast<FlatPtr const*>(buf); |     auto* raw_jmp_buf = reinterpret_cast<FlatPtr const*>(buf); | ||||||
| 
 | 
 | ||||||
|     auto add_possible_value = [&](FlatPtr data) { |  | ||||||
|         if constexpr (sizeof(FlatPtr*) == sizeof(Value)) { |  | ||||||
|             // Because Value stores pointers in non-canonical form we have to check if the top bytes
 |  | ||||||
|             // match any pointer-backed tag, in that case we have to extract the pointer to its
 |  | ||||||
|             // canonical form and add that as a possible pointer.
 |  | ||||||
|             if ((data & SHIFTED_IS_CELL_PATTERN) == SHIFTED_IS_CELL_PATTERN) |  | ||||||
|                 possible_pointers.set(Value::extract_pointer_bits(data)); |  | ||||||
|             else |  | ||||||
|                 possible_pointers.set(data); |  | ||||||
|         } else { |  | ||||||
|             static_assert((sizeof(Value) % sizeof(FlatPtr*)) == 0); |  | ||||||
|             // In the 32-bit case we will look at the top and bottom part of Value separately we just
 |  | ||||||
|             // add both the upper and lower bytes as possible pointers.
 |  | ||||||
|             possible_pointers.set(data); |  | ||||||
|         } |  | ||||||
|     }; |  | ||||||
| 
 |  | ||||||
|     for (size_t i = 0; i < ((size_t)sizeof(buf)) / sizeof(FlatPtr); ++i) |     for (size_t i = 0; i < ((size_t)sizeof(buf)) / sizeof(FlatPtr); ++i) | ||||||
|         add_possible_value(raw_jmp_buf[i]); |         add_possible_value(possible_pointers, raw_jmp_buf[i]); | ||||||
| 
 | 
 | ||||||
|     auto stack_reference = bit_cast<FlatPtr>(&dummy); |     auto stack_reference = bit_cast<FlatPtr>(&dummy); | ||||||
|     auto& stack_info = m_vm.stack_info(); |     auto& stack_info = m_vm.stack_info(); | ||||||
| 
 | 
 | ||||||
|     for (FlatPtr stack_address = stack_reference; stack_address < stack_info.top(); stack_address += sizeof(FlatPtr)) { |     for (FlatPtr stack_address = stack_reference; stack_address < stack_info.top(); stack_address += sizeof(FlatPtr)) { | ||||||
|         auto data = *reinterpret_cast<FlatPtr*>(stack_address); |         auto data = *reinterpret_cast<FlatPtr*>(stack_address); | ||||||
|         add_possible_value(data); |         add_possible_value(possible_pointers, data); | ||||||
|  |         gather_asan_fake_stack_roots(possible_pointers, data); | ||||||
|     } |     } | ||||||
| 
 | 
 | ||||||
|     // NOTE: If we have any custom ranges registered, scan those as well.
 |     // NOTE: If we have any custom ranges registered, scan those as well.
 | ||||||
|  | @ -178,7 +206,7 @@ __attribute__((no_sanitize("address"))) void Heap::gather_conservative_roots(Has | ||||||
|     if (s_custom_ranges_for_conservative_scan) { |     if (s_custom_ranges_for_conservative_scan) { | ||||||
|         for (auto& custom_range : *s_custom_ranges_for_conservative_scan) { |         for (auto& custom_range : *s_custom_ranges_for_conservative_scan) { | ||||||
|             for (size_t i = 0; i < (custom_range.value / sizeof(FlatPtr)); ++i) { |             for (size_t i = 0; i < (custom_range.value / sizeof(FlatPtr)); ++i) { | ||||||
|                 add_possible_value(custom_range.key[i]); |                 add_possible_value(possible_pointers, custom_range.key[i]); | ||||||
|             } |             } | ||||||
|         } |         } | ||||||
|     } |     } | ||||||
|  |  | ||||||
|  | @ -85,6 +85,7 @@ private: | ||||||
| 
 | 
 | ||||||
|     void gather_roots(HashTable<Cell*>&); |     void gather_roots(HashTable<Cell*>&); | ||||||
|     void gather_conservative_roots(HashTable<Cell*>&); |     void gather_conservative_roots(HashTable<Cell*>&); | ||||||
|  |     void gather_asan_fake_stack_roots(HashTable<FlatPtr>&, FlatPtr); | ||||||
|     void mark_live_cells(HashTable<Cell*> const& live_cells); |     void mark_live_cells(HashTable<Cell*> const& live_cells); | ||||||
|     void finalize_unmarked_cells(); |     void finalize_unmarked_cells(); | ||||||
|     void sweep_dead_cells(bool print_report, Core::ElapsedTimer const&); |     void sweep_dead_cells(bool print_report, Core::ElapsedTimer const&); | ||||||
|  |  | ||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue
	
	 Andrew Kaster
						Andrew Kaster