mirror of
				https://github.com/mborgerson/xemu.git
				synced 2025-10-29 22:51:05 +00:00 
			
		
		
		
	Use uint8_t for (unsigned) byte, and uint16_t for (unsigned) 16-bit word. Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com> Message-Id: <20210518183655.1711377-4-philmd@redhat.com> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
		
			
				
	
	
		
			534 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
			
		
		
	
	
			534 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
| /*
 | |
|  *  Physical memory access templates
 | |
|  *
 | |
|  *  Copyright (c) 2003 Fabrice Bellard
 | |
|  *  Copyright (c) 2015 Linaro, Inc.
 | |
|  *  Copyright (c) 2016 Red Hat, Inc.
 | |
|  *
 | |
|  * This library is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU Lesser General Public
 | |
|  * License as published by the Free Software Foundation; either
 | |
|  * version 2.1 of the License, or (at your option) any later version.
 | |
|  *
 | |
|  * This library is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|  * Lesser General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU Lesser General Public
 | |
|  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
 | |
|  */
 | |
| 
 | |
| /* warning: addr must be aligned */
 | |
| static inline uint32_t glue(address_space_ldl_internal, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result,
 | |
|     enum device_endian endian)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     uint64_t val;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 4;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, false, attrs);
 | |
|     if (l < 4 || !memory_access_is_direct(mr, false)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
| 
 | |
|         /* I/O case */
 | |
|         r = memory_region_dispatch_read(mr, addr1, &val,
 | |
|                                         MO_32 | devend_memop(endian), attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         fuzz_dma_read_cb(addr, 4, mr);
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         switch (endian) {
 | |
|         case DEVICE_LITTLE_ENDIAN:
 | |
|             val = ldl_le_p(ptr);
 | |
|             break;
 | |
|         case DEVICE_BIG_ENDIAN:
 | |
|             val = ldl_be_p(ptr);
 | |
|             break;
 | |
|         default:
 | |
|             val = ldl_p(ptr);
 | |
|             break;
 | |
|         }
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
|     return val;
 | |
| }
 | |
| 
 | |
| uint32_t glue(address_space_ldl, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_ldl_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                     DEVICE_NATIVE_ENDIAN);
 | |
| }
 | |
| 
 | |
| uint32_t glue(address_space_ldl_le, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_ldl_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                     DEVICE_LITTLE_ENDIAN);
 | |
| }
 | |
| 
 | |
| uint32_t glue(address_space_ldl_be, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_ldl_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                     DEVICE_BIG_ENDIAN);
 | |
| }
 | |
| 
 | |
| /* warning: addr must be aligned */
 | |
| static inline uint64_t glue(address_space_ldq_internal, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result,
 | |
|     enum device_endian endian)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     uint64_t val;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 8;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, false, attrs);
 | |
|     if (l < 8 || !memory_access_is_direct(mr, false)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
| 
 | |
|         /* I/O case */
 | |
|         r = memory_region_dispatch_read(mr, addr1, &val,
 | |
|                                         MO_64 | devend_memop(endian), attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         fuzz_dma_read_cb(addr, 8, mr);
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         switch (endian) {
 | |
|         case DEVICE_LITTLE_ENDIAN:
 | |
|             val = ldq_le_p(ptr);
 | |
|             break;
 | |
|         case DEVICE_BIG_ENDIAN:
 | |
|             val = ldq_be_p(ptr);
 | |
|             break;
 | |
|         default:
 | |
|             val = ldq_p(ptr);
 | |
|             break;
 | |
|         }
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
|     return val;
 | |
| }
 | |
| 
 | |
| uint64_t glue(address_space_ldq, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_ldq_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                     DEVICE_NATIVE_ENDIAN);
 | |
| }
 | |
| 
 | |
| uint64_t glue(address_space_ldq_le, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_ldq_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                     DEVICE_LITTLE_ENDIAN);
 | |
| }
 | |
| 
 | |
| uint64_t glue(address_space_ldq_be, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_ldq_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                     DEVICE_BIG_ENDIAN);
 | |
| }
 | |
| 
 | |
| uint8_t glue(address_space_ldub, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     uint64_t val;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 1;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, false, attrs);
 | |
|     if (!memory_access_is_direct(mr, false)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
| 
 | |
|         /* I/O case */
 | |
|         r = memory_region_dispatch_read(mr, addr1, &val, MO_8, attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         fuzz_dma_read_cb(addr, 1, mr);
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         val = ldub_p(ptr);
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
|     return val;
 | |
| }
 | |
| 
 | |
| /* warning: addr must be aligned */
 | |
| static inline uint16_t glue(address_space_lduw_internal, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result,
 | |
|     enum device_endian endian)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     uint64_t val;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 2;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, false, attrs);
 | |
|     if (l < 2 || !memory_access_is_direct(mr, false)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
| 
 | |
|         /* I/O case */
 | |
|         r = memory_region_dispatch_read(mr, addr1, &val,
 | |
|                                         MO_16 | devend_memop(endian), attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         fuzz_dma_read_cb(addr, 2, mr);
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         switch (endian) {
 | |
|         case DEVICE_LITTLE_ENDIAN:
 | |
|             val = lduw_le_p(ptr);
 | |
|             break;
 | |
|         case DEVICE_BIG_ENDIAN:
 | |
|             val = lduw_be_p(ptr);
 | |
|             break;
 | |
|         default:
 | |
|             val = lduw_p(ptr);
 | |
|             break;
 | |
|         }
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
|     return val;
 | |
| }
 | |
| 
 | |
| uint16_t glue(address_space_lduw, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_lduw_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                      DEVICE_NATIVE_ENDIAN);
 | |
| }
 | |
| 
 | |
| uint16_t glue(address_space_lduw_le, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_lduw_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                                      DEVICE_LITTLE_ENDIAN);
 | |
| }
 | |
| 
 | |
| uint16_t glue(address_space_lduw_be, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     return glue(address_space_lduw_internal, SUFFIX)(ARG1, addr, attrs, result,
 | |
|                                        DEVICE_BIG_ENDIAN);
 | |
| }
 | |
| 
 | |
| /* warning: addr must be aligned. The ram page is not masked as dirty
 | |
|    and the code inside is not invalidated. It is useful if the dirty
 | |
|    bits are used to track modified PTEs */
 | |
| void glue(address_space_stl_notdirty, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 4;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     uint8_t dirty_log_mask;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, true, attrs);
 | |
|     if (l < 4 || !memory_access_is_direct(mr, true)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
| 
 | |
|         r = memory_region_dispatch_write(mr, addr1, val, MO_32, attrs);
 | |
|     } else {
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         stl_p(ptr, val);
 | |
| 
 | |
|         dirty_log_mask = memory_region_get_dirty_log_mask(mr);
 | |
|         dirty_log_mask &= ~(1 << DIRTY_MEMORY_CODE);
 | |
|         cpu_physical_memory_set_dirty_range(memory_region_get_ram_addr(mr) + addr,
 | |
|                                             4, dirty_log_mask);
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
| }
 | |
| 
 | |
| /* warning: addr must be aligned */
 | |
| static inline void glue(address_space_stl_internal, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint32_t val, MemTxAttrs attrs,
 | |
|     MemTxResult *result, enum device_endian endian)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 4;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, true, attrs);
 | |
|     if (l < 4 || !memory_access_is_direct(mr, true)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
|         r = memory_region_dispatch_write(mr, addr1, val,
 | |
|                                          MO_32 | devend_memop(endian), attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         switch (endian) {
 | |
|         case DEVICE_LITTLE_ENDIAN:
 | |
|             stl_le_p(ptr, val);
 | |
|             break;
 | |
|         case DEVICE_BIG_ENDIAN:
 | |
|             stl_be_p(ptr, val);
 | |
|             break;
 | |
|         default:
 | |
|             stl_p(ptr, val);
 | |
|             break;
 | |
|         }
 | |
|         invalidate_and_set_dirty(mr, addr1, 4);
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
| }
 | |
| 
 | |
| void glue(address_space_stl, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stl_internal, SUFFIX)(ARG1, addr, val, attrs,
 | |
|                                              result, DEVICE_NATIVE_ENDIAN);
 | |
| }
 | |
| 
 | |
| void glue(address_space_stl_le, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stl_internal, SUFFIX)(ARG1, addr, val, attrs,
 | |
|                                              result, DEVICE_LITTLE_ENDIAN);
 | |
| }
 | |
| 
 | |
| void glue(address_space_stl_be, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stl_internal, SUFFIX)(ARG1, addr, val, attrs,
 | |
|                                              result, DEVICE_BIG_ENDIAN);
 | |
| }
 | |
| 
 | |
| void glue(address_space_stb, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint8_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 1;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, true, attrs);
 | |
|     if (!memory_access_is_direct(mr, true)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
|         r = memory_region_dispatch_write(mr, addr1, val, MO_8, attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         stb_p(ptr, val);
 | |
|         invalidate_and_set_dirty(mr, addr1, 1);
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
| }
 | |
| 
 | |
| /* warning: addr must be aligned */
 | |
| static inline void glue(address_space_stw_internal, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint16_t val, MemTxAttrs attrs,
 | |
|     MemTxResult *result, enum device_endian endian)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 2;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, true, attrs);
 | |
|     if (l < 2 || !memory_access_is_direct(mr, true)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
|         r = memory_region_dispatch_write(mr, addr1, val,
 | |
|                                          MO_16 | devend_memop(endian), attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         switch (endian) {
 | |
|         case DEVICE_LITTLE_ENDIAN:
 | |
|             stw_le_p(ptr, val);
 | |
|             break;
 | |
|         case DEVICE_BIG_ENDIAN:
 | |
|             stw_be_p(ptr, val);
 | |
|             break;
 | |
|         default:
 | |
|             stw_p(ptr, val);
 | |
|             break;
 | |
|         }
 | |
|         invalidate_and_set_dirty(mr, addr1, 2);
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
| }
 | |
| 
 | |
| void glue(address_space_stw, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint16_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stw_internal, SUFFIX)(ARG1, addr, val, attrs, result,
 | |
|                                              DEVICE_NATIVE_ENDIAN);
 | |
| }
 | |
| 
 | |
| void glue(address_space_stw_le, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint16_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stw_internal, SUFFIX)(ARG1, addr, val, attrs, result,
 | |
|                                              DEVICE_LITTLE_ENDIAN);
 | |
| }
 | |
| 
 | |
| void glue(address_space_stw_be, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint16_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stw_internal, SUFFIX)(ARG1, addr, val, attrs, result,
 | |
|                                DEVICE_BIG_ENDIAN);
 | |
| }
 | |
| 
 | |
| static void glue(address_space_stq_internal, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint64_t val, MemTxAttrs attrs,
 | |
|     MemTxResult *result, enum device_endian endian)
 | |
| {
 | |
|     uint8_t *ptr;
 | |
|     MemoryRegion *mr;
 | |
|     hwaddr l = 8;
 | |
|     hwaddr addr1;
 | |
|     MemTxResult r;
 | |
|     bool release_lock = false;
 | |
| 
 | |
|     RCU_READ_LOCK();
 | |
|     mr = TRANSLATE(addr, &addr1, &l, true, attrs);
 | |
|     if (l < 8 || !memory_access_is_direct(mr, true)) {
 | |
|         release_lock |= prepare_mmio_access(mr);
 | |
|         r = memory_region_dispatch_write(mr, addr1, val,
 | |
|                                          MO_64 | devend_memop(endian), attrs);
 | |
|     } else {
 | |
|         /* RAM case */
 | |
|         ptr = qemu_map_ram_ptr(mr->ram_block, addr1);
 | |
|         switch (endian) {
 | |
|         case DEVICE_LITTLE_ENDIAN:
 | |
|             stq_le_p(ptr, val);
 | |
|             break;
 | |
|         case DEVICE_BIG_ENDIAN:
 | |
|             stq_be_p(ptr, val);
 | |
|             break;
 | |
|         default:
 | |
|             stq_p(ptr, val);
 | |
|             break;
 | |
|         }
 | |
|         invalidate_and_set_dirty(mr, addr1, 8);
 | |
|         r = MEMTX_OK;
 | |
|     }
 | |
|     if (result) {
 | |
|         *result = r;
 | |
|     }
 | |
|     if (release_lock) {
 | |
|         qemu_mutex_unlock_iothread();
 | |
|     }
 | |
|     RCU_READ_UNLOCK();
 | |
| }
 | |
| 
 | |
| void glue(address_space_stq, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint64_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stq_internal, SUFFIX)(ARG1, addr, val, attrs, result,
 | |
|                                              DEVICE_NATIVE_ENDIAN);
 | |
| }
 | |
| 
 | |
| void glue(address_space_stq_le, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint64_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stq_internal, SUFFIX)(ARG1, addr, val, attrs, result,
 | |
|                                              DEVICE_LITTLE_ENDIAN);
 | |
| }
 | |
| 
 | |
| void glue(address_space_stq_be, SUFFIX)(ARG1_DECL,
 | |
|     hwaddr addr, uint64_t val, MemTxAttrs attrs, MemTxResult *result)
 | |
| {
 | |
|     glue(address_space_stq_internal, SUFFIX)(ARG1, addr, val, attrs, result,
 | |
|                                              DEVICE_BIG_ENDIAN);
 | |
| }
 | |
| 
 | |
| #undef ARG1_DECL
 | |
| #undef ARG1
 | |
| #undef SUFFIX
 | |
| #undef TRANSLATE
 | |
| #undef RCU_READ_LOCK
 | |
| #undef RCU_READ_UNLOCK
 |