Changeset b0c2075 in mainline
- Timestamp:
- 2013-09-10T17:48:57Z (12 years ago)
- Branches:
- lfn, master, serial, ticket/834-toolchain-update, topic/msim-upgrade, topic/simplify-dev-export
- Children:
- 85147f3
- Parents:
- 86733f3
- Location:
- kernel
- Files:
-
- 2 deleted
- 30 edited
Legend:
- Unmodified
- Added
- Removed
-
kernel/Makefile
r86733f3 rb0c2075 220 220 generic/src/mm/km.c \ 221 221 generic/src/mm/reserve.c \ 222 generic/src/mm/buddy.c \223 222 generic/src/mm/frame.c \ 224 223 generic/src/mm/page.c \ -
kernel/arch/abs32le/include/arch/mm/page.h
r86733f3 rb0c2075 57 57 58 58 /* Page table sizes for each level. */ 59 #define PTL0_ SIZE_ARCH ONE_FRAME60 #define PTL1_ SIZE_ARCH 061 #define PTL2_ SIZE_ARCH 062 #define PTL3_ SIZE_ARCH ONE_FRAME59 #define PTL0_FRAMES_ARCH 1 60 #define PTL1_FRAMES_ARCH 1 61 #define PTL2_FRAMES_ARCH 1 62 #define PTL3_FRAMES_ARCH 1 63 63 64 64 /* Macros calculating indices for each level. */ -
kernel/arch/amd64/include/arch/mm/page.h
r86733f3 rb0c2075 61 61 62 62 /* Page table sizes for each level. */ 63 #define PTL0_ SIZE_ARCH ONE_FRAME64 #define PTL1_ SIZE_ARCH ONE_FRAME65 #define PTL2_ SIZE_ARCH ONE_FRAME66 #define PTL3_ SIZE_ARCH ONE_FRAME63 #define PTL0_FRAMES_ARCH 1 64 #define PTL1_FRAMES_ARCH 1 65 #define PTL2_FRAMES_ARCH 1 66 #define PTL3_FRAMES_ARCH 1 67 67 68 68 /* Macros calculating indices into page tables in each level. */ -
kernel/arch/arm32/include/arch/mm/page.h
r86733f3 rb0c2075 73 73 74 74 /* Page table sizes for each level. */ 75 #define PTL0_ SIZE_ARCH FOUR_FRAMES76 #define PTL1_ SIZE_ARCH 077 #define PTL2_ SIZE_ARCH 078 #define PTL3_ SIZE_ARCH ONE_FRAME75 #define PTL0_FRAMES_ARCH 4 76 #define PTL1_FRAMES_ARCH 1 77 #define PTL2_FRAMES_ARCH 1 78 #define PTL3_FRAMES_ARCH 1 79 79 80 80 /* Macros calculating indices into page tables for each level. */ -
kernel/arch/arm32/src/mm/page.c
r86733f3 rb0c2075 69 69 #ifdef HIGH_EXCEPTION_VECTORS 70 70 /* Create mapping for exception table at high offset */ 71 uintptr_t ev_frame = frame_alloc( ONE_FRAME, FRAME_NONE, 0);71 uintptr_t ev_frame = frame_alloc(1, FRAME_NONE, 0); 72 72 page_mapping_insert(AS_KERNEL, EXC_BASE_ADDRESS, ev_frame, flags); 73 73 #else -
kernel/arch/arm32/src/ras.c
r86733f3 rb0c2075 51 51 void ras_init(void) 52 52 { 53 uintptr_t frame = frame_alloc(ONE_FRAME,54 FRAME_ATOMIC | FRAME_HIGHMEM, 0);53 uintptr_t frame = 54 frame_alloc(1, FRAME_ATOMIC | FRAME_HIGHMEM, 0); 55 55 if (!frame) 56 frame = frame_alloc( ONE_FRAME, FRAME_LOWMEM, 0);56 frame = frame_alloc(1, FRAME_LOWMEM, 0); 57 57 58 58 ras_page = (uintptr_t *) km_map(frame, -
kernel/arch/ia32/include/arch/mm/page.h
r86733f3 rb0c2075 66 66 67 67 /* Page table sizes for each level. */ 68 #define PTL0_ SIZE_ARCH ONE_FRAME69 #define PTL1_ SIZE_ARCH 070 #define PTL2_ SIZE_ARCH 071 #define PTL3_ SIZE_ARCH ONE_FRAME68 #define PTL0_FRAMES_ARCH 1 69 #define PTL1_FRAMES_ARCH 1 70 #define PTL2_FRAMES_ARCH 1 71 #define PTL3_FRAMES_ARCH 1 72 72 73 73 /* Macros calculating indices for each level. */ -
kernel/arch/ia64/src/mm/vhpt.c
r86733f3 rb0c2075 42 42 uintptr_t vhpt_set_up(void) 43 43 { 44 vhpt_base = (vhpt_entry_t *) PA2KA(frame_alloc(VHPT_WIDTH - FRAME_WIDTH,45 FRAME_ATOMIC, 0));44 vhpt_base = (vhpt_entry_t *) 45 PA2KA(frame_alloc(SIZE2FRAMES(VHPT_SIZE), FRAME_ATOMIC, 0)); 46 46 if (!vhpt_base) 47 47 panic("Kernel configured with VHPT but no memory for table."); … … 83 83 void vhpt_invalidate_all() 84 84 { 85 memsetb(vhpt_base, 1 << VHPT_WIDTH, 0);85 memsetb(vhpt_base, VHPT_SIZE, 0); 86 86 } 87 87 -
kernel/arch/mips32/include/arch/mm/page.h
r86733f3 rb0c2075 78 78 79 79 /* Macros describing size of page tables in each level. */ 80 #define PTL0_ SIZE_ARCH ONE_FRAME81 #define PTL1_ SIZE_ARCH 082 #define PTL2_ SIZE_ARCH 083 #define PTL3_ SIZE_ARCH ONE_FRAME80 #define PTL0_FRAMES_ARCH 1 81 #define PTL1_FRAMES_ARCH 1 82 #define PTL2_FRAMES_ARCH 1 83 #define PTL3_FRAMES_ARCH 1 84 84 85 85 /* Macros calculating entry indices for each level. */ -
kernel/arch/mips32/src/mm/tlb.c
r86733f3 rb0c2075 48 48 #include <symtab.h> 49 49 50 #define PFN_SHIFT 12 51 #define VPN_SHIFT 12 52 #define ADDR2VPN(a) ((a) >> VPN_SHIFT) 53 #define ADDR2VPN2(a) (ADDR2VPN((a)) >> 1) 54 #define VPN2ADDR(vpn) ((vpn) << VPN_SHIFT) 55 #define VPN22ADDR(vpn2) (VPN2ADDR(vpn2) << 1) 56 #define PFN2ADDR(pfn) ((pfn) << PFN_SHIFT) 57 58 #define BANK_SELECT_BIT(a) (((a) >> PAGE_WIDTH) & 1) 59 50 #define PFN_SHIFT 12 51 #define VPN_SHIFT 12 52 53 #define ADDR2HI_VPN(a) ((a) >> VPN_SHIFT) 54 #define ADDR2HI_VPN2(a) (ADDR2HI_VPN((a)) >> 1) 55 56 #define HI_VPN2ADDR(vpn) ((vpn) << VPN_SHIFT) 57 #define HI_VPN22ADDR(vpn2) (HI_VPN2ADDR(vpn2) << 1) 58 59 #define LO_PFN2ADDR(pfn) ((pfn) << PFN_SHIFT) 60 61 #define BANK_SELECT_BIT(a) (((a) >> PAGE_WIDTH) & 1) 60 62 61 63 /** Initialize TLB. … … 266 268 { 267 269 hi->value = 0; 268 hi->vpn2 = ADDR2 VPN2(ALIGN_DOWN(addr, PAGE_SIZE));270 hi->vpn2 = ADDR2HI_VPN2(ALIGN_DOWN(addr, PAGE_SIZE)); 269 271 hi->asid = asid; 270 272 } … … 295 297 296 298 printf("%-4u %-6u %0#10x %-#6x %1u%1u%1u%1u %0#10x\n", 297 i, hi.asid, VPN22ADDR(hi.vpn2), mask.mask,298 lo0.g, lo0.v, lo0.d, lo0.c, PFN2ADDR(lo0.pfn));299 i, hi.asid, HI_VPN22ADDR(hi.vpn2), mask.mask, 300 lo0.g, lo0.v, lo0.d, lo0.c, LO_PFN2ADDR(lo0.pfn)); 299 301 printf(" %1u%1u%1u%1u %0#10x\n", 300 lo1.g, lo1.v, lo1.d, lo1.c, PFN2ADDR(lo1.pfn));302 lo1.g, lo1.v, lo1.d, lo1.c, LO_PFN2ADDR(lo1.pfn)); 301 303 } 302 304 -
kernel/arch/ppc32/include/arch/mm/page.h
r86733f3 rb0c2075 70 70 71 71 /* Page table sizes for each level. */ 72 #define PTL0_ SIZE_ARCH ONE_FRAME73 #define PTL1_ SIZE_ARCH 074 #define PTL2_ SIZE_ARCH 075 #define PTL3_ SIZE_ARCH ONE_FRAME72 #define PTL0_FRAMES_ARCH 1 73 #define PTL1_FRAMES_ARCH 1 74 #define PTL2_FRAMES_ARCH 1 75 #define PTL3_FRAMES_ARCH 1 76 76 77 77 /* Macros calculating indices into page tables on each level. */ -
kernel/arch/sparc64/src/mm/sun4u/as.c
r86733f3 rb0c2075 63 63 { 64 64 #ifdef CONFIG_TSB 65 /* 66 * The order must be calculated with respect to the emulated 67 * 16K page size. 68 * 69 */ 70 uint8_t order = fnzb32(((ITSB_ENTRY_COUNT + DTSB_ENTRY_COUNT) * 71 sizeof(tsb_entry_t)) >> FRAME_WIDTH); 72 73 uintptr_t tsb = PA2KA(frame_alloc(order, flags, 0)); 65 size_t frames = SIZE2FRAMES((ITSB_ENTRY_COUNT + DTSB_ENTRY_COUNT) * 66 sizeof(tsb_entry_t)); 67 68 uintptr_t tsb = PA2KA(frame_alloc(frames, flags, 0)); 74 69 if (!tsb) 75 70 return -1; 76 71 77 72 as->arch.itsb = (tsb_entry_t *) tsb; 78 as->arch.dtsb = (tsb_entry_t *) (tsb + ITSB_ENTRY_COUNT *79 sizeof(tsb_entry_t));73 as->arch.dtsb = (tsb_entry_t *) (tsb + 74 ITSB_ENTRY_COUNT * sizeof(tsb_entry_t)); 80 75 81 76 memsetb(as->arch.itsb, -
kernel/arch/sparc64/src/mm/sun4v/as.c
r86733f3 rb0c2075 66 66 { 67 67 #ifdef CONFIG_TSB 68 uint8_t order = fnzb32(69 (TSB_ENTRY_COUNT * sizeof(tsb_entry_t)) >> FRAME_WIDTH);68 size_t frames = 69 SIZE2FRAMES(TSB_ENTRY_COUNT * sizeof(tsb_entry_t)); 70 70 71 uintptr_t tsb = frame_alloc( order, flags, 0);71 uintptr_t tsb = frame_alloc(frames, flags, 0); 72 72 if (!tsb) 73 73 return -1; -
kernel/genarch/include/genarch/mm/page_pt.h
r86733f3 rb0c2075 58 58 #define PTL3_ENTRIES PTL3_ENTRIES_ARCH 59 59 60 /* Table sizes in each level */61 #define PTL0_ SIZE PTL0_SIZE_ARCH62 #define PTL1_ SIZE PTL1_SIZE_ARCH63 #define PTL2_ SIZE PTL2_SIZE_ARCH64 #define PTL3_ SIZE PTL3_SIZE_ARCH60 /* Table sizes in each level (in frames) */ 61 #define PTL0_FRAMES PTL0_FRAMES_ARCH 62 #define PTL1_FRAMES PTL1_FRAMES_ARCH 63 #define PTL2_FRAMES PTL2_FRAMES_ARCH 64 #define PTL3_FRAMES PTL3_FRAMES_ARCH 65 65 66 66 /* -
kernel/genarch/src/mm/as_pt.c
r86733f3 rb0c2075 73 73 pte_t *ptl0_create(unsigned int flags) 74 74 { 75 pte_t *dst_ptl0 = (pte_t *) PA2KA(frame_alloc(PTL0_SIZE,76 FRAME_LOWMEM, 0));77 size_t table_size = FRAME _SIZE << PTL0_SIZE;75 pte_t *dst_ptl0 = (pte_t *) 76 PA2KA(frame_alloc(PTL0_FRAMES, FRAME_LOWMEM, 0)); 77 size_t table_size = FRAMES2SIZE(PTL0_FRAMES); 78 78 79 79 if (flags & FLAG_AS_KERNEL) -
kernel/genarch/src/mm/page_pt.c
r86733f3 rb0c2075 82 82 83 83 if (GET_PTL1_FLAGS(ptl0, PTL0_INDEX(page)) & PAGE_NOT_PRESENT) { 84 pte_t *newpt = (pte_t *) PA2KA(frame_alloc(PTL1_SIZE,85 FRAME_LOWMEM, 0));86 memsetb(newpt, FRAME _SIZE << PTL1_SIZE, 0);84 pte_t *newpt = (pte_t *) 85 PA2KA(frame_alloc(PTL1_FRAMES, FRAME_LOWMEM, 0)); 86 memsetb(newpt, FRAMES2SIZE(PTL1_FRAMES), 0); 87 87 SET_PTL1_ADDRESS(ptl0, PTL0_INDEX(page), KA2PA(newpt)); 88 88 SET_PTL1_FLAGS(ptl0, PTL0_INDEX(page), … … 101 101 102 102 if (GET_PTL2_FLAGS(ptl1, PTL1_INDEX(page)) & PAGE_NOT_PRESENT) { 103 pte_t *newpt = (pte_t *) PA2KA(frame_alloc(PTL2_SIZE,104 FRAME_LOWMEM, 0));105 memsetb(newpt, FRAME _SIZE << PTL2_SIZE, 0);103 pte_t *newpt = (pte_t *) 104 PA2KA(frame_alloc(PTL2_FRAMES, FRAME_LOWMEM, 0)); 105 memsetb(newpt, FRAMES2SIZE(PTL2_FRAMES), 0); 106 106 SET_PTL2_ADDRESS(ptl1, PTL1_INDEX(page), KA2PA(newpt)); 107 107 SET_PTL2_FLAGS(ptl1, PTL1_INDEX(page), … … 118 118 119 119 if (GET_PTL3_FLAGS(ptl2, PTL2_INDEX(page)) & PAGE_NOT_PRESENT) { 120 pte_t *newpt = (pte_t *) PA2KA(frame_alloc(PTL3_SIZE,121 FRAME_LOWMEM, 0));122 memsetb(newpt, FRAME _SIZE << PTL3_SIZE, 0);120 pte_t *newpt = (pte_t *) 121 PA2KA(frame_alloc(PTL3_FRAMES, FRAME_LOWMEM, 0)); 122 memsetb(newpt, FRAMES2SIZE(PTL3_FRAMES), 0); 123 123 SET_PTL3_ADDRESS(ptl2, PTL2_INDEX(page), KA2PA(newpt)); 124 124 SET_PTL3_FLAGS(ptl2, PTL2_INDEX(page), … … 371 371 uintptr_t ptl0 = PA2KA((uintptr_t) AS_KERNEL->genarch.page_table); 372 372 uintptr_t ptl0_step = ptl0_step_get(); 373 size_t order;373 size_t frames; 374 374 375 375 #if (PTL1_ENTRIES != 0) 376 order = PTL1_SIZE;376 frames = PTL1_FRAMES; 377 377 #elif (PTL2_ENTRIES != 0) 378 order = PTL2_SIZE;378 frames = PTL2_FRAMES; 379 379 #else 380 order = PTL3_SIZE;380 frames = PTL3_FRAMES; 381 381 #endif 382 382 … … 384 384 addr - 1 < base + size - 1; 385 385 addr += ptl0_step) { 386 uintptr_t l1 = PA2KA(frame_alloc( order, FRAME_LOWMEM, 0));387 memsetb((void *) l1, FRAME _SIZE << order, 0);386 uintptr_t l1 = PA2KA(frame_alloc(frames, FRAME_LOWMEM, 0)); 387 memsetb((void *) l1, FRAMES2SIZE(frames), 0); 388 388 SET_PTL1_ADDRESS(ptl0, PTL0_INDEX(addr), KA2PA(l1)); 389 389 SET_PTL1_FLAGS(ptl0, PTL0_INDEX(addr), -
kernel/generic/include/config.h
r86733f3 rb0c2075 37 37 38 38 #include <arch/mm/page.h> 39 #include <macros.h> 39 40 40 #define ONE_FRAME 0 41 #define TWO_FRAMES 1 42 #define FOUR_FRAMES 2 43 44 #define STACK_FRAMES TWO_FRAMES 45 #define STACK_SIZE ((1 << STACK_FRAMES) << PAGE_WIDTH) 41 #define STACK_FRAMES 2 42 #define STACK_SIZE FRAMES2SIZE(STACK_FRAMES) 46 43 47 44 #define STACK_SIZE_USER (1 * 1024 * 1024) -
kernel/generic/include/macros.h
r86733f3 rb0c2075 117 117 overlaps(KA2PA((x)), (szx), KA2PA((y)), (szy)) 118 118 119 #define PFN2ADDR(frame) ((frame) << FRAME_WIDTH) 120 #define ADDR2PFN(addr) ((addr) >> FRAME_WIDTH) 121 122 #define FRAMES2SIZE(frames) ((frames) << FRAME_WIDTH) 123 #define SIZE2FRAMES(size) \ 124 (((size) == 0) ? 0 : ((((size) - 1) >> FRAME_WIDTH) + 1)) 125 119 126 #define KiB2SIZE(kb) ((kb) << 10) 120 127 #define MiB2SIZE(mb) ((mb) << 20) -
kernel/generic/include/mm/frame.h
r86733f3 rb0c2075 39 39 #include <typedefs.h> 40 40 #include <trace.h> 41 #include <adt/bitmap.h> 41 42 #include <adt/list.h> 42 #include <mm/buddy.h>43 43 #include <synch/spinlock.h> 44 44 #include <arch/mm/page.h> … … 90 90 91 91 typedef struct { 92 size_t refcount; /**< Tracking of shared frames */ 93 link_t buddy_link; /**< Link to the next free block inside 94 one order */ 95 void *parent; /**< If allocated by slab, this points there */ 96 uint8_t buddy_order; /**< Buddy system block order */ 92 size_t refcount; /**< Tracking of shared frames */ 93 void *parent; /**< If allocated by slab, this points there */ 97 94 } frame_t; 98 95 99 96 typedef struct { 100 pfn_t base; /**< Frame_no of the first frame 101 in the frames array */ 102 size_t count; /**< Size of zone */ 103 size_t free_count; /**< Number of free frame_t 104 structures */ 105 size_t busy_count; /**< Number of busy frame_t 106 structures */ 107 zone_flags_t flags; /**< Type of the zone */ 97 /** Frame_no of the first frame in the frames array */ 98 pfn_t base; 108 99 109 frame_t *frames; /**< Array of frame_t structures 110 in this zone */ 111 buddy_system_t *buddy_system; /**< Buddy system for the zone */ 100 /** Size of zone */ 101 size_t count; 102 103 /** Number of free frame_t structures */ 104 size_t free_count; 105 106 /** Number of busy frame_t structures */ 107 size_t busy_count; 108 109 /** Type of the zone */ 110 zone_flags_t flags; 111 112 /** Frame bitmap */ 113 bitmap_t bitmap; 114 115 /** Array of frame_t structures in this zone */ 116 frame_t *frames; 112 117 } zone_t; 113 118 … … 124 129 extern zones_t zones; 125 130 126 NO_TRACE static inline uintptr_t PFN2ADDR(pfn_t frame)127 {128 return (uintptr_t) (frame << FRAME_WIDTH);129 }130 131 NO_TRACE static inline pfn_t ADDR2PFN(uintptr_t addr)132 {133 return (pfn_t) (addr >> FRAME_WIDTH);134 }135 136 NO_TRACE static inline size_t SIZE2FRAMES(size_t size)137 {138 if (size == 0)139 return 0;140 141 return (size_t) ((size - 1) >> FRAME_WIDTH) + 1;142 }143 144 NO_TRACE static inline size_t FRAMES2SIZE(size_t frames)145 {146 return (size_t) (frames << FRAME_WIDTH);147 }148 149 #define IS_BUDDY_ORDER_OK(index, order) \150 ((~(((sysarg_t) -1) << (order)) & (index)) == 0)151 #define IS_BUDDY_LEFT_BLOCK(zone, frame) \152 (((frame_index((zone), (frame)) >> (frame)->buddy_order) & 0x1) == 0)153 #define IS_BUDDY_RIGHT_BLOCK(zone, frame) \154 (((frame_index((zone), (frame)) >> (frame)->buddy_order) & 0x1) == 1)155 #define IS_BUDDY_LEFT_BLOCK_ABS(zone, frame) \156 (((frame_index_abs((zone), (frame)) >> (frame)->buddy_order) & 0x1) == 0)157 #define IS_BUDDY_RIGHT_BLOCK_ABS(zone, frame) \158 (((frame_index_abs((zone), (frame)) >> (frame)->buddy_order) & 0x1) == 1)159 160 131 extern void frame_init(void); 161 132 extern bool frame_adjust_zone_bounds(bool, uintptr_t *, size_t *); 162 extern uintptr_t frame_alloc_generic( uint8_t, frame_flags_t, uintptr_t, size_t *);163 extern uintptr_t frame_alloc( uint8_t, frame_flags_t, uintptr_t);164 extern uintptr_t frame_alloc_noreserve( uint8_t, frame_flags_t, uintptr_t);133 extern uintptr_t frame_alloc_generic(size_t, frame_flags_t, uintptr_t, size_t *); 134 extern uintptr_t frame_alloc(size_t, frame_flags_t, uintptr_t); 135 extern uintptr_t frame_alloc_noreserve(size_t, frame_flags_t, uintptr_t); 165 136 extern void frame_free_generic(uintptr_t, frame_flags_t); 166 137 extern void frame_free(uintptr_t); -
kernel/generic/include/mm/slab.h
r86733f3 rb0c2075 55 55 /** Maximum wasted space we allow for cache */ 56 56 #define SLAB_MAX_BADNESS(cache) \ 57 ( ((unsigned int) PAGE_SIZE << (cache)->order) >> 2)57 (FRAMES2SIZE((cache)->frames) >> 2) 58 58 59 59 /* slab_reclaim constants */ … … 101 101 102 102 /* Computed values */ 103 uint8_t order; /**< Order of frames to be allocated */103 size_t frames; /**< Number of frames to be allocated */ 104 104 size_t objects; /**< Number of objects that fit in */ 105 105 -
kernel/generic/src/cpu/cpu.c
r86733f3 rb0c2075 74 74 for (i = 0; i < config.cpu_count; i++) { 75 75 cpus[i].stack = (uint8_t *) PA2KA(frame_alloc(STACK_FRAMES, 76 FRAME_LOWMEM | FRAME_ATOMIC, 0));76 FRAME_LOWMEM | FRAME_ATOMIC, STACK_SIZE - 1)); 77 77 cpus[i].id = i; 78 78 -
kernel/generic/src/ddi/ddi.c
r86733f3 rb0c2075 322 322 } 323 323 324 NO_TRACE static int dmamem_map_anonymous(size_t size, unsigned int map_flags, 325 unsigned int flags, uintptr_t *phys, uintptr_t *virt, uintptr_t bound) 324 NO_TRACE static int dmamem_map_anonymous(size_t size, uintptr_t constraint, 325 unsigned int map_flags, unsigned int flags, uintptr_t *phys, 326 uintptr_t *virt, uintptr_t bound) 326 327 { 327 328 ASSERT(TASK); 328 329 329 330 size_t frames = SIZE2FRAMES(size); 330 uint8_t order; 331 332 /* We need the 2^order >= frames */ 333 if (frames == 1) 334 order = 0; 335 else 336 order = fnzb(frames - 1) + 1; 337 338 *phys = frame_alloc_noreserve(order, 0, 0); 331 *phys = frame_alloc_noreserve(frames, 0, constraint); 339 332 if (*phys == 0) 340 333 return ENOMEM; … … 390 383 */ 391 384 385 uintptr_t constraint; 386 int rc = copy_from_uspace(&constraint, phys_ptr, 387 sizeof(constraint)); 388 if (rc != EOK) 389 return rc; 390 392 391 uintptr_t phys; 393 392 uintptr_t virt = (uintptr_t) -1; 394 int rc = dmamem_map_anonymous(size, map_flags, flags,393 rc = dmamem_map_anonymous(size, constraint, map_flags, flags, 395 394 &phys, &virt, bound); 396 395 if (rc != EOK) -
kernel/generic/src/mm/frame.c
r86733f3 rb0c2075 38 38 * 39 39 * This file contains the physical frame allocator and memory zone management. 40 * The frame allocator is built on top of the buddy allocator. 41 * 42 * @see buddy.c 40 * The frame allocator is built on top of the two-level bitmap structure. 41 * 43 42 */ 44 43 … … 62 61 #include <str.h> 63 62 63 #define BITMAP_BLOCK_SIZE 1024 64 64 65 zones_t zones; 65 66 … … 92 93 } 93 94 94 NO_TRACE static inline size_t make_frame_index(zone_t *zone, frame_t *frame)95 {96 return (frame - zone->frames);97 }98 99 95 /** Initialize frame structure. 100 96 * … … 104 100 NO_TRACE static void frame_initialize(frame_t *frame) 105 101 { 106 frame->refcount = 1;107 frame-> buddy_order = 0;102 frame->refcount = 0; 103 frame->parent = NULL; 108 104 } 109 105 … … 161 157 162 158 /* Move other zones up */ 163 size_t j; 164 for (j = zones.count; j > i; j--) { 159 for (size_t j = zones.count; j > i; j--) 165 160 zones.info[j] = zones.info[j - 1]; 166 if (zones.info[j].buddy_system != NULL)167 zones.info[j].buddy_system->data =168 (void *) &zones.info[j];169 }170 161 171 162 zones.count++; … … 237 228 } 238 229 239 /** @return True if zone can allocate specified order */ 240 NO_TRACE static bool zone_can_alloc(zone_t *zone, uint8_t order) 241 { 230 /** @return True if zone can allocate specified number of frames */ 231 NO_TRACE static bool zone_can_alloc(zone_t *zone, size_t count, 232 pfn_t constraint) 233 { 234 /* 235 * The function bitmap_allocate_range() does not modify 236 * the bitmap if the last argument is NULL. 237 */ 242 238 return ((zone->flags & ZONE_AVAILABLE) && 243 buddy_system_can_alloc(zone->buddy_system, order)); 244 } 245 246 /** Find a zone that can allocate order frames. 239 bitmap_allocate_range(&zone->bitmap, count, zone->base, 240 constraint, NULL)); 241 } 242 243 /** Find a zone that can allocate specified number of frames 247 244 * 248 245 * Assume interrupts are disabled and zones lock is 249 246 * locked. 250 247 * 251 * @param order Size (2^order) of free space we are trying to find. 252 * @param flags Required flags of the target zone. 253 * @param hind Preferred zone. 254 * 255 */ 256 NO_TRACE static size_t find_free_zone(uint8_t order, zone_flags_t flags, 257 size_t hint) 248 * @param count Number of free frames we are trying to find. 249 * @param flags Required flags of the target zone. 250 * @param constraint Indication of bits that cannot be set in the 251 * physical frame number of the first allocated frame. 252 * @param hind Preferred zone. 253 * 254 */ 255 NO_TRACE static size_t find_free_zone(size_t count, zone_flags_t flags, 256 pfn_t constraint, size_t hint) 258 257 { 259 258 if (hint >= zones.count) … … 267 266 if (ZONE_FLAGS_MATCH(zones.info[i].flags, flags)) { 268 267 /* 269 * Check if the zone has 2^order frames area available.268 * Check if the zone can satisfy the allocation request. 270 269 */ 271 if (zone_can_alloc(&zones.info[i], order))270 if (zone_can_alloc(&zones.info[i], count, constraint)) 272 271 return i; 273 272 } … … 282 281 } 283 282 284 /**************************/285 /* Buddy system functions */286 /**************************/287 288 /** Buddy system find_block implementation.289 *290 * Find block that is parent of current list.291 * That means go to lower addresses, until such block is found292 *293 * @param order Order of parent must be different then this294 * parameter!!295 *296 */297 NO_TRACE static link_t *zone_buddy_find_block(buddy_system_t *buddy,298 link_t *child, uint8_t order)299 {300 frame_t *frame = list_get_instance(child, frame_t, buddy_link);301 zone_t *zone = (zone_t *) buddy->data;302 303 size_t index = frame_index(zone, frame);304 do {305 if (zone->frames[index].buddy_order != order)306 return &zone->frames[index].buddy_link;307 } while (index-- > 0);308 309 return NULL;310 }311 312 /** Buddy system find_buddy implementation.313 *314 * @param buddy Buddy system.315 * @param block Block for which buddy should be found.316 *317 * @return Buddy for given block if found.318 *319 */320 NO_TRACE static link_t *zone_buddy_find_buddy(buddy_system_t *buddy,321 link_t *block)322 {323 frame_t *frame = list_get_instance(block, frame_t, buddy_link);324 zone_t *zone = (zone_t *) buddy->data;325 ASSERT(IS_BUDDY_ORDER_OK(frame_index_abs(zone, frame),326 frame->buddy_order));327 328 bool is_left = IS_BUDDY_LEFT_BLOCK_ABS(zone, frame);329 330 size_t index;331 if (is_left) {332 index = (frame_index(zone, frame)) +333 (1 << frame->buddy_order);334 } else { /* is_right */335 index = (frame_index(zone, frame)) -336 (1 << frame->buddy_order);337 }338 339 if (frame_index_valid(zone, index)) {340 if ((zone->frames[index].buddy_order == frame->buddy_order) &&341 (zone->frames[index].refcount == 0)) {342 return &zone->frames[index].buddy_link;343 }344 }345 346 return NULL;347 }348 349 /** Buddy system bisect implementation.350 *351 * @param buddy Buddy system.352 * @param block Block to bisect.353 *354 * @return Right block.355 *356 */357 NO_TRACE static link_t *zone_buddy_bisect(buddy_system_t *buddy, link_t *block)358 {359 frame_t *frame_l = list_get_instance(block, frame_t, buddy_link);360 frame_t *frame_r = (frame_l + (1 << (frame_l->buddy_order - 1)));361 362 return &frame_r->buddy_link;363 }364 365 /** Buddy system coalesce implementation.366 *367 * @param buddy Buddy system.368 * @param block_1 First block.369 * @param block_2 First block's buddy.370 *371 * @return Coalesced block (actually block that represents lower372 * address).373 *374 */375 NO_TRACE static link_t *zone_buddy_coalesce(buddy_system_t *buddy,376 link_t *block_1, link_t *block_2)377 {378 frame_t *frame1 = list_get_instance(block_1, frame_t, buddy_link);379 frame_t *frame2 = list_get_instance(block_2, frame_t, buddy_link);380 381 return ((frame1 < frame2) ? block_1 : block_2);382 }383 384 /** Buddy system set_order implementation.385 *386 * @param buddy Buddy system.387 * @param block Buddy system block.388 * @param order Order to set.389 *390 */391 NO_TRACE static void zone_buddy_set_order(buddy_system_t *buddy, link_t *block,392 uint8_t order)393 {394 list_get_instance(block, frame_t, buddy_link)->buddy_order = order;395 }396 397 /** Buddy system get_order implementation.398 *399 * @param buddy Buddy system.400 * @param block Buddy system block.401 *402 * @return Order of block.403 *404 */405 NO_TRACE static uint8_t zone_buddy_get_order(buddy_system_t *buddy,406 link_t *block)407 {408 return list_get_instance(block, frame_t, buddy_link)->buddy_order;409 }410 411 /** Buddy system mark_busy implementation.412 *413 * @param buddy Buddy system.414 * @param block Buddy system block.415 *416 */417 NO_TRACE static void zone_buddy_mark_busy(buddy_system_t *buddy, link_t *block)418 {419 list_get_instance(block, frame_t, buddy_link)->refcount = 1;420 }421 422 /** Buddy system mark_available implementation.423 *424 * @param buddy Buddy system.425 * @param block Buddy system block.426 *427 */428 NO_TRACE static void zone_buddy_mark_available(buddy_system_t *buddy,429 link_t *block)430 {431 list_get_instance(block, frame_t, buddy_link)->refcount = 0;432 }433 434 static buddy_system_operations_t zone_buddy_system_operations = {435 .find_buddy = zone_buddy_find_buddy,436 .bisect = zone_buddy_bisect,437 .coalesce = zone_buddy_coalesce,438 .set_order = zone_buddy_set_order,439 .get_order = zone_buddy_get_order,440 .mark_busy = zone_buddy_mark_busy,441 .mark_available = zone_buddy_mark_available,442 .find_block = zone_buddy_find_block443 };444 445 283 /******************/ 446 284 /* Zone functions */ 447 285 /******************/ 448 286 287 /** Return frame from zone. */ 288 NO_TRACE static frame_t *zone_get_frame(zone_t *zone, size_t index) 289 { 290 ASSERT(index < zone->count); 291 292 return &zone->frames[index]; 293 } 294 449 295 /** Allocate frame in particular zone. 450 296 * … … 452 298 * Panics if allocation is impossible. 453 299 * 454 * @param zone Zone to allocate from. 455 * @param order Allocate exactly 2^order frames. 300 * @param zone Zone to allocate from. 301 * @param count Number of frames to allocate 302 * @param constraint Indication of bits that cannot be set in the 303 * physical frame number of the first allocated frame. 456 304 * 457 305 * @return Frame index in zone. 458 306 * 459 307 */ 460 NO_TRACE static pfn_t zone_frame_alloc(zone_t *zone, uint8_t order) 308 NO_TRACE static size_t zone_frame_alloc(zone_t *zone, size_t count, 309 pfn_t constraint) 461 310 { 462 311 ASSERT(zone->flags & ZONE_AVAILABLE); 463 312 464 /* Allocate frames from zone buddy system */ 465 link_t *link = buddy_system_alloc(zone->buddy_system, order); 466 467 ASSERT(link); 313 /* Allocate frames from zone */ 314 size_t index; 315 int avail = bitmap_allocate_range(&zone->bitmap, count, zone->base, 316 constraint, &index); 317 318 ASSERT(avail); 319 320 /* Update frame reference count */ 321 for (size_t i = 0; i < count; i++) { 322 frame_t *frame = zone_get_frame(zone, index + i); 323 324 ASSERT(frame->refcount == 0); 325 frame->refcount = 1; 326 } 468 327 469 328 /* Update zone information. */ 470 zone->free_count -= (1 << order); 471 zone->busy_count += (1 << order); 472 473 /* Frame will be actually a first frame of the block. */ 474 frame_t *frame = list_get_instance(link, frame_t, buddy_link); 475 476 /* Get frame address */ 477 return make_frame_index(zone, frame); 329 zone->free_count -= count; 330 zone->busy_count += count; 331 332 return index; 478 333 } 479 334 … … 482 337 * Assume zone is locked and is available for deallocation. 483 338 * 484 * @param zone 485 * @param frame_idx Frame index relative to zone.486 * 487 * @return 488 * 489 */ 490 NO_TRACE static size_t zone_frame_free(zone_t *zone, size_t frame_idx)339 * @param zone Pointer to zone from which the frame is to be freed. 340 * @param index Frame index relative to zone. 341 * 342 * @return Number of freed frames. 343 * 344 */ 345 NO_TRACE static size_t zone_frame_free(zone_t *zone, size_t index) 491 346 { 492 347 ASSERT(zone->flags & ZONE_AVAILABLE); 493 348 494 frame_t *frame = &zone->frames[frame_idx];349 frame_t *frame = zone_get_frame(zone, index); 495 350 size_t size = 0; 496 351 497 ASSERT(frame->refcount );352 ASSERT(frame->refcount > 0); 498 353 499 354 if (!--frame->refcount) { 500 size = 1 << frame->buddy_order;501 buddy_system_free(zone->buddy_system, &frame->buddy_link);355 bitmap_free_range(&zone->bitmap, index, 1); 356 502 357 /* Update zone information. */ 503 zone->free_count += size; 504 zone->busy_count -= size; 358 zone->free_count++; 359 zone->busy_count--; 360 361 size = FRAME_SIZE; 505 362 } 506 363 … … 508 365 } 509 366 510 /** Return frame from zone. */511 NO_TRACE static frame_t *zone_get_frame(zone_t *zone, size_t frame_idx)512 {513 ASSERT(frame_idx < zone->count);514 return &zone->frames[frame_idx];515 }516 517 367 /** Mark frame in zone unavailable to allocation. */ 518 NO_TRACE static void zone_mark_unavailable(zone_t *zone, size_t frame_idx)368 NO_TRACE static void zone_mark_unavailable(zone_t *zone, size_t index) 519 369 { 520 370 if (!(zone->flags & ZONE_AVAILABLE)) 521 371 return; 522 // ASSERT(zone->flags & ZONE_AVAILABLE); 523 524 frame_t *frame = zone_get_frame(zone, frame_idx); 525 if (frame->refcount) 372 373 frame_t *frame = zone_get_frame(zone, index); 374 if (frame->refcount > 0) 526 375 return; 527 376 528 link_t *link __attribute__ ((unused)); 529 530 link = buddy_system_alloc_block(zone->buddy_system, 531 &frame->buddy_link); 532 533 ASSERT(link); 377 bitmap_set_range(&zone->bitmap, index, 1); 378 534 379 zone->free_count--; 535 380 reserve_force_alloc(1); … … 538 383 /** Merge two zones. 539 384 * 540 * Expect buddy to point to space at least zone_conf_size large.541 385 * Assume z1 & z2 are locked and compatible and zones lock is 542 386 * locked. 543 387 * 544 * @param z1 First zone to merge.545 * @param z2 Second zone to merge.546 * @param old_z1 Original dateof the first zone.547 * @param buddy Merged zone buddy.388 * @param z1 First zone to merge. 389 * @param z2 Second zone to merge. 390 * @param old_z1 Original data of the first zone. 391 * @param confdata Merged zone configuration data. 548 392 * 549 393 */ 550 394 NO_TRACE static void zone_merge_internal(size_t z1, size_t z2, zone_t *old_z1, 551 buddy_system_t *buddy)395 void *confdata) 552 396 { 553 397 ASSERT(zones.info[z1].flags & ZONE_AVAILABLE); … … 564 408 zones.info[z1].free_count += zones.info[z2].free_count; 565 409 zones.info[z1].busy_count += zones.info[z2].busy_count; 566 zones.info[z1].buddy_system = buddy; 567 568 uint8_t order = fnzb(zones.info[z1].count); 569 buddy_system_create(zones.info[z1].buddy_system, order, 570 &zone_buddy_system_operations, (void *) &zones.info[z1]); 571 572 zones.info[z1].frames = 573 (frame_t *) ((uint8_t *) zones.info[z1].buddy_system 574 + buddy_conf_size(order)); 575 576 /* This marks all frames busy */ 577 size_t i; 578 for (i = 0; i < zones.info[z1].count; i++) 579 frame_initialize(&zones.info[z1].frames[i]); 580 581 /* Copy frames from both zones to preserve full frame orders, 582 * parents etc. Set all free frames with refcount = 0 to 1, because 583 * we add all free frames to buddy allocator later again, clearing 584 * order to 0. Don't set busy frames with refcount = 0, as they 585 * will not be reallocated during merge and it would make later 586 * problems with allocation/free. 410 411 bitmap_initialize(&zones.info[z1].bitmap, zones.info[z1].count, 412 BITMAP_BLOCK_SIZE, confdata + 413 (sizeof(frame_t) * zones.info[z1].count)); 414 415 zones.info[z1].frames = (frame_t *) confdata; 416 417 /* 418 * Copy frames and bits from both zones to preserve parents, etc. 587 419 */ 588 for (i = 0; i < old_z1->count; i++) 420 421 for (size_t i = 0; i < old_z1->count; i++) { 422 bitmap_set(&zones.info[z1].bitmap, i, 423 bitmap_get(&old_z1->bitmap, i)); 589 424 zones.info[z1].frames[i] = old_z1->frames[i]; 590 591 for (i = 0; i < zones.info[z2].count; i++) 592 zones.info[z1].frames[base_diff + i] 593 = zones.info[z2].frames[i]; 594 595 i = 0; 596 while (i < zones.info[z1].count) { 597 if (zones.info[z1].frames[i].refcount) { 598 /* Skip busy frames */ 599 i += 1 << zones.info[z1].frames[i].buddy_order; 600 } else { 601 /* Free frames, set refcount = 1 602 * (all free frames have refcount == 0, we need not 603 * to check the order) 604 */ 605 zones.info[z1].frames[i].refcount = 1; 606 zones.info[z1].frames[i].buddy_order = 0; 607 i++; 608 } 609 } 610 611 /* Add free blocks from the original zone z1 */ 612 while (zone_can_alloc(old_z1, 0)) { 613 /* Allocate from the original zone */ 614 pfn_t frame_idx = zone_frame_alloc(old_z1, 0); 615 616 /* Free the frame from the merged zone */ 617 frame_t *frame = &zones.info[z1].frames[frame_idx]; 618 frame->refcount = 0; 619 buddy_system_free(zones.info[z1].buddy_system, &frame->buddy_link); 620 } 621 622 /* Add free blocks from the original zone z2 */ 623 while (zone_can_alloc(&zones.info[z2], 0)) { 624 /* Allocate from the original zone */ 625 pfn_t frame_idx = zone_frame_alloc(&zones.info[z2], 0); 626 627 /* Free the frame from the merged zone */ 628 frame_t *frame = &zones.info[z1].frames[base_diff + frame_idx]; 629 frame->refcount = 0; 630 buddy_system_free(zones.info[z1].buddy_system, &frame->buddy_link); 425 } 426 427 for (size_t i = 0; i < zones.info[z2].count; i++) { 428 bitmap_set(&zones.info[z1].bitmap, base_diff + i, 429 bitmap_get(&zones.info[z2].bitmap, i)); 430 zones.info[z1].frames[base_diff + i] = 431 zones.info[z2].frames[i]; 631 432 } 632 433 } … … 651 452 size_t cframes = SIZE2FRAMES(zone_conf_size(count)); 652 453 653 if ((pfn < zones.info[znum].base) 654 ||(pfn >= zones.info[znum].base + zones.info[znum].count))454 if ((pfn < zones.info[znum].base) || 455 (pfn >= zones.info[znum].base + zones.info[znum].count)) 655 456 return; 656 457 657 frame_t *frame __attribute__ ((unused)); 658 659 frame = &zones.info[znum].frames[pfn - zones.info[znum].base]; 660 ASSERT(!frame->buddy_order); 661 662 size_t i; 663 for (i = 0; i < cframes; i++) { 664 zones.info[znum].busy_count++; 458 for (size_t i = 0; i < cframes; i++) 665 459 (void) zone_frame_free(&zones.info[znum], 666 460 pfn - zones.info[znum].base + i); 667 }668 }669 670 /** Reduce allocated block to count of order 0 frames.671 *672 * The allocated block needs 2^order frames. Reduce all frames673 * in the block to order 0 and free the unneeded frames. This means that674 * when freeing the previously allocated block starting with frame_idx,675 * you have to free every frame.676 *677 * @param znum Zone.678 * @param frame_idx Index the first frame of the block.679 * @param count Allocated frames in block.680 *681 */682 NO_TRACE static void zone_reduce_region(size_t znum, pfn_t frame_idx,683 size_t count)684 {685 ASSERT(zones.info[znum].flags & ZONE_AVAILABLE);686 ASSERT(frame_idx + count < zones.info[znum].count);687 688 uint8_t order = zones.info[znum].frames[frame_idx].buddy_order;689 ASSERT((size_t) (1 << order) >= count);690 691 /* Reduce all blocks to order 0 */692 size_t i;693 for (i = 0; i < (size_t) (1 << order); i++) {694 frame_t *frame = &zones.info[znum].frames[i + frame_idx];695 frame->buddy_order = 0;696 if (!frame->refcount)697 frame->refcount = 1;698 ASSERT(frame->refcount == 1);699 }700 701 /* Free unneeded frames */702 for (i = count; i < (size_t) (1 << order); i++)703 (void) zone_frame_free(&zones.info[znum], i + frame_idx);704 461 } 705 462 … … 721 478 bool ret = true; 722 479 723 /* We can join only 2 zones with none existing inbetween, 480 /* 481 * We can join only 2 zones with none existing inbetween, 724 482 * the zones have to be available and with the same 725 483 * set of flags … … 735 493 + zones.info[z2].count)); 736 494 737 uint8_t order;738 if (cframes == 1)739 order = 0;740 else741 order = fnzb(cframes - 1) + 1;742 743 495 /* Allocate merged zone data inside one of the zones */ 744 496 pfn_t pfn; 745 if (zone_can_alloc(&zones.info[z1], order)) { 746 pfn = zones.info[z1].base + zone_frame_alloc(&zones.info[z1], order); 747 } else if (zone_can_alloc(&zones.info[z2], order)) { 748 pfn = zones.info[z2].base + zone_frame_alloc(&zones.info[z2], order); 497 if (zone_can_alloc(&zones.info[z1], cframes, 0)) { 498 pfn = zones.info[z1].base + 499 zone_frame_alloc(&zones.info[z1], cframes, 0); 500 } else if (zone_can_alloc(&zones.info[z2], cframes, 0)) { 501 pfn = zones.info[z2].base + 502 zone_frame_alloc(&zones.info[z2], cframes, 0); 749 503 } else { 750 504 ret = false; … … 754 508 /* Preserve original data from z1 */ 755 509 zone_t old_z1 = zones.info[z1]; 756 old_z1.buddy_system->data = (void *) &old_z1;757 510 758 511 /* Do zone merging */ 759 buddy_system_t *buddy = (buddy_system_t *) PA2KA(PFN2ADDR(pfn)); 760 zone_merge_internal(z1, z2, &old_z1, buddy); 761 762 /* Free unneeded config frames */ 763 zone_reduce_region(z1, pfn - zones.info[z1].base, cframes); 512 zone_merge_internal(z1, z2, &old_z1, (void *) PA2KA(PFN2ADDR(pfn))); 764 513 765 514 /* Subtract zone information from busy frames */ … … 774 523 775 524 /* Move zones down */ 776 size_t i; 777 for (i = z2 + 1; i < zones.count; i++) { 525 for (size_t i = z2 + 1; i < zones.count; i++) 778 526 zones.info[i - 1] = zones.info[i]; 779 if (zones.info[i - 1].buddy_system != NULL)780 zones.info[i - 1].buddy_system->data =781 (void *) &zones.info[i - 1];782 }783 527 784 528 zones.count--; … … 799 543 void zone_merge_all(void) 800 544 { 801 size_t i = 0; 545 size_t i = 1; 546 802 547 while (i < zones.count) { 803 if (!zone_merge(i , i + 1))548 if (!zone_merge(i - 1, i)) 804 549 i++; 805 550 } … … 808 553 /** Create new frame zone. 809 554 * 810 * @param zone Zone to construct.811 * @param buddy Address of buddy system configuration information.812 * @param start Physical address of the first frame within thezone.813 * @param count Count of frames in zone.814 * @param flags Zone flags.555 * @param zone Zone to construct. 556 * @param start Physical address of the first frame within the zone. 557 * @param count Count of frames in zone. 558 * @param flags Zone flags. 559 * @param confdata Configuration data of the zone. 815 560 * 816 561 * @return Initialized zone. 817 562 * 818 563 */ 819 NO_TRACE static void zone_construct(zone_t *zone, buddy_system_t *buddy,820 pfn_t start, size_t count, zone_flags_t flags)564 NO_TRACE static void zone_construct(zone_t *zone, pfn_t start, size_t count, 565 zone_flags_t flags, void *confdata) 821 566 { 822 567 zone->base = start; … … 825 570 zone->free_count = count; 826 571 zone->busy_count = 0; 827 zone->buddy_system = buddy;828 572 829 573 if (flags & ZONE_AVAILABLE) { 830 574 /* 831 * Compute order for buddy system and initialize 575 * Initialize frame bitmap (located after the array of 576 * frame_t structures in the configuration space). 832 577 */ 833 uint8_t order = fnzb(count); 834 buddy_system_create(zone->buddy_system, order, 835 &zone_buddy_system_operations, (void *) zone); 836 837 /* Allocate frames _after_ the confframe */ 838 839 /* Check sizes */ 840 zone->frames = (frame_t *) ((uint8_t *) zone->buddy_system + 841 buddy_conf_size(order)); 842 843 size_t i; 844 for (i = 0; i < count; i++) 578 579 bitmap_initialize(&zone->bitmap, count, BITMAP_BLOCK_SIZE, 580 confdata + (sizeof(frame_t) * count)); 581 582 /* 583 * Initialize the array of frame_t structures. 584 */ 585 586 zone->frames = (frame_t *) confdata; 587 588 for (size_t i = 0; i < count; i++) 845 589 frame_initialize(&zone->frames[i]); 846 847 /* Stuffing frames */ 848 for (i = 0; i < count; i++) { 849 zone->frames[i].refcount = 0; 850 buddy_system_free(zone->buddy_system, &zone->frames[i].buddy_link); 851 } 852 } else 590 } else { 591 bitmap_initialize(&zone->bitmap, 0, 0, NULL); 853 592 zone->frames = NULL; 593 } 854 594 } 855 595 … … 863 603 size_t zone_conf_size(size_t count) 864 604 { 865 return (count * sizeof(frame_t) + buddy_conf_size(fnzb(count))); 605 return (count * sizeof(frame_t) + 606 bitmap_size(count, BITMAP_BLOCK_SIZE)); 866 607 } 867 608 … … 869 610 pfn_t zone_external_conf_alloc(size_t count) 870 611 { 871 size_t size = zone_conf_size(count); 872 size_t order = ispwr2(size) ? fnzb(size) : (fnzb(size) + 1); 873 874 return ADDR2PFN((uintptr_t) frame_alloc(order - FRAME_WIDTH, 875 FRAME_LOWMEM | FRAME_ATOMIC, 0)); 612 size_t frames = SIZE2FRAMES(zone_conf_size(count)); 613 614 return ADDR2PFN((uintptr_t) 615 frame_alloc(frames, FRAME_LOWMEM | FRAME_ATOMIC, 0)); 876 616 } 877 617 … … 881 621 * @param count Size of zone in frames. 882 622 * @param confframe Where configuration frames are supposed to be. 883 * Automatically checks ,that we will not disturb the623 * Automatically checks that we will not disturb the 884 624 * kernel and possibly init. If confframe is given 885 625 * _outside_ this zone, it is expected, that the area is … … 898 638 899 639 if (flags & ZONE_AVAILABLE) { /* Create available zone */ 900 /* Theoretically we could have NULL here, practically make sure 640 /* 641 * Theoretically we could have NULL here, practically make sure 901 642 * nobody tries to do that. If some platform requires, remove 902 643 * the assert 903 644 */ 904 645 ASSERT(confframe != ADDR2PFN((uintptr_t ) NULL)); 905 646 906 647 /* Update the known end of physical memory. */ 907 648 config.physmem_end = max(config.physmem_end, PFN2ADDR(start + count)); 908 649 909 /* If confframe is supposed to be inside our zone, then make sure 650 /* 651 * If confframe is supposed to be inside our zone, then make sure 910 652 * it does not span kernel & init 911 653 */ 912 654 size_t confcount = SIZE2FRAMES(zone_conf_size(count)); 655 913 656 if ((confframe >= start) && (confframe < start + count)) { 914 657 for (; confframe < start + count; confframe++) { … … 923 666 924 667 bool overlap = false; 925 size_t i; 926 for (i = 0; i < init.cnt; i++) 668 for (size_t i = 0; i < init.cnt; i++) { 927 669 if (overlaps(addr, PFN2ADDR(confcount), 928 670 init.tasks[i].paddr, … … 931 673 break; 932 674 } 675 } 676 933 677 if (overlap) 934 678 continue; … … 937 681 } 938 682 939 if (confframe >= start + count) { 940 flags &= ~ZONE_AVAILABLE; 941 goto nonavail; 942 // panic("Cannot find configuration data for zone."); 943 } 683 if (confframe >= start + count) 684 panic("Cannot find configuration data for zone."); 944 685 } 945 686 … … 950 691 } 951 692 952 buddy_system_t *buddy = (buddy_system_t*) PA2KA(PFN2ADDR(confframe));953 zone_construct(&zones.info[znum], buddy, start, count, flags);693 void *confdata = (void *) PA2KA(PFN2ADDR(confframe)); 694 zone_construct(&zones.info[znum], start, count, flags, confdata); 954 695 955 696 /* If confdata in zone, mark as unavailable */ 956 697 if ((confframe >= start) && (confframe < start + count)) { 957 size_t i; 958 for (i = confframe; i < confframe + confcount; i++) 698 for (size_t i = confframe; i < confframe + confcount; i++) 959 699 zone_mark_unavailable(&zones.info[znum], 960 700 i - zones.info[znum].base); … … 965 705 return znum; 966 706 } 967 nonavail: 968 (void)0; // label trick 707 969 708 /* Non-available zone */ 970 709 size_t znum = zones_insert_zone(start, count, flags); … … 973 712 return (size_t) -1; 974 713 } 975 zone_construct(&zones.info[znum], NULL, start, count, flags); 714 715 zone_construct(&zones.info[znum], start, count, flags, NULL); 976 716 977 717 irq_spinlock_unlock(&zones.lock, true); … … 1015 755 } 1016 756 1017 /** Allocate power-of-two frames of physical memory. 1018 * 1019 * @param order Allocate exactly 2^order frames. 1020 * @param flags Flags for host zone selection and address processing. 1021 * @param pzone Preferred zone. 757 /** Allocate frames of physical memory. 758 * 759 * @param count Number of continuous frames to allocate. 760 * @param flags Flags for host zone selection and address processing. 761 * @param constraint Indication of physical address bits that cannot be 762 * set in the address of the first allocated frame. 763 * @param pzone Preferred zone. 1022 764 * 1023 765 * @return Physical address of the allocated frame. 1024 766 * 1025 767 */ 1026 uintptr_t frame_alloc_generic( uint8_t order, frame_flags_t flags,768 uintptr_t frame_alloc_generic(size_t count, frame_flags_t flags, 1027 769 uintptr_t constraint, size_t *pzone) 1028 770 { 1029 size_t size = ((size_t) 1) << order; 771 ASSERT(count > 0); 772 1030 773 size_t hint = pzone ? (*pzone) : 0; 774 pfn_t frame_constraint = ADDR2PFN(constraint); 1031 775 1032 776 /* 1033 777 * If not told otherwise, we must first reserve the memory. 1034 778 */ 1035 if (!(flags & FRAME_NO_RESERVE)) 1036 reserve_force_alloc( size);1037 779 if (!(flags & FRAME_NO_RESERVE)) 780 reserve_force_alloc(count); 781 1038 782 loop: 1039 783 irq_spinlock_lock(&zones.lock, true); … … 1042 786 * First, find suitable frame zone. 1043 787 */ 1044 size_t znum = find_free_zone(order, 1045 FRAME_TO_ZONE_FLAGS(flags), hint); 1046 1047 /* If no memory, reclaim some slab memory, 1048 if it does not help, reclaim all */ 788 size_t znum = find_free_zone(count, FRAME_TO_ZONE_FLAGS(flags), 789 frame_constraint, hint); 790 791 /* 792 * If no memory, reclaim some slab memory, 793 * if it does not help, reclaim all. 794 */ 1049 795 if ((znum == (size_t) -1) && (!(flags & FRAME_NO_RECLAIM))) { 1050 796 irq_spinlock_unlock(&zones.lock, true); … … 1053 799 1054 800 if (freed > 0) 1055 znum = find_free_zone( order,1056 FRAME_TO_ZONE_FLAGS(flags), hint);801 znum = find_free_zone(count, FRAME_TO_ZONE_FLAGS(flags), 802 frame_constraint, hint); 1057 803 1058 804 if (znum == (size_t) -1) { … … 1062 808 1063 809 if (freed > 0) 1064 znum = find_free_zone( order,1065 FRAME_TO_ZONE_FLAGS(flags), hint);810 znum = find_free_zone(count, FRAME_TO_ZONE_FLAGS(flags), 811 frame_constraint, hint); 1066 812 } 1067 813 } … … 1070 816 if (flags & FRAME_ATOMIC) { 1071 817 irq_spinlock_unlock(&zones.lock, true); 818 1072 819 if (!(flags & FRAME_NO_RESERVE)) 1073 reserve_free(size); 820 reserve_free(count); 821 1074 822 return 0; 1075 823 } … … 1090 838 #ifdef CONFIG_DEBUG 1091 839 printf("Thread %" PRIu64 " waiting for %zu frames, " 1092 "%zu available.\n", THREAD->tid, size, avail);840 "%zu available.\n", THREAD->tid, count, avail); 1093 841 #endif 1094 842 1095 843 /* 1096 * Since the mem_avail_mtx is an active mutex, we need to disable interrupts1097 * to prevent deadlock with TLB shootdown.844 * Since the mem_avail_mtx is an active mutex, we need to 845 * disable interrupts to prevent deadlock with TLB shootdown. 1098 846 */ 1099 847 ipl_t ipl = interrupts_disable(); … … 1101 849 1102 850 if (mem_avail_req > 0) 1103 mem_avail_req = min(mem_avail_req, size);851 mem_avail_req = min(mem_avail_req, count); 1104 852 else 1105 mem_avail_req = size; 853 mem_avail_req = count; 854 1106 855 size_t gen = mem_avail_gen; 1107 856 … … 1119 868 } 1120 869 1121 pfn_t pfn = zone_frame_alloc(&zones.info[znum], order)1122 + zones.info[znum].base;870 pfn_t pfn = zone_frame_alloc(&zones.info[znum], count, 871 frame_constraint) + zones.info[znum].base; 1123 872 1124 873 irq_spinlock_unlock(&zones.lock, true); … … 1130 879 } 1131 880 1132 uintptr_t frame_alloc( uint8_t order, frame_flags_t flags, uintptr_t constraint)1133 { 1134 return frame_alloc_generic( order, flags, constraint, NULL);1135 } 1136 1137 uintptr_t frame_alloc_noreserve( uint8_t order, frame_flags_t flags,881 uintptr_t frame_alloc(size_t count, frame_flags_t flags, uintptr_t constraint) 882 { 883 return frame_alloc_generic(count, flags, constraint, NULL); 884 } 885 886 uintptr_t frame_alloc_noreserve(size_t count, frame_flags_t flags, 1138 887 uintptr_t constraint) 1139 888 { 1140 return frame_alloc_generic( order, flags | FRAME_NO_RESERVE, constraint,889 return frame_alloc_generic(count, flags | FRAME_NO_RESERVE, constraint, 1141 890 NULL); 1142 891 } … … 1154 903 void frame_free_generic(uintptr_t frame, frame_flags_t flags) 1155 904 { 1156 size_t size;1157 1158 905 irq_spinlock_lock(&zones.lock, true); 1159 906 … … 1163 910 pfn_t pfn = ADDR2PFN(frame); 1164 911 size_t znum = find_zone(pfn, 1, 0); 1165 912 1166 913 ASSERT(znum != (size_t) -1); 1167 914 1168 size = zone_frame_free(&zones.info[znum], pfn - zones.info[znum].base); 915 size_t size = 916 zone_frame_free(&zones.info[znum], pfn - zones.info[znum].base); 1169 917 1170 918 irq_spinlock_unlock(&zones.lock, true); … … 1173 921 * Signal that some memory has been freed. 1174 922 */ 1175 923 1176 924 1177 925 /* … … 1179 927 * to prevent deadlock with TLB shootdown. 1180 928 */ 929 1181 930 ipl_t ipl = interrupts_disable(); 1182 931 mutex_lock(&mem_avail_mtx); 932 1183 933 if (mem_avail_req > 0) 1184 934 mem_avail_req -= min(mem_avail_req, size); … … 1188 938 condvar_broadcast(&mem_avail_cv); 1189 939 } 940 1190 941 mutex_unlock(&mem_avail_mtx); 1191 942 interrupts_restore(ipl); … … 1236 987 irq_spinlock_lock(&zones.lock, true); 1237 988 1238 size_t i; 1239 for (i = 0; i < count; i++) { 989 for (size_t i = 0; i < count; i++) { 1240 990 size_t znum = find_zone(start + i, 1, 0); 991 1241 992 if (znum == (size_t) -1) /* PFN not found */ 1242 993 continue; … … 1263 1014 /* Tell the architecture to create some memory */ 1264 1015 frame_low_arch_init(); 1016 1265 1017 if (config.cpu_active == 1) { 1266 1018 frame_mark_unavailable(ADDR2PFN(KA2PA(config.base)), … … 1269 1021 SIZE2FRAMES(config.stack_size)); 1270 1022 1271 size_t i; 1272 for (i = 0; i < init.cnt; i++) { 1273 pfn_t pfn = ADDR2PFN(init.tasks[i].paddr); 1274 frame_mark_unavailable(pfn, 1023 for (size_t i = 0; i < init.cnt; i++) 1024 frame_mark_unavailable(ADDR2PFN(init.tasks[i].paddr), 1275 1025 SIZE2FRAMES(init.tasks[i].size)); 1276 }1277 1026 1278 1027 if (ballocs.size) … … 1280 1029 SIZE2FRAMES(ballocs.size)); 1281 1030 1282 /* Black list first frame, as allocating NULL would 1031 /* 1032 * Blacklist first frame, as allocating NULL would 1283 1033 * fail in some places 1284 1034 */ 1285 1035 frame_mark_unavailable(0, 1); 1286 1036 } 1037 1287 1038 frame_high_arch_init(); 1288 1039 } … … 1290 1041 /** Adjust bounds of physical memory region according to low/high memory split. 1291 1042 * 1292 * @param low[in] If true, the adjustment is performed to make the region 1293 * fit in the low memory. Otherwise the adjustment is 1294 * performed to make the region fit in the high memory. 1295 * @param basep[inout] Pointer to a variable which contains the region's base 1296 * address and which may receive the adjusted base address. 1297 * @param sizep[inout] Pointer to a variable which contains the region's size 1298 * and which may receive the adjusted size. 1299 * @retun True if the region still exists even after the 1300 * adjustment, false otherwise. 1043 * @param low[in] If true, the adjustment is performed to make the region 1044 * fit in the low memory. Otherwise the adjustment is 1045 * performed to make the region fit in the high memory. 1046 * @param basep[inout] Pointer to a variable which contains the region's base 1047 * address and which may receive the adjusted base address. 1048 * @param sizep[inout] Pointer to a variable which contains the region's size 1049 * and which may receive the adjusted size. 1050 * 1051 * @return True if the region still exists even after the adjustment. 1052 * @return False otherwise. 1053 * 1301 1054 */ 1302 1055 bool frame_adjust_zone_bounds(bool low, uintptr_t *basep, size_t *sizep) 1303 1056 { 1304 1057 uintptr_t limit = KA2PA(config.identity_base) + config.identity_size; 1305 1058 1306 1059 if (low) { 1307 1060 if (*basep > limit) 1308 1061 return false; 1062 1309 1063 if (*basep + *sizep > limit) 1310 1064 *sizep = limit - *basep; … … 1312 1066 if (*basep + *sizep <= limit) 1313 1067 return false; 1068 1314 1069 if (*basep <= limit) { 1315 1070 *sizep -= limit - *basep; … … 1317 1072 } 1318 1073 } 1074 1319 1075 return true; 1320 1076 } … … 1328 1084 1329 1085 uint64_t total = 0; 1330 size_t i;1331 for ( i = 0; i < zones.count; i++)1086 1087 for (size_t i = 0; i < zones.count; i++) 1332 1088 total += (uint64_t) FRAMES2SIZE(zones.info[i].count); 1333 1089 … … 1352 1108 *free = 0; 1353 1109 1354 size_t i; 1355 for (i = 0; i < zones.count; i++) { 1110 for (size_t i = 0; i < zones.count; i++) { 1356 1111 *total += (uint64_t) FRAMES2SIZE(zones.info[i].count); 1357 1112 … … 1390 1145 */ 1391 1146 1392 size_t i; 1393 for (i = 0;; i++) { 1147 for (size_t i = 0;; i++) { 1394 1148 irq_spinlock_lock(&zones.lock, true); 1395 1149 … … 1444 1198 size_t znum = (size_t) -1; 1445 1199 1446 size_t i; 1447 for (i = 0; i < zones.count; i++) { 1200 for (size_t i = 0; i < zones.count; i++) { 1448 1201 if ((i == num) || (PFN2ADDR(zones.info[i].base) == num)) { 1449 1202 znum = i; … … 1458 1211 } 1459 1212 1460 uintptr_t base = PFN2ADDR(zones.info[ i].base);1461 zone_flags_t flags = zones.info[ i].flags;1462 size_t count = zones.info[ i].count;1463 size_t free_count = zones.info[ i].free_count;1464 size_t busy_count = zones.info[ i].busy_count;1213 uintptr_t base = PFN2ADDR(zones.info[znum].base); 1214 zone_flags_t flags = zones.info[znum].flags; 1215 size_t count = zones.info[znum].count; 1216 size_t free_count = zones.info[znum].free_count; 1217 size_t busy_count = zones.info[znum].busy_count; 1465 1218 1466 1219 irq_spinlock_unlock(&zones.lock, true); -
kernel/generic/src/mm/km.c
r86733f3 rb0c2075 247 247 */ 248 248 uintptr_t page; 249 uintptr_t frame = frame_alloc(ONE_FRAME,250 FRAME_HIGHMEM | FRAME_ATOMIC | flags, 0);249 uintptr_t frame = 250 frame_alloc(1, FRAME_HIGHMEM | FRAME_ATOMIC | flags, 0); 251 251 if (frame) { 252 252 page = km_map(frame, PAGE_SIZE, … … 256 256 ASSERT(page); 257 257 } else { 258 frame = frame_alloc( ONE_FRAME, FRAME_LOWMEM | flags, 0);258 frame = frame_alloc(1, FRAME_LOWMEM | flags, 0); 259 259 if (!frame) 260 260 return (uintptr_t) NULL; -
kernel/generic/src/mm/slab.c
r86733f3 rb0c2075 183 183 184 184 void *data = (void *) 185 PA2KA(frame_alloc_generic(cache-> order, flags, 0, &zone));185 PA2KA(frame_alloc_generic(cache->frames, flags, 0, &zone)); 186 186 if (!data) 187 187 return NULL; … … 197 197 } 198 198 } else { 199 fsize = (PAGE_SIZE << cache->order);199 fsize = FRAMES2SIZE(cache->frames); 200 200 slab = data + fsize - sizeof(*slab); 201 201 } … … 203 203 /* Fill in slab structures */ 204 204 size_t i; 205 for (i = 0; i < ((size_t) 1 << cache->order); i++)205 for (i = 0; i < cache->frames; i++) 206 206 frame_set_parent(ADDR2PFN(KA2PA(data)) + i, slab, zone); 207 207 … … 231 231 atomic_dec(&cache->allocated_slabs); 232 232 233 return (1 << cache->order);233 return cache->frames; 234 234 } 235 235 … … 558 558 { 559 559 if (cache->flags & SLAB_CACHE_SLINSIDE) 560 return ( (PAGE_SIZE << cache->order)561 - sizeof(slab_t)) /cache->size;560 return (FRAMES2SIZE(cache->frames) - sizeof(slab_t)) / 561 cache->size; 562 562 else 563 return (PAGE_SIZE << cache->order) / cache->size;563 return FRAMES2SIZE(cache->frames) / cache->size; 564 564 } 565 565 … … 570 570 { 571 571 size_t objects = comp_objects(cache); 572 size_t ssize = PAGE_SIZE << cache->order;572 size_t ssize = FRAMES2SIZE(cache->frames); 573 573 574 574 if (cache->flags & SLAB_CACHE_SLINSIDE) … … 634 634 cache->flags |= SLAB_CACHE_SLINSIDE; 635 635 636 /* Minimum slab order */ 637 size_t pages = SIZE2FRAMES(cache->size); 638 639 /* We need the 2^order >= pages */ 640 if (pages == 1) 641 cache->order = 0; 642 else 643 cache->order = fnzb(pages - 1) + 1; 636 /* Minimum slab frames */ 637 cache->frames = SIZE2FRAMES(cache->size); 644 638 645 639 while (badness(cache) > SLAB_MAX_BADNESS(cache)) 646 cache-> order += 1;640 cache->frames <<= 1; 647 641 648 642 cache->objects = comp_objects(cache); … … 870 864 871 865 const char *name = cache->name; 872 uint8_t order = cache->order;866 size_t frames = cache->frames; 873 867 size_t size = cache->size; 874 868 size_t objects = cache->objects; … … 880 874 irq_spinlock_unlock(&slab_cache_lock, true); 881 875 882 printf("%-18s %8zu %8 u %8zu %8ld %8ld %8ld %-5s\n",883 name, size, (1 << order), objects, allocated_slabs,876 printf("%-18s %8zu %8zu %8zu %8ld %8ld %8ld %-5s\n", 877 name, size, frames, objects, allocated_slabs, 884 878 cached_objs, allocated_objs, 885 879 flags & SLAB_CACHE_SLINSIDE ? "in" : "out"); -
kernel/generic/src/proc/thread.c
r86733f3 rb0c2075 193 193 194 194 thread->kstack = (uint8_t *) 195 PA2KA(frame_alloc(STACK_FRAMES, kmflags, 0));195 PA2KA(frame_alloc(STACK_FRAMES, kmflags, STACK_SIZE - 1)); 196 196 if (!thread->kstack) { 197 197 #ifdef CONFIG_FPU -
kernel/generic/src/time/clock.c
r86733f3 rb0c2075 81 81 void clock_counter_init(void) 82 82 { 83 uintptr_t faddr = frame_alloc( ONE_FRAME, FRAME_ATOMIC, 0);83 uintptr_t faddr = frame_alloc(1, FRAME_ATOMIC, 0); 84 84 if (faddr == 0) 85 85 panic("Cannot allocate page for clock."); -
kernel/test/mm/falloc1.c
r86733f3 rb0c2075 37 37 #include <align.h> 38 38 39 #define MAX_FRAMES 1024 U39 #define MAX_FRAMES 1024 40 40 #define MAX_ORDER 8 41 41 #define TEST_RUNS 2 … … 51 51 return "Unable to allocate frames"; 52 52 53 unsigned int results[MAX_ORDER + 1]; 53 unsigned int results[MAX_FRAMES + 1]; 54 54 55 for (unsigned int run = 0; run < TEST_RUNS; run++) { 55 for (unsigned int order = 0; order <= MAX_ORDER; order++) { 56 TPRINTF("Allocating %u frames blocks ... ", 1 << order); 56 for (size_t count = 1; count <= MAX_FRAMES; count++) { 57 size_t bytes = FRAMES2SIZE(count); 58 59 TPRINTF("Allocating %zu frames blocks (%zu bytes) ... ", 60 count, bytes); 57 61 58 62 unsigned int allocated = 0; 59 for (unsigned int i = 0; i < (MAX_FRAMES >> order); i++) {63 for (unsigned int i = 0; i < (MAX_FRAMES / count); i++) { 60 64 frames[allocated] = 61 PA2KA(frame_alloc(order, FRAME_ATOMIC, 0)); 62 63 if (ALIGN_UP(frames[allocated], FRAME_SIZE << order) != 64 frames[allocated]) { 65 TPRINTF("Block at address %p (size %u) is not aligned\n", 66 (void *) frames[allocated], (FRAME_SIZE << order) >> 10); 67 return "Test failed"; 68 } 65 PA2KA(frame_alloc(count, FRAME_ATOMIC, 0)); 69 66 70 67 if (frames[allocated]) … … 78 75 TPRINTF("%d blocks allocated.\n", allocated); 79 76 80 if (run ) {81 if (results[ order] != allocated)77 if (run > 0) { 78 if (results[count] != allocated) 82 79 return "Possible frame leak"; 83 80 } else 84 results[ order] = allocated;81 results[count] = allocated; 85 82 86 83 TPRINTF("Deallocating ... "); -
kernel/test/mm/falloc2.c
r86733f3 rb0c2075 40 40 #include <arch.h> 41 41 42 #define MAX_FRAMES 256U 43 #define MAX_ORDER 8 42 #define MAX_FRAMES 256 44 43 45 44 #define THREAD_RUNS 1 … … 66 65 67 66 for (unsigned int run = 0; run < THREAD_RUNS; run++) { 68 for (unsigned int order = 0; order <= MAX_ORDER; order++) { 67 for (size_t count = 1; count <= MAX_FRAMES; count++) { 68 size_t bytes = FRAMES2SIZE(count); 69 69 70 TPRINTF("Thread #%" PRIu64 " (cpu%u): " 70 "Allocating % u frames blocks... \n", THREAD->tid,71 CPU->id, 1 << order);71 "Allocating %zu frames blocks (%zu bytes) ... \n", THREAD->tid, 72 CPU->id, count, bytes); 72 73 73 74 unsigned int allocated = 0; 74 for (unsigned int i = 0; i < (MAX_FRAMES >> order); i++) {75 for (unsigned int i = 0; i < (MAX_FRAMES / count); i++) { 75 76 frames[allocated] = 76 PA2KA(frame_alloc( order, FRAME_ATOMIC, 0));77 PA2KA(frame_alloc(count, FRAME_ATOMIC, 0)); 77 78 if (frames[allocated]) { 78 memsetb((void *) frames[allocated], FRAME_SIZE << order, val);79 memsetb((void *) frames[allocated], bytes, val); 79 80 allocated++; 80 81 } else … … 89 90 90 91 for (unsigned int i = 0; i < allocated; i++) { 91 for (size_t k = 0; k <= (((size_t) FRAME_SIZE << order) - 1); 92 k++) { 92 for (size_t k = 0; k < bytes; k++) { 93 93 if (((uint8_t *) frames[i])[k] != val) { 94 94 TPRINTF("Thread #%" PRIu64 " (cpu%u): " -
kernel/test/mm/mapping1.c
r86733f3 rb0c2075 41 41 const char *test_mapping1(void) 42 42 { 43 uintptr_t page0, page1; 44 uint32_t v; 43 uintptr_t frame = frame_alloc(1, FRAME_NONE, 0); 45 44 46 uintptr_t frame = frame_alloc(ONE_FRAME, FRAME_NONE, 0); 47 48 page0 = km_map(frame, FRAME_SIZE, 45 uintptr_t page0 = km_map(frame, FRAME_SIZE, 49 46 PAGE_READ | PAGE_WRITE | PAGE_CACHEABLE); 50 47 TPRINTF("Virtual address %p mapped to physical address %p.\n", 51 48 (void *) page0, (void *) frame); 52 page1 = km_map(frame, FRAME_SIZE, 49 50 uintptr_t page1 = km_map(frame, FRAME_SIZE, 53 51 PAGE_READ | PAGE_WRITE | PAGE_CACHEABLE); 54 52 TPRINTF("Virtual address %p mapped to physical address %p.\n", … … 62 60 TPRINTF("Reading magic using the second virtual address.\n"); 63 61 64 v = *((uint32_t *) page1);62 uint32_t v = *((uint32_t *) page1); 65 63 66 64 if (v != TEST_MAGIC) {
Note:
See TracChangeset
for help on using the changeset viewer.