drop safemap code
[minix.git] / servers / vm / region.c
blobb37729f38050745465749e858e90249005d269b9
2 #include <minix/com.h>
3 #include <minix/callnr.h>
4 #include <minix/type.h>
5 #include <minix/config.h>
6 #include <minix/const.h>
7 #include <minix/sysutil.h>
8 #include <minix/syslib.h>
9 #include <minix/debug.h>
10 #include <minix/bitmap.h>
11 #include <minix/hash.h>
12 #include <machine/multiboot.h>
14 #include <sys/mman.h>
16 #include <limits.h>
17 #include <string.h>
18 #include <assert.h>
19 #include <stdint.h>
20 #include <memory.h>
21 #include <sys/param.h>
23 #include "vm.h"
24 #include "proto.h"
25 #include "util.h"
26 #include "glo.h"
27 #include "region.h"
28 #include "sanitycheck.h"
29 #include "physravl.h"
30 #include "memlist.h"
31 #include "memtype.h"
33 /* LRU list. */
34 static yielded_t *lru_youngest = NULL, *lru_oldest = NULL;
36 static int map_ph_writept(struct vmproc *vmp, struct vir_region *vr,
37 struct phys_region *pr);
39 static phys_bytes freeyieldednode(yielded_t *node, int freemem);
41 static struct vir_region *map_copy_region(struct vmproc *vmp, struct
42 vir_region *vr);
44 #if SANITYCHECKS
45 static void lrucheck(void);
46 #endif
48 /* hash table of yielded blocks */
49 #define YIELD_HASHSIZE 65536
50 static yielded_avl vm_yielded_blocks[YIELD_HASHSIZE];
52 static int avl_inited = 0;
54 void map_region_init(void)
56 int h;
57 assert(!avl_inited);
58 for(h = 0; h < YIELD_HASHSIZE; h++)
59 yielded_init(&vm_yielded_blocks[h]);
60 avl_inited = 1;
63 static yielded_avl *get_yielded_avl(block_id_t id)
65 u32_t h;
67 assert(avl_inited);
69 hash_i_64(id.owner, id.id, h);
70 h = h % YIELD_HASHSIZE;
72 assert(h >= 0);
73 assert(h < YIELD_HASHSIZE);
75 return &vm_yielded_blocks[h];
78 void map_printregion(struct vmproc *vmp, struct vir_region *vr)
80 physr_iter iter;
81 struct phys_region *ph;
82 printf("map_printmap: map_name: %s\n", vr->memtype->name);
83 printf("\t%lx (len 0x%lx, %lukB), %p\n",
84 vr->vaddr, vr->length, vr->length/1024, vr->memtype->name);
85 printf("\t\tphysblocks:\n");
86 physr_start_iter_least(vr->phys, &iter);
87 while((ph = physr_get_iter(&iter))) {
88 printf("\t\t@ %lx (refs %d): phys 0x%lx\n",
89 (vr->vaddr + ph->offset),
90 ph->ph->refcount, ph->ph->phys);
91 physr_incr_iter(&iter);
95 /*===========================================================================*
96 * map_printmap *
97 *===========================================================================*/
98 void map_printmap(vmp)
99 struct vmproc *vmp;
101 struct vir_region *vr;
102 region_iter iter;
104 printf("memory regions in process %d:\n", vmp->vm_endpoint);
106 region_start_iter_least(&vmp->vm_regions_avl, &iter);
107 while((vr = region_get_iter(&iter))) {
108 map_printregion(vmp, vr);
109 region_incr_iter(&iter);
113 static struct vir_region *getnextvr(struct vir_region *vr)
115 struct vir_region *nextvr;
116 region_iter v_iter;
117 SLABSANE(vr);
118 region_start_iter(&vr->parent->vm_regions_avl, &v_iter, vr->vaddr, AVL_EQUAL);
119 assert(region_get_iter(&v_iter));
120 assert(region_get_iter(&v_iter) == vr);
121 region_incr_iter(&v_iter);
122 nextvr = region_get_iter(&v_iter);
123 if(!nextvr) return NULL;
124 SLABSANE(nextvr);
125 assert(vr->parent == nextvr->parent);
126 assert(vr->vaddr < nextvr->vaddr);
127 assert(vr->vaddr + vr->length <= nextvr->vaddr);
128 return nextvr;
131 int pr_writable(struct vir_region *vr, struct phys_region *pr)
133 assert(vr->memtype->writable);
134 return ((vr->flags & VR_WRITABLE) && vr->memtype->writable(pr));
137 #if SANITYCHECKS
139 /*===========================================================================*
140 * map_sanitycheck_pt *
141 *===========================================================================*/
142 static int map_sanitycheck_pt(struct vmproc *vmp,
143 struct vir_region *vr, struct phys_region *pr)
145 struct phys_block *pb = pr->ph;
146 int rw;
147 int r;
149 if(pr_writable(vr, pr))
150 rw = PTF_WRITE;
151 else
152 rw = PTF_READ;
154 r = pt_writemap(vmp, &vmp->vm_pt, vr->vaddr + pr->offset,
155 pb->phys, VM_PAGE_SIZE, PTF_PRESENT | PTF_USER | rw, WMF_VERIFY);
157 if(r != OK) {
158 printf("proc %d phys_region 0x%lx sanity check failed\n",
159 vmp->vm_endpoint, pr->offset);
160 map_printregion(vmp, vr);
163 return r;
166 /*===========================================================================*
167 * map_sanitycheck *
168 *===========================================================================*/
169 void map_sanitycheck(char *file, int line)
171 struct vmproc *vmp;
173 lrucheck();
175 /* Macro for looping over all physical blocks of all regions of
176 * all processes.
178 #define ALLREGIONS(regioncode, physcode) \
179 for(vmp = vmproc; vmp < &vmproc[VMP_NR]; vmp++) { \
180 region_iter v_iter; \
181 struct vir_region *vr; \
182 if(!(vmp->vm_flags & VMF_INUSE)) \
183 continue; \
184 region_start_iter_least(&vmp->vm_regions_avl, &v_iter); \
185 while((vr = region_get_iter(&v_iter))) { \
186 physr_iter iter; \
187 struct phys_region *pr; \
188 regioncode; \
189 physr_start_iter_least(vr->phys, &iter); \
190 while((pr = physr_get_iter(&iter))) { \
191 physcode; \
192 physr_incr_iter(&iter); \
194 region_incr_iter(&v_iter); \
198 #define MYSLABSANE(s) MYASSERT(slabsane_f(__FILE__, __LINE__, s, sizeof(*(s))))
199 /* Basic pointers check. */
200 ALLREGIONS(MYSLABSANE(vr),MYSLABSANE(pr); MYSLABSANE(pr->ph);MYSLABSANE(pr->parent));
201 ALLREGIONS(/* MYASSERT(vr->parent == vmp) */,MYASSERT(pr->parent == vr););
203 /* Do counting for consistency check. */
204 ALLREGIONS(;,USE(pr->ph, pr->ph->seencount = 0;););
205 ALLREGIONS(;,USE(pr->ph, pr->ph->seencount++;);
206 if(pr->ph->seencount == 1) {
207 if(pr->parent->memtype->ev_sanitycheck)
208 pr->parent->memtype->ev_sanitycheck(pr, file, line);
212 /* Do consistency check. */
213 ALLREGIONS({ struct vir_region *nextvr = getnextvr(vr);
214 if(nextvr) {
215 MYASSERT(vr->vaddr < nextvr->vaddr);
216 MYASSERT(vr->vaddr + vr->length <= nextvr->vaddr);
219 MYASSERT(!(vr->vaddr % VM_PAGE_SIZE));,
220 if(pr->ph->refcount != pr->ph->seencount) {
221 map_printmap(vmp);
222 printf("ph in vr %p: 0x%lx refcount %u "
223 "but seencount %u\n",
224 vr, pr->offset,
225 pr->ph->refcount, pr->ph->seencount);
228 int n_others = 0;
229 struct phys_region *others;
230 if(pr->ph->refcount > 0) {
231 MYASSERT(pr->ph->firstregion);
232 if(pr->ph->refcount == 1) {
233 MYASSERT(pr->ph->firstregion == pr);
235 } else {
236 MYASSERT(!pr->ph->firstregion);
238 for(others = pr->ph->firstregion; others;
239 others = others->next_ph_list) {
240 MYSLABSANE(others);
241 MYASSERT(others->ph == pr->ph);
242 n_others++;
244 MYASSERT(pr->ph->refcount == n_others);
246 MYASSERT(pr->ph->refcount == pr->ph->seencount);
247 MYASSERT(!(pr->offset % VM_PAGE_SIZE)););
248 ALLREGIONS(,MYASSERT(map_sanitycheck_pt(vmp, vr, pr) == OK));
251 #define LRUCHECK lrucheck()
253 static void lrucheck(void)
255 yielded_t *list;
257 /* list is empty and ok if both ends point to null. */
258 if(!lru_youngest && !lru_oldest)
259 return;
261 /* if not, both should point to something. */
262 SLABSANE(lru_youngest);
263 SLABSANE(lru_oldest);
265 assert(!lru_youngest->younger);
266 assert(!lru_oldest->older);
268 for(list = lru_youngest; list; list = list->older) {
269 SLABSANE(list);
270 if(list->younger) {
271 SLABSANE(list->younger);
272 assert(list->younger->older == list);
273 } else assert(list == lru_youngest);
274 if(list->older) {
275 SLABSANE(list->older);
276 assert(list->older->younger == list);
277 } else assert(list == lru_oldest);
281 void blockstats(void)
283 yielded_t *list;
284 int blocks = 0;
285 phys_bytes mem = 0;
286 clock_t ticks;
287 int s;
289 s = getuptime(&ticks);
291 assert(s == OK);
293 LRUCHECK;
295 for(list = lru_youngest; list; list = list->older) {
296 mem += VM_PAGE_SIZE;
297 blocks++;
300 if(blocks > 0)
301 printf("%d blocks, %lukB; ", blocks, mem/1024);
303 printmemstats();
305 #else
306 #define LRUCHECK
307 #endif
310 /*=========================================================================*
311 * map_ph_writept *
312 *=========================================================================*/
313 static int map_ph_writept(struct vmproc *vmp, struct vir_region *vr,
314 struct phys_region *pr)
316 int rw;
317 struct phys_block *pb = pr->ph;
319 assert(vr);
320 assert(pr);
321 assert(pb);
323 assert(!(vr->vaddr % VM_PAGE_SIZE));
324 assert(!(pr->offset % VM_PAGE_SIZE));
325 assert(pb->refcount > 0);
327 if(pr_writable(vr, pr))
328 rw = PTF_WRITE;
329 else
330 rw = PTF_READ;
332 if(pt_writemap(vmp, &vmp->vm_pt, vr->vaddr + pr->offset,
333 pb->phys, VM_PAGE_SIZE, PTF_PRESENT | PTF_USER | rw,
334 #if SANITYCHECKS
335 !pr->written ? 0 :
336 #endif
337 WMF_OVERWRITE) != OK) {
338 printf("VM: map_writept: pt_writemap failed\n");
339 return ENOMEM;
342 #if SANITYCHECKS
343 USE(pr, pr->written = 1;);
344 #endif
346 return OK;
349 #define SLOT_FAIL ((vir_bytes) -1)
351 /*===========================================================================*
352 * region_find_slot_range *
353 *===========================================================================*/
354 static vir_bytes region_find_slot_range(struct vmproc *vmp,
355 vir_bytes minv, vir_bytes maxv, vir_bytes length)
357 struct vir_region *lastregion;
358 vir_bytes startv = 0;
359 int foundflag = 0;
360 region_iter iter;
362 SANITYCHECK(SCL_FUNCTIONS);
364 /* Length must be reasonable. */
365 assert(length > 0);
367 /* Special case: allow caller to set maxv to 0 meaning 'I want
368 * it to be mapped in right here.'
370 if(maxv == 0) {
371 maxv = minv + length;
373 /* Sanity check. */
374 if(maxv <= minv) {
375 printf("region_find_slot: minv 0x%lx and bytes 0x%lx\n",
376 minv, length);
377 return SLOT_FAIL;
381 /* Basic input sanity checks. */
382 assert(!(length % VM_PAGE_SIZE));
383 if(minv >= maxv) {
384 printf("VM: 1 minv: 0x%lx maxv: 0x%lx length: 0x%lx\n",
385 minv, maxv, length);
388 assert(minv < maxv);
390 if(minv + length > maxv)
391 return SLOT_FAIL;
393 #define FREEVRANGE_TRY(rangestart, rangeend) { \
394 vir_bytes frstart = (rangestart), frend = (rangeend); \
395 frstart = MAX(frstart, minv); \
396 frend = MIN(frend, maxv); \
397 if(frend > frstart && (frend - frstart) >= length) { \
398 startv = frend-length; \
399 foundflag = 1; \
402 #define FREEVRANGE(start, end) { \
403 assert(!foundflag); \
404 FREEVRANGE_TRY(((start)+VM_PAGE_SIZE), ((end)-VM_PAGE_SIZE)); \
405 if(!foundflag) { \
406 FREEVRANGE_TRY((start), (end)); \
410 /* find region after maxv. */
411 region_start_iter(&vmp->vm_regions_avl, &iter, maxv, AVL_GREATER_EQUAL);
412 lastregion = region_get_iter(&iter);
414 if(!lastregion) {
415 /* This is the free virtual address space after the last region. */
416 region_start_iter(&vmp->vm_regions_avl, &iter, maxv, AVL_LESS);
417 lastregion = region_get_iter(&iter);
418 FREEVRANGE(lastregion ?
419 lastregion->vaddr+lastregion->length : 0, VM_DATATOP);
422 if(!foundflag) {
423 struct vir_region *vr;
424 while((vr = region_get_iter(&iter)) && !foundflag) {
425 struct vir_region *nextvr;
426 region_decr_iter(&iter);
427 nextvr = region_get_iter(&iter);
428 FREEVRANGE(nextvr ? nextvr->vaddr+nextvr->length : 0,
429 vr->vaddr);
433 if(!foundflag) {
434 printf("VM: region_find_slot: no 0x%lx bytes found for %d between 0x%lx and 0x%lx\n",
435 length, vmp->vm_endpoint, minv, maxv);
436 util_stacktrace();
437 return SLOT_FAIL;
440 /* However we got it, startv must be in the requested range. */
441 assert(startv >= minv);
442 assert(startv < maxv);
443 assert(startv + length <= maxv);
445 /* remember this position as a hint for next time. */
446 vmp->vm_region_top = startv + length;
448 return startv;
451 /*===========================================================================*
452 * region_find_slot *
453 *===========================================================================*/
454 static vir_bytes region_find_slot(struct vmproc *vmp,
455 vir_bytes minv, vir_bytes maxv, vir_bytes length)
457 vir_bytes v, hint = vmp->vm_region_top;
459 /* use the top of the last inserted region as a minv hint if
460 * possible. remember that a zero maxv is a special case.
463 if(maxv && hint < maxv && hint >= minv) {
464 v = region_find_slot_range(vmp, minv, hint, length);
466 if(v != SLOT_FAIL)
467 return v;
470 return region_find_slot_range(vmp, minv, maxv, length);
473 struct vir_region *region_new(struct vmproc *vmp, vir_bytes startv, vir_bytes length,
474 int flags, mem_type_t *memtype)
476 physr_avl *phavl;
477 struct vir_region *newregion;
478 static u32_t id;
480 if(!(SLABALLOC(newregion))) {
481 printf("vm: region_new: could not allocate\n");
482 return NULL;
485 /* Fill in node details. */
486 USE(newregion,
487 memset(newregion, 0, sizeof(*newregion));
488 newregion->vaddr = startv;
489 newregion->length = length;
490 newregion->flags = flags;
491 newregion->memtype = memtype;
492 newregion->remaps = 0;
493 newregion->id = id++;
494 newregion->lower = newregion->higher = NULL;
495 newregion->parent = vmp;);
497 SLABALLOC(phavl);
498 if(!phavl) {
499 printf("VM: region_new: allocating phys avl failed\n");
500 SLABFREE(newregion);
501 return NULL;
503 USE(newregion, newregion->phys = phavl;);
504 physr_init(newregion->phys);
506 return newregion;
509 /*===========================================================================*
510 * map_page_region *
511 *===========================================================================*/
512 struct vir_region *map_page_region(vmp, minv, maxv, length,
513 flags, mapflags, memtype)
514 struct vmproc *vmp;
515 vir_bytes minv;
516 vir_bytes maxv;
517 vir_bytes length;
518 u32_t flags;
519 int mapflags;
520 mem_type_t *memtype;
522 struct vir_region *newregion;
523 vir_bytes startv;
525 assert(!(length % VM_PAGE_SIZE));
527 SANITYCHECK(SCL_FUNCTIONS);
529 startv = region_find_slot(vmp, minv, maxv, length);
530 if (startv == SLOT_FAIL)
531 return NULL;
533 /* Now we want a new region. */
534 if(!(newregion = region_new(vmp, startv, length, flags, memtype))) {
535 printf("VM: map_page_region: allocating region failed\n");
536 return NULL;
539 /* If a new event is specified, invoke it. */
540 if(newregion->memtype->ev_new)
541 newregion->memtype->ev_new(newregion);
543 if(mapflags & MF_PREALLOC) {
544 if(map_handle_memory(vmp, newregion, 0, length, 1) != OK) {
545 printf("VM: map_page_region: prealloc failed\n");
546 USE(newregion,
547 SLABFREE(newregion->phys););
548 SLABFREE(newregion);
549 return NULL;
553 /* Pre-allocations should be uninitialized, but after that it's a
554 * different story.
556 USE(newregion, newregion->flags &= ~VR_UNINITIALIZED;);
558 /* Link it. */
559 region_insert(&vmp->vm_regions_avl, newregion);
561 #if SANITYCHECKS
562 assert(startv == newregion->vaddr);
564 struct vir_region *nextvr;
565 if((nextvr = getnextvr(newregion))) {
566 assert(newregion->vaddr < nextvr->vaddr);
569 #endif
571 SANITYCHECK(SCL_FUNCTIONS);
573 return newregion;
576 /*===========================================================================*
577 * map_subfree *
578 *===========================================================================*/
579 static int map_subfree(struct vir_region *region,
580 vir_bytes start, vir_bytes len)
582 struct phys_region *pr;
583 physr_iter iter;
584 vir_bytes end = start+len;
586 int full = 0;
588 #if SANITYCHECKS
590 SLABSANE(region);
591 SLABSANE(region->phys);
592 physr_start_iter_least(region->phys, &iter);
593 while((pr = physr_get_iter(&iter))) {
594 struct phys_region *others;
595 struct phys_block *pb;
597 pb = pr->ph;
599 for(others = pb->firstregion; others;
600 others = others->next_ph_list) {
601 assert(others->ph == pb);
603 physr_incr_iter(&iter);
606 #endif
608 if(start == 0 && len == region->length)
609 full = 1;
611 physr_init_iter(&iter);
612 physr_start_iter(region->phys, &iter, start, AVL_GREATER_EQUAL);
613 while((pr = physr_get_iter(&iter))) {
614 physr_incr_iter(&iter);
615 if(pr->offset >= end)
616 break;
617 pb_unreferenced(region, pr, !full);
618 if(!full) {
619 physr_start_iter(region->phys, &iter,
620 pr->offset, AVL_GREATER_EQUAL);
622 SLABFREE(pr);
625 if(full)
626 physr_init(region->phys);
628 return OK;
631 /*===========================================================================*
632 * map_free *
633 *===========================================================================*/
634 int map_free(struct vir_region *region)
636 int r;
638 if((r=map_subfree(region, 0, region->length)) != OK) {
639 printf("%d\n", __LINE__);
640 return r;
643 if(region->memtype->ev_delete)
644 region->memtype->ev_delete(region);
646 USE(region,
647 SLABFREE(region->phys););
648 SLABFREE(region);
650 return OK;
653 /*===========================================================================*
654 * yielded_block_cmp *
655 *===========================================================================*/
656 int yielded_block_cmp(struct block_id *id1, struct block_id *id2)
658 if(id1->owner < id2->owner)
659 return -1;
660 if(id1->owner > id2->owner)
661 return 1;
662 return cmp64(id1->id, id2->id);
666 /*===========================================================================*
667 * free_yielded_proc *
668 *===========================================================================*/
669 static vir_bytes free_yielded_proc(struct vmproc *vmp)
671 vir_bytes total = 0;
672 int h;
674 SANITYCHECK(SCL_FUNCTIONS);
676 /* Free associated regions. */
677 for(h = 0; h < YIELD_HASHSIZE && vmp->vm_yielded > 0; h++) {
678 yielded_t *yb;
679 yielded_iter iter;
680 yielded_avl *avl = &vm_yielded_blocks[h];
681 yielded_start_iter_least(avl, &iter);
682 while((yb = yielded_get_iter(&iter))) {
683 yielded_t *next_yb;
684 SLABSANE(yb);
685 yielded_incr_iter(&iter);
686 if(yb->id.owner != vmp->vm_endpoint)
687 continue;
688 next_yb = yielded_get_iter(&iter);
689 total += freeyieldednode(yb, 1);
690 /* the above removal invalidated our iter; restart it
691 * for the node we want to start at.
693 if(!next_yb) break;
694 yielded_start_iter(avl, &iter, next_yb->id, AVL_EQUAL);
695 assert(yielded_get_iter(&iter) == next_yb);
699 return total;
703 static phys_bytes freeyieldednode(yielded_t *node, int freemem)
705 yielded_t *older, *younger, *removed;
706 yielded_avl *avl;
707 int p;
709 SLABSANE(node);
711 LRUCHECK;
713 /* Update LRU. */
715 younger = node->younger;
716 older = node->older;
718 if(younger) {
719 SLABSANE(younger);
720 assert(younger->older == node);
721 USE(younger, younger->older = node->older;);
722 } else {
723 assert(node == lru_youngest);
724 lru_youngest = node->older;
727 if(older) {
728 SLABSANE(older);
729 assert(older->younger == node);
730 USE(older, older->younger = node->younger;);
731 } else {
732 assert(node == lru_oldest);
733 lru_oldest = node->younger;
736 LRUCHECK;
738 /* Update AVL. */
740 if(vm_isokendpt(node->id.owner, &p) != OK)
741 panic("out of date owner of yielded block %d", node->id.owner);
742 avl = get_yielded_avl(node->id);
743 removed = yielded_remove(avl, node->id);
744 assert(removed == node);
745 assert(vmproc[p].vm_yielded > 0);
746 vmproc[p].vm_yielded--;
748 /* Free associated memory if requested. */
750 if(freemem) {
751 free_mem(ABS2CLICK(node->physaddr), node->pages);
754 /* Free node. */
755 SLABFREE(node);
757 return VM_PAGE_SIZE;
760 /*========================================================================*
761 * free_yielded *
762 *========================================================================*/
763 vir_bytes free_yielded(vir_bytes max_bytes)
766 /* PRIVATE yielded_t *lru_youngest = NULL, *lru_oldest = NULL; */
767 vir_bytes freed = 0;
768 int blocks = 0;
770 while(freed < max_bytes && lru_oldest) {
771 SLABSANE(lru_oldest);
772 freed += freeyieldednode(lru_oldest, 1);
773 blocks++;
776 return freed;
779 /*========================================================================*
780 * map_free_proc *
781 *========================================================================*/
782 int map_free_proc(vmp)
783 struct vmproc *vmp;
785 struct vir_region *r;
787 while((r = region_search_root(&vmp->vm_regions_avl))) {
788 SANITYCHECK(SCL_DETAIL);
789 #if SANITYCHECKS
790 nocheck++;
791 #endif
792 region_remove(&vmp->vm_regions_avl, r->vaddr); /* For sanity checks. */
793 map_free(r);
794 #if SANITYCHECKS
795 nocheck--;
796 #endif
797 SANITYCHECK(SCL_DETAIL);
800 region_init(&vmp->vm_regions_avl);
802 /* Free associated yielded blocks. */
803 free_yielded_proc(vmp);
805 SANITYCHECK(SCL_FUNCTIONS);
807 return OK;
810 /*===========================================================================*
811 * map_lookup *
812 *===========================================================================*/
813 struct vir_region *map_lookup(vmp, offset, physr)
814 struct vmproc *vmp;
815 vir_bytes offset;
816 struct phys_region **physr;
818 struct vir_region *r;
820 SANITYCHECK(SCL_FUNCTIONS);
822 #if SANITYCHECKS
823 if(!region_search_root(&vmp->vm_regions_avl))
824 panic("process has no regions: %d", vmp->vm_endpoint);
825 #endif
827 if((r = region_search(&vmp->vm_regions_avl, offset, AVL_LESS_EQUAL))) {
828 vir_bytes ph;
829 if(offset >= r->vaddr && offset < r->vaddr + r->length) {
830 ph = offset - r->vaddr;
831 if(physr) {
832 *physr = physr_search(r->phys, ph, AVL_EQUAL);
833 if(*physr) assert((*physr)->offset == ph);
835 return r;
839 SANITYCHECK(SCL_FUNCTIONS);
841 return NULL;
844 u32_t vrallocflags(u32_t flags)
846 u32_t allocflags = 0;
848 if(flags & VR_PHYS64K)
849 allocflags |= PAF_ALIGN64K;
850 if(flags & VR_LOWER16MB)
851 allocflags |= PAF_LOWER16MB;
852 if(flags & VR_LOWER1MB)
853 allocflags |= PAF_LOWER1MB;
854 if(flags & VR_CONTIG)
855 allocflags |= PAF_CONTIG;
856 if(!(flags & VR_UNINITIALIZED))
857 allocflags |= PAF_CLEAR;
859 return allocflags;
862 /*===========================================================================*
863 * map_clone_ph_block *
864 *===========================================================================*/
865 struct phys_region *map_clone_ph_block(vmp, region, ph, iter)
866 struct vmproc *vmp;
867 struct vir_region *region;
868 struct phys_region *ph;
869 physr_iter *iter;
871 vir_bytes offset;
872 u32_t allocflags;
873 phys_bytes physaddr;
874 struct phys_region *newpr;
875 int region_has_single_block;
876 SANITYCHECK(SCL_FUNCTIONS);
878 /* Warning: this function will free the passed
879 * phys_region *ph and replace it (in the same offset)
880 * with another! So both the pointer to it
881 * and any iterators over the phys_regions in the vir_region
882 * will be invalid on successful return. (Iterators over
883 * the vir_region could be invalid on unsuccessful return too.)
886 /* This is only to be done if there is more than one copy. */
887 assert(ph->ph->refcount > 1);
889 /* This function takes a physical block, copies its contents
890 * into newly allocated memory, and replaces the single physical
891 * block by one or more physical blocks with refcount 1 with the
892 * same contents as the original. In other words, a fragmentable
893 * version of map_copy_ph_block().
896 /* Remember where and how much. */
897 offset = ph->offset;
898 physaddr = ph->ph->phys;
900 /* Now unlink the original physical block so we can replace
901 * it with new ones.
904 SLABSANE(ph);
905 SLABSANE(ph->ph);
906 assert(ph->ph->refcount > 1);
907 pb_unreferenced(region, ph, 1);
908 SLABFREE(ph);
910 SANITYCHECK(SCL_DETAIL);
912 /* Put new free memory in. */
913 allocflags = vrallocflags(region->flags | VR_UNINITIALIZED);
914 region_has_single_block = (offset == 0 && region->length == VM_PAGE_SIZE);
915 assert(region_has_single_block || !(allocflags & PAF_CONTIG));
916 assert(!(allocflags & PAF_CLEAR));
918 if(map_pf(vmp, region, offset, 1) != OK) {
919 /* XXX original range now gone. */
920 printf("VM: map_clone_ph_block: map_pf failed.\n");
921 return NULL;
924 /* Copy the block to the new memory.
925 * Can only fail if map_new_physblock didn't do what we asked.
927 if(copy_abs2region(physaddr, region, offset, VM_PAGE_SIZE) != OK)
928 panic("copy_abs2region failed, no good reason for that");
930 newpr = physr_search(region->phys, offset, AVL_EQUAL);
931 assert(newpr);
932 assert(newpr->offset == offset);
934 if(iter) {
935 physr_start_iter(region->phys, iter, offset, AVL_EQUAL);
936 assert(physr_get_iter(iter) == newpr);
939 SANITYCHECK(SCL_FUNCTIONS);
941 return newpr;
945 /*===========================================================================*
946 * map_pf *
947 *===========================================================================*/
948 int map_pf(vmp, region, offset, write)
949 struct vmproc *vmp;
950 struct vir_region *region;
951 vir_bytes offset;
952 int write;
954 struct phys_region *ph;
955 int r = OK;
957 offset -= offset % VM_PAGE_SIZE;
959 assert(offset >= 0);
960 assert(offset < region->length);
962 assert(!(region->vaddr % VM_PAGE_SIZE));
963 assert(!(write && !(region->flags & VR_WRITABLE)));
965 SANITYCHECK(SCL_FUNCTIONS);
967 if(!(ph = physr_search(region->phys, offset, AVL_EQUAL))) {
968 struct phys_block *pb;
970 /* New block. */
972 if(!(pb = pb_new(MAP_NONE))) {
973 printf("map_pf: pb_new failed\n");
974 return ENOMEM;
977 if(!(ph = pb_reference(pb, offset, region))) {
978 printf("map_pf: pb_reference failed\n");
979 pb_free(pb);
980 return ENOMEM;
984 assert(ph);
985 assert(ph->ph);
987 /* If we're writing and the block is already
988 * writable, nothing to do.
991 assert(region->memtype->writable);
993 if(!write || !region->memtype->writable(ph)) {
994 assert(region->memtype->ev_pagefault);
995 assert(ph->ph);
996 if((r = region->memtype->ev_pagefault(vmp,
997 region, ph, write)) == SUSPEND) {
998 panic("map_pf: memtype->ev_pagefault returned SUSPEND\n");
999 return SUSPEND;
1002 if(r != OK) {
1003 printf("map_pf: memtype->ev_pagefault failed\n");
1004 if(ph)
1005 pb_unreferenced(region, ph, 1);
1006 return r;
1009 assert(ph);
1010 assert(ph->ph);
1013 assert(ph->ph);
1015 if((r = map_ph_writept(vmp, region, ph)) != OK) {
1016 printf("map_pf: writept failed\n");
1017 return r;
1020 SANITYCHECK(SCL_FUNCTIONS);
1022 #if SANITYCHECKS
1023 if(OK != pt_checkrange(&vmp->vm_pt, region->vaddr+offset,
1024 VM_PAGE_SIZE, write)) {
1025 panic("map_pf: pt_checkrange failed: %d", r);
1027 #endif
1029 return r;
1032 int map_handle_memory(vmp, region, start_offset, length, write)
1033 struct vmproc *vmp;
1034 struct vir_region *region;
1035 vir_bytes start_offset;
1036 vir_bytes length;
1037 int write;
1039 vir_bytes offset, lim;
1040 int r;
1042 assert(length > 0);
1043 lim = start_offset + length;
1044 assert(lim > start_offset);
1046 for(offset = start_offset; offset < lim; offset += VM_PAGE_SIZE)
1047 if((r = map_pf(vmp, region, offset, write)) != OK)
1048 return r;
1050 return OK;
1053 /*===========================================================================*
1054 * map_pin_memory *
1055 *===========================================================================*/
1056 int map_pin_memory(struct vmproc *vmp)
1058 struct vir_region *vr;
1059 int r;
1060 region_iter iter;
1061 region_start_iter_least(&vmp->vm_regions_avl, &iter);
1062 /* Scan all memory regions. */
1063 while((vr = region_get_iter(&iter))) {
1064 /* Make sure region is mapped to physical memory and writable.*/
1065 r = map_handle_memory(vmp, vr, 0, vr->length, 1);
1066 if(r != OK) {
1067 panic("map_pin_memory: map_handle_memory failed: %d", r);
1069 region_incr_iter(&iter);
1071 return OK;
1074 #if SANITYCHECKS
1075 static int count_phys_regions(struct vir_region *vr)
1077 int n = 0;
1078 struct phys_region *ph;
1079 physr_iter iter;
1080 physr_start_iter_least(vr->phys, &iter);
1081 while((ph = physr_get_iter(&iter))) {
1082 n++;
1083 physr_incr_iter(&iter);
1085 return n;
1087 #endif
1089 /*===========================================================================*
1090 * map_copy_region *
1091 *===========================================================================*/
1092 static struct vir_region *map_copy_region(struct vmproc *vmp, struct vir_region *vr)
1094 /* map_copy_region creates a complete copy of the vir_region
1095 * data structure, linking in the same phys_blocks directly,
1096 * but all in limbo, i.e., the caller has to link the vir_region
1097 * to a process. Therefore it doesn't increase the refcount in
1098 * the phys_block; the caller has to do this once it's linked.
1099 * The reason for this is to keep the sanity checks working
1100 * within this function.
1102 struct vir_region *newvr;
1103 struct phys_region *ph;
1104 int r;
1105 physr_iter iter;
1106 #if SANITYCHECKS
1107 int cr;
1108 cr = count_phys_regions(vr);
1109 #endif
1111 if(!(newvr = region_new(vr->parent, vr->vaddr, vr->length, vr->flags, vr->memtype)))
1112 return NULL;
1114 if(vr->memtype->ev_copy && (r=vr->memtype->ev_copy(vr, newvr)) != OK) {
1115 map_free(newvr);
1116 printf("VM: memtype-specific copy failed (%d)\n", r);
1117 return NULL;
1120 physr_start_iter_least(vr->phys, &iter);
1121 while((ph = physr_get_iter(&iter))) {
1122 struct phys_region *newph = pb_reference(ph->ph, ph->offset, newvr);
1124 if(!newph) { map_free(newvr); return NULL; }
1126 #if SANITYCHECKS
1127 USE(newph, newph->written = 0;);
1128 assert(count_phys_regions(vr) == cr);
1129 #endif
1130 physr_incr_iter(&iter);
1133 #if SANITYCHECKS
1134 assert(count_phys_regions(vr) == count_phys_regions(newvr));
1135 #endif
1137 return newvr;
1140 /*===========================================================================*
1141 * copy_abs2region *
1142 *===========================================================================*/
1143 int copy_abs2region(phys_bytes abs, struct vir_region *destregion,
1144 phys_bytes offset, phys_bytes len)
1147 assert(destregion);
1148 assert(destregion->phys);
1149 while(len > 0) {
1150 phys_bytes sublen, suboffset;
1151 struct phys_region *ph;
1152 assert(destregion);
1153 assert(destregion->phys);
1154 if(!(ph = physr_search(destregion->phys, offset, AVL_LESS_EQUAL))) {
1155 printf("VM: copy_abs2region: no phys region found (1).\n");
1156 return EFAULT;
1158 assert(ph->offset <= offset);
1159 if(ph->offset+VM_PAGE_SIZE <= offset) {
1160 printf("VM: copy_abs2region: no phys region found (2).\n");
1161 return EFAULT;
1163 suboffset = offset - ph->offset;
1164 assert(suboffset < VM_PAGE_SIZE);
1165 sublen = len;
1166 if(sublen > VM_PAGE_SIZE - suboffset)
1167 sublen = VM_PAGE_SIZE - suboffset;
1168 assert(suboffset + sublen <= VM_PAGE_SIZE);
1169 if(ph->ph->refcount != 1) {
1170 printf("VM: copy_abs2region: refcount not 1.\n");
1171 return EFAULT;
1174 if(sys_abscopy(abs, ph->ph->phys + suboffset, sublen) != OK) {
1175 printf("VM: copy_abs2region: abscopy failed.\n");
1176 return EFAULT;
1178 abs += sublen;
1179 offset += sublen;
1180 len -= sublen;
1183 return OK;
1186 /*=========================================================================*
1187 * map_writept *
1188 *=========================================================================*/
1189 int map_writept(struct vmproc *vmp)
1191 struct vir_region *vr;
1192 struct phys_region *ph;
1193 int r;
1194 region_iter v_iter;
1195 region_start_iter_least(&vmp->vm_regions_avl, &v_iter);
1197 while((vr = region_get_iter(&v_iter))) {
1198 physr_iter ph_iter;
1199 physr_start_iter_least(vr->phys, &ph_iter);
1201 while((ph = physr_get_iter(&ph_iter))) {
1202 physr_incr_iter(&ph_iter);
1204 if((r=map_ph_writept(vmp, vr, ph)) != OK) {
1205 printf("VM: map_writept: failed\n");
1206 return r;
1209 region_incr_iter(&v_iter);
1212 return OK;
1215 /*========================================================================*
1216 * map_proc_copy *
1217 *========================================================================*/
1218 int map_proc_copy(dst, src)
1219 struct vmproc *dst;
1220 struct vmproc *src;
1222 /* Copy all the memory regions from the src process to the dst process. */
1223 region_init(&dst->vm_regions_avl);
1225 return map_proc_copy_from(dst, src, NULL);
1228 /*========================================================================*
1229 * map_proc_copy_from *
1230 *========================================================================*/
1231 int map_proc_copy_from(dst, src, start_src_vr)
1232 struct vmproc *dst;
1233 struct vmproc *src;
1234 struct vir_region *start_src_vr;
1236 struct vir_region *vr;
1237 region_iter v_iter;
1239 if(!start_src_vr)
1240 start_src_vr = region_search_least(&src->vm_regions_avl);
1242 assert(start_src_vr);
1243 assert(start_src_vr->parent == src);
1244 region_start_iter(&src->vm_regions_avl, &v_iter,
1245 start_src_vr->vaddr, AVL_EQUAL);
1246 assert(region_get_iter(&v_iter) == start_src_vr);
1248 /* Copy source regions after the destination's last region (if any). */
1250 SANITYCHECK(SCL_FUNCTIONS);
1252 while((vr = region_get_iter(&v_iter))) {
1253 physr_iter iter_orig, iter_new;
1254 struct vir_region *newvr;
1255 struct phys_region *orig_ph, *new_ph;
1256 if(!(newvr = map_copy_region(dst, vr))) {
1257 map_free_proc(dst);
1258 return ENOMEM;
1260 USE(newvr, newvr->parent = dst;);
1261 region_insert(&dst->vm_regions_avl, newvr);
1262 physr_start_iter_least(vr->phys, &iter_orig);
1263 physr_start_iter_least(newvr->phys, &iter_new);
1264 while((orig_ph = physr_get_iter(&iter_orig))) {
1265 struct phys_block *pb;
1266 new_ph = physr_get_iter(&iter_new);
1267 /* Check two physregions both are nonnull,
1268 * are different, and match physblocks.
1270 assert(new_ph);
1271 assert(orig_ph);
1272 assert(orig_ph != new_ph);
1273 pb = orig_ph->ph;
1274 assert(orig_ph->ph == new_ph->ph);
1276 /* Get next new physregion */
1277 physr_incr_iter(&iter_orig);
1278 physr_incr_iter(&iter_new);
1280 assert(!physr_get_iter(&iter_new));
1281 region_incr_iter(&v_iter);
1284 map_writept(src);
1285 map_writept(dst);
1287 SANITYCHECK(SCL_FUNCTIONS);
1288 return OK;
1291 int map_region_extend_upto_v(struct vmproc *vmp, vir_bytes v)
1293 vir_bytes offset = v;
1294 struct vir_region *vr, *nextvr;
1296 offset = roundup(offset, VM_PAGE_SIZE);
1298 if(!(vr = region_search(&vmp->vm_regions_avl, offset, AVL_LESS))) {
1299 printf("VM: nothing to extend\n");
1300 return ENOMEM;
1303 assert(vr->vaddr <= offset);
1304 if((nextvr = getnextvr(vr))) {
1305 assert(offset <= nextvr->vaddr);
1308 if(nextvr && nextvr->vaddr < offset) {
1309 printf("VM: can't grow into next region\n");
1310 return ENOMEM;
1313 if(!vr->memtype->ev_resize) {
1314 printf("VM: can't resize this type of memory\n");
1315 return ENOMEM;
1318 return vr->memtype->ev_resize(vmp, vr, offset - vr->vaddr);
1321 /*========================================================================*
1322 * map_unmap_region *
1323 *========================================================================*/
1324 int map_unmap_region(struct vmproc *vmp, struct vir_region *r,
1325 vir_bytes offset, vir_bytes len)
1327 /* Shrink the region by 'len' bytes, from the start. Unreference
1328 * memory it used to reference if any.
1330 vir_bytes regionstart;
1332 SANITYCHECK(SCL_FUNCTIONS);
1334 if(offset+len > r->length || (len % VM_PAGE_SIZE)) {
1335 printf("VM: bogus length 0x%lx\n", len);
1336 return EINVAL;
1339 regionstart = r->vaddr + offset;
1341 /* unreference its memory */
1342 map_subfree(r, offset, len);
1344 /* if unmap was at start/end of this region, it actually shrinks */
1345 if(offset == 0) {
1346 struct phys_region *pr;
1347 physr_iter iter;
1349 region_remove(&vmp->vm_regions_avl, r->vaddr);
1351 USE(r,
1352 r->vaddr += len;
1353 r->length -= len;);
1355 region_insert(&vmp->vm_regions_avl, r);
1357 /* vaddr has increased; to make all the phys_regions
1358 * point to the same addresses, make them shrink by the
1359 * same amount.
1361 physr_init_iter(&iter);
1362 physr_start_iter(r->phys, &iter, offset, AVL_GREATER_EQUAL);
1364 while((pr = physr_get_iter(&iter))) {
1365 assert(pr->offset >= offset);
1366 USE(pr, pr->offset -= len;);
1367 physr_incr_iter(&iter);
1369 } else if(offset + len == r->length) {
1370 assert(len <= r->length);
1371 r->length -= len;
1374 if(r->length == 0) {
1375 /* Whole region disappears. Unlink and free it. */
1376 region_remove(&vmp->vm_regions_avl, r->vaddr);
1377 map_free(r);
1380 SANITYCHECK(SCL_DETAIL);
1382 if(pt_writemap(vmp, &vmp->vm_pt, regionstart,
1383 MAP_NONE, len, 0, WMF_OVERWRITE) != OK) {
1384 printf("VM: map_unmap_region: pt_writemap failed\n");
1385 return ENOMEM;
1388 SANITYCHECK(SCL_FUNCTIONS);
1390 return OK;
1393 /*========================================================================*
1394 * map_get_phys *
1395 *========================================================================*/
1396 int map_get_phys(struct vmproc *vmp, vir_bytes addr, phys_bytes *r)
1398 struct vir_region *vr;
1400 if (!(vr = map_lookup(vmp, addr, NULL)) ||
1401 (vr->vaddr != addr))
1402 return EINVAL;
1404 if (!vr->memtype->regionid)
1405 return EINVAL;
1407 if(r)
1408 *r = vr->memtype->regionid(vr);
1410 return OK;
1413 /*========================================================================*
1414 * map_get_ref *
1415 *========================================================================*/
1416 int map_get_ref(struct vmproc *vmp, vir_bytes addr, u8_t *cnt)
1418 struct vir_region *vr;
1420 if (!(vr = map_lookup(vmp, addr, NULL)) ||
1421 (vr->vaddr != addr) || !vr->memtype->refcount)
1422 return EINVAL;
1424 if (cnt)
1425 *cnt = vr->memtype->refcount(vr);
1427 return OK;
1430 /*========================================================================*
1431 * get_stats_info *
1432 *========================================================================*/
1433 void get_stats_info(struct vm_stats_info *vsi)
1435 yielded_t *yb;
1437 vsi->vsi_cached = 0L;
1439 for(yb = lru_youngest; yb; yb = yb->older)
1440 vsi->vsi_cached++;
1443 void get_usage_info_kernel(struct vm_usage_info *vui)
1445 memset(vui, 0, sizeof(*vui));
1446 vui->vui_total = kernel_boot_info.kernel_allocated_bytes;
1449 static void get_usage_info_vm(struct vm_usage_info *vui)
1451 memset(vui, 0, sizeof(*vui));
1452 vui->vui_total = kernel_boot_info.vm_allocated_bytes +
1453 get_vm_self_pages() * VM_PAGE_SIZE;
1456 /*========================================================================*
1457 * get_usage_info *
1458 *========================================================================*/
1459 void get_usage_info(struct vmproc *vmp, struct vm_usage_info *vui)
1461 struct vir_region *vr;
1462 physr_iter iter;
1463 struct phys_region *ph;
1464 region_iter v_iter;
1465 region_start_iter_least(&vmp->vm_regions_avl, &v_iter);
1467 memset(vui, 0, sizeof(*vui));
1469 if(vmp->vm_endpoint == VM_PROC_NR) {
1470 get_usage_info_vm(vui);
1471 return;
1474 if(vmp->vm_endpoint < 0) {
1475 get_usage_info_kernel(vui);
1476 return;
1479 while((vr = region_get_iter(&v_iter))) {
1480 physr_start_iter_least(vr->phys, &iter);
1481 while((ph = physr_get_iter(&iter))) {
1482 /* All present pages are counted towards the total. */
1483 vui->vui_total += VM_PAGE_SIZE;
1485 if (ph->ph->refcount > 1) {
1486 /* Any page with a refcount > 1 is common. */
1487 vui->vui_common += VM_PAGE_SIZE;
1489 /* Any common, non-COW page is shared. */
1490 if (vr->flags & VR_SHARED)
1491 vui->vui_shared += VM_PAGE_SIZE;
1493 physr_incr_iter(&iter);
1495 region_incr_iter(&v_iter);
1499 /*===========================================================================*
1500 * get_region_info *
1501 *===========================================================================*/
1502 int get_region_info(struct vmproc *vmp, struct vm_region_info *vri,
1503 int max, vir_bytes *nextp)
1505 struct vir_region *vr;
1506 vir_bytes next;
1507 int count;
1508 region_iter v_iter;
1510 next = *nextp;
1512 if (!max) return 0;
1514 region_start_iter(&vmp->vm_regions_avl, &v_iter, next, AVL_GREATER_EQUAL);
1515 if(!(vr = region_get_iter(&v_iter))) return 0;
1517 for(count = 0; (vr = region_get_iter(&v_iter)) && count < max; count++, vri++) {
1518 struct phys_region *ph1, *ph2;
1520 /* Report part of the region that's actually in use. */
1522 /* Get first and last phys_regions, if any */
1523 ph1 = physr_search_least(vr->phys);
1524 ph2 = physr_search_greatest(vr->phys);
1525 if(!ph1 || !ph2) { assert(!ph1 && !ph2); continue; }
1527 /* Report start+length of region starting from lowest use. */
1528 vri->vri_addr = vr->vaddr + ph1->offset;
1529 vri->vri_prot = 0;
1530 vri->vri_length = ph2->offset + VM_PAGE_SIZE - ph1->offset;
1532 /* "AND" the provided protection with per-page protection. */
1533 if (!(vr->flags & VR_WRITABLE))
1534 vri->vri_prot &= ~PROT_WRITE;
1536 next = vr->vaddr + vr->length;
1537 region_incr_iter(&v_iter);
1540 *nextp = next;
1541 return count;
1544 /*========================================================================*
1545 * regionprintstats *
1546 *========================================================================*/
1547 void printregionstats(struct vmproc *vmp)
1549 struct vir_region *vr;
1550 struct phys_region *pr;
1551 physr_iter iter;
1552 vir_bytes used = 0, weighted = 0;
1553 region_iter v_iter;
1554 region_start_iter_least(&vmp->vm_regions_avl, &v_iter);
1556 while((vr = region_get_iter(&v_iter))) {
1557 region_incr_iter(&v_iter);
1558 if(vr->flags & VR_DIRECT)
1559 continue;
1560 physr_start_iter_least(vr->phys, &iter);
1561 while((pr = physr_get_iter(&iter))) {
1562 physr_incr_iter(&iter);
1563 used += VM_PAGE_SIZE;
1564 weighted += VM_PAGE_SIZE / pr->ph->refcount;
1568 printf("%6lukB %6lukB\n", used/1024, weighted/1024);
1570 return;
1573 /*===========================================================================*
1574 * get_clean_phys_region *
1575 *===========================================================================*/
1576 static struct phys_region *
1577 get_clean_phys_region(struct vmproc *vmp, vir_bytes vaddr, struct vir_region **ret_region)
1579 struct vir_region *region;
1580 vir_bytes mapaddr;
1581 struct phys_region *ph;
1583 mapaddr = vaddr;
1585 if(!(region = map_lookup(vmp, mapaddr, &ph)) || !ph) {
1586 printf("VM: get_clean_phys_region: 0x%lx not found\n", vaddr);
1587 return NULL;
1590 assert(mapaddr >= region->vaddr);
1591 assert(mapaddr < region->vaddr + region->length);
1593 /* If it's mapped more than once, make a copy. */
1594 assert(ph->ph->refcount > 0);
1595 if(ph->ph->refcount > 1) {
1596 if(!(ph = map_clone_ph_block(vmp, region,
1597 ph, NULL))) {
1598 printf("VM: get_clean_phys_region: ph copy failed\n");
1599 return NULL;
1603 assert(ph->ph->refcount == 1);
1605 *ret_region = region;
1607 return ph;
1610 static int getblock(struct vmproc *vmp, u64_t id, vir_bytes vaddr, int pages)
1612 yielded_t *yb;
1613 struct phys_region *ph;
1614 struct vir_region *region;
1615 yielded_avl *avl;
1616 block_id_t blockid;
1617 phys_bytes phaddr;
1618 int p;
1620 /* Try to get the yielded block */
1621 blockid.owner = vmp->vm_endpoint;
1622 blockid.id = id;
1623 avl = get_yielded_avl(blockid);
1624 if(!(yb = yielded_search(avl, blockid, AVL_EQUAL))) {
1625 return ESRCH;
1628 if(yb->pages != pages) {
1629 printf("VM: getblock: length mismatch (%d != %d)\n",
1630 pages, yb->pages);
1631 return EFAULT;
1634 phaddr = yb->physaddr;
1636 for(p = 0; p < pages; p++) {
1637 /* Get the intended phys region, make sure refcount is 1. */
1638 if(!(ph = get_clean_phys_region(vmp, vaddr, &region))) {
1639 printf("VM: getblock: not found for %d\n", vmp->vm_endpoint);
1640 return EINVAL;
1643 assert(ph->ph->refcount == 1);
1645 /* Free the block that is currently there. */
1646 free_mem(ABS2CLICK(ph->ph->phys), 1);
1648 /* Set the phys block to new addr and update pagetable. */
1649 USE(ph->ph, ph->ph->phys = phaddr;);
1650 if(map_ph_writept(vmp, region, ph) != OK) {
1651 /* Presumably it was mapped, so there is no reason
1652 * updating should fail.
1654 panic("do_get_block: couldn't write pt");
1657 vaddr += VM_PAGE_SIZE;
1658 phaddr += VM_PAGE_SIZE;
1661 /* Forget about the yielded block and free the struct. */
1662 freeyieldednode(yb, 0);
1664 return OK;
1667 static int yieldblock(struct vmproc *vmp, u64_t id,
1668 vir_bytes vaddr, yielded_t **retyb, int pages)
1670 yielded_t *newyb;
1671 vir_bytes mem_clicks, v, p, new_phaddr;
1672 struct vir_region *region;
1673 struct phys_region *ph = NULL, *prev_ph = NULL, *first_ph = NULL;
1674 yielded_avl *avl;
1675 block_id_t blockid;
1677 /* Makes no sense if yielded block ID already exists, and
1678 * is likely a serious bug in the caller.
1680 blockid.id = id;
1681 blockid.owner = vmp->vm_endpoint;
1682 avl = get_yielded_avl(blockid);
1683 if(yielded_search(avl, blockid, AVL_EQUAL)) {
1684 printf("!");
1685 return EINVAL;
1688 if((vaddr % VM_PAGE_SIZE) || pages < 1) return EFAULT;
1690 v = vaddr;
1691 for(p = 0; p < pages; p++) {
1692 if(!(region = map_lookup(vmp, v, &ph)) || !ph) {
1693 printf("VM: do_yield_block: not found for %d\n",
1694 vmp->vm_endpoint);
1695 return EINVAL;
1697 if(!(region->flags & VR_ANON)) {
1698 printf("VM: yieldblock: non-anon 0x%lx\n", v);
1699 return EFAULT;
1701 if(ph->ph->refcount != 1) {
1702 printf("VM: do_yield_block: mapped not once for %d\n",
1703 vmp->vm_endpoint);
1704 return EFAULT;
1706 if(prev_ph) {
1707 if(ph->ph->phys != prev_ph->ph->phys + VM_PAGE_SIZE) {
1708 printf("VM: physically discontiguous yield\n");
1709 return EINVAL;
1712 prev_ph = ph;
1713 if(!first_ph) first_ph = ph;
1714 v += VM_PAGE_SIZE;
1717 /* Make a new block to record the yielding in. */
1718 if(!SLABALLOC(newyb)) {
1719 return ENOMEM;
1722 assert(!(ph->ph->phys % VM_PAGE_SIZE));
1724 if((mem_clicks = alloc_mem(pages, PAF_CLEAR)) == NO_MEM) {
1725 SLABFREE(newyb);
1726 return ENOMEM;
1729 /* Update yielded block info. */
1730 USE(newyb,
1731 newyb->id = blockid;
1732 newyb->physaddr = first_ph->ph->phys;
1733 newyb->pages = pages;
1734 newyb->younger = NULL;);
1736 new_phaddr = CLICK2ABS(mem_clicks);
1738 /* Set new phys block to new addr and update pagetable. */
1739 v = vaddr;
1740 for(p = 0; p < pages; p++) {
1741 region = map_lookup(vmp, v, &ph);
1742 assert(region && ph);
1743 assert(ph->ph->refcount == 1);
1744 USE(ph->ph,
1745 ph->ph->phys = new_phaddr;);
1746 if(map_ph_writept(vmp, region, ph) != OK) {
1747 /* Presumably it was mapped, so there is no reason
1748 * updating should fail.
1750 panic("yield_block: couldn't write pt");
1752 v += VM_PAGE_SIZE;
1753 new_phaddr += VM_PAGE_SIZE;
1756 /* Remember yielded block. */
1758 yielded_insert(avl, newyb);
1759 vmp->vm_yielded++;
1761 /* Add to LRU list too. It's the youngest block. */
1762 LRUCHECK;
1764 if(lru_youngest) {
1765 USE(lru_youngest,
1766 lru_youngest->younger = newyb;);
1767 } else {
1768 lru_oldest = newyb;
1771 USE(newyb,
1772 newyb->older = lru_youngest;);
1774 lru_youngest = newyb;
1776 LRUCHECK;
1778 if(retyb)
1779 *retyb = newyb;
1781 return OK;
1784 /*===========================================================================*
1785 * do_forgetblocks *
1786 *===========================================================================*/
1787 int do_forgetblocks(message *m)
1789 int n;
1790 struct vmproc *vmp;
1791 endpoint_t caller = m->m_source;
1793 if(vm_isokendpt(caller, &n) != OK)
1794 panic("do_yield_block: message from strange source: %d",
1795 m->m_source);
1797 vmp = &vmproc[n];
1799 free_yielded_proc(vmp);
1801 return OK;
1804 /*===========================================================================*
1805 * do_forgetblock *
1806 *===========================================================================*/
1807 int do_forgetblock(message *m)
1809 int n;
1810 struct vmproc *vmp;
1811 endpoint_t caller = m->m_source;
1812 yielded_t *yb;
1813 u64_t id;
1814 block_id_t blockid;
1815 yielded_avl *avl;
1817 if(vm_isokendpt(caller, &n) != OK)
1818 panic("do_yield_block: message from strange source: %d",
1819 m->m_source);
1821 vmp = &vmproc[n];
1823 id = make64(m->VMFB_IDLO, m->VMFB_IDHI);
1825 blockid.id = id;
1826 blockid.owner = vmp->vm_endpoint;
1827 avl = get_yielded_avl(blockid);
1828 if((yb = yielded_search(avl, blockid, AVL_EQUAL))) {
1829 freeyieldednode(yb, 1);
1832 return OK;
1835 /*===========================================================================*
1836 * do_yieldblockgetblock *
1837 *===========================================================================*/
1838 int do_yieldblockgetblock(message *m)
1840 u64_t yieldid, getid;
1841 int n;
1842 endpoint_t caller = m->m_source;
1843 struct vmproc *vmp;
1844 yielded_t *yb = NULL;
1845 int r = ESRCH;
1846 int pages;
1848 if(vm_isokendpt(caller, &n) != OK)
1849 panic("do_yieldblockgetblock: message from strange source: %d",
1850 m->m_source);
1852 vmp = &vmproc[n];
1854 pages = m->VMYBGB_LEN / VM_PAGE_SIZE;
1856 if((m->VMYBGB_LEN % VM_PAGE_SIZE) || pages < 1) {
1857 static int printed;
1858 if(!printed) {
1859 printed = 1;
1860 printf("vm: non-page-aligned or short block length\n");
1862 return EFAULT;
1865 yieldid = make64(m->VMYBGB_YIELDIDLO, m->VMYBGB_YIELDIDHI);
1866 getid = make64(m->VMYBGB_GETIDLO, m->VMYBGB_GETIDHI);
1868 if(cmp64(yieldid, VM_BLOCKID_NONE) != 0) {
1869 /* A block was given to yield. */
1870 yieldblock(vmp, yieldid, (vir_bytes) m->VMYBGB_VADDR, &yb,
1871 pages);
1874 if(cmp64(getid, VM_BLOCKID_NONE) != 0) {
1875 /* A block was given to get. */
1876 r = getblock(vmp, getid, (vir_bytes) m->VMYBGB_VADDR, pages);
1879 return r;
1882 void map_setparent(struct vmproc *vmp)
1884 region_iter iter;
1885 struct vir_region *vr;
1886 region_start_iter_least(&vmp->vm_regions_avl, &iter);
1887 while((vr = region_get_iter(&iter))) {
1888 USE(vr, vr->parent = vmp;);
1889 region_incr_iter(&iter);
1893 int physregions(struct vir_region *vr)
1895 int n = 0;
1896 physr_iter iter;
1897 physr_start_iter_least(vr->phys, &iter);
1898 while(physr_get_iter(&iter)) {
1899 n++;
1900 physr_incr_iter(&iter);
1902 return n;