diff --git a/sys/amd64/amd64/pmap.c b/sys/amd64/amd64/pmap.c index c602cdc82096..53d6906a04f5 100644 --- a/sys/amd64/amd64/pmap.c +++ b/sys/amd64/amd64/pmap.c @@ -39,7 +39,7 @@ * SUCH DAMAGE. * * from: @(#)pmap.c 7.7 (Berkeley) 5/12/91 - * $Id: pmap.c,v 1.125 1996/10/12 21:35:03 dyson Exp $ + * $Id: pmap.c,v 1.126 1996/10/13 01:38:37 dyson Exp $ */ /* @@ -314,8 +314,8 @@ pmap_init(phys_start, phys_end) vm_offset_t phys_start, phys_end; { vm_offset_t addr; - vm_size_t npg, s; - int i; + vm_size_t s; + int i, npg; /* * calculate the number of pv_entries needed @@ -333,7 +333,7 @@ pmap_init(phys_start, phys_end) addr = (vm_offset_t) kmem_alloc(kernel_map, s); pv_table = (pv_table_t *) addr; - for(i=0;ip_vmspace->vm_pmap; + + /* + * allocate object for the upages + */ + upobj = vm_object_allocate( OBJT_DEFAULT, + UPAGES); + p->p_vmspace->vm_upages_obj = upobj; + + /* get a kernel virtual address for the UPAGES for this proc */ + up = (struct user *) kmem_alloc_pageable(u_map, UPAGES * PAGE_SIZE); + if (up == NULL) + panic("vm_fork: u_map allocation failed"); + + /* + * Allocate the ptp and incr the hold count appropriately + */ + m = pmap_allocpte(pmap, (vm_offset_t) kstack); + m->hold_count += (UPAGES - 1); + + ptep = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + ptek = (unsigned *) vtopte((vm_offset_t) up); + + for(i=0;iwire_count++; + ++cnt.v_wire_count; + + /* + * Enter the page into both the kernel and the process + * address space. + */ + *(ptep + i) = VM_PAGE_TO_PHYS(m) | PG_RW | PG_V; + *(ptek + i) = VM_PAGE_TO_PHYS(m) | PG_RW | PG_V; + + m->flags &= ~(PG_ZERO|PG_BUSY); + m->flags |= PG_MAPPED|PG_WRITEABLE; + m->valid = VM_PAGE_BITS_ALL; + } + + p->p_addr = up; +} + +/* + * Dispose the UPAGES for a process that has exited. + * This routine directly impacts the exit perf of a process. + */ +void +pmap_dispose_proc(p) + struct proc *p; +{ + int i; + vm_object_t upobj; + pmap_t pmap; + vm_page_t m; + unsigned *ptep, *ptek; + + pmap = &p->p_vmspace->vm_pmap; + ptep = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + ptek = (unsigned *) vtopte((vm_offset_t) p->p_addr); + + upobj = p->p_vmspace->vm_upages_obj; + + for(i=0;ip_addr, ctob(UPAGES)); +} + +/* + * Allow the UPAGES for a process to be prejudicially paged out. + */ +void +pmap_swapout_proc(p) + struct proc *p; +{ + int i; + vm_object_t upobj; + pmap_t pmap; + vm_page_t m; + unsigned *pte; + + pmap = &p->p_vmspace->vm_pmap; + pte = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + + upobj = p->p_vmspace->vm_upages_obj; + /* + * let the upages be paged + */ + for(i=0;idirty = VM_PAGE_BITS_ALL; + *(pte + i) = 0; + pmap_unuse_pt(pmap, (vm_offset_t) kstack + i * PAGE_SIZE, NULL); + + vm_page_unwire(m); + vm_page_deactivate(m); + pmap_kremove( (vm_offset_t) p->p_addr + PAGE_SIZE * i); + } +} + +/* + * Bring the UPAGES for a specified process back in. + */ +void +pmap_swapin_proc(p) + struct proc *p; +{ + int i; + vm_object_t upobj; + pmap_t pmap; + vm_page_t m; + unsigned *pte; + + pmap = &p->p_vmspace->vm_pmap; + /* + * Allocate the ptp and incr the hold count appropriately + */ + m = pmap_allocpte(pmap, (vm_offset_t) kstack); + m->hold_count += (UPAGES - 1); + pte = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + + upobj = p->p_vmspace->vm_upages_obj; + for(i=0;iflags & PG_BUSY) || m->busy) { + m->flags |= PG_WANTED; + tsleep(m, PVM, "swinuw",0); + goto retry; + } + m->flags |= PG_BUSY; + } + vm_page_wire(m); + splx(s); + + *(pte+i) = VM_PAGE_TO_PHYS(m) | PG_RW | PG_V; + pmap_kenter(((vm_offset_t) p->p_addr) + i * PAGE_SIZE, + VM_PAGE_TO_PHYS(m)); + + if (m->valid != VM_PAGE_BITS_ALL) { + int rv; + rv = vm_pager_get_pages(upobj, &m, 1, 0); + if (rv != VM_PAGER_OK) + panic("faultin: cannot get upages for proc: %d\n", p->p_pid); + m->valid = VM_PAGE_BITS_ALL; + } + PAGE_WAKEUP(m); + m->flags |= PG_MAPPED|PG_WRITEABLE; + } +} /*************************************************** * Page table page management routines..... diff --git a/sys/amd64/amd64/vm_machdep.c b/sys/amd64/amd64/vm_machdep.c index a96716872357..44b799d93256 100644 --- a/sys/amd64/amd64/vm_machdep.c +++ b/sys/amd64/amd64/vm_machdep.c @@ -38,7 +38,7 @@ * * from: @(#)vm_machdep.c 7.3 (Berkeley) 5/13/91 * Utah $Hdr: vm_machdep.c 1.16.1.1 89/06/23$ - * $Id: vm_machdep.c,v 1.68 1996/09/08 20:44:12 dyson Exp $ + * $Id: vm_machdep.c,v 1.69 1996/09/28 22:37:43 dyson Exp $ */ #include "npx.h" @@ -623,8 +623,7 @@ cpu_wait(p) struct proc *p; { /* drop per-process resources */ - pmap_qremove((vm_offset_t) p->p_addr, UPAGES); - kmem_free(u_map, (vm_offset_t)p->p_addr, ctob(UPAGES)); + pmap_dispose_proc(p); vmspace_free(p->p_vmspace); } diff --git a/sys/i386/i386/pmap.c b/sys/i386/i386/pmap.c index c602cdc82096..53d6906a04f5 100644 --- a/sys/i386/i386/pmap.c +++ b/sys/i386/i386/pmap.c @@ -39,7 +39,7 @@ * SUCH DAMAGE. * * from: @(#)pmap.c 7.7 (Berkeley) 5/12/91 - * $Id: pmap.c,v 1.125 1996/10/12 21:35:03 dyson Exp $ + * $Id: pmap.c,v 1.126 1996/10/13 01:38:37 dyson Exp $ */ /* @@ -314,8 +314,8 @@ pmap_init(phys_start, phys_end) vm_offset_t phys_start, phys_end; { vm_offset_t addr; - vm_size_t npg, s; - int i; + vm_size_t s; + int i, npg; /* * calculate the number of pv_entries needed @@ -333,7 +333,7 @@ pmap_init(phys_start, phys_end) addr = (vm_offset_t) kmem_alloc(kernel_map, s); pv_table = (pv_table_t *) addr; - for(i=0;ip_vmspace->vm_pmap; + + /* + * allocate object for the upages + */ + upobj = vm_object_allocate( OBJT_DEFAULT, + UPAGES); + p->p_vmspace->vm_upages_obj = upobj; + + /* get a kernel virtual address for the UPAGES for this proc */ + up = (struct user *) kmem_alloc_pageable(u_map, UPAGES * PAGE_SIZE); + if (up == NULL) + panic("vm_fork: u_map allocation failed"); + + /* + * Allocate the ptp and incr the hold count appropriately + */ + m = pmap_allocpte(pmap, (vm_offset_t) kstack); + m->hold_count += (UPAGES - 1); + + ptep = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + ptek = (unsigned *) vtopte((vm_offset_t) up); + + for(i=0;iwire_count++; + ++cnt.v_wire_count; + + /* + * Enter the page into both the kernel and the process + * address space. + */ + *(ptep + i) = VM_PAGE_TO_PHYS(m) | PG_RW | PG_V; + *(ptek + i) = VM_PAGE_TO_PHYS(m) | PG_RW | PG_V; + + m->flags &= ~(PG_ZERO|PG_BUSY); + m->flags |= PG_MAPPED|PG_WRITEABLE; + m->valid = VM_PAGE_BITS_ALL; + } + + p->p_addr = up; +} + +/* + * Dispose the UPAGES for a process that has exited. + * This routine directly impacts the exit perf of a process. + */ +void +pmap_dispose_proc(p) + struct proc *p; +{ + int i; + vm_object_t upobj; + pmap_t pmap; + vm_page_t m; + unsigned *ptep, *ptek; + + pmap = &p->p_vmspace->vm_pmap; + ptep = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + ptek = (unsigned *) vtopte((vm_offset_t) p->p_addr); + + upobj = p->p_vmspace->vm_upages_obj; + + for(i=0;ip_addr, ctob(UPAGES)); +} + +/* + * Allow the UPAGES for a process to be prejudicially paged out. + */ +void +pmap_swapout_proc(p) + struct proc *p; +{ + int i; + vm_object_t upobj; + pmap_t pmap; + vm_page_t m; + unsigned *pte; + + pmap = &p->p_vmspace->vm_pmap; + pte = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + + upobj = p->p_vmspace->vm_upages_obj; + /* + * let the upages be paged + */ + for(i=0;idirty = VM_PAGE_BITS_ALL; + *(pte + i) = 0; + pmap_unuse_pt(pmap, (vm_offset_t) kstack + i * PAGE_SIZE, NULL); + + vm_page_unwire(m); + vm_page_deactivate(m); + pmap_kremove( (vm_offset_t) p->p_addr + PAGE_SIZE * i); + } +} + +/* + * Bring the UPAGES for a specified process back in. + */ +void +pmap_swapin_proc(p) + struct proc *p; +{ + int i; + vm_object_t upobj; + pmap_t pmap; + vm_page_t m; + unsigned *pte; + + pmap = &p->p_vmspace->vm_pmap; + /* + * Allocate the ptp and incr the hold count appropriately + */ + m = pmap_allocpte(pmap, (vm_offset_t) kstack); + m->hold_count += (UPAGES - 1); + pte = (unsigned *) pmap_pte(pmap, (vm_offset_t) kstack); + + upobj = p->p_vmspace->vm_upages_obj; + for(i=0;iflags & PG_BUSY) || m->busy) { + m->flags |= PG_WANTED; + tsleep(m, PVM, "swinuw",0); + goto retry; + } + m->flags |= PG_BUSY; + } + vm_page_wire(m); + splx(s); + + *(pte+i) = VM_PAGE_TO_PHYS(m) | PG_RW | PG_V; + pmap_kenter(((vm_offset_t) p->p_addr) + i * PAGE_SIZE, + VM_PAGE_TO_PHYS(m)); + + if (m->valid != VM_PAGE_BITS_ALL) { + int rv; + rv = vm_pager_get_pages(upobj, &m, 1, 0); + if (rv != VM_PAGER_OK) + panic("faultin: cannot get upages for proc: %d\n", p->p_pid); + m->valid = VM_PAGE_BITS_ALL; + } + PAGE_WAKEUP(m); + m->flags |= PG_MAPPED|PG_WRITEABLE; + } +} /*************************************************** * Page table page management routines..... diff --git a/sys/i386/i386/vm_machdep.c b/sys/i386/i386/vm_machdep.c index a96716872357..44b799d93256 100644 --- a/sys/i386/i386/vm_machdep.c +++ b/sys/i386/i386/vm_machdep.c @@ -38,7 +38,7 @@ * * from: @(#)vm_machdep.c 7.3 (Berkeley) 5/13/91 * Utah $Hdr: vm_machdep.c 1.16.1.1 89/06/23$ - * $Id: vm_machdep.c,v 1.68 1996/09/08 20:44:12 dyson Exp $ + * $Id: vm_machdep.c,v 1.69 1996/09/28 22:37:43 dyson Exp $ */ #include "npx.h" @@ -623,8 +623,7 @@ cpu_wait(p) struct proc *p; { /* drop per-process resources */ - pmap_qremove((vm_offset_t) p->p_addr, UPAGES); - kmem_free(u_map, (vm_offset_t)p->p_addr, ctob(UPAGES)); + pmap_dispose_proc(p); vmspace_free(p->p_vmspace); } diff --git a/sys/vm/pmap.h b/sys/vm/pmap.h index 25b39539632b..2144280a8d06 100644 --- a/sys/vm/pmap.h +++ b/sys/vm/pmap.h @@ -61,7 +61,7 @@ * any improvements or extensions that they make and grant Carnegie the * rights to redistribute these changes. * - * $Id: pmap.h,v 1.14 1996/07/30 03:08:04 dyson Exp $ + * $Id: pmap.h,v 1.15 1996/09/08 20:44:31 dyson Exp $ */ /* @@ -126,6 +126,10 @@ void pmap_zero_page __P((vm_offset_t)); void pmap_prefault __P((pmap_t pmap, vm_offset_t addra, vm_map_entry_t entry, vm_object_t object)); int pmap_mincore __P((pmap_t pmap, vm_offset_t addr)); +void pmap_new_proc __P((struct proc *p)); +void pmap_dispose_proc __P((struct proc *p)); +void pmap_swapout_proc __P((struct proc *p)); +void pmap_swapin_proc __P((struct proc *p)); #endif /* KERNEL */ diff --git a/sys/vm/vm_glue.c b/sys/vm/vm_glue.c index fee37a2f08d3..a76bdeb528ac 100644 --- a/sys/vm/vm_glue.c +++ b/sys/vm/vm_glue.c @@ -59,7 +59,7 @@ * any improvements or extensions that they make and grant Carnegie the * rights to redistribute these changes. * - * $Id: vm_glue.c,v 1.52 1996/07/02 02:07:56 dyson Exp $ + * $Id: vm_glue.c,v 1.53 1996/09/15 11:24:21 bde Exp $ */ #include @@ -205,55 +205,9 @@ vm_fork(p1, p2) if (p1->p_vmspace->vm_shm) shmfork(p1, p2); - /* - * Allocate a wired-down (for now) pcb and kernel stack for the - * process - */ + pmap_new_proc(p2); - pvp = &p2->p_vmspace->vm_pmap; - - /* - * allocate object for the upages - */ - p2->p_vmspace->vm_upages_obj = upobj = vm_object_allocate( OBJT_DEFAULT, - UPAGES); - - /* get a kernel virtual address for the UPAGES for this proc */ - up = (struct user *) kmem_alloc_pageable(u_map, UPAGES * PAGE_SIZE); - if (up == NULL) - panic("vm_fork: u_map allocation failed"); - - for(i=0;iflags &= ~PG_ZERO; - m->flags |= PG_MAPPED|PG_WRITEABLE; - m->valid = VM_PAGE_BITS_ALL; - } - - p2->p_addr = up; + up = p2->p_addr; /* * p_stats and p_sigacts currently point at fields in the user struct @@ -318,51 +272,11 @@ faultin(p) int s; if ((p->p_flag & P_INMEM) == 0) { - pmap_t pmap = &p->p_vmspace->vm_pmap; - vm_page_t m; - vm_object_t upobj = p->p_vmspace->vm_upages_obj; ++p->p_lock; -#if defined(SWAP_DEBUG) - printf("swapping in %d\n", p->p_pid); -#endif - for(i=0;iflags & PG_BUSY) || m->busy) { - m->flags |= PG_WANTED; - tsleep(m, PVM, "swinuw",0); - goto retry; - } - m->flags |= PG_BUSY; - } - vm_page_wire(m); - splx(s); + pmap_swapin_proc(p); - pmap_enter( pmap, (vm_offset_t) kstack + i * PAGE_SIZE, - VM_PAGE_TO_PHYS(m), VM_PROT_READ|VM_PROT_WRITE, TRUE); - pmap_kenter(((vm_offset_t) p->p_addr) + i * PAGE_SIZE, - VM_PAGE_TO_PHYS(m)); - if (m->valid != VM_PAGE_BITS_ALL) { - int rv; - rv = vm_pager_get_pages(upobj, - &m, 1, 0); - if (rv != VM_PAGER_OK) - panic("faultin: cannot get upages for proc: %d\n", p->p_pid); - m->valid = VM_PAGE_BITS_ALL; - } - PAGE_WAKEUP(m); - m->flags |= PG_MAPPED|PG_WRITEABLE; - } - s = splhigh(); if (p->p_stat == SRUN) @@ -541,20 +455,7 @@ swapout(p) remrq(p); (void) spl0(); - /* - * let the upages be paged - */ - for(i=0;ip_vmspace->vm_upages_obj, i)) == NULL) - panic("swapout: upage already missing???"); - m->dirty = VM_PAGE_BITS_ALL; - vm_page_unwire(m); - vm_page_deactivate(m); - pmap_kremove( (vm_offset_t) p->p_addr + PAGE_SIZE * i); - } - pmap_remove(pmap, (vm_offset_t) kstack, - (vm_offset_t) kstack + PAGE_SIZE * UPAGES); + pmap_swapout_proc(p); p->p_flag &= ~P_SWAPPING; p->p_swtime = 0; diff --git a/sys/vm/vm_page.c b/sys/vm/vm_page.c index c145486a7548..ebf49e44dc38 100644 --- a/sys/vm/vm_page.c +++ b/sys/vm/vm_page.c @@ -34,7 +34,7 @@ * SUCH DAMAGE. * * from: @(#)vm_page.c 7.4 (Berkeley) 5/7/91 - * $Id: vm_page.c,v 1.67 1996/10/06 18:27:39 dyson Exp $ + * $Id: vm_page.c,v 1.68 1996/10/12 20:09:48 bde Exp $ */ /* @@ -684,9 +684,6 @@ vm_page_select_free(object, pindex, prefqueue) #if PQ_L2_SIZE > 1 index = pindex + object->pg_color; - /* - * These are special cased because of clock-arithemetic - */ for(j = 0; j < PQ_L1_SIZE; j++) { for(i = (PQ_L2_SIZE/2) - (PQ_L1_SIZE - 1); (i + j) >= 0;