/* * linux/ipc/shm.c * Copyright (C) 1992, 1993 Krishna Balasubramanian * Many improvements/fixes by Bruno Haible. * Replaced `struct shm_desc' by `struct vm_area_struct', July 1994. * Fixed the shm swap deallocation (shm_unuse()), August 1998 Andrea Arcangeli. * * /proc/sysvipc/shm support (c) 1999 Dragos Acostachioaie * BIGMEM support, Andrea Arcangeli * SMP thread shm, Jean-Luc Boyard */ #include #include #include #include #include #include #include #include #include #include #include #include extern int ipcperms (struct ipc_perm *ipcp, short shmflg); extern unsigned long get_swap_page (void); static int findkey (key_t key); static int newseg (key_t key, int shmflg, int size); static int shm_map (struct vm_area_struct *shmd); static void killseg (int id); static void shm_open (struct vm_area_struct *shmd); static void shm_close (struct vm_area_struct *shmd); static unsigned long shm_nopage(struct vm_area_struct *, unsigned long, int); static int shm_swapout(struct vm_area_struct *, struct page *); #ifdef CONFIG_PROC_FS static int sysvipc_shm_read_proc(char *buffer, char **start, off_t offset, int length, int *eof, void *data); #endif static int shm_tot = 0; /* total number of shared memory pages */ static int shm_rss = 0; /* number of shared memory pages that are in memory */ static int shm_swp = 0; /* number of shared memory pages that are in swap */ static int max_shmid = 0; /* every used id is <= max_shmid */ static DECLARE_WAIT_QUEUE_HEAD(shm_wait); /* calling findkey() may need to wait */ static struct shmid_kernel *shm_segs[SHMMNI]; static unsigned short shm_seq = 0; /* incremented, for recognizing stale ids */ spinlock_t shm_lock = SPIN_LOCK_UNLOCKED; /* some statistics */ static ulong swap_attempts = 0; static ulong swap_successes = 0; static ulong used_segs = 0; void __init shm_init (void) { int id; #ifdef CONFIG_PROC_FS struct proc_dir_entry *ent; #endif for (id = 0; id < SHMMNI; id++) shm_segs[id] = (struct shmid_kernel *) IPC_UNUSED; shm_tot = shm_rss = shm_seq = max_shmid = used_segs = 0; init_waitqueue_head(&shm_wait); #ifdef CONFIG_PROC_FS ent = create_proc_entry("sysvipc/shm", 0, 0); ent->read_proc = sysvipc_shm_read_proc; #endif return; } static int findkey (key_t key) { int id; struct shmid_kernel *shp; for (id = 0; id <= max_shmid; id++) { if ((shp = shm_segs[id]) == IPC_NOID) { DECLARE_WAITQUEUE(wait, current); add_wait_queue(&shm_wait, &wait); for(;;) { set_current_state(TASK_UNINTERRUPTIBLE); if ((shp = shm_segs[id]) != IPC_NOID) break; spin_unlock(&shm_lock); schedule(); spin_lock(&shm_lock); } __set_current_state(TASK_RUNNING); remove_wait_queue(&shm_wait, &wait); } if (shp == IPC_UNUSED) continue; if (key == shp->u.shm_perm.key) return id; } return -1; } /* * allocate new shmid_kernel and pgtable. protected by shm_segs[id] = NOID. */ static int newseg (key_t key, int shmflg, int size) { struct shmid_kernel *shp; int numpages = (size + PAGE_SIZE -1) >> PAGE_SHIFT; int id, i; if (size < SHMMIN) return -EINVAL; if (shm_tot + numpages >= SHMALL) return -ENOSPC; for (id = 0; id < SHMMNI; id++) if (shm_segs[id] == IPC_UNUSED) { shm_segs[id] = (struct shmid_kernel *) IPC_NOID; goto found; } return -ENOSPC; found: spin_unlock(&shm_lock); shp = (struct shmid_kernel *) kmalloc (sizeof (*shp), GFP_KERNEL); if (!shp) { spin_lock(&shm_lock); shm_segs[id] = (struct shmid_kernel *) IPC_UNUSED; wake_up (&shm_wait); return -ENOMEM; } lock_kernel(); shp->shm_pages = (ulong *) vmalloc (numpages*sizeof(ulong)); unlock_kernel(); if (!shp->shm_pages) { kfree(shp); spin_lock(&shm_lock); shm_segs[id] = (struct shmid_kernel *) IPC_UNUSED; wake_up (&shm_wait); return -ENOMEM; } for (i = 0; i < numpages; shp->shm_pages[i++] = 0); shp->u.shm_perm.key = key; shp->u.shm_perm.mode = (shmflg & S_IRWXUGO); shp->u.shm_perm.cuid = shp->u.shm_perm.uid = current->euid; shp->u.shm_perm.cgid = shp->u.shm_perm.gid = current->egid; shp->u.shm_segsz = size; shp->u.shm_cpid = current->pid; shp->attaches = NULL; shp->u.shm_lpid = shp->u.shm_nattch = 0; shp->u.shm_atime = shp->u.shm_dtime = 0; shp->u.shm_ctime = CURRENT_TIME; shp->shm_npages = numpages; spin_lock(&shm_lock); shm_tot += numpages; shp->u.shm_perm.seq = shm_seq; if (id > max_shmid) max_shmid = id; shm_segs[id] = shp; used_segs++; wake_up (&shm_wait); return (unsigned int) shp->u.shm_perm.seq * SHMMNI + id; } int shmmax = SHMMAX; asmlinkage long sys_shmget (key_t key, int size, int shmflg) { struct shmid_kernel *shp; int err, id = 0; down(¤t->mm->mmap_sem); spin_lock(&shm_lock); if (size < 0 || size > shmmax) { err = -EINVAL; } else if (key == IPC_PRIVATE) { err = newseg(key, shmflg, size); } else if ((id = findkey (key)) == -1) { if (!(shmflg & IPC_CREAT)) err = -ENOENT; else err = newseg(key, shmflg, size); } else if ((shmflg & IPC_CREAT) && (shmflg & IPC_EXCL)) { err = -EEXIST; } else { shp = shm_segs[id]; if (shp->u.shm_perm.mode & SHM_DEST) err = -EIDRM; else if (size > shp->u.shm_segsz) err = -EINVAL; else if (ipcperms (&shp->u.shm_perm, shmflg)) err = -EACCES; else err = (int) shp->u.shm_perm.seq * SHMMNI + id; } spin_unlock(&shm_lock); up(¤t->mm->mmap_sem); return err; } /* * Only called after testing nattch and SHM_DEST. * Here pages, pgtable and shmid_kernel are freed. */ static void killseg (int id) { struct shmid_kernel *shp; int i, numpages; int rss, swp; shp = shm_segs[id]; if (shp == IPC_NOID || shp == IPC_UNUSED) { printk ("shm nono: killseg called on unused seg id=%d\n", id); return; } shp->u.shm_perm.seq++; /* for shmat */ shm_seq = (shm_seq+1) % ((unsigned)(1<<31)/SHMMNI); /* increment, but avoid overflow */ shm_segs[id] = (struct shmid_kernel *) IPC_UNUSED; used_segs--; if (id == max_shmid) while (max_shmid && (shm_segs[--max_shmid] == IPC_UNUSED)); if (!shp->shm_pages) { printk ("shm nono: killseg shp->pages=NULL. id=%d\n", id); return; } spin_unlock(&shm_lock); numpages = shp->shm_npages; for (i = 0, rss = 0, swp = 0; i < numpages ; i++) { pte_t pte; pte = __pte(shp->shm_pages[i]); if (pte_none(pte)) continue; if (pte_present(pte)) { free_page (pte_page(pte)); rss++; } else { lock_kernel(); swap_free(pte_val(pte)); unlock_kernel(); swp++; } } lock_kernel(); vfree(shp->shm_pages); unlock_kernel(); kfree(shp); spin_lock(&shm_lock); shm_rss -= rss; shm_swp -= swp; shm_tot -= numpages; return; } asmlinkage long sys_shmctl (int shmid, int cmd, struct shmid_ds *buf) { struct shmid_ds tbuf; struct shmid_kernel *shp; struct ipc_perm *ipcp; int id, err = -EINVAL; if (cmd < 0 || shmid < 0) goto out_unlocked; if (cmd == IPC_SET) { err = -EFAULT; if(copy_from_user (&tbuf, buf, sizeof (*buf))) goto out_unlocked; } spin_lock(&shm_lock); switch (cmd) { /* replace with proc interface ? */ case IPC_INFO: { struct shminfo shminfo; err = -EFAULT; if (!buf) goto out; shminfo.shmmni = SHMMNI; shminfo.shmmax = shmmax; shminfo.shmmin = SHMMIN; shminfo.shmall = SHMALL; shminfo.shmseg = SHMSEG; spin_unlock(&shm_lock); if(copy_to_user (buf, &shminfo, sizeof(struct shminfo))) goto out_unlocked; spin_lock(&shm_lock); err = max_shmid; goto out; } case SHM_INFO: { struct shm_info shm_info; err = -EFAULT; shm_info.used_ids = used_segs; shm_info.shm_rss = shm_rss; shm_info.shm_tot = shm_tot; shm_info.shm_swp = shm_swp; shm_info.swap_attempts = swap_attempts; shm_info.swap_successes = swap_successes; spin_unlock(&shm_lock); if(copy_to_user (buf, &shm_info, sizeof(shm_info))) goto out_unlocked; spin_lock(&shm_lock); err = max_shmid; goto out; } case SHM_STAT: err = -EINVAL; if (shmid > max_shmid) goto out; shp = shm_segs[shmid]; if (shp == IPC_UNUSED || shp == IPC_NOID) goto out; if (ipcperms (&shp->u.shm_perm, S_IRUGO)) goto out; id = (unsigned int) shp->u.shm_perm.seq * SHMMNI + shmid; err = -EFAULT; spin_unlock(&shm_lock); if(copy_to_user (buf, &shp->u, sizeof(*buf))) goto out_unlocked; spin_lock(&shm_lock); err = id; goto out; } shp = shm_segs[id = (unsigned int) shmid % SHMMNI]; err = -EINVAL; if (shp == IPC_UNUSED || shp == IPC_NOID) goto out; err = -EIDRM; if (shp->u.shm_perm.seq != (unsigned int) shmid / SHMMNI) goto out; ipcp = &shp->u.shm_perm; switch (cmd) { case SHM_UNLOCK: err = -EPERM; if (!capable(CAP_IPC_LOCK)) goto out; err = -EINVAL; if (!(ipcp->mode & SHM_LOCKED)) goto out; ipcp->mode &= ~SHM_LOCKED; break; case SHM_LOCK: /* Allow superuser to lock segment in memory */ /* Should the pages be faulted in here or leave it to user? */ /* need to determine interaction with current->swappable */ err = -EPERM; if (!capable(CAP_IPC_LOCK)) goto out; err = -EINVAL; if (ipcp->mode & SHM_LOCKED) goto out; ipcp->mode |= SHM_LOCKED; break; case IPC_STAT: err = -EACCES; if (ipcperms (ipcp, S_IRUGO)) goto out; err = -EFAULT; spin_unlock(&shm_lock); if(copy_to_user (buf, &shp->u, sizeof(shp->u))) goto out_unlocked; spin_lock(&shm_lock); break; case IPC_SET: if (current->euid == shp->u.shm_perm.uid || current->euid == shp->u.shm_perm.cuid || capable(CAP_SYS_ADMIN)) { ipcp->uid = tbuf.shm_perm.uid; ipcp->gid = tbuf.shm_perm.gid; ipcp->mode = (ipcp->mode & ~S_IRWXUGO) | (tbuf.shm_perm.mode & S_IRWXUGO); shp->u.shm_ctime = CURRENT_TIME; break; } err = -EPERM; goto out; case IPC_RMID: if (current->euid == shp->u.shm_perm.uid || current->euid == shp->u.shm_perm.cuid || capable(CAP_SYS_ADMIN)) { shp->u.shm_perm.mode |= SHM_DEST; if (shp->u.shm_nattch <= 0) killseg (id); break; } err = -EPERM; goto out; default: err = -EINVAL; goto out; } err = 0; out: spin_unlock(&shm_lock); out_unlocked: return err; } /* * The per process internal structure for managing segments is * `struct vm_area_struct'. * A shmat will add to and shmdt will remove from the list. * shmd->vm_mm the attacher * shmd->vm_start virt addr of attach, multiple of SHMLBA * shmd->vm_end multiple of SHMLBA * shmd->vm_next next attach for task * shmd->vm_next_share next attach for segment * shmd->vm_offset offset into segment * shmd->vm_private_data signature for this attach */ static struct vm_operations_struct shm_vm_ops = { shm_open, /* open - callback for a new vm-area open */ shm_close, /* close - callback for when the vm-area is released */ NULL, /* no need to sync pages at unmap */ NULL, /* protect */ NULL, /* sync */ NULL, /* advise */ shm_nopage, /* nopage */ NULL, /* wppage */ shm_swapout /* swapout */ }; /* Insert shmd into the list shp->attaches */ static inline void insert_attach (struct shmid_kernel * shp, struct vm_area_struct * shmd) { if((shmd->vm_next_share = shp->attaches) != NULL) shp->attaches->vm_pprev_share = &shmd->vm_next_share; shp->attaches = shmd; shmd->vm_pprev_share = &shp->attaches; } /* Remove shmd from list shp->attaches */ static inline void remove_attach (struct shmid_kernel * shp, struct vm_area_struct * shmd) { if(shmd->vm_next_share) shmd->vm_next_share->vm_pprev_share = shmd->vm_pprev_share; *shmd->vm_pprev_share = shmd->vm_next_share; } /* * ensure page tables exist * mark page table entries with shm_sgn. */ static int shm_map (struct vm_area_struct *shmd) { unsigned long tmp; /* clear old mappings */ do_munmap(shmd->vm_start, shmd->vm_end - shmd->vm_start); /* add new mapping */ tmp = shmd->vm_end - shmd->vm_start; if((current->mm->total_vm << PAGE_SHIFT) + tmp > (unsigned long) current->rlim[RLIMIT_AS].rlim_cur) return -ENOMEM; current->mm->total_vm += tmp >> PAGE_SHIFT; insert_vm_struct(current->mm, shmd); merge_segments(current->mm, shmd->vm_start, shmd->vm_end); return 0; } /* * Fix shmaddr, allocate descriptor, map shm, add attach descriptor to lists. */ asmlinkage long sys_shmat (int shmid, char *shmaddr, int shmflg, ulong *raddr) { struct shmid_kernel *shp; struct vm_area_struct *shmd; int err = -EINVAL; unsigned int id; unsigned long addr; unsigned long len; down(¤t->mm->mmap_sem); spin_lock(&shm_lock); if (shmid < 0) { /* printk("shmat() -> EINVAL because shmid = %d < 0\n",shmid); */ goto out; } shp = shm_segs[id = (unsigned int) shmid % SHMMNI]; if (shp == IPC_UNUSED || shp == IPC_NOID) { /* printk("shmat() -> EINVAL because shmid = %d is invalid\n",shmid); */ goto out; } if (!(addr = (ulong) shmaddr)) { if (shmflg & SHM_REMAP) goto out; err = -ENOMEM; addr = 0; again: if (!(addr = get_unmapped_area(addr, shp->u.shm_segsz))) goto out; if(addr & (SHMLBA - 1)) { addr = (addr + (SHMLBA - 1)) & ~(SHMLBA - 1); goto again; } } else if (addr & (SHMLBA-1)) { if (shmflg & SHM_RND) addr &= ~(SHMLBA-1); /* round down */ else goto out; } /* * Check if addr exceeds TASK_SIZE (from do_mmap) */ len = PAGE_SIZE*shp->shm_npages; err = -EINVAL; if (addr >= TASK_SIZE || len > TASK_SIZE || addr > TASK_SIZE - len) goto out; /* * If shm segment goes below stack, make sure there is some * space left for the stack to grow (presently 4 pages). */ if (addr < current->mm->start_stack && addr > current->mm->start_stack - PAGE_SIZE*(shp->shm_npages + 4)) { /* printk("shmat() -> EINVAL because segment intersects stack\n"); */ goto out; } if (!(shmflg & SHM_REMAP)) if ((shmd = find_vma_intersection(current->mm, addr, addr + shp->u.shm_segsz))) { /* printk("shmat() -> EINVAL because the interval [0x%lx,0x%lx) intersects an already mapped interval [0x%lx,0x%lx).\n", addr, addr + shp->shm_segsz, shmd->vm_start, shmd->vm_end); */ goto out; } err = -EACCES; if (ipcperms(&shp->u.shm_perm, shmflg & SHM_RDONLY ? S_IRUGO : S_IRUGO|S_IWUGO)) goto out; err = -EIDRM; if (shp->u.shm_perm.seq != (unsigned int) shmid / SHMMNI) goto out; spin_unlock(&shm_lock); err = -ENOMEM; shmd = kmem_cache_alloc(vm_area_cachep, SLAB_KERNEL); spin_lock(&shm_lock); if (!shmd) goto out; if ((shp != shm_segs[id]) || (shp->u.shm_perm.seq != (unsigned int) shmid / SHMMNI)) { kmem_cache_free(vm_area_cachep, shmd); err = -EIDRM; goto out; } shmd->vm_private_data = shm_segs + id; shmd->vm_start = addr; shmd->vm_end = addr + shp->shm_npages * PAGE_SIZE; shmd->vm_mm = current->mm; shmd->vm_page_prot = (shmflg & SHM_RDONLY) ? PAGE_READONLY : PAGE_SHARED; shmd->vm_flags = VM_SHM | VM_MAYSHARE | VM_SHARED | VM_MAYREAD | VM_MAYEXEC | VM_READ | VM_EXEC | ((shmflg & SHM_RDONLY) ? 0 : VM_MAYWRITE | VM_WRITE); shmd->vm_file = NULL; shmd->vm_offset = 0; shmd->vm_ops = &shm_vm_ops; shp->u.shm_nattch++; /* prevent destruction */ spin_unlock(&shm_lock); err = shm_map (shmd); spin_lock(&shm_lock); if (err) goto failed_shm_map; insert_attach(shp,shmd); /* insert shmd into shp->attaches */ shp->u.shm_lpid = current->pid; shp->u.shm_atime = CURRENT_TIME; *raddr = addr; err = 0; out: spin_unlock(&shm_lock); up(¤t->mm->mmap_sem); return err; failed_shm_map: if (--shp->u.shm_nattch <= 0 && shp->u.shm_perm.mode & SHM_DEST) killseg(id); spin_unlock(&shm_lock); up(¤t->mm->mmap_sem); kmem_cache_free(vm_area_cachep, shmd); return err; } /* This is called by fork, once for every shm attach. */ static void shm_open (struct vm_area_struct *shmd) { struct shmid_kernel *shp; spin_lock(&shm_lock); shp = *(struct shmid_kernel **) shmd->vm_private_data; insert_attach(shp,shmd); /* insert shmd into shp->attaches */ shp->u.shm_nattch++; shp->u.shm_atime = CURRENT_TIME; shp->u.shm_lpid = current->pid; spin_unlock(&shm_lock); } /* * remove the attach descriptor shmd. * free memory for segment if it is marked destroyed. * The descriptor has already been removed from the current->mm->mmap list * and will later be kfree()d. */ static void shm_close (struct vm_area_struct *shmd) { struct shmid_kernel *shp; spin_lock(&shm_lock); /* remove from the list of attaches of the shm segment */ shp = *(struct shmid_kernel **) shmd->vm_private_data; remove_attach(shp,shmd); /* remove from shp->attaches */ shp->u.shm_lpid = current->pid; shp->u.shm_dtime = CURRENT_TIME; if (--shp->u.shm_nattch <= 0 && shp->u.shm_perm.mode & SHM_DEST) { unsigned int id = (struct shmid_kernel **)shmd->vm_private_data - shm_segs; killseg (id); } spin_unlock(&shm_lock); } /* * detach and kill segment if marked destroyed. * The work is done in shm_close. */ asmlinkage long sys_shmdt (char *shmaddr) { struct vm_area_struct *shmd, *shmdnext; down(¤t->mm->mmap_sem); for (shmd = current->mm->mmap; shmd; shmd = shmdnext) { shmdnext = shmd->vm_next; if (shmd->vm_ops == &shm_vm_ops && shmd->vm_start - shmd->vm_offset == (ulong) shmaddr) do_munmap(shmd->vm_start, shmd->vm_end - shmd->vm_start); } up(¤t->mm->mmap_sem); return 0; } /* * Enter the shm page into the SHM data structures. * * The way "nopage" is done, we don't actually have to * do anything here: nopage will have filled in the shm * data structures already, and shm_swap_out() will just * work off them.. */ static int shm_swapout(struct vm_area_struct * vma, struct page * page) { return 0; } /* * page not present ... go through shm_pages */ static unsigned long shm_nopage(struct vm_area_struct * shmd, unsigned long address, int no_share) { pte_t pte; struct shmid_kernel *shp; unsigned int idx; unsigned long page; struct page * page_map; shp = *(struct shmid_kernel **) shmd->vm_private_data; idx = (address - shmd->vm_start + shmd->vm_offset) >> PAGE_SHIFT; #ifdef DEBUG_SHM if (shp == IPC_UNUSED || shp == IPC_NOID) { printk ("shm_nopage: id=%d invalid. Race.\n", id); return 0; } if (idx >= shp->shm_npages) { printk ("shm_nopage : too large page index. id=%d\n", id); return 0; } #endif spin_lock(&shm_lock); again: pte = __pte(shp->shm_pages[idx]); if (!pte_present(pte)) { if (pte_none(pte)) { spin_unlock(&shm_lock); page = __get_free_page(GFP_BIGUSER); if (!page) goto oom; clear_bigpage(page); spin_lock(&shm_lock); if (pte_val(pte) != shp->shm_pages[idx]) goto changed; } else { unsigned long entry = pte_val(pte); spin_unlock(&shm_lock); page_map = lookup_swap_cache(entry); if (!page_map) { lock_kernel(); swapin_readahead(entry); page_map = read_swap_cache(entry); unlock_kernel(); if (!page_map) goto oom; } delete_from_swap_cache(page_map); page_map = replace_with_bigmem(page_map); page = page_address(page_map); lock_kernel(); swap_free(entry); unlock_kernel(); spin_lock(&shm_lock); shm_swp--; pte = __pte(shp->shm_pages[idx]); if (pte_present(pte)) goto present; } shm_rss++; pte = pte_mkdirty(mk_pte(page, PAGE_SHARED)); shp->shm_pages[idx] = pte_val(pte); } else --current->maj_flt; /* was incremented in do_no_page */ done: /* pte_val(pte) == shp->shm_pages[idx] */ get_page(mem_map + MAP_NR(pte_page(pte))); spin_unlock(&shm_lock); current->min_flt++; return pte_page(pte); changed: free_page(page); goto again; present: free_page(page); goto done; oom: return -1; } /* * Goes through counter = (shm_rss >> prio) present shm pages. */ static unsigned long swap_id = 0; /* currently being swapped */ static unsigned long swap_idx = 0; /* next to swap */ int shm_swap (int prio, int gfp_mask) { pte_t page; struct shmid_kernel *shp; unsigned long swap_nr; unsigned long id, idx; int loop = 0; int counter; struct page * page_map; counter = shm_rss >> prio; lock_kernel(); if (!counter || !(swap_nr = get_swap_page())) { unlock_kernel(); return 0; } unlock_kernel(); spin_lock(&shm_lock); check_id: shp = shm_segs[swap_id]; if (shp == IPC_UNUSED || shp == IPC_NOID || shp->u.shm_perm.mode & SHM_LOCKED ) { next_id: swap_idx = 0; if (++swap_id > max_shmid) { swap_id = 0; if (loop) goto failed; loop = 1; } goto check_id; } id = swap_id; check_table: idx = swap_idx++; if (idx >= shp->shm_npages) goto next_id; page = __pte(shp->shm_pages[idx]); if (!pte_present(page)) goto check_table; page_map = &mem_map[MAP_NR(pte_page(page))]; if ((gfp_mask & __GFP_DMA) && !PageDMA(page_map)) goto check_table; if (!(gfp_mask & __GFP_BIGMEM) && PageBIGMEM(page_map)) goto check_table; swap_attempts++; if (--counter < 0) { /* failed */ failed: spin_unlock(&shm_lock); lock_kernel(); swap_free (swap_nr); unlock_kernel(); return 0; } if (page_count(mem_map + MAP_NR(pte_page(page))) != 1) goto check_table; if (!(page_map = prepare_bigmem_swapout(page_map))) goto check_table; shp->shm_pages[idx] = swap_nr; swap_successes++; shm_swp++; shm_rss--; spin_unlock(&shm_lock); lock_kernel(); swap_duplicate(swap_nr); add_to_swap_cache(page_map, swap_nr); rw_swap_page(WRITE, page_map, 0); unlock_kernel(); __free_page(page_map); return 1; } /* * Free the swap entry and set the new pte for the shm page. */ static void shm_unuse_page(struct shmid_kernel *shp, unsigned long idx, unsigned long page, unsigned long entry) { pte_t pte; pte = pte_mkdirty(mk_pte(page, PAGE_SHARED)); shp->shm_pages[idx] = pte_val(pte); get_page(mem_map + MAP_NR(page)); shm_rss++; shm_swp--; spin_unlock(&shm_lock); lock_kernel(); swap_free(entry); unlock_kernel(); } /* * unuse_shm() search for an eventually swapped out shm page. */ void shm_unuse(unsigned long entry, unsigned long page) { int i, n; spin_lock(&shm_lock); for (i = 0; i < SHMMNI; i++) if (shm_segs[i] != IPC_UNUSED && shm_segs[i] != IPC_NOID) for (n = 0; n < shm_segs[i]->shm_npages; n++) if (shm_segs[i]->shm_pages[n] == entry) { shm_unuse_page(shm_segs[i], n, page, entry); return; } spin_unlock(&shm_lock); } #ifdef CONFIG_PROC_FS static int sysvipc_shm_read_proc(char *buffer, char **start, off_t offset, int length, int *eof, void *data) { off_t pos = 0; off_t begin = 0; int i, len = 0; len += sprintf(buffer, " key shmid perms size cpid lpid nattch uid gid cuid cgid atime dtime ctime\n"); spin_lock(&shm_lock); for(i = 0; i < SHMMNI; i++) if(shm_segs[i] != IPC_UNUSED) { len += sprintf(buffer + len, "%10d %10d %4o %10d %5u %5u %5d %5u %5u %5u %5u %10lu %10lu %10lu\n", shm_segs[i]->u.shm_perm.key, shm_segs[i]->u.shm_perm.seq * SHMMNI + i, shm_segs[i]->u.shm_perm.mode, shm_segs[i]->u.shm_segsz, shm_segs[i]->u.shm_cpid, shm_segs[i]->u.shm_lpid, shm_segs[i]->u.shm_nattch, shm_segs[i]->u.shm_perm.uid, shm_segs[i]->u.shm_perm.gid, shm_segs[i]->u.shm_perm.cuid, shm_segs[i]->u.shm_perm.cgid, shm_segs[i]->u.shm_atime, shm_segs[i]->u.shm_dtime, shm_segs[i]->u.shm_ctime); pos += len; if(pos < offset) { len = 0; begin = pos; } if(pos > offset + length) goto done; } *eof = 1; done: *start = buffer + (offset - begin); len -= (offset - begin); if(len > length) len = length; if(len < 0) len = 0; spin_unlock(&shm_lock); return len; } #endif