Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
util.c
Go to the documentation of this file.
1 /*
2  * linux/ipc/util.c
3  * Copyright (C) 1992 Krishna Balasubramanian
4  *
5  * Sep 1997 - Call suser() last after "normal" permission checks so we
6  * get BSD style process accounting right.
7  * Occurs in several places in the IPC code.
8  * Chris Evans, <[email protected]>
9  * Nov 1999 - ipc helper functions, unified SMP locking
10  * Manfred Spraul <[email protected]>
11  * Oct 2002 - One lock per IPC id. RCU ipc_free for lock-free grow_ary().
12  * Mingming Cao <[email protected]>
13  * Mar 2006 - support for audit of ipc object properties
14  * Dustin Kirkland <[email protected]>
15  * Jun 2006 - namespaces ssupport
16  * OpenVZ, SWsoft Inc.
17  * Pavel Emelianov <[email protected]>
18  */
19 
20 #include <linux/mm.h>
21 #include <linux/shm.h>
22 #include <linux/init.h>
23 #include <linux/msg.h>
24 #include <linux/vmalloc.h>
25 #include <linux/slab.h>
26 #include <linux/capability.h>
27 #include <linux/highuid.h>
28 #include <linux/security.h>
29 #include <linux/rcupdate.h>
30 #include <linux/workqueue.h>
31 #include <linux/seq_file.h>
32 #include <linux/proc_fs.h>
33 #include <linux/audit.h>
34 #include <linux/nsproxy.h>
35 #include <linux/rwsem.h>
36 #include <linux/memory.h>
37 #include <linux/ipc_namespace.h>
38 
39 #include <asm/unistd.h>
40 
41 #include "util.h"
42 
44  const char *path;
45  const char *header;
46  int ids;
47  int (*show)(struct seq_file *, void *);
48 };
49 
50 #ifdef CONFIG_MEMORY_HOTPLUG
51 
52 static void ipc_memory_notifier(struct work_struct *work)
53 {
55 }
56 
57 static DECLARE_WORK(ipc_memory_wq, ipc_memory_notifier);
58 
59 
60 static int ipc_memory_callback(struct notifier_block *self,
61  unsigned long action, void *arg)
62 {
63  switch (action) {
64  case MEM_ONLINE: /* memory successfully brought online */
65  case MEM_OFFLINE: /* or offline: it's time to recompute msgmni */
66  /*
67  * This is done by invoking the ipcns notifier chain with the
68  * IPC_MEMCHANGED event.
69  * In order not to keep the lock on the hotplug memory chain
70  * for too long, queue a work item that will, when waken up,
71  * activate the ipcns notification chain.
72  * No need to keep several ipc work items on the queue.
73  */
74  if (!work_pending(&ipc_memory_wq))
75  schedule_work(&ipc_memory_wq);
76  break;
77  case MEM_GOING_ONLINE:
78  case MEM_GOING_OFFLINE:
79  case MEM_CANCEL_ONLINE:
80  case MEM_CANCEL_OFFLINE:
81  default:
82  break;
83  }
84 
85  return NOTIFY_OK;
86 }
87 
88 #endif /* CONFIG_MEMORY_HOTPLUG */
89 
100 static int __init ipc_init(void)
101 {
102  sem_init();
103  msg_init();
104  shm_init();
105  hotplug_memory_notifier(ipc_memory_callback, IPC_CALLBACK_PRI);
107  return 0;
108 }
109 __initcall(ipc_init);
110 
119 void ipc_init_ids(struct ipc_ids *ids)
120 {
121  init_rwsem(&ids->rw_mutex);
122 
123  ids->in_use = 0;
124  ids->seq = 0;
125  {
126  int seq_limit = INT_MAX/SEQ_MULTIPLIER;
127  if (seq_limit > USHRT_MAX)
128  ids->seq_max = USHRT_MAX;
129  else
130  ids->seq_max = seq_limit;
131  }
132 
133  idr_init(&ids->ipcs_idr);
134 }
135 
136 #ifdef CONFIG_PROC_FS
137 static const struct file_operations sysvipc_proc_fops;
145 void __init ipc_init_proc_interface(const char *path, const char *header,
146  int ids, int (*show)(struct seq_file *, void *))
147 {
148  struct proc_dir_entry *pde;
149  struct ipc_proc_iface *iface;
150 
151  iface = kmalloc(sizeof(*iface), GFP_KERNEL);
152  if (!iface)
153  return;
154  iface->path = path;
155  iface->header = header;
156  iface->ids = ids;
157  iface->show = show;
158 
159  pde = proc_create_data(path,
160  S_IRUGO, /* world readable */
161  NULL, /* parent dir */
162  &sysvipc_proc_fops,
163  iface);
164  if (!pde) {
165  kfree(iface);
166  }
167 }
168 #endif
169 
181 static struct kern_ipc_perm *ipc_findkey(struct ipc_ids *ids, key_t key)
182 {
183  struct kern_ipc_perm *ipc;
184  int next_id;
185  int total;
186 
187  for (total = 0, next_id = 0; total < ids->in_use; next_id++) {
188  ipc = idr_find(&ids->ipcs_idr, next_id);
189 
190  if (ipc == NULL)
191  continue;
192 
193  if (ipc->key != key) {
194  total++;
195  continue;
196  }
197 
198  ipc_lock_by_ptr(ipc);
199  return ipc;
200  }
201 
202  return NULL;
203 }
204 
212 int ipc_get_maxid(struct ipc_ids *ids)
213 {
214  struct kern_ipc_perm *ipc;
215  int max_id = -1;
216  int total, id;
217 
218  if (ids->in_use == 0)
219  return -1;
220 
221  if (ids->in_use == IPCMNI)
222  return IPCMNI - 1;
223 
224  /* Look for the last assigned id */
225  total = 0;
226  for (id = 0; id < IPCMNI && total < ids->in_use; id++) {
227  ipc = idr_find(&ids->ipcs_idr, id);
228  if (ipc != NULL) {
229  max_id = id;
230  total++;
231  }
232  }
233  return max_id;
234 }
235 
250 int ipc_addid(struct ipc_ids* ids, struct kern_ipc_perm* new, int size)
251 {
252  kuid_t euid;
253  kgid_t egid;
254  int id, err;
255 
256  if (size > IPCMNI)
257  size = IPCMNI;
258 
259  if (ids->in_use >= size)
260  return -ENOSPC;
261 
262  spin_lock_init(&new->lock);
263  new->deleted = 0;
264  rcu_read_lock();
265  spin_lock(&new->lock);
266 
267  err = idr_get_new(&ids->ipcs_idr, new, &id);
268  if (err) {
269  spin_unlock(&new->lock);
270  rcu_read_unlock();
271  return err;
272  }
273 
274  ids->in_use++;
275 
276  current_euid_egid(&euid, &egid);
277  new->cuid = new->uid = euid;
278  new->gid = new->cgid = egid;
279 
280  new->seq = ids->seq++;
281  if(ids->seq > ids->seq_max)
282  ids->seq = 0;
283 
284  new->id = ipc_buildid(id, new->seq);
285  return id;
286 }
287 
298 static int ipcget_new(struct ipc_namespace *ns, struct ipc_ids *ids,
299  struct ipc_ops *ops, struct ipc_params *params)
300 {
301  int err;
302 retry:
303  err = idr_pre_get(&ids->ipcs_idr, GFP_KERNEL);
304 
305  if (!err)
306  return -ENOMEM;
307 
308  down_write(&ids->rw_mutex);
309  err = ops->getnew(ns, params);
310  up_write(&ids->rw_mutex);
311 
312  if (err == -EAGAIN)
313  goto retry;
314 
315  return err;
316 }
317 
333 static int ipc_check_perms(struct ipc_namespace *ns,
334  struct kern_ipc_perm *ipcp,
335  struct ipc_ops *ops,
336  struct ipc_params *params)
337 {
338  int err;
339 
340  if (ipcperms(ns, ipcp, params->flg))
341  err = -EACCES;
342  else {
343  err = ops->associate(ipcp, params->flg);
344  if (!err)
345  err = ipcp->id;
346  }
347 
348  return err;
349 }
350 
365 static int ipcget_public(struct ipc_namespace *ns, struct ipc_ids *ids,
366  struct ipc_ops *ops, struct ipc_params *params)
367 {
368  struct kern_ipc_perm *ipcp;
369  int flg = params->flg;
370  int err;
371 retry:
372  err = idr_pre_get(&ids->ipcs_idr, GFP_KERNEL);
373 
374  /*
375  * Take the lock as a writer since we are potentially going to add
376  * a new entry + read locks are not "upgradable"
377  */
378  down_write(&ids->rw_mutex);
379  ipcp = ipc_findkey(ids, params->key);
380  if (ipcp == NULL) {
381  /* key not used */
382  if (!(flg & IPC_CREAT))
383  err = -ENOENT;
384  else if (!err)
385  err = -ENOMEM;
386  else
387  err = ops->getnew(ns, params);
388  } else {
389  /* ipc object has been locked by ipc_findkey() */
390 
391  if (flg & IPC_CREAT && flg & IPC_EXCL)
392  err = -EEXIST;
393  else {
394  err = 0;
395  if (ops->more_checks)
396  err = ops->more_checks(ipcp, params);
397  if (!err)
398  /*
399  * ipc_check_perms returns the IPC id on
400  * success
401  */
402  err = ipc_check_perms(ns, ipcp, ops, params);
403  }
404  ipc_unlock(ipcp);
405  }
406  up_write(&ids->rw_mutex);
407 
408  if (err == -EAGAIN)
409  goto retry;
410 
411  return err;
412 }
413 
414 
424 void ipc_rmid(struct ipc_ids *ids, struct kern_ipc_perm *ipcp)
425 {
426  int lid = ipcid_to_idx(ipcp->id);
427 
428  idr_remove(&ids->ipcs_idr, lid);
429 
430  ids->in_use--;
431 
432  ipcp->deleted = 1;
433 
434  return;
435 }
436 
445 void* ipc_alloc(int size)
446 {
447  void* out;
448  if(size > PAGE_SIZE)
449  out = vmalloc(size);
450  else
451  out = kmalloc(size, GFP_KERNEL);
452  return out;
453 }
454 
464 void ipc_free(void* ptr, int size)
465 {
466  if(size > PAGE_SIZE)
467  vfree(ptr);
468  else
469  kfree(ptr);
470 }
471 
472 /*
473  * rcu allocations:
474  * There are three headers that are prepended to the actual allocation:
475  * - during use: ipc_rcu_hdr.
476  * - during the rcu grace period: ipc_rcu_grace.
477  * - [only if vmalloc]: ipc_rcu_sched.
478  * Their lifetime doesn't overlap, thus the headers share the same memory.
479  * Unlike a normal union, they are right-aligned, thus some container_of
480  * forward/backward casting is necessary:
481  */
483 {
484  int refcount;
486  void *data[0];
487 };
488 
489 
491 {
492  struct rcu_head rcu;
493  /* "void *" makes sure alignment of following data is sane. */
494  void *data[0];
495 };
496 
498 {
500  /* "void *" makes sure alignment of following data is sane. */
501  void *data[0];
502 };
503 
504 #define HDRLEN_KMALLOC (sizeof(struct ipc_rcu_grace) > sizeof(struct ipc_rcu_hdr) ? \
505  sizeof(struct ipc_rcu_grace) : sizeof(struct ipc_rcu_hdr))
506 #define HDRLEN_VMALLOC (sizeof(struct ipc_rcu_sched) > HDRLEN_KMALLOC ? \
507  sizeof(struct ipc_rcu_sched) : HDRLEN_KMALLOC)
508 
509 static inline int rcu_use_vmalloc(int size)
510 {
511  /* Too big for a single page? */
512  if (HDRLEN_KMALLOC + size > PAGE_SIZE)
513  return 1;
514  return 0;
515 }
516 
526 void* ipc_rcu_alloc(int size)
527 {
528  void* out;
529  /*
530  * We prepend the allocation with the rcu struct, and
531  * workqueue if necessary (for vmalloc).
532  */
533  if (rcu_use_vmalloc(size)) {
534  out = vmalloc(HDRLEN_VMALLOC + size);
535  if (out) {
536  out += HDRLEN_VMALLOC;
537  container_of(out, struct ipc_rcu_hdr, data)->is_vmalloc = 1;
538  container_of(out, struct ipc_rcu_hdr, data)->refcount = 1;
539  }
540  } else {
541  out = kmalloc(HDRLEN_KMALLOC + size, GFP_KERNEL);
542  if (out) {
543  out += HDRLEN_KMALLOC;
544  container_of(out, struct ipc_rcu_hdr, data)->is_vmalloc = 0;
545  container_of(out, struct ipc_rcu_hdr, data)->refcount = 1;
546  }
547  }
548 
549  return out;
550 }
551 
552 void ipc_rcu_getref(void *ptr)
553 {
554  container_of(ptr, struct ipc_rcu_hdr, data)->refcount++;
555 }
556 
557 static void ipc_do_vfree(struct work_struct *work)
558 {
559  vfree(container_of(work, struct ipc_rcu_sched, work));
560 }
561 
569 static void ipc_schedule_free(struct rcu_head *head)
570 {
571  struct ipc_rcu_grace *grace;
572  struct ipc_rcu_sched *sched;
573 
574  grace = container_of(head, struct ipc_rcu_grace, rcu);
575  sched = container_of(&(grace->data[0]), struct ipc_rcu_sched,
576  data[0]);
577 
578  INIT_WORK(&sched->work, ipc_do_vfree);
579  schedule_work(&sched->work);
580 }
581 
582 void ipc_rcu_putref(void *ptr)
583 {
584  if (--container_of(ptr, struct ipc_rcu_hdr, data)->refcount > 0)
585  return;
586 
587  if (container_of(ptr, struct ipc_rcu_hdr, data)->is_vmalloc) {
588  call_rcu(&container_of(ptr, struct ipc_rcu_grace, data)->rcu,
589  ipc_schedule_free);
590  } else {
591  kfree_rcu(container_of(ptr, struct ipc_rcu_grace, data), rcu);
592  }
593 }
594 
607 int ipcperms(struct ipc_namespace *ns, struct kern_ipc_perm *ipcp, short flag)
608 {
609  kuid_t euid = current_euid();
610  int requested_mode, granted_mode;
611 
612  audit_ipc_obj(ipcp);
613  requested_mode = (flag >> 6) | (flag >> 3) | flag;
614  granted_mode = ipcp->mode;
615  if (uid_eq(euid, ipcp->cuid) ||
616  uid_eq(euid, ipcp->uid))
617  granted_mode >>= 6;
618  else if (in_group_p(ipcp->cgid) || in_group_p(ipcp->gid))
619  granted_mode >>= 3;
620  /* is there some bit set in requested_mode but not in granted_mode? */
621  if ((requested_mode & ~granted_mode & 0007) &&
623  return -1;
624 
625  return security_ipc_permission(ipcp, flag);
626 }
627 
628 /*
629  * Functions to convert between the kern_ipc_perm structure and the
630  * old/new ipc_perm structures
631  */
632 
644 {
645  out->key = in->key;
646  out->uid = from_kuid_munged(current_user_ns(), in->uid);
647  out->gid = from_kgid_munged(current_user_ns(), in->gid);
648  out->cuid = from_kuid_munged(current_user_ns(), in->cuid);
649  out->cgid = from_kgid_munged(current_user_ns(), in->cgid);
650  out->mode = in->mode;
651  out->seq = in->seq;
652 }
653 
664 {
665  out->key = in->key;
666  SET_UID(out->uid, in->uid);
667  SET_GID(out->gid, in->gid);
668  SET_UID(out->cuid, in->cuid);
669  SET_GID(out->cgid, in->cgid);
670  out->mode = in->mode;
671  out->seq = in->seq;
672 }
673 
684 struct kern_ipc_perm *ipc_lock(struct ipc_ids *ids, int id)
685 {
686  struct kern_ipc_perm *out;
687  int lid = ipcid_to_idx(id);
688 
689  rcu_read_lock();
690  out = idr_find(&ids->ipcs_idr, lid);
691  if (out == NULL) {
692  rcu_read_unlock();
693  return ERR_PTR(-EINVAL);
694  }
695 
696  spin_lock(&out->lock);
697 
698  /* ipc_rmid() may have already freed the ID while ipc_lock
699  * was spinning: here verify that the structure is still valid
700  */
701  if (out->deleted) {
702  spin_unlock(&out->lock);
703  rcu_read_unlock();
704  return ERR_PTR(-EINVAL);
705  }
706 
707  return out;
708 }
709 
710 struct kern_ipc_perm *ipc_lock_check(struct ipc_ids *ids, int id)
711 {
712  struct kern_ipc_perm *out;
713 
714  out = ipc_lock(ids, id);
715  if (IS_ERR(out))
716  return out;
717 
718  if (ipc_checkid(out, id)) {
719  ipc_unlock(out);
720  return ERR_PTR(-EIDRM);
721  }
722 
723  return out;
724 }
725 
736 int ipcget(struct ipc_namespace *ns, struct ipc_ids *ids,
737  struct ipc_ops *ops, struct ipc_params *params)
738 {
739  if (params->key == IPC_PRIVATE)
740  return ipcget_new(ns, ids, ops, params);
741  else
742  return ipcget_public(ns, ids, ops, params);
743 }
744 
751 {
754  if (!uid_valid(uid) || !gid_valid(gid))
755  return -EINVAL;
756 
757  out->uid = uid;
758  out->gid = gid;
759  out->mode = (out->mode & ~S_IRWXUGO)
760  | (in->mode & S_IRWXUGO);
761 
762  return 0;
763 }
764 
783  struct ipc_ids *ids, int id, int cmd,
784  struct ipc64_perm *perm, int extra_perm)
785 {
786  struct kern_ipc_perm *ipcp;
787  kuid_t euid;
788  int err;
789 
790  down_write(&ids->rw_mutex);
791  ipcp = ipc_lock_check(ids, id);
792  if (IS_ERR(ipcp)) {
793  err = PTR_ERR(ipcp);
794  goto out_up;
795  }
796 
797  audit_ipc_obj(ipcp);
798  if (cmd == IPC_SET)
799  audit_ipc_set_perm(extra_perm, perm->uid,
800  perm->gid, perm->mode);
801 
802  euid = current_euid();
803  if (uid_eq(euid, ipcp->cuid) || uid_eq(euid, ipcp->uid) ||
805  return ipcp;
806 
807  err = -EPERM;
808  ipc_unlock(ipcp);
809 out_up:
810  up_write(&ids->rw_mutex);
811  return ERR_PTR(err);
812 }
813 
814 #ifdef CONFIG_ARCH_WANT_IPC_PARSE_VERSION
815 
816 
826 int ipc_parse_version (int *cmd)
827 {
828  if (*cmd & IPC_64) {
829  *cmd ^= IPC_64;
830  return IPC_64;
831  } else {
832  return IPC_OLD;
833  }
834 }
835 
836 #endif /* CONFIG_ARCH_WANT_IPC_PARSE_VERSION */
837 
838 #ifdef CONFIG_PROC_FS
839 struct ipc_proc_iter {
840  struct ipc_namespace *ns;
841  struct ipc_proc_iface *iface;
842 };
843 
844 /*
845  * This routine locks the ipc structure found at least at position pos.
846  */
847 static struct kern_ipc_perm *sysvipc_find_ipc(struct ipc_ids *ids, loff_t pos,
848  loff_t *new_pos)
849 {
850  struct kern_ipc_perm *ipc;
851  int total, id;
852 
853  total = 0;
854  for (id = 0; id < pos && total < ids->in_use; id++) {
855  ipc = idr_find(&ids->ipcs_idr, id);
856  if (ipc != NULL)
857  total++;
858  }
859 
860  if (total >= ids->in_use)
861  return NULL;
862 
863  for ( ; pos < IPCMNI; pos++) {
864  ipc = idr_find(&ids->ipcs_idr, pos);
865  if (ipc != NULL) {
866  *new_pos = pos + 1;
867  ipc_lock_by_ptr(ipc);
868  return ipc;
869  }
870  }
871 
872  /* Out of range - return NULL to terminate iteration */
873  return NULL;
874 }
875 
876 static void *sysvipc_proc_next(struct seq_file *s, void *it, loff_t *pos)
877 {
878  struct ipc_proc_iter *iter = s->private;
879  struct ipc_proc_iface *iface = iter->iface;
880  struct kern_ipc_perm *ipc = it;
881 
882  /* If we had an ipc id locked before, unlock it */
883  if (ipc && ipc != SEQ_START_TOKEN)
884  ipc_unlock(ipc);
885 
886  return sysvipc_find_ipc(&iter->ns->ids[iface->ids], *pos, pos);
887 }
888 
889 /*
890  * File positions: pos 0 -> header, pos n -> ipc id = n - 1.
891  * SeqFile iterator: iterator value locked ipc pointer or SEQ_TOKEN_START.
892  */
893 static void *sysvipc_proc_start(struct seq_file *s, loff_t *pos)
894 {
895  struct ipc_proc_iter *iter = s->private;
896  struct ipc_proc_iface *iface = iter->iface;
897  struct ipc_ids *ids;
898 
899  ids = &iter->ns->ids[iface->ids];
900 
901  /*
902  * Take the lock - this will be released by the corresponding
903  * call to stop().
904  */
905  down_read(&ids->rw_mutex);
906 
907  /* pos < 0 is invalid */
908  if (*pos < 0)
909  return NULL;
910 
911  /* pos == 0 means header */
912  if (*pos == 0)
913  return SEQ_START_TOKEN;
914 
915  /* Find the (pos-1)th ipc */
916  return sysvipc_find_ipc(ids, *pos - 1, pos);
917 }
918 
919 static void sysvipc_proc_stop(struct seq_file *s, void *it)
920 {
921  struct kern_ipc_perm *ipc = it;
922  struct ipc_proc_iter *iter = s->private;
923  struct ipc_proc_iface *iface = iter->iface;
924  struct ipc_ids *ids;
925 
926  /* If we had a locked structure, release it */
927  if (ipc && ipc != SEQ_START_TOKEN)
928  ipc_unlock(ipc);
929 
930  ids = &iter->ns->ids[iface->ids];
931  /* Release the lock we took in start() */
932  up_read(&ids->rw_mutex);
933 }
934 
935 static int sysvipc_proc_show(struct seq_file *s, void *it)
936 {
937  struct ipc_proc_iter *iter = s->private;
938  struct ipc_proc_iface *iface = iter->iface;
939 
940  if (it == SEQ_START_TOKEN)
941  return seq_puts(s, iface->header);
942 
943  return iface->show(s, it);
944 }
945 
946 static const struct seq_operations sysvipc_proc_seqops = {
947  .start = sysvipc_proc_start,
948  .stop = sysvipc_proc_stop,
949  .next = sysvipc_proc_next,
950  .show = sysvipc_proc_show,
951 };
952 
953 static int sysvipc_proc_open(struct inode *inode, struct file *file)
954 {
955  int ret;
956  struct seq_file *seq;
957  struct ipc_proc_iter *iter;
958 
959  ret = -ENOMEM;
960  iter = kmalloc(sizeof(*iter), GFP_KERNEL);
961  if (!iter)
962  goto out;
963 
964  ret = seq_open(file, &sysvipc_proc_seqops);
965  if (ret)
966  goto out_kfree;
967 
968  seq = file->private_data;
969  seq->private = iter;
970 
971  iter->iface = PDE(inode)->data;
972  iter->ns = get_ipc_ns(current->nsproxy->ipc_ns);
973 out:
974  return ret;
975 out_kfree:
976  kfree(iter);
977  goto out;
978 }
979 
980 static int sysvipc_proc_release(struct inode *inode, struct file *file)
981 {
982  struct seq_file *seq = file->private_data;
983  struct ipc_proc_iter *iter = seq->private;
984  put_ipc_ns(iter->ns);
985  return seq_release_private(inode, file);
986 }
987 
988 static const struct file_operations sysvipc_proc_fops = {
989  .open = sysvipc_proc_open,
990  .read = seq_read,
991  .llseek = seq_lseek,
992  .release = sysvipc_proc_release,
993 };
994 #endif /* CONFIG_PROC_FS */