Blame view

include/linux/rculist_bl.h 4.29 KB
4e35e6070   Nick Piggin   kernel: add bl_list
1
2
3
4
5
6
7
8
9
10
11
12
13
  #ifndef _LINUX_RCULIST_BL_H
  #define _LINUX_RCULIST_BL_H
  
  /*
   * RCU-protected bl list version. See include/linux/list_bl.h.
   */
  #include <linux/list_bl.h>
  #include <linux/rcupdate.h>
  
  static inline void hlist_bl_set_first_rcu(struct hlist_bl_head *h,
  					struct hlist_bl_node *n)
  {
  	LIST_BL_BUG_ON((unsigned long)n & LIST_BL_LOCKMASK);
2c6755988   Nick Piggin   fs: hlist UP debu...
14
15
  	LIST_BL_BUG_ON(((unsigned long)h->first & LIST_BL_LOCKMASK) !=
  							LIST_BL_LOCKMASK);
4e35e6070   Nick Piggin   kernel: add bl_list
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
  	rcu_assign_pointer(h->first,
  		(struct hlist_bl_node *)((unsigned long)n | LIST_BL_LOCKMASK));
  }
  
  static inline struct hlist_bl_node *hlist_bl_first_rcu(struct hlist_bl_head *h)
  {
  	return (struct hlist_bl_node *)
  		((unsigned long)rcu_dereference(h->first) & ~LIST_BL_LOCKMASK);
  }
  
  /**
   * hlist_bl_del_init_rcu - deletes entry from hash list with re-initialization
   * @n: the element to delete from the hash list.
   *
   * Note: hlist_bl_unhashed() on the node returns true after this. It is
   * useful for RCU based read lockfree traversal if the writer side
   * must know if the list entry is still hashed or already unhashed.
   *
   * In particular, it means that we can not poison the forward pointers
   * that may still be used for walking the hash list and we can only
   * zero the pprev pointer so list_unhashed() will return true after
   * this.
   *
   * The caller must take whatever precautions are necessary (such as
   * holding appropriate locks) to avoid racing with another
   * list-mutation primitive, such as hlist_bl_add_head_rcu() or
   * hlist_bl_del_rcu(), running on this same list.  However, it is
   * perfectly legal to run concurrently with the _rcu list-traversal
   * primitives, such as hlist_bl_for_each_entry_rcu().
   */
  static inline void hlist_bl_del_init_rcu(struct hlist_bl_node *n)
  {
  	if (!hlist_bl_unhashed(n)) {
  		__hlist_bl_del(n);
  		n->pprev = NULL;
  	}
  }
  
  /**
   * hlist_bl_del_rcu - deletes entry from hash list without re-initialization
   * @n: the element to delete from the hash list.
   *
   * Note: hlist_bl_unhashed() on entry does not return true after this,
   * the entry is in an undefined state. It is useful for RCU based
   * lockfree traversal.
   *
   * In particular, it means that we can not poison the forward
   * pointers that may still be used for walking the hash list.
   *
   * The caller must take whatever precautions are necessary
   * (such as holding appropriate locks) to avoid racing
   * with another list-mutation primitive, such as hlist_bl_add_head_rcu()
   * or hlist_bl_del_rcu(), running on this same list.
   * However, it is perfectly legal to run concurrently with
   * the _rcu list-traversal primitives, such as
   * hlist_bl_for_each_entry().
   */
  static inline void hlist_bl_del_rcu(struct hlist_bl_node *n)
  {
  	__hlist_bl_del(n);
  	n->pprev = LIST_POISON2;
  }
  
  /**
   * hlist_bl_add_head_rcu
   * @n: the element to add to the hash list.
   * @h: the list to add to.
   *
   * Description:
   * Adds the specified element to the specified hlist_bl,
   * while permitting racing traversals.
   *
   * The caller must take whatever precautions are necessary
   * (such as holding appropriate locks) to avoid racing
   * with another list-mutation primitive, such as hlist_bl_add_head_rcu()
   * or hlist_bl_del_rcu(), running on this same list.
   * However, it is perfectly legal to run concurrently with
   * the _rcu list-traversal primitives, such as
   * hlist_bl_for_each_entry_rcu(), used to prevent memory-consistency
   * problems on Alpha CPUs.  Regardless of the type of CPU, the
   * list-traversal primitive must be guarded by rcu_read_lock().
   */
  static inline void hlist_bl_add_head_rcu(struct hlist_bl_node *n,
  					struct hlist_bl_head *h)
  {
  	struct hlist_bl_node *first;
  
  	/* don't need hlist_bl_first_rcu because we're under lock */
  	first = hlist_bl_first(h);
  
  	n->next = first;
  	if (first)
  		first->pprev = &n->next;
  	n->pprev = &h->first;
  
  	/* need _rcu because we can have concurrent lock free readers */
  	hlist_bl_set_first_rcu(h, n);
  }
  /**
   * hlist_bl_for_each_entry_rcu - iterate over rcu list of given type
   * @tpos:	the type * to use as a loop cursor.
   * @pos:	the &struct hlist_bl_node to use as a loop cursor.
   * @head:	the head for your list.
   * @member:	the name of the hlist_bl_node within the struct.
   *
   */
  #define hlist_bl_for_each_entry_rcu(tpos, pos, head, member)		\
  	for (pos = hlist_bl_first_rcu(head);				\
  		pos &&							\
  		({ tpos = hlist_bl_entry(pos, typeof(*tpos), member); 1; }); \
  		pos = rcu_dereference_raw(pos->next))
  
  #endif