~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/kernel/bpf/bpf_lru_list.c

Version: ~ [ linux-6.4-rc3 ] ~ [ linux-6.3.4 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.30 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.113 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.180 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.243 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.283 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.315 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0-only
  2 /* Copyright (c) 2016 Facebook
  3  */
  4 #include <linux/cpumask.h>
  5 #include <linux/spinlock.h>
  6 #include <linux/percpu.h>
  7 
  8 #include "bpf_lru_list.h"
  9 
 10 #define LOCAL_FREE_TARGET               (128)
 11 #define LOCAL_NR_SCANS                  LOCAL_FREE_TARGET
 12 
 13 #define PERCPU_FREE_TARGET              (4)
 14 #define PERCPU_NR_SCANS                 PERCPU_FREE_TARGET
 15 
 16 /* Helpers to get the local list index */
 17 #define LOCAL_LIST_IDX(t)       ((t) - BPF_LOCAL_LIST_T_OFFSET)
 18 #define LOCAL_FREE_LIST_IDX     LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_FREE)
 19 #define LOCAL_PENDING_LIST_IDX  LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_PENDING)
 20 #define IS_LOCAL_LIST_TYPE(t)   ((t) >= BPF_LOCAL_LIST_T_OFFSET)
 21 
 22 static int get_next_cpu(int cpu)
 23 {
 24         cpu = cpumask_next(cpu, cpu_possible_mask);
 25         if (cpu >= nr_cpu_ids)
 26                 cpu = cpumask_first(cpu_possible_mask);
 27         return cpu;
 28 }
 29 
 30 /* Local list helpers */
 31 static struct list_head *local_free_list(struct bpf_lru_locallist *loc_l)
 32 {
 33         return &loc_l->lists[LOCAL_FREE_LIST_IDX];
 34 }
 35 
 36 static struct list_head *local_pending_list(struct bpf_lru_locallist *loc_l)
 37 {
 38         return &loc_l->lists[LOCAL_PENDING_LIST_IDX];
 39 }
 40 
 41 /* bpf_lru_node helpers */
 42 static bool bpf_lru_node_is_ref(const struct bpf_lru_node *node)
 43 {
 44         return node->ref;
 45 }
 46 
 47 static void bpf_lru_list_count_inc(struct bpf_lru_list *l,
 48                                    enum bpf_lru_list_type type)
 49 {
 50         if (type < NR_BPF_LRU_LIST_COUNT)
 51                 l->counts[type]++;
 52 }
 53 
 54 static void bpf_lru_list_count_dec(struct bpf_lru_list *l,
 55                                    enum bpf_lru_list_type type)
 56 {
 57         if (type < NR_BPF_LRU_LIST_COUNT)
 58                 l->counts[type]--;
 59 }
 60 
 61 static void __bpf_lru_node_move_to_free(struct bpf_lru_list *l,
 62                                         struct bpf_lru_node *node,
 63                                         struct list_head *free_list,
 64                                         enum bpf_lru_list_type tgt_free_type)
 65 {
 66         if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
 67                 return;
 68 
 69         /* If the removing node is the next_inactive_rotation candidate,
 70          * move the next_inactive_rotation pointer also.
 71          */
 72         if (&node->list == l->next_inactive_rotation)
 73                 l->next_inactive_rotation = l->next_inactive_rotation->prev;
 74 
 75         bpf_lru_list_count_dec(l, node->type);
 76 
 77         node->type = tgt_free_type;
 78         list_move(&node->list, free_list);
 79 }
 80 
 81 /* Move nodes from local list to the LRU list */
 82 static void __bpf_lru_node_move_in(struct bpf_lru_list *l,
 83                                    struct bpf_lru_node *node,
 84                                    enum bpf_lru_list_type tgt_type)
 85 {
 86         if (WARN_ON_ONCE(!IS_LOCAL_LIST_TYPE(node->type)) ||
 87             WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type)))
 88                 return;
 89 
 90         bpf_lru_list_count_inc(l, tgt_type);
 91         node->type = tgt_type;
 92         node->ref = 0;
 93         list_move(&node->list, &l->lists[tgt_type]);
 94 }
 95 
 96 /* Move nodes between or within active and inactive list (like
 97  * active to inactive, inactive to active or tail of active back to
 98  * the head of active).
 99  */
100 static void __bpf_lru_node_move(struct bpf_lru_list *l,
101                                 struct bpf_lru_node *node,
102                                 enum bpf_lru_list_type tgt_type)
103 {
104         if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)) ||
105             WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type)))
106                 return;
107 
108         if (node->type != tgt_type) {
109                 bpf_lru_list_count_dec(l, node->type);
110                 bpf_lru_list_count_inc(l, tgt_type);
111                 node->type = tgt_type;
112         }
113         node->ref = 0;
114 
115         /* If the moving node is the next_inactive_rotation candidate,
116          * move the next_inactive_rotation pointer also.
117          */
118         if (&node->list == l->next_inactive_rotation)
119                 l->next_inactive_rotation = l->next_inactive_rotation->prev;
120 
121         list_move(&node->list, &l->lists[tgt_type]);
122 }
123 
124 static bool bpf_lru_list_inactive_low(const struct bpf_lru_list *l)
125 {
126         return l->counts[BPF_LRU_LIST_T_INACTIVE] <
127                 l->counts[BPF_LRU_LIST_T_ACTIVE];
128 }
129 
130 /* Rotate the active list:
131  * 1. Start from tail
132  * 2. If the node has the ref bit set, it will be rotated
133  *    back to the head of active list with the ref bit cleared.
134  *    Give this node one more chance to survive in the active list.
135  * 3. If the ref bit is not set, move it to the head of the
136  *    inactive list.
137  * 4. It will at most scan nr_scans nodes
138  */
139 static void __bpf_lru_list_rotate_active(struct bpf_lru *lru,
140                                          struct bpf_lru_list *l)
141 {
142         struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE];
143         struct bpf_lru_node *node, *tmp_node, *first_node;
144         unsigned int i = 0;
145 
146         first_node = list_first_entry(active, struct bpf_lru_node, list);
147         list_for_each_entry_safe_reverse(node, tmp_node, active, list) {
148                 if (bpf_lru_node_is_ref(node))
149                         __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
150                 else
151                         __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
152 
153                 if (++i == lru->nr_scans || node == first_node)
154                         break;
155         }
156 }
157 
158 /* Rotate the inactive list.  It starts from the next_inactive_rotation
159  * 1. If the node has ref bit set, it will be moved to the head
160  *    of active list with the ref bit cleared.
161  * 2. If the node does not have ref bit set, it will leave it
162  *    at its current location (i.e. do nothing) so that it can
163  *    be considered during the next inactive_shrink.
164  * 3. It will at most scan nr_scans nodes
165  */
166 static void __bpf_lru_list_rotate_inactive(struct bpf_lru *lru,
167                                            struct bpf_lru_list *l)
168 {
169         struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE];
170         struct list_head *cur, *last, *next = inactive;
171         struct bpf_lru_node *node;
172         unsigned int i = 0;
173 
174         if (list_empty(inactive))
175                 return;
176 
177         last = l->next_inactive_rotation->next;
178         if (last == inactive)
179                 last = last->next;
180 
181         cur = l->next_inactive_rotation;
182         while (i < lru->nr_scans) {
183                 if (cur == inactive) {
184                         cur = cur->prev;
185                         continue;
186                 }
187 
188                 node = list_entry(cur, struct bpf_lru_node, list);
189                 next = cur->prev;
190                 if (bpf_lru_node_is_ref(node))
191                         __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
192                 if (cur == last)
193                         break;
194                 cur = next;
195                 i++;
196         }
197 
198         l->next_inactive_rotation = next;
199 }
200 
201 /* Shrink the inactive list.  It starts from the tail of the
202  * inactive list and only move the nodes without the ref bit
203  * set to the designated free list.
204  */
205 static unsigned int
206 __bpf_lru_list_shrink_inactive(struct bpf_lru *lru,
207                                struct bpf_lru_list *l,
208                                unsigned int tgt_nshrink,
209                                struct list_head *free_list,
210                                enum bpf_lru_list_type tgt_free_type)
211 {
212         struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE];
213         struct bpf_lru_node *node, *tmp_node;
214         unsigned int nshrinked = 0;
215         unsigned int i = 0;
216 
217         list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) {
218                 if (bpf_lru_node_is_ref(node)) {
219                         __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
220                 } else if (lru->del_from_htab(lru->del_arg, node)) {
221                         __bpf_lru_node_move_to_free(l, node, free_list,
222                                                     tgt_free_type);
223                         if (++nshrinked == tgt_nshrink)
224                                 break;
225                 }
226 
227                 if (++i == lru->nr_scans)
228                         break;
229         }
230 
231         return nshrinked;
232 }
233 
234 /* 1. Rotate the active list (if needed)
235  * 2. Always rotate the inactive list
236  */
237 static void __bpf_lru_list_rotate(struct bpf_lru *lru, struct bpf_lru_list *l)
238 {
239         if (bpf_lru_list_inactive_low(l))
240                 __bpf_lru_list_rotate_active(lru, l);
241 
242         __bpf_lru_list_rotate_inactive(lru, l);
243 }
244 
245 /* Calls __bpf_lru_list_shrink_inactive() to shrink some
246  * ref-bit-cleared nodes and move them to the designated
247  * free list.
248  *
249  * If it cannot get a free node after calling
250  * __bpf_lru_list_shrink_inactive().  It will just remove
251  * one node from either inactive or active list without
252  * honoring the ref-bit.  It prefers inactive list to active
253  * list in this situation.
254  */
255 static unsigned int __bpf_lru_list_shrink(struct bpf_lru *lru,
256                                           struct bpf_lru_list *l,
257                                           unsigned int tgt_nshrink,
258                                           struct list_head *free_list,
259                                           enum bpf_lru_list_type tgt_free_type)
260 
261 {
262         struct bpf_lru_node *node, *tmp_node;
263         struct list_head *force_shrink_list;
264         unsigned int nshrinked;
265 
266         nshrinked = __bpf_lru_list_shrink_inactive(lru, l, tgt_nshrink,
267                                                    free_list, tgt_free_type);
268         if (nshrinked)
269                 return nshrinked;
270 
271         /* Do a force shrink by ignoring the reference bit */
272         if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE]))
273                 force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE];
274         else
275                 force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIVE];
276 
277         list_for_each_entry_safe_reverse(node, tmp_node, force_shrink_list,
278                                          list) {
279                 if (lru->del_from_htab(lru->del_arg, node)) {
280                         __bpf_lru_node_move_to_free(l, node, free_list,
281                                                     tgt_free_type);
282                         return 1;
283                 }
284         }
285 
286         return 0;
287 }
288 
289 /* Flush the nodes from the local pending list to the LRU list */
290 static void __local_list_flush(struct bpf_lru_list *l,
291                                struct bpf_lru_locallist *loc_l)
292 {
293         struct bpf_lru_node *node, *tmp_node;
294 
295         list_for_each_entry_safe_reverse(node, tmp_node,
296                                          local_pending_list(loc_l), list) {
297                 if (bpf_lru_node_is_ref(node))
298                         __bpf_lru_node_move_in(l, node, BPF_LRU_LIST_T_ACTIVE);
299                 else
300                         __bpf_lru_node_move_in(l, node,
301                                                BPF_LRU_LIST_T_INACTIVE);
302         }
303 }
304 
305 static void bpf_lru_list_push_free(struct bpf_lru_list *l,
306                                    struct bpf_lru_node *node)
307 {
308         unsigned long flags;
309 
310         if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
311                 return;
312 
313         raw_spin_lock_irqsave(&l->lock, flags);
314         __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
315         raw_spin_unlock_irqrestore(&l->lock, flags);
316 }
317 
318 static void bpf_lru_list_pop_free_to_local(struct bpf_lru *lru,
319                                            struct bpf_lru_locallist *loc_l)
320 {
321         struct bpf_lru_list *l = &lru->common_lru.lru_list;
322         struct bpf_lru_node *node, *tmp_node;
323         unsigned int nfree = 0;
324 
325         raw_spin_lock(&l->lock);
326 
327         __local_list_flush(l, loc_l);
328 
329         __bpf_lru_list_rotate(lru, l);
330 
331         list_for_each_entry_safe(node, tmp_node, &l->lists[BPF_LRU_LIST_T_FREE],
332                                  list) {
333                 __bpf_lru_node_move_to_free(l, node, local_free_list(loc_l),
334                                             BPF_LRU_LOCAL_LIST_T_FREE);
335                 if (++nfree == LOCAL_FREE_TARGET)
336                         break;
337         }
338 
339         if (nfree < LOCAL_FREE_TARGET)
340                 __bpf_lru_list_shrink(lru, l, LOCAL_FREE_TARGET - nfree,
341                                       local_free_list(loc_l),
342                                       BPF_LRU_LOCAL_LIST_T_FREE);
343 
344         raw_spin_unlock(&l->lock);
345 }
346 
347 static void __local_list_add_pending(struct bpf_lru *lru,
348                                      struct bpf_lru_locallist *loc_l,
349                                      int cpu,
350                                      struct bpf_lru_node *node,
351                                      u32 hash)
352 {
353         *(u32 *)((void *)node + lru->hash_offset) = hash;
354         node->cpu = cpu;
355         node->type = BPF_LRU_LOCAL_LIST_T_PENDING;
356         node->ref = 0;
357         list_add(&node->list, local_pending_list(loc_l));
358 }
359 
360 static struct bpf_lru_node *
361 __local_list_pop_free(struct bpf_lru_locallist *loc_l)
362 {
363         struct bpf_lru_node *node;
364 
365         node = list_first_entry_or_null(local_free_list(loc_l),
366                                         struct bpf_lru_node,
367                                         list);
368         if (node)
369                 list_del(&node->list);
370 
371         return node;
372 }
373 
374 static struct bpf_lru_node *
375 __local_list_pop_pending(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l)
376 {
377         struct bpf_lru_node *node;
378         bool force = false;
379 
380 ignore_ref:
381         /* Get from the tail (i.e. older element) of the pending list. */
382         list_for_each_entry_reverse(node, local_pending_list(loc_l),
383                                     list) {
384                 if ((!bpf_lru_node_is_ref(node) || force) &&
385                     lru->del_from_htab(lru->del_arg, node)) {
386                         list_del(&node->list);
387                         return node;
388                 }
389         }
390 
391         if (!force) {
392                 force = true;
393                 goto ignore_ref;
394         }
395 
396         return NULL;
397 }
398 
399 static struct bpf_lru_node *bpf_percpu_lru_pop_free(struct bpf_lru *lru,
400                                                     u32 hash)
401 {
402         struct list_head *free_list;
403         struct bpf_lru_node *node = NULL;
404         struct bpf_lru_list *l;
405         unsigned long flags;
406         int cpu = raw_smp_processor_id();
407 
408         l = per_cpu_ptr(lru->percpu_lru, cpu);
409 
410         raw_spin_lock_irqsave(&l->lock, flags);
411 
412         __bpf_lru_list_rotate(lru, l);
413 
414         free_list = &l->lists[BPF_LRU_LIST_T_FREE];
415         if (list_empty(free_list))
416                 __bpf_lru_list_shrink(lru, l, PERCPU_FREE_TARGET, free_list,
417                                       BPF_LRU_LIST_T_FREE);
418 
419         if (!list_empty(free_list)) {
420                 node = list_first_entry(free_list, struct bpf_lru_node, list);
421                 *(u32 *)((void *)node + lru->hash_offset) = hash;
422                 node->ref = 0;
423                 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
424         }
425 
426         raw_spin_unlock_irqrestore(&l->lock, flags);
427 
428         return node;
429 }
430 
431 static struct bpf_lru_node *bpf_common_lru_pop_free(struct bpf_lru *lru,
432                                                     u32 hash)
433 {
434         struct bpf_lru_locallist *loc_l, *steal_loc_l;
435         struct bpf_common_lru *clru = &lru->common_lru;
436         struct bpf_lru_node *node;
437         int steal, first_steal;
438         unsigned long flags;
439         int cpu = raw_smp_processor_id();
440 
441         loc_l = per_cpu_ptr(clru->local_list, cpu);
442 
443         raw_spin_lock_irqsave(&loc_l->lock, flags);
444 
445         node = __local_list_pop_free(loc_l);
446         if (!node) {
447                 bpf_lru_list_pop_free_to_local(lru, loc_l);
448                 node = __local_list_pop_free(loc_l);
449         }
450 
451         if (node)
452                 __local_list_add_pending(lru, loc_l, cpu, node, hash);
453 
454         raw_spin_unlock_irqrestore(&loc_l->lock, flags);
455 
456         if (node)
457                 return node;
458 
459         /* No free nodes found from the local free list and
460          * the global LRU list.
461          *
462          * Steal from the local free/pending list of the
463          * current CPU and remote CPU in RR.  It starts
464          * with the loc_l->next_steal CPU.
465          */
466 
467         first_steal = loc_l->next_steal;
468         steal = first_steal;
469         do {
470                 steal_loc_l = per_cpu_ptr(clru->local_list, steal);
471 
472                 raw_spin_lock_irqsave(&steal_loc_l->lock, flags);
473 
474                 node = __local_list_pop_free(steal_loc_l);
475                 if (!node)
476                         node = __local_list_pop_pending(lru, steal_loc_l);
477 
478                 raw_spin_unlock_irqrestore(&steal_loc_l->lock, flags);
479 
480                 steal = get_next_cpu(steal);
481         } while (!node && steal != first_steal);
482 
483         loc_l->next_steal = steal;
484 
485         if (node) {
486                 raw_spin_lock_irqsave(&loc_l->lock, flags);
487                 __local_list_add_pending(lru, loc_l, cpu, node, hash);
488                 raw_spin_unlock_irqrestore(&loc_l->lock, flags);
489         }
490 
491         return node;
492 }
493 
494 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash)
495 {
496         if (lru->percpu)
497                 return bpf_percpu_lru_pop_free(lru, hash);
498         else
499                 return bpf_common_lru_pop_free(lru, hash);
500 }
501 
502 static void bpf_common_lru_push_free(struct bpf_lru *lru,
503                                      struct bpf_lru_node *node)
504 {
505         u8 node_type = READ_ONCE(node->type);
506         unsigned long flags;
507 
508         if (WARN_ON_ONCE(node_type == BPF_LRU_LIST_T_FREE) ||
509             WARN_ON_ONCE(node_type == BPF_LRU_LOCAL_LIST_T_FREE))
510                 return;
511 
512         if (node_type == BPF_LRU_LOCAL_LIST_T_PENDING) {
513                 struct bpf_lru_locallist *loc_l;
514 
515                 loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu);
516 
517                 raw_spin_lock_irqsave(&loc_l->lock, flags);
518 
519                 if (unlikely(node->type != BPF_LRU_LOCAL_LIST_T_PENDING)) {
520                         raw_spin_unlock_irqrestore(&loc_l->lock, flags);
521                         goto check_lru_list;
522                 }
523 
524                 node->type = BPF_LRU_LOCAL_LIST_T_FREE;
525                 node->ref = 0;
526                 list_move(&node->list, local_free_list(loc_l));
527 
528                 raw_spin_unlock_irqrestore(&loc_l->lock, flags);
529                 return;
530         }
531 
532 check_lru_list:
533         bpf_lru_list_push_free(&lru->common_lru.lru_list, node);
534 }
535 
536 static void bpf_percpu_lru_push_free(struct bpf_lru *lru,
537                                      struct bpf_lru_node *node)
538 {
539         struct bpf_lru_list *l;
540         unsigned long flags;
541 
542         l = per_cpu_ptr(lru->percpu_lru, node->cpu);
543 
544         raw_spin_lock_irqsave(&l->lock, flags);
545 
546         __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
547 
548         raw_spin_unlock_irqrestore(&l->lock, flags);
549 }
550 
551 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node)
552 {
553         if (lru->percpu)
554                 bpf_percpu_lru_push_free(lru, node);
555         else
556                 bpf_common_lru_push_free(lru, node);
557 }
558 
559 static void bpf_common_lru_populate(struct bpf_lru *lru, void *buf,
560                                     u32 node_offset, u32 elem_size,
561                                     u32 nr_elems)
562 {
563         struct bpf_lru_list *l = &lru->common_lru.lru_list;
564         u32 i;
565 
566         for (i = 0; i < nr_elems; i++) {
567                 struct bpf_lru_node *node;
568 
569                 node = (struct bpf_lru_node *)(buf + node_offset);
570                 node->type = BPF_LRU_LIST_T_FREE;
571                 node->ref = 0;
572                 list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
573                 buf += elem_size;
574         }
575 }
576 
577 static void bpf_percpu_lru_populate(struct bpf_lru *lru, void *buf,
578                                     u32 node_offset, u32 elem_size,
579                                     u32 nr_elems)
580 {
581         u32 i, pcpu_entries;
582         int cpu;
583         struct bpf_lru_list *l;
584 
585         pcpu_entries = nr_elems / num_possible_cpus();
586 
587         i = 0;
588 
589         for_each_possible_cpu(cpu) {
590                 struct bpf_lru_node *node;
591 
592                 l = per_cpu_ptr(lru->percpu_lru, cpu);
593 again:
594                 node = (struct bpf_lru_node *)(buf + node_offset);
595                 node->cpu = cpu;
596                 node->type = BPF_LRU_LIST_T_FREE;
597                 node->ref = 0;
598                 list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
599                 i++;
600                 buf += elem_size;
601                 if (i == nr_elems)
602                         break;
603                 if (i % pcpu_entries)
604                         goto again;
605         }
606 }
607 
608 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
609                       u32 elem_size, u32 nr_elems)
610 {
611         if (lru->percpu)
612                 bpf_percpu_lru_populate(lru, buf, node_offset, elem_size,
613                                         nr_elems);
614         else
615                 bpf_common_lru_populate(lru, buf, node_offset, elem_size,
616                                         nr_elems);
617 }
618 
619 static void bpf_lru_locallist_init(struct bpf_lru_locallist *loc_l, int cpu)
620 {
621         int i;
622 
623         for (i = 0; i < NR_BPF_LRU_LOCAL_LIST_T; i++)
624                 INIT_LIST_HEAD(&loc_l->lists[i]);
625 
626         loc_l->next_steal = cpu;
627 
628         raw_spin_lock_init(&loc_l->lock);
629 }
630 
631 static void bpf_lru_list_init(struct bpf_lru_list *l)
632 {
633         int i;
634 
635         for (i = 0; i < NR_BPF_LRU_LIST_T; i++)
636                 INIT_LIST_HEAD(&l->lists[i]);
637 
638         for (i = 0; i < NR_BPF_LRU_LIST_COUNT; i++)
639                 l->counts[i] = 0;
640 
641         l->next_inactive_rotation = &l->lists[BPF_LRU_LIST_T_INACTIVE];
642 
643         raw_spin_lock_init(&l->lock);
644 }
645 
646 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
647                  del_from_htab_func del_from_htab, void *del_arg)
648 {
649         int cpu;
650 
651         if (percpu) {
652                 lru->percpu_lru = alloc_percpu(struct bpf_lru_list);
653                 if (!lru->percpu_lru)
654                         return -ENOMEM;
655 
656                 for_each_possible_cpu(cpu) {
657                         struct bpf_lru_list *l;
658 
659                         l = per_cpu_ptr(lru->percpu_lru, cpu);
660                         bpf_lru_list_init(l);
661                 }
662                 lru->nr_scans = PERCPU_NR_SCANS;
663         } else {
664                 struct bpf_common_lru *clru = &lru->common_lru;
665 
666                 clru->local_list = alloc_percpu(struct bpf_lru_locallist);
667                 if (!clru->local_list)
668                         return -ENOMEM;
669 
670                 for_each_possible_cpu(cpu) {
671                         struct bpf_lru_locallist *loc_l;
672 
673                         loc_l = per_cpu_ptr(clru->local_list, cpu);
674                         bpf_lru_locallist_init(loc_l, cpu);
675                 }
676 
677                 bpf_lru_list_init(&clru->lru_list);
678                 lru->nr_scans = LOCAL_NR_SCANS;
679         }
680 
681         lru->percpu = percpu;
682         lru->del_from_htab = del_from_htab;
683         lru->del_arg = del_arg;
684         lru->hash_offset = hash_offset;
685 
686         return 0;
687 }
688 
689 void bpf_lru_destroy(struct bpf_lru *lru)
690 {
691         if (lru->percpu)
692                 free_percpu(lru->percpu_lru);
693         else
694                 free_percpu(lru->common_lru.local_list);
695 }
696 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | Wiki (Japanese) | Wiki (English) | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

osdn.jp