Linux-libre 5.3.12-gnu
[librecmc/linux-libre.git] / net / netfilter / ipvs / ip_vs_mh.c
1 // SPDX-License-Identifier: GPL-2.0
2 /* IPVS:        Maglev Hashing scheduling module
3  *
4  * Authors:     Inju Song <inju.song@navercorp.com>
5  *
6  */
7
8 /* The mh algorithm is to assign a preference list of all the lookup
9  * table positions to each destination and populate the table with
10  * the most-preferred position of destinations. Then it is to select
11  * destination with the hash key of source IP address through looking
12  * up a the lookup table.
13  *
14  * The algorithm is detailed in:
15  * [3.4 Consistent Hasing]
16 https://www.usenix.org/system/files/conference/nsdi16/nsdi16-paper-eisenbud.pdf
17  *
18  */
19
20 #define KMSG_COMPONENT "IPVS"
21 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
22
23 #include <linux/ip.h>
24 #include <linux/slab.h>
25 #include <linux/module.h>
26 #include <linux/kernel.h>
27 #include <linux/skbuff.h>
28
29 #include <net/ip_vs.h>
30
31 #include <linux/siphash.h>
32 #include <linux/bitops.h>
33 #include <linux/gcd.h>
34
35 #define IP_VS_SVC_F_SCHED_MH_FALLBACK   IP_VS_SVC_F_SCHED1 /* MH fallback */
36 #define IP_VS_SVC_F_SCHED_MH_PORT       IP_VS_SVC_F_SCHED2 /* MH use port */
37
38 struct ip_vs_mh_lookup {
39         struct ip_vs_dest __rcu *dest;  /* real server (cache) */
40 };
41
42 struct ip_vs_mh_dest_setup {
43         unsigned int    offset; /* starting offset */
44         unsigned int    skip;   /* skip */
45         unsigned int    perm;   /* next_offset */
46         int             turns;  /* weight / gcd() and rshift */
47 };
48
49 /* Available prime numbers for MH table */
50 static int primes[] = {251, 509, 1021, 2039, 4093,
51                        8191, 16381, 32749, 65521, 131071};
52
53 /* For IPVS MH entry hash table */
54 #ifndef CONFIG_IP_VS_MH_TAB_INDEX
55 #define CONFIG_IP_VS_MH_TAB_INDEX       12
56 #endif
57 #define IP_VS_MH_TAB_BITS               (CONFIG_IP_VS_MH_TAB_INDEX / 2)
58 #define IP_VS_MH_TAB_INDEX              (CONFIG_IP_VS_MH_TAB_INDEX - 8)
59 #define IP_VS_MH_TAB_SIZE               primes[IP_VS_MH_TAB_INDEX]
60
61 struct ip_vs_mh_state {
62         struct rcu_head                 rcu_head;
63         struct ip_vs_mh_lookup          *lookup;
64         struct ip_vs_mh_dest_setup      *dest_setup;
65         hsiphash_key_t                  hash1, hash2;
66         int                             gcd;
67         int                             rshift;
68 };
69
70 static inline void generate_hash_secret(hsiphash_key_t *hash1,
71                                         hsiphash_key_t *hash2)
72 {
73         hash1->key[0] = 2654435761UL;
74         hash1->key[1] = 2654435761UL;
75
76         hash2->key[0] = 2654446892UL;
77         hash2->key[1] = 2654446892UL;
78 }
79
80 /* Helper function to determine if server is unavailable */
81 static inline bool is_unavailable(struct ip_vs_dest *dest)
82 {
83         return atomic_read(&dest->weight) <= 0 ||
84                dest->flags & IP_VS_DEST_F_OVERLOAD;
85 }
86
87 /* Returns hash value for IPVS MH entry */
88 static inline unsigned int
89 ip_vs_mh_hashkey(int af, const union nf_inet_addr *addr,
90                  __be16 port, hsiphash_key_t *key, unsigned int offset)
91 {
92         unsigned int v;
93         __be32 addr_fold = addr->ip;
94
95 #ifdef CONFIG_IP_VS_IPV6
96         if (af == AF_INET6)
97                 addr_fold = addr->ip6[0] ^ addr->ip6[1] ^
98                             addr->ip6[2] ^ addr->ip6[3];
99 #endif
100         v = (offset + ntohs(port) + ntohl(addr_fold));
101         return hsiphash(&v, sizeof(v), key);
102 }
103
104 /* Reset all the hash buckets of the specified table. */
105 static void ip_vs_mh_reset(struct ip_vs_mh_state *s)
106 {
107         int i;
108         struct ip_vs_mh_lookup *l;
109         struct ip_vs_dest *dest;
110
111         l = &s->lookup[0];
112         for (i = 0; i < IP_VS_MH_TAB_SIZE; i++) {
113                 dest = rcu_dereference_protected(l->dest, 1);
114                 if (dest) {
115                         ip_vs_dest_put(dest);
116                         RCU_INIT_POINTER(l->dest, NULL);
117                 }
118                 l++;
119         }
120 }
121
122 static int ip_vs_mh_permutate(struct ip_vs_mh_state *s,
123                               struct ip_vs_service *svc)
124 {
125         struct list_head *p;
126         struct ip_vs_mh_dest_setup *ds;
127         struct ip_vs_dest *dest;
128         int lw;
129
130         /* If gcd is smaller then 1, number of dests or
131          * all last_weight of dests are zero. So, skip
132          * permutation for the dests.
133          */
134         if (s->gcd < 1)
135                 return 0;
136
137         /* Set dest_setup for the dests permutation */
138         p = &svc->destinations;
139         ds = &s->dest_setup[0];
140         while ((p = p->next) != &svc->destinations) {
141                 dest = list_entry(p, struct ip_vs_dest, n_list);
142
143                 ds->offset = ip_vs_mh_hashkey(svc->af, &dest->addr,
144                                               dest->port, &s->hash1, 0) %
145                                               IP_VS_MH_TAB_SIZE;
146                 ds->skip = ip_vs_mh_hashkey(svc->af, &dest->addr,
147                                             dest->port, &s->hash2, 0) %
148                                             (IP_VS_MH_TAB_SIZE - 1) + 1;
149                 ds->perm = ds->offset;
150
151                 lw = atomic_read(&dest->last_weight);
152                 ds->turns = ((lw / s->gcd) >> s->rshift) ? : (lw != 0);
153                 ds++;
154         }
155
156         return 0;
157 }
158
159 static int ip_vs_mh_populate(struct ip_vs_mh_state *s,
160                              struct ip_vs_service *svc)
161 {
162         int n, c, dt_count;
163         unsigned long *table;
164         struct list_head *p;
165         struct ip_vs_mh_dest_setup *ds;
166         struct ip_vs_dest *dest, *new_dest;
167
168         /* If gcd is smaller then 1, number of dests or
169          * all last_weight of dests are zero. So, skip
170          * the population for the dests and reset lookup table.
171          */
172         if (s->gcd < 1) {
173                 ip_vs_mh_reset(s);
174                 return 0;
175         }
176
177         table =  kcalloc(BITS_TO_LONGS(IP_VS_MH_TAB_SIZE),
178                          sizeof(unsigned long), GFP_KERNEL);
179         if (!table)
180                 return -ENOMEM;
181
182         p = &svc->destinations;
183         n = 0;
184         dt_count = 0;
185         while (n < IP_VS_MH_TAB_SIZE) {
186                 if (p == &svc->destinations)
187                         p = p->next;
188
189                 ds = &s->dest_setup[0];
190                 while (p != &svc->destinations) {
191                         /* Ignore added server with zero weight */
192                         if (ds->turns < 1) {
193                                 p = p->next;
194                                 ds++;
195                                 continue;
196                         }
197
198                         c = ds->perm;
199                         while (test_bit(c, table)) {
200                                 /* Add skip, mod IP_VS_MH_TAB_SIZE */
201                                 ds->perm += ds->skip;
202                                 if (ds->perm >= IP_VS_MH_TAB_SIZE)
203                                         ds->perm -= IP_VS_MH_TAB_SIZE;
204                                 c = ds->perm;
205                         }
206
207                         __set_bit(c, table);
208
209                         dest = rcu_dereference_protected(s->lookup[c].dest, 1);
210                         new_dest = list_entry(p, struct ip_vs_dest, n_list);
211                         if (dest != new_dest) {
212                                 if (dest)
213                                         ip_vs_dest_put(dest);
214                                 ip_vs_dest_hold(new_dest);
215                                 RCU_INIT_POINTER(s->lookup[c].dest, new_dest);
216                         }
217
218                         if (++n == IP_VS_MH_TAB_SIZE)
219                                 goto out;
220
221                         if (++dt_count >= ds->turns) {
222                                 dt_count = 0;
223                                 p = p->next;
224                                 ds++;
225                         }
226                 }
227         }
228
229 out:
230         kfree(table);
231         return 0;
232 }
233
234 /* Get ip_vs_dest associated with supplied parameters. */
235 static inline struct ip_vs_dest *
236 ip_vs_mh_get(struct ip_vs_service *svc, struct ip_vs_mh_state *s,
237              const union nf_inet_addr *addr, __be16 port)
238 {
239         unsigned int hash = ip_vs_mh_hashkey(svc->af, addr, port, &s->hash1, 0)
240                                              % IP_VS_MH_TAB_SIZE;
241         struct ip_vs_dest *dest = rcu_dereference(s->lookup[hash].dest);
242
243         return (!dest || is_unavailable(dest)) ? NULL : dest;
244 }
245
246 /* As ip_vs_mh_get, but with fallback if selected server is unavailable */
247 static inline struct ip_vs_dest *
248 ip_vs_mh_get_fallback(struct ip_vs_service *svc, struct ip_vs_mh_state *s,
249                       const union nf_inet_addr *addr, __be16 port)
250 {
251         unsigned int offset, roffset;
252         unsigned int hash, ihash;
253         struct ip_vs_dest *dest;
254
255         /* First try the dest it's supposed to go to */
256         ihash = ip_vs_mh_hashkey(svc->af, addr, port,
257                                  &s->hash1, 0) % IP_VS_MH_TAB_SIZE;
258         dest = rcu_dereference(s->lookup[ihash].dest);
259         if (!dest)
260                 return NULL;
261         if (!is_unavailable(dest))
262                 return dest;
263
264         IP_VS_DBG_BUF(6, "MH: selected unavailable server %s:%u, reselecting",
265                       IP_VS_DBG_ADDR(dest->af, &dest->addr), ntohs(dest->port));
266
267         /* If the original dest is unavailable, loop around the table
268          * starting from ihash to find a new dest
269          */
270         for (offset = 0; offset < IP_VS_MH_TAB_SIZE; offset++) {
271                 roffset = (offset + ihash) % IP_VS_MH_TAB_SIZE;
272                 hash = ip_vs_mh_hashkey(svc->af, addr, port, &s->hash1,
273                                         roffset) % IP_VS_MH_TAB_SIZE;
274                 dest = rcu_dereference(s->lookup[hash].dest);
275                 if (!dest)
276                         break;
277                 if (!is_unavailable(dest))
278                         return dest;
279                 IP_VS_DBG_BUF(6,
280                               "MH: selected unavailable server %s:%u (offset %u), reselecting",
281                               IP_VS_DBG_ADDR(dest->af, &dest->addr),
282                               ntohs(dest->port), roffset);
283         }
284
285         return NULL;
286 }
287
288 /* Assign all the hash buckets of the specified table with the service. */
289 static int ip_vs_mh_reassign(struct ip_vs_mh_state *s,
290                              struct ip_vs_service *svc)
291 {
292         int ret;
293
294         if (svc->num_dests > IP_VS_MH_TAB_SIZE)
295                 return -EINVAL;
296
297         if (svc->num_dests >= 1) {
298                 s->dest_setup = kcalloc(svc->num_dests,
299                                         sizeof(struct ip_vs_mh_dest_setup),
300                                         GFP_KERNEL);
301                 if (!s->dest_setup)
302                         return -ENOMEM;
303         }
304
305         ip_vs_mh_permutate(s, svc);
306
307         ret = ip_vs_mh_populate(s, svc);
308         if (ret < 0)
309                 goto out;
310
311         IP_VS_DBG_BUF(6, "MH: reassign lookup table of %s:%u\n",
312                       IP_VS_DBG_ADDR(svc->af, &svc->addr),
313                       ntohs(svc->port));
314
315 out:
316         if (svc->num_dests >= 1) {
317                 kfree(s->dest_setup);
318                 s->dest_setup = NULL;
319         }
320         return ret;
321 }
322
323 static int ip_vs_mh_gcd_weight(struct ip_vs_service *svc)
324 {
325         struct ip_vs_dest *dest;
326         int weight;
327         int g = 0;
328
329         list_for_each_entry(dest, &svc->destinations, n_list) {
330                 weight = atomic_read(&dest->last_weight);
331                 if (weight > 0) {
332                         if (g > 0)
333                                 g = gcd(weight, g);
334                         else
335                                 g = weight;
336                 }
337         }
338         return g;
339 }
340
341 /* To avoid assigning huge weight for the MH table,
342  * calculate shift value with gcd.
343  */
344 static int ip_vs_mh_shift_weight(struct ip_vs_service *svc, int gcd)
345 {
346         struct ip_vs_dest *dest;
347         int new_weight, weight = 0;
348         int mw, shift;
349
350         /* If gcd is smaller then 1, number of dests or
351          * all last_weight of dests are zero. So, return
352          * shift value as zero.
353          */
354         if (gcd < 1)
355                 return 0;
356
357         list_for_each_entry(dest, &svc->destinations, n_list) {
358                 new_weight = atomic_read(&dest->last_weight);
359                 if (new_weight > weight)
360                         weight = new_weight;
361         }
362
363         /* Because gcd is greater than zero,
364          * the maximum weight and gcd are always greater than zero
365          */
366         mw = weight / gcd;
367
368         /* shift = occupied bits of weight/gcd - MH highest bits */
369         shift = fls(mw) - IP_VS_MH_TAB_BITS;
370         return (shift >= 0) ? shift : 0;
371 }
372
373 static void ip_vs_mh_state_free(struct rcu_head *head)
374 {
375         struct ip_vs_mh_state *s;
376
377         s = container_of(head, struct ip_vs_mh_state, rcu_head);
378         kfree(s->lookup);
379         kfree(s);
380 }
381
382 static int ip_vs_mh_init_svc(struct ip_vs_service *svc)
383 {
384         int ret;
385         struct ip_vs_mh_state *s;
386
387         /* Allocate the MH table for this service */
388         s = kzalloc(sizeof(*s), GFP_KERNEL);
389         if (!s)
390                 return -ENOMEM;
391
392         s->lookup = kcalloc(IP_VS_MH_TAB_SIZE, sizeof(struct ip_vs_mh_lookup),
393                             GFP_KERNEL);
394         if (!s->lookup) {
395                 kfree(s);
396                 return -ENOMEM;
397         }
398
399         generate_hash_secret(&s->hash1, &s->hash2);
400         s->gcd = ip_vs_mh_gcd_weight(svc);
401         s->rshift = ip_vs_mh_shift_weight(svc, s->gcd);
402
403         IP_VS_DBG(6,
404                   "MH lookup table (memory=%zdbytes) allocated for current service\n",
405                   sizeof(struct ip_vs_mh_lookup) * IP_VS_MH_TAB_SIZE);
406
407         /* Assign the lookup table with current dests */
408         ret = ip_vs_mh_reassign(s, svc);
409         if (ret < 0) {
410                 ip_vs_mh_reset(s);
411                 ip_vs_mh_state_free(&s->rcu_head);
412                 return ret;
413         }
414
415         /* No more failures, attach state */
416         svc->sched_data = s;
417         return 0;
418 }
419
420 static void ip_vs_mh_done_svc(struct ip_vs_service *svc)
421 {
422         struct ip_vs_mh_state *s = svc->sched_data;
423
424         /* Got to clean up lookup entry here */
425         ip_vs_mh_reset(s);
426
427         call_rcu(&s->rcu_head, ip_vs_mh_state_free);
428         IP_VS_DBG(6, "MH lookup table (memory=%zdbytes) released\n",
429                   sizeof(struct ip_vs_mh_lookup) * IP_VS_MH_TAB_SIZE);
430 }
431
432 static int ip_vs_mh_dest_changed(struct ip_vs_service *svc,
433                                  struct ip_vs_dest *dest)
434 {
435         struct ip_vs_mh_state *s = svc->sched_data;
436
437         s->gcd = ip_vs_mh_gcd_weight(svc);
438         s->rshift = ip_vs_mh_shift_weight(svc, s->gcd);
439
440         /* Assign the lookup table with the updated service */
441         return ip_vs_mh_reassign(s, svc);
442 }
443
444 /* Helper function to get port number */
445 static inline __be16
446 ip_vs_mh_get_port(const struct sk_buff *skb, struct ip_vs_iphdr *iph)
447 {
448         __be16 _ports[2], *ports;
449
450         /* At this point we know that we have a valid packet of some kind.
451          * Because ICMP packets are only guaranteed to have the first 8
452          * bytes, let's just grab the ports.  Fortunately they're in the
453          * same position for all three of the protocols we care about.
454          */
455         switch (iph->protocol) {
456         case IPPROTO_TCP:
457         case IPPROTO_UDP:
458         case IPPROTO_SCTP:
459                 ports = skb_header_pointer(skb, iph->len, sizeof(_ports),
460                                            &_ports);
461                 if (unlikely(!ports))
462                         return 0;
463
464                 if (likely(!ip_vs_iph_inverse(iph)))
465                         return ports[0];
466                 else
467                         return ports[1];
468         default:
469                 return 0;
470         }
471 }
472
473 /* Maglev Hashing scheduling */
474 static struct ip_vs_dest *
475 ip_vs_mh_schedule(struct ip_vs_service *svc, const struct sk_buff *skb,
476                   struct ip_vs_iphdr *iph)
477 {
478         struct ip_vs_dest *dest;
479         struct ip_vs_mh_state *s;
480         __be16 port = 0;
481         const union nf_inet_addr *hash_addr;
482
483         hash_addr = ip_vs_iph_inverse(iph) ? &iph->daddr : &iph->saddr;
484
485         IP_VS_DBG(6, "%s : Scheduling...\n", __func__);
486
487         if (svc->flags & IP_VS_SVC_F_SCHED_MH_PORT)
488                 port = ip_vs_mh_get_port(skb, iph);
489
490         s = (struct ip_vs_mh_state *)svc->sched_data;
491
492         if (svc->flags & IP_VS_SVC_F_SCHED_MH_FALLBACK)
493                 dest = ip_vs_mh_get_fallback(svc, s, hash_addr, port);
494         else
495                 dest = ip_vs_mh_get(svc, s, hash_addr, port);
496
497         if (!dest) {
498                 ip_vs_scheduler_err(svc, "no destination available");
499                 return NULL;
500         }
501
502         IP_VS_DBG_BUF(6, "MH: source IP address %s:%u --> server %s:%u\n",
503                       IP_VS_DBG_ADDR(svc->af, hash_addr),
504                       ntohs(port),
505                       IP_VS_DBG_ADDR(dest->af, &dest->addr),
506                       ntohs(dest->port));
507
508         return dest;
509 }
510
511 /* IPVS MH Scheduler structure */
512 static struct ip_vs_scheduler ip_vs_mh_scheduler = {
513         .name =                 "mh",
514         .refcnt =               ATOMIC_INIT(0),
515         .module =               THIS_MODULE,
516         .n_list  =              LIST_HEAD_INIT(ip_vs_mh_scheduler.n_list),
517         .init_service =         ip_vs_mh_init_svc,
518         .done_service =         ip_vs_mh_done_svc,
519         .add_dest =             ip_vs_mh_dest_changed,
520         .del_dest =             ip_vs_mh_dest_changed,
521         .upd_dest =             ip_vs_mh_dest_changed,
522         .schedule =             ip_vs_mh_schedule,
523 };
524
525 static int __init ip_vs_mh_init(void)
526 {
527         return register_ip_vs_scheduler(&ip_vs_mh_scheduler);
528 }
529
530 static void __exit ip_vs_mh_cleanup(void)
531 {
532         unregister_ip_vs_scheduler(&ip_vs_mh_scheduler);
533         rcu_barrier();
534 }
535
536 module_init(ip_vs_mh_init);
537 module_exit(ip_vs_mh_cleanup);
538 MODULE_DESCRIPTION("Maglev hashing ipvs scheduler");
539 MODULE_LICENSE("GPL v2");
540 MODULE_AUTHOR("Inju Song <inju.song@navercorp.com>");