varnish-cache/vmod/vmod_directors_shard_cfg.c
0
/*-
1
 * Copyright 2009-2016 UPLEX - Nils Goroll Systemoptimierung
2
 * All rights reserved.
3
 *
4
 * Authors: Nils Goroll <nils.goroll@uplex.de>
5
 *          Geoffrey Simmons <geoff@uplex.de>
6
 *
7
 * SPDX-License-Identifier: BSD-2-Clause
8
 *
9
 * Redistribution and use in source and binary forms, with or without
10
 * modification, are permitted provided that the following conditions
11
 * are met:
12
 * 1. Redistributions of source code must retain the above copyright
13
 *    notice, this list of conditions and the following disclaimer.
14
 * 2. Redistributions in binary form must reproduce the above copyright
15
 *    notice, this list of conditions and the following disclaimer in the
16
 *    documentation and/or other materials provided with the distribution.
17
 *
18
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
19
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21
 * ARE DISCLAIMED.  IN NO EVENT SHALL AUTHOR OR CONTRIBUTORS BE LIABLE
22
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
23
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
24
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
25
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
26
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
27
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
28
 * SUCH DAMAGE.
29
 */
30
31
#include "config.h"
32
33
#include <limits.h>
34
#include <stdlib.h>
35
#include <stdio.h>
36
#include <string.h>
37
38
#include "cache/cache.h"
39
40
#include "vmod_directors_shard_dir.h"
41
#include "vmod_directors_shard_cfg.h"
42
43
/*lint -esym(749,  shard_change_task_e::*) */
44
enum shard_change_task_e {
45
        _SHARD_TASK_E_INVALID = 0,
46
        CLEAR,
47
        ADD_BE,
48
        REMOVE_BE,
49
        _SHARD_TASK_E_MAX
50
};
51
52
struct shard_change_task {
53
        unsigned                                magic;
54
#define SHARD_CHANGE_TASK_MAGIC                 0x1e1168af
55
        enum shard_change_task_e                task;
56
        void                                    *priv;
57
        VCL_REAL                                weight;
58
        VSTAILQ_ENTRY(shard_change_task)        list;
59
};
60
61
struct shard_change {
62
        unsigned                                magic;
63
#define SHARD_CHANGE_MAGIC                      0xdff5c9a6
64
        struct vsl_log                          *vsl;
65
        struct sharddir                         *shardd;
66
        VSTAILQ_HEAD(,shard_change_task)        tasks;
67
};
68
69
struct backend_reconfig {
70
        struct sharddir * const shardd;
71
        unsigned                hint;   // on number of backends after reconfig
72
        unsigned                hole_n; // number of holes in backends array
73
        unsigned                hole_i; // index hint on first hole
74
};
75
76
/* forward decl */
77
static VCL_BOOL
78
change_reconfigure(VRT_CTX, struct shard_change *change, VCL_INT replicas);
79
80
/*
81
 * ============================================================
82
 * change / task list
83
 *
84
 * for backend reconfiguration, we create a change list on the VCL workspace in
85
 * a PRIV_TASK state, which we work in reconfigure.
86
 */
87
88
static void v_matchproto_(vmod_priv_fini_f)
89 190
shard_change_fini(VRT_CTX, void * priv)
90
{
91
        struct shard_change *change;
92
93 190
        if (priv == NULL)
94 0
                return;
95
96 190
        CAST_OBJ_NOTNULL(change, priv, SHARD_CHANGE_MAGIC);
97
98 190
        (void) change_reconfigure(ctx, change, 67);
99 190
}
100
101
static const struct vmod_priv_methods shard_change_priv_methods[1] = {{
102
        .magic = VMOD_PRIV_METHODS_MAGIC,
103
        .type = "vmod_directors_shard_cfg",
104
        .fini = shard_change_fini
105
}};
106
107
static struct shard_change *
108 1970
shard_change_get(VRT_CTX, struct sharddir * const shardd)
109
{
110
        struct vmod_priv *task;
111
        struct shard_change *change;
112 1970
        const void *id = (const char *)shardd + task_off_cfg;
113
114 1970
        CHECK_OBJ_NOTNULL(ctx, VRT_CTX_MAGIC);
115
116 1970
        task = VRT_priv_task(ctx, id);
117 1970
        if (task == NULL) {
118 0
                shard_fail(ctx, shardd->name, "%s", "no priv_task");
119 0
                return (NULL);
120
        }
121
122 1970
        if (task->priv != NULL) {
123 1780
                CAST_OBJ_NOTNULL(change, task->priv, SHARD_CHANGE_MAGIC);
124 1780
                assert (change->vsl == ctx->vsl);
125 1780
                assert (change->shardd == shardd);
126 1780
                return (change);
127
        }
128
129 380
        WS_TASK_ALLOC_OBJ(ctx, change, SHARD_CHANGE_MAGIC);
130 190
        if (change == NULL)
131 0
                return (NULL);
132 190
        change->vsl = ctx->vsl;
133 190
        change->shardd = shardd;
134 190
        VSTAILQ_INIT(&change->tasks);
135 190
        task->priv = change;
136 190
        task->methods = shard_change_priv_methods;
137
138 190
        return (change);
139 1970
}
140
141
static void
142 400
shard_change_finish(struct shard_change *change)
143
{
144 400
        CHECK_OBJ_NOTNULL(change, SHARD_CHANGE_MAGIC);
145
146 400
        VSTAILQ_INIT(&change->tasks);
147 400
}
148
149
static struct shard_change_task *
150 1570
shard_change_task_add(VRT_CTX, struct shard_change *change,
151
    enum shard_change_task_e task_e, void *priv)
152
{
153
        struct shard_change_task *task;
154
155 1570
        CHECK_OBJ_NOTNULL(change, SHARD_CHANGE_MAGIC);
156
157 3140
        WS_TASK_ALLOC_OBJ(ctx, task, SHARD_CHANGE_TASK_MAGIC);
158 1570
        if (task == NULL)
159 0
                return (NULL);
160 1570
        task->task = task_e;
161 1570
        task->priv = priv;
162 1570
        VSTAILQ_INSERT_TAIL(&change->tasks, task, list);
163
164 1570
        return (task);
165 1570
}
166
167
static inline struct shard_change_task *
168 1420
shard_change_task_backend(VRT_CTX, struct sharddir *shardd,
169
    enum shard_change_task_e task_e, VCL_BACKEND be, VCL_STRING ident,
170
    VCL_DURATION rampup)
171
{
172
        struct shard_change *change;
173
        struct shard_backend *b;
174
175 1420
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
176 1420
        assert(task_e == ADD_BE || task_e == REMOVE_BE);
177
178 1420
        change = shard_change_get(ctx, shardd);
179 1420
        if (change == NULL)
180 0
                return (NULL);
181
182 1420
        b = WS_Alloc(ctx->ws, sizeof(*b));
183 1420
        if (b == NULL) {
184 0
                shard_fail(ctx, change->shardd->name, "%s",
185
                    "could not get workspace for task");
186 0
                return (NULL);
187
        }
188
189 1420
        b->backend = NULL;
190 1420
        VRT_Assign_Backend(&b->backend, be);
191 1420
        b->ident = ident != NULL && *ident != '\0' ? ident : NULL;
192 1420
        b->rampup = rampup;
193
194 1420
        return (shard_change_task_add(ctx, change, task_e, b));
195 1420
}
196
197
/*
198
 * ============================================================
199
 * director reconfiguration tasks
200
 */
201
VCL_BOOL
202 1130
shardcfg_add_backend(VRT_CTX, struct sharddir *shardd,
203
    VCL_BACKEND be, VCL_STRING ident, VCL_DURATION rampup, VCL_REAL weight)
204
{
205
        struct shard_change_task *task;
206
207 1130
        assert (weight >= 1);
208 1130
        AN(be);
209
210 2260
        task = shard_change_task_backend(ctx, shardd, ADD_BE,
211 1130
            be, ident, rampup);
212
213 1130
        if (task == NULL)
214 0
                return (0);
215
216 1130
        task->weight = weight;
217 1130
        return (1);
218 1130
}
219
220
VCL_BOOL
221 290
shardcfg_remove_backend(VRT_CTX, struct sharddir *shardd,
222
    VCL_BACKEND be, VCL_STRING ident)
223
{
224 870
        return (shard_change_task_backend(ctx, shardd, REMOVE_BE,
225 580
            be, ident, 0) != NULL);
226
}
227
228
VCL_BOOL
229 150
shardcfg_clear(VRT_CTX, struct sharddir *shardd)
230
{
231
        struct shard_change *change;
232
233 150
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
234
235 150
        change = shard_change_get(ctx, shardd);
236 150
        if (change == NULL)
237 0
                return (0);
238
239 150
        return (shard_change_task_add(ctx, change, CLEAR, NULL) != NULL);
240 150
}
241
242
/*
243
 * ============================================================
244
 * consistent hashing circle init
245
 */
246
247
typedef int (*compar)( const void*, const void* );
248
249
static int
250 104230
circlepoint_compare(const struct shard_circlepoint *a,
251
    const struct shard_circlepoint *b)
252
{
253 104230
        return ((a->point == b->point) ? 0 : ((a->point > b->point) ? 1 : -1));
254
}
255
256
static void
257 360
shardcfg_hashcircle(struct sharddir *shardd)
258
{
259
        const struct shard_backend *backends, *b;
260
        unsigned h;
261
        uint32_t i, j, n_points, r, rmax;
262
        const char *ident;
263
        char s[12]; // log10(UINT32_MAX) + 2;
264
265 360
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
266 360
        AZ(shardd->hashcircle);
267
268 360
        assert(shardd->n_backend > 0);
269 360
        backends=shardd->backend;
270 360
        AN(backends);
271
272 360
        n_points = 0;
273 360
        rmax = (UINT32_MAX - 1) / shardd->n_backend;
274 1870
        for (b = backends; b < backends + shardd->n_backend; b++) {
275 1510
                CHECK_OBJ_NOTNULL(b->backend, DIRECTOR_MAGIC);
276 1510
                n_points += vmin_t(uint32_t, b->replicas, rmax);
277 1510
        }
278
279 360
        assert(n_points < UINT32_MAX);
280
281 360
        shardd->n_points = n_points;
282 360
        shardd->hashcircle = calloc(n_points, sizeof(struct shard_circlepoint));
283 360
        AN(shardd->hashcircle);
284
285 360
        i = 0;
286 1870
        for (h = 0, b = backends; h < shardd->n_backend; h++, b++) {
287 1510
                ident = b->ident ? b->ident : VRT_BACKEND_string(b->backend);
288
289 1510
                AN(ident);
290 1510
                assert(ident[0] != '\0');
291
292 1510
                r = vmin_t(uint32_t, b->replicas, rmax);
293
294 18350
                for (j = 0; j < r; j++) {
295 16840
                        bprintf(s, "%d", j);
296 16840
                        assert (i < n_points);
297 16840
                        shardd->hashcircle[i].point =
298 16840
                            VRT_HashStrands32(TOSTRANDS(2, ident, s));
299 16840
                        shardd->hashcircle[i].host = h;
300 16840
                        i++;
301 16840
                }
302 1510
        }
303 360
        assert (i == n_points);
304 360
        qsort( (void *) shardd->hashcircle, n_points,
305
            sizeof (struct shard_circlepoint), (compar) circlepoint_compare);
306
307 360
        if ((shardd->debug_flags & SHDBG_CIRCLE) == 0)
308 130
                return;
309
310 6290
        for (i = 0; i < n_points; i++)
311 6060
                SHDBG(SHDBG_CIRCLE, shardd,
312
                    "hashcircle[%5jd] = {point = %8x, host = %2u}\n",
313
                    (intmax_t)i, shardd->hashcircle[i].point,
314
                    shardd->hashcircle[i].host);
315 360
}
316
317
/*
318
 * ============================================================
319
 * configure the director backends
320
 */
321
322
static void
323 460
shardcfg_backend_free(struct shard_backend *f)
324
{
325 460
        if (f->freeptr)
326 350
                free (f->freeptr);
327 460
        VRT_Assign_Backend(&f->backend, NULL);
328 460
        memset(f, 0, sizeof(*f));
329 460
}
330
331
static void
332 970
shardcfg_backend_copyin(struct shard_backend *dst,
333
    const struct shard_backend *src)
334
{
335 970
        dst->backend = src->backend;
336 970
        dst->ident = src->ident ? strdup(src->ident) : NULL;
337 970
        dst->rampup = src->rampup;
338 970
}
339
340
static int
341 7140
shardcfg_backend_cmp(const struct shard_backend *a,
342
    const struct shard_backend *b)
343
{
344
        const char *ai, *bi;
345
346 7140
        ai = a->ident;
347 7140
        bi = b->ident;
348
349 7140
        assert(ai || a->backend);
350 7140
        assert(bi || b->backend);
351
352
        /* vcl_names are unique, so we can compare the backend pointers */
353 7140
        if (ai == NULL && bi == NULL)
354 470
                return (a->backend != b->backend);
355
356 6670
        if (ai == NULL)
357 60
                ai = VRT_BACKEND_string(a->backend);
358
359 6670
        if (bi == NULL)
360 150
                bi = VRT_BACKEND_string(b->backend);
361
362 6670
        AN(ai);
363 6670
        AN(bi);
364 6670
        return (strcmp(ai, bi));
365 7140
}
366
367
/* for removal, we delete all instances if the backend matches */
368
static int
369 3250
shardcfg_backend_del_cmp(const struct shard_backend *task,
370
    const struct shard_backend *b)
371
{
372 3250
        assert(task->backend || task->ident);
373
374 3250
        if (task->ident == NULL)
375 90
                return (task->backend != b->backend);
376
377 3160
        return (shardcfg_backend_cmp(task, b));
378 3250
}
379
380
static const struct shard_backend *
381 1050
shardcfg_backend_lookup(const struct backend_reconfig *re,
382
    const struct shard_backend *b)
383
{
384 1050
        unsigned i, max = re->shardd->n_backend + re->hole_n;
385 1050
        const struct shard_backend *bb = re->shardd->backend;
386
387 1050
        if (max > 0)
388 800
                AN(bb);
389
390 5010
        for (i = 0; i < max; i++) {
391 4040
                if (bb[i].backend == NULL)
392 60
                        continue;       // hole
393 3980
                if (!shardcfg_backend_cmp(b, &bb[i]))
394 80
                        return (&bb[i]);
395 3900
        }
396 970
        return (NULL);
397 1050
}
398
399
static void
400 190
shardcfg_backend_expand(const struct backend_reconfig *re)
401
{
402 190
        unsigned min = re->hint;
403
404 190
        CHECK_OBJ_NOTNULL(re->shardd, SHARDDIR_MAGIC);
405
406 190
        min = vmax_t(unsigned, min, 16);
407
408 190
        if (re->shardd->l_backend < min)
409 190
                re->shardd->l_backend = min;
410
        else
411 0
                re->shardd->l_backend *= 2;
412
413 380
        re->shardd->backend = realloc(re->shardd->backend,
414 190
            re->shardd->l_backend * sizeof *re->shardd->backend);
415
416 190
        AN(re->shardd->backend);
417 190
}
418
419
static void
420 970
shardcfg_backend_add(struct backend_reconfig *re,
421
    const struct shard_backend *b, uint32_t replicas)
422
{
423
        unsigned i;
424 970
        struct shard_backend *bb = re->shardd->backend;
425
426 970
        if (re->hole_n == 0) {
427 930
                if (re->shardd->n_backend >= re->shardd->l_backend) {
428 190
                        shardcfg_backend_expand(re);
429 190
                        bb = re->shardd->backend;
430 190
                }
431 930
                assert(re->shardd->n_backend < re->shardd->l_backend);
432 930
                i = re->shardd->n_backend;
433 930
        } else {
434 40
                assert(re->hole_i != UINT_MAX);
435 40
                do {
436 40
                        if (!bb[re->hole_i].backend)
437 40
                                break;
438 0
                } while (++(re->hole_i) < re->shardd->n_backend + re->hole_n);
439 40
                assert(re->hole_i < re->shardd->n_backend + re->hole_n);
440
441 40
                i = (re->hole_i)++;
442 40
                (re->hole_n)--;
443
        }
444
445 970
        re->shardd->n_backend++;
446 970
        shardcfg_backend_copyin(&bb[i], b);
447 970
        bb[i].replicas = replicas;
448 970
}
449
450
void
451 180
shardcfg_backend_clear(struct sharddir *shardd)
452
{
453
        unsigned i;
454 330
        for (i = 0; i < shardd->n_backend; i++)
455 150
                shardcfg_backend_free(&shardd->backend[i]);
456 180
        shardd->n_backend = 0;
457 180
}
458
459
460
static void
461 290
shardcfg_backend_del(struct backend_reconfig *re, struct shard_backend *spec)
462
{
463 290
        unsigned i, max = re->shardd->n_backend + re->hole_n;
464 290
        struct shard_backend * const bb = re->shardd->backend;
465
466 4760
        for (i = 0; i < max; i++) {
467 4470
                if (bb[i].backend == NULL)
468 1220
                        continue;       // hole
469 3250
                if (shardcfg_backend_del_cmp(spec, &bb[i]))
470 2940
                        continue;
471
472 310
                shardcfg_backend_free(&bb[i]);
473 310
                re->shardd->n_backend--;
474 310
                if (i < re->shardd->n_backend + re->hole_n) {
475 270
                        (re->hole_n)++;
476 270
                        re->hole_i = vmin(re->hole_i, i);
477 270
                }
478 310
        }
479 290
        VRT_Assign_Backend(&spec->backend, NULL);
480 290
}
481
482
static void
483 360
shardcfg_backend_finalize(struct backend_reconfig *re)
484
{
485
        unsigned i;
486 360
        struct shard_backend * const bb = re->shardd->backend;
487
488 420
        while (re->hole_n > 0) {
489
                // trim end
490 90
                i = re->shardd->n_backend + re->hole_n - 1;
491 260
                while (re->hole_n && bb[i].backend == NULL) {
492 170
                        (re->hole_n)--;
493 170
                        i--;
494
                }
495
496 90
                if (re->hole_n == 0)
497 30
                        break;
498
499 60
                assert(re->hole_i < i);
500
501 60
                do {
502 60
                        if (!bb[re->hole_i].backend)
503 60
                                break;
504 0
                } while (++(re->hole_i) <= i);
505
506 60
                assert(re->hole_i < i);
507 60
                assert(bb[re->hole_i].backend == NULL);
508 60
                assert(bb[i].backend != NULL);
509
510 60
                memcpy(&bb[re->hole_i], &bb[i], sizeof(*bb));
511 60
                memset(&bb[i], 0, sizeof(*bb));
512
513 60
                (re->hole_n)--;
514 60
                (re->hole_i)++;
515
        }
516
517 360
        assert(re->hole_n == 0);
518 360
}
519
520
/*
521
 * ============================================================
522
 * work the change tasks
523
 */
524
525
static void
526 400
shardcfg_apply_change(struct vsl_log *vsl, struct sharddir *shardd,
527
    const struct shard_change *change, VCL_INT replicas)
528
{
529
        struct shard_change_task *task, *clear;
530
        const struct shard_backend *b;
531
        uint32_t b_replicas;
532
533 1200
        struct backend_reconfig re = {
534 400
                .shardd = shardd,
535 400
                .hint = shardd->n_backend,
536
                .hole_n = 0,
537
                .hole_i = UINT_MAX
538
        };
539
540
        // XXX assert sharddir_locked(shardd)
541
542 400
        clear = NULL;
543 1970
        VSTAILQ_FOREACH(task, &change->tasks, list) {
544 1570
                CHECK_OBJ_NOTNULL(task, SHARD_CHANGE_TASK_MAGIC);
545 1570
                switch (task->task) {
546
                case CLEAR:
547 150
                        clear = task;
548 150
                        re.hint = 0;
549 150
                        break;
550
                case ADD_BE:
551 1130
                        re.hint++;
552 1130
                        break;
553
                case REMOVE_BE:
554 290
                        break;
555
                default:
556 0
                        INCOMPL();
557 0
                }
558 1570
        }
559
560 400
        if (clear) {
561 130
                shardcfg_backend_clear(shardd);
562 130
                clear = VSTAILQ_NEXT(clear, list);
563 130
                if (clear == NULL)
564 40
                        return;
565 90
        }
566
567 360
        task = clear;
568 1700
        VSTAILQ_FOREACH_FROM(task, &change->tasks, list) {
569 1340
                CHECK_OBJ_NOTNULL(task, SHARD_CHANGE_TASK_MAGIC);
570 1340
                switch (task->task) {
571
                case CLEAR:
572 0
                        assert(task->task != CLEAR);
573 0
                        break;
574
                case ADD_BE:
575 1050
                        b = shardcfg_backend_lookup(&re, task->priv);
576
577 1050
                        if (b == NULL) {
578 970
                                assert (task->weight >= 1);
579 970
                                if (replicas * task->weight > UINT32_MAX)
580 0
                                        b_replicas = UINT32_MAX;
581
                                else
582 970
                                        b_replicas = (uint32_t) // flint
583 970
                                                (replicas * task->weight);
584
585 1940
                                shardcfg_backend_add(&re, task->priv,
586 970
                                    b_replicas);
587 970
                                break;
588
                        }
589
590 80
                        const char * const ident = b->ident;
591
592 80
                        shard_notice(vsl, shardd->name,
593
                            "backend %s%s%s already exists - skipping",
594
                            VRT_BACKEND_string(b->backend),
595
                            ident ? "/" : "",
596
                            ident ? ident : "");
597 80
                        break;
598
                case REMOVE_BE:
599 290
                        shardcfg_backend_del(&re, task->priv);
600 290
                        break;
601
                default:
602 0
                        INCOMPL();
603 0
                }
604 1340
        }
605 360
        shardcfg_backend_finalize(&re);
606 400
}
607
608
/*
609
 * ============================================================
610
 * top reconfiguration function
611
 */
612
613
static VCL_BOOL
614 590
change_reconfigure(VRT_CTX, struct shard_change *change, VCL_INT replicas)
615
{
616
        struct sharddir *shardd;
617
618 590
        CHECK_OBJ_NOTNULL(change, SHARD_CHANGE_MAGIC);
619 590
        assert (replicas > 0);
620 590
        shardd = change->shardd;
621 590
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
622
623 590
        if (VSTAILQ_FIRST(&change->tasks) == NULL)
624 190
                return (1);
625
626 400
        sharddir_wrlock(shardd);
627
628 400
        shardcfg_apply_change(ctx->vsl, shardd, change, replicas);
629 400
        shard_change_finish(change);
630
631 400
        if (shardd->hashcircle)
632 180
                free(shardd->hashcircle);
633 400
        shardd->hashcircle = NULL;
634
635 400
        if (shardd->n_backend == 0) {
636 40
                shard_err0(ctx->vsl, shardd->name,
637
                    ".reconfigure() no backends");
638 40
                sharddir_unlock(shardd);
639 40
                return (0);
640
        }
641
642 360
        shardcfg_hashcircle(shardd);
643 360
        sharddir_unlock(shardd);
644 360
        return (1);
645 590
}
646
647
VCL_BOOL
648 420
shardcfg_reconfigure(VRT_CTX, struct sharddir *shardd, VCL_INT replicas)
649
{
650
        struct shard_change *change;
651
652 420
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
653 420
        if (replicas <= 0) {
654 20
                shard_err(ctx->vsl, shardd->name,
655
                    ".reconfigure() invalid replicas argument %ld", replicas);
656 20
                return (0);
657
        }
658
659 400
        change = shard_change_get(ctx, shardd);
660 400
        if (change == NULL)
661 0
                return (0);
662
663 400
        return (change_reconfigure(ctx, change, replicas));
664 420
}
665
666
/*
667
 * ============================================================
668
 * misc config related
669
 */
670
671
/* only for sharddir_delete() */
672
void
673 50
shardcfg_delete(const struct sharddir *shardd)
674
{
675
676 50
        AZ(shardd->n_backend);
677 50
        if (shardd->backend)
678 0
                free(shardd->backend);
679 50
        if (shardd->hashcircle)
680 0
                free(shardd->hashcircle);
681 50
}
682
683
VCL_VOID
684 20
shardcfg_set_warmup(struct sharddir *shardd, VCL_REAL ratio)
685
{
686 20
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
687 20
        assert(ratio >= 0 && ratio < 1);
688 20
        sharddir_wrlock(shardd);
689 20
        shardd->warmup = ratio;
690 20
        sharddir_unlock(shardd);
691 20
}
692
693
VCL_VOID
694 20
shardcfg_set_rampup(struct sharddir *shardd, VCL_DURATION duration)
695
{
696 20
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
697 20
        assert(duration >= 0);
698 20
        sharddir_wrlock(shardd);
699 20
        shardd->rampup_duration = duration;
700 20
        sharddir_unlock(shardd);
701 20
}
702
703
VCL_DURATION
704 1620
shardcfg_get_rampup(const struct sharddir *shardd, unsigned host)
705
{
706
        VCL_DURATION r;
707
708 1620
        CHECK_OBJ_NOTNULL(shardd, SHARDDIR_MAGIC);
709
        // assert sharddir_rdlock_held(shardd);
710 1620
        assert (host < shardd->n_backend);
711
712 1620
        if (isnan(shardd->backend[host].rampup))
713 1580
                r = shardd->rampup_duration;
714
        else
715 40
                r = shardd->backend[host].rampup;
716
717 1620
        return (r);
718
}