Spaces:
Running
Running
File size: 6,023 Bytes
b110593 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 |
// _ _
// __ _____ __ ___ ___ __ _| |_ ___
// \ \ /\ / / _ \/ _` \ \ / / |/ _` | __/ _ \
// \ V V / __/ (_| |\ V /| | (_| | || __/
// \_/\_/ \___|\__,_| \_/ |_|\__,_|\__\___|
//
// Copyright © 2016 - 2024 Weaviate B.V. All rights reserved.
//
// CONTACT: [email protected]
//
package replica
import (
"context"
"fmt"
"sync"
"github.com/sirupsen/logrus"
)
type (
// readyOp asks a replica if it is ready to commit
readyOp func(_ context.Context, host, requestID string) error
// readyOp asks a replica to execute the actual operation
commitOp[T any] func(_ context.Context, host, requestID string) (T, error)
// readOp defines a generic read operation
readOp[T any] func(_ context.Context, host string, fullRead bool) (T, error)
// coordinator coordinates replication of write and read requests
coordinator[T any] struct {
Client
Resolver *resolver // node_name -> host_address
log logrus.FieldLogger
Class string
Shard string
TxID string // transaction ID
}
)
// newCoordinator used by the replicator
func newCoordinator[T any](r *Replicator, shard, requestID string, l logrus.FieldLogger,
) *coordinator[T] {
return &coordinator[T]{
Client: r.client,
Resolver: r.resolver,
log: l,
Class: r.class,
Shard: shard,
TxID: requestID,
}
}
// newCoordinator used by the Finder to read objects from replicas
func newReadCoordinator[T any](f *Finder, shard string) *coordinator[T] {
return &coordinator[T]{
Resolver: f.resolver,
Class: f.class,
Shard: shard,
}
}
// broadcast sends write request to all replicas (first phase of a two-phase commit)
func (c *coordinator[T]) broadcast(ctx context.Context,
replicas []string,
op readyOp, level int,
) <-chan string {
// prepare tells replicas to be ready
prepare := func() <-chan _Result[string] {
resChan := make(chan _Result[string], len(replicas))
go func() { // broadcast
defer close(resChan)
var wg sync.WaitGroup
wg.Add(len(replicas))
for _, replica := range replicas {
go func(replica string, candidateCh chan<- _Result[string]) error {
defer wg.Done()
err := op(ctx, replica, c.TxID)
candidateCh <- _Result[string]{replica, err}
return err
}(replica, resChan)
}
wg.Wait()
}()
return resChan
}
// handle responses to prepare requests
replicaCh := make(chan string, len(replicas))
go func(level int) {
defer close(replicaCh)
actives := make([]string, 0, level) // cache for active replicas
for r := range prepare() {
if r.Err != nil { // connection error
c.log.WithField("op", "broadcast").Error(r.Err)
continue
}
level--
if level > 0 { // cache since level has not been reached yet
actives = append(actives, r.Value)
continue
}
if level == 0 { // consistency level has been reached
for _, x := range actives {
replicaCh <- x
}
}
replicaCh <- r.Value
}
if level > 0 { // abort: nothing has been sent to the caller
fs := logrus.Fields{"op": "broadcast", "active": len(actives), "total": len(replicas)}
c.log.WithFields(fs).Error("abort")
for _, node := range replicas {
c.Abort(ctx, node, c.Class, c.Shard, c.TxID)
}
}
}(level)
return replicaCh
}
// commitAll tells replicas to commit pending updates related to a specific request
// (second phase of a two-phase commit)
func (c *coordinator[T]) commitAll(ctx context.Context,
replicaCh <-chan string,
op commitOp[T],
) <-chan _Result[T] {
replyCh := make(chan _Result[T], cap(replicaCh))
go func() { // tells active replicas to commit
wg := sync.WaitGroup{}
for replica := range replicaCh {
wg.Add(1)
go func(replica string) {
defer wg.Done()
resp, err := op(ctx, replica, c.TxID)
replyCh <- _Result[T]{resp, err}
}(replica)
}
wg.Wait()
close(replyCh)
}()
return replyCh
}
// Push pushes updates to all replicas of a specific shard
func (c *coordinator[T]) Push(ctx context.Context,
cl ConsistencyLevel,
ask readyOp,
com commitOp[T],
) (<-chan _Result[T], int, error) {
state, err := c.Resolver.State(c.Shard, cl, "")
if err != nil {
return nil, 0, fmt.Errorf("%w : class %q shard %q", err, c.Class, c.Shard)
}
level := state.Level
nodeCh := c.broadcast(ctx, state.Hosts, ask, level)
return c.commitAll(context.Background(), nodeCh, com), level, nil
}
// Pull data from replica depending on consistency level
// Pull involves just as many replicas to satisfy the consistency level.
//
// directCandidate when specified a direct request is set to this node (default to this node)
func (c *coordinator[T]) Pull(ctx context.Context,
cl ConsistencyLevel,
op readOp[T], directCandidate string,
) (<-chan _Result[T], rState, error) {
state, err := c.Resolver.State(c.Shard, cl, directCandidate)
if err != nil {
return nil, state, fmt.Errorf("%w : class %q shard %q", err, c.Class, c.Shard)
}
level := state.Level
replyCh := make(chan _Result[T], level)
candidates := state.Hosts[:level] // direct ones
candidatePool := make(chan string, len(state.Hosts)-level) // remaining ones
for _, replica := range state.Hosts[level:] {
candidatePool <- replica
}
close(candidatePool) // pool is ready
go func() {
wg := sync.WaitGroup{}
wg.Add(len(candidates))
for i := range candidates { // Ask direct candidate first
go func(idx int) {
defer wg.Done()
resp, err := op(ctx, candidates[idx], idx == 0)
// If node is not responding delegate request to another node
for err != nil {
if delegate, ok := <-candidatePool; ok {
resp, err = op(ctx, delegate, idx == 0)
} else {
break
}
}
replyCh <- _Result[T]{resp, err}
}(i)
}
wg.Wait()
close(replyCh)
}()
return replyCh, state, nil
}
|