main.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
// design touchstones // configured entirely via environment variables // minimal configurable options // a single global cluster // easy cluster formation // todo // dns resolver for services package main import ( "fmt" "os" "os/signal" "strings" "syscall" "time" "github.com/hashicorp/memberlist" "github.com/hashicorp/serf/serf" "golang.org/x/exp/slog" ) const ( // gracefulTimeout controls how long we wait before forcefully terminating // note that this value interacts with serfConfig.LeavePropagateDelay gracefulTimeout = 10 * time.Second ) func main() { config := readConfig() logger := setupLogger(config) agent := setupAgent(config, logger) if err := agent.Start(); err != nil { panic(err) } defer agent.Shutdown() // join any specified startup nodes if err := startupJoin(config, agent); err != nil { panic(err) } go debugPrints(agent) if err := handleSignals(config, agent); err != nil { panic(err) } } func setupLogger(config *Config) *slog.Logger { debugOn := false if config.LogLevel == "DEBUG" { debugOn = true } opts := slog.HandlerOptions{ Level: config.GetLogLevel(), AddSource: debugOn, } handler := opts.NewTextHandler(os.Stderr) logger := slog.New(handler) slog.SetDefault(logger) return logger } func debugPrints(a *Agent) { for { for _, m := range a.serf.Members() { a.logger.Debug("debug-loop", "name", m.Name, "addr", m.Addr, "status", m.Status) } time.Sleep(time.Second * 5) } } // handleSignals blocks until we get an exit-causing signal func handleSignals(config *Config, agent *Agent) error { signalCh := make(chan os.Signal, 4) signal.Notify(signalCh, os.Interrupt, syscall.SIGTERM) // Wait for a signal var sig os.Signal select { case s := <-signalCh: sig = s case <-agent.ShutdownCh(): // Agent is already shutdown! return nil } agent.logger.Info("caught signal", "signal", sig) // Check if we should do a graceful leave graceful := false if sig == os.Interrupt || sig == syscall.SIGTERM { graceful = true } // Bail fast if not doing a graceful leave if !graceful { agent.logger.Warn("leave cluster with zero grace") return nil } // Attempt a graceful leave gracefulCh := make(chan struct{}) agent.logger.Info("shut down agent gracefully") go func() { if err := agent.Leave(); err != nil { agent.logger.Error("Error: %s", err) return } close(gracefulCh) }() // Wait for leave or another signal select { case <-signalCh: return fmt.Errorf("idfk") case <-time.After(gracefulTimeout): return fmt.Errorf("leave timed out") case <-gracefulCh: return nil } } func readConfig() *Config { config := DefaultConfig() // CASCADE_LOGLEVEL=info if os.Getenv("CASCADE_LOGLEVEL") != "" { config.LogLevel = strings.ToUpper(os.Getenv("CASCADE_LOGLEVEL")) } // CASCADE_BIND=192.168.0.15:12345 if os.Getenv("CASCADE_BIND") != "" { config.BindAddr = os.Getenv("CASCADE_BIND") } // CASCADE_JOIN=127.0.0.1,127.0.0.5 if os.Getenv("CASCADE_JOIN") != "" { config.StartJoin = strings.Split(os.Getenv("CASCADE_JOIN"), ",") } // CASCADE_NAME=nostromo.j3s.sh if os.Getenv("CASCADE_NAME") != "" { config.NodeName = os.Getenv("CASCADE_NAME") } return config } func startupJoin(config *Config, agent *Agent) error { if len(config.StartJoin) == 0 { return nil } n, err := agent.Join(config.StartJoin) if err != nil { return err } if n > 0 { agent.logger.Info("issue join request", "nodes", n) } return nil } func setupAgent(config *Config, logger *slog.Logger) *Agent { bindIP, bindPort, err := config.AddrParts(config.BindAddr) if err != nil { panic(err) } serfConfig := serf.DefaultConfig() // logLogger is just a bridge from the old logger to the // new slog handler. logLogger := slog.NewLogLogger(logger.Handler(), config.GetLogLevel()) serfConfig.Logger = logLogger serfConfig.NodeName = config.NodeName serfConfig.ProtocolVersion = uint8(serf.ProtocolVersionMax) // TODO: how should cascade handle name conflicts? // defaulting to just blowing up for now, but // we _could_ take the tailscale route & append // -1 or whatever to the node. that would be more user friendly. // TODO: some of these serf settings were pulled // from consul[1]. re-examine them eventually. serfConfig.EnableNameConflictResolution = false serfConfig.LeavePropagateDelay = 3 * time.Second serfConfig.MinQueueDepth = 4096 serfConfig.QueueDepthWarning = 1000000 serfConfig.ReconnectTimeout = 3 * 24 * time.Hour serfConfig.MemberlistConfig = memberlist.DefaultWANConfig() serfConfig.MemberlistConfig.Logger = logLogger serfConfig.MemberlistConfig.DeadNodeReclaimTime = 30 * time.Second serfConfig.MemberlistConfig.BindAddr = bindIP serfConfig.MemberlistConfig.BindPort = bindPort agent := Create(config, serfConfig, logger) return agent } // [1]: sources for consul serf tweaks // https://github.com/hashicorp/consul/blob/v1.14.4/agent/consul/config.go // https://github.com/hashicorp/consul/blob/v1.14.4/lib/serf/serf.go