database.go 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442
  1. // Copyright (C) 2018 The Syncthing Authors.
  2. //
  3. // This Source Code Form is subject to the terms of the Mozilla Public
  4. // License, v. 2.0. If a copy of the MPL was not distributed with this file,
  5. // You can obtain one at https://mozilla.org/MPL/2.0/.
  6. package main
  7. import (
  8. "bufio"
  9. "cmp"
  10. "context"
  11. "encoding/binary"
  12. "errors"
  13. "io"
  14. "log"
  15. "os"
  16. "path"
  17. "runtime"
  18. "slices"
  19. "strings"
  20. "time"
  21. "github.com/puzpuzpuz/xsync/v3"
  22. "google.golang.org/protobuf/proto"
  23. "github.com/syncthing/syncthing/internal/gen/discosrv"
  24. "github.com/syncthing/syncthing/internal/protoutil"
  25. "github.com/syncthing/syncthing/lib/protocol"
  26. "github.com/syncthing/syncthing/lib/rand"
  27. "github.com/syncthing/syncthing/lib/s3"
  28. )
  29. type clock interface {
  30. Now() time.Time
  31. }
  32. type defaultClock struct{}
  33. func (defaultClock) Now() time.Time {
  34. return time.Now()
  35. }
  36. type database interface {
  37. put(key *protocol.DeviceID, rec *discosrv.DatabaseRecord) error
  38. merge(key *protocol.DeviceID, addrs []*discosrv.DatabaseAddress, seen int64) error
  39. get(key *protocol.DeviceID) (*discosrv.DatabaseRecord, error)
  40. }
  41. type inMemoryStore struct {
  42. m *xsync.MapOf[protocol.DeviceID, *discosrv.DatabaseRecord]
  43. dir string
  44. flushInterval time.Duration
  45. s3 *s3.Session
  46. objKey string
  47. clock clock
  48. }
  49. func newInMemoryStore(dir string, flushInterval time.Duration, s3sess *s3.Session) *inMemoryStore {
  50. hn, err := os.Hostname()
  51. if err != nil {
  52. hn = rand.String(8)
  53. }
  54. s := &inMemoryStore{
  55. m: xsync.NewMapOf[protocol.DeviceID, *discosrv.DatabaseRecord](),
  56. dir: dir,
  57. flushInterval: flushInterval,
  58. s3: s3sess,
  59. objKey: hn + ".db",
  60. clock: defaultClock{},
  61. }
  62. nr, err := s.read()
  63. if os.IsNotExist(err) && s3sess != nil {
  64. // Try to read from AWS
  65. latestKey, cerr := s3sess.LatestKey()
  66. if cerr != nil {
  67. log.Println("Error reading database from S3:", err)
  68. return s
  69. }
  70. fd, cerr := os.Create(path.Join(s.dir, "records.db"))
  71. if cerr != nil {
  72. log.Println("Error creating database file:", err)
  73. return s
  74. }
  75. if cerr := s3sess.Download(fd, latestKey); cerr != nil {
  76. log.Printf("Error reading database from S3: %v", err)
  77. }
  78. _ = fd.Close()
  79. nr, err = s.read()
  80. }
  81. if err != nil {
  82. log.Println("Error reading database:", err)
  83. }
  84. log.Printf("Read %d records from database", nr)
  85. s.expireAndCalculateStatistics()
  86. return s
  87. }
  88. func (s *inMemoryStore) put(key *protocol.DeviceID, rec *discosrv.DatabaseRecord) error {
  89. t0 := time.Now()
  90. s.m.Store(*key, rec)
  91. databaseOperations.WithLabelValues(dbOpPut, dbResSuccess).Inc()
  92. databaseOperationSeconds.WithLabelValues(dbOpPut).Observe(time.Since(t0).Seconds())
  93. return nil
  94. }
  95. func (s *inMemoryStore) merge(key *protocol.DeviceID, addrs []*discosrv.DatabaseAddress, seen int64) error {
  96. t0 := time.Now()
  97. newRec := &discosrv.DatabaseRecord{
  98. Addresses: addrs,
  99. Seen: seen,
  100. }
  101. if oldRec, ok := s.m.Load(*key); ok {
  102. newRec = merge(oldRec, newRec)
  103. }
  104. s.m.Store(*key, newRec)
  105. databaseOperations.WithLabelValues(dbOpMerge, dbResSuccess).Inc()
  106. databaseOperationSeconds.WithLabelValues(dbOpMerge).Observe(time.Since(t0).Seconds())
  107. return nil
  108. }
  109. func (s *inMemoryStore) get(key *protocol.DeviceID) (*discosrv.DatabaseRecord, error) {
  110. t0 := time.Now()
  111. defer func() {
  112. databaseOperationSeconds.WithLabelValues(dbOpGet).Observe(time.Since(t0).Seconds())
  113. }()
  114. rec, ok := s.m.Load(*key)
  115. if !ok {
  116. databaseOperations.WithLabelValues(dbOpGet, dbResNotFound).Inc()
  117. return &discosrv.DatabaseRecord{}, nil
  118. }
  119. rec.Addresses = expire(rec.Addresses, s.clock.Now())
  120. databaseOperations.WithLabelValues(dbOpGet, dbResSuccess).Inc()
  121. return rec, nil
  122. }
  123. func (s *inMemoryStore) Serve(ctx context.Context) error {
  124. if s.flushInterval <= 0 {
  125. <-ctx.Done()
  126. return nil
  127. }
  128. t := time.NewTimer(s.flushInterval)
  129. defer t.Stop()
  130. loop:
  131. for {
  132. select {
  133. case <-t.C:
  134. log.Println("Calculating statistics")
  135. s.expireAndCalculateStatistics()
  136. log.Println("Flushing database")
  137. if err := s.write(); err != nil {
  138. log.Println("Error writing database:", err)
  139. }
  140. log.Println("Finished flushing database")
  141. t.Reset(s.flushInterval)
  142. case <-ctx.Done():
  143. // We're done.
  144. break loop
  145. }
  146. }
  147. return s.write()
  148. }
  149. func (s *inMemoryStore) expireAndCalculateStatistics() {
  150. now := s.clock.Now()
  151. cutoff24h := now.Add(-24 * time.Hour).UnixNano()
  152. cutoff1w := now.Add(-7 * 24 * time.Hour).UnixNano()
  153. current, currentIPv4, currentIPv6, currentIPv6GUA, last24h, last1w := 0, 0, 0, 0, 0, 0
  154. n := 0
  155. s.m.Range(func(key protocol.DeviceID, rec *discosrv.DatabaseRecord) bool {
  156. if n%1000 == 0 {
  157. runtime.Gosched()
  158. }
  159. n++
  160. addresses := expire(rec.Addresses, now)
  161. if len(addresses) == 0 {
  162. rec.Addresses = nil
  163. s.m.Store(key, rec)
  164. } else if len(addresses) != len(rec.Addresses) {
  165. rec.Addresses = addresses
  166. s.m.Store(key, rec)
  167. }
  168. switch {
  169. case len(rec.Addresses) > 0:
  170. current++
  171. seenIPv4, seenIPv6, seenIPv6GUA := false, false, false
  172. for _, addr := range rec.Addresses {
  173. // We do fast and loose matching on strings here instead of
  174. // parsing the address and the IP and doing "proper" checks,
  175. // to keep things fast and generate less garbage.
  176. if strings.Contains(addr.Address, "[") {
  177. seenIPv6 = true
  178. if strings.Contains(addr.Address, "[2") {
  179. seenIPv6GUA = true
  180. }
  181. } else {
  182. seenIPv4 = true
  183. }
  184. if seenIPv4 && seenIPv6 && seenIPv6GUA {
  185. break
  186. }
  187. }
  188. if seenIPv4 {
  189. currentIPv4++
  190. }
  191. if seenIPv6 {
  192. currentIPv6++
  193. }
  194. if seenIPv6GUA {
  195. currentIPv6GUA++
  196. }
  197. case rec.Seen > cutoff24h:
  198. last24h++
  199. case rec.Seen > cutoff1w:
  200. last1w++
  201. default:
  202. // drop the record if it's older than a week
  203. s.m.Delete(key)
  204. }
  205. return true
  206. })
  207. databaseKeys.WithLabelValues("current").Set(float64(current))
  208. databaseKeys.WithLabelValues("currentIPv4").Set(float64(currentIPv4))
  209. databaseKeys.WithLabelValues("currentIPv6").Set(float64(currentIPv6))
  210. databaseKeys.WithLabelValues("currentIPv6GUA").Set(float64(currentIPv6GUA))
  211. databaseKeys.WithLabelValues("last24h").Set(float64(last24h))
  212. databaseKeys.WithLabelValues("last1w").Set(float64(last1w))
  213. databaseStatisticsSeconds.Set(time.Since(now).Seconds())
  214. }
  215. func (s *inMemoryStore) write() (err error) {
  216. t0 := time.Now()
  217. defer func() {
  218. if err == nil {
  219. databaseWriteSeconds.Set(time.Since(t0).Seconds())
  220. databaseLastWritten.Set(float64(t0.Unix()))
  221. }
  222. }()
  223. dbf := path.Join(s.dir, "records.db")
  224. fd, err := os.Create(dbf + ".tmp")
  225. if err != nil {
  226. return err
  227. }
  228. bw := bufio.NewWriter(fd)
  229. var buf []byte
  230. var rangeErr error
  231. now := s.clock.Now()
  232. cutoff1w := now.Add(-7 * 24 * time.Hour).UnixNano()
  233. n := 0
  234. s.m.Range(func(key protocol.DeviceID, value *discosrv.DatabaseRecord) bool {
  235. if n%1000 == 0 {
  236. runtime.Gosched()
  237. }
  238. n++
  239. if value.Seen < cutoff1w {
  240. // drop the record if it's older than a week
  241. return true
  242. }
  243. rec := &discosrv.ReplicationRecord{
  244. Key: key[:],
  245. Addresses: value.Addresses,
  246. Seen: value.Seen,
  247. }
  248. s := proto.Size(rec)
  249. if s+4 > len(buf) {
  250. buf = make([]byte, s+4)
  251. }
  252. n, err := protoutil.MarshalTo(buf[4:], rec)
  253. if err != nil {
  254. rangeErr = err
  255. return false
  256. }
  257. binary.BigEndian.PutUint32(buf, uint32(n))
  258. if _, err := bw.Write(buf[:n+4]); err != nil {
  259. rangeErr = err
  260. return false
  261. }
  262. return true
  263. })
  264. if rangeErr != nil {
  265. _ = fd.Close()
  266. return rangeErr
  267. }
  268. if err := bw.Flush(); err != nil {
  269. _ = fd.Close
  270. return err
  271. }
  272. if err := fd.Close(); err != nil {
  273. return err
  274. }
  275. if err := os.Rename(dbf+".tmp", dbf); err != nil {
  276. return err
  277. }
  278. // Upload to S3
  279. if s.s3 != nil {
  280. fd, err = os.Open(dbf)
  281. if err != nil {
  282. log.Printf("Error uploading database to S3: %v", err)
  283. return nil
  284. }
  285. defer fd.Close()
  286. if err := s.s3.Upload(fd, s.objKey); err != nil {
  287. log.Printf("Error uploading database to S3: %v", err)
  288. }
  289. log.Println("Finished uploading database")
  290. }
  291. return nil
  292. }
  293. func (s *inMemoryStore) read() (int, error) {
  294. fd, err := os.Open(path.Join(s.dir, "records.db"))
  295. if err != nil {
  296. return 0, err
  297. }
  298. defer fd.Close()
  299. br := bufio.NewReader(fd)
  300. var buf []byte
  301. nr := 0
  302. for {
  303. var n uint32
  304. if err := binary.Read(br, binary.BigEndian, &n); err != nil {
  305. if errors.Is(err, io.EOF) {
  306. break
  307. }
  308. return nr, err
  309. }
  310. if int(n) > len(buf) {
  311. buf = make([]byte, n)
  312. }
  313. if _, err := io.ReadFull(br, buf[:n]); err != nil {
  314. return nr, err
  315. }
  316. rec := &discosrv.ReplicationRecord{}
  317. if err := proto.Unmarshal(buf[:n], rec); err != nil {
  318. return nr, err
  319. }
  320. key, err := protocol.DeviceIDFromBytes(rec.Key)
  321. if err != nil {
  322. key, err = protocol.DeviceIDFromString(string(rec.Key))
  323. }
  324. if err != nil {
  325. log.Println("Bad device ID:", err)
  326. continue
  327. }
  328. slices.SortFunc(rec.Addresses, Cmp)
  329. rec.Addresses = slices.CompactFunc(rec.Addresses, Equal)
  330. s.m.Store(key, &discosrv.DatabaseRecord{
  331. Addresses: expire(rec.Addresses, s.clock.Now()),
  332. Seen: rec.Seen,
  333. })
  334. nr++
  335. }
  336. return nr, nil
  337. }
  338. // merge returns the merged result of the two database records a and b. The
  339. // result is the union of the two address sets, with the newer expiry time
  340. // chosen for any duplicates. The address list in a is overwritten and
  341. // reused for the result.
  342. func merge(a, b *discosrv.DatabaseRecord) *discosrv.DatabaseRecord {
  343. // Both lists must be sorted for this to work.
  344. a.Seen = max(a.Seen, b.Seen)
  345. aIdx := 0
  346. bIdx := 0
  347. for aIdx < len(a.Addresses) && bIdx < len(b.Addresses) {
  348. switch cmp.Compare(a.Addresses[aIdx].Address, b.Addresses[bIdx].Address) {
  349. case 0:
  350. // a == b, choose the newer expiry time
  351. a.Addresses[aIdx].Expires = max(a.Addresses[aIdx].Expires, b.Addresses[bIdx].Expires)
  352. aIdx++
  353. bIdx++
  354. case -1:
  355. // a < b, keep a and move on
  356. aIdx++
  357. case 1:
  358. // a > b, insert b before a
  359. a.Addresses = append(a.Addresses[:aIdx], append([]*discosrv.DatabaseAddress{b.Addresses[bIdx]}, a.Addresses[aIdx:]...)...)
  360. bIdx++
  361. }
  362. }
  363. if bIdx < len(b.Addresses) {
  364. a.Addresses = append(a.Addresses, b.Addresses[bIdx:]...)
  365. }
  366. return a
  367. }
  368. // expire returns the list of addresses after removing expired entries.
  369. // Expiration happen in place, so the slice given as the parameter is
  370. // destroyed. Internal order is preserved.
  371. func expire(addrs []*discosrv.DatabaseAddress, now time.Time) []*discosrv.DatabaseAddress {
  372. cutoff := now.UnixNano()
  373. naddrs := addrs[:0]
  374. for i := range addrs {
  375. if i > 0 && addrs[i].Address == addrs[i-1].Address {
  376. // Skip duplicates
  377. continue
  378. }
  379. if addrs[i].Expires >= cutoff {
  380. naddrs = append(naddrs, addrs[i])
  381. }
  382. }
  383. if len(naddrs) == 0 {
  384. return nil
  385. }
  386. return naddrs
  387. }
  388. func Cmp(d, other *discosrv.DatabaseAddress) (n int) {
  389. if c := cmp.Compare(d.Address, other.Address); c != 0 {
  390. return c
  391. }
  392. return cmp.Compare(d.Expires, other.Expires)
  393. }
  394. func Equal(d, other *discosrv.DatabaseAddress) bool {
  395. return d.Address == other.Address
  396. }