You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
192 lines
4.7 KiB
192 lines
4.7 KiB
package cache
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"time"
|
|
|
|
"github.com/OpenIMSDK/Open-IM-Server/pkg/common/log"
|
|
"github.com/OpenIMSDK/Open-IM-Server/pkg/errs"
|
|
"github.com/OpenIMSDK/Open-IM-Server/pkg/utils"
|
|
"github.com/dtm-labs/rockscache"
|
|
)
|
|
|
|
const (
|
|
scanCount = 3000
|
|
maxRetryTimes = 5
|
|
retryInterval = time.Millisecond * 100
|
|
)
|
|
|
|
var errIndex = errors.New("err index")
|
|
|
|
type metaCache interface {
|
|
ExecDel(ctx context.Context) error
|
|
// delete key rapid
|
|
DelKey(ctx context.Context, key string) error
|
|
AddKeys(keys ...string)
|
|
ClearKeys()
|
|
GetPreDelKeys() []string
|
|
}
|
|
|
|
func NewMetaCacheRedis(rcClient *rockscache.Client, keys ...string) metaCache {
|
|
return &metaCacheRedis{rcClient: rcClient, keys: keys, maxRetryTimes: maxRetryTimes, retryInterval: retryInterval}
|
|
}
|
|
|
|
type metaCacheRedis struct {
|
|
rcClient *rockscache.Client
|
|
keys []string
|
|
maxRetryTimes int
|
|
retryInterval time.Duration
|
|
}
|
|
|
|
func (m *metaCacheRedis) ExecDel(ctx context.Context) error {
|
|
if len(m.keys) > 0 {
|
|
log.ZDebug(ctx, "delete cache", "keys", m.keys)
|
|
retryTimes := 0
|
|
for {
|
|
if err := m.rcClient.TagAsDeletedBatch2(ctx, m.keys); err != nil {
|
|
if retryTimes >= m.maxRetryTimes {
|
|
err = errs.ErrInternalServer.Wrap(fmt.Sprintf("delete cache error: %v, keys: %v, retry times %d, please check redis server", err, m.keys, retryTimes))
|
|
log.ZWarn(ctx, "delete cache failed, please handle keys", err, "keys", m.keys)
|
|
return err
|
|
}
|
|
retryTimes++
|
|
} else {
|
|
break
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (m *metaCacheRedis) DelKey(ctx context.Context, key string) error {
|
|
return m.rcClient.TagAsDeleted2(ctx, key)
|
|
}
|
|
|
|
func (m *metaCacheRedis) AddKeys(keys ...string) {
|
|
m.keys = append(m.keys, keys...)
|
|
}
|
|
|
|
func (m *metaCacheRedis) ClearKeys() {
|
|
m.keys = []string{}
|
|
}
|
|
|
|
func (m *metaCacheRedis) GetPreDelKeys() []string {
|
|
return m.keys
|
|
}
|
|
|
|
func GetDefaultOpt() rockscache.Options {
|
|
opts := rockscache.NewDefaultOptions()
|
|
opts.StrongConsistency = true
|
|
opts.RandomExpireAdjustment = 0.2
|
|
return opts
|
|
}
|
|
|
|
func getCache[T any](ctx context.Context, rcClient *rockscache.Client, key string, expire time.Duration, fn func(ctx context.Context) (T, error)) (T, error) {
|
|
var t T
|
|
var write bool
|
|
v, err := rcClient.Fetch2(ctx, key, expire, func() (s string, err error) {
|
|
t, err = fn(ctx)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
bs, err := json.Marshal(t)
|
|
if err != nil {
|
|
return "", utils.Wrap(err, "")
|
|
}
|
|
write = true
|
|
return string(bs), nil
|
|
})
|
|
if err != nil {
|
|
return t, err
|
|
}
|
|
if write {
|
|
return t, nil
|
|
}
|
|
if v == "" {
|
|
return t, errs.ErrRecordNotFound.Wrap("cache is not found")
|
|
}
|
|
err = json.Unmarshal([]byte(v), &t)
|
|
if err != nil {
|
|
log.ZError(ctx, "cache json.Unmarshal failed", err, "key", key, "value", v, "expire", expire)
|
|
return t, utils.Wrap(err, "")
|
|
}
|
|
return t, nil
|
|
}
|
|
|
|
func batchGetCache[T any](ctx context.Context, rcClient *rockscache.Client, keys []string, expire time.Duration, keyIndexFn func(t T, keys []string) (int, error), fn func(ctx context.Context) ([]T, error)) ([]T, error) {
|
|
batchMap, err := rcClient.FetchBatch2(ctx, keys, expire, func(idxs []int) (m map[int]string, err error) {
|
|
values := make(map[int]string)
|
|
tArrays, err := fn(ctx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
for _, v := range tArrays {
|
|
index, err := keyIndexFn(v, keys)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
bs, err := json.Marshal(v)
|
|
if err != nil {
|
|
return nil, utils.Wrap(err, "marshal failed")
|
|
}
|
|
values[index] = string(bs)
|
|
}
|
|
return values, nil
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
var tArrays []T
|
|
for _, v := range batchMap {
|
|
if v != "" {
|
|
var t T
|
|
err = json.Unmarshal([]byte(v), &t)
|
|
if err != nil {
|
|
return nil, utils.Wrap(err, "unmarshal failed")
|
|
}
|
|
tArrays = append(tArrays, t)
|
|
}
|
|
}
|
|
return tArrays, nil
|
|
}
|
|
|
|
func batchGetCacheMap[T any](ctx context.Context, rcClient *rockscache.Client, keys, originKeys []string, expire time.Duration, keyIndexFn func(s string, keys []string) (int, error), fn func(ctx context.Context) (map[string]T, error)) (map[string]T, error) {
|
|
batchMap, err := rcClient.FetchBatch2(ctx, keys, expire, func(idxs []int) (m map[int]string, err error) {
|
|
tArrays, err := fn(ctx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
values := make(map[int]string)
|
|
for k, v := range tArrays {
|
|
index, err := keyIndexFn(k, originKeys)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
bs, err := json.Marshal(v)
|
|
if err != nil {
|
|
return nil, utils.Wrap(err, "marshal failed")
|
|
}
|
|
values[index] = string(bs)
|
|
}
|
|
return values, nil
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
tMap := make(map[string]T)
|
|
for i, v := range batchMap {
|
|
if v != "" {
|
|
var t T
|
|
err = json.Unmarshal([]byte(v), &t)
|
|
if err != nil {
|
|
return nil, utils.Wrap(err, "unmarshal failed")
|
|
}
|
|
tMap[originKeys[i]] = t
|
|
}
|
|
}
|
|
return tMap, nil
|
|
}
|