All checks were successful
Build and Release / release (push) Successful in 1m49s
242 lines
6.2 KiB
Go
242 lines
6.2 KiB
Go
package repositories
|
|
|
|
import (
|
|
"context"
|
|
"crypto/md5"
|
|
"encoding/json"
|
|
"fmt"
|
|
"history-api/internal/gen/sqlc"
|
|
"history-api/internal/models"
|
|
"history-api/pkg/cache"
|
|
"history-api/pkg/constants"
|
|
"history-api/pkg/convert"
|
|
"time"
|
|
|
|
"github.com/jackc/pgx/v5"
|
|
"github.com/jackc/pgx/v5/pgtype"
|
|
)
|
|
|
|
type StatisticRepository interface {
|
|
Search(ctx context.Context, params sqlc.SearchSystemStatisticsParams) ([]*models.StatisticEntity, error)
|
|
GetByDate(ctx context.Context, date time.Time) (*models.StatisticEntity, error)
|
|
GetByID(ctx context.Context, id pgtype.UUID) (*models.StatisticEntity, error)
|
|
Upsert(ctx context.Context, date time.Time) (*models.StatisticEntity, error)
|
|
WithTx(tx pgx.Tx) StatisticRepository
|
|
}
|
|
|
|
type statisticRepository struct {
|
|
q *sqlc.Queries
|
|
c cache.Cache
|
|
db sqlc.DBTX
|
|
}
|
|
|
|
func NewStatisticRepository(db sqlc.DBTX, c cache.Cache) StatisticRepository {
|
|
return &statisticRepository{
|
|
q: sqlc.New(db),
|
|
c: c,
|
|
db: db,
|
|
}
|
|
}
|
|
|
|
func (r *statisticRepository) WithTx(tx pgx.Tx) StatisticRepository {
|
|
return &statisticRepository{
|
|
q: r.q.WithTx(tx),
|
|
c: r.c,
|
|
db: tx,
|
|
}
|
|
}
|
|
|
|
func (r *statisticRepository) generateQueryKey(prefix string, params any) string {
|
|
b, _ := json.Marshal(params)
|
|
hash := fmt.Sprintf("%x", md5.Sum(b))
|
|
return fmt.Sprintf("%s:%s", prefix, hash)
|
|
}
|
|
|
|
func mapToEntity(row sqlc.SystemStatistic) *models.StatisticEntity {
|
|
return &models.StatisticEntity{
|
|
ID: convert.UUIDToString(row.ID),
|
|
Date: row.Date.Time,
|
|
TotalUsers: row.TotalUsers,
|
|
TotalProjects: row.TotalProjects,
|
|
TotalCommits: row.TotalCommits,
|
|
TotalSubmissions: row.TotalSubmissions,
|
|
TotalMedias: row.TotalMedias,
|
|
TotalWikis: row.TotalWikis,
|
|
TotalEntities: row.TotalEntities,
|
|
TotalGeometries: row.TotalGeometries,
|
|
TotalStorageBytes: row.TotalStorageBytes,
|
|
NewUsers: row.NewUsers,
|
|
NewProjects: row.NewProjects,
|
|
NewCommits: row.NewCommits,
|
|
NewSubmissions: row.NewSubmissions,
|
|
NewMedias: row.NewMedias,
|
|
NewWikis: row.NewWikis,
|
|
NewEntities: row.NewEntities,
|
|
NewGeometries: row.NewGeometries,
|
|
NewStorageBytes: row.NewStorageBytes,
|
|
CreatedAt: convert.TimeToPtr(row.CreatedAt),
|
|
}
|
|
}
|
|
|
|
func (r *statisticRepository) getByIDsWithFallback(ctx context.Context, ids []string) ([]*models.StatisticEntity, error) {
|
|
if len(ids) == 0 {
|
|
return []*models.StatisticEntity{}, nil
|
|
}
|
|
keys := make([]string, len(ids))
|
|
for i, id := range ids {
|
|
keys[i] = fmt.Sprintf("statistic:id:%s", id)
|
|
}
|
|
raws := r.c.MGet(ctx, keys...)
|
|
|
|
var stats []*models.StatisticEntity
|
|
missingStatsToCache := make(map[string]any)
|
|
|
|
var missingPgIds []pgtype.UUID
|
|
for i, b := range raws {
|
|
if len(b) == 0 {
|
|
pgId := pgtype.UUID{}
|
|
err := pgId.Scan(ids[i])
|
|
if err == nil {
|
|
missingPgIds = append(missingPgIds, pgId)
|
|
}
|
|
}
|
|
}
|
|
|
|
dbMap := make(map[string]*models.StatisticEntity)
|
|
if len(missingPgIds) > 0 {
|
|
dbRows, err := r.q.GetSystemStatisticsByIDs(ctx, missingPgIds)
|
|
if err == nil {
|
|
for _, row := range dbRows {
|
|
entity := mapToEntity(row)
|
|
dbMap[entity.ID] = entity
|
|
}
|
|
}
|
|
}
|
|
|
|
for i, b := range raws {
|
|
if len(b) > 0 {
|
|
var s models.StatisticEntity
|
|
if err := json.Unmarshal(b, &s); err == nil {
|
|
stats = append(stats, &s)
|
|
}
|
|
} else {
|
|
if item, ok := dbMap[ids[i]]; ok {
|
|
stats = append(stats, item)
|
|
missingStatsToCache[keys[i]] = item
|
|
}
|
|
}
|
|
}
|
|
|
|
if len(missingStatsToCache) > 0 {
|
|
_ = r.c.MSet(ctx, missingStatsToCache, constants.NormalCacheDuration)
|
|
}
|
|
|
|
return stats, nil
|
|
}
|
|
|
|
func (r *statisticRepository) Search(ctx context.Context, params sqlc.SearchSystemStatisticsParams) ([]*models.StatisticEntity, error) {
|
|
queryKey := r.generateQueryKey("statistic:search", params)
|
|
|
|
var cachedIDs []string
|
|
if err := r.c.Get(ctx, queryKey, &cachedIDs); err == nil && len(cachedIDs) > 0 {
|
|
return r.getByIDsWithFallback(ctx, cachedIDs)
|
|
}
|
|
|
|
rows, err := r.q.SearchSystemStatistics(ctx, params)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var ids []string
|
|
statsToCache := make(map[string]any)
|
|
var stats []*models.StatisticEntity
|
|
|
|
for _, row := range rows {
|
|
entity := mapToEntity(row)
|
|
ids = append(ids, entity.ID)
|
|
stats = append(stats, entity)
|
|
statsToCache[fmt.Sprintf("statistic:id:%s", entity.ID)] = entity
|
|
}
|
|
|
|
if len(statsToCache) > 0 {
|
|
_ = r.c.MSet(ctx, statsToCache, constants.NormalCacheDuration)
|
|
}
|
|
if len(ids) > 0 {
|
|
_ = r.c.Set(ctx, queryKey, ids, constants.ListCacheDuration)
|
|
}
|
|
|
|
return stats, nil
|
|
}
|
|
|
|
func (r *statisticRepository) GetByID(ctx context.Context, id pgtype.UUID) (*models.StatisticEntity, error) {
|
|
cacheId := fmt.Sprintf("statistic:id:%s", convert.UUIDToString(id))
|
|
var stat models.StatisticEntity
|
|
err := r.c.Get(ctx, cacheId, &stat)
|
|
if err == nil {
|
|
return &stat, nil
|
|
}
|
|
|
|
rows, err := r.q.GetSystemStatisticsByIDs(ctx, []pgtype.UUID{id})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if len(rows) == 0 {
|
|
return nil, nil
|
|
}
|
|
|
|
entity := mapToEntity(rows[0])
|
|
_ = r.c.Set(ctx, cacheId, entity, constants.NormalCacheDuration)
|
|
|
|
return entity, nil
|
|
}
|
|
|
|
func (r *statisticRepository) GetByDate(ctx context.Context, date time.Time) (*models.StatisticEntity, error) {
|
|
dateStr := date.Format("2006-01-02")
|
|
cacheId := fmt.Sprintf("statistic:date:%s", dateStr)
|
|
|
|
var stat models.StatisticEntity
|
|
err := r.c.Get(ctx, cacheId, &stat)
|
|
if err == nil {
|
|
return &stat, nil
|
|
}
|
|
|
|
row, err := r.q.GetSystemStatisticsByDate(ctx, pgtype.Date{Time: date, Valid: true})
|
|
if err != nil {
|
|
if err == pgx.ErrNoRows {
|
|
return nil, nil
|
|
}
|
|
return nil, err
|
|
}
|
|
|
|
entity := mapToEntity(row)
|
|
|
|
_ = r.c.Set(ctx, cacheId, entity, constants.NormalCacheDuration)
|
|
_ = r.c.Set(ctx, fmt.Sprintf("statistic:id:%s", entity.ID), entity, constants.NormalCacheDuration)
|
|
|
|
return entity, nil
|
|
}
|
|
|
|
func (r *statisticRepository) Upsert(ctx context.Context, date time.Time) (*models.StatisticEntity, error) {
|
|
row, err := r.q.UpsertSystemStatistics(ctx, pgtype.Date{Time: date, Valid: true})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
entity := mapToEntity(row)
|
|
|
|
// Clear search cache and the specific date cache
|
|
go func() {
|
|
bgCtx := context.Background()
|
|
_ = r.c.DelByPattern(bgCtx, "statistic:search*")
|
|
_ = r.c.Del(
|
|
bgCtx,
|
|
fmt.Sprintf("statistic:id:%s", entity.ID),
|
|
fmt.Sprintf("statistic:date:%s", date.Format("2006-01-02")),
|
|
)
|
|
}()
|
|
|
|
return entity, nil
|
|
}
|
|
|
|
|