Implements Phase 1 part 2 of ADR-0022 (Rate Limiting and Cache Strategy): in-memory cache service using github.com/patrickmn/go-cache. Wired onto Server struct and used by handleVersion to memoize the response for 60 seconds. Companion to PR #22 (per-IP rate limit middleware). Changes: - New: pkg/cache/cache.go (58 lines, Service interface + InMemoryService) - New: pkg/cache/cache_test.go (125 lines, 6 unit tests, all passing) - Modified: pkg/config/config.go (CacheConfig struct + 3 SetDefault + 3 BindEnv + 3 getters) - Modified: pkg/server/server.go (cacheService field + init in NewServer + use in handleVersion) - Modified: config.yaml (cache section with defaults) - go.mod / go.sum (github.com/patrickmn/go-cache v2.1.0+incompatible) Closes #13 (Phase 1 fully complete: rate limit in PR #22, cache here). Phase 2 (Redis-compatible shared cache via Dragonfly/KeyDB) deferred. BDD scenario not added: cache hit is hard to test via the existing testserver (same architectural limitation as the rate limit BDD - testserver pre-started, env vars don't propagate). Behavior is fully covered by unit tests (6/6 PASS). TODO: BDD scenario can be added once testserver supports per-scenario config. Generated ~95% in autonomy by Mistral Vibe via ICM workspace ~/Work/Vibe/workspaces/cache-service-inmemory/. T6 cost €2.50 for stages 01-02 (50% reduction vs T5, thanks to pre-extracted snippets in shared/). Trainer (Claude) finalized commit/PR (Mistral hit max-turns). 🤖 Co-Authored-By: Mistral Vibe (devstral-2 / mistral-medium-3.5) Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
57 lines
1.5 KiB
Go
57 lines
1.5 KiB
Go
package cache
|
|
|
|
import (
|
|
"time"
|
|
|
|
gocache "github.com/patrickmn/go-cache"
|
|
)
|
|
|
|
// Service defines the interface for cache operations
|
|
type Service interface {
|
|
Set(key string, value interface{}, ttl time.Duration)
|
|
Get(key string) (interface{}, bool)
|
|
Delete(key string)
|
|
Flush()
|
|
ItemCount() int
|
|
}
|
|
|
|
// InMemoryService implements Service using go-cache library
|
|
type InMemoryService struct {
|
|
cache *gocache.Cache
|
|
}
|
|
|
|
// NewInMemoryService creates a new in-memory cache service
|
|
// defaultTTL: default time-to-live for cache items
|
|
// cleanupInterval: interval at which expired items are cleaned up
|
|
func NewInMemoryService(defaultTTL, cleanupInterval time.Duration) Service {
|
|
c := gocache.New(defaultTTL, cleanupInterval)
|
|
return &InMemoryService{cache: c}
|
|
}
|
|
|
|
// Set stores a value in the cache with the specified TTL
|
|
func (s *InMemoryService) Set(key string, value interface{}, ttl time.Duration) {
|
|
s.cache.Set(key, value, ttl)
|
|
}
|
|
|
|
// Get retrieves a value from the cache
|
|
// Returns the value and true if found, nil and false if not found or expired
|
|
func (s *InMemoryService) Get(key string) (interface{}, bool) {
|
|
val, found := s.cache.Get(key)
|
|
return val, found
|
|
}
|
|
|
|
// Delete removes an item from the cache
|
|
func (s *InMemoryService) Delete(key string) {
|
|
s.cache.Delete(key)
|
|
}
|
|
|
|
// Flush clears all items from the cache
|
|
func (s *InMemoryService) Flush() {
|
|
s.cache.Flush()
|
|
}
|
|
|
|
// ItemCount returns the number of items currently in the cache
|
|
func (s *InMemoryService) ItemCount() int {
|
|
return s.cache.ItemCount()
|
|
}
|