mirror of
https://github.com/coder/coder.git
synced 2025-07-03 16:13:58 +00:00
chore: ensure proper rbac permissions on 'Acquire' file in the cache (#18348)
The file cache was caching the `Unauthorized` errors if a user without the right perms opened the file first. So all future opens would fail. Now the cache always opens with a subject that can read files. And authz is checked on the Acquire per user.
This commit is contained in:
266
coderd/files/cache_test.go
Normal file
266
coderd/files/cache_test.go
Normal file
@ -0,0 +1,266 @@
|
||||
package files_test
|
||||
|
||||
import (
|
||||
"context"
|
||||
"sync/atomic"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/google/uuid"
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
"github.com/spf13/afero"
|
||||
"github.com/stretchr/testify/require"
|
||||
"golang.org/x/sync/errgroup"
|
||||
|
||||
"cdr.dev/slog/sloggers/slogtest"
|
||||
"github.com/coder/coder/v2/coderd/coderdtest"
|
||||
"github.com/coder/coder/v2/coderd/coderdtest/promhelp"
|
||||
"github.com/coder/coder/v2/coderd/database"
|
||||
"github.com/coder/coder/v2/coderd/database/dbauthz"
|
||||
"github.com/coder/coder/v2/coderd/database/dbgen"
|
||||
"github.com/coder/coder/v2/coderd/database/dbtestutil"
|
||||
"github.com/coder/coder/v2/coderd/files"
|
||||
"github.com/coder/coder/v2/coderd/rbac"
|
||||
"github.com/coder/coder/v2/coderd/rbac/policy"
|
||||
"github.com/coder/coder/v2/testutil"
|
||||
)
|
||||
|
||||
// nolint:paralleltest,tparallel // Serially testing is easier
|
||||
func TestCacheRBAC(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
db, cache, rec := cacheAuthzSetup(t)
|
||||
ctx := testutil.Context(t, testutil.WaitMedium)
|
||||
|
||||
file := dbgen.File(t, db, database.File{})
|
||||
|
||||
nobodyID := uuid.New()
|
||||
nobody := dbauthz.As(ctx, rbac.Subject{
|
||||
ID: nobodyID.String(),
|
||||
Roles: rbac.Roles{},
|
||||
Scope: rbac.ScopeAll,
|
||||
})
|
||||
|
||||
userID := uuid.New()
|
||||
userReader := dbauthz.As(ctx, rbac.Subject{
|
||||
ID: userID.String(),
|
||||
Roles: rbac.Roles{
|
||||
must(rbac.RoleByName(rbac.RoleTemplateAdmin())),
|
||||
},
|
||||
Scope: rbac.ScopeAll,
|
||||
})
|
||||
|
||||
//nolint:gocritic // Unit testing
|
||||
cacheReader := dbauthz.AsFileReader(ctx)
|
||||
|
||||
t.Run("NoRolesOpen", func(t *testing.T) {
|
||||
// Ensure start is clean
|
||||
require.Equal(t, 0, cache.Count())
|
||||
rec.Reset()
|
||||
|
||||
_, err := cache.Acquire(nobody, file.ID)
|
||||
require.Error(t, err)
|
||||
require.True(t, rbac.IsUnauthorizedError(err))
|
||||
|
||||
// Ensure that the cache is empty
|
||||
require.Equal(t, 0, cache.Count())
|
||||
|
||||
// Check the assertions
|
||||
rec.AssertActorID(t, nobodyID.String(), rec.Pair(policy.ActionRead, file))
|
||||
rec.AssertActorID(t, rbac.SubjectTypeFileReaderID, rec.Pair(policy.ActionRead, file))
|
||||
})
|
||||
|
||||
t.Run("CacheHasFile", func(t *testing.T) {
|
||||
rec.Reset()
|
||||
require.Equal(t, 0, cache.Count())
|
||||
|
||||
// Read the file with a file reader to put it into the cache.
|
||||
_, err := cache.Acquire(cacheReader, file.ID)
|
||||
require.NoError(t, err)
|
||||
require.Equal(t, 1, cache.Count())
|
||||
|
||||
// "nobody" should not be able to read the file.
|
||||
_, err = cache.Acquire(nobody, file.ID)
|
||||
require.Error(t, err)
|
||||
require.True(t, rbac.IsUnauthorizedError(err))
|
||||
require.Equal(t, 1, cache.Count())
|
||||
|
||||
// UserReader can
|
||||
_, err = cache.Acquire(userReader, file.ID)
|
||||
require.NoError(t, err)
|
||||
require.Equal(t, 1, cache.Count())
|
||||
|
||||
cache.Release(file.ID)
|
||||
cache.Release(file.ID)
|
||||
require.Equal(t, 0, cache.Count())
|
||||
|
||||
rec.AssertActorID(t, nobodyID.String(), rec.Pair(policy.ActionRead, file))
|
||||
rec.AssertActorID(t, rbac.SubjectTypeFileReaderID, rec.Pair(policy.ActionRead, file))
|
||||
rec.AssertActorID(t, userID.String(), rec.Pair(policy.ActionRead, file))
|
||||
})
|
||||
}
|
||||
|
||||
func cachePromMetricName(metric string) string {
|
||||
return "coderd_file_cache_" + metric
|
||||
}
|
||||
|
||||
func TestConcurrency(t *testing.T) {
|
||||
t.Parallel()
|
||||
//nolint:gocritic // Unit testing
|
||||
ctx := dbauthz.AsFileReader(t.Context())
|
||||
|
||||
const fileSize = 10
|
||||
emptyFS := afero.NewIOFS(afero.NewReadOnlyFs(afero.NewMemMapFs()))
|
||||
var fetches atomic.Int64
|
||||
reg := prometheus.NewRegistry()
|
||||
c := files.New(func(_ context.Context, _ uuid.UUID) (files.CacheEntryValue, error) {
|
||||
fetches.Add(1)
|
||||
// Wait long enough before returning to make sure that all of the goroutines
|
||||
// will be waiting in line, ensuring that no one duplicated a fetch.
|
||||
time.Sleep(testutil.IntervalMedium)
|
||||
return files.CacheEntryValue{FS: emptyFS, Size: fileSize}, nil
|
||||
}, reg, &coderdtest.FakeAuthorizer{})
|
||||
|
||||
batches := 1000
|
||||
groups := make([]*errgroup.Group, 0, batches)
|
||||
for range batches {
|
||||
groups = append(groups, new(errgroup.Group))
|
||||
}
|
||||
|
||||
// Call Acquire with a unique ID per batch, many times per batch, with many
|
||||
// batches all in parallel. This is pretty much the worst-case scenario:
|
||||
// thousands of concurrent reads, with both warm and cold loads happening.
|
||||
batchSize := 10
|
||||
for _, g := range groups {
|
||||
id := uuid.New()
|
||||
for range batchSize {
|
||||
g.Go(func() error {
|
||||
// We don't bother to Release these references because the Cache will be
|
||||
// released at the end of the test anyway.
|
||||
_, err := c.Acquire(ctx, id)
|
||||
return err
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
for _, g := range groups {
|
||||
require.NoError(t, g.Wait())
|
||||
}
|
||||
require.Equal(t, int64(batches), fetches.Load())
|
||||
|
||||
// Verify all the counts & metrics are correct.
|
||||
require.Equal(t, batches, c.Count())
|
||||
require.Equal(t, batches*fileSize, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_size_bytes_current"), nil))
|
||||
require.Equal(t, batches*fileSize, promhelp.CounterValue(t, reg, cachePromMetricName("open_files_size_bytes_total"), nil))
|
||||
require.Equal(t, batches, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_current"), nil))
|
||||
require.Equal(t, batches, promhelp.CounterValue(t, reg, cachePromMetricName("open_files_total"), nil))
|
||||
require.Equal(t, batches*batchSize, promhelp.GaugeValue(t, reg, cachePromMetricName("open_file_refs_current"), nil))
|
||||
require.Equal(t, batches*batchSize, promhelp.CounterValue(t, reg, cachePromMetricName("open_file_refs_total"), nil))
|
||||
}
|
||||
|
||||
func TestRelease(t *testing.T) {
|
||||
t.Parallel()
|
||||
//nolint:gocritic // Unit testing
|
||||
ctx := dbauthz.AsFileReader(t.Context())
|
||||
|
||||
const fileSize = 10
|
||||
emptyFS := afero.NewIOFS(afero.NewReadOnlyFs(afero.NewMemMapFs()))
|
||||
reg := prometheus.NewRegistry()
|
||||
c := files.New(func(_ context.Context, _ uuid.UUID) (files.CacheEntryValue, error) {
|
||||
return files.CacheEntryValue{
|
||||
FS: emptyFS,
|
||||
Size: fileSize,
|
||||
}, nil
|
||||
}, reg, &coderdtest.FakeAuthorizer{})
|
||||
|
||||
batches := 100
|
||||
ids := make([]uuid.UUID, 0, batches)
|
||||
for range batches {
|
||||
ids = append(ids, uuid.New())
|
||||
}
|
||||
|
||||
// Acquire a bunch of references
|
||||
batchSize := 10
|
||||
for openedIdx, id := range ids {
|
||||
for batchIdx := range batchSize {
|
||||
it, err := c.Acquire(ctx, id)
|
||||
require.NoError(t, err)
|
||||
require.Equal(t, emptyFS, it)
|
||||
|
||||
// Each time a new file is opened, the metrics should be updated as so:
|
||||
opened := openedIdx + 1
|
||||
// Number of unique files opened is equal to the idx of the ids.
|
||||
require.Equal(t, opened, c.Count())
|
||||
require.Equal(t, opened, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_current"), nil))
|
||||
// Current file size is unique files * file size.
|
||||
require.Equal(t, opened*fileSize, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_size_bytes_current"), nil))
|
||||
// The number of refs is the current iteration of both loops.
|
||||
require.Equal(t, ((opened-1)*batchSize)+(batchIdx+1), promhelp.GaugeValue(t, reg, cachePromMetricName("open_file_refs_current"), nil))
|
||||
}
|
||||
}
|
||||
|
||||
// Make sure cache is fully loaded
|
||||
require.Equal(t, c.Count(), batches)
|
||||
|
||||
// Now release all of the references
|
||||
for closedIdx, id := range ids {
|
||||
stillOpen := len(ids) - closedIdx
|
||||
for closingIdx := range batchSize {
|
||||
c.Release(id)
|
||||
|
||||
// Each time a file is released, the metrics should decrement the file refs
|
||||
require.Equal(t, (stillOpen*batchSize)-(closingIdx+1), promhelp.GaugeValue(t, reg, cachePromMetricName("open_file_refs_current"), nil))
|
||||
|
||||
closed := closingIdx+1 == batchSize
|
||||
if closed {
|
||||
continue
|
||||
}
|
||||
|
||||
// File ref still exists, so the counts should not change yet.
|
||||
require.Equal(t, stillOpen, c.Count())
|
||||
require.Equal(t, stillOpen, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_current"), nil))
|
||||
require.Equal(t, stillOpen*fileSize, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_size_bytes_current"), nil))
|
||||
}
|
||||
}
|
||||
|
||||
// ...and make sure that the cache has emptied itself.
|
||||
require.Equal(t, c.Count(), 0)
|
||||
|
||||
// Verify all the counts & metrics are correct.
|
||||
// All existing files are closed
|
||||
require.Equal(t, 0, c.Count())
|
||||
require.Equal(t, 0, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_size_bytes_current"), nil))
|
||||
require.Equal(t, 0, promhelp.GaugeValue(t, reg, cachePromMetricName("open_files_current"), nil))
|
||||
require.Equal(t, 0, promhelp.GaugeValue(t, reg, cachePromMetricName("open_file_refs_current"), nil))
|
||||
|
||||
// Total counts remain
|
||||
require.Equal(t, batches*fileSize, promhelp.CounterValue(t, reg, cachePromMetricName("open_files_size_bytes_total"), nil))
|
||||
require.Equal(t, batches, promhelp.CounterValue(t, reg, cachePromMetricName("open_files_total"), nil))
|
||||
require.Equal(t, batches*batchSize, promhelp.CounterValue(t, reg, cachePromMetricName("open_file_refs_total"), nil))
|
||||
}
|
||||
|
||||
func cacheAuthzSetup(t *testing.T) (database.Store, *files.Cache, *coderdtest.RecordingAuthorizer) {
|
||||
t.Helper()
|
||||
|
||||
logger := slogtest.Make(t, &slogtest.Options{})
|
||||
reg := prometheus.NewRegistry()
|
||||
|
||||
db, _ := dbtestutil.NewDB(t)
|
||||
authz := rbac.NewAuthorizer(reg)
|
||||
rec := &coderdtest.RecordingAuthorizer{
|
||||
Called: nil,
|
||||
Wrapped: authz,
|
||||
}
|
||||
|
||||
// Dbauthz wrap the db
|
||||
db = dbauthz.New(db, rec, logger, coderdtest.AccessControlStorePointer())
|
||||
c := files.NewFromStore(db, reg, rec)
|
||||
return db, c, rec
|
||||
}
|
||||
|
||||
func must[T any](t T, err error) T {
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
return t
|
||||
}
|
Reference in New Issue
Block a user