mirror of
https://github.com/ivuorinen/gibidify.git
synced 2026-01-26 03:24:05 +00:00
Add overflow checks before converting uint64 memory values to int64 to prevent potential integer overflow issues identified by gosec (G115). - Add math.MaxInt64 checks in fileproc/backpressure.go - Add math.MaxInt64 checks in fileproc/resource_monitor_validation.go - Add math.MaxInt64 checks in fileproc/resource_monitor_metrics.go - Add math.MaxInt64 check in benchmark/benchmark.go with nosec annotation Co-authored-by: ivuorinen <11024+ivuorinen@users.noreply.github.com>
96 lines
2.5 KiB
Go
96 lines
2.5 KiB
Go
package fileproc
|
|
|
|
import (
|
|
"context"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/spf13/viper"
|
|
|
|
"github.com/ivuorinen/gibidify/testutil"
|
|
)
|
|
|
|
func TestResourceMonitor_ConcurrentReadsLimit(t *testing.T) {
|
|
testutil.ResetViperConfig(t, "")
|
|
|
|
// Set a low concurrent reads limit for testing
|
|
viper.Set("resourceLimits.enabled", true)
|
|
viper.Set("resourceLimits.maxConcurrentReads", 2)
|
|
|
|
rm := NewResourceMonitor()
|
|
defer rm.Close()
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 100*time.Millisecond)
|
|
defer cancel()
|
|
|
|
// First read slot should succeed
|
|
err := rm.AcquireReadSlot(ctx)
|
|
if err != nil {
|
|
t.Errorf("Expected no error for first read slot, got %v", err)
|
|
}
|
|
|
|
// Second read slot should succeed
|
|
err = rm.AcquireReadSlot(ctx)
|
|
if err != nil {
|
|
t.Errorf("Expected no error for second read slot, got %v", err)
|
|
}
|
|
|
|
// Third read slot should timeout (context deadline exceeded)
|
|
err = rm.AcquireReadSlot(ctx)
|
|
if err == nil {
|
|
t.Error("Expected timeout error for third read slot, got nil")
|
|
}
|
|
|
|
// Release one slot and try again
|
|
rm.ReleaseReadSlot()
|
|
|
|
// Create new context for the next attempt
|
|
ctx2, cancel2 := context.WithTimeout(context.Background(), 100*time.Millisecond)
|
|
defer cancel2()
|
|
|
|
err = rm.AcquireReadSlot(ctx2)
|
|
if err != nil {
|
|
t.Errorf("Expected no error after releasing a slot, got %v", err)
|
|
}
|
|
|
|
// Clean up remaining slots
|
|
rm.ReleaseReadSlot()
|
|
rm.ReleaseReadSlot()
|
|
}
|
|
|
|
func TestResourceMonitor_TimeoutContexts(t *testing.T) {
|
|
testutil.ResetViperConfig(t, "")
|
|
|
|
// Set short timeouts for testing
|
|
viper.Set("resourceLimits.enabled", true)
|
|
viper.Set("resourceLimits.fileProcessingTimeoutSec", 1) // 1 second
|
|
viper.Set("resourceLimits.overallTimeoutSec", 2) // 2 seconds
|
|
|
|
rm := NewResourceMonitor()
|
|
defer rm.Close()
|
|
|
|
parentCtx := context.Background()
|
|
|
|
// Test file processing context
|
|
fileCtx, fileCancel := rm.CreateFileProcessingContext(parentCtx)
|
|
defer fileCancel()
|
|
|
|
deadline, ok := fileCtx.Deadline()
|
|
if !ok {
|
|
t.Error("Expected file processing context to have a deadline")
|
|
} else if time.Until(deadline) > time.Second+100*time.Millisecond {
|
|
t.Error("File processing timeout appears to be too long")
|
|
}
|
|
|
|
// Test overall processing context
|
|
overallCtx, overallCancel := rm.CreateOverallProcessingContext(parentCtx)
|
|
defer overallCancel()
|
|
|
|
deadline, ok = overallCtx.Deadline()
|
|
if !ok {
|
|
t.Error("Expected overall processing context to have a deadline")
|
|
} else if time.Until(deadline) > 2*time.Second+100*time.Millisecond {
|
|
t.Error("Overall processing timeout appears to be too long")
|
|
}
|
|
}
|