mirror of
https://gitee.com/wanwujie/sub2api
synced 2026-04-09 09:34:46 +08:00
When default failover retries are exhausted, continue retrying with Antigravity accounts only (up to 10 times, configurable via GATEWAY_ANTIGRAVITY_EXTRA_RETRIES). Each extra retry uses a fixed 500ms delay. Non-Antigravity accounts are skipped during the extra retry phase. Applied to all three endpoints: Gemini compat, Claude, and Gemini native API paths.
418 lines
12 KiB
Go
418 lines
12 KiB
Go
//go:build unit
|
|
|
|
package handler
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"net/http"
|
|
"net/http/httptest"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/Wei-Shaw/sub2api/internal/config"
|
|
"github.com/Wei-Shaw/sub2api/internal/service"
|
|
"github.com/gin-gonic/gin"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
// --- sleepFixedDelay ---
|
|
|
|
func TestSleepFixedDelay_ZeroDelay(t *testing.T) {
|
|
got := sleepFixedDelay(context.Background(), 0)
|
|
require.True(t, got, "zero delay should return true immediately")
|
|
}
|
|
|
|
func TestSleepFixedDelay_NegativeDelay(t *testing.T) {
|
|
got := sleepFixedDelay(context.Background(), -1*time.Second)
|
|
require.True(t, got, "negative delay should return true immediately")
|
|
}
|
|
|
|
func TestSleepFixedDelay_NormalDelay(t *testing.T) {
|
|
start := time.Now()
|
|
got := sleepFixedDelay(context.Background(), 50*time.Millisecond)
|
|
elapsed := time.Since(start)
|
|
require.True(t, got, "normal delay should return true")
|
|
require.GreaterOrEqual(t, elapsed, 40*time.Millisecond, "should sleep at least ~50ms")
|
|
}
|
|
|
|
func TestSleepFixedDelay_ContextCancelled(t *testing.T) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
cancel() // cancel immediately
|
|
got := sleepFixedDelay(ctx, 10*time.Second)
|
|
require.False(t, got, "cancelled context should return false")
|
|
}
|
|
|
|
func TestSleepFixedDelay_ContextTimeout(t *testing.T) {
|
|
ctx, cancel := context.WithTimeout(context.Background(), 10*time.Millisecond)
|
|
defer cancel()
|
|
got := sleepFixedDelay(ctx, 5*time.Second)
|
|
require.False(t, got, "context timeout should return false before delay completes")
|
|
}
|
|
|
|
// --- antigravityExtraRetryDelay constant ---
|
|
|
|
func TestAntigravityExtraRetryDelayValue(t *testing.T) {
|
|
require.Equal(t, 500*time.Millisecond, antigravityExtraRetryDelay)
|
|
}
|
|
|
|
// --- NewGatewayHandler antigravityExtraRetries field ---
|
|
|
|
func TestNewGatewayHandler_AntigravityExtraRetries_Default(t *testing.T) {
|
|
h := NewGatewayHandler(nil, nil, nil, nil, nil, nil, nil, nil, nil, nil)
|
|
require.Equal(t, 10, h.antigravityExtraRetries, "default should be 10 when cfg is nil")
|
|
}
|
|
|
|
func TestNewGatewayHandler_AntigravityExtraRetries_FromConfig(t *testing.T) {
|
|
cfg := &config.Config{
|
|
Gateway: config.GatewayConfig{
|
|
AntigravityExtraRetries: 5,
|
|
},
|
|
}
|
|
h := NewGatewayHandler(nil, nil, nil, nil, nil, nil, nil, nil, nil, cfg)
|
|
require.Equal(t, 5, h.antigravityExtraRetries, "should use config value")
|
|
}
|
|
|
|
func TestNewGatewayHandler_AntigravityExtraRetries_ZeroDisables(t *testing.T) {
|
|
cfg := &config.Config{
|
|
Gateway: config.GatewayConfig{
|
|
AntigravityExtraRetries: 0,
|
|
},
|
|
}
|
|
h := NewGatewayHandler(nil, nil, nil, nil, nil, nil, nil, nil, nil, cfg)
|
|
require.Equal(t, 0, h.antigravityExtraRetries, "zero should disable extra retries")
|
|
}
|
|
|
|
// --- handleFailoverAllAccountsExhausted (renamed: using handleFailoverExhausted) ---
|
|
// We test the error response format helpers that the extra retry path uses.
|
|
|
|
func TestHandleFailoverExhausted_JSON(t *testing.T) {
|
|
gin.SetMode(gin.TestMode)
|
|
rec := httptest.NewRecorder()
|
|
c, _ := gin.CreateTestContext(rec)
|
|
|
|
h := &GatewayHandler{}
|
|
failoverErr := &service.UpstreamFailoverError{StatusCode: 429}
|
|
h.handleFailoverExhausted(c, failoverErr, service.PlatformAntigravity, false)
|
|
|
|
require.Equal(t, http.StatusTooManyRequests, rec.Code)
|
|
|
|
var body map[string]any
|
|
err := json.Unmarshal(rec.Body.Bytes(), &body)
|
|
require.NoError(t, err)
|
|
errObj, ok := body["error"].(map[string]any)
|
|
require.True(t, ok)
|
|
require.Equal(t, "rate_limit_error", errObj["type"])
|
|
}
|
|
|
|
func TestHandleFailoverExhaustedSimple_JSON(t *testing.T) {
|
|
gin.SetMode(gin.TestMode)
|
|
rec := httptest.NewRecorder()
|
|
c, _ := gin.CreateTestContext(rec)
|
|
|
|
h := &GatewayHandler{}
|
|
h.handleFailoverExhaustedSimple(c, 502, false)
|
|
|
|
require.Equal(t, http.StatusBadGateway, rec.Code)
|
|
|
|
var body map[string]any
|
|
err := json.Unmarshal(rec.Body.Bytes(), &body)
|
|
require.NoError(t, err)
|
|
errObj, ok := body["error"].(map[string]any)
|
|
require.True(t, ok)
|
|
require.Equal(t, "upstream_error", errObj["type"])
|
|
}
|
|
|
|
// --- Extra retry platform filter logic ---
|
|
|
|
func TestExtraRetryPlatformFilter(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
switchCount int
|
|
maxAccountSwitch int
|
|
platform string
|
|
expectSkip bool
|
|
}{
|
|
{
|
|
name: "default_retry_phase_antigravity_not_skipped",
|
|
switchCount: 1,
|
|
maxAccountSwitch: 3,
|
|
platform: service.PlatformAntigravity,
|
|
expectSkip: false,
|
|
},
|
|
{
|
|
name: "default_retry_phase_gemini_not_skipped",
|
|
switchCount: 1,
|
|
maxAccountSwitch: 3,
|
|
platform: service.PlatformGemini,
|
|
expectSkip: false,
|
|
},
|
|
{
|
|
name: "extra_retry_phase_antigravity_not_skipped",
|
|
switchCount: 3,
|
|
maxAccountSwitch: 3,
|
|
platform: service.PlatformAntigravity,
|
|
expectSkip: false,
|
|
},
|
|
{
|
|
name: "extra_retry_phase_gemini_skipped",
|
|
switchCount: 3,
|
|
maxAccountSwitch: 3,
|
|
platform: service.PlatformGemini,
|
|
expectSkip: true,
|
|
},
|
|
{
|
|
name: "extra_retry_phase_anthropic_skipped",
|
|
switchCount: 3,
|
|
maxAccountSwitch: 3,
|
|
platform: service.PlatformAnthropic,
|
|
expectSkip: true,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
// Replicate the filter condition from the handler
|
|
shouldSkip := tt.switchCount >= tt.maxAccountSwitch && tt.platform != service.PlatformAntigravity
|
|
require.Equal(t, tt.expectSkip, shouldSkip)
|
|
})
|
|
}
|
|
}
|
|
|
|
// --- Extra retry counter logic ---
|
|
|
|
func TestExtraRetryCounterExhaustion(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
maxExtraRetries int
|
|
currentExtraCount int
|
|
expectExhausted bool
|
|
}{
|
|
{
|
|
name: "first_extra_retry",
|
|
maxExtraRetries: 10,
|
|
currentExtraCount: 1,
|
|
expectExhausted: false,
|
|
},
|
|
{
|
|
name: "at_limit",
|
|
maxExtraRetries: 10,
|
|
currentExtraCount: 10,
|
|
expectExhausted: false,
|
|
},
|
|
{
|
|
name: "exceeds_limit",
|
|
maxExtraRetries: 10,
|
|
currentExtraCount: 11,
|
|
expectExhausted: true,
|
|
},
|
|
{
|
|
name: "zero_disables_extra_retry",
|
|
maxExtraRetries: 0,
|
|
currentExtraCount: 1,
|
|
expectExhausted: true,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
// Replicate the exhaustion condition: antigravityExtraCount > h.antigravityExtraRetries
|
|
exhausted := tt.currentExtraCount > tt.maxExtraRetries
|
|
require.Equal(t, tt.expectExhausted, exhausted)
|
|
})
|
|
}
|
|
}
|
|
|
|
// --- mapUpstreamError (used by handleFailoverExhausted) ---
|
|
|
|
func TestMapUpstreamError(t *testing.T) {
|
|
h := &GatewayHandler{}
|
|
tests := []struct {
|
|
name string
|
|
statusCode int
|
|
expectedStatus int
|
|
expectedType string
|
|
}{
|
|
{"429", 429, http.StatusTooManyRequests, "rate_limit_error"},
|
|
{"529", 529, http.StatusServiceUnavailable, "overloaded_error"},
|
|
{"500", 500, http.StatusBadGateway, "upstream_error"},
|
|
{"502", 502, http.StatusBadGateway, "upstream_error"},
|
|
{"503", 503, http.StatusBadGateway, "upstream_error"},
|
|
{"504", 504, http.StatusBadGateway, "upstream_error"},
|
|
{"401", 401, http.StatusBadGateway, "upstream_error"},
|
|
{"403", 403, http.StatusBadGateway, "upstream_error"},
|
|
{"unknown", 418, http.StatusBadGateway, "upstream_error"},
|
|
}
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
status, errType, _ := h.mapUpstreamError(tt.statusCode)
|
|
require.Equal(t, tt.expectedStatus, status)
|
|
require.Equal(t, tt.expectedType, errType)
|
|
})
|
|
}
|
|
}
|
|
|
|
// --- Gemini native path: handleGeminiFailoverExhausted ---
|
|
|
|
func TestHandleGeminiFailoverExhausted_NilError(t *testing.T) {
|
|
gin.SetMode(gin.TestMode)
|
|
rec := httptest.NewRecorder()
|
|
c, _ := gin.CreateTestContext(rec)
|
|
|
|
h := &GatewayHandler{}
|
|
h.handleGeminiFailoverExhausted(c, nil)
|
|
|
|
require.Equal(t, http.StatusBadGateway, rec.Code)
|
|
var body map[string]any
|
|
err := json.Unmarshal(rec.Body.Bytes(), &body)
|
|
require.NoError(t, err)
|
|
errObj, ok := body["error"].(map[string]any)
|
|
require.True(t, ok)
|
|
require.Equal(t, "Upstream request failed", errObj["message"])
|
|
}
|
|
|
|
func TestHandleGeminiFailoverExhausted_429(t *testing.T) {
|
|
gin.SetMode(gin.TestMode)
|
|
rec := httptest.NewRecorder()
|
|
c, _ := gin.CreateTestContext(rec)
|
|
|
|
h := &GatewayHandler{}
|
|
failoverErr := &service.UpstreamFailoverError{StatusCode: 429}
|
|
h.handleGeminiFailoverExhausted(c, failoverErr)
|
|
|
|
require.Equal(t, http.StatusTooManyRequests, rec.Code)
|
|
}
|
|
|
|
// --- handleStreamingAwareError streaming mode ---
|
|
|
|
func TestHandleStreamingAwareError_StreamStarted(t *testing.T) {
|
|
gin.SetMode(gin.TestMode)
|
|
rec := httptest.NewRecorder()
|
|
c, _ := gin.CreateTestContext(rec)
|
|
|
|
// Simulate stream already started: set content type and write initial data
|
|
c.Writer.Header().Set("Content-Type", "text/event-stream")
|
|
c.Writer.WriteHeaderNow()
|
|
|
|
h := &GatewayHandler{}
|
|
h.handleStreamingAwareError(c, http.StatusTooManyRequests, "rate_limit_error", "test error", true)
|
|
|
|
body := rec.Body.String()
|
|
require.Contains(t, body, "rate_limit_error")
|
|
require.Contains(t, body, "test error")
|
|
require.Contains(t, body, "data: ")
|
|
}
|
|
|
|
func TestHandleStreamingAwareError_NotStreaming(t *testing.T) {
|
|
gin.SetMode(gin.TestMode)
|
|
rec := httptest.NewRecorder()
|
|
c, _ := gin.CreateTestContext(rec)
|
|
|
|
h := &GatewayHandler{}
|
|
h.handleStreamingAwareError(c, http.StatusServiceUnavailable, "api_error", "no model", false)
|
|
|
|
require.Equal(t, http.StatusServiceUnavailable, rec.Code)
|
|
var body map[string]any
|
|
err := json.Unmarshal(rec.Body.Bytes(), &body)
|
|
require.NoError(t, err)
|
|
errObj, ok := body["error"].(map[string]any)
|
|
require.True(t, ok)
|
|
require.Equal(t, "api_error", errObj["type"])
|
|
require.Equal(t, "no model", errObj["message"])
|
|
}
|
|
|
|
// --- Integration: extra retry flow simulation ---
|
|
|
|
func TestExtraRetryFlowSimulation(t *testing.T) {
|
|
// Simulate the full extra retry flow logic
|
|
maxAccountSwitches := 3
|
|
maxExtraRetries := 2
|
|
switchCount := 0
|
|
antigravityExtraCount := 0
|
|
|
|
type attempt struct {
|
|
platform string
|
|
isFailover bool
|
|
}
|
|
|
|
// Simulate: 3 default retries (all fail), then 2 extra retries (all fail), then exhausted
|
|
attempts := []attempt{
|
|
{service.PlatformAntigravity, true}, // switchCount 0 -> 1
|
|
{service.PlatformGemini, true}, // switchCount 1 -> 2
|
|
{service.PlatformAntigravity, true}, // switchCount 2 -> 3 (reaches max)
|
|
{service.PlatformAntigravity, true}, // extra retry 1
|
|
{service.PlatformAntigravity, true}, // extra retry 2
|
|
{service.PlatformAntigravity, true}, // extra retry 3 -> exhausted
|
|
}
|
|
|
|
var exhausted bool
|
|
var skipped int
|
|
|
|
for _, a := range attempts {
|
|
if exhausted {
|
|
break
|
|
}
|
|
|
|
// Extra retry phase: skip non-Antigravity
|
|
if switchCount >= maxAccountSwitches && a.platform != service.PlatformAntigravity {
|
|
skipped++
|
|
continue
|
|
}
|
|
|
|
if a.isFailover {
|
|
if switchCount >= maxAccountSwitches {
|
|
antigravityExtraCount++
|
|
if antigravityExtraCount > maxExtraRetries {
|
|
exhausted = true
|
|
continue
|
|
}
|
|
// extra retry delay + continue
|
|
continue
|
|
}
|
|
switchCount++
|
|
}
|
|
}
|
|
|
|
require.Equal(t, 3, switchCount, "should have 3 default retries")
|
|
require.Equal(t, 3, antigravityExtraCount, "counter incremented 3 times")
|
|
require.True(t, exhausted, "should be exhausted after exceeding max extra retries")
|
|
require.Equal(t, 0, skipped, "no non-antigravity accounts in this simulation")
|
|
}
|
|
|
|
func TestExtraRetryFlowSimulation_SkipsNonAntigravity(t *testing.T) {
|
|
maxAccountSwitches := 2
|
|
switchCount := 2 // already past default retries
|
|
antigravityExtraCount := 0
|
|
maxExtraRetries := 5
|
|
|
|
type accountSelection struct {
|
|
platform string
|
|
}
|
|
|
|
selections := []accountSelection{
|
|
{service.PlatformGemini}, // should be skipped
|
|
{service.PlatformAnthropic}, // should be skipped
|
|
{service.PlatformAntigravity}, // should be attempted
|
|
}
|
|
|
|
var skippedCount int
|
|
var attemptedCount int
|
|
|
|
for _, sel := range selections {
|
|
if switchCount >= maxAccountSwitches && sel.platform != service.PlatformAntigravity {
|
|
skippedCount++
|
|
continue
|
|
}
|
|
// Simulate failover
|
|
antigravityExtraCount++
|
|
if antigravityExtraCount > maxExtraRetries {
|
|
break
|
|
}
|
|
attemptedCount++
|
|
}
|
|
|
|
require.Equal(t, 2, skippedCount, "gemini and anthropic accounts should be skipped")
|
|
require.Equal(t, 1, attemptedCount, "only antigravity account should be attempted")
|
|
require.Equal(t, 1, antigravityExtraCount)
|
|
}
|