Skip to content

feat(benchmark): add MiniMax as LLM provider for HomeSec-Bench#176

Open
octo-patch wants to merge 1 commit intoSharpAI:masterfrom
octo-patch:feature/add-minimax-provider
Open

feat(benchmark): add MiniMax as LLM provider for HomeSec-Bench#176
octo-patch wants to merge 1 commit intoSharpAI:masterfrom
octo-patch:feature/add-minimax-provider

Conversation

@octo-patch
Copy link

Summary

Add MiniMax Cloud API as a first-class LLM provider for the Home Security AI Benchmark (HomeSec-Bench), enabling users to benchmark MiniMax models (M2.7, M2.7-highspeed, M2.5, M2.5-highspeed) against local Qwen/DeepSeek and other cloud providers like OpenAI.

Changes

  • Provider presets system with auto-configured base URL and default model for MiniMax
  • MiniMax auto-detection via API type, base URL pattern, or MINIMAX_API_KEY env var
  • Temperature clamping [0, 1.0] for MiniMax API compatibility
  • Cloud API recognition enables stream_options for token tracking
  • config.yaml llmProvider selector and minimaxModel dropdown for Aegis UI
  • Documentation updated SKILL.md with providers table, README.md mention
  • 52 tests (49 unit + 3 integration) covering all provider logic

Files Changed (5 files, 538 additions)

File Changes
scripts/run-benchmark.cjs Provider presets, auto-detection, temp clamping
config.yaml llmProvider and minimaxModel parameters
SKILL.md Provider docs, env vars, standalone examples
README.md MiniMax benchmark mention
tests/minimax-provider.test.cjs 49 unit + 3 integration tests

Usage

AEGIS_LLM_API_TYPE=minimax MINIMAX_API_KEY=your-key node scripts/run-benchmark.cjs

MiniMax Models

Model Context Notes
MiniMax-M2.7 1M tokens Latest flagship
MiniMax-M2.7-highspeed 1M tokens Optimized for throughput
MiniMax-M2.5 204K tokens Previous generation
MiniMax-M2.5-highspeed 204K tokens Fast inference

Test Plan

  • All 49 unit tests pass
  • All 3 integration tests pass with live MiniMax API
  • Existing benchmark functionality unchanged
  • Run full HomeSec-Bench suite with MiniMax-M2.7

…Bench

Add MiniMax Cloud API (M2.7, M2.7-highspeed, M2.5, M2.5-highspeed) as a
built-in provider preset for the Home Security AI Benchmark, enabling users
to benchmark MiniMax models against local and other cloud LLMs.

Changes:
- Provider presets system with auto-configured base URL for MiniMax
  (api.minimax.io) and OpenAI
- MiniMax auto-detection via AEGIS_LLM_API_TYPE=minimax or base URL
- Temperature clamping [0, 1.0] for MiniMax API compatibility
- MINIMAX_API_KEY env var as fallback when AEGIS_LLM_API_KEY is not set
- MiniMax recognized as cloud API for stream_options support
- config.yaml: llmProvider selector (builtin/openai/minimax) and
  minimaxModel selector (M2.7, M2.7-highspeed, M2.5, M2.5-highspeed)
- Updated SKILL.md with provider docs, supported providers table,
  standalone usage examples
- README.md: mention MiniMax as benchmark provider option
- 52 tests (49 unit + 3 integration) covering provider resolution,
  model defaults, detection, temperature clamping, config validation
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant