ArXiv TLDR

Bian Que: An Agentic Framework with Flexible Skill Arrangement for Online System Operations

🐦 Tweet
2604.26805

Bochao Liu, Zhipeng Qian, Yang Zhao, Xinyuan Jiang, Zihan Liang + 8 more

cs.AIcs.MA

TLDR

Bian Que is an LLM agent framework that automates online system operations by intelligently orchestrating data and knowledge, reducing human effort and improving efficiency.

Key contributions

  • Introduces a unified operational paradigm for release interception, proactive inspection, and root cause analysis.
  • Features Flexible Skill Arrangement for context-aware data and knowledge retrieval, auto-generated by LLMs.
  • Implements a self-evolving mechanism for continuous improvement through case-memory distillation and skill refinement.
  • Achieves significant results: 75% alert reduction, 80% root-cause accuracy, and 50%+ MTTR reduction.

Why it matters

This paper tackles the critical orchestration bottleneck for LLM-based agents in large-scale online system operations. Bian Que's framework significantly reduces human effort and improves efficiency in O&M tasks. Its proven effectiveness in a real-world e-commerce search engine demonstrates substantial practical impact.

Original Abstract

Operating and maintaining (O&M) large-scale online engine systems (search, recommendation, advertising) demands substantial human effort for release monitoring, alert response, and root cause analysis. While LLM-based agents are a natural fit for these tasks, the deployment bottleneck is not reasoning capability but orchestration: selecting, for each operational event, the relevant data (metrics, logs, change events) and the applicable operational knowledge (handbook rules and practitioner experience). Feeding all signals indiscriminately causes dilution and hallucination, while manually curating the event-to-(data, knowledge) mapping is intractable under dozens of daily releases. We present Bian Que, an agentic framework with three contributions: (i) a \emph{unified operational paradigm} abstracting day-to-day O&M into three canonical patterns: release interception, proactive inspection, and alert root cause analysis; (ii) \emph{Flexible Skill Arrangement}, where each Skill specifies which data and knowledge to retrieve for a given business-module context and can be automatically generated and updated by LLMs or iteratively refined through natural-language instructions from on-call engineers; (iii) a \emph{unified self-evolving mechanism} in which one correction signal drives two parallel pathways, case-memory-to-knowledge distillation and targeted Skill refinement. Deployed on the e-commerce search engine of KuaiShou, the major short-video platform in China, Bian Que reduces alert volume by 75%, achieves 80% root-cause analysis accuracy, and cuts mean time to resolution by over 50%. Our framework achieves 99.0% pass rate on offline evaluations. Our code is available at https://github.com/benchen4395/BianQue_Assistant.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.