智算算力交易分层定价模型研究
Research on the Tiered Pricing Model for Intelligent Computing Power Trading
DOI: 10.12677/sea.2026.152014, PDF,    科研立项经费支持
作者: 梁秉豪, 张传刚*, 袁明明:浪潮通信信息系统有限公司,山东 济南
关键词: 算力交易定价模型基础服务增值服务分层定价Computing Power Trading Pricing Model Basic Services Value-Added Services Tiered Pricing
摘要: 随着生成式人工智能、大模型训练、智能视频分析等应用的快速发展,全球智能算力需求呈现指数级增长。传统的算力定价模式主要基于硬件配置或使用时长,难以适应多样化、高并发的智算任务需求,存在任务适配性差与服务等级协议覆盖不足等问题。本文提出一种面向智算算力交易的分层定价模型,将定价模型划分为基础服务与增值服务两部分。基础层基于任务类型(如图像检测、自然语言处理、语音识别等)的核心先验参数构建定价模型;增值层则根据用户对时效、并发、多节点协同等服务等级需求进行浮动。结合真实智算任务日志开展实证研究,通过对比传统计费与分层计费的成本、任务完成率等指标,验证了模型的优越性。该模型在提升资源利用率、满足差异化服务需求方面具有显著优势,为算力资源的市场化配置提供了可行的技术路径。
Abstract: With the rapid development of applications such as generative AI, large model training, and intelligent video analytics, the global demand for intelligent computing power is growing exponentially. Traditional computing power pricing models, primarily based on hardware configuration or usage duration, struggle to accommodate the requirements of diverse, highly concurrent intelligent computing tasks. These models face challenges including poor task adaptability and insufficient Service Level Agreement (SLA) coverage. This paper proposes a tiered pricing model for intelligent computing power transactions, dividing the pricing structure into basic services and value-added services. The basic service tier establishes pricing models based on core prior parameters of task types (such as image detection, natural language processing, and speech recognition). The value-added tier incorporates dynamic adjustments according to users’ SLA requirements including timeliness, concurrency, and multi-node collaboration. Empirical research is conducted using real intelligent computing task logs, and the superiority of the model is verified by comparing costs, task completion rates, and other indicators between traditional billing and layered billing. This model demonstrates significant advantages in enhancing resource utilization and meeting differentiated service demands, providing a viable technical pathway for the market-based allocation of computing power resources.
文章引用:梁秉豪, 张传刚, 袁明明. 智算算力交易分层定价模型研究[J]. 软件工程与应用, 2026, 15(2): 137-142. https://doi.org/10.12677/sea.2026.152014

参考文献

[1] Armbrust, M., Fox, A., Griffith, R., Joseph, A.D., Katz, R., Konwinski, A., et al. (2010) A View of Cloud Computing. Communications of the ACM, 53, 50-58. [Google Scholar] [CrossRef
[2] Li, Z., OBrien, L., Cai, R. and Zhang, H. (2012) Towards a Taxonomy of Performance Evaluation of Commercial Cloud Services. 2012 IEEE Fifth International Conference on Cloud Computing, Honolulu, 24-29 June 2012, 344-351. [Google Scholar] [CrossRef
[3] Luo, W.Q., Tian, Z., Li, Y.F., et al. (2025) Task-Aware Resolution Optimization for Visual Large Language Models.
[4] Amazon Web Services (2025) Using Amazon SageMaker Serverless Inference.
https://docs.aws.amazon.com/zh_cn/sagemaker/latest/dg/serverless-endpoints.html
[5] 黄潇洁. 面向算力网络的算网资源定价方法研究[D]: [硕士学位论文]. 北京: 北京邮电大学, 2024.
[6] Hu, J., Li, K., Liu, C. and Li, K. (2021) A Game-Based Price Bidding Algorithm for Multi-Attribute Cloud Resource Provision. IEEE Transactions on Services Computing, 14, 1111-1122. [Google Scholar] [CrossRef
[7] Chen, Y., Li, Z., Yang, B., Nai, K. and Li, K. (2020) A Stackelberg Game Approach to Multiple Resources Allocation and Pricing in Mobile Edge Computing. Future Generation Computer Systems, 108, 273-287. [Google Scholar] [CrossRef
[8] Lloret-Batlle, R. and Jayakrishnan, R. (2017) Envy-Free Pricing for Collaborative Consumption of Supply in Transportation Systems. Transportation Research Procedia, 23, 772-789. [Google Scholar] [CrossRef
[9] Lu, R., Hong, S.H. and Zhang, X. (2018) A Dynamic Pricing Demand Response Algorithm for Smart Grid: Reinforcement Learning Approach. Applied Energy, 220, 220-230. [Google Scholar] [CrossRef
[10] Narayanan, D., Shoeybi, M., Casper, J., et al. (2021) Efficient Large-Scale Language Model Training on GPU Clusters. [Google Scholar] [CrossRef
[11] Kaplan, J., Mccandlish, S., Henighan, T., et al. (2020) Scaling Laws for Neural Language Models.
[12] Hoffmann, J., Borgeaud, S., Mensch, A., et al. (2022) Training Compute-Optimal Large Language Models.