< 上一个 | 内容 | 下一个 >

2 难解问题的智能算法:第 4 章参考文献

[1] Hutson M. DeepMind AI creates algorithms that sort data faster than those built by people. Nature, 2023, 618(7965): 443-444.

[2] Kool W, van Hoof H, Welling M. Attention, Learn to solve routing problems!. Proceedings of International Conference on Learning Representations. USA (New Orleans, LA): ICLR, 2019.

[3] Chen X, Tian Y. Learning to perform local rewriting for combinatorial optimization. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett eds. Proceedings of the Advances in Neural Information Processing Systems. Canada (Vancouver, BC): NeurIPS, 2019. 6278–6289.

[4] Lu H, Zhang X, Yang S. A learning-based iterative method for solving vehicle routing problems. Proceedings of International Conference on Learning Representations. Ethiopia (Addis Ababa): ICLR, 2020.

[5] Fu Z H, Qiu K B, Zha H. Generalize a small pre-trained model to arbitrarily large TSP instances. Proceedings of the AAAI Conference on Artificial Intelligence. Virtual: AAAI, 2021. 7474-7482.

[6] Jin Y, Ding Y, Pan X, He K, Zhao L, Qin T, Song L, Bian J. Pointerformer: deep reinforced multi-pointer transformer for the traveling salesman problem. Proceedings of the AAAI Conference on Artificial Intelligence. USA (Washington, DC): AAAI, 2023.

[7] Hottung A, Tierney K. Neural large neighborhood search for the capacitated vehicle routing problem. In Giuseppe De Giacomo, Alejandro Catala, Bistra Dilkina, Michela Milano, Senen Barro, Alberto Bugarin, and Jerome Lang eds. Proceedings of the European


image


Conference on Artificial Intelligence. Spain (Santiago de Compostela): ECAI, 2020. 443-450.

[8] Delarue A, Anderson R, Tjandraatmadja C. Reinforcement learning with combinatorial actions: an application to vehicle routing. In Hugo Larochelle, Marc'Aurelio Ranzato, Raia Hadsell, Maria- Florina Balcan, and Hsuan-Tien Lin eds. Proceedings of the Advances in Neural Information Processing Systems. Virtual: NeurIPS, 2020.

[9] Hottung A, Bhandari B, Tierney K. Learning a latent search space for routing problems using variational autoencoders. Proceedings of International Conference on Learning Representations. Austria: ICLR, 2021.

[10] Li S, Yan Z, Wu C. Learning to delegate for large-scale vehicle routing. In Marc'Aurelio Ranzato, Alina Beygelzimer,Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan eds. Proceedings of the Advances in Neural Information Processing Systems. Virtual: NeurIPS, 2021. 26198-26211.

[11] Choo J, Kwon Y D, Kim J, Jae J, Hottung A, Tierney K, Gwon Y. Simulation-guided beam search for neural combinatorial optimization. In Jinho Choo, Yeong-Dae Kwon, Jihoon Kim, Jeongwoo Jae, Andre Hottung, Kevin Tierney, and Youngjune Gwon eds. Proceedings of the Advances in Neural Information Processing Systems. Canada (Vancouver, BC): NeurIPS, 2022. 8760-8772.

[12] Hou Q, Yang J, Su Y, Wang X, Deng Y. Generalize learned heuristics to solve large-scale vehicle routing problems in real-time. Proceedings of International Conference on Learning Representations. Rwanda (Kigali): ICLR, 2023.


[13] Xin L, Song W, Cao Z, Zhang J. Multi-decoder attention model with embedding glimpse for solving vehicle routing problems. Proceedings of the AAAI Conference on Artificial Intelligence. USA (Washington, DC): AAAI, 2023. 12042-12049.

[14] Joshi C K, Laurent T, Bresson X. An efficient graph convolutional network technique for the travelling salesman problem. arXiv preprint arXiv:1906.01227, 2019.

[15] Kool W, van Hoof H, Gromicho J, Welling M. Deep policy dynamic programming for vehicle routing problems. In Pierre Schaus ed. Proceedings of Constraint Programming, Artificial Intelligence, and Operations Research, and Operations Research. USA (Los Angeles, CA): CPAIOR, 2022. 190-213.

[16] Barrett T, Clements W, Foerster J, Lvovsky A. Exploratory combinatorial optimization with reinforcement learning. Proceedings of the AAAI Conference on Artificial Intelligence. USA (New York, NY): AAAI, 2020. 3243-3250.

[17] Khalil E, Dai H, Zhang Y, Dilkina B, Song L. Learning combinatorial optimization algorithms over graphs. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett eds. Proceedings of the Advances in Neural Information Processing Systems. USA (Long Beach, CA): NeurIPS, 2017. 26198-26211.

[18] Barrett TD, Parsonson CW, Laterre A. Learning to solve combinatorial graph partitioning problems via efficient exploration. Proceedings of International Conference on Learning Representations. ICLR, 2022.

[19] Ireland D, Montana G. Lense: Learning to navigate subgraph


embeddings for large-scale combinatorial optimisation. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato eds. Proceedings of International Conference on Machine Learning, USA (Baltimore, Maryland): ICML, 2022. 9622- 9638.

[20] Yao F, Cai R, Wang H. Reversible action design for combinatorial optimization with reinforcement learning. Proceedings of the AAAI Conference on Artificial Intelligence. Canada (Vancouver): AAAI, 2020.

[21] Zhang D, Dai H, Malkin N, et al. Let the flows tell: solving graph combinatorial optimization problems with GFlowNets. arXiv preprint. arXiv:2305.17010, 2023.

[22] Li Z, Chen Q, Koltun V. Combinatorial optimization with graph convolutional networks and guided tree search. In Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicolo Cesa-Bianchi and Roman Garnett eds. Proceedings of the Advances in Neural Information Processing Systems. Canada (Montreal): NeurIPS, 2018. 537-546.

[23] Karalias N, Loukas A. Erdos goes neural: an unsupervised learning framework for combinatorial optimization on graphs. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin eds. Proceedings of the Advances in Neural Information Processing Systems. NeurIPS, 2020. 6659-6672.

[24] Wang L, Hu X, Wang Y, et al. Dynamic job-shop scheduling in smart manufacturing using deep reinforcement learning. Computer Networks, 2021, 190: 107969.

[25] Zhang C, Song W, Cao Z, et al. Learning to dispatch for job shop


scheduling via deep reinforcement learning. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin eds. Proceedings of the Advances in Neural Information Processing Systems. NeurIPS, 2020. 1621-1632.

[26] Park J, Chun J, Kim SH, Kim Y, Park J. Learning to schedule job- shop problems: representation and policy learning using graph neural network and reinforcement learning. International Journal of Production Research. 2021, 59(11):3360-3377.

[27] Jeon W, Gagrani M, Bartan B, et al. Neural DAG scheduling via one- shot priority sampling. Proceedings of the International Conference on Learning Representations. Rwanda (Kigali): ICLR, 2023.

[28] Park J, Bakhtiyar S, Park J. Schedulenet: learn to solve multi-agent scheduling problems with reinforcement learning. arXiv preprint. arXiv:2106.03051, 2021.

[29] Malherbe C, Grosnit A, Tutunov R, et al. Optimistic tree searches for combinatorial black-box optimization. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho eds. Proceedings of the Advances in Neural Information Processing Systems. USA (Louisiana): NeurIPS, 2022. 33080-33092.

[30] Amizadeh S, Matusevych S, Weimer M. Learning to solve circuit- sat: an unsupervised differentiable approach. Proceedings of the International Conference on Learning Representations. USA (Louisiana): ICLR, 2018.

[31] Jaszczur S, Łuszczyk M, Michalewski H. Neural heuristics for SAT solving. arXiv preprint. arXiv:2005.13406, 2020.

[32] Ozolins E, Freivalds K, Draguns A, et al. Goal-aware neural SAT solver. Proceedings of the International Joint Conference on Neural


Networks. Italy (Padua): IEEE, 2022. 1-8.

[33] Li M, Shi Z, Lai Q, et al. DeepSAT: an EDA-driven learning framework for SAT. arXiv preprint. arXiv:2205.13745, 2022.

[34] Karalias N, Robinson J, Loukas A, et al. Neural set function extensions: learning with discrete functions in high dimensions. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho eds. Proceedings of the Advances in Neural Information Processing Systems. USA (Louisiana): NeurIPS, 2022. 15338- 15352.

[35] Shi Z, Li M, Khan S, et al. Satformer: transformers for SAT solving. arXiv preprint. arXiv:2209.00953, 2022.

[36] Duan H, Nejati S, Trimponias G, et al. Online bayesian moment matching based SAT solver heuristics. In Hal Daumé III and Aarti Singh eds. Proceedings of the International Conference on Machine Learning. PMLR, 2020. 2710-2719.

[37] Wang R, Hua Z, Liu G, et al. A bi-level framework for learning to solve combinatorial optimization on graphs. In Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan eds. Proceedings of the Advances in Neural Information Processing Systems. NeurIPS, 2021. 21453- 21466.

[38] Lu H, Li Z, Wang R, et al. ROCO: a general framework for evaluating robustness of combinatorial optimization solvers on graphs. Proceedings of the International Conference on Learning Representations. Rwanda (Kigali): ICLR, 2023.

[39] He Y, Wu G, Chen Y, et al. A two-stage framework and reinforcement learning-based optimization algorithms for complex


scheduling problems. arXiv preprint. arXiv:2103.05847, 2021.

[40] Mao H, Alizadeh M, Menache I, et al. Resource management with deep reinforcement learning. In Bryan Ford, Alex C. Snoeren, and Ellen W. Zegura eds. Proceedings of the ACM Workshop on Hot Topics in Networks. USA (GA): HotNets, 2016. 50-56.

[41] Shao Z, Yang J, Shen C, et al. Learning for robust combinatorial optimization: Algorithm and application. Proceedings of IEEE Conference on Computer Communications. London (United Kingdom): INFOCOM, 2022. 930-939.

[42] Fu ZH, Qiu KB, Zha H. Generalize a small pre-trained model to arbitrarily large TSP instances. Proceedings of the 35th AAAI Conference on Artificial Intelligence. Virtual Event: AAAI, 2021. 7474-7482.

[43] Verma R, Singhal A, Khadilkar H, et al. A generalized reinforcement learning algorithm for online 3d bin-packing. arXiv preprint arXiv:2007.00463, 2020.

[44] Pejic I, van den Berg D. Monte Carlo tree search on perfect rectangle packing problem instance. In: Carlos Artemio Coello Coello ed. Proceedings of the Genetic and Evolutionary Computation Conference. Mexico: GECCO, 2020. 1697-1703.

[45] Zhu Q, Li X, Zhang Z, et al. Learning to Pack: A data-driven tree search algorithm for large-scale 3d bin packing problem. In: Gianluca Demartini, Guido Zuccon, J. Shane Culpepper, Zi Huang, and Hanghang Tong eds. Proceedings of the International Conference on Information and Knowledge Management. Australia (Queensland): CIKM, 2021. 4393-4402.

[46] Bai Y, Ding H, Bian S, et al. Simgnn: A neural network approach to


fast graph similarity computation. In: J. Shane Culpepper, Alistair Moffat, Paul N. Bennett, and Kristina Lerman eds. Proceedings of the ACM International Conference on Web Search and Data Mining. Australia (Melbourne): WSDM, 2019. 384-392.

[47] Li Y, Gu C, Dullien T, et al. Graph matching networks for learning the similarity of graph structured object. In: Kamalika Chaudhuri and Ruslan Salakhutdinov eds. Proceedings of the International Conference on Machine Learning. USA (California): ICML, 2019. 3835-3845.

[48] Wang R, Zhang T, Yu T, et al. Combinatorial learning of graph edit distance via dynamic embedding. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Virtual: CVPR, 2021. 5241-5250.

[49] Bai Y, Ding H, Gu K, et al. Learning-based efficient graph similarity computation via multi-scale convolutional set matching. Proceedings of the AAAI Conference on Artificial Intelligence. USA (New York): AAAI, 2020. 3219-3226.

[50] Dai X, Yan X, Zhou K, et al. Convolutional embedding for edit distance. In: Jimmy X. Huang, Yi Chang, Xueqi Cheng, Jaap Kamps, Vanessa Murdock, JiRong Wen, and Yiqun Liu eds. Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval. Virtual: SIGIR, 2020. 599- 608.

[51] Wang R, Shen L, Chen Y, et al. Towards one-shot neural combinatorial solvers: theoretical and empirical notes on the cardinality-constrained case. Proceedings of the Eleventh International Conference on Learning Representations. Rwanda


(Kigali): ICLR, 2023.

[52] Meirom E, Maron H, Mannor S, et al. Controlling graph dynamics with reinforcement learning and graph neural networks. In: Marina Meila and Tong Zhang eds. Proceedings of the International Conference on Machine Learning. Virtual Event: ICML, 2021. 7565-7577.

[53] Grover A, Wang E, Zweig A, et al. Stochastic optimization of sorting networks via continuous relaxations. Proceedings of the International Conference on Learning Representations. USA (New Orleans): ICLR, 2019.

[54] Xie Y, Dai H, Chen M, et al. Differentiable top-k with optimal

transport. In Hugo Larochelle, Marc'Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin eds. Proceedings of the Advances in Neural Information Processing

Systems. Virtual: NeurIPS, 2020. 20520-20531.

[55] Swezey R, Grover A, Charron B, et al. Pirank: Scalable learning to rank via differentiable sorting. In: Marc'Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan eds. Proceedings of the Advances in Neural Information Processing Systems. Virtual: NeurIPS, 2021. 21644-21654.

[56] Silver D, Hubert T, Schrittwieser J, et al. Mastering chess and shogi by self-play with a general reinforcement learning algorithm. arXiv preprint arXiv:1712.01815, 2017.

[57] Silver D, Schrittwieser J, Simonyan K, et al. Mastering the game of Go without human knowledge. Nature, 2017. 354-359.

[58] Huang J, Patwary M, Diamos G. Coloring Big Graphs with


AlphaGoZero. arXiv preprint arXiv:1902.10162, 2019.


[59] Laterre A, Fu Y, Jabri M K, et al. Ranked reward: enabling self-play reinforcement learning for combinatorial optimization. arXiv preprint arXiv: 1807.01672, 2018.

[60] Li Z, Chen Q, Koltun V. Combinatorial optimization with graph convolutional networks and guided tree search. In: Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicolo Cesa-Bianchi, and Roman Garnett eds. Proceedings of the Advances in Neural Information Processing Systems. Canada (Montreal): NeurIPS, 2018. 537-546.

[61] Qi W. Alpha-T: learning to traverse over graphs with an AlphaZero- inspired Self-Play framework. Research Square, 2021.

[62] Watkins C J C H, Dayan P. Q-learning. Machine learning, 1992, 8: 279-292.

[63] Osband I, Blundell C, Pritzel A, et al. Deep exploration via bootstrapped DQN. In: Daniel D. Lee, Masashi Sugiyama, Ulrike von Luxburg, Isabelle Guyon and Roman Garnett eds. Proceeding of the Advances in Neural Information Processing Systems. Spain (Barcelona): NeurIPS, 2016. 4026-4034.

[64] Anschel O, Baram N, Shimkin N. Averaged-dqn: variance reduction and stabilization for deep reinforcement learning. In: Doina Precup and Yee Whye Teh eds. Proceeding of the International Conference on Machine Learning. Australia (Sydney): PMLR, 2017. 176-185.

[65] Khalil E, Dai H, Zhang Y, et al. Learning combinatorial optimization algorithms over graphs. In: Isabelle Guyon, Ulrike von Luxburg,


image


Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan and Roman Garnett eds. Proceeding of the Advances in Neural Information Processing Systems. USA (Long Beach, CA): NeurIPS, 2017. 6348-6358.

[66] Cappart Q, Goutierre E, Bergman D, et al. Improving optimization bounds using machine learning: decision diagrams meet deep reinforcement learning. Proceedings of the AAAI Conference on Artificial Intelligence. USA (Honolulu, Hawaii): AAAI, 2019. 1443- 1451.

[67] Bai Y, Xu D, Wang A, et al. Fast detection of maximum common subgraph via deep Q-learning. arXiv preprint. arXiv:2002.03129, 2020.

[68] Song J, Lanka R, Yue Y, et al. Co-training for policy learning. In: Amir Globerson and Ricardo Silva eds. Proceedings of the Conference on Uncertainty in Artificial Intelligence. Israel (Tel Aviv): AUAI, 2019. 1191-1201.

[69] Barrett T, Clements W, Foerster J, et al. Exploratory combinatorial optimization with reinforcement learning. Proceedings of the AAAI Conference on Artificial Intelligence. USA (New York, NY): AAAI, 2020. 3243-3250.

[70] Liao H, Zhang W, Dong X, et al. A deep reinforcement learning approach for global routing. Journal of Mechanical Design, 2020, 142(6): 061701.

[71] Scavuzzo L, Chen F, Chételat D, et al. Learning to branch with tree mdps. In: Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and


Kyunghyun Cho eds. Proceedings of the Advances in Neural Information Processing Systems. USA (Louisiana): NeurIPS, 2022. 18514-18526.

[72] Qu Q, Li X, Zhou Y, et al. An improved reinforcement learning algorithm for learning to branch. arXiv preprint. arXiv:2201.06213, 2022.

[73] Wang J, Zhao L, Liu J, et al. Smart resource allocation for mobile edge computing: A deep reinforcement learning approach. IEEE Transactions on Emerging Topics in Computing, 2019, 9(3): 1529- 1541.

[74] He Y, Wu G, Chen Y, et al. A two-stage framework and reinforcement learning-based optimization algorithms for complex scheduling problems. arXiv preprint. arXiv:2103.05847, 2021.

[75] Jacobs T, Alesiani F, Ermis G, et al. Reinforcement learning for route optimization with robustness guarantees. In: Zhihua Z ed. Proceedings of the International Joint Conference on Artificial Intelligence. Canada (Montreal): IJCAI, 2021. 2592-2598.

[76] Tang Y, Agrawal S, Faenza Y. Reinforcement learning for integer programming: learning to cut. arXiv preprint arXiv:1906.04859, 2019.

[77] Yolcu E, Póczos B. Learning local search heuristics for boolean satisfiability. Proceedings of the International Conference on Machine Learning. Virtual Event: PMLR, 2020. 9367-9376.

[78] Ma Q, Ge S, He D, et al. Combinatorial optimization by graph pointer networks and hierarchical reinforcement learning. arXiv


preprint. arXiv:1911.04936, 2019.


[79] Nazari M, Oroojlooy A, Snyder L, et al. Reinforcement learning for solving the vehicle routing problem. In: Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicolo Cesa-Bianchi, and Roman Garnett eds. Proceeding of the Advances in Neural Information Processing Systems. Canada (Montreal): NeurIPS, 2018. 9839-9849.

[80] Deudon M, Cournut P, Lacoste A, et al. Learning heuristics for the TSP by policy gradient. In: Michel Deudon, Pierre Cournut, Alexandre Lacoste, Yossiri Adulyasak, and Louis-Martin Rousseau eds. Integration of Constraint Programming, Artificial Intelligence, and Operations Research. Delft, The Netherlands: Springer, 2018. 170-181.

[81] Kool W, van Hoof H, Welling M. Attention, learn to solve routing problems. Proceedings of the International Conference on Learning Representations. USA (New Orleans, LA): ICLR, 2018.

[82] Bello I, Pham H, Le Q V, et al. Neural combinatorial optimization with reinforcement learning. Proceedings of the International Conference on Learning Representations. France (Toulon): ICLR, 2017.

[83] Hu H, Zhang X, Yan X, et al. Solving a new 3d bin packing problem with deep reinforcement learning method. arXiv preprint arXiv:1708.05930, 2017.

[84] Lu H, Zhang X, Yang S. A learning-based iterative method for solving vehicle routing problems. Proceedings of the International


Conference on Learning Representations. Ethiopia (Addis Ababa): ICLR, 2020.

[85] Sun H, Chen W, Li H, et al. Improving learning to branch via reinforcement learning. Learning Meets Combinatorial Algorithms at NeurIPS 2020.

[86] Emami P, Ranka S. Learning permutations with sinkhorn policy gradient. arXiv preprint arXiv:1805.07010, 2018.

[87] Chen X, Tian Y. Learning to perform local rewriting for combinatorial optimization. In: Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett eds. Proceedings of the Advances in Neural Information Processing Systems. Canada (Vancouver, BC): NeurIPS, 2019. 6278–6289.

[88] Malazgirt G A, Unsal O S, Kestelman A C. Tauriel: Targeting traveling salesman problem with a deep reinforcement learning inspired architecture. arXiv preprint arXiv:1905.05567, 2019.

[89] Cappart Q, Moisan T, Rousseau L M, et al. Combining reinforcement learning and constraint programming for combinatorial optimization. Proceedings of the AAAI Conference on Artificial Intelligence. Virtual: AAAI, 2021. 3677-3687.

[90] Gao L, Chen M, Chen Q, et al. Learn to design the heuristics for vehicle routing problem. arXiv preprint arXiv:2002.08539, 2020.

[91] Silver D, Hubert T, Schrittwieser J, et al. Mastering chess and shogi by self-play with a general reinforcement learning algorithm. arXiv preprint arXiv:1712.01815, 2017.


[92] Laterre A, Fu Y, Jabri M K, et al. Ranked reward: Enabling self-play reinforcement learning for combinatorial optimization. arXiv preprint arXiv:1807.01672, 2018.

[93] Silver D, Schrittwieser J, Simonyan K, et al. Mastering the game of Go without human knowledge. Nature, 2017, 550(7676): 354-359.

[94] Abe K, Xu Z, Sato I, et al. Solving np-hard problems on graphs with extended alphago zero. arXiv preprint arXiv:1905.11623, 2019.

[95] Li Z, Chen Q, Koltun V. Combinatorial optimization with graph convolutional networks and guided tree search. In: Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, NicoloCesa-Bianchi, and Roman Garnett eds. Proceedings of the International Conference on Neural Information Processing Systems. Canada (Montreal): NeurIPS, 2018. 537-546.

[96] Huang J, Patwary M, Diamos G. Coloring big graphs with alphagozero. arXiv preprint arXiv:1902.10162, 2019.

[97] Wang Q. Alpha-T: learning to traverse over graphs with an AlphaZero-inspired Self-Play framework. Research Square, 2021.

[98] Pierrot T, Ligner G, Reed S, et al. Learning compositional neural programs with recursive tree search and planning. In: Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alche- Buc, Emily B. Fox, and Roman Garnett eds. Proceedings of the Advances in Neural Information Processing Systems. Canada (Vancouver, BC): NeurIPS 2019. 14673-14683

[99] Xu R, Lieberherr K. Learning self-game-play agents for combinatorial optimization problems. In: Edith Elkind, Manuela


Veloso, Noa Agmon, and Matthew E. Taylor eds. Proceedings of the International Conference on Autonomous Agents and MultiAgent Systems. Canada (Montreal): AAMAS, 2019. 2276-2278.

[100] Zeng X, Peng H, Li A. Effective and stable role-based multi-agent collaboration by structural information principles. arXiv preprint arXiv:2304.00755, 2023.

[101] Yang Z, Zhang G, Wu J, et al. Minimum entropy principle guided graph neural networks. In: Tat-Seng Chua, Hady W. Lauw, Luo Si, Evimaria Terzi, and Panayiotis Tsapara eds. Proceedings of the ACM International Conference on Web Search and Data Mining. Singapore: WSDM, 2023. 114-122.

[102] Wu J, Li S, Li J, et al. A simple yet effective method for graph classification. In: Luc De Raedt ed. Proceedings of the International Joint Conference on Artificial Intelligence. Austria (Vienna): IJCAI, 2022. 3580-3586.