| Issue |
ITM Web Conf.
Volume 80, 2025
2025 2nd International Conference on Advanced Computer Applications and Artificial Intelligence (ACAAI 2025)
|
|
|---|---|---|
| Article Number | 01026 | |
| Number of page(s) | 7 | |
| Section | Machine Learning & Deep Learning Algorithms | |
| DOI | https://doi.org/10.1051/itmconf/20258001026 | |
| Published online | 16 December 2025 | |
A residual GRU-Attention hybrid architecture for high-fidelity Chinese classical poetry generation
Department of Computer Science, University of Utah, Salt Lake City, UT, USA
* Corresponding author: This email address is being protected from spambots. You need JavaScript enabled to view it.
Automatic generation of classical poetry is a highly challenging constrained text generation task, demanding strict adherence to stylistic rules like meter, rhyme, and parallelism alongside high-level semantic coherence. While hybrid architectures combining Recurrent Neural Networks (RNNs) and Attention mechanisms show promise, they often suffer from significant training instability, hindering their performance. This paper systematically investigates these stability challenges in RNN-Attention hybrid architectures and demonstrates through comparative experiments the decisive impact of architectural integration strategies on model convergence and generation quality. We validate our findings on Chinese classical poetry generation, a domain characterized by its intricate constraints. Experimental results show that our proposed model (V3), which employs a novel integration strategy, achieves a comprehensive quality score of 8.1, marking a significant improvement of 24.6% over a strong GRU baseline (V2). Furthermore, ablation studies confirm that a naive concatenation of modules leads to mode collapse, highlighting that the proposed integration strategy— rather than the simple stacking of components—is the key to ensuring model stability and achieving superior performance. Our work offers valuable insights for designing robust hybrid models for other complex text generation tasks.
© The Authors, published by EDP Sciences, 2025
This is an Open Access article distributed under the terms of the Creative Commons Attribution License 4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.

