Automated Generation of International Trade Documents via Fusion of Large Language Models and Deterministic Rules
Abstract
numerical or formatting errors. While Large Language Models have demonstrated exceptional capabilities in natural language understanding, their direct application in high-compliance scenarios for end-to-end generation remains hindered by numerical inconsistency,
logical instability, and uncontrollable outputs. This paper proposes an automated generation method that integrates LLMs with deterministic rules and algorithms. By constructing a hybrid pipeline consisting of "Semantic ParsingRule ConstraintTemplate Generation, " the method leverages LLMs to handle semantic uncertainty while delegating numerical allocation and rigid logic to deterministic
algorithms, ensuring structural consistency and numerical validity. The proposed method has been deployed in real-world trade scenarios for three years. Experimental results indicate that this approach significantly outperforms both manual processes and pure LLMbased solutions in terms of end-to-end efficiency and reliability, providing a viable engineering path for LLM applications in highreliability document automation.
Keywords
Full Text:
PDFReferences
[1] Wei, X., Cui, X., Cheng, N., Wang, X., & Zhang, X. (2023). Zero-Shot Information Extraction via Chatting with ChatGPT. arXiv
preprint arXiv:2302.10205.
[2] Xu, D., Chen, Y., Yuan, S., et al. (2024). Large Language Models for Information Extraction: A Survey. arXiv preprint
arXiv:2308.07107.
[3] Ji, Z., Lee, N., Frieske, R., et al. (2023). Survey of Hallucination in Natural Language Generation. ACM Computing Surveys, 55(12),
1-38.
[4] Patel, A., Bhattamishra, S., & Goyal, N. (2021). Are NLP Models really able to Solve Simple Math Word Problems?. Proceedings of
the 2021 Conference of the North American Chapter of the Association for Computational Linguistics.
[5] Nye, M., Andreassen, A. J., Gur-Ari, G., et al. (2021). Show Your Work: Scratchpads for Intermediate Computation with Language
Models. arXiv preprint arXiv:2112.00114.
DOI: http://dx.doi.org/10.70711/aitr.v3i7.8874
Refbacks
- There are currently no refbacks.