
Postgraduate research project focused on domain-specific natural language generation for creating coherent and engaging Dungeons & Dragons (D&D) fantasy adventures using Large Language Models (LLMs).
The work addressed challenges in conditional text generation, domain adaptation, and entity-aware narrative consistency, resulting in publicly released datasets, models, and peer-reviewed publications.
A large, structured dataset was curated from D&D source materials (adventures, characters, locations, quests) and evaluated for coverage and quality.
This dataset serves as a foundation for domain-adapted language model training.
SHADE (Semantic Hypernym Annotator for Domain-Specific entities) automatically annotates domain-specific entities with semantic relationships, improving contextual understanding and generation fidelity in LLMs.
Multiple LLM architectures were evaluated, including domain-fine-tuned Mistral, Llama, GPTNeo, and OpenAI GPT models, with emphasis on narrative coherence, stylistic control, and domain authenticity.
[1] Forgotten Realms Wiki Dataset (FRW)
Akila Peiris. A domain-specific dataset for Dungeons & Dragons content generation.
Hugging Face Datasets, 2022.
https://huggingface.co/datasets/Akila/ForgottenRealmsWikiDataset
[2] Forgotten Realms Free Text Generator
Domain-adapted language model for unconstrained fantasy text generation.
Hugging Face Models.
https://huggingface.co/Akila/ForgottenRealmsFreeTextGenerator
[3] Mistral of Realms 7B Instruct (GGUF)
Instruction-tuned Mistral model fine-tuned on D&D domain data.
Hugging Face Models.
https://huggingface.co/Akila/Mistral-of-Realms-7b-Instruct-gguf
[4] Mistral of Realms 7B Instruct v0.2 (GGUF)
Improved iteration with refined instruction-following behavior.
Hugging Face Models.
https://huggingface.co/Akila/Mistral-of-Realms-7b-Instruct-v0.2-gguf
[5] A. Peiris, N. de Silva.
SHADE: Semantic Hypernym Annotator for Domain-Specific Entities – DnD Domain Use Case.
Proceedings of the IEEE 17th International Conference on Industrial and Information Systems (ICIIS), 2023.
[6] A. Peiris, N. de Silva.
Synthesis and Evaluation of a Domain-Specific Large Dataset for Dungeons & Dragons.
Proceedings of the 36th Pacific Asia Conference on Language, Information and Computation (PACLIC), ACL, 2022.