Supporting Meta-model-based Language Evolution and Rapid Prototyping with Automated Grammar Optimization
SSRN Electronic Journal(2024)
摘要
In model-driven engineering, developing a textual domain-specific language
(DSL) involves constructing a meta-model, which defines an underlying abstract
syntax, and a grammar, which defines the concrete syntax for the DSL. Language
workbenches such as Xtext allow the grammar to be automatically generated from
the meta-model, yet the generated grammar usually needs to be manually
optimized to improve its usability. When the meta-model changes during rapid
prototyping or language evolution, it can become necessary to re-generate the
grammar and optimize it again, causing repeated effort and potential for
errors. In this paper, we present GrammarOptimizer, an approach for optimizing
generated grammars in the context of meta-model-based language evolution. To
reduce the effort for language engineers during rapid prototyping and language
evolution, it offers a catalog of configurable grammar optimization rules. Once
configured, these rules can be automatically applied and re-applied after
future evolution steps, greatly reducing redundant manual effort. In addition,
some of the supported optimizations can globally change the style of concrete
syntax elements, further significantly reducing the effort for manual
optimizations. The grammar optimization rules were extracted from a comparison
of generated and existing, expert-created grammars, based on seven available
DSLs.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要