Transformer with Tree-order Encoding for Neural Program Generation

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Standard

Transformer with Tree-order Encoding for Neural Program Generation. / Thellmann, Klaudia-Doris; Stadler, Bernhard; Usbeck, Ricardo et al.
Conference XXX. 2022.

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Harvard

APA

Vancouver

Thellmann KD, Stadler B, Usbeck R, Lehmann J. Transformer with Tree-order Encoding for Neural Program Generation. In Conference XXX. 2022 doi: 10.48550/arXiv.2206.13354

Bibtex

@inbook{56b3c87266394178bbbbf7097e04d0af,
title = "Transformer with Tree-order Encoding for Neural Program Generation",
abstract = " While a considerable amount of semantic parsing approaches have employed RNN architectures for code generation tasks, there have been only few attempts to investigate the applicability of Transformers for this task. Including hierarchical information of the underlying programming language syntax has proven to be effective for code generation. Since the positional encoding of the Transformer can only represent positions in a flat sequence, we have extended the encoding scheme to allow the attention mechanism to also attend over hierarchical positions in the input. Furthermore, we have realized a decoder based on a restrictive grammar graph model to improve the generation accuracy and ensure the well-formedness of the generated code. While we did not surpass the state of the art, our findings suggest that employing a tree-based positional encoding in combination with a shared natural-language subword vocabulary improves generation performance over sequential positional encodings. ",
keywords = "cs.CL, cs.AI, 68T07, 68T50, I.2.7, Informatics",
author = "Klaudia-Doris Thellmann and Bernhard Stadler and Ricardo Usbeck and Jens Lehmann",
note = "This paper was authored in late 2020 and early 2021 for the most part",
year = "2022",
month = may,
day = "30",
doi = "10.48550/arXiv.2206.13354",
language = "English",
booktitle = "Conference XXX",

}

RIS

TY - CHAP

T1 - Transformer with Tree-order Encoding for Neural Program Generation

AU - Thellmann, Klaudia-Doris

AU - Stadler, Bernhard

AU - Usbeck, Ricardo

AU - Lehmann, Jens

N1 - This paper was authored in late 2020 and early 2021 for the most part

PY - 2022/5/30

Y1 - 2022/5/30

N2 - While a considerable amount of semantic parsing approaches have employed RNN architectures for code generation tasks, there have been only few attempts to investigate the applicability of Transformers for this task. Including hierarchical information of the underlying programming language syntax has proven to be effective for code generation. Since the positional encoding of the Transformer can only represent positions in a flat sequence, we have extended the encoding scheme to allow the attention mechanism to also attend over hierarchical positions in the input. Furthermore, we have realized a decoder based on a restrictive grammar graph model to improve the generation accuracy and ensure the well-formedness of the generated code. While we did not surpass the state of the art, our findings suggest that employing a tree-based positional encoding in combination with a shared natural-language subword vocabulary improves generation performance over sequential positional encodings.

AB - While a considerable amount of semantic parsing approaches have employed RNN architectures for code generation tasks, there have been only few attempts to investigate the applicability of Transformers for this task. Including hierarchical information of the underlying programming language syntax has proven to be effective for code generation. Since the positional encoding of the Transformer can only represent positions in a flat sequence, we have extended the encoding scheme to allow the attention mechanism to also attend over hierarchical positions in the input. Furthermore, we have realized a decoder based on a restrictive grammar graph model to improve the generation accuracy and ensure the well-formedness of the generated code. While we did not surpass the state of the art, our findings suggest that employing a tree-based positional encoding in combination with a shared natural-language subword vocabulary improves generation performance over sequential positional encodings.

KW - cs.CL

KW - cs.AI

KW - 68T07, 68T50

KW - I.2.7

KW - Informatics

U2 - 10.48550/arXiv.2206.13354

DO - 10.48550/arXiv.2206.13354

M3 - Article in conference proceedings

BT - Conference XXX

ER -

Recently viewed

Publications

  1. Recurrence Quantification Analysis of Processes and Products of Discourse
  2. Structure and dynamics laboratory testing of an indirectly controlled full variable valve train for camless engines
  3. Using Wikipedia for Cross-Language Named Entity Recognition
  4. Legitimizing Digital Transformation: From System Integration to Platformization
  5. Using Heider’s Epistemology of Thing and Medium for Unpacking the Conception of Documents: Gantt Charts and Boundary Objects
  6. Introduction to ‘Exploring the frontiers: unveiling new horizons in carbon efficient biomass utilization’
  7. Knowledge transfer during the integration of knowledge-intensive acquisitions
  8. Representation for interactive exercises
  9. Feature Extraction and Aggregation for Predicting the Euro 2016
  10. Likelihood-based panel cointegration test in the presence of a linear time trend and cross-sectional dependence
  11. Overview of Non-Apis Bees
  12. Creativity in the ‘spaces of hope’
  13. Lengthscale-dependent modelling of ductile failure in metallic microstructures
  14. A community of shared values? Dimensions and dynamics of cultural integration in the European Union
  15. Can Geodesign Be Used to Facilitate Boundary Management for Planning and Implementation of Nature-based Solutions?
  16. The relationship between values and knowledge in visioning for landscape management
  17. Long-term population dynamics of Dactylorhiza incarnata (L.) Soo after abandonment and re-introduction of mowing
  18. Reading instruction in 5th grade: teachers’ perspectives on promoting self-regulated reading in language and content area teaching
  19. Registered Replication Report on Srull and Wyer (1979)