Transformer with Tree-order Encoding for Neural Program Generation

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Standard

Transformer with Tree-order Encoding for Neural Program Generation. / Thellmann, Klaudia-Doris; Stadler, Bernhard; Usbeck, Ricardo et al.
Conference XXX. 2022.

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Harvard

APA

Vancouver

Thellmann KD, Stadler B, Usbeck R, Lehmann J. Transformer with Tree-order Encoding for Neural Program Generation. In Conference XXX. 2022 doi: 10.48550/arXiv.2206.13354

Bibtex

@inbook{56b3c87266394178bbbbf7097e04d0af,
title = "Transformer with Tree-order Encoding for Neural Program Generation",
abstract = " While a considerable amount of semantic parsing approaches have employed RNN architectures for code generation tasks, there have been only few attempts to investigate the applicability of Transformers for this task. Including hierarchical information of the underlying programming language syntax has proven to be effective for code generation. Since the positional encoding of the Transformer can only represent positions in a flat sequence, we have extended the encoding scheme to allow the attention mechanism to also attend over hierarchical positions in the input. Furthermore, we have realized a decoder based on a restrictive grammar graph model to improve the generation accuracy and ensure the well-formedness of the generated code. While we did not surpass the state of the art, our findings suggest that employing a tree-based positional encoding in combination with a shared natural-language subword vocabulary improves generation performance over sequential positional encodings. ",
keywords = "cs.CL, cs.AI, 68T07, 68T50, I.2.7, Informatics",
author = "Klaudia-Doris Thellmann and Bernhard Stadler and Ricardo Usbeck and Jens Lehmann",
note = "This paper was authored in late 2020 and early 2021 for the most part",
year = "2022",
month = may,
day = "30",
doi = "10.48550/arXiv.2206.13354",
language = "English",
booktitle = "Conference XXX",

}

RIS

TY - CHAP

T1 - Transformer with Tree-order Encoding for Neural Program Generation

AU - Thellmann, Klaudia-Doris

AU - Stadler, Bernhard

AU - Usbeck, Ricardo

AU - Lehmann, Jens

N1 - This paper was authored in late 2020 and early 2021 for the most part

PY - 2022/5/30

Y1 - 2022/5/30

N2 - While a considerable amount of semantic parsing approaches have employed RNN architectures for code generation tasks, there have been only few attempts to investigate the applicability of Transformers for this task. Including hierarchical information of the underlying programming language syntax has proven to be effective for code generation. Since the positional encoding of the Transformer can only represent positions in a flat sequence, we have extended the encoding scheme to allow the attention mechanism to also attend over hierarchical positions in the input. Furthermore, we have realized a decoder based on a restrictive grammar graph model to improve the generation accuracy and ensure the well-formedness of the generated code. While we did not surpass the state of the art, our findings suggest that employing a tree-based positional encoding in combination with a shared natural-language subword vocabulary improves generation performance over sequential positional encodings.

AB - While a considerable amount of semantic parsing approaches have employed RNN architectures for code generation tasks, there have been only few attempts to investigate the applicability of Transformers for this task. Including hierarchical information of the underlying programming language syntax has proven to be effective for code generation. Since the positional encoding of the Transformer can only represent positions in a flat sequence, we have extended the encoding scheme to allow the attention mechanism to also attend over hierarchical positions in the input. Furthermore, we have realized a decoder based on a restrictive grammar graph model to improve the generation accuracy and ensure the well-formedness of the generated code. While we did not surpass the state of the art, our findings suggest that employing a tree-based positional encoding in combination with a shared natural-language subword vocabulary improves generation performance over sequential positional encodings.

KW - cs.CL

KW - cs.AI

KW - 68T07, 68T50

KW - I.2.7

KW - Informatics

U2 - 10.48550/arXiv.2206.13354

DO - 10.48550/arXiv.2206.13354

M3 - Article in conference proceedings

BT - Conference XXX

ER -

Recently viewed

Activities

  1. Computer Simulations in Design. How Social Media meet Computational Methods in Design Processes
  2. Learning written argumentation in mathematic´s contexts
  3. Challenges And Pitfalls Of Designing, Conducting, And Evaluating Web-Based Interventions
  4. archiDART: an R package for the automated 2D computation of root architectural traits
  5. Changes of teachers' knowledge, beliefs and attitudes regarding inquiry-based learning in the framework of a professional development programme
  6. Co-supervisor of the dissertation "Multi-trophic interactions and functional diversity in biodiversity experiments."
  7. Workshop on Stochastic Models, Statistics and Their Applications 2017
  8. Is a better understanding of assembly a way to help reassemble communities for restoration?
  9. An axiomatic foundation of entropic preferences under Knightian uncertainty
  10. Nordic Seminar on Computational Mechanics - NSCM 2016
  11. How to observe Culture?
  12. It's how, not what we use that matters - Communications Modes in the Internet
  13. Bsc-Thesis: Functional diversity of leaf-litter ants in a tree diversity experiment
  14. Knowledge Spaces
  15. Dealing with temporal tensions in drug development processes
  16. LC-MS identification of the photo-transformation products of desipramine with studying the effect of different environmental variables on the kinetics of their formation
  17. Alternative Mouse – Alternative User? Towards a History of Assistive Media
  18. Coherent behavior in geophysical flows
  19. Coupled component modelling for inter- and transdisciplinariy climate change impact reserach: dimensions of integration and examples of interface design
  20. Monitoring Phenological phases of the crops using TerraSAR-X Spotlight Images based on Factor Analysis
  21. Institute for Advanced Study Princeton

Publications

  1. Enhancing the Building Information Modeling Lifecycle of Complex Structures with IoT
  2. Modelling the Complexity of Measurement Estimation Situations - A Theoretical Framework for the Estimation of Lengths
  3. Optimized neural networks for modeling of loudspeaker directivity diagrams
  4. A comparison of ML, WLSMV and Bayesian methods for multilevel structural equation models in small samples: A simulation study
  5. Human–learning–machines: introduction to a special section on how cybernetics and constructivism inspired new forms of learning
  6. Fast, Fully Automated Analysis of Voriconazole from Serum by LC-LC-ESI-MS-MS with Parallel Column-Switching Technique
  7. The Influence of Note-taking on Mathematical Solution Processes while Working on Reality-Based Tasks
  8. Binary Random Nets II
  9. A cognitive mapping approach to understanding public objection to energy infrastructure
  10. Nonlinear PD fault-tolerant control for dynamic positioning of ships with actuator constraints
  11. A framework for business model development in technology-driven start-ups
  12. Evaluation of standard ERP software implementation approaches in terms of their capability for business process optimization
  13. Understanding the modes of use and availability of critical metals-An expert-based scenario analysis for the case of indium
  14. Machine Learning Applications
  15. Do guided internet-based interventions result in clinically relevant changes for patients with depression?
  16. Networking for the environment
  17. Design of an Information-Based Distributed Production Planning System
  18. Estimation and interpretation of a Heckman selection model with endogenous covariates
  19. The buffering effect of selection, optimization, and compensation strategy use on the relationship between problem solving demands and occupational well-being
  20. Participatory energy scenario development as dramatic scripting
  21. Polar Coordinates and Interactive Learning
  22. Clause identification using entropy guided transformation learning