Transformer with Tree-order Encoding for Neural Program Generation

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Authors

While a considerable amount of semantic parsing approaches have employed RNN architectures for code generation tasks, there have been only few attempts to investigate the applicability of Transformers for this task. Including hierarchical information of the underlying programming language syntax has proven to be effective for code generation. Since the positional encoding of the Transformer can only represent positions in a flat sequence, we have extended the encoding scheme to allow the attention mechanism to also attend over hierarchical positions in the input. Furthermore, we have realized a decoder based on a restrictive grammar graph model to improve the generation accuracy and ensure the well-formedness of the generated code. While we did not surpass the state of the art, our findings suggest that employing a tree-based positional encoding in combination with a shared natural-language subword vocabulary improves generation performance over sequential positional encodings.
Original languageEnglish
Title of host publicationConference XXX
Number of pages10
DOIs
Publication statusIn preparation - 30.05.2022
Externally publishedYes

Bibliographical note

This paper was authored in late 2020 and early 2021 for the most part

    Research areas

  • cs.CL, cs.AI, 68T07, 68T50, I.2.7
  • Informatics

Recently viewed

Publications

  1. A Multilevel CFA-MTMM Model for Nested Structurally Different Methods
  2. Closed-loop control of product geometry by using an artificial neural network in incremental sheet forming with active medium
  3. A Framework for Anomaly Classification and Segmentation in Remanufacturing using Autoencoders and Simulated Data
  4. Inverting the Large Lecture Class: Active Learning in an Introductory International Relations Course
  5. Application of non-convex rate dependent gradient plasticity to the modeling and simulation of inelastic microstructure development and inhomogeneous material behavior
  6. Neural network-based adaptive fault-tolerant control for strict-feedback nonlinear systems with input dead zone and saturation
  7. N3 - A collection of datasets for named entity recognition and disambiguation in the NLP interchange format
  8. Managing Business Process in Distributed Systems: Requirements, Models, and Implementation
  9. Fostering Circularity: Building a Local Community and Implementing Circular Processes
  10. Modeling and Performance Analysis of a Node in Fault Tolerant Wireless Sensor Networks
  11. An on-line orthogonal wavelet denoising algorithm for high-resolution surface scans
  12. ACL–adaptive correction of learning parameters for backpropagation based algorithms
  13. Development of a quality assurance framework for the open source development model
  14. Preventive Emergency Detection Based on the Probabilistic Evaluation of Distributed, Embedded Sensor Networks
  15. Throttle valve control using an inverse local linear model tree based on a Fuzzy neural network
  16. Finding Similar Movements in Positional Data Streams
  17. Neural Network-Based Finite-Time Control for Stochastic Nonlinear Systems with Input Dead-Zone and Saturation
  18. N-term approximation in anisotropic function spaces
  19. A change of values is in the air
  20. Design of a Real Time Path of Motion Using a Sliding Mode Control with a Switching Surface
  21. The scaled boundary finite element method for computational homogenization of heterogeneous media
  22. Digital Control of a Camless Engine Using Lyapunov Approach with Backward Euler Approximation
  23. Diffusion-driven microstructure evolution in OpenCalphad
  24. Evaluating OWL 2 reasoners in the context of checking entity-relationship diagrams during software development
  25. Using trait-based filtering as a predictive framework for conservation
  26. Tracing exploratory modes in digital collections of museum Web sites using reverse information architecture