3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in method-name-prediction-10
Use these libraries to find method-name-prediction-10 models and implementations
No subtasks available.
In a preliminary experiment on a neural model recently proposed in the literature, it is found that the model is very brittle, and simple perturbations in the input can cause the model to make mistakes in its prediction.
The effect of obfuscating variable names during training of a code2vec model is investigated to force it to rely on the structure of the code rather than specific names and a simple approach to creating class-level embeddings by aggregating sets of method embeddeddings is considered.
The proposed approach, Sivand, uses simplification techniques that reduce the size of input programs of a CI model while preserving the predictions of the model, and is broadly applicable across many model architectures and problem domains.
This work evaluates the memorization and generalization tendencies in neural code intelligence models through a case study across several benchmarks and model families by leveraging established approaches from other fields that use DNNs, such as introducing targeted noise into the training dataset.
The code intelligence (CI) models are often black-box and do not offer any insights on the input features that they learn for making correct predictions. This opacity may lead to distrust in their prediction and hamper their wider adoption in safety-critical applications. In recent, the program reduction technique is widely being used to identify key input features in order to explain the prediction of CI models. The approach removes irrelevant parts from an input program and keeps the minimal snippets that a CI model needs to maintain its prediction. However, the state-of-the-art approaches mainly use a syntax-unaware program reduction technique that does not follow the syntax of programs, which adds significant overhead to the reduction of input programs and explainability of models. In this paper, we apply a syntax-guided program reduction technique that follows the syntax of input programs during reduction. Our experiments on multiple models across different types of input programs show that the syntax-guided program reduction technique significantly outperforms the syntax-unaware program reduction technique in reducing the size of input programs. Extracting key input features from reduced programs reveals that the syntax-guided reduced programs contain more label-specific key input features and are more vulnerable to adversarial transformation when renaming the key tokens in programs. These label-specific key input features may help to understand the reasoning of models' prediction from different perspectives and increase the trustworthiness to correct classification given by CI models.
A syntax-guided program reduction technique that considers the grammar of the input programs during reduction that is faster and provides smaller sets of key tokens in reduced programs is applied.
This work develops a framework to assess quality improvements that models can get after fine-tuning for the method name prediction task on a particular project, and shows that per-project fine- tuning can greatly improve the models' quality as they capture the project's domain and naming conventions.
A large-scale evaluation of the generalizability of two popular neural program analyzers using seven semantically-equivalent transformations of programs to provide the initial stepping stones for quantifying robustness in neural program Analyzers.
Contracode is proposed: a contrastive pre-training task that learns code functionality, not form, and improves summarization and TypeScript type inference accuracy by 2 to 13 percentage points over competitive baselines.
The results show that even with small semantically preserving changes to the programs, these neural program models often fail to generalize their performance, and suggest that Neural program models based on data and control dependencies in programs generalize better than neural program model based only on abstract syntax trees (ASTs).
Adding a benchmark result helps the community track progress.