Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ Repository for CodeBERT, fine-tuned on Stack Overflow snippets with respect to N
|
|
16 |
This model is initialized with [CodeBERT-base](https://huggingface.co/microsoft/codebert-base) and trained to classify whether a user will drop out given their posts and code snippets.
|
17 |
## Training Regime
|
18 |
Preprocessing methods for input texts include unicode normalisation (NFC form), removal of extraneous whitespaces, removal of punctuations (except within links), lowercasing and removal of stopwords.
|
19 |
-
Code snippets were also removed of their in-line comments or docstrings (cf. the main manuscript). RoBERTa tokenizer was used, as the built-in tokenizer for the original
|
20 |
|
21 |
Training was done across 8 epochs with a batch size of 8, learning rate of 1e-5, epsilon (weight update denominator) of 1e-8.
|
22 |
A random 20% sample of the entire dataset was used as the validation set.
|
|
|
16 |
This model is initialized with [CodeBERT-base](https://huggingface.co/microsoft/codebert-base) and trained to classify whether a user will drop out given their posts and code snippets.
|
17 |
## Training Regime
|
18 |
Preprocessing methods for input texts include unicode normalisation (NFC form), removal of extraneous whitespaces, removal of punctuations (except within links), lowercasing and removal of stopwords.
|
19 |
+
Code snippets were also removed of their in-line comments or docstrings (cf. the main manuscript). RoBERTa tokenizer was used, as the built-in tokenizer for the original CodeBERT implementation.
|
20 |
|
21 |
Training was done across 8 epochs with a batch size of 8, learning rate of 1e-5, epsilon (weight update denominator) of 1e-8.
|
22 |
A random 20% sample of the entire dataset was used as the validation set.
|