This project is the implementation of the system submitted to the SLAM 2018 (Second Language Acquisition Modeling 2018) shared task.
This page gives instructions for replicating the results in our system.
Our project is built on python. We have ensured python 2 and 3 compatibility. In this section, we describe the installation procedure for Ubuntu 16.04.
shell
git clone https://github.com/iampuntre/slam18.git
cd slam18
virtualenv env
source env/bin/activate
pip install -r requirements.txt
mkdir data
Note: Follow equivalent instructions for your Operating System
In our experiments, we use the SLAM 2018 Dataset. To download the dataset, download from https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/8SWHNO.
After downloading the data, unzip it in the data
directory.
In order to train the model, you will have to configure the parameters.ini
file. You can find this file in src/parameters.ini
.
We have three sections in the file - model
, options
and context_features
. The model
section is used to point to the train and test files. options
section is used to manipulate the various hyperparameters while training the model. Lastly, we have the context_features
section, which is used to activate/deactivate various context based features.
Change the appropriate values for train, dev and test files. We have preset the values of the hyperparameters that we have used in our experiments. By default, all the context features are activated.
After you have successfully configured the parameters.ini
file, you should prepare the data for training. This is an intermediate step, where we extract the tokens and part of speech present in the surrounding of the context. For more details, read our paper.
To prepare the data, execute the following -
shell
python src/prepare_data.py --params_file src/parameters.ini
You should be able to see three .json
files in your data directory.
To train the model, type the following command in your terminal -
shell
python src/train_model.py --params_file src/parameters.ini
Note: It is recommended you run this step only if you have sufficient memory (atleast 16GB)
To evaluate your predictions, execute the following command -
shell
python src/eval.py --params_file src/parameters.ini
If you make use of this work in your paper, please cite our paper
``` @InProceedings{nayak-rao:2018:W18-05, author = {Nayak, Nihal V. and Rao, Arjun R.}, title = {Context Based Approach for Second Language Acquisition}, booktitle = {Proceedings of the Thirteenth Workshop on Innovative Use of NLP for Building Educational Applications}, month = {June}, year = {2018}, address = {New Orleans, Louisiana}, publisher = {Association for Computational Linguistics}, pages = {212--216}, url = {http://www.aclweb.org/anthology/W18-0524} }
```
Bumps ipython from 7.16.3 to 8.10.0.
Sourced from ipython's releases.
See https://pypi.org/project/ipython/
We do not use GitHub release anymore. Please see PyPI https://pypi.org/project/ipython/
15ea1ed
release 8.10.0560ad10
DOC: Update what's new for 8.10 (#13939)7557ade
DOC: Update what's new for 8.10385d693
Merge pull request from GHSA-29gw-9793-fvw7e548ee2
Swallow potential exceptions from showtraceback() (#13934)0694b08
MAINT: mock slowest test. (#13885)8655912
MAINT: mock slowest test.a011765
Isolate the attack tests with setUp and tearDown methodsc7a9470
Add some regression tests for this changefd34cf5
Swallow potential exceptions from showtraceback()Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase
.
natural-language-processing machine-learning machine-learning-models language-learning nlp duolingo cognitive-science