Skip to content

Assignment submission for the course Fundamentals of Deep Learning (CS6910) in the Spring 2022 Semester, under Prof. Mitesh Khapra

Notifications You must be signed in to change notification settings

vamsimalineni96/Indic_Transliteration_Model

Repository files navigation

Recurrent Neural network to build a transliteration system

Course Project submission for the course CS6910 Fundamentals of Deep Learning.
Check this link for the task description: Problem Statement link
Check this link for the report : Report link

Team Members : Vamsi Sai Krishna Malineni (OE20S302), Mohammed Safi Ur Rahman Khan (CS21M035)


Overview :

The purpose of this course project are :

  1. Building a transliteration system using Recurrent Neural Networks.
  2. Comparing different cells such as vanilla RNN, LSTM and GRU.
  3. Implementing attention mechanism and understand how these overcome the limitations of vanilla seq2seq models.
  4. Fine tuning GPT2 transformer model to generate lyrics based on the prompt given.

The link to the wandb report:


General Instructions:

  1. Install the required libraries using the following command :
pip install -r requirements.txt
  1. The is presented in the following notebooks :

    1. RNN.ipynb : This notebook corresponds to training an Seq2Seq Model without attention
    2. RNN_with_Attention.ipynb : This notebook corresponds to training an Seq2Seq Model with attention, visualizing the connectivity and plotting the attention maps
    3. Lyrics_Generation.ipynb : This notebook corresponds to training GPT2 transformer model for lyrics generation Along with jupyter notebooks we have also given python code filed (.py) files. These contain the code to direclty train and test the model in a non interactive way.
  2. For seeing the outputs and the various explanations, and how the code has developed, please check the jupyter notebooks . For training and testing the model from command line, run the (.py) file by following the instructions given in below sections.

  3. If you are running the jupyter notebooks on colab, the libraries from the requirements.txt file are preinstalled, with the exception of the following:

    • wandb
    • transformers
    • datasets

      You can install wandb by using the following command :
!pip install wandb
!pip install transformers
!pip install datasets
  1. The dataset for the RNN part of the assignment can be found at : RNN Dataset Link
  2. The dataset for the Transformers part of the assignment can be found at : Transformer Dataset Link

Building and training a Seq2Seq model :

Running the code

As mentioned earlier, there are two files. One is a jupyter notebook and the other is the python code file.
The jupyter notebook has the outputs still intact so that can be used for reference.
The python file has all the functions and the code used in the jupyter file (along with some additional code that can be used to run from the command line)

Running the python file from the terminal

The python file can be run from the terminal by passing the various command line arguments. Please make sure that the three files of the dataset are as it is present in the same directory as this python file
There are two modes of running this file

1. Running the hyperparameter sweeps using wandb

python rnn_without_attention.py --sweep yes

The code will now run in the sweep mode and will enable wandb integration and log all the data in wand. Make sure you have wandb installed if you want to run in this mode. Also, change the entity name and project name in the code before running in this mode

2. Running in normal mode

python rnn_without_attention.py --sweep XXX --cell XXX --embedSize XXX --dropout XXX --numLayers XXX --hiddenLayerSize XXX --numEpochs XXX --batchSize XXX --optimizer XXX

Replace XXX in above with the appropriate parameter you want to train the model with
For example:

python rnn_without_attention.py --sweep no --cell LSTM --embedSize 64 --dropout 0.2 --numLayers 1 --hiddenLayerSize 32 --numEpochs 2 --batchSize 32 --optimizer adam

Description of various command line arguments

  1. --sweep : Do you want to sweep or not: Enter 'yes' or 'no'. If this is 'yes' then below arguments are not required. Enter below arguments only if this is 'no'
  2. --cell : Cell Type: LSTM or RNN or GRU
  3. --embedSize : Input Embedding Size: integer value
  4. --dropout : Dropout: float value
  5. --numLayers : Number of Encoder/Decoder Layers: integer value
  6. --hiddenLayerSize : Hidden units in cell: integer value
  7. --numEpochs : Number of Epochs: integer value
  8. --batchSize : Batch Size: integer value
  9. --optimizer : Optimizer function: adam or nadam or rmsprop

Running the jupyter notebook

This can be run in a sequential manner. i.e., one cell at a time. This notebook also has the code for plotting the various images required for the assignment.

General Functions Description

1. Getting training data :


The training and validation data in encoded form is obtained using the following code snippet:

encoder_input_data, decoder_input_data, decoder_target_data, encoder_val_input_data, decoder_val_input_data, decoder_val_target_data = one_hot_encoding(input, target, val_input, val_target, input_tokens, target_tokens)

where :

  1. input : array of words of the input language (i.e. English) present in training dataset
  2. target : array of words of the target language (i.e. Telugu) present in training dataset
  3. val_input : array of words of the input language (i.e. English) present in validation dataset
  4. val_target : array of words of the target language (i.e. Telugu) present in validation dataset
  5. input_tokens : list of charaters in the input language
  6. target_tokens : list of characters in the target language

2. Building Model :


The following function is used to build the RNN model:

build_model(num_encoders, num_decoders, cell, embed_size, dropout, hidden_layer_size)

where :

  1. num_encoders : number of encoder layers
  2. num_decoders : number of decoder layers
  3. cell : type of the rnn cell ( simpleRNN, LSTM ,GRU )
  4. embed_size : size of the embedding
  5. dropout : % of dropout (in decimals)
  6. hidden_layer_size : dimensionalilty of output space

3. Hyperparameter sweeps :


Use the following function to perform hyperparameter sweeps in wandb:

sweeper(project_name,entity_name)

where:

  1. project_name : Enter the wandb project name
  2. entity_name : Enter the wandb entity name

The various hyperparameters used are :

hyperparameters = {
          "cell":   {"values":  ["RNN","GRU","LSTM"]},
          "embed_size": {"values":   [16,32,64,256]},
          "hidden_layer_size":  {"values":  [16,32,64,256]},
          "num_layers": {"values":   [1,2,3]},
          "dropout":    {"values":  [0.2,0.3,0.4]},
          "epochs": {"values":   [5,10,15,20]},
          "batch_size": {"values":  [32,64]},
          "optimizer":  {"values":    ["adam", "rmsprop", "nadam"]}
      }

4. Getting Test data :

The test data in encoded form is obtained by using the following code snipppet :

encoder_test_input_data, decoder_test_input_data, decoder_test_target_data = one_hot_encoding_test(test_input, test_target, input_tokens, target_tokens)

where :

  1. test_input : array of words of the input language (i.e. English) present in test dataset
  2. test_target : array of words of the target language (i.e. Telugu) present in test dataset
  3. input_tokens : list of charaters in the input language
  4. target_tokens : list of characters in the target language

5. Testing the best performing model :

The best performing model from the hyperparameter sweeps is fed with test data. Use the following code snippet to find the test accuracy :

testing()

Building and training a Seq2Seq model with attention mechanism:

Running the code

As mentioned earlier, there are two files. One is a jupyter notebook and the other is the python code file.
The jupyter notebook has the outputs still intact so that can be used for reference.
The python file has all the functions and the code used in the jupyter file (along with some additional code that can be used to run from the command line)

Running the python file from the terminal

The python file can be run from the terminal by passing the various command line arguments. Please make sure that the three files of the dataset are as it is present in the same directory as this python file
There are two modes of running this file

1. Running the hyperparameter sweeps using wandb

python rnn_with_attention.py --sweep yes

The code will now run in the sweep mode and will enable wandb integration and log all the data in wand. Make sure you have wandb installed if you want to run in this mode. Also, change the entity name and project name in the code before running in this mode

2. Running in normal mode

python rnn_with_attention.py --sweep XXX --cell XXX --embedSize XXX --dropout XXX --numLayers XXX --hiddenLayerSize XXX --numEpochs XXX --batchSize XXX --optimizer XXX

Replace XXX in above with the appropriate parameter you want to train the model with
For example:

python rnn_with_attention.py --sweep no --cell LSTM --embedSize 64 --dropout 0.2 --numLayers 1 --hiddenLayerSize 32 --numEpochs 2 --batchSize 32 --optimizer adam

Description of various command line arguments

  1. --sweep : Do you want to sweep or not: Enter 'yes' or 'no'. If this is 'yes' then below arguments are not required. Enter below arguments only if this is 'no'
  2. --cell : Cell Type: LSTM or RNN or GRU
  3. --embedSize : Input Embedding Size: integer value
  4. --dropout : Dropout: float value
  5. --numLayers : Number of Encoder/Decoder Layers: integer value
  6. --hiddenLayerSize : Hidden units in cell: integer value
  7. --numEpochs : Number of Epochs: integer value
  8. --batchSize : Batch Size: integer value
  9. --optimizer : Optimizer function: adam or nadam or rmsprop

Running the jupyter notebook

This can be run in a sequential manner. i.e., one cell at a time. This notebook also has the code for plotting the various images required for the assignment.

General Functions Description

1. Getting training data :


The training and validation data in encoded form is obtained using the following code snippet:

encoder_input_data, decoder_input_data, decoder_target_data, encoder_val_input_data, decoder_val_input_data, decoder_val_target_data = one_hot_encoding(input, target, val_input, val_target, input_tokens, target_tokens)

where :

  1. input : array of words of the input language (i.e. English) present in training dataset
  2. target : array of words of the target language (i.e. Telugu) present in training dataset
  3. val_input : array of words of the input language (i.e. English) present in validation dataset
  4. val_target : array of words of the target language (i.e. Telugu) present in validation dataset
  5. input_tokens : list of charaters in the input language
  6. target_tokens : list of characters in the target language

2. Building Model :


The following function is used to build the RNN model:

build_model(num_encoders, num_decoders, cell, embed_size, dropout, hidden_layer_size)

where :

  1. num_encoders : number of encoder layers
  2. num_decoders : number of decoder layers
  3. cell : type of the rnn cell ( simpleRNN, LSTM ,GRU )
  4. embed_size : size of the embedding
  5. dropout : % of dropout (in decimals)
  6. hidden_layer_size : dimensionalilty of output space

3. Hyperparameter sweeps :


Use the following function to perform hyperparameter sweeps in wandb:

sweeper(project_name,entity_name)

where:

  1. project_name : Enter the wandb project name
  2. entity_name : Enter the wandb entity name

The various hyperparameters used are :

hyperparameters = {
          "cell":{ "values":["RNN","GRU","LSTM"]},
          "embed_size":{ "values":[16,32,64,256]},
          "hidden_layer_size":{"values":[16,32,64,256]},
          "num_layers":{"values":[1,2,3]},
          "dropout":{"values":[0.2,0.3,0.4]},
          "epochs":{"values":[5,10,15,20]},
          "batch_size":{"values":[32,64]},
          "optimizer":{"values":["adam", "rmsprop", "nadam"]}
      }

4. Getting Test data :


The test data in encoded form is obtained by using the following code snipppet :

encoder_test_input_data, decoder_test_input_data, decoder_test_target_data = one_hot_encoding_test(test_input, test_target, input_tokens, target_tokens)

where :

  1. test_input : array of words of the input language (i.e. English) present in test dataset
  2. test_target : array of words of the target language (i.e. Telugu) present in test dataset
  3. input_tokens : list of charaters in the input language
  4. target_tokens : list of characters in the target language

5. Testing the best performing model :


The best performing model from the hyperparameter sweeps is fed with test data. Use the following code snippet to find the test accuracy :

testing()

6. Plotting Attention maps :


The attention maps for the model can be built running the following section of ipynb notebook :

Plotting attention maps

7. Visualising Connectivity :


The connectivity can be visualized using either of the following functions :

visualize_gif(index)
visualize(index)

where :

  1. visualize_gif(index) displays a gif format of connectivity visualization of the word at the given index
  2. visulaize(index) displays an image format of connectivity visualization of the word at the given index

The resources used are :

  1. https://towardsdatascience.com/visualising-lstm-activations-in-keras-b50206da96ff
  2. https://machinelearningmastery.com/define-encoder-decoder-sequence-sequence-model-neural-machine-translation-keras/
  3. https://keras.io/examples/nlp/lstm_seq2seq/

Training and fine tuning GPT2 Transformers:


The dataset used for fine tuning GPT2 model can be found here : https://www.kaggle.com/neisse/scrapped-lyrics-from-6-genres
This dataset contains lyrics from 79 muscial genres (data is scraped from the website vagalume.com.br)
The notebook is split into these following segments :

  1. Library imports : This section imports the required libraries for the task.
  2. Data Preperation: This section builds the train,test and validation datasets.
  3. Model Training : This section deals with the model training by running run_clm.py file.
  4. Lyrics Generation : This section generates lyrics by running run_generation.py file. The lyrics generation is based on the prompt given : " I love deep learning ".

The resources used for this question are as follows:

  1. https://github.com/sfs0126/Lyric-Generator-fine-tuned-GPT-2
  2. https://github.com/huggingface/transformers/tree/master/examples/pytorch
  3. https://towardsdatascience.com/fine-tuning-gpt2-for-text-generation-using-pytorch-2ee61a4f1ba7

About

Assignment submission for the course Fundamentals of Deep Learning (CS6910) in the Spring 2022 Semester, under Prof. Mitesh Khapra

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published