project has two main objectives. Firstly, we aim to build a model that achieves a reasonable
performance while keeping the number of trainable parameters to a minimum. In this regard,
we insert task-specific modules inside the pre-trained BERT model to control the flow of
information between transformer blocks. Our proposed method for fine-tuning BERT
achieves comparable performance to fine-tuning all BERT parameters while only training …