Loading...
HF多模态

voidism/diffcse-roberta-base-sts

DiffCSE: Difference-based C...

标签:


DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings

GitHub Stars

Open In Colab

arXiv link: https://arxiv.org/abs/2204.10298
To be published in NAACL 2022

Authors:
Yung-Sung Chuang,
Rumen Dangovski,
Hongyin Luo,
Yang Zhang,
Shiyu Chang,
Marin Soljačić,
Shang-Wen Li,
Scott Wen-tau Yih,
Yoon Kim,
James Glass

Our code is mainly based on the code of SimCSE. Please refer to their repository for more detailed information.


Overview

DiffCSE

We propose DiffCSE, an unsupervised contrastive learning framework for learning sentence embeddings. DiffCSE learns sentence embeddings that are sensitive to the difference between the original sentence and an edited sentence, where the edited sentence is obtained by stochastically masking out the original sentence and then sampling from a masked language model. We show that DiffSCE is an instance of equivariant contrastive learning (Dangovski et al., 2021), which generalizes contrastive learning and learns representations that are insensitive to certain types of augmentations and sensitive to other “harmful” types of augmentations. Our experiments show that DiffCSE achieves state-of-the-art results among unsupervised sentence representation learning methods, outperforming unsupervised SimCSE by 2.3 absolute points on semantic textual similarity tasks.


Setups

Python


Requirements

  • Python 3.9.5


Install our customized Transformers package

cd transformers-4.2.1
pip install .

If you have already installed transformers==4.2.1 through pip, you need to put modeling_bert.py into <your_python_env>/site-packages/transformers/models/bert/modeling_bert.py and modeling_roberta.py into <your_python_env>/site-packages/transformers/models/bert/modeling_roberta.py.
We modify these two files in the package so that we can perform conditional pretraining tasks using BERT/RoBERTa. If possible, please directly pip install our customized Transformers package.


Install other packages

pip install -r requirements.txt


Download the pretraining dataset

cd data
bash download_wiki.sh


Download the downstream dataset

cd SentEval/data/downstream/
bash download_dataset.sh


Training

(The same as run_diffcse.sh.)

python train.py \
    --model_name_or_path bert-base-uncased \
    --generator_name distilbert-base-uncased \
    --train_file data/wiki1m_for_simcse.txt \
    --output_dir <your_output_model_dir> \
    --num_train_epochs 2 \
    --per_device_train_batch_size 64 \
    --learning_rate 7e-6 \
    --max_seq_length 32 \
    --evaluation_strategy steps \
    --metric_for_best_model stsb_spearman \
    --load_best_model_at_end \
    --eval_steps 125 \
    --pooler_type cls \
    --mlp_only_train \
    --overwrite_output_dir \
    --logging_first_step \
    --logging_dir <your_logging_dir> \
    --temp 0.05 \
    --do_train \
    --do_eval \
    --batchnorm \
    --lambda_weight 0.005 \
    --fp16 --masking_ratio 0.30

Our new arguments:

  • --lambda_weight: the lambda coefficient mentioned in Section 3 of our paper.
  • --masking_ratio: the masking ratio for MLM generator to randomly replace tokens.
  • --generator_name: the model name of generator. For bert-base-uncased, we use distilbert-base-uncased. For roberta-base, we use distilroberta-base.

Arguments from SimCSE:

  • --train_file: Training file path (data/wiki1m_for_simcse.txt).
  • --model_name_or_path: Pre-trained checkpoints to start with such as BERT-based models (bert-base-uncased, bert-large-uncased, etc.) and RoBERTa-based models (RoBERTa-base, RoBERTa-large).
  • --temp: Temperature for the contrastive loss. We always use 0.05.
  • --pooler_type: Pooling method.
  • --mlp_only_train: For unsupervised SimCSE or DiffCSE, it works better to train the model with MLP layer but test the model without it. You should use this argument when training unsupervised SimCSE/DiffCSE models.

For the results in our paper, we use a NVidia 2080Ti GPU with CUDA 11.2. Using different types of devices or different versions of CUDA/Python/PyTorch may lead to slightly different performance.


Evaluation

Open In Colab
We provide a simple colab notebook to reproduce our results easily. We can also run the commands below for evaluation:

python evaluation.py \
    --model_name_or_path <your_output_model_dir> \
    --pooler cls_before_pooler \
    --task_set <sts|transfer|full> \
    --mode test

To evaluate our pretrained DiffCSE checkpoints, we can use the following scripts:


BERT


STS

python evaluation.py \
    --model_name_or_path voidism/diffcse-bert-base-uncased-sts \
    --pooler cls_before_pooler \
    --task_set sts \
    --mode test


Transfer Tasks

python evaluation.py \
    --model_name_or_path voidism/diffcse-bert-base-uncased-trans \
    --pooler cls_before_pooler \
    --task_set transfer \
    --mode test


RoBERTa


STS

python evaluation.py \
    --model_name_or_path voidism/diffcse-roberta-base-sts \
    --pooler cls_before_pooler \
    --task_set sts \
    --mode test


Transfer Tasks

python evaluation.py \
    --model_name_or_path voidism/diffcse-roberta-base-trans \
    --pooler cls_before_pooler \
    --task_set transfer \
    --mode test

For more detailed information, please check SimCSE’s GitHub repo.


Pretrained models

Hugging Face Models

  • DiffCSE-BERT-base (STS): https://huggingface.co/voidism/diffcse-bert-base-uncased-sts
  • DiffCSE-BERT-base (transfer tasks): https://huggingface.co/voidism/diffcse-bert-base-uncased-trans
  • DiffCSE-RoBERTa-base (STS): https://huggingface.co/voidism/diffcse-roberta-base-sts
  • DiffCSE-RoBERTa-base (transfer tasks): https://huggingface.co/voidism/diffcse-roberta-base-trans

We can load the models using the API provided by SimCSE.
See Getting Started for more information.

from diffcse import DiffCSE
model_bert_sts = DiffCSE("voidism/diffcse-bert-base-uncased-sts")
model_bert_trans = DiffCSE("voidism/diffcse-bert-base-uncased-trans")
model_roberta_sts = DiffCSE("voidism/diffcse-roberta-base-sts")
model_roberta_trans = DiffCSE("voidism/diffcse-roberta-base-trans")


Citations

DOI

Please cite our paper and the SimCSE paper if they are helpful to your work!

@inproceedings{chuang2022diffcse,
   title={{DiffCSE}: Difference-based Contrastive Learning for Sentence Embeddings},
   author={Chuang, Yung-Sung and Dangovski, Rumen and Luo, Hongyin and Zhang, Yang and Chang, Shiyu and Soljacic, Marin and Li, Shang-Wen and Yih, Wen-tau and Kim, Yoon and Glass, James},
   booktitle={Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL)},
   year={2022}
}
@inproceedings{gao2021simcse,
   title={{SimCSE}: Simple Contrastive Learning of Sentence Embeddings},
   author={Gao, Tianyu and Yao, Xingcheng and Chen, Danqi},
   booktitle={Empirical Methods in Natural Language Processing (EMNLP)},
   year={2021}
}

数据统计

数据评估

voidism/diffcse-roberta-base-sts浏览人数已经达到1,078,如你需要查询该站的相关权重信息,可以点击"5118数据""爱站数据""Chinaz数据"进入;以目前的网站数据参考,建议大家请以爱站数据为准,更多网站价值评估因素如:voidism/diffcse-roberta-base-sts的访问速度、搜索引擎收录以及索引量、用户体验等;当然要评估一个站的价值,最主要还是需要根据您自身的需求以及需要,一些确切的数据则需要找voidism/diffcse-roberta-base-sts的站长进行洽谈提供。如该站的IP、PV、跳出率等!

关于voidism/diffcse-roberta-base-sts特别声明

本站Ai导航提供的voidism/diffcse-roberta-base-sts都来源于网络,不保证外部链接的准确性和完整性,同时,对于该外部链接的指向,不由Ai导航实际控制,在2023年5月9日 下午7:16收录时,该网页上的内容,都属于合规合法,后期网页的内容如出现违规,可以直接联系网站管理员进行删除,Ai导航不承担任何责任。

相关导航

暂无评论

暂无评论...