Перейти к файлу
Yaqing Wang 787b565ed2
Create LICENSE
2022-03-02 14:06:01 -05:00
dataset version 0.1 2021-08-15 19:42:24 +00:00
docker version 0.1 2021-08-15 19:42:24 +00:00
figs Add files via upload 2021-08-15 16:00:29 -04:00
scripts version 0.1 2021-08-15 19:42:24 +00:00
src version 0.1 2021-08-15 19:42:24 +00:00
CODE_OF_CONDUCT.md CODE_OF_CONDUCT.md committed 2021-06-03 14:15:34 -07:00
LICENSE Create LICENSE 2022-03-02 14:06:01 -05:00
README.md Update README.md 2021-08-16 12:54:15 -04:00
SECURITY.md SECURITY.md committed 2021-06-03 14:15:35 -07:00
SUPPORT.md SUPPORT.md committed 2021-06-03 14:15:36 -07:00
requirements.txt Create requirements.txt 2021-12-07 14:38:28 -05:00

README.md

Meta Self-training for Few-shot Neural Sequence Labeling

This is the implementation of the paper Meta Self-training for Few-shot Neural Sequence Labeling. MetaST is short for meta-learning for self-training.

Overview

In this work we present MetaST, a self-training framework with auto data selection technique to prevent error probagation. You can find more details of this work in our paper.

Setup Environment

Install via pip:

  1. create a conda environment running Python 3.6:
conda create --name MetaST python=3.6
conda activate MetaST
  1. install the required dependencies:
pip install -r requirements.txt

Use docker:

  1. Pull docker
docker pull yaqing/pytorch-meta-st:v0.3
  1. Run docker
docker run -it --rm --runtime nvidia  yaqing/pytorch-meta-st:v0.3 bash

Please refer to the following link if you first use docker: https://docs.docker.com/

Quick start

Run MetaST

Training on SNIPS
> bash ./scripts/run_snips.sh

Note that we ran experiments on V100 GPU (32GB) for base MetaST models. You may need to reduce batch size for other GPUs.

Supported datasets:

SNIPS, CONLL03 (En / multilingulal), MIT Movie, MIT restaurant, Wikiann (En / multilingulal) with shot of 5, 10, 20, 100.

Supported models:

  1. BERT: BERT-based classifier

We will soon add more language model families.

Notes and Acknowledgments

The implementation is based on https://github.com/huggingface/transformers
We also used some code from: https://github.com/cliang1453/BOND

How do I cite MetaST?

@inproceedings{10.1145/3447548.3467235,
author = {Wang, Yaqing and Mukherjee, Subhabrata and Chu, Haoda and Tu, Yuancheng and Wu, Ming and Gao, Jing and Awadallah, Ahmed Hassan},
title = {Meta Self-Training for Few-Shot Neural Sequence Labeling},
year = {2021},
publisher = {Association for Computing Machinery},
booktitle = {Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining},
pages = {1737–1747},
series = {KDD '21}
}