Граф коммитов

41 Коммитов

Автор SHA1 Сообщение Дата
Julianne Knott 7b19f8fe5f
Fix Axure pipeline for unit tests (#122)
Fixing pipeline functionality for automatically running tests.

Note: some unit tests now fail, and will be addressed in a future pr.
2022-08-17 09:38:06 -07:00
Julianne Knott 0075c11814
Support outputting multiple sequences for each sample in hf models. (#113)
Add --num_return_sequences parameter
Fix bug introduced by --beam parameter
Updated Benchmarks
2022-03-30 16:25:39 -07:00
Julianne Knott df98314faa
Transformers v4.12.x compatible (#107)
Update to support Huggingface Transformers v4.12.0 - v4.12.5
Added compatibility with ProphetNet
2021-12-14 15:06:42 -08:00
Fei Hu 6dece93193
Update the required version of FairSeq 2021-10-28 08:56:01 -07:00
Julianne Knott 1974223378
Fairseq v0.10.2 compatible (#104)
Updated Fastseq to be compatible with the latest version of Fairseq (0.10.2).
2021-10-27 16:43:34 -07:00
Fei Hu c70b715dec
Fix an issue in installation 2021-09-08 11:09:02 -07:00
Yu Yan 252d61aa7e
Edit cite (#103)
* Update README.md
2021-08-27 10:46:17 -07:00
Fei Hu 957c15f01d
Add 'Open in VS Code' badge (#102) 2021-07-22 14:13:44 -07:00
Yu Yan 00df21aad9
Revise README and add doc for El attention (#88)
Add paper links and demo page.
2021-06-10 15:56:11 -07:00
Jangwon Park 91bb698ee8
doc: minor typo (#91) 2021-06-10 09:53:55 -07:00
Jack Lindamood d7c7467034
Update README.md spelling (#90)
Small spelling correction on README
2021-06-10 00:02:08 -07:00
NickNickGo 2d8c7817b3
Including El Attn optimization (#85)
* Including El Attn optimization

* minor changes

* Update benchmark_fs.sh

* Update benchmark_fs.sh

* minor changes

* minor changes

* env variable for testing fairseq optimizer

* CI tests to gpu3

* Trigger Build
2021-05-19 21:35:26 -07:00
Fei Hu b62e727395
Update README (#84) 2021-05-06 13:46:06 -07:00
Fei Hu 0b7afdfa74
Optimize GPT2 (#81)
* Optimize GPT2

* Add benchmarking scripts for GPT2

* Replace the line break in the generation hypo and update the benchmarking data

* Update README and benchmark script

* Disable transformers-test_gpt2_model_att_mask_past. The currect cache behavior is not compatible with that unit test because the cache key and value will be updated if past is none when the model is called. This unit test will work well if switching the order of calling the second and third model.

* Add readme file for gpt2

* Minor updates
2021-01-25 09:30:11 -08:00
Jiusheng Chen 4fbb2211b9
Benchmark refinement (#74)
* Use bigger data for prophetnet. Reformat benchmarks.

* Use real hf baseline.

* Fix Transformers rouge metric scale.

* Support ngram in hf.

* Fix blue score thresholds.
2020-12-26 17:04:28 -08:00
Jiusheng Chen 0bf9be7f85
Clean and fix benchmarks. (#72) 2020-12-16 22:03:51 -08:00
NickNickGo 8d217ee846
Detokenization parallelization public repo (#69)
* moving to public repo

* moving to public repo

* moving to public repo

* minor change

* minor changes

* benchmarks on larger dataset
2020-12-15 11:42:35 -08:00
Yu Yan 8ef23f1090
Update benchmark for bart (#68)
Switch to full cnndm valid dataset. The previous subset with 1024 sample is too small, when use E2E time to measure speed.
2020-12-09 14:50:31 -08:00
Jiusheng Chen b68715efcb
Update wmt benchmark. (#65)
* Update wmt benchmark.

* Fix dockerfile.

* WMT post process worker from 1 to 5.
2020-12-03 14:30:15 -08:00
Fei Hu a55e9ded44
Update the instructions for installation (#64) 2020-11-25 13:19:50 -08:00
Fei Hu 01e7d492be
Update install_requires and enable fairseq to work with torch 1.6&1.7 (#59)
* Update install_requires and enable fairseq to work with torch 1.6&1.7

* Better error message and address some warnings in torch1.7

* Raise the error if fairseq/transformers are installed but the optmizations can not be applied

* Move transformers/fairseq to extra_require

* Remove the out-of-dated build files for ngram cuda op

* Run fastseq units before transformers and fairseq
2020-11-20 22:14:25 -08:00
Jiusheng Chen cc74935fff
Update speed in benchmark (#57)
* Update benchmark perf numbers.

* Fix benchmark perf numbers. Should be higher.
2020-11-16 21:45:59 -08:00
NickNickGo 6b3c0cb058
Cuda op for ngram repeat blocking (#40)
* Cuda op for ngram repeat blocking

* clean up

* Unit test for cuda op

* unit test updated, minor updates in cpp/cu code

* Rebased on new codebase , updated all benchmarks

* Update README.md

* Update README.md

* Update README.md

* minor change in kernel

* changing install order
2020-11-12 17:21:14 -08:00
NickNickGo 853f57be8a
updating benchmarks of latest repo (#52)
* updating benchmarks of latest repo

* Updating numbers after 3 consecutive runs and cuda driver update

* Update README.md

* Update README.md

* Update README.md
2020-11-12 14:35:57 -08:00
JiushengChen 8151de4f13
Add READMEs for models. (#53)
* Add examples folder. Refine readmes.

* Refine readme.

* Add readme for bart, distilbart, t5, wmt.

* Fix main readme typos.

* Update prophetnet model link. Update dependency version.

* Add code example.

* Fix typo.

Co-authored-by: Jiusheng Chen <jiuchen@microsoft.com>
2020-11-10 11:26:00 -08:00
Fei Hu f9c982f71d
Update docs (#42) 2020-10-14 14:49:26 -07:00
Fei Hu f3a21234e7
Add Prophetnet (#34) 2020-09-15 15:04:48 -07:00
Yu Yan e645065aa4
Simplify ngram block algo (#18)
Simplify ngram block algorithm. Bump speed from 11.0 to 14.8 samples/s.

Before change: generate all ngram pair, and pick banned tokens by look up ngram with last n-1 token.
After change: generate banned tokens directly

For example, the previous generate tokens are 1 2 3 4 2 3. token need to be banned is 4.
Before change, it generate all pair in dict {"1 2": 3, "2 3": 4, "3 4": 2, "4 2" : 3}, and do look up by "2 3", finally find 4 should be banned.
After change, it put 4 in list, and banned it.
2020-08-31 16:09:40 -07:00
JiushengChen d07633eef6
Fix wmt benchmark. (#21)
* Fix wmt benchmark.

* Fix a typo in hf_distibart (#22)

* Refine benchmarks.

* Optimize Transformers-T5 (#15)

* Fix wmt benchmark.

* Refine benchmarks.

* Resovle conflict. Update benchmark numbers based on docker.

* Fix wmt benchmark.

* Refine benchmarks.

* Fix wmt benchmark.

* Refine benchmarks.

* Resovle conflict. Update benchmark numbers based on docker.

Co-authored-by: Fei Hu <hufei68@gmail.com>
2020-08-31 15:54:37 -07:00
Fei Hu 8a797a0fe6
Optimize Transformers-T5 (#15) 2020-08-31 12:16:49 -07:00
Fei Hu 59ffa18d2f
Support BART model in Transformers-2.11.0 (#16)
* Support BART model in Transformers-2.11.0

* Add the benchmarking results for transformers-2.11.0

* Directly call calc_banned_ngram_tokens_v2 instead of replacing calc_banned_ngram_tokens because we change the function signature and may break other places which use this function
2020-08-28 14:09:58 -07:00
Jiusheng Chen 0eaf0c9693 Rename to fastseq-generate-for-fairseq, no tailing s. 2020-08-25 05:08:06 +00:00
Jiusheng Chen 7e62020063 Add dockerfile and update readme. 2020-08-25 02:05:12 +00:00
Jiusheng Chen 18fd179b31 Rename to fastseq-generation-for-(fairseqs|transformers). 2020-08-24 22:31:02 +00:00
Jiusheng Chen 98e1578dd4 Benchmark supports transformers w/wo fastseq. 2020-08-21 18:57:49 +00:00
Fei Hu 8ff7210ecd
Optimize transformers-beam-search related parts (#6)
- Avoid the frequent small data copying between GPU and CPU when computing ngrams;
- Avoid the sorting of the cached key and values for encoder-decoder-attention;
- Reduce the cache memory of the encoder-decoder-attention by beam_size times so that we can run a larger batch size;
- Optimize the implementation of updating scores for banned_ngram tokens and banned_bad_word_token;
2020-08-20 11:08:25 -07:00
Yu Yan ca19fc6fa0
parallel postprocess and computation (#5)
parallel postprocess and computation, bump speed from 14.3 samples/s to 19.0 samples/s
2020-08-19 14:42:52 -07:00
Yu Yan e0db4f20d0
reduce memory copy in postprocess (#3)
* reduce memory copy in postprocess

* support both 0.9.0 and latest version
2020-08-12 17:23:04 -07:00
Fei Hu a5b144f40d
Update README.md 2020-07-31 15:54:51 -07:00
Fei Hu 53246ef9fb
Initial code commit 2020-07-31 15:46:58 -07:00
Microsoft Open Source ba974a07c6 Initial README.md commit 2020-07-15 10:53:32 -07:00