Граф коммитов

98 Коммитов

Автор SHA1 Сообщение Дата
Julianne Knott 1974223378
Fairseq v0.10.2 compatible (#104)
Updated Fastseq to be compatible with the latest version of Fairseq (0.10.2).
2021-10-27 16:43:34 -07:00
Fei Hu c70b715dec
Fix an issue in installation 2021-09-08 11:09:02 -07:00
Yu Yan 252d61aa7e
Edit cite (#103)
* Update README.md
2021-08-27 10:46:17 -07:00
Fei Hu 957c15f01d
Add 'Open in VS Code' badge (#102) 2021-07-22 14:13:44 -07:00
Fei Hu 9d1b243a28
Fix the parser issue in supporting prophetnet task (#101)
* Fix the parser issue in supporting prophetnet task

* Disable HfApiPublicTest as content of https://moon-staging.huggingface.co/api/models changed
2021-07-03 15:47:46 -07:00
Yu Yan 2080ee3a5f
Fix Colab link (#100) 2021-06-28 15:48:20 -07:00
Yu Yan 00df21aad9
Revise README and add doc for El attention (#88)
Add paper links and demo page.
2021-06-10 15:56:11 -07:00
Jangwon Park 91bb698ee8
doc: minor typo (#91) 2021-06-10 09:53:55 -07:00
Jack Lindamood d7c7467034
Update README.md spelling (#90)
Small spelling correction on README
2021-06-10 00:02:08 -07:00
拂柳残声 fbdb59ebbc
fix unilm token type, parameters & tests issues (#86) 2021-05-20 15:19:50 -07:00
NickNickGo 2d8c7817b3
Including El Attn optimization (#85)
* Including El Attn optimization

* minor changes

* Update benchmark_fs.sh

* Update benchmark_fs.sh

* minor changes

* minor changes

* env variable for testing fairseq optimizer

* CI tests to gpu3

* Trigger Build
2021-05-19 21:35:26 -07:00
Fei Hu b62e727395
Update README (#84) 2021-05-06 13:46:06 -07:00
拂柳残声 e057adf77d
Unilm Model Support & GenerationMixinV3 (#80)
* add unilm to fastseq

* add xsum unilm base uncased model & unittests

* add benchmark test & code refine

* fix unilm unittest

* integrate generation mixin v3 to v2

* checkout transformers_generate

* remove xsum unilm weight

* fix bugs in beam search optimizer

* fix bugs in beam search optimizer

* fix hf_unilm benchmark

* fix build issues

* for recheck

* add readme & comments & code refine

* resolve comments

* benchmark indent consistent

* fix bugs

* fix bugs & test benchmarks
2021-01-31 20:18:36 -08:00
Fei Hu 0b7afdfa74
Optimize GPT2 (#81)
* Optimize GPT2

* Add benchmarking scripts for GPT2

* Replace the line break in the generation hypo and update the benchmarking data

* Update README and benchmark script

* Disable transformers-test_gpt2_model_att_mask_past. The currect cache behavior is not compatible with that unit test because the cache key and value will be updated if past is none when the model is called. This unit test will work well if switching the order of calling the second and third model.

* Add readme file for gpt2

* Minor updates
2021-01-25 09:30:11 -08:00
Jiusheng Chen 83e8eaaebd
Early exit for ngram blocking. (#79)
* Early exit for ngram blocking.

* Early return.
2021-01-17 21:21:30 -08:00
Fei Hu c5549069cb
Enable fastseq generate CLI to handle the empty shard correctly (#82)
* Handle the empty shard

* Remove unused variable
2021-01-15 20:46:46 -08:00
Jiusheng Chen 4fbb2211b9
Benchmark refinement (#74)
* Use bigger data for prophetnet. Reformat benchmarks.

* Use real hf baseline.

* Fix Transformers rouge metric scale.

* Support ngram in hf.

* Fix blue score thresholds.
2020-12-26 17:04:28 -08:00
Fei Hu b50caf1a7f
Add scripts for cooking binary dataset for ProphetNet (#73) 2020-12-17 18:04:24 -08:00
Jiusheng Chen 0bf9be7f85
Clean and fix benchmarks. (#72) 2020-12-16 22:03:51 -08:00
NickNickGo 8d217ee846
Detokenization parallelization public repo (#69)
* moving to public repo

* moving to public repo

* moving to public repo

* minor change

* minor changes

* benchmarks on larger dataset
2020-12-15 11:42:35 -08:00
Yu Yan 8ef23f1090
Update benchmark for bart (#68)
Switch to full cnndm valid dataset. The previous subset with 1024 sample is too small, when use E2E time to measure speed.
2020-12-09 14:50:31 -08:00
Jiusheng Chen b68715efcb
Update wmt benchmark. (#65)
* Update wmt benchmark.

* Fix dockerfile.

* WMT post process worker from 1 to 5.
2020-12-03 14:30:15 -08:00
Fei Hu a55e9ded44
Update the instructions for installation (#64) 2020-11-25 13:19:50 -08:00
Fei Hu 7558a5c59c
Add missing init files (#62) 2020-11-21 10:35:00 -08:00
Fei Hu 01e7d492be
Update install_requires and enable fairseq to work with torch 1.6&1.7 (#59)
* Update install_requires and enable fairseq to work with torch 1.6&1.7

* Better error message and address some warnings in torch1.7

* Raise the error if fairseq/transformers are installed but the optmizations can not be applied

* Move transformers/fairseq to extra_require

* Remove the out-of-dated build files for ngram cuda op

* Run fastseq units before transformers and fairseq
2020-11-20 22:14:25 -08:00
Fei Hu 4c58e5a0f0
Generate the XML log file for each fastseq unit test (#56)
* Generate the XML log file for each unit tests

* run all fastseq unit tests

* Add Nikhil's changes on pipeline to publish XML

* Just use a small unit test to test pipeline

* Change the xml folder path

* Add more tests

* Add env var for xml log dir and test the failures

* Enable all fastseq unit tests

* Enable all tests

* Generate xml files for fairseq and transformers unit tests

* Fix an issue in pytest command

* Trigger the CI pipeline
2020-11-20 16:40:53 -08:00
NickNickGo 62b6657391
made ngram op device agnostic, unit test cleaned (#61) 2020-11-20 10:10:39 -08:00
Jiusheng Chen 3ac6c2c5ab
Fix prophenet dict loading. (#58)
* Fix prophenet dict loading.

* Use logger.

* Fix import.
2020-11-17 21:54:44 -08:00
Jiusheng Chen cc74935fff
Update speed in benchmark (#57)
* Update benchmark perf numbers.

* Fix benchmark perf numbers. Should be higher.
2020-11-16 21:45:59 -08:00
Jiusheng Chen 97b7e60e03
Speedup hs distilbart download. (#55)
* Speedup hs distilbart download.

* Touch perf.
2020-11-13 19:23:55 -08:00
NickNickGo 6b3c0cb058
Cuda op for ngram repeat blocking (#40)
* Cuda op for ngram repeat blocking

* clean up

* Unit test for cuda op

* unit test updated, minor updates in cpp/cu code

* Rebased on new codebase , updated all benchmarks

* Update README.md

* Update README.md

* Update README.md

* minor change in kernel

* changing install order
2020-11-12 17:21:14 -08:00
NickNickGo 853f57be8a
updating benchmarks of latest repo (#52)
* updating benchmarks of latest repo

* Updating numbers after 3 consecutive runs and cuda driver update

* Update README.md

* Update README.md

* Update README.md
2020-11-12 14:35:57 -08:00
JiushengChen 8151de4f13
Add READMEs for models. (#53)
* Add examples folder. Refine readmes.

* Refine readme.

* Add readme for bart, distilbart, t5, wmt.

* Fix main readme typos.

* Update prophetnet model link. Update dependency version.

* Add code example.

* Fix typo.

Co-authored-by: Jiusheng Chen <jiuchen@microsoft.com>
2020-11-10 11:26:00 -08:00
Fei Hu a5c1e27e6a
Remove some trailing spaces (#51) 2020-11-04 21:26:51 -08:00
Fei Hu 0b337c251e
Target at only one major version of transformers + fairseq in the main branch (#44)
* Refactoring to target only one version of transformers + fairseq in the
main branch

* Address the comments

* Add the error handling for applying optimizations

* Use a same version var for code and setup

* Refactor the class-replace-related parts

* Change log error to warning

* Skip the registration of ProphetNet model if fairseq can not be imported.
2020-11-03 14:11:10 -08:00
NickNickGo 2a0e090040
minor change in CI pipeline (#50)
* minot change

* gpu argument for docker
2020-11-02 18:47:46 -08:00
NickNickGo 9f301ca811
azure pipeline spec (#48)
* azure pipeline spec

* comments added

* minor changes

* simplifying pipeline

* fastseq unittests added

* minor docker file change
2020-10-29 13:49:39 -07:00
JiushengChen f48fcc7e79
Decode hypothesis in Fairseq v0.9.0. (#46)
* Decode hypothesis in Fairseq v0.9.0.

* Adjust import order.

Co-authored-by: Jiusheng Chen <jiuchen@microsoft.com>
2020-10-27 21:12:32 -07:00
Fei Hu 42f9296beb
Add the missing __init__.py for models module (#45) 2020-10-26 11:22:59 -07:00
Fei Hu f5c6fef106
Fix transformers failures (#41) 2020-10-15 13:49:05 -07:00
Fei Hu f9c982f71d
Update docs (#42) 2020-10-14 14:49:26 -07:00
NickNickGo 807c161075
Guarding user env with test wrappers (#33)
* Guarding user env with test wrappers

* minor changes

* cleaning up tests

* minor change

* minor change

* Ensurign zero errors, changing dir for transformers

* Using Pytest instead of Unittest

* Remove reformer test

Co-authored-by: Jiusheng Chen <chenjiusheng@outlook.com>
2020-09-16 10:06:23 -07:00
NickNickGo 76a002c162
Added parallelization , speed bump in decoding approx 1.83 (#20)
* Added parellelization in hypo collection code , speed bump in inference approx 2x

* Guarding user env with bash wrapper

* Revert "Guarding user env with bash wrapper"

This reverts commit d020cc56c6.

* Removed small tensors copy from GPU to CPU. added linting corrections.

* small changes after running transformers Unit tests

* creating gpu tensor

* reverting all formatting changes due to YAPF

* linting checks

* liniting checks
2020-09-16 10:05:21 -07:00
Fei Hu f3a21234e7
Add Prophetnet (#34) 2020-09-15 15:04:48 -07:00
Fei Hu f0b2755d50
Enhance pylintrc (#38) 2020-09-14 10:29:01 -07:00
Fei Hu 24b4fdc17e
Fix error of dic changed size during iteration (#35) 2020-09-08 15:48:52 -07:00
Fei Hu 67b734e904
Add the configs for cache and log envs (#31) 2020-09-03 14:24:04 -07:00
Fei Hu d901cfba5b
update the speed range for t5 (#30) 2020-09-02 18:29:27 -07:00
Fei Hu e52c598624
Fix the fairseq test (#29) 2020-09-02 16:29:01 -07:00
Fei Hu c21c47e79b
Fix an issue in replace decorator (#27) 2020-09-02 09:29:12 -07:00