Граф коммитов

196 Коммитов

Автор SHA1 Сообщение Дата
Debadeepta Dey f0528dfdf8 Made gpt2 flex config have much finer grained d_model range. Added script for visualizing how memory/latency change as we vary a particular dimension, keeping others constant. 2022-12-16 16:45:50 -03:00
Debadeepta Dey 794cffedf3 Added script for analyzing a corpus of fully trained architectures and evaluating proxy measures like decoder params and total params. 2022-12-16 16:45:48 -03:00
Debadeepta Dey 846f6cb2e0 At the end of training, save summaries to work dir. 2022-12-16 16:45:26 -03:00
Debadeepta Dey f14267f997 Minor. 2022-12-16 16:44:31 -03:00
Debadeepta Dey 69095c8959 Made vocab size of gpt2 models 10k. Added analytical check in check_constraints to speed up candidate finding. 2022-12-16 16:43:31 -03:00
Debadeepta Dey 05db6b8e2c Minor fixes. 2022-12-16 16:42:14 -03:00
Debadeepta Dey 003096b431 Added GPT2 launcher. 2022-12-16 16:42:07 -03:00
Debadeepta Dey ad1a2d39f6 Added total params vs memory, latency plots to search. 2022-12-16 16:42:07 -03:00
Debadeepta Dey b8fdea477f Fixed semi-brute force. 2022-12-16 16:42:02 -03:00
Debadeepta Dey 0877ccae4e Further simplified search. 2022-12-16 16:41:48 -03:00
Debadeepta Dey 911f55e84d Minor. 2022-12-16 16:41:47 -03:00
Debadeepta Dey 34f128aadb Further cleanup. 2022-12-16 16:41:46 -03:00
Debadeepta Dey b924c73015 Cleaned up search code. 2022-12-16 16:41:46 -03:00
Debadeepta Dey f199a28433 Minor. 2022-12-16 16:41:18 -03:00
Debadeepta Dey 415215d9df Typo fixes. 2022-12-16 16:40:38 -03:00
Debadeepta Dey 38ba29c3ea Minor edits and comments. 2022-12-16 16:40:29 -03:00
Debadeepta Dey dda6f5e0e2 Added launch.json entry for distributed launch debug target with environment variable for optionally disabling NCCL P2P. 2022-12-16 16:28:39 -03:00
Debadeepta Dey 917daeed56 Added launch target for transformerxl training. 2022-12-16 16:28:34 -03:00
Shital Shah 23005fb5d1 gpt2 training refactor 2022-12-16 16:27:30 -03:00
Shital Shah 5ff4063fc9 gpt2 script runnable 2022-12-16 16:27:26 -03:00
Shital Shah 4c09c3dc98 imagenet tensor shape mismatch fix 2022-12-16 16:24:48 -03:00
Debadeepta Dey 3de997ef12 Fixed settings.json to not have hardcoded pythonpath. 2022-12-16 16:24:46 -03:00
Debadeepta Dey fee9ac86ea Fixed division by zero edge case in petridish sampler. 2022-12-16 16:24:33 -03:00
Shital Shah 31a000eddb remove python path 2022-12-16 16:24:09 -03:00
Debadeepta Dey 8f83a8f545 Fixed typo. 2022-12-16 16:23:58 -03:00
Shital Shah ce6e015e04 delete exp folder, manual run ignores copy files from search 2022-12-16 16:23:11 -03:00
Shital Shah 840a08fe31 post merge sync with refactoring 2022-12-16 16:21:39 -03:00
Ubuntu 232efb4bf5 Now multiple seed models will be trained in petridish distributed. 2022-12-16 16:19:34 -03:00
Ubuntu e4e1591acd Fixed settings by removing pythonpath. 2022-12-16 16:19:31 -03:00
Shital Shah 484cfea190 remove explicit python path 2022-12-16 16:17:49 -03:00
Debadeepta Dey 7579f9bcfd Moved petridish_ray_mock.py file to misc folder. 2022-12-16 16:17:49 -03:00
Debadeepta Dey 7c3e3d4b2c Removed redundant petridish debug launcher. 2022-12-16 16:17:02 -03:00
Shital Shah 30ed18c0f4 Updated .md files 2022-12-16 16:13:54 -03:00
Debadeepta Dey 0bd82db1c5 Fixed gs finalizer. Running some jobs on cluster to baseline. 2022-12-16 16:13:46 -03:00
Shital Shah 78aaddb6e2 max_batches bug fix, added Divnas evall full run config 2022-12-16 16:12:25 -03:00
Shital Shah 57a10a2dac general arch params implementation, support for multiple optimizers 2020-05-18 03:12:44 -07:00
Shital Shah fe8d3efc45 diversenad branch merge 2020-05-18 03:12:44 -07:00
Shital Shah 38f921cded Fix gradual wamup. enable dataset specific toy mode batch sizes, imagenet toy mode working, disable decay_bn for now 2020-05-18 03:12:43 -07:00
Shital Shah 5a4ef14edb fix windows run 2020-05-18 03:12:42 -07:00
Shital Shah cec9c16780 imagenet support for handcrafted models 2020-05-18 03:12:42 -07:00
Shital Shah 0a5d3cf8ac ImageNet toy mode cell outs verified, better readibility in macro builder, move aux tower stride to yaml, prev reduction based on model stems to support imagenet, remove adaptive avg pool from aux tower, remove bn from pool op, converted original darts genotype to archai format 2020-05-18 03:12:42 -07:00
Shital Shah 6720187057 Added ImageNet toy run, multi_step LRs, separate out imagenet folder class, add hue in imagenet aug, module name guess for pre-crafted models, add reduction property in stems, remove imagenet from stem/pool names, combine s4 stem in s2 stem, tune toy.yaml for more realistic toy mode 2020-05-18 03:12:42 -07:00
Shital Shah e40777ef35 resnet eval working in toy 2020-05-18 03:12:41 -07:00
Shital Shah cd6a01a5f4 added eval only test 2020-05-18 03:12:41 -07:00
Shital Shah 0f9f83d59a fix log message, use trainer title in log, 2020-05-18 03:12:41 -07:00
Shital Shah af1d639c6e initial 2020-05-18 03:11:07 -07:00