57 Commits

Author SHA1 Message Date
a74da5d99a move to azure 2020-12-02 12:56:19 -08:00
0574c5708b delete 2020-01-03 18:06:25 -08:00
03fce0a080 Update README.md
per cullen's request
2020-01-03 14:07:30 -08:00
0f97760ebe Update LICENSE
Hi. David Lansky and Jack Clark asked me to update the GPT-2 License accordingly.
2020-01-03 14:04:53 -08:00
ebdba20a19 updated g_form contact 2019-11-26 13:33:24 -08:00
d98291d2ae update model card 2019-11-05 09:01:23 -08:00
fbae7db92a update readmes 2019-11-05 08:53:40 -08:00
ac5d52295f nucleus sampling 2019-08-26 21:20:33 -07:00
f35fa1d920 push 774M model 2019-08-20 09:10:36 -07:00
cb415376c3 add model card 2019-08-19 17:34:30 -07:00
e9378792c4 Merge pull request #161 from openai/christopherhesse-patch-1
Update README.md
2019-07-26 17:55:04 -07:00
41a6793dc6 Update README.md 2019-07-26 17:02:46 -07:00
c0859d7523 Fix TODO in sample.sample_sequences- Avoid 'leaving last token calculation to while loop' (#119)
* do initial run on full context

* decrement while loop iterations

* add context to output

* remove first param

* removing first param: change shape invariant
2019-05-30 21:49:18 -07:00
e5c5054474 allow models to be in a separate folder via models_dir argument (#129)
* models_dir argument to allow models in a separate folder

* default value for models_dir to be same as before

* allow environment variables and user home in models_dir
2019-05-16 09:42:58 -07:00
dd75299dfe remove samples 2019-05-03 15:43:08 -07:00
b5ef71a922 reference dataset 2019-05-03 15:26:08 -07:00
0503b1b249 updates for 345M model 2019-05-02 20:39:33 -07:00
d14501aade Update CONTRIBUTORS.md 2019-03-18 14:27:10 -07:00
86378284e1 fix for windows (thanks to chrothenbach) 2019-03-07 11:26:58 -08:00
79a246a58e add contributors md and move dev docs out 2019-03-06 15:06:30 -08:00
953530fc24 update readme with usage caveats and calls for research
This write-up was loosely inspired in part by Mitchell et al.’s work on
[Model Cards for Model Reporting](https://arxiv.org/abs/1810.03993).
Adding such model usage sections could be good practice in general for
open source research projects with potentially broad applications.
2019-03-06 15:06:30 -08:00
ed0dedcd55 update download stuff 2019-03-04 10:46:37 -08:00
8eb67930d7 Python download script (#89)
added python download script and modified requirements to add the modules needed. Tested in Windows Version 10.0.17134 Build 17134  and Ubuntu 18.04.1 LTS
2019-03-04 09:57:42 -08:00
0465394bd3 updates 2019-02-28 15:51:34 -08:00
9d1e704fec slight fix to batch size description 2019-02-26 23:05:11 -08:00
ed49f03778 Add documentation for help flags (#81)
add description for flags
2019-02-26 23:01:38 -08:00
c314ddab12 Minor: update readme
Example will `tee` stdout to `/tmp/samples` from conditional and
unconditional generation scripts.
2019-02-26 18:54:07 -08:00
c5b9c8924b Minor: update readme
Add note about setting PYTHONIOENCODING=UTF-8 env var for running
examples
2019-02-21 12:31:13 -08:00
68bf7a0036 add .gitattributes file to ensure files copied to docker container have LF line endings and all files stay unix on commit 2019-02-21 10:27:27 -08:00
a3aa7de9e1 add conditional samples with default settings 2019-02-20 20:51:19 -08:00
b6f943d37e update readme 2019-02-20 15:00:44 -08:00
946facf551 fixed seed arg to ensure reproducibility in conditional-samples model 2019-02-20 14:51:14 -08:00
2cf46d997d fixed unconditional sampling reproducibility issue 2019-02-20 14:50:52 -08:00
99af6d7092 Add a Dockerfile and document usage in README 2019-02-19 18:58:27 -08:00
23ed990b03 instructinos mention git clone 2019-02-19 18:05:57 -08:00
bf43e732f8 more warning 2019-02-19 17:57:33 -08:00
92ce9f2d9e shuffle headings 2019-02-19 17:57:01 -08:00
825aa3d338 separate out tensorflow install 2019-02-19 17:56:23 -08:00
fc0ee6da2f add conditional samples 2019-02-19 17:22:28 -08:00
aae26abd60 add license 2019-02-19 17:06:49 -08:00
6dab221dad reorganize and add temp 0.7 2019-02-19 00:45:05 -08:00
5b6468456e update README 2019-02-17 17:24:52 -08:00
9c3a78da74 Better example parameters for conditional sample command (#41)
This PR adds better initial parameters to the conditional sample generation command in the docs.

The results are pretty poor in the interactive script with the default settings.  

Now, you'll get better results if you run the interactive samples.
2019-02-16 14:23:13 -08:00
0aad2ab3f4 Fetch model using curl, add shebang to download_files.sh and mark it executable 2019-02-16 11:07:01 -08:00
2d0b62225c Update encoder.py to work on windows
This fixes https://github.com/openai/gpt-2/issues/26

```  File "C:\Users\James Pollack\Desktop\gpt-2\src\encoder.py", line 112, in get_encoder
    bpe_data = f.read()
  File "C:\Anaconda\envs\gpt-2\lib\encodings\cp1252.py", line 23, in decode
    return codecs.charmap_decode(input,self.errors,decoding_table)[0]
UnicodeDecodeError: 'charmap' codec can't decode byte 0x81 in position 66951: character maps to <undefined>```
2019-02-15 11:55:23 -08:00
ebd5894d22 error message 2019-02-15 10:48:19 -08:00
16095a6139 gpt-2 truncated samples 2019-02-15 10:32:58 -08:00
77c7ae5ee5 Fix ModuleNotFoundError: No module named 'src' in the interactive_conditional_samples.py 2019-02-14 13:20:24 -08:00
bb617c436e Fix ModuleNotFoundError: No module named 'src' 2019-02-14 13:20:24 -08:00
69354b7661 mkdir -p 2019-02-14 11:37:20 -08:00