Created
May 1, 2020 06:57
-
-
Save fo40225/d39d67cf4d5c99640bec9b1138649d30 to your computer and use it in GitHub Desktop.
Radeon VII ROCm pytorch transformers bert
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
user@ubuntu:~$ sudo docker run -it -v $HOME:/data --privileged --rm \ | |
> --device=/dev/kfd --device=/dev/dri --group-add video \ | |
> rocm/pytorch:rocm3.3_ubuntu16.04_py3.6_pytorch | |
root@4373d0a3dd33:/# apt install cabextract | |
Reading package lists... Done | |
Building dependency tree | |
Reading state information... Done | |
The following additional packages will be installed: | |
libmspack0 | |
The following NEW packages will be installed: | |
cabextract libmspack0 | |
0 upgraded, 2 newly installed, 0 to remove and 32 not upgraded. | |
Need to get 58.9 kB of archives. | |
After this operation, 205 kB of additional disk space will be used. | |
Do you want to continue? [Y/n] y | |
Get:1 http://archive.ubuntu.com/ubuntu xenial-updates/main amd64 libmspack0 amd64 0.5-1ubuntu0.16.04.4 [37.4 kB] | |
Get:2 http://archive.ubuntu.com/ubuntu xenial/universe amd64 cabextract amd64 1.6-1 [21.4 kB] | |
Fetched 58.9 kB in 1s (48.4 kB/s) | |
debconf: delaying package configuration, since apt-utils is not installed | |
Selecting previously unselected package libmspack0:amd64. | |
(Reading database ... 30303 files and directories currently installed.) | |
Preparing to unpack .../libmspack0_0.5-1ubuntu0.16.04.4_amd64.deb ... | |
Unpacking libmspack0:amd64 (0.5-1ubuntu0.16.04.4) ... | |
Selecting previously unselected package cabextract. | |
Preparing to unpack .../cabextract_1.6-1_amd64.deb ... | |
Unpacking cabextract (1.6-1) ... | |
Processing triggers for libc-bin (2.23-0ubuntu10) ... | |
Setting up libmspack0:amd64 (0.5-1ubuntu0.16.04.4) ... | |
Setting up cabextract (1.6-1) ... | |
Processing triggers for libc-bin (2.23-0ubuntu10) ... | |
root@4373d0a3dd33:/# wget https://download.microsoft.com/download/D/4/6/D46FF87A-F6B9-4252-AA8B-3604ED519838/MSRParaphraseCorpus.msi | |
--2020-05-01 06:49:41-- https://download.microsoft.com/download/D/4/6/D46FF87A-F6B9-4252-AA8B-3604ED519838/MSRParaphraseCorpus.msi | |
Resolving download.microsoft.com (download.microsoft.com)... 163.28.224.147 | |
Connecting to download.microsoft.com (download.microsoft.com)|163.28.224.147|:443... connected. | |
HTTP request sent, awaiting response... 200 OK | |
Length: 1359872 (1.3M) [application/octet-stream] | |
Saving to: ‘MSRParaphraseCorpus.msi’ | |
MSRParaphraseCorpus 100%[===================>] 1.30M --.-KB/s in 0.02s | |
2020-05-01 06:49:42 (60.7 MB/s) - ‘MSRParaphraseCorpus.msi’ saved [1359872/1359872] | |
root@4373d0a3dd33:/# mkdir MRPC | |
root@4373d0a3dd33:/# cabextract MSRParaphraseCorpus.msi -d MRPC | |
Extracting cabinet: MSRParaphraseCorpus.msi | |
extracting MRPC/_2D65ED66D69C42A28B021C3E24C1D8C0 | |
extracting MRPC/_2DEC3DBE877E4DB192D17C0256E90F1D | |
extracting MRPC/_63DE49D9E7214609BE7E38DD145D8081 | |
extracting MRPC/_B3CFEFE1C368459BA1D1B8A2FA07A16D | |
extracting MRPC/_C5BC91AAB1554DF3AF5E4105DE57C85A | |
extracting MRPC/_D18B15DC041F43D7925309EFFCFE0236 | |
extracting MRPC/_D7B391F9EAFF4B1B8BCE8F21B20B1B61 | |
All done, no errors. | |
root@4373d0a3dd33:/# cat MRPC/_2DEC3DBE877E4DB192D17C0256E90F1D | tr -d $'\r' > MRPC/msr_paraphrase_train.txt | |
root@4373d0a3dd33:/# cat MRPC/_D7B391F9EAFF4B1B8BCE8F21B20B1B61 | tr -d $'\r' > MRPC/msr_paraphrase_test.txt | |
root@4373d0a3dd33:/# rm MRPC/_* | |
root@4373d0a3dd33:/# rm MSRParaphraseCorpus.msi | |
root@4373d0a3dd33:/# wget https://gist.githubusercontent.com/W4ngatang/60c2bdb54d156a41194446737ce03e2e/raw/17b8dd0d724281ed7c3b2aeeda662b92809aadd5/download_glue_data.py | |
--2020-05-01 06:49:50-- https://gist.githubusercontent.com/W4ngatang/60c2bdb54d156a41194446737ce03e2e/raw/17b8dd0d724281ed7c3b2aeeda662b92809aadd5/download_glue_data.py | |
Resolving gist.githubusercontent.com (gist.githubusercontent.com)... 151.101.76.133 | |
Connecting to gist.githubusercontent.com (gist.githubusercontent.com)|151.101.76.133|:443... connected. | |
HTTP request sent, awaiting response... 200 OK | |
Length: 8225 (8.0K) [text/plain] | |
Saving to: ‘download_glue_data.py’ | |
download_glue_data. 100%[===================>] 8.03K --.-KB/s in 0s | |
2020-05-01 06:49:51 (102 MB/s) - ‘download_glue_data.py’ saved [8225/8225] | |
root@4373d0a3dd33:/# python3.6 download_glue_data.py --path_to_mrpc MRPC | |
Downloading and extracting CoLA... | |
Completed! | |
Downloading and extracting SST... | |
Completed! | |
Processing MRPC... | |
Completed! | |
Downloading and extracting QQP... | |
Completed! | |
Downloading and extracting STS... | |
Completed! | |
Downloading and extracting MNLI... | |
Completed! | |
Downloading and extracting SNLI... | |
Completed! | |
Downloading and extracting QNLI... | |
Completed! | |
Downloading and extracting RTE... | |
Completed! | |
Downloading and extracting WNLI... | |
Completed! | |
Downloading and extracting diagnostic... | |
Completed! | |
root@4373d0a3dd33:/# git clone https://github.com/huggingface/transformers.git -b v2.8.0 | |
Cloning into 'transformers'... | |
remote: Enumerating objects: 76, done. | |
remote: Counting objects: 100% (76/76), done. | |
remote: Compressing objects: 100% (62/62), done. | |
remote: Total 25885 (delta 37), reused 25 (delta 7), pack-reused 25809 | |
Receiving objects: 100% (25885/25885), 15.27 MiB | 5.85 MiB/s, done. | |
Resolving deltas: 100% (18087/18087), done. | |
Checking connectivity... done. | |
Note: checking out '11c3257a18c4b5e1a3c1746eefd96f180358397b'. | |
You are in 'detached HEAD' state. You can look around, make experimental | |
changes and commit them, and you can discard any commits you make in this | |
state without impacting any branches by performing another checkout. | |
If you want to create a new branch to retain commits you create, you may | |
do so (now or later) by using -b with the checkout command again. Example: | |
git checkout -b <new-branch-name> | |
root@4373d0a3dd33:/# cd transformers | |
root@4373d0a3dd33:/transformers# pip install . | |
Processing /transformers | |
Requirement already satisfied: numpy in /usr/local/lib/python3.6/dist-packages (from transformers==2.8.0) | |
Collecting tokenizers==0.5.2 (from transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/d1/3f/73c881ea4723e43c1e9acf317cf407fab3a278daab3a69c98dcac511c04f/tokenizers-0.5.2-cp36-cp36m-manylinux1_x86_64.whl (3.7MB) | |
100% |████████████████████████████████| 3.7MB 654kB/s | |
Collecting boto3 (from transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/73/db/c8902630b9a8ad3d17cb3de9ee347fe917094ec0fecd8e451076bbe1b4e0/boto3-1.13.0-py2.py3-none-any.whl (128kB) | |
100% |████████████████████████████████| 133kB 9.9MB/s | |
Requirement already satisfied: filelock in /usr/local/lib/python3.6/dist-packages (from transformers==2.8.0) | |
Requirement already satisfied: requests in /usr/local/lib/python3.6/dist-packages (from transformers==2.8.0) | |
Collecting tqdm>=4.27 (from transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/4a/1c/6359be64e8301b84160f6f6f7936bbfaaa5e9a4eab6cbc681db07600b949/tqdm-4.45.0-py2.py3-none-any.whl (60kB) | |
100% |████████████████████████████████| 61kB 10.1MB/s | |
Collecting regex!=2019.12.17 (from transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/1d/c1/c90beb2dbbfbf19f3634e16a441d5f11fa787bdf0748a35b8b88452c0e78/regex-2020.4.4-cp36-cp36m-manylinux1_x86_64.whl (679kB) | |
100% |████████████████████████████████| 686kB 3.6MB/s | |
Collecting sentencepiece (from transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/98/2c/8df20f3ac6c22ac224fff307ebc102818206c53fc454ecd37d8ac2060df5/sentencepiece-0.1.86-cp36-cp36m-manylinux1_x86_64.whl (1.0MB) | |
100% |████████████████████████████████| 1.0MB 2.5MB/s | |
Collecting sacremoses (from transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/99/50/93509f906a40bffd7d175f97fd75ea328ad9bd91f48f59c4bd084c94a25e/sacremoses-0.0.41.tar.gz (883kB) | |
100% |████████████████████████████████| 890kB 4.1MB/s | |
Collecting dataclasses (from transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/e1/d2/6f02df2616fd4016075f60157c7a0452b38d8f7938ae94343911e0fb0b09/dataclasses-0.7-py3-none-any.whl | |
Collecting s3transfer<0.4.0,>=0.3.0 (from boto3->transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/69/79/e6afb3d8b0b4e96cefbdc690f741d7dd24547ff1f94240c997a26fa908d3/s3transfer-0.3.3-py2.py3-none-any.whl (69kB) | |
100% |████████████████████████████████| 71kB 11.7MB/s | |
Collecting jmespath<1.0.0,>=0.7.1 (from boto3->transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/a3/43/1e939e1fcd87b827fe192d0c9fc25b48c5b3368902bfb913de7754b0dc03/jmespath-0.9.5-py2.py3-none-any.whl | |
Collecting botocore<1.17.0,>=1.16.0 (from boto3->transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/c6/06/ddf104c9da4de837dd0d871341e74c9f21542e0428318597c35bbf04bb2f/botocore-1.16.0-py2.py3-none-any.whl (6.2MB) | |
100% |████████████████████████████████| 6.2MB 403kB/s | |
Requirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0) | |
Requirement already satisfied: chardet<4,>=3.0.2 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0) | |
Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0) | |
Requirement already satisfied: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0) | |
Requirement already satisfied: six in /usr/local/lib/python3.6/dist-packages (from sacremoses->transformers==2.8.0) | |
Requirement already satisfied: click in /usr/local/lib/python3.6/dist-packages (from sacremoses->transformers==2.8.0) | |
Requirement already satisfied: joblib in /usr/local/lib/python3.6/dist-packages (from sacremoses->transformers==2.8.0) | |
Collecting docutils<0.16,>=0.10 (from botocore<1.17.0,>=1.16.0->boto3->transformers==2.8.0) | |
Downloading https://files.pythonhosted.org/packages/22/cd/a6aa959dca619918ccb55023b4cb151949c64d4d5d55b3f4ffd7eee0c6e8/docutils-0.15.2-py3-none-any.whl (547kB) | |
100% |████████████████████████████████| 552kB 4.0MB/s | |
Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/local/lib/python3.6/dist-packages (from botocore<1.17.0,>=1.16.0->boto3->transformers==2.8.0) | |
Building wheels for collected packages: sacremoses | |
Running setup.py bdist_wheel for sacremoses ... done | |
Stored in directory: /root/.cache/pip/wheels/22/5a/d4/b020a81249de7dc63758a34222feaa668dbe8ebfe9170cc9b1 | |
Successfully built sacremoses | |
Installing collected packages: tokenizers, docutils, jmespath, botocore, s3transfer, boto3, tqdm, regex, sentencepiece, sacremoses, dataclasses, transformers | |
Running setup.py install for transformers ... done | |
Successfully installed boto3-1.13.0 botocore-1.16.0 dataclasses-0.7 docutils-0.15.2 jmespath-0.9.5 regex-2020.4.4 s3transfer-0.3.3 sacremoses-0.0.41 sentencepiece-0.1.86 tokenizers-0.5.2 tqdm-4.45.0 transformers-2.8.0 | |
You are using pip version 9.0.1, however version 20.1 is available. | |
You should consider upgrading via the 'pip install --upgrade pip' command. | |
root@4373d0a3dd33:/transformers# pip install -r ./examples/requirements.txt | |
Collecting tensorboardX (from -r ./examples/requirements.txt (line 1)) | |
Downloading https://files.pythonhosted.org/packages/35/f1/5843425495765c8c2dd0784a851a93ef204d314fc87bcc2bbb9f662a3ad1/tensorboardX-2.0-py2.py3-none-any.whl (195kB) | |
100% |████████████████████████████████| 204kB 2.8MB/s | |
Collecting tensorboard (from -r ./examples/requirements.txt (line 2)) | |
Downloading https://files.pythonhosted.org/packages/1d/fd/4f3ca1516cbb3713259ef229abd9314bba0077ef6070285dde0dd1ed21b2/tensorboard-2.2.1-py3-none-any.whl (3.0MB) | |
100% |████████████████████████████████| 3.0MB 845kB/s | |
Requirement already satisfied: scikit-learn in /usr/local/lib/python3.6/dist-packages (from -r ./examples/requirements.txt (line 3)) | |
Collecting seqeval (from -r ./examples/requirements.txt (line 4)) | |
Downloading https://files.pythonhosted.org/packages/34/91/068aca8d60ce56dd9ba4506850e876aba5e66a6f2f29aa223224b50df0de/seqeval-0.0.12.tar.gz | |
Requirement already satisfied: psutil in /usr/local/lib/python3.6/dist-packages (from -r ./examples/requirements.txt (line 5)) | |
Collecting sacrebleu (from -r ./examples/requirements.txt (line 6)) | |
Downloading https://files.pythonhosted.org/packages/6e/9d/9846507837ca50ae20917f59d83b79246b8313bd19d4f5bf575ecb98132b/sacrebleu-1.4.9-py3-none-any.whl (60kB) | |
100% |████████████████████████████████| 61kB 10.7MB/s | |
Collecting rouge-score (from -r ./examples/requirements.txt (line 7)) | |
Downloading https://files.pythonhosted.org/packages/d1/6d/2b9a64cba1e4e6ecd4effbf6834b2592b54dc813654f84029758e5daeeb5/rouge_score-0.0.3-py3-none-any.whl | |
Collecting tensorflow_datasets (from -r ./examples/requirements.txt (line 8)) | |
Downloading https://files.pythonhosted.org/packages/bd/99/996b15ff5d11166c3516012838f569f78d57b71d4aac051caea826f6c7e0/tensorflow_datasets-3.1.0-py3-none-any.whl (3.3MB) | |
100% |████████████████████████████████| 3.3MB 740kB/s | |
Requirement already satisfied: protobuf>=3.8.0 in /usr/local/lib/python3.6/dist-packages (from tensorboardX->-r ./examples/requirements.txt (line 1)) | |
Requirement already satisfied: numpy in /usr/local/lib/python3.6/dist-packages (from tensorboardX->-r ./examples/requirements.txt (line 1)) | |
Requirement already satisfied: six in /usr/local/lib/python3.6/dist-packages (from tensorboardX->-r ./examples/requirements.txt (line 1)) | |
Requirement already satisfied: requests<3,>=2.21.0 in /usr/local/lib/python3.6/dist-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: google-auth<2,>=1.6.3 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: markdown>=2.6.8 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: absl-py>=0.4 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: grpcio>=1.24.3 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: tensorboard-plugin-wit>=1.6.0 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: setuptools>=41.0.0 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: werkzeug>=0.11.15 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: wheel>=0.26; python_version >= "3" in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: scipy>=0.17.0 in /usr/local/lib/python3.6/dist-packages (from scikit-learn->-r ./examples/requirements.txt (line 3)) | |
Requirement already satisfied: joblib>=0.11 in /usr/local/lib/python3.6/dist-packages (from scikit-learn->-r ./examples/requirements.txt (line 3)) | |
Collecting Keras>=2.2.4 (from seqeval->-r ./examples/requirements.txt (line 4)) | |
Downloading https://files.pythonhosted.org/packages/ad/fd/6bfe87920d7f4fd475acd28500a42482b6b84479832bdc0fe9e589a60ceb/Keras-2.3.1-py2.py3-none-any.whl (377kB) | |
100% |████████████████████████████████| 378kB 4.8MB/s | |
Requirement already satisfied: typing in /usr/local/lib/python3.6/dist-packages (from sacrebleu->-r ./examples/requirements.txt (line 6)) | |
Collecting portalocker (from sacrebleu->-r ./examples/requirements.txt (line 6)) | |
Downloading https://files.pythonhosted.org/packages/53/84/7b3146ec6378d28abc73ab484f09f47dfa008ad6f03f33d90a369f880e25/portalocker-1.7.0-py2.py3-none-any.whl | |
Collecting nltk (from rouge-score->-r ./examples/requirements.txt (line 7)) | |
Downloading https://files.pythonhosted.org/packages/92/75/ce35194d8e3022203cca0d2f896dbb88689f9b3fce8e9f9cff942913519d/nltk-3.5.zip (1.4MB) | |
100% |████████████████████████████████| 1.4MB 1.7MB/s | |
Requirement already satisfied: attrs>=18.1.0 in /usr/local/lib/python3.6/dist-packages (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Requirement already satisfied: future in /usr/local/lib/python3.6/dist-packages (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Collecting promise (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Downloading https://files.pythonhosted.org/packages/cf/9c/fb5d48abfe5d791cd496e4242ebcf87a4bb2e0c3dcd6e0ae68c11426a528/promise-2.3.tar.gz | |
Requirement already satisfied: tqdm in /usr/local/lib/python3.6/dist-packages (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Collecting tensorflow-metadata (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Downloading https://files.pythonhosted.org/packages/57/12/213dc5982e45283591ee0cb535b08ff603200ba84643bbea0aaa2109ed7c/tensorflow_metadata-0.21.2-py2.py3-none-any.whl | |
Collecting wrapt (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Downloading https://files.pythonhosted.org/packages/82/f7/e43cefbe88c5fd371f4cf0cf5eb3feccd07515af9fd6cf7dbf1d1793a797/wrapt-1.12.1.tar.gz | |
Collecting termcolor (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Downloading https://files.pythonhosted.org/packages/8a/48/a76be51647d0eb9f10e2a4511bf3ffb8cc1e6b14e9e4fab46173aa79f981/termcolor-1.1.0.tar.gz | |
Collecting dill (from tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Downloading https://files.pythonhosted.org/packages/c7/11/345f3173809cea7f1a193bfbf02403fff250a3360e0e118a1630985e547d/dill-0.3.1.1.tar.gz (151kB) | |
100% |████████████████████████████████| 153kB 8.1MB/s | |
Requirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: chardet<4,>=3.0.2 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: rsa<4.1,>=3.1.4 in /root/.local/lib/python3.6/site-packages (from google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: cachetools<5.0,>=2.0.0 in /root/.local/lib/python3.6/site-packages (from google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: pyasn1-modules>=0.2.1 in /root/.local/lib/python3.6/site-packages (from google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: requests-oauthlib>=0.7.0 in /root/.local/lib/python3.6/site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Collecting keras-preprocessing>=1.0.5 (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4)) | |
Downloading https://files.pythonhosted.org/packages/28/6a/8c1f62c37212d9fc441a7e26736df51ce6f0e38455816445471f10da4f0a/Keras_Preprocessing-1.1.0-py2.py3-none-any.whl (41kB) | |
100% |████████████████████████████████| 51kB 12.3MB/s | |
Requirement already satisfied: pyyaml in /usr/local/lib/python3.6/dist-packages (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4)) | |
Collecting keras-applications>=1.0.6 (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4)) | |
Downloading https://files.pythonhosted.org/packages/71/e3/19762fdfc62877ae9102edf6342d71b28fbfd9dea3d2f96a882ce099b03f/Keras_Applications-1.0.8-py3-none-any.whl (50kB) | |
100% |████████████████████████████████| 51kB 10.0MB/s | |
Collecting h5py (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4)) | |
Downloading https://files.pythonhosted.org/packages/60/06/cafdd44889200e5438b897388f3075b52a8ef01f28a17366d91de0fa2d05/h5py-2.10.0-cp36-cp36m-manylinux1_x86_64.whl (2.9MB) | |
100% |████████████████████████████████| 2.9MB 855kB/s | |
Requirement already satisfied: click in /usr/local/lib/python3.6/dist-packages (from nltk->rouge-score->-r ./examples/requirements.txt (line 7)) | |
Requirement already satisfied: regex in /usr/local/lib/python3.6/dist-packages (from nltk->rouge-score->-r ./examples/requirements.txt (line 7)) | |
Collecting googleapis-common-protos (from tensorflow-metadata->tensorflow_datasets->-r ./examples/requirements.txt (line 8)) | |
Downloading https://files.pythonhosted.org/packages/05/46/168fd780f594a4d61122f7f3dc0561686084319ad73b4febbf02ae8b32cf/googleapis-common-protos-1.51.0.tar.gz | |
Requirement already satisfied: pyasn1>=0.1.3 in /root/.local/lib/python3.6/site-packages (from rsa<4.1,>=3.1.4->google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Requirement already satisfied: oauthlib>=3.0.0 in /root/.local/lib/python3.6/site-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard->-r ./examples/requirements.txt (line 2)) | |
Building wheels for collected packages: seqeval, nltk, promise, wrapt, termcolor, dill, googleapis-common-protos | |
Running setup.py bdist_wheel for seqeval ... done | |
Stored in directory: /root/.cache/pip/wheels/4f/32/0a/df3b340a82583566975377d65e724895b3fad101a3fb729f68 | |
Running setup.py bdist_wheel for nltk ... done | |
Stored in directory: /root/.cache/pip/wheels/ae/8c/3f/b1fe0ba04555b08b57ab52ab7f86023639a526d8bc8d384306 | |
Running setup.py bdist_wheel for promise ... done | |
Stored in directory: /root/.cache/pip/wheels/19/49/34/c3c1e78bcb954c49e5ec0d31784fe63d14d427f316b12fbde9 | |
Running setup.py bdist_wheel for wrapt ... done | |
Stored in directory: /root/.cache/pip/wheels/b1/c2/ed/d62208260edbd3fa7156545c00ef966f45f2063d0a84f8208a | |
Running setup.py bdist_wheel for termcolor ... done | |
Stored in directory: /root/.cache/pip/wheels/7c/06/54/bc84598ba1daf8f970247f550b175aaaee85f68b4b0c5ab2c6 | |
Running setup.py bdist_wheel for dill ... done | |
Stored in directory: /root/.cache/pip/wheels/59/b1/91/f02e76c732915c4015ab4010f3015469866c1eb9b14058d8e7 | |
Running setup.py bdist_wheel for googleapis-common-protos ... done | |
Stored in directory: /root/.cache/pip/wheels/2c/f9/7f/6eb87e636072bf467e25348bbeb96849333e6a080dca78f706 | |
Successfully built seqeval nltk promise wrapt termcolor dill googleapis-common-protos | |
Installing collected packages: tensorboardX, tensorboard, keras-preprocessing, h5py, keras-applications, Keras, seqeval, portalocker, sacrebleu, nltk, rouge-score, promise, googleapis-common-protos, tensorflow-metadata, wrapt, termcolor, dill, tensorflow-datasets | |
Successfully installed Keras-2.3.1 dill-0.3.1.1 googleapis-common-protos-1.51.0 h5py-2.10.0 keras-applications-1.0.8 keras-preprocessing-1.1.0 nltk-3.5 portalocker-1.7.0 promise-2.3 rouge-score-0.0.3 sacrebleu-1.4.9 seqeval-0.0.12 tensorboard-2.2.1 tensorboardX-2.0 tensorflow-datasets-3.1.0 tensorflow-metadata-0.21.2 termcolor-1.1.0 wrapt-1.12.1 | |
You are using pip version 9.0.1, however version 20.1 is available. | |
You should consider upgrading via the 'pip install --upgrade pip' command. | |
root@4373d0a3dd33:/transformers# export GLUE_DIR=../glue_data | |
root@4373d0a3dd33:/transformers# export TASK_NAME=MRPC | |
root@4373d0a3dd33:/transformers# python3.6 ./examples/run_glue.py \ | |
> --model_type bert \ | |
> --model_name_or_path bert-base-uncased \ | |
> --task_name $TASK_NAME \ | |
> --do_train \ | |
> --do_eval \ | |
> --do_lower_case \ | |
> --data_dir $GLUE_DIR/$TASK_NAME \ | |
> --max_seq_length 128 \ | |
> --per_gpu_eval_batch_size=8 \ | |
> --per_gpu_train_batch_size=8 \ | |
> --learning_rate 2e-5 \ | |
> --num_train_epochs 3.0 \ | |
> --output_dir /tmp/$TASK_NAME/ \ | |
> | |
05/01/2020 06:51:05 - WARNING - __main__ - Process rank: -1, device: cuda, n_gpu: 1, distributed training: False, 16-bits training: False | |
05/01/2020 06:51:06 - INFO - filelock - Lock 140686679170408 acquired on /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517.lock | |
05/01/2020 06:51:06 - INFO - transformers.file_utils - https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json not found in cache or force_download set to True, downloading to /root/.cache/torch/transformers/tmpza_j_4i1 | |
Downloading: 100%|██████████████████████████████| 433/433 [00:00<00:00, 401kB/s] | |
05/01/2020 06:51:07 - INFO - transformers.file_utils - storing https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json in cache at /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517 | |
05/01/2020 06:51:07 - INFO - transformers.file_utils - creating metadata file for /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517 | |
05/01/2020 06:51:07 - INFO - filelock - Lock 140686679170408 released on /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517.lock | |
05/01/2020 06:51:07 - INFO - transformers.configuration_utils - loading configuration file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json from cache at /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517 | |
05/01/2020 06:51:07 - INFO - transformers.configuration_utils - Model config BertConfig { | |
"_num_labels": 2, | |
"architectures": [ | |
"BertForMaskedLM" | |
], | |
"attention_probs_dropout_prob": 0.1, | |
"bad_words_ids": null, | |
"bos_token_id": null, | |
"decoder_start_token_id": null, | |
"do_sample": false, | |
"early_stopping": false, | |
"eos_token_id": null, | |
"finetuning_task": "mrpc", | |
"hidden_act": "gelu", | |
"hidden_dropout_prob": 0.1, | |
"hidden_size": 768, | |
"id2label": { | |
"0": "LABEL_0", | |
"1": "LABEL_1" | |
}, | |
"initializer_range": 0.02, | |
"intermediate_size": 3072, | |
"is_decoder": false, | |
"is_encoder_decoder": false, | |
"label2id": { | |
"LABEL_0": 0, | |
"LABEL_1": 1 | |
}, | |
"layer_norm_eps": 1e-12, | |
"length_penalty": 1.0, | |
"max_length": 20, | |
"max_position_embeddings": 512, | |
"min_length": 0, | |
"model_type": "bert", | |
"no_repeat_ngram_size": 0, | |
"num_attention_heads": 12, | |
"num_beams": 1, | |
"num_hidden_layers": 12, | |
"num_return_sequences": 1, | |
"output_attentions": false, | |
"output_hidden_states": false, | |
"output_past": true, | |
"pad_token_id": 0, | |
"prefix": null, | |
"pruned_heads": {}, | |
"repetition_penalty": 1.0, | |
"task_specific_params": null, | |
"temperature": 1.0, | |
"top_k": 50, | |
"top_p": 1.0, | |
"torchscript": false, | |
"type_vocab_size": 2, | |
"use_bfloat16": false, | |
"vocab_size": 30522 | |
} | |
05/01/2020 06:51:08 - INFO - transformers.configuration_utils - loading configuration file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json from cache at /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517 | |
05/01/2020 06:51:08 - INFO - transformers.configuration_utils - Model config BertConfig { | |
"_num_labels": 2, | |
"architectures": [ | |
"BertForMaskedLM" | |
], | |
"attention_probs_dropout_prob": 0.1, | |
"bad_words_ids": null, | |
"bos_token_id": null, | |
"decoder_start_token_id": null, | |
"do_sample": false, | |
"early_stopping": false, | |
"eos_token_id": null, | |
"finetuning_task": null, | |
"hidden_act": "gelu", | |
"hidden_dropout_prob": 0.1, | |
"hidden_size": 768, | |
"id2label": { | |
"0": "LABEL_0", | |
"1": "LABEL_1" | |
}, | |
"initializer_range": 0.02, | |
"intermediate_size": 3072, | |
"is_decoder": false, | |
"is_encoder_decoder": false, | |
"label2id": { | |
"LABEL_0": 0, | |
"LABEL_1": 1 | |
}, | |
"layer_norm_eps": 1e-12, | |
"length_penalty": 1.0, | |
"max_length": 20, | |
"max_position_embeddings": 512, | |
"min_length": 0, | |
"model_type": "bert", | |
"no_repeat_ngram_size": 0, | |
"num_attention_heads": 12, | |
"num_beams": 1, | |
"num_hidden_layers": 12, | |
"num_return_sequences": 1, | |
"output_attentions": false, | |
"output_hidden_states": false, | |
"output_past": true, | |
"pad_token_id": 0, | |
"prefix": null, | |
"pruned_heads": {}, | |
"repetition_penalty": 1.0, | |
"task_specific_params": null, | |
"temperature": 1.0, | |
"top_k": 50, | |
"top_p": 1.0, | |
"torchscript": false, | |
"type_vocab_size": 2, | |
"use_bfloat16": false, | |
"vocab_size": 30522 | |
} | |
05/01/2020 06:51:09 - INFO - filelock - Lock 140686678173008 acquired on /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084.lock | |
05/01/2020 06:51:09 - INFO - transformers.file_utils - https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-vocab.txt not found in cache or force_download set to True, downloading to /root/.cache/torch/transformers/tmp959pqfe0 | |
Downloading: 100%|████████████████████████████| 232k/232k [00:01<00:00, 207kB/s] | |
05/01/2020 06:51:11 - INFO - transformers.file_utils - storing https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-vocab.txt in cache at /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084 | |
05/01/2020 06:51:11 - INFO - transformers.file_utils - creating metadata file for /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084 | |
05/01/2020 06:51:11 - INFO - filelock - Lock 140686678173008 released on /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084.lock | |
05/01/2020 06:51:11 - INFO - transformers.tokenization_utils - loading file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-vocab.txt from cache at /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084 | |
05/01/2020 06:51:12 - INFO - filelock - Lock 140687539655008 acquired on /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157.lock | |
05/01/2020 06:51:12 - INFO - transformers.file_utils - https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-pytorch_model.bin not found in cache or force_download set to True, downloading to /root/.cache/torch/transformers/tmp6iqjs947 | |
Downloading: 100%|███████████████████████████| 440M/440M [01:34<00:00, 4.68MB/s] | |
05/01/2020 06:52:47 - INFO - transformers.file_utils - storing https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-pytorch_model.bin in cache at /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157 | |
05/01/2020 06:52:47 - INFO - transformers.file_utils - creating metadata file for /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157 | |
05/01/2020 06:52:47 - INFO - filelock - Lock 140687539655008 released on /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157.lock | |
05/01/2020 06:52:47 - INFO - transformers.modeling_utils - loading weights file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-pytorch_model.bin from cache at /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157 | |
05/01/2020 06:52:49 - INFO - transformers.modeling_utils - Weights of BertForSequenceClassification not initialized from pretrained model: ['classifier.weight', 'classifier.bias'] | |
05/01/2020 06:52:49 - INFO - transformers.modeling_utils - Weights from pretrained model not used in BertForSequenceClassification: ['cls.predictions.bias', 'cls.predictions.transform.dense.weight', 'cls.predictions.transform.dense.bias', 'cls.predictions.decoder.weight', 'cls.seq_relationship.weight', 'cls.seq_relationship.bias', 'cls.predictions.transform.LayerNorm.weight', 'cls.predictions.transform.LayerNorm.bias'] | |
/root/.local/lib/python3.6/site-packages/torch/cuda/__init__.py:87: UserWarning: | |
Found GPU0 Device 66af which is of cuda capability 3.0. | |
PyTorch no longer supports this GPU because it is too old. | |
The minimum cuda capability that we support is 3.5. | |
warnings.warn(old_gpu_warn % (d, name, major, capability[1])) | |
05/01/2020 06:52:49 - INFO - __main__ - Training/evaluation parameters Namespace(adam_epsilon=1e-08, cache_dir='', config_name='', data_dir='../glue_data/MRPC', device=device(type='cuda'), do_eval=True, do_lower_case=True, do_train=True, eval_all_checkpoints=False, evaluate_during_training=False, fp16=False, fp16_opt_level='O1', gradient_accumulation_steps=1, learning_rate=2e-05, local_rank=-1, logging_steps=500, max_grad_norm=1.0, max_seq_length=128, max_steps=-1, model_name_or_path='bert-base-uncased', model_type='bert', n_gpu=1, no_cuda=False, num_train_epochs=3.0, output_dir='/tmp/MRPC/', output_mode='classification', overwrite_cache=False, overwrite_output_dir=False, per_gpu_eval_batch_size=8, per_gpu_train_batch_size=8, save_steps=500, seed=42, server_ip='', server_port='', task_name='mrpc', tokenizer_name='', warmup_steps=0, weight_decay=0.0) | |
05/01/2020 06:52:49 - INFO - __main__ - Creating features from dataset file at ../glue_data/MRPC | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - LOOKING AT ../glue_data/MRPC/train.tsv | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - Writing example 0/3668 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-1 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 2572 3217 5831 5496 2010 2567 1010 3183 2002 2170 1000 1996 7409 1000 1010 1997 9969 4487 23809 3436 2010 3350 1012 102 7727 2000 2032 2004 2069 1000 1996 7409 1000 1010 2572 3217 5831 5496 2010 2567 1997 9969 4487 23809 3436 2010 3350 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 1 (id = 1) | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-2 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 9805 3540 11514 2050 3079 11282 2243 1005 1055 2077 4855 1996 4677 2000 3647 4576 1999 2687 2005 1002 1016 1012 1019 4551 1012 102 9805 3540 11514 2050 4149 11282 2243 1005 1055 1999 2786 2005 1002 6353 2509 2454 1998 2853 2009 2000 3647 4576 2005 1002 1015 1012 1022 4551 1999 2687 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 0 (id = 0) | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-3 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 2027 2018 2405 2019 15147 2006 1996 4274 2006 2238 2184 1010 5378 1996 6636 2005 5096 1010 2002 2794 1012 102 2006 2238 2184 1010 1996 2911 1005 1055 5608 2018 2405 2019 15147 2006 1996 4274 1010 5378 1996 14792 2005 5096 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 1 (id = 1) | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-4 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 2105 6021 19481 13938 2102 1010 21628 6661 2020 2039 2539 16653 1010 2030 1018 1012 1018 1003 1010 2012 1037 1002 1018 1012 5179 1010 2383 3041 2275 1037 2501 2152 1997 1037 1002 1018 1012 5401 1012 102 21628 6661 5598 2322 16653 1010 2030 1018 1012 1020 1003 1010 2000 2275 1037 2501 5494 2152 2012 1037 1002 1018 1012 5401 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 0 (id = 0) | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-5 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 1996 4518 3123 1002 1016 1012 2340 1010 2030 2055 2340 3867 1010 2000 2485 5958 2012 1002 2538 1012 4868 2006 1996 2047 2259 4518 3863 1012 102 18720 1004 1041 13058 1012 6661 5598 1002 1015 1012 6191 2030 1022 3867 2000 1002 2538 1012 6021 2006 1996 2047 2259 4518 3863 2006 5958 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 1 (id = 1) | |
05/01/2020 06:52:51 - INFO - __main__ - Saving features into cached file ../glue_data/MRPC/cached_train_bert-base-uncased_128_mrpc | |
05/01/2020 06:52:52 - INFO - __main__ - ***** Running training ***** | |
05/01/2020 06:52:52 - INFO - __main__ - Num examples = 3668 | |
05/01/2020 06:52:52 - INFO - __main__ - Num Epochs = 3 | |
05/01/2020 06:52:52 - INFO - __main__ - Instantaneous batch size per GPU = 8 | |
05/01/2020 06:52:52 - INFO - __main__ - Total train batch size (w. parallel, distributed & accumulation) = 8 | |
05/01/2020 06:52:52 - INFO - __main__ - Gradient Accumulation steps = 1 | |
05/01/2020 06:52:52 - INFO - __main__ - Total optimization steps = 1377 | |
Epoch: 0%| | 0/3 [00:00<?, ?it/s/pytorch/torch/csrc/utils/python_arg_parser.cpp:750: UserWarning: This overload of add_ is deprecated: | |
add_(Number alpha, Tensor other) | |
Consider using one of the following signatures instead: | |
add_(Tensor other, Number alpha) | |
TORCH_WARN_ONCE(msg); | |
Iteration: 100%|██████████████████████████████| 459/459 [01:07<00:00, 6.81it/s] | |
Epoch: 33%|████████████▋ | 1/3 [01:07<02:14, 67.42s/it/root/.local/lib/python3.6/site-packages/torch/optim/lr_scheduler.py:231: UserWarning: To get the last learning rate computed by the scheduler, please use `get_last_lr()`. | |
warnings.warn("To get the last learning rate computed by the scheduler, " | |
{"learning_rate": 1.2737835875090777e-05, "loss": 0.5262352135330438, "step": 500} | |
05/01/2020 06:54:04 - INFO - transformers.configuration_utils - Configuration saved in /tmp/MRPC/checkpoint-500/config.json | |
05/01/2020 06:54:04 - INFO - transformers.modeling_utils - Model weights saved in /tmp/MRPC/checkpoint-500/pytorch_model.bin | |
05/01/2020 06:54:04 - INFO - __main__ - Saving model checkpoint to /tmp/MRPC/checkpoint-500 | |
/root/.local/lib/python3.6/site-packages/torch/optim/lr_scheduler.py:200: UserWarning: Please also save or load the state of the optimzer when saving or loading the scheduler. | |
warnings.warn(SAVE_STATE_WARNING, UserWarning) | |
05/01/2020 06:54:05 - INFO - __main__ - Saving optimizer and scheduler states to /tmp/MRPC/checkpoint-500 | |
Iteration: 100%|██████████████████████████████| 459/459 [00:52<00:00, 8.73it/s] | |
Epoch: 67%|█████████████████████████▎ | 2/3 [02:00<01:02, 62.97s/it{"learning_rate": 5.475671750181555e-06, "loss": 0.3086102132201195, "step": 1000} | |
05/01/2020 06:55:01 - INFO - transformers.configuration_utils - Configuration saved in /tmp/MRPC/checkpoint-1000/config.json | |
05/01/2020 06:55:01 - INFO - transformers.modeling_utils - Model weights saved in /tmp/MRPC/checkpoint-1000/pytorch_model.bin | |
05/01/2020 06:55:02 - INFO - __main__ - Saving model checkpoint to /tmp/MRPC/checkpoint-1000 | |
05/01/2020 06:55:02 - INFO - __main__ - Saving optimizer and scheduler states to /tmp/MRPC/checkpoint-1000 | |
Iteration: 100%|██████████████████████████████| 459/459 [00:52<00:00, 8.74it/s] | |
Epoch: 100%|██████████████████████████████████████| 3/3 [02:52<00:00, 57.51s/it] | |
05/01/2020 06:55:45 - INFO - __main__ - global_step = 1377, average loss = 0.3641815590892006 | |
05/01/2020 06:55:45 - INFO - __main__ - Saving model checkpoint to /tmp/MRPC/ | |
05/01/2020 06:55:45 - INFO - transformers.configuration_utils - Configuration saved in /tmp/MRPC/config.json | |
05/01/2020 06:55:45 - INFO - transformers.modeling_utils - Model weights saved in /tmp/MRPC/pytorch_model.bin | |
05/01/2020 06:55:45 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json | |
05/01/2020 06:55:45 - INFO - transformers.configuration_utils - Model config BertConfig { | |
"_num_labels": 2, | |
"architectures": [ | |
"BertForSequenceClassification" | |
], | |
"attention_probs_dropout_prob": 0.1, | |
"bad_words_ids": null, | |
"bos_token_id": null, | |
"decoder_start_token_id": null, | |
"do_sample": false, | |
"early_stopping": false, | |
"eos_token_id": null, | |
"finetuning_task": "mrpc", | |
"hidden_act": "gelu", | |
"hidden_dropout_prob": 0.1, | |
"hidden_size": 768, | |
"id2label": { | |
"0": "LABEL_0", | |
"1": "LABEL_1" | |
}, | |
"initializer_range": 0.02, | |
"intermediate_size": 3072, | |
"is_decoder": false, | |
"is_encoder_decoder": false, | |
"label2id": { | |
"LABEL_0": 0, | |
"LABEL_1": 1 | |
}, | |
"layer_norm_eps": 1e-12, | |
"length_penalty": 1.0, | |
"max_length": 20, | |
"max_position_embeddings": 512, | |
"min_length": 0, | |
"model_type": "bert", | |
"no_repeat_ngram_size": 0, | |
"num_attention_heads": 12, | |
"num_beams": 1, | |
"num_hidden_layers": 12, | |
"num_return_sequences": 1, | |
"output_attentions": false, | |
"output_hidden_states": false, | |
"output_past": true, | |
"pad_token_id": 0, | |
"prefix": null, | |
"pruned_heads": {}, | |
"repetition_penalty": 1.0, | |
"task_specific_params": null, | |
"temperature": 1.0, | |
"top_k": 50, | |
"top_p": 1.0, | |
"torchscript": false, | |
"type_vocab_size": 2, | |
"use_bfloat16": false, | |
"vocab_size": 30522 | |
} | |
05/01/2020 06:55:45 - INFO - transformers.modeling_utils - loading weights file /tmp/MRPC/pytorch_model.bin | |
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json | |
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - Model config BertConfig { | |
"_num_labels": 2, | |
"architectures": [ | |
"BertForSequenceClassification" | |
], | |
"attention_probs_dropout_prob": 0.1, | |
"bad_words_ids": null, | |
"bos_token_id": null, | |
"decoder_start_token_id": null, | |
"do_sample": false, | |
"early_stopping": false, | |
"eos_token_id": null, | |
"finetuning_task": "mrpc", | |
"hidden_act": "gelu", | |
"hidden_dropout_prob": 0.1, | |
"hidden_size": 768, | |
"id2label": { | |
"0": "LABEL_0", | |
"1": "LABEL_1" | |
}, | |
"initializer_range": 0.02, | |
"intermediate_size": 3072, | |
"is_decoder": false, | |
"is_encoder_decoder": false, | |
"label2id": { | |
"LABEL_0": 0, | |
"LABEL_1": 1 | |
}, | |
"layer_norm_eps": 1e-12, | |
"length_penalty": 1.0, | |
"max_length": 20, | |
"max_position_embeddings": 512, | |
"min_length": 0, | |
"model_type": "bert", | |
"no_repeat_ngram_size": 0, | |
"num_attention_heads": 12, | |
"num_beams": 1, | |
"num_hidden_layers": 12, | |
"num_return_sequences": 1, | |
"output_attentions": false, | |
"output_hidden_states": false, | |
"output_past": true, | |
"pad_token_id": 0, | |
"prefix": null, | |
"pruned_heads": {}, | |
"repetition_penalty": 1.0, | |
"task_specific_params": null, | |
"temperature": 1.0, | |
"top_k": 50, | |
"top_p": 1.0, | |
"torchscript": false, | |
"type_vocab_size": 2, | |
"use_bfloat16": false, | |
"vocab_size": 30522 | |
} | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Model name '/tmp/MRPC/' not found in model shortcut name list (bert-base-uncased, bert-large-uncased, bert-base-cased, bert-large-cased, bert-base-multilingual-uncased, bert-base-multilingual-cased, bert-base-chinese, bert-base-german-cased, bert-large-uncased-whole-word-masking, bert-large-cased-whole-word-masking, bert-large-uncased-whole-word-masking-finetuned-squad, bert-large-cased-whole-word-masking-finetuned-squad, bert-base-cased-finetuned-mrpc, bert-base-german-dbmdz-cased, bert-base-german-dbmdz-uncased, bert-base-finnish-cased-v1, bert-base-finnish-uncased-v1, bert-base-dutch-cased). Assuming '/tmp/MRPC/' is a path, a model identifier, or url to a directory containing tokenizer files. | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Didn't find file /tmp/MRPC/added_tokens.json. We won't load it. | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/vocab.txt | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file None | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/special_tokens_map.json | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/tokenizer_config.json | |
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json | |
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - Model config BertConfig { | |
"_num_labels": 2, | |
"architectures": [ | |
"BertForSequenceClassification" | |
], | |
"attention_probs_dropout_prob": 0.1, | |
"bad_words_ids": null, | |
"bos_token_id": null, | |
"decoder_start_token_id": null, | |
"do_sample": false, | |
"early_stopping": false, | |
"eos_token_id": null, | |
"finetuning_task": "mrpc", | |
"hidden_act": "gelu", | |
"hidden_dropout_prob": 0.1, | |
"hidden_size": 768, | |
"id2label": { | |
"0": "LABEL_0", | |
"1": "LABEL_1" | |
}, | |
"initializer_range": 0.02, | |
"intermediate_size": 3072, | |
"is_decoder": false, | |
"is_encoder_decoder": false, | |
"label2id": { | |
"LABEL_0": 0, | |
"LABEL_1": 1 | |
}, | |
"layer_norm_eps": 1e-12, | |
"length_penalty": 1.0, | |
"max_length": 20, | |
"max_position_embeddings": 512, | |
"min_length": 0, | |
"model_type": "bert", | |
"no_repeat_ngram_size": 0, | |
"num_attention_heads": 12, | |
"num_beams": 1, | |
"num_hidden_layers": 12, | |
"num_return_sequences": 1, | |
"output_attentions": false, | |
"output_hidden_states": false, | |
"output_past": true, | |
"pad_token_id": 0, | |
"prefix": null, | |
"pruned_heads": {}, | |
"repetition_penalty": 1.0, | |
"task_specific_params": null, | |
"temperature": 1.0, | |
"top_k": 50, | |
"top_p": 1.0, | |
"torchscript": false, | |
"type_vocab_size": 2, | |
"use_bfloat16": false, | |
"vocab_size": 30522 | |
} | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Model name '/tmp/MRPC/' not found in model shortcut name list (bert-base-uncased, bert-large-uncased, bert-base-cased, bert-large-cased, bert-base-multilingual-uncased, bert-base-multilingual-cased, bert-base-chinese, bert-base-german-cased, bert-large-uncased-whole-word-masking, bert-large-cased-whole-word-masking, bert-large-uncased-whole-word-masking-finetuned-squad, bert-large-cased-whole-word-masking-finetuned-squad, bert-base-cased-finetuned-mrpc, bert-base-german-dbmdz-cased, bert-base-german-dbmdz-uncased, bert-base-finnish-cased-v1, bert-base-finnish-uncased-v1, bert-base-dutch-cased). Assuming '/tmp/MRPC/' is a path, a model identifier, or url to a directory containing tokenizer files. | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Didn't find file /tmp/MRPC/added_tokens.json. We won't load it. | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/vocab.txt | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file None | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/special_tokens_map.json | |
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/tokenizer_config.json | |
05/01/2020 06:55:46 - INFO - __main__ - Evaluate the following checkpoints: ['/tmp/MRPC/'] | |
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json | |
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - Model config BertConfig { | |
"_num_labels": 2, | |
"architectures": [ | |
"BertForSequenceClassification" | |
], | |
"attention_probs_dropout_prob": 0.1, | |
"bad_words_ids": null, | |
"bos_token_id": null, | |
"decoder_start_token_id": null, | |
"do_sample": false, | |
"early_stopping": false, | |
"eos_token_id": null, | |
"finetuning_task": "mrpc", | |
"hidden_act": "gelu", | |
"hidden_dropout_prob": 0.1, | |
"hidden_size": 768, | |
"id2label": { | |
"0": "LABEL_0", | |
"1": "LABEL_1" | |
}, | |
"initializer_range": 0.02, | |
"intermediate_size": 3072, | |
"is_decoder": false, | |
"is_encoder_decoder": false, | |
"label2id": { | |
"LABEL_0": 0, | |
"LABEL_1": 1 | |
}, | |
"layer_norm_eps": 1e-12, | |
"length_penalty": 1.0, | |
"max_length": 20, | |
"max_position_embeddings": 512, | |
"min_length": 0, | |
"model_type": "bert", | |
"no_repeat_ngram_size": 0, | |
"num_attention_heads": 12, | |
"num_beams": 1, | |
"num_hidden_layers": 12, | |
"num_return_sequences": 1, | |
"output_attentions": false, | |
"output_hidden_states": false, | |
"output_past": true, | |
"pad_token_id": 0, | |
"prefix": null, | |
"pruned_heads": {}, | |
"repetition_penalty": 1.0, | |
"task_specific_params": null, | |
"temperature": 1.0, | |
"top_k": 50, | |
"top_p": 1.0, | |
"torchscript": false, | |
"type_vocab_size": 2, | |
"use_bfloat16": false, | |
"vocab_size": 30522 | |
} | |
05/01/2020 06:55:46 - INFO - transformers.modeling_utils - loading weights file /tmp/MRPC/pytorch_model.bin | |
05/01/2020 06:55:48 - INFO - __main__ - Creating features from dataset file at ../glue_data/MRPC | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - Writing example 0/408 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-1 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 2002 2056 1996 9440 2121 7903 2063 11345 2449 2987 1005 1056 4906 1996 2194 1005 1055 2146 1011 2744 3930 5656 1012 102 1000 1996 9440 2121 7903 2063 11345 2449 2515 2025 4906 2256 2146 1011 2744 3930 5656 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 1 (id = 1) | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-2 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 20201 22948 2056 10958 19053 4140 6283 1996 8956 6939 1998 2246 2830 2000 2478 2010 2146 2086 1997 2731 1999 1996 2162 1012 102 2010 2564 2056 2002 2001 1000 2531 3867 2369 2577 5747 1000 1998 2246 2830 2000 2478 2010 2086 1997 2731 1999 1996 2162 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 0 (id = 0) | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-3 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 1996 7922 2001 2012 12904 1012 6227 18371 2114 1996 18371 1010 4257 2006 1996 5219 1010 1998 2012 1015 1012 27054 2487 2114 1996 5364 23151 2278 1010 2036 4257 1012 102 1996 7922 2001 2012 12904 1012 6275 18371 16545 2100 1027 1010 8990 4257 2006 1996 5219 1010 1998 2012 1015 1012 23090 2487 2114 1996 5364 23151 2278 10381 2546 1027 1010 2091 1014 1012 1015 3867 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 0 (id = 0) | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-4 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 1996 10028 1011 25022 2080 2003 3403 2127 2255 2000 5630 2065 2009 2097 2203 5668 2063 1037 4018 1012 102 1996 10028 1011 25022 2080 2623 9317 2008 2009 2097 5630 1999 2255 3251 2000 2203 5668 2063 1037 4018 2077 1996 27419 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 1 (id = 1) | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example *** | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-5 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 2053 5246 2031 2042 2275 2005 1996 2942 2030 1996 4735 3979 1012 102 2053 5246 2031 2042 2275 2005 1996 4735 2030 2942 3572 1010 2021 17137 3051 2038 12254 2025 5905 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | |
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 0 (id = 0) | |
05/01/2020 06:55:48 - INFO - __main__ - Saving features into cached file ../glue_data/MRPC/cached_dev_bert-base-uncased_128_mrpc | |
05/01/2020 06:55:48 - INFO - __main__ - ***** Running evaluation ***** | |
05/01/2020 06:55:48 - INFO - __main__ - Num examples = 408 | |
05/01/2020 06:55:48 - INFO - __main__ - Batch size = 8 | |
Evaluating: 100%|███████████████████████████████| 51/51 [00:01<00:00, 31.97it/s] | |
05/01/2020 06:55:50 - INFO - __main__ - ***** Eval results ***** | |
05/01/2020 06:55:50 - INFO - __main__ - acc = 0.8774509803921569 | |
05/01/2020 06:55:50 - INFO - __main__ - acc_and_f1 = 0.8962084833933572 | |
05/01/2020 06:55:50 - INFO - __main__ - f1 = 0.9149659863945577 | |
root@4373d0a3dd33:/transformers# |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment