Skip to content

Instantly share code, notes, and snippets.

@fo40225
Created May 1, 2020 06:57
Show Gist options
  • Save fo40225/d39d67cf4d5c99640bec9b1138649d30 to your computer and use it in GitHub Desktop.
Save fo40225/d39d67cf4d5c99640bec9b1138649d30 to your computer and use it in GitHub Desktop.
Radeon VII ROCm pytorch transformers bert
user@ubuntu:~$ sudo docker run -it -v $HOME:/data --privileged --rm \
> --device=/dev/kfd --device=/dev/dri --group-add video \
> rocm/pytorch:rocm3.3_ubuntu16.04_py3.6_pytorch
root@4373d0a3dd33:/# apt install cabextract
Reading package lists... Done
Building dependency tree
Reading state information... Done
The following additional packages will be installed:
libmspack0
The following NEW packages will be installed:
cabextract libmspack0
0 upgraded, 2 newly installed, 0 to remove and 32 not upgraded.
Need to get 58.9 kB of archives.
After this operation, 205 kB of additional disk space will be used.
Do you want to continue? [Y/n] y
Get:1 http://archive.ubuntu.com/ubuntu xenial-updates/main amd64 libmspack0 amd64 0.5-1ubuntu0.16.04.4 [37.4 kB]
Get:2 http://archive.ubuntu.com/ubuntu xenial/universe amd64 cabextract amd64 1.6-1 [21.4 kB]
Fetched 58.9 kB in 1s (48.4 kB/s)
debconf: delaying package configuration, since apt-utils is not installed
Selecting previously unselected package libmspack0:amd64.
(Reading database ... 30303 files and directories currently installed.)
Preparing to unpack .../libmspack0_0.5-1ubuntu0.16.04.4_amd64.deb ...
Unpacking libmspack0:amd64 (0.5-1ubuntu0.16.04.4) ...
Selecting previously unselected package cabextract.
Preparing to unpack .../cabextract_1.6-1_amd64.deb ...
Unpacking cabextract (1.6-1) ...
Processing triggers for libc-bin (2.23-0ubuntu10) ...
Setting up libmspack0:amd64 (0.5-1ubuntu0.16.04.4) ...
Setting up cabextract (1.6-1) ...
Processing triggers for libc-bin (2.23-0ubuntu10) ...
root@4373d0a3dd33:/# wget https://download.microsoft.com/download/D/4/6/D46FF87A-F6B9-4252-AA8B-3604ED519838/MSRParaphraseCorpus.msi
--2020-05-01 06:49:41-- https://download.microsoft.com/download/D/4/6/D46FF87A-F6B9-4252-AA8B-3604ED519838/MSRParaphraseCorpus.msi
Resolving download.microsoft.com (download.microsoft.com)... 163.28.224.147
Connecting to download.microsoft.com (download.microsoft.com)|163.28.224.147|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 1359872 (1.3M) [application/octet-stream]
Saving to: ‘MSRParaphraseCorpus.msi’
MSRParaphraseCorpus 100%[===================>] 1.30M --.-KB/s in 0.02s
2020-05-01 06:49:42 (60.7 MB/s) - ‘MSRParaphraseCorpus.msi’ saved [1359872/1359872]
root@4373d0a3dd33:/# mkdir MRPC
root@4373d0a3dd33:/# cabextract MSRParaphraseCorpus.msi -d MRPC
Extracting cabinet: MSRParaphraseCorpus.msi
extracting MRPC/_2D65ED66D69C42A28B021C3E24C1D8C0
extracting MRPC/_2DEC3DBE877E4DB192D17C0256E90F1D
extracting MRPC/_63DE49D9E7214609BE7E38DD145D8081
extracting MRPC/_B3CFEFE1C368459BA1D1B8A2FA07A16D
extracting MRPC/_C5BC91AAB1554DF3AF5E4105DE57C85A
extracting MRPC/_D18B15DC041F43D7925309EFFCFE0236
extracting MRPC/_D7B391F9EAFF4B1B8BCE8F21B20B1B61
All done, no errors.
root@4373d0a3dd33:/# cat MRPC/_2DEC3DBE877E4DB192D17C0256E90F1D | tr -d $'\r' > MRPC/msr_paraphrase_train.txt
root@4373d0a3dd33:/# cat MRPC/_D7B391F9EAFF4B1B8BCE8F21B20B1B61 | tr -d $'\r' > MRPC/msr_paraphrase_test.txt
root@4373d0a3dd33:/# rm MRPC/_*
root@4373d0a3dd33:/# rm MSRParaphraseCorpus.msi
root@4373d0a3dd33:/# wget https://gist.githubusercontent.com/W4ngatang/60c2bdb54d156a41194446737ce03e2e/raw/17b8dd0d724281ed7c3b2aeeda662b92809aadd5/download_glue_data.py
--2020-05-01 06:49:50-- https://gist.githubusercontent.com/W4ngatang/60c2bdb54d156a41194446737ce03e2e/raw/17b8dd0d724281ed7c3b2aeeda662b92809aadd5/download_glue_data.py
Resolving gist.githubusercontent.com (gist.githubusercontent.com)... 151.101.76.133
Connecting to gist.githubusercontent.com (gist.githubusercontent.com)|151.101.76.133|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 8225 (8.0K) [text/plain]
Saving to: ‘download_glue_data.py’
download_glue_data. 100%[===================>] 8.03K --.-KB/s in 0s
2020-05-01 06:49:51 (102 MB/s) - ‘download_glue_data.py’ saved [8225/8225]
root@4373d0a3dd33:/# python3.6 download_glue_data.py --path_to_mrpc MRPC
Downloading and extracting CoLA...
Completed!
Downloading and extracting SST...
Completed!
Processing MRPC...
Completed!
Downloading and extracting QQP...
Completed!
Downloading and extracting STS...
Completed!
Downloading and extracting MNLI...
Completed!
Downloading and extracting SNLI...
Completed!
Downloading and extracting QNLI...
Completed!
Downloading and extracting RTE...
Completed!
Downloading and extracting WNLI...
Completed!
Downloading and extracting diagnostic...
Completed!
root@4373d0a3dd33:/# git clone https://github.com/huggingface/transformers.git -b v2.8.0
Cloning into 'transformers'...
remote: Enumerating objects: 76, done.
remote: Counting objects: 100% (76/76), done.
remote: Compressing objects: 100% (62/62), done.
remote: Total 25885 (delta 37), reused 25 (delta 7), pack-reused 25809
Receiving objects: 100% (25885/25885), 15.27 MiB | 5.85 MiB/s, done.
Resolving deltas: 100% (18087/18087), done.
Checking connectivity... done.
Note: checking out '11c3257a18c4b5e1a3c1746eefd96f180358397b'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
root@4373d0a3dd33:/# cd transformers
root@4373d0a3dd33:/transformers# pip install .
Processing /transformers
Requirement already satisfied: numpy in /usr/local/lib/python3.6/dist-packages (from transformers==2.8.0)
Collecting tokenizers==0.5.2 (from transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/d1/3f/73c881ea4723e43c1e9acf317cf407fab3a278daab3a69c98dcac511c04f/tokenizers-0.5.2-cp36-cp36m-manylinux1_x86_64.whl (3.7MB)
100% |████████████████████████████████| 3.7MB 654kB/s
Collecting boto3 (from transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/73/db/c8902630b9a8ad3d17cb3de9ee347fe917094ec0fecd8e451076bbe1b4e0/boto3-1.13.0-py2.py3-none-any.whl (128kB)
100% |████████████████████████████████| 133kB 9.9MB/s
Requirement already satisfied: filelock in /usr/local/lib/python3.6/dist-packages (from transformers==2.8.0)
Requirement already satisfied: requests in /usr/local/lib/python3.6/dist-packages (from transformers==2.8.0)
Collecting tqdm>=4.27 (from transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/4a/1c/6359be64e8301b84160f6f6f7936bbfaaa5e9a4eab6cbc681db07600b949/tqdm-4.45.0-py2.py3-none-any.whl (60kB)
100% |████████████████████████████████| 61kB 10.1MB/s
Collecting regex!=2019.12.17 (from transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/1d/c1/c90beb2dbbfbf19f3634e16a441d5f11fa787bdf0748a35b8b88452c0e78/regex-2020.4.4-cp36-cp36m-manylinux1_x86_64.whl (679kB)
100% |████████████████████████████████| 686kB 3.6MB/s
Collecting sentencepiece (from transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/98/2c/8df20f3ac6c22ac224fff307ebc102818206c53fc454ecd37d8ac2060df5/sentencepiece-0.1.86-cp36-cp36m-manylinux1_x86_64.whl (1.0MB)
100% |████████████████████████████████| 1.0MB 2.5MB/s
Collecting sacremoses (from transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/99/50/93509f906a40bffd7d175f97fd75ea328ad9bd91f48f59c4bd084c94a25e/sacremoses-0.0.41.tar.gz (883kB)
100% |████████████████████████████████| 890kB 4.1MB/s
Collecting dataclasses (from transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/e1/d2/6f02df2616fd4016075f60157c7a0452b38d8f7938ae94343911e0fb0b09/dataclasses-0.7-py3-none-any.whl
Collecting s3transfer<0.4.0,>=0.3.0 (from boto3->transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/69/79/e6afb3d8b0b4e96cefbdc690f741d7dd24547ff1f94240c997a26fa908d3/s3transfer-0.3.3-py2.py3-none-any.whl (69kB)
100% |████████████████████████████████| 71kB 11.7MB/s
Collecting jmespath<1.0.0,>=0.7.1 (from boto3->transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/a3/43/1e939e1fcd87b827fe192d0c9fc25b48c5b3368902bfb913de7754b0dc03/jmespath-0.9.5-py2.py3-none-any.whl
Collecting botocore<1.17.0,>=1.16.0 (from boto3->transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/c6/06/ddf104c9da4de837dd0d871341e74c9f21542e0428318597c35bbf04bb2f/botocore-1.16.0-py2.py3-none-any.whl (6.2MB)
100% |████████████████████████████████| 6.2MB 403kB/s
Requirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0)
Requirement already satisfied: chardet<4,>=3.0.2 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0)
Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0)
Requirement already satisfied: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /usr/local/lib/python3.6/dist-packages (from requests->transformers==2.8.0)
Requirement already satisfied: six in /usr/local/lib/python3.6/dist-packages (from sacremoses->transformers==2.8.0)
Requirement already satisfied: click in /usr/local/lib/python3.6/dist-packages (from sacremoses->transformers==2.8.0)
Requirement already satisfied: joblib in /usr/local/lib/python3.6/dist-packages (from sacremoses->transformers==2.8.0)
Collecting docutils<0.16,>=0.10 (from botocore<1.17.0,>=1.16.0->boto3->transformers==2.8.0)
Downloading https://files.pythonhosted.org/packages/22/cd/a6aa959dca619918ccb55023b4cb151949c64d4d5d55b3f4ffd7eee0c6e8/docutils-0.15.2-py3-none-any.whl (547kB)
100% |████████████████████████████████| 552kB 4.0MB/s
Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/local/lib/python3.6/dist-packages (from botocore<1.17.0,>=1.16.0->boto3->transformers==2.8.0)
Building wheels for collected packages: sacremoses
Running setup.py bdist_wheel for sacremoses ... done
Stored in directory: /root/.cache/pip/wheels/22/5a/d4/b020a81249de7dc63758a34222feaa668dbe8ebfe9170cc9b1
Successfully built sacremoses
Installing collected packages: tokenizers, docutils, jmespath, botocore, s3transfer, boto3, tqdm, regex, sentencepiece, sacremoses, dataclasses, transformers
Running setup.py install for transformers ... done
Successfully installed boto3-1.13.0 botocore-1.16.0 dataclasses-0.7 docutils-0.15.2 jmespath-0.9.5 regex-2020.4.4 s3transfer-0.3.3 sacremoses-0.0.41 sentencepiece-0.1.86 tokenizers-0.5.2 tqdm-4.45.0 transformers-2.8.0
You are using pip version 9.0.1, however version 20.1 is available.
You should consider upgrading via the 'pip install --upgrade pip' command.
root@4373d0a3dd33:/transformers# pip install -r ./examples/requirements.txt
Collecting tensorboardX (from -r ./examples/requirements.txt (line 1))
Downloading https://files.pythonhosted.org/packages/35/f1/5843425495765c8c2dd0784a851a93ef204d314fc87bcc2bbb9f662a3ad1/tensorboardX-2.0-py2.py3-none-any.whl (195kB)
100% |████████████████████████████████| 204kB 2.8MB/s
Collecting tensorboard (from -r ./examples/requirements.txt (line 2))
Downloading https://files.pythonhosted.org/packages/1d/fd/4f3ca1516cbb3713259ef229abd9314bba0077ef6070285dde0dd1ed21b2/tensorboard-2.2.1-py3-none-any.whl (3.0MB)
100% |████████████████████████████████| 3.0MB 845kB/s
Requirement already satisfied: scikit-learn in /usr/local/lib/python3.6/dist-packages (from -r ./examples/requirements.txt (line 3))
Collecting seqeval (from -r ./examples/requirements.txt (line 4))
Downloading https://files.pythonhosted.org/packages/34/91/068aca8d60ce56dd9ba4506850e876aba5e66a6f2f29aa223224b50df0de/seqeval-0.0.12.tar.gz
Requirement already satisfied: psutil in /usr/local/lib/python3.6/dist-packages (from -r ./examples/requirements.txt (line 5))
Collecting sacrebleu (from -r ./examples/requirements.txt (line 6))
Downloading https://files.pythonhosted.org/packages/6e/9d/9846507837ca50ae20917f59d83b79246b8313bd19d4f5bf575ecb98132b/sacrebleu-1.4.9-py3-none-any.whl (60kB)
100% |████████████████████████████████| 61kB 10.7MB/s
Collecting rouge-score (from -r ./examples/requirements.txt (line 7))
Downloading https://files.pythonhosted.org/packages/d1/6d/2b9a64cba1e4e6ecd4effbf6834b2592b54dc813654f84029758e5daeeb5/rouge_score-0.0.3-py3-none-any.whl
Collecting tensorflow_datasets (from -r ./examples/requirements.txt (line 8))
Downloading https://files.pythonhosted.org/packages/bd/99/996b15ff5d11166c3516012838f569f78d57b71d4aac051caea826f6c7e0/tensorflow_datasets-3.1.0-py3-none-any.whl (3.3MB)
100% |████████████████████████████████| 3.3MB 740kB/s
Requirement already satisfied: protobuf>=3.8.0 in /usr/local/lib/python3.6/dist-packages (from tensorboardX->-r ./examples/requirements.txt (line 1))
Requirement already satisfied: numpy in /usr/local/lib/python3.6/dist-packages (from tensorboardX->-r ./examples/requirements.txt (line 1))
Requirement already satisfied: six in /usr/local/lib/python3.6/dist-packages (from tensorboardX->-r ./examples/requirements.txt (line 1))
Requirement already satisfied: requests<3,>=2.21.0 in /usr/local/lib/python3.6/dist-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: google-auth<2,>=1.6.3 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: markdown>=2.6.8 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: absl-py>=0.4 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: grpcio>=1.24.3 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: tensorboard-plugin-wit>=1.6.0 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: setuptools>=41.0.0 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: werkzeug>=0.11.15 in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: wheel>=0.26; python_version >= "3" in /root/.local/lib/python3.6/site-packages (from tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: scipy>=0.17.0 in /usr/local/lib/python3.6/dist-packages (from scikit-learn->-r ./examples/requirements.txt (line 3))
Requirement already satisfied: joblib>=0.11 in /usr/local/lib/python3.6/dist-packages (from scikit-learn->-r ./examples/requirements.txt (line 3))
Collecting Keras>=2.2.4 (from seqeval->-r ./examples/requirements.txt (line 4))
Downloading https://files.pythonhosted.org/packages/ad/fd/6bfe87920d7f4fd475acd28500a42482b6b84479832bdc0fe9e589a60ceb/Keras-2.3.1-py2.py3-none-any.whl (377kB)
100% |████████████████████████████████| 378kB 4.8MB/s
Requirement already satisfied: typing in /usr/local/lib/python3.6/dist-packages (from sacrebleu->-r ./examples/requirements.txt (line 6))
Collecting portalocker (from sacrebleu->-r ./examples/requirements.txt (line 6))
Downloading https://files.pythonhosted.org/packages/53/84/7b3146ec6378d28abc73ab484f09f47dfa008ad6f03f33d90a369f880e25/portalocker-1.7.0-py2.py3-none-any.whl
Collecting nltk (from rouge-score->-r ./examples/requirements.txt (line 7))
Downloading https://files.pythonhosted.org/packages/92/75/ce35194d8e3022203cca0d2f896dbb88689f9b3fce8e9f9cff942913519d/nltk-3.5.zip (1.4MB)
100% |████████████████████████████████| 1.4MB 1.7MB/s
Requirement already satisfied: attrs>=18.1.0 in /usr/local/lib/python3.6/dist-packages (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Requirement already satisfied: future in /usr/local/lib/python3.6/dist-packages (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Collecting promise (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Downloading https://files.pythonhosted.org/packages/cf/9c/fb5d48abfe5d791cd496e4242ebcf87a4bb2e0c3dcd6e0ae68c11426a528/promise-2.3.tar.gz
Requirement already satisfied: tqdm in /usr/local/lib/python3.6/dist-packages (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Collecting tensorflow-metadata (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Downloading https://files.pythonhosted.org/packages/57/12/213dc5982e45283591ee0cb535b08ff603200ba84643bbea0aaa2109ed7c/tensorflow_metadata-0.21.2-py2.py3-none-any.whl
Collecting wrapt (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Downloading https://files.pythonhosted.org/packages/82/f7/e43cefbe88c5fd371f4cf0cf5eb3feccd07515af9fd6cf7dbf1d1793a797/wrapt-1.12.1.tar.gz
Collecting termcolor (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Downloading https://files.pythonhosted.org/packages/8a/48/a76be51647d0eb9f10e2a4511bf3ffb8cc1e6b14e9e4fab46173aa79f981/termcolor-1.1.0.tar.gz
Collecting dill (from tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Downloading https://files.pythonhosted.org/packages/c7/11/345f3173809cea7f1a193bfbf02403fff250a3360e0e118a1630985e547d/dill-0.3.1.1.tar.gz (151kB)
100% |████████████████████████████████| 153kB 8.1MB/s
Requirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: chardet<4,>=3.0.2 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.21.0->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: rsa<4.1,>=3.1.4 in /root/.local/lib/python3.6/site-packages (from google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: cachetools<5.0,>=2.0.0 in /root/.local/lib/python3.6/site-packages (from google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: pyasn1-modules>=0.2.1 in /root/.local/lib/python3.6/site-packages (from google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: requests-oauthlib>=0.7.0 in /root/.local/lib/python3.6/site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard->-r ./examples/requirements.txt (line 2))
Collecting keras-preprocessing>=1.0.5 (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4))
Downloading https://files.pythonhosted.org/packages/28/6a/8c1f62c37212d9fc441a7e26736df51ce6f0e38455816445471f10da4f0a/Keras_Preprocessing-1.1.0-py2.py3-none-any.whl (41kB)
100% |████████████████████████████████| 51kB 12.3MB/s
Requirement already satisfied: pyyaml in /usr/local/lib/python3.6/dist-packages (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4))
Collecting keras-applications>=1.0.6 (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4))
Downloading https://files.pythonhosted.org/packages/71/e3/19762fdfc62877ae9102edf6342d71b28fbfd9dea3d2f96a882ce099b03f/Keras_Applications-1.0.8-py3-none-any.whl (50kB)
100% |████████████████████████████████| 51kB 10.0MB/s
Collecting h5py (from Keras>=2.2.4->seqeval->-r ./examples/requirements.txt (line 4))
Downloading https://files.pythonhosted.org/packages/60/06/cafdd44889200e5438b897388f3075b52a8ef01f28a17366d91de0fa2d05/h5py-2.10.0-cp36-cp36m-manylinux1_x86_64.whl (2.9MB)
100% |████████████████████████████████| 2.9MB 855kB/s
Requirement already satisfied: click in /usr/local/lib/python3.6/dist-packages (from nltk->rouge-score->-r ./examples/requirements.txt (line 7))
Requirement already satisfied: regex in /usr/local/lib/python3.6/dist-packages (from nltk->rouge-score->-r ./examples/requirements.txt (line 7))
Collecting googleapis-common-protos (from tensorflow-metadata->tensorflow_datasets->-r ./examples/requirements.txt (line 8))
Downloading https://files.pythonhosted.org/packages/05/46/168fd780f594a4d61122f7f3dc0561686084319ad73b4febbf02ae8b32cf/googleapis-common-protos-1.51.0.tar.gz
Requirement already satisfied: pyasn1>=0.1.3 in /root/.local/lib/python3.6/site-packages (from rsa<4.1,>=3.1.4->google-auth<2,>=1.6.3->tensorboard->-r ./examples/requirements.txt (line 2))
Requirement already satisfied: oauthlib>=3.0.0 in /root/.local/lib/python3.6/site-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard->-r ./examples/requirements.txt (line 2))
Building wheels for collected packages: seqeval, nltk, promise, wrapt, termcolor, dill, googleapis-common-protos
Running setup.py bdist_wheel for seqeval ... done
Stored in directory: /root/.cache/pip/wheels/4f/32/0a/df3b340a82583566975377d65e724895b3fad101a3fb729f68
Running setup.py bdist_wheel for nltk ... done
Stored in directory: /root/.cache/pip/wheels/ae/8c/3f/b1fe0ba04555b08b57ab52ab7f86023639a526d8bc8d384306
Running setup.py bdist_wheel for promise ... done
Stored in directory: /root/.cache/pip/wheels/19/49/34/c3c1e78bcb954c49e5ec0d31784fe63d14d427f316b12fbde9
Running setup.py bdist_wheel for wrapt ... done
Stored in directory: /root/.cache/pip/wheels/b1/c2/ed/d62208260edbd3fa7156545c00ef966f45f2063d0a84f8208a
Running setup.py bdist_wheel for termcolor ... done
Stored in directory: /root/.cache/pip/wheels/7c/06/54/bc84598ba1daf8f970247f550b175aaaee85f68b4b0c5ab2c6
Running setup.py bdist_wheel for dill ... done
Stored in directory: /root/.cache/pip/wheels/59/b1/91/f02e76c732915c4015ab4010f3015469866c1eb9b14058d8e7
Running setup.py bdist_wheel for googleapis-common-protos ... done
Stored in directory: /root/.cache/pip/wheels/2c/f9/7f/6eb87e636072bf467e25348bbeb96849333e6a080dca78f706
Successfully built seqeval nltk promise wrapt termcolor dill googleapis-common-protos
Installing collected packages: tensorboardX, tensorboard, keras-preprocessing, h5py, keras-applications, Keras, seqeval, portalocker, sacrebleu, nltk, rouge-score, promise, googleapis-common-protos, tensorflow-metadata, wrapt, termcolor, dill, tensorflow-datasets
Successfully installed Keras-2.3.1 dill-0.3.1.1 googleapis-common-protos-1.51.0 h5py-2.10.0 keras-applications-1.0.8 keras-preprocessing-1.1.0 nltk-3.5 portalocker-1.7.0 promise-2.3 rouge-score-0.0.3 sacrebleu-1.4.9 seqeval-0.0.12 tensorboard-2.2.1 tensorboardX-2.0 tensorflow-datasets-3.1.0 tensorflow-metadata-0.21.2 termcolor-1.1.0 wrapt-1.12.1
You are using pip version 9.0.1, however version 20.1 is available.
You should consider upgrading via the 'pip install --upgrade pip' command.
root@4373d0a3dd33:/transformers# export GLUE_DIR=../glue_data
root@4373d0a3dd33:/transformers# export TASK_NAME=MRPC
root@4373d0a3dd33:/transformers# python3.6 ./examples/run_glue.py \
> --model_type bert \
> --model_name_or_path bert-base-uncased \
> --task_name $TASK_NAME \
> --do_train \
> --do_eval \
> --do_lower_case \
> --data_dir $GLUE_DIR/$TASK_NAME \
> --max_seq_length 128 \
> --per_gpu_eval_batch_size=8 \
> --per_gpu_train_batch_size=8 \
> --learning_rate 2e-5 \
> --num_train_epochs 3.0 \
> --output_dir /tmp/$TASK_NAME/ \
>
05/01/2020 06:51:05 - WARNING - __main__ - Process rank: -1, device: cuda, n_gpu: 1, distributed training: False, 16-bits training: False
05/01/2020 06:51:06 - INFO - filelock - Lock 140686679170408 acquired on /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517.lock
05/01/2020 06:51:06 - INFO - transformers.file_utils - https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json not found in cache or force_download set to True, downloading to /root/.cache/torch/transformers/tmpza_j_4i1
Downloading: 100%|██████████████████████████████| 433/433 [00:00<00:00, 401kB/s]
05/01/2020 06:51:07 - INFO - transformers.file_utils - storing https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json in cache at /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517
05/01/2020 06:51:07 - INFO - transformers.file_utils - creating metadata file for /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517
05/01/2020 06:51:07 - INFO - filelock - Lock 140686679170408 released on /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517.lock
05/01/2020 06:51:07 - INFO - transformers.configuration_utils - loading configuration file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json from cache at /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517
05/01/2020 06:51:07 - INFO - transformers.configuration_utils - Model config BertConfig {
"_num_labels": 2,
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"bad_words_ids": null,
"bos_token_id": null,
"decoder_start_token_id": null,
"do_sample": false,
"early_stopping": false,
"eos_token_id": null,
"finetuning_task": "mrpc",
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"is_decoder": false,
"is_encoder_decoder": false,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"layer_norm_eps": 1e-12,
"length_penalty": 1.0,
"max_length": 20,
"max_position_embeddings": 512,
"min_length": 0,
"model_type": "bert",
"no_repeat_ngram_size": 0,
"num_attention_heads": 12,
"num_beams": 1,
"num_hidden_layers": 12,
"num_return_sequences": 1,
"output_attentions": false,
"output_hidden_states": false,
"output_past": true,
"pad_token_id": 0,
"prefix": null,
"pruned_heads": {},
"repetition_penalty": 1.0,
"task_specific_params": null,
"temperature": 1.0,
"top_k": 50,
"top_p": 1.0,
"torchscript": false,
"type_vocab_size": 2,
"use_bfloat16": false,
"vocab_size": 30522
}
05/01/2020 06:51:08 - INFO - transformers.configuration_utils - loading configuration file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-config.json from cache at /root/.cache/torch/transformers/4dad0251492946e18ac39290fcfe91b89d370fee250efe9521476438fe8ca185.7156163d5fdc189c3016baca0775ffce230789d7fa2a42ef516483e4ca884517
05/01/2020 06:51:08 - INFO - transformers.configuration_utils - Model config BertConfig {
"_num_labels": 2,
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"bad_words_ids": null,
"bos_token_id": null,
"decoder_start_token_id": null,
"do_sample": false,
"early_stopping": false,
"eos_token_id": null,
"finetuning_task": null,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"is_decoder": false,
"is_encoder_decoder": false,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"layer_norm_eps": 1e-12,
"length_penalty": 1.0,
"max_length": 20,
"max_position_embeddings": 512,
"min_length": 0,
"model_type": "bert",
"no_repeat_ngram_size": 0,
"num_attention_heads": 12,
"num_beams": 1,
"num_hidden_layers": 12,
"num_return_sequences": 1,
"output_attentions": false,
"output_hidden_states": false,
"output_past": true,
"pad_token_id": 0,
"prefix": null,
"pruned_heads": {},
"repetition_penalty": 1.0,
"task_specific_params": null,
"temperature": 1.0,
"top_k": 50,
"top_p": 1.0,
"torchscript": false,
"type_vocab_size": 2,
"use_bfloat16": false,
"vocab_size": 30522
}
05/01/2020 06:51:09 - INFO - filelock - Lock 140686678173008 acquired on /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084.lock
05/01/2020 06:51:09 - INFO - transformers.file_utils - https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-vocab.txt not found in cache or force_download set to True, downloading to /root/.cache/torch/transformers/tmp959pqfe0
Downloading: 100%|████████████████████████████| 232k/232k [00:01<00:00, 207kB/s]
05/01/2020 06:51:11 - INFO - transformers.file_utils - storing https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-vocab.txt in cache at /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084
05/01/2020 06:51:11 - INFO - transformers.file_utils - creating metadata file for /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084
05/01/2020 06:51:11 - INFO - filelock - Lock 140686678173008 released on /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084.lock
05/01/2020 06:51:11 - INFO - transformers.tokenization_utils - loading file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-vocab.txt from cache at /root/.cache/torch/transformers/26bc1ad6c0ac742e9b52263248f6d0f00068293b33709fae12320c0e35ccfbbb.542ce4285a40d23a559526243235df47c5f75c197f04f37d1a0c124c32c9a084
05/01/2020 06:51:12 - INFO - filelock - Lock 140687539655008 acquired on /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157.lock
05/01/2020 06:51:12 - INFO - transformers.file_utils - https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-pytorch_model.bin not found in cache or force_download set to True, downloading to /root/.cache/torch/transformers/tmp6iqjs947
Downloading: 100%|███████████████████████████| 440M/440M [01:34<00:00, 4.68MB/s]
05/01/2020 06:52:47 - INFO - transformers.file_utils - storing https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-pytorch_model.bin in cache at /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157
05/01/2020 06:52:47 - INFO - transformers.file_utils - creating metadata file for /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157
05/01/2020 06:52:47 - INFO - filelock - Lock 140687539655008 released on /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157.lock
05/01/2020 06:52:47 - INFO - transformers.modeling_utils - loading weights file https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased-pytorch_model.bin from cache at /root/.cache/torch/transformers/aa1ef1aede4482d0dbcd4d52baad8ae300e60902e88fcb0bebdec09afd232066.36ca03ab34a1a5d5fa7bc3d03d55c4fa650fed07220e2eeebc06ce58d0e9a157
05/01/2020 06:52:49 - INFO - transformers.modeling_utils - Weights of BertForSequenceClassification not initialized from pretrained model: ['classifier.weight', 'classifier.bias']
05/01/2020 06:52:49 - INFO - transformers.modeling_utils - Weights from pretrained model not used in BertForSequenceClassification: ['cls.predictions.bias', 'cls.predictions.transform.dense.weight', 'cls.predictions.transform.dense.bias', 'cls.predictions.decoder.weight', 'cls.seq_relationship.weight', 'cls.seq_relationship.bias', 'cls.predictions.transform.LayerNorm.weight', 'cls.predictions.transform.LayerNorm.bias']
/root/.local/lib/python3.6/site-packages/torch/cuda/__init__.py:87: UserWarning:
Found GPU0 Device 66af which is of cuda capability 3.0.
PyTorch no longer supports this GPU because it is too old.
The minimum cuda capability that we support is 3.5.
warnings.warn(old_gpu_warn % (d, name, major, capability[1]))
05/01/2020 06:52:49 - INFO - __main__ - Training/evaluation parameters Namespace(adam_epsilon=1e-08, cache_dir='', config_name='', data_dir='../glue_data/MRPC', device=device(type='cuda'), do_eval=True, do_lower_case=True, do_train=True, eval_all_checkpoints=False, evaluate_during_training=False, fp16=False, fp16_opt_level='O1', gradient_accumulation_steps=1, learning_rate=2e-05, local_rank=-1, logging_steps=500, max_grad_norm=1.0, max_seq_length=128, max_steps=-1, model_name_or_path='bert-base-uncased', model_type='bert', n_gpu=1, no_cuda=False, num_train_epochs=3.0, output_dir='/tmp/MRPC/', output_mode='classification', overwrite_cache=False, overwrite_output_dir=False, per_gpu_eval_batch_size=8, per_gpu_train_batch_size=8, save_steps=500, seed=42, server_ip='', server_port='', task_name='mrpc', tokenizer_name='', warmup_steps=0, weight_decay=0.0)
05/01/2020 06:52:49 - INFO - __main__ - Creating features from dataset file at ../glue_data/MRPC
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - LOOKING AT ../glue_data/MRPC/train.tsv
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - Writing example 0/3668
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-1
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 2572 3217 5831 5496 2010 2567 1010 3183 2002 2170 1000 1996 7409 1000 1010 1997 9969 4487 23809 3436 2010 3350 1012 102 7727 2000 2032 2004 2069 1000 1996 7409 1000 1010 2572 3217 5831 5496 2010 2567 1997 9969 4487 23809 3436 2010 3350 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 1 (id = 1)
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-2
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 9805 3540 11514 2050 3079 11282 2243 1005 1055 2077 4855 1996 4677 2000 3647 4576 1999 2687 2005 1002 1016 1012 1019 4551 1012 102 9805 3540 11514 2050 4149 11282 2243 1005 1055 1999 2786 2005 1002 6353 2509 2454 1998 2853 2009 2000 3647 4576 2005 1002 1015 1012 1022 4551 1999 2687 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 0 (id = 0)
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-3
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 2027 2018 2405 2019 15147 2006 1996 4274 2006 2238 2184 1010 5378 1996 6636 2005 5096 1010 2002 2794 1012 102 2006 2238 2184 1010 1996 2911 1005 1055 5608 2018 2405 2019 15147 2006 1996 4274 1010 5378 1996 14792 2005 5096 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 1 (id = 1)
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-4
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 2105 6021 19481 13938 2102 1010 21628 6661 2020 2039 2539 16653 1010 2030 1018 1012 1018 1003 1010 2012 1037 1002 1018 1012 5179 1010 2383 3041 2275 1037 2501 2152 1997 1037 1002 1018 1012 5401 1012 102 21628 6661 5598 2322 16653 1010 2030 1018 1012 1020 1003 1010 2000 2275 1037 2501 5494 2152 2012 1037 1002 1018 1012 5401 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 0 (id = 0)
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - guid: train-5
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - input_ids: 101 1996 4518 3123 1002 1016 1012 2340 1010 2030 2055 2340 3867 1010 2000 2485 5958 2012 1002 2538 1012 4868 2006 1996 2047 2259 4518 3863 1012 102 18720 1004 1041 13058 1012 6661 5598 1002 1015 1012 6191 2030 1022 3867 2000 1002 2538 1012 6021 2006 1996 2047 2259 4518 3863 2006 5958 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:52:49 - INFO - transformers.data.processors.glue - label: 1 (id = 1)
05/01/2020 06:52:51 - INFO - __main__ - Saving features into cached file ../glue_data/MRPC/cached_train_bert-base-uncased_128_mrpc
05/01/2020 06:52:52 - INFO - __main__ - ***** Running training *****
05/01/2020 06:52:52 - INFO - __main__ - Num examples = 3668
05/01/2020 06:52:52 - INFO - __main__ - Num Epochs = 3
05/01/2020 06:52:52 - INFO - __main__ - Instantaneous batch size per GPU = 8
05/01/2020 06:52:52 - INFO - __main__ - Total train batch size (w. parallel, distributed & accumulation) = 8
05/01/2020 06:52:52 - INFO - __main__ - Gradient Accumulation steps = 1
05/01/2020 06:52:52 - INFO - __main__ - Total optimization steps = 1377
Epoch: 0%| | 0/3 [00:00<?, ?it/s/pytorch/torch/csrc/utils/python_arg_parser.cpp:750: UserWarning: This overload of add_ is deprecated:
add_(Number alpha, Tensor other)
Consider using one of the following signatures instead:
add_(Tensor other, Number alpha)
TORCH_WARN_ONCE(msg);
Iteration: 100%|██████████████████████████████| 459/459 [01:07<00:00, 6.81it/s]
Epoch: 33%|████████████▋ | 1/3 [01:07<02:14, 67.42s/it/root/.local/lib/python3.6/site-packages/torch/optim/lr_scheduler.py:231: UserWarning: To get the last learning rate computed by the scheduler, please use `get_last_lr()`.
warnings.warn("To get the last learning rate computed by the scheduler, "
{"learning_rate": 1.2737835875090777e-05, "loss": 0.5262352135330438, "step": 500}
05/01/2020 06:54:04 - INFO - transformers.configuration_utils - Configuration saved in /tmp/MRPC/checkpoint-500/config.json
05/01/2020 06:54:04 - INFO - transformers.modeling_utils - Model weights saved in /tmp/MRPC/checkpoint-500/pytorch_model.bin
05/01/2020 06:54:04 - INFO - __main__ - Saving model checkpoint to /tmp/MRPC/checkpoint-500
/root/.local/lib/python3.6/site-packages/torch/optim/lr_scheduler.py:200: UserWarning: Please also save or load the state of the optimzer when saving or loading the scheduler.
warnings.warn(SAVE_STATE_WARNING, UserWarning)
05/01/2020 06:54:05 - INFO - __main__ - Saving optimizer and scheduler states to /tmp/MRPC/checkpoint-500
Iteration: 100%|██████████████████████████████| 459/459 [00:52<00:00, 8.73it/s]
Epoch: 67%|█████████████████████████▎ | 2/3 [02:00<01:02, 62.97s/it{"learning_rate": 5.475671750181555e-06, "loss": 0.3086102132201195, "step": 1000}
05/01/2020 06:55:01 - INFO - transformers.configuration_utils - Configuration saved in /tmp/MRPC/checkpoint-1000/config.json
05/01/2020 06:55:01 - INFO - transformers.modeling_utils - Model weights saved in /tmp/MRPC/checkpoint-1000/pytorch_model.bin
05/01/2020 06:55:02 - INFO - __main__ - Saving model checkpoint to /tmp/MRPC/checkpoint-1000
05/01/2020 06:55:02 - INFO - __main__ - Saving optimizer and scheduler states to /tmp/MRPC/checkpoint-1000
Iteration: 100%|██████████████████████████████| 459/459 [00:52<00:00, 8.74it/s]
Epoch: 100%|██████████████████████████████████████| 3/3 [02:52<00:00, 57.51s/it]
05/01/2020 06:55:45 - INFO - __main__ - global_step = 1377, average loss = 0.3641815590892006
05/01/2020 06:55:45 - INFO - __main__ - Saving model checkpoint to /tmp/MRPC/
05/01/2020 06:55:45 - INFO - transformers.configuration_utils - Configuration saved in /tmp/MRPC/config.json
05/01/2020 06:55:45 - INFO - transformers.modeling_utils - Model weights saved in /tmp/MRPC/pytorch_model.bin
05/01/2020 06:55:45 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json
05/01/2020 06:55:45 - INFO - transformers.configuration_utils - Model config BertConfig {
"_num_labels": 2,
"architectures": [
"BertForSequenceClassification"
],
"attention_probs_dropout_prob": 0.1,
"bad_words_ids": null,
"bos_token_id": null,
"decoder_start_token_id": null,
"do_sample": false,
"early_stopping": false,
"eos_token_id": null,
"finetuning_task": "mrpc",
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"is_decoder": false,
"is_encoder_decoder": false,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"layer_norm_eps": 1e-12,
"length_penalty": 1.0,
"max_length": 20,
"max_position_embeddings": 512,
"min_length": 0,
"model_type": "bert",
"no_repeat_ngram_size": 0,
"num_attention_heads": 12,
"num_beams": 1,
"num_hidden_layers": 12,
"num_return_sequences": 1,
"output_attentions": false,
"output_hidden_states": false,
"output_past": true,
"pad_token_id": 0,
"prefix": null,
"pruned_heads": {},
"repetition_penalty": 1.0,
"task_specific_params": null,
"temperature": 1.0,
"top_k": 50,
"top_p": 1.0,
"torchscript": false,
"type_vocab_size": 2,
"use_bfloat16": false,
"vocab_size": 30522
}
05/01/2020 06:55:45 - INFO - transformers.modeling_utils - loading weights file /tmp/MRPC/pytorch_model.bin
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - Model config BertConfig {
"_num_labels": 2,
"architectures": [
"BertForSequenceClassification"
],
"attention_probs_dropout_prob": 0.1,
"bad_words_ids": null,
"bos_token_id": null,
"decoder_start_token_id": null,
"do_sample": false,
"early_stopping": false,
"eos_token_id": null,
"finetuning_task": "mrpc",
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"is_decoder": false,
"is_encoder_decoder": false,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"layer_norm_eps": 1e-12,
"length_penalty": 1.0,
"max_length": 20,
"max_position_embeddings": 512,
"min_length": 0,
"model_type": "bert",
"no_repeat_ngram_size": 0,
"num_attention_heads": 12,
"num_beams": 1,
"num_hidden_layers": 12,
"num_return_sequences": 1,
"output_attentions": false,
"output_hidden_states": false,
"output_past": true,
"pad_token_id": 0,
"prefix": null,
"pruned_heads": {},
"repetition_penalty": 1.0,
"task_specific_params": null,
"temperature": 1.0,
"top_k": 50,
"top_p": 1.0,
"torchscript": false,
"type_vocab_size": 2,
"use_bfloat16": false,
"vocab_size": 30522
}
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Model name '/tmp/MRPC/' not found in model shortcut name list (bert-base-uncased, bert-large-uncased, bert-base-cased, bert-large-cased, bert-base-multilingual-uncased, bert-base-multilingual-cased, bert-base-chinese, bert-base-german-cased, bert-large-uncased-whole-word-masking, bert-large-cased-whole-word-masking, bert-large-uncased-whole-word-masking-finetuned-squad, bert-large-cased-whole-word-masking-finetuned-squad, bert-base-cased-finetuned-mrpc, bert-base-german-dbmdz-cased, bert-base-german-dbmdz-uncased, bert-base-finnish-cased-v1, bert-base-finnish-uncased-v1, bert-base-dutch-cased). Assuming '/tmp/MRPC/' is a path, a model identifier, or url to a directory containing tokenizer files.
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Didn't find file /tmp/MRPC/added_tokens.json. We won't load it.
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/vocab.txt
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file None
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/special_tokens_map.json
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/tokenizer_config.json
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - Model config BertConfig {
"_num_labels": 2,
"architectures": [
"BertForSequenceClassification"
],
"attention_probs_dropout_prob": 0.1,
"bad_words_ids": null,
"bos_token_id": null,
"decoder_start_token_id": null,
"do_sample": false,
"early_stopping": false,
"eos_token_id": null,
"finetuning_task": "mrpc",
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"is_decoder": false,
"is_encoder_decoder": false,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"layer_norm_eps": 1e-12,
"length_penalty": 1.0,
"max_length": 20,
"max_position_embeddings": 512,
"min_length": 0,
"model_type": "bert",
"no_repeat_ngram_size": 0,
"num_attention_heads": 12,
"num_beams": 1,
"num_hidden_layers": 12,
"num_return_sequences": 1,
"output_attentions": false,
"output_hidden_states": false,
"output_past": true,
"pad_token_id": 0,
"prefix": null,
"pruned_heads": {},
"repetition_penalty": 1.0,
"task_specific_params": null,
"temperature": 1.0,
"top_k": 50,
"top_p": 1.0,
"torchscript": false,
"type_vocab_size": 2,
"use_bfloat16": false,
"vocab_size": 30522
}
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Model name '/tmp/MRPC/' not found in model shortcut name list (bert-base-uncased, bert-large-uncased, bert-base-cased, bert-large-cased, bert-base-multilingual-uncased, bert-base-multilingual-cased, bert-base-chinese, bert-base-german-cased, bert-large-uncased-whole-word-masking, bert-large-cased-whole-word-masking, bert-large-uncased-whole-word-masking-finetuned-squad, bert-large-cased-whole-word-masking-finetuned-squad, bert-base-cased-finetuned-mrpc, bert-base-german-dbmdz-cased, bert-base-german-dbmdz-uncased, bert-base-finnish-cased-v1, bert-base-finnish-uncased-v1, bert-base-dutch-cased). Assuming '/tmp/MRPC/' is a path, a model identifier, or url to a directory containing tokenizer files.
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - Didn't find file /tmp/MRPC/added_tokens.json. We won't load it.
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/vocab.txt
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file None
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/special_tokens_map.json
05/01/2020 06:55:46 - INFO - transformers.tokenization_utils - loading file /tmp/MRPC/tokenizer_config.json
05/01/2020 06:55:46 - INFO - __main__ - Evaluate the following checkpoints: ['/tmp/MRPC/']
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - loading configuration file /tmp/MRPC/config.json
05/01/2020 06:55:46 - INFO - transformers.configuration_utils - Model config BertConfig {
"_num_labels": 2,
"architectures": [
"BertForSequenceClassification"
],
"attention_probs_dropout_prob": 0.1,
"bad_words_ids": null,
"bos_token_id": null,
"decoder_start_token_id": null,
"do_sample": false,
"early_stopping": false,
"eos_token_id": null,
"finetuning_task": "mrpc",
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"is_decoder": false,
"is_encoder_decoder": false,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"layer_norm_eps": 1e-12,
"length_penalty": 1.0,
"max_length": 20,
"max_position_embeddings": 512,
"min_length": 0,
"model_type": "bert",
"no_repeat_ngram_size": 0,
"num_attention_heads": 12,
"num_beams": 1,
"num_hidden_layers": 12,
"num_return_sequences": 1,
"output_attentions": false,
"output_hidden_states": false,
"output_past": true,
"pad_token_id": 0,
"prefix": null,
"pruned_heads": {},
"repetition_penalty": 1.0,
"task_specific_params": null,
"temperature": 1.0,
"top_k": 50,
"top_p": 1.0,
"torchscript": false,
"type_vocab_size": 2,
"use_bfloat16": false,
"vocab_size": 30522
}
05/01/2020 06:55:46 - INFO - transformers.modeling_utils - loading weights file /tmp/MRPC/pytorch_model.bin
05/01/2020 06:55:48 - INFO - __main__ - Creating features from dataset file at ../glue_data/MRPC
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - Writing example 0/408
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-1
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 2002 2056 1996 9440 2121 7903 2063 11345 2449 2987 1005 1056 4906 1996 2194 1005 1055 2146 1011 2744 3930 5656 1012 102 1000 1996 9440 2121 7903 2063 11345 2449 2515 2025 4906 2256 2146 1011 2744 3930 5656 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 1 (id = 1)
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-2
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 20201 22948 2056 10958 19053 4140 6283 1996 8956 6939 1998 2246 2830 2000 2478 2010 2146 2086 1997 2731 1999 1996 2162 1012 102 2010 2564 2056 2002 2001 1000 2531 3867 2369 2577 5747 1000 1998 2246 2830 2000 2478 2010 2086 1997 2731 1999 1996 2162 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 0 (id = 0)
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-3
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 1996 7922 2001 2012 12904 1012 6227 18371 2114 1996 18371 1010 4257 2006 1996 5219 1010 1998 2012 1015 1012 27054 2487 2114 1996 5364 23151 2278 1010 2036 4257 1012 102 1996 7922 2001 2012 12904 1012 6275 18371 16545 2100 1027 1010 8990 4257 2006 1996 5219 1010 1998 2012 1015 1012 23090 2487 2114 1996 5364 23151 2278 10381 2546 1027 1010 2091 1014 1012 1015 3867 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 0 (id = 0)
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-4
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 1996 10028 1011 25022 2080 2003 3403 2127 2255 2000 5630 2065 2009 2097 2203 5668 2063 1037 4018 1012 102 1996 10028 1011 25022 2080 2623 9317 2008 2009 2097 5630 1999 2255 3251 2000 2203 5668 2063 1037 4018 2077 1996 27419 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 1 (id = 1)
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - *** Example ***
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - guid: dev-5
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - input_ids: 101 2053 5246 2031 2042 2275 2005 1996 2942 2030 1996 4735 3979 1012 102 2053 5246 2031 2042 2275 2005 1996 4735 2030 2942 3572 1010 2021 17137 3051 2038 12254 2025 5905 1012 102 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - attention_mask: 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - token_type_ids: 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
05/01/2020 06:55:48 - INFO - transformers.data.processors.glue - label: 0 (id = 0)
05/01/2020 06:55:48 - INFO - __main__ - Saving features into cached file ../glue_data/MRPC/cached_dev_bert-base-uncased_128_mrpc
05/01/2020 06:55:48 - INFO - __main__ - ***** Running evaluation *****
05/01/2020 06:55:48 - INFO - __main__ - Num examples = 408
05/01/2020 06:55:48 - INFO - __main__ - Batch size = 8
Evaluating: 100%|███████████████████████████████| 51/51 [00:01<00:00, 31.97it/s]
05/01/2020 06:55:50 - INFO - __main__ - ***** Eval results *****
05/01/2020 06:55:50 - INFO - __main__ - acc = 0.8774509803921569
05/01/2020 06:55:50 - INFO - __main__ - acc_and_f1 = 0.8962084833933572
05/01/2020 06:55:50 - INFO - __main__ - f1 = 0.9149659863945577
root@4373d0a3dd33:/transformers#
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment