osiria commited on
Commit
be7ed73
1 Parent(s): faefc92

Upload 6 files

Browse files
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "XLMRobertaForQuestionAnswering"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "xlm-roberta",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 1,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.31.0.dev0",
23
+ "type_vocab_size": 1,
24
+ "use_cache": true,
25
+ "vocab_size": 30498
26
+ }
osiria_minilm_l6_h384_italian_qa_evaluation.ipynb ADDED
@@ -0,0 +1,839 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "nbformat": 4,
3
+ "nbformat_minor": 0,
4
+ "metadata": {
5
+ "colab": {
6
+ "provenance": [],
7
+ "gpuType": "T4"
8
+ },
9
+ "kernelspec": {
10
+ "name": "python3",
11
+ "display_name": "Python 3"
12
+ },
13
+ "language_info": {
14
+ "name": "python"
15
+ },
16
+ "accelerator": "GPU"
17
+ },
18
+ "cells": [
19
+ {
20
+ "cell_type": "markdown",
21
+ "source": [
22
+ "# osiria/minilm-italian-l6-h384-question-answering\n",
23
+ "## test set metrics"
24
+ ],
25
+ "metadata": {
26
+ "id": "oub2ir8ZDkrH"
27
+ }
28
+ },
29
+ {
30
+ "cell_type": "code",
31
+ "execution_count": 1,
32
+ "metadata": {
33
+ "colab": {
34
+ "base_uri": "https://localhost:8080/"
35
+ },
36
+ "id": "02K9xO6Gvto1",
37
+ "outputId": "3f72da34-6228-451a-c656-3f9fbb358e7f"
38
+ },
39
+ "outputs": [
40
+ {
41
+ "output_type": "stream",
42
+ "name": "stdout",
43
+ "text": [
44
+ "Collecting git+https://github.com/huggingface/transformers\n",
45
+ " Cloning https://github.com/huggingface/transformers to /tmp/pip-req-build-zw2jat5c\n",
46
+ " Running command git clone --filter=blob:none --quiet https://github.com/huggingface/transformers /tmp/pip-req-build-zw2jat5c\n",
47
+ " Resolved https://github.com/huggingface/transformers to commit df5c5c62ae253055336f5bb0828ca8e3e15ab6bd\n",
48
+ " Installing build dependencies ... \u001b[?25l\u001b[?25hdone\n",
49
+ " Getting requirements to build wheel ... \u001b[?25l\u001b[?25hdone\n",
50
+ " Preparing metadata (pyproject.toml) ... \u001b[?25l\u001b[?25hdone\n",
51
+ "Requirement already satisfied: filelock in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (3.13.1)\n",
52
+ "Requirement already satisfied: huggingface-hub<1.0,>=0.19.3 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (0.19.4)\n",
53
+ "Requirement already satisfied: numpy>=1.17 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (1.23.5)\n",
54
+ "Requirement already satisfied: packaging>=20.0 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (23.2)\n",
55
+ "Requirement already satisfied: pyyaml>=5.1 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (6.0.1)\n",
56
+ "Requirement already satisfied: regex!=2019.12.17 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (2023.6.3)\n",
57
+ "Requirement already satisfied: requests in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (2.31.0)\n",
58
+ "Requirement already satisfied: tokenizers<0.19,>=0.14 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (0.15.0)\n",
59
+ "Requirement already satisfied: safetensors>=0.3.1 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (0.4.1)\n",
60
+ "Requirement already satisfied: tqdm>=4.27 in /usr/local/lib/python3.10/dist-packages (from transformers==4.36.0.dev0) (4.66.1)\n",
61
+ "Requirement already satisfied: fsspec>=2023.5.0 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub<1.0,>=0.19.3->transformers==4.36.0.dev0) (2023.6.0)\n",
62
+ "Requirement already satisfied: typing-extensions>=3.7.4.3 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub<1.0,>=0.19.3->transformers==4.36.0.dev0) (4.5.0)\n",
63
+ "Requirement already satisfied: charset-normalizer<4,>=2 in /usr/local/lib/python3.10/dist-packages (from requests->transformers==4.36.0.dev0) (3.3.2)\n",
64
+ "Requirement already satisfied: idna<4,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests->transformers==4.36.0.dev0) (3.6)\n",
65
+ "Requirement already satisfied: urllib3<3,>=1.21.1 in /usr/local/lib/python3.10/dist-packages (from requests->transformers==4.36.0.dev0) (2.0.7)\n",
66
+ "Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests->transformers==4.36.0.dev0) (2023.11.17)\n",
67
+ "Building wheels for collected packages: transformers\n",
68
+ " Building wheel for transformers (pyproject.toml) ... \u001b[?25l\u001b[?25hdone\n",
69
+ " Created wheel for transformers: filename=transformers-4.36.0.dev0-py3-none-any.whl size=8195433 sha256=8bf5185dab843815883ed25caa99c87242dfa4ee86bdfdc0be3e42a6cfa8a17c\n",
70
+ " Stored in directory: /tmp/pip-ephem-wheel-cache-_xtobubu/wheels/c0/14/d6/6c9a5582d2ac191ec0a483be151a4495fe1eb2a6706ca49f1b\n",
71
+ "Successfully built transformers\n",
72
+ "Installing collected packages: transformers\n",
73
+ " Attempting uninstall: transformers\n",
74
+ " Found existing installation: transformers 4.35.2\n",
75
+ " Uninstalling transformers-4.35.2:\n",
76
+ " Successfully uninstalled transformers-4.35.2\n",
77
+ "Successfully installed transformers-4.36.0.dev0\n"
78
+ ]
79
+ }
80
+ ],
81
+ "source": [
82
+ "!pip install git+https://github.com/huggingface/transformers"
83
+ ]
84
+ },
85
+ {
86
+ "cell_type": "code",
87
+ "source": [
88
+ "!pip install datasets"
89
+ ],
90
+ "metadata": {
91
+ "colab": {
92
+ "base_uri": "https://localhost:8080/"
93
+ },
94
+ "id": "VKgybrUswa1q",
95
+ "outputId": "033a0aae-2528-4994-90b0-06c6d01bfdb3"
96
+ },
97
+ "execution_count": 2,
98
+ "outputs": [
99
+ {
100
+ "output_type": "stream",
101
+ "name": "stdout",
102
+ "text": [
103
+ "Collecting datasets\n",
104
+ " Downloading datasets-2.15.0-py3-none-any.whl (521 kB)\n",
105
+ "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m521.2/521.2 kB\u001b[0m \u001b[31m3.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
106
+ "\u001b[?25hRequirement already satisfied: numpy>=1.17 in /usr/local/lib/python3.10/dist-packages (from datasets) (1.23.5)\n",
107
+ "Requirement already satisfied: pyarrow>=8.0.0 in /usr/local/lib/python3.10/dist-packages (from datasets) (9.0.0)\n",
108
+ "Collecting pyarrow-hotfix (from datasets)\n",
109
+ " Downloading pyarrow_hotfix-0.6-py3-none-any.whl (7.9 kB)\n",
110
+ "Collecting dill<0.3.8,>=0.3.0 (from datasets)\n",
111
+ " Downloading dill-0.3.7-py3-none-any.whl (115 kB)\n",
112
+ "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m115.3/115.3 kB\u001b[0m \u001b[31m15.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
113
+ "\u001b[?25hRequirement already satisfied: pandas in /usr/local/lib/python3.10/dist-packages (from datasets) (1.5.3)\n",
114
+ "Requirement already satisfied: requests>=2.19.0 in /usr/local/lib/python3.10/dist-packages (from datasets) (2.31.0)\n",
115
+ "Requirement already satisfied: tqdm>=4.62.1 in /usr/local/lib/python3.10/dist-packages (from datasets) (4.66.1)\n",
116
+ "Requirement already satisfied: xxhash in /usr/local/lib/python3.10/dist-packages (from datasets) (3.4.1)\n",
117
+ "Collecting multiprocess (from datasets)\n",
118
+ " Downloading multiprocess-0.70.15-py310-none-any.whl (134 kB)\n",
119
+ "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m134.8/134.8 kB\u001b[0m \u001b[31m19.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
120
+ "\u001b[?25hRequirement already satisfied: fsspec[http]<=2023.10.0,>=2023.1.0 in /usr/local/lib/python3.10/dist-packages (from datasets) (2023.6.0)\n",
121
+ "Requirement already satisfied: aiohttp in /usr/local/lib/python3.10/dist-packages (from datasets) (3.9.1)\n",
122
+ "Requirement already satisfied: huggingface-hub>=0.18.0 in /usr/local/lib/python3.10/dist-packages (from datasets) (0.19.4)\n",
123
+ "Requirement already satisfied: packaging in /usr/local/lib/python3.10/dist-packages (from datasets) (23.2)\n",
124
+ "Requirement already satisfied: pyyaml>=5.1 in /usr/local/lib/python3.10/dist-packages (from datasets) (6.0.1)\n",
125
+ "Requirement already satisfied: attrs>=17.3.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets) (23.1.0)\n",
126
+ "Requirement already satisfied: multidict<7.0,>=4.5 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets) (6.0.4)\n",
127
+ "Requirement already satisfied: yarl<2.0,>=1.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets) (1.9.3)\n",
128
+ "Requirement already satisfied: frozenlist>=1.1.1 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets) (1.4.0)\n",
129
+ "Requirement already satisfied: aiosignal>=1.1.2 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets) (1.3.1)\n",
130
+ "Requirement already satisfied: async-timeout<5.0,>=4.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets) (4.0.3)\n",
131
+ "Requirement already satisfied: filelock in /usr/local/lib/python3.10/dist-packages (from huggingface-hub>=0.18.0->datasets) (3.13.1)\n",
132
+ "Requirement already satisfied: typing-extensions>=3.7.4.3 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub>=0.18.0->datasets) (4.5.0)\n",
133
+ "Requirement already satisfied: charset-normalizer<4,>=2 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->datasets) (3.3.2)\n",
134
+ "Requirement already satisfied: idna<4,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->datasets) (3.6)\n",
135
+ "Requirement already satisfied: urllib3<3,>=1.21.1 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->datasets) (2.0.7)\n",
136
+ "Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->datasets) (2023.11.17)\n",
137
+ "Requirement already satisfied: python-dateutil>=2.8.1 in /usr/local/lib/python3.10/dist-packages (from pandas->datasets) (2.8.2)\n",
138
+ "Requirement already satisfied: pytz>=2020.1 in /usr/local/lib/python3.10/dist-packages (from pandas->datasets) (2023.3.post1)\n",
139
+ "Requirement already satisfied: six>=1.5 in /usr/local/lib/python3.10/dist-packages (from python-dateutil>=2.8.1->pandas->datasets) (1.16.0)\n",
140
+ "Installing collected packages: pyarrow-hotfix, dill, multiprocess, datasets\n",
141
+ "Successfully installed datasets-2.15.0 dill-0.3.7 multiprocess-0.70.15 pyarrow-hotfix-0.6\n"
142
+ ]
143
+ }
144
+ ]
145
+ },
146
+ {
147
+ "cell_type": "code",
148
+ "source": [
149
+ "!pip install accelerate"
150
+ ],
151
+ "metadata": {
152
+ "colab": {
153
+ "base_uri": "https://localhost:8080/"
154
+ },
155
+ "id": "xYEW0Hjqz2lo",
156
+ "outputId": "a0e36c52-c360-4a01-8618-1489551286f8"
157
+ },
158
+ "execution_count": 3,
159
+ "outputs": [
160
+ {
161
+ "output_type": "stream",
162
+ "name": "stdout",
163
+ "text": [
164
+ "Collecting accelerate\n",
165
+ " Downloading accelerate-0.25.0-py3-none-any.whl (265 kB)\n",
166
+ "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m265.7/265.7 kB\u001b[0m \u001b[31m2.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
167
+ "\u001b[?25hRequirement already satisfied: numpy>=1.17 in /usr/local/lib/python3.10/dist-packages (from accelerate) (1.23.5)\n",
168
+ "Requirement already satisfied: packaging>=20.0 in /usr/local/lib/python3.10/dist-packages (from accelerate) (23.2)\n",
169
+ "Requirement already satisfied: psutil in /usr/local/lib/python3.10/dist-packages (from accelerate) (5.9.5)\n",
170
+ "Requirement already satisfied: pyyaml in /usr/local/lib/python3.10/dist-packages (from accelerate) (6.0.1)\n",
171
+ "Requirement already satisfied: torch>=1.10.0 in /usr/local/lib/python3.10/dist-packages (from accelerate) (2.1.0+cu118)\n",
172
+ "Requirement already satisfied: huggingface-hub in /usr/local/lib/python3.10/dist-packages (from accelerate) (0.19.4)\n",
173
+ "Requirement already satisfied: safetensors>=0.3.1 in /usr/local/lib/python3.10/dist-packages (from accelerate) (0.4.1)\n",
174
+ "Requirement already satisfied: filelock in /usr/local/lib/python3.10/dist-packages (from torch>=1.10.0->accelerate) (3.13.1)\n",
175
+ "Requirement already satisfied: typing-extensions in /usr/local/lib/python3.10/dist-packages (from torch>=1.10.0->accelerate) (4.5.0)\n",
176
+ "Requirement already satisfied: sympy in /usr/local/lib/python3.10/dist-packages (from torch>=1.10.0->accelerate) (1.12)\n",
177
+ "Requirement already satisfied: networkx in /usr/local/lib/python3.10/dist-packages (from torch>=1.10.0->accelerate) (3.2.1)\n",
178
+ "Requirement already satisfied: jinja2 in /usr/local/lib/python3.10/dist-packages (from torch>=1.10.0->accelerate) (3.1.2)\n",
179
+ "Requirement already satisfied: fsspec in /usr/local/lib/python3.10/dist-packages (from torch>=1.10.0->accelerate) (2023.6.0)\n",
180
+ "Requirement already satisfied: triton==2.1.0 in /usr/local/lib/python3.10/dist-packages (from torch>=1.10.0->accelerate) (2.1.0)\n",
181
+ "Requirement already satisfied: requests in /usr/local/lib/python3.10/dist-packages (from huggingface-hub->accelerate) (2.31.0)\n",
182
+ "Requirement already satisfied: tqdm>=4.42.1 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub->accelerate) (4.66.1)\n",
183
+ "Requirement already satisfied: MarkupSafe>=2.0 in /usr/local/lib/python3.10/dist-packages (from jinja2->torch>=1.10.0->accelerate) (2.1.3)\n",
184
+ "Requirement already satisfied: charset-normalizer<4,>=2 in /usr/local/lib/python3.10/dist-packages (from requests->huggingface-hub->accelerate) (3.3.2)\n",
185
+ "Requirement already satisfied: idna<4,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests->huggingface-hub->accelerate) (3.6)\n",
186
+ "Requirement already satisfied: urllib3<3,>=1.21.1 in /usr/local/lib/python3.10/dist-packages (from requests->huggingface-hub->accelerate) (2.0.7)\n",
187
+ "Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests->huggingface-hub->accelerate) (2023.11.17)\n",
188
+ "Requirement already satisfied: mpmath>=0.19 in /usr/local/lib/python3.10/dist-packages (from sympy->torch>=1.10.0->accelerate) (1.3.0)\n",
189
+ "Installing collected packages: accelerate\n",
190
+ "Successfully installed accelerate-0.25.0\n"
191
+ ]
192
+ }
193
+ ]
194
+ },
195
+ {
196
+ "cell_type": "code",
197
+ "source": [
198
+ "!pip install evaluate"
199
+ ],
200
+ "metadata": {
201
+ "colab": {
202
+ "base_uri": "https://localhost:8080/"
203
+ },
204
+ "id": "9OFz3JdR0cQ_",
205
+ "outputId": "137e11b1-d6ce-4ba7-89b9-1ce61fac9964"
206
+ },
207
+ "execution_count": 4,
208
+ "outputs": [
209
+ {
210
+ "output_type": "stream",
211
+ "name": "stdout",
212
+ "text": [
213
+ "Collecting evaluate\n",
214
+ " Downloading evaluate-0.4.1-py3-none-any.whl (84 kB)\n",
215
+ "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m84.1/84.1 kB\u001b[0m \u001b[31m965.1 kB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
216
+ "\u001b[?25hRequirement already satisfied: datasets>=2.0.0 in /usr/local/lib/python3.10/dist-packages (from evaluate) (2.15.0)\n",
217
+ "Requirement already satisfied: numpy>=1.17 in /usr/local/lib/python3.10/dist-packages (from evaluate) (1.23.5)\n",
218
+ "Requirement already satisfied: dill in /usr/local/lib/python3.10/dist-packages (from evaluate) (0.3.7)\n",
219
+ "Requirement already satisfied: pandas in /usr/local/lib/python3.10/dist-packages (from evaluate) (1.5.3)\n",
220
+ "Requirement already satisfied: requests>=2.19.0 in /usr/local/lib/python3.10/dist-packages (from evaluate) (2.31.0)\n",
221
+ "Requirement already satisfied: tqdm>=4.62.1 in /usr/local/lib/python3.10/dist-packages (from evaluate) (4.66.1)\n",
222
+ "Requirement already satisfied: xxhash in /usr/local/lib/python3.10/dist-packages (from evaluate) (3.4.1)\n",
223
+ "Requirement already satisfied: multiprocess in /usr/local/lib/python3.10/dist-packages (from evaluate) (0.70.15)\n",
224
+ "Requirement already satisfied: fsspec[http]>=2021.05.0 in /usr/local/lib/python3.10/dist-packages (from evaluate) (2023.6.0)\n",
225
+ "Requirement already satisfied: huggingface-hub>=0.7.0 in /usr/local/lib/python3.10/dist-packages (from evaluate) (0.19.4)\n",
226
+ "Requirement already satisfied: packaging in /usr/local/lib/python3.10/dist-packages (from evaluate) (23.2)\n",
227
+ "Collecting responses<0.19 (from evaluate)\n",
228
+ " Downloading responses-0.18.0-py3-none-any.whl (38 kB)\n",
229
+ "Requirement already satisfied: pyarrow>=8.0.0 in /usr/local/lib/python3.10/dist-packages (from datasets>=2.0.0->evaluate) (9.0.0)\n",
230
+ "Requirement already satisfied: pyarrow-hotfix in /usr/local/lib/python3.10/dist-packages (from datasets>=2.0.0->evaluate) (0.6)\n",
231
+ "Requirement already satisfied: aiohttp in /usr/local/lib/python3.10/dist-packages (from datasets>=2.0.0->evaluate) (3.9.1)\n",
232
+ "Requirement already satisfied: pyyaml>=5.1 in /usr/local/lib/python3.10/dist-packages (from datasets>=2.0.0->evaluate) (6.0.1)\n",
233
+ "Requirement already satisfied: filelock in /usr/local/lib/python3.10/dist-packages (from huggingface-hub>=0.7.0->evaluate) (3.13.1)\n",
234
+ "Requirement already satisfied: typing-extensions>=3.7.4.3 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub>=0.7.0->evaluate) (4.5.0)\n",
235
+ "Requirement already satisfied: charset-normalizer<4,>=2 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->evaluate) (3.3.2)\n",
236
+ "Requirement already satisfied: idna<4,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->evaluate) (3.6)\n",
237
+ "Requirement already satisfied: urllib3<3,>=1.21.1 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->evaluate) (2.0.7)\n",
238
+ "Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests>=2.19.0->evaluate) (2023.11.17)\n",
239
+ "Requirement already satisfied: python-dateutil>=2.8.1 in /usr/local/lib/python3.10/dist-packages (from pandas->evaluate) (2.8.2)\n",
240
+ "Requirement already satisfied: pytz>=2020.1 in /usr/local/lib/python3.10/dist-packages (from pandas->evaluate) (2023.3.post1)\n",
241
+ "Requirement already satisfied: attrs>=17.3.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets>=2.0.0->evaluate) (23.1.0)\n",
242
+ "Requirement already satisfied: multidict<7.0,>=4.5 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets>=2.0.0->evaluate) (6.0.4)\n",
243
+ "Requirement already satisfied: yarl<2.0,>=1.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets>=2.0.0->evaluate) (1.9.3)\n",
244
+ "Requirement already satisfied: frozenlist>=1.1.1 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets>=2.0.0->evaluate) (1.4.0)\n",
245
+ "Requirement already satisfied: aiosignal>=1.1.2 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets>=2.0.0->evaluate) (1.3.1)\n",
246
+ "Requirement already satisfied: async-timeout<5.0,>=4.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets>=2.0.0->evaluate) (4.0.3)\n",
247
+ "Requirement already satisfied: six>=1.5 in /usr/local/lib/python3.10/dist-packages (from python-dateutil>=2.8.1->pandas->evaluate) (1.16.0)\n",
248
+ "Installing collected packages: responses, evaluate\n",
249
+ "Successfully installed evaluate-0.4.1 responses-0.18.0\n"
250
+ ]
251
+ }
252
+ ]
253
+ },
254
+ {
255
+ "cell_type": "code",
256
+ "source": [
257
+ "!wget https://raw.githubusercontent.com/huggingface/transformers/main/examples/pytorch/question-answering/run_qa.py\n",
258
+ "!wget https://raw.githubusercontent.com/huggingface/transformers/main/examples/pytorch/question-answering/trainer_qa.py\n",
259
+ "!wget https://raw.githubusercontent.com/huggingface/transformers/main/examples/pytorch/question-answering/utils_qa.py"
260
+ ],
261
+ "metadata": {
262
+ "colab": {
263
+ "base_uri": "https://localhost:8080/"
264
+ },
265
+ "id": "JJhpBNwpxOTK",
266
+ "outputId": "e656d5a0-4f25-4e7b-8200-039389ad88ae"
267
+ },
268
+ "execution_count": 5,
269
+ "outputs": [
270
+ {
271
+ "output_type": "stream",
272
+ "name": "stdout",
273
+ "text": [
274
+ "--2023-12-09 13:04:52-- https://raw.githubusercontent.com/huggingface/transformers/main/examples/pytorch/question-answering/run_qa.py\n",
275
+ "Resolving raw.githubusercontent.com (raw.githubusercontent.com)... 185.199.108.133, 185.199.109.133, 185.199.110.133, ...\n",
276
+ "Connecting to raw.githubusercontent.com (raw.githubusercontent.com)|185.199.108.133|:443... connected.\n",
277
+ "HTTP request sent, awaiting response... 200 OK\n",
278
+ "Length: 32883 (32K) [text/plain]\n",
279
+ "Saving to: ‘run_qa.py’\n",
280
+ "\n",
281
+ "\rrun_qa.py 0%[ ] 0 --.-KB/s \rrun_qa.py 100%[===================>] 32.11K --.-KB/s in 0.01s \n",
282
+ "\n",
283
+ "2023-12-09 13:04:52 (3.06 MB/s) - ‘run_qa.py’ saved [32883/32883]\n",
284
+ "\n",
285
+ "--2023-12-09 13:04:52-- https://raw.githubusercontent.com/huggingface/transformers/main/examples/pytorch/question-answering/trainer_qa.py\n",
286
+ "Resolving raw.githubusercontent.com (raw.githubusercontent.com)... 185.199.108.133, 185.199.109.133, 185.199.110.133, ...\n",
287
+ "Connecting to raw.githubusercontent.com (raw.githubusercontent.com)|185.199.108.133|:443... connected.\n",
288
+ "HTTP request sent, awaiting response... 200 OK\n",
289
+ "Length: 6253 (6.1K) [text/plain]\n",
290
+ "Saving to: ‘trainer_qa.py’\n",
291
+ "\n",
292
+ "trainer_qa.py 100%[===================>] 6.11K --.-KB/s in 0s \n",
293
+ "\n",
294
+ "2023-12-09 13:04:52 (95.6 MB/s) - ‘trainer_qa.py’ saved [6253/6253]\n",
295
+ "\n",
296
+ "--2023-12-09 13:04:52-- https://raw.githubusercontent.com/huggingface/transformers/main/examples/pytorch/question-answering/utils_qa.py\n",
297
+ "Resolving raw.githubusercontent.com (raw.githubusercontent.com)... 185.199.111.133, 185.199.108.133, 185.199.109.133, ...\n",
298
+ "Connecting to raw.githubusercontent.com (raw.githubusercontent.com)|185.199.111.133|:443... connected.\n",
299
+ "HTTP request sent, awaiting response... 200 OK\n",
300
+ "Length: 22777 (22K) [text/plain]\n",
301
+ "Saving to: ‘utils_qa.py’\n",
302
+ "\n",
303
+ "utils_qa.py 100%[===================>] 22.24K --.-KB/s in 0.001s \n",
304
+ "\n",
305
+ "2023-12-09 13:04:52 (14.6 MB/s) - ‘utils_qa.py’ saved [22777/22777]\n",
306
+ "\n"
307
+ ]
308
+ }
309
+ ]
310
+ },
311
+ {
312
+ "cell_type": "code",
313
+ "source": [
314
+ "!ls"
315
+ ],
316
+ "metadata": {
317
+ "colab": {
318
+ "base_uri": "https://localhost:8080/"
319
+ },
320
+ "id": "Loa6y9XqzXJ4",
321
+ "outputId": "0aa57423-b1a3-4961-b7ce-8eeb9b7ea175"
322
+ },
323
+ "execution_count": 6,
324
+ "outputs": [
325
+ {
326
+ "output_type": "stream",
327
+ "name": "stdout",
328
+ "text": [
329
+ "run_qa.py sample_data\ttrainer_qa.py utils_qa.py\n"
330
+ ]
331
+ }
332
+ ]
333
+ },
334
+ {
335
+ "cell_type": "code",
336
+ "source": [
337
+ "!mkdir eval_results"
338
+ ],
339
+ "metadata": {
340
+ "id": "QDNxV31P0reW"
341
+ },
342
+ "execution_count": 7,
343
+ "outputs": []
344
+ },
345
+ {
346
+ "cell_type": "code",
347
+ "source": [
348
+ "!python -m run_qa --model_name_or_path osiria/minilm-italian-l6-h384-question-answering --dataset_name squad_it --do_predict --per_device_eval_batch_size=5 --output_dir eval_results"
349
+ ],
350
+ "metadata": {
351
+ "colab": {
352
+ "base_uri": "https://localhost:8080/"
353
+ },
354
+ "id": "B5O5Lom8x_c4",
355
+ "outputId": "285a1739-b358-4127-8429-5437f35682ae"
356
+ },
357
+ "execution_count": 8,
358
+ "outputs": [
359
+ {
360
+ "output_type": "stream",
361
+ "name": "stdout",
362
+ "text": [
363
+ "2023-12-09 13:05:00.034232: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n",
364
+ "2023-12-09 13:05:00.034295: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n",
365
+ "2023-12-09 13:05:00.034336: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n",
366
+ "2023-12-09 13:05:01.133400: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n",
367
+ "12/09/2023 13:05:04 - WARNING - __main__ - Process rank: 0, device: cuda:0, n_gpu: 1, distributed training: False, 16-bits training: False\n",
368
+ "12/09/2023 13:05:04 - INFO - __main__ - Training/evaluation parameters TrainingArguments(\n",
369
+ "_n_gpu=1,\n",
370
+ "adafactor=False,\n",
371
+ "adam_beta1=0.9,\n",
372
+ "adam_beta2=0.999,\n",
373
+ "adam_epsilon=1e-08,\n",
374
+ "auto_find_batch_size=False,\n",
375
+ "bf16=False,\n",
376
+ "bf16_full_eval=False,\n",
377
+ "data_seed=None,\n",
378
+ "dataloader_drop_last=False,\n",
379
+ "dataloader_num_workers=0,\n",
380
+ "dataloader_persistent_workers=False,\n",
381
+ "dataloader_pin_memory=True,\n",
382
+ "ddp_backend=None,\n",
383
+ "ddp_broadcast_buffers=None,\n",
384
+ "ddp_bucket_cap_mb=None,\n",
385
+ "ddp_find_unused_parameters=None,\n",
386
+ "ddp_timeout=1800,\n",
387
+ "debug=[],\n",
388
+ "deepspeed=None,\n",
389
+ "disable_tqdm=False,\n",
390
+ "dispatch_batches=None,\n",
391
+ "do_eval=False,\n",
392
+ "do_predict=True,\n",
393
+ "do_train=False,\n",
394
+ "eval_accumulation_steps=None,\n",
395
+ "eval_delay=0,\n",
396
+ "eval_steps=None,\n",
397
+ "evaluation_strategy=no,\n",
398
+ "fp16=False,\n",
399
+ "fp16_backend=auto,\n",
400
+ "fp16_full_eval=False,\n",
401
+ "fp16_opt_level=O1,\n",
402
+ "fsdp=[],\n",
403
+ "fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False},\n",
404
+ "fsdp_min_num_params=0,\n",
405
+ "fsdp_transformer_layer_cls_to_wrap=None,\n",
406
+ "full_determinism=False,\n",
407
+ "gradient_accumulation_steps=1,\n",
408
+ "gradient_checkpointing=False,\n",
409
+ "gradient_checkpointing_kwargs=None,\n",
410
+ "greater_is_better=None,\n",
411
+ "group_by_length=False,\n",
412
+ "half_precision_backend=auto,\n",
413
+ "hub_always_push=False,\n",
414
+ "hub_model_id=None,\n",
415
+ "hub_private_repo=False,\n",
416
+ "hub_strategy=every_save,\n",
417
+ "hub_token=<HUB_TOKEN>,\n",
418
+ "ignore_data_skip=False,\n",
419
+ "include_inputs_for_metrics=False,\n",
420
+ "include_num_input_tokens_seen=False,\n",
421
+ "include_tokens_per_second=False,\n",
422
+ "jit_mode_eval=False,\n",
423
+ "label_names=None,\n",
424
+ "label_smoothing_factor=0.0,\n",
425
+ "learning_rate=5e-05,\n",
426
+ "length_column_name=length,\n",
427
+ "load_best_model_at_end=False,\n",
428
+ "local_rank=0,\n",
429
+ "log_level=passive,\n",
430
+ "log_level_replica=warning,\n",
431
+ "log_on_each_node=True,\n",
432
+ "logging_dir=eval_results/runs/Dec09_13-05-04_f36df394933d,\n",
433
+ "logging_first_step=False,\n",
434
+ "logging_nan_inf_filter=True,\n",
435
+ "logging_steps=500,\n",
436
+ "logging_strategy=steps,\n",
437
+ "lr_scheduler_kwargs={},\n",
438
+ "lr_scheduler_type=linear,\n",
439
+ "max_grad_norm=1.0,\n",
440
+ "max_steps=-1,\n",
441
+ "metric_for_best_model=None,\n",
442
+ "mp_parameters=,\n",
443
+ "neftune_noise_alpha=None,\n",
444
+ "no_cuda=False,\n",
445
+ "num_train_epochs=3.0,\n",
446
+ "optim=adamw_torch,\n",
447
+ "optim_args=None,\n",
448
+ "output_dir=eval_results,\n",
449
+ "overwrite_output_dir=False,\n",
450
+ "past_index=-1,\n",
451
+ "per_device_eval_batch_size=5,\n",
452
+ "per_device_train_batch_size=8,\n",
453
+ "prediction_loss_only=False,\n",
454
+ "push_to_hub=False,\n",
455
+ "push_to_hub_model_id=None,\n",
456
+ "push_to_hub_organization=None,\n",
457
+ "push_to_hub_token=<PUSH_TO_HUB_TOKEN>,\n",
458
+ "ray_scope=last,\n",
459
+ "remove_unused_columns=True,\n",
460
+ "report_to=['tensorboard'],\n",
461
+ "resume_from_checkpoint=None,\n",
462
+ "run_name=eval_results,\n",
463
+ "save_on_each_node=False,\n",
464
+ "save_only_model=False,\n",
465
+ "save_safetensors=True,\n",
466
+ "save_steps=500,\n",
467
+ "save_strategy=steps,\n",
468
+ "save_total_limit=None,\n",
469
+ "seed=42,\n",
470
+ "skip_memory_metrics=True,\n",
471
+ "split_batches=False,\n",
472
+ "tf32=None,\n",
473
+ "torch_compile=False,\n",
474
+ "torch_compile_backend=None,\n",
475
+ "torch_compile_mode=None,\n",
476
+ "torchdynamo=None,\n",
477
+ "tpu_metrics_debug=False,\n",
478
+ "tpu_num_cores=None,\n",
479
+ "use_cpu=False,\n",
480
+ "use_ipex=False,\n",
481
+ "use_legacy_prediction_loop=False,\n",
482
+ "use_mps_device=False,\n",
483
+ "warmup_ratio=0.0,\n",
484
+ "warmup_steps=0,\n",
485
+ "weight_decay=0.0,\n",
486
+ ")\n",
487
+ "https://huggingface.co/datasets/squad_it/resolve/main/squad_it.py not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/2f7a246f661e6acb3a9ad072ed14241ea343265e7d863b88717897941c3ab1dd.671c3c3daf2f05e0355b498b0f5a7f2167fa94dc6de36195fe1ba0503e6dcc30.py.incomplete\n",
488
+ "12/09/2023 13:05:04 - INFO - datasets.utils.file_utils - https://huggingface.co/datasets/squad_it/resolve/main/squad_it.py not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/2f7a246f661e6acb3a9ad072ed14241ea343265e7d863b88717897941c3ab1dd.671c3c3daf2f05e0355b498b0f5a7f2167fa94dc6de36195fe1ba0503e6dcc30.py.incomplete\n",
489
+ "Downloading builder script: 100% 4.74k/4.74k [00:00<00:00, 17.6MB/s]\n",
490
+ "storing https://huggingface.co/datasets/squad_it/resolve/main/squad_it.py in cache at /root/.cache/huggingface/datasets/downloads/2f7a246f661e6acb3a9ad072ed14241ea343265e7d863b88717897941c3ab1dd.671c3c3daf2f05e0355b498b0f5a7f2167fa94dc6de36195fe1ba0503e6dcc30.py\n",
491
+ "12/09/2023 13:05:04 - INFO - datasets.utils.file_utils - storing https://huggingface.co/datasets/squad_it/resolve/main/squad_it.py in cache at /root/.cache/huggingface/datasets/downloads/2f7a246f661e6acb3a9ad072ed14241ea343265e7d863b88717897941c3ab1dd.671c3c3daf2f05e0355b498b0f5a7f2167fa94dc6de36195fe1ba0503e6dcc30.py\n",
492
+ "creating metadata file for /root/.cache/huggingface/datasets/downloads/2f7a246f661e6acb3a9ad072ed14241ea343265e7d863b88717897941c3ab1dd.671c3c3daf2f05e0355b498b0f5a7f2167fa94dc6de36195fe1ba0503e6dcc30.py\n",
493
+ "12/09/2023 13:05:04 - INFO - datasets.utils.file_utils - creating metadata file for /root/.cache/huggingface/datasets/downloads/2f7a246f661e6acb3a9ad072ed14241ea343265e7d863b88717897941c3ab1dd.671c3c3daf2f05e0355b498b0f5a7f2167fa94dc6de36195fe1ba0503e6dcc30.py\n",
494
+ "https://huggingface.co/datasets/squad_it/resolve/main/dataset_infos.json not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/f6d7650a1e4a9f0e42b54f34ac614f7c1142ad26c1bd19a448eb606fe09f7515.b459489a7b1a4d859decbeef851034c7846ef6cce8ec2cf192cf4b4f40d0865d.incomplete\n",
495
+ "12/09/2023 13:05:05 - INFO - datasets.utils.file_utils - https://huggingface.co/datasets/squad_it/resolve/main/dataset_infos.json not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/f6d7650a1e4a9f0e42b54f34ac614f7c1142ad26c1bd19a448eb606fe09f7515.b459489a7b1a4d859decbeef851034c7846ef6cce8ec2cf192cf4b4f40d0865d.incomplete\n",
496
+ "Downloading metadata: 100% 2.51k/2.51k [00:00<00:00, 13.8MB/s]\n",
497
+ "storing https://huggingface.co/datasets/squad_it/resolve/main/dataset_infos.json in cache at /root/.cache/huggingface/datasets/downloads/f6d7650a1e4a9f0e42b54f34ac614f7c1142ad26c1bd19a448eb606fe09f7515.b459489a7b1a4d859decbeef851034c7846ef6cce8ec2cf192cf4b4f40d0865d\n",
498
+ "12/09/2023 13:05:05 - INFO - datasets.utils.file_utils - storing https://huggingface.co/datasets/squad_it/resolve/main/dataset_infos.json in cache at /root/.cache/huggingface/datasets/downloads/f6d7650a1e4a9f0e42b54f34ac614f7c1142ad26c1bd19a448eb606fe09f7515.b459489a7b1a4d859decbeef851034c7846ef6cce8ec2cf192cf4b4f40d0865d\n",
499
+ "creating metadata file for /root/.cache/huggingface/datasets/downloads/f6d7650a1e4a9f0e42b54f34ac614f7c1142ad26c1bd19a448eb606fe09f7515.b459489a7b1a4d859decbeef851034c7846ef6cce8ec2cf192cf4b4f40d0865d\n",
500
+ "12/09/2023 13:05:05 - INFO - datasets.utils.file_utils - creating metadata file for /root/.cache/huggingface/datasets/downloads/f6d7650a1e4a9f0e42b54f34ac614f7c1142ad26c1bd19a448eb606fe09f7515.b459489a7b1a4d859decbeef851034c7846ef6cce8ec2cf192cf4b4f40d0865d\n",
501
+ "https://huggingface.co/datasets/squad_it/resolve/main/README.md not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/b664892e5965a1ef2e173eadc7e9cbf6ad4aeb6240d7b3bb6ab9f32850363a6f.71fe7dbf5f3b574d85b57b5a722ebc49fe8ec8b03d655344891fb1b05c4634fb.incomplete\n",
502
+ "12/09/2023 13:05:05 - INFO - datasets.utils.file_utils - https://huggingface.co/datasets/squad_it/resolve/main/README.md not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/b664892e5965a1ef2e173eadc7e9cbf6ad4aeb6240d7b3bb6ab9f32850363a6f.71fe7dbf5f3b574d85b57b5a722ebc49fe8ec8b03d655344891fb1b05c4634fb.incomplete\n",
503
+ "Downloading readme: 100% 7.27k/7.27k [00:00<00:00, 32.7MB/s]\n",
504
+ "storing https://huggingface.co/datasets/squad_it/resolve/main/README.md in cache at /root/.cache/huggingface/datasets/downloads/b664892e5965a1ef2e173eadc7e9cbf6ad4aeb6240d7b3bb6ab9f32850363a6f.71fe7dbf5f3b574d85b57b5a722ebc49fe8ec8b03d655344891fb1b05c4634fb\n",
505
+ "12/09/2023 13:05:05 - INFO - datasets.utils.file_utils - storing https://huggingface.co/datasets/squad_it/resolve/main/README.md in cache at /root/.cache/huggingface/datasets/downloads/b664892e5965a1ef2e173eadc7e9cbf6ad4aeb6240d7b3bb6ab9f32850363a6f.71fe7dbf5f3b574d85b57b5a722ebc49fe8ec8b03d655344891fb1b05c4634fb\n",
506
+ "creating metadata file for /root/.cache/huggingface/datasets/downloads/b664892e5965a1ef2e173eadc7e9cbf6ad4aeb6240d7b3bb6ab9f32850363a6f.71fe7dbf5f3b574d85b57b5a722ebc49fe8ec8b03d655344891fb1b05c4634fb\n",
507
+ "12/09/2023 13:05:05 - INFO - datasets.utils.file_utils - creating metadata file for /root/.cache/huggingface/datasets/downloads/b664892e5965a1ef2e173eadc7e9cbf6ad4aeb6240d7b3bb6ab9f32850363a6f.71fe7dbf5f3b574d85b57b5a722ebc49fe8ec8b03d655344891fb1b05c4634fb\n",
508
+ "Loading Dataset Infos from /root/.cache/huggingface/modules/datasets_modules/datasets/squad_it/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71\n",
509
+ "12/09/2023 13:05:05 - INFO - datasets.info - Loading Dataset Infos from /root/.cache/huggingface/modules/datasets_modules/datasets/squad_it/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71\n",
510
+ "Generating dataset squad_it (/root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71)\n",
511
+ "12/09/2023 13:05:05 - INFO - datasets.builder - Generating dataset squad_it (/root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71)\n",
512
+ "Downloading and preparing dataset squad_it/default to /root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71...\n",
513
+ "12/09/2023 13:05:05 - INFO - datasets.builder - Downloading and preparing dataset squad_it/default to /root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71...\n",
514
+ "Dataset not on Hf google storage. Downloading and preparing it from source\n",
515
+ "12/09/2023 13:05:05 - INFO - datasets.builder - Dataset not on Hf google storage. Downloading and preparing it from source\n",
516
+ "Downloading data files: 0% 0/2 [00:00<?, ?it/s]https://github.com/crux82/squad-it/raw/master/SQuAD_it-train.json.gz not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/44ac0548c3284061d1356ff7f5212fe42b923b0aa0854c886ce93dfeb7ab629b.incomplete\n",
517
+ "12/09/2023 13:05:06 - INFO - datasets.utils.file_utils - https://github.com/crux82/squad-it/raw/master/SQuAD_it-train.json.gz not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/44ac0548c3284061d1356ff7f5212fe42b923b0aa0854c886ce93dfeb7ab629b.incomplete\n",
518
+ "\n",
519
+ "Downloading data: 0% 0.00/7.73M [00:00<?, ?B/s]\u001b[A\n",
520
+ "Downloading data: 100% 7.73M/7.73M [00:00<00:00, 57.4MB/s]\n",
521
+ "storing https://github.com/crux82/squad-it/raw/master/SQuAD_it-train.json.gz in cache at /root/.cache/huggingface/datasets/downloads/44ac0548c3284061d1356ff7f5212fe42b923b0aa0854c886ce93dfeb7ab629b\n",
522
+ "12/09/2023 13:05:07 - INFO - datasets.utils.file_utils - storing https://github.com/crux82/squad-it/raw/master/SQuAD_it-train.json.gz in cache at /root/.cache/huggingface/datasets/downloads/44ac0548c3284061d1356ff7f5212fe42b923b0aa0854c886ce93dfeb7ab629b\n",
523
+ "creating metadata file for /root/.cache/huggingface/datasets/downloads/44ac0548c3284061d1356ff7f5212fe42b923b0aa0854c886ce93dfeb7ab629b\n",
524
+ "12/09/2023 13:05:07 - INFO - datasets.utils.file_utils - creating metadata file for /root/.cache/huggingface/datasets/downloads/44ac0548c3284061d1356ff7f5212fe42b923b0aa0854c886ce93dfeb7ab629b\n",
525
+ "Downloading data files: 50% 1/2 [00:01<00:01, 1.55s/it]https://github.com/crux82/squad-it/raw/master/SQuAD_it-test.json.gz not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/e560a4a54eb929ce542e945627942afca80859dcbf499866a59984eb9026d31f.incomplete\n",
526
+ "12/09/2023 13:05:07 - INFO - datasets.utils.file_utils - https://github.com/crux82/squad-it/raw/master/SQuAD_it-test.json.gz not found in cache or force_download set to True, downloading to /root/.cache/huggingface/datasets/downloads/e560a4a54eb929ce542e945627942afca80859dcbf499866a59984eb9026d31f.incomplete\n",
527
+ "\n",
528
+ "Downloading data: 100% 1.05M/1.05M [00:00<00:00, 15.8MB/s]\n",
529
+ "storing https://github.com/crux82/squad-it/raw/master/SQuAD_it-test.json.gz in cache at /root/.cache/huggingface/datasets/downloads/e560a4a54eb929ce542e945627942afca80859dcbf499866a59984eb9026d31f\n",
530
+ "12/09/2023 13:05:08 - INFO - datasets.utils.file_utils - storing https://github.com/crux82/squad-it/raw/master/SQuAD_it-test.json.gz in cache at /root/.cache/huggingface/datasets/downloads/e560a4a54eb929ce542e945627942afca80859dcbf499866a59984eb9026d31f\n",
531
+ "creating metadata file for /root/.cache/huggingface/datasets/downloads/e560a4a54eb929ce542e945627942afca80859dcbf499866a59984eb9026d31f\n",
532
+ "12/09/2023 13:05:08 - INFO - datasets.utils.file_utils - creating metadata file for /root/.cache/huggingface/datasets/downloads/e560a4a54eb929ce542e945627942afca80859dcbf499866a59984eb9026d31f\n",
533
+ "Downloading data files: 100% 2/2 [00:02<00:00, 1.20s/it]\n",
534
+ "Downloading took 0.0 min\n",
535
+ "12/09/2023 13:05:08 - INFO - datasets.download.download_manager - Downloading took 0.0 min\n",
536
+ "Checksum Computation took 0.0 min\n",
537
+ "12/09/2023 13:05:08 - INFO - datasets.download.download_manager - Checksum Computation took 0.0 min\n",
538
+ "Extracting data files: 100% 2/2 [00:00<00:00, 8.59it/s]\n",
539
+ "Generating train split\n",
540
+ "12/09/2023 13:05:08 - INFO - datasets.builder - Generating train split\n",
541
+ "Generating train split: 100% 54159/54159 [00:06<00:00, 8532.57 examples/s] \n",
542
+ "Generating test split\n",
543
+ "12/09/2023 13:05:14 - INFO - datasets.builder - Generating test split\n",
544
+ "Generating test split: 100% 7609/7609 [00:00<00:00, 11548.29 examples/s]\n",
545
+ "All the splits matched successfully.\n",
546
+ "12/09/2023 13:05:15 - INFO - datasets.utils.info_utils - All the splits matched successfully.\n",
547
+ "Dataset squad_it downloaded and prepared to /root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71. Subsequent calls will reuse this data.\n",
548
+ "12/09/2023 13:05:15 - INFO - datasets.builder - Dataset squad_it downloaded and prepared to /root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71. Subsequent calls will reuse this data.\n",
549
+ "config.json: 100% 652/652 [00:00<00:00, 3.76MB/s]\n",
550
+ "[INFO|configuration_utils.py:739] 2023-12-09 13:05:15,608 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--osiria--minilm-italian-l6-h384-question-answering/snapshots/1faf21988c80be33048fe9d3e0bb98aafec858bb/config.json\n",
551
+ "[INFO|configuration_utils.py:802] 2023-12-09 13:05:15,613 >> Model config XLMRobertaConfig {\n",
552
+ " \"_name_or_path\": \"osiria/minilm-italian-l6-h384-question-answering\",\n",
553
+ " \"architectures\": [\n",
554
+ " \"XLMRobertaForQuestionAnswering\"\n",
555
+ " ],\n",
556
+ " \"attention_probs_dropout_prob\": 0.1,\n",
557
+ " \"bos_token_id\": 0,\n",
558
+ " \"classifier_dropout\": null,\n",
559
+ " \"eos_token_id\": 2,\n",
560
+ " \"hidden_act\": \"gelu\",\n",
561
+ " \"hidden_dropout_prob\": 0.1,\n",
562
+ " \"hidden_size\": 384,\n",
563
+ " \"initializer_range\": 0.02,\n",
564
+ " \"intermediate_size\": 1536,\n",
565
+ " \"layer_norm_eps\": 1e-05,\n",
566
+ " \"max_position_embeddings\": 514,\n",
567
+ " \"model_type\": \"xlm-roberta\",\n",
568
+ " \"num_attention_heads\": 12,\n",
569
+ " \"num_hidden_layers\": 6,\n",
570
+ " \"pad_token_id\": 1,\n",
571
+ " \"position_embedding_type\": \"absolute\",\n",
572
+ " \"torch_dtype\": \"float32\",\n",
573
+ " \"transformers_version\": \"4.36.0.dev0\",\n",
574
+ " \"type_vocab_size\": 1,\n",
575
+ " \"use_cache\": true,\n",
576
+ " \"vocab_size\": 30498\n",
577
+ "}\n",
578
+ "\n",
579
+ "tokenizer_config.json: 100% 418/418 [00:00<00:00, 2.73MB/s]\n",
580
+ "tokenizer.json: 100% 1.39M/1.39M [00:00<00:00, 18.7MB/s]\n",
581
+ "special_tokens_map.json: 100% 280/280 [00:00<00:00, 1.75MB/s]\n",
582
+ "[INFO|tokenization_utils_base.py:2026] 2023-12-09 13:05:16,227 >> loading file sentencepiece.bpe.model from cache at None\n",
583
+ "[INFO|tokenization_utils_base.py:2026] 2023-12-09 13:05:16,227 >> loading file tokenizer.json from cache at /root/.cache/huggingface/hub/models--osiria--minilm-italian-l6-h384-question-answering/snapshots/1faf21988c80be33048fe9d3e0bb98aafec858bb/tokenizer.json\n",
584
+ "[INFO|tokenization_utils_base.py:2026] 2023-12-09 13:05:16,227 >> loading file added_tokens.json from cache at None\n",
585
+ "[INFO|tokenization_utils_base.py:2026] 2023-12-09 13:05:16,227 >> loading file special_tokens_map.json from cache at /root/.cache/huggingface/hub/models--osiria--minilm-italian-l6-h384-question-answering/snapshots/1faf21988c80be33048fe9d3e0bb98aafec858bb/special_tokens_map.json\n",
586
+ "[INFO|tokenization_utils_base.py:2026] 2023-12-09 13:05:16,228 >> loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--osiria--minilm-italian-l6-h384-question-answering/snapshots/1faf21988c80be33048fe9d3e0bb98aafec858bb/tokenizer_config.json\n",
587
+ "pytorch_model.bin: 100% 90.3M/90.3M [00:01<00:00, 57.7MB/s]\n",
588
+ "[INFO|modeling_utils.py:3299] 2023-12-09 13:05:18,490 >> loading weights file pytorch_model.bin from cache at /root/.cache/huggingface/hub/models--osiria--minilm-italian-l6-h384-question-answering/snapshots/1faf21988c80be33048fe9d3e0bb98aafec858bb/pytorch_model.bin\n",
589
+ "[INFO|modeling_utils.py:4139] 2023-12-09 13:05:18,858 >> All model checkpoint weights were used when initializing XLMRobertaForQuestionAnswering.\n",
590
+ "\n",
591
+ "[INFO|modeling_utils.py:4147] 2023-12-09 13:05:18,858 >> All the weights of XLMRobertaForQuestionAnswering were initialized from the model checkpoint at osiria/minilm-italian-l6-h384-question-answering.\n",
592
+ "If your task is similar to the task the model of the checkpoint was trained on, you can already use XLMRobertaForQuestionAnswering for predictions without further training.\n",
593
+ "Running tokenizer on prediction dataset: 0% 0/7609 [00:00<?, ? examples/s]Caching processed dataset at /root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71/cache-817e084ce623e156.arrow\n",
594
+ "12/09/2023 13:05:19 - INFO - datasets.arrow_dataset - Caching processed dataset at /root/.cache/huggingface/datasets/squad_it/default/0.1.0/d442bdb4794b4bae227ab19105b76d706ed7cf2ac342e4c9da4a5c36bde19d71/cache-817e084ce623e156.arrow\n",
595
+ "Running tokenizer on prediction dataset: 100% 7609/7609 [00:08<00:00, 880.77 examples/s]\n",
596
+ "Downloading builder script: 100% 4.53k/4.53k [00:00<00:00, 18.0MB/s]\n",
597
+ "Downloading extra modules: 100% 3.32k/3.32k [00:00<00:00, 17.0MB/s]\n",
598
+ "12/09/2023 13:05:33 - INFO - __main__ - *** Predict ***\n",
599
+ "[INFO|trainer.py:712] 2023-12-09 13:05:33,267 >> The following columns in the test set don't have a corresponding argument in `XLMRobertaForQuestionAnswering.forward` and have been ignored: offset_mapping, example_id. If offset_mapping, example_id are not expected by `XLMRobertaForQuestionAnswering.forward`, you can safely ignore this message.\n",
600
+ "[INFO|trainer.py:3154] 2023-12-09 13:05:33,270 >> ***** Running Prediction *****\n",
601
+ "[INFO|trainer.py:3156] 2023-12-09 13:05:33,270 >> Num examples = 7988\n",
602
+ "[INFO|trainer.py:3159] 2023-12-09 13:05:33,270 >> Batch size = 5\n",
603
+ "100% 1598/1598 [00:38<00:00, 43.23it/s]12/09/2023 13:06:23 - INFO - utils_qa - Post-processing 7609 example predictions split into 7988 features.\n",
604
+ "\n",
605
+ " 0% 0/7609 [00:00<?, ?it/s]\u001b[A\n",
606
+ " 1% 41/7609 [00:00<00:18, 401.05it/s]\u001b[A\n",
607
+ " 1% 82/7609 [00:00<00:20, 373.83it/s]\u001b[A\n",
608
+ " 2% 120/7609 [00:00<00:19, 375.54it/s]\u001b[A\n",
609
+ " 2% 158/7609 [00:00<00:19, 373.21it/s]\u001b[A\n",
610
+ " 3% 196/7609 [00:00<00:19, 375.08it/s]\u001b[A\n",
611
+ " 3% 237/7609 [00:00<00:19, 385.93it/s]\u001b[A\n",
612
+ " 4% 277/7609 [00:00<00:18, 389.16it/s]\u001b[A\n",
613
+ " 4% 316/7609 [00:00<00:19, 373.73it/s]\u001b[A\n",
614
+ " 5% 354/7609 [00:00<00:19, 371.48it/s]\u001b[A\n",
615
+ " 5% 394/7609 [00:01<00:19, 379.36it/s]\u001b[A\n",
616
+ " 6% 433/7609 [00:01<00:19, 373.82it/s]\u001b[A\n",
617
+ " 6% 474/7609 [00:01<00:18, 384.33it/s]\u001b[A\n",
618
+ " 7% 513/7609 [00:01<00:18, 382.05it/s]\u001b[A\n",
619
+ " 7% 552/7609 [00:01<00:19, 361.93it/s]\u001b[A\n",
620
+ "100% 1598/1598 [00:49<00:00, 43.23it/s]\n",
621
+ " 8% 634/7609 [00:01<00:18, 383.75it/s]\u001b[A\n",
622
+ " 9% 674/7609 [00:01<00:17, 385.73it/s]\u001b[A\n",
623
+ " 9% 715/7609 [00:01<00:17, 391.92it/s]\u001b[A\n",
624
+ " 10% 755/7609 [00:01<00:17, 384.72it/s]\u001b[A\n",
625
+ " 10% 795/7609 [00:02<00:17, 387.04it/s]\u001b[A\n",
626
+ " 11% 834/7609 [00:02<00:17, 377.84it/s]\u001b[A\n",
627
+ " 11% 872/7609 [00:02<00:17, 377.50it/s]\u001b[A\n",
628
+ " 12% 910/7609 [00:02<00:17, 373.54it/s]\u001b[A\n",
629
+ " 12% 948/7609 [00:02<00:17, 371.87it/s]\u001b[A\n",
630
+ " 13% 988/7609 [00:02<00:17, 378.28it/s]\u001b[A\n",
631
+ " 14% 1029/7609 [00:02<00:17, 385.26it/s]\u001b[A\n",
632
+ " 14% 1070/7609 [00:02<00:16, 390.90it/s]\u001b[A\n",
633
+ " 15% 1110/7609 [00:02<00:16, 388.54it/s]\u001b[A\n",
634
+ " 15% 1149/7609 [00:03<00:16, 382.61it/s]\u001b[A\n",
635
+ " 16% 1188/7609 [00:03<00:17, 367.95it/s]\u001b[A\n",
636
+ " 16% 1229/7609 [00:03<00:16, 378.84it/s]\u001b[A\n",
637
+ " 17% 1268/7609 [00:03<00:16, 380.23it/s]\u001b[A\n",
638
+ " 17% 1308/7609 [00:03<00:16, 383.91it/s]\u001b[A\n",
639
+ " 18% 1347/7609 [00:03<00:16, 378.43it/s]\u001b[A\n",
640
+ " 18% 1385/7609 [00:03<00:16, 374.90it/s]\u001b[A\n",
641
+ " 19% 1424/7609 [00:03<00:16, 376.95it/s]\u001b[A\n",
642
+ " 19% 1462/7609 [00:03<00:16, 370.65it/s]\u001b[A\n",
643
+ " 20% 1500/7609 [00:03<00:17, 350.38it/s]\u001b[A\n",
644
+ " 20% 1538/7609 [00:04<00:16, 358.03it/s]\u001b[A\n",
645
+ " 21% 1579/7609 [00:04<00:16, 370.54it/s]\u001b[A\n",
646
+ " 21% 1617/7609 [00:04<00:16, 357.26it/s]\u001b[A\n",
647
+ " 22% 1653/7609 [00:04<00:16, 355.85it/s]\u001b[A\n",
648
+ " 22% 1689/7609 [00:04<00:16, 356.50it/s]\u001b[A\n",
649
+ " 23% 1729/7609 [00:04<00:15, 368.06it/s]\u001b[A\n",
650
+ " 23% 1766/7609 [00:04<00:16, 362.43it/s]\u001b[A\n",
651
+ " 24% 1806/7609 [00:04<00:15, 371.78it/s]\u001b[A\n",
652
+ " 24% 1844/7609 [00:04<00:15, 374.09it/s]\u001b[A\n",
653
+ " 25% 1884/7609 [00:05<00:15, 379.97it/s]\u001b[A\n",
654
+ " 25% 1924/7609 [00:05<00:14, 383.62it/s]\u001b[A\n",
655
+ " 26% 1963/7609 [00:05<00:15, 354.44it/s]\u001b[A\n",
656
+ " 26% 1999/7609 [00:05<00:17, 317.18it/s]\u001b[A\n",
657
+ " 27% 2032/7609 [00:05<00:21, 265.40it/s]\u001b[A\n",
658
+ " 27% 2066/7609 [00:05<00:19, 280.18it/s]\u001b[A\n",
659
+ " 28% 2096/7609 [00:05<00:25, 214.52it/s]\u001b[A\n",
660
+ " 28% 2122/7609 [00:06<00:24, 223.59it/s]\u001b[A\n",
661
+ " 28% 2162/7609 [00:06<00:20, 263.47it/s]\u001b[A\n",
662
+ " 29% 2196/7609 [00:06<00:19, 281.93it/s]\u001b[A\n",
663
+ " 29% 2232/7609 [00:06<00:17, 301.40it/s]\u001b[A\n",
664
+ " 30% 2265/7609 [00:06<00:18, 295.49it/s]\u001b[A\n",
665
+ " 30% 2303/7609 [00:06<00:16, 317.40it/s]\u001b[A\n",
666
+ " 31% 2341/7609 [00:06<00:15, 334.53it/s]\u001b[A\n",
667
+ " 31% 2381/7609 [00:06<00:14, 353.03it/s]\u001b[A\n",
668
+ " 32% 2420/7609 [00:06<00:14, 362.45it/s]\u001b[A\n",
669
+ " 32% 2460/7609 [00:06<00:13, 368.51it/s]\u001b[A\n",
670
+ " 33% 2498/7609 [00:07<00:14, 349.37it/s]\u001b[A\n",
671
+ " 33% 2540/7609 [00:07<00:13, 366.50it/s]\u001b[A\n",
672
+ " 34% 2579/7609 [00:07<00:13, 370.66it/s]\u001b[A\n",
673
+ " 34% 2617/7609 [00:07<00:14, 351.73it/s]\u001b[A\n",
674
+ " 35% 2657/7609 [00:07<00:13, 363.13it/s]\u001b[A\n",
675
+ " 35% 2694/7609 [00:07<00:13, 356.81it/s]\u001b[A\n",
676
+ " 36% 2731/7609 [00:07<00:13, 359.78it/s]\u001b[A\n",
677
+ " 36% 2772/7609 [00:07<00:12, 373.10it/s]\u001b[A\n",
678
+ " 37% 2813/7609 [00:07<00:12, 382.65it/s]\u001b[A\n",
679
+ " 37% 2852/7609 [00:07<00:12, 382.80it/s]\u001b[A\n",
680
+ " 38% 2894/7609 [00:08<00:12, 392.22it/s]\u001b[A\n",
681
+ " 39% 2937/7609 [00:08<00:11, 402.25it/s]\u001b[A\n",
682
+ " 39% 2978/7609 [00:08<00:11, 393.04it/s]\u001b[A\n",
683
+ " 40% 3018/7609 [00:08<00:11, 389.48it/s]\u001b[A\n",
684
+ " 40% 3058/7609 [00:08<00:12, 378.03it/s]\u001b[A\n",
685
+ " 41% 3096/7609 [00:08<00:12, 370.92it/s]\u001b[A\n",
686
+ " 41% 3134/7609 [00:08<00:12, 360.59it/s]\u001b[A\n",
687
+ " 42% 3173/7609 [00:08<00:12, 366.35it/s]\u001b[A\n",
688
+ " 42% 3210/7609 [00:08<00:12, 361.80it/s]\u001b[A\n",
689
+ " 43% 3247/7609 [00:09<00:12, 357.73it/s]\u001b[A\n",
690
+ " 43% 3285/7609 [00:09<00:11, 363.47it/s]\u001b[A\n",
691
+ " 44% 3322/7609 [00:09<00:11, 363.00it/s]\u001b[A\n",
692
+ " 44% 3360/7609 [00:09<00:11, 367.90it/s]\u001b[A\n",
693
+ " 45% 3399/7609 [00:09<00:11, 373.00it/s]\u001b[A\n",
694
+ " 45% 3439/7609 [00:09<00:10, 380.60it/s]\u001b[A\n",
695
+ " 46% 3478/7609 [00:09<00:12, 336.39it/s]\u001b[A\n",
696
+ " 46% 3513/7609 [00:09<00:13, 296.52it/s]\u001b[A\n",
697
+ " 47% 3545/7609 [00:10<00:14, 272.10it/s]\u001b[A\n",
698
+ " 47% 3574/7609 [00:10<00:15, 259.31it/s]\u001b[A\n",
699
+ " 47% 3601/7609 [00:10<00:16, 248.17it/s]\u001b[A\n",
700
+ " 48% 3627/7609 [00:10<00:17, 227.36it/s]\u001b[A\n",
701
+ " 48% 3651/7609 [00:10<00:17, 220.80it/s]\u001b[A\n",
702
+ " 48% 3674/7609 [00:10<00:18, 218.57it/s]\u001b[A\n",
703
+ " 49% 3697/7609 [00:10<00:17, 218.89it/s]\u001b[A\n",
704
+ " 49% 3720/7609 [00:10<00:17, 216.68it/s]\u001b[A\n",
705
+ " 49% 3742/7609 [00:10<00:18, 210.65it/s]\u001b[A\n",
706
+ " 49% 3764/7609 [00:11<00:18, 209.35it/s]\u001b[A\n",
707
+ " 50% 3786/7609 [00:11<00:18, 211.83it/s]\u001b[A\n",
708
+ " 50% 3809/7609 [00:11<00:17, 214.51it/s]\u001b[A\n",
709
+ " 50% 3834/7609 [00:11<00:16, 222.59it/s]\u001b[A\n",
710
+ " 51% 3858/7609 [00:11<00:16, 225.59it/s]\u001b[A\n",
711
+ " 51% 3881/7609 [00:11<00:16, 222.47it/s]\u001b[A\n",
712
+ " 51% 3904/7609 [00:11<00:16, 222.57it/s]\u001b[A\n",
713
+ " 52% 3927/7609 [00:11<00:16, 221.91it/s]\u001b[A\n",
714
+ " 52% 3950/7609 [00:11<00:16, 223.75it/s]\u001b[A\n",
715
+ " 52% 3973/7609 [00:11<00:16, 218.35it/s]\u001b[A\n",
716
+ " 53% 3995/7609 [00:12<00:16, 217.07it/s]\u001b[A\n",
717
+ " 53% 4017/7609 [00:12<00:16, 214.83it/s]\u001b[A\n",
718
+ " 53% 4040/7609 [00:12<00:16, 219.00it/s]\u001b[A\n",
719
+ " 53% 4062/7609 [00:12<00:17, 205.86it/s]\u001b[A\n",
720
+ " 54% 4083/7609 [00:12<00:17, 206.14it/s]\u001b[A\n",
721
+ " 54% 4104/7609 [00:12<00:17, 196.27it/s]\u001b[A\n",
722
+ " 54% 4127/7609 [00:12<00:17, 204.40it/s]\u001b[A\n",
723
+ " 55% 4151/7609 [00:12<00:16, 213.31it/s]\u001b[A\n",
724
+ " 55% 4174/7609 [00:12<00:15, 218.00it/s]\u001b[A\n",
725
+ " 55% 4196/7609 [00:13<00:15, 217.17it/s]\u001b[A\n",
726
+ " 55% 4219/7609 [00:13<00:15, 219.75it/s]\u001b[A\n",
727
+ " 56% 4242/7609 [00:13<00:15, 218.32it/s]\u001b[A\n",
728
+ " 56% 4264/7609 [00:13<00:15, 218.25it/s]\u001b[A\n",
729
+ " 56% 4286/7609 [00:13<00:15, 216.53it/s]\u001b[A\n",
730
+ " 57% 4310/7609 [00:13<00:14, 221.19it/s]\u001b[A\n",
731
+ " 57% 4344/7609 [00:13<00:12, 254.04it/s]\u001b[A\n",
732
+ " 58% 4380/7609 [00:13<00:11, 285.06it/s]\u001b[A\n",
733
+ " 58% 4424/7609 [00:13<00:09, 329.34it/s]\u001b[A\n",
734
+ " 59% 4458/7609 [00:13<00:09, 324.38it/s]\u001b[A\n",
735
+ " 59% 4502/7609 [00:14<00:08, 356.17it/s]\u001b[A\n",
736
+ " 60% 4543/7609 [00:14<00:08, 370.47it/s]\u001b[A\n",
737
+ " 60% 4583/7609 [00:14<00:08, 376.88it/s]\u001b[A\n",
738
+ " 61% 4623/7609 [00:14<00:07, 383.12it/s]\u001b[A\n",
739
+ " 61% 4666/7609 [00:14<00:07, 394.89it/s]\u001b[A\n",
740
+ " 62% 4707/7609 [00:14<00:07, 399.20it/s]\u001b[A\n",
741
+ " 62% 4751/7609 [00:14<00:06, 408.94it/s]\u001b[A\n",
742
+ " 63% 4792/7609 [00:14<00:07, 390.51it/s]\u001b[A\n",
743
+ " 64% 4834/7609 [00:14<00:06, 398.83it/s]\u001b[A\n",
744
+ " 64% 4875/7609 [00:15<00:06, 394.43it/s]\u001b[A\n",
745
+ " 65% 4916/7609 [00:15<00:06, 398.01it/s]\u001b[A\n",
746
+ " 65% 4956/7609 [00:15<00:07, 377.41it/s]\u001b[A\n",
747
+ " 66% 4998/7609 [00:15<00:06, 388.68it/s]\u001b[A\n",
748
+ " 66% 5038/7609 [00:15<00:06, 386.58it/s]\u001b[A\n",
749
+ " 67% 5077/7609 [00:15<00:06, 386.95it/s]\u001b[A\n",
750
+ " 67% 5118/7609 [00:15<00:06, 393.49it/s]\u001b[A\n",
751
+ " 68% 5158/7609 [00:16<00:11, 222.22it/s]\u001b[A\n",
752
+ " 68% 5200/7609 [00:16<00:09, 259.28it/s]\u001b[A\n",
753
+ " 69% 5242/7609 [00:16<00:08, 292.90it/s]\u001b[A\n",
754
+ " 69% 5282/7609 [00:16<00:07, 316.37it/s]\u001b[A\n",
755
+ " 70% 5322/7609 [00:16<00:06, 334.94it/s]\u001b[A\n",
756
+ " 70% 5362/7609 [00:16<00:06, 350.70it/s]\u001b[A\n",
757
+ " 71% 5401/7609 [00:16<00:06, 355.94it/s]\u001b[A\n",
758
+ " 72% 5443/7609 [00:16<00:05, 371.93it/s]\u001b[A\n",
759
+ " 72% 5482/7609 [00:16<00:05, 368.15it/s]\u001b[A\n",
760
+ " 73% 5522/7609 [00:16<00:05, 376.18it/s]\u001b[A\n",
761
+ " 73% 5563/7609 [00:17<00:05, 384.39it/s]\u001b[A\n",
762
+ " 74% 5605/7609 [00:17<00:05, 392.32it/s]\u001b[A\n",
763
+ " 74% 5645/7609 [00:17<00:04, 393.87it/s]\u001b[A\n",
764
+ " 75% 5687/7609 [00:17<00:04, 399.15it/s]\u001b[A\n",
765
+ " 75% 5728/7609 [00:17<00:04, 399.67it/s]\u001b[A\n",
766
+ " 76% 5769/7609 [00:17<00:04, 401.90it/s]\u001b[A\n",
767
+ " 76% 5815/7609 [00:17<00:04, 417.79it/s]\u001b[A\n",
768
+ " 77% 5857/7609 [00:17<00:04, 412.49it/s]\u001b[A\n",
769
+ " 78% 5899/7609 [00:17<00:04, 396.45it/s]\u001b[A\n",
770
+ " 78% 5939/7609 [00:17<00:04, 397.23it/s]\u001b[A\n",
771
+ " 79% 5979/7609 [00:18<00:04, 380.92it/s]\u001b[A\n",
772
+ " 79% 6019/7609 [00:18<00:04, 384.73it/s]\u001b[A\n",
773
+ " 80% 6060/7609 [00:18<00:03, 390.43it/s]\u001b[A\n",
774
+ " 80% 6100/7609 [00:18<00:03, 389.08it/s]\u001b[A\n",
775
+ " 81% 6139/7609 [00:18<00:04, 357.51it/s]\u001b[A\n",
776
+ " 81% 6176/7609 [00:18<00:04, 343.01it/s]\u001b[A\n",
777
+ " 82% 6219/7609 [00:18<00:03, 366.55it/s]\u001b[A\n",
778
+ " 82% 6261/7609 [00:18<00:03, 380.56it/s]\u001b[A\n",
779
+ " 83% 6302/7609 [00:18<00:03, 387.32it/s]\u001b[A\n",
780
+ " 83% 6342/7609 [00:19<00:03, 387.59it/s]\u001b[A\n",
781
+ " 84% 6385/7609 [00:19<00:03, 399.56it/s]\u001b[A\n",
782
+ " 84% 6429/7609 [00:19<00:02, 409.05it/s]\u001b[A\n",
783
+ " 85% 6471/7609 [00:19<00:02, 410.40it/s]\u001b[A\n",
784
+ " 86% 6513/7609 [00:19<00:02, 412.63it/s]\u001b[A\n",
785
+ " 86% 6555/7609 [00:19<00:02, 410.10it/s]\u001b[A\n",
786
+ " 87% 6597/7609 [00:19<00:02, 411.45it/s]\u001b[A\n",
787
+ " 87% 6639/7609 [00:19<00:02, 400.74it/s]\u001b[A\n",
788
+ " 88% 6680/7609 [00:19<00:02, 381.14it/s]\u001b[A\n",
789
+ " 88% 6720/7609 [00:19<00:02, 383.86it/s]\u001b[A\n",
790
+ " 89% 6759/7609 [00:20<00:02, 382.16it/s]\u001b[A\n",
791
+ " 89% 6798/7609 [00:20<00:02, 379.85it/s]\u001b[A\n",
792
+ " 90% 6837/7609 [00:20<00:02, 374.69it/s]\u001b[A\n",
793
+ " 90% 6875/7609 [00:20<00:02, 361.02it/s]\u001b[A\n",
794
+ " 91% 6917/7609 [00:20<00:01, 375.71it/s]\u001b[A\n",
795
+ " 91% 6959/7609 [00:20<00:01, 386.62it/s]\u001b[A\n",
796
+ " 92% 7001/7609 [00:20<00:01, 393.89it/s]\u001b[A\n",
797
+ " 93% 7041/7609 [00:20<00:01, 392.62it/s]\u001b[A\n",
798
+ " 93% 7081/7609 [00:20<00:01, 363.88it/s]\u001b[A\n",
799
+ " 94% 7118/7609 [00:21<00:01, 357.49it/s]\u001b[A\n",
800
+ " 94% 7155/7609 [00:21<00:01, 351.27it/s]\u001b[A\n",
801
+ " 95% 7193/7609 [00:21<00:01, 355.56it/s]\u001b[A\n",
802
+ " 95% 7232/7609 [00:21<00:01, 361.94it/s]\u001b[A\n",
803
+ " 96% 7269/7609 [00:21<00:00, 362.41it/s]\u001b[A\n",
804
+ " 96% 7311/7609 [00:21<00:00, 377.50it/s]\u001b[A\n",
805
+ " 97% 7349/7609 [00:21<00:00, 364.61it/s]\u001b[A\n",
806
+ " 97% 7389/7609 [00:21<00:00, 374.59it/s]\u001b[A\n",
807
+ " 98% 7429/7609 [00:21<00:00, 377.78it/s]\u001b[A\n",
808
+ " 98% 7467/7609 [00:22<00:00, 360.20it/s]\u001b[A\n",
809
+ " 99% 7504/7609 [00:22<00:00, 353.22it/s]\u001b[A\n",
810
+ " 99% 7540/7609 [00:22<00:00, 343.34it/s]\u001b[A\n",
811
+ "100% 7575/7609 [00:22<00:00, 335.47it/s]\u001b[A\n",
812
+ "100% 7609/7609 [00:22<00:00, 338.94it/s]\n",
813
+ "12/09/2023 13:06:45 - INFO - utils_qa - Saving predictions to eval_results/predict_predictions.json.\n",
814
+ "12/09/2023 13:06:45 - INFO - utils_qa - Saving nbest_preds to eval_results/predict_nbest_predictions.json.\n",
815
+ "***** predict metrics *****\n",
816
+ " predict_samples = 7988\n",
817
+ " test_exact_match = 60.7307\n",
818
+ " test_f1 = 72.4316\n",
819
+ " test_runtime = 0:00:40.66\n",
820
+ " test_samples_per_second = 196.422\n",
821
+ " test_steps_per_second = 39.294\n",
822
+ "[INFO|modelcard.py:452] 2023-12-09 13:06:49,887 >> Dropping the following result as it does not have all the necessary fields:\n",
823
+ "{'task': {'name': 'Question Answering', 'type': 'question-answering'}, 'dataset': {'name': 'squad_it', 'type': 'squad_it'}}\n",
824
+ "100% 1598/1598 [01:14<00:00, 21.57it/s]\n"
825
+ ]
826
+ }
827
+ ]
828
+ },
829
+ {
830
+ "cell_type": "code",
831
+ "source": [],
832
+ "metadata": {
833
+ "id": "ij8YkY0xyNXI"
834
+ },
835
+ "execution_count": null,
836
+ "outputs": []
837
+ }
838
+ ]
839
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fec99672ee8429584c79e03f6ed4a8357870a2b4d54d939fb277a3cd9edbb141
3
+ size 90262889
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "clean_up_tokenization_spaces": true,
4
+ "cls_token": "<s>",
5
+ "eos_token": "</s>",
6
+ "mask_token": {
7
+ "__type": "AddedToken",
8
+ "content": "<mask>",
9
+ "lstrip": true,
10
+ "normalized": true,
11
+ "rstrip": false,
12
+ "single_word": false
13
+ },
14
+ "model_max_length": 512,
15
+ "pad_token": "<pad>",
16
+ "sep_token": "</s>",
17
+ "tokenizer_class": "XLMRobertaTokenizer",
18
+ "unk_token": "<unk>"
19
+ }