|
94 | 94 | "name": "stderr", |
95 | 95 | "output_type": "stream", |
96 | 96 | "text": [ |
97 | | - "2022-01-05 20:42:12 INFO: Load pretrained SentenceTransformer: allenai-specter\n", |
98 | | - "2022-01-05 20:42:36 INFO: Use pytorch device: cuda\n", |
99 | | - "2022-01-05 20:42:36 INFO: Missing data detected. Dropping them\n", |
100 | | - "2022-01-05 20:42:36 INFO: ========== Step1: Calculating Embeddings ==========\n", |
101 | | - "Batches: 100%|██████████| 3/3 [00:02<00:00, 1.17it/s]\n", |
102 | | - "2022-01-05 20:42:41 INFO: ========== Step2: Topic modeling ==========\n", |
103 | | - "2022-01-05 20:42:41 INFO: Initializing the topic model\n", |
104 | | - "2022-01-05 20:42:41 INFO: Training the topic model\n", |
105 | | - "2022-01-05 20:42:50,425 - BERTopic - Reduced dimensionality with UMAP\n", |
106 | | - "2022-01-05 20:42:50,437 - BERTopic - Clustered UMAP embeddings with HDBSCAN\n", |
107 | | - "2022-01-05 20:42:50 INFO: Populating Topic Results\n", |
108 | | - "2022-01-05 20:42:50 INFO: ========== Step3: STriP Network ==========\n", |
109 | | - "2022-01-05 20:42:50 INFO: Cosine similarity\n", |
110 | | - "2022-01-05 20:42:50 INFO: Calculating optimal threshold\n", |
111 | | - "2022-01-05 20:42:50 INFO: Number of connections: 126\n", |
112 | | - "2022-01-05 20:42:50 INFO: Calculating Network Plot\n", |
113 | | - "2022-01-05 20:42:50 INFO: ========== Model Fit Successfully! ==========\n" |
| 97 | + "2022-01-06 12:16:44 INFO: Load pretrained SentenceTransformer: allenai-specter\n", |
| 98 | + "2022-01-06 12:17:07 INFO: Use pytorch device: cuda\n", |
| 99 | + "2022-01-06 12:17:07 INFO: Missing data detected. Dropping them\n", |
| 100 | + "2022-01-06 12:17:07 INFO: ========== Step1: Calculating Embeddings ==========\n", |
| 101 | + "Batches: 100%|██████████| 3/3 [00:02<00:00, 1.11it/s]\n", |
| 102 | + "2022-01-06 12:17:12 INFO: ========== Step2: Topic modeling ==========\n", |
| 103 | + "2022-01-06 12:17:12 INFO: Initializing the topic model\n", |
| 104 | + "2022-01-06 12:17:12 INFO: Training the topic model\n", |
| 105 | + "2022-01-06 12:17:21,291 - BERTopic - Reduced dimensionality with UMAP\n", |
| 106 | + "2022-01-06 12:17:21,304 - BERTopic - Clustered UMAP embeddings with HDBSCAN\n", |
| 107 | + "2022-01-06 12:17:21 INFO: Populating Topic Results\n", |
| 108 | + "2022-01-06 12:17:21 INFO: ========== Step3: STriP Network ==========\n", |
| 109 | + "2022-01-06 12:17:21 INFO: Cosine similarity\n", |
| 110 | + "2022-01-06 12:17:21 INFO: Calculating optimal threshold\n", |
| 111 | + "2022-01-06 12:17:21 INFO: Number of connections: 126\n", |
| 112 | + "2022-01-06 12:17:21 INFO: Calculating Network Plot\n", |
| 113 | + "2022-01-06 12:17:21 INFO: ========== Model Fit Successfully! ==========\n" |
114 | 114 | ] |
115 | 115 | }, |
116 | 116 | { |
|
139 | 139 | "name": "stderr", |
140 | 140 | "output_type": "stream", |
141 | 141 | "text": [ |
142 | | - "2022-01-05 20:07:00 INFO: Calculating Network Centrality\n" |
| 142 | + "2022-01-06 12:18:23 INFO: Calculating Network Centrality\n" |
143 | 143 | ] |
144 | 144 | }, |
145 | 145 | { |
|
161 | 161 | [ |
162 | 162 | "Want To Reduce Labeling Cost? GPT-3 Can<br>Help<br><br>Data annotation is a time-consuming<br>and labor-intensive process for many NLP tasks.<br>Although there exist various methods to produce<br>pseudo data labels, they are often task-specific<br>and require a decent amount of labeled data to<br>start with. Recently, the immense language model<br>GPT-3 with 175 billion parameters has achieved<br>tremendous improvement across many few-shot<br>learning tasks. In this paper, we explore ways to<br>..." |
163 | 163 | ], |
| 164 | + [ |
| 165 | + "FNet: Mixing Tokens with Fourier<br>Transforms<br><br>We show that Transformer encoder<br>architec-tures can be massively sped up, with<br>limited accuracy costs, by replacing the self-<br>attention sublayers with simple linear<br>transformations that \"mix\" input tokens. These<br>linear transformations , along with simple<br>nonlinearities in feed-forward layers, are<br>sufficient to model semantic relationships in<br>several text classification tasks. Perhaps most<br>surprisingly, we find that ..." |
| 166 | + ], |
164 | 167 | [ |
165 | 168 | "Neural Machine Translation of Rare Words with<br>Subword Units<br><br>Neural machine translation<br>(NMT) models typically operate with a fixed<br>vocabulary , but translation is an open-vocabulary<br>problem. Previous work addresses the translation<br>of out-of-vocabulary words by backing off to a<br>dictionary. In this paper , we introduce a simpler<br>and more effective approach, making the NMT model<br>capable of open-vocabulary translation by encoding<br>rare and unknown words as sequences ..." |
166 | 169 | ] |
|
183 | 186 | 0.10673251529415916, |
184 | 187 | 0.09934333324744282, |
185 | 188 | 0.07184737801176154, |
| 189 | + 0.05168250442223043, |
186 | 190 | 0.042019416676950916 |
187 | 191 | ], |
188 | 192 | "xaxis": "x", |
189 | 193 | "y": [ |
190 | 194 | "5", |
191 | 195 | "24", |
192 | 196 | "59", |
| 197 | + "8", |
193 | 198 | "7" |
194 | 199 | ], |
195 | 200 | "yaxis": "y" |
|
200 | 205 | [ |
201 | 206 | "An Image is Worth 16x16 Words: Transformers for<br>Image Recognition at Scale<br><br>While the<br>Transformer architecture has become the de-facto<br>standard for natural language processing tasks,<br>its applications to computer vision remain<br>limited. In vision, attention is either applied in<br>conjunction with convolutional networks, or used<br>to replace certain components of convolutional<br>networks while keeping their overall structure in<br>place. We show that this reliance on CNNs is..." |
202 | 207 | ], |
203 | | - [ |
204 | | - "Unsupervised Data Augmentation for Consistency<br>Training<br><br>Semi-supervised learning lately<br>has shown much promise in improving deep learning<br>models when labeled data is scarce. Common among<br>recent approaches is the use of consistency<br>training on a large amount of unlabeled data to<br>constrain model predictions to be invariant to<br>input noise. In this work, we present a new<br>perspective on how to effectively noise unlabeled<br>examples and argue that the quality of noising..." |
205 | | - ], |
206 | 208 | [ |
207 | 209 | "The 2021 Image Similarity Dataset and<br>Challenge<br><br>This paper introduces a new<br>benchmark for large-scale image similarity<br>detection. This benchmark is used for the Image<br>Similarity Challenge at NeurIPS'21 (ISC2021). The<br>goal is to determine whether a query image is a<br>modified copy of any image in a reference corpus<br>of size 1~million. The benchmark features a<br>variety of image transformations such as automated<br>transformations, hand-crafted image edits and<br>machine-..." |
208 | 210 | ], |
|
213 | 215 | "Learning Transferable Visual Models From Natural<br>Language Supervision<br><br>State-of-the-art<br>computer vision systems are trained to predict a<br>fixed set of predetermined object categories. This<br>restricted form of supervision limits their<br>generality and usability since additional labeled<br>data is needed to specify any other visual<br>concept. Learning directly from raw text about<br>images is a promising alternative which leverages<br>a much broader source of supervision. We<br>d..." |
214 | 216 | ] |
215 | 217 | ], |
216 | | - "hovertemplate": "Topic_Name=image, vision, learning, visual<br>Betweenness Centrality=%{x}<br>index=%{y}<br>Text=%{customdata[0]}<extra></extra>", |
217 | | - "legendgroup": "image, vision, learning, visual", |
| 218 | + "hovertemplate": "Topic_Name=image, learning, contrastive, vision<br>Betweenness Centrality=%{x}<br>index=%{y}<br>Text=%{customdata[0]}<extra></extra>", |
| 219 | + "legendgroup": "image, learning, contrastive, vision", |
218 | 220 | "marker": { |
219 | 221 | "color": "#EF553B", |
220 | 222 | "pattern": { |
221 | 223 | "shape": "" |
222 | 224 | } |
223 | 225 | }, |
224 | | - "name": "image, vision, learning, visual", |
225 | | - "offsetgroup": "image, vision, learning, visual", |
| 226 | + "name": "image, learning, contrastive, vision", |
| 227 | + "offsetgroup": "image, learning, contrastive, vision", |
226 | 228 | "orientation": "h", |
227 | 229 | "showlegend": true, |
228 | 230 | "textposition": "auto", |
229 | 231 | "type": "bar", |
230 | 232 | "x": [ |
231 | 233 | 0.07375869019704637, |
232 | | - 0.06015364679748239, |
233 | 234 | 0.059237319511292116, |
234 | 235 | 0.053904344657769304, |
235 | 236 | 0.03623643212684308 |
236 | 237 | ], |
237 | 238 | "xaxis": "x", |
238 | 239 | "y": [ |
239 | 240 | "57", |
240 | | - "18", |
241 | 241 | "42", |
242 | 242 | "1", |
243 | 243 | "34" |
|
248 | 248 | "alignmentgroup": "True", |
249 | 249 | "customdata": [ |
250 | 250 | [ |
251 | | - "FNet: Mixing Tokens with Fourier<br>Transforms<br><br>We show that Transformer encoder<br>architec-tures can be massively sped up, with<br>limited accuracy costs, by replacing the self-<br>attention sublayers with simple linear<br>transformations that \"mix\" input tokens. These<br>linear transformations , along with simple<br>nonlinearities in feed-forward layers, are<br>sufficient to model semantic relationships in<br>several text classification tasks. Perhaps most<br>surprisingly, we find that ..." |
| 251 | + "Unsupervised Data Augmentation for Consistency<br>Training<br><br>Semi-supervised learning lately<br>has shown much promise in improving deep learning<br>models when labeled data is scarce. Common among<br>recent approaches is the use of consistency<br>training on a large amount of unlabeled data to<br>constrain model predictions to be invariant to<br>input noise. In this work, we present a new<br>perspective on how to effectively noise unlabeled<br>examples and argue that the quality of noising..." |
252 | 252 | ] |
253 | 253 | ], |
254 | | - "hovertemplate": "Topic_Name=image, matching, similarity, copy<br>Betweenness Centrality=%{x}<br>index=%{y}<br>Text=%{customdata[0]}<extra></extra>", |
255 | | - "legendgroup": "image, matching, similarity, copy", |
| 254 | + "hovertemplate": "Topic_Name=learning, image, titles, product<br>Betweenness Centrality=%{x}<br>index=%{y}<br>Text=%{customdata[0]}<extra></extra>", |
| 255 | + "legendgroup": "learning, image, titles, product", |
256 | 256 | "marker": { |
257 | 257 | "color": "#00cc96", |
258 | 258 | "pattern": { |
259 | 259 | "shape": "" |
260 | 260 | } |
261 | 261 | }, |
262 | | - "name": "image, matching, similarity, copy", |
263 | | - "offsetgroup": "image, matching, similarity, copy", |
| 262 | + "name": "learning, image, titles, product", |
| 263 | + "offsetgroup": "learning, image, titles, product", |
264 | 264 | "orientation": "h", |
265 | 265 | "showlegend": true, |
266 | 266 | "textposition": "auto", |
267 | 267 | "type": "bar", |
268 | 268 | "x": [ |
269 | | - 0.05168250442223043 |
| 269 | + 0.06015364679748239 |
270 | 270 | ], |
271 | 271 | "xaxis": "x", |
272 | 272 | "y": [ |
273 | | - "8" |
| 273 | + "18" |
274 | 274 | ], |
275 | 275 | "yaxis": "y" |
276 | 276 | } |
|
1138 | 1138 | } |
1139 | 1139 | ], |
1140 | 1140 | "source": [ |
1141 | | - "stripnet.most_important()" |
1142 | | - ] |
1143 | | - }, |
1144 | | - { |
1145 | | - "cell_type": "code", |
1146 | | - "execution_count": 3, |
1147 | | - "metadata": {}, |
1148 | | - "outputs": [ |
1149 | | - { |
1150 | | - "data": { |
1151 | | - "text/plain": [ |
1152 | | - "['bertopic==0.9.4',\n", |
1153 | | - " 'networkx==2.6.3',\n", |
1154 | | - " 'numpy==1.22.0',\n", |
1155 | | - " 'pandas==1.3.5',\n", |
1156 | | - " 'plotly==5.5.0',\n", |
1157 | | - " 'pyvis==0.1.9',\n", |
1158 | | - " 'scikit_learn==1.0.2',\n", |
1159 | | - " 'sentence_transformers==2.1.0',\n", |
1160 | | - " 'setuptools==58.0.4']" |
1161 | | - ] |
1162 | | - }, |
1163 | | - "execution_count": 3, |
1164 | | - "metadata": {}, |
1165 | | - "output_type": "execute_result" |
1166 | | - } |
1167 | | - ], |
1168 | | - "source": [ |
1169 | | - "import pathlib\n", |
1170 | | - "pathlib.Path(\"../requirements.txt\").read_text().splitlines()" |
| 1141 | + "stripnet.most_important_docs()" |
1171 | 1142 | ] |
1172 | 1143 | }, |
1173 | 1144 | { |
|
1183 | 1154 | "hash": "165d1ae889830a583229da7bcb4f0175182080283a5d782889056a279531f3b2" |
1184 | 1155 | }, |
1185 | 1156 | "kernelspec": { |
1186 | | - "display_name": "Python 3.8.12 64-bit ('stripnet': conda)", |
| 1157 | + "display_name": "Python 3 (ipykernel)", |
1187 | 1158 | "language": "python", |
1188 | 1159 | "name": "python3" |
1189 | 1160 | }, |
|
1198 | 1169 | "nbconvert_exporter": "python", |
1199 | 1170 | "pygments_lexer": "ipython3", |
1200 | 1171 | "version": "3.8.12" |
1201 | | - }, |
1202 | | - "orig_nbformat": 4 |
| 1172 | + } |
1203 | 1173 | }, |
1204 | 1174 | "nbformat": 4, |
1205 | | - "nbformat_minor": 2 |
| 1175 | + "nbformat_minor": 4 |
1206 | 1176 | } |
0 commit comments