generated from slds-lmu/seminar_website_skeleton
-
Notifications
You must be signed in to change notification settings - Fork 27
/
book.bib.backup
3606 lines (3146 loc) · 164 KB
/
book.bib.backup
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@ARTICLE{Glide2021,
author = {Alex Nichol and
Prafulla Dhariwal and
Aditya Ramesh and
Pranav Shyam and
Pamela Mishkin and
Bob McGrew and
Ilya Sutskever and
Mark Chen},
title = {{GLIDE:} Towards Photorealistic Image Generation and Editing with
Text-Guided Diffusion Models},
journal = {CoRR},
volume = {abs/2112.10741},
year = {2021},
url = {https://arxiv.org/abs/2112.10741},
eprinttype = {arXiv},
eprint = {2112.10741},
timestamp = {Tue, 04 Jan 2022 15:59:27 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2112-10741.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
}
@ARTICLE{Lu2020,
AUTHOR = {Yu, Jun and Li, Jing and Yu, Zhou and Huang, Qingming},
YEAR = {2020},
DOI = {10.1109/TCSVT.2019.2947482},
JOURNALTITLE = {IEEE Transactions on Circuits and Systems for Video Technology},
NUMBER = {12},
PAGES = {4467--4480},
TITLE = {Multimodal Transformer With Multi-View Visual Representation for Image Captioning},
VOLUME = {30},
}
@MISC{Fedus2021,
AUTHOR = {Fedus, William and Zoph, Barret and Shazeer, Noam},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/2101.03961},
YEAR = {2021},
DOI = {10.48550/ARXIV.2101.03961},
KEYWORDS = {Machine Learning (cs.LG),Artificial Intelligence (cs.AI),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity},
}
@MISC{Mustafa2022,
YEAR = {2022},
AUTHOR = {Mustafa, Basil and Riquelme, Carlos and Puigcerver, Joan and Jenatton, Rodolphe and Houlsby, Neil},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/2206.02770},
DOI = {10.48550/ARXIV.2206.02770},
KEYWORDS = {Computer Vision and Pattern Recognition (cs.CV),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Multimodal Contrastive Learning with LIMoE: the Language-Image Mixture of Experts},
}
@ARTICLE{Carion2020,
YEAR = {2020},
AUTHOR = {Carion, Nicolas and Massa, Francisco and Synnaeve, Gabriel and Usunier, Nicolas and Kirillov, Alexander and Zagoruyko, Sergey},
URL = {https://arxiv.org/abs/2005.12872},
EPRINT = {2005.12872},
EPRINTTYPE = {arXiv},
JOURNALTITLE = {CoRR},
TITLE = {End-to-End Object Detection with Transformers},
}
@MISC{Crawshaw2020,
YEAR = {2020},
AUTHOR = {Crawshaw, Michael},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/2009.09796},
DOI = {10.48550/ARXIV.2009.09796},
KEYWORDS = {Machine Learning (cs.LG),Computer Vision and Pattern Recognition (cs.CV),Machine Learning (stat.ML),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Multi-Task Learning with Deep Neural Networks: A Survey},
}
@ARTICLE{Baltrusaitis2019,
YEAR = {2019},
AUTHOR = {Baltrušaitis, Tadas and Ahuja, Chaitanya and Morency, Louis-Philippe},
DOI = {10.1109/TPAMI.2018.2798607},
JOURNALTITLE = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
NUMBER = {2},
PAGES = {423--443},
TITLE = {Multimodal Machine Learning: A Survey and Taxonomy},
VOLUME = {41},
}
@ARTICLE{Kaiser2017,
YEAR = {2017},
AUTHOR = {Kaiser, Lukasz and Gomez, Aidan N. and Shazeer, Noam and Vaswani, Ashish and Parmar, Niki and Jones, Llion and Uszkoreit, Jakob},
URL = {https://arxiv.org/pdf/1706.05137.pdf},
JOURNALTITLE = {arXiv},
TITLE = {One Model To Learn Them All},
}
@INPROCEEDINGS{Hu2021,
YEAR = {2021},
AUTHOR = {Hu, Ronghang and Singh, Amanpreet},
BOOKTITLE = {2021 IEEE/CVF International Conference on Computer Vision (ICCV)},
DOI = {10.1109/ICCV48922.2021.00147},
PAGES = {1419--1429},
TITLE = {UniT: Multimodal Multitask Learning with a Unified Transformer},
}
@MISC{Li2019,
AUTHOR = {Li, Liunian Harold and Yatskar, Mark and Yin, Da and Hsieh, Cho-Jui and Chang, Kai-Wei},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/1908.03557},
YEAR = {2019},
DOI = {10.48550/ARXIV.1908.03557},
KEYWORDS = {Computer Vision and Pattern Recognition (cs.CV),Computation and Language (cs.CL),Machine Learning (cs.LG),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {VisualBERT: A Simple and Performant Baseline for Vision and Language},
}
@ONLINE{Dean21,
AUTHOR = {Dean, Jeff},
URL = {https://blog.google/technology/ai/introducing-pathways-next-generation-ai-architecture/},
YEAR = {2021},
TITLE = {Introducing Pathways: A next-generation AI architecture},
}
@ARTICLE{Krishna2017,
AUTHOR = {Krishna, Ranjay and Zhu, Yuke and Groth, Oliver and Johnson, Justin and Hata, Kenji and Kravitz, Joshua and Chen, Stephanie and Kalantidis, Yannis and Li, Li-Jia and Shamma, David A. and Bernstein, Michael S. and Fei-Fei, Li},
LOCATION = {USA},
PUBLISHER = {Kluwer Academic Publishers},
URL = {https://DOI.org/10.1007/s11263-016-0981-7},
YEAR = {2017},
DOI = {10.1007/s11263-016-0981-7},
ISSN = {0920-5691},
JOURNALTITLE = {Int. J. Comput. Vision},
KEYWORDS = {Language,Relationships,Attributes,Question answering,Scene graph,Crowdsourcing,Computer vision,Knowledge,Image,Objects,Dataset},
NUMBER = {1},
PAGES = {32--73},
TITLE = {Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations},
VOLUME = {123},
}
@INPROCEEDINGS{Wang2022,
AUTHOR = {Wang, Peng and Yang, An and Men, Rui and Lin, Junyang and Bai, Shuai and Li, Zhikang and Ma, Jianxin and Zhou, Chang and Zhou, Jingren and Yang, Hongxia},
EDITOR = {Chaudhuri, Kamalika and Jegelka, Stefanie and Song, Le and Szepesvari, Csaba and Niu, Gang and Sabato, Sivan},
PUBLISHER = {PMLR},
URL = {https://proceedings.mlr.press/v162/wang22al.html},
BOOKTITLE = {Proceedings of the 39th International Conference on Machine Learning},
YEAR = {2022},
FILE = {https://proceedings.mlr.press/v162/wang22al/wang22al.pdf},
PAGES = {23318--23340},
SERIES = {Proceedings of Machine Learning Research},
TITLE = {{OFA}: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework},
VOLUME = {162},
}
@MISC{Reed2022,
AUTHOR = {Reed, Scott and Zolna, Konrad and Parisotto, Emilio and Colmenarejo, Sergio Gomez and Novikov, Alexander and Barth-Maron, Gabriel and Gimenez, Mai and Sulsky, Yury and Kay, Jackie and Springenberg, Jost Tobias and Eccles, Tom and Bruce, Jake and Razavi, Ali and Edwards, Ashley and Heess, Nicolas and Chen, Yutian and Hadsell, Raia and Vinyals, Oriol and Bordbar, Mahyar and de Freitas, Nando},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/2205.06175},
YEAR = {2022},
DOI = {10.48550/ARXIV.2205.06175},
KEYWORDS = {Artificial Intelligence (cs.AI),Computation and Language (cs.CL),Machine Learning (cs.LG),Robotics (cs.RO),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {A Generalist Agent},
}
@ARTICLE{Chowdhery2022,
YEAR = {2022},
AUTHOR = {Chowdhery, Aakanksha and Narang, Sharan and Devlin, Jacob and Bosma, Maarten and Mishra, Gaurav and Roberts, Adam and Barham, Paul and Chung, Hyung Won and Sutton, Charles and Gehrmann, Sebastian and Schuh, Parker and Shi, Kensen and Tsvyashchenko, Sasha and Maynez, Joshua and Rao, Abhishek and Barnes, Parker and Tay, Yi and Shazeer, Noam and Prabhakaran, Vinodkumar and Reif, Emily and Du, Nan and Hutchinson, Ben and Pope, Reiner and Bradbury, James and Austin, Jacob and Isard, Michael and Gur-Ari, Guy and Yin, Pengcheng and Duke, Toju and Levskaya, Anselm and Ghemawat, Sanjay and Dev, Sunipa and Michalewski, Henryk and Garcia, Xavier and Misra, Vedant and Robinson, Kevin and Fedus, Liam and Zhou, Denny and Ippolito, Daphne and Luan, David and Lim, Hyeontaek and Zoph, Barret and Spiridonov, Alexander and Sepassi, Ryan and Dohan, David and Agrawal, Shivani and Omernick, Mark and Dai, Andrew M. and Pillai, Thanumalayan Sankaranarayana and Pellat, Marie and Lewkowycz, Aitor and Moreira, Erica and Child, Rewon and Polozov, Oleksandr and Lee, Katherine and Zhou, Zongwei and Wang, Xuezhi and Saeta, Brennan and Diaz, Mark and Firat, Orhan and Catasta, Michele and Wei, Jason and Meier-Hellstern, Kathy and Eck, Douglas and Dean, Jeff and Petrov, Slav and Fiedel, Noah},
URL = {https://arxiv.org/abs/2204.02311},
JOURNALTITLE = {arxiv:2204.02311},
TITLE = {PaLM: Scaling Language Modeling with Pathways},
}
@MISC{Yu2022,
AUTHOR = {Yu, Jiahui and Xu, Yuanzhong and Koh, Jing Yu and Luong, Thang and Baid, Gunjan and Wang, Zirui and Vasudevan, Vijay and Ku, Alexander and Yang, Yinfei and Ayan, Burcu Karagol and Hutchinson, Ben and Han, Wei and Parekh, Zarana and Li, Xin and Zhang, Han and Baldridge, Jason and Wu, Yonghui},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/2206.10789},
YEAR = {2022},
DOI = {10.48550/ARXIV.2206.10789},
KEYWORDS = {Computer Vision and Pattern Recognition (cs.CV),Machine Learning (cs.LG),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Scaling Autoregressive Models for Content-Rich Text-to-Image Generation},
}
@INPROCEEDINGS{Fernando2017,
AUTHOR = {Fernando, Chrisantha and Banarse, Dylan and Blundell, Charles and Zwols, Yori and Ha, David and Rusu, Andrei A. and Pritzel, Alexander and Wierstra, Daan},
URL = {https://arxiv.org/abs/1701.08734},
YEAR = {2017},
TITLE = {PathNet: Evolution Channels Gradient Descent in Super Neural Networks},
}
@INPROCEEDINGS{He2016b,
AUTHOR = {He, Kaiming and Zhang, Xiangyu and Ren, Shaoqing and Sun, Jian},
EDITOR = {Leibe, Bastian and Matas, Jiri and Sebe, Nicu and Welling, Max},
LOCATION = {Cham},
PUBLISHER = {Springer International Publishing},
BOOKTITLE = {Computer Vision -- ECCV 2016},
YEAR = {2016},
ISBN = {978-3-319-46493-0},
PAGES = {630--645},
TITLE = {Identity Mappings in Deep Residual Networks},
}
@INPROCEEDINGS{Dean20,
AUTHOR = {Dean, Jeffrey},
BOOKTITLE = {2020 IEEE International Solid- State Circuits Conference - (ISSCC)},
YEAR = {2020},
DOI = {10.1109/ISSCC19947.2020.9063049},
PAGES = {8--14},
TITLE = {1.1 The Deep Learning Revolution and Its Implications for Computer Architecture and Chip Design},
}
@REPORT{Lewkowycz2022,
AUTHOR = {Lewkowycz, Aitor and Andreassen, Anders and Dohan, David Martin and Dyer, Ethan S and Michalewski, Henryk and Ramasesh, Vinay and Slone, Ambrose and Anil, Cem and Schlag, Imanol and Gutman-Solo, Theo and Wu, Yuhuai and Neyshabur, Behnam and Gur-Ari, Guy and Misra, Vedant},
URL = {https://arxiv.org/abs/2206.14858},
YEAR = {2022},
TITLE = {Solving Quantitative Reasoning Problems with Language Models},
TYPE = {techreport},
}
@INPROCEEDINGS{Riquelme2021,
AUTHOR = {Riquelme, Carlos and Puigcerver, Joan and Mustafa, Basil and Neumann, Maxim and Jenatton, Rodolphe and Susano Pinto, André and Keysers, Daniel and Houlsby, Neil},
EDITOR = {Ranzato, M. and Beygelzimer, A. and Dauphin, Y. and Liang, P.S. and Vaughan, J. Wortman},
PUBLISHER = {Curran Associates, Inc.},
URL = {https://proceedings.neurips.cc/paper/2021/file/48237d9f2dea8c74c2a72126cf63d933-Paper.pdf},
BOOKTITLE = {Advances in Neural Information Processing Systems},
YEAR = {2021},
PAGES = {8583--8595},
TITLE = {Scaling Vision with Sparse Mixture of Experts},
VOLUME = {34},
}
@MISC{Gesmundo2022a,
AUTHOR = {Gesmundo, Andrea and Dean, Jeff},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/2205.10937},
YEAR = {2022},
DOI = {10.48550/ARXIV.2205.10937},
KEYWORDS = {Machine Learning (cs.LG),Artificial Intelligence (cs.AI),Computer Vision and Pattern Recognition (cs.CV),Neural and Evolutionary Computing (cs.NE),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {muNet: Evolving Pretrained Deep Neural Networks into Scalable Auto-tuning Multitask Systems},
}
@ARTICLE{Steiner2021,
YEAR = {2021},
AUTHOR = {Steiner, Andreas and Kolesnikov, Alexander and Zhai, Xiaohua and Wightman, Ross and Uszkoreit, Jakob and Beyer, Lucas},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/2106.10270},
DOI = {10.48550/ARXIV.2106.10270},
KEYWORDS = {Computer Vision and Pattern Recognition (cs.CV),Artificial Intelligence (cs.AI),Machine Learning (cs.LG),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers},
}
@INPROCEEDINGS{Houlsby2019,
ABSTRACT = {Fine-tuning large pretrained models is an effective transfer mechanism in NLP. However, in the presence of many downstream tasks, fine-tuning is parameter inefficient: an entire new model is required for every task. As an alternative, we propose transfer with adapter modules. Adapter modules yield a compact and extensible model; they add only a few trainable parameters per task, and new tasks can be added without revisiting previous ones. The parameters of the original network remain fixed, yielding a high degree of parameter sharing. To demonstrate adapter’s effectiveness, we transfer the recently proposed BERT Transformer model to $26$ diverse text classification tasks, including the GLUE benchmark. Adapters attain near state-of-the-art performance, whilst adding only a few parameters per task. On GLUE, we attain within $0.8%$ of the performance of full fine-tuning, adding only $3.6%$ parameters per task. By contrast, fine-tuning trains $100%$ of the parameters per task.},
AUTHOR = {Houlsby, Neil and Giurgiu, Andrei and Jastrzebski, Stanislaw and Morrone, Bruna and De Laroussilhe, Quentin and Gesmundo, Andrea and Attariyan, Mona and Gelly, Sylvain},
EDITOR = {Chaudhuri, Kamalika and Salakhutdinov, Ruslan},
PUBLISHER = {PMLR},
URL = {https://proceedings.mlr.press/v97/houlsby19a.html},
BOOKTITLE = {Proceedings of the 36th International Conference on Machine Learning},
YEAR = {2019},
FILE = {http://proceedings.mlr.press/v97/houlsby19a/houlsby19a.pdf},
PAGES = {2790--2799},
SERIES = {Proceedings of Machine Learning Research},
TITLE = {Parameter-Efficient Transfer Learning for {NLP}},
VOLUME = {97},
}
@INPROCEEDINGS{Rebuffi2017,
AUTHOR = {Rebuffi, Sylvestre-Alvise and Bilen, Hakan and Vedaldi, Andrea},
EDITOR = {Guyon, I. and Luxburg, U. Von and Bengio, S. and Wallach, H. and Fergus, R. and Vishwanathan, S. and Garnett, R.},
PUBLISHER = {Curran Associates, Inc.},
URL = {https://proceedings.neurips.cc/paper/2017/file/e7b24b112a44fdd9ee93bdf998c6ca0e-Paper.pdf},
BOOKTITLE = {Advances in Neural Information Processing Systems},
YEAR = {2017},
TITLE = {Learning multiple visual domains with residual adapters},
VOLUME = {30},
}
@MISC{Bilen2017,
AUTHOR = {Bilen, Hakan and Rebuffi, SSylvestre and Jakab, Tomas},
YEAR = {2017},
TITLE = {Visual domain decathlon},
}
@ARTICLE{Doerr2021,
YEAR = {2021},
AUTHOR = {Doerr, Benjamin and Neumann, Frank},
LOCATION = {New York, NY, USA},
PUBLISHER = {Association for Computing Machinery},
URL = {https://doi.org/10.1145/3472304},
DOI = {10.1145/3472304},
ISSN = {2688-299X},
JOURNALTITLE = {ACM Trans. Evol. Learn. Optim.},
KEYWORDS = {parameterized complexity,discrete optimization,evolutionary algorithms,estimation of distribution algorithms,Theory},
NUMBER = {4},
TITLE = {A Survey on Recent Progress in the Theory of Evolutionary Algorithms for Discrete Optimization},
VOLUME = {1},
}
@ARTICLE{Baeck1993,
YEAR = {1993},
AUTHOR = {B{\"a}ck, Thomas and Schwefel, Hans-Paul},
DOI = {10.1162/evco.1993.1.1.1},
JOURNALTITLE = {Evolutionary Computation},
NUMBER = {1},
PAGES = {1--23},
TITLE = {An Overview of Evolutionary Algorithms for Parameter Optimization},
VOLUME = {1},
}
@MISC{Hinton2015,
YEAR = {2015},
AUTHOR = {Hinton, Geoffrey and Vinyals, Oriol and Dean, Jeff},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/1503.02531},
DOI = {10.48550/ARXIV.1503.02531},
KEYWORDS = {Machine Learning (stat.ML),Machine Learning (cs.LG),Neural and Evolutionary Computing (cs.NE),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Distilling the Knowledge in a Neural Network},
}
@INPROCEEDINGS{Shaazer2017,
YEAR = {2017},
AUTHOR = {Shazeer, Noam and Mirhoseini, Azalia and Maziarz, Krzysztof and Davis, Andy and Le, Quoc and Hinton, Geoffrey and Dean, Jeff},
URL = {https://openreview.net/pdf?id=B1ckMDqlg},
TITLE = {Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer},
}
@ARTICLE{Jordan1994,
YEAR = {1994},
AUTHOR = {Jordan, Michael I. and Jacobs, Robert A.},
DOI = {10.1162/neco.1994.6.2.181},
JOURNALTITLE = {Neural Computation},
NUMBER = {2},
PAGES = {181--214},
TITLE = {Hierarchical Mixtures of Experts and the EM Algorithm},
VOLUME = {6},
}
@ARTICLE{Jacobs1991,
YEAR = {1991},
AUTHOR = {Jacobs, Robert A. and Jordan, Michael I. and Nowlan, Steven J. and Hinton, Geoffrey E.},
DOI = {10.1162/neco.1991.3.1.79},
JOURNALTITLE = {Neural Computation},
NUMBER = {1},
PAGES = {79--87},
TITLE = {Adaptive Mixtures of Local Experts},
VOLUME = {3},
}
@INPROCEEDINGS{sennrich-etal-2016-neural,
YEAR = {2016},
AUTHOR = {Sennrich, Rico and Haddow, Barry and Birch, Alexandra},
LOCATION = {Berlin, Germany},
PUBLISHER = {Association for Computational Linguistics},
URL = {https://aclanthology.org/P16-1162},
BOOKTITLE = {Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
DOI = {10.18653/v1/P16-1162},
PAGES = {1715--1725},
TITLE = {Neural Machine Translation of Rare Words with Subword Units},
}
@INPROCEEDINGS{pmlr-v139-ramesh21a,
YEAR = {2021},
AUTHOR = {Ramesh, Aditya and Pavlov, Mikhail and Goh, Gabriel and Gray, Scott and Voss, Chelsea and Radford, Alec and Chen, Mark and Sutskever, Ilya},
EDITOR = {Meila, Marina and Zhang, Tong},
PUBLISHER = {PMLR},
URL = {https://proceedings.mlr.press/v139/ramesh21a.html},
BOOKTITLE = {Proceedings of the 38th International Conference on Machine Learning},
FILE = {http://proceedings.mlr.press/v139/ramesh21a/ramesh21a.pdf},
PAGES = {8821--8831},
SERIES = {Proceedings of Machine Learning Research},
TITLE = {Zero-Shot Text-to-Image Generation},
VOLUME = {139},
}
@ARTICLE{ResNet,
YEAR = {2015},
AUTHOR = {He, Kaiming and Zhang, Xiangyu and Ren, Shaoqing and Sun, Jian},
URL = {http://arxiv.org/abs/1512.03385},
EPRINT = {1512.03385},
EPRINTTYPE = {arXiv},
JOURNALTITLE = {CoRR},
TITLE = {Deep Residual Learning for Image Recognition},
}
@INPROCEEDINGS{mccoco,
YEAR = {2014},
AUTHOR = {Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James and Perona, Pietro and Ramanan, Deva and Dollár, Piotr and Zitnick, C. Lawrence},
PUBLISHER = {Springer International Publishing},
BOOKTITLE = {Computer Vision -- ECCV 2014},
ISBN = {978-3-319-10602-1},
PAGES = {740--755},
TITLE = {Microsoft COCO: Common Objects in Context},
}
@INPROCEEDINGS{kudo-richardson-2018-sentencepiece,
YEAR = {2018},
AUTHOR = {Kudo, Taku and Richardson, John},
LOCATION = {Brussels, Belgium},
PUBLISHER = {Association for Computational Linguistics},
URL = {https://aclanthology.org/D18-2012},
BOOKTITLE = {Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations},
DOI = {10.18653/v1/D18-2012},
PAGES = {66--71},
TITLE = {{S}entence{P}iece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing},
}
@ARTICLE{Devlin2018,
YEAR = {2018},
AUTHOR = {Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
EPRINT = {1810.04805},
EPRINTCLASS = {cs.CL},
EPRINTTYPE = {arXiv},
FILE = {:http\://arxiv.org/pdf/1810.04805v2:PDF},
KEYWORDS = {cs.CL},
TITLE = {BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding},
}
@ARTICLE{brown2020language,
YEAR = {2020},
AUTHOR = {Brown, Tom and Mann, Benjamin and Ryder, Nick and Subbiah, Melanie and Kaplan, Jared D and Dhariwal, Prafulla and Neelakantan, Arvind and Shyam, Pranav and Sastry, Girish and Askell, Amanda and others},
JOURNALTITLE = {Advances in neural information processing systems},
PAGES = {1877--1901},
TITLE = {Language models are few-shot learners},
VOLUME = {33},
}
@ARTICLE{ImageNet,
YEAR = {2015},
AUTHOR = {Russakovsky, Olga and Deng, Jia and Su, Hao and Krause, Jonathan and Satheesh, Sanjeev and Ma, Sean and Huang, Zhiheng and Karpathy, Andrej and Khosla, Aditya and Bernstein, Michael and Berg, Alexander C. and Fei-Fei, Li},
LOCATION = {USA},
PUBLISHER = {Kluwer Academic Publishers},
URL = {https://doi.org/10.1007/s11263-015-0816-y},
DOI = {10.1007/s11263-015-0816-y},
ISSN = {0920-5691},
JOURNALTITLE = {Int. J. Comput. Vision},
KEYWORDS = {Benchmark,Object detection,Large-scale,Object recognition,Dataset},
NUMBER = {3},
PAGES = {211--252},
TITLE = {ImageNet Large Scale Visual Recognition Challenge},
VOLUME = {115},
}
@ARTICLE{dosovitskiy2020image,
YEAR = {2020},
AUTHOR = {Dosovitskiy, Alexey and Beyer, Lucas and Kolesnikov, Alexander and Weissenborn, Dirk and Zhai, Xiaohua and Unterthiner, Thomas and Dehghani, Mostafa and Minderer, Matthias and Heigold, Georg and Gelly, Sylvain and others},
JOURNALTITLE = {arXiv preprint arXiv:2010.11929},
TITLE = {An image is worth 16x16 words: Transformers for image recognition at scale},
}
@ARTICLE{vaswani2017attention,
YEAR = {2017},
AUTHOR = {Vaswani, Ashish and Shazeer, Noam and Parmar, Niki and Uszkoreit, Jakob and Jones, Llion and Gomez, Aidan N and Kaiser, {Ł}ukasz and Polosukhin, Illia},
JOURNALTITLE = {Advances in neural information processing systems},
TITLE = {Attention is all you need},
VOLUME = {30},
}
@INPROCEEDINGS{deng2009imagenet,
YEAR = {2009},
AUTHOR = {Deng, Jia and Dong, Wei and Socher, Richard and Li, Li-Jia and Li, Kai and Fei-Fei, Li},
ORGANIZATION = {Ieee},
BOOKTITLE = {2009 IEEE conference on computer vision and pattern recognition},
PAGES = {248--255},
TITLE = {Imagenet: A large-scale hierarchical image database},
}
@ARTICLE{parti,
YEAR = {2022},
AUTHOR = {Yu, Jiahui and Xu, Yuanzhong and Koh, Jing and Luong, Thang and Baid, Gunjan and Vasudevan, Vijay and Ku, Alexander and Yang, Yinfei and Ayan, Burcu and Hutchinson, Ben and Han, Wei and Parekh, Zarana and Li, Xin and Zhang, Han and Baldridge, Jason and Wu, Yonghui},
DOI = {10.48550/arXiv.2206.10789},
TITLE = {Scaling Autoregressive Models for Content-Rich Text-to-Image Generation},
}
@INPROCEEDINGS{lewis-etal-2020-bart,
YEAR = {2020},
AUTHOR = {Lewis, Mike and Liu, Yinhan and Goyal, Naman and Ghazvininejad, Marjan and Mohamed, Abdelrahman and Levy, Omer and Stoyanov, Veselin and Zettlemoyer, Luke},
LOCATION = {Online},
PUBLISHER = {Association for Computational Linguistics},
URL = {https://aclanthology.org/2020.acl-main.703},
BOOKTITLE = {Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics},
DOI = {10.18653/v1/2020.acl-main.703},
PAGES = {7871--7880},
TITLE = {{BART}: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension},
}
@ARTICLE{atari,
YEAR = {2013},
AUTHOR = {Bellemare, Marc G. and Naddaf, Yavar and Veness, Joel and Bowling, Michael},
LOCATION = {El Segundo, CA, USA},
PUBLISHER = {AI Access Foundation},
ISSN = {1076-9757},
JOURNALTITLE = {J. Artif. Int. Res.},
NUMBER = {1},
PAGES = {253--279},
TITLE = {The Arcade Learning Environment: An Evaluation Platform for General Agents},
VOLUME = {47},
}
@ONLINE{darkMatter,
YEAR = {2021},
AUTHOR = {Yann, Lecun and Ishan, Misra},
URL = {https://ai.facebook.com/blog/self-supervised-learning-the-dark-matter-of-intelligence/},
TITLE = {Self-supervised learning: The dark matter of intelligence},
URLDATE = {2022-06-26},
}
@ONLINE{redditUsers,
YEAR = {2016},
AUTHOR = {MICHAEL BARTHEL, GALEN STOCKING, JESSE HOLCOMB and MITCHELL, AMY},
URL = {https://www.pewresearch.org/journalism/2016/02/25/reddit-news-users-more-likely-to-be-male-young-and-digital-in-their-news-preferences/},
TITLE = {Reddit news users more likely to be male, young and digital in their news preferences},
URLDATE = {2022-08-07},
}
@ONLINE{coco_eval,
YEAR = {2019},
AUTHOR = {Mircosoft},
URL = {https://cocodataset.org/#detection-eval},
TITLE = {Evaluate:Detection},
URLDATE = {2022-07-09},
}
@ONLINE{unsupBrain,
YEAR = {2021},
AUTHOR = {Mineault, Patrick},
URL = {https://xcorr.net/2021/12/31/2021-in-review-unsupervised-brain-models/},
TITLE = {Unsupervised models of the brain},
URLDATE = {2022-06-26},
}
@ARTICLE{zhuang2021unsupervised,
YEAR = {2021},
AUTHOR = {Zhuang, Chengxu and Yan, Siming and Nayebi, Aran and Schrimpf, Martin and Frank, Michael C and DiCarlo, James J and Yamins, Daniel LK},
PUBLISHER = {National Acad Sciences},
JOURNALTITLE = {Proceedings of the National Academy of Sciences},
NUMBER = {3},
PAGES = {e2014196118},
TITLE = {Unsupervised neural network models of the ventral visual stream},
VOLUME = {118},
}
@ARTICLE{liu2019roberta,
YEAR = {2019},
AUTHOR = {Liu, Yinhan and Ott, Myle and Goyal, Naman and Du, Jingfei and Joshi, Mandar and Chen, Danqi and Levy, Omer and Lewis, Mike and Zettlemoyer, Luke and Stoyanov, Veselin},
JOURNALTITLE = {arXiv preprint arXiv:1907.11692},
TITLE = {Roberta: A robustly optimized bert pretraining approach},
}
@ARTICLE{bromley1993signature,
YEAR = {1993},
AUTHOR = {Bromley, Jane and Guyon, Isabelle and LeCun, Yann and S{\"a}ckinger, Eduard and Shah, Roopak},
JOURNALTITLE = {Advances in neural information processing systems},
TITLE = {Signature verification using a" siamese" time delay neural network},
VOLUME = {6},
}
@INPROCEEDINGS{caron2021emerging,
YEAR = {2021},
AUTHOR = {Caron, Mathilde and Touvron, Hugo and Misra, Ishan and Jégou, Hervé and Mairal, Julien and Bojanowski, Piotr and Joulin, Armand},
BOOKTITLE = {Proceedings of the IEEE/CVF International Conference on Computer Vision},
PAGES = {9650--9660},
TITLE = {Emerging properties in self-supervised vision transformers},
}
@INPROCEEDINGS{mahajan2018exploring,
YEAR = {2018},
AUTHOR = {Mahajan, Dhruv and Girshick, Ross and Ramanathan, Vignesh and He, Kaiming and Paluri, Manohar and Li, Yixuan and Bharambe, Ashwin and Van Der Maaten, Laurens},
BOOKTITLE = {Proceedings of the European conference on computer vision (ECCV)},
PAGES = {181--196},
TITLE = {Exploring the limits of weakly supervised pretraining},
}
@ARTICLE{kolesnikov2019large,
YEAR = {2019},
AUTHOR = {Kolesnikov, Alexander and Beyer, Lucas and Zhai, Xiaohua and Puigcerver, Joan and Yung, Jessica and Gelly, Sylvain and Houlsby, Neil},
PUBLISHER = {arXiv},
JOURNALTITLE = {arXiv preprint arXiv:1912.11370},
NUMBER = {8},
TITLE = {Large scale learning of general visual representations for transfer},
VOLUME = {2},
}
@ARTICLE{rajpurkar2016squad,
YEAR = {2016},
AUTHOR = {Rajpurkar, Pranav and Zhang, Jian and Lopyrev, Konstantin and Liang, Percy},
JOURNALTITLE = {arXiv preprint arXiv:1606.05250},
TITLE = {Squad: 100,000+ questions for machine comprehension of text},
}
@ARTICLE{rajpurkar2018know,
YEAR = {2018},
AUTHOR = {Rajpurkar, Pranav and Jia, Robin and Liang, Percy},
JOURNALTITLE = {arXiv preprint arXiv:1806.03822},
TITLE = {Know what you don't know: Unanswerable questions for SQuAD},
}
@ARTICLE{srivastava2022beyond,
YEAR = {2022},
AUTHOR = {Srivastava, Aarohi and Rastogi, Abhinav and Rao, Abhishek and Shoeb, Abu Awal Md and Abid, Abubakar and Fisch, Adam and Brown, Adam R and Santoro, Adam and Gupta, Aditya and Garriga-Alonso, Adrià and others},
JOURNALTITLE = {arXiv preprint arXiv:2206.04615},
TITLE = {Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models},
}
@ARTICLE{bowman2021will,
YEAR = {2021},
AUTHOR = {Bowman, Samuel R and Dahl, George E},
JOURNALTITLE = {arXiv preprint arXiv:2104.02145},
TITLE = {What Will it Take to Fix Benchmarking in Natural Language Understanding?},
}
@ARTICLE{goodfellow2014explaining,
YEAR = {2014},
AUTHOR = {Goodfellow, Ian J and Shlens, Jonathon and Szegedy, Christian},
JOURNALTITLE = {arXiv preprint arXiv:1412.6572},
TITLE = {Explaining and harnessing adversarial examples},
}
@INPROCEEDINGS{recht2019imagenet,
YEAR = {2019},
AUTHOR = {Recht, Benjamin and Roelofs, Rebecca and Schmidt, Ludwig and Shankar, Vaishaal},
ORGANIZATION = {PMLR},
BOOKTITLE = {International Conference on Machine Learning},
PAGES = {5389--5400},
TITLE = {Do imagenet classifiers generalize to imagenet?},
}
@ARTICLE{beyer2020we,
YEAR = {2020},
AUTHOR = {Beyer, Lucas and Hénaff, Olivier J and Kolesnikov, Alexander and Zhai, Xiaohua and Oord, A{\"a}ron van den},
JOURNALTITLE = {arXiv preprint arXiv:2006.07159},
TITLE = {Are we done with imagenet?},
}
@ARTICLE{li2022mask,
AUTHOR = {Li, Feng and Zhang, Hao and Liu, Shilong and Zhang, Lei and Ni, Lionel M and Shum, Heung-Yeung and others},
YEAR = {2022},
JOURNALTITLE = {arXiv preprint arXiv:2206.02777},
TITLE = {Mask DINO: Towards A Unified Transformer-based Framework for Object Detection and Segmentation},
}
@INPROCEEDINGS{koehn2005europarl,
YEAR = {2005},
AUTHOR = {Koehn, Philipp},
BOOKTITLE = {Proceedings of machine translation summit x: papers},
PAGES = {79--86},
TITLE = {Europarl: A parallel corpus for statistical machine translation},
}
@MISC{Gokaslan2019OpenWeb,
YEAR = {2019},
AUTHOR = {Gokaslan, Aaron and Cohen, Vanya},
TITLE = {OpenWebText Corpus},
}
@ARTICLE{xue2020mt5,
YEAR = {2020},
AUTHOR = {Xue, Linting and Constant, Noah and Roberts, Adam and Kale, Mihir and Al-Rfou, Rami and Siddhant, Aditya and Barua, Aditya and Raffel, Colin},
JOURNALTITLE = {arXiv preprint arXiv:2010.11934},
TITLE = {mT5: A massively multilingual pre-trained text-to-text transformer},
}
@ARTICLE{wenzek2019ccnet,
YEAR = {2019},
AUTHOR = {Wenzek, Guillaume and Lachaux, Marie-Anne and Conneau, Alexis and Chaudhary, Vishrav and Guzmán, Francisco and Joulin, Armand and Grave, Edouard},
JOURNALTITLE = {arXiv preprint arXiv:1911.00359},
TITLE = {Ccnet: Extracting high quality monolingual datasets from web crawl data},
}
@ARTICLE{bandy2021addressing,
YEAR = {2021},
AUTHOR = {Bandy, Jack and Vincent, Nicholas},
JOURNALTITLE = {arXiv preprint arXiv:2105.05241},
TITLE = {Addressing" documentation debt" in machine learning research: A retrospective datasheet for bookcorpus},
}
@ARTICLE{gao2017knowledge,
YEAR = {2017},
AUTHOR = {Gao, Jiyang and Li, Zhen and Nevatia, Ram and others},
JOURNALTITLE = {arXiv preprint arXiv:1711.07607},
TITLE = {Knowledge concentration: Learning 100k object classifiers in a single CNN},
}
@INPROCEEDINGS{shao2019objects365,
YEAR = {2019},
AUTHOR = {Shao, Shuai and Li, Zeming and Zhang, Tianyuan and Peng, Chao and Yu, Gang and Zhang, Xiangyu and Li, Jing and Sun, Jian},
BOOKTITLE = {Proceedings of the IEEE/CVF international conference on computer vision},
PAGES = {8430--8439},
TITLE = {Objects365: A large-scale, high-quality dataset for object detection},
}
@ARTICLE{yuan2022wudaomm,
YEAR = {2022},
AUTHOR = {Yuan, Sha and Shuai, Zhao and Jiahong, Leng and Zhao, Xue and Hanyu, Zhao and Jie, Tang},
JOURNALTITLE = {arXiv preprint arXiv:2203.11480},
TITLE = {WuDaoMM: A large-scale Multi-Modal Dataset for Pre-training models},
}
@INPROCEEDINGS{srinivasan2021wit,
YEAR = {2021},
AUTHOR = {Srinivasan, Krishna and Raman, Karthik and Chen, Jiecao and Bendersky, Michael and Najork, Marc},
BOOKTITLE = {Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval},
PAGES = {2443--2449},
TITLE = {Wit: Wikipedia-based image text dataset for multimodal multilingual machine learning},
}
@ARTICLE{tiedemann2018emerging,
AUTHOR = {Tiedemann, J{\"o}rg},
YEAR = {2018},
JOURNALTITLE = {arXiv preprint arXiv:1802.00273},
TITLE = {Emerging language spaces learned from massively multilingual corpora},
}
@ARTICLE{mayer2014creating,
AUTHOR = {Mayer, Thomas and Cysouw, Michael},
YEAR = {2014},
JOURNALTITLE = {Oceania},
NUMBER = {273},
PAGES = {40},
TITLE = {Creating a massively parallel Bible corpus},
VOLUME = {135},
}
@INPROCEEDINGS{zellers2019recognition,
YEAR = {2019},
AUTHOR = {Zellers, Rowan and Bisk, Yonatan and Farhadi, Ali and Choi, Yejin},
BOOKTITLE = {Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
PAGES = {6720--6731},
TITLE = {From recognition to cognition: Visual commonsense reasoning},
}
@INPROCEEDINGS{antol2015vqa,
YEAR = {2015},
AUTHOR = {Antol, Stanislaw and Agrawal, Aishwarya and Lu, Jiasen and Mitchell, Margaret and Batra, Dhruv and Zitnick, C Lawrence and Parikh, Devi},
BOOKTITLE = {Proceedings of the IEEE international conference on computer vision},
PAGES = {2425--2433},
TITLE = {Vqa: Visual question answering},
}
@INPROCEEDINGS{zhang2016yin,
YEAR = {2016},
AUTHOR = {Zhang, Peng and Goyal, Yash and Summers-Stay, Douglas and Batra, Dhruv and Parikh, Devi},
BOOKTITLE = {Proceedings of the IEEE conference on computer vision and pattern recognition},
PAGES = {5014--5022},
TITLE = {Yin and yang: Balancing and answering binary visual questions},
}
@INPROCEEDINGS{goyal2017making,
YEAR = {2017},
AUTHOR = {Goyal, Yash and Khot, Tejas and Summers-Stay, Douglas and Batra, Dhruv and Parikh, Devi},
BOOKTITLE = {Proceedings of the IEEE conference on computer vision and pattern recognition},
PAGES = {6904--6913},
TITLE = {Making the v in vqa matter: Elevating the role of image understanding in visual question answering},
}
@INPROCEEDINGS{hudson2019gqa,
YEAR = {2019},
AUTHOR = {Hudson, Drew A and Manning, Christopher D},
BOOKTITLE = {Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
PAGES = {6700--6709},
TITLE = {Gqa: A new dataset for real-world visual reasoning and compositional question answering},
}
@ARTICLE{shekhar2017foil,
YEAR = {2017},
AUTHOR = {Shekhar, Ravi and Pezzelle, Sandro and Klimovich, Yauhen and Herbelot, Aurélie and Nabi, Moin and Sangineto, Enver and Bernardi, Raffaella},
JOURNALTITLE = {arXiv preprint arXiv:1705.01359},
TITLE = {Foil it! find one mismatch between image and language caption},
}
@ARTICLE{ribeiro2020beyond,
YEAR = {2020},
AUTHOR = {Ribeiro, Marco Tulio and Wu, Tongshuang and Guestrin, Carlos and Singh, Sameer},
JOURNALTITLE = {arXiv preprint arXiv:2005.04118},
TITLE = {Beyond accuracy: Behavioral testing of NLP models with CheckList},
}
@INPROCEEDINGS{parcalabescu-etal-2022-valse,
YEAR = {2022},
AUTHOR = {Parcalabescu, Letitia and Cafagna, Michele and Muradjan, Lilitta and Frank, Anette and Calixto, Iacer and Gatt, Albert},
PUBLISHER = {Association for Computational Linguistics},
URL = {https://aclanthology.org/2022.acl-long.567},
BOOKTITLE = {Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
PAGES = {8253--8280},
TITLE = {{VALSE}: A Task-Independent Benchmark for Vision and Language Models Centered on Linguistic Phenomena},
}
@ARTICLE{sheng2019woman,
YEAR = {2019},
AUTHOR = {Sheng, Emily and Chang, Kai-Wei and Natarajan, Premkumar and Peng, Nanyun},
JOURNALTITLE = {arXiv preprint arXiv:1909.01326},
TITLE = {The woman worked as a babysitter: On biases in language generation},
}
@INPROCEEDINGS{dhamala2021bold,
YEAR = {2021},
AUTHOR = {Dhamala, Jwala and Sun, Tony and Kumar, Varun and Krishna, Satyapriya and Pruksachatkun, Yada and Chang, Kai-Wei and Gupta, Rahul},
BOOKTITLE = {Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency},
PAGES = {862--872},
TITLE = {Bold: Dataset and metrics for measuring biases in open-ended language generation},
}
@ARTICLE{prabhu2020large,
YEAR = {2020},
AUTHOR = {Prabhu, Vinay Uday and Birhane, Abeba},
JOURNALTITLE = {arXiv preprint arXiv:2006.16923},
TITLE = {Large image datasets: A pyrrhic win for computer vision?},
}
@ARTICLE{birhane2021multimodal,
YEAR = {2021},
AUTHOR = {Birhane, Abeba and Prabhu, Vinay Uday and Kahembwe, Emmanuel},
JOURNALTITLE = {arXiv preprint arXiv:2110.01963},
TITLE = {Multimodal datasets: misogyny, pornography, and malignant stereotypes},
}
@ARTICLE{strubell2019energy,
YEAR = {2019},
AUTHOR = {Strubell, Emma and Ganesh, Ananya and McCallum, Andrew},
JOURNALTITLE = {arXiv preprint arXiv:1906.02243},
TITLE = {Energy and policy considerations for deep learning in NLP},
}
@ARTICLE{lottick2019energy,
YEAR = {2019},
AUTHOR = {Lottick, Kadan and Susai, Silvia and Friedler, Sorelle A and Wilson, Jonathan P},
JOURNALTITLE = {arXiv preprint arXiv:1911.08354},
TITLE = {Energy Usage Reports: Environmental awareness as part of algorithmic accountability},
}
@ARTICLE{henderson2020towards,
YEAR = {2020},
AUTHOR = {Henderson, Peter and Hu, Jieru and Romoff, Joshua and Brunskill, Emma and Jurafsky, Dan and Pineau, Joelle},
JOURNALTITLE = {Journal of Machine Learning Research},
NUMBER = {248},
PAGES = {1--43},
TITLE = {Towards the systematic reporting of the energy and carbon footprints of machine learning},
VOLUME = {21},
}
@INPROCEEDINGS{guo2016ms,
YEAR = {2016},
AUTHOR = {Guo, Yandong and Zhang, Lei and Hu, Yuxiao and He, Xiaodong and Gao, Jianfeng},
ORGANIZATION = {Springer},
BOOKTITLE = {European conference on computer vision},
PAGES = {87--102},
TITLE = {Ms-celeb-1m: A dataset and benchmark for large-scale face recognition},
}
@INPROCEEDINGS{sun,
YEAR = {2010},
AUTHOR = {Xiao, Jianxiong and Hays, James and Ehinger, Krista A. and Oliva, Aude and Torralba, Antonio},
BOOKTITLE = {2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition},
DOI = {10.1109/CVPR.2010.5539970},
PAGES = {3485--3492},
TITLE = {SUN database: Large-scale scene recognition from abbey to zoo},
}
@ARTICLE{pascalvoc,
YEAR = {2010},
AUTHOR = {Everingham, Mark and {van Gool}, Luc and Williams, {Christopher K. I.} and Winn, John and Zisserman, Andrew},
LANGUAGE = {English},
PUBLISHER = {Springer Netherlands},
DOI = {10.1007/s11263-009-0275-4},
ISSN = {0920-5691},
JOURNALTITLE = {International Journal of Computer Vision},
KEYWORDS = {Benchmark,Database,Object detection,Object recognition},
NUMBER = {2},
PAGES = {303--338},
TITLE = {The PASCAL Visual Object Classes (VOC) Challenge},
VOLUME = {88},
}
@ARTICLE{WordNet,
YEAR = {2000},
AUTHOR = {Fellbaum, Christiane D.},
JOURNALTITLE = {Language},
PAGES = {706},
TITLE = {WordNet : an electronic lexical database},
VOLUME = {76},
}
@INPROCEEDINGS{Socher10connectingmodalities,
YEAR = {2010},
AUTHOR = {Socher, Richard and Fei-fei, Li},
BOOKTITLE = {In IEEE Computer Society Conference on Computer Vision and Pattern Recognition},
TITLE = {Connecting modalities: Semi-supervised segmentation and annotation of images using unaligned text corpora},
}
@ARTICLE{5487377,
YEAR = {2010},
AUTHOR = {Yao, Benjamin Z. and Yang, Xiong and Lin, Liang and Lee, Mun Wai and Zhu, Song-Chun},
DOI = {10.1109/JPROC.2010.2050411},
JOURNALTITLE = {Proceedings of the IEEE},
NUMBER = {8},
PAGES = {1485--1508},
TITLE = {I2T: Image Parsing to Text Description},
VOLUME = {98},
}
@INPROCEEDINGS{vinyals,
YEAR = {2015},
AUTHOR = {Vinyals, Oriol and Toshev, Alexander and Bengio, Samy and Erhan, Dumitru},
DOI = {10.1109/CVPR.2015.7298935},
PAGES = {3156--3164},
TITLE = {Show and tell: A neural image caption generator},
}
@MISC{karpthy1,
YEAR = {2014},
AUTHOR = {Karpathy, Andrej and Fei-Fei, Li},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/1412.2306},
DOI = {10.48550/ARXIV.1412.2306},
KEYWORDS = {Computer Vision and Pattern Recognition (cs.CV),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Deep Visual-Semantic Alignments for Generating Image Descriptions},
}
@MISC{xu1,
YEAR = {2015},
AUTHOR = {Xu, Kelvin and Ba, Jimmy and Kiros, Ryan and Cho, Kyunghyun and Courville, Aaron and Salakhutdinov, Ruslan and Zemel, Richard and Bengio, Yoshua},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/1502.03044},
DOI = {10.48550/ARXIV.1502.03044},
KEYWORDS = {Machine Learning (cs.LG),Computer Vision and Pattern Recognition (cs.CV),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Show, Attend and Tell: Neural Image Caption Generation with Visual Attention},
}
@MISC{yao1,
YEAR = {2018},
AUTHOR = {Yao, Ting and Pan, Yingwei and Li, Yehao and Mei, Tao},
PUBLISHER = {arXiv},
URL = {https://arxiv.org/abs/1809.07041},
DOI = {10.48550/ARXIV.1809.07041},
KEYWORDS = {Computer Vision and Pattern Recognition (cs.CV),FOS: Computer and information sciences,FOS: Computer and information sciences},
TITLE = {Exploring Visual Relationship for Image Captioning},
}
@INPROCEEDINGS{devlin-etal-2019-bert,
YEAR = {2019},
AUTHOR = {Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
LOCATION = {Minneapolis, Minnesota},
PUBLISHER = {Association for Computational Linguistics},
URL = {https://aclanthology.org/N19-1423},
DOI = {10.18653/v1/N19-1423},
PAGES = {4171--4186},
TITLE = {{BERT}: Pre-training of Deep Bidirectional Transformers for Language Understanding},
}
@INPROCEEDINGS{HerdadeKBS19,
YEAR = {2019},
AUTHOR = {Herdade, Simao and Kappeler, Armin and Boakye, Kofi and Soares, Joao},
URL = {http://papers.nips.cc/paper/9293-image-captioning-transforming-objects-into-words},
PAGES = {11135--11145},
TITLE = {Image Captioning: Transforming Objects into Words},
}
@INPROCEEDINGS{huang1,
YEAR = {2019},
AUTHOR = {Huang, Lun and Wang, Wenmin and Chen, Jie and Wei, Xiao-Yong},
DOI = {10.1109/ICCV.2019.00473},
PAGES = {4633--4642},
TITLE = {Attention on Attention for Image Captioning},
}
@INPROCEEDINGS{NIPS2017_3f5ee243,
YEAR = {2017},
AUTHOR = {Vaswani, Ashish and Shazeer, Noam and Parmar, Niki and Uszkoreit, Jakob and Jones, Llion and Gomez, Aidan N and Kaiser, Łukasz and Polosukhin, Illia},
EDITOR = {Guyon, I. and Luxburg, U. Von and Bengio, S. and Wallach, H. and Fergus, R. and Vishwanathan, S. and Garnett, R.},
PUBLISHER = {Curran Associates, Inc.},
URL = {https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf},
BOOKTITLE = {Advances in Neural Information Processing Systems},
TITLE = {Attention is All you Need},
VOLUME = {30},
}
@INPROCEEDINGS{spice,
YEAR = {2016},
AUTHOR = {Anderson, Peter and Fernando, Basura and Johnson, Mark and Gould, Stephen},
EDITOR = {Leibe, Bastian and Matas, Jiri and Sebe, Nicu and Welling, Max},
LOCATION = {Cham},
PUBLISHER = {Springer International Publishing},
BOOKTITLE = {Computer Vision -- ECCV 2016},
ISBN = {978-3-319-46454-1},
PAGES = {382--398},
TITLE = {SPICE: Semantic Propositional Image Caption Evaluation},
}
@INPROCEEDINGS{meteor,
YEAR = {2005},
AUTHOR = {Banerjee, Satanjeev and Lavie, Alon},
LOCATION = {Ann Arbor, Michigan},
PUBLISHER = {Association for Computational Linguistics},
URL = {https://aclanthology.org/W05-0909},
BOOKTITLE = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},
PAGES = {65--72},
TITLE = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},
}
@INPROCEEDINGS{lin-2004-rouge,
YEAR = {2004},
AUTHOR = {Lin, Chin-Yew},
LOCATION = {Barcelona, Spain},
PUBLISHER = {Association for Computational Linguistics},
URL = {https://aclanthology.org/W04-1013},
BOOKTITLE = {Text Summarization Branches Out},
PAGES = {74--81},
TITLE = {{ROUGE}: A Package for Automatic Evaluation of Summaries},