-
Notifications
You must be signed in to change notification settings - Fork 4
/
reference.bib
2788 lines (2463 loc) · 122 KB
/
reference.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@article{Hinton06,
author = {Hinton, Geoffrey E. and Osindero, Simon and Teh, Yee Whye},
journal = {Neural Computation},
pages = {1527--1554},
title = {A Fast Learning Algorithm for Deep Belief Nets},
volume = {18},
year = {2006}
}
@book{goodfellow2016deep,
title = {Deep learning},
author = {Goodfellow, Ian and Bengio, Yoshua and Courville, Aaron and Bengio, Yoshua},
volume = {1},
year = {2016},
publisher = {MIT Press}
}
@article{wang2023far,
title = {How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources},
author = {Wang, Yizhong and Ivison, Hamish and Dasigi, Pradeep and Hessel, Jack and Khot, Tushar and Chandu, Khyathi Raghavi and Wadden, David and MacMillan, Kelsey and Smith, Noah A and Beltagy, Iz and others},
journal = {arXiv preprint arXiv:2306.04751},
year = {2023}
}
@article{kuznetsova2020open,
title = {The open images dataset v4: Unified image classification, object detection, and visual relationship detection at scale},
author = {Kuznetsova, Alina and Rom, Hassan and Alldrin, Neil and Uijlings, Jasper and Krasin, Ivan and Pont-Tuset, Jordi and Kamali, Shahab and Popov, Stefan and Malloci, Matteo and Kolesnikov, Alexander and others},
journal = {International Journal of Computer Vision},
volume = {128},
number = {7},
pages = {1956--1981},
year = {2020},
publisher = {Springer}
}
@inproceedings{biten2022let,
title = {Let there be a clock on the beach: Reducing object hallucination in image captioning},
author = {Biten, Ali Furkan and G{\'o}mez, Llu{\'\i}s and Karatzas, Dimosthenis},
booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision},
pages = {1381--1390},
year = {2022}
}
@inproceedings{macleod2017understanding,
title = {Understanding blind people's experiences with computer-generated captions of social media images},
author = {MacLeod, Haley and Bennett, Cynthia L and Morris, Meredith Ringel and Cutrell, Edward},
booktitle = {proceedings of the 2017 CHI conference on human factors in computing systems},
pages = {5988--5999},
year = {2017}
}
@article{shi2023replug,
title = {Replug: Retrieval-augmented black-box language models},
author = {Shi, Weijia and Min, Sewon and Yasunaga, Michihiro and Seo, Minjoon and James, Rich and Lewis, Mike and Zettlemoyer, Luke and Yih, Wen-tau},
journal = {arXiv preprint arXiv:2301.12652},
year = {2023}
}
@article{mckenna2023sources,
title = {Sources of Hallucination by Large Language Models on Inference Tasks},
author = {McKenna, Nick and Li, Tianyi and Cheng, Liang and Hosseini, Mohammad Javad and Johnson, Mark and Steedman, Mark},
journal = {arXiv preprint arXiv:2305.14552},
year = {2023}
}
@article{glaese2022improving,
title = {Improving alignment of dialogue agents via targeted human judgements},
author = {Glaese, Amelia and McAleese, Nat and Tr{\k{e}}bacz, Maja and Aslanides, John and Firoiu, Vlad and Ewalds, Timo and Rauh, Maribeth and Weidinger, Laura and Chadwick, Martin and Thacker, Phoebe and others},
journal = {arXiv preprint arXiv:2209.14375},
year = {2022}
}
@article{ouyang2022training,
title = {Training language models to follow instructions with human feedback},
author = {Ouyang, Long and Wu, Jeffrey and Jiang, Xu and Almeida, Diogo and Wainwright, Carroll and Mishkin, Pamela and Zhang, Chong and Agarwal, Sandhini and Slama, Katarina and Ray, Alex and others},
journal = {Advances in Neural Information Processing Systems},
volume = {35},
pages = {27730--27744},
year = {2022}
}
@article{bai2022training,
title = {Training a helpful and harmless assistant with reinforcement learning from human feedback},
author = {Bai, Yuntao and Jones, Andy and Ndousse, Kamal and Askell, Amanda and Chen, Anna and DasSarma, Nova and Drain, Dawn and Fort, Stanislav and Ganguli, Deep and Henighan, Tom and others},
journal = {arXiv preprint arXiv:2204.05862},
year = {2022}
}
@article{OpenAI2023GPT4TR,
title = {GPT-4 Technical Report},
author = {OpenAI},
journal = {ArXiv},
year = {2023},
volume = {abs/2303.08774}
}
@article{zhu2023minigpt,
title = {MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models},
author = {Zhu, Deyao and Chen, Jun and Shen, Xiaoqian and Li, Xiang and Elhoseiny, Mohamed},
journal = {arXiv preprint arXiv:2304.10592},
year = {2023}
}
@article{li2023otter,
title = {Otter: A Multi-Modal Model with In-Context Instruction Tuning},
author = {Li, Bo and Zhang, Yuanhan and Chen, Liangyu and Wang, Jinghao and Yang, Jingkang and Liu, Ziwei},
journal = {arXiv preprint arXiv:2305.03726},
year = {2023}
}
@article{gao2023llama,
title = {Llama-adapter v2: Parameter-efficient visual instruction model},
author = {Gao, Peng and Han, Jiaming and Zhang, Renrui and Lin, Ziyi and Geng, Shijie and Zhou, Aojun and Zhang, Wei and Lu, Pan and He, Conghui and Yue, Xiangyu and others},
journal = {arXiv preprint arXiv:2304.15010},
year = {2023}
}
Ask AI to edit or generate...
@misc{zhang2023llamaadapter,
title = {LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention},
author = {Renrui Zhang and Jiaming Han and Chris Liu and Peng Gao and Aojun Zhou and Xiangfei Hu and Shilin Yan and Pan Lu and Hongsheng Li and Yu Qiao},
year = {2023},
eprint = {2303.16199},
archiveprefix = {arXiv},
primaryclass = {cs.CV}
}
@article{brown2020language,
title = {Language models are few-shot learners},
author = {Brown, Tom and Mann, Benjamin and Ryder, Nick and Subbiah, Melanie and Kaplan, Jared D and Dhariwal, Prafulla and Neelakantan, Arvind and Shyam, Pranav and Sastry, Girish and Askell, Amanda and others},
journal = {Advances in neural information processing systems},
volume = {33},
pages = {1877--1901},
year = {2020}
}
@article{chiang2023vicuna,
title = {Vicuna: An open-source chatbot impressing gpt-4 with 90\%* chatgpt quality},
author = {Chiang, Wei-Lin and Li, Zhuohan and Lin, Zi and Sheng, Ying and Wu, Zhanghao and Zhang, Hao and Zheng, Lianmin and Zhuang, Siyuan and Zhuang, Yonghao and Gonzalez, Joseph E and others},
journal = {See https://vicuna. lmsys. org (accessed 14 April 2023)},
year = {2023}
}
@article{dai2023instructblip,
title = {Instructblip: Towards general-purpose vision-language models with instruction tuning},
author = {Dai, Wenliang and Li, Junnan and Li, Dongxu and Tiong, Anthony Meng Huat and Zhao, Junqi and Wang, Weisheng and Li, Boyang and Fung, Pascale and Hoi, Steven},
journal = {arXiv preprint arXiv:2305.06500},
year = {2023}
}
@inproceedings{du2022glm,
title = {GLM: General Language Model Pretraining with Autoregressive Blank Infilling},
author = {Du, Zhengxiao and Qian, Yujie and Liu, Xiao and Ding, Ming and Qiu, Jiezhong and Yang, Zhilin and Tang, Jie},
booktitle = {Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
pages = {320--335},
year = {2022}
}
@article{longpre2023flan,
title = {The flan collection: Designing data and methods for effective instruction tuning},
author = {Longpre, Shayne and Hou, Le and Vu, Tu and Webson, Albert and Chung, Hyung Won and Tay, Yi and Zhou, Denny and Le, Quoc V and Zoph, Barret and Wei, Jason and others},
journal = {arXiv preprint arXiv:2301.13688},
year = {2023}
}
@article{liu2023mmbench,
title = {MMBench: Is Your Multi-modal Model an All-around Player?},
author = {Liu, Yuan and Duan, Haodong and Zhang, Yuanhan and Li, Bo and Zhang, Songyang and Zhao, Wangbo and Yuan, Yike and Wang, Jiaqi and He, Conghui and Liu, Ziwei and others},
journal = {arXiv preprint arXiv:2307.06281},
year = {2023}
}
@article{hu2021lora,
title = {Lora: Low-rank adaptation of large language models},
author = {Hu, Edward J and Shen, Yelong and Wallis, Phillip and Allen-Zhu, Zeyuan and Li, Yuanzhi and Wang, Shean and Wang, Lu and Chen, Weizhu},
journal = {arXiv preprint arXiv:2106.09685},
year = {2021}
}
@article{ye2023mplug,
title = {mplug-owl: Modularization empowers large language models with multimodality},
author = {Ye, Qinghao and Xu, Haiyang and Xu, Guohai and Ye, Jiabo and Yan, Ming and Zhou, Yiyang and Wang, Junyang and Hu, Anwen and Shi, Pengcheng and Shi, Yaya and others},
journal = {arXiv preprint arXiv:2304.14178},
year = {2023}
}
@article{li2023blip,
title = {Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models},
author = {Li, Junnan and Li, Dongxu and Savarese, Silvio and Hoi, Steven},
journal = {arXiv preprint arXiv:2301.12597},
year = {2023}
}
@article{radford2019language,
title = {Language models are unsupervised multitask learners},
author = {Radford, Alec and Wu, Jeffrey and Child, Rewon and Luan, David and Amodei, Dario and Sutskever, Ilya and others},
journal = {OpenAI blog},
volume = {1},
number = {8},
pages = {9},
year = {2019}
}
@article{li2023mimic,
title = {MIMIC-IT: Multi-Modal In-Context Instruction Tuning},
author = {Bo Li and Yuanhan Zhang and Liangyu Chen and Jinghao Wang and Fanyi Pu and Jingkang Yang and Chunyuan Li and Ziwei Liu},
year = {2023},
eprint = {2306.05425},
archiveprefix = {arXiv},
primaryclass = {cs.CV}
}
@inproceedings{vqav2,
title = {Making the v in vqa matter: Elevating the role of image understanding in visual question answering},
author = {Goyal, Yash and Khot, Tejas and Summers-Stay, Douglas and Batra, Dhruv and Parikh, Devi},
booktitle = {Proceedings of the IEEE conference on computer vision and pattern recognition},
pages = {6904--6913},
year = {2017}
}
@article{coco_caption,
title = {Microsoft coco captions: Data collection and evaluation server},
author = {Chen, Xinlei and Fang, Hao and Lin, Tsung-Yi and Vedantam, Ramakrishna and Gupta, Saurabh and Doll{\'a}r, Piotr and Zitnick, C Lawrence},
journal = {arXiv preprint arXiv:1504.00325},
year = {2015}
}
@inproceedings{hudson2019gqa,
title = {Gqa: A new dataset for real-world visual reasoning and compositional question answering},
author = {Hudson, Drew A and Manning, Christopher D},
booktitle = {Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
pages = {6700--6709},
year = {2019}
}
@inproceedings{ok-vqa,
title = {Ok-vqa: A visual question answering benchmark requiring external knowledge},
author = {Marino, Kenneth and Rastegari, Mohammad and Farhadi, Ali and Mottaghi, Roozbeh},
booktitle = {Proceedings of the IEEE/cvf conference on computer vision and pattern recognition},
pages = {3195--3204},
year = {2019}
}
@inproceedings{Xu2023LVLMeHubAC,
title = {LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models},
author = {Peng Xu and Wenqi Shao and Kaipeng Zhang and Peng Gao and Shuo Liu and Meng Lei and Fanqing Meng and Siyuan Huang and Yu Jiao Qiao and Ping Luo},
year = {2023}
}
@misc{zheng2023judging,
title = {Judging LLM-as-a-judge with MT-Bench and Chatbot Arena},
author = {Lianmin Zheng and Wei-Lin Chiang and Ying Sheng and Siyuan Zhuang and Zhanghao Wu and Yonghao Zhuang and Zi Lin and Zhuohan Li and Dacheng Li and Eric. P Xing and Hao Zhang and Joseph E. Gonzalez and Ion Stoica},
year = {2023},
eprint = {2306.05685},
archiveprefix = {arXiv},
primaryclass = {cs.CL}
}
@book{oaksford2007bayesian,
title = {Bayesian rationality: The probabilistic approach to human reasoning},
author = {Oaksford, Mike and Chater, Nick},
year = {2007},
publisher = {Oxford University Press}
}
@book{fodor1983modularity,
title = {The modularity of mind},
author = {Fodor, Jerry A},
year = {1983},
publisher = {MIT press}
}
@inproceedings{nocaps,
title = {Nocaps: Novel object captioning at scale},
author = {Agrawal, Harsh and Desai, Karan and Wang, Yufei and Chen, Xinlei and Jain, Rishabh and Johnson, Mark and Batra, Dhruv and Parikh, Devi and Lee, Stefan and Anderson, Peter},
booktitle = {Proceedings of the IEEE/CVF international conference on computer vision},
pages = {8948--8957},
year = {2019}
}
@article{flickr30k,
title = {From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions},
author = {Young, Peter and Lai, Alice and Hodosh, Micah and Hockenmaier, Julia},
journal = {Transactions of the Association for Computational Linguistics},
volume = {2},
pages = {67--78},
year = {2014},
publisher = {MIT Press}
}
@inproceedings{vizwiz,
title = {Vizwiz grand challenge: Answering visual questions from blind people},
author = {Gurari, Danna and Li, Qing and Stangl, Abigale J and Guo, Anhong and Lin, Chi and Grauman, Kristen and Luo, Jiebo and Bigham, Jeffrey P},
booktitle = {Proceedings of the IEEE conference on computer vision and pattern recognition},
pages = {3608--3617},
year = {2018}
}
@inproceedings{textvqa,
title = {Towards vqa models that can read},
author = {Singh, Amanpreet and Natarajan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Batra, Dhruv and Parikh, Devi and Rohrbach, Marcus},
booktitle = {Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
pages = {8317--8326},
year = {2019}
}
@article{scienceqa,
title = {Learn to explain: Multimodal reasoning via thought chains for science question answering},
author = {Lu, Pan and Mishra, Swaroop and Xia, Tanglin and Qiu, Liang and Chang, Kai-Wei and Zhu, Song-Chun and Tafjord, Oyvind and Clark, Peter and Kalyan, Ashwin},
journal = {Advances in Neural Information Processing Systems},
volume = {35},
pages = {2507--2521},
year = {2022}
}
@inproceedings{youcook2,
title = {Towards automatic learning of procedures from web instructional videos},
author = {Zhou, Luowei and Xu, Chenliang and Corso, Jason},
booktitle = {Proceedings of the AAAI Conference on Artificial Intelligence},
volume = {32},
number = {1},
year = {2018}
}
@article{laurenccon2023obelisc,
title = {OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents},
author = {Lauren{\c{c}}on, Hugo and Saulnier, Lucile and Tronchon, L{\'e}o and Bekman, Stas and Singh, Amanpreet and Lozhkov, Anton and Wang, Thomas and Karamcheti, Siddharth and Rush, Alexander M and Kiela, Douwe and others},
journal = {arXiv preprint arXiv:2306.16527},
year = {2023}
}
@inproceedings{singh2019towards,
title = {Towards VQA Models That Can Read},
author = {Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Parikh, Devi and Rohrbach, Marcus},
booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages = {8317-8326},
year = {2019}
}
@article{zhou2017places,
title = {Places: A 10 million Image Database for Scene Recognition},
author = {Zhou, Bolei and Lapedriza, Agata and Khosla, Aditya and Oliva, Aude and Torralba, Antonio},
journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
year = {2017},
publisher = {IEEE}
}
@misc{2023internlm,
title = {InternLM: A Multilingual Language Model with Progressively Enhanced Capabilities},
author = {InternLM Team},
howpublished = {\url{https://github.com/InternLM/InternLM-techreport}},
year = {2023}
}
@inproceedings{W3Cschool,
title = {W3C school},
booktitle = {https://www.w3schools.com/},
year = {2023}
}
@misc{ARAS,
doi = {10.48550/ARXIV.2209.09393},
url = {https://arxiv.org/abs/2209.09393},
author = {Duan, Haodong and Zhao, Yue and Chen, Kai and Xiong, Yuanjun and Lin, Dahua},
keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Mitigating Representation Bias in Action Recognition: Algorithms and Benchmarks},
publisher = {arXiv},
year = {2022},
copyright = {arXiv.org perpetual, non-exclusive license}
}
@inproceedings{johnson2017clevr,
title = {Clevr: A diagnostic dataset for compositional language and elementary visual reasoning},
author = {Johnson, Justin and Hariharan, Bharath and Van Der Maaten, Laurens and Fei-Fei, Li and Lawrence Zitnick, C and Girshick, Ross},
booktitle = {Proceedings of the IEEE conference on computer vision and pattern recognition},
pages = {2901--2910},
year = {2017}
}
@inproceedings{li2017dual,
title = {Dual-glance model for deciphering social relationships},
author = {Li, Junnan and Wong, Yongkang and Zhao, Qi and Kankanhalli, Mohan S},
booktitle = {Proceedings of the IEEE international conference on computer vision},
pages = {2650--2659},
year = {2017}
}
@article{koniq10k,
author = {V. {Hosu} and H. {Lin} and T. {Sziranyi} and D.
{Saupe}},
journal = {IEEE Transactions on Image Processing},
title = {KonIQ-10k: An Ecologically Valid Database for Deep
Learning of Blind Image Quality Assessment},
year = {2020},
volume = {29},
pages = {4041-4056}
}
@article{Liu2022VisualSR,
title = {Visual Spatial Reasoning},
author = {Fangyu Liu and Guy Edward Toh Emerson and Nigel Collier},
journal = {Transactions of the Association for Computational Linguistics},
year = {2023}
}
@article{huang2023c,
title = {C-eval: A multi-level multi-discipline chinese evaluation suite for foundation models},
author = {Huang, Yuzhen and Bai, Yuzhuo and Zhu, Zhihao and Zhang, Junlei and Zhang, Jinghan and Su, Tangjun and Liu, Junteng and Lv, Chuancheng and Zhang, Yikai and Lei, Jiayi and others},
journal = {arXiv preprint arXiv:2305.08322},
year = {2023}
}
@article{hendrycks2020measuring,
title = {Measuring massive multitask language understanding},
author = {Hendrycks, Dan and Burns, Collin and Basart, Steven and Zou, Andy and Mazeika, Mantas and Song, Dawn and Steinhardt, Jacob},
journal = {arXiv preprint arXiv:2009.03300},
year = {2020}
}
@misc{gong2023multimodalgpt,
title = {MultiModal-GPT: A Vision and Language Model for Dialogue with Humans},
author = {Tao Gong and Chengqi Lyu and Shilong Zhang and Yudong Wang and Miao Zheng and Qian Zhao and Kuikun Liu and Wenwei Zhang and Ping Luo and Kai Chen},
year = {2023},
eprint = {2305.04790},
archiveprefix = {arXiv},
primaryclass = {cs.CV}
}
@article{chen2023shikra,
title = {Shikra: Unleashing Multimodal LLM's Referential Dialogue Magic},
author = {Chen, Keqin and Zhang, Zhao and Zeng, Weili and Zhang, Richong and Zhu, Feng and Zhao, Rui},
journal = {arXiv preprint arXiv:2306.15195},
year = {2023}
}
@article{peng2023kosmos,
title = {Kosmos-2: Grounding Multimodal Large Language Models to the World},
author = {Peng, Zhiliang and Wang, Wenhui and Dong, Li and Hao, Yaru and Huang, Shaohan and Ma, Shuming and Wei, Furu},
journal = {arXiv preprint arXiv:2306.14824},
year = {2023}
}
@article{su2023pandagpt,
title = {PandaGPT: One Model To Instruction-Follow Them All},
author = {Su, Yixuan and Lan, Tian and Li, Huayang and Xu, Jialu and Wang, Yan and Cai, Deng},
journal = {arXiv preprint arXiv:2305.16355},
year = {2023}
}
@article{Fu2023MMEAC,
title = {MME: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models},
author = {Chaoyou Fu and Peixian Chen and Yunhang Shen and Yulei Qin and Mengdan Zhang and Xu Lin and Zhenyu Qiu and Wei Lin and Jinrui Yang and Xiawu Zheng and Ke Li and Xing Sun and Rongrong Ji},
journal = {ArXiv},
year = {2023},
volume = {abs/2306.13394},
url = {https://api.semanticscholar.org/CorpusID:259243928}
}
@article{schuhmann2021laion,
title = {Laion-400m: Open dataset of clip-filtered 400 million image-text pairs},
author = {Schuhmann, Christoph and Vencu, Richard and Beaumont, Romain and Kaczmarczyk, Robert and Mullis, Clayton and Katta, Aarush and Coombes, Theo and Jitsev, Jenia and Komatsuzaki, Aran},
journal = {arXiv preprint arXiv:2111.02114},
year = {2021}
}
@incollection{Bengio+chapter2007,
author = {Bengio, Yoshua and LeCun, Yann},
booktitle = {Large Scale Kernel Machines},
publisher = {MIT Press},
title = {Scaling Learning Algorithms Towards {AI}},
year = {2007}
}
@article{sun2023principle,
title = {Principle-driven self-alignment of language models from scratch with minimal human supervision},
author = {Sun, Zhiqing and Shen, Yikang and Zhou, Qinhong and Zhang, Hongxin and Chen, Zhenfang and Cox, David and Yang, Yiming and Gan, Chuang},
journal = {arXiv preprint arXiv:2305.03047},
year = {2023}
}
@misc{alpaca,
author = {Rohan Taori and Ishaan Gulrajani and Tianyi Zhang and Yann Dubois and Xuechen Li and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto },
title = {Stanford Alpaca: An Instruction-following LLaMA model},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/tatsu-lab/stanford_alpaca}}
}
@misc{textdavinci,
author = {OpenAI},
title = {Model index for researchers},
year = 2023,
howpublished = {\url{https://platform.openai.com/docs/model-index-for-researchers}}
}
@article{devlin2018bert,
title = {{BERT}: {P}re-training of deep bidirectional transformers for language understanding},
author = {Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
journal = {arXiv preprint arXiv:1810.04805},
year = {2018}
}
@article{liu2019roberta,
title = {{Ro{BERT}a: A robustly optimized {BERT} pretraining approach}},
author = {Liu, Yinhan and Ott, Myle and Goyal, Naman and Du, Jingfei and Joshi, Mandar and Chen, Danqi and Levy, Omer and Lewis, Mike and Zettlemoyer, Luke and Stoyanov, Veselin},
journal = {arXiv preprint arXiv:1907.11692},
year = {2019}
}
@article{wei2022chain,
title = {Chain-of-thought prompting elicits reasoning in large language models},
author = {Wei, Jason and Wang, Xuezhi and Schuurmans, Dale and Bosma, Maarten and Chi, Ed and Le, Quoc and Zhou, Denny},
journal = {NeurIPS},
year = {2022}
}
@article{raffel2019exploring,
title = {Exploring the limits of transfer learning with a unified text-to-text transformer},
author = {Raffel, Colin and Shazeer, Noam and Roberts, Adam and Lee, Katherine and Narang, Sharan and Matena, Michael and Zhou, Yanqi and Li, Wei and Liu, Peter J},
journal = {arXiv preprint arXiv:1910.10683},
year = {2019}
}
@article{wei2022emergent,
title = {Emergent abilities of large language models},
author = {Wei, Jason and Tay, Yi and Bommasani, Rishi and Raffel, Colin and Zoph, Barret and Borgeaud, Sebastian and Yogatama, Dani and Bosma, Maarten and Zhou, Denny and Metzler, Donald and others},
journal = {TMLR},
year = {2022}
}
@article{srivastava2022beyond,
title = {Beyond the imitation game: Quantifying and extrapolating the capabilities of language models},
author = {Srivastava, Aarohi and Rastogi, Abhinav and Rao, Abhishek and Shoeb, Abu Awal Md and Abid, Abubakar and Fisch, Adam and Brown, Adam R. and Santoro, Adam and Gupta, Aditya and Garriga-Alonso, Adri{\`a} and others},
journal = {arXiv preprint arXiv:2206.04615},
year = {2022}
}
@article{xu2023baize,
title = {Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data},
author = {Xu, Canwen and Guo, Daya and Duan, Nan and McAuley, Julian},
journal = {arXiv preprint arXiv:2304.01196},
year = {2023}
}
@incollection{pytorch,
title = {Py{T}orch: An Imperative Style, High-Performance Deep Learning Library},
author = {Paszke, Adam and Gross, Sam and Massa, Francisco and Lerer, Adam and Bradbury, James and Chanan, Gregory and Killeen, Trevor and Lin, Zeming and Gimelshein, Natalia and Antiga, Luca and Desmaison, Alban and Kopf, Andreas and Yang, Edward and DeVito, Zachary and Raison, Martin and Tejani, Alykhan and Chilamkurthy, Sasank and Steiner, Benoit and Fang, Lu and Bai, Junjie and Chintala, Soumith},
booktitle = {Advances in Neural Information Processing Systems 32},
editor = {H. Wallach and H. Larochelle and A. Beygelzimer and F. d\textquotesingle Alch\'{e}-Buc and E. Fox and R. Garnett},
pages = {8024--8035},
year = {2019},
publisher = {Curran Associates, Inc.},
url = {http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf}
}
@article{zhou2022least,
title = {Least-to-most prompting enables complex reasoning in large language models},
author = {Zhou, Denny and Sch{\"a}rli, Nathanael and Hou, Le and Wei, Jason and Scales, Nathan and Wang, Xuezhi and Schuurmans, Dale and Bousquet, Olivier and Le, Quoc and Chi, Ed},
journal = {arXiv preprint arXiv:2205.10625},
year = {2022}
}
@article{dai2019transformer,
title = {Transformer-{XL}: Attentive language models beyond a fixed-length context},
author = {Dai, Zihang and Yang, Zhilin and Yang, Yiming and Carbonell, Jaime and Le, Quoc V. and Salakhutdinov, Ruslan},
journal = {arXiv preprint arXiv:1901.02860},
year = {2019}
}
@article{kingma2014adam,
title = {Adam: A method for stochastic optimization},
author = {Kingma, Diederik P. and Ba, Jimmy},
journal = {arXiv preprint arXiv:1412.6980},
year = {2014}
}
@incollection{lecun2012efficient,
title = {Efficient backprop},
author = {LeCun, Yann A. and Bottou, L{\'e}on and Orr, Genevieve B and M{\"u}ller, Klaus-Robert},
booktitle = {Neural networks: Tricks of the trade},
pages = {9--48},
year = {2012},
publisher = {Springer}
}
@article{ba2016layer,
title = {Layer normalization},
author = {Ba, Jimmy Lei and Kiros, Jamie Ryan and Hinton, Geoffrey E.},
journal = {arXiv preprint arXiv:1607.06450},
year = {2016}
}
@article{kirkpatrick2017overcoming,
title = {Overcoming catastrophic forgetting in neural networks},
author = {Kirkpatrick, James and Pascanu, Razvan and Rabinowitz, Neil and Veness, Joel and Desjardins, Guillaume and Rusu, Andrei A and Milan, Kieran and Quan, John and Ramalho, Tiago and Grabska-Barwinska, Agnieszka and others},
journal = {Proceedings of the national academy of sciences},
volume = {114},
number = {13},
pages = {3521--3526},
year = {2017},
publisher = {National Acad Sciences}
}
@article{vaswani2017attention,
title = {Attention is all you need},
author = {Vaswani, Ashish and Shazeer, Noam and Parmar, Niki and Uszkoreit, Jakob and Jones, Llion and Gomez, Aidan N and Kaiser, {\L}ukasz and Polosukhin, Illia},
journal = {NeurIPS},
year = {2017}
}
@inproceedings{he2016deep,
title = {Deep residual learning for image recognition},
author = {He, Kaiming and Zhang, Xiangyu and Ren, Shaoqing and Sun, Jian},
booktitle = {Proceedings of the IEEE conference on computer vision and pattern recognition},
pages = {770--778},
year = {2016}
}
@article{steel1960principles,
title = {Principles and procedures of statistics},
author = {Steel, Robert George Douglas and Torrie, James Hiram and others},
journal = {Principles and procedures of statistics.},
year = {1960},
publisher = {McGraw-Hill Book Company, Inc., New York, Toronto, London}
}
@article{humphreys2022data,
title = {A data-driven approach for learning to control computers},
author = {Humphreys, Peter C and Raposo, David and Pohlen, Toby and Thornton, Gregory and Chhaparia, Rachita and Muldal, Alistair and Abramson, Josh and Georgiev, Petko and Goldin, Alex and Santoro, Adam and others},
journal = {arXiv preprint arXiv:2202.08137},
year = {2022}
}
@article{ho2016generative,
title = {Generative adversarial imitation learning},
author = {Ho, Jonathan and Ermon, Stefano},
journal = {Advances in Neural Information Processing Systems},
volume = {29},
year = {2016}
}
@inproceedings{codevilla2018end,
title = {End-to-end driving via conditional imitation learning},
author = {Codevilla, Felipe and M{\"u}ller, Matthias and L{\'o}pez, Antonio and Koltun, Vladlen and Dosovitskiy, Alexey},
booktitle = {2018 IEEE international conference on robotics and automation (ICRA)},
pages = {4693--4700},
year = {2018},
organization = {IEEE}
}
// Nividia Self Driving
@article{bojarski2016end,
title = {End to end learning for self-driving cars},
author = {Bojarski, Mariusz and Del Testa, Davide and Dworakowski, Daniel and Firner, Bernhard and Flepp, Beat and Goyal, Prasoon and Jackel, Lawrence D and Monfort, Mathew and Muller, Urs and Zhang, Jiakai and others},
journal = {arXiv preprint arXiv:1604.07316},
year = {2016}
}
@article{coulom2007computing,
title = {Computing “{E}lo ratings” of move patterns in the game of {G}o},
author = {Coulom, R{\'e}mi},
journal = {ICGA journal},
volume = {30},
number = {4},
pages = {198--208},
year = {2007},
publisher = {IOS Press}
}
@inproceedings{hester2018deep,
title = {Deep q-learning from demonstrations},
author = {Hester, Todd and Vecerik, Matej and Pietquin, Olivier and Lanctot, Marc and Schaul, Tom and Piot, Bilal and Horgan, Dan and Quan, John and Sendonaris, Andrew and Osband, Ian and others},
booktitle = {Proceedings of the AAAI Conference on Artificial Intelligence},
volume = {32},
number = {1},
year = {2018}
}
@inproceedings{oh2016control,
title = {Control of memory, active perception, and action in {M}inecraft},
author = {Oh, Junhyuk and Chockalingam, Valliappa and Lee, Honglak and others},
booktitle = {International Conference on Machine Learning},
pages = {2790--2799},
year = {2016},
organization = {PMLR}
}
@article{kanitscheider2021multi,
title = {Multi-task curriculum learning in a complex, visual, hard-exploration domain: {M}inecraft},
author = {Kanitscheider, Ingmar and Huizinga, Joost and Farhi, David and Guss, William Hebgen and Houghton, Brandon and Sampedro, Raul and Zhokhov, Peter and Baker, Bowen and Ecoffet, Adrien and Tang, Jie and others},
journal = {arXiv preprint arXiv:2106.14876},
year = {2021}
}
@inproceedings{scheller2020sample,
title = {Sample efficient reinforcement learning through learning from demonstrations in {M}inecraft},
author = {Scheller, Christian and Schraner, Yanick and Vogel, Manfred},
booktitle = {NeurIPS 2019 Competition and Demonstration Track},
pages = {67--76},
year = {2020},
organization = {PMLR}
}
@article{guss2019minerl,
title = {Mine{RL}: A large-scale dataset of {M}inecraft demonstrations},
author = {Guss, William H and Houghton, Brandon and Topin, Nicholay and Wang, Phillip and Codel, Cayden and Veloso, Manuela and Salakhutdinov, Ruslan},
journal = {arXiv preprint arXiv:1907.13440},
year = {2019}
}
@inproceedings{tessler2017deep,
title = {A deep hierarchical approach to lifelong learning in {M}inecraft},
author = {Tessler, Chen and Givony, Shahar and Zahavy, Tom and Mankowitz, Daniel and Mannor, Shie},
booktitle = {Proceedings of the AAAI Conference on Artificial Intelligence},
volume = {31},
number = {1},
year = {2017}
}
@article{twinfinite2021played,
author = {Twinfinite Staff},
date = {2021-12-18},
title = {Most Played Games in 2021, Ranked by Peak Concurrent Players},
journal = {Twinfinite},
url = {https://twinfinite.net/2021/12/most-played-games-in-2020-ranked-by-peak-concurrent-players/},
urldate = {2021-12-18}
}
@article{burda2018exploration,
title = {Exploration by random network distillation},
author = {Burda, Yuri and Edwards, Harrison and Storkey, Amos and Klimov, Oleg},
journal = {arXiv preprint arXiv:1810.12894},
year = {2018}
}
@article{bellemare2016unifying,
title = {Unifying count-based exploration and intrinsic motivation},
author = {Bellemare, Marc and Srinivasan, Sriram and Ostrovski, Georg and Schaul, Tom and Saxton, David and Munos, Remi},
journal = {Advances in Neural Information Processing Systems},
volume = {29},
year = {2016}
}
@article{ecoffet2021first,
title = {First return, then explore},
author = {Ecoffet, Adrien and Huizinga, Joost and Lehman, Joel and Stanley, Kenneth O and Clune, Jeff},
journal = {Nature},
volume = {590},
number = {7847},
pages = {580--586},
year = {2021},
publisher = {Nature Publishing Group}
}
@book{sutton2018reinforcement,
title = {Reinforcement learning: An introduction},
author = {Sutton, Richard S and Barto, Andrew G},
year = {2018},
publisher = {MIT press}
}
@article{jaderberg2019human,
title = {Human-level performance in 3{D} multiplayer games with population-based reinforcement learning},
author = {Jaderberg, Max and Czarnecki, Wojciech M and Dunning, Iain and Marris, Luke and Lever, Guy and Castaneda, Antonio Garcia and Beattie, Charles and Rabinowitz, Neil C and Morcos, Ari S and Ruderman, Avraham and others},
journal = {Science},
volume = {364},
number = {6443},
pages = {859--865},
year = {2019},
publisher = {American Association for the Advancement of Science}
}
@inproceedings{bender2021dangers,
title = {On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?},
author = {Bender, Emily M and Gebru, Timnit and McMillan-Major, Angelina and Shmitchell, Shmargaret},
booktitle = {Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency},
pages = {610--623},
year = {2021}
}
@article{dauphin2014identifying,
title = {Identifying and attacking the saddle point problem in high-dimensional non-convex optimization},
author = {Dauphin, Yann N and Pascanu, Razvan and Gulcehre, Caglar and Cho, Kyunghyun and Ganguli, Surya and Bengio, Yoshua},
journal = {Advances in Neural Information Processing Systems},
volume = {27},
year = {2014}
}
@article{yosinski2014transferable,
title = {How transferable are features in deep neural networks?},
author = {Yosinski, Jason and Clune, Jeff and Bengio, Yoshua and Lipson, Hod},
journal = {Advances in Neural Information Processing Systems},
volume = {27},
year = {2014}
}
@article{baker2019emergent,
title = {Emergent tool use from multi-agent autocurricula},
author = {Baker, Bowen and Kanitscheider, Ingmar and Markov, Todor and Wu, Yi and Powell, Glenn and McGrew, Bob and Mordatch, Igor},
journal = {arXiv preprint arXiv:1909.07528},
year = {2019}
}
@article{berner2019dota,
title = {Dota 2 with large scale deep reinforcement learning},
author = {Berner, Christopher and Brockman, Greg and Chan, Brooke and Cheung, Vicki and D{\k{e}}biak, Przemys{\l}aw and Dennison, Christy and Farhi, David and Fischer, Quirin and Hashme, Shariq and Hesse, Chris and others},
journal = {arXiv preprint arXiv:1912.06680},
year = {2019}
}
@article{team2021open,
title = {Open-ended learning leads to generally capable agents},
author = {Team, Open Ended Learning and Stooke, Adam and Mahajan, Anuj and Barros, Catarina and Deck, Charlie and Bauer, Jakob and Sygnowski, Jakub and Trebacz, Maja and Jaderberg, Max and Mathieu, Michael and others},
journal = {arXiv preprint arXiv:2107.12808},
year = {2021}
}
@inproceedings{radford2021learning,
title = {Learning transferable visual models from natural language supervision},
author = {Radford, Alec and Kim, Jong Wook and Hallacy, Chris and Ramesh, Aditya and Goh, Gabriel and Agarwal, Sandhini and Sastry, Girish and Askell, Amanda and Mishkin, Pamela and Clark, Jack and others},
booktitle = {International Conference on Machine Learning},
pages = {8748--8763},
year = {2021},
organization = {PMLR}
}
@article{ramesh2022hierarchical,
title = {Hierarchical text-conditional image generation with {CLIP} latents},
author = {Ramesh, Aditya and Dhariwal, Prafulla and Nichol, Alex and Chu, Casey and Chen, Mark},
journal = {arXiv preprint arXiv:2204.06125},
year = {2022}
}
@article{giusti2015machine,
title = {A machine learning approach to visual perception of forest trails for mobile robots},
author = {Giusti, Alessandro and Guzzi, J{\'e}r{\^o}me and Cire{\c{s}}an, Dan C and He, Fang-Lin and Rodr{\'\i}guez, Juan P and Fontana, Flavio and Faessler, Matthias and Forster, Christian and Schmidhuber, J{\"u}rgen and Di Caro, Gianni and others},
journal = {IEEE Robotics and Automation Letters},
volume = {1},
number = {2},
pages = {661--667},
year = {2015},
publisher = {IEEE}
}
@incollection{sammut1992learning,
title = {Learning to fly},
author = {Sammut, Claude and Hurst, Scott and Kedzier, Dana and Michie, Donald},
booktitle = {Machine Learning Proceedings 1992},
pages = {385--393},
year = {1992},
publisher = {Elsevier}
}
@article{hussein2017imitation,
title = {Imitation learning: A survey of learning methods},
author = {Hussein, Ahmed and Gaber, Mohamed Medhat and Elyan, Eyad and Jayne, Chrisina},
journal = {ACM Computing Surveys (CSUR)},
volume = {50},
number = {2},
pages = {1--35},
year = {2017},
publisher = {ACM New York, NY, USA}
}
@inproceedings{behbahani2019learning,
title = {Learning from demonstration in the wild},
author = {Behbahani, Feryal and Shiarlis, Kyriacos and Chen, Xi and Kurin, Vitaly and Kasewa, Sudhanshu and Stirbu, Ciprian and Gomes, Joao and Paul, Supratik and Oliehoek, Frans A and Messias, Joao and others},
booktitle = {2019 International Conference on Robotics and Automation (ICRA)},
pages = {775--781},
year = {2019},
organization = {IEEE}
}
@inproceedings{edwards2019imitating,
title = {Imitating latent policies from observation},
author = {Edwards, Ashley and Sahni, Himanshu and Schroecker, Yannick and Isbell, Charles},
booktitle = {International conference on machine learning},
pages = {1755--1763},
year = {2019},
organization = {PMLR}
}
@inproceedings{liu2018imitation,
title = {Imitation from observation: Learning to imitate behaviors from raw video via context translation},
author = {Liu, YuXuan and Gupta, Abhishek and Abbeel, Pieter and Levine, Sergey},
booktitle = {2018 IEEE International Conference on Robotics and Automation (ICRA)},
pages = {1118--1125},
year = {2018},
organization = {IEEE}
}
\\ AlphaGo – uses BC on expert games to initialize
@article{silver2016mastering,
title = {Mastering the game of {G}o with deep neural networks and tree search},
author = {Silver, David and Huang, Aja and Maddison, Chris J and Guez, Arthur and Sifre, Laurent and Van Den Driessche, George and Schrittwieser, Julian and Antonoglou, Ioannis and Panneershelvam, Veda and Lanctot, Marc and others},
journal = {Nature},
volume = {529},
number = {7587},
pages = {484--489},
year = {2016},
publisher = {Nature Publishing Group}
}
\\ Recent ViT scaling law paper from google
@article{zhai2021scaling,
author = {Xiaohua Zhai and
Alexander Kolesnikov and
Neil Houlsby and
Lucas Beyer},
title = {Scaling Vision Transformers},
journal = {CoRR},
volume = {abs/2106.04560},
year = {2021},
url = {https://arxiv.org/abs/2106.04560},
eprinttype = {arXiv},
eprint = {2106.04560},
timestamp = {Fri, 11 Jun 2021 11:04:16 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2106-04560.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
@article{kudithipudi2022biological,
title = {Biological underpinnings for lifelong learning machines},
author = {Kudithipudi, Dhireesha and Aguilar-Simon, Mario and Babb, Jonathan and Bazhenov, Maxim and Blackiston, Douglas and Bongard, Josh and Brna, Andrew P and Chakravarthi Raja, Suraj and Cheney, Nick and Clune, Jeff and others},
journal = {Nature Machine Intelligence},
volume = {4},
number = {3},
pages = {196--210},
year = {2022},
publisher = {Nature Publishing Group}
}
// Instagram pretraining paper
@inproceedings{mahajan2018exploring,
title = {Exploring the limits of weakly supervised pretraining},
author = {Mahajan, Dhruv and Girshick, Ross and Ramanathan, Vignesh and He, Kaiming and Paluri, Manohar and Li, Yixuan and Bharambe, Ashwin and Van Der Maaten, Laurens},
booktitle = {Proceedings of the European conference on computer vision (ECCV)},
pages = {181--196},
year = {2018}
}
@article{vinyals2019grandmaster,
title = {Grandmaster level in {S}tar{C}raft {II} using multi-agent reinforcement learning},
author = {Vinyals, Oriol and Babuschkin, Igor and Czarnecki, Wojciech M and Mathieu, Micha{\"e}l and Dudzik, Andrew and Chung, Junyoung and Choi, David H and Powell, Richard and Ewalds, Timo and Georgiev, Petko and others},
journal = {Nature},
volume = {575},
number = {7782},
pages = {350--354},
year = {2019},
publisher = {Nature Publishing Group}
}
@article{bommasani2021opportunities,
title = {On the opportunities and risks of foundation models},
author = {Bommasani, Rishi and Hudson, Drew A and Adeli, Ehsan and Altman, Russ and Arora, Simran and von Arx, Sydney and Bernstein, Michael S and Bohg, Jeannette and Bosselut, Antoine and Brunskill, Emma and others},
journal = {arXiv preprint arXiv:2108.07258},
year = {2021}
}
@article{argall2009survey,
title = {A survey of robot learning from demonstration},
author = {Argall, Brenna D and Chernova, Sonia and Veloso, Manuela and Browning, Brett},
journal = {Robotics and autonomous systems},
volume = {57},
number = {5},
pages = {469--483},
year = {2009},
publisher = {Elsevier}
}
@article{schaal1999imitation,
title = {Is imitation learning the route to humanoid robots?},
author = {Schaal, Stefan},
journal = {Trends in cognitive sciences},
volume = {3},
number = {6},
pages = {233--242},
year = {1999},
publisher = {Elsevier}
}
@inproceedings{ng2000algorithms,
title = {Algorithms for inverse reinforcement learning.},
author = {Ng, Andrew Y and Russell, Stuart J and others},
booktitle = {Icml},
volume = {1},
pages = {2},
year = {2000}
}
@article{pomerleau1988alvinn,
title = {Alvinn: An autonomous land vehicle in a neural network},
author = {Pomerleau, Dean A},
journal = {Advances in Neural Information Processing Systems},
volume = {1},
year = {1988}
}
@inproceedings{sermanet2018time,
title = {Time-contrastive networks: Self-supervised learning from video},
author = {Sermanet, Pierre and Lynch, Corey and Chebotar, Yevgen and Hsu, Jasmine and Jang, Eric and Schaal, Stefan and Levine, Sergey and Brain, Google},
booktitle = {2018 IEEE international conference on robotics and automation (ICRA)},
pages = {1134--1141},
year = {2018},
organization = {IEEE}
}
@inproceedings{calandra2015learning,
title = {Learning inverse dynamics models with contacts},
author = {Calandra, Roberto and Ivaldi, Serena and Deisenroth, Marc Peter and Rueckert, Elmar and Peters, Jan},
booktitle = {2015 IEEE International Conference on Robotics and Automation (ICRA)},
pages = {3186--3191},
year = {2015},
organization = {IEEE}
}
@article{christiano2016transfer,
title = {Transfer from simulation to real world through learning deep inverse dynamics model},