-
Notifications
You must be signed in to change notification settings - Fork 121
/
references.bib
1818 lines (1784 loc) · 79.3 KB
/
references.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@article{martins2012bbb,
author = {Martins, Ines Filipa and Teixeira, Ana L and Pinheiro, Luis and Falcao, Andre O},
doi = {10.1021/ci300124c},
journal = {Journal of Chemical Information and Modeling},
number = {6},
pages = {1686--1697},
title = {{A Bayesian Approach to in Silico Blood-Brain Barrier Penetration Modeling}},
url = {https://doi.org/10.1021/ci300124c},
volume = {52},
year = {2012}
}
@article{ramakrishnan2014quantum,
title = {Quantum chemistry structures and properties of 134 kilo molecules},
author = {Ramakrishnan, Raghunathan and Dral, Pavlo O and Rupp, Matthias and Von Lilienfeld, O Anatole},
year = 2014,
journal = {Scientific data},
publisher = {Nature Publishing Group},
volume = 1,
number = 1,
pages = {1--7}
}
@article{krenn2020self,
title = {Self-Referencing Embedded Strings (SELFIES): A 100\% robust molecular string representation},
author = {Krenn, Mario and H{\"a}se, Florian and Nigam, AkshatKumar and Friederich, Pascal and Aspuru-Guzik, Alan},
year = 2020,
journal = {Machine Learning: Science and Technology},
publisher = {IOP Publishing},
volume = 1,
number = 4,
pages = {045024}
}
@article{sterling2015zinc,
title = {ZINC 15--ligand discovery for everyone},
author = {Sterling, Teague and Irwin, John J},
year = 2015,
journal = {Journal of chemical information and modeling},
publisher = {ACS Publications},
volume = 55,
number = 11,
pages = {2324--2337}
}
@book{alpaydin2020introduction,
title = {Introduction to machine learning},
author = {Alpaydin, Ethem},
year = 2020,
publisher = {MIT press}
}
@article{balachandran2019machine,
title = {Machine learning guided design of functional materials with targeted properties},
author = {Balachandran, Prasanna V},
year = 2019,
journal = {Computational Materials Science},
publisher = {Elsevier},
volume = 164,
pages = {82--90}
}
@article{barber2019predictive,
title = {Predictive inference with the jackknife+},
author = {Barber, Rina Foygel and Candes, Emmanuel J and Ramdas, Aaditya and Tibshirani, Ryan J},
year = 2019,
journal = {arXiv preprint arXiv:1905.02928}
}
@article{barrett2018classifying,
title = {Classifying antimicrobial and multifunctional peptides with Bayesian network models},
author = {Barrett, Rainier and Jiang, Shaoyi and White, Andrew D},
year = 2018,
journal = {Peptide Science},
publisher = {Wiley Online Library},
volume = 110,
number = 4,
pages = {e24079}
}
@book{bishop2006pattern,
title = {Pattern recognition and machine learning},
author = {Bishop, Christopher M},
year = 2006,
publisher = {springer}
}
@article{chawla2002smote,
title = {SMOTE: synthetic minority over-sampling technique},
author = {Chawla, Nitesh V and Bowyer, Kevin W and Hall, Lawrence O and Kegelmeyer, W Philip},
year = 2002,
journal = {Journal of artificial intelligence research},
volume = 16,
pages = {321--357}
}
@article{fung2021benchmarking,
title = {Benchmarking graph neural networks for materials chemistry},
author = {Victor Fung and Jiaxin Zhang and Eric Juarez and Bobby G. Sumpter},
year = 2021,
month = jun,
journal = {npj Computational Materials},
publisher = {Springer Science and Business Media {LLC}},
volume = 7,
number = 1,
doi = {10.1038/s41524-021-00554-0},
url = {https://doi.org/10.1038/s41524-021-00554-0}
}
@article{gomez2020machine,
title = {Machine Learning and Big-Data in Computational Chemistry},
author = {G{\'o}mez-Bombarelli, Rafael and Aspuru-Guzik, Al{\'a}n},
year = 2020,
journal = {Handbook of Materials Modeling: Methods: Theory and Modeling},
publisher = {Springer},
pages = {1939--1962}
}
@article{moriwaki2018mordred,
title = {Mordred: a molecular descriptor calculator},
author = {Moriwaki, Hirotomo and Tian, Yu-Shi and Kawashita, Norihito and Takagi, Tatsuya},
year = 2018,
journal = {Journal of cheminformatics},
publisher = {Springer},
volume = 10,
number = 1,
pages = 4
}
@article{nandy2018strategies,
title = {Strategies and software for machine learning accelerated discovery in transition metal chemistry},
author = {Nandy, Aditya and Duan, Chenru and Janet, Jon Paul and Gugler, Stefan and Kulik, Heather J},
year = 2018,
journal = {Industrial \& Engineering Chemistry Research},
publisher = {ACS Publications},
volume = 57,
number = 42,
pages = {13973--13986}
}
@article{rupp2012fast,
title = {Fast and accurate modeling of molecular atomization energies with machine learning},
author = {Rupp, Matthias and Tkatchenko, Alexandre and M{\"u}ller, Klaus-Robert and Von Lilienfeld, O Anatole},
year = 2012,
journal = {Physical review letters},
publisher = {APS},
volume = 108,
number = 5,
pages = {058301}
}
@article{scherer2020kernel,
title = {Kernel-based machine learning for efficient simulations of molecular liquids},
author = {Scherer, Christoph and Scheid, René and Andrienko, Denis and Bereau, Tristan},
year = 2020,
journal = {Journal of Chemical Theory and Computation},
publisher = {ACS Publications},
volume = 16,
number = 5,
pages = {3194--3204}
}
@article{Sorkun2019,
title = {{AqSolDB, a curated reference set of aqueous solubility and 2D descriptors for a diverse set of compounds}},
author = {Sorkun, Murat Cihan and Khetan, Abhishek and Er, S{\"{u}}leyman},
year = 2019,
journal = {Sci. Data},
volume = 6,
number = 1,
pages = 143,
doi = {10.1038/s41597-019-0151-1},
issn = {2052-4463},
abstract = {Water is a ubiquitous solvent in chemistry and life. It is therefore no surprise that the aqueous solubility of compounds has a key role in various domains, including but not limited to drug discovery, paint, coating, and battery materials design. Measurement and prediction of aqueous solubility is a complex and prevailing challenge in chemistry. For the latter, different data-driven prediction models have recently been developed to augment the physics-based modeling approaches. To construct accurate data-driven estimation models, it is essential that the underlying experimental calibration data used by these models is of high fidelity and quality. Existing solubility datasets show variance in the chemical space of compounds covered, measurement methods, experimental conditions, but also in the non-standard representations, size, and accessibility of data. To address this problem, we generated a new database of compounds, AqSolDB, by merging a total of nine different aqueous solubility datasets, curating the merged data, standardizing and validating the compound representation formats, marking with reliability labels, and providing 2D descriptors of compounds as a Supplementary Resource.}
}
@article{sun2019machine,
title = {Machine learning--assisted molecular design and efficiency prediction for high-performance organic photovoltaic materials},
author = {Sun, Wenbo and Zheng, Yujie and Yang, Ke and Zhang, Qi and Shah, Akeel A and Wu, Zhou and Sun, Yuyang and Feng, Liang and Chen, Dongyang and Xiao, Zeyun and others},
year = 2019,
journal = {Science advances},
publisher = {American Association for the Advancement of Science},
volume = 5,
number = 11,
pages = {eaay4275}
}
@article{sutton2020identifying,
title = {Identifying domains of applicability of machine learning models for materials science},
author = {Sutton, Christopher and Boley, Mario and Ghiringhelli, Luca M and Rupp, Matthias and Vreeken, Jilles and Scheffler, Matthias},
year = 2020,
journal = {Nature Communications},
publisher = {Nature Publishing Group},
volume = 11,
number = 1,
pages = {1--9}
}
@article{wu2018moleculenet,
title = {MoleculeNet: a benchmark for molecular machine learning},
author = {Wu, Zhenqin and Ramsundar, Bharath and Feinberg, Evan N and Gomes, Joseph and Geniesse, Caleb and Pappu, Aneesh S and Leswing, Karl and Pande, Vijay},
year = 2018,
journal = {Chemical science},
publisher = {Royal Society of Chemistry},
volume = 9,
number = 2,
pages = {513--530}
}
@article{xie2019graph,
title = {Graph dynamical networks for unsupervised learning of atomic scale dynamics in materials},
author = {Xie, Tian and France-Lanord, Arthur and Wang, Yanming and Shao-Horn, Yang and Grossman, Jeffrey C},
year = 2019,
journal = {Nature communications},
publisher = {Nature Publishing Group},
volume = 10,
number = 1,
pages = {1--9}
}
@inproceedings{byrd2019effect,
title = {What is the effect of importance weighting in deep learning?},
author = {Byrd, Jonathon and Lipton, Zachary},
year = 2019,
booktitle = {International Conference on Machine Learning},
pages = {872--881},
organization = {PMLR}
}
@article{song2021inferring,
title = {Inferring protein sequence-function relationships with large-scale positive-unlabeled learning},
author = {Song, Hyebin and Bremer, Bennett J and Hinds, Emily C and Raskutti, Garvesh and Romero, Philip A},
year = 2021,
journal = {Cell Systems},
publisher = {Elsevier},
volume = 12,
number = 1,
pages = {92--101}
}
@article{pham2005selection,
title = {Selection of K in K-means clustering},
author = {Pham, Duc Truong and Dimov, Stefan S and Nguyen, Chi D},
year = 2005,
journal = {Proceedings of the Institution of Mechanical Engineers, Part C: Journal of Mechanical Engineering Science},
publisher = {SAGE Publications Sage UK: London, England},
volume = 219,
number = 1,
pages = {103--119}
}
@article{raschka2018model,
title = {Model evaluation, model selection, and algorithm selection in machine learning},
author = {Raschka, Sebastian},
year = 2018,
journal = {arXiv preprint arXiv:1811.12808}
}
@article{he2009learning,
title = {Learning from imbalanced data},
author = {He, Haibo and Garcia, Edwardo A},
year = 2009,
journal = {IEEE Transactions on knowledge and data engineering},
publisher = {Ieee},
volume = 21,
number = 9,
pages = {1263--1284}
}
@article{youbi2021simple,
title = {Simple data balancing achieves competitive worst-group-accuracy},
author = {Youbi Idrissi, Badr and Arjovsky, Martin and Pezeshki, Mohammad and Lopez-Paz, David},
year = 2021,
journal = {arXiv e-prints},
pages = {arXiv--2110}
}
@article{45826,
title = {Neural Architecture Search with Reinforcement Learning},
author = {Barret Zoph and Quoc V. Le},
year = 2017,
journal = {arXiv preprint arXiv:1611.01578},
url = {https://arxiv.org/abs/1611.01578}
}
@article{LeNet,
title = {Gradient-based learning applied to document recognition},
author = {Y. {Lecun} and L. {Bottou} and Y. {Bengio} and P. {Haffner}},
year = 1998,
journal = {Proceedings of the IEEE},
volume = 86,
number = 11,
pages = {2278--2324}
}
@article{hyperband,
title = {Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization},
author = {Lisha Li and Kevin Jamieson and Giulia DeSalvo and Afshin Rostamizadeh and Ameet Talwalkar},
year = 2018,
journal = {Journal of Machine Learning Research},
volume = 18,
number = 185,
pages = {1--52},
url = {http://jmlr.org/papers/v18/16-558.html}
}
@inproceedings{zhao2020exploring,
title = {Exploring self-attention for image recognition},
author = {Zhao, Hengshuang and Jia, Jiaya and Koltun, Vladlen},
year = 2020,
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages = {10076--10085}
}
@inproceedings{glorot2010understanding,
title = {Understanding the difficulty of training deep feedforward neural networks},
author = {Glorot, Xavier and Bengio, Yoshua},
year = 2010,
booktitle = {Proceedings of the thirteenth international conference on artificial intelligence and statistics},
pages = {249--256}
}
@article{noconv,
title = {Deep, big, simple neural nets for handwritten digit recognition},
author = {Cire{\c{s}}an, Dan Claudiu and Meier, Ueli and Gambardella, Luca Maria and Schmidhuber, J{\"u}rgen},
year = 2010,
journal = {Neural computation},
publisher = {MIT Press One Rogers Street, Cambridge, MA 02142-1209, USA journals-info~…},
volume = 22,
number = 12,
pages = {3207--3220}
}
@article{solubility,
title = {Bioinformatics approaches for improved recombinant protein production in Escherichia coli: protein solubility prediction},
author = {Chang, Catherine Ching Han and Song, Jiangning and Tey, Beng Ti and Ramanan, Ramakrishnan Nagasundara},
year = 2014,
journal = {Briefings in bioinformatics},
publisher = {Oxford University Press},
volume = 15,
number = 6,
pages = {953--962}
}
@article{deepsol,
title = {{DeepSol: a deep learning framework for sequence-based protein solubility prediction}},
author = {Khurana, Sameer and Rawi, Reda and Kunji, Khalid and Chuang, Gwo-Yu and Bensmail, Halima and Mall, Raghvendra},
year = 2018,
month = {03},
journal = {Bioinformatics},
volume = 34,
number = 15,
pages = {2605--2613},
doi = {10.1093/bioinformatics/bty166},
issn = {1367-4803},
url = {https://doi.org/10.1093/bioinformatics/bty166},
abstract = {{Protein solubility plays a vital role in pharmaceutical research and production yield. For a given protein, the extent of its solubility can represent the quality of its function, and is ultimately defined by its sequence. Thus, it is imperative to develop novel, highly accurate in silico sequence-based protein solubility predictors. In this work we propose, DeepSol, a novel Deep Learning-based protein solubility predictor. The backbone of our framework is a convolutional neural network that exploits k-mer structure and additional sequence and structural features extracted from the protein sequence.DeepSol outperformed all known sequence-based state-of-the-art solubility prediction methods and attained an accuracy of 0.77 and Matthew’s correlation coefficient of 0.55. The superior prediction accuracy of DeepSol allows to screen for sequences with enhanced production capacity and can more reliably predict solubility of novel proteins.DeepSol’s best performing models and results are publicly deposited at https://doi.org/10.5281/zenodo.1162886 (Khurana and Mall, 2018).Supplementary data are available at Bioinformatics online.}}
}
@article{neal2018modern,
title = {A modern take on the bias-variance tradeoff in neural networks},
author = {Neal, Brady and Mittal, Sarthak and Baratin, Aristide and Tantia, Vinayak and Scicluna, Matthew and Lacoste-Julien, Simon and Mitliagkas, Ioannis},
year = 2018,
journal = {arXiv preprint arXiv:1810.08591},
abstract = {{Protein solubility plays a vital role in pharmaceutical research and production yield. For a given protein, the extent of its solubility can represent the quality of its function, and is ultimately defined by its sequence. Thus, it is imperative to develop novel, highly accurate in silico sequence-based protein solubility predictors. In this work we propose, DeepSol, a novel Deep Learning-based protein solubility predictor. The backbone of our framework is a convolutional neural network that exploits k-mer structure and additional sequence and structural features extracted from the protein sequence.DeepSol outperformed all known sequence-based state-of-the-art solubility prediction methods and attained an accuracy of 0.77 and Matthew's correlation coefficient of 0.55. The superior prediction accuracy of DeepSol allows to screen for sequences with enhanced production capacity and can more reliably predict solubility of novel proteins.DeepSol's best performing models and results are publicly deposited at https://doi.org/10.5281/zenodo.1162886 (Khurana and Mall, 2018).Supplementary data are available at Bioinformatics online.}}
}
@inproceedings{gal2016dropout,
title = {Dropout as a bayesian approximation: Representing model uncertainty in deep learning},
author = {Gal, Yarin and Ghahramani, Zoubin},
year = 2016,
booktitle = {international conference on machine learning},
pages = {1050--1059}
}
@article{kipf2016semi,
title = {Semi-supervised classification with graph convolutional networks},
author = {Kipf, Thomas N and Welling, Max},
year = 2016,
journal = {arXiv preprint arXiv:1609.02907}
}
@article{battaglia2018relational,
title = {Relational inductive biases, deep learning, and graph networks},
author = {Battaglia, Peter W and Hamrick, Jessica B and Bapst, Victor and Sanchez-Gonzalez, Alvaro and Zambaldi, Vinicius and Malinowski, Mateusz and Tacchetti, Andrea and Raposo, David and Santoro, Adam and Faulkner, Ryan and others},
year = 2018,
journal = {arXiv preprint arXiv:1806.01261}
}
@article{li2015gated,
title = {Gated graph sequence neural networks},
author = {Li, Yujia and Tarlow, Daniel and Brockschmidt, Marc and Zemel, Richard},
year = 2015,
journal = {arXiv preprint arXiv:1511.05493}
}
@article{chung2014empirical,
title = {Empirical evaluation of gated recurrent neural networks on sequence modeling},
author = {Chung, Junyoung and Gulcehre, Caglar and Cho, KyungHyun and Bengio, Yoshua},
year = 2014,
journal = {arXiv preprint arXiv:1412.3555}
}
@article{gilmer2017neural,
title = {Neural message passing for quantum chemistry},
author = {Gilmer, Justin and Schoenholz, Samuel S and Riley, Patrick F and Vinyals, Oriol and Dahl, George E},
year = 2017,
journal = {arXiv preprint arXiv:1704.01212}
}
@inproceedings{vaswani2017attention,
title = {Attention is all you need},
author = {Vaswani, Ashish and Shazeer, Noam and Parmar, Niki and Uszkoreit, Jakob and Jones, Llion and Gomez, Aidan N and Kaiser, {\L}ukasz and Polosukhin, Illia},
year = 2017,
booktitle = {Advances in neural information processing systems},
pages = {5998--6008}
}
@article{luong2015effective,
title = {Effective approaches to attention-based neural machine translation},
author = {Luong, Minh-Thang and Pham, Hieu and Manning, Christopher D},
year = 2015,
journal = {arXiv preprint arXiv:1508.04025}
}
@article{zhang2018gaan,
title = {Gaan: Gated attention networks for learning on large and spatiotemporal graphs},
author = {Zhang, Jiani and Shi, Xingjian and Xie, Junyuan and Ma, Hao and King, Irwin and Yeung, Dit-Yan},
year = 2018,
journal = {arXiv preprint arXiv:1803.07294}
}
@article{treisman1980feature,
title = {A feature-integration theory of attention},
author = {Treisman, Anne M and Gelade, Garry},
year = 1980,
journal = {Cognitive psychology},
publisher = {Elsevier},
volume = 12,
number = 1,
pages = {97--136}
}
@article{BALUJA1997329,
title = {Expectation-based selective attention for visual monitoring and control of a robot vehicle},
author = {Shumeet Baluja and Dean A. Pomerleau},
year = 1997,
journal = {Robotics and Autonomous Systems},
volume = 22,
number = 3,
pages = {329--344},
doi = {https://doi.org/10.1016/S0921-8890(97)00046-8},
issn = {0921-8890},
url = {http://www.sciencedirect.com/science/article/pii/S0921889097000468},
note = {Robot Learning: The New Wave},
keywords = {Expectation-based selective attention, Autonomous navigation, Temporal coherence, Saliency map, Artificial neural networks},
abstract = {Reliable vision-based control of an autonomous vehicle requires the ability to focus attention on the important features in an input scene. Previous work with an autonomous lane following system, ALVINN (Pomerleau, 1993), has yielded good results in uncluttered conditions. This paper presents an artificial neural network based learning approach for handling difficult scenes which will confuse the ALVINN system. This work presents a mechanism for achieving task-specific focus of attention by exploiting temporal coherence. A saliency map, which is based upon a computed expectation of the contents of the inputs in the next time step, indicates which regions of the input retina are important for performing the task. The saliency map can be used to accentuate the features which are important for the task, and de-emphasize those which are not.}
}
@article{thomas2018tensor,
title = {Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds},
author = {Thomas, Nathaniel and Smidt, Tess and Kearnes, Steven and Yang, Lusann and Li, Li and Kohlhoff, Kai and Riley, Patrick},
year = 2018,
journal = {arXiv preprint arXiv:1802.08219}
}
@inproceedings{weiler20183d,
title = {3d steerable cnns: Learning rotationally equivariant features in volumetric data},
author = {Weiler, Maurice and Geiger, Mario and Welling, Max and Boomsma, Wouter and Cohen, Taco S},
year = 2018,
booktitle = {Advances in Neural Information Processing Systems},
pages = {10381--10392}
}
@article{Bart,
title = {On representing chemical environments},
author = {Bart\'ok, Albert P. and Kondor, Risi and Cs\'anyi, G\'abor},
year = 2013,
month = {May},
journal = {Phys. Rev. B},
publisher = {American Physical Society},
volume = 87,
pages = 184115,
doi = {10.1103/PhysRevB.87.184115},
url = {https://link.aps.org/doi/10.1103/PhysRevB.87.184115},
issue = 18,
numpages = 16
}
@article{behler2011atom,
title = {Atom-centered symmetry functions for constructing high-dimensional neural network potentials},
author = {Behler, J{\"o}rg},
year = 2011,
journal = {The Journal of chemical physics},
publisher = {American Institute of Physics},
volume = 134,
number = 7,
pages = {074106}
}
@inproceedings{ravanbakhsh2017equivariance,
title = {Equivariance Through Parameter-Sharing},
author = {Ravanbakhsh, Siamak and Schneider, Jeff and P{\'o}czos, Barnab{\'a}s},
year = 2017,
booktitle = {International Conference on Machine Learning},
pages = {2892--2901}
}
@article{mesquita2020rethinking,
title = {Rethinking pooling in graph neural networks},
author = {Mesquita, Diego and Souza, Amauri and Kaski, Samuel},
year = 2020,
journal = {Advances in Neural Information Processing Systems},
volume = 33
}
@article{luzhnica2019graph,
title = {On graph classification networks, datasets and baselines},
author = {Luzhnica, Enxhell and Day, Ben and Li{\`o}, Pietro},
year = 2019,
journal = {arXiv preprint arXiv:1905.04682}
}
@inproceedings{xu2018powerful,
title = {How Powerful are Graph Neural Networks?},
author = {Xu, Keyulu and Hu, Weihua and Leskovec, Jure and Jegelka, Stefanie},
year = 2018,
booktitle = {International Conference on Learning Representations}
}
@inproceedings{hamilton2017inductive,
title = {Inductive representation learning on large graphs},
author = {Hamilton, Will and Ying, Zhitao and Leskovec, Jure},
year = 2017,
booktitle = {Advances in neural information processing systems},
pages = {1024--1034}
}
@article{shchur2018pitfalls,
title = {Pitfalls of graph neural network evaluation},
author = {Shchur, Oleksandr and Mumme, Maximilian and Bojchevski, Aleksandar and G{\"u}nnemann, Stephan},
year = 2018,
journal = {arXiv preprint arXiv:1811.05868}
}
@inproceedings{errica2019fair,
title = {A Fair Comparison of Graph Neural Networks for Graph Classification},
author = {Errica, Federico and Podda, Marco and Bacciu, Davide and Micheli, Alessio},
year = 2019,
booktitle = {International Conference on Learning Representations}
}
@article{dwivedi2020benchmarking,
title = {Benchmarking graph neural networks},
author = {Dwivedi, Vijay Prakash and Joshi, Chaitanya K and Laurent, Thomas and Bengio, Yoshua and Bresson, Xavier},
year = 2020,
journal = {arXiv preprint arXiv:2003.00982}
}
@article{bronstein2017geometric,
title = {Geometric deep learning: going beyond euclidean data},
author = {Bronstein, Michael M and Bruna, Joan and LeCun, Yann and Szlam, Arthur and Vandergheynst, Pierre},
year = 2017,
journal = {IEEE Signal Processing Magazine},
publisher = {IEEE},
volume = 34,
number = 4,
pages = {18--42}
}
@article{wu2020comprehensive,
title = {A comprehensive survey on graph neural networks},
author = {Wu, Zonghan and Pan, Shirui and Chen, Fengwen and Long, Guodong and Zhang, Chengqi and Philip, S Yu},
year = 2020,
journal = {IEEE Transactions on Neural Networks and Learning Systems},
publisher = {IEEE}
}
@article{kobyzev2020normalizing,
title = {Normalizing flows: An introduction and review of current methods},
author = {Kobyzev, Ivan and Prince, Simon and Brubaker, Marcus},
year = 2020,
journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
publisher = {IEEE}
}
@inproceedings{papamakarios2017masked,
title = {Masked autoregressive flow for density estimation},
author = {Papamakarios, George and Pavlakou, Theo and Murray, Iain},
year = 2017,
booktitle = {Advances in Neural Information Processing Systems},
pages = {2338--2347}
}
@inproceedings{kingma2016improved,
title = {Improved variational inference with inverse autoregressive flow},
author = {Kingma, Durk P and Salimans, Tim and Jozefowicz, Rafal and Chen, Xi and Sutskever, Ilya and Welling, Max},
year = 2016,
booktitle = {Advances in neural information processing systems},
pages = {4743--4751}
}
@article{kim2018flowavenet,
title = {FloWaveNet: A generative flow for raw audio},
author = {Kim, Sungwon and Lee, Sang-gil and Song, Jongyoon and Kim, Jaehyeon and Yoon, Sungroh},
year = 2018,
journal = {arXiv preprint arXiv:1811.02155}
}
@article{dinh2016density,
title = {Density estimation using real nvp},
author = {Dinh, Laurent and Sohl-Dickstein, Jascha and Bengio, Samy},
year = 2016,
journal = {arXiv preprint arXiv:1605.08803}
}
@article{das2019dimensionality,
title = {Dimensionality reduction flows},
author = {Das, Hari Prasanna and Abbeel, Pieter and Spanos, Costas J},
year = 2019,
journal = {arXiv preprint arXiv:1908.01686}
}
@article{foote2000relation,
title = {A relation between the principal axes of inertia and ligand binding},
author = {Foote, Jefferson and Raman, Anandi},
year = 2000,
journal = {Proceedings of the National Academy of Sciences},
publisher = {National Acad Sciences},
volume = 97,
number = 3,
pages = {978--983}
}
@article{esteves2020theoretical,
title = {Theoretical Aspects of Group Equivariant Neural Networks},
author = {Esteves, Carlos},
year = 2020,
journal = {arXiv preprint arXiv:2004.05154}
}
@inproceedings{mathieu2019disentangling,
title = {Disentangling disentanglement in variational autoencoders},
author = {Mathieu, Emile and Rainforth, Tom and Siddharth, N and Teh, Yee Whye},
year = 2019,
booktitle = {International Conference on Machine Learning},
publisher = {PMLR},
pages = {4402--4412}
}
@inproceedings{klicpera2019directional,
title = {Directional Message Passing for Molecular Graphs},
author = {Klicpera, Johannes and Gro{\ss}, Janek and G{\"u}nnemann, Stephan},
year = 2020,
booktitle = {International Conference on Learning Representations}
}
@article{jing2020learning,
title = {Learning from Protein Structure with Geometric Vector Perceptrons},
author = {Jing, Bowen and Eismann, Stephan and Suriana, Patricia and Townshend, Raphael JL and Dror, Ron},
year = 2020,
journal = {arXiv preprint arXiv:2009.01411}
}
@article{li2020graph,
title = {Graph neural network based coarse-grained mapping prediction},
author = {Li, Zhiheng and Wellawatte, Geemi P and Chakraborty, Maghesree and Gandhi, Heta A and Xu, Chenliang and White, Andrew D},
year = 2020,
journal = {Chemical Science},
publisher = {Royal Society of Chemistry},
volume = 11,
number = 35,
pages = {9524--9531}
}
@article{yang2020predicting,
title = {Predicting Chemical Shifts with Graph Neural Networks},
author = {Yang, Ziyue and Chakraborty, Maghesree and White, Andrew D},
year = 2020,
journal = {bioRxiv},
publisher = {Cold Spring Harbor Laboratory}
}
@article{maziarka2020molecule,
title = {Molecule Attention Transformer},
author = {Maziarka, {\L}ukasz and Danel, Tomasz and Mucha, S{\l}awomir and Rataj, Krzysztof and Tabor, Jacek and Jastrz{\k{e}}bski, Stanis{\l}aw},
year = 2020,
journal = {arXiv preprint arXiv:2002.08264}
}
@article{miller2020relevance,
title = {Relevance of rotationally equivariant convolutions for predicting molecular properties},
author = {Miller, Benjamin Kurt and Geiger, Mario and Smidt, Tess E and No{\'e}, Frank},
year = 2020,
journal = {arXiv preprint arXiv:2008.08461}
}
@article{wu2020stochastic,
title = {Stochastic Normalizing Flows},
author = {Wu, Hao and K{\"o}hler, Jonas and No{\'e}, Frank},
year = 2020,
journal = {arXiv preprint arXiv:2002.06707}
}
@inproceedings{papamakarios2019sequential,
title = {Sequential neural likelihood: Fast likelihood-free inference with autoregressive flows},
author = {Papamakarios, George and Sterratt, David and Murray, Iain},
year = 2019,
booktitle = {The 22nd International Conference on Artificial Intelligence and Statistics},
pages = {837--848},
organization = {PMLR}
}
@article{papamakarios2019normalizing,
title = {Normalizing flows for probabilistic modeling and inference},
author = {Papamakarios, George and Nalisnick, Eric and Rezende, Danilo Jimenez and Mohamed, Shakir and Lakshminarayanan, Balaji},
year = 2019,
journal = {arXiv preprint arXiv:1912.02762}
}
@inproceedings{zaheer2017deep,
title = {Deep sets},
author = {Zaheer, Manzil and Kottur, Satwik and Ravanbakhsh, Siamak and Poczos, Barnabas and Salakhutdinov, Russ R and Smola, Alexander J},
year = 2017,
booktitle = {Advances in neural information processing systems},
pages = {3391--3401}
}
@inproceedings{kondor2018generalization,
title = {On the Generalization of Equivariance and Convolution in Neural Networks to the Action of Compact Groups},
author = {Kondor, Risi and Trivedi, Shubhendu},
year = 2018,
booktitle = {International Conference on Machine Learning},
pages = {2747--2755}
}
@article{cohen2019general,
title = {A general theory of equivariant cnns on homogeneous spaces},
author = {Cohen, Taco S and Geiger, Mario and Weiler, Maurice},
year = 2019,
journal = {Advances in neural information processing systems},
volume = 32,
pages = {9145--9156}
}
@article{winter2021auto,
title = {Auto-Encoding Molecular Conformations},
author = {Winter, Robin and No{\'e}, Frank and Clevert, Djork-Arn{\'e}},
year = 2021,
journal = {arXiv preprint arXiv:2101.01618}
}
@article{finzi2020generalizing,
title = {Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data},
author = {Finzi, Marc and Stanton, Samuel and Izmailov, Pavel and Wilson, Andrew Gordon},
year = 2020,
journal = {arXiv preprint arXiv:2002.12880}
}
@article{romero2020attentive,
title = {Attentive Group Equivariant Convolutional Networks},
author = {Romero, David W and Bekkers, Erik J and Tomczak, Jakub M and Hoogendoorn, Mark},
year = 2020,
journal = {arXiv},
pages = {arXiv--2002}
}
@inproceedings{cohen2016group,
title = {Group equivariant convolutional networks},
author = {Cohen, Taco and Welling, Max},
year = 2016,
booktitle = {International conference on machine learning},
pages = {2990--2999}
}
@article{wang2020equivariant,
title = {Equivariant Maps for Hierarchical Structures},
author = {Wang, Renhao and Albooyeh, Marjan and Ravanbakhsh, Siamak},
year = 2020,
journal = {arXiv preprint arXiv:2006.03627}
}
@article{batzner2021se3equivariant,
title = {SE(3)-Equivariant Graph Neural Networks for Data-Efficient and Accurate Interatomic Potentials},
author = {Simon Batzner and Tess E. Smidt and Lixin Sun and Jonathan P. Mailoa and Mordechai Kornbluth and Nicola Molinari and Boris Kozinsky},
year = 2021,
journal = {arXiv preprint arXiv:2101.03164}
}
@article{klicpera2020directional,
title = {Directional message passing for molecular graphs},
author = {Klicpera, Johannes and Gro{\ss}, Janek and G{\"u}nnemann, Stephan},
year = 2020,
journal = {arXiv preprint arXiv:2003.03123}
}
@book{serre1977linear,
title = {Linear representations of finite groups},
author = {Serre, Jean-Pierre},
year = 1977,
publisher = {Springer},
volume = 42
}
@book{zee2016,
title = {Group theory in a nutshell for physicists},
author = {Zee, Anthony},
year = 2016,
publisher = {Princeton University Press}
}
@article{musil2021physicsinspired,
title = {Physics-inspired structural representations for molecules and materials},
author = {Felix Musil and Andrea Grisafi and Albert P. Bartók and Christoph Ortner and Gábor Csányi and Michele Ceriotti},
year = 2021,
journal = {arXiv preprint arXiv:2101.04673}
}
@article{chew2020fast,
title = {Fast predictions of liquid-phase acid-catalyzed reaction rates using molecular dynamics simulations and convolutional neural networks},
author = {Chew, Alex K and Jiang, Shengli and Zhang, Weiqi and Zavala, Victor M and Van Lehn, Reid C},
year = 2020,
journal = {Chemical Science},
publisher = {Royal Society of Chemistry},
volume = 11,
number = 46,
pages = {12464--12476}
}
@article{lang2020wigner,
title = {A Wigner-Eckart Theorem for Group Equivariant Convolution Kernels},
author = {Lang, Leon and Weiler, Maurice},
year = 2020,
journal = {arXiv preprint arXiv:2010.10952}
}
@article{kondor2018clebsch,
title = {Clebsch-gordan nets: a fully fourier space spherical convolutional neural network},
author = {Kondor, Risi and Lin, Zhen and Trivedi, Shubhendu},
year = 2018,
journal = {arXiv preprint arXiv:1806.09231}
}
@inproceedings{hoogeboom2021argmax,
title = {Argmax Flows: Learning Categorical Distributions with Normalizing Flows},
author = {Emiel Hoogeboom and Didrik Nielsen and Priyank Jaini and Patrick Forr{\'e} and Max Welling},
year = 2021,
booktitle = {Third Symposium on Advances in Approximate Bayesian Inference},
url = {https://openreview.net/forum?id=fdsXhAy5Cp}
}
@article{Xie2018Crystal,
title = {Crystal Graph Convolutional Neural Networks for an Accurate and Interpretable Prediction of Material Properties},
author = {Xie, Tian and Grossman, Jeffrey C.},
year = 2018,
month = {Apr},
journal = {Phys. Rev. Lett.},
publisher = {American Physical Society},
volume = 120,
pages = 145301,
doi = {10.1103/PhysRevLett.120.145301},
url = {https://link.aps.org/doi/10.1103/PhysRevLett.120.145301},
issue = 14,
numpages = 6
}
@article{Krenn_2020,
title = {Self-referencing embedded strings ({SELFIES}): A 100{\%} robust molecular string representation},
author = {Mario Krenn and Florian Häse and AkshatKumar Nigam and Pascal Friederich and Alan Aspuru-Guzik},
year = 2020,
month = {nov},
journal = {Machine Learning: Science and Technology},
publisher = {{IOP} Publishing},
volume = 1,
number = 4,
pages = {045024},
doi = {10.1088/2632-2153/aba947},
url = {https://doi.org/10.1088/2632-2153/aba947},
abstract = {The discovery of novel materials and functional molecules can help to solve some of society’s most urgent challenges, ranging from efficient energy harvesting and storage to uncovering novel pharmaceutical drug candidates. Traditionally matter engineering–generally denoted as inverse design–was based massively on human intuition and high-throughput virtual screening. The last few years have seen the emergence of significant interest in computer-inspired designs based on evolutionary or deep learning methods. The major challenge here is that the standard strings molecular representation SMILES shows substantial weaknesses in that task because large fractions of strings do not correspond to valid molecules. Here, we solve this problem at a fundamental level and introduce SELFIES (SELF-referencIng Embedded Strings), a string-based representation of molecules which is 100% robust. Every SELFIES string corresponds to a valid molecule, and SELFIES can represent every molecule. SELFIES can be directly applied in arbitrary machine learning models without the adaptation of the models; each of the generated molecule candidates is valid. In our experiments, the model’s internal memory stores two orders of magnitude more diverse molecules than a similar test with SMILES. Furthermore, as all molecules are valid, it allows for explanation and interpretation of the internal working of the generative models.}
}
@article{rajan2020decimer,
title = {DECIMER: towards deep learning for chemical image recognition},
author = {Rajan, Kohulan and Zielesny, Achim and Steinbeck, Christoph},
year = 2020,
journal = {Journal of Cheminformatics},
publisher = {Springer},
volume = 12,
number = 1,
pages = {1--9}
}
@article{segler2018generating,
title = {Generating focused molecule libraries for drug discovery with recurrent neural networks},
author = {Segler, Marwin HS and Kogej, Thierry and Tyrchan, Christian and Waller, Mark P},
year = 2018,
journal = {ACS central science},
publisher = {ACS Publications},
volume = 4,
number = 1,
pages = {120--131}
}
@article{gomez2018automatic,
title = {Automatic chemical design using a data-driven continuous representation of molecules},
author = {G{\'o}mez-Bombarelli, Rafael and Wei, Jennifer N and Duvenaud, David and Hern{\'a}ndez-Lobato, Jos{\'e} Miguel and S{\'a}nchez-Lengeling, Benjam{\'\i}n and Sheberla, Dennis and Aguilera-Iparraguirre, Jorge and Hirzel, Timothy D and Adams, Ryan P and Aspuru-Guzik, Al{\'a}n},
year = 2018,
journal = {ACS central science},
publisher = {ACS Publications},
volume = 4,
number = 2,
pages = {268--276}
}
@article{chithrananda2020chemberta,
title = {ChemBERTa: Large-Scale Self-Supervised Pretraining for Molecular Property Prediction},
author = {Chithrananda, Seyone and Grand, Gabe and Ramsundar, Bharath},
year = 2020,
journal = {arXiv preprint arXiv:2010.09885}
}
@article{brown2020language,
title = {Language models are few-shot learners},
author = {Brown, Tom B and Mann, Benjamin and Ryder, Nick and Subbiah, Melanie and Kaplan, Jared and Dhariwal, Prafulla and Neelakantan, Arvind and Shyam, Pranav and Sastry, Girish and Askell, Amanda and others},
year = 2020,
journal = {arXiv preprint arXiv:2005.14165}
}
@article{liu2019roberta,
title = {Roberta: A robustly optimized bert pretraining approach},
author = {Liu, Yinhan and Ott, Myle and Goyal, Naman and Du, Jingfei and Joshi, Mandar and Chen, Danqi and Levy, Omer and Lewis, Mike and Zettlemoyer, Luke and Stoyanov, Veselin},
year = 2019,
journal = {arXiv preprint arXiv:1907.11692}
}
@article{reynolds2021prompt,
title = {Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm},
author = {Laria Reynolds and Kyle McDonell},
year = 2021,
journal = {arXiv preprint arXiv:2102.07350}
}
@article{tshitoyan2019unsupervised,
title = {Unsupervised word embeddings capture latent knowledge from materials science literature},
author = {Tshitoyan, Vahe and Dagdelen, John and Weston, Leigh and Dunn, Alexander and Rong, Ziqin and Kononova, Olga and Persson, Kristin A and Ceder, Gerbrand and Jain, Anubhav},
year = 2019,
journal = {Nature},
publisher = {Nature Publishing Group},
volume = 571,
number = 7763,
pages = {95--98}
}
@article{friedrich2020sofc,
title = {The SOFC-Exp corpus and neural approaches to information extraction in the materials science domain},
author = {Friedrich, Annemarie and Adel, Heike and Tomazic, Federico and Hingerl, Johannes and Benteau, Renou and Maruscyk, Anika and Lange, Lukas},
year = 2020,
journal = {arXiv preprint arXiv:2006.03039}
}
@article{weininger1988smiles,
title = {SMILES, a chemical language and information system. 1. Introduction to methodology and encoding rules},
author = {Weininger, David},
year = 1988,
journal = {Journal of chemical information and computer sciences},
publisher = {ACS Publications},
volume = 28,
number = 1,
pages = {31--36}
}
@article{wang2021molclr,
title = {MolCLR: Molecular Contrastive Learning of Representations via Graph Neural Networks},
author = {Yuyang Wang and Jianren Wang and Zhonglin Cao and Amir Barati Farimani},
year = 2021,
journal = {arXiv preprint arXiv:2102.10056}
}
@misc{geiger2022e3nn,
doi = {10.48550/ARXIV.2207.09453},
url = {https://arxiv.org/abs/2207.09453},
author = {Geiger, Mario and Smidt, Tess},
keywords = {Machine Learning (cs.LG), Artificial Intelligence (cs.AI), Neural and Evolutionary Computing (cs.NE), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {e3nn: Euclidean Neural Networks},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
@article{heller2015inchi,
title = {InChI, the IUPAC international chemical identifier},
author = {Heller, Stephen R and McNaught, Alan and Pletnev, Igor and Stein, Stephen and Tchekhovskoi, Dmitrii},
year = 2015,
journal = {Journal of cheminformatics},
publisher = {BioMed Central},
volume = 7,
number = 1,
pages = {1--34}
}
@article{butler2018machine,
title = {Machine learning for molecular and materials science},
author = {Butler, Keith T and Davies, Daniel W and Cartwright, Hugh and Isayev, Olexandr and Walsh, Aron},
year = 2018,
journal = {Nature},
publisher = {Nature Publishing Group},
volume = 559,
number = 7715,
pages = {547--555}
}
@article{swain2016chemdataextractor,
title = {ChemDataExtractor: a toolkit for automated extraction of chemical information from the scientific literature},
author = {Swain, Matthew C and Cole, Jacqueline M},
year = 2016,
journal = {Journal of chemical information and modeling},
publisher = {ACS Publications},
volume = 56,
number = 10,
pages = {1894--1904}
}
@article{merk2018novo,
title = {De novo design of bioactive small molecules by artificial intelligence},
author = {Merk, Daniel and Friedrich, Lukas and Grisoni, Francesca and Schneider, Gisbert},
year = 2018,
journal = {Molecular informatics},
publisher = {Wiley Online Library},
volume = 37,
number = {1-2},
pages = 1700153
}
@article{schwaller2020prediction,
title = {Prediction of chemical reaction yields using deep learning},
author = {Schwaller, Philippe and Vaucher, Alain C and Laino, Teodoro and Reymond, Jean-Louis},
year = 2020,
journal = {ChemRxiv Preprint},
url = {https://doi.org/10.26434/chemrxiv.12758474.v2}
}
@article{schwaller2020predicting,
title = {Predicting retrosynthetic pathways using transformer-based models and a hyper-graph exploration strategy},
author = {Schwaller, Philippe and Petraglia, Riccardo and Zullo, Valerio and Nair, Vishnu H and Haeuselmann, Rico Andreas and Pisoni, Riccardo and Bekas, Costas and Iuliano, Anna and Laino, Teodoro},
year = 2020,
journal = {Chemical Science},
publisher = {Royal Society of Chemistry},
volume = 11,
number = 12,
pages = {3316--3325}
}
@article{schwaller2019molecular,
title = {Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction},
author = {Schwaller, Philippe and Laino, Teodoro and Gaudin, Th{\'e}ophile and Bolgar, Peter and Hunter, Christopher A and Bekas, Costas and Lee, Alpha A},
year = 2019,
journal = {ACS central science},
publisher = {ACS Publications},
volume = 5,
number = 9,
pages = {1572--1583}
}
@article{vaucher2020automated,
title = {Automated extraction of chemical synthesis actions from experimental procedures},
author = {Vaucher, Alain C and Zipoli, Federico and Geluykens, Joppe and Nair, Vishnu H and Schwaller, Philippe and Laino, Teodoro},
year = 2020,
journal = {Nature communications},
publisher = {Nature Publishing Group},
volume = 11,
number = 1,
pages = {1--11}
}
@article{schwaller2021mapping,
title = {Mapping the space of chemical reactions using attention-based neural networks},
author = {Schwaller, Philippe and Probst, Daniel and Vaucher, Alain C and Nair, Vishnu H and Kreutter, David and Laino, Teodoro and Reymond, Jean-Louis},
year = 2021,
journal = {Nature Machine Intelligence},
publisher = {Nature Publishing Group},
pages = {1--9}
}
@article{brown2019guacamol,
title = {GuacaMol: benchmarking models for de novo molecular design},
author = {Brown, Nathan and Fiscato, Marco and Segler, Marwin HS and Vaucher, Alain C},
year = 2019,
journal = {Journal of chemical information and modeling},
publisher = {ACS Publications},
volume = 59,
number = 3,
pages = {1096--1108}
}
@article{shmilovich2020discovery,
title = {Discovery of self-assembling $\pi$-conjugated peptides by active learning-directed coarse-grained molecular simulation},
author = {Shmilovich, Kirill and Mansbach, Rachael A and Sidky, Hythem and Dunne, Olivia E and Panda, Sayak Subhra and Tovar, John D and Ferguson, Andrew L},
year = 2020,
journal = {The Journal of Physical Chemistry B},
publisher = {ACS Publications},
volume = 124,
number = 19,
pages = {3873--3891}
}
@article{wang2019coarse,
title = {Coarse-graining auto-encoders for molecular dynamics},
author = {Wang, Wujie and G{\'o}mez-Bombarelli, Rafael},
year = 2019,
journal = {npj Computational Materials},
publisher = {Nature Publishing Group},
volume = 5,
number = 1,
pages = {1--9}
}
@article{ribeiro2018reweighted,
title = {Reweighted autoencoded variational Bayes for enhanced sampling (RAVE)},
author = {Ribeiro, Jo{\~a}o Marcelo Lamim and Bravo, Pablo and Wang, Yihang and Tiwary, Pratyush},
year = 2018,
journal = {The Journal of chemical physics},
publisher = {AIP Publishing LLC},