This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale | 12042 | ||
---|---|---|---|
A Simple Framework for Contrastive Learning of Visual Representations | 8476 | ||
Language Models are Few-Shot Learners | 7903 | OpenAI | |
YOLOv4: Optimal Speed and Accuracy of Object Detection | 7860 | Academia Sinica | |
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. | 6362 | ||
Momentum Contrast for Unsupervised Visual Representation Learning | 6060 | Meta | |
End-to-End Object Detection with Transformers | 4998 | Meta, Paris Dauphine University | |
Analyzing and Improving the Image Quality of StyleGAN | 3101 | Aalto University, NVIDIA | |
EfficientDet: Scalable and Efficient Object Detection | 3081 | ||
Advances and Open Problems in Federated Learning | 2921 | Australian National University, Carnegie Mellon University, Cornell University, Emory University, École Polytechnique Fédérale de Lausanne, Georgia Institute of Technology, Google, Hong Kong University of Science and Technology, INRIA, IT University of Copenhagen, MIT, |
We can make this file beautiful and searchable if this error is corrected: It looks like row 6 should actually have 6 columns, instead of 4 in line 5.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Title Tweets Citations Organization Country Org Type | |
Highly accurate protein structure prediction with AlphaFold 8783 DeepMind, Seoul National University South Korea, UK industry | |
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows 383 5389 Microsoft USA industry | |
Learning Transferable Visual Models From Natural Language Supervision 178 3658 OpenAI USA industry | |
Accurate prediction of protein structures and interactions using a three-track neural network 1659 Harvard University, Lawrence Berkeley National Laboratory, North-West University, Stanford University, UC Berkeley, University of British Columbia, University of Cambridge, University of Graz, University of Texas Southwestern Medical Center, University of Victoria, University of Washington, University of the Free State Austria, Canada, South Africa, UK, USA | |
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions 69 1306 Inception Institute of AI, Nanjing University, Nanjing University of Science and |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Title | Tweets | Citations | Organization | Country | Org Type | |
---|---|---|---|---|---|---|
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale | 142 | 12042 | USA | industry | ||
A Simple Framework for Contrastive Learning of Visual Representations | 16 | 8476 | USA | industry | ||
Language Models are Few-Shot Learners | 331 | 7903 | OpenAI | USA | industry | |
YOLOv4: Optimal Speed and Accuracy of Object Detection | 20 | 7860 | Academia Sinica | Taiwan | industry | |
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. | 53 | 6362 | USA | industry | ||
Momentum Contrast for Unsupervised Visual Representation Learning | 8 | 6060 | Meta | USA | industry | |
End-to-End Object Detection with Transformers | 43 | 4998 | Meta, Paris Dauphine University | France, USA | industry | |
Analyzing and Improving the Image Quality of StyleGAN | 44 | 3101 | Aalto University, NVIDIA | Finland, USA | industry | |
EfficientDet: Scalable and Efficient Object Detection | 7 | 3081 | USA | industry |
We can make this file beautiful and searchable if this error is corrected: It looks like row 9 should actually have 6 columns, instead of 3 in line 8.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Title Tweets Citations Organization Country Org Type | |
AlphaFold Protein Structure Database: massively expanding the structural coverage of protein-sequence space with high-accuracy models 1331 DeepMind, European Molecular Biology Laboratory UK academia | |
ColabFold: making protein folding accessible to all 1138 Harvard University, Max Planck Institute for Multidisciplinary Sciences, Michigan State University, Seoul National University, University of Tokyo Germany, Japan, South Korea, USA academia | |
A ConvNet for the 2020s 857 835 Meta, UC Berkeley USA industry | |
Hierarchical Text-Conditional Image Generation with CLIP Latents 105 718 OpenAI USA industry | |
PaLM: Scaling Language Modeling with Pathways 445 426 Google USA industry | |
Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding 2462 390 Google USA industry | |
Instant Neural Graphics Primitives with a Multiresolution Hash Encoding 11 342 NVIDIA USA industry | |
SignalP 6.0 predicts all five types of signal peptides using protein language models 2 |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Citations | Title | Authors | |
---|---|---|---|
13 | Augmented Sliced Wasserstein Distances | Xiongjie Chen et al. | |
9 | Bayesian Neural Network Priors Revisited | Vincent Fortuin et al. | |
6 | Finetuned Language Models are Zero-Shot Learners | Jason Wei et al. | |
5 | SimVLM: Simple Visual Language Model Pretraining with Weak Supervision | Zirui Wang et al. | |
4 | Exploring the Limits of Large Scale Pre-training | Samira Abnar et al. | |
4 | LoRA: Low-Rank Adaptation of Large Language Models | Edward J Hu et al. | |
4 | Iterative Refinement Graph Neural Network for Antibody Sequence-Structure Co-design | Wengong Jin et al. | |
4 | Geometry-Consistent Neural Shape Representation with Implicit Displacement Fields | Wang Yifan et al. | |
4 | Equivalent Convex Optimization Models and Implicit Regularization | Tolga Ergen et al. |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Cited | Title | Authors | |
---|---|---|---|
121 | Unsupervised data augmentation for consistency training | Qizhe Xie et al. | |
95 | Fixmatch: Simplifying semi-supervised learning with consistency and confidence | Kihyuk Sohn et al. | |
77 | Language Models are Few-Shot Learners | Tom B. Brown et al. | |
55 | On adaptive attacks to adversarial example defenses | Ekin D. Cubuk et al. | |
54 | Randaugment: Practical automated data augmentation with a reduced search space | Florian Tramèr et al. | |
46 | What makes for good views for contrastive learning | Yonglong Tian et al. | |
46 | Debiased Contrastive Learning | Ching-Yao Chuang et al. | |
44 | Big Self-Supervised Models are Strong Semi-Supervised Learners | Ting Chen et al. | |
37 | Unsupervised Learning of Visual Features by Contrasting Cluster Assignments | Mathilde Caron et al. |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
{ | |
"embeddings": [ | |
{ | |
"tensorName": "Contextual-263-278-187", | |
"tensorShape": [ | |
51, | |
768 | |
], | |
"tensorPath": "https://gist.githubusercontent.com/sergicastellasape/aac0b1877636492c04229bc2b2ddd5ab", | |
"metadataPath": "https://gist.githubusercontent.com/sergicastellasape/e3860a436d5998a90adcf981b32413a0" |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
token | sentence_idx | |
---|---|---|
[CLS] | 263 | |
which | 263 | |
are | 263 | |
the | 263 | |
best | 263 | |
image | 263 | |
editing | 263 | |
companies | 263 | |
that | 263 |
We can't make this file beautiful and searchable because it's too large.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
-1.025043964385986328e+00 7.829327881336212158e-02 4.816381931304931641e-01 -1.908442527055740356e-01 2.205026745796203613e-01 -1.147374659776687622e-01 1.540432274341583252e-01 -1.012257695198059082e+00 3.716341778635978699e-02 3.808288276195526123e-01 -6.373665332794189453e-01 4.972777888178825378e-02 -6.241744160652160645e-01 -3.645479977130889893e-01 -5.662599802017211914e-01 -4.851117134094238281e-01 -2.727096557617187500e+00 1.045854762196540833e-02 1.004694581031799316e+00 -1.256615877151489258e+00 9.553827643394470215e-01 -2.682465314865112305e-01 -9.519978165626525879e-01 -4.256964325904846191e-01 1.507839322090148926e+00 1.278128147125244141e+00 -2.082836180925369263e-01 -2.918414175510406494e-01 -9.696484804153442383e-01 -4.932737648487091064e-01 -4.334076941013336182e-01 -3.406805917620658875e-02 -7.126870751380920410e-01 -9.347224831581115723e-01 -4.260569810867309570e-01 5.294006317853927612e-02 2.762570083141326904e-01 -5.656169056892395020e-01 8.140192031860351562e-01 9.236034750938415527e-01 |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
token | sentence_idx | |
---|---|---|
[CLS] | 263 | |
which | 263 | |
are | 263 | |
the | 263 | |
best | 263 | |
image | 263 | |
editing | 263 | |
companies | 263 | |
that | 263 |