Spaces:
Sleeping
Sleeping
Update text and title
Browse files- hexviz/app.py +12 -2
hexviz/app.py
CHANGED
@@ -5,12 +5,18 @@ from stmol import showmol
|
|
5 |
|
6 |
from hexviz.attention import Model, ModelType, get_attention_pairs
|
7 |
|
8 |
-
st.title("
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
|
10 |
# Define list of model types
|
11 |
models = [
|
12 |
-
Model(name=ModelType.ZymCTRL, layers=36, heads=16),
|
13 |
Model(name=ModelType.TAPE_BERT, layers=12, heads=12),
|
|
|
14 |
# Model(name=ModelType.PROT_T5, layers=24, heads=32),
|
15 |
]
|
16 |
|
@@ -42,3 +48,7 @@ def get_3dview(pdb):
|
|
42 |
|
43 |
xyzview = get_3dview(pdb_id)
|
44 |
showmol(xyzview, height=500, width=800)
|
|
|
|
|
|
|
|
|
|
5 |
|
6 |
from hexviz.attention import Model, ModelType, get_attention_pairs
|
7 |
|
8 |
+
st.title("Attention Visualization on proteins")
|
9 |
+
|
10 |
+
"""
|
11 |
+
Visualize attention weights on protein structures for the protein language models ZymCTRL and TAPE-BERT.
|
12 |
+
Pick a PDB ID, layer and head to visualize attention.
|
13 |
+
"""
|
14 |
+
|
15 |
|
16 |
# Define list of model types
|
17 |
models = [
|
|
|
18 |
Model(name=ModelType.TAPE_BERT, layers=12, heads=12),
|
19 |
+
Model(name=ModelType.ZymCTRL, layers=36, heads=16),
|
20 |
# Model(name=ModelType.PROT_T5, layers=24, heads=32),
|
21 |
]
|
22 |
|
|
|
48 |
|
49 |
xyzview = get_3dview(pdb_id)
|
50 |
showmol(xyzview, height=500, width=800)
|
51 |
+
|
52 |
+
"""
|
53 |
+
More models will be added soon.
|
54 |
+
"""
|