Update Home.py
Browse files
Home.py
CHANGED
|
@@ -27,17 +27,6 @@ df = connect_gsheet()
|
|
| 27 |
|
| 28 |
st.write(df)
|
| 29 |
|
| 30 |
-
# image dictionary
|
| 31 |
-
image_dict = {
|
| 32 |
-
"Atom Indonesia": "https://github.com/faizhalas/Search4All/blob/main/images/journal/atom.png?raw=true",
|
| 33 |
-
"EKSPLORIUM": "https://github.com/faizhalas/Search4All/blob/main/images/journal/eksplorium.jpg?raw=true",
|
| 34 |
-
"GANENDRA": "https://github.com/faizhalas/Search4All/blob/main/images/journal/ganendra.png?raw=true",
|
| 35 |
-
"Jurnal Pengembangan Energi Nuklir": "https://github.com/faizhalas/Search4All/blob/main/images/journal/jpen.jpg?raw=true",
|
| 36 |
-
"Jurnal Sains dan Teknologi Nuklir Indonesia": "https://github.com/faizhalas/Search4All/blob/main/images/journal/jstni.jpg?raw=true",
|
| 37 |
-
"Jurnal Teknologi Reaktor Nuklir Tri Dasa Mega": "https://github.com/faizhalas/Search4All/blob/main/images/journal/tridasa.jpg?raw=true",
|
| 38 |
-
"URANIA": "https://github.com/faizhalas/Search4All/blob/main/images/journal/urania.jpg?raw=true"
|
| 39 |
-
}
|
| 40 |
-
|
| 41 |
#Title
|
| 42 |
st.title('PerpusDB: Article')
|
| 43 |
|
|
@@ -73,14 +62,12 @@ if keyword_list_j is not None:
|
|
| 73 |
column_result = df[np.logical_and.reduce(conditions)]
|
| 74 |
key_df_j = pd.concat([key_df_j, column_result]).drop_duplicates()
|
| 75 |
|
| 76 |
-
st.write(key_df_j)
|
| 77 |
-
|
| 78 |
# creating result
|
| 79 |
if text_search:
|
| 80 |
st.caption(f"We found **{key_df_j.shape[0]}** article(s) available for you.")
|
| 81 |
|
| 82 |
for _, row in key_df_j.reset_index().iterrows():
|
| 83 |
-
with st.container():
|
| 84 |
st.caption(f"{row['Year'].strip()} - {row['Volume'].strip()}, {row['Issue'].strip()}")
|
| 85 |
st.markdown(f"**{row['Authors'].strip()}**")
|
| 86 |
st.markdown(f"**Title:** *{row['Title'].strip()}*")
|
|
|
|
| 27 |
|
| 28 |
st.write(df)
|
| 29 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
#Title
|
| 31 |
st.title('PerpusDB: Article')
|
| 32 |
|
|
|
|
| 62 |
column_result = df[np.logical_and.reduce(conditions)]
|
| 63 |
key_df_j = pd.concat([key_df_j, column_result]).drop_duplicates()
|
| 64 |
|
|
|
|
|
|
|
| 65 |
# creating result
|
| 66 |
if text_search:
|
| 67 |
st.caption(f"We found **{key_df_j.shape[0]}** article(s) available for you.")
|
| 68 |
|
| 69 |
for _, row in key_df_j.reset_index().iterrows():
|
| 70 |
+
with st.container(border=True):
|
| 71 |
st.caption(f"{row['Year'].strip()} - {row['Volume'].strip()}, {row['Issue'].strip()}")
|
| 72 |
st.markdown(f"**{row['Authors'].strip()}**")
|
| 73 |
st.markdown(f"**Title:** *{row['Title'].strip()}*")
|