Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -952,8 +952,14 @@ if menu == "📊 So sánh model":
|
|
| 952 |
df_sym = df_train[df_train["symbol"] == symbol]
|
| 953 |
|
| 954 |
series = df_sym["close"].astype(float).values
|
| 955 |
-
|
| 956 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 957 |
|
| 958 |
with torch.no_grad():
|
| 959 |
outputs = model(
|
|
@@ -978,6 +984,7 @@ if menu == "📊 So sánh model":
|
|
| 978 |
"upper_95": hi
|
| 979 |
})
|
| 980 |
|
|
|
|
| 981 |
def forecast_timegpt(symbol, horizon):
|
| 982 |
client = load_timegpt()
|
| 983 |
df_sym = df_train[df_train["symbol"] == symbol].copy()
|
|
@@ -1791,8 +1798,14 @@ if menu == "🧪 Demo sản phẩm":
|
|
| 1791 |
df_sym = df_train[df_train["symbol"] == symbol]
|
| 1792 |
|
| 1793 |
series = df_sym["close"].astype(float).values
|
| 1794 |
-
|
| 1795 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1796 |
|
| 1797 |
with torch.no_grad():
|
| 1798 |
outputs = model(
|
|
|
|
| 952 |
df_sym = df_train[df_train["symbol"] == symbol]
|
| 953 |
|
| 954 |
series = df_sym["close"].astype(float).values
|
| 955 |
+
|
| 956 |
+
past_tensor = (
|
| 957 |
+
torch.from_numpy(series)
|
| 958 |
+
.to(dtype=torch.bfloat16, device=model.device)
|
| 959 |
+
.unsqueeze(0)
|
| 960 |
+
)
|
| 961 |
+
|
| 962 |
+
freq_tensor = torch.tensor([0], dtype=torch.long, device=model.device)
|
| 963 |
|
| 964 |
with torch.no_grad():
|
| 965 |
outputs = model(
|
|
|
|
| 984 |
"upper_95": hi
|
| 985 |
})
|
| 986 |
|
| 987 |
+
|
| 988 |
def forecast_timegpt(symbol, horizon):
|
| 989 |
client = load_timegpt()
|
| 990 |
df_sym = df_train[df_train["symbol"] == symbol].copy()
|
|
|
|
| 1798 |
df_sym = df_train[df_train["symbol"] == symbol]
|
| 1799 |
|
| 1800 |
series = df_sym["close"].astype(float).values
|
| 1801 |
+
|
| 1802 |
+
past_tensor = (
|
| 1803 |
+
torch.from_numpy(series)
|
| 1804 |
+
.to(dtype=torch.bfloat16, device=model.device)
|
| 1805 |
+
.unsqueeze(0)
|
| 1806 |
+
)
|
| 1807 |
+
|
| 1808 |
+
freq_tensor = torch.tensor([0], dtype=torch.long, device=model.device)
|
| 1809 |
|
| 1810 |
with torch.no_grad():
|
| 1811 |
outputs = model(
|