Spaces:
Sleeping
Sleeping
Manus AI
Fix: Final runtime errors (backtest_output NameError and TA-Lib TSI attribute error).
a743675 | import pandas as pd | |
| import numpy as np | |
| import yfinance as yf | |
| import os | |
| import finnhub | |
| from twelvedata import TDClient | |
| try: | |
| import talib as ta | |
| except ImportError: | |
| ta = None | |
| from datetime import datetime, timedelta | |
| from newsapi import NewsApiClient | |
| from vaderSentiment.vaderSentiment import SentimentIntensityAnalyzer | |
| from sklearn.preprocessing import MinMaxScaler | |
| from alpha_vantage.timeseries import TimeSeries | |
| import time | |
| import logging | |
| import requests | |
| # Configure logging | |
| logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s') | |
| def print_log(message, level='INFO'): | |
| if level == 'INFO': | |
| logging.info(message) | |
| elif level == 'WARNING': | |
| logging.warning(message) | |
| elif level == 'ERROR': | |
| logging.error(message) | |
| else: | |
| logging.debug(message) | |
| analyzer = SentimentIntensityAnalyzer() | |
| def load_data_finnhub(ticker, start, end, interval, api_key): | |
| print_log(f"Fetching data for {ticker} from Finnhub") | |
| try: | |
| if not api_key: | |
| raise ValueError("Finnhub API key is required.") | |
| finnhub_client = finnhub.Client(api_key=api_key) | |
| start_ts = int(time.mktime(pd.to_datetime(start).timetuple())) | |
| end_ts = int(time.mktime(pd.to_datetime(end).timetuple())) | |
| # Finnhub interval mapping | |
| finnhub_interval_map = { | |
| "1m": "1", "5m": "5", "15m": "15", "30m": "30", "60m": "60", | |
| "1h": "60", "1d": "D", "1wk": "W", "1mo": "M" | |
| } | |
| fh_interval = finnhub_interval_map.get(interval, "D") | |
| res = finnhub_client.stock_candles(ticker, fh_interval, start_ts, end_ts) | |
| if res['s'] == 'no_data': | |
| raise ValueError(f"No data for {ticker} from Finnhub") | |
| df = pd.DataFrame(res) | |
| df['Date'] = pd.to_datetime(df['t'], unit='s') | |
| df = df.rename(columns={'o': 'Open', 'h': 'High', 'l': 'Low', 'c': 'value', 'v': 'Volume'}) | |
| df = df.set_index('Date') | |
| df = df[['Open', 'High', 'Low', 'value', 'Volume']] | |
| return df | |
| except Exception as e: | |
| print_log(f"Error in load_data_finnhub for {ticker}: {str(e)}", 'ERROR') | |
| raise ValueError(f"Failed to load data for {ticker} from Finnhub: {str(e)}") | |
| def load_data_twelvedata(ticker, start, end, interval, api_key): | |
| print_log(f"Fetching data for {ticker} from Twelve Data") | |
| try: | |
| if not api_key: | |
| raise ValueError("Twelve Data API key is required.") | |
| td = TDClient(apikey=api_key) | |
| # Twelve Data interval mapping | |
| twelvedata_interval_map = { | |
| "1m": "1min", "5m": "5min", "15m": "15min", "30m": "30min", "60m": "1h", | |
| "1h": "1h", "1d": "1day", "1wk": "1week", "1mo": "1month" | |
| } | |
| td_interval = twelvedata_interval_map.get(interval, "1day") | |
| ts = td.time_series(symbol=ticker, interval=td_interval, start_date=start, end_date=end, outputsize=5000) | |
| df = ts.as_pandas() | |
| if df is None or df.empty: | |
| raise ValueError(f"No data for {ticker} from Twelve Data") | |
| df = df.rename(columns={'close': 'value', 'open': 'Open', 'high': 'High', 'low': 'Low', 'volume': 'Volume'}) | |
| df.index.name = 'Date' | |
| return df | |
| except Exception as e: | |
| print_log(f"Error in load_data_twelvedata for {ticker}: {str(e)}", 'ERROR') | |
| raise ValueError(f"Failed to load data for {ticker} from Twelve Data: {str(e)}") | |
| def load_data(data_src='yahoo', ticker='AAPL', start='2020-01-01', end='2023-01-01', interval='1d', file_upload=None, alpha_api_key=None, finnhub_api_key=None, twelvedata_api_key=None): | |
| print_log(f"Loading data: source={data_src}, ticker={ticker}, start={start}, end={end}, interval={interval}, file_upload={'set' if file_upload else 'unset'}, alpha_api_key={'set' if alpha_api_key else 'unset'}, finnhub_api_key={'set' if finnhub_api_key else 'unset'}, twelvedata_api_key={'set' if twelvedata_api_key else 'unset'}") | |
| start_date = pd.to_datetime(start) | |
| end_date = pd.to_datetime(end) | |
| if start_date >= end_date: | |
| raise ValueError(f"Start date {start} must be before end date {end}") | |
| if end_date > datetime.now(): | |
| print_log(f"End date {end} is in the future. Using current date as end date.", 'WARNING') | |
| end_date = datetime.now() | |
| df = pd.DataFrame() | |
| if data_src == 'csv' and file_upload: | |
| try: | |
| file_path = getattr(file_upload, 'name', file_upload) | |
| print_log(f"Loading CSV from {file_path}") | |
| df = pd.read_csv(file_path) | |
| if 'Date' not in df.columns: | |
| raise ValueError("CSV must contain a 'Date' column") | |
| df['Date'] = pd.to_datetime(df['Date']).dt.tz_localize(None) | |
| df = df.set_index('Date') | |
| if 'Close' not in df.columns and 'value' not in df.columns: | |
| raise ValueError("CSV must contain 'Close' or 'value' column") | |
| if 'Close' in df.columns: | |
| df = df.rename(columns={'Close': 'value'}) | |
| if df.empty: | |
| raise ValueError(f"CSV data is empty for {ticker}") | |
| if df['value'].isna().all(): | |
| raise ValueError(f"CSV 'value' column contains only NaNs for {ticker}") | |
| except Exception as e: | |
| print_log(f"Failed to load CSV {file_path}: {str(e)}", 'ERROR') | |
| raise ValueError(f"Failed to load CSV: {str(e)}") | |
| elif data_src == 'yahoo': | |
| print_log(f"Fetching data for {ticker} from Yahoo Finance") | |
| try: | |
| # Adjust start_date for yfinance intraday limitations | |
| if interval in ["1m"]: | |
| max_days = 7 | |
| elif interval in ["2m", "5m", "15m", "30m", "60m", "90m", "1h"]: | |
| max_days = 60 | |
| else: | |
| max_days = None | |
| if max_days: | |
| adjusted_end_date = end_date + timedelta(days=1) # yfinance end date is exclusive | |
| adjusted_start_date = adjusted_end_date - timedelta(days=max_days) | |
| if start_date < adjusted_start_date: | |
| print_log(f"Adjusting start date for {interval} interval from {start_date.strftime('%Y-%m-%d')} to {adjusted_start_date.strftime('%Y-%m-%d')} due to yfinance limitations.", 'WARNING') | |
| start_date = adjusted_start_date | |
| df = yf.download(ticker, start=start_date, end=end_date, interval=interval, progress=False, auto_adjust=False) | |
| if isinstance(df.columns, pd.MultiIndex): | |
| df.columns = df.columns.droplevel(1) | |
| if df.empty: | |
| raise ValueError(f"No data returned from Yahoo Finance for {ticker}") | |
| if 'Close' not in df.columns: | |
| raise ValueError(f"Yahoo Finance data missing 'Close' column for {ticker}") | |
| df = df.rename(columns={'Close': 'value'}) | |
| # The index is already datetime, no need to create a 'Date' column and then reset | |
| if df['value'].isna().all(): | |
| raise ValueError(f"Yahoo Finance 'value' column contains only NaNs for {ticker}") | |
| if df['value'].empty: | |
| raise ValueError(f"Yahoo Finance 'value' column is empty for {ticker}") | |
| except Exception as e: | |
| print_log(f"Yahoo Finance failed for {ticker}: {str(e)}", 'ERROR') | |
| raise ValueError(f"Yahoo Finance failed for {ticker}: {str(e)}") | |
| elif data_src == 'alpha_vantage' and alpha_api_key: | |
| print_log(f"Attempting Alpha Vantage for {ticker}, interval {interval}") | |
| try: | |
| ts = TimeSeries(key=alpha_api_key, output_format='pandas') | |
| # Alpha Vantage interval mapping | |
| av_interval_map = { | |
| "1m": "1min", "5m": "5min", "15m": "15min", "30m": "30min", "60m": "60min", | |
| "1h": "60min" | |
| } | |
| av_interval = av_interval_map.get(interval) | |
| if av_interval: | |
| df_av, _ = ts.get_intraday(symbol=ticker, interval=av_interval, outputsize='full') | |
| elif interval == "1d": | |
| df_av, _ = ts.get_daily(symbol=ticker, outputsize='full') | |
| elif interval == "1wk": | |
| df_av, _ = ts.get_weekly(symbol=ticker) | |
| elif interval == "1mo": | |
| df_av, _ = ts.get_monthly(symbol=ticker) | |
| else: | |
| raise ValueError(f"Unsupported interval for Alpha Vantage: {interval}") | |
| if df_av.empty: | |
| raise ValueError(f"No data returned from Alpha Vantage for {ticker}") | |
| # Standardize column names | |
| df_av = df_av.rename(columns={ | |
| '4. close': 'value', '1. open': 'Open', '2. high': 'High', | |
| '3. low': 'Low', '5. volume': 'Volume' | |
| }) | |
| # For daily/weekly/monthly, index is already datetime. For intraday, it's also datetime. | |
| # Ensure consistent column order and index type | |
| df = df_av[['Open', 'High', 'Low', 'value', 'Volume']] | |
| df.index = pd.to_datetime(df.index) | |
| df = df.sort_index() | |
| if df['value'].isna().all(): | |
| raise ValueError(f"Alpha Vantage 'value' column contains only NaNs for {ticker}") | |
| if df['value'].empty: | |
| raise ValueError(f"Alpha Vantage 'value' column is empty for {ticker}") | |
| print_log(f"Data loaded for {ticker} from Alpha Vantage with date range: {df.index.min()} to {df.index.max()}, shape: {df.shape}") | |
| except Exception as e: | |
| print_log(f"Alpha Vantage failed for {ticker}: {str(e)}", 'ERROR') | |
| raise ValueError(f"Alpha Vantage failed for {ticker}: {str(e)}") | |
| elif data_src == 'finnhub' and finnhub_api_key: | |
| df = load_data_finnhub(ticker, start, end, interval, finnhub_api_key) | |
| elif data_src == 'twelvedata' and twelvedata_api_key: | |
| df = load_data_twelvedata(ticker, start, end, interval, twelvedata_api_key) | |
| if df.empty: | |
| raise ValueError(f"No data loaded for {ticker} from {data_src}") | |
| # Ensure index is DatetimeIndex and sorted | |
| if not isinstance(df.index, pd.DatetimeIndex): | |
| df.index = pd.to_datetime(df.index) | |
| df = df.sort_index() | |
| required_cols = ['Open', 'High', 'Low', 'value', 'Volume'] | |
| for col in required_cols: | |
| if col not in df.columns: | |
| df[col] = np.nan | |
| if 'value' not in df.columns: | |
| raise ValueError(f"Target column 'value' is missing for {ticker}") | |
| if df['value'].isna().all(): | |
| raise ValueError(f"Target column 'value' contains only NaNs for {ticker}") | |
| if df['value'].empty: | |
| raise ValueError(f"Target column 'value' is empty for {ticker}") | |
| print_log(f"Data loaded for {ticker} with date range: {df.index.min()} to {df.index.max()}, shape: {df.shape}") | |
| return df | |
| def add_technical_indicators(df, selected_indicators): | |
| try: | |
| print_log(f"Starting add_technical_indicators with indicators: {selected_indicators}") | |
| if df.empty: | |
| print_log("DataFrame is empty, skipping technical indicator calculation.", "WARNING") | |
| return df, [] | |
| for col in ['Open', 'High', 'Low', 'value', 'Volume']: | |
| if col not in df.columns: | |
| df[col] = np.nan | |
| df[col] = pd.to_numeric(df[col], errors='coerce') | |
| df.dropna(subset=['Open', 'High', 'Low', 'value', 'Volume'], inplace=True) | |
| if df.empty: | |
| print_log("DataFrame is empty after dropping NaNs for technical indicators.", "WARNING") | |
| return df, [] | |
| if ta is None: | |
| print_log("TA-Lib not available. Cannot compute indicators. Falling back to 'value'.", 'ERROR') | |
| return df, [] | |
| close = df['value'].values | |
| high = df['High'].values | |
| low = df['Low'].values | |
| volume = df['Volume'].values | |
| open_ = df['Open'].values | |
| indicator_map = { | |
| 'rsi': {'func': ta.RSI, 'inputs': ['close'], 'params': {'timeperiod': 14}, 'output': ['rsi_14']}, | |
| 'macd': {'func': ta.MACD, 'inputs': ['close'], 'params': {'fastperiod': 12, 'slowperiod': 26, 'signalperiod': 9}, 'output': ['macd_12_26_9', 'macds_12_26_9', 'macdhist_12_26_9']}, | |
| 'bbands': {'func': ta.BBANDS, 'inputs': ['close'], 'params': {'timeperiod': 20}, 'output': ['upperband_20', 'middleband_20', 'lowerband_20']}, | |
| 'stoch': {'func': ta.STOCH, 'inputs': ['high', 'low', 'close'], 'params': {'fastk_period': 14, 'slowk_period': 3, 'slowd_period': 3}, 'output': ['slowk_14_3_3', 'slowd_14_3_3']}, | |
| 'adx': {'func': ta.ADX, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['adx_14']}, | |
| 'atr': {'func': ta.ATR, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['atr_14']}, | |
| 'cci': {'func': ta.CCI, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['cci_14']}, | |
| 'ema': {'func': ta.EMA, 'inputs': ['close'], 'params': {'timeperiod': 14}, 'output': ['ema_14']}, | |
| 'sma': {'func': ta.SMA, 'inputs': ['close'], 'params': {'timeperiod': 14}, 'output': ['sma_14']}, | |
| 'mom': {'func': ta.MOM, 'inputs': ['close'], 'params': {'timeperiod': 10}, 'output': ['mom_10']}, | |
| 'roc': {'func': ta.ROC, 'inputs': ['close'], 'params': {'timeperiod': 10}, 'output': ['roc_10']}, | |
| 'willr': {'func': ta.WILLR, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['willr_14']}, | |
| 'ultosc': {'func': ta.ULTOSC, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod1': 7, 'timeperiod2': 14, 'timeperiod3': 28}, 'output': ['ultosc_7_14_28']}, | |
| 'dx': {'func': ta.DX, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['dx_14']}, | |
| 'minus_di': {'func': ta.MINUS_DI, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['minus_di_14']}, | |
| 'plus_di': {'func': ta.PLUS_DI, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['plus_di_14']}, | |
| 'mfi': {'func': ta.MFI, 'inputs': ['high', 'low', 'close', 'volume'], 'params': {'timeperiod': 14}, 'output': ['mfi_14']}, | |
| 'obv': {'func': ta.OBV, 'inputs': ['close', 'volume'], 'params': {}, 'output': ['obv']}, | |
| 'ad': {'func': ta.AD, 'inputs': ['high', 'low', 'close', 'volume'], 'params': {}, 'output': ['ad']}, | |
| 'adosc': {'func': ta.ADOSC, 'inputs': ['high', 'low', 'close', 'volume'], 'params': {'fastperiod': 3, 'slowperiod': 10}, 'output': ['adosc_3_10']}, | |
| 'aroon': {'func': ta.AROON, 'inputs': ['high', 'low'], 'params': {'timeperiod': 14}, 'output': ['aroon_down_14', 'aroon_up_14']}, | |
| 'aroonosc': {'func': ta.AROONOSC, 'inputs': ['high', 'low'], 'params': {'timeperiod': 14}, 'output': ['aroonosc_14']}, | |
| 'bop': {'func': ta.BOP, 'inputs': ['open_', 'high', 'low', 'close'], 'params': {}, 'output': ['bop']}, | |
| 'cmo': {'func': ta.CMO, 'inputs': ['close'], 'params': {'timeperiod': 14}, 'output': ['cmo_14']}, | |
| 'dema': {'func': ta.DEMA, 'inputs': ['close'], 'params': {'timeperiod': 30}, 'output': ['dema_30']}, | |
| 'kama': {'func': ta.KAMA, 'inputs': ['close'], 'params': {'timeperiod': 30}, 'output': ['kama_30']}, | |
| 'ppo': {'func': ta.PPO, 'inputs': ['close'], 'params': {'fastperiod': 12, 'slowperiod': 26, 'matype': 0}, 'output': ['ppo_12_26_0']}, | |
| 'rocp': {'func': ta.ROCP, 'inputs': ['close'], 'params': {'timeperiod': 10}, 'output': ['rocp_10']}, | |
| 'rocr': {'func': ta.ROCR, 'inputs': ['close'], 'params': {'timeperiod': 10}, 'output': ['rocr_10']}, | |
| 'rocr100': {'func': ta.ROCR100, 'inputs': ['close'], 'params': {'timeperiod': 10}, 'output': ['rocr100_10']}, | |
| 'trix': {'func': ta.TRIX, 'inputs': ['close'], 'params': {'timeperiod': 14}, 'output': ['trix_14']}, | |
| # 'tsi': {'func': ta.TSI, 'inputs': ['close'], 'params': {'fastperiod': 13, 'slowperiod': 25}, 'output': ['tsi_13_25']}, # Removed due to TA-Lib attribute error | |
| 'uo': {'func': ta.ULTOSC, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod1': 7, 'timeperiod2': 14, 'timeperiod3': 28}, 'output': ['ultosc_7_14_28']}, | |
| 'willr': {'func': ta.WILLR, 'inputs': ['high', 'low', 'close'], 'params': {'timeperiod': 14}, 'output': ['willr_14']}, | |
| 'wma': {'func': ta.WMA, 'inputs': ['close'], 'params': {'timeperiod': 30}, 'output': ['wma_30']}, | |
| } | |
| added_features = [] | |
| for indicator_name in selected_indicators: | |
| if indicator_name in indicator_map: | |
| indicator_info = indicator_map[indicator_name] | |
| func = indicator_info['func'] | |
| inputs = [] | |
| for input_name in indicator_info['inputs']: | |
| if input_name == 'close': | |
| inputs.append(close) | |
| elif input_name == 'high': | |
| inputs.append(high) | |
| elif input_name == 'low': | |
| inputs.append(low) | |
| elif input_name == 'volume': | |
| inputs.append(volume) | |
| elif input_name == 'open_': | |
| inputs.append(open_) | |
| # Ensure inputs are numpy arrays and not empty | |
| if not all(len(arr) > 0 for arr in inputs): | |
| print_log(f"Skipping {indicator_name}: insufficient data for inputs.", "WARNING") | |
| continue | |
| try: | |
| output_values = func(*inputs, **indicator_info['params']) | |
| if not isinstance(output_values, tuple): | |
| output_values = (output_values,) | |
| for i, col_name in enumerate(indicator_info['output']): | |
| df[col_name] = np.nan | |
| # Ensure the output array matches the DataFrame length | |
| if len(output_values[i]) == len(df): | |
| df[col_name] = output_values[i] | |
| else: | |
| # Align output to DataFrame by padding with NaNs at the beginning | |
| nan_padding = np.full(len(df) - len(output_values[i]), np.nan) | |
| df[col_name] = np.concatenate((nan_padding, output_values[i])) | |
| added_features.append(col_name) | |
| except Exception as e: | |
| print_log(f"Error calculating indicator {indicator_name}: {str(e)}", "ERROR") | |
| else: | |
| print_log(f"Unknown indicator: {indicator_name}", "WARNING") | |
| df.dropna(inplace=True) | |
| print_log(f"Finished add_technical_indicators. New features added: {added_features}") | |
| return df, added_features | |
| except Exception as e: | |
| print_log(f"Error in add_technical_indicators: {str(e)}", 'ERROR') | |
| return df, [] | |
| def add_sentiment(df, ticker, news_api_key, start_date, end_date): | |
| try: | |
| sentiment_text, sentiment_score = sentiment_analysis(ticker, start_date, end_date, news_api_key) | |
| df['sentiment'] = sentiment_score if sentiment_score is not None else 0.0 | |
| return df, sentiment_text | |
| except Exception as e: | |
| print_log(f"Error adding sentiment: {str(e)}", 'ERROR') | |
| df['sentiment'] = 0.0 # Default to neutral sentiment on error | |
| return df, f"Error adding sentiment: {str(e)}" | |
| def sentiment_analysis(ticker, start_date, end_date, api_key): | |
| try: | |
| if not api_key: | |
| print_log("News API key not provided for sentiment analysis.", 'WARNING') | |
| return "No API key provided", None | |
| newsapi = NewsApiClient(api_key=api_key) | |
| start = pd.to_datetime(start_date) | |
| end = pd.to_datetime(end_date) | |
| articles = newsapi.get_everything( | |
| q=ticker, from_param=start.strftime("%Y-%m-%d"), to=end.strftime("%Y-%m-%d"), | |
| language='en', sort_by='relevancy' | |
| ) | |
| sentiments = [analyzer.polarity_scores(article["title"])['compound'] for article in articles["articles"] if "title" in article and article["title"] is not None] | |
| avg_sentiment = np.mean(sentiments) if sentiments else 0.0 | |
| sentiment_text = f"Average sentiment for {ticker}: {avg_sentiment:.2f}" | |
| return sentiment_text, avg_sentiment | |
| except Exception as e: | |
| print_log(f"Sentiment analysis failed: {str(e)}", 'ERROR') | |
| return f"Sentiment analysis failed: {str(e)}", None | |
| def preprocess_data(df, features, target, window_size, horizon): | |
| try: | |
| print_log(f"Starting preprocessing: features={features}, target={target}, window={window_size}, horizon={horizon}") | |
| if not isinstance(df.index, pd.DatetimeIndex): | |
| raise ValueError("DataFrame index must be a DatetimeIndex for preprocessing.") | |
| # The target column must be included in the features for scaling | |
| all_features = list(set(features + [target])) | |
| updated_feature_cols = [f for f in all_features if f in df.columns] | |
| if not updated_feature_cols: | |
| raise ValueError("None of the selected features are available in the data.") | |
| data = df[updated_feature_cols].values | |
| # Scale all features | |
| feature_scaler = MinMaxScaler(feature_range=(0, 1)) | |
| scaled_data = feature_scaler.fit_transform(data) | |
| # Scale the target column separately for inverse transform | |
| target_scaler = MinMaxScaler(feature_range=(0, 1)) | |
| target_scaler.fit(df[[target]].values) | |
| # Get the index of the target column in the scaled data | |
| target_idx = updated_feature_cols.index(target) | |
| X, y = [], [] | |
| for i in range(len(scaled_data) - window_size - horizon + 1): | |
| X.append(scaled_data[i:(i + window_size), :]) | |
| # The target is the 'value' at the end of the window + horizon | |
| y.append(scaled_data[i + window_size + horizon - 1, target_idx]) | |
| X, y = np.array(X), np.array(y) | |
| if X.shape[1] != window_size or X.shape[2] != len(updated_feature_cols): | |
| raise ValueError(f"Shape mismatch in X: expected ({len(scaled_data) - window_size - horizon + 1}, {window_size}, {len(updated_feature_cols)}), got {X.shape}") | |
| print_log(f"Preprocessing complete. X shape: {X.shape}, y shape: {y.shape}") | |
| return X, y, feature_scaler, target_scaler, updated_feature_cols, target_idx | |
| except Exception as e: | |
| print_log(f"Error in preprocess_data: {str(e)}", 'ERROR') | |
| raise ValueError(f"Failed to preprocess data: {str(e)}") | |