--- title: "Copper" output: html_document date: "2025-06-23" --- ```{r setup, include=FALSE} knitr::opts_chunk$set(echo = TRUE) library(readr) library(tidyverse) library(readxl) library(ggplot2) library(dplyr) library(httr) library(jsonlite) ``` ```{r} # # res <- GET("https://eodhd.com/api/screener", # query = list( # industries = "Metals & Mining", # query = "copper", # api_token = " 5f3afd582bd7b4.95720069" # )) # # companies <- fromJSON(content(res, "text")) # # copper = companies$data ``` ```{r} api_token <- "5f3afd582bd7b4.95720069" ``` ```{r} res <- GET("https://eodhd.com/api/exchanges-list", query = list(api_token = api_token)) exchange_list <- fromJSON(content(res, "text", encoding = "UTF-8")) exchange_df <- as.data.frame(exchange_list) # Step 2: Initialize storage for results all_stocks <- list() # Step 3: Loop over exchange codes and download stocks for (exchange in exchange_df$Code) { message("Fetching data for exchange: ", exchange) res <- GET( paste0("https://eodhd.com/api/exchange-symbol-list/", exchange), query = list(api_token = api_token) ) if (status_code(res) == 200) { json_text <- content(res, "text", encoding = "UTF-8") stock_df <- tryCatch( { read.csv(text = json_text) }, error = function(e) { message("Skipping ", exchange, " due to error: ", e$message) return(NULL) } ) if (!is.null(stock_df) && nrow(stock_df) > 0) { stock_df$ExchangeCode <- exchange all_stocks[[length(all_stocks) + 1]] <- stock_df } } else { message("Failed to fetch data for ", exchange, " (", status_code(res), ")") } Sys.sleep(1) # avoid rate limit } # Step 3.5: Force all columns to character for binding all_stocks_clean <- lapply(all_stocks, function(df) { if (is.data.frame(df)) { df[] <- lapply(df, as.character) # convert every column to character return(df) } else { return(NULL) } }) all_stocks_clean <- Filter(Negate(is.null), all_stocks_clean) # remove NULLs # Step 4: Combine cleaned data frames stock_metadata_df <- dplyr::bind_rows(all_stocks_clean) # Optional: Save or view View(stock_metadata_df) # write.csv(stock_metadata_df, "global_stock_metadata.csv", row.names = FALSE) ``` # Stock MetaData ```{r} merge_fields_deep <- function(base_row, ...) { lists <- list(...) for (list_item in lists) { if (is.null(list_item)) next for (field in names(list_item)) { value <- list_item[[field]] if (is.list(value)) { value <- paste(unlist(value), collapse = ", ") } base_row[[field]] <- value } } return(base_row) } extract_institution_fields <- function(holder_list) { if (is.null(holder_list$Institutions)) return(list()) holders <- holder_list$Institutions fields <- list() for (i in seq_along(holders)) { h <- holders[[i]] prefix <- paste0("Holder_", i, "_") fields[[paste0(prefix, "name")]] <- h$name fields[[paste0(prefix, "date")]] <- h$date fields[[paste0(prefix, "totalShares")]] <- h$totalShares fields[[paste0(prefix, "currentShares")]] <- h$currentShares fields[[paste0(prefix, "change_p")]] <- h$change_p } return(fields) } extract_fund_fields <- function(holder_list) { if (is.null(holder_list$Funds)) return(list()) funds <- holder_list$Funds fields <- list() for (i in seq_along(funds)) { f <- funds[[i]] prefix <- paste0("Fund_", i, "_") fields[[paste0(prefix, "name")]] <- f$name fields[[paste0(prefix, "date")]] <- f$date fields[[paste0(prefix, "totalShares")]] <- f$totalShares fields[[paste0(prefix, "currentShares")]] <- f$currentShares fields[[paste0(prefix, "change_p")]] <- f$change_p } return(fields) } extract_dividend_fields <- function(div_list) { if (is.null(div_list) || length(div_list) == 0) return(list()) div_raw <- lapply(div_list, function(x) { if (!is.null(x$Year) && !is.null(x$Count)) { data.frame(Year = as.integer(x$Year), Count = as.integer(x$Count)) } else { NULL } }) div_df <- do.call(rbind, div_raw) if (is.null(div_df) || nrow(div_df) == 0) return(list()) # 保留近5年(最多到 2024) current_year <- as.integer(format(Sys.Date(), "%Y")) target_years <- (current_year - 4):current_year div_df <- div_df[div_df$Year %in% target_years, ] if (nrow(div_df) == 0) return(list()) # ✅ 加这一句! out <- setNames(as.list(div_df$Count), paste0("Dividend_", div_df$Year)) return(out) } # 主循环开始 stock_only <- stock_metadata_df %>% filter(Type %in% c("Common Stock", "Preferred Stock", "ETF")) #stocks_subset <- head(stock_only, 5) stocks_subset <- stock_only[94334:96494, ] %>% filter(Type == "Common Stock") #stocks_subset$Code <- sprintf("%06d", as.integer(stocks_subset$Code)) #stocks_subset$Code <- as.character(stocks_subset$Code) #stocks_subset <- tail(stocks_subset) #stocks_subset <- stock_only[50000:nrow(stock_only), ] #stocks_subset <- stock_only enriched_data <- list() for (i in 1:nrow(stocks_subset)) { symbol <- as.character(stocks_subset[i, "Code"]) exchange <- as.character(stocks_subset[i, "ExchangeCode"]) full_symbol <- paste0(symbol, ".", gsub(" ", "", exchange)) message(sprintf("🔄 Processing %d / %d: %s", i, nrow(stocks_subset), full_symbol)) url <- paste0("https://eodhd.com/api/fundamentals/", full_symbol, "?api_token=", api_token) res <- tryCatch(GET(url), error = function(e) NULL) if (!is.null(res) && status_code(res) == 200) { json_data <- tryCatch(fromJSON(content(res, "text", encoding = "UTF-8")), error = function(e) NULL) if (!is.null(json_data)) { # 提取三部分数据 inst_fields <- extract_institution_fields(json_data$Holders) fund_fields <- extract_fund_fields(json_data$Holders) dividend_data <- extract_dividend_fields(json_data$SplitsDividends$NumberDividendsByYear) merged_fields <- c( json_data$General, json_data$Highlights, json_data$Valuation, json_data$SharesStats, json_data$Technicals, json_data$SplitsDividends[names(json_data$SplitsDividends) != "NumberDividendsByYear"], dividend_data, inst_fields, fund_fields ) enriched_row <- merge_fields_deep(stocks_subset[i, ], merged_fields) enriched_data[[length(enriched_data) + 1]] <- enriched_row } else { message("⚠️ No JSON data for ", full_symbol) } } else { message("❌ Failed request for ", full_symbol) } Sys.sleep(1) # 避免过快请求 } # 合并输出结果 if (length(enriched_data) > 0) { result_df <- bind_rows(enriched_data) } else { warning("No data enriched.") result_df <- data.frame() } #View(result_df) #write.csv(result_df, "metadata(1-1w).csv", row.names = FALSE) #write.csv(result_df, "metadata(1w-2w).csv", row.names = FALSE) #write.csv(result_df, "metadata_Taiwan(94324:96485).csv", row.names = FALSE) ``` ```{r} w1 = read_csv("metadata(1-1w).csv") w2 = read_csv("metadata(1w-2w).csv") w3 = read_csv("metadata(2w-4w).csv") w4 = read_csv("metadata(4w-6w).csv") w5 = read_csv("metadata(6w-8w).csv") w6 = read_csv("metadata(8w-82080).csv") w7 = read_csv("metadata(85250-94323).csv") w8 = read_csv("metadata(96486-end).csv") w9 = read_csv("metadata_China.csv") w10 = read_csv("metadata_Taiwan.csv") x1 = w1 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x2 = w2 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x3 = w3 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x4 = w4 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x5 = w5 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x6 = w6 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x7 = w7 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x8 = w8 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x9 = w9 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) x10 = w10 %>% group_by(Industry) %>% filter(Type == "Common Stock")%>% summarize(total_count = n()) ``` ```{r} yy = rbind(x1,x2,x3,x4,x5,x6,x7,x8,x9,x10) yy = yy %>% group_by(Industry) %>% summarize(count = sum(total_count)) %>% arrange(desc(count)) write.csv(yy, "Industry_Count.csv", row.names = FALSE) ``` ```{r} x1 = w1 %>% group_by(Industry) %>% filter(Type == "Common Stock") x2 = w2 %>% group_by(Industry) %>% filter(Type == "Common Stock") x3 = w3 %>% group_by(Industry) %>% filter(Type == "Common Stock") x4 = w4 %>% group_by(Industry) %>% filter(Type == "Common Stock") x5 = w5 %>% group_by(Industry) %>% filter(Type == "Common Stock") x6 = w6 %>% group_by(Industry) %>% filter(Type == "Common Stock") x7 = w7 %>% group_by(Industry) %>% filter(Type == "Common Stock") x8 = w8 %>% group_by(Industry) %>% filter(Type == "Common Stock") x9 = w9 %>% group_by(Industry) %>% filter(Type == "Common Stock") x10 = w10 %>% group_by(Industry) %>% filter(Type == "Common Stock") # write.csv(x1, "metadata_1.csv", row.names = FALSE) # write.csv(x2, "metadata_2.csv", row.names = FALSE) # write.csv(x3, "metadata_3.csv", row.names = FALSE) # write.csv(x4, "metadata_4.csv", row.names = FALSE) # write.csv(x5, "metadata_5.csv", row.names = FALSE) # write.csv(x6, "metadata_6.csv", row.names = FALSE) # write.csv(x7, "metadata_7.csv", row.names = FALSE) # write.csv(x8, "metadata_8.csv", row.names = FALSE) # write.csv(x9, "metadata_9.csv", row.names = FALSE) # write.csv(x10, "metadata_10.csv", row.names = FALSE) ``` ## Clean Data ```{r} extract_top3_holders_clean <- function(df) { # Step 1: 找出 totalShares 列 share_cols <- grep("^Holder_\\d+_totalShares$", names(df), value = TRUE) holder_fields <- c("_name", "_date", "_totalShares", "_currentShares") # Step 2: 找出每行中最大的 3 个 holder top3_indices <- apply(df[ , share_cols, drop = FALSE], 1, function(row) { non_na <- which(!is.na(row)) if (length(non_na) == 0) return(rep(NA, 3)) top <- order(row[non_na], decreasing = TRUE)[1:min(3, length(non_na))] return(non_na[top]) }) # Step 3: 抽出列名 top3_colnames <- lapply(top3_indices, function(idxs) { if (all(is.na(idxs))) return(rep(NA, 3)) return(share_cols[idxs]) }) top3_colnames <- as.data.frame(do.call(rbind, top3_colnames), stringsAsFactors = FALSE) # Step 4: 抽取每行 holder 对应数据并重命名 row_extracts <- lapply(1:nrow(top3_colnames), function(i) { bases <- top3_colnames[i, ] if (all(is.na(bases))) { empty_names <- unlist(lapply(1:3, function(j) paste0("Holder_", j, holder_fields))) return(as.data.frame(matrix(NA, nrow = 1, ncol = length(empty_names), dimnames = list(NULL, empty_names)))) } bases <- bases[!is.na(bases)] cols <- unlist(lapply(bases, function(base) paste0(gsub("_totalShares$", "", base), holder_fields))) out <- df[i, cols, drop = FALSE] new_names <- unlist(lapply(seq_along(bases), function(j) paste0("Holder_", j, holder_fields))) names(out) <- new_names full_names <- unlist(lapply(1:3, function(j) paste0("Holder_", j, holder_fields))) for (nm in setdiff(full_names, names(out))) { out[[nm]] <- NA } out <- out[ , full_names] return(out) }) top3_df <- do.call(rbind, row_extracts) # Step 5: 删除原始 holder/fund 列并合并 df <- df[ , !grepl("^Holder_|^Fund_", names(df))] df <- cbind(df, top3_df) # Step 6: 删除无关字段 filter_out <- c( "CurrencyCode", "CurrencyName", "CurrencySymbol", "CountryISO", "ISIN", "LEI", "Listings", "Officers", "LogoURL", "ShortPercent", "ForwardAnnualDividendRate", "ForwardAnnualDividendYield", "DividendDate", "ExDividendDate", "LastSplitFactor", "LastSplitDate", "Dividend_2021", "Dividend_2022", "Dividend_2023", "Dividend_2024", "Dividend_2025","DelistedDate","Category" ) df <- df %>% select(-any_of(filter_out)) return(df) } ``` ```{r} z1 = extract_top3_holders_clean(x1) z2 = extract_top3_holders_clean(x2) z3 = extract_top3_holders_clean(x3) #z5 = extract_top3_holders_clean(x5) z7 = extract_top3_holders_clean(x7) filter_out <- c( "CurrencyCode", "CurrencyName", "CurrencySymbol", "CountryISO", "ISIN", "LEI", "Listings", "Officers", "LogoURL", "ShortPercent", "ForwardAnnualDividendRate", "ForwardAnnualDividendYield", "DividendDate", "ExDividendDate", "LastSplitFactor", "LastSplitDate", "Dividend_2021", "Dividend_2022", "Dividend_2023", "Dividend_2024", "Dividend_2025", "DelistedDate","Category" ) rm(list = ls(pattern = "^Fund_")) x5 <- x5[ , !grepl("^Fund_", names(x5))] rm(list = ls(pattern = "^Holder_")) x5 <- x5[ , !grepl("^Holder_", names(x5))] z5 = x5%>% select(-any_of(filter_out)) z4 = x4 %>% select(-any_of(filter_out)) z6 = x6 %>% select(-any_of(filter_out)) z8 = x8 %>% select(-any_of(filter_out)) z9 = x9 %>% select(-any_of(filter_out)) z10 = x10 %>% select(-any_of(filter_out)) ``` ```{r} # 需要补齐的 Holder 列名 holder_fields <- c("_name", "_date", "_totalShares", "_currentShares") required_holder_cols <- unlist(lapply(1:3, function(j) paste0("Holder_", j, holder_fields))) # 获取 x1 到 x10 的数据框 data_list <- mget(paste0("z", 1:10)) # 用 lapply 批量补全 data_list_fixed <- lapply(data_list, function(df) { missing_cols <- setdiff(required_holder_cols, names(df)) for (col in missing_cols) { df[[col]] <- NA } return(df) }) # 还原到环境中(x1 到 x10 被更新) list2env(data_list_fixed, .GlobalEnv) date_cols <- grep("_date$", names(z1), value = TRUE) # 强制转换 z1/z2 的日期列为 Date 类型 for (col in date_cols) { if (col %in% names(z1)) z1[[col]] <- as.Date(z1[[col]]) if (col %in% names(z2)) z2[[col]] <- as.Date(z2[[col]]) if (col %in% names(z3)) z3[[col]] <- as.Date(z3[[col]]) if (col %in% names(z4)) z4[[col]] <- as.Date(z4[[col]]) if (col %in% names(z5)) z5[[col]] <- as.Date(z5[[col]]) if (col %in% names(z6)) z6[[col]] <- as.Date(z6[[col]]) if (col %in% names(z7)) z7[[col]] <- as.Date(z7[[col]]) if (col %in% names(z8)) z8[[col]] <- as.Date(z8[[col]]) if (col %in% names(z9)) z9[[col]] <- as.Date(z9[[col]]) if (col %in% names(z10)) z10[[col]] <- as.Date(z10[[col]]) } stock_meta_data <- rbind(z1, z2,z3,z4,z5,z6,z7,z8,z9,z10) ``` # Filter Industry ```{r} ind = c("Other Industrial Metals & Mining", "Gold", "Oil & Gas E&P", "Other Precious Metals & Mining", "Semiconductors", "Oil & Gas Equipment & Services", "Semiconductor Equipment & Materials", "Metal Fabrication", "Oil & Gas Refining & Marketing", "Copper", "Oil & Gas Midstream", "Thermal Coal", "Oil & Gas Integrated", "Uranium", "Silver", "Oil & Gas Drilling", "Metals & Mining", "Oil & Gas", "Coal", "Oil & Gas Related Equipment and Services", "Oil, Gas & Consumable Fuels" ) cleaned_data = stock_meta_data %>% filter(Industry %in% ind) write.csv(cleaned_data, "cleaned_stock_data.csv", row.names = FALSE) ``` # Merge Financial ```{r} fin = read.csv("financials_filtered.csv") stock_meta_data$PrimaryTicker <- paste0(stock_meta_data$Code, ".", stock_meta_data$ExchangeCode) merged = stock_meta_data %>% left_join(fin, by = c("PrimaryTicker" = "Ticker")) merged_industry = merged %>% filter(Industry %in% ind) ``` # EDA ```{r} # 选择相关列 equity_cols <- paste0("totalStockholderEquity_", 2020:2024) df_equity <- merged_industry[, c("PrimaryTicker", equity_cols)] # 定义 CAGR 函数 calc_cagr <- function(start, end, years = 4) { if (is.na(start) || is.na(end) || start <= 0) return(NA) return((end / start)^(1 / years) - 1) } # 逐行计算 CAGR(2020~2024) df_equity$Equity_CAGR_2020_2024 <- mapply( calc_cagr, df_equity$totalStockholderEquity_2020, df_equity$totalStockholderEquity_2024 ) ``` ```{r} library(tidyverse) # 假设 df_equity 已包含 Ticker, totalStockholderEquity_2020~2024 和 CAGR 列 # 选出 CAGR 最大的前 5 家公司 top5 <- df_equity %>% arrange(desc(Equity_CAGR_2020_2024)) %>% slice(1:5) # 把数据转成长格式,适合 ggplot2 绘图 df_long <- top5 %>% select(PrimaryTicker, starts_with("totalStockholderEquity_")) %>% pivot_longer( cols = -PrimaryTicker, names_to = "Year", values_to = "Equity" ) %>% mutate(Year = as.numeric(gsub("totalStockholderEquity_", "", Year))) # 画折线图 ggplot(df_long, aes(x = Year, y = Equity, color = PrimaryTicker)) + geom_line(size = 1) + geom_point(size = 2) + labs( title = "Top 5 Companies by Equity CAGR (2020–2024)", x = "Year", y = "Total Stockholder Equity" ) + theme_minimal() + scale_y_continuous(labels = scales::comma) ```