repo
stringlengths 7
55
| path
stringlengths 4
223
| func_name
stringlengths 1
134
| original_string
stringlengths 75
104k
| language
stringclasses 1
value | code
stringlengths 75
104k
| code_tokens
listlengths 19
28.4k
| docstring
stringlengths 1
46.9k
| docstring_tokens
listlengths 1
1.97k
| sha
stringlengths 40
40
| url
stringlengths 87
315
| partition
stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/main.py
|
QA_SU_save_stock_xdxr
|
def QA_SU_save_stock_xdxr(engine, client=DATABASE):
"""save stock_xdxr
Arguments:
engine {[type]} -- [description]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
engine = select_save_engine(engine)
engine.QA_SU_save_stock_xdxr(client=client)
|
python
|
def QA_SU_save_stock_xdxr(engine, client=DATABASE):
"""save stock_xdxr
Arguments:
engine {[type]} -- [description]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
engine = select_save_engine(engine)
engine.QA_SU_save_stock_xdxr(client=client)
|
[
"def",
"QA_SU_save_stock_xdxr",
"(",
"engine",
",",
"client",
"=",
"DATABASE",
")",
":",
"engine",
"=",
"select_save_engine",
"(",
"engine",
")",
"engine",
".",
"QA_SU_save_stock_xdxr",
"(",
"client",
"=",
"client",
")"
] |
save stock_xdxr
Arguments:
engine {[type]} -- [description]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_xdxr"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/main.py#L315-L326
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/main.py
|
QA_SU_save_stock_block
|
def QA_SU_save_stock_block(engine, client=DATABASE):
"""save stock_block
Arguments:
engine {[type]} -- [description]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
engine = select_save_engine(engine)
engine.QA_SU_save_stock_block(client=client)
|
python
|
def QA_SU_save_stock_block(engine, client=DATABASE):
"""save stock_block
Arguments:
engine {[type]} -- [description]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
engine = select_save_engine(engine)
engine.QA_SU_save_stock_block(client=client)
|
[
"def",
"QA_SU_save_stock_block",
"(",
"engine",
",",
"client",
"=",
"DATABASE",
")",
":",
"engine",
"=",
"select_save_engine",
"(",
"engine",
")",
"engine",
".",
"QA_SU_save_stock_block",
"(",
"client",
"=",
"client",
")"
] |
save stock_block
Arguments:
engine {[type]} -- [description]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_block"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/main.py#L329-L340
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/main.py
|
select_save_engine
|
def select_save_engine(engine, paralleled=False):
'''
select save_engine , tushare ts Tushare 使用 Tushare 免费数据接口, tdx 使用通达信数据接口
:param engine: 字符串Str
:param paralleled: 是否并行处理;默认为False
:return: sts means save_tushare_py or stdx means save_tdx_py
'''
if engine in ['tushare', 'ts', 'Tushare']:
return sts
elif engine in ['tdx']:
if paralleled:
return stdx_parallelism
else:
return stdx
elif engine in ['gm', 'goldenminer']:
return sgm
elif engine in ['jq', 'joinquant']:
return sjq
else:
print('QA Error QASU.main.py call select_save_engine with parameter %s is None of thshare, ts, Thshare, or tdx', engine)
|
python
|
def select_save_engine(engine, paralleled=False):
'''
select save_engine , tushare ts Tushare 使用 Tushare 免费数据接口, tdx 使用通达信数据接口
:param engine: 字符串Str
:param paralleled: 是否并行处理;默认为False
:return: sts means save_tushare_py or stdx means save_tdx_py
'''
if engine in ['tushare', 'ts', 'Tushare']:
return sts
elif engine in ['tdx']:
if paralleled:
return stdx_parallelism
else:
return stdx
elif engine in ['gm', 'goldenminer']:
return sgm
elif engine in ['jq', 'joinquant']:
return sjq
else:
print('QA Error QASU.main.py call select_save_engine with parameter %s is None of thshare, ts, Thshare, or tdx', engine)
|
[
"def",
"select_save_engine",
"(",
"engine",
",",
"paralleled",
"=",
"False",
")",
":",
"if",
"engine",
"in",
"[",
"'tushare'",
",",
"'ts'",
",",
"'Tushare'",
"]",
":",
"return",
"sts",
"elif",
"engine",
"in",
"[",
"'tdx'",
"]",
":",
"if",
"paralleled",
":",
"return",
"stdx_parallelism",
"else",
":",
"return",
"stdx",
"elif",
"engine",
"in",
"[",
"'gm'",
",",
"'goldenminer'",
"]",
":",
"return",
"sgm",
"elif",
"engine",
"in",
"[",
"'jq'",
",",
"'joinquant'",
"]",
":",
"return",
"sjq",
"else",
":",
"print",
"(",
"'QA Error QASU.main.py call select_save_engine with parameter %s is None of thshare, ts, Thshare, or tdx'",
",",
"engine",
")"
] |
select save_engine , tushare ts Tushare 使用 Tushare 免费数据接口, tdx 使用通达信数据接口
:param engine: 字符串Str
:param paralleled: 是否并行处理;默认为False
:return: sts means save_tushare_py or stdx means save_tdx_py
|
[
"select",
"save_engine",
"tushare",
"ts",
"Tushare",
"使用",
"Tushare",
"免费数据接口,",
"tdx",
"使用通达信数据接口",
":",
"param",
"engine",
":",
"字符串Str",
":",
"param",
"paralleled",
":",
"是否并行处理;默认为False",
":",
"return",
":",
"sts",
"means",
"save_tushare_py",
"or",
"stdx",
"means",
"save_tdx_py"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/main.py#L343-L362
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_day
|
def QA_fetch_stock_day(code, start, end, format='numpy', frequence='day', collections=DATABASE.stock_day):
"""'获取股票日线'
Returns:
[type] -- [description]
感谢@几何大佬的提示
https://docs.mongodb.com/manual/tutorial/project-fields-from-query-results/#return-the-specified-fields-and-the-id-field-only
"""
start = str(start)[0:10]
end = str(end)[0:10]
#code= [code] if isinstance(code,str) else code
# code checking
code = QA_util_code_tolist(code)
if QA_util_date_valid(end):
cursor = collections.find({
'code': {'$in': code}, "date_stamp": {
"$lte": QA_util_date_stamp(end),
"$gte": QA_util_date_stamp(start)}}, {"_id": 0}, batch_size=10000)
#res=[QA_util_dict_remove_key(data, '_id') for data in cursor]
res = pd.DataFrame([item for item in cursor])
try:
res = res.assign(volume=res.vol, date=pd.to_datetime(
res.date)).drop_duplicates((['date', 'code'])).query('volume>1').set_index('date', drop=False)
res = res.ix[:, ['code', 'open', 'high', 'low',
'close', 'volume', 'amount', 'date']]
except:
res = None
if format in ['P', 'p', 'pandas', 'pd']:
return res
elif format in ['json', 'dict']:
return QA_util_to_json_from_pandas(res)
# 多种数据格式
elif format in ['n', 'N', 'numpy']:
return numpy.asarray(res)
elif format in ['list', 'l', 'L']:
return numpy.asarray(res).tolist()
else:
print("QA Error QA_fetch_stock_day format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return None
else:
QA_util_log_info(
'QA Error QA_fetch_stock_day data parameter start=%s end=%s is not right' % (start, end))
|
python
|
def QA_fetch_stock_day(code, start, end, format='numpy', frequence='day', collections=DATABASE.stock_day):
"""'获取股票日线'
Returns:
[type] -- [description]
感谢@几何大佬的提示
https://docs.mongodb.com/manual/tutorial/project-fields-from-query-results/#return-the-specified-fields-and-the-id-field-only
"""
start = str(start)[0:10]
end = str(end)[0:10]
#code= [code] if isinstance(code,str) else code
# code checking
code = QA_util_code_tolist(code)
if QA_util_date_valid(end):
cursor = collections.find({
'code': {'$in': code}, "date_stamp": {
"$lte": QA_util_date_stamp(end),
"$gte": QA_util_date_stamp(start)}}, {"_id": 0}, batch_size=10000)
#res=[QA_util_dict_remove_key(data, '_id') for data in cursor]
res = pd.DataFrame([item for item in cursor])
try:
res = res.assign(volume=res.vol, date=pd.to_datetime(
res.date)).drop_duplicates((['date', 'code'])).query('volume>1').set_index('date', drop=False)
res = res.ix[:, ['code', 'open', 'high', 'low',
'close', 'volume', 'amount', 'date']]
except:
res = None
if format in ['P', 'p', 'pandas', 'pd']:
return res
elif format in ['json', 'dict']:
return QA_util_to_json_from_pandas(res)
# 多种数据格式
elif format in ['n', 'N', 'numpy']:
return numpy.asarray(res)
elif format in ['list', 'l', 'L']:
return numpy.asarray(res).tolist()
else:
print("QA Error QA_fetch_stock_day format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return None
else:
QA_util_log_info(
'QA Error QA_fetch_stock_day data parameter start=%s end=%s is not right' % (start, end))
|
[
"def",
"QA_fetch_stock_day",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'numpy'",
",",
"frequence",
"=",
"'day'",
",",
"collections",
"=",
"DATABASE",
".",
"stock_day",
")",
":",
"start",
"=",
"str",
"(",
"start",
")",
"[",
"0",
":",
"10",
"]",
"end",
"=",
"str",
"(",
"end",
")",
"[",
"0",
":",
"10",
"]",
"#code= [code] if isinstance(code,str) else code",
"# code checking",
"code",
"=",
"QA_util_code_tolist",
"(",
"code",
")",
"if",
"QA_util_date_valid",
"(",
"end",
")",
":",
"cursor",
"=",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
",",
"\"date_stamp\"",
":",
"{",
"\"$lte\"",
":",
"QA_util_date_stamp",
"(",
"end",
")",
",",
"\"$gte\"",
":",
"QA_util_date_stamp",
"(",
"start",
")",
"}",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"#res=[QA_util_dict_remove_key(data, '_id') for data in cursor]",
"res",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"cursor",
"]",
")",
"try",
":",
"res",
"=",
"res",
".",
"assign",
"(",
"volume",
"=",
"res",
".",
"vol",
",",
"date",
"=",
"pd",
".",
"to_datetime",
"(",
"res",
".",
"date",
")",
")",
".",
"drop_duplicates",
"(",
"(",
"[",
"'date'",
",",
"'code'",
"]",
")",
")",
".",
"query",
"(",
"'volume>1'",
")",
".",
"set_index",
"(",
"'date'",
",",
"drop",
"=",
"False",
")",
"res",
"=",
"res",
".",
"ix",
"[",
":",
",",
"[",
"'code'",
",",
"'open'",
",",
"'high'",
",",
"'low'",
",",
"'close'",
",",
"'volume'",
",",
"'amount'",
",",
"'date'",
"]",
"]",
"except",
":",
"res",
"=",
"None",
"if",
"format",
"in",
"[",
"'P'",
",",
"'p'",
",",
"'pandas'",
",",
"'pd'",
"]",
":",
"return",
"res",
"elif",
"format",
"in",
"[",
"'json'",
",",
"'dict'",
"]",
":",
"return",
"QA_util_to_json_from_pandas",
"(",
"res",
")",
"# 多种数据格式",
"elif",
"format",
"in",
"[",
"'n'",
",",
"'N'",
",",
"'numpy'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"res",
")",
"elif",
"format",
"in",
"[",
"'list'",
",",
"'l'",
",",
"'L'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"res",
")",
".",
"tolist",
"(",
")",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_day format parameter %s is none of \\\"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\\\" \"",
"%",
"format",
")",
"return",
"None",
"else",
":",
"QA_util_log_info",
"(",
"'QA Error QA_fetch_stock_day data parameter start=%s end=%s is not right'",
"%",
"(",
"start",
",",
"end",
")",
")"
] |
获取股票日线'
Returns:
[type] -- [description]
感谢@几何大佬的提示
https://docs.mongodb.com/manual/tutorial/project-fields-from-query-results/#return-the-specified-fields-and-the-id-field-only
|
[
"获取股票日线"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L48-L96
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_min
|
def QA_fetch_stock_min(code, start, end, format='numpy', frequence='1min', collections=DATABASE.stock_min):
'获取股票分钟线'
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
else:
print("QA Error QA_fetch_stock_min parameter frequence=%s is none of 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m" % frequence)
__data = []
# code checking
code = QA_util_code_tolist(code)
cursor = collections.find({
'code': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, {"_id": 0}, batch_size=10000)
res = pd.DataFrame([item for item in cursor])
try:
res = res.assign(volume=res.vol, datetime=pd.to_datetime(
res.datetime)).query('volume>1').drop_duplicates(['datetime', 'code']).set_index('datetime', drop=False)
# return res
except:
res = None
if format in ['P', 'p', 'pandas', 'pd']:
return res
elif format in ['json', 'dict']:
return QA_util_to_json_from_pandas(res)
# 多种数据格式
elif format in ['n', 'N', 'numpy']:
return numpy.asarray(res)
elif format in ['list', 'l', 'L']:
return numpy.asarray(res).tolist()
else:
print("QA Error QA_fetch_stock_min format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return None
|
python
|
def QA_fetch_stock_min(code, start, end, format='numpy', frequence='1min', collections=DATABASE.stock_min):
'获取股票分钟线'
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
else:
print("QA Error QA_fetch_stock_min parameter frequence=%s is none of 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m" % frequence)
__data = []
# code checking
code = QA_util_code_tolist(code)
cursor = collections.find({
'code': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, {"_id": 0}, batch_size=10000)
res = pd.DataFrame([item for item in cursor])
try:
res = res.assign(volume=res.vol, datetime=pd.to_datetime(
res.datetime)).query('volume>1').drop_duplicates(['datetime', 'code']).set_index('datetime', drop=False)
# return res
except:
res = None
if format in ['P', 'p', 'pandas', 'pd']:
return res
elif format in ['json', 'dict']:
return QA_util_to_json_from_pandas(res)
# 多种数据格式
elif format in ['n', 'N', 'numpy']:
return numpy.asarray(res)
elif format in ['list', 'l', 'L']:
return numpy.asarray(res).tolist()
else:
print("QA Error QA_fetch_stock_min format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return None
|
[
"def",
"QA_fetch_stock_min",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'numpy'",
",",
"frequence",
"=",
"'1min'",
",",
"collections",
"=",
"DATABASE",
".",
"stock_min",
")",
":",
"if",
"frequence",
"in",
"[",
"'1min'",
",",
"'1m'",
"]",
":",
"frequence",
"=",
"'1min'",
"elif",
"frequence",
"in",
"[",
"'5min'",
",",
"'5m'",
"]",
":",
"frequence",
"=",
"'5min'",
"elif",
"frequence",
"in",
"[",
"'15min'",
",",
"'15m'",
"]",
":",
"frequence",
"=",
"'15min'",
"elif",
"frequence",
"in",
"[",
"'30min'",
",",
"'30m'",
"]",
":",
"frequence",
"=",
"'30min'",
"elif",
"frequence",
"in",
"[",
"'60min'",
",",
"'60m'",
"]",
":",
"frequence",
"=",
"'60min'",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_min parameter frequence=%s is none of 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m\"",
"%",
"frequence",
")",
"__data",
"=",
"[",
"]",
"# code checking",
"code",
"=",
"QA_util_code_tolist",
"(",
"code",
")",
"cursor",
"=",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
",",
"\"time_stamp\"",
":",
"{",
"\"$gte\"",
":",
"QA_util_time_stamp",
"(",
"start",
")",
",",
"\"$lte\"",
":",
"QA_util_time_stamp",
"(",
"end",
")",
"}",
",",
"'type'",
":",
"frequence",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"res",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"cursor",
"]",
")",
"try",
":",
"res",
"=",
"res",
".",
"assign",
"(",
"volume",
"=",
"res",
".",
"vol",
",",
"datetime",
"=",
"pd",
".",
"to_datetime",
"(",
"res",
".",
"datetime",
")",
")",
".",
"query",
"(",
"'volume>1'",
")",
".",
"drop_duplicates",
"(",
"[",
"'datetime'",
",",
"'code'",
"]",
")",
".",
"set_index",
"(",
"'datetime'",
",",
"drop",
"=",
"False",
")",
"# return res",
"except",
":",
"res",
"=",
"None",
"if",
"format",
"in",
"[",
"'P'",
",",
"'p'",
",",
"'pandas'",
",",
"'pd'",
"]",
":",
"return",
"res",
"elif",
"format",
"in",
"[",
"'json'",
",",
"'dict'",
"]",
":",
"return",
"QA_util_to_json_from_pandas",
"(",
"res",
")",
"# 多种数据格式",
"elif",
"format",
"in",
"[",
"'n'",
",",
"'N'",
",",
"'numpy'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"res",
")",
"elif",
"format",
"in",
"[",
"'list'",
",",
"'l'",
",",
"'L'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"res",
")",
".",
"tolist",
"(",
")",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_min format parameter %s is none of \\\"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\\\" \"",
"%",
"format",
")",
"return",
"None"
] |
获取股票分钟线
|
[
"获取股票分钟线"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L99-L143
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_list
|
def QA_fetch_stock_list(collections=DATABASE.stock_list):
'获取股票列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
python
|
def QA_fetch_stock_list(collections=DATABASE.stock_list):
'获取股票列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
[
"def",
"QA_fetch_stock_list",
"(",
"collections",
"=",
"DATABASE",
".",
"stock_list",
")",
":",
"return",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
")",
"]",
")",
".",
"drop",
"(",
"'_id'",
",",
"axis",
"=",
"1",
",",
"inplace",
"=",
"False",
")",
".",
"set_index",
"(",
"'code'",
",",
"drop",
"=",
"False",
")"
] |
获取股票列表
|
[
"获取股票列表"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L151-L154
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_etf_list
|
def QA_fetch_etf_list(collections=DATABASE.etf_list):
'获取ETF列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
python
|
def QA_fetch_etf_list(collections=DATABASE.etf_list):
'获取ETF列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
[
"def",
"QA_fetch_etf_list",
"(",
"collections",
"=",
"DATABASE",
".",
"etf_list",
")",
":",
"return",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
")",
"]",
")",
".",
"drop",
"(",
"'_id'",
",",
"axis",
"=",
"1",
",",
"inplace",
"=",
"False",
")",
".",
"set_index",
"(",
"'code'",
",",
"drop",
"=",
"False",
")"
] |
获取ETF列表
|
[
"获取ETF列表"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L157-L160
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_index_list
|
def QA_fetch_index_list(collections=DATABASE.index_list):
'获取指数列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
python
|
def QA_fetch_index_list(collections=DATABASE.index_list):
'获取指数列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
[
"def",
"QA_fetch_index_list",
"(",
"collections",
"=",
"DATABASE",
".",
"index_list",
")",
":",
"return",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
")",
"]",
")",
".",
"drop",
"(",
"'_id'",
",",
"axis",
"=",
"1",
",",
"inplace",
"=",
"False",
")",
".",
"set_index",
"(",
"'code'",
",",
"drop",
"=",
"False",
")"
] |
获取指数列表
|
[
"获取指数列表"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L163-L165
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_terminated
|
def QA_fetch_stock_terminated(collections=DATABASE.stock_terminated):
'获取股票基本信息 , 已经退市的股票列表'
# 🛠todo 转变成 dataframe 类型数据
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
python
|
def QA_fetch_stock_terminated(collections=DATABASE.stock_terminated):
'获取股票基本信息 , 已经退市的股票列表'
# 🛠todo 转变成 dataframe 类型数据
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
[
"def",
"QA_fetch_stock_terminated",
"(",
"collections",
"=",
"DATABASE",
".",
"stock_terminated",
")",
":",
"# 🛠todo 转变成 dataframe 类型数据",
"return",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
")",
"]",
")",
".",
"drop",
"(",
"'_id'",
",",
"axis",
"=",
"1",
",",
"inplace",
"=",
"False",
")",
".",
"set_index",
"(",
"'code'",
",",
"drop",
"=",
"False",
")"
] |
获取股票基本信息 , 已经退市的股票列表
|
[
"获取股票基本信息",
"已经退市的股票列表"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L168-L171
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_basic_info_tushare
|
def QA_fetch_stock_basic_info_tushare(collections=DATABASE.stock_info_tushare):
'''
purpose:
tushare 股票列表数据库
code,代码
name,名称
industry,所属行业
area,地区
pe,市盈率
outstanding,流通股本(亿)
totals,总股本(亿)
totalAssets,总资产(万)
liquidAssets,流动资产
fixedAssets,固定资产
reserved,公积金
reservedPerShare,每股公积金
esp,每股收益
bvps,每股净资
pb,市净率
timeToMarket,上市日期
undp,未分利润
perundp, 每股未分配
rev,收入同比(%)
profit,利润同比(%)
gpr,毛利率(%)
npr,净利润率(%)
holders,股东人数
add by tauruswang,
:param collections: stock_info_tushare 集合
:return:
'''
'获取股票基本信息'
items = [item for item in collections.find()]
# 🛠todo 转变成 dataframe 类型数据
return items
|
python
|
def QA_fetch_stock_basic_info_tushare(collections=DATABASE.stock_info_tushare):
'''
purpose:
tushare 股票列表数据库
code,代码
name,名称
industry,所属行业
area,地区
pe,市盈率
outstanding,流通股本(亿)
totals,总股本(亿)
totalAssets,总资产(万)
liquidAssets,流动资产
fixedAssets,固定资产
reserved,公积金
reservedPerShare,每股公积金
esp,每股收益
bvps,每股净资
pb,市净率
timeToMarket,上市日期
undp,未分利润
perundp, 每股未分配
rev,收入同比(%)
profit,利润同比(%)
gpr,毛利率(%)
npr,净利润率(%)
holders,股东人数
add by tauruswang,
:param collections: stock_info_tushare 集合
:return:
'''
'获取股票基本信息'
items = [item for item in collections.find()]
# 🛠todo 转变成 dataframe 类型数据
return items
|
[
"def",
"QA_fetch_stock_basic_info_tushare",
"(",
"collections",
"=",
"DATABASE",
".",
"stock_info_tushare",
")",
":",
"'获取股票基本信息'",
"items",
"=",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
")",
"]",
"# 🛠todo 转变成 dataframe 类型数据",
"return",
"items"
] |
purpose:
tushare 股票列表数据库
code,代码
name,名称
industry,所属行业
area,地区
pe,市盈率
outstanding,流通股本(亿)
totals,总股本(亿)
totalAssets,总资产(万)
liquidAssets,流动资产
fixedAssets,固定资产
reserved,公积金
reservedPerShare,每股公积金
esp,每股收益
bvps,每股净资
pb,市净率
timeToMarket,上市日期
undp,未分利润
perundp, 每股未分配
rev,收入同比(%)
profit,利润同比(%)
gpr,毛利率(%)
npr,净利润率(%)
holders,股东人数
add by tauruswang,
:param collections: stock_info_tushare 集合
:return:
|
[
"purpose",
":",
"tushare",
"股票列表数据库"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L174-L211
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_full
|
def QA_fetch_stock_full(date, format='numpy', collections=DATABASE.stock_day):
'获取全市场的某一日的数据'
Date = str(date)[0:10]
if QA_util_date_valid(Date) is True:
__data = []
for item in collections.find({
"date_stamp": QA_util_date_stamp(Date)}, batch_size=10000):
__data.append([str(item['code']), float(item['open']), float(item['high']), float(
item['low']), float(item['close']), float(item['vol']), item['date']])
# 多种数据格式
if format in ['n', 'N', 'numpy']:
__data = numpy.asarray(__data)
elif format in ['list', 'l', 'L']:
__data = __data
elif format in ['P', 'p', 'pandas', 'pd']:
__data = DataFrame(__data, columns=[
'code', 'open', 'high', 'low', 'close', 'volume', 'date'])
__data['date'] = pd.to_datetime(__data['date'])
__data = __data.set_index('date', drop=False)
else:
print("QA Error QA_fetch_stock_full format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return __data
else:
QA_util_log_info(
'QA Error QA_fetch_stock_full data parameter date=%s not right' % date)
|
python
|
def QA_fetch_stock_full(date, format='numpy', collections=DATABASE.stock_day):
'获取全市场的某一日的数据'
Date = str(date)[0:10]
if QA_util_date_valid(Date) is True:
__data = []
for item in collections.find({
"date_stamp": QA_util_date_stamp(Date)}, batch_size=10000):
__data.append([str(item['code']), float(item['open']), float(item['high']), float(
item['low']), float(item['close']), float(item['vol']), item['date']])
# 多种数据格式
if format in ['n', 'N', 'numpy']:
__data = numpy.asarray(__data)
elif format in ['list', 'l', 'L']:
__data = __data
elif format in ['P', 'p', 'pandas', 'pd']:
__data = DataFrame(__data, columns=[
'code', 'open', 'high', 'low', 'close', 'volume', 'date'])
__data['date'] = pd.to_datetime(__data['date'])
__data = __data.set_index('date', drop=False)
else:
print("QA Error QA_fetch_stock_full format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return __data
else:
QA_util_log_info(
'QA Error QA_fetch_stock_full data parameter date=%s not right' % date)
|
[
"def",
"QA_fetch_stock_full",
"(",
"date",
",",
"format",
"=",
"'numpy'",
",",
"collections",
"=",
"DATABASE",
".",
"stock_day",
")",
":",
"Date",
"=",
"str",
"(",
"date",
")",
"[",
"0",
":",
"10",
"]",
"if",
"QA_util_date_valid",
"(",
"Date",
")",
"is",
"True",
":",
"__data",
"=",
"[",
"]",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"\"date_stamp\"",
":",
"QA_util_date_stamp",
"(",
"Date",
")",
"}",
",",
"batch_size",
"=",
"10000",
")",
":",
"__data",
".",
"append",
"(",
"[",
"str",
"(",
"item",
"[",
"'code'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'open'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'high'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'low'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'close'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'vol'",
"]",
")",
",",
"item",
"[",
"'date'",
"]",
"]",
")",
"# 多种数据格式",
"if",
"format",
"in",
"[",
"'n'",
",",
"'N'",
",",
"'numpy'",
"]",
":",
"__data",
"=",
"numpy",
".",
"asarray",
"(",
"__data",
")",
"elif",
"format",
"in",
"[",
"'list'",
",",
"'l'",
",",
"'L'",
"]",
":",
"__data",
"=",
"__data",
"elif",
"format",
"in",
"[",
"'P'",
",",
"'p'",
",",
"'pandas'",
",",
"'pd'",
"]",
":",
"__data",
"=",
"DataFrame",
"(",
"__data",
",",
"columns",
"=",
"[",
"'code'",
",",
"'open'",
",",
"'high'",
",",
"'low'",
",",
"'close'",
",",
"'volume'",
",",
"'date'",
"]",
")",
"__data",
"[",
"'date'",
"]",
"=",
"pd",
".",
"to_datetime",
"(",
"__data",
"[",
"'date'",
"]",
")",
"__data",
"=",
"__data",
".",
"set_index",
"(",
"'date'",
",",
"drop",
"=",
"False",
")",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_full format parameter %s is none of \\\"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\\\" \"",
"%",
"format",
")",
"return",
"__data",
"else",
":",
"QA_util_log_info",
"(",
"'QA Error QA_fetch_stock_full data parameter date=%s not right'",
"%",
"date",
")"
] |
获取全市场的某一日的数据
|
[
"获取全市场的某一日的数据"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L226-L252
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_index_min
|
def QA_fetch_index_min(
code,
start, end,
format='numpy',
frequence='1min',
collections=DATABASE.index_min):
'获取股票分钟线'
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
__data = []
code = QA_util_code_tolist(code)
cursor = collections.find({
'code': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, {"_id": 0}, batch_size=10000)
if format in ['dict', 'json']:
return [data for data in cursor]
# for item in cursor:
__data = pd.DataFrame([item for item in cursor])
__data = __data.assign(datetime=pd.to_datetime(__data['datetime']))
# __data.append([str(item['code']), float(item['open']), float(item['high']), float(
# item['low']), float(item['close']), int(item['up_count']), int(item['down_count']), float(item['vol']), float(item['amount']), item['datetime'], item['time_stamp'], item['date'], item['type']])
# __data = DataFrame(__data, columns=[
# 'code', 'open', 'high', 'low', 'close', 'up_count', 'down_count', 'volume', 'amount', 'datetime', 'time_stamp', 'date', 'type'])
# __data['datetime'] = pd.to_datetime(__data['datetime'])
__data = __data.set_index('datetime', drop=False)
if format in ['numpy', 'np', 'n']:
return numpy.asarray(__data)
elif format in ['list', 'l', 'L']:
return numpy.asarray(__data).tolist()
elif format in ['P', 'p', 'pandas', 'pd']:
return __data
|
python
|
def QA_fetch_index_min(
code,
start, end,
format='numpy',
frequence='1min',
collections=DATABASE.index_min):
'获取股票分钟线'
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
__data = []
code = QA_util_code_tolist(code)
cursor = collections.find({
'code': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, {"_id": 0}, batch_size=10000)
if format in ['dict', 'json']:
return [data for data in cursor]
# for item in cursor:
__data = pd.DataFrame([item for item in cursor])
__data = __data.assign(datetime=pd.to_datetime(__data['datetime']))
# __data.append([str(item['code']), float(item['open']), float(item['high']), float(
# item['low']), float(item['close']), int(item['up_count']), int(item['down_count']), float(item['vol']), float(item['amount']), item['datetime'], item['time_stamp'], item['date'], item['type']])
# __data = DataFrame(__data, columns=[
# 'code', 'open', 'high', 'low', 'close', 'up_count', 'down_count', 'volume', 'amount', 'datetime', 'time_stamp', 'date', 'type'])
# __data['datetime'] = pd.to_datetime(__data['datetime'])
__data = __data.set_index('datetime', drop=False)
if format in ['numpy', 'np', 'n']:
return numpy.asarray(__data)
elif format in ['list', 'l', 'L']:
return numpy.asarray(__data).tolist()
elif format in ['P', 'p', 'pandas', 'pd']:
return __data
|
[
"def",
"QA_fetch_index_min",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'numpy'",
",",
"frequence",
"=",
"'1min'",
",",
"collections",
"=",
"DATABASE",
".",
"index_min",
")",
":",
"if",
"frequence",
"in",
"[",
"'1min'",
",",
"'1m'",
"]",
":",
"frequence",
"=",
"'1min'",
"elif",
"frequence",
"in",
"[",
"'5min'",
",",
"'5m'",
"]",
":",
"frequence",
"=",
"'5min'",
"elif",
"frequence",
"in",
"[",
"'15min'",
",",
"'15m'",
"]",
":",
"frequence",
"=",
"'15min'",
"elif",
"frequence",
"in",
"[",
"'30min'",
",",
"'30m'",
"]",
":",
"frequence",
"=",
"'30min'",
"elif",
"frequence",
"in",
"[",
"'60min'",
",",
"'60m'",
"]",
":",
"frequence",
"=",
"'60min'",
"__data",
"=",
"[",
"]",
"code",
"=",
"QA_util_code_tolist",
"(",
"code",
")",
"cursor",
"=",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
",",
"\"time_stamp\"",
":",
"{",
"\"$gte\"",
":",
"QA_util_time_stamp",
"(",
"start",
")",
",",
"\"$lte\"",
":",
"QA_util_time_stamp",
"(",
"end",
")",
"}",
",",
"'type'",
":",
"frequence",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"if",
"format",
"in",
"[",
"'dict'",
",",
"'json'",
"]",
":",
"return",
"[",
"data",
"for",
"data",
"in",
"cursor",
"]",
"# for item in cursor:",
"__data",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"cursor",
"]",
")",
"__data",
"=",
"__data",
".",
"assign",
"(",
"datetime",
"=",
"pd",
".",
"to_datetime",
"(",
"__data",
"[",
"'datetime'",
"]",
")",
")",
"# __data.append([str(item['code']), float(item['open']), float(item['high']), float(",
"# item['low']), float(item['close']), int(item['up_count']), int(item['down_count']), float(item['vol']), float(item['amount']), item['datetime'], item['time_stamp'], item['date'], item['type']])",
"# __data = DataFrame(__data, columns=[",
"# 'code', 'open', 'high', 'low', 'close', 'up_count', 'down_count', 'volume', 'amount', 'datetime', 'time_stamp', 'date', 'type'])",
"# __data['datetime'] = pd.to_datetime(__data['datetime'])",
"__data",
"=",
"__data",
".",
"set_index",
"(",
"'datetime'",
",",
"drop",
"=",
"False",
")",
"if",
"format",
"in",
"[",
"'numpy'",
",",
"'np'",
",",
"'n'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"__data",
")",
"elif",
"format",
"in",
"[",
"'list'",
",",
"'l'",
",",
"'L'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"__data",
")",
".",
"tolist",
"(",
")",
"elif",
"format",
"in",
"[",
"'P'",
",",
"'p'",
",",
"'pandas'",
",",
"'pd'",
"]",
":",
"return",
"__data"
] |
获取股票分钟线
|
[
"获取股票分钟线"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L291-L334
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_future_min
|
def QA_fetch_future_min(
code,
start, end,
format='numpy',
frequence='1min',
collections=DATABASE.future_min):
'获取股票分钟线'
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
__data = []
code = QA_util_code_tolist(code, auto_fill=False)
cursor = collections.find({
'code': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, batch_size=10000)
if format in ['dict', 'json']:
return [data for data in cursor]
for item in cursor:
__data.append([str(item['code']), float(item['open']), float(item['high']), float(
item['low']), float(item['close']), float(item['position']), float(item['price']), float(item['trade']),
item['datetime'], item['tradetime'], item['time_stamp'], item['date'], item['type']])
__data = DataFrame(__data, columns=[
'code', 'open', 'high', 'low', 'close', 'position', 'price', 'trade', 'datetime', 'tradetime', 'time_stamp', 'date', 'type'])
__data['datetime'] = pd.to_datetime(__data['datetime'])
__data = __data.set_index('datetime', drop=False)
if format in ['numpy', 'np', 'n']:
return numpy.asarray(__data)
elif format in ['list', 'l', 'L']:
return numpy.asarray(__data).tolist()
elif format in ['P', 'p', 'pandas', 'pd']:
return __data
|
python
|
def QA_fetch_future_min(
code,
start, end,
format='numpy',
frequence='1min',
collections=DATABASE.future_min):
'获取股票分钟线'
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
__data = []
code = QA_util_code_tolist(code, auto_fill=False)
cursor = collections.find({
'code': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, batch_size=10000)
if format in ['dict', 'json']:
return [data for data in cursor]
for item in cursor:
__data.append([str(item['code']), float(item['open']), float(item['high']), float(
item['low']), float(item['close']), float(item['position']), float(item['price']), float(item['trade']),
item['datetime'], item['tradetime'], item['time_stamp'], item['date'], item['type']])
__data = DataFrame(__data, columns=[
'code', 'open', 'high', 'low', 'close', 'position', 'price', 'trade', 'datetime', 'tradetime', 'time_stamp', 'date', 'type'])
__data['datetime'] = pd.to_datetime(__data['datetime'])
__data = __data.set_index('datetime', drop=False)
if format in ['numpy', 'np', 'n']:
return numpy.asarray(__data)
elif format in ['list', 'l', 'L']:
return numpy.asarray(__data).tolist()
elif format in ['P', 'p', 'pandas', 'pd']:
return __data
|
[
"def",
"QA_fetch_future_min",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'numpy'",
",",
"frequence",
"=",
"'1min'",
",",
"collections",
"=",
"DATABASE",
".",
"future_min",
")",
":",
"if",
"frequence",
"in",
"[",
"'1min'",
",",
"'1m'",
"]",
":",
"frequence",
"=",
"'1min'",
"elif",
"frequence",
"in",
"[",
"'5min'",
",",
"'5m'",
"]",
":",
"frequence",
"=",
"'5min'",
"elif",
"frequence",
"in",
"[",
"'15min'",
",",
"'15m'",
"]",
":",
"frequence",
"=",
"'15min'",
"elif",
"frequence",
"in",
"[",
"'30min'",
",",
"'30m'",
"]",
":",
"frequence",
"=",
"'30min'",
"elif",
"frequence",
"in",
"[",
"'60min'",
",",
"'60m'",
"]",
":",
"frequence",
"=",
"'60min'",
"__data",
"=",
"[",
"]",
"code",
"=",
"QA_util_code_tolist",
"(",
"code",
",",
"auto_fill",
"=",
"False",
")",
"cursor",
"=",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
",",
"\"time_stamp\"",
":",
"{",
"\"$gte\"",
":",
"QA_util_time_stamp",
"(",
"start",
")",
",",
"\"$lte\"",
":",
"QA_util_time_stamp",
"(",
"end",
")",
"}",
",",
"'type'",
":",
"frequence",
"}",
",",
"batch_size",
"=",
"10000",
")",
"if",
"format",
"in",
"[",
"'dict'",
",",
"'json'",
"]",
":",
"return",
"[",
"data",
"for",
"data",
"in",
"cursor",
"]",
"for",
"item",
"in",
"cursor",
":",
"__data",
".",
"append",
"(",
"[",
"str",
"(",
"item",
"[",
"'code'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'open'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'high'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'low'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'close'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'position'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'price'",
"]",
")",
",",
"float",
"(",
"item",
"[",
"'trade'",
"]",
")",
",",
"item",
"[",
"'datetime'",
"]",
",",
"item",
"[",
"'tradetime'",
"]",
",",
"item",
"[",
"'time_stamp'",
"]",
",",
"item",
"[",
"'date'",
"]",
",",
"item",
"[",
"'type'",
"]",
"]",
")",
"__data",
"=",
"DataFrame",
"(",
"__data",
",",
"columns",
"=",
"[",
"'code'",
",",
"'open'",
",",
"'high'",
",",
"'low'",
",",
"'close'",
",",
"'position'",
",",
"'price'",
",",
"'trade'",
",",
"'datetime'",
",",
"'tradetime'",
",",
"'time_stamp'",
",",
"'date'",
",",
"'type'",
"]",
")",
"__data",
"[",
"'datetime'",
"]",
"=",
"pd",
".",
"to_datetime",
"(",
"__data",
"[",
"'datetime'",
"]",
")",
"__data",
"=",
"__data",
".",
"set_index",
"(",
"'datetime'",
",",
"drop",
"=",
"False",
")",
"if",
"format",
"in",
"[",
"'numpy'",
",",
"'np'",
",",
"'n'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"__data",
")",
"elif",
"format",
"in",
"[",
"'list'",
",",
"'l'",
",",
"'L'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"__data",
")",
".",
"tolist",
"(",
")",
"elif",
"format",
"in",
"[",
"'P'",
",",
"'p'",
",",
"'pandas'",
",",
"'pd'",
"]",
":",
"return",
"__data"
] |
获取股票分钟线
|
[
"获取股票分钟线"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L373-L416
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_future_list
|
def QA_fetch_future_list(collections=DATABASE.future_list):
'获取期货列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
python
|
def QA_fetch_future_list(collections=DATABASE.future_list):
'获取期货列表'
return pd.DataFrame([item for item in collections.find()]).drop('_id', axis=1, inplace=False).set_index('code', drop=False)
|
[
"def",
"QA_fetch_future_list",
"(",
"collections",
"=",
"DATABASE",
".",
"future_list",
")",
":",
"return",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
")",
"]",
")",
".",
"drop",
"(",
"'_id'",
",",
"axis",
"=",
"1",
",",
"inplace",
"=",
"False",
")",
".",
"set_index",
"(",
"'code'",
",",
"drop",
"=",
"False",
")"
] |
获取期货列表
|
[
"获取期货列表"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L419-L421
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_ctp_tick
|
def QA_fetch_ctp_tick(code, start, end, frequence, format='pd', collections=DATABASE.ctp_tick):
"""仅供存储的ctp tick使用
Arguments:
code {[type]} -- [description]
Keyword Arguments:
format {str} -- [description] (default: {'pd'})
collections {[type]} -- [description] (default: {DATABASE.ctp_tick})
Returns:
[type] -- [description]
"""
code = QA_util_code_tolist(code, auto_fill=False)
cursor = collections.find({
'InstrumentID': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, {"_id": 0}, batch_size=10000)
hq = pd.DataFrame([data for data in cursor]).replace(1.7976931348623157e+308,
numpy.nan).replace('', numpy.nan).dropna(axis=1)
p1 = hq.loc[:, ['ActionDay', 'AskPrice1', 'AskVolume1', 'AveragePrice', 'BidPrice1',
'BidVolume1', 'HighestPrice', 'InstrumentID', 'LastPrice',
'OpenInterest', 'TradingDay', 'UpdateMillisec',
'UpdateTime', 'Volume']]
p1 = p1.assign(datetime=p1.ActionDay.apply(QA_util_date_int2str)+' '+p1.UpdateTime + (p1.UpdateMillisec/1000000).apply(lambda x: str('%.6f' % x)[1:]),
code=p1.InstrumentID)
p1.datetime = pd.to_datetime(p1.datetime)
return p1.set_index(p1.datetime)
|
python
|
def QA_fetch_ctp_tick(code, start, end, frequence, format='pd', collections=DATABASE.ctp_tick):
"""仅供存储的ctp tick使用
Arguments:
code {[type]} -- [description]
Keyword Arguments:
format {str} -- [description] (default: {'pd'})
collections {[type]} -- [description] (default: {DATABASE.ctp_tick})
Returns:
[type] -- [description]
"""
code = QA_util_code_tolist(code, auto_fill=False)
cursor = collections.find({
'InstrumentID': {'$in': code}, "time_stamp": {
"$gte": QA_util_time_stamp(start),
"$lte": QA_util_time_stamp(end)
}, 'type': frequence
}, {"_id": 0}, batch_size=10000)
hq = pd.DataFrame([data for data in cursor]).replace(1.7976931348623157e+308,
numpy.nan).replace('', numpy.nan).dropna(axis=1)
p1 = hq.loc[:, ['ActionDay', 'AskPrice1', 'AskVolume1', 'AveragePrice', 'BidPrice1',
'BidVolume1', 'HighestPrice', 'InstrumentID', 'LastPrice',
'OpenInterest', 'TradingDay', 'UpdateMillisec',
'UpdateTime', 'Volume']]
p1 = p1.assign(datetime=p1.ActionDay.apply(QA_util_date_int2str)+' '+p1.UpdateTime + (p1.UpdateMillisec/1000000).apply(lambda x: str('%.6f' % x)[1:]),
code=p1.InstrumentID)
p1.datetime = pd.to_datetime(p1.datetime)
return p1.set_index(p1.datetime)
|
[
"def",
"QA_fetch_ctp_tick",
"(",
"code",
",",
"start",
",",
"end",
",",
"frequence",
",",
"format",
"=",
"'pd'",
",",
"collections",
"=",
"DATABASE",
".",
"ctp_tick",
")",
":",
"code",
"=",
"QA_util_code_tolist",
"(",
"code",
",",
"auto_fill",
"=",
"False",
")",
"cursor",
"=",
"collections",
".",
"find",
"(",
"{",
"'InstrumentID'",
":",
"{",
"'$in'",
":",
"code",
"}",
",",
"\"time_stamp\"",
":",
"{",
"\"$gte\"",
":",
"QA_util_time_stamp",
"(",
"start",
")",
",",
"\"$lte\"",
":",
"QA_util_time_stamp",
"(",
"end",
")",
"}",
",",
"'type'",
":",
"frequence",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"hq",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"data",
"for",
"data",
"in",
"cursor",
"]",
")",
".",
"replace",
"(",
"1.7976931348623157e+308",
",",
"numpy",
".",
"nan",
")",
".",
"replace",
"(",
"''",
",",
"numpy",
".",
"nan",
")",
".",
"dropna",
"(",
"axis",
"=",
"1",
")",
"p1",
"=",
"hq",
".",
"loc",
"[",
":",
",",
"[",
"'ActionDay'",
",",
"'AskPrice1'",
",",
"'AskVolume1'",
",",
"'AveragePrice'",
",",
"'BidPrice1'",
",",
"'BidVolume1'",
",",
"'HighestPrice'",
",",
"'InstrumentID'",
",",
"'LastPrice'",
",",
"'OpenInterest'",
",",
"'TradingDay'",
",",
"'UpdateMillisec'",
",",
"'UpdateTime'",
",",
"'Volume'",
"]",
"]",
"p1",
"=",
"p1",
".",
"assign",
"(",
"datetime",
"=",
"p1",
".",
"ActionDay",
".",
"apply",
"(",
"QA_util_date_int2str",
")",
"+",
"' '",
"+",
"p1",
".",
"UpdateTime",
"+",
"(",
"p1",
".",
"UpdateMillisec",
"/",
"1000000",
")",
".",
"apply",
"(",
"lambda",
"x",
":",
"str",
"(",
"'%.6f'",
"%",
"x",
")",
"[",
"1",
":",
"]",
")",
",",
"code",
"=",
"p1",
".",
"InstrumentID",
")",
"p1",
".",
"datetime",
"=",
"pd",
".",
"to_datetime",
"(",
"p1",
".",
"datetime",
")",
"return",
"p1",
".",
"set_index",
"(",
"p1",
".",
"datetime",
")"
] |
仅供存储的ctp tick使用
Arguments:
code {[type]} -- [description]
Keyword Arguments:
format {str} -- [description] (default: {'pd'})
collections {[type]} -- [description] (default: {DATABASE.ctp_tick})
Returns:
[type] -- [description]
|
[
"仅供存储的ctp",
"tick使用"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L428-L459
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_xdxr
|
def QA_fetch_stock_xdxr(code, format='pd', collections=DATABASE.stock_xdxr):
'获取股票除权信息/数据库'
code = QA_util_code_tolist(code)
data = pd.DataFrame([item for item in collections.find(
{'code': {'$in': code}}, batch_size=10000)]).drop(['_id'], axis=1)
data['date'] = pd.to_datetime(data['date'])
return data.set_index('date', drop=False)
|
python
|
def QA_fetch_stock_xdxr(code, format='pd', collections=DATABASE.stock_xdxr):
'获取股票除权信息/数据库'
code = QA_util_code_tolist(code)
data = pd.DataFrame([item for item in collections.find(
{'code': {'$in': code}}, batch_size=10000)]).drop(['_id'], axis=1)
data['date'] = pd.to_datetime(data['date'])
return data.set_index('date', drop=False)
|
[
"def",
"QA_fetch_stock_xdxr",
"(",
"code",
",",
"format",
"=",
"'pd'",
",",
"collections",
"=",
"DATABASE",
".",
"stock_xdxr",
")",
":",
"code",
"=",
"QA_util_code_tolist",
"(",
"code",
")",
"data",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
"}",
",",
"batch_size",
"=",
"10000",
")",
"]",
")",
".",
"drop",
"(",
"[",
"'_id'",
"]",
",",
"axis",
"=",
"1",
")",
"data",
"[",
"'date'",
"]",
"=",
"pd",
".",
"to_datetime",
"(",
"data",
"[",
"'date'",
"]",
")",
"return",
"data",
".",
"set_index",
"(",
"'date'",
",",
"drop",
"=",
"False",
")"
] |
获取股票除权信息/数据库
|
[
"获取股票除权信息",
"/",
"数据库"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L462-L468
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_quotations
|
def QA_fetch_quotations(date=datetime.date.today(), db=DATABASE):
'获取全部实时5档行情的存储结果'
try:
collections = db.get_collection(
'realtime_{}'.format(date))
data = pd.DataFrame([item for item in collections.find(
{}, {"_id": 0}, batch_size=10000)])
return data.assign(date=pd.to_datetime(data.datetime.apply(lambda x: str(x)[0:10]))).assign(datetime=pd.to_datetime(data.datetime)).set_index(['datetime', 'code'], drop=False).sort_index()
except Exception as e:
raise e
|
python
|
def QA_fetch_quotations(date=datetime.date.today(), db=DATABASE):
'获取全部实时5档行情的存储结果'
try:
collections = db.get_collection(
'realtime_{}'.format(date))
data = pd.DataFrame([item for item in collections.find(
{}, {"_id": 0}, batch_size=10000)])
return data.assign(date=pd.to_datetime(data.datetime.apply(lambda x: str(x)[0:10]))).assign(datetime=pd.to_datetime(data.datetime)).set_index(['datetime', 'code'], drop=False).sort_index()
except Exception as e:
raise e
|
[
"def",
"QA_fetch_quotations",
"(",
"date",
"=",
"datetime",
".",
"date",
".",
"today",
"(",
")",
",",
"db",
"=",
"DATABASE",
")",
":",
"try",
":",
"collections",
"=",
"db",
".",
"get_collection",
"(",
"'realtime_{}'",
".",
"format",
"(",
"date",
")",
")",
"data",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"]",
")",
"return",
"data",
".",
"assign",
"(",
"date",
"=",
"pd",
".",
"to_datetime",
"(",
"data",
".",
"datetime",
".",
"apply",
"(",
"lambda",
"x",
":",
"str",
"(",
"x",
")",
"[",
"0",
":",
"10",
"]",
")",
")",
")",
".",
"assign",
"(",
"datetime",
"=",
"pd",
".",
"to_datetime",
"(",
"data",
".",
"datetime",
")",
")",
".",
"set_index",
"(",
"[",
"'datetime'",
",",
"'code'",
"]",
",",
"drop",
"=",
"False",
")",
".",
"sort_index",
"(",
")",
"except",
"Exception",
"as",
"e",
":",
"raise",
"e"
] |
获取全部实时5档行情的存储结果
|
[
"获取全部实时5档行情的存储结果"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L524-L533
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_account
|
def QA_fetch_account(message={}, db=DATABASE):
"""get the account
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
"""
collection = DATABASE.account
return [res for res in collection.find(message, {"_id": 0})]
|
python
|
def QA_fetch_account(message={}, db=DATABASE):
"""get the account
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
"""
collection = DATABASE.account
return [res for res in collection.find(message, {"_id": 0})]
|
[
"def",
"QA_fetch_account",
"(",
"message",
"=",
"{",
"}",
",",
"db",
"=",
"DATABASE",
")",
":",
"collection",
"=",
"DATABASE",
".",
"account",
"return",
"[",
"res",
"for",
"res",
"in",
"collection",
".",
"find",
"(",
"message",
",",
"{",
"\"_id\"",
":",
"0",
"}",
")",
"]"
] |
get the account
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
|
[
"get",
"the",
"account"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L536-L549
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_risk
|
def QA_fetch_risk(message={}, params={"_id": 0, 'assets': 0, 'timeindex': 0, 'totaltimeindex': 0, 'benchmark_assets': 0, 'month_profit': 0}, db=DATABASE):
"""get the risk message
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
"""
collection = DATABASE.risk
return [res for res in collection.find(message, params)]
|
python
|
def QA_fetch_risk(message={}, params={"_id": 0, 'assets': 0, 'timeindex': 0, 'totaltimeindex': 0, 'benchmark_assets': 0, 'month_profit': 0}, db=DATABASE):
"""get the risk message
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
"""
collection = DATABASE.risk
return [res for res in collection.find(message, params)]
|
[
"def",
"QA_fetch_risk",
"(",
"message",
"=",
"{",
"}",
",",
"params",
"=",
"{",
"\"_id\"",
":",
"0",
",",
"'assets'",
":",
"0",
",",
"'timeindex'",
":",
"0",
",",
"'totaltimeindex'",
":",
"0",
",",
"'benchmark_assets'",
":",
"0",
",",
"'month_profit'",
":",
"0",
"}",
",",
"db",
"=",
"DATABASE",
")",
":",
"collection",
"=",
"DATABASE",
".",
"risk",
"return",
"[",
"res",
"for",
"res",
"in",
"collection",
".",
"find",
"(",
"message",
",",
"params",
")",
"]"
] |
get the risk message
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
|
[
"get",
"the",
"risk",
"message"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L552-L565
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_user
|
def QA_fetch_user(user_cookie, db=DATABASE):
"""
get the user
Arguments:
user_cookie : str the unique cookie_id for a user
Keyword Arguments:
db: database for query
Returns:
list --- [ACCOUNT]
"""
collection = DATABASE.account
return [res for res in collection.find({'user_cookie': user_cookie}, {"_id": 0})]
|
python
|
def QA_fetch_user(user_cookie, db=DATABASE):
"""
get the user
Arguments:
user_cookie : str the unique cookie_id for a user
Keyword Arguments:
db: database for query
Returns:
list --- [ACCOUNT]
"""
collection = DATABASE.account
return [res for res in collection.find({'user_cookie': user_cookie}, {"_id": 0})]
|
[
"def",
"QA_fetch_user",
"(",
"user_cookie",
",",
"db",
"=",
"DATABASE",
")",
":",
"collection",
"=",
"DATABASE",
".",
"account",
"return",
"[",
"res",
"for",
"res",
"in",
"collection",
".",
"find",
"(",
"{",
"'user_cookie'",
":",
"user_cookie",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
")",
"]"
] |
get the user
Arguments:
user_cookie : str the unique cookie_id for a user
Keyword Arguments:
db: database for query
Returns:
list --- [ACCOUNT]
|
[
"get",
"the",
"user"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L568-L582
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_strategy
|
def QA_fetch_strategy(message={}, db=DATABASE):
"""get the account
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
"""
collection = DATABASE.strategy
return [res for res in collection.find(message, {"_id": 0})]
|
python
|
def QA_fetch_strategy(message={}, db=DATABASE):
"""get the account
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
"""
collection = DATABASE.strategy
return [res for res in collection.find(message, {"_id": 0})]
|
[
"def",
"QA_fetch_strategy",
"(",
"message",
"=",
"{",
"}",
",",
"db",
"=",
"DATABASE",
")",
":",
"collection",
"=",
"DATABASE",
".",
"strategy",
"return",
"[",
"res",
"for",
"res",
"in",
"collection",
".",
"find",
"(",
"message",
",",
"{",
"\"_id\"",
":",
"0",
"}",
")",
"]"
] |
get the account
Arguments:
query_mes {[type]} -- [description]
Keyword Arguments:
collection {[type]} -- [description] (default: {DATABASE})
Returns:
[type] -- [description]
|
[
"get",
"the",
"account"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L585-L598
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_lhb
|
def QA_fetch_lhb(date, db=DATABASE):
'获取某一天龙虎榜数据'
try:
collections = db.lhb
return pd.DataFrame([item for item in collections.find(
{'date': date}, {"_id": 0})]).set_index('code', drop=False).sort_index()
except Exception as e:
raise e
|
python
|
def QA_fetch_lhb(date, db=DATABASE):
'获取某一天龙虎榜数据'
try:
collections = db.lhb
return pd.DataFrame([item for item in collections.find(
{'date': date}, {"_id": 0})]).set_index('code', drop=False).sort_index()
except Exception as e:
raise e
|
[
"def",
"QA_fetch_lhb",
"(",
"date",
",",
"db",
"=",
"DATABASE",
")",
":",
"try",
":",
"collections",
"=",
"db",
".",
"lhb",
"return",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"'date'",
":",
"date",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
")",
"]",
")",
".",
"set_index",
"(",
"'code'",
",",
"drop",
"=",
"False",
")",
".",
"sort_index",
"(",
")",
"except",
"Exception",
"as",
"e",
":",
"raise",
"e"
] |
获取某一天龙虎榜数据
|
[
"获取某一天龙虎榜数据"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L601-L608
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_financial_report
|
def QA_fetch_financial_report(code, report_date, ltype='EN', db=DATABASE):
"""获取专业财务报表
Arguments:
code {[type]} -- [description]
report_date {[type]} -- [description]
Keyword Arguments:
ltype {str} -- [description] (default: {'EN'})
db {[type]} -- [description] (default: {DATABASE})
Raises:
e -- [description]
Returns:
pd.DataFrame -- [description]
"""
if isinstance(code, str):
code = [code]
if isinstance(report_date, str):
report_date = [QA_util_date_str2int(report_date)]
elif isinstance(report_date, int):
report_date = [report_date]
elif isinstance(report_date, list):
report_date = [QA_util_date_str2int(item) for item in report_date]
collection = db.financial
num_columns = [item[:3] for item in list(financial_dict.keys())]
CH_columns = [item[3:] for item in list(financial_dict.keys())]
EN_columns = list(financial_dict.values())
#num_columns.extend(['283', '_id', 'code', 'report_date'])
# CH_columns.extend(['283', '_id', 'code', 'report_date'])
#CH_columns = pd.Index(CH_columns)
#EN_columns = list(financial_dict.values())
#EN_columns.extend(['283', '_id', 'code', 'report_date'])
#EN_columns = pd.Index(EN_columns)
try:
if code is not None and report_date is not None:
data = [item for item in collection.find(
{'code': {'$in': code}, 'report_date': {'$in': report_date}}, {"_id": 0}, batch_size=10000)]
elif code is None and report_date is not None:
data = [item for item in collection.find(
{'report_date': {'$in': report_date}}, {"_id": 0}, batch_size=10000)]
elif code is not None and report_date is None:
data = [item for item in collection.find(
{'code': {'$in': code}}, {"_id": 0}, batch_size=10000)]
else:
data = [item for item in collection.find({}, {"_id": 0})]
if len(data) > 0:
res_pd = pd.DataFrame(data)
if ltype in ['CH', 'CN']:
cndict = dict(zip(num_columns, CH_columns))
cndict['283'] = '283'
try:
cndict['284'] = '284'
cndict['285'] = '285'
cndict['286'] = '286'
except:
pass
cndict['code'] = 'code'
cndict['report_date'] = 'report_date'
res_pd.columns = res_pd.columns.map(lambda x: cndict[x])
elif ltype is 'EN':
endict = dict(zip(num_columns, EN_columns))
endict['283'] = '283'
try:
endict['284'] = '284'
endict['285'] = '285'
endict['286'] = '286'
except:
pass
endict['code'] = 'code'
endict['report_date'] = 'report_date'
res_pd.columns = res_pd.columns.map(lambda x: endict[x])
if res_pd.report_date.dtype == numpy.int64:
res_pd.report_date = pd.to_datetime(
res_pd.report_date.apply(QA_util_date_int2str))
else:
res_pd.report_date = pd.to_datetime(res_pd.report_date)
return res_pd.replace(-4.039810335e+34, numpy.nan).set_index(['report_date', 'code'], drop=False)
else:
return None
except Exception as e:
raise e
|
python
|
def QA_fetch_financial_report(code, report_date, ltype='EN', db=DATABASE):
"""获取专业财务报表
Arguments:
code {[type]} -- [description]
report_date {[type]} -- [description]
Keyword Arguments:
ltype {str} -- [description] (default: {'EN'})
db {[type]} -- [description] (default: {DATABASE})
Raises:
e -- [description]
Returns:
pd.DataFrame -- [description]
"""
if isinstance(code, str):
code = [code]
if isinstance(report_date, str):
report_date = [QA_util_date_str2int(report_date)]
elif isinstance(report_date, int):
report_date = [report_date]
elif isinstance(report_date, list):
report_date = [QA_util_date_str2int(item) for item in report_date]
collection = db.financial
num_columns = [item[:3] for item in list(financial_dict.keys())]
CH_columns = [item[3:] for item in list(financial_dict.keys())]
EN_columns = list(financial_dict.values())
#num_columns.extend(['283', '_id', 'code', 'report_date'])
# CH_columns.extend(['283', '_id', 'code', 'report_date'])
#CH_columns = pd.Index(CH_columns)
#EN_columns = list(financial_dict.values())
#EN_columns.extend(['283', '_id', 'code', 'report_date'])
#EN_columns = pd.Index(EN_columns)
try:
if code is not None and report_date is not None:
data = [item for item in collection.find(
{'code': {'$in': code}, 'report_date': {'$in': report_date}}, {"_id": 0}, batch_size=10000)]
elif code is None and report_date is not None:
data = [item for item in collection.find(
{'report_date': {'$in': report_date}}, {"_id": 0}, batch_size=10000)]
elif code is not None and report_date is None:
data = [item for item in collection.find(
{'code': {'$in': code}}, {"_id": 0}, batch_size=10000)]
else:
data = [item for item in collection.find({}, {"_id": 0})]
if len(data) > 0:
res_pd = pd.DataFrame(data)
if ltype in ['CH', 'CN']:
cndict = dict(zip(num_columns, CH_columns))
cndict['283'] = '283'
try:
cndict['284'] = '284'
cndict['285'] = '285'
cndict['286'] = '286'
except:
pass
cndict['code'] = 'code'
cndict['report_date'] = 'report_date'
res_pd.columns = res_pd.columns.map(lambda x: cndict[x])
elif ltype is 'EN':
endict = dict(zip(num_columns, EN_columns))
endict['283'] = '283'
try:
endict['284'] = '284'
endict['285'] = '285'
endict['286'] = '286'
except:
pass
endict['code'] = 'code'
endict['report_date'] = 'report_date'
res_pd.columns = res_pd.columns.map(lambda x: endict[x])
if res_pd.report_date.dtype == numpy.int64:
res_pd.report_date = pd.to_datetime(
res_pd.report_date.apply(QA_util_date_int2str))
else:
res_pd.report_date = pd.to_datetime(res_pd.report_date)
return res_pd.replace(-4.039810335e+34, numpy.nan).set_index(['report_date', 'code'], drop=False)
else:
return None
except Exception as e:
raise e
|
[
"def",
"QA_fetch_financial_report",
"(",
"code",
",",
"report_date",
",",
"ltype",
"=",
"'EN'",
",",
"db",
"=",
"DATABASE",
")",
":",
"if",
"isinstance",
"(",
"code",
",",
"str",
")",
":",
"code",
"=",
"[",
"code",
"]",
"if",
"isinstance",
"(",
"report_date",
",",
"str",
")",
":",
"report_date",
"=",
"[",
"QA_util_date_str2int",
"(",
"report_date",
")",
"]",
"elif",
"isinstance",
"(",
"report_date",
",",
"int",
")",
":",
"report_date",
"=",
"[",
"report_date",
"]",
"elif",
"isinstance",
"(",
"report_date",
",",
"list",
")",
":",
"report_date",
"=",
"[",
"QA_util_date_str2int",
"(",
"item",
")",
"for",
"item",
"in",
"report_date",
"]",
"collection",
"=",
"db",
".",
"financial",
"num_columns",
"=",
"[",
"item",
"[",
":",
"3",
"]",
"for",
"item",
"in",
"list",
"(",
"financial_dict",
".",
"keys",
"(",
")",
")",
"]",
"CH_columns",
"=",
"[",
"item",
"[",
"3",
":",
"]",
"for",
"item",
"in",
"list",
"(",
"financial_dict",
".",
"keys",
"(",
")",
")",
"]",
"EN_columns",
"=",
"list",
"(",
"financial_dict",
".",
"values",
"(",
")",
")",
"#num_columns.extend(['283', '_id', 'code', 'report_date'])",
"# CH_columns.extend(['283', '_id', 'code', 'report_date'])",
"#CH_columns = pd.Index(CH_columns)",
"#EN_columns = list(financial_dict.values())",
"#EN_columns.extend(['283', '_id', 'code', 'report_date'])",
"#EN_columns = pd.Index(EN_columns)",
"try",
":",
"if",
"code",
"is",
"not",
"None",
"and",
"report_date",
"is",
"not",
"None",
":",
"data",
"=",
"[",
"item",
"for",
"item",
"in",
"collection",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
",",
"'report_date'",
":",
"{",
"'$in'",
":",
"report_date",
"}",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"]",
"elif",
"code",
"is",
"None",
"and",
"report_date",
"is",
"not",
"None",
":",
"data",
"=",
"[",
"item",
"for",
"item",
"in",
"collection",
".",
"find",
"(",
"{",
"'report_date'",
":",
"{",
"'$in'",
":",
"report_date",
"}",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"]",
"elif",
"code",
"is",
"not",
"None",
"and",
"report_date",
"is",
"None",
":",
"data",
"=",
"[",
"item",
"for",
"item",
"in",
"collection",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"]",
"else",
":",
"data",
"=",
"[",
"item",
"for",
"item",
"in",
"collection",
".",
"find",
"(",
"{",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
")",
"]",
"if",
"len",
"(",
"data",
")",
">",
"0",
":",
"res_pd",
"=",
"pd",
".",
"DataFrame",
"(",
"data",
")",
"if",
"ltype",
"in",
"[",
"'CH'",
",",
"'CN'",
"]",
":",
"cndict",
"=",
"dict",
"(",
"zip",
"(",
"num_columns",
",",
"CH_columns",
")",
")",
"cndict",
"[",
"'283'",
"]",
"=",
"'283'",
"try",
":",
"cndict",
"[",
"'284'",
"]",
"=",
"'284'",
"cndict",
"[",
"'285'",
"]",
"=",
"'285'",
"cndict",
"[",
"'286'",
"]",
"=",
"'286'",
"except",
":",
"pass",
"cndict",
"[",
"'code'",
"]",
"=",
"'code'",
"cndict",
"[",
"'report_date'",
"]",
"=",
"'report_date'",
"res_pd",
".",
"columns",
"=",
"res_pd",
".",
"columns",
".",
"map",
"(",
"lambda",
"x",
":",
"cndict",
"[",
"x",
"]",
")",
"elif",
"ltype",
"is",
"'EN'",
":",
"endict",
"=",
"dict",
"(",
"zip",
"(",
"num_columns",
",",
"EN_columns",
")",
")",
"endict",
"[",
"'283'",
"]",
"=",
"'283'",
"try",
":",
"endict",
"[",
"'284'",
"]",
"=",
"'284'",
"endict",
"[",
"'285'",
"]",
"=",
"'285'",
"endict",
"[",
"'286'",
"]",
"=",
"'286'",
"except",
":",
"pass",
"endict",
"[",
"'code'",
"]",
"=",
"'code'",
"endict",
"[",
"'report_date'",
"]",
"=",
"'report_date'",
"res_pd",
".",
"columns",
"=",
"res_pd",
".",
"columns",
".",
"map",
"(",
"lambda",
"x",
":",
"endict",
"[",
"x",
"]",
")",
"if",
"res_pd",
".",
"report_date",
".",
"dtype",
"==",
"numpy",
".",
"int64",
":",
"res_pd",
".",
"report_date",
"=",
"pd",
".",
"to_datetime",
"(",
"res_pd",
".",
"report_date",
".",
"apply",
"(",
"QA_util_date_int2str",
")",
")",
"else",
":",
"res_pd",
".",
"report_date",
"=",
"pd",
".",
"to_datetime",
"(",
"res_pd",
".",
"report_date",
")",
"return",
"res_pd",
".",
"replace",
"(",
"-",
"4.039810335e+34",
",",
"numpy",
".",
"nan",
")",
".",
"set_index",
"(",
"[",
"'report_date'",
",",
"'code'",
"]",
",",
"drop",
"=",
"False",
")",
"else",
":",
"return",
"None",
"except",
"Exception",
"as",
"e",
":",
"raise",
"e"
] |
获取专业财务报表
Arguments:
code {[type]} -- [description]
report_date {[type]} -- [description]
Keyword Arguments:
ltype {str} -- [description] (default: {'EN'})
db {[type]} -- [description] (default: {DATABASE})
Raises:
e -- [description]
Returns:
pd.DataFrame -- [description]
|
[
"获取专业财务报表",
"Arguments",
":",
"code",
"{",
"[",
"type",
"]",
"}",
"--",
"[",
"description",
"]",
"report_date",
"{",
"[",
"type",
"]",
"}",
"--",
"[",
"description",
"]",
"Keyword",
"Arguments",
":",
"ltype",
"{",
"str",
"}",
"--",
"[",
"description",
"]",
"(",
"default",
":",
"{",
"EN",
"}",
")",
"db",
"{",
"[",
"type",
"]",
"}",
"--",
"[",
"description",
"]",
"(",
"default",
":",
"{",
"DATABASE",
"}",
")",
"Raises",
":",
"e",
"--",
"[",
"description",
"]",
"Returns",
":",
"pd",
".",
"DataFrame",
"--",
"[",
"description",
"]"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L713-L801
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery.py
|
QA_fetch_stock_divyield
|
def QA_fetch_stock_divyield(code, start, end=None, format='pd', collections=DATABASE.stock_divyield):
'获取股票日线'
#code= [code] if isinstance(code,str) else code
# code checking
code = QA_util_code_tolist(code)
if QA_util_date_valid(end):
__data = []
cursor = collections.find({
'a_stockcode': {'$in': code}, "dir_dcl_date": {
"$lte": end,
"$gte": start}}, {"_id": 0}, batch_size=10000)
#res=[QA_util_dict_remove_key(data, '_id') for data in cursor]
res = pd.DataFrame([item for item in cursor])
try:
res = res.drop_duplicates(
(['dir_dcl_date', 'a_stockcode']))
res = res.ix[:, ['a_stockcode', 'a_stocksname', 'div_info', 'div_type_code', 'bonus_shr',
'cash_bt', 'cap_shr', 'epsp', 'ps_cr', 'ps_up', 'reg_date', 'dir_dcl_date',
'a_stockcode1', 'ex_divi_date', 'prg']]
except:
res = None
if format in ['P', 'p', 'pandas', 'pd']:
return res
elif format in ['json', 'dict']:
return QA_util_to_json_from_pandas(res)
# 多种数据格式
elif format in ['n', 'N', 'numpy']:
return numpy.asarray(res)
elif format in ['list', 'l', 'L']:
return numpy.asarray(res).tolist()
else:
print("QA Error QA_fetch_stock_divyield format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return None
else:
QA_util_log_info(
'QA Error QA_fetch_stock_divyield data parameter start=%s end=%s is not right' % (start, end))
|
python
|
def QA_fetch_stock_divyield(code, start, end=None, format='pd', collections=DATABASE.stock_divyield):
'获取股票日线'
#code= [code] if isinstance(code,str) else code
# code checking
code = QA_util_code_tolist(code)
if QA_util_date_valid(end):
__data = []
cursor = collections.find({
'a_stockcode': {'$in': code}, "dir_dcl_date": {
"$lte": end,
"$gte": start}}, {"_id": 0}, batch_size=10000)
#res=[QA_util_dict_remove_key(data, '_id') for data in cursor]
res = pd.DataFrame([item for item in cursor])
try:
res = res.drop_duplicates(
(['dir_dcl_date', 'a_stockcode']))
res = res.ix[:, ['a_stockcode', 'a_stocksname', 'div_info', 'div_type_code', 'bonus_shr',
'cash_bt', 'cap_shr', 'epsp', 'ps_cr', 'ps_up', 'reg_date', 'dir_dcl_date',
'a_stockcode1', 'ex_divi_date', 'prg']]
except:
res = None
if format in ['P', 'p', 'pandas', 'pd']:
return res
elif format in ['json', 'dict']:
return QA_util_to_json_from_pandas(res)
# 多种数据格式
elif format in ['n', 'N', 'numpy']:
return numpy.asarray(res)
elif format in ['list', 'l', 'L']:
return numpy.asarray(res).tolist()
else:
print("QA Error QA_fetch_stock_divyield format parameter %s is none of \"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\" " % format)
return None
else:
QA_util_log_info(
'QA Error QA_fetch_stock_divyield data parameter start=%s end=%s is not right' % (start, end))
|
[
"def",
"QA_fetch_stock_divyield",
"(",
"code",
",",
"start",
",",
"end",
"=",
"None",
",",
"format",
"=",
"'pd'",
",",
"collections",
"=",
"DATABASE",
".",
"stock_divyield",
")",
":",
"#code= [code] if isinstance(code,str) else code",
"# code checking",
"code",
"=",
"QA_util_code_tolist",
"(",
"code",
")",
"if",
"QA_util_date_valid",
"(",
"end",
")",
":",
"__data",
"=",
"[",
"]",
"cursor",
"=",
"collections",
".",
"find",
"(",
"{",
"'a_stockcode'",
":",
"{",
"'$in'",
":",
"code",
"}",
",",
"\"dir_dcl_date\"",
":",
"{",
"\"$lte\"",
":",
"end",
",",
"\"$gte\"",
":",
"start",
"}",
"}",
",",
"{",
"\"_id\"",
":",
"0",
"}",
",",
"batch_size",
"=",
"10000",
")",
"#res=[QA_util_dict_remove_key(data, '_id') for data in cursor]",
"res",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"cursor",
"]",
")",
"try",
":",
"res",
"=",
"res",
".",
"drop_duplicates",
"(",
"(",
"[",
"'dir_dcl_date'",
",",
"'a_stockcode'",
"]",
")",
")",
"res",
"=",
"res",
".",
"ix",
"[",
":",
",",
"[",
"'a_stockcode'",
",",
"'a_stocksname'",
",",
"'div_info'",
",",
"'div_type_code'",
",",
"'bonus_shr'",
",",
"'cash_bt'",
",",
"'cap_shr'",
",",
"'epsp'",
",",
"'ps_cr'",
",",
"'ps_up'",
",",
"'reg_date'",
",",
"'dir_dcl_date'",
",",
"'a_stockcode1'",
",",
"'ex_divi_date'",
",",
"'prg'",
"]",
"]",
"except",
":",
"res",
"=",
"None",
"if",
"format",
"in",
"[",
"'P'",
",",
"'p'",
",",
"'pandas'",
",",
"'pd'",
"]",
":",
"return",
"res",
"elif",
"format",
"in",
"[",
"'json'",
",",
"'dict'",
"]",
":",
"return",
"QA_util_to_json_from_pandas",
"(",
"res",
")",
"# 多种数据格式",
"elif",
"format",
"in",
"[",
"'n'",
",",
"'N'",
",",
"'numpy'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"res",
")",
"elif",
"format",
"in",
"[",
"'list'",
",",
"'l'",
",",
"'L'",
"]",
":",
"return",
"numpy",
".",
"asarray",
"(",
"res",
")",
".",
"tolist",
"(",
")",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_divyield format parameter %s is none of \\\"P, p, pandas, pd , json, dict , n, N, numpy, list, l, L, !\\\" \"",
"%",
"format",
")",
"return",
"None",
"else",
":",
"QA_util_log_info",
"(",
"'QA Error QA_fetch_stock_divyield data parameter start=%s end=%s is not right'",
"%",
"(",
"start",
",",
"end",
")",
")"
] |
获取股票日线
|
[
"获取股票日线"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery.py#L844-L882
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_day
|
def QA_SU_save_stock_day(client=DATABASE, ui_log=None, ui_progress=None):
'''
save stock_day
保存日线数据
:param client:
:param ui_log: 给GUI qt 界面使用
:param ui_progress: 给GUI qt 界面使用
:param ui_progress_int_value: 给GUI qt 界面使用
'''
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll_stock_day = client.stock_day
coll_stock_day.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll_stock_day):
try:
QA_util_log_info(
'##JOB01 Now Saving STOCK_DAY==== {}'.format(str(code)),
ui_log
)
# 首选查找数据库 是否 有 这个代码的数据
ref = coll_stock_day.find({'code': str(code)[0:6]})
end_date = str(now_time())[0:10]
# 当前数据库已经包含了这个代码的数据, 继续增量更新
# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现
if ref.count() > 0:
# 接着上次获取的日期继续更新
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
'UPDATE_STOCK_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_stock_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
QA_util_get_next_day(start_date),
end_date,
'00'
)
)
)
# 当前数据库中没有这个代码的股票数据, 从1990-01-01 开始下载所有的数据
else:
start_date = '1990-01-01'
QA_util_log_info(
'UPDATE_STOCK_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_stock_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
start_date,
end_date,
'00'
)
)
)
except Exception as error0:
print(error0)
err.append(str(code))
for item in range(len(stock_list)):
QA_util_log_info('The {} of Total {}'.format(item, len(stock_list)))
strProgressToLog = 'DOWNLOAD PROGRESS {} {}'.format(
str(float(item / len(stock_list) * 100))[0:4] + '%',
ui_log
)
intProgressToLog = int(float(item / len(stock_list) * 100))
QA_util_log_info(
strProgressToLog,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgressToLog
)
__saving_work(stock_list[item], coll_stock_day)
if len(err) < 1:
QA_util_log_info('SUCCESS save stock day ^_^', ui_log)
else:
QA_util_log_info('ERROR CODE \n ', ui_log)
QA_util_log_info(err, ui_log)
|
python
|
def QA_SU_save_stock_day(client=DATABASE, ui_log=None, ui_progress=None):
'''
save stock_day
保存日线数据
:param client:
:param ui_log: 给GUI qt 界面使用
:param ui_progress: 给GUI qt 界面使用
:param ui_progress_int_value: 给GUI qt 界面使用
'''
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll_stock_day = client.stock_day
coll_stock_day.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll_stock_day):
try:
QA_util_log_info(
'##JOB01 Now Saving STOCK_DAY==== {}'.format(str(code)),
ui_log
)
# 首选查找数据库 是否 有 这个代码的数据
ref = coll_stock_day.find({'code': str(code)[0:6]})
end_date = str(now_time())[0:10]
# 当前数据库已经包含了这个代码的数据, 继续增量更新
# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现
if ref.count() > 0:
# 接着上次获取的日期继续更新
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
'UPDATE_STOCK_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_stock_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
QA_util_get_next_day(start_date),
end_date,
'00'
)
)
)
# 当前数据库中没有这个代码的股票数据, 从1990-01-01 开始下载所有的数据
else:
start_date = '1990-01-01'
QA_util_log_info(
'UPDATE_STOCK_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_stock_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
start_date,
end_date,
'00'
)
)
)
except Exception as error0:
print(error0)
err.append(str(code))
for item in range(len(stock_list)):
QA_util_log_info('The {} of Total {}'.format(item, len(stock_list)))
strProgressToLog = 'DOWNLOAD PROGRESS {} {}'.format(
str(float(item / len(stock_list) * 100))[0:4] + '%',
ui_log
)
intProgressToLog = int(float(item / len(stock_list) * 100))
QA_util_log_info(
strProgressToLog,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgressToLog
)
__saving_work(stock_list[item], coll_stock_day)
if len(err) < 1:
QA_util_log_info('SUCCESS save stock day ^_^', ui_log)
else:
QA_util_log_info('ERROR CODE \n ', ui_log)
QA_util_log_info(err, ui_log)
|
[
"def",
"QA_SU_save_stock_day",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"stock_list",
"=",
"QA_fetch_get_stock_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"coll_stock_day",
"=",
"client",
".",
"stock_day",
"coll_stock_day",
".",
"create_index",
"(",
"[",
"(",
"\"code\"",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"\"date_stamp\"",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll_stock_day",
")",
":",
"try",
":",
"QA_util_log_info",
"(",
"'##JOB01 Now Saving STOCK_DAY==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
")",
"# 首选查找数据库 是否 有 这个代码的数据",
"ref",
"=",
"coll_stock_day",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"6",
"]",
"}",
")",
"end_date",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"10",
"]",
"# 当前数据库已经包含了这个代码的数据, 继续增量更新",
"# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现",
"if",
"ref",
".",
"count",
"(",
")",
">",
"0",
":",
"# 接着上次获取的日期继续更新",
"start_date",
"=",
"ref",
"[",
"ref",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'date'",
"]",
"QA_util_log_info",
"(",
"'UPDATE_STOCK_DAY \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"coll_stock_day",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_day",
"(",
"str",
"(",
"code",
")",
",",
"QA_util_get_next_day",
"(",
"start_date",
")",
",",
"end_date",
",",
"'00'",
")",
")",
")",
"# 当前数据库中没有这个代码的股票数据, 从1990-01-01 开始下载所有的数据",
"else",
":",
"start_date",
"=",
"'1990-01-01'",
"QA_util_log_info",
"(",
"'UPDATE_STOCK_DAY \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"coll_stock_day",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_day",
"(",
"str",
"(",
"code",
")",
",",
"start_date",
",",
"end_date",
",",
"'00'",
")",
")",
")",
"except",
"Exception",
"as",
"error0",
":",
"print",
"(",
"error0",
")",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"for",
"item",
"in",
"range",
"(",
"len",
"(",
"stock_list",
")",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"item",
",",
"len",
"(",
"stock_list",
")",
")",
")",
"strProgressToLog",
"=",
"'DOWNLOAD PROGRESS {} {}'",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
",",
"ui_log",
")",
"intProgressToLog",
"=",
"int",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"QA_util_log_info",
"(",
"strProgressToLog",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intProgressToLog",
")",
"__saving_work",
"(",
"stock_list",
"[",
"item",
"]",
",",
"coll_stock_day",
")",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS save stock day ^_^'",
",",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"'ERROR CODE \\n '",
",",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
")"
] |
save stock_day
保存日线数据
:param client:
:param ui_log: 给GUI qt 界面使用
:param ui_progress: 给GUI qt 界面使用
:param ui_progress_int_value: 给GUI qt 界面使用
|
[
"save",
"stock_day",
"保存日线数据",
":",
"param",
"client",
":",
":",
"param",
"ui_log",
":",
"给GUI",
"qt",
"界面使用",
":",
"param",
"ui_progress",
":",
"给GUI",
"qt",
"界面使用",
":",
"param",
"ui_progress_int_value",
":",
"给GUI",
"qt",
"界面使用"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L82-L184
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_week
|
def QA_SU_save_stock_week(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_week
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll_stock_week = client.stock_week
coll_stock_week.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll_stock_week):
try:
QA_util_log_info(
'##JOB01 Now Saving STOCK_WEEK==== {}'.format(str(code)),
ui_log=ui_log
)
ref = coll_stock_week.find({'code': str(code)[0:6]})
end_date = str(now_time())[0:10]
if ref.count() > 0:
# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
'UPDATE_STOCK_WEEK \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
coll_stock_week.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
QA_util_get_next_day(start_date),
end_date,
'00',
frequence='week'
)
)
)
else:
start_date = '1990-01-01'
QA_util_log_info(
'UPDATE_STOCK_WEEK \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
coll_stock_week.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
start_date,
end_date,
'00',
frequence='week'
)
)
)
except:
err.append(str(code))
for item in range(len(stock_list)):
QA_util_log_info(
'The {} of Total {}'.format(item,
len(stock_list)),
ui_log=ui_log
)
strProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(item / len(stock_list) * 100))[0:4] + '%'
)
intProgress = int(float(item / len(stock_list) * 100))
QA_util_log_info(
strProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgress
)
__saving_work(stock_list[item], coll_stock_week)
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_stock_week(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_week
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll_stock_week = client.stock_week
coll_stock_week.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll_stock_week):
try:
QA_util_log_info(
'##JOB01 Now Saving STOCK_WEEK==== {}'.format(str(code)),
ui_log=ui_log
)
ref = coll_stock_week.find({'code': str(code)[0:6]})
end_date = str(now_time())[0:10]
if ref.count() > 0:
# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
'UPDATE_STOCK_WEEK \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
coll_stock_week.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
QA_util_get_next_day(start_date),
end_date,
'00',
frequence='week'
)
)
)
else:
start_date = '1990-01-01'
QA_util_log_info(
'UPDATE_STOCK_WEEK \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
coll_stock_week.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_stock_day(
str(code),
start_date,
end_date,
'00',
frequence='week'
)
)
)
except:
err.append(str(code))
for item in range(len(stock_list)):
QA_util_log_info(
'The {} of Total {}'.format(item,
len(stock_list)),
ui_log=ui_log
)
strProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(item / len(stock_list) * 100))[0:4] + '%'
)
intProgress = int(float(item / len(stock_list) * 100))
QA_util_log_info(
strProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgress
)
__saving_work(stock_list[item], coll_stock_week)
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_stock_week",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"stock_list",
"=",
"QA_fetch_get_stock_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"coll_stock_week",
"=",
"client",
".",
"stock_week",
"coll_stock_week",
".",
"create_index",
"(",
"[",
"(",
"\"code\"",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"\"date_stamp\"",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll_stock_week",
")",
":",
"try",
":",
"QA_util_log_info",
"(",
"'##JOB01 Now Saving STOCK_WEEK==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"ref",
"=",
"coll_stock_week",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"6",
"]",
"}",
")",
"end_date",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"10",
"]",
"if",
"ref",
".",
"count",
"(",
")",
">",
"0",
":",
"# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现",
"start_date",
"=",
"ref",
"[",
"ref",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'date'",
"]",
"QA_util_log_info",
"(",
"'UPDATE_STOCK_WEEK \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"coll_stock_week",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_day",
"(",
"str",
"(",
"code",
")",
",",
"QA_util_get_next_day",
"(",
"start_date",
")",
",",
"end_date",
",",
"'00'",
",",
"frequence",
"=",
"'week'",
")",
")",
")",
"else",
":",
"start_date",
"=",
"'1990-01-01'",
"QA_util_log_info",
"(",
"'UPDATE_STOCK_WEEK \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"coll_stock_week",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_day",
"(",
"str",
"(",
"code",
")",
",",
"start_date",
",",
"end_date",
",",
"'00'",
",",
"frequence",
"=",
"'week'",
")",
")",
")",
"except",
":",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"for",
"item",
"in",
"range",
"(",
"len",
"(",
"stock_list",
")",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"item",
",",
"len",
"(",
"stock_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intProgress",
"=",
"int",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"QA_util_log_info",
"(",
"strProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intProgress",
")",
"__saving_work",
"(",
"stock_list",
"[",
"item",
"]",
",",
"coll_stock_week",
")",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
save stock_week
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_week"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L196-L292
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_xdxr
|
def QA_SU_save_stock_xdxr(client=DATABASE, ui_log=None, ui_progress=None):
"""[summary]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
# client.drop_collection('stock_xdxr')
try:
coll = client.stock_xdxr
coll.create_index(
[('code',
pymongo.ASCENDING),
('date',
pymongo.ASCENDING)],
unique=True
)
except:
client.drop_collection('stock_xdxr')
coll = client.stock_xdxr
coll.create_index(
[('code',
pymongo.ASCENDING),
('date',
pymongo.ASCENDING)],
unique=True
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB02 Now Saving XDXR INFO ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_xdxr(str(code))),
ordered=False
)
except:
err.append(str(code))
for i_ in range(len(stock_list)):
QA_util_log_info(
'The {} of Total {}'.format(i_,
len(stock_list)),
ui_log=ui_log
)
strLogInfo = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(stock_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(stock_list) * 100))
QA_util_log_info(
strLogInfo,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(stock_list[i_], coll)
|
python
|
def QA_SU_save_stock_xdxr(client=DATABASE, ui_log=None, ui_progress=None):
"""[summary]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
# client.drop_collection('stock_xdxr')
try:
coll = client.stock_xdxr
coll.create_index(
[('code',
pymongo.ASCENDING),
('date',
pymongo.ASCENDING)],
unique=True
)
except:
client.drop_collection('stock_xdxr')
coll = client.stock_xdxr
coll.create_index(
[('code',
pymongo.ASCENDING),
('date',
pymongo.ASCENDING)],
unique=True
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB02 Now Saving XDXR INFO ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_xdxr(str(code))),
ordered=False
)
except:
err.append(str(code))
for i_ in range(len(stock_list)):
QA_util_log_info(
'The {} of Total {}'.format(i_,
len(stock_list)),
ui_log=ui_log
)
strLogInfo = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(stock_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(stock_list) * 100))
QA_util_log_info(
strLogInfo,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(stock_list[i_], coll)
|
[
"def",
"QA_SU_save_stock_xdxr",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"stock_list",
"=",
"QA_fetch_get_stock_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"# client.drop_collection('stock_xdxr')",
"try",
":",
"coll",
"=",
"client",
".",
"stock_xdxr",
"coll",
".",
"create_index",
"(",
"[",
"(",
"'code'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'date'",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
",",
"unique",
"=",
"True",
")",
"except",
":",
"client",
".",
"drop_collection",
"(",
"'stock_xdxr'",
")",
"coll",
"=",
"client",
".",
"stock_xdxr",
"coll",
".",
"create_index",
"(",
"[",
"(",
"'code'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'date'",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
",",
"unique",
"=",
"True",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll",
")",
":",
"QA_util_log_info",
"(",
"'##JOB02 Now Saving XDXR INFO ==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"try",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_xdxr",
"(",
"str",
"(",
"code",
")",
")",
")",
",",
"ordered",
"=",
"False",
")",
"except",
":",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"stock_list",
")",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"i_",
",",
"len",
"(",
"stock_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strLogInfo",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"QA_util_log_info",
"(",
"strLogInfo",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"__saving_work",
"(",
"stock_list",
"[",
"i_",
"]",
",",
"coll",
")"
] |
[summary]
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"[",
"summary",
"]"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L494-L555
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_min
|
def QA_SU_save_stock_min(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll = client.stock_min
coll.create_index(
[
('code',
pymongo.ASCENDING),
('time_stamp',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)
]
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB03 Now Saving STOCK_MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:6], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB03.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_stock_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)[1::]
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB03.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_stock_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
err.append(code)
QA_util_log_info(err, ui_log=ui_log)
executor = ThreadPoolExecutor(max_workers=4)
# executor.map((__saving_work, stock_list[i_], coll),URLS)
res = {
executor.submit(__saving_work,
stock_list[i_],
coll)
for i_ in range(len(stock_list))
}
count = 0
for i_ in concurrent.futures.as_completed(res):
QA_util_log_info(
'The {} of Total {}'.format(count,
len(stock_list)),
ui_log=ui_log
)
strProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(stock_list) * 100))[0:4] + '%'
)
intProgress = int(count / len(stock_list) * 10000.0)
QA_util_log_info(
strProgress,
ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_stock_min(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll = client.stock_min
coll.create_index(
[
('code',
pymongo.ASCENDING),
('time_stamp',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)
]
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB03 Now Saving STOCK_MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:6], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB03.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_stock_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)[1::]
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB03.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_stock_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
err.append(code)
QA_util_log_info(err, ui_log=ui_log)
executor = ThreadPoolExecutor(max_workers=4)
# executor.map((__saving_work, stock_list[i_], coll),URLS)
res = {
executor.submit(__saving_work,
stock_list[i_],
coll)
for i_ in range(len(stock_list))
}
count = 0
for i_ in concurrent.futures.as_completed(res):
QA_util_log_info(
'The {} of Total {}'.format(count,
len(stock_list)),
ui_log=ui_log
)
strProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(stock_list) * 100))[0:4] + '%'
)
intProgress = int(count / len(stock_list) * 10000.0)
QA_util_log_info(
strProgress,
ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_stock_min",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"stock_list",
"=",
"QA_fetch_get_stock_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"coll",
"=",
"client",
".",
"stock_min",
"coll",
".",
"create_index",
"(",
"[",
"(",
"'code'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'time_stamp'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'date_stamp'",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll",
")",
":",
"QA_util_log_info",
"(",
"'##JOB03 Now Saving STOCK_MIN ==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"try",
":",
"for",
"type",
"in",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
":",
"ref_",
"=",
"coll",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"6",
"]",
",",
"'type'",
":",
"type",
"}",
")",
"end_time",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"19",
"]",
"if",
"ref_",
".",
"count",
"(",
")",
">",
"0",
":",
"start_time",
"=",
"ref_",
"[",
"ref_",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'datetime'",
"]",
"QA_util_log_info",
"(",
"'##JOB03.{} Now Saving {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_stock_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
")",
"[",
"1",
":",
":",
"]",
")",
"else",
":",
"start_time",
"=",
"'2015-01-01'",
"QA_util_log_info",
"(",
"'##JOB03.{} Now Saving {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_stock_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
")",
")",
"except",
"Exception",
"as",
"e",
":",
"QA_util_log_info",
"(",
"e",
",",
"ui_log",
"=",
"ui_log",
")",
"err",
".",
"append",
"(",
"code",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")",
"executor",
"=",
"ThreadPoolExecutor",
"(",
"max_workers",
"=",
"4",
")",
"# executor.map((__saving_work, stock_list[i_], coll),URLS)",
"res",
"=",
"{",
"executor",
".",
"submit",
"(",
"__saving_work",
",",
"stock_list",
"[",
"i_",
"]",
",",
"coll",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"stock_list",
")",
")",
"}",
"count",
"=",
"0",
"for",
"i_",
"in",
"concurrent",
".",
"futures",
".",
"as_completed",
"(",
"res",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"count",
",",
"len",
"(",
"stock_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"count",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intProgress",
"=",
"int",
"(",
"count",
"/",
"len",
"(",
"stock_list",
")",
"*",
"10000.0",
")",
"QA_util_log_info",
"(",
"strProgress",
",",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intProgress",
")",
"count",
"=",
"count",
"+",
"1",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
save stock_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_min"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L558-L679
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_index_day
|
def QA_SU_save_index_day(client=DATABASE, ui_log=None, ui_progress=None):
"""save index_day
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
__index_list = QA_fetch_get_stock_list('index')
coll = client.index_day
coll.create_index(
[('code',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll):
try:
ref_ = coll.find({'code': str(code)[0:6]})
end_time = str(now_time())[0:10]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['date']
QA_util_log_info(
'##JOB04 Now Saving INDEX_DAY==== \n Trying updating {} from {} to {}'
.format(code,
start_time,
end_time),
ui_log=ui_log
)
if start_time != end_time:
coll.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_index_day(
str(code),
QA_util_get_next_day(start_time),
end_time
)
)
)
else:
try:
start_time = '1990-01-01'
QA_util_log_info(
'##JOB04 Now Saving INDEX_DAY==== \n Trying updating {} from {} to {}'
.format(code,
start_time,
end_time),
ui_log=ui_log
)
coll.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_index_day(
str(code),
start_time,
end_time
)
)
)
except:
start_time = '2009-01-01'
QA_util_log_info(
'##JOB04 Now Saving INDEX_DAY==== \n Trying updating {} from {} to {}'
.format(code,
start_time,
end_time),
ui_log=ui_log
)
coll.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_index_day(
str(code),
start_time,
end_time
)
)
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
err.append(str(code))
QA_util_log_info(err, ui_log=ui_log)
for i_ in range(len(__index_list)):
# __saving_work('000001')
QA_util_log_info(
'The {} of Total {}'.format(i_,
len(__index_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(__index_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(__index_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(__index_list.index[i_][0], coll)
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_index_day(client=DATABASE, ui_log=None, ui_progress=None):
"""save index_day
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
__index_list = QA_fetch_get_stock_list('index')
coll = client.index_day
coll.create_index(
[('code',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll):
try:
ref_ = coll.find({'code': str(code)[0:6]})
end_time = str(now_time())[0:10]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['date']
QA_util_log_info(
'##JOB04 Now Saving INDEX_DAY==== \n Trying updating {} from {} to {}'
.format(code,
start_time,
end_time),
ui_log=ui_log
)
if start_time != end_time:
coll.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_index_day(
str(code),
QA_util_get_next_day(start_time),
end_time
)
)
)
else:
try:
start_time = '1990-01-01'
QA_util_log_info(
'##JOB04 Now Saving INDEX_DAY==== \n Trying updating {} from {} to {}'
.format(code,
start_time,
end_time),
ui_log=ui_log
)
coll.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_index_day(
str(code),
start_time,
end_time
)
)
)
except:
start_time = '2009-01-01'
QA_util_log_info(
'##JOB04 Now Saving INDEX_DAY==== \n Trying updating {} from {} to {}'
.format(code,
start_time,
end_time),
ui_log=ui_log
)
coll.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_index_day(
str(code),
start_time,
end_time
)
)
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
err.append(str(code))
QA_util_log_info(err, ui_log=ui_log)
for i_ in range(len(__index_list)):
# __saving_work('000001')
QA_util_log_info(
'The {} of Total {}'.format(i_,
len(__index_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(__index_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(__index_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(__index_list.index[i_][0], coll)
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_index_day",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"__index_list",
"=",
"QA_fetch_get_stock_list",
"(",
"'index'",
")",
"coll",
"=",
"client",
".",
"index_day",
"coll",
".",
"create_index",
"(",
"[",
"(",
"'code'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'date_stamp'",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll",
")",
":",
"try",
":",
"ref_",
"=",
"coll",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"6",
"]",
"}",
")",
"end_time",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"10",
"]",
"if",
"ref_",
".",
"count",
"(",
")",
">",
"0",
":",
"start_time",
"=",
"ref_",
"[",
"ref_",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'date'",
"]",
"QA_util_log_info",
"(",
"'##JOB04 Now Saving INDEX_DAY==== \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_time",
",",
"end_time",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_index_day",
"(",
"str",
"(",
"code",
")",
",",
"QA_util_get_next_day",
"(",
"start_time",
")",
",",
"end_time",
")",
")",
")",
"else",
":",
"try",
":",
"start_time",
"=",
"'1990-01-01'",
"QA_util_log_info",
"(",
"'##JOB04 Now Saving INDEX_DAY==== \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_time",
",",
"end_time",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_index_day",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
")",
")",
")",
"except",
":",
"start_time",
"=",
"'2009-01-01'",
"QA_util_log_info",
"(",
"'##JOB04 Now Saving INDEX_DAY==== \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_time",
",",
"end_time",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_index_day",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
")",
")",
")",
"except",
"Exception",
"as",
"e",
":",
"QA_util_log_info",
"(",
"e",
",",
"ui_log",
"=",
"ui_log",
")",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"__index_list",
")",
")",
":",
"# __saving_work('000001')",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"i_",
",",
"len",
"(",
"__index_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strLogProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"__index_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"__index_list",
")",
"*",
"10000.0",
")",
")",
"QA_util_log_info",
"(",
"strLogProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"__saving_work",
"(",
"__index_list",
".",
"index",
"[",
"i_",
"]",
"[",
"0",
"]",
",",
"coll",
")",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
save index_day
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"index_day"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L682-L790
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_index_min
|
def QA_SU_save_index_min(client=DATABASE, ui_log=None, ui_progress=None):
"""save index_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
__index_list = QA_fetch_get_stock_list('index')
coll = client.index_min
coll.create_index(
[
('code',
pymongo.ASCENDING),
('time_stamp',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)
]
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB05 Now Saving Index_MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:6], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB05.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_index_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data[1::])
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB05.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_index_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except:
err.append(code)
executor = ThreadPoolExecutor(max_workers=4)
res = {
executor.submit(__saving_work,
__index_list.index[i_][0],
coll)
for i_ in range(len(__index_list))
} # multi index ./.
count = 0
for i_ in concurrent.futures.as_completed(res):
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(__index_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(count / len(__index_list) * 10000.0))
QA_util_log_info(
'The {} of Total {}'.format(count,
len(__index_list)),
ui_log=ui_log
)
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_index_min(client=DATABASE, ui_log=None, ui_progress=None):
"""save index_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
__index_list = QA_fetch_get_stock_list('index')
coll = client.index_min
coll.create_index(
[
('code',
pymongo.ASCENDING),
('time_stamp',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)
]
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB05 Now Saving Index_MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:6], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB05.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_index_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data[1::])
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB05.{} Now Saving {} from {} to {} =={} '.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_index_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except:
err.append(code)
executor = ThreadPoolExecutor(max_workers=4)
res = {
executor.submit(__saving_work,
__index_list.index[i_][0],
coll)
for i_ in range(len(__index_list))
} # multi index ./.
count = 0
for i_ in concurrent.futures.as_completed(res):
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(__index_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(count / len(__index_list) * 10000.0))
QA_util_log_info(
'The {} of Total {}'.format(count,
len(__index_list)),
ui_log=ui_log
)
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_index_min",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"__index_list",
"=",
"QA_fetch_get_stock_list",
"(",
"'index'",
")",
"coll",
"=",
"client",
".",
"index_min",
"coll",
".",
"create_index",
"(",
"[",
"(",
"'code'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'time_stamp'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'date_stamp'",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll",
")",
":",
"QA_util_log_info",
"(",
"'##JOB05 Now Saving Index_MIN ==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"try",
":",
"for",
"type",
"in",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
":",
"ref_",
"=",
"coll",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"6",
"]",
",",
"'type'",
":",
"type",
"}",
")",
"end_time",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"19",
"]",
"if",
"ref_",
".",
"count",
"(",
")",
">",
"0",
":",
"start_time",
"=",
"ref_",
"[",
"ref_",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'datetime'",
"]",
"QA_util_log_info",
"(",
"'##JOB05.{} Now Saving {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_index_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
"[",
"1",
":",
":",
"]",
")",
")",
"else",
":",
"start_time",
"=",
"'2015-01-01'",
"QA_util_log_info",
"(",
"'##JOB05.{} Now Saving {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_index_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
")",
")",
"except",
":",
"err",
".",
"append",
"(",
"code",
")",
"executor",
"=",
"ThreadPoolExecutor",
"(",
"max_workers",
"=",
"4",
")",
"res",
"=",
"{",
"executor",
".",
"submit",
"(",
"__saving_work",
",",
"__index_list",
".",
"index",
"[",
"i_",
"]",
"[",
"0",
"]",
",",
"coll",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"__index_list",
")",
")",
"}",
"# multi index ./.",
"count",
"=",
"0",
"for",
"i_",
"in",
"concurrent",
".",
"futures",
".",
"as_completed",
"(",
"res",
")",
":",
"strLogProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"count",
"/",
"len",
"(",
"__index_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"count",
"/",
"len",
"(",
"__index_list",
")",
"*",
"10000.0",
")",
")",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"count",
",",
"len",
"(",
"__index_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"strLogProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"count",
"=",
"count",
"+",
"1",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
save index_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"index_min"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L793-L916
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_list
|
def QA_SU_save_stock_list(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_list
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
client.drop_collection('stock_list')
coll = client.stock_list
coll.create_index('code')
try:
# 🛠todo 这个应该是第一个任务 JOB01, 先更新股票列表!!
QA_util_log_info(
'##JOB08 Now Saving STOCK_LIST ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
stock_list_from_tdx = QA_fetch_get_stock_list()
pandas_data = QA_util_to_json_from_pandas(stock_list_from_tdx)
coll.insert_many(pandas_data)
QA_util_log_info(
"完成股票列表获取",
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=10000
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
print(" Error save_tdx.QA_SU_save_stock_list exception!")
pass
|
python
|
def QA_SU_save_stock_list(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_list
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
client.drop_collection('stock_list')
coll = client.stock_list
coll.create_index('code')
try:
# 🛠todo 这个应该是第一个任务 JOB01, 先更新股票列表!!
QA_util_log_info(
'##JOB08 Now Saving STOCK_LIST ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
stock_list_from_tdx = QA_fetch_get_stock_list()
pandas_data = QA_util_to_json_from_pandas(stock_list_from_tdx)
coll.insert_many(pandas_data)
QA_util_log_info(
"完成股票列表获取",
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=10000
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
print(" Error save_tdx.QA_SU_save_stock_list exception!")
pass
|
[
"def",
"QA_SU_save_stock_list",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"client",
".",
"drop_collection",
"(",
"'stock_list'",
")",
"coll",
"=",
"client",
".",
"stock_list",
"coll",
".",
"create_index",
"(",
"'code'",
")",
"try",
":",
"# 🛠todo 这个应该是第一个任务 JOB01, 先更新股票列表!!",
"QA_util_log_info",
"(",
"'##JOB08 Now Saving STOCK_LIST ===='",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"5000",
")",
"stock_list_from_tdx",
"=",
"QA_fetch_get_stock_list",
"(",
")",
"pandas_data",
"=",
"QA_util_to_json_from_pandas",
"(",
"stock_list_from_tdx",
")",
"coll",
".",
"insert_many",
"(",
"pandas_data",
")",
"QA_util_log_info",
"(",
"\"完成股票列表获取\",",
"",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"10000",
")",
"except",
"Exception",
"as",
"e",
":",
"QA_util_log_info",
"(",
"e",
",",
"ui_log",
"=",
"ui_log",
")",
"print",
"(",
"\" Error save_tdx.QA_SU_save_stock_list exception!\"",
")",
"pass"
] |
save stock_list
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_list"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L1140-L1171
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_etf_list
|
def QA_SU_save_etf_list(client=DATABASE, ui_log=None, ui_progress=None):
"""save etf_list
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
try:
QA_util_log_info(
'##JOB16 Now Saving ETF_LIST ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
etf_list_from_tdx = QA_fetch_get_stock_list(type_="etf")
pandas_data = QA_util_to_json_from_pandas(etf_list_from_tdx)
if len(pandas_data) > 0:
# 获取到数据后才进行drop collection 操作
client.drop_collection('etf_list')
coll = client.etf_list
coll.create_index('code')
coll.insert_many(pandas_data)
QA_util_log_info(
"完成ETF列表获取",
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=10000
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
print(" Error save_tdx.QA_SU_save_etf_list exception!")
pass
|
python
|
def QA_SU_save_etf_list(client=DATABASE, ui_log=None, ui_progress=None):
"""save etf_list
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
try:
QA_util_log_info(
'##JOB16 Now Saving ETF_LIST ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
etf_list_from_tdx = QA_fetch_get_stock_list(type_="etf")
pandas_data = QA_util_to_json_from_pandas(etf_list_from_tdx)
if len(pandas_data) > 0:
# 获取到数据后才进行drop collection 操作
client.drop_collection('etf_list')
coll = client.etf_list
coll.create_index('code')
coll.insert_many(pandas_data)
QA_util_log_info(
"完成ETF列表获取",
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=10000
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
print(" Error save_tdx.QA_SU_save_etf_list exception!")
pass
|
[
"def",
"QA_SU_save_etf_list",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"try",
":",
"QA_util_log_info",
"(",
"'##JOB16 Now Saving ETF_LIST ===='",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"5000",
")",
"etf_list_from_tdx",
"=",
"QA_fetch_get_stock_list",
"(",
"type_",
"=",
"\"etf\"",
")",
"pandas_data",
"=",
"QA_util_to_json_from_pandas",
"(",
"etf_list_from_tdx",
")",
"if",
"len",
"(",
"pandas_data",
")",
">",
"0",
":",
"# 获取到数据后才进行drop collection 操作",
"client",
".",
"drop_collection",
"(",
"'etf_list'",
")",
"coll",
"=",
"client",
".",
"etf_list",
"coll",
".",
"create_index",
"(",
"'code'",
")",
"coll",
".",
"insert_many",
"(",
"pandas_data",
")",
"QA_util_log_info",
"(",
"\"完成ETF列表获取\",",
"",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"10000",
")",
"except",
"Exception",
"as",
"e",
":",
"QA_util_log_info",
"(",
"e",
",",
"ui_log",
"=",
"ui_log",
")",
"print",
"(",
"\" Error save_tdx.QA_SU_save_etf_list exception!\"",
")",
"pass"
] |
save etf_list
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"etf_list"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L1174-L1205
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_block
|
def QA_SU_save_stock_block(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_block
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
client.drop_collection('stock_block')
coll = client.stock_block
coll.create_index('code')
try:
QA_util_log_info(
'##JOB09 Now Saving STOCK_BlOCK ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_block('tdx'))
)
QA_util_log_info(
'tdx Block ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
# 🛠todo fixhere here 获取同花顺板块, 还是调用tdx的
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_block('ths'))
)
QA_util_log_info(
'ths Block ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=8000
)
QA_util_log_info(
'完成股票板块获取=',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=10000
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
print(" Error save_tdx.QA_SU_save_stock_block exception!")
pass
|
python
|
def QA_SU_save_stock_block(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_block
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
client.drop_collection('stock_block')
coll = client.stock_block
coll.create_index('code')
try:
QA_util_log_info(
'##JOB09 Now Saving STOCK_BlOCK ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_block('tdx'))
)
QA_util_log_info(
'tdx Block ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=5000
)
# 🛠todo fixhere here 获取同花顺板块, 还是调用tdx的
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_block('ths'))
)
QA_util_log_info(
'ths Block ====',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=8000
)
QA_util_log_info(
'完成股票板块获取=',
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=10000
)
except Exception as e:
QA_util_log_info(e, ui_log=ui_log)
print(" Error save_tdx.QA_SU_save_stock_block exception!")
pass
|
[
"def",
"QA_SU_save_stock_block",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"client",
".",
"drop_collection",
"(",
"'stock_block'",
")",
"coll",
"=",
"client",
".",
"stock_block",
"coll",
".",
"create_index",
"(",
"'code'",
")",
"try",
":",
"QA_util_log_info",
"(",
"'##JOB09 Now Saving STOCK_BlOCK ===='",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"5000",
")",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_block",
"(",
"'tdx'",
")",
")",
")",
"QA_util_log_info",
"(",
"'tdx Block ===='",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"5000",
")",
"# 🛠todo fixhere here 获取同花顺板块, 还是调用tdx的",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_block",
"(",
"'ths'",
")",
")",
")",
"QA_util_log_info",
"(",
"'ths Block ===='",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"8000",
")",
"QA_util_log_info",
"(",
"'完成股票板块获取=',",
"",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"10000",
")",
"except",
"Exception",
"as",
"e",
":",
"QA_util_log_info",
"(",
"e",
",",
"ui_log",
"=",
"ui_log",
")",
"print",
"(",
"\" Error save_tdx.QA_SU_save_stock_block exception!\"",
")",
"pass"
] |
save stock_block
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_block"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L1208-L1257
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_info
|
def QA_SU_save_stock_info(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_info
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
client.drop_collection('stock_info')
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll = client.stock_info
coll.create_index('code')
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB10 Now Saving STOCK INFO ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_info(str(code)))
)
except:
err.append(str(code))
for i_ in range(len(stock_list)):
# __saving_work('000001')
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(stock_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(stock_list) * 10000.0))
QA_util_log_info('The {} of Total {}'.format(i_, len(stock_list)))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(stock_list[i_], coll)
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_stock_info(client=DATABASE, ui_log=None, ui_progress=None):
"""save stock_info
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
client.drop_collection('stock_info')
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll = client.stock_info
coll.create_index('code')
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB10 Now Saving STOCK INFO ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
coll.insert_many(
QA_util_to_json_from_pandas(QA_fetch_get_stock_info(str(code)))
)
except:
err.append(str(code))
for i_ in range(len(stock_list)):
# __saving_work('000001')
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(stock_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(stock_list) * 10000.0))
QA_util_log_info('The {} of Total {}'.format(i_, len(stock_list)))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(stock_list[i_], coll)
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_stock_info",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"client",
".",
"drop_collection",
"(",
"'stock_info'",
")",
"stock_list",
"=",
"QA_fetch_get_stock_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"coll",
"=",
"client",
".",
"stock_info",
"coll",
".",
"create_index",
"(",
"'code'",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll",
")",
":",
"QA_util_log_info",
"(",
"'##JOB10 Now Saving STOCK INFO ==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"try",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_stock_info",
"(",
"str",
"(",
"code",
")",
")",
")",
")",
"except",
":",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"stock_list",
")",
")",
":",
"# __saving_work('000001')",
"strLogProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"stock_list",
")",
"*",
"10000.0",
")",
")",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"i_",
",",
"len",
"(",
"stock_list",
")",
")",
")",
"QA_util_log_info",
"(",
"strLogProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"__saving_work",
"(",
"stock_list",
"[",
"i_",
"]",
",",
"coll",
")",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
save stock_info
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_info"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L1260-L1306
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_stock_transaction
|
def QA_SU_save_stock_transaction(
client=DATABASE,
ui_log=None,
ui_progress=None
):
"""save stock_transaction
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll = client.stock_transaction
coll.create_index('code')
err = []
def __saving_work(code):
QA_util_log_info(
'##JOB11 Now Saving STOCK_TRANSACTION ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
coll.insert_many(
QA_util_to_json_from_pandas(
# 🛠todo str(stock_list[code]) 参数不对?
QA_fetch_get_stock_transaction(
str(code),
'1990-01-01',
str(now_time())[0:10]
)
)
)
except:
err.append(str(code))
for i_ in range(len(stock_list)):
# __saving_work('000001')
QA_util_log_info(
'The {} of Total {}'.format(i_,
len(stock_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(stock_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(stock_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(stock_list[i_])
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_stock_transaction(
client=DATABASE,
ui_log=None,
ui_progress=None
):
"""save stock_transaction
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
stock_list = QA_fetch_get_stock_list().code.unique().tolist()
coll = client.stock_transaction
coll.create_index('code')
err = []
def __saving_work(code):
QA_util_log_info(
'##JOB11 Now Saving STOCK_TRANSACTION ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
coll.insert_many(
QA_util_to_json_from_pandas(
# 🛠todo str(stock_list[code]) 参数不对?
QA_fetch_get_stock_transaction(
str(code),
'1990-01-01',
str(now_time())[0:10]
)
)
)
except:
err.append(str(code))
for i_ in range(len(stock_list)):
# __saving_work('000001')
QA_util_log_info(
'The {} of Total {}'.format(i_,
len(stock_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(i_ / len(stock_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(i_ / len(stock_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(stock_list[i_])
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_stock_transaction",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"stock_list",
"=",
"QA_fetch_get_stock_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"coll",
"=",
"client",
".",
"stock_transaction",
"coll",
".",
"create_index",
"(",
"'code'",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
")",
":",
"QA_util_log_info",
"(",
"'##JOB11 Now Saving STOCK_TRANSACTION ==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"try",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"# 🛠todo str(stock_list[code]) 参数不对?",
"QA_fetch_get_stock_transaction",
"(",
"str",
"(",
"code",
")",
",",
"'1990-01-01'",
",",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"10",
"]",
")",
")",
")",
"except",
":",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"stock_list",
")",
")",
":",
"# __saving_work('000001')",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"i_",
",",
"len",
"(",
"stock_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strLogProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"stock_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"i_",
"/",
"len",
"(",
"stock_list",
")",
"*",
"10000.0",
")",
")",
"QA_util_log_info",
"(",
"strLogProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"__saving_work",
"(",
"stock_list",
"[",
"i_",
"]",
")",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
save stock_transaction
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"stock_transaction"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L1309-L1368
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_option_commodity_day
|
def QA_SU_save_option_commodity_day(
client=DATABASE,
ui_log=None,
ui_progress=None
):
'''
:param client:
:return:
'''
_save_option_commodity_cu_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_m_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_sr_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_ru_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_cf_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_c_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
|
python
|
def QA_SU_save_option_commodity_day(
client=DATABASE,
ui_log=None,
ui_progress=None
):
'''
:param client:
:return:
'''
_save_option_commodity_cu_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_m_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_sr_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_ru_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_cf_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_c_day(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
|
[
"def",
"QA_SU_save_option_commodity_day",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"_save_option_commodity_cu_day",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_m_day",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_sr_day",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_ru_day",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_cf_day",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_c_day",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")"
] |
:param client:
:return:
|
[
":",
"param",
"client",
":",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L2292-L2331
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_option_commodity_min
|
def QA_SU_save_option_commodity_min(
client=DATABASE,
ui_log=None,
ui_progress=None
):
'''
:param client:
:return:
'''
# 测试中发现, 一起回去,容易出现错误,每次获取一个品种后 ,更换服务ip继续获取 ?
_save_option_commodity_cu_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_sr_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_m_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_ru_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_cf_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_c_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
|
python
|
def QA_SU_save_option_commodity_min(
client=DATABASE,
ui_log=None,
ui_progress=None
):
'''
:param client:
:return:
'''
# 测试中发现, 一起回去,容易出现错误,每次获取一个品种后 ,更换服务ip继续获取 ?
_save_option_commodity_cu_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_sr_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_m_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_ru_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_cf_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
_save_option_commodity_c_min(
client=client,
ui_log=ui_log,
ui_progress=ui_progress
)
|
[
"def",
"QA_SU_save_option_commodity_min",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"# 测试中发现, 一起回去,容易出现错误,每次获取一个品种后 ,更换服务ip继续获取 ?",
"_save_option_commodity_cu_min",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_sr_min",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_m_min",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_ru_min",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_cf_min",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")",
"_save_option_commodity_c_min",
"(",
"client",
"=",
"client",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
")"
] |
:param client:
:return:
|
[
":",
"param",
"client",
":",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L3384-L3427
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_option_min
|
def QA_SU_save_option_min(client=DATABASE, ui_log=None, ui_progress=None):
'''
:param client:
:return:
'''
option_contract_list = QA_fetch_get_option_contract_time_to_market()
coll_option_min = client.option_day_min
coll_option_min.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
# 索引 code
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB13 Now Saving Option 50ETF MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:8], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB13.{} Now Saving Option 50ETF {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
QA_util_log_info(
" 写入 新增历史合约记录数 {} ".format(len(__data))
)
coll.insert_many(
QA_util_to_json_from_pandas(__data[1::])
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB13.{} Now Option 50ETF {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
QA_util_log_info(
" 写入 新增合约记录数 {} ".format(len(__data))
)
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except:
err.append(code)
executor = ThreadPoolExecutor(max_workers=4)
res = {
executor.submit(
__saving_work,
option_contract_list[i_]["code"],
coll_option_min
)
for i_ in range(len(option_contract_list))
} # multi index ./.
count = 0
for i_ in concurrent.futures.as_completed(res):
QA_util_log_info(
'The {} of Total {}'.format(count,
len(option_contract_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(option_contract_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(count / len(option_contract_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_option_min(client=DATABASE, ui_log=None, ui_progress=None):
'''
:param client:
:return:
'''
option_contract_list = QA_fetch_get_option_contract_time_to_market()
coll_option_min = client.option_day_min
coll_option_min.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
# 索引 code
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB13 Now Saving Option 50ETF MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:8], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB13.{} Now Saving Option 50ETF {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
QA_util_log_info(
" 写入 新增历史合约记录数 {} ".format(len(__data))
)
coll.insert_many(
QA_util_to_json_from_pandas(__data[1::])
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB13.{} Now Option 50ETF {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
QA_util_log_info(
" 写入 新增合约记录数 {} ".format(len(__data))
)
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except:
err.append(code)
executor = ThreadPoolExecutor(max_workers=4)
res = {
executor.submit(
__saving_work,
option_contract_list[i_]["code"],
coll_option_min
)
for i_ in range(len(option_contract_list))
} # multi index ./.
count = 0
for i_ in concurrent.futures.as_completed(res):
QA_util_log_info(
'The {} of Total {}'.format(count,
len(option_contract_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(option_contract_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(count / len(option_contract_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_option_min",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"option_contract_list",
"=",
"QA_fetch_get_option_contract_time_to_market",
"(",
")",
"coll_option_min",
"=",
"client",
".",
"option_day_min",
"coll_option_min",
".",
"create_index",
"(",
"[",
"(",
"\"code\"",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"\"date_stamp\"",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"# 索引 code",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll",
")",
":",
"QA_util_log_info",
"(",
"'##JOB13 Now Saving Option 50ETF MIN ==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"try",
":",
"for",
"type",
"in",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
":",
"ref_",
"=",
"coll",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"8",
"]",
",",
"'type'",
":",
"type",
"}",
")",
"end_time",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"19",
"]",
"if",
"ref_",
".",
"count",
"(",
")",
">",
"0",
":",
"start_time",
"=",
"ref_",
"[",
"ref_",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'datetime'",
"]",
"QA_util_log_info",
"(",
"'##JOB13.{} Now Saving Option 50ETF {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_future_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"QA_util_log_info",
"(",
"\" 写入 新增历史合约记录数 {} \".format(len(__data))",
"",
"",
"",
"",
"",
"",
"",
"",
")",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
"[",
"1",
":",
":",
"]",
")",
")",
"else",
":",
"start_time",
"=",
"'2015-01-01'",
"QA_util_log_info",
"(",
"'##JOB13.{} Now Option 50ETF {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_future_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"QA_util_log_info",
"(",
"\" 写入 新增合约记录数 {} \".format(len(__data",
")",
")",
"",
"",
"",
"",
"",
"",
")",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
")",
")",
"except",
":",
"err",
".",
"append",
"(",
"code",
")",
"executor",
"=",
"ThreadPoolExecutor",
"(",
"max_workers",
"=",
"4",
")",
"res",
"=",
"{",
"executor",
".",
"submit",
"(",
"__saving_work",
",",
"option_contract_list",
"[",
"i_",
"]",
"[",
"\"code\"",
"]",
",",
"coll_option_min",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"option_contract_list",
")",
")",
"}",
"# multi index ./.",
"count",
"=",
"0",
"for",
"i_",
"in",
"concurrent",
".",
"futures",
".",
"as_completed",
"(",
"res",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"count",
",",
"len",
"(",
"option_contract_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strLogProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"count",
"/",
"len",
"(",
"option_contract_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"count",
"/",
"len",
"(",
"option_contract_list",
")",
"*",
"10000.0",
")",
")",
"QA_util_log_info",
"(",
"strLogProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"count",
"=",
"count",
"+",
"1",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
:param client:
:return:
|
[
":",
"param",
"client",
":",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L3430-L3563
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_option_day
|
def QA_SU_save_option_day(client=DATABASE, ui_log=None, ui_progress=None):
'''
:param client:
:return:
'''
option_contract_list = QA_fetch_get_option_50etf_contract_time_to_market()
coll_option_day = client.option_day
coll_option_day.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
# 索引 code
def __saving_work(code, coll_option_day):
try:
QA_util_log_info(
'##JOB12 Now Saving OPTION_DAY==== {}'.format(str(code)),
ui_log=ui_log
)
# 首选查找数据库 是否 有 这个代码的数据
# 期权代码 从 10000001 开始编码 10001228
ref = coll_option_day.find({'code': str(code)[0:8]})
end_date = str(now_time())[0:10]
# 当前数据库已经包含了这个代码的数据, 继续增量更新
# 加入这个判断的原因是因为如果是刚上市的 数据库会没有数据 所以会有负索引问题出现
if ref.count() > 0:
# 接着上次获取的日期继续更新
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
' 上次获取期权日线数据的最后日期是 {}'.format(start_date),
ui_log=ui_log
)
QA_util_log_info(
'UPDATE_OPTION_DAY \n 从上一次下载数据开始继续 Trying update {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
start_date0 = QA_util_get_next_day(start_date)
df0 = QA_fetch_get_option_day(
code=code,
start_date=start_date0,
end_date=end_date,
frequence='day',
ip=None,
port=None
)
retCount = df0.iloc[:, 0].size
QA_util_log_info(
"日期从开始{}-结束{} , 合约代码{} , 返回了{}条记录 , 准备写入数据库".format(
start_date0,
end_date,
code,
retCount
),
ui_log=ui_log
)
coll_option_day.insert_many(
QA_util_to_json_from_pandas(df0)
)
else:
QA_util_log_info(
"^已经获取过这天的数据了^ {}".format(start_date),
ui_log=ui_log
)
else:
start_date = '1990-01-01'
QA_util_log_info(
'UPDATE_OPTION_DAY \n 从新开始下载数据 Trying update {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
df0 = QA_fetch_get_option_day(
code=code,
start_date=start_date,
end_date=end_date,
frequence='day',
ip=None,
port=None
)
retCount = df0.iloc[:, 0].size
QA_util_log_info(
"日期从开始{}-结束{} , 合约代码{} , 获取了{}条记录 , 准备写入数据库^_^ ".format(
start_date,
end_date,
code,
retCount
),
ui_log=ui_log
)
coll_option_day.insert_many(
QA_util_to_json_from_pandas(df0)
)
else:
QA_util_log_info(
"*已经获取过这天的数据了* {}".format(start_date),
ui_log=ui_log
)
except Exception as error0:
print(error0)
err.append(str(code))
for item in range(len(option_contract_list)):
QA_util_log_info(
'The {} of Total {}'.format(item,
len(option_contract_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(item / len(option_contract_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(item / len(option_contract_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(option_contract_list[item].code, coll_option_day)
if len(err) < 1:
QA_util_log_info('SUCCESS save option day ^_^ ', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_option_day(client=DATABASE, ui_log=None, ui_progress=None):
'''
:param client:
:return:
'''
option_contract_list = QA_fetch_get_option_50etf_contract_time_to_market()
coll_option_day = client.option_day
coll_option_day.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
# 索引 code
def __saving_work(code, coll_option_day):
try:
QA_util_log_info(
'##JOB12 Now Saving OPTION_DAY==== {}'.format(str(code)),
ui_log=ui_log
)
# 首选查找数据库 是否 有 这个代码的数据
# 期权代码 从 10000001 开始编码 10001228
ref = coll_option_day.find({'code': str(code)[0:8]})
end_date = str(now_time())[0:10]
# 当前数据库已经包含了这个代码的数据, 继续增量更新
# 加入这个判断的原因是因为如果是刚上市的 数据库会没有数据 所以会有负索引问题出现
if ref.count() > 0:
# 接着上次获取的日期继续更新
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
' 上次获取期权日线数据的最后日期是 {}'.format(start_date),
ui_log=ui_log
)
QA_util_log_info(
'UPDATE_OPTION_DAY \n 从上一次下载数据开始继续 Trying update {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
start_date0 = QA_util_get_next_day(start_date)
df0 = QA_fetch_get_option_day(
code=code,
start_date=start_date0,
end_date=end_date,
frequence='day',
ip=None,
port=None
)
retCount = df0.iloc[:, 0].size
QA_util_log_info(
"日期从开始{}-结束{} , 合约代码{} , 返回了{}条记录 , 准备写入数据库".format(
start_date0,
end_date,
code,
retCount
),
ui_log=ui_log
)
coll_option_day.insert_many(
QA_util_to_json_from_pandas(df0)
)
else:
QA_util_log_info(
"^已经获取过这天的数据了^ {}".format(start_date),
ui_log=ui_log
)
else:
start_date = '1990-01-01'
QA_util_log_info(
'UPDATE_OPTION_DAY \n 从新开始下载数据 Trying update {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log=ui_log
)
if start_date != end_date:
df0 = QA_fetch_get_option_day(
code=code,
start_date=start_date,
end_date=end_date,
frequence='day',
ip=None,
port=None
)
retCount = df0.iloc[:, 0].size
QA_util_log_info(
"日期从开始{}-结束{} , 合约代码{} , 获取了{}条记录 , 准备写入数据库^_^ ".format(
start_date,
end_date,
code,
retCount
),
ui_log=ui_log
)
coll_option_day.insert_many(
QA_util_to_json_from_pandas(df0)
)
else:
QA_util_log_info(
"*已经获取过这天的数据了* {}".format(start_date),
ui_log=ui_log
)
except Exception as error0:
print(error0)
err.append(str(code))
for item in range(len(option_contract_list)):
QA_util_log_info(
'The {} of Total {}'.format(item,
len(option_contract_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(item / len(option_contract_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(item / len(option_contract_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
__saving_work(option_contract_list[item].code, coll_option_day)
if len(err) < 1:
QA_util_log_info('SUCCESS save option day ^_^ ', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_option_day",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"option_contract_list",
"=",
"QA_fetch_get_option_50etf_contract_time_to_market",
"(",
")",
"coll_option_day",
"=",
"client",
".",
"option_day",
"coll_option_day",
".",
"create_index",
"(",
"[",
"(",
"\"code\"",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"\"date_stamp\"",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"# 索引 code",
"def",
"__saving_work",
"(",
"code",
",",
"coll_option_day",
")",
":",
"try",
":",
"QA_util_log_info",
"(",
"'##JOB12 Now Saving OPTION_DAY==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"# 首选查找数据库 是否 有 这个代码的数据",
"# 期权代码 从 10000001 开始编码 10001228",
"ref",
"=",
"coll_option_day",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"8",
"]",
"}",
")",
"end_date",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"10",
"]",
"# 当前数据库已经包含了这个代码的数据, 继续增量更新",
"# 加入这个判断的原因是因为如果是刚上市的 数据库会没有数据 所以会有负索引问题出现",
"if",
"ref",
".",
"count",
"(",
")",
">",
"0",
":",
"# 接着上次获取的日期继续更新",
"start_date",
"=",
"ref",
"[",
"ref",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'date'",
"]",
"QA_util_log_info",
"(",
"' 上次获取期权日线数据的最后日期是 {}'.format(start_date),",
"",
"",
"",
"",
"",
"",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"'UPDATE_OPTION_DAY \\n 从上一次下载数据开始继续 Trying update {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"start_date0",
"=",
"QA_util_get_next_day",
"(",
"start_date",
")",
"df0",
"=",
"QA_fetch_get_option_day",
"(",
"code",
"=",
"code",
",",
"start_date",
"=",
"start_date0",
",",
"end_date",
"=",
"end_date",
",",
"frequence",
"=",
"'day'",
",",
"ip",
"=",
"None",
",",
"port",
"=",
"None",
")",
"retCount",
"=",
"df0",
".",
"iloc",
"[",
":",
",",
"0",
"]",
".",
"size",
"QA_util_log_info",
"(",
"\"日期从开始{}-结束{} , 合约代码{} , 返回了{}条记录 , 准备写入数据库\".format(",
"",
"",
"",
"start_date0",
",",
"end_date",
",",
"code",
",",
"retCount",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"coll_option_day",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"df0",
")",
")",
"else",
":",
"QA_util_log_info",
"(",
"\"^已经获取过这天的数据了^ {}\".format(start_date),",
"",
"",
"",
"",
"",
"",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"start_date",
"=",
"'1990-01-01'",
"QA_util_log_info",
"(",
"'UPDATE_OPTION_DAY \\n 从新开始下载数据 Trying update {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"df0",
"=",
"QA_fetch_get_option_day",
"(",
"code",
"=",
"code",
",",
"start_date",
"=",
"start_date",
",",
"end_date",
"=",
"end_date",
",",
"frequence",
"=",
"'day'",
",",
"ip",
"=",
"None",
",",
"port",
"=",
"None",
")",
"retCount",
"=",
"df0",
".",
"iloc",
"[",
":",
",",
"0",
"]",
".",
"size",
"QA_util_log_info",
"(",
"\"日期从开始{}-结束{} , 合约代码{} , 获取了{}条记录 , 准备写入数据库^_^ \".format(",
"",
"",
"",
"start_date",
",",
"end_date",
",",
"code",
",",
"retCount",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"coll_option_day",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"df0",
")",
")",
"else",
":",
"QA_util_log_info",
"(",
"\"*已经获取过这天的数据了* {}\".format(start_date),",
"",
"",
"",
"",
"",
"",
"ui_log",
"=",
"ui_log",
")",
"except",
"Exception",
"as",
"error0",
":",
"print",
"(",
"error0",
")",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"for",
"item",
"in",
"range",
"(",
"len",
"(",
"option_contract_list",
")",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"item",
",",
"len",
"(",
"option_contract_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strLogProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"option_contract_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"option_contract_list",
")",
"*",
"10000.0",
")",
")",
"QA_util_log_info",
"(",
"strLogProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"__saving_work",
"(",
"option_contract_list",
"[",
"item",
"]",
".",
"code",
",",
"coll_option_day",
")",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS save option day ^_^ '",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
:param client:
:return:
|
[
":",
"param",
"client",
":",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L3566-L3710
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_future_day
|
def QA_SU_save_future_day(client=DATABASE, ui_log=None, ui_progress=None):
'''
save future_day
保存日线数据
:param client:
:param ui_log: 给GUI qt 界面使用
:param ui_progress: 给GUI qt 界面使用
:param ui_progress_int_value: 给GUI qt 界面使用
:return:
'''
future_list = [
item for item in QA_fetch_get_future_list().code.unique().tolist()
if str(item)[-2:] in ['L8',
'L9']
]
coll_future_day = client.future_day
coll_future_day.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll_future_day):
try:
QA_util_log_info(
'##JOB12 Now Saving Future_DAY==== {}'.format(str(code)),
ui_log
)
# 首选查找数据库 是否 有 这个代码的数据
ref = coll_future_day.find({'code': str(code)[0:4]})
end_date = str(now_time())[0:10]
# 当前数据库已经包含了这个代码的数据, 继续增量更新
# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现
if ref.count() > 0:
# 接着上次获取的日期继续更新
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
'UPDATE_Future_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_future_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_future_day(
str(code),
QA_util_get_next_day(start_date),
end_date
)
)
)
# 当前数据库中没有这个代码的股票数据, 从1990-01-01 开始下载所有的数据
else:
start_date = '2001-01-01'
QA_util_log_info(
'UPDATE_Future_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_future_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_future_day(
str(code),
start_date,
end_date
)
)
)
except Exception as error0:
print(error0)
err.append(str(code))
for item in range(len(future_list)):
QA_util_log_info('The {} of Total {}'.format(item, len(future_list)))
strProgressToLog = 'DOWNLOAD PROGRESS {} {}'.format(
str(float(item / len(future_list) * 100))[0:4] + '%',
ui_log
)
intProgressToLog = int(float(item / len(future_list) * 100))
QA_util_log_info(
strProgressToLog,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgressToLog
)
__saving_work(future_list[item], coll_future_day)
if len(err) < 1:
QA_util_log_info('SUCCESS save future day ^_^', ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log)
QA_util_log_info(err, ui_log)
|
python
|
def QA_SU_save_future_day(client=DATABASE, ui_log=None, ui_progress=None):
'''
save future_day
保存日线数据
:param client:
:param ui_log: 给GUI qt 界面使用
:param ui_progress: 给GUI qt 界面使用
:param ui_progress_int_value: 给GUI qt 界面使用
:return:
'''
future_list = [
item for item in QA_fetch_get_future_list().code.unique().tolist()
if str(item)[-2:] in ['L8',
'L9']
]
coll_future_day = client.future_day
coll_future_day.create_index(
[("code",
pymongo.ASCENDING),
("date_stamp",
pymongo.ASCENDING)]
)
err = []
def __saving_work(code, coll_future_day):
try:
QA_util_log_info(
'##JOB12 Now Saving Future_DAY==== {}'.format(str(code)),
ui_log
)
# 首选查找数据库 是否 有 这个代码的数据
ref = coll_future_day.find({'code': str(code)[0:4]})
end_date = str(now_time())[0:10]
# 当前数据库已经包含了这个代码的数据, 继续增量更新
# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现
if ref.count() > 0:
# 接着上次获取的日期继续更新
start_date = ref[ref.count() - 1]['date']
QA_util_log_info(
'UPDATE_Future_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_future_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_future_day(
str(code),
QA_util_get_next_day(start_date),
end_date
)
)
)
# 当前数据库中没有这个代码的股票数据, 从1990-01-01 开始下载所有的数据
else:
start_date = '2001-01-01'
QA_util_log_info(
'UPDATE_Future_DAY \n Trying updating {} from {} to {}'
.format(code,
start_date,
end_date),
ui_log
)
if start_date != end_date:
coll_future_day.insert_many(
QA_util_to_json_from_pandas(
QA_fetch_get_future_day(
str(code),
start_date,
end_date
)
)
)
except Exception as error0:
print(error0)
err.append(str(code))
for item in range(len(future_list)):
QA_util_log_info('The {} of Total {}'.format(item, len(future_list)))
strProgressToLog = 'DOWNLOAD PROGRESS {} {}'.format(
str(float(item / len(future_list) * 100))[0:4] + '%',
ui_log
)
intProgressToLog = int(float(item / len(future_list) * 100))
QA_util_log_info(
strProgressToLog,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intProgressToLog
)
__saving_work(future_list[item], coll_future_day)
if len(err) < 1:
QA_util_log_info('SUCCESS save future day ^_^', ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log)
QA_util_log_info(err, ui_log)
|
[
"def",
"QA_SU_save_future_day",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"future_list",
"=",
"[",
"item",
"for",
"item",
"in",
"QA_fetch_get_future_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"if",
"str",
"(",
"item",
")",
"[",
"-",
"2",
":",
"]",
"in",
"[",
"'L8'",
",",
"'L9'",
"]",
"]",
"coll_future_day",
"=",
"client",
".",
"future_day",
"coll_future_day",
".",
"create_index",
"(",
"[",
"(",
"\"code\"",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"\"date_stamp\"",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll_future_day",
")",
":",
"try",
":",
"QA_util_log_info",
"(",
"'##JOB12 Now Saving Future_DAY==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
")",
"# 首选查找数据库 是否 有 这个代码的数据",
"ref",
"=",
"coll_future_day",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"4",
"]",
"}",
")",
"end_date",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"10",
"]",
"# 当前数据库已经包含了这个代码的数据, 继续增量更新",
"# 加入这个判断的原因是因为如果股票是刚上市的 数据库会没有数据 所以会有负索引问题出现",
"if",
"ref",
".",
"count",
"(",
")",
">",
"0",
":",
"# 接着上次获取的日期继续更新",
"start_date",
"=",
"ref",
"[",
"ref",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'date'",
"]",
"QA_util_log_info",
"(",
"'UPDATE_Future_DAY \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"coll_future_day",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_future_day",
"(",
"str",
"(",
"code",
")",
",",
"QA_util_get_next_day",
"(",
"start_date",
")",
",",
"end_date",
")",
")",
")",
"# 当前数据库中没有这个代码的股票数据, 从1990-01-01 开始下载所有的数据",
"else",
":",
"start_date",
"=",
"'2001-01-01'",
"QA_util_log_info",
"(",
"'UPDATE_Future_DAY \\n Trying updating {} from {} to {}'",
".",
"format",
"(",
"code",
",",
"start_date",
",",
"end_date",
")",
",",
"ui_log",
")",
"if",
"start_date",
"!=",
"end_date",
":",
"coll_future_day",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"QA_fetch_get_future_day",
"(",
"str",
"(",
"code",
")",
",",
"start_date",
",",
"end_date",
")",
")",
")",
"except",
"Exception",
"as",
"error0",
":",
"print",
"(",
"error0",
")",
"err",
".",
"append",
"(",
"str",
"(",
"code",
")",
")",
"for",
"item",
"in",
"range",
"(",
"len",
"(",
"future_list",
")",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"item",
",",
"len",
"(",
"future_list",
")",
")",
")",
"strProgressToLog",
"=",
"'DOWNLOAD PROGRESS {} {}'",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"future_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
",",
"ui_log",
")",
"intProgressToLog",
"=",
"int",
"(",
"float",
"(",
"item",
"/",
"len",
"(",
"future_list",
")",
"*",
"100",
")",
")",
"QA_util_log_info",
"(",
"strProgressToLog",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intProgressToLog",
")",
"__saving_work",
"(",
"future_list",
"[",
"item",
"]",
",",
"coll_future_day",
")",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS save future day ^_^'",
",",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
")"
] |
save future_day
保存日线数据
:param client:
:param ui_log: 给GUI qt 界面使用
:param ui_progress: 给GUI qt 界面使用
:param ui_progress_int_value: 给GUI qt 界面使用
:return:
|
[
"save",
"future_day",
"保存日线数据",
":",
"param",
"client",
":",
":",
"param",
"ui_log",
":",
"给GUI",
"qt",
"界面使用",
":",
"param",
"ui_progress",
":",
"给GUI",
"qt",
"界面使用",
":",
"param",
"ui_progress_int_value",
":",
"给GUI",
"qt",
"界面使用",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L3804-L3909
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QASU/save_tdx.py
|
QA_SU_save_future_min
|
def QA_SU_save_future_min(client=DATABASE, ui_log=None, ui_progress=None):
"""save future_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
future_list = [
item for item in QA_fetch_get_future_list().code.unique().tolist()
if str(item)[-2:] in ['L8',
'L9']
]
coll = client.future_min
coll.create_index(
[
('code',
pymongo.ASCENDING),
('time_stamp',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)
]
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB13 Now Saving Future_MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:6], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB13.{} Now Saving Future {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data[1::])
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB13.{} Now Saving Future {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except:
err.append(code)
executor = ThreadPoolExecutor(max_workers=4)
res = {
executor.submit(__saving_work,
future_list[i_],
coll)
for i_ in range(len(future_list))
} # multi index ./.
count = 0
for i_ in concurrent.futures.as_completed(res):
QA_util_log_info(
'The {} of Total {}'.format(count,
len(future_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(future_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(count / len(future_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
python
|
def QA_SU_save_future_min(client=DATABASE, ui_log=None, ui_progress=None):
"""save future_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
"""
future_list = [
item for item in QA_fetch_get_future_list().code.unique().tolist()
if str(item)[-2:] in ['L8',
'L9']
]
coll = client.future_min
coll.create_index(
[
('code',
pymongo.ASCENDING),
('time_stamp',
pymongo.ASCENDING),
('date_stamp',
pymongo.ASCENDING)
]
)
err = []
def __saving_work(code, coll):
QA_util_log_info(
'##JOB13 Now Saving Future_MIN ==== {}'.format(str(code)),
ui_log=ui_log
)
try:
for type in ['1min', '5min', '15min', '30min', '60min']:
ref_ = coll.find({'code': str(code)[0:6], 'type': type})
end_time = str(now_time())[0:19]
if ref_.count() > 0:
start_time = ref_[ref_.count() - 1]['datetime']
QA_util_log_info(
'##JOB13.{} Now Saving Future {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data[1::])
)
else:
start_time = '2015-01-01'
QA_util_log_info(
'##JOB13.{} Now Saving Future {} from {} to {} =={} '
.format(
['1min',
'5min',
'15min',
'30min',
'60min'].index(type),
str(code),
start_time,
end_time,
type
),
ui_log=ui_log
)
if start_time != end_time:
__data = QA_fetch_get_future_min(
str(code),
start_time,
end_time,
type
)
if len(__data) > 1:
coll.insert_many(
QA_util_to_json_from_pandas(__data)
)
except:
err.append(code)
executor = ThreadPoolExecutor(max_workers=4)
res = {
executor.submit(__saving_work,
future_list[i_],
coll)
for i_ in range(len(future_list))
} # multi index ./.
count = 0
for i_ in concurrent.futures.as_completed(res):
QA_util_log_info(
'The {} of Total {}'.format(count,
len(future_list)),
ui_log=ui_log
)
strLogProgress = 'DOWNLOAD PROGRESS {} '.format(
str(float(count / len(future_list) * 100))[0:4] + '%'
)
intLogProgress = int(float(count / len(future_list) * 10000.0))
QA_util_log_info(
strLogProgress,
ui_log=ui_log,
ui_progress=ui_progress,
ui_progress_int_value=intLogProgress
)
count = count + 1
if len(err) < 1:
QA_util_log_info('SUCCESS', ui_log=ui_log)
else:
QA_util_log_info(' ERROR CODE \n ', ui_log=ui_log)
QA_util_log_info(err, ui_log=ui_log)
|
[
"def",
"QA_SU_save_future_min",
"(",
"client",
"=",
"DATABASE",
",",
"ui_log",
"=",
"None",
",",
"ui_progress",
"=",
"None",
")",
":",
"future_list",
"=",
"[",
"item",
"for",
"item",
"in",
"QA_fetch_get_future_list",
"(",
")",
".",
"code",
".",
"unique",
"(",
")",
".",
"tolist",
"(",
")",
"if",
"str",
"(",
"item",
")",
"[",
"-",
"2",
":",
"]",
"in",
"[",
"'L8'",
",",
"'L9'",
"]",
"]",
"coll",
"=",
"client",
".",
"future_min",
"coll",
".",
"create_index",
"(",
"[",
"(",
"'code'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'time_stamp'",
",",
"pymongo",
".",
"ASCENDING",
")",
",",
"(",
"'date_stamp'",
",",
"pymongo",
".",
"ASCENDING",
")",
"]",
")",
"err",
"=",
"[",
"]",
"def",
"__saving_work",
"(",
"code",
",",
"coll",
")",
":",
"QA_util_log_info",
"(",
"'##JOB13 Now Saving Future_MIN ==== {}'",
".",
"format",
"(",
"str",
"(",
"code",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"try",
":",
"for",
"type",
"in",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
":",
"ref_",
"=",
"coll",
".",
"find",
"(",
"{",
"'code'",
":",
"str",
"(",
"code",
")",
"[",
"0",
":",
"6",
"]",
",",
"'type'",
":",
"type",
"}",
")",
"end_time",
"=",
"str",
"(",
"now_time",
"(",
")",
")",
"[",
"0",
":",
"19",
"]",
"if",
"ref_",
".",
"count",
"(",
")",
">",
"0",
":",
"start_time",
"=",
"ref_",
"[",
"ref_",
".",
"count",
"(",
")",
"-",
"1",
"]",
"[",
"'datetime'",
"]",
"QA_util_log_info",
"(",
"'##JOB13.{} Now Saving Future {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_future_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
"[",
"1",
":",
":",
"]",
")",
")",
"else",
":",
"start_time",
"=",
"'2015-01-01'",
"QA_util_log_info",
"(",
"'##JOB13.{} Now Saving Future {} from {} to {} =={} '",
".",
"format",
"(",
"[",
"'1min'",
",",
"'5min'",
",",
"'15min'",
",",
"'30min'",
",",
"'60min'",
"]",
".",
"index",
"(",
"type",
")",
",",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"if",
"start_time",
"!=",
"end_time",
":",
"__data",
"=",
"QA_fetch_get_future_min",
"(",
"str",
"(",
"code",
")",
",",
"start_time",
",",
"end_time",
",",
"type",
")",
"if",
"len",
"(",
"__data",
")",
">",
"1",
":",
"coll",
".",
"insert_many",
"(",
"QA_util_to_json_from_pandas",
"(",
"__data",
")",
")",
"except",
":",
"err",
".",
"append",
"(",
"code",
")",
"executor",
"=",
"ThreadPoolExecutor",
"(",
"max_workers",
"=",
"4",
")",
"res",
"=",
"{",
"executor",
".",
"submit",
"(",
"__saving_work",
",",
"future_list",
"[",
"i_",
"]",
",",
"coll",
")",
"for",
"i_",
"in",
"range",
"(",
"len",
"(",
"future_list",
")",
")",
"}",
"# multi index ./.",
"count",
"=",
"0",
"for",
"i_",
"in",
"concurrent",
".",
"futures",
".",
"as_completed",
"(",
"res",
")",
":",
"QA_util_log_info",
"(",
"'The {} of Total {}'",
".",
"format",
"(",
"count",
",",
"len",
"(",
"future_list",
")",
")",
",",
"ui_log",
"=",
"ui_log",
")",
"strLogProgress",
"=",
"'DOWNLOAD PROGRESS {} '",
".",
"format",
"(",
"str",
"(",
"float",
"(",
"count",
"/",
"len",
"(",
"future_list",
")",
"*",
"100",
")",
")",
"[",
"0",
":",
"4",
"]",
"+",
"'%'",
")",
"intLogProgress",
"=",
"int",
"(",
"float",
"(",
"count",
"/",
"len",
"(",
"future_list",
")",
"*",
"10000.0",
")",
")",
"QA_util_log_info",
"(",
"strLogProgress",
",",
"ui_log",
"=",
"ui_log",
",",
"ui_progress",
"=",
"ui_progress",
",",
"ui_progress_int_value",
"=",
"intLogProgress",
")",
"count",
"=",
"count",
"+",
"1",
"if",
"len",
"(",
"err",
")",
"<",
"1",
":",
"QA_util_log_info",
"(",
"'SUCCESS'",
",",
"ui_log",
"=",
"ui_log",
")",
"else",
":",
"QA_util_log_info",
"(",
"' ERROR CODE \\n '",
",",
"ui_log",
"=",
"ui_log",
")",
"QA_util_log_info",
"(",
"err",
",",
"ui_log",
"=",
"ui_log",
")"
] |
save future_min
Keyword Arguments:
client {[type]} -- [description] (default: {DATABASE})
|
[
"save",
"future_min"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QASU/save_tdx.py#L4016-L4146
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QACmd/__init__.py
|
CLI.do_shell
|
def do_shell(self, arg):
"run a shell commad"
print(">", arg)
sub_cmd = subprocess.Popen(arg, shell=True, stdout=subprocess.PIPE)
print(sub_cmd.communicate()[0])
|
python
|
def do_shell(self, arg):
"run a shell commad"
print(">", arg)
sub_cmd = subprocess.Popen(arg, shell=True, stdout=subprocess.PIPE)
print(sub_cmd.communicate()[0])
|
[
"def",
"do_shell",
"(",
"self",
",",
"arg",
")",
":",
"print",
"(",
"\">\"",
",",
"arg",
")",
"sub_cmd",
"=",
"subprocess",
".",
"Popen",
"(",
"arg",
",",
"shell",
"=",
"True",
",",
"stdout",
"=",
"subprocess",
".",
"PIPE",
")",
"print",
"(",
"sub_cmd",
".",
"communicate",
"(",
")",
"[",
"0",
"]",
")"
] |
run a shell commad
|
[
"run",
"a",
"shell",
"commad"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QACmd/__init__.py#L84-L88
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_stock_min_adv
|
def QA_fetch_stock_min_adv(
code,
start, end=None,
frequence='1min',
if_drop_index=True,
# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉
collections=DATABASE.stock_min):
'''
'获取股票分钟线'
:param code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param frequence: 字符串str 分钟线的类型 支持 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m 类型
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return: QA_DataStruct_Stock_min 类型
'''
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
else:
print("QA Error QA_fetch_stock_min_adv parameter frequence=%s is none of 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m" % frequence)
return None
# __data = [] 未使用
end = start if end is None else end
if len(start) == 10:
start = '{} 09:30:00'.format(start)
if len(end) == 10:
end = '{} 15:00:00'.format(end)
if start == end:
# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_stock_min, 不支持start end是相等的
print("QA Error QA_fetch_stock_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! " % (
code, start, end))
return None
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
res = QA_fetch_stock_min(
code, start, end, format='pd', frequence=frequence)
if res is None:
print("QA Error QA_fetch_stock_min_adv parameter code=%s , start=%s, end=%s frequence=%s call QA_fetch_stock_min return None" % (
code, start, end, frequence))
return None
else:
res_set_index = res.set_index(['datetime', 'code'], drop=if_drop_index)
# if res_set_index is None:
# print("QA Error QA_fetch_stock_min_adv set index 'datetime, code' return None")
# return None
return QA_DataStruct_Stock_min(res_set_index)
|
python
|
def QA_fetch_stock_min_adv(
code,
start, end=None,
frequence='1min',
if_drop_index=True,
# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉
collections=DATABASE.stock_min):
'''
'获取股票分钟线'
:param code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param frequence: 字符串str 分钟线的类型 支持 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m 类型
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return: QA_DataStruct_Stock_min 类型
'''
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
else:
print("QA Error QA_fetch_stock_min_adv parameter frequence=%s is none of 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m" % frequence)
return None
# __data = [] 未使用
end = start if end is None else end
if len(start) == 10:
start = '{} 09:30:00'.format(start)
if len(end) == 10:
end = '{} 15:00:00'.format(end)
if start == end:
# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_stock_min, 不支持start end是相等的
print("QA Error QA_fetch_stock_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! " % (
code, start, end))
return None
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
res = QA_fetch_stock_min(
code, start, end, format='pd', frequence=frequence)
if res is None:
print("QA Error QA_fetch_stock_min_adv parameter code=%s , start=%s, end=%s frequence=%s call QA_fetch_stock_min return None" % (
code, start, end, frequence))
return None
else:
res_set_index = res.set_index(['datetime', 'code'], drop=if_drop_index)
# if res_set_index is None:
# print("QA Error QA_fetch_stock_min_adv set index 'datetime, code' return None")
# return None
return QA_DataStruct_Stock_min(res_set_index)
|
[
"def",
"QA_fetch_stock_min_adv",
"(",
"code",
",",
"start",
",",
"end",
"=",
"None",
",",
"frequence",
"=",
"'1min'",
",",
"if_drop_index",
"=",
"True",
",",
"# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉",
"collections",
"=",
"DATABASE",
".",
"stock_min",
")",
":",
"if",
"frequence",
"in",
"[",
"'1min'",
",",
"'1m'",
"]",
":",
"frequence",
"=",
"'1min'",
"elif",
"frequence",
"in",
"[",
"'5min'",
",",
"'5m'",
"]",
":",
"frequence",
"=",
"'5min'",
"elif",
"frequence",
"in",
"[",
"'15min'",
",",
"'15m'",
"]",
":",
"frequence",
"=",
"'15min'",
"elif",
"frequence",
"in",
"[",
"'30min'",
",",
"'30m'",
"]",
":",
"frequence",
"=",
"'30min'",
"elif",
"frequence",
"in",
"[",
"'60min'",
",",
"'60m'",
"]",
":",
"frequence",
"=",
"'60min'",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_min_adv parameter frequence=%s is none of 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m\"",
"%",
"frequence",
")",
"return",
"None",
"# __data = [] 未使用",
"end",
"=",
"start",
"if",
"end",
"is",
"None",
"else",
"end",
"if",
"len",
"(",
"start",
")",
"==",
"10",
":",
"start",
"=",
"'{} 09:30:00'",
".",
"format",
"(",
"start",
")",
"if",
"len",
"(",
"end",
")",
"==",
"10",
":",
"end",
"=",
"'{} 15:00:00'",
".",
"format",
"(",
"end",
")",
"if",
"start",
"==",
"end",
":",
"# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_stock_min, 不支持start end是相等的",
"print",
"(",
"\"QA Error QA_fetch_stock_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! \"",
"%",
"(",
"code",
",",
"start",
",",
"end",
")",
")",
"return",
"None",
"# 🛠 todo 报告错误 如果开始时间 在 结束时间之后",
"res",
"=",
"QA_fetch_stock_min",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'pd'",
",",
"frequence",
"=",
"frequence",
")",
"if",
"res",
"is",
"None",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_min_adv parameter code=%s , start=%s, end=%s frequence=%s call QA_fetch_stock_min return None\"",
"%",
"(",
"code",
",",
"start",
",",
"end",
",",
"frequence",
")",
")",
"return",
"None",
"else",
":",
"res_set_index",
"=",
"res",
".",
"set_index",
"(",
"[",
"'datetime'",
",",
"'code'",
"]",
",",
"drop",
"=",
"if_drop_index",
")",
"# if res_set_index is None:",
"# print(\"QA Error QA_fetch_stock_min_adv set index 'datetime, code' return None\")",
"# return None",
"return",
"QA_DataStruct_Stock_min",
"(",
"res_set_index",
")"
] |
'获取股票分钟线'
:param code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param frequence: 字符串str 分钟线的类型 支持 1min 1m 5min 5m 15min 15m 30min 30m 60min 60m 类型
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return: QA_DataStruct_Stock_min 类型
|
[
"获取股票分钟线",
":",
"param",
"code",
":",
"字符串str",
"eg",
"600085",
":",
"param",
"start",
":",
"字符串str",
"开始日期",
"eg",
"2011",
"-",
"01",
"-",
"01",
":",
"param",
"end",
":",
"字符串str",
"结束日期",
"eg",
"2011",
"-",
"05",
"-",
"01",
":",
"param",
"frequence",
":",
"字符串str",
"分钟线的类型",
"支持",
"1min",
"1m",
"5min",
"5m",
"15min",
"15m",
"30min",
"30m",
"60min",
"60m",
"类型",
":",
"param",
"if_drop_index",
":",
"Ture",
"False",
",",
"dataframe",
"drop",
"index",
"or",
"not",
":",
"param",
"collections",
":",
"mongodb",
"数据库",
":",
"return",
":",
"QA_DataStruct_Stock_min",
"类型"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L123-L182
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_stock_day_full_adv
|
def QA_fetch_stock_day_full_adv(date):
'''
'返回全市场某一天的数据'
:param date:
:return: QA_DataStruct_Stock_day类 型数据
'''
# 🛠 todo 检查日期data参数
res = QA_fetch_stock_full(date, 'pd')
if res is None:
print("QA Error QA_fetch_stock_day_full_adv parameter date=%s call QA_fetch_stock_full return None" % (date))
return None
else:
res_set_index = res.set_index(['date', 'code'])
# if res_set_index is None:
# print("QA Error QA_fetch_stock_day_full set index 'date, code' return None")
return QA_DataStruct_Stock_day(res_set_index)
|
python
|
def QA_fetch_stock_day_full_adv(date):
'''
'返回全市场某一天的数据'
:param date:
:return: QA_DataStruct_Stock_day类 型数据
'''
# 🛠 todo 检查日期data参数
res = QA_fetch_stock_full(date, 'pd')
if res is None:
print("QA Error QA_fetch_stock_day_full_adv parameter date=%s call QA_fetch_stock_full return None" % (date))
return None
else:
res_set_index = res.set_index(['date', 'code'])
# if res_set_index is None:
# print("QA Error QA_fetch_stock_day_full set index 'date, code' return None")
return QA_DataStruct_Stock_day(res_set_index)
|
[
"def",
"QA_fetch_stock_day_full_adv",
"(",
"date",
")",
":",
"# 🛠 todo 检查日期data参数",
"res",
"=",
"QA_fetch_stock_full",
"(",
"date",
",",
"'pd'",
")",
"if",
"res",
"is",
"None",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_day_full_adv parameter date=%s call QA_fetch_stock_full return None\"",
"%",
"(",
"date",
")",
")",
"return",
"None",
"else",
":",
"res_set_index",
"=",
"res",
".",
"set_index",
"(",
"[",
"'date'",
",",
"'code'",
"]",
")",
"# if res_set_index is None:",
"# print(\"QA Error QA_fetch_stock_day_full set index 'date, code' return None\")",
"return",
"QA_DataStruct_Stock_day",
"(",
"res_set_index",
")"
] |
'返回全市场某一天的数据'
:param date:
:return: QA_DataStruct_Stock_day类 型数据
|
[
"返回全市场某一天的数据",
":",
"param",
"date",
":",
":",
"return",
":",
"QA_DataStruct_Stock_day类",
"型数据"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L185-L200
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_index_day_adv
|
def QA_fetch_index_day_adv(
code,
start, end=None,
if_drop_index=True,
# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉
collections=DATABASE.index_day):
'''
:param code: code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return:
'''
'获取指数日线'
end = start if end is None else end
start = str(start)[0:10]
end = str(end)[0:10]
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# 🛠 todo 如果相等
res = QA_fetch_index_day(code, start, end, format='pd')
if res is None:
print("QA Error QA_fetch_index_day_adv parameter code=%s start=%s end=%s call QA_fetch_index_day return None" % (
code, start, end))
return None
else:
res_set_index = res.set_index(['date', 'code'], drop=if_drop_index)
# if res_set_index is None:
# print("QA Error QA_fetch_index_day_adv set index 'date, code' return None")
# return None
return QA_DataStruct_Index_day(res_set_index)
|
python
|
def QA_fetch_index_day_adv(
code,
start, end=None,
if_drop_index=True,
# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉
collections=DATABASE.index_day):
'''
:param code: code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return:
'''
'获取指数日线'
end = start if end is None else end
start = str(start)[0:10]
end = str(end)[0:10]
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# 🛠 todo 如果相等
res = QA_fetch_index_day(code, start, end, format='pd')
if res is None:
print("QA Error QA_fetch_index_day_adv parameter code=%s start=%s end=%s call QA_fetch_index_day return None" % (
code, start, end))
return None
else:
res_set_index = res.set_index(['date', 'code'], drop=if_drop_index)
# if res_set_index is None:
# print("QA Error QA_fetch_index_day_adv set index 'date, code' return None")
# return None
return QA_DataStruct_Index_day(res_set_index)
|
[
"def",
"QA_fetch_index_day_adv",
"(",
"code",
",",
"start",
",",
"end",
"=",
"None",
",",
"if_drop_index",
"=",
"True",
",",
"# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉",
"collections",
"=",
"DATABASE",
".",
"index_day",
")",
":",
"'获取指数日线'",
"end",
"=",
"start",
"if",
"end",
"is",
"None",
"else",
"end",
"start",
"=",
"str",
"(",
"start",
")",
"[",
"0",
":",
"10",
"]",
"end",
"=",
"str",
"(",
"end",
")",
"[",
"0",
":",
"10",
"]",
"# 🛠 todo 报告错误 如果开始时间 在 结束时间之后",
"# 🛠 todo 如果相等",
"res",
"=",
"QA_fetch_index_day",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'pd'",
")",
"if",
"res",
"is",
"None",
":",
"print",
"(",
"\"QA Error QA_fetch_index_day_adv parameter code=%s start=%s end=%s call QA_fetch_index_day return None\"",
"%",
"(",
"code",
",",
"start",
",",
"end",
")",
")",
"return",
"None",
"else",
":",
"res_set_index",
"=",
"res",
".",
"set_index",
"(",
"[",
"'date'",
",",
"'code'",
"]",
",",
"drop",
"=",
"if_drop_index",
")",
"# if res_set_index is None:",
"# print(\"QA Error QA_fetch_index_day_adv set index 'date, code' return None\")",
"# return None",
"return",
"QA_DataStruct_Index_day",
"(",
"res_set_index",
")"
] |
:param code: code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return:
|
[
":",
"param",
"code",
":",
"code",
":",
"字符串str",
"eg",
"600085",
":",
"param",
"start",
":",
"字符串str",
"开始日期",
"eg",
"2011",
"-",
"01",
"-",
"01",
":",
"param",
"end",
":",
"字符串str",
"结束日期",
"eg",
"2011",
"-",
"05",
"-",
"01",
":",
"param",
"if_drop_index",
":",
"Ture",
"False",
",",
"dataframe",
"drop",
"index",
"or",
"not",
":",
"param",
"collections",
":",
"mongodb",
"数据库",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L203-L235
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_index_min_adv
|
def QA_fetch_index_min_adv(
code,
start, end=None,
frequence='1min',
if_drop_index=True,
collections=DATABASE.index_min):
'''
'获取股票分钟线'
:param code:
:param start:
:param end:
:param frequence:
:param if_drop_index:
:param collections:
:return:
'''
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
# __data = [] 没有使用
end = start if end is None else end
if len(start) == 10:
start = '{} 09:30:00'.format(start)
if len(end) == 10:
end = '{} 15:00:00'.format(end)
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# if start == end:
# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_index_min_adv, 不支持start end是相等的
#print("QA Error QA_fetch_index_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! " % (code, start, end))
# return None
res = QA_fetch_index_min(
code, start, end, format='pd', frequence=frequence)
if res is None:
print("QA Error QA_fetch_index_min_adv parameter code=%s start=%s end=%s frequence=%s call QA_fetch_index_min return None" % (
code, start, end, frequence))
else:
res_reset_index = res.set_index(
['datetime', 'code'], drop=if_drop_index)
# if res_reset_index is None:
# print("QA Error QA_fetch_index_min_adv set index 'date, code' return None")
return QA_DataStruct_Index_min(res_reset_index)
|
python
|
def QA_fetch_index_min_adv(
code,
start, end=None,
frequence='1min',
if_drop_index=True,
collections=DATABASE.index_min):
'''
'获取股票分钟线'
:param code:
:param start:
:param end:
:param frequence:
:param if_drop_index:
:param collections:
:return:
'''
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
# __data = [] 没有使用
end = start if end is None else end
if len(start) == 10:
start = '{} 09:30:00'.format(start)
if len(end) == 10:
end = '{} 15:00:00'.format(end)
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# if start == end:
# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_index_min_adv, 不支持start end是相等的
#print("QA Error QA_fetch_index_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! " % (code, start, end))
# return None
res = QA_fetch_index_min(
code, start, end, format='pd', frequence=frequence)
if res is None:
print("QA Error QA_fetch_index_min_adv parameter code=%s start=%s end=%s frequence=%s call QA_fetch_index_min return None" % (
code, start, end, frequence))
else:
res_reset_index = res.set_index(
['datetime', 'code'], drop=if_drop_index)
# if res_reset_index is None:
# print("QA Error QA_fetch_index_min_adv set index 'date, code' return None")
return QA_DataStruct_Index_min(res_reset_index)
|
[
"def",
"QA_fetch_index_min_adv",
"(",
"code",
",",
"start",
",",
"end",
"=",
"None",
",",
"frequence",
"=",
"'1min'",
",",
"if_drop_index",
"=",
"True",
",",
"collections",
"=",
"DATABASE",
".",
"index_min",
")",
":",
"if",
"frequence",
"in",
"[",
"'1min'",
",",
"'1m'",
"]",
":",
"frequence",
"=",
"'1min'",
"elif",
"frequence",
"in",
"[",
"'5min'",
",",
"'5m'",
"]",
":",
"frequence",
"=",
"'5min'",
"elif",
"frequence",
"in",
"[",
"'15min'",
",",
"'15m'",
"]",
":",
"frequence",
"=",
"'15min'",
"elif",
"frequence",
"in",
"[",
"'30min'",
",",
"'30m'",
"]",
":",
"frequence",
"=",
"'30min'",
"elif",
"frequence",
"in",
"[",
"'60min'",
",",
"'60m'",
"]",
":",
"frequence",
"=",
"'60min'",
"# __data = [] 没有使用",
"end",
"=",
"start",
"if",
"end",
"is",
"None",
"else",
"end",
"if",
"len",
"(",
"start",
")",
"==",
"10",
":",
"start",
"=",
"'{} 09:30:00'",
".",
"format",
"(",
"start",
")",
"if",
"len",
"(",
"end",
")",
"==",
"10",
":",
"end",
"=",
"'{} 15:00:00'",
".",
"format",
"(",
"end",
")",
"# 🛠 todo 报告错误 如果开始时间 在 结束时间之后",
"# if start == end:",
"# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_index_min_adv, 不支持start end是相等的",
"#print(\"QA Error QA_fetch_index_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! \" % (code, start, end))",
"# return None",
"res",
"=",
"QA_fetch_index_min",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'pd'",
",",
"frequence",
"=",
"frequence",
")",
"if",
"res",
"is",
"None",
":",
"print",
"(",
"\"QA Error QA_fetch_index_min_adv parameter code=%s start=%s end=%s frequence=%s call QA_fetch_index_min return None\"",
"%",
"(",
"code",
",",
"start",
",",
"end",
",",
"frequence",
")",
")",
"else",
":",
"res_reset_index",
"=",
"res",
".",
"set_index",
"(",
"[",
"'datetime'",
",",
"'code'",
"]",
",",
"drop",
"=",
"if_drop_index",
")",
"# if res_reset_index is None:",
"# print(\"QA Error QA_fetch_index_min_adv set index 'date, code' return None\")",
"return",
"QA_DataStruct_Index_min",
"(",
"res_reset_index",
")"
] |
'获取股票分钟线'
:param code:
:param start:
:param end:
:param frequence:
:param if_drop_index:
:param collections:
:return:
|
[
"获取股票分钟线",
":",
"param",
"code",
":",
":",
"param",
"start",
":",
":",
"param",
"end",
":",
":",
"param",
"frequence",
":",
":",
"param",
"if_drop_index",
":",
":",
"param",
"collections",
":",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L238-L290
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_stock_list_adv
|
def QA_fetch_stock_list_adv(collections=DATABASE.stock_list):
'''
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
'''
stock_list_items = QA_fetch_stock_list(collections)
if len(stock_list_items) == 0:
print("QA Error QA_fetch_stock_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.stock_list is empty!")
return None
return stock_list_items
|
python
|
def QA_fetch_stock_list_adv(collections=DATABASE.stock_list):
'''
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
'''
stock_list_items = QA_fetch_stock_list(collections)
if len(stock_list_items) == 0:
print("QA Error QA_fetch_stock_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.stock_list is empty!")
return None
return stock_list_items
|
[
"def",
"QA_fetch_stock_list_adv",
"(",
"collections",
"=",
"DATABASE",
".",
"stock_list",
")",
":",
"stock_list_items",
"=",
"QA_fetch_stock_list",
"(",
"collections",
")",
"if",
"len",
"(",
"stock_list_items",
")",
"==",
"0",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.stock_list is empty!\"",
")",
"return",
"None",
"return",
"stock_list_items"
] |
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
|
[
"获取股票列表",
":",
"param",
"collections",
":",
"mongodb",
"数据库",
":",
"return",
":",
"DataFrame"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L319-L329
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_index_list_adv
|
def QA_fetch_index_list_adv(collections=DATABASE.index_list):
'''
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
'''
index_list_items = QA_fetch_index_list(collections)
if len(index_list_items) == 0:
print("QA Error QA_fetch_index_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.index_list is empty!")
return None
return index_list_items
|
python
|
def QA_fetch_index_list_adv(collections=DATABASE.index_list):
'''
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
'''
index_list_items = QA_fetch_index_list(collections)
if len(index_list_items) == 0:
print("QA Error QA_fetch_index_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.index_list is empty!")
return None
return index_list_items
|
[
"def",
"QA_fetch_index_list_adv",
"(",
"collections",
"=",
"DATABASE",
".",
"index_list",
")",
":",
"index_list_items",
"=",
"QA_fetch_index_list",
"(",
"collections",
")",
"if",
"len",
"(",
"index_list_items",
")",
"==",
"0",
":",
"print",
"(",
"\"QA Error QA_fetch_index_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.index_list is empty!\"",
")",
"return",
"None",
"return",
"index_list_items"
] |
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
|
[
"获取股票列表",
":",
"param",
"collections",
":",
"mongodb",
"数据库",
":",
"return",
":",
"DataFrame"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L332-L342
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_future_day_adv
|
def QA_fetch_future_day_adv(
code,
start, end=None,
if_drop_index=True,
# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉
collections=DATABASE.index_day):
'''
:param code: code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return:
'''
'获取期货日线'
end = start if end is None else end
start = str(start)[0:10]
end = str(end)[0:10]
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# 🛠 todo 如果相等
res = QA_fetch_future_day(code, start, end, format='pd')
if res is None:
print("QA Error QA_fetch_future_day_adv parameter code=%s start=%s end=%s call QA_fetch_future_day return None" % (
code, start, end))
else:
res_set_index = res.set_index(['date', 'code'])
# if res_set_index is None:
# print("QA Error QA_fetch_index_day_adv set index 'date, code' return None")
# return None
return QA_DataStruct_Future_day(res_set_index)
|
python
|
def QA_fetch_future_day_adv(
code,
start, end=None,
if_drop_index=True,
# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉
collections=DATABASE.index_day):
'''
:param code: code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return:
'''
'获取期货日线'
end = start if end is None else end
start = str(start)[0:10]
end = str(end)[0:10]
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# 🛠 todo 如果相等
res = QA_fetch_future_day(code, start, end, format='pd')
if res is None:
print("QA Error QA_fetch_future_day_adv parameter code=%s start=%s end=%s call QA_fetch_future_day return None" % (
code, start, end))
else:
res_set_index = res.set_index(['date', 'code'])
# if res_set_index is None:
# print("QA Error QA_fetch_index_day_adv set index 'date, code' return None")
# return None
return QA_DataStruct_Future_day(res_set_index)
|
[
"def",
"QA_fetch_future_day_adv",
"(",
"code",
",",
"start",
",",
"end",
"=",
"None",
",",
"if_drop_index",
"=",
"True",
",",
"# 🛠 todo collections 参数没有用到, 且数据库是固定的, 这个变量后期去掉",
"collections",
"=",
"DATABASE",
".",
"index_day",
")",
":",
"'获取期货日线'",
"end",
"=",
"start",
"if",
"end",
"is",
"None",
"else",
"end",
"start",
"=",
"str",
"(",
"start",
")",
"[",
"0",
":",
"10",
"]",
"end",
"=",
"str",
"(",
"end",
")",
"[",
"0",
":",
"10",
"]",
"# 🛠 todo 报告错误 如果开始时间 在 结束时间之后",
"# 🛠 todo 如果相等",
"res",
"=",
"QA_fetch_future_day",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'pd'",
")",
"if",
"res",
"is",
"None",
":",
"print",
"(",
"\"QA Error QA_fetch_future_day_adv parameter code=%s start=%s end=%s call QA_fetch_future_day return None\"",
"%",
"(",
"code",
",",
"start",
",",
"end",
")",
")",
"else",
":",
"res_set_index",
"=",
"res",
".",
"set_index",
"(",
"[",
"'date'",
",",
"'code'",
"]",
")",
"# if res_set_index is None:",
"# print(\"QA Error QA_fetch_index_day_adv set index 'date, code' return None\")",
"# return None",
"return",
"QA_DataStruct_Future_day",
"(",
"res_set_index",
")"
] |
:param code: code: 字符串str eg 600085
:param start: 字符串str 开始日期 eg 2011-01-01
:param end: 字符串str 结束日期 eg 2011-05-01
:param if_drop_index: Ture False , dataframe drop index or not
:param collections: mongodb 数据库
:return:
|
[
":",
"param",
"code",
":",
"code",
":",
"字符串str",
"eg",
"600085",
":",
"param",
"start",
":",
"字符串str",
"开始日期",
"eg",
"2011",
"-",
"01",
"-",
"01",
":",
"param",
"end",
":",
"字符串str",
"结束日期",
"eg",
"2011",
"-",
"05",
"-",
"01",
":",
"param",
"if_drop_index",
":",
"Ture",
"False",
",",
"dataframe",
"drop",
"index",
"or",
"not",
":",
"param",
"collections",
":",
"mongodb",
"数据库",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L344-L375
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_future_min_adv
|
def QA_fetch_future_min_adv(
code,
start, end=None,
frequence='1min',
if_drop_index=True,
collections=DATABASE.future_min):
'''
'获取股票分钟线'
:param code:
:param start:
:param end:
:param frequence:
:param if_drop_index:
:param collections:
:return:
'''
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
# __data = [] 没有使用
end = start if end is None else end
if len(start) == 10:
start = '{} 00:00:00'.format(start)
if len(end) == 10:
end = '{} 15:00:00'.format(end)
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# if start == end:
# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_index_min_adv, 不支持start end是相等的
#print("QA Error QA_fetch_index_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! " % (code, start, end))
# return None
res = QA_fetch_future_min(
code, start, end, format='pd', frequence=frequence)
if res is None:
print("QA Error QA_fetch_future_min_adv parameter code=%s start=%s end=%s frequence=%s call QA_fetch_future_min return None" % (
code, start, end, frequence))
else:
res_reset_index = res.set_index(
['datetime', 'code'], drop=if_drop_index)
# if res_reset_index is None:
# print("QA Error QA_fetch_index_min_adv set index 'date, code' return None")
return QA_DataStruct_Future_min(res_reset_index)
|
python
|
def QA_fetch_future_min_adv(
code,
start, end=None,
frequence='1min',
if_drop_index=True,
collections=DATABASE.future_min):
'''
'获取股票分钟线'
:param code:
:param start:
:param end:
:param frequence:
:param if_drop_index:
:param collections:
:return:
'''
if frequence in ['1min', '1m']:
frequence = '1min'
elif frequence in ['5min', '5m']:
frequence = '5min'
elif frequence in ['15min', '15m']:
frequence = '15min'
elif frequence in ['30min', '30m']:
frequence = '30min'
elif frequence in ['60min', '60m']:
frequence = '60min'
# __data = [] 没有使用
end = start if end is None else end
if len(start) == 10:
start = '{} 00:00:00'.format(start)
if len(end) == 10:
end = '{} 15:00:00'.format(end)
# 🛠 todo 报告错误 如果开始时间 在 结束时间之后
# if start == end:
# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_index_min_adv, 不支持start end是相等的
#print("QA Error QA_fetch_index_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! " % (code, start, end))
# return None
res = QA_fetch_future_min(
code, start, end, format='pd', frequence=frequence)
if res is None:
print("QA Error QA_fetch_future_min_adv parameter code=%s start=%s end=%s frequence=%s call QA_fetch_future_min return None" % (
code, start, end, frequence))
else:
res_reset_index = res.set_index(
['datetime', 'code'], drop=if_drop_index)
# if res_reset_index is None:
# print("QA Error QA_fetch_index_min_adv set index 'date, code' return None")
return QA_DataStruct_Future_min(res_reset_index)
|
[
"def",
"QA_fetch_future_min_adv",
"(",
"code",
",",
"start",
",",
"end",
"=",
"None",
",",
"frequence",
"=",
"'1min'",
",",
"if_drop_index",
"=",
"True",
",",
"collections",
"=",
"DATABASE",
".",
"future_min",
")",
":",
"if",
"frequence",
"in",
"[",
"'1min'",
",",
"'1m'",
"]",
":",
"frequence",
"=",
"'1min'",
"elif",
"frequence",
"in",
"[",
"'5min'",
",",
"'5m'",
"]",
":",
"frequence",
"=",
"'5min'",
"elif",
"frequence",
"in",
"[",
"'15min'",
",",
"'15m'",
"]",
":",
"frequence",
"=",
"'15min'",
"elif",
"frequence",
"in",
"[",
"'30min'",
",",
"'30m'",
"]",
":",
"frequence",
"=",
"'30min'",
"elif",
"frequence",
"in",
"[",
"'60min'",
",",
"'60m'",
"]",
":",
"frequence",
"=",
"'60min'",
"# __data = [] 没有使用",
"end",
"=",
"start",
"if",
"end",
"is",
"None",
"else",
"end",
"if",
"len",
"(",
"start",
")",
"==",
"10",
":",
"start",
"=",
"'{} 00:00:00'",
".",
"format",
"(",
"start",
")",
"if",
"len",
"(",
"end",
")",
"==",
"10",
":",
"end",
"=",
"'{} 15:00:00'",
".",
"format",
"(",
"end",
")",
"# 🛠 todo 报告错误 如果开始时间 在 结束时间之后",
"# if start == end:",
"# 🛠 todo 如果相等,根据 frequence 获取开始时间的 时间段 QA_fetch_index_min_adv, 不支持start end是相等的",
"#print(\"QA Error QA_fetch_index_min_adv parameter code=%s , start=%s, end=%s is equal, should have time span! \" % (code, start, end))",
"# return None",
"res",
"=",
"QA_fetch_future_min",
"(",
"code",
",",
"start",
",",
"end",
",",
"format",
"=",
"'pd'",
",",
"frequence",
"=",
"frequence",
")",
"if",
"res",
"is",
"None",
":",
"print",
"(",
"\"QA Error QA_fetch_future_min_adv parameter code=%s start=%s end=%s frequence=%s call QA_fetch_future_min return None\"",
"%",
"(",
"code",
",",
"start",
",",
"end",
",",
"frequence",
")",
")",
"else",
":",
"res_reset_index",
"=",
"res",
".",
"set_index",
"(",
"[",
"'datetime'",
",",
"'code'",
"]",
",",
"drop",
"=",
"if_drop_index",
")",
"# if res_reset_index is None:",
"# print(\"QA Error QA_fetch_index_min_adv set index 'date, code' return None\")",
"return",
"QA_DataStruct_Future_min",
"(",
"res_reset_index",
")"
] |
'获取股票分钟线'
:param code:
:param start:
:param end:
:param frequence:
:param if_drop_index:
:param collections:
:return:
|
[
"获取股票分钟线",
":",
"param",
"code",
":",
":",
"param",
"start",
":",
":",
"param",
"end",
":",
":",
"param",
"frequence",
":",
":",
"param",
"if_drop_index",
":",
":",
"param",
"collections",
":",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L378-L430
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_future_list_adv
|
def QA_fetch_future_list_adv(collections=DATABASE.future_list):
'''
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
'''
future_list_items = QA_fetch_future_list()
if len(future_list_items) == 0:
print("QA Error QA_fetch_future_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.future_list is empty!")
return None
return future_list_items
|
python
|
def QA_fetch_future_list_adv(collections=DATABASE.future_list):
'''
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
'''
future_list_items = QA_fetch_future_list()
if len(future_list_items) == 0:
print("QA Error QA_fetch_future_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.future_list is empty!")
return None
return future_list_items
|
[
"def",
"QA_fetch_future_list_adv",
"(",
"collections",
"=",
"DATABASE",
".",
"future_list",
")",
":",
"future_list_items",
"=",
"QA_fetch_future_list",
"(",
")",
"if",
"len",
"(",
"future_list_items",
")",
"==",
"0",
":",
"print",
"(",
"\"QA Error QA_fetch_future_list_adv call item for item in collections.find() return 0 item, maybe the DATABASE.future_list is empty!\"",
")",
"return",
"None",
"return",
"future_list_items"
] |
'获取股票列表'
:param collections: mongodb 数据库
:return: DataFrame
|
[
"获取股票列表",
":",
"param",
"collections",
":",
"mongodb",
"数据库",
":",
"return",
":",
"DataFrame"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L433-L443
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_stock_block_adv
|
def QA_fetch_stock_block_adv(code=None, blockname=None, collections=DATABASE.stock_block):
'''
返回板块 ❌
:param code:
:param blockname:
:param collections: 默认数据库 stock_block
:return: QA_DataStruct_Stock_block
'''
if code is not None and blockname is None:
# 返回这个股票代码所属的板块
data = pd.DataFrame([item for item in collections.find(
{'code': {'$in': code}})])
data = data.drop(['_id'], axis=1)
return QA_DataStruct_Stock_block(data.set_index(['blockname', 'code'], drop=True).drop_duplicates())
elif blockname is not None and code is None:
#
# 🛠 todo fnished 返回 这个板块所有的股票
# 返回该板块所属的股票
# print("QA Error blockname is Not none code none, return all code from its block name have not implemented yet !")
items_from_collections = [item for item in collections.find(
{'blockname': re.compile(blockname)})]
data = pd.DataFrame(items_from_collections).drop(['_id'], axis=1)
data_set_index = data.set_index(['blockname', 'code'], drop=True)
return QA_DataStruct_Stock_block(data_set_index)
else:
# 🛠 todo 返回 判断 这个股票是否和属于该板块
data = pd.DataFrame(
[item for item in collections.find()]).drop(['_id'], axis=1)
data_set_index = data.set_index(['blockname', 'code'], drop=True)
return QA_DataStruct_Stock_block(data_set_index)
|
python
|
def QA_fetch_stock_block_adv(code=None, blockname=None, collections=DATABASE.stock_block):
'''
返回板块 ❌
:param code:
:param blockname:
:param collections: 默认数据库 stock_block
:return: QA_DataStruct_Stock_block
'''
if code is not None and blockname is None:
# 返回这个股票代码所属的板块
data = pd.DataFrame([item for item in collections.find(
{'code': {'$in': code}})])
data = data.drop(['_id'], axis=1)
return QA_DataStruct_Stock_block(data.set_index(['blockname', 'code'], drop=True).drop_duplicates())
elif blockname is not None and code is None:
#
# 🛠 todo fnished 返回 这个板块所有的股票
# 返回该板块所属的股票
# print("QA Error blockname is Not none code none, return all code from its block name have not implemented yet !")
items_from_collections = [item for item in collections.find(
{'blockname': re.compile(blockname)})]
data = pd.DataFrame(items_from_collections).drop(['_id'], axis=1)
data_set_index = data.set_index(['blockname', 'code'], drop=True)
return QA_DataStruct_Stock_block(data_set_index)
else:
# 🛠 todo 返回 判断 这个股票是否和属于该板块
data = pd.DataFrame(
[item for item in collections.find()]).drop(['_id'], axis=1)
data_set_index = data.set_index(['blockname', 'code'], drop=True)
return QA_DataStruct_Stock_block(data_set_index)
|
[
"def",
"QA_fetch_stock_block_adv",
"(",
"code",
"=",
"None",
",",
"blockname",
"=",
"None",
",",
"collections",
"=",
"DATABASE",
".",
"stock_block",
")",
":",
"if",
"code",
"is",
"not",
"None",
"and",
"blockname",
"is",
"None",
":",
"# 返回这个股票代码所属的板块",
"data",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
"}",
")",
"]",
")",
"data",
"=",
"data",
".",
"drop",
"(",
"[",
"'_id'",
"]",
",",
"axis",
"=",
"1",
")",
"return",
"QA_DataStruct_Stock_block",
"(",
"data",
".",
"set_index",
"(",
"[",
"'blockname'",
",",
"'code'",
"]",
",",
"drop",
"=",
"True",
")",
".",
"drop_duplicates",
"(",
")",
")",
"elif",
"blockname",
"is",
"not",
"None",
"and",
"code",
"is",
"None",
":",
"#",
"# 🛠 todo fnished 返回 这个板块所有的股票",
"# 返回该板块所属的股票",
"# print(\"QA Error blockname is Not none code none, return all code from its block name have not implemented yet !\")",
"items_from_collections",
"=",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"'blockname'",
":",
"re",
".",
"compile",
"(",
"blockname",
")",
"}",
")",
"]",
"data",
"=",
"pd",
".",
"DataFrame",
"(",
"items_from_collections",
")",
".",
"drop",
"(",
"[",
"'_id'",
"]",
",",
"axis",
"=",
"1",
")",
"data_set_index",
"=",
"data",
".",
"set_index",
"(",
"[",
"'blockname'",
",",
"'code'",
"]",
",",
"drop",
"=",
"True",
")",
"return",
"QA_DataStruct_Stock_block",
"(",
"data_set_index",
")",
"else",
":",
"# 🛠 todo 返回 判断 这个股票是否和属于该板块",
"data",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
")",
"]",
")",
".",
"drop",
"(",
"[",
"'_id'",
"]",
",",
"axis",
"=",
"1",
")",
"data_set_index",
"=",
"data",
".",
"set_index",
"(",
"[",
"'blockname'",
",",
"'code'",
"]",
",",
"drop",
"=",
"True",
")",
"return",
"QA_DataStruct_Stock_block",
"(",
"data_set_index",
")"
] |
返回板块 ❌
:param code:
:param blockname:
:param collections: 默认数据库 stock_block
:return: QA_DataStruct_Stock_block
|
[
"返回板块",
"❌",
":",
"param",
"code",
":",
":",
"param",
"blockname",
":",
":",
"param",
"collections",
":",
"默认数据库",
"stock_block",
":",
"return",
":",
"QA_DataStruct_Stock_block"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L446-L478
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_stock_realtime_adv
|
def QA_fetch_stock_realtime_adv(code=None,
num=1,
collections=DATABASE.get_collection('realtime_{}'.format(datetime.date.today()))):
'''
返回当日的上下五档, code可以是股票可以是list, num是每个股票获取的数量
:param code:
:param num:
:param collections: realtime_XXXX-XX-XX 每天实时时间
:return: DataFrame
'''
if code is not None:
# code 必须转换成list 去查询数据库
if isinstance(code, str):
code = [code]
elif isinstance(code, list):
pass
else:
print(
"QA Error QA_fetch_stock_realtime_adv parameter code is not List type or String type")
items_from_collections = [item for item in collections.find(
{'code': {'$in': code}}, limit=num*len(code), sort=[('datetime', pymongo.DESCENDING)])]
if items_from_collections is None:
print("QA Error QA_fetch_stock_realtime_adv find parameter code={} num={} collection={} return NOne".format(
code, num, collections))
return
data = pd.DataFrame(items_from_collections)
data_set_index = data.set_index(
['datetime', 'code'], drop=False).drop(['_id'], axis=1)
return data_set_index
else:
print("QA Error QA_fetch_stock_realtime_adv parameter code is None")
|
python
|
def QA_fetch_stock_realtime_adv(code=None,
num=1,
collections=DATABASE.get_collection('realtime_{}'.format(datetime.date.today()))):
'''
返回当日的上下五档, code可以是股票可以是list, num是每个股票获取的数量
:param code:
:param num:
:param collections: realtime_XXXX-XX-XX 每天实时时间
:return: DataFrame
'''
if code is not None:
# code 必须转换成list 去查询数据库
if isinstance(code, str):
code = [code]
elif isinstance(code, list):
pass
else:
print(
"QA Error QA_fetch_stock_realtime_adv parameter code is not List type or String type")
items_from_collections = [item for item in collections.find(
{'code': {'$in': code}}, limit=num*len(code), sort=[('datetime', pymongo.DESCENDING)])]
if items_from_collections is None:
print("QA Error QA_fetch_stock_realtime_adv find parameter code={} num={} collection={} return NOne".format(
code, num, collections))
return
data = pd.DataFrame(items_from_collections)
data_set_index = data.set_index(
['datetime', 'code'], drop=False).drop(['_id'], axis=1)
return data_set_index
else:
print("QA Error QA_fetch_stock_realtime_adv parameter code is None")
|
[
"def",
"QA_fetch_stock_realtime_adv",
"(",
"code",
"=",
"None",
",",
"num",
"=",
"1",
",",
"collections",
"=",
"DATABASE",
".",
"get_collection",
"(",
"'realtime_{}'",
".",
"format",
"(",
"datetime",
".",
"date",
".",
"today",
"(",
")",
")",
")",
")",
":",
"if",
"code",
"is",
"not",
"None",
":",
"# code 必须转换成list 去查询数据库",
"if",
"isinstance",
"(",
"code",
",",
"str",
")",
":",
"code",
"=",
"[",
"code",
"]",
"elif",
"isinstance",
"(",
"code",
",",
"list",
")",
":",
"pass",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_realtime_adv parameter code is not List type or String type\"",
")",
"items_from_collections",
"=",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"{",
"'$in'",
":",
"code",
"}",
"}",
",",
"limit",
"=",
"num",
"*",
"len",
"(",
"code",
")",
",",
"sort",
"=",
"[",
"(",
"'datetime'",
",",
"pymongo",
".",
"DESCENDING",
")",
"]",
")",
"]",
"if",
"items_from_collections",
"is",
"None",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_realtime_adv find parameter code={} num={} collection={} return NOne\"",
".",
"format",
"(",
"code",
",",
"num",
",",
"collections",
")",
")",
"return",
"data",
"=",
"pd",
".",
"DataFrame",
"(",
"items_from_collections",
")",
"data_set_index",
"=",
"data",
".",
"set_index",
"(",
"[",
"'datetime'",
",",
"'code'",
"]",
",",
"drop",
"=",
"False",
")",
".",
"drop",
"(",
"[",
"'_id'",
"]",
",",
"axis",
"=",
"1",
")",
"return",
"data_set_index",
"else",
":",
"print",
"(",
"\"QA Error QA_fetch_stock_realtime_adv parameter code is None\"",
")"
] |
返回当日的上下五档, code可以是股票可以是list, num是每个股票获取的数量
:param code:
:param num:
:param collections: realtime_XXXX-XX-XX 每天实时时间
:return: DataFrame
|
[
"返回当日的上下五档",
"code可以是股票可以是list",
"num是每个股票获取的数量",
":",
"param",
"code",
":",
":",
"param",
"num",
":",
":",
"param",
"collections",
":",
"realtime_XXXX",
"-",
"XX",
"-",
"XX",
"每天实时时间",
":",
"return",
":",
"DataFrame"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L481-L513
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_financial_report_adv
|
def QA_fetch_financial_report_adv(code, start, end=None, ltype='EN'):
"""高级财务查询接口
Arguments:
code {[type]} -- [description]
start {[type]} -- [description]
Keyword Arguments:
end {[type]} -- [description] (default: {None})
"""
if end is None:
return QA_DataStruct_Financial(QA_fetch_financial_report(code, start, ltype=ltype))
else:
series = pd.Series(
data=month_data, index=pd.to_datetime(month_data), name='date')
timerange = series.loc[start:end].tolist()
return QA_DataStruct_Financial(QA_fetch_financial_report(code, timerange, ltype=ltype))
|
python
|
def QA_fetch_financial_report_adv(code, start, end=None, ltype='EN'):
"""高级财务查询接口
Arguments:
code {[type]} -- [description]
start {[type]} -- [description]
Keyword Arguments:
end {[type]} -- [description] (default: {None})
"""
if end is None:
return QA_DataStruct_Financial(QA_fetch_financial_report(code, start, ltype=ltype))
else:
series = pd.Series(
data=month_data, index=pd.to_datetime(month_data), name='date')
timerange = series.loc[start:end].tolist()
return QA_DataStruct_Financial(QA_fetch_financial_report(code, timerange, ltype=ltype))
|
[
"def",
"QA_fetch_financial_report_adv",
"(",
"code",
",",
"start",
",",
"end",
"=",
"None",
",",
"ltype",
"=",
"'EN'",
")",
":",
"if",
"end",
"is",
"None",
":",
"return",
"QA_DataStruct_Financial",
"(",
"QA_fetch_financial_report",
"(",
"code",
",",
"start",
",",
"ltype",
"=",
"ltype",
")",
")",
"else",
":",
"series",
"=",
"pd",
".",
"Series",
"(",
"data",
"=",
"month_data",
",",
"index",
"=",
"pd",
".",
"to_datetime",
"(",
"month_data",
")",
",",
"name",
"=",
"'date'",
")",
"timerange",
"=",
"series",
".",
"loc",
"[",
"start",
":",
"end",
"]",
".",
"tolist",
"(",
")",
"return",
"QA_DataStruct_Financial",
"(",
"QA_fetch_financial_report",
"(",
"code",
",",
"timerange",
",",
"ltype",
"=",
"ltype",
")",
")"
] |
高级财务查询接口
Arguments:
code {[type]} -- [description]
start {[type]} -- [description]
Keyword Arguments:
end {[type]} -- [description] (default: {None})
|
[
"高级财务查询接口",
"Arguments",
":",
"code",
"{",
"[",
"type",
"]",
"}",
"--",
"[",
"description",
"]",
"start",
"{",
"[",
"type",
"]",
"}",
"--",
"[",
"description",
"]",
"Keyword",
"Arguments",
":",
"end",
"{",
"[",
"type",
"]",
"}",
"--",
"[",
"description",
"]",
"(",
"default",
":",
"{",
"None",
"}",
")"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L516-L532
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QAQuery_Advance.py
|
QA_fetch_stock_financial_calendar_adv
|
def QA_fetch_stock_financial_calendar_adv(code, start="all", end=None, format='pd', collections=DATABASE.report_calendar):
'获取股票日线'
#code= [code] if isinstance(code,str) else code
end = start if end is None else end
start = str(start)[0:10]
end = str(end)[0:10]
# code checking
if start == 'all':
start = '1990-01-01'
end = str(datetime.date.today())
if end is None:
return QA_DataStruct_Financial(QA_fetch_stock_financial_calendar(code, start, str(datetime.date.today())))
else:
series = pd.Series(
data=month_data, index=pd.to_datetime(month_data), name='date')
timerange = series.loc[start:end].tolist()
return QA_DataStruct_Financial(QA_fetch_stock_financial_calendar(code, start, end))
|
python
|
def QA_fetch_stock_financial_calendar_adv(code, start="all", end=None, format='pd', collections=DATABASE.report_calendar):
'获取股票日线'
#code= [code] if isinstance(code,str) else code
end = start if end is None else end
start = str(start)[0:10]
end = str(end)[0:10]
# code checking
if start == 'all':
start = '1990-01-01'
end = str(datetime.date.today())
if end is None:
return QA_DataStruct_Financial(QA_fetch_stock_financial_calendar(code, start, str(datetime.date.today())))
else:
series = pd.Series(
data=month_data, index=pd.to_datetime(month_data), name='date')
timerange = series.loc[start:end].tolist()
return QA_DataStruct_Financial(QA_fetch_stock_financial_calendar(code, start, end))
|
[
"def",
"QA_fetch_stock_financial_calendar_adv",
"(",
"code",
",",
"start",
"=",
"\"all\"",
",",
"end",
"=",
"None",
",",
"format",
"=",
"'pd'",
",",
"collections",
"=",
"DATABASE",
".",
"report_calendar",
")",
":",
"#code= [code] if isinstance(code,str) else code",
"end",
"=",
"start",
"if",
"end",
"is",
"None",
"else",
"end",
"start",
"=",
"str",
"(",
"start",
")",
"[",
"0",
":",
"10",
"]",
"end",
"=",
"str",
"(",
"end",
")",
"[",
"0",
":",
"10",
"]",
"# code checking",
"if",
"start",
"==",
"'all'",
":",
"start",
"=",
"'1990-01-01'",
"end",
"=",
"str",
"(",
"datetime",
".",
"date",
".",
"today",
"(",
")",
")",
"if",
"end",
"is",
"None",
":",
"return",
"QA_DataStruct_Financial",
"(",
"QA_fetch_stock_financial_calendar",
"(",
"code",
",",
"start",
",",
"str",
"(",
"datetime",
".",
"date",
".",
"today",
"(",
")",
")",
")",
")",
"else",
":",
"series",
"=",
"pd",
".",
"Series",
"(",
"data",
"=",
"month_data",
",",
"index",
"=",
"pd",
".",
"to_datetime",
"(",
"month_data",
")",
",",
"name",
"=",
"'date'",
")",
"timerange",
"=",
"series",
".",
"loc",
"[",
"start",
":",
"end",
"]",
".",
"tolist",
"(",
")",
"return",
"QA_DataStruct_Financial",
"(",
"QA_fetch_stock_financial_calendar",
"(",
"code",
",",
"start",
",",
"end",
")",
")"
] |
获取股票日线
|
[
"获取股票日线"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QAQuery_Advance.py#L572-L591
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAFetch/QATradeFile.py
|
QA_fetch_get_tdxtraderecord
|
def QA_fetch_get_tdxtraderecord(file):
"""
QUANTAXIS 读取历史交易记录 通达信 历史成交-输出-xlsfile--转换csvfile
"""
try:
with open('./20180606.csv', 'r') as f:
l = csv.reader(f)
data = [item for item in l]
res = pd.DataFrame(data[1:], columns=data[0])
return res
except:
raise IOError('QA CANNOT READ THIS RECORD')
|
python
|
def QA_fetch_get_tdxtraderecord(file):
"""
QUANTAXIS 读取历史交易记录 通达信 历史成交-输出-xlsfile--转换csvfile
"""
try:
with open('./20180606.csv', 'r') as f:
l = csv.reader(f)
data = [item for item in l]
res = pd.DataFrame(data[1:], columns=data[0])
return res
except:
raise IOError('QA CANNOT READ THIS RECORD')
|
[
"def",
"QA_fetch_get_tdxtraderecord",
"(",
"file",
")",
":",
"try",
":",
"with",
"open",
"(",
"'./20180606.csv'",
",",
"'r'",
")",
"as",
"f",
":",
"l",
"=",
"csv",
".",
"reader",
"(",
"f",
")",
"data",
"=",
"[",
"item",
"for",
"item",
"in",
"l",
"]",
"res",
"=",
"pd",
".",
"DataFrame",
"(",
"data",
"[",
"1",
":",
"]",
",",
"columns",
"=",
"data",
"[",
"0",
"]",
")",
"return",
"res",
"except",
":",
"raise",
"IOError",
"(",
"'QA CANNOT READ THIS RECORD'",
")"
] |
QUANTAXIS 读取历史交易记录 通达信 历史成交-输出-xlsfile--转换csvfile
|
[
"QUANTAXIS",
"读取历史交易记录",
"通达信",
"历史成交",
"-",
"输出",
"-",
"xlsfile",
"--",
"转换csvfile"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAFetch/QATradeFile.py#L39-L51
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAIndicator/talib_indicators.py
|
AROON
|
def AROON(DataFrame, N=14):
"""阿隆指标
Arguments:
DataFrame {[type]} -- [description]
Keyword Arguments:
N {int} -- [description] (default: {14})
Returns:
[type] -- [description]
"""
ar_up, ar_down = talib.AROON(DataFrame.high.values, DataFrame.low.values, N)
return pd.DataFrame({'AROON_UP': ar_up,'AROON_DOWN': ar_down}, index=DataFrame.index)
|
python
|
def AROON(DataFrame, N=14):
"""阿隆指标
Arguments:
DataFrame {[type]} -- [description]
Keyword Arguments:
N {int} -- [description] (default: {14})
Returns:
[type] -- [description]
"""
ar_up, ar_down = talib.AROON(DataFrame.high.values, DataFrame.low.values, N)
return pd.DataFrame({'AROON_UP': ar_up,'AROON_DOWN': ar_down}, index=DataFrame.index)
|
[
"def",
"AROON",
"(",
"DataFrame",
",",
"N",
"=",
"14",
")",
":",
"ar_up",
",",
"ar_down",
"=",
"talib",
".",
"AROON",
"(",
"DataFrame",
".",
"high",
".",
"values",
",",
"DataFrame",
".",
"low",
".",
"values",
",",
"N",
")",
"return",
"pd",
".",
"DataFrame",
"(",
"{",
"'AROON_UP'",
":",
"ar_up",
",",
"'AROON_DOWN'",
":",
"ar_down",
"}",
",",
"index",
"=",
"DataFrame",
".",
"index",
")"
] |
阿隆指标
Arguments:
DataFrame {[type]} -- [description]
Keyword Arguments:
N {int} -- [description] (default: {14})
Returns:
[type] -- [description]
|
[
"阿隆指标",
"Arguments",
":",
"DataFrame",
"{",
"[",
"type",
"]",
"}",
"--",
"[",
"description",
"]",
"Keyword",
"Arguments",
":",
"N",
"{",
"int",
"}",
"--",
"[",
"description",
"]",
"(",
"default",
":",
"{",
"14",
"}",
")",
"Returns",
":",
"[",
"type",
"]",
"--",
"[",
"description",
"]"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAIndicator/talib_indicators.py#L55-L69
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAARP/market_preset.py
|
MARKET_PRESET.get_commission_coeff
|
def get_commission_coeff(self, code):
"""
当前无法区分是百分比还是按手数收费,不过可以拿到以后自行判断
"""
return max(self.get_code(code).get('commission_coeff_peramount'),
self.get_code(code).get('commission_coeff_pervol'))
|
python
|
def get_commission_coeff(self, code):
"""
当前无法区分是百分比还是按手数收费,不过可以拿到以后自行判断
"""
return max(self.get_code(code).get('commission_coeff_peramount'),
self.get_code(code).get('commission_coeff_pervol'))
|
[
"def",
"get_commission_coeff",
"(",
"self",
",",
"code",
")",
":",
"return",
"max",
"(",
"self",
".",
"get_code",
"(",
"code",
")",
".",
"get",
"(",
"'commission_coeff_peramount'",
")",
",",
"self",
".",
"get_code",
"(",
"code",
")",
".",
"get",
"(",
"'commission_coeff_pervol'",
")",
")"
] |
当前无法区分是百分比还是按手数收费,不过可以拿到以后自行判断
|
[
"当前无法区分是百分比还是按手数收费",
"不过可以拿到以后自行判断"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAARP/market_preset.py#L638-L643
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAAnalysis/QAAnalysis_trade.py
|
QAAnalysis_trade.import_trade
|
def import_trade(self, trade):
"""
trade是一个可迭代的list/generator
"""
for item in trade:
self.make_deal(item.code, item.datetime, item.amount,
item.towards, item.price.item.order_model, item.amount_model)
|
python
|
def import_trade(self, trade):
"""
trade是一个可迭代的list/generator
"""
for item in trade:
self.make_deal(item.code, item.datetime, item.amount,
item.towards, item.price.item.order_model, item.amount_model)
|
[
"def",
"import_trade",
"(",
"self",
",",
"trade",
")",
":",
"for",
"item",
"in",
"trade",
":",
"self",
".",
"make_deal",
"(",
"item",
".",
"code",
",",
"item",
".",
"datetime",
",",
"item",
".",
"amount",
",",
"item",
".",
"towards",
",",
"item",
".",
"price",
".",
"item",
".",
"order_model",
",",
"item",
".",
"amount_model",
")"
] |
trade是一个可迭代的list/generator
|
[
"trade是一个可迭代的list",
"/",
"generator"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAAnalysis/QAAnalysis_trade.py#L54-L60
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAAnalysis/QAAnalysis_trade.py
|
QAAnalysis_trade.make_deal
|
def make_deal(self, code, datetime, amount=100, towards=ORDER_DIRECTION.BUY, price=0, order_model=ORDER_MODEL.MARKET, amount_model=AMOUNT_MODEL.BY_AMOUNT):
"""
这是一个一定会成交,并且立刻结转(及t+0)的交易入口
"""
self.account.receive_deal(self.backtest_broker.receive_order(QA_Event(order=self.account.send_order(
code=code, time=datetime, amount=amount, towards=towards, price=price, order_model=order_model, amount_model=amount_model
))))
self.account.settle()
|
python
|
def make_deal(self, code, datetime, amount=100, towards=ORDER_DIRECTION.BUY, price=0, order_model=ORDER_MODEL.MARKET, amount_model=AMOUNT_MODEL.BY_AMOUNT):
"""
这是一个一定会成交,并且立刻结转(及t+0)的交易入口
"""
self.account.receive_deal(self.backtest_broker.receive_order(QA_Event(order=self.account.send_order(
code=code, time=datetime, amount=amount, towards=towards, price=price, order_model=order_model, amount_model=amount_model
))))
self.account.settle()
|
[
"def",
"make_deal",
"(",
"self",
",",
"code",
",",
"datetime",
",",
"amount",
"=",
"100",
",",
"towards",
"=",
"ORDER_DIRECTION",
".",
"BUY",
",",
"price",
"=",
"0",
",",
"order_model",
"=",
"ORDER_MODEL",
".",
"MARKET",
",",
"amount_model",
"=",
"AMOUNT_MODEL",
".",
"BY_AMOUNT",
")",
":",
"self",
".",
"account",
".",
"receive_deal",
"(",
"self",
".",
"backtest_broker",
".",
"receive_order",
"(",
"QA_Event",
"(",
"order",
"=",
"self",
".",
"account",
".",
"send_order",
"(",
"code",
"=",
"code",
",",
"time",
"=",
"datetime",
",",
"amount",
"=",
"amount",
",",
"towards",
"=",
"towards",
",",
"price",
"=",
"price",
",",
"order_model",
"=",
"order_model",
",",
"amount_model",
"=",
"amount_model",
")",
")",
")",
")",
"self",
".",
"account",
".",
"settle",
"(",
")"
] |
这是一个一定会成交,并且立刻结转(及t+0)的交易入口
|
[
"这是一个一定会成交",
"并且立刻结转",
"(",
"及t",
"+",
"0",
")",
"的交易入口"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAAnalysis/QAAnalysis_trade.py#L62-L69
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_Order.cancel
|
def cancel(self):
"""撤单
Arguments:
amount {int} -- 撤单数量
"""
self.cancel_amount = self.amount - self.trade_amount
if self.trade_amount == 0:
# 未交易 直接订单全撤
self._status = ORDER_STATUS.CANCEL_ALL
else:
# 部分交易 剩余订单全撤
self._status = ORDER_STATUS.CANCEL_PART
|
python
|
def cancel(self):
"""撤单
Arguments:
amount {int} -- 撤单数量
"""
self.cancel_amount = self.amount - self.trade_amount
if self.trade_amount == 0:
# 未交易 直接订单全撤
self._status = ORDER_STATUS.CANCEL_ALL
else:
# 部分交易 剩余订单全撤
self._status = ORDER_STATUS.CANCEL_PART
|
[
"def",
"cancel",
"(",
"self",
")",
":",
"self",
".",
"cancel_amount",
"=",
"self",
".",
"amount",
"-",
"self",
".",
"trade_amount",
"if",
"self",
".",
"trade_amount",
"==",
"0",
":",
"# 未交易 直接订单全撤",
"self",
".",
"_status",
"=",
"ORDER_STATUS",
".",
"CANCEL_ALL",
"else",
":",
"# 部分交易 剩余订单全撤",
"self",
".",
"_status",
"=",
"ORDER_STATUS",
".",
"CANCEL_PART"
] |
撤单
Arguments:
amount {int} -- 撤单数量
|
[
"撤单"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L246-L259
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_Order.failed
|
def failed(self, reason=None):
"""失败订单(未成功创建入broker)
Arguments:
reason {str} -- 失败原因
"""
# 订单创建失败(如废单/场外废单/价格高于涨停价/价格低于跌停价/通讯失败)
self._status = ORDER_STATUS.FAILED
self.reason = str(reason)
|
python
|
def failed(self, reason=None):
"""失败订单(未成功创建入broker)
Arguments:
reason {str} -- 失败原因
"""
# 订单创建失败(如废单/场外废单/价格高于涨停价/价格低于跌停价/通讯失败)
self._status = ORDER_STATUS.FAILED
self.reason = str(reason)
|
[
"def",
"failed",
"(",
"self",
",",
"reason",
"=",
"None",
")",
":",
"# 订单创建失败(如废单/场外废单/价格高于涨停价/价格低于跌停价/通讯失败)",
"self",
".",
"_status",
"=",
"ORDER_STATUS",
".",
"FAILED",
"self",
".",
"reason",
"=",
"str",
"(",
"reason",
")"
] |
失败订单(未成功创建入broker)
Arguments:
reason {str} -- 失败原因
|
[
"失败订单",
"(",
"未成功创建入broker",
")"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L261-L269
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_Order.trade
|
def trade(self, trade_id, trade_price, trade_amount, trade_time):
"""trade 状态
Arguments:
amount {[type]} -- [description]
"""
if self.status in [ORDER_STATUS.SUCCESS_PART, ORDER_STATUS.QUEUED]:
trade_amount = int(trade_amount)
trade_id = str(trade_id)
if trade_amount < 1:
self._status = ORDER_STATUS.NEXT
else:
if trade_id not in self.trade_id:
trade_price = float(trade_price)
trade_time = str(trade_time)
self.trade_id.append(trade_id)
self.trade_price = (
self.trade_price * self.trade_amount +
trade_price * trade_amount
) / (
self.trade_amount + trade_amount
)
self.trade_amount += trade_amount
self.trade_time.append(trade_time)
self.callback(
self.code,
trade_id,
self.order_id,
self.realorder_id,
trade_price,
trade_amount,
self.towards,
trade_time
)
else:
pass
else:
raise RuntimeError(
'ORDER STATUS {} CANNNOT TRADE'.format(self.status)
)
|
python
|
def trade(self, trade_id, trade_price, trade_amount, trade_time):
"""trade 状态
Arguments:
amount {[type]} -- [description]
"""
if self.status in [ORDER_STATUS.SUCCESS_PART, ORDER_STATUS.QUEUED]:
trade_amount = int(trade_amount)
trade_id = str(trade_id)
if trade_amount < 1:
self._status = ORDER_STATUS.NEXT
else:
if trade_id not in self.trade_id:
trade_price = float(trade_price)
trade_time = str(trade_time)
self.trade_id.append(trade_id)
self.trade_price = (
self.trade_price * self.trade_amount +
trade_price * trade_amount
) / (
self.trade_amount + trade_amount
)
self.trade_amount += trade_amount
self.trade_time.append(trade_time)
self.callback(
self.code,
trade_id,
self.order_id,
self.realorder_id,
trade_price,
trade_amount,
self.towards,
trade_time
)
else:
pass
else:
raise RuntimeError(
'ORDER STATUS {} CANNNOT TRADE'.format(self.status)
)
|
[
"def",
"trade",
"(",
"self",
",",
"trade_id",
",",
"trade_price",
",",
"trade_amount",
",",
"trade_time",
")",
":",
"if",
"self",
".",
"status",
"in",
"[",
"ORDER_STATUS",
".",
"SUCCESS_PART",
",",
"ORDER_STATUS",
".",
"QUEUED",
"]",
":",
"trade_amount",
"=",
"int",
"(",
"trade_amount",
")",
"trade_id",
"=",
"str",
"(",
"trade_id",
")",
"if",
"trade_amount",
"<",
"1",
":",
"self",
".",
"_status",
"=",
"ORDER_STATUS",
".",
"NEXT",
"else",
":",
"if",
"trade_id",
"not",
"in",
"self",
".",
"trade_id",
":",
"trade_price",
"=",
"float",
"(",
"trade_price",
")",
"trade_time",
"=",
"str",
"(",
"trade_time",
")",
"self",
".",
"trade_id",
".",
"append",
"(",
"trade_id",
")",
"self",
".",
"trade_price",
"=",
"(",
"self",
".",
"trade_price",
"*",
"self",
".",
"trade_amount",
"+",
"trade_price",
"*",
"trade_amount",
")",
"/",
"(",
"self",
".",
"trade_amount",
"+",
"trade_amount",
")",
"self",
".",
"trade_amount",
"+=",
"trade_amount",
"self",
".",
"trade_time",
".",
"append",
"(",
"trade_time",
")",
"self",
".",
"callback",
"(",
"self",
".",
"code",
",",
"trade_id",
",",
"self",
".",
"order_id",
",",
"self",
".",
"realorder_id",
",",
"trade_price",
",",
"trade_amount",
",",
"self",
".",
"towards",
",",
"trade_time",
")",
"else",
":",
"pass",
"else",
":",
"raise",
"RuntimeError",
"(",
"'ORDER STATUS {} CANNNOT TRADE'",
".",
"format",
"(",
"self",
".",
"status",
")",
")"
] |
trade 状态
Arguments:
amount {[type]} -- [description]
|
[
"trade",
"状态"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L271-L314
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_Order.to_otgdict
|
def to_otgdict(self):
"""{
"aid": "insert_order", # //必填, 下单请求
# //必填, 需要与登录用户名一致, 或为登录用户的子账户(例如登录用户为user1, 则报单 user_id 应当为 user1 或 user1.some_unit)
"user_id": account_cookie,
# //必填, 委托单号, 需确保在一个账号中不重复, 限长512字节
"order_id": order_id if order_id else QA.QA_util_random_with_topic('QAOTG'),
"exchange_id": exchange_id, # //必填, 下单到哪个交易所
"instrument_id": code, # //必填, 下单合约代码
"direction": order_direction, # //必填, 下单买卖方向
# //必填, 下单开平方向, 仅当指令相关对象不支持开平机制(例如股票)时可不填写此字段
"offset": order_offset,
"volume": volume, # //必填, 下单手数
"price_type": "LIMIT", # //必填, 报单价格类型
"limit_price": price, # //当 price_type == LIMIT 时需要填写此字段, 报单价格
"volume_condition": "ANY",
"time_condition": "GFD",
}
"""
return {
"aid": "insert_order", # //必填, 下单请求
# //必填, 需要与登录用户名一致, 或为登录用户的子账户(例如登录用户为user1, 则报单 user_id 应当为 user1 或 user1.some_unit)
"user_id": self.account_cookie,
# //必填, 委托单号, 需确保在一个账号中不重复, 限长512字节
"order_id": self.order_id,
"exchange_id": self.exchange_id, # //必填, 下单到哪个交易所
"instrument_id": self.code, # //必填, 下单合约代码
"direction": self.direction, # //必填, 下单买卖方向
# //必填, 下单开平方向, 仅当指令相关对象不支持开平机制(例如股票)时可不填写此字段
"offset": self.offset,
"volume": self.amount, # //必填, 下单手数
"price_type": self.order_model, # //必填, 报单价格类型
"limit_price": self.price, # //当 price_type == LIMIT 时需要填写此字段, 报单价格
"volume_condition": "ANY",
"time_condition": "GFD",
}
|
python
|
def to_otgdict(self):
"""{
"aid": "insert_order", # //必填, 下单请求
# //必填, 需要与登录用户名一致, 或为登录用户的子账户(例如登录用户为user1, 则报单 user_id 应当为 user1 或 user1.some_unit)
"user_id": account_cookie,
# //必填, 委托单号, 需确保在一个账号中不重复, 限长512字节
"order_id": order_id if order_id else QA.QA_util_random_with_topic('QAOTG'),
"exchange_id": exchange_id, # //必填, 下单到哪个交易所
"instrument_id": code, # //必填, 下单合约代码
"direction": order_direction, # //必填, 下单买卖方向
# //必填, 下单开平方向, 仅当指令相关对象不支持开平机制(例如股票)时可不填写此字段
"offset": order_offset,
"volume": volume, # //必填, 下单手数
"price_type": "LIMIT", # //必填, 报单价格类型
"limit_price": price, # //当 price_type == LIMIT 时需要填写此字段, 报单价格
"volume_condition": "ANY",
"time_condition": "GFD",
}
"""
return {
"aid": "insert_order", # //必填, 下单请求
# //必填, 需要与登录用户名一致, 或为登录用户的子账户(例如登录用户为user1, 则报单 user_id 应当为 user1 或 user1.some_unit)
"user_id": self.account_cookie,
# //必填, 委托单号, 需确保在一个账号中不重复, 限长512字节
"order_id": self.order_id,
"exchange_id": self.exchange_id, # //必填, 下单到哪个交易所
"instrument_id": self.code, # //必填, 下单合约代码
"direction": self.direction, # //必填, 下单买卖方向
# //必填, 下单开平方向, 仅当指令相关对象不支持开平机制(例如股票)时可不填写此字段
"offset": self.offset,
"volume": self.amount, # //必填, 下单手数
"price_type": self.order_model, # //必填, 报单价格类型
"limit_price": self.price, # //当 price_type == LIMIT 时需要填写此字段, 报单价格
"volume_condition": "ANY",
"time_condition": "GFD",
}
|
[
"def",
"to_otgdict",
"(",
"self",
")",
":",
"return",
"{",
"\"aid\"",
":",
"\"insert_order\"",
",",
"# //必填, 下单请求",
"# //必填, 需要与登录用户名一致, 或为登录用户的子账户(例如登录用户为user1, 则报单 user_id 应当为 user1 或 user1.some_unit)",
"\"user_id\"",
":",
"self",
".",
"account_cookie",
",",
"# //必填, 委托单号, 需确保在一个账号中不重复, 限长512字节",
"\"order_id\"",
":",
"self",
".",
"order_id",
",",
"\"exchange_id\"",
":",
"self",
".",
"exchange_id",
",",
"# //必填, 下单到哪个交易所",
"\"instrument_id\"",
":",
"self",
".",
"code",
",",
"# //必填, 下单合约代码",
"\"direction\"",
":",
"self",
".",
"direction",
",",
"# //必填, 下单买卖方向",
"# //必填, 下单开平方向, 仅当指令相关对象不支持开平机制(例如股票)时可不填写此字段",
"\"offset\"",
":",
"self",
".",
"offset",
",",
"\"volume\"",
":",
"self",
".",
"amount",
",",
"# //必填, 下单手数",
"\"price_type\"",
":",
"self",
".",
"order_model",
",",
"# //必填, 报单价格类型",
"\"limit_price\"",
":",
"self",
".",
"price",
",",
"# //当 price_type == LIMIT 时需要填写此字段, 报单价格",
"\"volume_condition\"",
":",
"\"ANY\"",
",",
"\"time_condition\"",
":",
"\"GFD\"",
",",
"}"
] |
{
"aid": "insert_order", # //必填, 下单请求
# //必填, 需要与登录用户名一致, 或为登录用户的子账户(例如登录用户为user1, 则报单 user_id 应当为 user1 或 user1.some_unit)
"user_id": account_cookie,
# //必填, 委托单号, 需确保在一个账号中不重复, 限长512字节
"order_id": order_id if order_id else QA.QA_util_random_with_topic('QAOTG'),
"exchange_id": exchange_id, # //必填, 下单到哪个交易所
"instrument_id": code, # //必填, 下单合约代码
"direction": order_direction, # //必填, 下单买卖方向
# //必填, 下单开平方向, 仅当指令相关对象不支持开平机制(例如股票)时可不填写此字段
"offset": order_offset,
"volume": volume, # //必填, 下单手数
"price_type": "LIMIT", # //必填, 报单价格类型
"limit_price": price, # //当 price_type == LIMIT 时需要填写此字段, 报单价格
"volume_condition": "ANY",
"time_condition": "GFD",
}
|
[
"{",
"aid",
":",
"insert_order",
"#",
"//",
"必填",
"下单请求",
"#",
"//",
"必填",
"需要与登录用户名一致",
"或为登录用户的子账户",
"(",
"例如登录用户为user1",
"则报单",
"user_id",
"应当为",
"user1",
"或",
"user1",
".",
"some_unit",
")",
"user_id",
":",
"account_cookie",
"#",
"//",
"必填",
"委托单号",
"需确保在一个账号中不重复",
"限长512字节",
"order_id",
":",
"order_id",
"if",
"order_id",
"else",
"QA",
".",
"QA_util_random_with_topic",
"(",
"QAOTG",
")",
"exchange_id",
":",
"exchange_id",
"#",
"//",
"必填",
"下单到哪个交易所",
"instrument_id",
":",
"code",
"#",
"//",
"必填",
"下单合约代码",
"direction",
":",
"order_direction",
"#",
"//",
"必填",
"下单买卖方向",
"#",
"//",
"必填",
"下单开平方向",
"仅当指令相关对象不支持开平机制",
"(",
"例如股票",
")",
"时可不填写此字段",
"offset",
":",
"order_offset",
"volume",
":",
"volume",
"#",
"//",
"必填",
"下单手数",
"price_type",
":",
"LIMIT",
"#",
"//",
"必填",
"报单价格类型",
"limit_price",
":",
"price",
"#",
"//",
"当",
"price_type",
"==",
"LIMIT",
"时需要填写此字段",
"报单价格",
"volume_condition",
":",
"ANY",
"time_condition",
":",
"GFD",
"}"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L365-L400
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_Order.from_otgformat
|
def from_otgformat(self, otgOrder):
"""[summary]
Arguments:
otgOrder {[type]} -- [description]
{'seqno': 6,
'user_id': '106184',
'order_id': 'WDRB_QA01_FtNlyBem',
'exchange_id': 'SHFE',
'instrument_id': 'rb1905',
'direction': 'SELL',
'offset': 'OPEN',
'volume_orign': 50, #(总报单手数)
'price_type': 'LIMIT', # "LIMIT" (价格类型, ANY=市价, LIMIT=限价)
'limit_price': 3432.0, # 4500.0 (委托价格, 仅当 price_type = LIMIT 时有效)
'time_condition': 'GFD',# "GFD" (时间条件, IOC=立即完成,否则撤销, GFS=本节有效, GFD=当日有效, GTC=撤销前有效, GFA=集合竞价有效)
'volume_condition': 'ANY', # "ANY" (手数条件, ANY=任何数量, MIN=最小数量, ALL=全部数量)
'insert_date_time': 1545656460000000000,# 1501074872000000000 (下单时间(按北京时间),自unix epoch(1970-01-01 00:00:00 GMT)以来的纳秒数)
'exchange_order_id': ' 3738',
'status': 'FINISHED', # "ALIVE" (委托单状态, ALIVE=有效, FINISHED=已完)
'volume_left': 0,
'last_msg': '全部成交报单已提交'} # "报单成功" (委托单状态信息)
"""
self.order_id = otgOrder.get('order_id')
self.account_cookie = otgOrder.get('user_id')
self.exchange_id = otgOrder.get('exchange_id')
self.code = str(otgOrder.get('instrument_id')).upper()
self.offset = otgOrder.get('offset')
self.direction = otgOrder.get('direction')
self.towards = eval('ORDER_DIRECTION.{}_{}'.format(
self.direction,
self.offset
))
self.amount = otgOrder.get('volume_orign')
self.trade_amount = self.amount - otgOrder.get('volume_left')
self.price = otgOrder.get('limit_price')
self.order_model = eval(
'ORDER_MODEL.{}'.format(otgOrder.get('price_type'))
)
self.time_condition = otgOrder.get('time_condition')
if otgOrder.get('insert_date_time') == 0:
self.datetime = 0
else:
self.datetime = QA_util_stamp2datetime(
int(otgOrder.get('insert_date_time'))
)
self.sending_time = self.datetime
self.volume_condition = otgOrder.get('volume_condition')
self.message = otgOrder.get('last_msg')
self._status = ORDER_STATUS.NEW
if '已撤单' in self.message or '拒绝' in self.message or '仓位不足' in self.message:
# 仓位不足: 一般是平今/平昨仓位不足
self._status = ORDER_STATUS.FAILED
self.realorder_id = otgOrder.get('exchange_order_id')
return self
|
python
|
def from_otgformat(self, otgOrder):
"""[summary]
Arguments:
otgOrder {[type]} -- [description]
{'seqno': 6,
'user_id': '106184',
'order_id': 'WDRB_QA01_FtNlyBem',
'exchange_id': 'SHFE',
'instrument_id': 'rb1905',
'direction': 'SELL',
'offset': 'OPEN',
'volume_orign': 50, #(总报单手数)
'price_type': 'LIMIT', # "LIMIT" (价格类型, ANY=市价, LIMIT=限价)
'limit_price': 3432.0, # 4500.0 (委托价格, 仅当 price_type = LIMIT 时有效)
'time_condition': 'GFD',# "GFD" (时间条件, IOC=立即完成,否则撤销, GFS=本节有效, GFD=当日有效, GTC=撤销前有效, GFA=集合竞价有效)
'volume_condition': 'ANY', # "ANY" (手数条件, ANY=任何数量, MIN=最小数量, ALL=全部数量)
'insert_date_time': 1545656460000000000,# 1501074872000000000 (下单时间(按北京时间),自unix epoch(1970-01-01 00:00:00 GMT)以来的纳秒数)
'exchange_order_id': ' 3738',
'status': 'FINISHED', # "ALIVE" (委托单状态, ALIVE=有效, FINISHED=已完)
'volume_left': 0,
'last_msg': '全部成交报单已提交'} # "报单成功" (委托单状态信息)
"""
self.order_id = otgOrder.get('order_id')
self.account_cookie = otgOrder.get('user_id')
self.exchange_id = otgOrder.get('exchange_id')
self.code = str(otgOrder.get('instrument_id')).upper()
self.offset = otgOrder.get('offset')
self.direction = otgOrder.get('direction')
self.towards = eval('ORDER_DIRECTION.{}_{}'.format(
self.direction,
self.offset
))
self.amount = otgOrder.get('volume_orign')
self.trade_amount = self.amount - otgOrder.get('volume_left')
self.price = otgOrder.get('limit_price')
self.order_model = eval(
'ORDER_MODEL.{}'.format(otgOrder.get('price_type'))
)
self.time_condition = otgOrder.get('time_condition')
if otgOrder.get('insert_date_time') == 0:
self.datetime = 0
else:
self.datetime = QA_util_stamp2datetime(
int(otgOrder.get('insert_date_time'))
)
self.sending_time = self.datetime
self.volume_condition = otgOrder.get('volume_condition')
self.message = otgOrder.get('last_msg')
self._status = ORDER_STATUS.NEW
if '已撤单' in self.message or '拒绝' in self.message or '仓位不足' in self.message:
# 仓位不足: 一般是平今/平昨仓位不足
self._status = ORDER_STATUS.FAILED
self.realorder_id = otgOrder.get('exchange_order_id')
return self
|
[
"def",
"from_otgformat",
"(",
"self",
",",
"otgOrder",
")",
":",
"self",
".",
"order_id",
"=",
"otgOrder",
".",
"get",
"(",
"'order_id'",
")",
"self",
".",
"account_cookie",
"=",
"otgOrder",
".",
"get",
"(",
"'user_id'",
")",
"self",
".",
"exchange_id",
"=",
"otgOrder",
".",
"get",
"(",
"'exchange_id'",
")",
"self",
".",
"code",
"=",
"str",
"(",
"otgOrder",
".",
"get",
"(",
"'instrument_id'",
")",
")",
".",
"upper",
"(",
")",
"self",
".",
"offset",
"=",
"otgOrder",
".",
"get",
"(",
"'offset'",
")",
"self",
".",
"direction",
"=",
"otgOrder",
".",
"get",
"(",
"'direction'",
")",
"self",
".",
"towards",
"=",
"eval",
"(",
"'ORDER_DIRECTION.{}_{}'",
".",
"format",
"(",
"self",
".",
"direction",
",",
"self",
".",
"offset",
")",
")",
"self",
".",
"amount",
"=",
"otgOrder",
".",
"get",
"(",
"'volume_orign'",
")",
"self",
".",
"trade_amount",
"=",
"self",
".",
"amount",
"-",
"otgOrder",
".",
"get",
"(",
"'volume_left'",
")",
"self",
".",
"price",
"=",
"otgOrder",
".",
"get",
"(",
"'limit_price'",
")",
"self",
".",
"order_model",
"=",
"eval",
"(",
"'ORDER_MODEL.{}'",
".",
"format",
"(",
"otgOrder",
".",
"get",
"(",
"'price_type'",
")",
")",
")",
"self",
".",
"time_condition",
"=",
"otgOrder",
".",
"get",
"(",
"'time_condition'",
")",
"if",
"otgOrder",
".",
"get",
"(",
"'insert_date_time'",
")",
"==",
"0",
":",
"self",
".",
"datetime",
"=",
"0",
"else",
":",
"self",
".",
"datetime",
"=",
"QA_util_stamp2datetime",
"(",
"int",
"(",
"otgOrder",
".",
"get",
"(",
"'insert_date_time'",
")",
")",
")",
"self",
".",
"sending_time",
"=",
"self",
".",
"datetime",
"self",
".",
"volume_condition",
"=",
"otgOrder",
".",
"get",
"(",
"'volume_condition'",
")",
"self",
".",
"message",
"=",
"otgOrder",
".",
"get",
"(",
"'last_msg'",
")",
"self",
".",
"_status",
"=",
"ORDER_STATUS",
".",
"NEW",
"if",
"'已撤单' in se",
"f.",
"essa",
"g",
"e or '拒",
"' ",
"n self.m",
"ss",
"ge o",
"r",
" '仓位不足'",
"in",
"self.message:",
"",
"",
"",
"",
"",
"# 仓位不足: 一般是平今/平昨仓位不足",
"self",
".",
"_status",
"=",
"ORDER_STATUS",
".",
"FAILED",
"self",
".",
"realorder_id",
"=",
"otgOrder",
".",
"get",
"(",
"'exchange_order_id'",
")",
"return",
"self"
] |
[summary]
Arguments:
otgOrder {[type]} -- [description]
{'seqno': 6,
'user_id': '106184',
'order_id': 'WDRB_QA01_FtNlyBem',
'exchange_id': 'SHFE',
'instrument_id': 'rb1905',
'direction': 'SELL',
'offset': 'OPEN',
'volume_orign': 50, #(总报单手数)
'price_type': 'LIMIT', # "LIMIT" (价格类型, ANY=市价, LIMIT=限价)
'limit_price': 3432.0, # 4500.0 (委托价格, 仅当 price_type = LIMIT 时有效)
'time_condition': 'GFD',# "GFD" (时间条件, IOC=立即完成,否则撤销, GFS=本节有效, GFD=当日有效, GTC=撤销前有效, GFA=集合竞价有效)
'volume_condition': 'ANY', # "ANY" (手数条件, ANY=任何数量, MIN=最小数量, ALL=全部数量)
'insert_date_time': 1545656460000000000,# 1501074872000000000 (下单时间(按北京时间),自unix epoch(1970-01-01 00:00:00 GMT)以来的纳秒数)
'exchange_order_id': ' 3738',
'status': 'FINISHED', # "ALIVE" (委托单状态, ALIVE=有效, FINISHED=已完)
'volume_left': 0,
'last_msg': '全部成交报单已提交'} # "报单成功" (委托单状态信息)
|
[
"[",
"summary",
"]"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L419-L476
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_Order.from_dict
|
def from_dict(self, order_dict):
'''
从字段类型的字段 填充 对象的字段
:param order_dict: dict 类型
:return: self QA_Order
'''
try:
# QA_util_log_info('QA_ORDER CHANGE: from {} change to {}'.format(
# self.order_id, order['order_id']))
self.price = order_dict['price']
self.date = order_dict['date']
self.datetime = order_dict['datetime']
self.sending_time = order_dict['sending_time'] # 下单时间
self.trade_time = order_dict['trade_time']
self.amount = order_dict['amount']
self.frequence = order_dict['frequence']
self.market_type = order_dict['market_type']
self.towards = order_dict['towards']
self.code = order_dict['code']
self.user = order_dict['user']
self.account_cookie = order_dict['account_cookie']
self.strategy = order_dict['strategy']
self.type = order_dict['type']
self.order_model = order_dict['order_model']
self.amount_model = order_dict['amount_model']
self.order_id = order_dict['order_id']
self.realorder_id = order_dict['realorder_id']
self.trade_id = order_dict['trade_id']
self.callback = order_dict['callback']
self.commission_coeff = order_dict['commission_coeff']
self.tax_coeff = order_dict['tax_coeff']
self.money = order_dict['money']
self._status = order_dict['_status']
self.cancel_amount = order_dict['cancel_amount']
self.trade_amount = order_dict['trade_amount']
self.trade_price = order_dict['trade_price']
self.reason = order_dict['reason']
return self
except Exception as e:
QA_util_log_info('Failed to tran from dict {}'.format(e))
|
python
|
def from_dict(self, order_dict):
'''
从字段类型的字段 填充 对象的字段
:param order_dict: dict 类型
:return: self QA_Order
'''
try:
# QA_util_log_info('QA_ORDER CHANGE: from {} change to {}'.format(
# self.order_id, order['order_id']))
self.price = order_dict['price']
self.date = order_dict['date']
self.datetime = order_dict['datetime']
self.sending_time = order_dict['sending_time'] # 下单时间
self.trade_time = order_dict['trade_time']
self.amount = order_dict['amount']
self.frequence = order_dict['frequence']
self.market_type = order_dict['market_type']
self.towards = order_dict['towards']
self.code = order_dict['code']
self.user = order_dict['user']
self.account_cookie = order_dict['account_cookie']
self.strategy = order_dict['strategy']
self.type = order_dict['type']
self.order_model = order_dict['order_model']
self.amount_model = order_dict['amount_model']
self.order_id = order_dict['order_id']
self.realorder_id = order_dict['realorder_id']
self.trade_id = order_dict['trade_id']
self.callback = order_dict['callback']
self.commission_coeff = order_dict['commission_coeff']
self.tax_coeff = order_dict['tax_coeff']
self.money = order_dict['money']
self._status = order_dict['_status']
self.cancel_amount = order_dict['cancel_amount']
self.trade_amount = order_dict['trade_amount']
self.trade_price = order_dict['trade_price']
self.reason = order_dict['reason']
return self
except Exception as e:
QA_util_log_info('Failed to tran from dict {}'.format(e))
|
[
"def",
"from_dict",
"(",
"self",
",",
"order_dict",
")",
":",
"try",
":",
"# QA_util_log_info('QA_ORDER CHANGE: from {} change to {}'.format(",
"# self.order_id, order['order_id']))",
"self",
".",
"price",
"=",
"order_dict",
"[",
"'price'",
"]",
"self",
".",
"date",
"=",
"order_dict",
"[",
"'date'",
"]",
"self",
".",
"datetime",
"=",
"order_dict",
"[",
"'datetime'",
"]",
"self",
".",
"sending_time",
"=",
"order_dict",
"[",
"'sending_time'",
"]",
"# 下单时间",
"self",
".",
"trade_time",
"=",
"order_dict",
"[",
"'trade_time'",
"]",
"self",
".",
"amount",
"=",
"order_dict",
"[",
"'amount'",
"]",
"self",
".",
"frequence",
"=",
"order_dict",
"[",
"'frequence'",
"]",
"self",
".",
"market_type",
"=",
"order_dict",
"[",
"'market_type'",
"]",
"self",
".",
"towards",
"=",
"order_dict",
"[",
"'towards'",
"]",
"self",
".",
"code",
"=",
"order_dict",
"[",
"'code'",
"]",
"self",
".",
"user",
"=",
"order_dict",
"[",
"'user'",
"]",
"self",
".",
"account_cookie",
"=",
"order_dict",
"[",
"'account_cookie'",
"]",
"self",
".",
"strategy",
"=",
"order_dict",
"[",
"'strategy'",
"]",
"self",
".",
"type",
"=",
"order_dict",
"[",
"'type'",
"]",
"self",
".",
"order_model",
"=",
"order_dict",
"[",
"'order_model'",
"]",
"self",
".",
"amount_model",
"=",
"order_dict",
"[",
"'amount_model'",
"]",
"self",
".",
"order_id",
"=",
"order_dict",
"[",
"'order_id'",
"]",
"self",
".",
"realorder_id",
"=",
"order_dict",
"[",
"'realorder_id'",
"]",
"self",
".",
"trade_id",
"=",
"order_dict",
"[",
"'trade_id'",
"]",
"self",
".",
"callback",
"=",
"order_dict",
"[",
"'callback'",
"]",
"self",
".",
"commission_coeff",
"=",
"order_dict",
"[",
"'commission_coeff'",
"]",
"self",
".",
"tax_coeff",
"=",
"order_dict",
"[",
"'tax_coeff'",
"]",
"self",
".",
"money",
"=",
"order_dict",
"[",
"'money'",
"]",
"self",
".",
"_status",
"=",
"order_dict",
"[",
"'_status'",
"]",
"self",
".",
"cancel_amount",
"=",
"order_dict",
"[",
"'cancel_amount'",
"]",
"self",
".",
"trade_amount",
"=",
"order_dict",
"[",
"'trade_amount'",
"]",
"self",
".",
"trade_price",
"=",
"order_dict",
"[",
"'trade_price'",
"]",
"self",
".",
"reason",
"=",
"order_dict",
"[",
"'reason'",
"]",
"return",
"self",
"except",
"Exception",
"as",
"e",
":",
"QA_util_log_info",
"(",
"'Failed to tran from dict {}'",
".",
"format",
"(",
"e",
")",
")"
] |
从字段类型的字段 填充 对象的字段
:param order_dict: dict 类型
:return: self QA_Order
|
[
"从字段类型的字段",
"填充",
"对象的字段",
":",
"param",
"order_dict",
":",
"dict",
"类型",
":",
"return",
":",
"self",
"QA_Order"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L478-L520
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_OrderQueue.insert_order
|
def insert_order(self, order):
'''
:param order: QA_Order类型
:return:
'''
#print(" *>> QAOrder!insert_order {}".format(order))
# QUEUED = 300 # queued 用于表示在order_queue中 实际表达的意思是订单存活 待成交
#order.status = ORDER_STATUS.QUEUED
# 🛠 todo 是为了速度快把order对象转换成 df 对象的吗?
#self.queue_df = self.queue_df.append(order.to_df(), ignore_index=True)
#self.queue_df.set_index('order_id', drop=True, inplace=True)
if order is not None:
self.order_list[order.order_id] = order
return order
else:
print('QAERROR Wrong for get None type while insert order to Queue')
|
python
|
def insert_order(self, order):
'''
:param order: QA_Order类型
:return:
'''
#print(" *>> QAOrder!insert_order {}".format(order))
# QUEUED = 300 # queued 用于表示在order_queue中 实际表达的意思是订单存活 待成交
#order.status = ORDER_STATUS.QUEUED
# 🛠 todo 是为了速度快把order对象转换成 df 对象的吗?
#self.queue_df = self.queue_df.append(order.to_df(), ignore_index=True)
#self.queue_df.set_index('order_id', drop=True, inplace=True)
if order is not None:
self.order_list[order.order_id] = order
return order
else:
print('QAERROR Wrong for get None type while insert order to Queue')
|
[
"def",
"insert_order",
"(",
"self",
",",
"order",
")",
":",
"#print(\" *>> QAOrder!insert_order {}\".format(order))",
"# QUEUED = 300 # queued 用于表示在order_queue中 实际表达的意思是订单存活 待成交",
"#order.status = ORDER_STATUS.QUEUED",
"# 🛠 todo 是为了速度快把order对象转换成 df 对象的吗?",
"#self.queue_df = self.queue_df.append(order.to_df(), ignore_index=True)",
"#self.queue_df.set_index('order_id', drop=True, inplace=True)",
"if",
"order",
"is",
"not",
"None",
":",
"self",
".",
"order_list",
"[",
"order",
".",
"order_id",
"]",
"=",
"order",
"return",
"order",
"else",
":",
"print",
"(",
"'QAERROR Wrong for get None type while insert order to Queue'",
")"
] |
:param order: QA_Order类型
:return:
|
[
":",
"param",
"order",
":",
"QA_Order类型",
":",
"return",
":"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L557-L572
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAMarket/QAOrder.py
|
QA_OrderQueue.pending
|
def pending(self):
'''
600 废单 未委托成功
200 委托成功,完全交易
203 委托成功,未完全成功
300 委托队列 待成交
400 已撤单
500 服务器撤单/每日结算
订单生成(100) -- 废单(600)
订单生成(100) -- 进入待成交队列(300) -- 完全成交(200) -- 每日结算(500)-- 死亡
订单生成(100) -- 进入待成交队列(300) -- 部分成交(203) -- 未成交(300) -- 每日结算(500) -- 死亡
订单生成(100) -- 进入待成交队列(300) -- 主动撤单(400) -- 每日结算(500) -- 死亡
选择待成交列表
:return: dataframe
'''
try:
return [
item for item in self.order_list.values() if item.status in [
ORDER_STATUS.QUEUED,
ORDER_STATUS.NEXT,
ORDER_STATUS.SUCCESS_PART
]
]
except:
return []
|
python
|
def pending(self):
'''
600 废单 未委托成功
200 委托成功,完全交易
203 委托成功,未完全成功
300 委托队列 待成交
400 已撤单
500 服务器撤单/每日结算
订单生成(100) -- 废单(600)
订单生成(100) -- 进入待成交队列(300) -- 完全成交(200) -- 每日结算(500)-- 死亡
订单生成(100) -- 进入待成交队列(300) -- 部分成交(203) -- 未成交(300) -- 每日结算(500) -- 死亡
订单生成(100) -- 进入待成交队列(300) -- 主动撤单(400) -- 每日结算(500) -- 死亡
选择待成交列表
:return: dataframe
'''
try:
return [
item for item in self.order_list.values() if item.status in [
ORDER_STATUS.QUEUED,
ORDER_STATUS.NEXT,
ORDER_STATUS.SUCCESS_PART
]
]
except:
return []
|
[
"def",
"pending",
"(",
"self",
")",
":",
"try",
":",
"return",
"[",
"item",
"for",
"item",
"in",
"self",
".",
"order_list",
".",
"values",
"(",
")",
"if",
"item",
".",
"status",
"in",
"[",
"ORDER_STATUS",
".",
"QUEUED",
",",
"ORDER_STATUS",
".",
"NEXT",
",",
"ORDER_STATUS",
".",
"SUCCESS_PART",
"]",
"]",
"except",
":",
"return",
"[",
"]"
] |
600 废单 未委托成功
200 委托成功,完全交易
203 委托成功,未完全成功
300 委托队列 待成交
400 已撤单
500 服务器撤单/每日结算
订单生成(100) -- 废单(600)
订单生成(100) -- 进入待成交队列(300) -- 完全成交(200) -- 每日结算(500)-- 死亡
订单生成(100) -- 进入待成交队列(300) -- 部分成交(203) -- 未成交(300) -- 每日结算(500) -- 死亡
订单生成(100) -- 进入待成交队列(300) -- 主动撤单(400) -- 每日结算(500) -- 死亡
选择待成交列表
:return: dataframe
|
[
"600",
"废单",
"未委托成功",
"200",
"委托成功",
"完全交易",
"203",
"委托成功",
"未完全成功",
"300",
"委托队列",
"待成交",
"400",
"已撤单",
"500",
"服务器撤单",
"/",
"每日结算"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAMarket/QAOrder.py#L593-L619
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAData/data_fq.py
|
_QA_data_stock_to_fq
|
def _QA_data_stock_to_fq(bfq_data, xdxr_data, fqtype):
'使用数据库数据进行复权'
info = xdxr_data.query('category==1')
bfq_data = bfq_data.assign(if_trade=1)
if len(info) > 0:
data = pd.concat(
[
bfq_data,
info.loc[bfq_data.index[0]:bfq_data.index[-1],
['category']]
],
axis=1
)
data['if_trade'].fillna(value=0, inplace=True)
data = data.fillna(method='ffill')
data = pd.concat(
[
data,
info.loc[bfq_data.index[0]:bfq_data.index[-1],
['fenhong',
'peigu',
'peigujia',
'songzhuangu']]
],
axis=1
)
else:
data = pd.concat(
[
bfq_data,
info.
loc[:,
['category',
'fenhong',
'peigu',
'peigujia',
'songzhuangu']]
],
axis=1
)
data = data.fillna(0)
data['preclose'] = (
data['close'].shift(1) * 10 - data['fenhong'] +
data['peigu'] * data['peigujia']
) / (10 + data['peigu'] + data['songzhuangu'])
if fqtype in ['01', 'qfq']:
data['adj'] = (data['preclose'].shift(-1) /
data['close']).fillna(1)[::-1].cumprod()
else:
data['adj'] = (data['close'] /
data['preclose'].shift(-1)).cumprod().shift(1).fillna(1)
for col in ['open', 'high', 'low', 'close', 'preclose']:
data[col] = data[col] * data['adj']
data['volume'] = data['volume'] / \
data['adj'] if 'volume' in data.columns else data['vol']/data['adj']
try:
data['high_limit'] = data['high_limit'] * data['adj']
data['low_limit'] = data['high_limit'] * data['adj']
except:
pass
return data.query('if_trade==1 and open != 0').drop(
['fenhong',
'peigu',
'peigujia',
'songzhuangu',
'if_trade',
'category'],
axis=1,
errors='ignore'
)
|
python
|
def _QA_data_stock_to_fq(bfq_data, xdxr_data, fqtype):
'使用数据库数据进行复权'
info = xdxr_data.query('category==1')
bfq_data = bfq_data.assign(if_trade=1)
if len(info) > 0:
data = pd.concat(
[
bfq_data,
info.loc[bfq_data.index[0]:bfq_data.index[-1],
['category']]
],
axis=1
)
data['if_trade'].fillna(value=0, inplace=True)
data = data.fillna(method='ffill')
data = pd.concat(
[
data,
info.loc[bfq_data.index[0]:bfq_data.index[-1],
['fenhong',
'peigu',
'peigujia',
'songzhuangu']]
],
axis=1
)
else:
data = pd.concat(
[
bfq_data,
info.
loc[:,
['category',
'fenhong',
'peigu',
'peigujia',
'songzhuangu']]
],
axis=1
)
data = data.fillna(0)
data['preclose'] = (
data['close'].shift(1) * 10 - data['fenhong'] +
data['peigu'] * data['peigujia']
) / (10 + data['peigu'] + data['songzhuangu'])
if fqtype in ['01', 'qfq']:
data['adj'] = (data['preclose'].shift(-1) /
data['close']).fillna(1)[::-1].cumprod()
else:
data['adj'] = (data['close'] /
data['preclose'].shift(-1)).cumprod().shift(1).fillna(1)
for col in ['open', 'high', 'low', 'close', 'preclose']:
data[col] = data[col] * data['adj']
data['volume'] = data['volume'] / \
data['adj'] if 'volume' in data.columns else data['vol']/data['adj']
try:
data['high_limit'] = data['high_limit'] * data['adj']
data['low_limit'] = data['high_limit'] * data['adj']
except:
pass
return data.query('if_trade==1 and open != 0').drop(
['fenhong',
'peigu',
'peigujia',
'songzhuangu',
'if_trade',
'category'],
axis=1,
errors='ignore'
)
|
[
"def",
"_QA_data_stock_to_fq",
"(",
"bfq_data",
",",
"xdxr_data",
",",
"fqtype",
")",
":",
"info",
"=",
"xdxr_data",
".",
"query",
"(",
"'category==1'",
")",
"bfq_data",
"=",
"bfq_data",
".",
"assign",
"(",
"if_trade",
"=",
"1",
")",
"if",
"len",
"(",
"info",
")",
">",
"0",
":",
"data",
"=",
"pd",
".",
"concat",
"(",
"[",
"bfq_data",
",",
"info",
".",
"loc",
"[",
"bfq_data",
".",
"index",
"[",
"0",
"]",
":",
"bfq_data",
".",
"index",
"[",
"-",
"1",
"]",
",",
"[",
"'category'",
"]",
"]",
"]",
",",
"axis",
"=",
"1",
")",
"data",
"[",
"'if_trade'",
"]",
".",
"fillna",
"(",
"value",
"=",
"0",
",",
"inplace",
"=",
"True",
")",
"data",
"=",
"data",
".",
"fillna",
"(",
"method",
"=",
"'ffill'",
")",
"data",
"=",
"pd",
".",
"concat",
"(",
"[",
"data",
",",
"info",
".",
"loc",
"[",
"bfq_data",
".",
"index",
"[",
"0",
"]",
":",
"bfq_data",
".",
"index",
"[",
"-",
"1",
"]",
",",
"[",
"'fenhong'",
",",
"'peigu'",
",",
"'peigujia'",
",",
"'songzhuangu'",
"]",
"]",
"]",
",",
"axis",
"=",
"1",
")",
"else",
":",
"data",
"=",
"pd",
".",
"concat",
"(",
"[",
"bfq_data",
",",
"info",
".",
"loc",
"[",
":",
",",
"[",
"'category'",
",",
"'fenhong'",
",",
"'peigu'",
",",
"'peigujia'",
",",
"'songzhuangu'",
"]",
"]",
"]",
",",
"axis",
"=",
"1",
")",
"data",
"=",
"data",
".",
"fillna",
"(",
"0",
")",
"data",
"[",
"'preclose'",
"]",
"=",
"(",
"data",
"[",
"'close'",
"]",
".",
"shift",
"(",
"1",
")",
"*",
"10",
"-",
"data",
"[",
"'fenhong'",
"]",
"+",
"data",
"[",
"'peigu'",
"]",
"*",
"data",
"[",
"'peigujia'",
"]",
")",
"/",
"(",
"10",
"+",
"data",
"[",
"'peigu'",
"]",
"+",
"data",
"[",
"'songzhuangu'",
"]",
")",
"if",
"fqtype",
"in",
"[",
"'01'",
",",
"'qfq'",
"]",
":",
"data",
"[",
"'adj'",
"]",
"=",
"(",
"data",
"[",
"'preclose'",
"]",
".",
"shift",
"(",
"-",
"1",
")",
"/",
"data",
"[",
"'close'",
"]",
")",
".",
"fillna",
"(",
"1",
")",
"[",
":",
":",
"-",
"1",
"]",
".",
"cumprod",
"(",
")",
"else",
":",
"data",
"[",
"'adj'",
"]",
"=",
"(",
"data",
"[",
"'close'",
"]",
"/",
"data",
"[",
"'preclose'",
"]",
".",
"shift",
"(",
"-",
"1",
")",
")",
".",
"cumprod",
"(",
")",
".",
"shift",
"(",
"1",
")",
".",
"fillna",
"(",
"1",
")",
"for",
"col",
"in",
"[",
"'open'",
",",
"'high'",
",",
"'low'",
",",
"'close'",
",",
"'preclose'",
"]",
":",
"data",
"[",
"col",
"]",
"=",
"data",
"[",
"col",
"]",
"*",
"data",
"[",
"'adj'",
"]",
"data",
"[",
"'volume'",
"]",
"=",
"data",
"[",
"'volume'",
"]",
"/",
"data",
"[",
"'adj'",
"]",
"if",
"'volume'",
"in",
"data",
".",
"columns",
"else",
"data",
"[",
"'vol'",
"]",
"/",
"data",
"[",
"'adj'",
"]",
"try",
":",
"data",
"[",
"'high_limit'",
"]",
"=",
"data",
"[",
"'high_limit'",
"]",
"*",
"data",
"[",
"'adj'",
"]",
"data",
"[",
"'low_limit'",
"]",
"=",
"data",
"[",
"'high_limit'",
"]",
"*",
"data",
"[",
"'adj'",
"]",
"except",
":",
"pass",
"return",
"data",
".",
"query",
"(",
"'if_trade==1 and open != 0'",
")",
".",
"drop",
"(",
"[",
"'fenhong'",
",",
"'peigu'",
",",
"'peigujia'",
",",
"'songzhuangu'",
",",
"'if_trade'",
",",
"'category'",
"]",
",",
"axis",
"=",
"1",
",",
"errors",
"=",
"'ignore'",
")"
] |
使用数据库数据进行复权
|
[
"使用数据库数据进行复权"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAData/data_fq.py#L102-L176
|
train
|
QUANTAXIS/QUANTAXIS
|
QUANTAXIS/QAData/data_fq.py
|
QA_data_stock_to_fq
|
def QA_data_stock_to_fq(__data, type_='01'):
def __QA_fetch_stock_xdxr(
code,
format_='pd',
collections=DATABASE.stock_xdxr
):
'获取股票除权信息/数据库'
try:
data = pd.DataFrame(
[item for item in collections.find({'code': code})]
).drop(['_id'],
axis=1)
data['date'] = pd.to_datetime(data['date'])
return data.set_index(['date', 'code'], drop=False)
except:
return pd.DataFrame(
data=[],
columns=[
'category',
'category_meaning',
'code',
'date',
'fenhong',
'fenshu',
'liquidity_after',
'liquidity_before',
'name',
'peigu',
'peigujia',
'shares_after',
'shares_before',
'songzhuangu',
'suogu',
'xingquanjia'
]
)
'股票 日线/分钟线 动态复权接口'
code = __data.index.remove_unused_levels().levels[1][0] if isinstance(
__data.index,
pd.core.indexes.multi.MultiIndex
) else __data['code'][0]
return _QA_data_stock_to_fq(
bfq_data=__data,
xdxr_data=__QA_fetch_stock_xdxr(code),
fqtype=type_
)
|
python
|
def QA_data_stock_to_fq(__data, type_='01'):
def __QA_fetch_stock_xdxr(
code,
format_='pd',
collections=DATABASE.stock_xdxr
):
'获取股票除权信息/数据库'
try:
data = pd.DataFrame(
[item for item in collections.find({'code': code})]
).drop(['_id'],
axis=1)
data['date'] = pd.to_datetime(data['date'])
return data.set_index(['date', 'code'], drop=False)
except:
return pd.DataFrame(
data=[],
columns=[
'category',
'category_meaning',
'code',
'date',
'fenhong',
'fenshu',
'liquidity_after',
'liquidity_before',
'name',
'peigu',
'peigujia',
'shares_after',
'shares_before',
'songzhuangu',
'suogu',
'xingquanjia'
]
)
'股票 日线/分钟线 动态复权接口'
code = __data.index.remove_unused_levels().levels[1][0] if isinstance(
__data.index,
pd.core.indexes.multi.MultiIndex
) else __data['code'][0]
return _QA_data_stock_to_fq(
bfq_data=__data,
xdxr_data=__QA_fetch_stock_xdxr(code),
fqtype=type_
)
|
[
"def",
"QA_data_stock_to_fq",
"(",
"__data",
",",
"type_",
"=",
"'01'",
")",
":",
"def",
"__QA_fetch_stock_xdxr",
"(",
"code",
",",
"format_",
"=",
"'pd'",
",",
"collections",
"=",
"DATABASE",
".",
"stock_xdxr",
")",
":",
"'获取股票除权信息/数据库'",
"try",
":",
"data",
"=",
"pd",
".",
"DataFrame",
"(",
"[",
"item",
"for",
"item",
"in",
"collections",
".",
"find",
"(",
"{",
"'code'",
":",
"code",
"}",
")",
"]",
")",
".",
"drop",
"(",
"[",
"'_id'",
"]",
",",
"axis",
"=",
"1",
")",
"data",
"[",
"'date'",
"]",
"=",
"pd",
".",
"to_datetime",
"(",
"data",
"[",
"'date'",
"]",
")",
"return",
"data",
".",
"set_index",
"(",
"[",
"'date'",
",",
"'code'",
"]",
",",
"drop",
"=",
"False",
")",
"except",
":",
"return",
"pd",
".",
"DataFrame",
"(",
"data",
"=",
"[",
"]",
",",
"columns",
"=",
"[",
"'category'",
",",
"'category_meaning'",
",",
"'code'",
",",
"'date'",
",",
"'fenhong'",
",",
"'fenshu'",
",",
"'liquidity_after'",
",",
"'liquidity_before'",
",",
"'name'",
",",
"'peigu'",
",",
"'peigujia'",
",",
"'shares_after'",
",",
"'shares_before'",
",",
"'songzhuangu'",
",",
"'suogu'",
",",
"'xingquanjia'",
"]",
")",
"code",
"=",
"__data",
".",
"index",
".",
"remove_unused_levels",
"(",
")",
".",
"levels",
"[",
"1",
"]",
"[",
"0",
"]",
"if",
"isinstance",
"(",
"__data",
".",
"index",
",",
"pd",
".",
"core",
".",
"indexes",
".",
"multi",
".",
"MultiIndex",
")",
"else",
"__data",
"[",
"'code'",
"]",
"[",
"0",
"]",
"return",
"_QA_data_stock_to_fq",
"(",
"bfq_data",
"=",
"__data",
",",
"xdxr_data",
"=",
"__QA_fetch_stock_xdxr",
"(",
"code",
")",
",",
"fqtype",
"=",
"type_",
")"
] |
股票 日线/分钟线 动态复权接口
|
[
"股票",
"日线",
"/",
"分钟线",
"动态复权接口"
] |
bb1fe424e4108b62a1f712b81a05cf829297a5c0
|
https://github.com/QUANTAXIS/QUANTAXIS/blob/bb1fe424e4108b62a1f712b81a05cf829297a5c0/QUANTAXIS/QAData/data_fq.py#L179-L228
|
train
|
gunthercox/ChatterBot
|
chatterbot/chatterbot.py
|
ChatBot.get_response
|
def get_response(self, statement=None, **kwargs):
"""
Return the bot's response based on the input.
:param statement: An statement object or string.
:returns: A response to the input.
:rtype: Statement
:param additional_response_selection_parameters: Parameters to pass to the
chat bot's logic adapters to control response selection.
:type additional_response_selection_parameters: dict
:param persist_values_to_response: Values that should be saved to the response
that the chat bot generates.
:type persist_values_to_response: dict
"""
Statement = self.storage.get_object('statement')
additional_response_selection_parameters = kwargs.pop('additional_response_selection_parameters', {})
persist_values_to_response = kwargs.pop('persist_values_to_response', {})
if isinstance(statement, str):
kwargs['text'] = statement
if isinstance(statement, dict):
kwargs.update(statement)
if statement is None and 'text' not in kwargs:
raise self.ChatBotException(
'Either a statement object or a "text" keyword '
'argument is required. Neither was provided.'
)
if hasattr(statement, 'serialize'):
kwargs.update(**statement.serialize())
tags = kwargs.pop('tags', [])
text = kwargs.pop('text')
input_statement = Statement(text=text, **kwargs)
input_statement.add_tags(*tags)
# Preprocess the input statement
for preprocessor in self.preprocessors:
input_statement = preprocessor(input_statement)
# Make sure the input statement has its search text saved
if not input_statement.search_text:
input_statement.search_text = self.storage.tagger.get_bigram_pair_string(input_statement.text)
if not input_statement.search_in_response_to and input_statement.in_response_to:
input_statement.search_in_response_to = self.storage.tagger.get_bigram_pair_string(input_statement.in_response_to)
response = self.generate_response(input_statement, additional_response_selection_parameters)
# Update any response data that needs to be changed
if persist_values_to_response:
for response_key in persist_values_to_response:
response_value = persist_values_to_response[response_key]
if response_key == 'tags':
input_statement.add_tags(*response_value)
response.add_tags(*response_value)
else:
setattr(input_statement, response_key, response_value)
setattr(response, response_key, response_value)
if not self.read_only:
self.learn_response(input_statement)
# Save the response generated for the input
self.storage.create(**response.serialize())
return response
|
python
|
def get_response(self, statement=None, **kwargs):
"""
Return the bot's response based on the input.
:param statement: An statement object or string.
:returns: A response to the input.
:rtype: Statement
:param additional_response_selection_parameters: Parameters to pass to the
chat bot's logic adapters to control response selection.
:type additional_response_selection_parameters: dict
:param persist_values_to_response: Values that should be saved to the response
that the chat bot generates.
:type persist_values_to_response: dict
"""
Statement = self.storage.get_object('statement')
additional_response_selection_parameters = kwargs.pop('additional_response_selection_parameters', {})
persist_values_to_response = kwargs.pop('persist_values_to_response', {})
if isinstance(statement, str):
kwargs['text'] = statement
if isinstance(statement, dict):
kwargs.update(statement)
if statement is None and 'text' not in kwargs:
raise self.ChatBotException(
'Either a statement object or a "text" keyword '
'argument is required. Neither was provided.'
)
if hasattr(statement, 'serialize'):
kwargs.update(**statement.serialize())
tags = kwargs.pop('tags', [])
text = kwargs.pop('text')
input_statement = Statement(text=text, **kwargs)
input_statement.add_tags(*tags)
# Preprocess the input statement
for preprocessor in self.preprocessors:
input_statement = preprocessor(input_statement)
# Make sure the input statement has its search text saved
if not input_statement.search_text:
input_statement.search_text = self.storage.tagger.get_bigram_pair_string(input_statement.text)
if not input_statement.search_in_response_to and input_statement.in_response_to:
input_statement.search_in_response_to = self.storage.tagger.get_bigram_pair_string(input_statement.in_response_to)
response = self.generate_response(input_statement, additional_response_selection_parameters)
# Update any response data that needs to be changed
if persist_values_to_response:
for response_key in persist_values_to_response:
response_value = persist_values_to_response[response_key]
if response_key == 'tags':
input_statement.add_tags(*response_value)
response.add_tags(*response_value)
else:
setattr(input_statement, response_key, response_value)
setattr(response, response_key, response_value)
if not self.read_only:
self.learn_response(input_statement)
# Save the response generated for the input
self.storage.create(**response.serialize())
return response
|
[
"def",
"get_response",
"(",
"self",
",",
"statement",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"Statement",
"=",
"self",
".",
"storage",
".",
"get_object",
"(",
"'statement'",
")",
"additional_response_selection_parameters",
"=",
"kwargs",
".",
"pop",
"(",
"'additional_response_selection_parameters'",
",",
"{",
"}",
")",
"persist_values_to_response",
"=",
"kwargs",
".",
"pop",
"(",
"'persist_values_to_response'",
",",
"{",
"}",
")",
"if",
"isinstance",
"(",
"statement",
",",
"str",
")",
":",
"kwargs",
"[",
"'text'",
"]",
"=",
"statement",
"if",
"isinstance",
"(",
"statement",
",",
"dict",
")",
":",
"kwargs",
".",
"update",
"(",
"statement",
")",
"if",
"statement",
"is",
"None",
"and",
"'text'",
"not",
"in",
"kwargs",
":",
"raise",
"self",
".",
"ChatBotException",
"(",
"'Either a statement object or a \"text\" keyword '",
"'argument is required. Neither was provided.'",
")",
"if",
"hasattr",
"(",
"statement",
",",
"'serialize'",
")",
":",
"kwargs",
".",
"update",
"(",
"*",
"*",
"statement",
".",
"serialize",
"(",
")",
")",
"tags",
"=",
"kwargs",
".",
"pop",
"(",
"'tags'",
",",
"[",
"]",
")",
"text",
"=",
"kwargs",
".",
"pop",
"(",
"'text'",
")",
"input_statement",
"=",
"Statement",
"(",
"text",
"=",
"text",
",",
"*",
"*",
"kwargs",
")",
"input_statement",
".",
"add_tags",
"(",
"*",
"tags",
")",
"# Preprocess the input statement",
"for",
"preprocessor",
"in",
"self",
".",
"preprocessors",
":",
"input_statement",
"=",
"preprocessor",
"(",
"input_statement",
")",
"# Make sure the input statement has its search text saved",
"if",
"not",
"input_statement",
".",
"search_text",
":",
"input_statement",
".",
"search_text",
"=",
"self",
".",
"storage",
".",
"tagger",
".",
"get_bigram_pair_string",
"(",
"input_statement",
".",
"text",
")",
"if",
"not",
"input_statement",
".",
"search_in_response_to",
"and",
"input_statement",
".",
"in_response_to",
":",
"input_statement",
".",
"search_in_response_to",
"=",
"self",
".",
"storage",
".",
"tagger",
".",
"get_bigram_pair_string",
"(",
"input_statement",
".",
"in_response_to",
")",
"response",
"=",
"self",
".",
"generate_response",
"(",
"input_statement",
",",
"additional_response_selection_parameters",
")",
"# Update any response data that needs to be changed",
"if",
"persist_values_to_response",
":",
"for",
"response_key",
"in",
"persist_values_to_response",
":",
"response_value",
"=",
"persist_values_to_response",
"[",
"response_key",
"]",
"if",
"response_key",
"==",
"'tags'",
":",
"input_statement",
".",
"add_tags",
"(",
"*",
"response_value",
")",
"response",
".",
"add_tags",
"(",
"*",
"response_value",
")",
"else",
":",
"setattr",
"(",
"input_statement",
",",
"response_key",
",",
"response_value",
")",
"setattr",
"(",
"response",
",",
"response_key",
",",
"response_value",
")",
"if",
"not",
"self",
".",
"read_only",
":",
"self",
".",
"learn_response",
"(",
"input_statement",
")",
"# Save the response generated for the input",
"self",
".",
"storage",
".",
"create",
"(",
"*",
"*",
"response",
".",
"serialize",
"(",
")",
")",
"return",
"response"
] |
Return the bot's response based on the input.
:param statement: An statement object or string.
:returns: A response to the input.
:rtype: Statement
:param additional_response_selection_parameters: Parameters to pass to the
chat bot's logic adapters to control response selection.
:type additional_response_selection_parameters: dict
:param persist_values_to_response: Values that should be saved to the response
that the chat bot generates.
:type persist_values_to_response: dict
|
[
"Return",
"the",
"bot",
"s",
"response",
"based",
"on",
"the",
"input",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/chatterbot.py#L57-L133
|
train
|
gunthercox/ChatterBot
|
chatterbot/chatterbot.py
|
ChatBot.generate_response
|
def generate_response(self, input_statement, additional_response_selection_parameters=None):
"""
Return a response based on a given input statement.
:param input_statement: The input statement to be processed.
"""
Statement = self.storage.get_object('statement')
results = []
result = None
max_confidence = -1
for adapter in self.logic_adapters:
if adapter.can_process(input_statement):
output = adapter.process(input_statement, additional_response_selection_parameters)
results.append(output)
self.logger.info(
'{} selected "{}" as a response with a confidence of {}'.format(
adapter.class_name, output.text, output.confidence
)
)
if output.confidence > max_confidence:
result = output
max_confidence = output.confidence
else:
self.logger.info(
'Not processing the statement using {}'.format(adapter.class_name)
)
class ResultOption:
def __init__(self, statement, count=1):
self.statement = statement
self.count = count
# If multiple adapters agree on the same statement,
# then that statement is more likely to be the correct response
if len(results) >= 3:
result_options = {}
for result_option in results:
result_string = result_option.text + ':' + (result_option.in_response_to or '')
if result_string in result_options:
result_options[result_string].count += 1
if result_options[result_string].statement.confidence < result_option.confidence:
result_options[result_string].statement = result_option
else:
result_options[result_string] = ResultOption(
result_option
)
most_common = list(result_options.values())[0]
for result_option in result_options.values():
if result_option.count > most_common.count:
most_common = result_option
if most_common.count > 1:
result = most_common.statement
response = Statement(
text=result.text,
in_response_to=input_statement.text,
conversation=input_statement.conversation,
persona='bot:' + self.name
)
response.confidence = result.confidence
return response
|
python
|
def generate_response(self, input_statement, additional_response_selection_parameters=None):
"""
Return a response based on a given input statement.
:param input_statement: The input statement to be processed.
"""
Statement = self.storage.get_object('statement')
results = []
result = None
max_confidence = -1
for adapter in self.logic_adapters:
if adapter.can_process(input_statement):
output = adapter.process(input_statement, additional_response_selection_parameters)
results.append(output)
self.logger.info(
'{} selected "{}" as a response with a confidence of {}'.format(
adapter.class_name, output.text, output.confidence
)
)
if output.confidence > max_confidence:
result = output
max_confidence = output.confidence
else:
self.logger.info(
'Not processing the statement using {}'.format(adapter.class_name)
)
class ResultOption:
def __init__(self, statement, count=1):
self.statement = statement
self.count = count
# If multiple adapters agree on the same statement,
# then that statement is more likely to be the correct response
if len(results) >= 3:
result_options = {}
for result_option in results:
result_string = result_option.text + ':' + (result_option.in_response_to or '')
if result_string in result_options:
result_options[result_string].count += 1
if result_options[result_string].statement.confidence < result_option.confidence:
result_options[result_string].statement = result_option
else:
result_options[result_string] = ResultOption(
result_option
)
most_common = list(result_options.values())[0]
for result_option in result_options.values():
if result_option.count > most_common.count:
most_common = result_option
if most_common.count > 1:
result = most_common.statement
response = Statement(
text=result.text,
in_response_to=input_statement.text,
conversation=input_statement.conversation,
persona='bot:' + self.name
)
response.confidence = result.confidence
return response
|
[
"def",
"generate_response",
"(",
"self",
",",
"input_statement",
",",
"additional_response_selection_parameters",
"=",
"None",
")",
":",
"Statement",
"=",
"self",
".",
"storage",
".",
"get_object",
"(",
"'statement'",
")",
"results",
"=",
"[",
"]",
"result",
"=",
"None",
"max_confidence",
"=",
"-",
"1",
"for",
"adapter",
"in",
"self",
".",
"logic_adapters",
":",
"if",
"adapter",
".",
"can_process",
"(",
"input_statement",
")",
":",
"output",
"=",
"adapter",
".",
"process",
"(",
"input_statement",
",",
"additional_response_selection_parameters",
")",
"results",
".",
"append",
"(",
"output",
")",
"self",
".",
"logger",
".",
"info",
"(",
"'{} selected \"{}\" as a response with a confidence of {}'",
".",
"format",
"(",
"adapter",
".",
"class_name",
",",
"output",
".",
"text",
",",
"output",
".",
"confidence",
")",
")",
"if",
"output",
".",
"confidence",
">",
"max_confidence",
":",
"result",
"=",
"output",
"max_confidence",
"=",
"output",
".",
"confidence",
"else",
":",
"self",
".",
"logger",
".",
"info",
"(",
"'Not processing the statement using {}'",
".",
"format",
"(",
"adapter",
".",
"class_name",
")",
")",
"class",
"ResultOption",
":",
"def",
"__init__",
"(",
"self",
",",
"statement",
",",
"count",
"=",
"1",
")",
":",
"self",
".",
"statement",
"=",
"statement",
"self",
".",
"count",
"=",
"count",
"# If multiple adapters agree on the same statement,",
"# then that statement is more likely to be the correct response",
"if",
"len",
"(",
"results",
")",
">=",
"3",
":",
"result_options",
"=",
"{",
"}",
"for",
"result_option",
"in",
"results",
":",
"result_string",
"=",
"result_option",
".",
"text",
"+",
"':'",
"+",
"(",
"result_option",
".",
"in_response_to",
"or",
"''",
")",
"if",
"result_string",
"in",
"result_options",
":",
"result_options",
"[",
"result_string",
"]",
".",
"count",
"+=",
"1",
"if",
"result_options",
"[",
"result_string",
"]",
".",
"statement",
".",
"confidence",
"<",
"result_option",
".",
"confidence",
":",
"result_options",
"[",
"result_string",
"]",
".",
"statement",
"=",
"result_option",
"else",
":",
"result_options",
"[",
"result_string",
"]",
"=",
"ResultOption",
"(",
"result_option",
")",
"most_common",
"=",
"list",
"(",
"result_options",
".",
"values",
"(",
")",
")",
"[",
"0",
"]",
"for",
"result_option",
"in",
"result_options",
".",
"values",
"(",
")",
":",
"if",
"result_option",
".",
"count",
">",
"most_common",
".",
"count",
":",
"most_common",
"=",
"result_option",
"if",
"most_common",
".",
"count",
">",
"1",
":",
"result",
"=",
"most_common",
".",
"statement",
"response",
"=",
"Statement",
"(",
"text",
"=",
"result",
".",
"text",
",",
"in_response_to",
"=",
"input_statement",
".",
"text",
",",
"conversation",
"=",
"input_statement",
".",
"conversation",
",",
"persona",
"=",
"'bot:'",
"+",
"self",
".",
"name",
")",
"response",
".",
"confidence",
"=",
"result",
".",
"confidence",
"return",
"response"
] |
Return a response based on a given input statement.
:param input_statement: The input statement to be processed.
|
[
"Return",
"a",
"response",
"based",
"on",
"a",
"given",
"input",
"statement",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/chatterbot.py#L135-L206
|
train
|
gunthercox/ChatterBot
|
chatterbot/chatterbot.py
|
ChatBot.learn_response
|
def learn_response(self, statement, previous_statement=None):
"""
Learn that the statement provided is a valid response.
"""
if not previous_statement:
previous_statement = statement.in_response_to
if not previous_statement:
previous_statement = self.get_latest_response(statement.conversation)
if previous_statement:
previous_statement = previous_statement.text
previous_statement_text = previous_statement
if not isinstance(previous_statement, (str, type(None), )):
statement.in_response_to = previous_statement.text
elif isinstance(previous_statement, str):
statement.in_response_to = previous_statement
self.logger.info('Adding "{}" as a response to "{}"'.format(
statement.text,
previous_statement_text
))
# Save the input statement
return self.storage.create(**statement.serialize())
|
python
|
def learn_response(self, statement, previous_statement=None):
"""
Learn that the statement provided is a valid response.
"""
if not previous_statement:
previous_statement = statement.in_response_to
if not previous_statement:
previous_statement = self.get_latest_response(statement.conversation)
if previous_statement:
previous_statement = previous_statement.text
previous_statement_text = previous_statement
if not isinstance(previous_statement, (str, type(None), )):
statement.in_response_to = previous_statement.text
elif isinstance(previous_statement, str):
statement.in_response_to = previous_statement
self.logger.info('Adding "{}" as a response to "{}"'.format(
statement.text,
previous_statement_text
))
# Save the input statement
return self.storage.create(**statement.serialize())
|
[
"def",
"learn_response",
"(",
"self",
",",
"statement",
",",
"previous_statement",
"=",
"None",
")",
":",
"if",
"not",
"previous_statement",
":",
"previous_statement",
"=",
"statement",
".",
"in_response_to",
"if",
"not",
"previous_statement",
":",
"previous_statement",
"=",
"self",
".",
"get_latest_response",
"(",
"statement",
".",
"conversation",
")",
"if",
"previous_statement",
":",
"previous_statement",
"=",
"previous_statement",
".",
"text",
"previous_statement_text",
"=",
"previous_statement",
"if",
"not",
"isinstance",
"(",
"previous_statement",
",",
"(",
"str",
",",
"type",
"(",
"None",
")",
",",
")",
")",
":",
"statement",
".",
"in_response_to",
"=",
"previous_statement",
".",
"text",
"elif",
"isinstance",
"(",
"previous_statement",
",",
"str",
")",
":",
"statement",
".",
"in_response_to",
"=",
"previous_statement",
"self",
".",
"logger",
".",
"info",
"(",
"'Adding \"{}\" as a response to \"{}\"'",
".",
"format",
"(",
"statement",
".",
"text",
",",
"previous_statement_text",
")",
")",
"# Save the input statement",
"return",
"self",
".",
"storage",
".",
"create",
"(",
"*",
"*",
"statement",
".",
"serialize",
"(",
")",
")"
] |
Learn that the statement provided is a valid response.
|
[
"Learn",
"that",
"the",
"statement",
"provided",
"is",
"a",
"valid",
"response",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/chatterbot.py#L208-L233
|
train
|
gunthercox/ChatterBot
|
chatterbot/conversation.py
|
StatementMixin.serialize
|
def serialize(self):
"""
:returns: A dictionary representation of the statement object.
:rtype: dict
"""
data = {}
for field_name in self.get_statement_field_names():
format_method = getattr(self, 'get_{}'.format(
field_name
), None)
if format_method:
data[field_name] = format_method()
else:
data[field_name] = getattr(self, field_name)
return data
|
python
|
def serialize(self):
"""
:returns: A dictionary representation of the statement object.
:rtype: dict
"""
data = {}
for field_name in self.get_statement_field_names():
format_method = getattr(self, 'get_{}'.format(
field_name
), None)
if format_method:
data[field_name] = format_method()
else:
data[field_name] = getattr(self, field_name)
return data
|
[
"def",
"serialize",
"(",
"self",
")",
":",
"data",
"=",
"{",
"}",
"for",
"field_name",
"in",
"self",
".",
"get_statement_field_names",
"(",
")",
":",
"format_method",
"=",
"getattr",
"(",
"self",
",",
"'get_{}'",
".",
"format",
"(",
"field_name",
")",
",",
"None",
")",
"if",
"format_method",
":",
"data",
"[",
"field_name",
"]",
"=",
"format_method",
"(",
")",
"else",
":",
"data",
"[",
"field_name",
"]",
"=",
"getattr",
"(",
"self",
",",
"field_name",
")",
"return",
"data"
] |
:returns: A dictionary representation of the statement object.
:rtype: dict
|
[
":",
"returns",
":",
"A",
"dictionary",
"representation",
"of",
"the",
"statement",
"object",
".",
":",
"rtype",
":",
"dict"
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/conversation.py#L44-L61
|
train
|
gunthercox/ChatterBot
|
chatterbot/utils.py
|
import_module
|
def import_module(dotted_path):
"""
Imports the specified module based on the
dot notated import path for the module.
"""
import importlib
module_parts = dotted_path.split('.')
module_path = '.'.join(module_parts[:-1])
module = importlib.import_module(module_path)
return getattr(module, module_parts[-1])
|
python
|
def import_module(dotted_path):
"""
Imports the specified module based on the
dot notated import path for the module.
"""
import importlib
module_parts = dotted_path.split('.')
module_path = '.'.join(module_parts[:-1])
module = importlib.import_module(module_path)
return getattr(module, module_parts[-1])
|
[
"def",
"import_module",
"(",
"dotted_path",
")",
":",
"import",
"importlib",
"module_parts",
"=",
"dotted_path",
".",
"split",
"(",
"'.'",
")",
"module_path",
"=",
"'.'",
".",
"join",
"(",
"module_parts",
"[",
":",
"-",
"1",
"]",
")",
"module",
"=",
"importlib",
".",
"import_module",
"(",
"module_path",
")",
"return",
"getattr",
"(",
"module",
",",
"module_parts",
"[",
"-",
"1",
"]",
")"
] |
Imports the specified module based on the
dot notated import path for the module.
|
[
"Imports",
"the",
"specified",
"module",
"based",
"on",
"the",
"dot",
"notated",
"import",
"path",
"for",
"the",
"module",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/utils.py#L6-L17
|
train
|
gunthercox/ChatterBot
|
chatterbot/utils.py
|
initialize_class
|
def initialize_class(data, *args, **kwargs):
"""
:param data: A string or dictionary containing a import_path attribute.
"""
if isinstance(data, dict):
import_path = data.get('import_path')
data.update(kwargs)
Class = import_module(import_path)
return Class(*args, **data)
else:
Class = import_module(data)
return Class(*args, **kwargs)
|
python
|
def initialize_class(data, *args, **kwargs):
"""
:param data: A string or dictionary containing a import_path attribute.
"""
if isinstance(data, dict):
import_path = data.get('import_path')
data.update(kwargs)
Class = import_module(import_path)
return Class(*args, **data)
else:
Class = import_module(data)
return Class(*args, **kwargs)
|
[
"def",
"initialize_class",
"(",
"data",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"isinstance",
"(",
"data",
",",
"dict",
")",
":",
"import_path",
"=",
"data",
".",
"get",
"(",
"'import_path'",
")",
"data",
".",
"update",
"(",
"kwargs",
")",
"Class",
"=",
"import_module",
"(",
"import_path",
")",
"return",
"Class",
"(",
"*",
"args",
",",
"*",
"*",
"data",
")",
"else",
":",
"Class",
"=",
"import_module",
"(",
"data",
")",
"return",
"Class",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] |
:param data: A string or dictionary containing a import_path attribute.
|
[
":",
"param",
"data",
":",
"A",
"string",
"or",
"dictionary",
"containing",
"a",
"import_path",
"attribute",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/utils.py#L20-L33
|
train
|
gunthercox/ChatterBot
|
chatterbot/utils.py
|
validate_adapter_class
|
def validate_adapter_class(validate_class, adapter_class):
"""
Raises an exception if validate_class is not a
subclass of adapter_class.
:param validate_class: The class to be validated.
:type validate_class: class
:param adapter_class: The class type to check against.
:type adapter_class: class
:raises: Adapter.InvalidAdapterTypeException
"""
from chatterbot.adapters import Adapter
# If a dictionary was passed in, check if it has an import_path attribute
if isinstance(validate_class, dict):
if 'import_path' not in validate_class:
raise Adapter.InvalidAdapterTypeException(
'The dictionary {} must contain a value for "import_path"'.format(
str(validate_class)
)
)
# Set the class to the import path for the next check
validate_class = validate_class.get('import_path')
if not issubclass(import_module(validate_class), adapter_class):
raise Adapter.InvalidAdapterTypeException(
'{} must be a subclass of {}'.format(
validate_class,
adapter_class.__name__
)
)
|
python
|
def validate_adapter_class(validate_class, adapter_class):
"""
Raises an exception if validate_class is not a
subclass of adapter_class.
:param validate_class: The class to be validated.
:type validate_class: class
:param adapter_class: The class type to check against.
:type adapter_class: class
:raises: Adapter.InvalidAdapterTypeException
"""
from chatterbot.adapters import Adapter
# If a dictionary was passed in, check if it has an import_path attribute
if isinstance(validate_class, dict):
if 'import_path' not in validate_class:
raise Adapter.InvalidAdapterTypeException(
'The dictionary {} must contain a value for "import_path"'.format(
str(validate_class)
)
)
# Set the class to the import path for the next check
validate_class = validate_class.get('import_path')
if not issubclass(import_module(validate_class), adapter_class):
raise Adapter.InvalidAdapterTypeException(
'{} must be a subclass of {}'.format(
validate_class,
adapter_class.__name__
)
)
|
[
"def",
"validate_adapter_class",
"(",
"validate_class",
",",
"adapter_class",
")",
":",
"from",
"chatterbot",
".",
"adapters",
"import",
"Adapter",
"# If a dictionary was passed in, check if it has an import_path attribute",
"if",
"isinstance",
"(",
"validate_class",
",",
"dict",
")",
":",
"if",
"'import_path'",
"not",
"in",
"validate_class",
":",
"raise",
"Adapter",
".",
"InvalidAdapterTypeException",
"(",
"'The dictionary {} must contain a value for \"import_path\"'",
".",
"format",
"(",
"str",
"(",
"validate_class",
")",
")",
")",
"# Set the class to the import path for the next check",
"validate_class",
"=",
"validate_class",
".",
"get",
"(",
"'import_path'",
")",
"if",
"not",
"issubclass",
"(",
"import_module",
"(",
"validate_class",
")",
",",
"adapter_class",
")",
":",
"raise",
"Adapter",
".",
"InvalidAdapterTypeException",
"(",
"'{} must be a subclass of {}'",
".",
"format",
"(",
"validate_class",
",",
"adapter_class",
".",
"__name__",
")",
")"
] |
Raises an exception if validate_class is not a
subclass of adapter_class.
:param validate_class: The class to be validated.
:type validate_class: class
:param adapter_class: The class type to check against.
:type adapter_class: class
:raises: Adapter.InvalidAdapterTypeException
|
[
"Raises",
"an",
"exception",
"if",
"validate_class",
"is",
"not",
"a",
"subclass",
"of",
"adapter_class",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/utils.py#L36-L70
|
train
|
gunthercox/ChatterBot
|
chatterbot/utils.py
|
get_response_time
|
def get_response_time(chatbot, statement='Hello'):
"""
Returns the amount of time taken for a given
chat bot to return a response.
:param chatbot: A chat bot instance.
:type chatbot: ChatBot
:returns: The response time in seconds.
:rtype: float
"""
import time
start_time = time.time()
chatbot.get_response(statement)
return time.time() - start_time
|
python
|
def get_response_time(chatbot, statement='Hello'):
"""
Returns the amount of time taken for a given
chat bot to return a response.
:param chatbot: A chat bot instance.
:type chatbot: ChatBot
:returns: The response time in seconds.
:rtype: float
"""
import time
start_time = time.time()
chatbot.get_response(statement)
return time.time() - start_time
|
[
"def",
"get_response_time",
"(",
"chatbot",
",",
"statement",
"=",
"'Hello'",
")",
":",
"import",
"time",
"start_time",
"=",
"time",
".",
"time",
"(",
")",
"chatbot",
".",
"get_response",
"(",
"statement",
")",
"return",
"time",
".",
"time",
"(",
")",
"-",
"start_time"
] |
Returns the amount of time taken for a given
chat bot to return a response.
:param chatbot: A chat bot instance.
:type chatbot: ChatBot
:returns: The response time in seconds.
:rtype: float
|
[
"Returns",
"the",
"amount",
"of",
"time",
"taken",
"for",
"a",
"given",
"chat",
"bot",
"to",
"return",
"a",
"response",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/utils.py#L73-L90
|
train
|
gunthercox/ChatterBot
|
chatterbot/utils.py
|
print_progress_bar
|
def print_progress_bar(description, iteration_counter, total_items, progress_bar_length=20):
"""
Print progress bar
:param description: Training description
:type description: str
:param iteration_counter: Incremental counter
:type iteration_counter: int
:param total_items: total number items
:type total_items: int
:param progress_bar_length: Progress bar length
:type progress_bar_length: int
:returns: void
:rtype: void
"""
import sys
percent = float(iteration_counter) / total_items
hashes = '#' * int(round(percent * progress_bar_length))
spaces = ' ' * (progress_bar_length - len(hashes))
sys.stdout.write('\r{0}: [{1}] {2}%'.format(description, hashes + spaces, int(round(percent * 100))))
sys.stdout.flush()
if total_items == iteration_counter:
print('\r')
|
python
|
def print_progress_bar(description, iteration_counter, total_items, progress_bar_length=20):
"""
Print progress bar
:param description: Training description
:type description: str
:param iteration_counter: Incremental counter
:type iteration_counter: int
:param total_items: total number items
:type total_items: int
:param progress_bar_length: Progress bar length
:type progress_bar_length: int
:returns: void
:rtype: void
"""
import sys
percent = float(iteration_counter) / total_items
hashes = '#' * int(round(percent * progress_bar_length))
spaces = ' ' * (progress_bar_length - len(hashes))
sys.stdout.write('\r{0}: [{1}] {2}%'.format(description, hashes + spaces, int(round(percent * 100))))
sys.stdout.flush()
if total_items == iteration_counter:
print('\r')
|
[
"def",
"print_progress_bar",
"(",
"description",
",",
"iteration_counter",
",",
"total_items",
",",
"progress_bar_length",
"=",
"20",
")",
":",
"import",
"sys",
"percent",
"=",
"float",
"(",
"iteration_counter",
")",
"/",
"total_items",
"hashes",
"=",
"'#'",
"*",
"int",
"(",
"round",
"(",
"percent",
"*",
"progress_bar_length",
")",
")",
"spaces",
"=",
"' '",
"*",
"(",
"progress_bar_length",
"-",
"len",
"(",
"hashes",
")",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"'\\r{0}: [{1}] {2}%'",
".",
"format",
"(",
"description",
",",
"hashes",
"+",
"spaces",
",",
"int",
"(",
"round",
"(",
"percent",
"*",
"100",
")",
")",
")",
")",
"sys",
".",
"stdout",
".",
"flush",
"(",
")",
"if",
"total_items",
"==",
"iteration_counter",
":",
"print",
"(",
"'\\r'",
")"
] |
Print progress bar
:param description: Training description
:type description: str
:param iteration_counter: Incremental counter
:type iteration_counter: int
:param total_items: total number items
:type total_items: int
:param progress_bar_length: Progress bar length
:type progress_bar_length: int
:returns: void
:rtype: void
|
[
"Print",
"progress",
"bar",
":",
"param",
"description",
":",
"Training",
"description",
":",
"type",
"description",
":",
"str"
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/utils.py#L93-L119
|
train
|
gunthercox/ChatterBot
|
chatterbot/logic/unit_conversion.py
|
UnitConversion.get_unit
|
def get_unit(self, ureg, unit_variations):
"""
Get the first match unit metric object supported by pint library
given a variation of unit metric names (Ex:['HOUR', 'hour']).
:param ureg: unit registry which units are defined and handled
:type ureg: pint.registry.UnitRegistry object
:param unit_variations: A list of strings with names of units
:type unit_variations: str
"""
for unit in unit_variations:
try:
return getattr(ureg, unit)
except Exception:
continue
return None
|
python
|
def get_unit(self, ureg, unit_variations):
"""
Get the first match unit metric object supported by pint library
given a variation of unit metric names (Ex:['HOUR', 'hour']).
:param ureg: unit registry which units are defined and handled
:type ureg: pint.registry.UnitRegistry object
:param unit_variations: A list of strings with names of units
:type unit_variations: str
"""
for unit in unit_variations:
try:
return getattr(ureg, unit)
except Exception:
continue
return None
|
[
"def",
"get_unit",
"(",
"self",
",",
"ureg",
",",
"unit_variations",
")",
":",
"for",
"unit",
"in",
"unit_variations",
":",
"try",
":",
"return",
"getattr",
"(",
"ureg",
",",
"unit",
")",
"except",
"Exception",
":",
"continue",
"return",
"None"
] |
Get the first match unit metric object supported by pint library
given a variation of unit metric names (Ex:['HOUR', 'hour']).
:param ureg: unit registry which units are defined and handled
:type ureg: pint.registry.UnitRegistry object
:param unit_variations: A list of strings with names of units
:type unit_variations: str
|
[
"Get",
"the",
"first",
"match",
"unit",
"metric",
"object",
"supported",
"by",
"pint",
"library",
"given",
"a",
"variation",
"of",
"unit",
"metric",
"names",
"(",
"Ex",
":",
"[",
"HOUR",
"hour",
"]",
")",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/logic/unit_conversion.py#L67-L83
|
train
|
gunthercox/ChatterBot
|
chatterbot/logic/unit_conversion.py
|
UnitConversion.get_valid_units
|
def get_valid_units(self, ureg, from_unit, target_unit):
"""
Returns the firt match `pint.unit.Unit` object for from_unit and
target_unit strings from a possible variation of metric unit names
supported by pint library.
:param ureg: unit registry which units are defined and handled
:type ureg: `pint.registry.UnitRegistry`
:param from_unit: source metric unit
:type from_unit: str
:param from_unit: target metric unit
:type from_unit: str
"""
from_unit_variations = [from_unit.lower(), from_unit.upper()]
target_unit_variations = [target_unit.lower(), target_unit.upper()]
from_unit = self.get_unit(ureg, from_unit_variations)
target_unit = self.get_unit(ureg, target_unit_variations)
return from_unit, target_unit
|
python
|
def get_valid_units(self, ureg, from_unit, target_unit):
"""
Returns the firt match `pint.unit.Unit` object for from_unit and
target_unit strings from a possible variation of metric unit names
supported by pint library.
:param ureg: unit registry which units are defined and handled
:type ureg: `pint.registry.UnitRegistry`
:param from_unit: source metric unit
:type from_unit: str
:param from_unit: target metric unit
:type from_unit: str
"""
from_unit_variations = [from_unit.lower(), from_unit.upper()]
target_unit_variations = [target_unit.lower(), target_unit.upper()]
from_unit = self.get_unit(ureg, from_unit_variations)
target_unit = self.get_unit(ureg, target_unit_variations)
return from_unit, target_unit
|
[
"def",
"get_valid_units",
"(",
"self",
",",
"ureg",
",",
"from_unit",
",",
"target_unit",
")",
":",
"from_unit_variations",
"=",
"[",
"from_unit",
".",
"lower",
"(",
")",
",",
"from_unit",
".",
"upper",
"(",
")",
"]",
"target_unit_variations",
"=",
"[",
"target_unit",
".",
"lower",
"(",
")",
",",
"target_unit",
".",
"upper",
"(",
")",
"]",
"from_unit",
"=",
"self",
".",
"get_unit",
"(",
"ureg",
",",
"from_unit_variations",
")",
"target_unit",
"=",
"self",
".",
"get_unit",
"(",
"ureg",
",",
"target_unit_variations",
")",
"return",
"from_unit",
",",
"target_unit"
] |
Returns the firt match `pint.unit.Unit` object for from_unit and
target_unit strings from a possible variation of metric unit names
supported by pint library.
:param ureg: unit registry which units are defined and handled
:type ureg: `pint.registry.UnitRegistry`
:param from_unit: source metric unit
:type from_unit: str
:param from_unit: target metric unit
:type from_unit: str
|
[
"Returns",
"the",
"firt",
"match",
"pint",
".",
"unit",
".",
"Unit",
"object",
"for",
"from_unit",
"and",
"target_unit",
"strings",
"from",
"a",
"possible",
"variation",
"of",
"metric",
"unit",
"names",
"supported",
"by",
"pint",
"library",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/logic/unit_conversion.py#L85-L104
|
train
|
gunthercox/ChatterBot
|
chatterbot/logic/unit_conversion.py
|
UnitConversion.handle_matches
|
def handle_matches(self, match):
"""
Returns a response statement from a matched input statement.
:param match: It is a valid matched pattern from the input statement
:type: `_sre.SRE_Match`
"""
response = Statement(text='')
from_parsed = match.group("from")
target_parsed = match.group("target")
n_statement = match.group("number")
if n_statement == 'a' or n_statement == 'an':
n_statement = '1.0'
n = mathparse.parse(n_statement, self.language.ISO_639.upper())
ureg = UnitRegistry()
from_parsed, target_parsed = self.get_valid_units(ureg, from_parsed, target_parsed)
if from_parsed is None or target_parsed is None:
response.confidence = 0.0
else:
from_value = ureg.Quantity(float(n), from_parsed)
target_value = from_value.to(target_parsed)
response.confidence = 1.0
response.text = str(target_value.magnitude)
return response
|
python
|
def handle_matches(self, match):
"""
Returns a response statement from a matched input statement.
:param match: It is a valid matched pattern from the input statement
:type: `_sre.SRE_Match`
"""
response = Statement(text='')
from_parsed = match.group("from")
target_parsed = match.group("target")
n_statement = match.group("number")
if n_statement == 'a' or n_statement == 'an':
n_statement = '1.0'
n = mathparse.parse(n_statement, self.language.ISO_639.upper())
ureg = UnitRegistry()
from_parsed, target_parsed = self.get_valid_units(ureg, from_parsed, target_parsed)
if from_parsed is None or target_parsed is None:
response.confidence = 0.0
else:
from_value = ureg.Quantity(float(n), from_parsed)
target_value = from_value.to(target_parsed)
response.confidence = 1.0
response.text = str(target_value.magnitude)
return response
|
[
"def",
"handle_matches",
"(",
"self",
",",
"match",
")",
":",
"response",
"=",
"Statement",
"(",
"text",
"=",
"''",
")",
"from_parsed",
"=",
"match",
".",
"group",
"(",
"\"from\"",
")",
"target_parsed",
"=",
"match",
".",
"group",
"(",
"\"target\"",
")",
"n_statement",
"=",
"match",
".",
"group",
"(",
"\"number\"",
")",
"if",
"n_statement",
"==",
"'a'",
"or",
"n_statement",
"==",
"'an'",
":",
"n_statement",
"=",
"'1.0'",
"n",
"=",
"mathparse",
".",
"parse",
"(",
"n_statement",
",",
"self",
".",
"language",
".",
"ISO_639",
".",
"upper",
"(",
")",
")",
"ureg",
"=",
"UnitRegistry",
"(",
")",
"from_parsed",
",",
"target_parsed",
"=",
"self",
".",
"get_valid_units",
"(",
"ureg",
",",
"from_parsed",
",",
"target_parsed",
")",
"if",
"from_parsed",
"is",
"None",
"or",
"target_parsed",
"is",
"None",
":",
"response",
".",
"confidence",
"=",
"0.0",
"else",
":",
"from_value",
"=",
"ureg",
".",
"Quantity",
"(",
"float",
"(",
"n",
")",
",",
"from_parsed",
")",
"target_value",
"=",
"from_value",
".",
"to",
"(",
"target_parsed",
")",
"response",
".",
"confidence",
"=",
"1.0",
"response",
".",
"text",
"=",
"str",
"(",
"target_value",
".",
"magnitude",
")",
"return",
"response"
] |
Returns a response statement from a matched input statement.
:param match: It is a valid matched pattern from the input statement
:type: `_sre.SRE_Match`
|
[
"Returns",
"a",
"response",
"statement",
"from",
"a",
"matched",
"input",
"statement",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/logic/unit_conversion.py#L106-L135
|
train
|
gunthercox/ChatterBot
|
chatterbot/logic/logic_adapter.py
|
LogicAdapter.get_default_response
|
def get_default_response(self, input_statement):
"""
This method is called when a logic adapter is unable to generate any
other meaningful response.
"""
from random import choice
if self.default_responses:
response = choice(self.default_responses)
else:
try:
response = self.chatbot.storage.get_random()
except StorageAdapter.EmptyDatabaseException:
response = input_statement
self.chatbot.logger.info(
'No known response to the input was found. Selecting a random response.'
)
# Set confidence to zero because a random response is selected
response.confidence = 0
return response
|
python
|
def get_default_response(self, input_statement):
"""
This method is called when a logic adapter is unable to generate any
other meaningful response.
"""
from random import choice
if self.default_responses:
response = choice(self.default_responses)
else:
try:
response = self.chatbot.storage.get_random()
except StorageAdapter.EmptyDatabaseException:
response = input_statement
self.chatbot.logger.info(
'No known response to the input was found. Selecting a random response.'
)
# Set confidence to zero because a random response is selected
response.confidence = 0
return response
|
[
"def",
"get_default_response",
"(",
"self",
",",
"input_statement",
")",
":",
"from",
"random",
"import",
"choice",
"if",
"self",
".",
"default_responses",
":",
"response",
"=",
"choice",
"(",
"self",
".",
"default_responses",
")",
"else",
":",
"try",
":",
"response",
"=",
"self",
".",
"chatbot",
".",
"storage",
".",
"get_random",
"(",
")",
"except",
"StorageAdapter",
".",
"EmptyDatabaseException",
":",
"response",
"=",
"input_statement",
"self",
".",
"chatbot",
".",
"logger",
".",
"info",
"(",
"'No known response to the input was found. Selecting a random response.'",
")",
"# Set confidence to zero because a random response is selected",
"response",
".",
"confidence",
"=",
"0",
"return",
"response"
] |
This method is called when a logic adapter is unable to generate any
other meaningful response.
|
[
"This",
"method",
"is",
"called",
"when",
"a",
"logic",
"adapter",
"is",
"unable",
"to",
"generate",
"any",
"other",
"meaningful",
"response",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/logic/logic_adapter.py#L103-L125
|
train
|
gunthercox/ChatterBot
|
chatterbot/logic/time_adapter.py
|
TimeLogicAdapter.time_question_features
|
def time_question_features(self, text):
"""
Provide an analysis of significant features in the string.
"""
features = {}
# A list of all words from the known sentences
all_words = " ".join(self.positive + self.negative).split()
# A list of the first word in each of the known sentence
all_first_words = []
for sentence in self.positive + self.negative:
all_first_words.append(
sentence.split(' ', 1)[0]
)
for word in text.split():
features['first_word({})'.format(word)] = (word in all_first_words)
for word in text.split():
features['contains({})'.format(word)] = (word in all_words)
for letter in 'abcdefghijklmnopqrstuvwxyz':
features['count({})'.format(letter)] = text.lower().count(letter)
features['has({})'.format(letter)] = (letter in text.lower())
return features
|
python
|
def time_question_features(self, text):
"""
Provide an analysis of significant features in the string.
"""
features = {}
# A list of all words from the known sentences
all_words = " ".join(self.positive + self.negative).split()
# A list of the first word in each of the known sentence
all_first_words = []
for sentence in self.positive + self.negative:
all_first_words.append(
sentence.split(' ', 1)[0]
)
for word in text.split():
features['first_word({})'.format(word)] = (word in all_first_words)
for word in text.split():
features['contains({})'.format(word)] = (word in all_words)
for letter in 'abcdefghijklmnopqrstuvwxyz':
features['count({})'.format(letter)] = text.lower().count(letter)
features['has({})'.format(letter)] = (letter in text.lower())
return features
|
[
"def",
"time_question_features",
"(",
"self",
",",
"text",
")",
":",
"features",
"=",
"{",
"}",
"# A list of all words from the known sentences",
"all_words",
"=",
"\" \"",
".",
"join",
"(",
"self",
".",
"positive",
"+",
"self",
".",
"negative",
")",
".",
"split",
"(",
")",
"# A list of the first word in each of the known sentence",
"all_first_words",
"=",
"[",
"]",
"for",
"sentence",
"in",
"self",
".",
"positive",
"+",
"self",
".",
"negative",
":",
"all_first_words",
".",
"append",
"(",
"sentence",
".",
"split",
"(",
"' '",
",",
"1",
")",
"[",
"0",
"]",
")",
"for",
"word",
"in",
"text",
".",
"split",
"(",
")",
":",
"features",
"[",
"'first_word({})'",
".",
"format",
"(",
"word",
")",
"]",
"=",
"(",
"word",
"in",
"all_first_words",
")",
"for",
"word",
"in",
"text",
".",
"split",
"(",
")",
":",
"features",
"[",
"'contains({})'",
".",
"format",
"(",
"word",
")",
"]",
"=",
"(",
"word",
"in",
"all_words",
")",
"for",
"letter",
"in",
"'abcdefghijklmnopqrstuvwxyz'",
":",
"features",
"[",
"'count({})'",
".",
"format",
"(",
"letter",
")",
"]",
"=",
"text",
".",
"lower",
"(",
")",
".",
"count",
"(",
"letter",
")",
"features",
"[",
"'has({})'",
".",
"format",
"(",
"letter",
")",
"]",
"=",
"(",
"letter",
"in",
"text",
".",
"lower",
"(",
")",
")",
"return",
"features"
] |
Provide an analysis of significant features in the string.
|
[
"Provide",
"an",
"analysis",
"of",
"significant",
"features",
"in",
"the",
"string",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/logic/time_adapter.py#L56-L82
|
train
|
gunthercox/ChatterBot
|
chatterbot/logic/mathematical_evaluation.py
|
MathematicalEvaluation.can_process
|
def can_process(self, statement):
"""
Determines whether it is appropriate for this
adapter to respond to the user input.
"""
response = self.process(statement)
self.cache[statement.text] = response
return response.confidence == 1
|
python
|
def can_process(self, statement):
"""
Determines whether it is appropriate for this
adapter to respond to the user input.
"""
response = self.process(statement)
self.cache[statement.text] = response
return response.confidence == 1
|
[
"def",
"can_process",
"(",
"self",
",",
"statement",
")",
":",
"response",
"=",
"self",
".",
"process",
"(",
"statement",
")",
"self",
".",
"cache",
"[",
"statement",
".",
"text",
"]",
"=",
"response",
"return",
"response",
".",
"confidence",
"==",
"1"
] |
Determines whether it is appropriate for this
adapter to respond to the user input.
|
[
"Determines",
"whether",
"it",
"is",
"appropriate",
"for",
"this",
"adapter",
"to",
"respond",
"to",
"the",
"user",
"input",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/logic/mathematical_evaluation.py#L28-L35
|
train
|
gunthercox/ChatterBot
|
chatterbot/logic/mathematical_evaluation.py
|
MathematicalEvaluation.process
|
def process(self, statement, additional_response_selection_parameters=None):
"""
Takes a statement string.
Returns the equation from the statement with the mathematical terms solved.
"""
from mathparse import mathparse
input_text = statement.text
# Use the result cached by the process method if it exists
if input_text in self.cache:
cached_result = self.cache[input_text]
self.cache = {}
return cached_result
# Getting the mathematical terms within the input statement
expression = mathparse.extract_expression(input_text, language=self.language.ISO_639.upper())
response = Statement(text=expression)
try:
response.text += ' = ' + str(
mathparse.parse(expression, language=self.language.ISO_639.upper())
)
# The confidence is 1 if the expression could be evaluated
response.confidence = 1
except mathparse.PostfixTokenEvaluationException:
response.confidence = 0
return response
|
python
|
def process(self, statement, additional_response_selection_parameters=None):
"""
Takes a statement string.
Returns the equation from the statement with the mathematical terms solved.
"""
from mathparse import mathparse
input_text = statement.text
# Use the result cached by the process method if it exists
if input_text in self.cache:
cached_result = self.cache[input_text]
self.cache = {}
return cached_result
# Getting the mathematical terms within the input statement
expression = mathparse.extract_expression(input_text, language=self.language.ISO_639.upper())
response = Statement(text=expression)
try:
response.text += ' = ' + str(
mathparse.parse(expression, language=self.language.ISO_639.upper())
)
# The confidence is 1 if the expression could be evaluated
response.confidence = 1
except mathparse.PostfixTokenEvaluationException:
response.confidence = 0
return response
|
[
"def",
"process",
"(",
"self",
",",
"statement",
",",
"additional_response_selection_parameters",
"=",
"None",
")",
":",
"from",
"mathparse",
"import",
"mathparse",
"input_text",
"=",
"statement",
".",
"text",
"# Use the result cached by the process method if it exists",
"if",
"input_text",
"in",
"self",
".",
"cache",
":",
"cached_result",
"=",
"self",
".",
"cache",
"[",
"input_text",
"]",
"self",
".",
"cache",
"=",
"{",
"}",
"return",
"cached_result",
"# Getting the mathematical terms within the input statement",
"expression",
"=",
"mathparse",
".",
"extract_expression",
"(",
"input_text",
",",
"language",
"=",
"self",
".",
"language",
".",
"ISO_639",
".",
"upper",
"(",
")",
")",
"response",
"=",
"Statement",
"(",
"text",
"=",
"expression",
")",
"try",
":",
"response",
".",
"text",
"+=",
"' = '",
"+",
"str",
"(",
"mathparse",
".",
"parse",
"(",
"expression",
",",
"language",
"=",
"self",
".",
"language",
".",
"ISO_639",
".",
"upper",
"(",
")",
")",
")",
"# The confidence is 1 if the expression could be evaluated",
"response",
".",
"confidence",
"=",
"1",
"except",
"mathparse",
".",
"PostfixTokenEvaluationException",
":",
"response",
".",
"confidence",
"=",
"0",
"return",
"response"
] |
Takes a statement string.
Returns the equation from the statement with the mathematical terms solved.
|
[
"Takes",
"a",
"statement",
"string",
".",
"Returns",
"the",
"equation",
"from",
"the",
"statement",
"with",
"the",
"mathematical",
"terms",
"solved",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/logic/mathematical_evaluation.py#L37-L67
|
train
|
gunthercox/ChatterBot
|
chatterbot/filters.py
|
get_recent_repeated_responses
|
def get_recent_repeated_responses(chatbot, conversation, sample=10, threshold=3, quantity=3):
"""
A filter that eliminates possibly repetitive responses to prevent
a chat bot from repeating statements that it has recently said.
"""
from collections import Counter
# Get the most recent statements from the conversation
conversation_statements = list(chatbot.storage.filter(
conversation=conversation,
order_by=['id']
))[sample * -1:]
text_of_recent_responses = [
statement.text for statement in conversation_statements
]
counter = Counter(text_of_recent_responses)
# Find the n most common responses from the conversation
most_common = counter.most_common(quantity)
return [
counted[0] for counted in most_common
if counted[1] >= threshold
]
|
python
|
def get_recent_repeated_responses(chatbot, conversation, sample=10, threshold=3, quantity=3):
"""
A filter that eliminates possibly repetitive responses to prevent
a chat bot from repeating statements that it has recently said.
"""
from collections import Counter
# Get the most recent statements from the conversation
conversation_statements = list(chatbot.storage.filter(
conversation=conversation,
order_by=['id']
))[sample * -1:]
text_of_recent_responses = [
statement.text for statement in conversation_statements
]
counter = Counter(text_of_recent_responses)
# Find the n most common responses from the conversation
most_common = counter.most_common(quantity)
return [
counted[0] for counted in most_common
if counted[1] >= threshold
]
|
[
"def",
"get_recent_repeated_responses",
"(",
"chatbot",
",",
"conversation",
",",
"sample",
"=",
"10",
",",
"threshold",
"=",
"3",
",",
"quantity",
"=",
"3",
")",
":",
"from",
"collections",
"import",
"Counter",
"# Get the most recent statements from the conversation",
"conversation_statements",
"=",
"list",
"(",
"chatbot",
".",
"storage",
".",
"filter",
"(",
"conversation",
"=",
"conversation",
",",
"order_by",
"=",
"[",
"'id'",
"]",
")",
")",
"[",
"sample",
"*",
"-",
"1",
":",
"]",
"text_of_recent_responses",
"=",
"[",
"statement",
".",
"text",
"for",
"statement",
"in",
"conversation_statements",
"]",
"counter",
"=",
"Counter",
"(",
"text_of_recent_responses",
")",
"# Find the n most common responses from the conversation",
"most_common",
"=",
"counter",
".",
"most_common",
"(",
"quantity",
")",
"return",
"[",
"counted",
"[",
"0",
"]",
"for",
"counted",
"in",
"most_common",
"if",
"counted",
"[",
"1",
"]",
">=",
"threshold",
"]"
] |
A filter that eliminates possibly repetitive responses to prevent
a chat bot from repeating statements that it has recently said.
|
[
"A",
"filter",
"that",
"eliminates",
"possibly",
"repetitive",
"responses",
"to",
"prevent",
"a",
"chat",
"bot",
"from",
"repeating",
"statements",
"that",
"it",
"has",
"recently",
"said",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/filters.py#L1-L26
|
train
|
gunthercox/ChatterBot
|
chatterbot/response_selection.py
|
get_most_frequent_response
|
def get_most_frequent_response(input_statement, response_list, storage=None):
"""
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: The response statement with the greatest number of occurrences.
:rtype: Statement
"""
matching_response = None
occurrence_count = -1
logger = logging.getLogger(__name__)
logger.info('Selecting response with greatest number of occurrences.')
for statement in response_list:
count = len(list(storage.filter(
text=statement.text,
in_response_to=input_statement.text)
))
# Keep the more common statement
if count >= occurrence_count:
matching_response = statement
occurrence_count = count
# Choose the most commonly occuring matching response
return matching_response
|
python
|
def get_most_frequent_response(input_statement, response_list, storage=None):
"""
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: The response statement with the greatest number of occurrences.
:rtype: Statement
"""
matching_response = None
occurrence_count = -1
logger = logging.getLogger(__name__)
logger.info('Selecting response with greatest number of occurrences.')
for statement in response_list:
count = len(list(storage.filter(
text=statement.text,
in_response_to=input_statement.text)
))
# Keep the more common statement
if count >= occurrence_count:
matching_response = statement
occurrence_count = count
# Choose the most commonly occuring matching response
return matching_response
|
[
"def",
"get_most_frequent_response",
"(",
"input_statement",
",",
"response_list",
",",
"storage",
"=",
"None",
")",
":",
"matching_response",
"=",
"None",
"occurrence_count",
"=",
"-",
"1",
"logger",
"=",
"logging",
".",
"getLogger",
"(",
"__name__",
")",
"logger",
".",
"info",
"(",
"'Selecting response with greatest number of occurrences.'",
")",
"for",
"statement",
"in",
"response_list",
":",
"count",
"=",
"len",
"(",
"list",
"(",
"storage",
".",
"filter",
"(",
"text",
"=",
"statement",
".",
"text",
",",
"in_response_to",
"=",
"input_statement",
".",
"text",
")",
")",
")",
"# Keep the more common statement",
"if",
"count",
">=",
"occurrence_count",
":",
"matching_response",
"=",
"statement",
"occurrence_count",
"=",
"count",
"# Choose the most commonly occuring matching response",
"return",
"matching_response"
] |
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: The response statement with the greatest number of occurrences.
:rtype: Statement
|
[
":",
"param",
"input_statement",
":",
"A",
"statement",
"that",
"closely",
"matches",
"an",
"input",
"to",
"the",
"chat",
"bot",
".",
":",
"type",
"input_statement",
":",
"Statement"
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/response_selection.py#L8-L41
|
train
|
gunthercox/ChatterBot
|
chatterbot/response_selection.py
|
get_first_response
|
def get_first_response(input_statement, response_list, storage=None):
"""
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: Return the first statement in the response list.
:rtype: Statement
"""
logger = logging.getLogger(__name__)
logger.info('Selecting first response from list of {} options.'.format(
len(response_list)
))
return response_list[0]
|
python
|
def get_first_response(input_statement, response_list, storage=None):
"""
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: Return the first statement in the response list.
:rtype: Statement
"""
logger = logging.getLogger(__name__)
logger.info('Selecting first response from list of {} options.'.format(
len(response_list)
))
return response_list[0]
|
[
"def",
"get_first_response",
"(",
"input_statement",
",",
"response_list",
",",
"storage",
"=",
"None",
")",
":",
"logger",
"=",
"logging",
".",
"getLogger",
"(",
"__name__",
")",
"logger",
".",
"info",
"(",
"'Selecting first response from list of {} options.'",
".",
"format",
"(",
"len",
"(",
"response_list",
")",
")",
")",
"return",
"response_list",
"[",
"0",
"]"
] |
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: Return the first statement in the response list.
:rtype: Statement
|
[
":",
"param",
"input_statement",
":",
"A",
"statement",
"that",
"closely",
"matches",
"an",
"input",
"to",
"the",
"chat",
"bot",
".",
":",
"type",
"input_statement",
":",
"Statement"
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/response_selection.py#L44-L63
|
train
|
gunthercox/ChatterBot
|
chatterbot/response_selection.py
|
get_random_response
|
def get_random_response(input_statement, response_list, storage=None):
"""
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: Choose a random response from the selection.
:rtype: Statement
"""
from random import choice
logger = logging.getLogger(__name__)
logger.info('Selecting a response from list of {} options.'.format(
len(response_list)
))
return choice(response_list)
|
python
|
def get_random_response(input_statement, response_list, storage=None):
"""
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: Choose a random response from the selection.
:rtype: Statement
"""
from random import choice
logger = logging.getLogger(__name__)
logger.info('Selecting a response from list of {} options.'.format(
len(response_list)
))
return choice(response_list)
|
[
"def",
"get_random_response",
"(",
"input_statement",
",",
"response_list",
",",
"storage",
"=",
"None",
")",
":",
"from",
"random",
"import",
"choice",
"logger",
"=",
"logging",
".",
"getLogger",
"(",
"__name__",
")",
"logger",
".",
"info",
"(",
"'Selecting a response from list of {} options.'",
".",
"format",
"(",
"len",
"(",
"response_list",
")",
")",
")",
"return",
"choice",
"(",
"response_list",
")"
] |
:param input_statement: A statement, that closely matches an input to the chat bot.
:type input_statement: Statement
:param response_list: A list of statement options to choose a response from.
:type response_list: list
:param storage: An instance of a storage adapter to allow the response selection
method to access other statements if needed.
:type storage: StorageAdapter
:return: Choose a random response from the selection.
:rtype: Statement
|
[
":",
"param",
"input_statement",
":",
"A",
"statement",
"that",
"closely",
"matches",
"an",
"input",
"to",
"the",
"chat",
"bot",
".",
":",
"type",
"input_statement",
":",
"Statement"
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/response_selection.py#L66-L86
|
train
|
gunthercox/ChatterBot
|
chatterbot/comparisons.py
|
LevenshteinDistance.compare
|
def compare(self, statement_a, statement_b):
"""
Compare the two input statements.
:return: The percent of similarity between the text of the statements.
:rtype: float
"""
# Return 0 if either statement has a falsy text value
if not statement_a.text or not statement_b.text:
return 0
# Get the lowercase version of both strings
statement_a_text = str(statement_a.text.lower())
statement_b_text = str(statement_b.text.lower())
similarity = SequenceMatcher(
None,
statement_a_text,
statement_b_text
)
# Calculate a decimal percent of the similarity
percent = round(similarity.ratio(), 2)
return percent
|
python
|
def compare(self, statement_a, statement_b):
"""
Compare the two input statements.
:return: The percent of similarity between the text of the statements.
:rtype: float
"""
# Return 0 if either statement has a falsy text value
if not statement_a.text or not statement_b.text:
return 0
# Get the lowercase version of both strings
statement_a_text = str(statement_a.text.lower())
statement_b_text = str(statement_b.text.lower())
similarity = SequenceMatcher(
None,
statement_a_text,
statement_b_text
)
# Calculate a decimal percent of the similarity
percent = round(similarity.ratio(), 2)
return percent
|
[
"def",
"compare",
"(",
"self",
",",
"statement_a",
",",
"statement_b",
")",
":",
"# Return 0 if either statement has a falsy text value",
"if",
"not",
"statement_a",
".",
"text",
"or",
"not",
"statement_b",
".",
"text",
":",
"return",
"0",
"# Get the lowercase version of both strings",
"statement_a_text",
"=",
"str",
"(",
"statement_a",
".",
"text",
".",
"lower",
"(",
")",
")",
"statement_b_text",
"=",
"str",
"(",
"statement_b",
".",
"text",
".",
"lower",
"(",
")",
")",
"similarity",
"=",
"SequenceMatcher",
"(",
"None",
",",
"statement_a_text",
",",
"statement_b_text",
")",
"# Calculate a decimal percent of the similarity",
"percent",
"=",
"round",
"(",
"similarity",
".",
"ratio",
"(",
")",
",",
"2",
")",
"return",
"percent"
] |
Compare the two input statements.
:return: The percent of similarity between the text of the statements.
:rtype: float
|
[
"Compare",
"the",
"two",
"input",
"statements",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/comparisons.py#L32-L57
|
train
|
gunthercox/ChatterBot
|
chatterbot/comparisons.py
|
SpacySimilarity.compare
|
def compare(self, statement_a, statement_b):
"""
Compare the two input statements.
:return: The percent of similarity between the closest synset distance.
:rtype: float
"""
document_a = self.nlp(statement_a.text)
document_b = self.nlp(statement_b.text)
return document_a.similarity(document_b)
|
python
|
def compare(self, statement_a, statement_b):
"""
Compare the two input statements.
:return: The percent of similarity between the closest synset distance.
:rtype: float
"""
document_a = self.nlp(statement_a.text)
document_b = self.nlp(statement_b.text)
return document_a.similarity(document_b)
|
[
"def",
"compare",
"(",
"self",
",",
"statement_a",
",",
"statement_b",
")",
":",
"document_a",
"=",
"self",
".",
"nlp",
"(",
"statement_a",
".",
"text",
")",
"document_b",
"=",
"self",
".",
"nlp",
"(",
"statement_b",
".",
"text",
")",
"return",
"document_a",
".",
"similarity",
"(",
"document_b",
")"
] |
Compare the two input statements.
:return: The percent of similarity between the closest synset distance.
:rtype: float
|
[
"Compare",
"the",
"two",
"input",
"statements",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/comparisons.py#L70-L80
|
train
|
gunthercox/ChatterBot
|
chatterbot/comparisons.py
|
JaccardSimilarity.compare
|
def compare(self, statement_a, statement_b):
"""
Return the calculated similarity of two
statements based on the Jaccard index.
"""
# Make both strings lowercase
document_a = self.nlp(statement_a.text.lower())
document_b = self.nlp(statement_b.text.lower())
statement_a_lemmas = set([
token.lemma_ for token in document_a if not token.is_stop
])
statement_b_lemmas = set([
token.lemma_ for token in document_b if not token.is_stop
])
# Calculate Jaccard similarity
numerator = len(statement_a_lemmas.intersection(statement_b_lemmas))
denominator = float(len(statement_a_lemmas.union(statement_b_lemmas)))
ratio = numerator / denominator
return ratio
|
python
|
def compare(self, statement_a, statement_b):
"""
Return the calculated similarity of two
statements based on the Jaccard index.
"""
# Make both strings lowercase
document_a = self.nlp(statement_a.text.lower())
document_b = self.nlp(statement_b.text.lower())
statement_a_lemmas = set([
token.lemma_ for token in document_a if not token.is_stop
])
statement_b_lemmas = set([
token.lemma_ for token in document_b if not token.is_stop
])
# Calculate Jaccard similarity
numerator = len(statement_a_lemmas.intersection(statement_b_lemmas))
denominator = float(len(statement_a_lemmas.union(statement_b_lemmas)))
ratio = numerator / denominator
return ratio
|
[
"def",
"compare",
"(",
"self",
",",
"statement_a",
",",
"statement_b",
")",
":",
"# Make both strings lowercase",
"document_a",
"=",
"self",
".",
"nlp",
"(",
"statement_a",
".",
"text",
".",
"lower",
"(",
")",
")",
"document_b",
"=",
"self",
".",
"nlp",
"(",
"statement_b",
".",
"text",
".",
"lower",
"(",
")",
")",
"statement_a_lemmas",
"=",
"set",
"(",
"[",
"token",
".",
"lemma_",
"for",
"token",
"in",
"document_a",
"if",
"not",
"token",
".",
"is_stop",
"]",
")",
"statement_b_lemmas",
"=",
"set",
"(",
"[",
"token",
".",
"lemma_",
"for",
"token",
"in",
"document_b",
"if",
"not",
"token",
".",
"is_stop",
"]",
")",
"# Calculate Jaccard similarity",
"numerator",
"=",
"len",
"(",
"statement_a_lemmas",
".",
"intersection",
"(",
"statement_b_lemmas",
")",
")",
"denominator",
"=",
"float",
"(",
"len",
"(",
"statement_a_lemmas",
".",
"union",
"(",
"statement_b_lemmas",
")",
")",
")",
"ratio",
"=",
"numerator",
"/",
"denominator",
"return",
"ratio"
] |
Return the calculated similarity of two
statements based on the Jaccard index.
|
[
"Return",
"the",
"calculated",
"similarity",
"of",
"two",
"statements",
"based",
"on",
"the",
"Jaccard",
"index",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/comparisons.py#L114-L135
|
train
|
gunthercox/ChatterBot
|
chatterbot/storage/mongodb.py
|
MongoDatabaseAdapter.get_statement_model
|
def get_statement_model(self):
"""
Return the class for the statement model.
"""
from chatterbot.conversation import Statement
# Create a storage-aware statement
statement = Statement
statement.storage = self
return statement
|
python
|
def get_statement_model(self):
"""
Return the class for the statement model.
"""
from chatterbot.conversation import Statement
# Create a storage-aware statement
statement = Statement
statement.storage = self
return statement
|
[
"def",
"get_statement_model",
"(",
"self",
")",
":",
"from",
"chatterbot",
".",
"conversation",
"import",
"Statement",
"# Create a storage-aware statement",
"statement",
"=",
"Statement",
"statement",
".",
"storage",
"=",
"self",
"return",
"statement"
] |
Return the class for the statement model.
|
[
"Return",
"the",
"class",
"for",
"the",
"statement",
"model",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/storage/mongodb.py#L44-L54
|
train
|
gunthercox/ChatterBot
|
chatterbot/storage/mongodb.py
|
MongoDatabaseAdapter.mongo_to_object
|
def mongo_to_object(self, statement_data):
"""
Return Statement object when given data
returned from Mongo DB.
"""
Statement = self.get_model('statement')
statement_data['id'] = statement_data['_id']
return Statement(**statement_data)
|
python
|
def mongo_to_object(self, statement_data):
"""
Return Statement object when given data
returned from Mongo DB.
"""
Statement = self.get_model('statement')
statement_data['id'] = statement_data['_id']
return Statement(**statement_data)
|
[
"def",
"mongo_to_object",
"(",
"self",
",",
"statement_data",
")",
":",
"Statement",
"=",
"self",
".",
"get_model",
"(",
"'statement'",
")",
"statement_data",
"[",
"'id'",
"]",
"=",
"statement_data",
"[",
"'_id'",
"]",
"return",
"Statement",
"(",
"*",
"*",
"statement_data",
")"
] |
Return Statement object when given data
returned from Mongo DB.
|
[
"Return",
"Statement",
"object",
"when",
"given",
"data",
"returned",
"from",
"Mongo",
"DB",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/storage/mongodb.py#L59-L68
|
train
|
gunthercox/ChatterBot
|
chatterbot/storage/mongodb.py
|
MongoDatabaseAdapter.filter
|
def filter(self, **kwargs):
"""
Returns a list of statements in the database
that match the parameters specified.
"""
import pymongo
page_size = kwargs.pop('page_size', 1000)
order_by = kwargs.pop('order_by', None)
tags = kwargs.pop('tags', [])
exclude_text = kwargs.pop('exclude_text', None)
exclude_text_words = kwargs.pop('exclude_text_words', [])
persona_not_startswith = kwargs.pop('persona_not_startswith', None)
search_text_contains = kwargs.pop('search_text_contains', None)
if tags:
kwargs['tags'] = {
'$in': tags
}
if exclude_text:
if 'text' not in kwargs:
kwargs['text'] = {}
elif 'text' in kwargs and isinstance(kwargs['text'], str):
text = kwargs.pop('text')
kwargs['text'] = {
'$eq': text
}
kwargs['text']['$nin'] = exclude_text
if exclude_text_words:
if 'text' not in kwargs:
kwargs['text'] = {}
elif 'text' in kwargs and isinstance(kwargs['text'], str):
text = kwargs.pop('text')
kwargs['text'] = {
'$eq': text
}
exclude_word_regex = '|'.join([
'.*{}.*'.format(word) for word in exclude_text_words
])
kwargs['text']['$not'] = re.compile(exclude_word_regex)
if persona_not_startswith:
if 'persona' not in kwargs:
kwargs['persona'] = {}
elif 'persona' in kwargs and isinstance(kwargs['persona'], str):
persona = kwargs.pop('persona')
kwargs['persona'] = {
'$eq': persona
}
kwargs['persona']['$not'] = re.compile('^bot:*')
if search_text_contains:
or_regex = '|'.join([
'{}'.format(word) for word in search_text_contains.split(' ')
])
kwargs['search_text'] = re.compile(or_regex)
mongo_ordering = []
if order_by:
# Sort so that newer datetimes appear first
if 'created_at' in order_by:
order_by.remove('created_at')
mongo_ordering.append(('created_at', pymongo.DESCENDING, ))
for order in order_by:
mongo_ordering.append((order, pymongo.ASCENDING))
total_statements = self.statements.find(kwargs).count()
for start_index in range(0, total_statements, page_size):
if mongo_ordering:
for match in self.statements.find(kwargs).sort(mongo_ordering).skip(start_index).limit(page_size):
yield self.mongo_to_object(match)
else:
for match in self.statements.find(kwargs).skip(start_index).limit(page_size):
yield self.mongo_to_object(match)
|
python
|
def filter(self, **kwargs):
"""
Returns a list of statements in the database
that match the parameters specified.
"""
import pymongo
page_size = kwargs.pop('page_size', 1000)
order_by = kwargs.pop('order_by', None)
tags = kwargs.pop('tags', [])
exclude_text = kwargs.pop('exclude_text', None)
exclude_text_words = kwargs.pop('exclude_text_words', [])
persona_not_startswith = kwargs.pop('persona_not_startswith', None)
search_text_contains = kwargs.pop('search_text_contains', None)
if tags:
kwargs['tags'] = {
'$in': tags
}
if exclude_text:
if 'text' not in kwargs:
kwargs['text'] = {}
elif 'text' in kwargs and isinstance(kwargs['text'], str):
text = kwargs.pop('text')
kwargs['text'] = {
'$eq': text
}
kwargs['text']['$nin'] = exclude_text
if exclude_text_words:
if 'text' not in kwargs:
kwargs['text'] = {}
elif 'text' in kwargs and isinstance(kwargs['text'], str):
text = kwargs.pop('text')
kwargs['text'] = {
'$eq': text
}
exclude_word_regex = '|'.join([
'.*{}.*'.format(word) for word in exclude_text_words
])
kwargs['text']['$not'] = re.compile(exclude_word_regex)
if persona_not_startswith:
if 'persona' not in kwargs:
kwargs['persona'] = {}
elif 'persona' in kwargs and isinstance(kwargs['persona'], str):
persona = kwargs.pop('persona')
kwargs['persona'] = {
'$eq': persona
}
kwargs['persona']['$not'] = re.compile('^bot:*')
if search_text_contains:
or_regex = '|'.join([
'{}'.format(word) for word in search_text_contains.split(' ')
])
kwargs['search_text'] = re.compile(or_regex)
mongo_ordering = []
if order_by:
# Sort so that newer datetimes appear first
if 'created_at' in order_by:
order_by.remove('created_at')
mongo_ordering.append(('created_at', pymongo.DESCENDING, ))
for order in order_by:
mongo_ordering.append((order, pymongo.ASCENDING))
total_statements = self.statements.find(kwargs).count()
for start_index in range(0, total_statements, page_size):
if mongo_ordering:
for match in self.statements.find(kwargs).sort(mongo_ordering).skip(start_index).limit(page_size):
yield self.mongo_to_object(match)
else:
for match in self.statements.find(kwargs).skip(start_index).limit(page_size):
yield self.mongo_to_object(match)
|
[
"def",
"filter",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"import",
"pymongo",
"page_size",
"=",
"kwargs",
".",
"pop",
"(",
"'page_size'",
",",
"1000",
")",
"order_by",
"=",
"kwargs",
".",
"pop",
"(",
"'order_by'",
",",
"None",
")",
"tags",
"=",
"kwargs",
".",
"pop",
"(",
"'tags'",
",",
"[",
"]",
")",
"exclude_text",
"=",
"kwargs",
".",
"pop",
"(",
"'exclude_text'",
",",
"None",
")",
"exclude_text_words",
"=",
"kwargs",
".",
"pop",
"(",
"'exclude_text_words'",
",",
"[",
"]",
")",
"persona_not_startswith",
"=",
"kwargs",
".",
"pop",
"(",
"'persona_not_startswith'",
",",
"None",
")",
"search_text_contains",
"=",
"kwargs",
".",
"pop",
"(",
"'search_text_contains'",
",",
"None",
")",
"if",
"tags",
":",
"kwargs",
"[",
"'tags'",
"]",
"=",
"{",
"'$in'",
":",
"tags",
"}",
"if",
"exclude_text",
":",
"if",
"'text'",
"not",
"in",
"kwargs",
":",
"kwargs",
"[",
"'text'",
"]",
"=",
"{",
"}",
"elif",
"'text'",
"in",
"kwargs",
"and",
"isinstance",
"(",
"kwargs",
"[",
"'text'",
"]",
",",
"str",
")",
":",
"text",
"=",
"kwargs",
".",
"pop",
"(",
"'text'",
")",
"kwargs",
"[",
"'text'",
"]",
"=",
"{",
"'$eq'",
":",
"text",
"}",
"kwargs",
"[",
"'text'",
"]",
"[",
"'$nin'",
"]",
"=",
"exclude_text",
"if",
"exclude_text_words",
":",
"if",
"'text'",
"not",
"in",
"kwargs",
":",
"kwargs",
"[",
"'text'",
"]",
"=",
"{",
"}",
"elif",
"'text'",
"in",
"kwargs",
"and",
"isinstance",
"(",
"kwargs",
"[",
"'text'",
"]",
",",
"str",
")",
":",
"text",
"=",
"kwargs",
".",
"pop",
"(",
"'text'",
")",
"kwargs",
"[",
"'text'",
"]",
"=",
"{",
"'$eq'",
":",
"text",
"}",
"exclude_word_regex",
"=",
"'|'",
".",
"join",
"(",
"[",
"'.*{}.*'",
".",
"format",
"(",
"word",
")",
"for",
"word",
"in",
"exclude_text_words",
"]",
")",
"kwargs",
"[",
"'text'",
"]",
"[",
"'$not'",
"]",
"=",
"re",
".",
"compile",
"(",
"exclude_word_regex",
")",
"if",
"persona_not_startswith",
":",
"if",
"'persona'",
"not",
"in",
"kwargs",
":",
"kwargs",
"[",
"'persona'",
"]",
"=",
"{",
"}",
"elif",
"'persona'",
"in",
"kwargs",
"and",
"isinstance",
"(",
"kwargs",
"[",
"'persona'",
"]",
",",
"str",
")",
":",
"persona",
"=",
"kwargs",
".",
"pop",
"(",
"'persona'",
")",
"kwargs",
"[",
"'persona'",
"]",
"=",
"{",
"'$eq'",
":",
"persona",
"}",
"kwargs",
"[",
"'persona'",
"]",
"[",
"'$not'",
"]",
"=",
"re",
".",
"compile",
"(",
"'^bot:*'",
")",
"if",
"search_text_contains",
":",
"or_regex",
"=",
"'|'",
".",
"join",
"(",
"[",
"'{}'",
".",
"format",
"(",
"word",
")",
"for",
"word",
"in",
"search_text_contains",
".",
"split",
"(",
"' '",
")",
"]",
")",
"kwargs",
"[",
"'search_text'",
"]",
"=",
"re",
".",
"compile",
"(",
"or_regex",
")",
"mongo_ordering",
"=",
"[",
"]",
"if",
"order_by",
":",
"# Sort so that newer datetimes appear first",
"if",
"'created_at'",
"in",
"order_by",
":",
"order_by",
".",
"remove",
"(",
"'created_at'",
")",
"mongo_ordering",
".",
"append",
"(",
"(",
"'created_at'",
",",
"pymongo",
".",
"DESCENDING",
",",
")",
")",
"for",
"order",
"in",
"order_by",
":",
"mongo_ordering",
".",
"append",
"(",
"(",
"order",
",",
"pymongo",
".",
"ASCENDING",
")",
")",
"total_statements",
"=",
"self",
".",
"statements",
".",
"find",
"(",
"kwargs",
")",
".",
"count",
"(",
")",
"for",
"start_index",
"in",
"range",
"(",
"0",
",",
"total_statements",
",",
"page_size",
")",
":",
"if",
"mongo_ordering",
":",
"for",
"match",
"in",
"self",
".",
"statements",
".",
"find",
"(",
"kwargs",
")",
".",
"sort",
"(",
"mongo_ordering",
")",
".",
"skip",
"(",
"start_index",
")",
".",
"limit",
"(",
"page_size",
")",
":",
"yield",
"self",
".",
"mongo_to_object",
"(",
"match",
")",
"else",
":",
"for",
"match",
"in",
"self",
".",
"statements",
".",
"find",
"(",
"kwargs",
")",
".",
"skip",
"(",
"start_index",
")",
".",
"limit",
"(",
"page_size",
")",
":",
"yield",
"self",
".",
"mongo_to_object",
"(",
"match",
")"
] |
Returns a list of statements in the database
that match the parameters specified.
|
[
"Returns",
"a",
"list",
"of",
"statements",
"in",
"the",
"database",
"that",
"match",
"the",
"parameters",
"specified",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/storage/mongodb.py#L70-L149
|
train
|
gunthercox/ChatterBot
|
chatterbot/storage/mongodb.py
|
MongoDatabaseAdapter.create
|
def create(self, **kwargs):
"""
Creates a new statement matching the keyword arguments specified.
Returns the created statement.
"""
Statement = self.get_model('statement')
if 'tags' in kwargs:
kwargs['tags'] = list(set(kwargs['tags']))
if 'search_text' not in kwargs:
kwargs['search_text'] = self.tagger.get_bigram_pair_string(kwargs['text'])
if 'search_in_response_to' not in kwargs:
if kwargs.get('in_response_to'):
kwargs['search_in_response_to'] = self.tagger.get_bigram_pair_string(kwargs['in_response_to'])
inserted = self.statements.insert_one(kwargs)
kwargs['id'] = inserted.inserted_id
return Statement(**kwargs)
|
python
|
def create(self, **kwargs):
"""
Creates a new statement matching the keyword arguments specified.
Returns the created statement.
"""
Statement = self.get_model('statement')
if 'tags' in kwargs:
kwargs['tags'] = list(set(kwargs['tags']))
if 'search_text' not in kwargs:
kwargs['search_text'] = self.tagger.get_bigram_pair_string(kwargs['text'])
if 'search_in_response_to' not in kwargs:
if kwargs.get('in_response_to'):
kwargs['search_in_response_to'] = self.tagger.get_bigram_pair_string(kwargs['in_response_to'])
inserted = self.statements.insert_one(kwargs)
kwargs['id'] = inserted.inserted_id
return Statement(**kwargs)
|
[
"def",
"create",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"Statement",
"=",
"self",
".",
"get_model",
"(",
"'statement'",
")",
"if",
"'tags'",
"in",
"kwargs",
":",
"kwargs",
"[",
"'tags'",
"]",
"=",
"list",
"(",
"set",
"(",
"kwargs",
"[",
"'tags'",
"]",
")",
")",
"if",
"'search_text'",
"not",
"in",
"kwargs",
":",
"kwargs",
"[",
"'search_text'",
"]",
"=",
"self",
".",
"tagger",
".",
"get_bigram_pair_string",
"(",
"kwargs",
"[",
"'text'",
"]",
")",
"if",
"'search_in_response_to'",
"not",
"in",
"kwargs",
":",
"if",
"kwargs",
".",
"get",
"(",
"'in_response_to'",
")",
":",
"kwargs",
"[",
"'search_in_response_to'",
"]",
"=",
"self",
".",
"tagger",
".",
"get_bigram_pair_string",
"(",
"kwargs",
"[",
"'in_response_to'",
"]",
")",
"inserted",
"=",
"self",
".",
"statements",
".",
"insert_one",
"(",
"kwargs",
")",
"kwargs",
"[",
"'id'",
"]",
"=",
"inserted",
".",
"inserted_id",
"return",
"Statement",
"(",
"*",
"*",
"kwargs",
")"
] |
Creates a new statement matching the keyword arguments specified.
Returns the created statement.
|
[
"Creates",
"a",
"new",
"statement",
"matching",
"the",
"keyword",
"arguments",
"specified",
".",
"Returns",
"the",
"created",
"statement",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/storage/mongodb.py#L151-L172
|
train
|
gunthercox/ChatterBot
|
chatterbot/storage/mongodb.py
|
MongoDatabaseAdapter.create_many
|
def create_many(self, statements):
"""
Creates multiple statement entries.
"""
create_statements = []
for statement in statements:
statement_data = statement.serialize()
tag_data = list(set(statement_data.pop('tags', [])))
statement_data['tags'] = tag_data
if not statement.search_text:
statement_data['search_text'] = self.tagger.get_bigram_pair_string(statement.text)
if not statement.search_in_response_to and statement.in_response_to:
statement_data['search_in_response_to'] = self.tagger.get_bigram_pair_string(statement.in_response_to)
create_statements.append(statement_data)
self.statements.insert_many(create_statements)
|
python
|
def create_many(self, statements):
"""
Creates multiple statement entries.
"""
create_statements = []
for statement in statements:
statement_data = statement.serialize()
tag_data = list(set(statement_data.pop('tags', [])))
statement_data['tags'] = tag_data
if not statement.search_text:
statement_data['search_text'] = self.tagger.get_bigram_pair_string(statement.text)
if not statement.search_in_response_to and statement.in_response_to:
statement_data['search_in_response_to'] = self.tagger.get_bigram_pair_string(statement.in_response_to)
create_statements.append(statement_data)
self.statements.insert_many(create_statements)
|
[
"def",
"create_many",
"(",
"self",
",",
"statements",
")",
":",
"create_statements",
"=",
"[",
"]",
"for",
"statement",
"in",
"statements",
":",
"statement_data",
"=",
"statement",
".",
"serialize",
"(",
")",
"tag_data",
"=",
"list",
"(",
"set",
"(",
"statement_data",
".",
"pop",
"(",
"'tags'",
",",
"[",
"]",
")",
")",
")",
"statement_data",
"[",
"'tags'",
"]",
"=",
"tag_data",
"if",
"not",
"statement",
".",
"search_text",
":",
"statement_data",
"[",
"'search_text'",
"]",
"=",
"self",
".",
"tagger",
".",
"get_bigram_pair_string",
"(",
"statement",
".",
"text",
")",
"if",
"not",
"statement",
".",
"search_in_response_to",
"and",
"statement",
".",
"in_response_to",
":",
"statement_data",
"[",
"'search_in_response_to'",
"]",
"=",
"self",
".",
"tagger",
".",
"get_bigram_pair_string",
"(",
"statement",
".",
"in_response_to",
")",
"create_statements",
".",
"append",
"(",
"statement_data",
")",
"self",
".",
"statements",
".",
"insert_many",
"(",
"create_statements",
")"
] |
Creates multiple statement entries.
|
[
"Creates",
"multiple",
"statement",
"entries",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/storage/mongodb.py#L174-L193
|
train
|
gunthercox/ChatterBot
|
chatterbot/storage/mongodb.py
|
MongoDatabaseAdapter.get_random
|
def get_random(self):
"""
Returns a random statement from the database
"""
from random import randint
count = self.count()
if count < 1:
raise self.EmptyDatabaseException()
random_integer = randint(0, count - 1)
statements = self.statements.find().limit(1).skip(random_integer)
return self.mongo_to_object(list(statements)[0])
|
python
|
def get_random(self):
"""
Returns a random statement from the database
"""
from random import randint
count = self.count()
if count < 1:
raise self.EmptyDatabaseException()
random_integer = randint(0, count - 1)
statements = self.statements.find().limit(1).skip(random_integer)
return self.mongo_to_object(list(statements)[0])
|
[
"def",
"get_random",
"(",
"self",
")",
":",
"from",
"random",
"import",
"randint",
"count",
"=",
"self",
".",
"count",
"(",
")",
"if",
"count",
"<",
"1",
":",
"raise",
"self",
".",
"EmptyDatabaseException",
"(",
")",
"random_integer",
"=",
"randint",
"(",
"0",
",",
"count",
"-",
"1",
")",
"statements",
"=",
"self",
".",
"statements",
".",
"find",
"(",
")",
".",
"limit",
"(",
"1",
")",
".",
"skip",
"(",
"random_integer",
")",
"return",
"self",
".",
"mongo_to_object",
"(",
"list",
"(",
"statements",
")",
"[",
"0",
"]",
")"
] |
Returns a random statement from the database
|
[
"Returns",
"a",
"random",
"statement",
"from",
"the",
"database"
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/storage/mongodb.py#L235-L250
|
train
|
gunthercox/ChatterBot
|
chatterbot/ext/sqlalchemy_app/models.py
|
Statement.add_tags
|
def add_tags(self, *tags):
"""
Add a list of strings to the statement as tags.
"""
self.tags.extend([
Tag(name=tag) for tag in tags
])
|
python
|
def add_tags(self, *tags):
"""
Add a list of strings to the statement as tags.
"""
self.tags.extend([
Tag(name=tag) for tag in tags
])
|
[
"def",
"add_tags",
"(",
"self",
",",
"*",
"tags",
")",
":",
"self",
".",
"tags",
".",
"extend",
"(",
"[",
"Tag",
"(",
"name",
"=",
"tag",
")",
"for",
"tag",
"in",
"tags",
"]",
")"
] |
Add a list of strings to the statement as tags.
|
[
"Add",
"a",
"list",
"of",
"strings",
"to",
"the",
"statement",
"as",
"tags",
"."
] |
1a03dcb45cba7bdc24d3db5e750582e0cb1518e2
|
https://github.com/gunthercox/ChatterBot/blob/1a03dcb45cba7bdc24d3db5e750582e0cb1518e2/chatterbot/ext/sqlalchemy_app/models.py#L108-L114
|
train
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.