file_name large_stringlengths 4 140 | prefix large_stringlengths 0 39k | suffix large_stringlengths 0 36.1k | middle large_stringlengths 0 29.4k | fim_type large_stringclasses 4
values |
|---|---|---|---|---|
sevencow.py | # -*- coding: utf-8 -*-
import os
import hmac
import time
import json
import mimetypes
import functools
from urlparse import urlparse
from urllib import urlencode
from base64 import urlsafe_b64encode
from hashlib import sha1
import requests
version_info = (0, 1, 2)
VERSION = __version__ = '.'.join( map(str, version_info) )
"""
Usage:
cow = Cow(ACCESS_KEY, SECRET_KEY)
b = cow.get_bucket(BUCKET)
b.put('a')
b.put('a', 'b')
b.put('a', names={'a': 'x'})
b.put('a', 'b', names={'a': 'x', 'b': 'y'})
b.stat('a')
b.stat('a', 'b')
b.delete('a')
b.delete('a', 'b')
b.copy('a', 'c')
b.copy(('a', 'c'), ('b', 'd'))
b.move('a', 'c')
b.move(('a', 'c'), ('b', 'd'))
"""
RS_HOST = 'http://rs.qbox.me'
UP_HOST = 'http://up.qbox.me'
RSF_HOST = 'http://rsf.qbox.me'
class CowException(Exception):
def __init__(self, url, status_code, reason, content):
self.url = url
self.status_code = status_code
self.reason = reason
self.content = content
Exception.__init__(self, '%s, %s' % (reason, content))
def signing(secret_key, data):
return urlsafe_b64encode(
hmac.new(secret_key, data, sha1).digest()
)
def requests_error_handler(func):
@functools.wraps(func)
def deco(*args, **kwargs):
try:
return func(*args, **kwargs)
except AssertionError as e:
req = e.args[0]
raise CowException(
req.url, req.status_code, req.reason, req.content
)
return deco
def expected_argument_type(pos, types):
def deco(func):
@functools.wraps(func)
def wrap(*args, **kwargs):
if not isinstance(args[pos], types):
raise TypeError(
"{0} Type error, Expected {1}".format(args[pos], types)
)
return func(*args, **kwargs)
return wrap
return deco
class UploadToken(object):
def __init__(self, access_key, secret_key, scope, ttl=3600):
self.access_key = access_key
self.secret_key = secret_key
self.scope = scope
self.ttl = ttl
self._token = None
self.generated = int(time.time())
@property
def token(self):
if int(time.time()) - self.generated > self.ttl - 60:
# 还有一分钟也认为过期了, make new token
self._token = self._make_token()
if not self._token:
self._token = self._make_token()
return self._token
def _make_token(self):
self.generated = int(time.time())
info = {
'scope': self.scope,
'deadline': self.generated + self.ttl
}
info = urlsafe_b64encode(json.dumps(info))
token = signing(self.secret_key, info)
return '%s:%s:%s' % (self.access_key, token, info)
class Cow(object):
def __init__(self, access_key, secret_key):
self.access_key = access_key
self.secret_key = secret_key |
self.stat = functools.partial(self._stat_rm_handler, 'stat')
self.delete = functools.partial(self._stat_rm_handler, 'delete')
self.copy = functools.partial(self._cp_mv_handler, 'copy')
self.move = functools.partial(self._cp_mv_handler, 'move')
def get_bucket(self, bucket):
"""对一个bucket的文件进行操作,
推荐使用此方法得到一个bucket对象,
然后对此bucket的操作就只用传递文件名即可
"""
return Bucket(self, bucket)
def generate_access_token(self, url, params=None):
uri = urlparse(url)
token = uri.path
if uri.query:
token = '%s?%s' % (token, uri.query)
token = '%s\n' % token
if params:
if isinstance(params, basestring):
token += params
else:
token += urlencode(params)
return '%s:%s' % (self.access_key, signing(self.secret_key, token))
def build_requests_headers(self, token):
return {
'Content-Type': 'application/x-www-form-urlencoded',
'Authorization': 'QBox %s' % token
}
@requests_error_handler
def api_call(self, url, params=None):
token = self.generate_access_token(url, params=params)
if params:
res = requests.post(url, data=params, headers=self.build_requests_headers(token))
else:
res = requests.post(url, headers=self.build_requests_headers(token))
assert res.status_code == 200, res
return res.json() if res.text else ''
def list_buckets(self):
"""列出所有的buckets"""
url = '%s/buckets' % RS_HOST
return self.api_call(url)
def create_bucket(self, name):
"""不建议使用API建立bucket
测试发现API建立的bucket默认无法设置<bucket_name>.qiniudn.com的二级域名
请直接到web界面建立
"""
url = '%s/mkbucket/%s' % (RS_HOST, name)
return self.api_call(url)
def drop_bucket(self, bucket):
"""删除整个bucket"""
url = '%s/drop/%s' % (RS_HOST, bucket)
return self.api_call(url)
def list_files(self, bucket, marker=None, limit=None, prefix=None):
"""列出bucket中的文件"""
query = ['bucket=%s' % bucket]
if marker:
query.append('marker=%s' % marker)
if limit:
query.append('limit=%s' % limit)
if prefix:
query.append('prefix=%s' % prefix)
url = '%s/list?%s' % (RSF_HOST, '&'.join(query))
return self.api_call(url)
def generate_upload_token(self, scope, ttl=3600):
"""上传文件的uploadToken"""
if scope not in self.upload_tokens:
self.upload_tokens[scope] = UploadToken(self.access_key, self.secret_key, scope, ttl=ttl)
return self.upload_tokens[scope].token
@requests_error_handler
@expected_argument_type(2, (basestring, list, tuple))
def put(self, scope, filename, names=None):
"""上传文件
filename 如果是字符串,表示上传单个文件,
如果是list或者tuple,表示上传多个文件
names 是dict,key为filename, value为上传后的名字
如果不设置,默认为文件名
"""
url = '%s/upload' % UP_HOST
token = self.generate_upload_token(scope)
names = names or {}
def _uploaded_name(filename):
return names.get(filename, None) or os.path.basename(filename)
def _put(filename):
files = {
'file': (filename, open(filename, 'rb')),
}
action = '/rs-put/%s' % urlsafe_b64encode(
'%s:%s' % (scope, _uploaded_name(filename))
)
_type, _encoding = mimetypes.guess_type(filename)
if _type:
action += '/mimeType/%s' % urlsafe_b64encode(_type)
data = {
'auth': token,
'action': action,
}
res = requests.post(url, files=files, data=data)
assert res.status_code == 200, res
return res.json()
if isinstance(filename, basestring):
# 单个文件
return _put(filename)
# 多文件
return [_put(f) for f in filename]
@expected_argument_type(2, (list, tuple))
def _cp_mv_handler(self, action, args):
"""copy move方法
action: 'copy' or 'move'
args: [src_bucket, src_filename, des_bucket, des_filename]
or [(src_bucket, src_filename, des_bucket, des_filename), (), ...]
args 第一种形式就是对一个文件进行操作,第二种形式是多个文件批量操作
用户不用直接调用这个方法
"""
if isinstance(args[0], basestring):
return self._cp_mv_single(action, args)
if isinstance(args[0], (list, tuple)):
return self._cp_mv_batch(action, args)
@expected_argument_type(3, (basestring, list, tuple))
def _stat_rm_handler(self, action, bucket, filename):
"""stat delete方法
action: 'stat' or 'delete'
bucket: 哪个bucket
filenmae: 'aabb' or ['aabb', 'ccdd', ...]
filename 第一种形式就是对一个文件进行操作,第二种形式是多个文件批量操作
用户不用直接调用这个方法
"""
if isinstance(filename, basestring):
return self._stat_rm_single(action, bucket, filename)
if isinstance(filename, (list, tuple)):
return self._stat_rm_batch(action, bucket, filename)
def _cp_mv_single(self, action, args):
src_bucket, src_filename, des_bucket, des_filename = args
url = '%s/%s/%s/%s' % (
RS_HOST,
action,
urlsafe_b64encode('%s:%s' % (src_bucket, src_filename)),
urlsafe_b64encode('%s:%s' % (des_bucket, des_filename)),
)
return self.api_call(url)
def _cp_mv_batch(self, action, args):
url = '%s/batch' % RS_HOST
def _one_param(arg):
return 'op=/%s/%s/%s' % (
action,
urlsafe_b64encode('%s:%s' % (arg[0], arg[1])),
urlsafe_b64encode('%s:%s' % (arg[2], arg[3])),
)
param = '&'.join( map(_one_param, args) )
return self.api_call(url, param)
def _stat_rm_single(self, action, bucket, filename):
url = '%s/%s/%s' % (
RS_HOST, action, urlsafe_b64encode('%s:%s' % (bucket, filename))
)
return self.api_call(url)
def _stat_rm_batch(self, action, bucket, filenames):
url = '%s/batch' % RS_HOST
param = [
'op=/%s/%s' % (
action, urlsafe_b64encode('%s:%s' % (bucket, f))
) for f in filenames
]
param = '&'.join(param)
return self.api_call(url, param)
def transform_argument(func):
@functools.wraps(func)
def deco(self, *args, **kwargs):
filename = args[0] if len(args) == 1 else args
return func(self, filename, **kwargs)
return deco
class Bucket(object):
def __init__(self, cow, bucket):
self.cow = cow
self.bucket = bucket
@transform_argument
def put(self, *args, **kwargs):
names = kwargs.get('names', None)
if names and not isinstance(names, dict):
raise TypeError(
"names Type error, Expected dict, But got Type of {0}".format(type(names))
)
return self.cow.put(self.bucket, args[0], names=names)
@transform_argument
def stat(self, *args):
return self.cow.stat(self.bucket, args[0])
@transform_argument
def delete(self, *args):
return self.cow.delete(self.bucket, args[0])
@transform_argument
def copy(self, *args):
return self.cow.copy(self._build_cp_mv_args(args[0]))
@transform_argument
def move(self, *args):
return self.cow.move(self._build_cp_mv_args(args[0]))
def list_files(self, marker=None, limit=None, prefix=None):
return self.cow.list_files(self.bucket, marker=marker, limit=limit, prefix=prefix)
def _build_cp_mv_args(self, filename):
if isinstance(filename[0], basestring):
args = [self.bucket, filename[0], self.bucket, filename[1]]
else:
args = []
for src, des in filename:
args.append( (self.bucket, src, self.bucket, des) )
return args | self.upload_tokens = {} | random_line_split |
sevencow.py | # -*- coding: utf-8 -*-
import os
import hmac
import time
import json
import mimetypes
import functools
from urlparse import urlparse
from urllib import urlencode
from base64 import urlsafe_b64encode
from hashlib import sha1
import requests
version_info = (0, 1, 2)
VERSION = __version__ = '.'.join( map(str, version_info) )
"""
Usage:
cow = Cow(ACCESS_KEY, SECRET_KEY)
b = cow.get_bucket(BUCKET)
b.put('a')
b.put('a', 'b')
b.put('a', names={'a': 'x'})
b.put('a', 'b', names={'a': 'x', 'b': 'y'})
b.stat('a')
b.stat('a', 'b')
b.delete('a')
b.delete('a', 'b')
b.copy('a', 'c')
b.copy(('a', 'c'), ('b', 'd'))
b.move('a', 'c')
b.move(('a', 'c'), ('b', 'd'))
"""
RS_HOST = 'http://rs.qbox.me'
UP_HOST = 'http://up.qbox.me'
RSF_HOST = 'http://rsf.qbox.me'
class CowException(Exception):
def __init__(self, url, status_code, reason, content):
self.url = url
self.status_code = status_code
self.reason = reason
self.content = content
Exception.__init__(self, '%s, %s' % (reason, content))
def signing(secret_key, data):
return urlsafe_b64encode(
hmac.new(secret_key, data, sha1).digest()
)
def requests_error_handler(func):
@functools.wraps(func)
def deco(*args, **kwargs):
try:
return func(*args, **kwargs)
except AssertionError as e:
req = e.args[0]
raise CowException(
req.url, req.status_code, req.reason, req.content
)
return deco
def expected_argument_type(pos, types):
def deco(func):
@functools.wraps(func)
def wrap(*args, **kwargs):
if not isinstance(args[pos], types):
raise TypeError(
"{0} Type error, Expected {1}".format(args[pos], types)
)
return func(*args, **kwargs)
return wrap
return deco
class UploadToken(object):
def __init__(self, access_key, secret_key, scope, ttl=3600):
self.access_key = access_key
self.secret_key = secret_key
self.scope = scope
self.ttl = ttl
self._token = None
self.generated = int(time.time())
@property
def token(self):
if int(time.time()) - self.generated > self.ttl - 60:
# 还有一分钟也认为过期了, make new token
self._token = self._make_token()
if not self._token:
self._token = self._make_token()
return self._token
def _make_token(self):
self.generated = int(time.time())
info = {
'scope': self.scope,
'deadline': self.generated + self.ttl
}
info = urlsafe_b64encode(json.dumps(info))
token = signing(self.secret_key, info)
return '%s:%s:%s' % (self.access_key, token, info)
class Cow(object):
def __init__(self, access_key, secret_key):
self.access_key = access_key
self.secret_key = secret_key
self.upload_tokens = {}
self.stat = functools.partial(self._stat_rm_handler, 'stat')
self.delete = functools.partial(self._stat_rm_handler, 'delete')
self.copy = functools.partial(self._cp_mv_handler, 'copy')
self.move = functools.partial(self._cp_mv_handler, 'move')
def get_bucket(self, bucket):
"""对一个bucket的文件进行操作,
推荐使用此方法得到一个bucket对象,
然后对此bucket的操作就只用传递文件名即可
"""
return Bucket(self, bucket)
def generate_access_token(self, url, params=None):
uri = urlparse(url)
token = uri.path
if uri.query:
token = '%s?%s' % (token, uri.query)
token = '%s\n' % token
if params:
if isinstance(params, basestring):
token += params
else:
token += urlencode(params)
return '%s:%s' % (self.access_key, signing(self.secret_key, token))
def build_requests_headers(self, token):
return {
'Content-Type': 'application/x-www-form-urlencoded',
'Authorization': 'QBox %s' % token
}
@requests_error_handler
def api_call(self, url, params=None):
token = self.generate_access_token(url, params=params)
if params:
res = requests.post(url, data=params, headers=self.build_requests_headers(token))
else:
res = requests.post(url, headers=self.build_requests_headers(token))
assert res.status_code == 200, res
return res.json() if res.text else ''
def list_buckets(self):
"""列出所有的buckets"""
url = '%s/buckets' % RS_HOST
return self.api_call(url)
def create_bucket(self, name):
"""不建议使用API建立bucket
测试发现API建立的bucket默认无法设置<bucket_name>.qiniudn.com的二级域名
请直接到web界面建立
"""
url = '%s/mkbucket/%s' % (RS_HOST, name)
return self.api_call(url)
def drop_bucket(self, bucket):
"""删除整个bucket"""
url = '%s/drop/%s' % (RS_HOST, bucket)
return self.api_call(url)
def list_files(self, bucket, marker=None, limit=None, prefix=None):
"""列出bucket中的文件"""
query = ['bucket=%s' % bucket]
if marker:
query.append('marker=%s' % marker)
if limit:
query.append('limit=%s' % limit)
if prefix:
query.append('prefix=%s' % prefix)
url = '%s/list?%s' % (RSF_HOST, '&'.join(query))
return self.api_call(url)
def generate_upload_token(self, scope, ttl=3600):
"""上传文件的uploadToken"""
if scope not in self.upload_tokens:
self.upload_tokens[scope] = UploadToken(self.access_key, self.secret_key, scope, ttl=ttl)
return self.upload_tokens[scope].token
@requests_error_handler
@expected_argument_type(2, (basestring, list, tuple))
def put(self, scope, filename, names=None):
"""上传文件
filename 如果是字符串,表示上传单个文件,
如果是list或者tuple,表示上传多个文件
names 是dict,key为filename, value为上传后的名字
如果不设置,默认为文件名
"""
url = '%s/upload' % UP_HOST
token = self.generate_upload_token(scope)
names = names or {}
def _uploaded_name(filename):
return names.get(filename, None) or os.path.basename(filename)
def _put(filename):
files = {
'file': (filename, open(filename, 'rb')),
}
action = '/rs-put/%s' % urlsafe_b64encode(
'%s:%s' % (scope, _uploaded_name(filename))
)
_type, _encoding = mimetypes.guess_type(filename)
if _type:
action += '/mimeType/%s' % urlsafe_b64encode(_type)
data = {
'auth': token,
'action': action,
}
res = requests.post(url, files=files, data=data)
assert res.status_code == 200, res
return res.json()
if isinstance(filename, basestring):
# 单个文件
return _put(filename)
# 多文件
return [_put(f) for f in filename]
@expected_argument_type(2, (list, tuple))
def _cp_mv_handler(self, action, args):
"""copy move方法
action: 'copy' or 'move'
args: [src_bucket, src_filename, des_bucket, des_filename]
or [(src_bucket, src_filename, des_bucket, des_filename), (), ...]
args 第一种形式就是对一个文件进行操作,第二种形式是多个文件批量操作
用户不用直接调用这个方法
"""
if isinstance(args[0], basestring):
return self._cp_mv_single(action, args)
if isinstance(args[0], (list, tuple)):
return self._cp_mv_batch(action, args)
@expected_argument_type(3, (basestring, list, tuple))
def _stat_rm_handler(self, action, bucket, filename):
"""stat delete方法
action: 'stat' or 'delete'
bucket: 哪个bucket
filenmae: 'aabb' or ['aabb', 'ccdd', ...]
filename 第一种形式就是对一个文件进行操作,第二种形式是多个文件批量操作
用户不用直接调用这个方法
"""
if isinstance(filename, basestring):
return self._stat_rm_single(action, bucket, filename)
if isinstance(filename, (list, tuple)):
return self._stat_rm_batch(action, bucket, filename)
def _cp_mv_single(self, action, args):
src_bucket, src_filename, des_bucket, des_filename = args
url = '%s/%s/%s/%s' % (
RS_HOST,
action,
urlsafe_b64encode('%s:%s' % (src_bucket, src_filename)),
urlsafe_b64encode('%s:%s' % (des_bucket, des_filename)),
)
return self.api_call(url)
def _cp_mv_batch(self, action, args):
url = '%s/batch' % RS_HOST
def _one_param(arg):
return 'op=/%s/%s/%s' % (
action,
urlsafe_b64encode('%s:%s' % (arg[0], arg[1])),
urlsafe_b64encode('%s:%s' % (arg[2], arg[3])),
)
param = '&'.join( map(_one_param, args) )
return self.api_call(url, param)
def _stat_rm_single(self, action, bucket, filename):
url = '%s/%s/%s' % (
RS_HOST, action, urlsafe_b64encode('%s:%s' % (bucket, filename))
)
return self.api_call(url)
def _stat_rm_batch(self, action, bucket, filenames):
url = '%s/batch' % RS_HOST
param = [
'op=/%s/%s' % (
action, urlsafe_b64encode('%s:%s' % (bucket, f))
) for f in filenames
]
param = '&'.join(param)
return self.api_call(url, param)
def transform_argument(func):
@functools.wraps(func)
def deco(self, *args, **kwargs):
filename = args[0] if len(args) == 1 else args
return func(self, filename, **kwargs)
return deco
class Bucket(object):
def __init__(self, cow, bucket):
self.cow = cow
self.bucket = bucket
@transform_argument
def put(self, *args, **kwargs):
names = kwargs.get('names', None)
if names and not isinstance(names, dict):
raise TypeError(
"names Type error, Expected dict, But got Type of {0}".format(type(names))
)
return self.cow.put(self.bucket, args[0], names=names)
@transform_argument
def stat(self, *args):
return self.cow.stat(self.b | ucket, args[0])
@transform_argument
def delete(self, *args):
return self.cow.delete(self.bucket, args[0])
@transform_argument
def copy(self, *args):
return self.cow.copy(self._build_cp_mv_args(args[0]))
@transform_argument
def move(self, *args):
return self.cow.move(self._build_cp_mv_args(args[0]))
def list_files(self, marker=None, limit=None, prefix=None):
return self.cow.list_files(self.bucket, marker=marker, limit=limit, prefix=prefix)
def _build_cp_mv_args(self, filename):
if isinstance(filename[0], basestring):
args = [self.bucket, filename[0], self.bucket, filename[1]]
else:
args = []
for src, des in filename:
args.append( (self.bucket, src, self.bucket, des) )
return args
| identifier_body | |
sevencow.py | # -*- coding: utf-8 -*-
import os
import hmac
import time
import json
import mimetypes
import functools
from urlparse import urlparse
from urllib import urlencode
from base64 import urlsafe_b64encode
from hashlib import sha1
import requests
version_info = (0, 1, 2)
VERSION = __version__ = '.'.join( map(str, version_info) )
"""
Usage:
cow = Cow(ACCESS_KEY, SECRET_KEY)
b = cow.get_bucket(BUCKET)
b.put('a')
b.put('a', 'b')
b.put('a', names={'a': 'x'})
b.put('a', 'b', names={'a': 'x', 'b': 'y'})
b.stat('a')
b.stat('a', 'b')
b.delete('a')
b.delete('a', 'b')
b.copy('a', 'c')
b.copy(('a', 'c'), ('b', 'd'))
b.move('a', 'c')
b.move(('a', 'c'), ('b', 'd'))
"""
RS_HOST = 'http://rs.qbox.me'
UP_HOST = 'http://up.qbox.me'
RSF_HOST = 'http://rsf.qbox.me'
class CowException(Exception):
def __init__(self, url, status_code, reason, content):
self.url = url
self.status_code = status_code
self.reason = reason
self.content = content
Exception.__init__(self, '%s, %s' % (reason, content))
def signing(secret_key, data):
return urlsafe_b64encode(
hmac.new(secret_key, data, sha1).digest()
)
def requests_error_handler(func):
@functools.wraps(func)
def deco(*args, **kwargs):
try:
return func(*args, **kwargs)
except AssertionError as e:
req = e.args[0]
raise CowException(
req.url, req.status_code, req.reason, req.content
)
return deco
def expected_argument_type(pos, types):
def deco(func):
@functools.wraps(func)
def wrap(*args, **kwargs):
if not isinstance(args[pos], types):
raise TypeError(
"{0} Type error, Expected {1}".format(args[pos], types)
)
return func(*args, **kwargs)
return wrap
return deco
class UploadToken(object):
def __init__(self, access_key, secret_key, scope, ttl=3600):
self.access_key = access_key
self.secret_key = secret_key
self.scope = scope
self.ttl = ttl
self._token = None
self.generated = int(time.time())
@property
def token(self):
if int(time.time()) - self.generated > self.ttl - 60:
# 还有一分钟也认为过期了, make new token
self._token = self._make_token()
if not self._token:
self._token = self._make_token()
return self._token
def _make_token(self):
self.generated = int(time.time())
info = {
'scope': self.scope,
'deadline': self.generated + self.ttl
}
info = urlsafe_b64encode(json.dumps(info))
token = signing(self.secret_key, info)
return '%s:%s:%s' % (self.access_key, token, info)
class Cow(object):
def __init__(self, access_key, secret_key):
self.access_key = access_key
self.secret_key = secret_key
self.upload_tokens = {}
self.stat = functools.partial(self._stat_rm_handler, 'stat')
self.delete = functools.partial(self._stat_rm_handler, 'delete')
self.copy = functools.partial(self._cp_mv_handler, 'copy')
self.move = functools.partial(self._cp_mv_handler, 'move')
def get_bucket(self, bucket):
"""对一个bucket的文件进行操作,
推荐使用此方法得到一个bucket对象,
然后对此bucket的操作就只用传递文件名即可
"""
return Bucket(self, bucket)
def generate_access_token(self, url, params=None):
uri = urlparse(url)
token = uri.path
if uri.query:
token = '%s?%s' % (token, uri.query)
token = '%s\n' % token
if params:
if isinstance(params, basestring):
token += params
else:
token += urlencode(params)
return '%s:%s' % (self.access_key, signing(self.secret_key, token))
def build_requests_headers(self, token):
return {
'Content-Type': 'application/x-www-form-urlencoded',
'Authorization': 'QBox %s' % token
}
@requests_error_handler
def api_call(self, url, params=None):
token = self.generate_access_token(url, params=params)
if params:
res = requests.post(url, data=params, headers=self.build_requests_headers(token))
else:
res = requests.post(url, headers=self.build_requests_headers(token))
assert res.status_code == 200, res
return res.json() if res.text else ''
def list_buckets(self):
"""列出所有的buckets"""
url = '%s/buckets' % RS_HOST
return self.api_call(url)
def create_bucket(self, name):
"""不建议使用API建立bucket
测试发现API建立的bucket默认无法设置<bucket_name>.qiniudn.com的二级域名
请直接到web界面建立
"""
url = '%s/mkbucket/%s' % (RS_HOST, name)
return self.api_call(url)
def drop_bucket(self, bucket):
"""删除整个bucket"""
url = '%s/drop/%s' % (RS_HOST, bucket)
return self.api_call(url)
def list_files(self, bucket, marker=None, limit=None, prefix=None):
"""列出bucket中的文件"""
query = ['bucket=%s' % bucket]
if marker:
query.append('marker=%s' % marker)
if limit:
query.append('limit=%s' % limit)
if prefix:
query.append('prefix=%s' % prefix)
url = '%s/list?%s' % (RSF_HOST, '&'.join(query))
return self.api_call(url)
def generate_upload_token(self, scope, ttl=3600):
"""上传文件的uploadToken"""
if scope not in self.upload_tokens:
self.upload_tokens[scope] = UploadToken(self.access_key, self.secret_key, scope, ttl=ttl)
return self.upload_tokens[scope].token
@requests_error_handler
@expected_argument_type(2, (basestring, list, tuple))
def put(self, scope, filename, names=None):
"""上传文件
filename 如果是字符串,表示上传单个文件,
如果是list或者tuple,表示上传多个文件
names 是dict,key为filename, value为上传后的名字
如果不设置,默认为文件名
"""
u | = '%s/upload' % UP_HOST
token = self.generate_upload_token(scope)
names = names or {}
def _uploaded_name(filename):
return names.get(filename, None) or os.path.basename(filename)
def _put(filename):
files = {
'file': (filename, open(filename, 'rb')),
}
action = '/rs-put/%s' % urlsafe_b64encode(
'%s:%s' % (scope, _uploaded_name(filename))
)
_type, _encoding = mimetypes.guess_type(filename)
if _type:
action += '/mimeType/%s' % urlsafe_b64encode(_type)
data = {
'auth': token,
'action': action,
}
res = requests.post(url, files=files, data=data)
assert res.status_code == 200, res
return res.json()
if isinstance(filename, basestring):
# 单个文件
return _put(filename)
# 多文件
return [_put(f) for f in filename]
@expected_argument_type(2, (list, tuple))
def _cp_mv_handler(self, action, args):
"""copy move方法
action: 'copy' or 'move'
args: [src_bucket, src_filename, des_bucket, des_filename]
or [(src_bucket, src_filename, des_bucket, des_filename), (), ...]
args 第一种形式就是对一个文件进行操作,第二种形式是多个文件批量操作
用户不用直接调用这个方法
"""
if isinstance(args[0], basestring):
return self._cp_mv_single(action, args)
if isinstance(args[0], (list, tuple)):
return self._cp_mv_batch(action, args)
@expected_argument_type(3, (basestring, list, tuple))
def _stat_rm_handler(self, action, bucket, filename):
"""stat delete方法
action: 'stat' or 'delete'
bucket: 哪个bucket
filenmae: 'aabb' or ['aabb', 'ccdd', ...]
filename 第一种形式就是对一个文件进行操作,第二种形式是多个文件批量操作
用户不用直接调用这个方法
"""
if isinstance(filename, basestring):
return self._stat_rm_single(action, bucket, filename)
if isinstance(filename, (list, tuple)):
return self._stat_rm_batch(action, bucket, filename)
def _cp_mv_single(self, action, args):
src_bucket, src_filename, des_bucket, des_filename = args
url = '%s/%s/%s/%s' % (
RS_HOST,
action,
urlsafe_b64encode('%s:%s' % (src_bucket, src_filename)),
urlsafe_b64encode('%s:%s' % (des_bucket, des_filename)),
)
return self.api_call(url)
def _cp_mv_batch(self, action, args):
url = '%s/batch' % RS_HOST
def _one_param(arg):
return 'op=/%s/%s/%s' % (
action,
urlsafe_b64encode('%s:%s' % (arg[0], arg[1])),
urlsafe_b64encode('%s:%s' % (arg[2], arg[3])),
)
param = '&'.join( map(_one_param, args) )
return self.api_call(url, param)
def _stat_rm_single(self, action, bucket, filename):
url = '%s/%s/%s' % (
RS_HOST, action, urlsafe_b64encode('%s:%s' % (bucket, filename))
)
return self.api_call(url)
def _stat_rm_batch(self, action, bucket, filenames):
url = '%s/batch' % RS_HOST
param = [
'op=/%s/%s' % (
action, urlsafe_b64encode('%s:%s' % (bucket, f))
) for f in filenames
]
param = '&'.join(param)
return self.api_call(url, param)
def transform_argument(func):
@functools.wraps(func)
def deco(self, *args, **kwargs):
filename = args[0] if len(args) == 1 else args
return func(self, filename, **kwargs)
return deco
class Bucket(object):
def __init__(self, cow, bucket):
self.cow = cow
self.bucket = bucket
@transform_argument
def put(self, *args, **kwargs):
names = kwargs.get('names', None)
if names and not isinstance(names, dict):
raise TypeError(
"names Type error, Expected dict, But got Type of {0}".format(type(names))
)
return self.cow.put(self.bucket, args[0], names=names)
@transform_argument
def stat(self, *args):
return self.cow.stat(self.bucket, args[0])
@transform_argument
def delete(self, *args):
return self.cow.delete(self.bucket, args[0])
@transform_argument
def copy(self, *args):
return self.cow.copy(self._build_cp_mv_args(args[0]))
@transform_argument
def move(self, *args):
return self.cow.move(self._build_cp_mv_args(args[0]))
def list_files(self, marker=None, limit=None, prefix=None):
return self.cow.list_files(self.bucket, marker=marker, limit=limit, prefix=prefix)
def _build_cp_mv_args(self, filename):
if isinstance(filename[0], basestring):
args = [self.bucket, filename[0], self.bucket, filename[1]]
else:
args = []
for src, des in filename:
args.append( (self.bucket, src, self.bucket, des) )
return args
| rl | identifier_name |
class_vertice_graph.py | """ The goal of this file is to have all the information on a graph. """
import sys, os, inspect
currentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))
parentdir = os.path.dirname(currentdir)
sys.path.insert(0, parentdir)
import numpy as np
import matplotlib.pyplot as plt
import copy
import exceptions
inf = np.inf
#Change this bool according to the situation
non_oriented = False
class Vertice:
"""" All the information of one vertice are contained here. """
def __init__(self, index, coordinates):
""" Entry : index of the vertice in list_of_vertices
and the coordinates of the vertice. """
self.index = index
self.coordinates = np.array([coordinates[0],coordinates[1]])
self._edges_list = [] # no neighbour by default
self.priority = inf # priority by default
self.visited = False # vertice is by default not visited
self.cost_dijkstra = inf # cost is by default inf
self.antecedent = -inf # antecedent not defined before using Dijkstra
#implemente apres passage par panda
self.index_edges_list = []#seulement implemente apres passage par panda
#database
self.id = None #identifiant idfm de la gare
self.gare_name = None #nom de la gare
self.color = None #couleur de la gare
self.is_a_station= True # boolean True, si le noeud est veritablement une gare. False sinon
def get_lines_connected(self):
list_of_line = []
for edge in self._edges_list:
if edge.id not in list_of_line:
list_of_line.append(edge.id)
return list_of_line
@property
def edges_list(self):
""" Returns the list of neighbour. """
return self._edges_list
# We suppose that the index and the coordinates never change.
# The other properties can.
@edges_list.setter
def | (self, edges_list):
""" An element of edges_list is an edge """
for e in edges_list:
exceptions.check_pertinent_edge(self, e)
self._edges_list = edges_list
def neighbours_list(self, list_tuple, id=0):
self._edges_list.clear()
"""interface with old constructor , tuple=(neighbour_vertice,cost) is an element of list_tuple """
for tuple in list_tuple:
E = Edge(self, tuple[0], id, tuple[1])
self._edges_list.append(E)
def number_of_neighbours(self):
return len(self._edges_list)
def is_linked(self, other):
"""returns True if there is an edge between self and other"""
for edge in self._edges_list:
if other.index == edge.linked[1].index:
return True
return False
def push_edge(self, edge, coords_verif=False):
if coords_verif:
exceptions.check_pertinent_edge_coords_verif(self, edge)
else:
exceptions.check_pertinent_edge(self, edge)
self._edges_list.append(edge)
"""
def cost_between(self, other):
for edge in self._edges_list:
[vertice, vertice_voisin] = edge.linked
if vertice_voisin == other:
return edge.given_cost"""
def __repr__(self):
return f"Vertice {str(self.index)}"
def __lt__(self, other):
return self.priority < other.priority
class Edge:
def __init__(self, vertice1, vertice2, id, given_cost=0):
self.linked = [vertice1,vertice2]
self.id = id #identifiant de la liaison. ici id=nom de la ligne a laqualle appartient la liaison
self._given_cost = given_cost #cout de deplacement de la liason donne par l'utilisateur ou la base de donnee
#data_base
self.color=None #couleur de la liason
self.connection_with_displayable=None #indice de la liason developpee( trace reel) dans la table de connection connection_table_edge_and_diplayable_edge de la classe graph
self.index=None
def set_given_cost(self,cost):
self._given_cost=cost
#ne pas mettre @property ici, on veut une methode pas un attribut
def euclidian_cost(self):
return np.sqrt(self.square_euclidian_cost())
def square_euclidian_cost(self):
return np.dot(np.transpose(self.linked[0].coordinates-self.linked[1].coordinates),(self.linked[0].coordinates-self.linked[1].coordinates))
def customized_cost1(self):
V_metro = 25.1 / 3.6 #vitesse moyenne en km/h /3.6 -> vitesse moyenne en m/s
V_train = 49.6 / 3.6
V_tram = 18 / 3.6
V_pieton = 4 / 3.6
if self.id in ["A","B","C","D","E","H","I","J","K","L","M","N","P","R","U","TER","GL"]:
return self._given_cost/V_train
if self.id in [str(i) for i in range(1,15)]+["ORL","CDG","3b","7b"]:
return self._given_cost/V_metro
if self.id in ["T"+str(i) for i in range(1,12)]+["T3A","T3B","FUN"]:
return self._given_cost/V_tram
if self.id in ["RER Walk"]:
return self._given_cost/V_pieton
raise ValueError(" Dans customized_cost1 " +self.id+" non pris en compte dans le calcul de distance")
def __eq__(self,other):
"""2 edges are equal iff same cordinates and same id """
boul0 = self.linked[0].coordinates[0]==other.linked[0].coordinates[0] and self.linked[0].coordinates[1]==other.linked[0].coordinates[1]
boul1 = self.linked[1].coordinates[0]==other.linked[1].coordinates[0] and self.linked[1].coordinates[1]==other.linked[1].coordinates[1]
boulid = self.id==other.id
return boul0 and boul1 and boulid
def __ne__(self,other):
"""2 edges are not equal iff they are not equal :) """
return (self==other)==False
#ne pas mettre @property ici, on veut une methode pas un attribut
def given_cost(self):
return self._given_cost
def __repr__(self):
return f"Edge [{str(self.linked[0].index)}, {str(self.linked[1].index)}] !oriented!"
class Graph:
""" All the information of a graph are contained here. """
def __init__(self,list_of_vertices):
""" Entry : the list of vertices. """
self.list_of_vertices = list_of_vertices
self.number_of_vertices = len(list_of_vertices)
self.connection_table_edge_and_diplayable_edge=[]
self.list_of_edges=[]
self.number_of_disp_edges=0
self.number_of_edges=0
def push_diplayable_edge(self,bidim_array):
self.connection_table_edge_and_diplayable_edge.append(copy.deepcopy(bidim_array))
self.number_of_disp_edges+=1
def push_edge(self,e):
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_edge_without_doublons(self, e):
if e not in self.list_of_edges:
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_vertice(self,vertice):
self.list_of_vertices.append(vertice)
self.number_of_vertices += 1
def push_vertice_without_doublons(self, vertice):
bool,index = self.is_vertice_in_graph_based_on_xy_with_tolerance(vertice,10**(-8))
#bool,index = self.is_vertice_in_graph_based_on_xy(vertice)
if bool == False:
self.push_vertice(vertice)
else:
vertice.coordinates=self.list_of_vertices[index].coordinates
for edge in vertice.edges_list:
if edge not in self.list_of_vertices[index].edges_list:
self.list_of_vertices[index].push_edge(edge,True)
def is_vertice_in_graph_based_on_xy(self,vertice):
for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if v.coordinates[0] == vertice.coordinates[0] and v.coordinates[1] == vertice.coordinates[1]:
return True,i
return False,None
def is_vertice_in_graph_based_on_xy_with_tolerance(self, vertice, epsilon):
for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if ((v.coordinates[0] - vertice.coordinates[0])**2) + ((v.coordinates[1] - vertice.coordinates[1])**2) < epsilon:
return True, i
return False, None
def __getitem__(self, key):#implement instance[key]
if key >= 0 and key < self.number_of_vertices:
return self.list_of_vertices[key]
else :
raise IndexError
def laplace_matrix(self):
""" Returns the laplace matrix. """
n = self.number_of_vertices
laplace_matrix = np.zeros((n, n))
for i in range(n):
laplace_matrix[i][i] = 1
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
laplace_matrix[i][edge.linked[1].index] = 1
return laplace_matrix
def A_matrix(self,type_cost=Edge.given_cost):
""" Returns the laplace matrix. """
n = self.number_of_vertices
A_matrix = np.zeros((n, n))
for i in range(n):
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
cost = type_cost(edge)
A_matrix[i][edge.linked[1].index] = cost
A_matrix[edge.linked[1].index][i] = cost
return A_matrix
def pairs_of_vertices(self):
"""Returns the pairs of connected vertices.
Beware ! There might be non-connected vertices in the graph. """
pairs_of_vertices = []
for vertice in self.list_of_vertices:
for edge in vertice.edges_list:
if non_oriented:
if (vertice, edge.linked[1]) and (edge.linked[1], vertice) not in pairs_of_vertices:
pairs_of_vertices.append((vertice, edge.linked[1]))
if not non_oriented:
if (vertice, edge.linked[1]) not in pairs_of_vertices:
pairs_of_vertices.append((vertice, edge.linked[1]))
return pairs_of_vertices
def number_of_edges(self):
a=self.pairs_of_vertices()
assert self.number_of_edges==len(a), "problem in Graph.pairs_of_vertices"
return self.number_of_edges
def search_index_by_coordinates(self,coord):
"""search the index of vertice at coordinates: """
for i in range(len(self.list_of_vertices)):
if self[i].coordinates[0]==coord[0] and self[i].coordinates[1]==coord[1]:
return i
def set_right_edges(self):
"""verify that the graph is coherent """
for v in self:
for e in v.edges_list:
e.linked[0]=v
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
for e in self.list_of_edges:
e.linked[0]=self[self.search_index_by_coordinates(e.linked[0].coordinates)]
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
def plot(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
x = e.linked[0].coordinates[0]
y = e.linked[0].coordinates[1]
dx = e.linked[1].coordinates[0] - x
dy = e.linked[1].coordinates[1] - y
plt.plot([x,x+dx], [y,y+dy], color=c)
# plt.arrow(x,y,dx,dy)
plt.axis = 'off'
plt.show()
def plot_dev(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
for i in range(len(self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable])-1):
x = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][0]
y = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][1]
dx = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][0]-x
dy = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][1]-y
plt.plot([x,x+dx], [y,y+dy], color=c)
plt.axis = 'off'
plt.show()
| edges_list | identifier_name |
class_vertice_graph.py | """ The goal of this file is to have all the information on a graph. """
import sys, os, inspect
currentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))
parentdir = os.path.dirname(currentdir)
sys.path.insert(0, parentdir)
import numpy as np
import matplotlib.pyplot as plt
import copy
import exceptions
inf = np.inf
#Change this bool according to the situation
non_oriented = False
class Vertice:
"""" All the information of one vertice are contained here. """
def __init__(self, index, coordinates):
""" Entry : index of the vertice in list_of_vertices
and the coordinates of the vertice. """
self.index = index
self.coordinates = np.array([coordinates[0],coordinates[1]])
self._edges_list = [] # no neighbour by default
self.priority = inf # priority by default
self.visited = False # vertice is by default not visited
self.cost_dijkstra = inf # cost is by default inf
self.antecedent = -inf # antecedent not defined before using Dijkstra
#implemente apres passage par panda
self.index_edges_list = []#seulement implemente apres passage par panda
#database
self.id = None #identifiant idfm de la gare
self.gare_name = None #nom de la gare
self.color = None #couleur de la gare
self.is_a_station= True # boolean True, si le noeud est veritablement une gare. False sinon
def get_lines_connected(self):
list_of_line = []
for edge in self._edges_list:
if edge.id not in list_of_line:
list_of_line.append(edge.id)
return list_of_line
@property
def edges_list(self):
""" Returns the list of neighbour. """
return self._edges_list
# We suppose that the index and the coordinates never change.
# The other properties can.
@edges_list.setter
def edges_list(self, edges_list):
""" An element of edges_list is an edge """
for e in edges_list:
exceptions.check_pertinent_edge(self, e)
self._edges_list = edges_list
def neighbours_list(self, list_tuple, id=0):
self._edges_list.clear()
"""interface with old constructor , tuple=(neighbour_vertice,cost) is an element of list_tuple """
for tuple in list_tuple:
E = Edge(self, tuple[0], id, tuple[1])
self._edges_list.append(E)
def number_of_neighbours(self):
return len(self._edges_list)
def is_linked(self, other):
"""returns True if there is an edge between self and other"""
for edge in self._edges_list:
if other.index == edge.linked[1].index:
return True
return False
def push_edge(self, edge, coords_verif=False):
if coords_verif:
exceptions.check_pertinent_edge_coords_verif(self, edge)
else:
exceptions.check_pertinent_edge(self, edge)
self._edges_list.append(edge)
"""
def cost_between(self, other):
for edge in self._edges_list:
[vertice, vertice_voisin] = edge.linked
if vertice_voisin == other:
return edge.given_cost"""
def __repr__(self):
return f"Vertice {str(self.index)}"
def __lt__(self, other):
return self.priority < other.priority
class Edge:
def __init__(self, vertice1, vertice2, id, given_cost=0):
self.linked = [vertice1,vertice2]
self.id = id #identifiant de la liaison. ici id=nom de la ligne a laqualle appartient la liaison
self._given_cost = given_cost #cout de deplacement de la liason donne par l'utilisateur ou la base de donnee
#data_base
self.color=None #couleur de la liason
self.connection_with_displayable=None #indice de la liason developpee( trace reel) dans la table de connection connection_table_edge_and_diplayable_edge de la classe graph
self.index=None
def set_given_cost(self,cost):
self._given_cost=cost
#ne pas mettre @property ici, on veut une methode pas un attribut
def euclidian_cost(self):
return np.sqrt(self.square_euclidian_cost())
def square_euclidian_cost(self):
return np.dot(np.transpose(self.linked[0].coordinates-self.linked[1].coordinates),(self.linked[0].coordinates-self.linked[1].coordinates))
def customized_cost1(self):
V_metro = 25.1 / 3.6 #vitesse moyenne en km/h /3.6 -> vitesse moyenne en m/s
V_train = 49.6 / 3.6
V_tram = 18 / 3.6
V_pieton = 4 / 3.6
if self.id in ["A","B","C","D","E","H","I","J","K","L","M","N","P","R","U","TER","GL"]:
return self._given_cost/V_train
if self.id in [str(i) for i in range(1,15)]+["ORL","CDG","3b","7b"]:
return self._given_cost/V_metro
if self.id in ["T"+str(i) for i in range(1,12)]+["T3A","T3B","FUN"]:
return self._given_cost/V_tram
if self.id in ["RER Walk"]:
return self._given_cost/V_pieton
raise ValueError(" Dans customized_cost1 " +self.id+" non pris en compte dans le calcul de distance")
def __eq__(self,other):
"""2 edges are equal iff same cordinates and same id """
boul0 = self.linked[0].coordinates[0]==other.linked[0].coordinates[0] and self.linked[0].coordinates[1]==other.linked[0].coordinates[1]
boul1 = self.linked[1].coordinates[0]==other.linked[1].coordinates[0] and self.linked[1].coordinates[1]==other.linked[1].coordinates[1]
boulid = self.id==other.id
return boul0 and boul1 and boulid
def __ne__(self,other):
"""2 edges are not equal iff they are not equal :) """
return (self==other)==False
#ne pas mettre @property ici, on veut une methode pas un attribut
def given_cost(self):
return self._given_cost
def __repr__(self):
return f"Edge [{str(self.linked[0].index)}, {str(self.linked[1].index)}] !oriented!"
class Graph:
""" All the information of a graph are contained here. """
def __init__(self,list_of_vertices):
""" Entry : the list of vertices. """
self.list_of_vertices = list_of_vertices | self.number_of_edges=0
def push_diplayable_edge(self,bidim_array):
self.connection_table_edge_and_diplayable_edge.append(copy.deepcopy(bidim_array))
self.number_of_disp_edges+=1
def push_edge(self,e):
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_edge_without_doublons(self, e):
if e not in self.list_of_edges:
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_vertice(self,vertice):
self.list_of_vertices.append(vertice)
self.number_of_vertices += 1
def push_vertice_without_doublons(self, vertice):
bool,index = self.is_vertice_in_graph_based_on_xy_with_tolerance(vertice,10**(-8))
#bool,index = self.is_vertice_in_graph_based_on_xy(vertice)
if bool == False:
self.push_vertice(vertice)
else:
vertice.coordinates=self.list_of_vertices[index].coordinates
for edge in vertice.edges_list:
if edge not in self.list_of_vertices[index].edges_list:
self.list_of_vertices[index].push_edge(edge,True)
def is_vertice_in_graph_based_on_xy(self,vertice):
for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if v.coordinates[0] == vertice.coordinates[0] and v.coordinates[1] == vertice.coordinates[1]:
return True,i
return False,None
def is_vertice_in_graph_based_on_xy_with_tolerance(self, vertice, epsilon):
for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if ((v.coordinates[0] - vertice.coordinates[0])**2) + ((v.coordinates[1] - vertice.coordinates[1])**2) < epsilon:
return True, i
return False, None
def __getitem__(self, key):#implement instance[key]
if key >= 0 and key < self.number_of_vertices:
return self.list_of_vertices[key]
else :
raise IndexError
def laplace_matrix(self):
""" Returns the laplace matrix. """
n = self.number_of_vertices
laplace_matrix = np.zeros((n, n))
for i in range(n):
laplace_matrix[i][i] = 1
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
laplace_matrix[i][edge.linked[1].index] = 1
return laplace_matrix
def A_matrix(self,type_cost=Edge.given_cost):
""" Returns the laplace matrix. """
n = self.number_of_vertices
A_matrix = np.zeros((n, n))
for i in range(n):
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
cost = type_cost(edge)
A_matrix[i][edge.linked[1].index] = cost
A_matrix[edge.linked[1].index][i] = cost
return A_matrix
def pairs_of_vertices(self):
"""Returns the pairs of connected vertices.
Beware ! There might be non-connected vertices in the graph. """
pairs_of_vertices = []
for vertice in self.list_of_vertices:
for edge in vertice.edges_list:
if non_oriented:
if (vertice, edge.linked[1]) and (edge.linked[1], vertice) not in pairs_of_vertices:
pairs_of_vertices.append((vertice, edge.linked[1]))
if not non_oriented:
if (vertice, edge.linked[1]) not in pairs_of_vertices:
pairs_of_vertices.append((vertice, edge.linked[1]))
return pairs_of_vertices
def number_of_edges(self):
a=self.pairs_of_vertices()
assert self.number_of_edges==len(a), "problem in Graph.pairs_of_vertices"
return self.number_of_edges
def search_index_by_coordinates(self,coord):
"""search the index of vertice at coordinates: """
for i in range(len(self.list_of_vertices)):
if self[i].coordinates[0]==coord[0] and self[i].coordinates[1]==coord[1]:
return i
def set_right_edges(self):
"""verify that the graph is coherent """
for v in self:
for e in v.edges_list:
e.linked[0]=v
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
for e in self.list_of_edges:
e.linked[0]=self[self.search_index_by_coordinates(e.linked[0].coordinates)]
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
def plot(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
x = e.linked[0].coordinates[0]
y = e.linked[0].coordinates[1]
dx = e.linked[1].coordinates[0] - x
dy = e.linked[1].coordinates[1] - y
plt.plot([x,x+dx], [y,y+dy], color=c)
# plt.arrow(x,y,dx,dy)
plt.axis = 'off'
plt.show()
def plot_dev(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
for i in range(len(self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable])-1):
x = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][0]
y = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][1]
dx = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][0]-x
dy = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][1]-y
plt.plot([x,x+dx], [y,y+dy], color=c)
plt.axis = 'off'
plt.show() | self.number_of_vertices = len(list_of_vertices)
self.connection_table_edge_and_diplayable_edge=[]
self.list_of_edges=[]
self.number_of_disp_edges=0 | random_line_split |
class_vertice_graph.py | """ The goal of this file is to have all the information on a graph. """
import sys, os, inspect
currentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))
parentdir = os.path.dirname(currentdir)
sys.path.insert(0, parentdir)
import numpy as np
import matplotlib.pyplot as plt
import copy
import exceptions
inf = np.inf
#Change this bool according to the situation
non_oriented = False
class Vertice:
"""" All the information of one vertice are contained here. """
def __init__(self, index, coordinates):
""" Entry : index of the vertice in list_of_vertices
and the coordinates of the vertice. """
self.index = index
self.coordinates = np.array([coordinates[0],coordinates[1]])
self._edges_list = [] # no neighbour by default
self.priority = inf # priority by default
self.visited = False # vertice is by default not visited
self.cost_dijkstra = inf # cost is by default inf
self.antecedent = -inf # antecedent not defined before using Dijkstra
#implemente apres passage par panda
self.index_edges_list = []#seulement implemente apres passage par panda
#database
self.id = None #identifiant idfm de la gare
self.gare_name = None #nom de la gare
self.color = None #couleur de la gare
self.is_a_station= True # boolean True, si le noeud est veritablement une gare. False sinon
def get_lines_connected(self):
list_of_line = []
for edge in self._edges_list:
if edge.id not in list_of_line:
list_of_line.append(edge.id)
return list_of_line
@property
def edges_list(self):
""" Returns the list of neighbour. """
return self._edges_list
# We suppose that the index and the coordinates never change.
# The other properties can.
@edges_list.setter
def edges_list(self, edges_list):
""" An element of edges_list is an edge """
for e in edges_list:
exceptions.check_pertinent_edge(self, e)
self._edges_list = edges_list
def neighbours_list(self, list_tuple, id=0):
self._edges_list.clear()
"""interface with old constructor , tuple=(neighbour_vertice,cost) is an element of list_tuple """
for tuple in list_tuple:
E = Edge(self, tuple[0], id, tuple[1])
self._edges_list.append(E)
def number_of_neighbours(self):
return len(self._edges_list)
def is_linked(self, other):
"""returns True if there is an edge between self and other"""
for edge in self._edges_list:
if other.index == edge.linked[1].index:
return True
return False
def push_edge(self, edge, coords_verif=False):
if coords_verif:
exceptions.check_pertinent_edge_coords_verif(self, edge)
else:
exceptions.check_pertinent_edge(self, edge)
self._edges_list.append(edge)
"""
def cost_between(self, other):
for edge in self._edges_list:
[vertice, vertice_voisin] = edge.linked
if vertice_voisin == other:
return edge.given_cost"""
def __repr__(self):
return f"Vertice {str(self.index)}"
def __lt__(self, other):
return self.priority < other.priority
class Edge:
def __init__(self, vertice1, vertice2, id, given_cost=0):
self.linked = [vertice1,vertice2]
self.id = id #identifiant de la liaison. ici id=nom de la ligne a laqualle appartient la liaison
self._given_cost = given_cost #cout de deplacement de la liason donne par l'utilisateur ou la base de donnee
#data_base
self.color=None #couleur de la liason
self.connection_with_displayable=None #indice de la liason developpee( trace reel) dans la table de connection connection_table_edge_and_diplayable_edge de la classe graph
self.index=None
def set_given_cost(self,cost):
self._given_cost=cost
#ne pas mettre @property ici, on veut une methode pas un attribut
def euclidian_cost(self):
return np.sqrt(self.square_euclidian_cost())
def square_euclidian_cost(self):
return np.dot(np.transpose(self.linked[0].coordinates-self.linked[1].coordinates),(self.linked[0].coordinates-self.linked[1].coordinates))
def customized_cost1(self):
V_metro = 25.1 / 3.6 #vitesse moyenne en km/h /3.6 -> vitesse moyenne en m/s
V_train = 49.6 / 3.6
V_tram = 18 / 3.6
V_pieton = 4 / 3.6
if self.id in ["A","B","C","D","E","H","I","J","K","L","M","N","P","R","U","TER","GL"]:
return self._given_cost/V_train
if self.id in [str(i) for i in range(1,15)]+["ORL","CDG","3b","7b"]:
return self._given_cost/V_metro
if self.id in ["T"+str(i) for i in range(1,12)]+["T3A","T3B","FUN"]:
return self._given_cost/V_tram
if self.id in ["RER Walk"]:
return self._given_cost/V_pieton
raise ValueError(" Dans customized_cost1 " +self.id+" non pris en compte dans le calcul de distance")
def __eq__(self,other):
"""2 edges are equal iff same cordinates and same id """
boul0 = self.linked[0].coordinates[0]==other.linked[0].coordinates[0] and self.linked[0].coordinates[1]==other.linked[0].coordinates[1]
boul1 = self.linked[1].coordinates[0]==other.linked[1].coordinates[0] and self.linked[1].coordinates[1]==other.linked[1].coordinates[1]
boulid = self.id==other.id
return boul0 and boul1 and boulid
def __ne__(self,other):
"""2 edges are not equal iff they are not equal :) """
return (self==other)==False
#ne pas mettre @property ici, on veut une methode pas un attribut
def given_cost(self):
return self._given_cost
def __repr__(self):
return f"Edge [{str(self.linked[0].index)}, {str(self.linked[1].index)}] !oriented!"
class Graph:
""" All the information of a graph are contained here. """
def __init__(self,list_of_vertices):
""" Entry : the list of vertices. """
self.list_of_vertices = list_of_vertices
self.number_of_vertices = len(list_of_vertices)
self.connection_table_edge_and_diplayable_edge=[]
self.list_of_edges=[]
self.number_of_disp_edges=0
self.number_of_edges=0
def push_diplayable_edge(self,bidim_array):
self.connection_table_edge_and_diplayable_edge.append(copy.deepcopy(bidim_array))
self.number_of_disp_edges+=1
def push_edge(self,e):
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_edge_without_doublons(self, e):
if e not in self.list_of_edges:
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_vertice(self,vertice):
self.list_of_vertices.append(vertice)
self.number_of_vertices += 1
def push_vertice_without_doublons(self, vertice):
bool,index = self.is_vertice_in_graph_based_on_xy_with_tolerance(vertice,10**(-8))
#bool,index = self.is_vertice_in_graph_based_on_xy(vertice)
if bool == False:
self.push_vertice(vertice)
else:
vertice.coordinates=self.list_of_vertices[index].coordinates
for edge in vertice.edges_list:
if edge not in self.list_of_vertices[index].edges_list:
self.list_of_vertices[index].push_edge(edge,True)
def is_vertice_in_graph_based_on_xy(self,vertice):
for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if v.coordinates[0] == vertice.coordinates[0] and v.coordinates[1] == vertice.coordinates[1]:
return True,i
return False,None
def is_vertice_in_graph_based_on_xy_with_tolerance(self, vertice, epsilon):
for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if ((v.coordinates[0] - vertice.coordinates[0])**2) + ((v.coordinates[1] - vertice.coordinates[1])**2) < epsilon:
return True, i
return False, None
def __getitem__(self, key):#implement instance[key]
if key >= 0 and key < self.number_of_vertices:
return self.list_of_vertices[key]
else :
raise IndexError
def laplace_matrix(self):
""" Returns the laplace matrix. """
n = self.number_of_vertices
laplace_matrix = np.zeros((n, n))
for i in range(n):
laplace_matrix[i][i] = 1
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
laplace_matrix[i][edge.linked[1].index] = 1
return laplace_matrix
def A_matrix(self,type_cost=Edge.given_cost):
""" Returns the laplace matrix. """
n = self.number_of_vertices
A_matrix = np.zeros((n, n))
for i in range(n):
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
cost = type_cost(edge)
A_matrix[i][edge.linked[1].index] = cost
A_matrix[edge.linked[1].index][i] = cost
return A_matrix
def pairs_of_vertices(self):
"""Returns the pairs of connected vertices.
Beware ! There might be non-connected vertices in the graph. """
pairs_of_vertices = []
for vertice in self.list_of_vertices:
for edge in vertice.edges_list:
if non_oriented:
if (vertice, edge.linked[1]) and (edge.linked[1], vertice) not in pairs_of_vertices:
|
if not non_oriented:
if (vertice, edge.linked[1]) not in pairs_of_vertices:
pairs_of_vertices.append((vertice, edge.linked[1]))
return pairs_of_vertices
def number_of_edges(self):
a=self.pairs_of_vertices()
assert self.number_of_edges==len(a), "problem in Graph.pairs_of_vertices"
return self.number_of_edges
def search_index_by_coordinates(self,coord):
"""search the index of vertice at coordinates: """
for i in range(len(self.list_of_vertices)):
if self[i].coordinates[0]==coord[0] and self[i].coordinates[1]==coord[1]:
return i
def set_right_edges(self):
"""verify that the graph is coherent """
for v in self:
for e in v.edges_list:
e.linked[0]=v
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
for e in self.list_of_edges:
e.linked[0]=self[self.search_index_by_coordinates(e.linked[0].coordinates)]
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
def plot(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
x = e.linked[0].coordinates[0]
y = e.linked[0].coordinates[1]
dx = e.linked[1].coordinates[0] - x
dy = e.linked[1].coordinates[1] - y
plt.plot([x,x+dx], [y,y+dy], color=c)
# plt.arrow(x,y,dx,dy)
plt.axis = 'off'
plt.show()
def plot_dev(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
for i in range(len(self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable])-1):
x = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][0]
y = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][1]
dx = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][0]-x
dy = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][1]-y
plt.plot([x,x+dx], [y,y+dy], color=c)
plt.axis = 'off'
plt.show()
| pairs_of_vertices.append((vertice, edge.linked[1])) | conditional_block |
class_vertice_graph.py | """ The goal of this file is to have all the information on a graph. """
import sys, os, inspect
currentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))
parentdir = os.path.dirname(currentdir)
sys.path.insert(0, parentdir)
import numpy as np
import matplotlib.pyplot as plt
import copy
import exceptions
inf = np.inf
#Change this bool according to the situation
non_oriented = False
class Vertice:
"""" All the information of one vertice are contained here. """
def __init__(self, index, coordinates):
""" Entry : index of the vertice in list_of_vertices
and the coordinates of the vertice. """
self.index = index
self.coordinates = np.array([coordinates[0],coordinates[1]])
self._edges_list = [] # no neighbour by default
self.priority = inf # priority by default
self.visited = False # vertice is by default not visited
self.cost_dijkstra = inf # cost is by default inf
self.antecedent = -inf # antecedent not defined before using Dijkstra
#implemente apres passage par panda
self.index_edges_list = []#seulement implemente apres passage par panda
#database
self.id = None #identifiant idfm de la gare
self.gare_name = None #nom de la gare
self.color = None #couleur de la gare
self.is_a_station= True # boolean True, si le noeud est veritablement une gare. False sinon
def get_lines_connected(self):
list_of_line = []
for edge in self._edges_list:
if edge.id not in list_of_line:
list_of_line.append(edge.id)
return list_of_line
@property
def edges_list(self):
""" Returns the list of neighbour. """
return self._edges_list
# We suppose that the index and the coordinates never change.
# The other properties can.
@edges_list.setter
def edges_list(self, edges_list):
""" An element of edges_list is an edge """
for e in edges_list:
exceptions.check_pertinent_edge(self, e)
self._edges_list = edges_list
def neighbours_list(self, list_tuple, id=0):
self._edges_list.clear()
"""interface with old constructor , tuple=(neighbour_vertice,cost) is an element of list_tuple """
for tuple in list_tuple:
E = Edge(self, tuple[0], id, tuple[1])
self._edges_list.append(E)
def number_of_neighbours(self):
return len(self._edges_list)
def is_linked(self, other):
"""returns True if there is an edge between self and other"""
for edge in self._edges_list:
if other.index == edge.linked[1].index:
return True
return False
def push_edge(self, edge, coords_verif=False):
if coords_verif:
exceptions.check_pertinent_edge_coords_verif(self, edge)
else:
exceptions.check_pertinent_edge(self, edge)
self._edges_list.append(edge)
"""
def cost_between(self, other):
for edge in self._edges_list:
[vertice, vertice_voisin] = edge.linked
if vertice_voisin == other:
return edge.given_cost"""
def __repr__(self):
return f"Vertice {str(self.index)}"
def __lt__(self, other):
return self.priority < other.priority
class Edge:
def __init__(self, vertice1, vertice2, id, given_cost=0):
self.linked = [vertice1,vertice2]
self.id = id #identifiant de la liaison. ici id=nom de la ligne a laqualle appartient la liaison
self._given_cost = given_cost #cout de deplacement de la liason donne par l'utilisateur ou la base de donnee
#data_base
self.color=None #couleur de la liason
self.connection_with_displayable=None #indice de la liason developpee( trace reel) dans la table de connection connection_table_edge_and_diplayable_edge de la classe graph
self.index=None
def set_given_cost(self,cost):
self._given_cost=cost
#ne pas mettre @property ici, on veut une methode pas un attribut
def euclidian_cost(self):
return np.sqrt(self.square_euclidian_cost())
def square_euclidian_cost(self):
return np.dot(np.transpose(self.linked[0].coordinates-self.linked[1].coordinates),(self.linked[0].coordinates-self.linked[1].coordinates))
def customized_cost1(self):
V_metro = 25.1 / 3.6 #vitesse moyenne en km/h /3.6 -> vitesse moyenne en m/s
V_train = 49.6 / 3.6
V_tram = 18 / 3.6
V_pieton = 4 / 3.6
if self.id in ["A","B","C","D","E","H","I","J","K","L","M","N","P","R","U","TER","GL"]:
return self._given_cost/V_train
if self.id in [str(i) for i in range(1,15)]+["ORL","CDG","3b","7b"]:
return self._given_cost/V_metro
if self.id in ["T"+str(i) for i in range(1,12)]+["T3A","T3B","FUN"]:
return self._given_cost/V_tram
if self.id in ["RER Walk"]:
return self._given_cost/V_pieton
raise ValueError(" Dans customized_cost1 " +self.id+" non pris en compte dans le calcul de distance")
def __eq__(self,other):
"""2 edges are equal iff same cordinates and same id """
boul0 = self.linked[0].coordinates[0]==other.linked[0].coordinates[0] and self.linked[0].coordinates[1]==other.linked[0].coordinates[1]
boul1 = self.linked[1].coordinates[0]==other.linked[1].coordinates[0] and self.linked[1].coordinates[1]==other.linked[1].coordinates[1]
boulid = self.id==other.id
return boul0 and boul1 and boulid
def __ne__(self,other):
"""2 edges are not equal iff they are not equal :) """
return (self==other)==False
#ne pas mettre @property ici, on veut une methode pas un attribut
def given_cost(self):
return self._given_cost
def __repr__(self):
return f"Edge [{str(self.linked[0].index)}, {str(self.linked[1].index)}] !oriented!"
class Graph:
""" All the information of a graph are contained here. """
def __init__(self,list_of_vertices):
""" Entry : the list of vertices. """
self.list_of_vertices = list_of_vertices
self.number_of_vertices = len(list_of_vertices)
self.connection_table_edge_and_diplayable_edge=[]
self.list_of_edges=[]
self.number_of_disp_edges=0
self.number_of_edges=0
def push_diplayable_edge(self,bidim_array):
self.connection_table_edge_and_diplayable_edge.append(copy.deepcopy(bidim_array))
self.number_of_disp_edges+=1
def push_edge(self,e):
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_edge_without_doublons(self, e):
if e not in self.list_of_edges:
self.number_of_edges+=1
self.list_of_edges.append(e)
def push_vertice(self,vertice):
self.list_of_vertices.append(vertice)
self.number_of_vertices += 1
def push_vertice_without_doublons(self, vertice):
bool,index = self.is_vertice_in_graph_based_on_xy_with_tolerance(vertice,10**(-8))
#bool,index = self.is_vertice_in_graph_based_on_xy(vertice)
if bool == False:
self.push_vertice(vertice)
else:
vertice.coordinates=self.list_of_vertices[index].coordinates
for edge in vertice.edges_list:
if edge not in self.list_of_vertices[index].edges_list:
self.list_of_vertices[index].push_edge(edge,True)
def is_vertice_in_graph_based_on_xy(self,vertice):
|
def is_vertice_in_graph_based_on_xy_with_tolerance(self, vertice, epsilon):
for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if ((v.coordinates[0] - vertice.coordinates[0])**2) + ((v.coordinates[1] - vertice.coordinates[1])**2) < epsilon:
return True, i
return False, None
def __getitem__(self, key):#implement instance[key]
if key >= 0 and key < self.number_of_vertices:
return self.list_of_vertices[key]
else :
raise IndexError
def laplace_matrix(self):
""" Returns the laplace matrix. """
n = self.number_of_vertices
laplace_matrix = np.zeros((n, n))
for i in range(n):
laplace_matrix[i][i] = 1
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
laplace_matrix[i][edge.linked[1].index] = 1
return laplace_matrix
def A_matrix(self,type_cost=Edge.given_cost):
""" Returns the laplace matrix. """
n = self.number_of_vertices
A_matrix = np.zeros((n, n))
for i in range(n):
vertice = self.list_of_vertices[i]
for edge in vertice.edges_list:
cost = type_cost(edge)
A_matrix[i][edge.linked[1].index] = cost
A_matrix[edge.linked[1].index][i] = cost
return A_matrix
def pairs_of_vertices(self):
"""Returns the pairs of connected vertices.
Beware ! There might be non-connected vertices in the graph. """
pairs_of_vertices = []
for vertice in self.list_of_vertices:
for edge in vertice.edges_list:
if non_oriented:
if (vertice, edge.linked[1]) and (edge.linked[1], vertice) not in pairs_of_vertices:
pairs_of_vertices.append((vertice, edge.linked[1]))
if not non_oriented:
if (vertice, edge.linked[1]) not in pairs_of_vertices:
pairs_of_vertices.append((vertice, edge.linked[1]))
return pairs_of_vertices
def number_of_edges(self):
a=self.pairs_of_vertices()
assert self.number_of_edges==len(a), "problem in Graph.pairs_of_vertices"
return self.number_of_edges
def search_index_by_coordinates(self,coord):
"""search the index of vertice at coordinates: """
for i in range(len(self.list_of_vertices)):
if self[i].coordinates[0]==coord[0] and self[i].coordinates[1]==coord[1]:
return i
def set_right_edges(self):
"""verify that the graph is coherent """
for v in self:
for e in v.edges_list:
e.linked[0]=v
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
for e in self.list_of_edges:
e.linked[0]=self[self.search_index_by_coordinates(e.linked[0].coordinates)]
e.linked[1]=self[self.search_index_by_coordinates(e.linked[1].coordinates)]
def plot(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
x = e.linked[0].coordinates[0]
y = e.linked[0].coordinates[1]
dx = e.linked[1].coordinates[0] - x
dy = e.linked[1].coordinates[1] - y
plt.plot([x,x+dx], [y,y+dy], color=c)
# plt.arrow(x,y,dx,dy)
plt.axis = 'off'
plt.show()
def plot_dev(self):
plt.clf()
for v in self._list_of_vertices:
c = f"#{v.color}"
plt.scatter(v.coordinates[0], v.coordinates[1], color=c)
for e in v.edges_list:
c = f"#{e.color}"
for i in range(len(self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable])-1):
x = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][0]
y = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i][1]
dx = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][0]-x
dy = self.connection_table_edge_and_diplayable_edge[e.connection_with_displayable][i+1][1]-y
plt.plot([x,x+dx], [y,y+dy], color=c)
plt.axis = 'off'
plt.show()
| for i in range(self.number_of_vertices):
v = self.list_of_vertices[i]
if v.coordinates[0] == vertice.coordinates[0] and v.coordinates[1] == vertice.coordinates[1]:
return True,i
return False,None | identifier_body |
build.rs | #[cfg(not(feature = "binary"))]
fn main() {}
#[cfg(feature = "binary")]
fn main() {
binary::main();
}
#[cfg(feature = "binary")]
mod binary {
use quote::quote;
use std::convert::TryInto;
pub fn main() {
use llvm_tools_build as llvm_tools;
use std::{
env,
fs::{self, File},
io::Write,
path::{Path, PathBuf},
process::{self, Command},
};
use toml::Value;
let target = env::var("TARGET").expect("TARGET not set");
let (firmware, expected_target) = if cfg!(feature = "uefi_bin") {
("UEFI", "x86_64-unknown-uefi")
} else if cfg!(feature = "bios_bin") {
("BIOS", "x86_64-bootloader")
} else {
panic!(
"Either the `uefi_bin` or `bios_bin` feature must be enabled when \
the `binary` feature is enabled"
);
};
if Path::new(&target)
.file_stem()
.expect("target has no file stem")
!= expected_target
|
let out_dir = PathBuf::from(env::var("OUT_DIR").expect("OUT_DIR not set"));
let kernel = PathBuf::from(match env::var("KERNEL") {
Ok(kernel) => kernel,
Err(_) => {
eprintln!(
"The KERNEL environment variable must be set for building the bootloader.\n\n\
Please use the `cargo builder` command for building."
);
process::exit(1);
}
});
let kernel_file_name = kernel
.file_name()
.expect("KERNEL has no valid file name")
.to_str()
.expect("kernel file name not valid utf8");
// check that the kernel file exists
assert!(
kernel.exists(),
"KERNEL does not exist: {}",
kernel.display()
);
// get access to llvm tools shipped in the llvm-tools-preview rustup component
let llvm_tools = match llvm_tools::LlvmTools::new() {
Ok(tools) => tools,
Err(llvm_tools::Error::NotFound) => {
eprintln!("Error: llvm-tools not found");
eprintln!("Maybe the rustup component `llvm-tools-preview` is missing?");
eprintln!(" Install it through: `rustup component add llvm-tools-preview`");
process::exit(1);
}
Err(err) => {
eprintln!("Failed to retrieve llvm-tools component: {:?}", err);
process::exit(1);
}
};
// check that kernel executable has code in it
let llvm_size = llvm_tools
.tool(&llvm_tools::exe("llvm-size"))
.expect("llvm-size not found in llvm-tools");
let mut cmd = Command::new(llvm_size);
cmd.arg(&kernel);
let output = cmd.output().expect("failed to run llvm-size");
let output_str = String::from_utf8_lossy(&output.stdout);
let second_line_opt = output_str.lines().skip(1).next();
let second_line = second_line_opt.expect(&format!(
"unexpected llvm-size line output:\n{}",
output_str
));
let text_size_opt = second_line.split_ascii_whitespace().next();
let text_size =
text_size_opt.expect(&format!("unexpected llvm-size output:\n{}", output_str));
if text_size == "0" {
panic!("Kernel executable has an empty text section. Perhaps the entry point was set incorrectly?\n\n\
Kernel executable at `{}`\n", kernel.display());
}
// strip debug symbols from kernel for faster loading
let stripped_kernel_file_name = format!("kernel_stripped-{}", kernel_file_name);
let stripped_kernel = out_dir.join(&stripped_kernel_file_name);
let objcopy = llvm_tools
.tool(&llvm_tools::exe("llvm-objcopy"))
.expect("llvm-objcopy not found in llvm-tools");
let mut cmd = Command::new(&objcopy);
cmd.arg("--strip-debug");
cmd.arg(&kernel);
cmd.arg(&stripped_kernel);
let exit_status = cmd
.status()
.expect("failed to run objcopy to strip debug symbols");
if !exit_status.success() {
eprintln!("Error: Stripping debug symbols failed");
process::exit(1);
}
if cfg!(feature = "uefi_bin") {
// write file for including kernel in binary
let file_path = out_dir.join("kernel_info.rs");
let mut file = File::create(file_path).expect("failed to create kernel_info.rs");
let kernel_size = fs::metadata(&stripped_kernel)
.expect("Failed to read file metadata of stripped kernel")
.len();
file.write_all(
format!(
"const KERNEL_SIZE: usize = {}; const KERNEL_BYTES: [u8; KERNEL_SIZE] = *include_bytes!(r\"{}\");",
kernel_size,
stripped_kernel.display(),
)
.as_bytes(),
)
.expect("write to kernel_info.rs failed");
}
if cfg!(feature = "bios_bin") {
// wrap the kernel executable as binary in a new ELF file
let stripped_kernel_file_name_replaced = stripped_kernel_file_name
.replace('-', "_")
.replace('.', "_");
let kernel_bin = out_dir.join(format!("kernel_bin-{}.o", kernel_file_name));
let kernel_archive = out_dir.join(format!("libkernel_bin-{}.a", kernel_file_name));
let mut cmd = Command::new(&objcopy);
cmd.arg("-I").arg("binary");
cmd.arg("-O").arg("elf64-x86-64");
cmd.arg("--binary-architecture=i386:x86-64");
cmd.arg("--rename-section").arg(".data=.kernel");
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_start=_kernel_start_addr",
stripped_kernel_file_name_replaced
));
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_end=_kernel_end_addr",
stripped_kernel_file_name_replaced
));
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_size=_kernel_size",
stripped_kernel_file_name_replaced
));
cmd.current_dir(&out_dir);
cmd.arg(&stripped_kernel_file_name);
cmd.arg(&kernel_bin);
let exit_status = cmd.status().expect("failed to run objcopy");
if !exit_status.success() {
eprintln!("Error: Running objcopy failed");
process::exit(1);
}
// create an archive for linking
let ar = llvm_tools
.tool(&llvm_tools::exe("llvm-ar"))
.unwrap_or_else(|| {
eprintln!("Failed to retrieve llvm-ar component");
eprint!("This component is available since nightly-2019-03-29,");
eprintln!("so try updating your toolchain if you're using an older nightly");
process::exit(1);
});
let mut cmd = Command::new(ar);
cmd.arg("crs");
cmd.arg(&kernel_archive);
cmd.arg(&kernel_bin);
let exit_status = cmd.status().expect("failed to run ar");
if !exit_status.success() {
eprintln!("Error: Running ar failed");
process::exit(1);
}
// pass link arguments to rustc
println!("cargo:rustc-link-search=native={}", out_dir.display());
println!(
"cargo:rustc-link-lib=static=kernel_bin-{}",
kernel_file_name
);
}
// Parse configuration from the kernel's Cargo.toml
let config = match env::var("KERNEL_MANIFEST") {
Err(env::VarError::NotPresent) => {
panic!("The KERNEL_MANIFEST environment variable must be set for building the bootloader.\n\n\
Please use `cargo builder` for building.");
}
Err(env::VarError::NotUnicode(_)) => {
panic!("The KERNEL_MANIFEST environment variable contains invalid unicode")
}
Ok(path)
if Path::new(&path).file_name().and_then(|s| s.to_str()) != Some("Cargo.toml") =>
{
let err = format!(
"The given `--kernel-manifest` path `{}` does not \
point to a `Cargo.toml`",
path,
);
quote! { compile_error!(#err) }
}
Ok(path) if !Path::new(&path).exists() => {
let err = format!(
"The given `--kernel-manifest` path `{}` does not exist.",
path
);
quote! {
compile_error!(#err)
}
}
Ok(path) => {
println!("cargo:rerun-if-changed={}", path);
let contents = fs::read_to_string(&path).expect(&format!(
"failed to read kernel manifest file (path: {})",
path
));
let manifest = contents
.parse::<Value>()
.expect("failed to parse kernel's Cargo.toml");
if manifest
.get("dependencies")
.and_then(|d| d.get("bootloader"))
.is_some()
{
// it seems to be the correct Cargo.toml
let config_table = manifest
.get("package")
.and_then(|table| table.get("metadata"))
.and_then(|table| table.get("bootloader"))
.cloned()
.unwrap_or_else(|| toml::Value::Table(toml::map::Map::new()));
config_table
.try_into::<ParsedConfig>()
.map(|c| quote! { #c })
.unwrap_or_else(|err| {
let err = format!(
"failed to parse bootloader config in {}:\n\n{}",
path,
err.to_string()
);
quote! {
compile_error!(#err)
}
})
} else {
let err = format!(
"no bootloader dependency in {}\n\n The \
`--kernel-manifest` path should point to the `Cargo.toml` \
of the kernel.",
path
);
quote! {
compile_error!(#err)
}
}
}
};
// Write config to file
let file_path = out_dir.join("bootloader_config.rs");
let mut file = File::create(file_path).expect("failed to create bootloader_config.rs");
file.write_all(
quote::quote! {
mod parsed_config {
use crate::config::Config;
pub const CONFIG: Config = #config;
}
}
.to_string()
.as_bytes(),
)
.expect("write to bootloader_config.rs failed");
println!("cargo:rerun-if-env-changed=KERNEL");
println!("cargo:rerun-if-env-changed=KERNEL_MANIFEST");
println!("cargo:rerun-if-changed={}", kernel.display());
println!("cargo:rerun-if-changed=build.rs");
}
fn val_true() -> bool {
true
}
/// Must be always identical with the struct in `src/config.rs`
///
/// This copy is needed because we can't derive Deserialize in the `src/config.rs`
/// module itself, since cargo currently unifies dependencies (the `toml` crate enables
/// serde's standard feature). Also, it allows to separate the parsing special cases
/// such as `AlignedAddress` more cleanly.
#[derive(Debug, serde::Deserialize)]
#[serde(rename_all = "kebab-case", deny_unknown_fields)]
struct ParsedConfig {
#[serde(default)]
pub map_physical_memory: bool,
#[serde(default)]
pub map_page_table_recursively: bool,
#[serde(default = "val_true")]
pub map_framebuffer: bool,
pub kernel_stack_size: Option<AlignedAddress>,
pub physical_memory_offset: Option<AlignedAddress>,
pub recursive_index: Option<u16>,
pub kernel_stack_address: Option<AlignedAddress>,
pub boot_info_address: Option<AlignedAddress>,
pub framebuffer_address: Option<AlignedAddress>,
}
/// Convert to tokens suitable for initializing the `Config` struct.
impl quote::ToTokens for ParsedConfig {
fn to_tokens(&self, tokens: &mut proc_macro2::TokenStream) {
fn optional(value: Option<impl quote::ToTokens>) -> proc_macro2::TokenStream {
value.map(|v| quote!(Some(#v))).unwrap_or(quote!(None))
}
let map_physical_memory = self.map_physical_memory;
let map_page_table_recursively = self.map_page_table_recursively;
let map_framebuffer = self.map_framebuffer;
let kernel_stack_size = optional(self.kernel_stack_size);
let physical_memory_offset = optional(self.physical_memory_offset);
let recursive_index = optional(self.recursive_index);
let kernel_stack_address = optional(self.kernel_stack_address);
let boot_info_address = optional(self.boot_info_address);
let framebuffer_address = optional(self.framebuffer_address);
tokens.extend(quote! { Config {
map_physical_memory: #map_physical_memory,
map_page_table_recursively: #map_page_table_recursively,
map_framebuffer: #map_framebuffer,
kernel_stack_size: #kernel_stack_size,
physical_memory_offset: #physical_memory_offset,
recursive_index: #recursive_index,
kernel_stack_address: #kernel_stack_address,
boot_info_address: #boot_info_address,
framebuffer_address: #framebuffer_address,
}});
}
}
#[derive(Debug, Clone, Copy)]
struct AlignedAddress(u64);
impl quote::ToTokens for AlignedAddress {
fn to_tokens(&self, tokens: &mut proc_macro2::TokenStream) {
self.0.to_tokens(tokens);
}
}
impl<'de> serde::Deserialize<'de> for AlignedAddress {
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
where
D: serde::Deserializer<'de>,
{
deserializer.deserialize_str(AlignedAddressVisitor)
}
}
/// Helper struct for implementing the `optional_version_deserialize` function.
struct AlignedAddressVisitor;
impl serde::de::Visitor<'_> for AlignedAddressVisitor {
type Value = AlignedAddress;
fn expecting(&self, formatter: &mut std::fmt::Formatter) -> std::fmt::Result {
write!(
formatter,
"a page-aligned memory address, either as integer or as decimal or hexadecimal \
string (e.g. \"0xffff0000\"); large addresses must be given as string because \
TOML does not support unsigned 64-bit integers"
)
}
fn visit_u64<E>(self, num: u64) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
if num % 0x1000 == 0 {
Ok(AlignedAddress(num))
} else {
Err(serde::de::Error::custom(format!(
"address {:#x} is not page aligned",
num
)))
}
}
fn visit_i64<E>(self, num: i64) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
let unsigned: u64 = num
.try_into()
.map_err(|_| serde::de::Error::custom(format!("address {} is negative", num)))?;
self.visit_u64(unsigned)
}
fn visit_str<E>(self, value: &str) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
// ignore any `_` (used for digit grouping)
let value = &value.replace('_', "");
let num = if value.starts_with("0x") {
u64::from_str_radix(&value[2..], 16)
} else {
u64::from_str_radix(&value, 10)
}
.map_err(|_err| {
serde::de::Error::custom(format!(
"string \"{}\" is not a valid memory address",
value
))
})?;
self.visit_u64(num)
}
}
}
| {
panic!(
"The {} bootloader must be compiled for the `{}` target.",
firmware, expected_target,
);
} | conditional_block |
build.rs | #[cfg(not(feature = "binary"))]
fn main() {}
#[cfg(feature = "binary")]
fn main() {
binary::main();
}
#[cfg(feature = "binary")]
mod binary {
use quote::quote;
use std::convert::TryInto;
pub fn main() {
use llvm_tools_build as llvm_tools;
use std::{
env,
fs::{self, File},
io::Write,
path::{Path, PathBuf},
process::{self, Command},
};
use toml::Value;
let target = env::var("TARGET").expect("TARGET not set");
let (firmware, expected_target) = if cfg!(feature = "uefi_bin") {
("UEFI", "x86_64-unknown-uefi")
} else if cfg!(feature = "bios_bin") {
("BIOS", "x86_64-bootloader")
} else {
panic!(
"Either the `uefi_bin` or `bios_bin` feature must be enabled when \
the `binary` feature is enabled"
);
};
if Path::new(&target)
.file_stem()
.expect("target has no file stem")
!= expected_target
{
panic!(
"The {} bootloader must be compiled for the `{}` target.",
firmware, expected_target,
);
}
let out_dir = PathBuf::from(env::var("OUT_DIR").expect("OUT_DIR not set"));
let kernel = PathBuf::from(match env::var("KERNEL") {
Ok(kernel) => kernel,
Err(_) => {
eprintln!(
"The KERNEL environment variable must be set for building the bootloader.\n\n\
Please use the `cargo builder` command for building."
);
process::exit(1);
}
});
let kernel_file_name = kernel
.file_name()
.expect("KERNEL has no valid file name")
.to_str()
.expect("kernel file name not valid utf8");
// check that the kernel file exists
assert!(
kernel.exists(),
"KERNEL does not exist: {}",
kernel.display()
);
// get access to llvm tools shipped in the llvm-tools-preview rustup component
let llvm_tools = match llvm_tools::LlvmTools::new() {
Ok(tools) => tools,
Err(llvm_tools::Error::NotFound) => {
eprintln!("Error: llvm-tools not found");
eprintln!("Maybe the rustup component `llvm-tools-preview` is missing?");
eprintln!(" Install it through: `rustup component add llvm-tools-preview`");
process::exit(1);
}
Err(err) => {
eprintln!("Failed to retrieve llvm-tools component: {:?}", err);
process::exit(1);
}
};
// check that kernel executable has code in it
let llvm_size = llvm_tools
.tool(&llvm_tools::exe("llvm-size"))
.expect("llvm-size not found in llvm-tools");
let mut cmd = Command::new(llvm_size);
cmd.arg(&kernel);
let output = cmd.output().expect("failed to run llvm-size");
let output_str = String::from_utf8_lossy(&output.stdout);
let second_line_opt = output_str.lines().skip(1).next();
let second_line = second_line_opt.expect(&format!(
"unexpected llvm-size line output:\n{}",
output_str
));
let text_size_opt = second_line.split_ascii_whitespace().next();
let text_size =
text_size_opt.expect(&format!("unexpected llvm-size output:\n{}", output_str));
if text_size == "0" {
panic!("Kernel executable has an empty text section. Perhaps the entry point was set incorrectly?\n\n\
Kernel executable at `{}`\n", kernel.display());
}
// strip debug symbols from kernel for faster loading
let stripped_kernel_file_name = format!("kernel_stripped-{}", kernel_file_name);
let stripped_kernel = out_dir.join(&stripped_kernel_file_name);
let objcopy = llvm_tools
.tool(&llvm_tools::exe("llvm-objcopy"))
.expect("llvm-objcopy not found in llvm-tools");
let mut cmd = Command::new(&objcopy);
cmd.arg("--strip-debug");
cmd.arg(&kernel);
cmd.arg(&stripped_kernel);
let exit_status = cmd
.status()
.expect("failed to run objcopy to strip debug symbols");
if !exit_status.success() {
eprintln!("Error: Stripping debug symbols failed");
process::exit(1);
}
if cfg!(feature = "uefi_bin") {
// write file for including kernel in binary
let file_path = out_dir.join("kernel_info.rs");
let mut file = File::create(file_path).expect("failed to create kernel_info.rs");
let kernel_size = fs::metadata(&stripped_kernel)
.expect("Failed to read file metadata of stripped kernel")
.len();
file.write_all(
format!(
"const KERNEL_SIZE: usize = {}; const KERNEL_BYTES: [u8; KERNEL_SIZE] = *include_bytes!(r\"{}\");",
kernel_size,
stripped_kernel.display(),
)
.as_bytes(),
)
.expect("write to kernel_info.rs failed");
}
if cfg!(feature = "bios_bin") {
// wrap the kernel executable as binary in a new ELF file
let stripped_kernel_file_name_replaced = stripped_kernel_file_name
.replace('-', "_")
.replace('.', "_");
let kernel_bin = out_dir.join(format!("kernel_bin-{}.o", kernel_file_name));
let kernel_archive = out_dir.join(format!("libkernel_bin-{}.a", kernel_file_name));
let mut cmd = Command::new(&objcopy);
cmd.arg("-I").arg("binary");
cmd.arg("-O").arg("elf64-x86-64");
cmd.arg("--binary-architecture=i386:x86-64");
cmd.arg("--rename-section").arg(".data=.kernel");
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_start=_kernel_start_addr",
stripped_kernel_file_name_replaced
));
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_end=_kernel_end_addr",
stripped_kernel_file_name_replaced
));
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_size=_kernel_size",
stripped_kernel_file_name_replaced
));
cmd.current_dir(&out_dir);
cmd.arg(&stripped_kernel_file_name);
cmd.arg(&kernel_bin);
let exit_status = cmd.status().expect("failed to run objcopy");
if !exit_status.success() {
eprintln!("Error: Running objcopy failed");
process::exit(1);
}
// create an archive for linking
let ar = llvm_tools
.tool(&llvm_tools::exe("llvm-ar"))
.unwrap_or_else(|| {
eprintln!("Failed to retrieve llvm-ar component");
eprint!("This component is available since nightly-2019-03-29,");
eprintln!("so try updating your toolchain if you're using an older nightly");
process::exit(1);
});
let mut cmd = Command::new(ar);
cmd.arg("crs");
cmd.arg(&kernel_archive);
cmd.arg(&kernel_bin);
let exit_status = cmd.status().expect("failed to run ar");
if !exit_status.success() {
eprintln!("Error: Running ar failed");
process::exit(1);
}
// pass link arguments to rustc
println!("cargo:rustc-link-search=native={}", out_dir.display());
println!(
"cargo:rustc-link-lib=static=kernel_bin-{}",
kernel_file_name
);
}
// Parse configuration from the kernel's Cargo.toml
let config = match env::var("KERNEL_MANIFEST") {
Err(env::VarError::NotPresent) => {
panic!("The KERNEL_MANIFEST environment variable must be set for building the bootloader.\n\n\
Please use `cargo builder` for building.");
}
Err(env::VarError::NotUnicode(_)) => {
panic!("The KERNEL_MANIFEST environment variable contains invalid unicode")
}
Ok(path)
if Path::new(&path).file_name().and_then(|s| s.to_str()) != Some("Cargo.toml") =>
{
let err = format!(
"The given `--kernel-manifest` path `{}` does not \
point to a `Cargo.toml`",
path,
);
quote! { compile_error!(#err) }
}
Ok(path) if !Path::new(&path).exists() => {
let err = format!(
"The given `--kernel-manifest` path `{}` does not exist.",
path
);
quote! {
compile_error!(#err)
}
}
Ok(path) => {
println!("cargo:rerun-if-changed={}", path);
let contents = fs::read_to_string(&path).expect(&format!(
"failed to read kernel manifest file (path: {})",
path
));
let manifest = contents
.parse::<Value>()
.expect("failed to parse kernel's Cargo.toml");
if manifest
.get("dependencies")
.and_then(|d| d.get("bootloader"))
.is_some()
{
// it seems to be the correct Cargo.toml
let config_table = manifest
.get("package")
.and_then(|table| table.get("metadata"))
.and_then(|table| table.get("bootloader"))
.cloned()
.unwrap_or_else(|| toml::Value::Table(toml::map::Map::new()));
config_table
.try_into::<ParsedConfig>()
.map(|c| quote! { #c })
.unwrap_or_else(|err| {
let err = format!(
"failed to parse bootloader config in {}:\n\n{}",
path,
err.to_string()
);
quote! {
compile_error!(#err)
}
})
} else {
let err = format!(
"no bootloader dependency in {}\n\n The \
`--kernel-manifest` path should point to the `Cargo.toml` \
of the kernel.",
path
);
quote! {
compile_error!(#err)
}
}
}
};
// Write config to file
let file_path = out_dir.join("bootloader_config.rs");
let mut file = File::create(file_path).expect("failed to create bootloader_config.rs");
file.write_all(
quote::quote! {
mod parsed_config {
use crate::config::Config;
pub const CONFIG: Config = #config;
}
}
.to_string()
.as_bytes(),
)
.expect("write to bootloader_config.rs failed");
println!("cargo:rerun-if-env-changed=KERNEL");
println!("cargo:rerun-if-env-changed=KERNEL_MANIFEST");
println!("cargo:rerun-if-changed={}", kernel.display());
println!("cargo:rerun-if-changed=build.rs");
}
fn val_true() -> bool {
true
}
/// Must be always identical with the struct in `src/config.rs`
///
/// This copy is needed because we can't derive Deserialize in the `src/config.rs`
/// module itself, since cargo currently unifies dependencies (the `toml` crate enables
/// serde's standard feature). Also, it allows to separate the parsing special cases
/// such as `AlignedAddress` more cleanly.
#[derive(Debug, serde::Deserialize)]
#[serde(rename_all = "kebab-case", deny_unknown_fields)]
struct ParsedConfig {
#[serde(default)]
pub map_physical_memory: bool,
#[serde(default)]
pub map_page_table_recursively: bool,
#[serde(default = "val_true")]
pub map_framebuffer: bool,
pub kernel_stack_size: Option<AlignedAddress>,
pub physical_memory_offset: Option<AlignedAddress>,
pub recursive_index: Option<u16>,
pub kernel_stack_address: Option<AlignedAddress>,
pub boot_info_address: Option<AlignedAddress>,
pub framebuffer_address: Option<AlignedAddress>,
}
/// Convert to tokens suitable for initializing the `Config` struct.
impl quote::ToTokens for ParsedConfig {
fn to_tokens(&self, tokens: &mut proc_macro2::TokenStream) {
fn optional(value: Option<impl quote::ToTokens>) -> proc_macro2::TokenStream {
value.map(|v| quote!(Some(#v))).unwrap_or(quote!(None))
}
let map_physical_memory = self.map_physical_memory;
let map_page_table_recursively = self.map_page_table_recursively;
let map_framebuffer = self.map_framebuffer;
let kernel_stack_size = optional(self.kernel_stack_size);
let physical_memory_offset = optional(self.physical_memory_offset);
let recursive_index = optional(self.recursive_index);
let kernel_stack_address = optional(self.kernel_stack_address);
let boot_info_address = optional(self.boot_info_address);
let framebuffer_address = optional(self.framebuffer_address);
tokens.extend(quote! { Config {
map_physical_memory: #map_physical_memory,
map_page_table_recursively: #map_page_table_recursively,
map_framebuffer: #map_framebuffer,
kernel_stack_size: #kernel_stack_size,
physical_memory_offset: #physical_memory_offset,
recursive_index: #recursive_index,
kernel_stack_address: #kernel_stack_address,
boot_info_address: #boot_info_address,
framebuffer_address: #framebuffer_address,
}});
}
}
#[derive(Debug, Clone, Copy)]
struct AlignedAddress(u64);
impl quote::ToTokens for AlignedAddress {
fn to_tokens(&self, tokens: &mut proc_macro2::TokenStream) {
self.0.to_tokens(tokens);
}
}
impl<'de> serde::Deserialize<'de> for AlignedAddress {
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
where
D: serde::Deserializer<'de>,
{
deserializer.deserialize_str(AlignedAddressVisitor)
}
}
/// Helper struct for implementing the `optional_version_deserialize` function.
struct | ;
impl serde::de::Visitor<'_> for AlignedAddressVisitor {
type Value = AlignedAddress;
fn expecting(&self, formatter: &mut std::fmt::Formatter) -> std::fmt::Result {
write!(
formatter,
"a page-aligned memory address, either as integer or as decimal or hexadecimal \
string (e.g. \"0xffff0000\"); large addresses must be given as string because \
TOML does not support unsigned 64-bit integers"
)
}
fn visit_u64<E>(self, num: u64) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
if num % 0x1000 == 0 {
Ok(AlignedAddress(num))
} else {
Err(serde::de::Error::custom(format!(
"address {:#x} is not page aligned",
num
)))
}
}
fn visit_i64<E>(self, num: i64) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
let unsigned: u64 = num
.try_into()
.map_err(|_| serde::de::Error::custom(format!("address {} is negative", num)))?;
self.visit_u64(unsigned)
}
fn visit_str<E>(self, value: &str) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
// ignore any `_` (used for digit grouping)
let value = &value.replace('_', "");
let num = if value.starts_with("0x") {
u64::from_str_radix(&value[2..], 16)
} else {
u64::from_str_radix(&value, 10)
}
.map_err(|_err| {
serde::de::Error::custom(format!(
"string \"{}\" is not a valid memory address",
value
))
})?;
self.visit_u64(num)
}
}
}
| AlignedAddressVisitor | identifier_name |
build.rs | #[cfg(not(feature = "binary"))]
fn main() {}
#[cfg(feature = "binary")]
fn main() {
binary::main();
}
#[cfg(feature = "binary")]
mod binary {
use quote::quote;
use std::convert::TryInto;
pub fn main() {
use llvm_tools_build as llvm_tools;
use std::{
env,
fs::{self, File},
io::Write,
path::{Path, PathBuf},
process::{self, Command},
};
use toml::Value;
let target = env::var("TARGET").expect("TARGET not set");
let (firmware, expected_target) = if cfg!(feature = "uefi_bin") {
("UEFI", "x86_64-unknown-uefi")
} else if cfg!(feature = "bios_bin") {
("BIOS", "x86_64-bootloader")
} else {
panic!(
"Either the `uefi_bin` or `bios_bin` feature must be enabled when \
the `binary` feature is enabled"
);
};
if Path::new(&target)
.file_stem()
.expect("target has no file stem")
!= expected_target
{
panic!(
"The {} bootloader must be compiled for the `{}` target.",
firmware, expected_target,
);
}
let out_dir = PathBuf::from(env::var("OUT_DIR").expect("OUT_DIR not set"));
let kernel = PathBuf::from(match env::var("KERNEL") {
Ok(kernel) => kernel,
Err(_) => {
eprintln!(
"The KERNEL environment variable must be set for building the bootloader.\n\n\
Please use the `cargo builder` command for building."
);
process::exit(1);
}
});
let kernel_file_name = kernel
.file_name()
.expect("KERNEL has no valid file name")
.to_str()
.expect("kernel file name not valid utf8");
// check that the kernel file exists
assert!(
kernel.exists(),
"KERNEL does not exist: {}",
kernel.display()
);
// get access to llvm tools shipped in the llvm-tools-preview rustup component
let llvm_tools = match llvm_tools::LlvmTools::new() {
Ok(tools) => tools,
Err(llvm_tools::Error::NotFound) => {
eprintln!("Error: llvm-tools not found");
eprintln!("Maybe the rustup component `llvm-tools-preview` is missing?");
eprintln!(" Install it through: `rustup component add llvm-tools-preview`");
process::exit(1);
}
Err(err) => {
eprintln!("Failed to retrieve llvm-tools component: {:?}", err);
process::exit(1);
}
};
// check that kernel executable has code in it
let llvm_size = llvm_tools
.tool(&llvm_tools::exe("llvm-size"))
.expect("llvm-size not found in llvm-tools");
let mut cmd = Command::new(llvm_size);
cmd.arg(&kernel);
let output = cmd.output().expect("failed to run llvm-size");
let output_str = String::from_utf8_lossy(&output.stdout);
let second_line_opt = output_str.lines().skip(1).next();
let second_line = second_line_opt.expect(&format!(
"unexpected llvm-size line output:\n{}",
output_str
));
let text_size_opt = second_line.split_ascii_whitespace().next();
let text_size =
text_size_opt.expect(&format!("unexpected llvm-size output:\n{}", output_str));
if text_size == "0" {
panic!("Kernel executable has an empty text section. Perhaps the entry point was set incorrectly?\n\n\
Kernel executable at `{}`\n", kernel.display());
}
// strip debug symbols from kernel for faster loading
let stripped_kernel_file_name = format!("kernel_stripped-{}", kernel_file_name);
let stripped_kernel = out_dir.join(&stripped_kernel_file_name);
let objcopy = llvm_tools
.tool(&llvm_tools::exe("llvm-objcopy"))
.expect("llvm-objcopy not found in llvm-tools");
let mut cmd = Command::new(&objcopy);
cmd.arg("--strip-debug");
cmd.arg(&kernel);
cmd.arg(&stripped_kernel);
let exit_status = cmd
.status()
.expect("failed to run objcopy to strip debug symbols");
if !exit_status.success() {
eprintln!("Error: Stripping debug symbols failed");
process::exit(1);
}
if cfg!(feature = "uefi_bin") {
// write file for including kernel in binary
let file_path = out_dir.join("kernel_info.rs");
let mut file = File::create(file_path).expect("failed to create kernel_info.rs");
let kernel_size = fs::metadata(&stripped_kernel)
.expect("Failed to read file metadata of stripped kernel")
.len();
file.write_all(
format!(
"const KERNEL_SIZE: usize = {}; const KERNEL_BYTES: [u8; KERNEL_SIZE] = *include_bytes!(r\"{}\");",
kernel_size,
stripped_kernel.display(),
)
.as_bytes(),
)
.expect("write to kernel_info.rs failed");
}
if cfg!(feature = "bios_bin") {
// wrap the kernel executable as binary in a new ELF file
let stripped_kernel_file_name_replaced = stripped_kernel_file_name
.replace('-', "_")
.replace('.', "_");
let kernel_bin = out_dir.join(format!("kernel_bin-{}.o", kernel_file_name));
let kernel_archive = out_dir.join(format!("libkernel_bin-{}.a", kernel_file_name));
let mut cmd = Command::new(&objcopy);
cmd.arg("-I").arg("binary");
cmd.arg("-O").arg("elf64-x86-64");
cmd.arg("--binary-architecture=i386:x86-64");
cmd.arg("--rename-section").arg(".data=.kernel");
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_start=_kernel_start_addr",
stripped_kernel_file_name_replaced
));
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_end=_kernel_end_addr",
stripped_kernel_file_name_replaced
));
cmd.arg("--redefine-sym").arg(format!(
"_binary_{}_size=_kernel_size",
stripped_kernel_file_name_replaced
));
cmd.current_dir(&out_dir);
cmd.arg(&stripped_kernel_file_name);
cmd.arg(&kernel_bin);
let exit_status = cmd.status().expect("failed to run objcopy");
if !exit_status.success() {
eprintln!("Error: Running objcopy failed");
process::exit(1);
}
// create an archive for linking
let ar = llvm_tools
.tool(&llvm_tools::exe("llvm-ar"))
.unwrap_or_else(|| {
eprintln!("Failed to retrieve llvm-ar component");
eprint!("This component is available since nightly-2019-03-29,");
eprintln!("so try updating your toolchain if you're using an older nightly");
process::exit(1);
});
let mut cmd = Command::new(ar);
cmd.arg("crs");
cmd.arg(&kernel_archive);
cmd.arg(&kernel_bin);
let exit_status = cmd.status().expect("failed to run ar");
if !exit_status.success() {
eprintln!("Error: Running ar failed");
process::exit(1);
}
// pass link arguments to rustc
println!("cargo:rustc-link-search=native={}", out_dir.display());
println!(
"cargo:rustc-link-lib=static=kernel_bin-{}", | // Parse configuration from the kernel's Cargo.toml
let config = match env::var("KERNEL_MANIFEST") {
Err(env::VarError::NotPresent) => {
panic!("The KERNEL_MANIFEST environment variable must be set for building the bootloader.\n\n\
Please use `cargo builder` for building.");
}
Err(env::VarError::NotUnicode(_)) => {
panic!("The KERNEL_MANIFEST environment variable contains invalid unicode")
}
Ok(path)
if Path::new(&path).file_name().and_then(|s| s.to_str()) != Some("Cargo.toml") =>
{
let err = format!(
"The given `--kernel-manifest` path `{}` does not \
point to a `Cargo.toml`",
path,
);
quote! { compile_error!(#err) }
}
Ok(path) if !Path::new(&path).exists() => {
let err = format!(
"The given `--kernel-manifest` path `{}` does not exist.",
path
);
quote! {
compile_error!(#err)
}
}
Ok(path) => {
println!("cargo:rerun-if-changed={}", path);
let contents = fs::read_to_string(&path).expect(&format!(
"failed to read kernel manifest file (path: {})",
path
));
let manifest = contents
.parse::<Value>()
.expect("failed to parse kernel's Cargo.toml");
if manifest
.get("dependencies")
.and_then(|d| d.get("bootloader"))
.is_some()
{
// it seems to be the correct Cargo.toml
let config_table = manifest
.get("package")
.and_then(|table| table.get("metadata"))
.and_then(|table| table.get("bootloader"))
.cloned()
.unwrap_or_else(|| toml::Value::Table(toml::map::Map::new()));
config_table
.try_into::<ParsedConfig>()
.map(|c| quote! { #c })
.unwrap_or_else(|err| {
let err = format!(
"failed to parse bootloader config in {}:\n\n{}",
path,
err.to_string()
);
quote! {
compile_error!(#err)
}
})
} else {
let err = format!(
"no bootloader dependency in {}\n\n The \
`--kernel-manifest` path should point to the `Cargo.toml` \
of the kernel.",
path
);
quote! {
compile_error!(#err)
}
}
}
};
// Write config to file
let file_path = out_dir.join("bootloader_config.rs");
let mut file = File::create(file_path).expect("failed to create bootloader_config.rs");
file.write_all(
quote::quote! {
mod parsed_config {
use crate::config::Config;
pub const CONFIG: Config = #config;
}
}
.to_string()
.as_bytes(),
)
.expect("write to bootloader_config.rs failed");
println!("cargo:rerun-if-env-changed=KERNEL");
println!("cargo:rerun-if-env-changed=KERNEL_MANIFEST");
println!("cargo:rerun-if-changed={}", kernel.display());
println!("cargo:rerun-if-changed=build.rs");
}
fn val_true() -> bool {
true
}
/// Must be always identical with the struct in `src/config.rs`
///
/// This copy is needed because we can't derive Deserialize in the `src/config.rs`
/// module itself, since cargo currently unifies dependencies (the `toml` crate enables
/// serde's standard feature). Also, it allows to separate the parsing special cases
/// such as `AlignedAddress` more cleanly.
#[derive(Debug, serde::Deserialize)]
#[serde(rename_all = "kebab-case", deny_unknown_fields)]
struct ParsedConfig {
#[serde(default)]
pub map_physical_memory: bool,
#[serde(default)]
pub map_page_table_recursively: bool,
#[serde(default = "val_true")]
pub map_framebuffer: bool,
pub kernel_stack_size: Option<AlignedAddress>,
pub physical_memory_offset: Option<AlignedAddress>,
pub recursive_index: Option<u16>,
pub kernel_stack_address: Option<AlignedAddress>,
pub boot_info_address: Option<AlignedAddress>,
pub framebuffer_address: Option<AlignedAddress>,
}
/// Convert to tokens suitable for initializing the `Config` struct.
impl quote::ToTokens for ParsedConfig {
fn to_tokens(&self, tokens: &mut proc_macro2::TokenStream) {
fn optional(value: Option<impl quote::ToTokens>) -> proc_macro2::TokenStream {
value.map(|v| quote!(Some(#v))).unwrap_or(quote!(None))
}
let map_physical_memory = self.map_physical_memory;
let map_page_table_recursively = self.map_page_table_recursively;
let map_framebuffer = self.map_framebuffer;
let kernel_stack_size = optional(self.kernel_stack_size);
let physical_memory_offset = optional(self.physical_memory_offset);
let recursive_index = optional(self.recursive_index);
let kernel_stack_address = optional(self.kernel_stack_address);
let boot_info_address = optional(self.boot_info_address);
let framebuffer_address = optional(self.framebuffer_address);
tokens.extend(quote! { Config {
map_physical_memory: #map_physical_memory,
map_page_table_recursively: #map_page_table_recursively,
map_framebuffer: #map_framebuffer,
kernel_stack_size: #kernel_stack_size,
physical_memory_offset: #physical_memory_offset,
recursive_index: #recursive_index,
kernel_stack_address: #kernel_stack_address,
boot_info_address: #boot_info_address,
framebuffer_address: #framebuffer_address,
}});
}
}
#[derive(Debug, Clone, Copy)]
struct AlignedAddress(u64);
impl quote::ToTokens for AlignedAddress {
fn to_tokens(&self, tokens: &mut proc_macro2::TokenStream) {
self.0.to_tokens(tokens);
}
}
impl<'de> serde::Deserialize<'de> for AlignedAddress {
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
where
D: serde::Deserializer<'de>,
{
deserializer.deserialize_str(AlignedAddressVisitor)
}
}
/// Helper struct for implementing the `optional_version_deserialize` function.
struct AlignedAddressVisitor;
impl serde::de::Visitor<'_> for AlignedAddressVisitor {
type Value = AlignedAddress;
fn expecting(&self, formatter: &mut std::fmt::Formatter) -> std::fmt::Result {
write!(
formatter,
"a page-aligned memory address, either as integer or as decimal or hexadecimal \
string (e.g. \"0xffff0000\"); large addresses must be given as string because \
TOML does not support unsigned 64-bit integers"
)
}
fn visit_u64<E>(self, num: u64) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
if num % 0x1000 == 0 {
Ok(AlignedAddress(num))
} else {
Err(serde::de::Error::custom(format!(
"address {:#x} is not page aligned",
num
)))
}
}
fn visit_i64<E>(self, num: i64) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
let unsigned: u64 = num
.try_into()
.map_err(|_| serde::de::Error::custom(format!("address {} is negative", num)))?;
self.visit_u64(unsigned)
}
fn visit_str<E>(self, value: &str) -> Result<Self::Value, E>
where
E: serde::de::Error,
{
// ignore any `_` (used for digit grouping)
let value = &value.replace('_', "");
let num = if value.starts_with("0x") {
u64::from_str_radix(&value[2..], 16)
} else {
u64::from_str_radix(&value, 10)
}
.map_err(|_err| {
serde::de::Error::custom(format!(
"string \"{}\" is not a valid memory address",
value
))
})?;
self.visit_u64(num)
}
}
} | kernel_file_name
);
}
| random_line_split |
state.rs | use std::collections::HashMap;
use num_traits::{AsPrimitive, Float};
use sdl2::{
keyboard::Keycode,
pixels::Color,
rect::{Point, Rect},
render::Canvas,
video::Window,
};
use crate::{
ext::ColorExt, key_state_handler::KeyStateHandler, map::Map, math::vector::Vec2D,
WINDOW_HEIGHT, WINDOW_WIDTH,
};
pub fn remap<
T: 'static + Float + Copy,
ON: AsPrimitive<T> + Copy,
OX: AsPrimitive<T> + Copy,
NN: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_min: ON,
old_max: OX,
new_min: NN,
new_max: NX,
) -> T {
new_min.as_()
+ (new_max.as_() - new_min.as_())
* ((value - old_min.as_()) / (old_max.as_() - old_min.as_()))
}
pub fn remap_minz<
T: 'static + Float + Copy + AsPrimitive<T>,
OX: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_max: OX,
new_max: NX,
) -> T {
remap(value, T::zero(), old_max, T::zero(), new_max)
}
const MOUSE_SENSITIVITY: f64 = 0.01;
const MOVE_SPEED: f64 = 2.5;
const PLAYER_WALL_PADDING: f64 = 10.;
const WALL_ACTUAL_HEIGHT: f64 = 48.;
pub struct State {
pub(crate) position: Vec2D,
pub(crate) angle: f64,
pub(crate) fov: f64,
pub(crate) wall_colors: Vec<Color>,
pub(crate) map: Map,
pub(crate) keys: KeyStateHandler,
pub(crate) columns: Vec<(usize, u32)>,
pub(crate) resolution: usize,
pub(crate) projection_factor: f64,
pub(crate) radian_per_column: f64,
pub(crate) column_width: u32,
}
impl State {
pub fn new() -> Self {
let wall_colors = vec![
Color::RGB(128, 255, 0),
Color::RGB(0, 128, 255),
Color::RGB(255, 0, 128),
Color::RGB(0, 255, 0),
Color::RGB(0, 0, 255),
Color::WHITE,
];
// let map = Map::default();
let map = Map::load("./assets/maps/many_walls.json").unwrap();
let (w, h) = map.dims;
// let movement_vector = Line::new(origin, origin + geo::Point::new(delta_x, delta_y));
let position = Vec2D::new(w as f64 / 2., 50. + h as f64 / 2.);
let fov = 60.;
let projection_plane_distance =
((WINDOW_WIDTH / 2) as f64 / (fov.to_radians() / 2.).tan()) as f64;
let resolution = WINDOW_WIDTH as usize;
Self {
position,
angle: std::f64::consts::PI,
fov,
wall_colors,
map,
keys: KeyStateHandler::new(),
columns: Vec::with_capacity(resolution),
resolution,
projection_factor: projection_plane_distance * WALL_ACTUAL_HEIGHT,
radian_per_column: fov.to_radians() / resolution as f64,
column_width: WINDOW_WIDTH / resolution as u32,
}
}
fn get_color(&self, index: usize) -> Color |
pub fn mouse_motion(&mut self, dx: i32) {
self.angle += MOUSE_SENSITIVITY * dx as f64;
}
fn calculate_collisions(&mut self) {
let mut current_angle = self.angle - (self.fov.to_radians() / 2.);
let end_angle = current_angle + (self.radian_per_column * self.resolution as f64);
self.columns.clear();
for _ in 0..self.resolution {
let mut ray = Vec2D::from_angle(current_angle);
ray.translate(&self.position);
let mut max_height = f64::NEG_INFINITY;
let mut wall_color_index = 0;
for wall in self.map.walls.iter() {
if let Some(intersection_vector) = ray.intersects(wall) {
let raw_distance = ray.dist(&intersection_vector);
let delta = current_angle - self.angle;
let corrected_distance = raw_distance * (delta.cos() as f64);
let projected_height = self.projection_factor / corrected_distance;
if projected_height > max_height {
max_height = projected_height;
wall_color_index = wall.color_index;
}
}
}
if max_height.is_infinite() {
self.columns.push((0, 0));
} else {
self.columns
.push((wall_color_index, max_height.round() as u32));
}
current_angle += self.radian_per_column;
}
}
fn update_camera(&mut self) {
let mut delta = Vec2D::Origin;
let par = Vec2D::from_angle(self.angle as f64);
let perp = Vec2D::from_angle(self.angle + (90f64).to_radians());
if self.keys.is_pressed(Keycode::W) {
delta += par;
}
if self.keys.is_pressed(Keycode::S) {
delta += -par;
}
if self.keys.is_pressed(Keycode::A) {
delta += -perp;
}
if self.keys.is_pressed(Keycode::D) {
delta += perp;
}
// Normalize delta so that the player doesn't move faster moving in a diagonal direction
if !delta.is_origin() {
delta = delta.normalize() * MOVE_SPEED;
}
self.position.add_x_y_raw(delta.x_y());
self.position.clamp(self.map.dims, PLAYER_WALL_PADDING);
}
pub fn update(&mut self) {
self.update_camera();
self.calculate_collisions();
}
pub fn draw_minimap(
&self,
canvas: &mut Canvas<Window>,
dims: (f64, f64),
) -> Result<(), String> {
let minimap_offset = (dims.0.max(dims.1) / 4.);
let minimap_base = Vec2D::new(minimap_offset, minimap_offset);
// Background
canvas.set_draw_color(Color::BLACK);
canvas.fill_rect(Rect::new(
minimap_offset as i32,
minimap_offset as i32,
dims.0 as u32,
dims.1 as u32,
))?;
canvas.set_draw_color(Color::WHITE);
// Player position
let position_mapped = self.position.remap(self.map.dims, dims) + minimap_base;
canvas.fill_rect(Rect::from_center(position_mapped, 8, 8))?;
// Player lines
let ray_scale = dims.0.max(dims.1) / 2.;
let half_fov = self.fov.to_radians() / 2.;
let forward_end = position_mapped + (Vec2D::from_angle(self.angle) * ray_scale);
let left_end = position_mapped + (Vec2D::from_angle(self.angle - half_fov) * ray_scale);
let right_end = position_mapped + (Vec2D::from_angle(self.angle + half_fov) * ray_scale);
canvas.draw_lines(&[
position_mapped.into(),
forward_end.into(),
position_mapped.into(),
left_end.into(),
position_mapped.into(),
right_end.into(),
] as &[Point])?;
// TODO: FOV lines
// Walls
for wall in self.map.walls.iter() {
canvas.set_draw_color(self.get_color(wall.color_index));
let start = wall.a.remap(self.map.dims, dims) + minimap_base;
let end = wall.b.remap(self.map.dims, dims) + minimap_base;
canvas.draw_line(start, end)?;
}
// let mut current_angle = self.angle + (self.fov.to_radians() / 2.);
// for _ in 0..self.resolution {
// let mut ray = self.position;
// ray.set_angle(current_angle);
// ray += self.position;
// let mut max_height = f64::NEG_INFINITY;
// let mut collisions: Vec<(bool, Vec2D)> = vec![];
// let mut collision = Vec2D::Origin;
// for wall in self.map.walls.iter() {
// if let Some(intersection_vector) = ray.intersects(wall) {
// let raw_distance = ray.dist(&intersection_vector);
// let delta = current_angle - self.angle;
// let corrected_distance = raw_distance * (delta.cos() as f64);
// let projected_height = self.projection_factor / corrected_distance;
// if projected_height > max_height {
// max_height = projected_height;
// collisions = collisions.into_iter().fold(vec![], |mut acc, cur| {
// acc.push((false, cur.1));
// acc
// });
// collisions.push((true, collision));
// collision = intersection_vector;
// } else {
// collisions.push((false, intersection_vector));
// }
// }
// }
// if !max_height.is_infinite() {
// canvas.set_draw_color(Color::RED);
// canvas.draw_rects(
// collisions
// .into_iter()
// .map(|(_, v)| {
// Rect::from_center(
// Point::new(
// remap(v.x(), 0., self.map.dims.0, 0., dims.0).floor() as i32,
// remap(v.y(), 0., self.map.dims.1, 0., dims.1).floor() as i32,
// ) + minimap_base,
// 2,
// 2,
// )
// })
// .collect::<Vec<Rect>>()
// .as_slice(),
// );
// }
// current_angle -= self.radian_per_column;
// }
Ok(())
}
fn render_frame(&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
let column_width_signed = self.column_width as i32;
// TODO: Draw background
let mut current_index = usize::MAX;
let mut current_color = Color::BLACK;
for (idx, (color_index, height)) in self.columns.iter().copied().enumerate() {
if color_index != current_index {
current_color = self.get_color(color_index);
current_index = color_index;
}
let dim_amt = remap(height as f64, 0, WINDOW_HEIGHT, 255, 0).floor() as u8;
canvas.set_draw_color(current_color.dim(dim_amt));
canvas.fill_rect(Rect::from_center(
Point::new(
idx as i32 * column_width_signed + (column_width_signed / 2),
WINDOW_HEIGHT as i32 / 2,
),
self.column_width,
height,
))?;
}
Ok(())
}
pub fn draw(&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
self.render_frame(canvas)?;
self.draw_minimap(canvas, (WINDOW_WIDTH as f64 / 5., WINDOW_WIDTH as f64 / 5.))?;
Ok(())
}
}
| {
self.wall_colors.get(index).copied().unwrap_or(Color::WHITE)
} | identifier_body |
state.rs | use std::collections::HashMap;
use num_traits::{AsPrimitive, Float};
use sdl2::{
keyboard::Keycode,
pixels::Color,
rect::{Point, Rect},
render::Canvas,
video::Window,
};
use crate::{
ext::ColorExt, key_state_handler::KeyStateHandler, map::Map, math::vector::Vec2D,
WINDOW_HEIGHT, WINDOW_WIDTH,
};
pub fn remap<
T: 'static + Float + Copy,
ON: AsPrimitive<T> + Copy,
OX: AsPrimitive<T> + Copy,
NN: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_min: ON,
old_max: OX,
new_min: NN,
new_max: NX,
) -> T {
new_min.as_()
+ (new_max.as_() - new_min.as_())
* ((value - old_min.as_()) / (old_max.as_() - old_min.as_()))
}
pub fn remap_minz<
T: 'static + Float + Copy + AsPrimitive<T>,
OX: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_max: OX,
new_max: NX,
) -> T {
remap(value, T::zero(), old_max, T::zero(), new_max)
}
const MOUSE_SENSITIVITY: f64 = 0.01;
const MOVE_SPEED: f64 = 2.5;
const PLAYER_WALL_PADDING: f64 = 10.;
const WALL_ACTUAL_HEIGHT: f64 = 48.;
pub struct State {
pub(crate) position: Vec2D,
pub(crate) angle: f64,
pub(crate) fov: f64,
pub(crate) wall_colors: Vec<Color>,
pub(crate) map: Map,
pub(crate) keys: KeyStateHandler,
pub(crate) columns: Vec<(usize, u32)>,
pub(crate) resolution: usize,
pub(crate) projection_factor: f64,
pub(crate) radian_per_column: f64,
pub(crate) column_width: u32,
}
impl State {
pub fn new() -> Self {
let wall_colors = vec![
Color::RGB(128, 255, 0),
Color::RGB(0, 128, 255),
Color::RGB(255, 0, 128),
Color::RGB(0, 255, 0),
Color::RGB(0, 0, 255),
Color::WHITE,
];
// let map = Map::default();
let map = Map::load("./assets/maps/many_walls.json").unwrap();
let (w, h) = map.dims;
// let movement_vector = Line::new(origin, origin + geo::Point::new(delta_x, delta_y));
let position = Vec2D::new(w as f64 / 2., 50. + h as f64 / 2.);
let fov = 60.;
let projection_plane_distance =
((WINDOW_WIDTH / 2) as f64 / (fov.to_radians() / 2.).tan()) as f64;
let resolution = WINDOW_WIDTH as usize;
Self {
position,
angle: std::f64::consts::PI,
fov,
wall_colors,
map,
keys: KeyStateHandler::new(),
columns: Vec::with_capacity(resolution),
resolution,
projection_factor: projection_plane_distance * WALL_ACTUAL_HEIGHT,
radian_per_column: fov.to_radians() / resolution as f64,
column_width: WINDOW_WIDTH / resolution as u32,
}
}
fn get_color(&self, index: usize) -> Color {
self.wall_colors.get(index).copied().unwrap_or(Color::WHITE)
}
pub fn mouse_motion(&mut self, dx: i32) {
self.angle += MOUSE_SENSITIVITY * dx as f64;
}
fn calculate_collisions(&mut self) {
let mut current_angle = self.angle - (self.fov.to_radians() / 2.);
let end_angle = current_angle + (self.radian_per_column * self.resolution as f64);
self.columns.clear();
for _ in 0..self.resolution {
let mut ray = Vec2D::from_angle(current_angle);
ray.translate(&self.position);
let mut max_height = f64::NEG_INFINITY;
let mut wall_color_index = 0;
for wall in self.map.walls.iter() {
if let Some(intersection_vector) = ray.intersects(wall) {
let raw_distance = ray.dist(&intersection_vector);
let delta = current_angle - self.angle;
let corrected_distance = raw_distance * (delta.cos() as f64);
let projected_height = self.projection_factor / corrected_distance;
if projected_height > max_height {
max_height = projected_height;
wall_color_index = wall.color_index;
}
}
}
if max_height.is_infinite() {
self.columns.push((0, 0));
} else {
self.columns
.push((wall_color_index, max_height.round() as u32));
}
current_angle += self.radian_per_column;
}
}
fn update_camera(&mut self) {
let mut delta = Vec2D::Origin;
let par = Vec2D::from_angle(self.angle as f64);
let perp = Vec2D::from_angle(self.angle + (90f64).to_radians());
if self.keys.is_pressed(Keycode::W) {
delta += par;
}
if self.keys.is_pressed(Keycode::S) {
delta += -par;
}
if self.keys.is_pressed(Keycode::A) {
delta += -perp;
}
if self.keys.is_pressed(Keycode::D) {
delta += perp;
}
// Normalize delta so that the player doesn't move faster moving in a diagonal direction
if !delta.is_origin() {
delta = delta.normalize() * MOVE_SPEED;
} | pub fn update(&mut self) {
self.update_camera();
self.calculate_collisions();
}
pub fn draw_minimap(
&self,
canvas: &mut Canvas<Window>,
dims: (f64, f64),
) -> Result<(), String> {
let minimap_offset = (dims.0.max(dims.1) / 4.);
let minimap_base = Vec2D::new(minimap_offset, minimap_offset);
// Background
canvas.set_draw_color(Color::BLACK);
canvas.fill_rect(Rect::new(
minimap_offset as i32,
minimap_offset as i32,
dims.0 as u32,
dims.1 as u32,
))?;
canvas.set_draw_color(Color::WHITE);
// Player position
let position_mapped = self.position.remap(self.map.dims, dims) + minimap_base;
canvas.fill_rect(Rect::from_center(position_mapped, 8, 8))?;
// Player lines
let ray_scale = dims.0.max(dims.1) / 2.;
let half_fov = self.fov.to_radians() / 2.;
let forward_end = position_mapped + (Vec2D::from_angle(self.angle) * ray_scale);
let left_end = position_mapped + (Vec2D::from_angle(self.angle - half_fov) * ray_scale);
let right_end = position_mapped + (Vec2D::from_angle(self.angle + half_fov) * ray_scale);
canvas.draw_lines(&[
position_mapped.into(),
forward_end.into(),
position_mapped.into(),
left_end.into(),
position_mapped.into(),
right_end.into(),
] as &[Point])?;
// TODO: FOV lines
// Walls
for wall in self.map.walls.iter() {
canvas.set_draw_color(self.get_color(wall.color_index));
let start = wall.a.remap(self.map.dims, dims) + minimap_base;
let end = wall.b.remap(self.map.dims, dims) + minimap_base;
canvas.draw_line(start, end)?;
}
// let mut current_angle = self.angle + (self.fov.to_radians() / 2.);
// for _ in 0..self.resolution {
// let mut ray = self.position;
// ray.set_angle(current_angle);
// ray += self.position;
// let mut max_height = f64::NEG_INFINITY;
// let mut collisions: Vec<(bool, Vec2D)> = vec![];
// let mut collision = Vec2D::Origin;
// for wall in self.map.walls.iter() {
// if let Some(intersection_vector) = ray.intersects(wall) {
// let raw_distance = ray.dist(&intersection_vector);
// let delta = current_angle - self.angle;
// let corrected_distance = raw_distance * (delta.cos() as f64);
// let projected_height = self.projection_factor / corrected_distance;
// if projected_height > max_height {
// max_height = projected_height;
// collisions = collisions.into_iter().fold(vec![], |mut acc, cur| {
// acc.push((false, cur.1));
// acc
// });
// collisions.push((true, collision));
// collision = intersection_vector;
// } else {
// collisions.push((false, intersection_vector));
// }
// }
// }
// if !max_height.is_infinite() {
// canvas.set_draw_color(Color::RED);
// canvas.draw_rects(
// collisions
// .into_iter()
// .map(|(_, v)| {
// Rect::from_center(
// Point::new(
// remap(v.x(), 0., self.map.dims.0, 0., dims.0).floor() as i32,
// remap(v.y(), 0., self.map.dims.1, 0., dims.1).floor() as i32,
// ) + minimap_base,
// 2,
// 2,
// )
// })
// .collect::<Vec<Rect>>()
// .as_slice(),
// );
// }
// current_angle -= self.radian_per_column;
// }
Ok(())
}
fn render_frame(&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
let column_width_signed = self.column_width as i32;
// TODO: Draw background
let mut current_index = usize::MAX;
let mut current_color = Color::BLACK;
for (idx, (color_index, height)) in self.columns.iter().copied().enumerate() {
if color_index != current_index {
current_color = self.get_color(color_index);
current_index = color_index;
}
let dim_amt = remap(height as f64, 0, WINDOW_HEIGHT, 255, 0).floor() as u8;
canvas.set_draw_color(current_color.dim(dim_amt));
canvas.fill_rect(Rect::from_center(
Point::new(
idx as i32 * column_width_signed + (column_width_signed / 2),
WINDOW_HEIGHT as i32 / 2,
),
self.column_width,
height,
))?;
}
Ok(())
}
pub fn draw(&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
self.render_frame(canvas)?;
self.draw_minimap(canvas, (WINDOW_WIDTH as f64 / 5., WINDOW_WIDTH as f64 / 5.))?;
Ok(())
}
} |
self.position.add_x_y_raw(delta.x_y());
self.position.clamp(self.map.dims, PLAYER_WALL_PADDING);
}
| random_line_split |
state.rs | use std::collections::HashMap;
use num_traits::{AsPrimitive, Float};
use sdl2::{
keyboard::Keycode,
pixels::Color,
rect::{Point, Rect},
render::Canvas,
video::Window,
};
use crate::{
ext::ColorExt, key_state_handler::KeyStateHandler, map::Map, math::vector::Vec2D,
WINDOW_HEIGHT, WINDOW_WIDTH,
};
pub fn remap<
T: 'static + Float + Copy,
ON: AsPrimitive<T> + Copy,
OX: AsPrimitive<T> + Copy,
NN: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_min: ON,
old_max: OX,
new_min: NN,
new_max: NX,
) -> T {
new_min.as_()
+ (new_max.as_() - new_min.as_())
* ((value - old_min.as_()) / (old_max.as_() - old_min.as_()))
}
pub fn remap_minz<
T: 'static + Float + Copy + AsPrimitive<T>,
OX: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_max: OX,
new_max: NX,
) -> T {
remap(value, T::zero(), old_max, T::zero(), new_max)
}
const MOUSE_SENSITIVITY: f64 = 0.01;
const MOVE_SPEED: f64 = 2.5;
const PLAYER_WALL_PADDING: f64 = 10.;
const WALL_ACTUAL_HEIGHT: f64 = 48.;
pub struct State {
pub(crate) position: Vec2D,
pub(crate) angle: f64,
pub(crate) fov: f64,
pub(crate) wall_colors: Vec<Color>,
pub(crate) map: Map,
pub(crate) keys: KeyStateHandler,
pub(crate) columns: Vec<(usize, u32)>,
pub(crate) resolution: usize,
pub(crate) projection_factor: f64,
pub(crate) radian_per_column: f64,
pub(crate) column_width: u32,
}
impl State {
pub fn new() -> Self {
let wall_colors = vec![
Color::RGB(128, 255, 0),
Color::RGB(0, 128, 255),
Color::RGB(255, 0, 128),
Color::RGB(0, 255, 0),
Color::RGB(0, 0, 255),
Color::WHITE,
];
// let map = Map::default();
let map = Map::load("./assets/maps/many_walls.json").unwrap();
let (w, h) = map.dims;
// let movement_vector = Line::new(origin, origin + geo::Point::new(delta_x, delta_y));
let position = Vec2D::new(w as f64 / 2., 50. + h as f64 / 2.);
let fov = 60.;
let projection_plane_distance =
((WINDOW_WIDTH / 2) as f64 / (fov.to_radians() / 2.).tan()) as f64;
let resolution = WINDOW_WIDTH as usize;
Self {
position,
angle: std::f64::consts::PI,
fov,
wall_colors,
map,
keys: KeyStateHandler::new(),
columns: Vec::with_capacity(resolution),
resolution,
projection_factor: projection_plane_distance * WALL_ACTUAL_HEIGHT,
radian_per_column: fov.to_radians() / resolution as f64,
column_width: WINDOW_WIDTH / resolution as u32,
}
}
fn get_color(&self, index: usize) -> Color {
self.wall_colors.get(index).copied().unwrap_or(Color::WHITE)
}
pub fn mouse_motion(&mut self, dx: i32) {
self.angle += MOUSE_SENSITIVITY * dx as f64;
}
fn calculate_collisions(&mut self) {
let mut current_angle = self.angle - (self.fov.to_radians() / 2.);
let end_angle = current_angle + (self.radian_per_column * self.resolution as f64);
self.columns.clear();
for _ in 0..self.resolution {
let mut ray = Vec2D::from_angle(current_angle);
ray.translate(&self.position);
let mut max_height = f64::NEG_INFINITY;
let mut wall_color_index = 0;
for wall in self.map.walls.iter() {
if let Some(intersection_vector) = ray.intersects(wall) |
}
if max_height.is_infinite() {
self.columns.push((0, 0));
} else {
self.columns
.push((wall_color_index, max_height.round() as u32));
}
current_angle += self.radian_per_column;
}
}
fn update_camera(&mut self) {
let mut delta = Vec2D::Origin;
let par = Vec2D::from_angle(self.angle as f64);
let perp = Vec2D::from_angle(self.angle + (90f64).to_radians());
if self.keys.is_pressed(Keycode::W) {
delta += par;
}
if self.keys.is_pressed(Keycode::S) {
delta += -par;
}
if self.keys.is_pressed(Keycode::A) {
delta += -perp;
}
if self.keys.is_pressed(Keycode::D) {
delta += perp;
}
// Normalize delta so that the player doesn't move faster moving in a diagonal direction
if !delta.is_origin() {
delta = delta.normalize() * MOVE_SPEED;
}
self.position.add_x_y_raw(delta.x_y());
self.position.clamp(self.map.dims, PLAYER_WALL_PADDING);
}
pub fn update(&mut self) {
self.update_camera();
self.calculate_collisions();
}
pub fn draw_minimap(
&self,
canvas: &mut Canvas<Window>,
dims: (f64, f64),
) -> Result<(), String> {
let minimap_offset = (dims.0.max(dims.1) / 4.);
let minimap_base = Vec2D::new(minimap_offset, minimap_offset);
// Background
canvas.set_draw_color(Color::BLACK);
canvas.fill_rect(Rect::new(
minimap_offset as i32,
minimap_offset as i32,
dims.0 as u32,
dims.1 as u32,
))?;
canvas.set_draw_color(Color::WHITE);
// Player position
let position_mapped = self.position.remap(self.map.dims, dims) + minimap_base;
canvas.fill_rect(Rect::from_center(position_mapped, 8, 8))?;
// Player lines
let ray_scale = dims.0.max(dims.1) / 2.;
let half_fov = self.fov.to_radians() / 2.;
let forward_end = position_mapped + (Vec2D::from_angle(self.angle) * ray_scale);
let left_end = position_mapped + (Vec2D::from_angle(self.angle - half_fov) * ray_scale);
let right_end = position_mapped + (Vec2D::from_angle(self.angle + half_fov) * ray_scale);
canvas.draw_lines(&[
position_mapped.into(),
forward_end.into(),
position_mapped.into(),
left_end.into(),
position_mapped.into(),
right_end.into(),
] as &[Point])?;
// TODO: FOV lines
// Walls
for wall in self.map.walls.iter() {
canvas.set_draw_color(self.get_color(wall.color_index));
let start = wall.a.remap(self.map.dims, dims) + minimap_base;
let end = wall.b.remap(self.map.dims, dims) + minimap_base;
canvas.draw_line(start, end)?;
}
// let mut current_angle = self.angle + (self.fov.to_radians() / 2.);
// for _ in 0..self.resolution {
// let mut ray = self.position;
// ray.set_angle(current_angle);
// ray += self.position;
// let mut max_height = f64::NEG_INFINITY;
// let mut collisions: Vec<(bool, Vec2D)> = vec![];
// let mut collision = Vec2D::Origin;
// for wall in self.map.walls.iter() {
// if let Some(intersection_vector) = ray.intersects(wall) {
// let raw_distance = ray.dist(&intersection_vector);
// let delta = current_angle - self.angle;
// let corrected_distance = raw_distance * (delta.cos() as f64);
// let projected_height = self.projection_factor / corrected_distance;
// if projected_height > max_height {
// max_height = projected_height;
// collisions = collisions.into_iter().fold(vec![], |mut acc, cur| {
// acc.push((false, cur.1));
// acc
// });
// collisions.push((true, collision));
// collision = intersection_vector;
// } else {
// collisions.push((false, intersection_vector));
// }
// }
// }
// if !max_height.is_infinite() {
// canvas.set_draw_color(Color::RED);
// canvas.draw_rects(
// collisions
// .into_iter()
// .map(|(_, v)| {
// Rect::from_center(
// Point::new(
// remap(v.x(), 0., self.map.dims.0, 0., dims.0).floor() as i32,
// remap(v.y(), 0., self.map.dims.1, 0., dims.1).floor() as i32,
// ) + minimap_base,
// 2,
// 2,
// )
// })
// .collect::<Vec<Rect>>()
// .as_slice(),
// );
// }
// current_angle -= self.radian_per_column;
// }
Ok(())
}
fn render_frame(&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
let column_width_signed = self.column_width as i32;
// TODO: Draw background
let mut current_index = usize::MAX;
let mut current_color = Color::BLACK;
for (idx, (color_index, height)) in self.columns.iter().copied().enumerate() {
if color_index != current_index {
current_color = self.get_color(color_index);
current_index = color_index;
}
let dim_amt = remap(height as f64, 0, WINDOW_HEIGHT, 255, 0).floor() as u8;
canvas.set_draw_color(current_color.dim(dim_amt));
canvas.fill_rect(Rect::from_center(
Point::new(
idx as i32 * column_width_signed + (column_width_signed / 2),
WINDOW_HEIGHT as i32 / 2,
),
self.column_width,
height,
))?;
}
Ok(())
}
pub fn draw(&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
self.render_frame(canvas)?;
self.draw_minimap(canvas, (WINDOW_WIDTH as f64 / 5., WINDOW_WIDTH as f64 / 5.))?;
Ok(())
}
}
| {
let raw_distance = ray.dist(&intersection_vector);
let delta = current_angle - self.angle;
let corrected_distance = raw_distance * (delta.cos() as f64);
let projected_height = self.projection_factor / corrected_distance;
if projected_height > max_height {
max_height = projected_height;
wall_color_index = wall.color_index;
}
} | conditional_block |
state.rs | use std::collections::HashMap;
use num_traits::{AsPrimitive, Float};
use sdl2::{
keyboard::Keycode,
pixels::Color,
rect::{Point, Rect},
render::Canvas,
video::Window,
};
use crate::{
ext::ColorExt, key_state_handler::KeyStateHandler, map::Map, math::vector::Vec2D,
WINDOW_HEIGHT, WINDOW_WIDTH,
};
pub fn remap<
T: 'static + Float + Copy,
ON: AsPrimitive<T> + Copy,
OX: AsPrimitive<T> + Copy,
NN: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_min: ON,
old_max: OX,
new_min: NN,
new_max: NX,
) -> T {
new_min.as_()
+ (new_max.as_() - new_min.as_())
* ((value - old_min.as_()) / (old_max.as_() - old_min.as_()))
}
pub fn remap_minz<
T: 'static + Float + Copy + AsPrimitive<T>,
OX: AsPrimitive<T> + Copy,
NX: AsPrimitive<T> + Copy,
>(
value: T,
old_max: OX,
new_max: NX,
) -> T {
remap(value, T::zero(), old_max, T::zero(), new_max)
}
const MOUSE_SENSITIVITY: f64 = 0.01;
const MOVE_SPEED: f64 = 2.5;
const PLAYER_WALL_PADDING: f64 = 10.;
const WALL_ACTUAL_HEIGHT: f64 = 48.;
pub struct State {
pub(crate) position: Vec2D,
pub(crate) angle: f64,
pub(crate) fov: f64,
pub(crate) wall_colors: Vec<Color>,
pub(crate) map: Map,
pub(crate) keys: KeyStateHandler,
pub(crate) columns: Vec<(usize, u32)>,
pub(crate) resolution: usize,
pub(crate) projection_factor: f64,
pub(crate) radian_per_column: f64,
pub(crate) column_width: u32,
}
impl State {
pub fn new() -> Self {
let wall_colors = vec![
Color::RGB(128, 255, 0),
Color::RGB(0, 128, 255),
Color::RGB(255, 0, 128),
Color::RGB(0, 255, 0),
Color::RGB(0, 0, 255),
Color::WHITE,
];
// let map = Map::default();
let map = Map::load("./assets/maps/many_walls.json").unwrap();
let (w, h) = map.dims;
// let movement_vector = Line::new(origin, origin + geo::Point::new(delta_x, delta_y));
let position = Vec2D::new(w as f64 / 2., 50. + h as f64 / 2.);
let fov = 60.;
let projection_plane_distance =
((WINDOW_WIDTH / 2) as f64 / (fov.to_radians() / 2.).tan()) as f64;
let resolution = WINDOW_WIDTH as usize;
Self {
position,
angle: std::f64::consts::PI,
fov,
wall_colors,
map,
keys: KeyStateHandler::new(),
columns: Vec::with_capacity(resolution),
resolution,
projection_factor: projection_plane_distance * WALL_ACTUAL_HEIGHT,
radian_per_column: fov.to_radians() / resolution as f64,
column_width: WINDOW_WIDTH / resolution as u32,
}
}
fn get_color(&self, index: usize) -> Color {
self.wall_colors.get(index).copied().unwrap_or(Color::WHITE)
}
pub fn mouse_motion(&mut self, dx: i32) {
self.angle += MOUSE_SENSITIVITY * dx as f64;
}
fn calculate_collisions(&mut self) {
let mut current_angle = self.angle - (self.fov.to_radians() / 2.);
let end_angle = current_angle + (self.radian_per_column * self.resolution as f64);
self.columns.clear();
for _ in 0..self.resolution {
let mut ray = Vec2D::from_angle(current_angle);
ray.translate(&self.position);
let mut max_height = f64::NEG_INFINITY;
let mut wall_color_index = 0;
for wall in self.map.walls.iter() {
if let Some(intersection_vector) = ray.intersects(wall) {
let raw_distance = ray.dist(&intersection_vector);
let delta = current_angle - self.angle;
let corrected_distance = raw_distance * (delta.cos() as f64);
let projected_height = self.projection_factor / corrected_distance;
if projected_height > max_height {
max_height = projected_height;
wall_color_index = wall.color_index;
}
}
}
if max_height.is_infinite() {
self.columns.push((0, 0));
} else {
self.columns
.push((wall_color_index, max_height.round() as u32));
}
current_angle += self.radian_per_column;
}
}
fn update_camera(&mut self) {
let mut delta = Vec2D::Origin;
let par = Vec2D::from_angle(self.angle as f64);
let perp = Vec2D::from_angle(self.angle + (90f64).to_radians());
if self.keys.is_pressed(Keycode::W) {
delta += par;
}
if self.keys.is_pressed(Keycode::S) {
delta += -par;
}
if self.keys.is_pressed(Keycode::A) {
delta += -perp;
}
if self.keys.is_pressed(Keycode::D) {
delta += perp;
}
// Normalize delta so that the player doesn't move faster moving in a diagonal direction
if !delta.is_origin() {
delta = delta.normalize() * MOVE_SPEED;
}
self.position.add_x_y_raw(delta.x_y());
self.position.clamp(self.map.dims, PLAYER_WALL_PADDING);
}
pub fn update(&mut self) {
self.update_camera();
self.calculate_collisions();
}
pub fn draw_minimap(
&self,
canvas: &mut Canvas<Window>,
dims: (f64, f64),
) -> Result<(), String> {
let minimap_offset = (dims.0.max(dims.1) / 4.);
let minimap_base = Vec2D::new(minimap_offset, minimap_offset);
// Background
canvas.set_draw_color(Color::BLACK);
canvas.fill_rect(Rect::new(
minimap_offset as i32,
minimap_offset as i32,
dims.0 as u32,
dims.1 as u32,
))?;
canvas.set_draw_color(Color::WHITE);
// Player position
let position_mapped = self.position.remap(self.map.dims, dims) + minimap_base;
canvas.fill_rect(Rect::from_center(position_mapped, 8, 8))?;
// Player lines
let ray_scale = dims.0.max(dims.1) / 2.;
let half_fov = self.fov.to_radians() / 2.;
let forward_end = position_mapped + (Vec2D::from_angle(self.angle) * ray_scale);
let left_end = position_mapped + (Vec2D::from_angle(self.angle - half_fov) * ray_scale);
let right_end = position_mapped + (Vec2D::from_angle(self.angle + half_fov) * ray_scale);
canvas.draw_lines(&[
position_mapped.into(),
forward_end.into(),
position_mapped.into(),
left_end.into(),
position_mapped.into(),
right_end.into(),
] as &[Point])?;
// TODO: FOV lines
// Walls
for wall in self.map.walls.iter() {
canvas.set_draw_color(self.get_color(wall.color_index));
let start = wall.a.remap(self.map.dims, dims) + minimap_base;
let end = wall.b.remap(self.map.dims, dims) + minimap_base;
canvas.draw_line(start, end)?;
}
// let mut current_angle = self.angle + (self.fov.to_radians() / 2.);
// for _ in 0..self.resolution {
// let mut ray = self.position;
// ray.set_angle(current_angle);
// ray += self.position;
// let mut max_height = f64::NEG_INFINITY;
// let mut collisions: Vec<(bool, Vec2D)> = vec![];
// let mut collision = Vec2D::Origin;
// for wall in self.map.walls.iter() {
// if let Some(intersection_vector) = ray.intersects(wall) {
// let raw_distance = ray.dist(&intersection_vector);
// let delta = current_angle - self.angle;
// let corrected_distance = raw_distance * (delta.cos() as f64);
// let projected_height = self.projection_factor / corrected_distance;
// if projected_height > max_height {
// max_height = projected_height;
// collisions = collisions.into_iter().fold(vec![], |mut acc, cur| {
// acc.push((false, cur.1));
// acc
// });
// collisions.push((true, collision));
// collision = intersection_vector;
// } else {
// collisions.push((false, intersection_vector));
// }
// }
// }
// if !max_height.is_infinite() {
// canvas.set_draw_color(Color::RED);
// canvas.draw_rects(
// collisions
// .into_iter()
// .map(|(_, v)| {
// Rect::from_center(
// Point::new(
// remap(v.x(), 0., self.map.dims.0, 0., dims.0).floor() as i32,
// remap(v.y(), 0., self.map.dims.1, 0., dims.1).floor() as i32,
// ) + minimap_base,
// 2,
// 2,
// )
// })
// .collect::<Vec<Rect>>()
// .as_slice(),
// );
// }
// current_angle -= self.radian_per_column;
// }
Ok(())
}
fn render_frame(&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
let column_width_signed = self.column_width as i32;
// TODO: Draw background
let mut current_index = usize::MAX;
let mut current_color = Color::BLACK;
for (idx, (color_index, height)) in self.columns.iter().copied().enumerate() {
if color_index != current_index {
current_color = self.get_color(color_index);
current_index = color_index;
}
let dim_amt = remap(height as f64, 0, WINDOW_HEIGHT, 255, 0).floor() as u8;
canvas.set_draw_color(current_color.dim(dim_amt));
canvas.fill_rect(Rect::from_center(
Point::new(
idx as i32 * column_width_signed + (column_width_signed / 2),
WINDOW_HEIGHT as i32 / 2,
),
self.column_width,
height,
))?;
}
Ok(())
}
pub fn | (&self, canvas: &mut Canvas<Window>) -> Result<(), String> {
self.render_frame(canvas)?;
self.draw_minimap(canvas, (WINDOW_WIDTH as f64 / 5., WINDOW_WIDTH as f64 / 5.))?;
Ok(())
}
}
| draw | identifier_name |
opdb.rs | /*
@author: xiao cai niao
@datetime: 2019/11/6
*/
use actix_web::{web};
use crate::webroute::route::{HostInfo, PostUserInfo, EditInfo, EditMainTain};
use crate::storage::rocks::{DbInfo, KeyValue, CfNameTypeCode, PrefixTypeCode};
use crate::ha::procotol::{DownNodeCheck, RecoveryInfo, ReplicationState, MysqlMonitorStatus};
use std::error::Error;
use crate::ha::nodes_manager::{SlaveInfo};
use serde::{Serialize, Deserialize};
use crate::rand_string;
use crate::ha::procotol::MysqlState;
use crate::ha::sys_manager::MonitorSetting;
use crate::webroute::new_route::ResponseMonitorStatic;
use std::str::from_utf8;
///
/// mysql node info, insert to rocksdb
///
///
pub fn insert_mysql_host_info(data: web::Data<DbInfo>, info: &web::Json<HostInfo>) -> Result<(), Box<dyn Error>> {
let check_unique = data.get(&info.host, &CfNameTypeCode::HaNodesInfo.get());
match check_unique {
Ok(v) => {
if v.value.len() > 0 {
let a = format!("this key: ({}) already exists in the database",&info.host);
return Err(a.into());
}
}
_ => {}
}
let v = HostInfoValue::new(info)?;
v.save(&data)?;
//初始化节点监控配置
let monitor_info = MonitorSetting::new(&info.host);
monitor_info.save(&data)?;
Ok(())
}
#[derive(Serialize, Deserialize, Debug)]
pub struct HaChangeLog {
pub key: String, //格式 host_timestamp host为宕机节点
pub cluster_name: String,
pub old_master_info: DownNodeCheck, //宕机节点信息
pub new_master_binlog_info: SlaveInfo, //如果宕机节点在切换之前未进行binlog追加将保存新master读取到的binlog信息,在宕机节点恢复时会进行判断回滚
pub recovery_info: RecoveryInfo, //宕机恢复同步所需的新master信息
pub recovery_status: bool, //是否已恢复
pub switch_status: bool, //切换状态
}
impl HaChangeLog {
pub fn new() -> HaChangeLog {
HaChangeLog{
key: "".to_string(),
cluster_name: "".to_string(),
old_master_info: DownNodeCheck { host: "".to_string(), dbport: 0 },
new_master_binlog_info: SlaveInfo {
host: "".to_string(),
dbport: 0,
slave_info: ReplicationState {
log_name: "".to_string(),
read_log_pos: 0,
exec_log_pos: 0
},
new_master: false
},
recovery_info: RecoveryInfo {
binlog: "".to_string(),
position: 0,
gtid: "".to_string(),
masterhost: "".to_string(),
masterport: 0,
read_binlog: "".to_string(),
read_position: 0
},
recovery_status: false,
switch_status: false
}
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let key = format!("{}_{}",self.key.clone(), crate::timestamp());
let value = serde_json::to_string(self)?;
let row = KeyValue{key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
pub fn update(&mut self, db: &web::Data<DbInfo>, row_key: String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(self)?;
let row = KeyValue{key: row_key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
}
///
///
///
/// 用户信息结构
#[derive(Serialize, Deserialize, Clone, Debug)]
pub struct UserInfo {
pub user_name: String,
pub password: String,
pub hook_id: String,
pub create_time: i64,
pub update_time: i64
}
impl UserInfo {
pub fn new(info: &PostUserInfo) -> UserInfo {
let create_time = crate::timestamp();
let update_time = crate::timestamp();
UserInfo{
user_name: info.user_name.clone(),
password: info.password.clone(),
hook_id: rand_string(),
create_time,
update_time
}
}
}
///
///
///
/// 节点基础信息, host做为key
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct HostInfoValue {
pub host: String, //127.0.0.1:3306
pub dbport: usize, //default 3306
pub rtype: String, //db、route
pub cluster_name: String, //集群名称,route类型默认default
pub online: bool, //db是否在线, true、false
pub insert_time: i64,
pub update_time: i64,
pub maintain: bool, //是否处于维护模式,true、false
}
impl HostInfoValue {
pub fn new(info: &HostInfo) -> Result<HostInfoValue, Box<dyn Error>> {
let h = HostInfoValue{
host: info.host.clone(),
rtype: info.rtype.clone(),
dbport: info.dbport.clone(),
cluster_name: info.cluster_name.clone(),
online: false,
insert_time: crate::timestamp(),
update_time: crate::timestamp(),
maintain: false
};
Ok(h)
}
///
/// 写入db
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let row = KeyValue{key: self.host.clone(), value};
db.put(&row, &CfNameTypeCode::HaNodesInfo.get())?;
Ok(())
}
///
/// 编辑节点信息
pub fn edit(&mut self, info: &web::Json<EditInfo>) {
self.host = info.host.clone();
self.dbport = info.dbport.clone();
self.cluster_name = info.cluster_name.clone();
self.update_time = crate::timestamp();
}
///
/// 设置节点维护模式状态
pub fn maintain(&mut self, info: &web::Json<EditMainTain>) {
if info.maintain == "true".to_string() {
self.maintain = false;
}else {
self.maintain = true;
}
self.update_time = crate::timestamp();
}
///
/// 获取当前节点在db中保存的状态信息
pub fn get_state(&self, db: &web::Data<DbInfo>) -> Result<MysqlState, Box<dyn Error>> {
let kv = db.get(&self.host, &CfNameTypeCode::NodesState.get())?;
if kv.value.len() > 0 {
let state: MysqlState = serde_json::from_str(&kv.value)?;
return Ok(state);
}else {
//let err = format!("this host: {} no state data", &self.host);
//return Err(err.into());
let state = MysqlState::new();
return Ok(state);
}
}
pub fn get_role(&self, db: &web::Data<DbInfo>) -> Result<String, Box<dyn Error>> {
let state = self.get_state(db)?;
Ok(state.role)
}
}
///
///
///
///
/// 获取db中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct NodeClusterList{
pub cluster_name_list: Vec<String>
}
impl NodeClusterList{
pub fn new() -> NodeClusterList{
NodeClusterList { cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let value: HostInfoValue = serde_json::from_str(&row.value)?;
if self.is_exists(&value.cluster_name){continue;}
self.cluster_name_list.push(value.cluster_name.clone());
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name{
return true;
}
}
return false;
}
}
/// 获取route信息中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct RouteClusterList{
pub cluster_name_list: Vec<String>
}
impl RouteClusterList{
pub fn new() -> RouteClusterList {
RouteClusterList{ cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let route_all = db.get_route_all()?;
for route in &route_all {
if !self.is_exists(&route.value.cluster_name){
self.cluster_name_list.push(route.value.cluster_name.clone());
}
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name {
return true;
}
}
return false;
}
}
///
///
///
///
/// node节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct NodeInfo{
pub cluster_name: String,
pub host: String,
pub dbport: usize,
pub online: bool, //是否在线, true、false
pub maintain: bool, //是否处于维护模式,true、false
pub role: String, //主从角色
pub master: String,
pub sql_thread: bool,
pub io_thread: bool,
pub seconds_behind: usize,
pub read_only: bool,
pub version: String,
pub executed_gtid_set: String,
pub innodb_flush_log_at_trx_commit: usize,
pub sync_binlog: usize,
pub server_id: usize,
pub event_scheduler: String,
pub sql_error: String
}
impl NodeInfo{
pub fn new(state: &MysqlState, node: &HostInfoValue) -> NodeInfo {
let mut ni = NodeInfo{
cluster_name: node.cluster_name.clone(),
host: node.host.clone(),
dbport: node.dbport.clone(),
online: node.online.clone(),
maintain: node.maintain.clone(),
role: state.role.clone(),
master: state.master.clone(),
sql_thread: state.sql_thread.clone(),
io_thread: state.io_thread.clone(),
seconds_behind: state.seconds_behind.clone(),
read_only: state.read_only.clone(),
version: state.version.clone(),
executed_gtid_set: state.executed_gtid_set.clone(),
innodb_flush_log_at_trx_commit: state.innodb_flush_log_at_trx_commit.clone(),
sync_binlog: state.sync_binlog.clone(),
server_id: state.server_id.clone(),
event_scheduler: state.event_scheduler.clone(),
sql_error: "".to_string()
};
if state.last_io_error.len() > 0{
ni.sql_error = state.last_io_error.clone();
}else if state.last_sql_error.len() > 0 {
ni.sql_error = state.last_sql_error.clone();
}
return ni;
}
}
///
///
///
///
/// 每个集群节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct ClusterNodeInfo{
pub cluster_name: String,
pub total: usize,
pub nodes_info: Vec<NodeInfo>
}
impl ClusterNodeInfo{
pub fn new(cluster_name: &String) -> ClusterNodeInfo{
ClusterNodeInfo{
cluster_name: cluster_name.clone(),
total: 0,
nodes_info: vec![]
}
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let node: HostInfoValue = serde_json::from_str(&row.value)?;
if &node.cluster_name == &self.cluster_name{
let state = node.get_state(db)?;
let node_info = NodeInfo::new(&state, &node);
self.total += 1;
self.nodes_info.push(node_info);
}
}
Ok(())
}
///
/// 统计所有节点监控信息, 用于首页展示
///
/// 倒叙迭代获取每个节点最后一条数据, 如果每个节点都已获取最后一条数据就退出迭代
pub fn static_monitor(&self, db: &web::Data<DbInfo>, rsm: &mut ResponseMonitorStatic) -> Result<(), Box<dyn Error>> {
let cf_name = CfNameTypeCode::SystemData.get();
let mut tmp: Vec<String> = vec![];
//首先检查是否开启监控
for node in &self.nodes_info{
if !self.check_monitor_setting(db, &node.host){
tmp.push(node.host.clone());
}
}
if let Some(cf) = db.db.cf_handle(&cf_name){
let mut iter = db.db.raw_iterator_cf(cf)?;
iter.seek_to_last();
iter.prev();
'all: while iter.valid() {
if tmp.len() == self.nodes_info.len(){
break 'all;
}
if let Some(s) = iter.key(){
let key: String = from_utf8(&s.to_vec())?.parse()?;
if key.starts_with(&PrefixTypeCode::NodeMonitorData.prefix()){
'b: for n in &self.nodes_info{
if key.contains(n.host.as_str()){
for t in &tmp{
if t == &n.host{
break 'b;
}
}
if let Some(v) = iter.value(){
let v: MysqlMonitorStatus = serde_json::from_slice(&v)?;
rsm.update(&v);
tmp.push(n.host.clone());
}
break 'b;
}
}
}
}
//
//
iter.prev();
}
}
Ok(())
}
fn check_monitor_setting(&self, db: &web::Data<DbInfo>, host: &String) -> bool{
let a = db.prefix_get(&PrefixTypeCode::NodeMonitorSeting, host);
match a {
Ok(v) => {
if v.value.len() > 0{
let value: MonitorSetting = serde_json::from_str(&v.value).unwrap();
return value.monitor.clone();
}
}
Err(e) => {
info!("{}", e.to_string());
}
}
return false;
}
}
///
///
impl MysqlState{
pub fn save(&self, db: &web::Data<DbInfo>, key: &String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let a = KeyValue{key: key.clone(), value };
db.put(&a, &CfNameTyp | pub cluster_name: String,
pub delay: usize
}
impl SlaveBehindSetting{
pub fn new(cluster_name: &String) -> SlaveBehindSetting {
SlaveBehindSetting{ cluster_name: cluster_name.clone(), delay: 100 }
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
db.prefix_put(&PrefixTypeCode::SlaveDelaySeting, &self.cluster_name, &self)?;
Ok(())
}
}
| eCode::NodesState.get())?;
Ok(())
}
}
///
///
/// slave behind 配置结构体
#[derive(Serialize, Deserialize, Debug)]
pub struct SlaveBehindSetting{
| conditional_block |
opdb.rs | /*
@author: xiao cai niao
@datetime: 2019/11/6
*/
use actix_web::{web};
use crate::webroute::route::{HostInfo, PostUserInfo, EditInfo, EditMainTain};
use crate::storage::rocks::{DbInfo, KeyValue, CfNameTypeCode, PrefixTypeCode};
use crate::ha::procotol::{DownNodeCheck, RecoveryInfo, ReplicationState, MysqlMonitorStatus};
use std::error::Error;
use crate::ha::nodes_manager::{SlaveInfo};
use serde::{Serialize, Deserialize};
use crate::rand_string;
use crate::ha::procotol::MysqlState;
use crate::ha::sys_manager::MonitorSetting;
use crate::webroute::new_route::ResponseMonitorStatic;
use std::str::from_utf8;
///
/// mysql node info, insert to rocksdb
///
///
pub fn in | ata: web::Data<DbInfo>, info: &web::Json<HostInfo>) -> Result<(), Box<dyn Error>> {
let check_unique = data.get(&info.host, &CfNameTypeCode::HaNodesInfo.get());
match check_unique {
Ok(v) => {
if v.value.len() > 0 {
let a = format!("this key: ({}) already exists in the database",&info.host);
return Err(a.into());
}
}
_ => {}
}
let v = HostInfoValue::new(info)?;
v.save(&data)?;
//初始化节点监控配置
let monitor_info = MonitorSetting::new(&info.host);
monitor_info.save(&data)?;
Ok(())
}
#[derive(Serialize, Deserialize, Debug)]
pub struct HaChangeLog {
pub key: String, //格式 host_timestamp host为宕机节点
pub cluster_name: String,
pub old_master_info: DownNodeCheck, //宕机节点信息
pub new_master_binlog_info: SlaveInfo, //如果宕机节点在切换之前未进行binlog追加将保存新master读取到的binlog信息,在宕机节点恢复时会进行判断回滚
pub recovery_info: RecoveryInfo, //宕机恢复同步所需的新master信息
pub recovery_status: bool, //是否已恢复
pub switch_status: bool, //切换状态
}
impl HaChangeLog {
pub fn new() -> HaChangeLog {
HaChangeLog{
key: "".to_string(),
cluster_name: "".to_string(),
old_master_info: DownNodeCheck { host: "".to_string(), dbport: 0 },
new_master_binlog_info: SlaveInfo {
host: "".to_string(),
dbport: 0,
slave_info: ReplicationState {
log_name: "".to_string(),
read_log_pos: 0,
exec_log_pos: 0
},
new_master: false
},
recovery_info: RecoveryInfo {
binlog: "".to_string(),
position: 0,
gtid: "".to_string(),
masterhost: "".to_string(),
masterport: 0,
read_binlog: "".to_string(),
read_position: 0
},
recovery_status: false,
switch_status: false
}
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let key = format!("{}_{}",self.key.clone(), crate::timestamp());
let value = serde_json::to_string(self)?;
let row = KeyValue{key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
pub fn update(&mut self, db: &web::Data<DbInfo>, row_key: String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(self)?;
let row = KeyValue{key: row_key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
}
///
///
///
/// 用户信息结构
#[derive(Serialize, Deserialize, Clone, Debug)]
pub struct UserInfo {
pub user_name: String,
pub password: String,
pub hook_id: String,
pub create_time: i64,
pub update_time: i64
}
impl UserInfo {
pub fn new(info: &PostUserInfo) -> UserInfo {
let create_time = crate::timestamp();
let update_time = crate::timestamp();
UserInfo{
user_name: info.user_name.clone(),
password: info.password.clone(),
hook_id: rand_string(),
create_time,
update_time
}
}
}
///
///
///
/// 节点基础信息, host做为key
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct HostInfoValue {
pub host: String, //127.0.0.1:3306
pub dbport: usize, //default 3306
pub rtype: String, //db、route
pub cluster_name: String, //集群名称,route类型默认default
pub online: bool, //db是否在线, true、false
pub insert_time: i64,
pub update_time: i64,
pub maintain: bool, //是否处于维护模式,true、false
}
impl HostInfoValue {
pub fn new(info: &HostInfo) -> Result<HostInfoValue, Box<dyn Error>> {
let h = HostInfoValue{
host: info.host.clone(),
rtype: info.rtype.clone(),
dbport: info.dbport.clone(),
cluster_name: info.cluster_name.clone(),
online: false,
insert_time: crate::timestamp(),
update_time: crate::timestamp(),
maintain: false
};
Ok(h)
}
///
/// 写入db
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let row = KeyValue{key: self.host.clone(), value};
db.put(&row, &CfNameTypeCode::HaNodesInfo.get())?;
Ok(())
}
///
/// 编辑节点信息
pub fn edit(&mut self, info: &web::Json<EditInfo>) {
self.host = info.host.clone();
self.dbport = info.dbport.clone();
self.cluster_name = info.cluster_name.clone();
self.update_time = crate::timestamp();
}
///
/// 设置节点维护模式状态
pub fn maintain(&mut self, info: &web::Json<EditMainTain>) {
if info.maintain == "true".to_string() {
self.maintain = false;
}else {
self.maintain = true;
}
self.update_time = crate::timestamp();
}
///
/// 获取当前节点在db中保存的状态信息
pub fn get_state(&self, db: &web::Data<DbInfo>) -> Result<MysqlState, Box<dyn Error>> {
let kv = db.get(&self.host, &CfNameTypeCode::NodesState.get())?;
if kv.value.len() > 0 {
let state: MysqlState = serde_json::from_str(&kv.value)?;
return Ok(state);
}else {
//let err = format!("this host: {} no state data", &self.host);
//return Err(err.into());
let state = MysqlState::new();
return Ok(state);
}
}
pub fn get_role(&self, db: &web::Data<DbInfo>) -> Result<String, Box<dyn Error>> {
let state = self.get_state(db)?;
Ok(state.role)
}
}
///
///
///
///
/// 获取db中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct NodeClusterList{
pub cluster_name_list: Vec<String>
}
impl NodeClusterList{
pub fn new() -> NodeClusterList{
NodeClusterList { cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let value: HostInfoValue = serde_json::from_str(&row.value)?;
if self.is_exists(&value.cluster_name){continue;}
self.cluster_name_list.push(value.cluster_name.clone());
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name{
return true;
}
}
return false;
}
}
/// 获取route信息中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct RouteClusterList{
pub cluster_name_list: Vec<String>
}
impl RouteClusterList{
pub fn new() -> RouteClusterList {
RouteClusterList{ cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let route_all = db.get_route_all()?;
for route in &route_all {
if !self.is_exists(&route.value.cluster_name){
self.cluster_name_list.push(route.value.cluster_name.clone());
}
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name {
return true;
}
}
return false;
}
}
///
///
///
///
/// node节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct NodeInfo{
pub cluster_name: String,
pub host: String,
pub dbport: usize,
pub online: bool, //是否在线, true、false
pub maintain: bool, //是否处于维护模式,true、false
pub role: String, //主从角色
pub master: String,
pub sql_thread: bool,
pub io_thread: bool,
pub seconds_behind: usize,
pub read_only: bool,
pub version: String,
pub executed_gtid_set: String,
pub innodb_flush_log_at_trx_commit: usize,
pub sync_binlog: usize,
pub server_id: usize,
pub event_scheduler: String,
pub sql_error: String
}
impl NodeInfo{
pub fn new(state: &MysqlState, node: &HostInfoValue) -> NodeInfo {
let mut ni = NodeInfo{
cluster_name: node.cluster_name.clone(),
host: node.host.clone(),
dbport: node.dbport.clone(),
online: node.online.clone(),
maintain: node.maintain.clone(),
role: state.role.clone(),
master: state.master.clone(),
sql_thread: state.sql_thread.clone(),
io_thread: state.io_thread.clone(),
seconds_behind: state.seconds_behind.clone(),
read_only: state.read_only.clone(),
version: state.version.clone(),
executed_gtid_set: state.executed_gtid_set.clone(),
innodb_flush_log_at_trx_commit: state.innodb_flush_log_at_trx_commit.clone(),
sync_binlog: state.sync_binlog.clone(),
server_id: state.server_id.clone(),
event_scheduler: state.event_scheduler.clone(),
sql_error: "".to_string()
};
if state.last_io_error.len() > 0{
ni.sql_error = state.last_io_error.clone();
}else if state.last_sql_error.len() > 0 {
ni.sql_error = state.last_sql_error.clone();
}
return ni;
}
}
///
///
///
///
/// 每个集群节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct ClusterNodeInfo{
pub cluster_name: String,
pub total: usize,
pub nodes_info: Vec<NodeInfo>
}
impl ClusterNodeInfo{
pub fn new(cluster_name: &String) -> ClusterNodeInfo{
ClusterNodeInfo{
cluster_name: cluster_name.clone(),
total: 0,
nodes_info: vec![]
}
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let node: HostInfoValue = serde_json::from_str(&row.value)?;
if &node.cluster_name == &self.cluster_name{
let state = node.get_state(db)?;
let node_info = NodeInfo::new(&state, &node);
self.total += 1;
self.nodes_info.push(node_info);
}
}
Ok(())
}
///
/// 统计所有节点监控信息, 用于首页展示
///
/// 倒叙迭代获取每个节点最后一条数据, 如果每个节点都已获取最后一条数据就退出迭代
pub fn static_monitor(&self, db: &web::Data<DbInfo>, rsm: &mut ResponseMonitorStatic) -> Result<(), Box<dyn Error>> {
let cf_name = CfNameTypeCode::SystemData.get();
let mut tmp: Vec<String> = vec![];
//首先检查是否开启监控
for node in &self.nodes_info{
if !self.check_monitor_setting(db, &node.host){
tmp.push(node.host.clone());
}
}
if let Some(cf) = db.db.cf_handle(&cf_name){
let mut iter = db.db.raw_iterator_cf(cf)?;
iter.seek_to_last();
iter.prev();
'all: while iter.valid() {
if tmp.len() == self.nodes_info.len(){
break 'all;
}
if let Some(s) = iter.key(){
let key: String = from_utf8(&s.to_vec())?.parse()?;
if key.starts_with(&PrefixTypeCode::NodeMonitorData.prefix()){
'b: for n in &self.nodes_info{
if key.contains(n.host.as_str()){
for t in &tmp{
if t == &n.host{
break 'b;
}
}
if let Some(v) = iter.value(){
let v: MysqlMonitorStatus = serde_json::from_slice(&v)?;
rsm.update(&v);
tmp.push(n.host.clone());
}
break 'b;
}
}
}
}
//
//
iter.prev();
}
}
Ok(())
}
fn check_monitor_setting(&self, db: &web::Data<DbInfo>, host: &String) -> bool{
let a = db.prefix_get(&PrefixTypeCode::NodeMonitorSeting, host);
match a {
Ok(v) => {
if v.value.len() > 0{
let value: MonitorSetting = serde_json::from_str(&v.value).unwrap();
return value.monitor.clone();
}
}
Err(e) => {
info!("{}", e.to_string());
}
}
return false;
}
}
///
///
impl MysqlState{
pub fn save(&self, db: &web::Data<DbInfo>, key: &String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let a = KeyValue{key: key.clone(), value };
db.put(&a, &CfNameTypeCode::NodesState.get())?;
Ok(())
}
}
///
///
/// slave behind 配置结构体
#[derive(Serialize, Deserialize, Debug)]
pub struct SlaveBehindSetting{
pub cluster_name: String,
pub delay: usize
}
impl SlaveBehindSetting{
pub fn new(cluster_name: &String) -> SlaveBehindSetting {
SlaveBehindSetting{ cluster_name: cluster_name.clone(), delay: 100 }
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
db.prefix_put(&PrefixTypeCode::SlaveDelaySeting, &self.cluster_name, &self)?;
Ok(())
}
}
| sert_mysql_host_info(d | identifier_name |
opdb.rs | /*
@author: xiao cai niao
@datetime: 2019/11/6
*/
use actix_web::{web};
use crate::webroute::route::{HostInfo, PostUserInfo, EditInfo, EditMainTain};
use crate::storage::rocks::{DbInfo, KeyValue, CfNameTypeCode, PrefixTypeCode};
use crate::ha::procotol::{DownNodeCheck, RecoveryInfo, ReplicationState, MysqlMonitorStatus};
use std::error::Error;
use crate::ha::nodes_manager::{SlaveInfo};
use serde::{Serialize, Deserialize};
use crate::rand_string;
use crate::ha::procotol::MysqlState;
use crate::ha::sys_manager::MonitorSetting;
use crate::webroute::new_route::ResponseMonitorStatic;
use std::str::from_utf8;
///
/// mysql node info, insert to rocksdb
///
///
pub fn insert_mysql_host_info(data: web::Data<DbInfo>, info: &web::Json<HostInfo>) -> Result<(), Box<dyn Error>> {
let check_unique = data.get(&info.host, &CfNameTypeCode::HaNodesInfo.get());
match check_unique {
Ok(v) => {
if v.value.len() > 0 {
let a = format!("this key: ({}) already exists in the database",&info.host);
return Err(a.into());
}
}
_ => {}
}
let v = HostInfoValue::new(info)?;
v.save(&data)?;
//初始化节点监控配置
let monitor_info = MonitorSetting::new(&info.host);
monitor_info.save(&data)?;
Ok(())
}
#[derive(Serialize, Deserialize, Debug)]
pub struct HaChangeLog {
pub key: String, //格式 host_timestamp host为宕机节点
pub cluster_name: String,
pub old_master_info: DownNodeCheck, //宕机节点信息
pub new_master_binlog_info: SlaveInfo, //如果宕机节点在切换之前未进行binlog追加将保存新master读取到的binlog信息,在宕机节点恢复时会进行判断回滚
pub recovery_info: RecoveryInfo, //宕机恢复同步所需的新master信息
pub recovery_status: bool, //是否已恢复
pub switch_status: bool, //切换状态
}
impl HaChangeLog {
pub fn new() -> HaChangeLog {
HaChangeLog{
key: "".to_string(),
cluster_name: "".to_string(),
old_master_info: DownNodeCheck { host: "".to_string(), dbport: 0 },
new_master_binlog_info: SlaveInfo {
host: "".to_string(),
dbport: 0,
slave_info: ReplicationState {
log_name: "".to_string(),
read_log_pos: 0,
exec_log_pos: 0
},
new_master: false
},
recovery_info: RecoveryInfo {
binlog: "".to_string(),
position: 0,
gtid: "".to_string(),
masterhost: "".to_string(),
masterport: 0,
read_binlog: "".to_string(),
read_position: 0
},
recovery_status: false,
switch_status: false
}
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let key = format!("{}_{}",self.key.clone(), crate::timestamp());
let value = serde_json::to_string(self)?;
let row = KeyValue{key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
pub fn update(&mut self, db: &web::Data<DbInfo>, row_key: String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(self)?;
let row = KeyValue{key: row_key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
}
///
///
///
/// 用户信息结构
#[derive(Serialize, Deserialize, Clone, Debug)]
pub struct UserInfo {
pub user_name: String,
pub password: String,
pub hook_id: String,
pub create_time: i64,
pub update_time: i64
}
impl UserInfo {
pub fn new(info: &PostUserInfo) -> UserInfo {
let create_time = crate::timestamp();
let update_time = crate::timestamp();
UserInfo{
user_name: info.user_name.clone(),
password: info.password.clone(),
hook_id: rand_string(),
create_time,
update_time
}
}
}
///
///
///
/// 节点基础信息, host做为key
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct HostInfoValue {
pub host: String, //127.0.0.1:3306
pub dbport: usize, //default 3306
pub rtype: String, //db、route
pub cluster_name: String, //集群名称,route类型默认default
pub online: bool, //db是否在线, true、false
pub insert_time: i64,
pub update_time: i64,
pub maintain: bool, //是否处于维护模式,true、false
}
impl HostInfoValue {
pub fn new(info: &HostInfo) -> Result<HostInfoValue, Box<dyn Error>> {
let h = HostInfoValue{
host: info.host.clone(),
rtype: info.rtype.clone(),
dbport: info.dbport.clone(),
cluster_name: info.cluster_name.clone(),
online: false,
insert_time: crate::timestamp(),
update_time: crate::timestamp(),
maintain: false
};
Ok(h)
}
///
/// 写入db
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let row = KeyValue{key: self.host.clone(), value};
db.put(&row, &CfNameTypeCode::HaNodesInfo.get())?;
Ok(())
}
///
/// 编辑节点信息
pub fn edit(&mut self, info: &web::Json<EditInfo>) {
self.host = info.host.clone();
self.dbport = info.dbport.clone();
self.cluster_name = info.cluster_name.clone();
self.update_time = crate::timestamp();
}
///
/// 设置节点维护模式状态
pub fn maintain(&mut self, info: &web::Json<EditMainTain>) { | }
self.update_time = crate::timestamp();
}
///
/// 获取当前节点在db中保存的状态信息
pub fn get_state(&self, db: &web::Data<DbInfo>) -> Result<MysqlState, Box<dyn Error>> {
let kv = db.get(&self.host, &CfNameTypeCode::NodesState.get())?;
if kv.value.len() > 0 {
let state: MysqlState = serde_json::from_str(&kv.value)?;
return Ok(state);
}else {
//let err = format!("this host: {} no state data", &self.host);
//return Err(err.into());
let state = MysqlState::new();
return Ok(state);
}
}
pub fn get_role(&self, db: &web::Data<DbInfo>) -> Result<String, Box<dyn Error>> {
let state = self.get_state(db)?;
Ok(state.role)
}
}
///
///
///
///
/// 获取db中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct NodeClusterList{
pub cluster_name_list: Vec<String>
}
impl NodeClusterList{
pub fn new() -> NodeClusterList{
NodeClusterList { cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let value: HostInfoValue = serde_json::from_str(&row.value)?;
if self.is_exists(&value.cluster_name){continue;}
self.cluster_name_list.push(value.cluster_name.clone());
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name{
return true;
}
}
return false;
}
}
/// 获取route信息中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct RouteClusterList{
pub cluster_name_list: Vec<String>
}
impl RouteClusterList{
pub fn new() -> RouteClusterList {
RouteClusterList{ cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let route_all = db.get_route_all()?;
for route in &route_all {
if !self.is_exists(&route.value.cluster_name){
self.cluster_name_list.push(route.value.cluster_name.clone());
}
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name {
return true;
}
}
return false;
}
}
///
///
///
///
/// node节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct NodeInfo{
pub cluster_name: String,
pub host: String,
pub dbport: usize,
pub online: bool, //是否在线, true、false
pub maintain: bool, //是否处于维护模式,true、false
pub role: String, //主从角色
pub master: String,
pub sql_thread: bool,
pub io_thread: bool,
pub seconds_behind: usize,
pub read_only: bool,
pub version: String,
pub executed_gtid_set: String,
pub innodb_flush_log_at_trx_commit: usize,
pub sync_binlog: usize,
pub server_id: usize,
pub event_scheduler: String,
pub sql_error: String
}
impl NodeInfo{
pub fn new(state: &MysqlState, node: &HostInfoValue) -> NodeInfo {
let mut ni = NodeInfo{
cluster_name: node.cluster_name.clone(),
host: node.host.clone(),
dbport: node.dbport.clone(),
online: node.online.clone(),
maintain: node.maintain.clone(),
role: state.role.clone(),
master: state.master.clone(),
sql_thread: state.sql_thread.clone(),
io_thread: state.io_thread.clone(),
seconds_behind: state.seconds_behind.clone(),
read_only: state.read_only.clone(),
version: state.version.clone(),
executed_gtid_set: state.executed_gtid_set.clone(),
innodb_flush_log_at_trx_commit: state.innodb_flush_log_at_trx_commit.clone(),
sync_binlog: state.sync_binlog.clone(),
server_id: state.server_id.clone(),
event_scheduler: state.event_scheduler.clone(),
sql_error: "".to_string()
};
if state.last_io_error.len() > 0{
ni.sql_error = state.last_io_error.clone();
}else if state.last_sql_error.len() > 0 {
ni.sql_error = state.last_sql_error.clone();
}
return ni;
}
}
///
///
///
///
/// 每个集群节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct ClusterNodeInfo{
pub cluster_name: String,
pub total: usize,
pub nodes_info: Vec<NodeInfo>
}
impl ClusterNodeInfo{
pub fn new(cluster_name: &String) -> ClusterNodeInfo{
ClusterNodeInfo{
cluster_name: cluster_name.clone(),
total: 0,
nodes_info: vec![]
}
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let node: HostInfoValue = serde_json::from_str(&row.value)?;
if &node.cluster_name == &self.cluster_name{
let state = node.get_state(db)?;
let node_info = NodeInfo::new(&state, &node);
self.total += 1;
self.nodes_info.push(node_info);
}
}
Ok(())
}
///
/// 统计所有节点监控信息, 用于首页展示
///
/// 倒叙迭代获取每个节点最后一条数据, 如果每个节点都已获取最后一条数据就退出迭代
pub fn static_monitor(&self, db: &web::Data<DbInfo>, rsm: &mut ResponseMonitorStatic) -> Result<(), Box<dyn Error>> {
let cf_name = CfNameTypeCode::SystemData.get();
let mut tmp: Vec<String> = vec![];
//首先检查是否开启监控
for node in &self.nodes_info{
if !self.check_monitor_setting(db, &node.host){
tmp.push(node.host.clone());
}
}
if let Some(cf) = db.db.cf_handle(&cf_name){
let mut iter = db.db.raw_iterator_cf(cf)?;
iter.seek_to_last();
iter.prev();
'all: while iter.valid() {
if tmp.len() == self.nodes_info.len(){
break 'all;
}
if let Some(s) = iter.key(){
let key: String = from_utf8(&s.to_vec())?.parse()?;
if key.starts_with(&PrefixTypeCode::NodeMonitorData.prefix()){
'b: for n in &self.nodes_info{
if key.contains(n.host.as_str()){
for t in &tmp{
if t == &n.host{
break 'b;
}
}
if let Some(v) = iter.value(){
let v: MysqlMonitorStatus = serde_json::from_slice(&v)?;
rsm.update(&v);
tmp.push(n.host.clone());
}
break 'b;
}
}
}
}
//
//
iter.prev();
}
}
Ok(())
}
fn check_monitor_setting(&self, db: &web::Data<DbInfo>, host: &String) -> bool{
let a = db.prefix_get(&PrefixTypeCode::NodeMonitorSeting, host);
match a {
Ok(v) => {
if v.value.len() > 0{
let value: MonitorSetting = serde_json::from_str(&v.value).unwrap();
return value.monitor.clone();
}
}
Err(e) => {
info!("{}", e.to_string());
}
}
return false;
}
}
///
///
impl MysqlState{
pub fn save(&self, db: &web::Data<DbInfo>, key: &String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let a = KeyValue{key: key.clone(), value };
db.put(&a, &CfNameTypeCode::NodesState.get())?;
Ok(())
}
}
///
///
/// slave behind 配置结构体
#[derive(Serialize, Deserialize, Debug)]
pub struct SlaveBehindSetting{
pub cluster_name: String,
pub delay: usize
}
impl SlaveBehindSetting{
pub fn new(cluster_name: &String) -> SlaveBehindSetting {
SlaveBehindSetting{ cluster_name: cluster_name.clone(), delay: 100 }
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
db.prefix_put(&PrefixTypeCode::SlaveDelaySeting, &self.cluster_name, &self)?;
Ok(())
}
} | if info.maintain == "true".to_string() {
self.maintain = false;
}else {
self.maintain = true; | random_line_split |
opdb.rs | /*
@author: xiao cai niao
@datetime: 2019/11/6
*/
use actix_web::{web};
use crate::webroute::route::{HostInfo, PostUserInfo, EditInfo, EditMainTain};
use crate::storage::rocks::{DbInfo, KeyValue, CfNameTypeCode, PrefixTypeCode};
use crate::ha::procotol::{DownNodeCheck, RecoveryInfo, ReplicationState, MysqlMonitorStatus};
use std::error::Error;
use crate::ha::nodes_manager::{SlaveInfo};
use serde::{Serialize, Deserialize};
use crate::rand_string;
use crate::ha::procotol::MysqlState;
use crate::ha::sys_manager::MonitorSetting;
use crate::webroute::new_route::ResponseMonitorStatic;
use std::str::from_utf8;
///
/// mysql node info, insert to rocksdb
///
///
pub fn insert_mysql_host_info(data: web::Data<DbInfo>, info: &web::Json<HostInfo>) -> Result<(), Box<dyn Error>> {
let check_unique = data.get(&info.host, &CfNameTypeCode::HaNodesInfo.get());
match check_unique {
Ok(v) => {
if v.value.len() > 0 {
let a = format!("this key: ({}) already exists in the database",&info.host);
return Err(a.into());
}
}
_ => {}
}
let v = HostInfoValue::new(info)?;
v.save(&data)?;
//初始化节点监控配置
let monitor_info = MonitorSetting::new(&info.host);
monitor_info.save(&data)?;
Ok(())
}
#[derive(Serialize, Deserialize, Debug)]
pub struct HaChangeLog {
pub key: String, //格式 host_timestamp host为宕机节点
pub cluster_name: String,
pub old_master_info: DownNodeCheck, //宕机节点信息
pub new_master_binlog_info: SlaveInfo, //如果宕机节点在切换之前未进行binlog追加将保存新master读取到的binlog信息,在宕机节点恢复时会进行判断回滚
pub recovery_info: RecoveryInfo, //宕机恢复同步所需的新master信息
pub recovery_status: bool, //是否已恢复
pub switch_status: bool, //切换状态
}
impl HaChangeLog {
pub fn new() -> HaChangeLog {
HaChangeLog{
key: "".to_string(),
cluster_name: "".to_string(),
old_master_info: DownNodeCheck { host: "".to_string(), dbport: 0 },
new_master_binlog_info: SlaveInfo {
host: "".to_string(),
dbport: 0,
slave_info: ReplicationState {
log_name: "".to_string(),
read_log_pos: 0,
exec_log_pos: 0
},
new_master: false
},
recovery_info: RecoveryInfo {
binlog: "".to_string(),
position: 0,
gtid: "".to_string(),
masterhost: "".to_string(),
masterport: 0,
read_binlog: "".to_string(),
read_position: 0
},
recovery_status: false,
switch_status: false
}
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let key = format!("{}_{}",self.key.clone(), crate::timestamp());
let value = serde_json::to_string(self)?;
let row = KeyValue{key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
pub fn update(&mut self, db: &web::Data<DbInfo>, row_key: String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(self)?;
let row = KeyValue{key: row_key, value};
db.put(&row, &CfNameTypeCode::HaChangeLog.get())?;
return Ok(());
}
}
///
///
///
/// 用户信息结构
#[derive(Serialize, Deserialize, Clone, Debug)]
pub struct UserInfo {
pub user_name: String,
pub password: String,
pub hook_id: String,
pub create_time: i64,
pub update_time: i64
}
impl UserInfo {
pub fn new(info: &PostUserInfo) -> UserInfo {
let create_time = crate::timestamp();
let update_time = crate::timestamp();
UserInfo{
user_name: info.user_name.clone(),
password: inf | ult 3306
pub rtype: String, //db、route
pub cluster_name: String, //集群名称,route类型默认default
pub online: bool, //db是否在线, true、false
pub insert_time: i64,
pub update_time: i64,
pub maintain: bool, //是否处于维护模式,true、false
}
impl HostInfoValue {
pub fn new(info: &HostInfo) -> Result<HostInfoValue, Box<dyn Error>> {
let h = HostInfoValue{
host: info.host.clone(),
rtype: info.rtype.clone(),
dbport: info.dbport.clone(),
cluster_name: info.cluster_name.clone(),
online: false,
insert_time: crate::timestamp(),
update_time: crate::timestamp(),
maintain: false
};
Ok(h)
}
///
/// 写入db
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let row = KeyValue{key: self.host.clone(), value};
db.put(&row, &CfNameTypeCode::HaNodesInfo.get())?;
Ok(())
}
///
/// 编辑节点信息
pub fn edit(&mut self, info: &web::Json<EditInfo>) {
self.host = info.host.clone();
self.dbport = info.dbport.clone();
self.cluster_name = info.cluster_name.clone();
self.update_time = crate::timestamp();
}
///
/// 设置节点维护模式状态
pub fn maintain(&mut self, info: &web::Json<EditMainTain>) {
if info.maintain == "true".to_string() {
self.maintain = false;
}else {
self.maintain = true;
}
self.update_time = crate::timestamp();
}
///
/// 获取当前节点在db中保存的状态信息
pub fn get_state(&self, db: &web::Data<DbInfo>) -> Result<MysqlState, Box<dyn Error>> {
let kv = db.get(&self.host, &CfNameTypeCode::NodesState.get())?;
if kv.value.len() > 0 {
let state: MysqlState = serde_json::from_str(&kv.value)?;
return Ok(state);
}else {
//let err = format!("this host: {} no state data", &self.host);
//return Err(err.into());
let state = MysqlState::new();
return Ok(state);
}
}
pub fn get_role(&self, db: &web::Data<DbInfo>) -> Result<String, Box<dyn Error>> {
let state = self.get_state(db)?;
Ok(state.role)
}
}
///
///
///
///
/// 获取db中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct NodeClusterList{
pub cluster_name_list: Vec<String>
}
impl NodeClusterList{
pub fn new() -> NodeClusterList{
NodeClusterList { cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>> {
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let value: HostInfoValue = serde_json::from_str(&row.value)?;
if self.is_exists(&value.cluster_name){continue;}
self.cluster_name_list.push(value.cluster_name.clone());
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name{
return true;
}
}
return false;
}
}
/// 获取route信息中现有的cluster列表
#[derive(Serialize, Deserialize, Debug)]
pub struct RouteClusterList{
pub cluster_name_list: Vec<String>
}
impl RouteClusterList{
pub fn new() -> RouteClusterList {
RouteClusterList{ cluster_name_list: vec![] }
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let route_all = db.get_route_all()?;
for route in &route_all {
if !self.is_exists(&route.value.cluster_name){
self.cluster_name_list.push(route.value.cluster_name.clone());
}
}
Ok(())
}
fn is_exists(&self, cluster_name: &String) -> bool {
for cl in &self.cluster_name_list {
if cl == cluster_name {
return true;
}
}
return false;
}
}
///
///
///
///
/// node节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct NodeInfo{
pub cluster_name: String,
pub host: String,
pub dbport: usize,
pub online: bool, //是否在线, true、false
pub maintain: bool, //是否处于维护模式,true、false
pub role: String, //主从角色
pub master: String,
pub sql_thread: bool,
pub io_thread: bool,
pub seconds_behind: usize,
pub read_only: bool,
pub version: String,
pub executed_gtid_set: String,
pub innodb_flush_log_at_trx_commit: usize,
pub sync_binlog: usize,
pub server_id: usize,
pub event_scheduler: String,
pub sql_error: String
}
impl NodeInfo{
pub fn new(state: &MysqlState, node: &HostInfoValue) -> NodeInfo {
let mut ni = NodeInfo{
cluster_name: node.cluster_name.clone(),
host: node.host.clone(),
dbport: node.dbport.clone(),
online: node.online.clone(),
maintain: node.maintain.clone(),
role: state.role.clone(),
master: state.master.clone(),
sql_thread: state.sql_thread.clone(),
io_thread: state.io_thread.clone(),
seconds_behind: state.seconds_behind.clone(),
read_only: state.read_only.clone(),
version: state.version.clone(),
executed_gtid_set: state.executed_gtid_set.clone(),
innodb_flush_log_at_trx_commit: state.innodb_flush_log_at_trx_commit.clone(),
sync_binlog: state.sync_binlog.clone(),
server_id: state.server_id.clone(),
event_scheduler: state.event_scheduler.clone(),
sql_error: "".to_string()
};
if state.last_io_error.len() > 0{
ni.sql_error = state.last_io_error.clone();
}else if state.last_sql_error.len() > 0 {
ni.sql_error = state.last_sql_error.clone();
}
return ni;
}
}
///
///
///
///
/// 每个集群节点信息
#[derive(Deserialize, Serialize, Debug)]
pub struct ClusterNodeInfo{
pub cluster_name: String,
pub total: usize,
pub nodes_info: Vec<NodeInfo>
}
impl ClusterNodeInfo{
pub fn new(cluster_name: &String) -> ClusterNodeInfo{
ClusterNodeInfo{
cluster_name: cluster_name.clone(),
total: 0,
nodes_info: vec![]
}
}
pub fn init(&mut self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
let result = db.iterator(&CfNameTypeCode::HaNodesInfo.get(), &String::from(""))?;
for row in &result{
let node: HostInfoValue = serde_json::from_str(&row.value)?;
if &node.cluster_name == &self.cluster_name{
let state = node.get_state(db)?;
let node_info = NodeInfo::new(&state, &node);
self.total += 1;
self.nodes_info.push(node_info);
}
}
Ok(())
}
///
/// 统计所有节点监控信息, 用于首页展示
///
/// 倒叙迭代获取每个节点最后一条数据, 如果每个节点都已获取最后一条数据就退出迭代
pub fn static_monitor(&self, db: &web::Data<DbInfo>, rsm: &mut ResponseMonitorStatic) -> Result<(), Box<dyn Error>> {
let cf_name = CfNameTypeCode::SystemData.get();
let mut tmp: Vec<String> = vec![];
//首先检查是否开启监控
for node in &self.nodes_info{
if !self.check_monitor_setting(db, &node.host){
tmp.push(node.host.clone());
}
}
if let Some(cf) = db.db.cf_handle(&cf_name){
let mut iter = db.db.raw_iterator_cf(cf)?;
iter.seek_to_last();
iter.prev();
'all: while iter.valid() {
if tmp.len() == self.nodes_info.len(){
break 'all;
}
if let Some(s) = iter.key(){
let key: String = from_utf8(&s.to_vec())?.parse()?;
if key.starts_with(&PrefixTypeCode::NodeMonitorData.prefix()){
'b: for n in &self.nodes_info{
if key.contains(n.host.as_str()){
for t in &tmp{
if t == &n.host{
break 'b;
}
}
if let Some(v) = iter.value(){
let v: MysqlMonitorStatus = serde_json::from_slice(&v)?;
rsm.update(&v);
tmp.push(n.host.clone());
}
break 'b;
}
}
}
}
//
//
iter.prev();
}
}
Ok(())
}
fn check_monitor_setting(&self, db: &web::Data<DbInfo>, host: &String) -> bool{
let a = db.prefix_get(&PrefixTypeCode::NodeMonitorSeting, host);
match a {
Ok(v) => {
if v.value.len() > 0{
let value: MonitorSetting = serde_json::from_str(&v.value).unwrap();
return value.monitor.clone();
}
}
Err(e) => {
info!("{}", e.to_string());
}
}
return false;
}
}
///
///
impl MysqlState{
pub fn save(&self, db: &web::Data<DbInfo>, key: &String) -> Result<(), Box<dyn Error>> {
let value = serde_json::to_string(&self)?;
let a = KeyValue{key: key.clone(), value };
db.put(&a, &CfNameTypeCode::NodesState.get())?;
Ok(())
}
}
///
///
/// slave behind 配置结构体
#[derive(Serialize, Deserialize, Debug)]
pub struct SlaveBehindSetting{
pub cluster_name: String,
pub delay: usize
}
impl SlaveBehindSetting{
pub fn new(cluster_name: &String) -> SlaveBehindSetting {
SlaveBehindSetting{ cluster_name: cluster_name.clone(), delay: 100 }
}
pub fn save(&self, db: &web::Data<DbInfo>) -> Result<(), Box<dyn Error>>{
db.prefix_put(&PrefixTypeCode::SlaveDelaySeting, &self.cluster_name, &self)?;
Ok(())
}
}
| o.password.clone(),
hook_id: rand_string(),
create_time,
update_time
}
}
}
///
///
///
/// 节点基础信息, host做为key
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct HostInfoValue {
pub host: String, //127.0.0.1:3306
pub dbport: usize, //defa | identifier_body |
server_impl.go | // Contains the implementation of a LSP server.
package lsp
import "errors"
import "strconv"
import "github.com/cmu440/lspnet"
import "container/list"
import "encoding/json"
import "fmt"
import "time"
type msgPackage struct {
msg *Message
addr *lspnet.UDPAddr
}
type abstractClient struct {
addr *lspnet.UDPAddr
nextSN int
nextDataRead int
// window size variable
// server send msg to client
nextSmallestAck int
mapAck map[int]bool
mapNeedAck map[int]*Message
// server receive msg from client
nextSmallestData int
mapReceived map[int]*Message
// variables of epoch
epochNum int
lostConn bool
// close
closeConn bool
}
type requestRead struct {
ask chan int
response chan *Message
}
type requestWrite struct {
ask chan []byte
connId chan int
response chan error
}
type requestClose struct {
ask chan int
//response chan int
getError chan error
}
type requestCloseConn struct {
ask chan int
getError chan error
}
type server struct {
nextConnectId int
clients map[int]*abstractClient
conn *lspnet.UDPConn
readFromClientChan chan *msgPackage
writeToClientChan chan *Message
readRequest *requestRead
writeRequest *requestWrite
readList *list.List
writeList *list.List
flag bool
// variables for window size
windowSize int
mapNeedSend *list.List
// variables for epoch
epochChan chan int
epochMillis int
epochLimit int
// close
deleteClient chan int
closeConnRequest *requestCloseConn
// close every thing
waitToWriteFinish bool
writeFinished chan int
waitToAckFinish bool
ackFinished chan int
closeRead chan int
closeEpoch chan int
closeEvent chan int
// close connection
closeConn chan int
}
// NewServer creates, initiates, and returns a new server. This function should
// NOT block. Instead, it should spawn one or more goroutines (to handle things
// like accepting incoming client connections, triggering epoch events at
// fixed intervals, synchronizing events using a for-select loop like you saw in
// project 0, etc.) and immediately return. It should return a non-nil error if
// there was an error resolving or listening on the specified port number.
func NewServer(port int, params *Params) (Server, error) {
s := &server{
nextConnectId: 1,
clients: make(map[int]*abstractClient),
readFromClientChan: make(chan *msgPackage),
writeToClientChan: make(chan *Message),
readRequest: &requestRead{
ask: make(chan int),
response: make(chan *Message),
},
writeRequest: &requestWrite{
ask: make(chan []byte),
connId: make(chan int),
response: make(chan error),
},
readList: list.New(),
writeList: list.New(),
flag: false,
// variables for window size
windowSize: params.WindowSize,
mapNeedSend: list.New(),
// variables for epoch
epochChan: make(chan int),
epochMillis: params.EpochMillis,
epochLimit: params.EpochLimit,
// close
deleteClient: make(chan int),
closeConnRequest: &requestCloseConn{
ask: make(chan int),
getError: make(chan error),
},
waitToWriteFinish: false,
writeFinished: make(chan int),
waitToAckFinish: false,
ackFinished: make(chan int),
closeRead: make(chan int, 1),
closeEpoch: make(chan int, 1),
closeEvent: make(chan int, 1),
// close conn
closeConn: make(chan int, 1),
}
// start server
addr, err := lspnet.ResolveUDPAddr("udp", "localhost:"+strconv.Itoa(port))
if err != nil {
return nil, err
}
conn, err := lspnet.ListenUDP("udp", addr)
if err != nil {
return nil, err
}
s.conn = conn
go s.readMessage()
go s.handleMessage()
go s.epochFire()
fmt.Println("new server")
return s, nil
}
func (s *server) epochFire() {
for {
select {
case <-s.closeEpoch:
return
default:
////fmt.Println("Server EpochFire!!!!!!!!!!!!!!!!!!!!!!!!!!")
<-time.After(time.Duration(s.epochMillis) * time.Millisecond)
s.epochChan <- 1
}
}
}
func (s *server) handleMessage() {
for {
select {
case <-s.writeFinished:
if s.writeList.Len() == 0 {
s.writeFinished <- 1
} else {
s.waitToWriteFinish = true
}
case <-s.ackFinished:
//fmt.Println("Server 192: receive ack finished")
s.checkAckFinished()
case msg := <-s.readFromClientChan:
////fmt.Println("Server 162: read from client: ", msg)
s.readFromClient(msg)
case msg := <-s.writeToClientChan:
////fmt.Println("Server 165: write to client: ", msg)
if s.clients[msg.ConnID] == nil {
continue
}
if msg.Type == MsgData {
s.clients[msg.ConnID].mapAck[msg.SeqNum] = false
s.clients[msg.ConnID].mapNeedAck[msg.SeqNum] = msg
// data type: need to consider order
if (msg.SeqNum - s.clients[msg.ConnID].nextSmallestAck) < s.windowSize {
s.writeToClient(msg)
} else {
s.writeList.PushBack(msg)
}
} else {
// ack/conn type: don't need to consider order
s.writeToClient(msg)
}
if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
case <-s.readRequest.ask:
s.handleReadRequest()
case payload := <-s.writeRequest.ask:
connId := <-s.writeRequest.connId
var response = s.handleWriteRequest(connId, payload)
s.writeRequest.response <- response
case <-s.epochChan:
s.addEpochNum()
s.handleEpochEvent()
case id := <-s.deleteClient:
delete(s.clients, id)
case id := <-s.closeConnRequest.ask:
if s.clients[id] == nil {
s.closeConnRequest.getError <- errors.New("No client")
}
s.sendDeadMsg(id)
s.closeConnRequest.getError <- nil
}
}
}
func (s *server) checkAckFinished() {
for _, c := range s.clients {
if len(c.mapAck) != 0 {
s.waitToAckFinish = true
//fmt.Println("Server: need to wait ack finish")
return
}
}
s.ackFinished <- 1
}
func (s *server) handleWriteRequest(connID int, payload []byte) error |
func (s *server) sendData(dataMsg *Message) {
s.writeToClientChan <- dataMsg
}
func (s *server) addEpochNum() {
for connId, c := range s.clients {
c.epochNum += 1
if c.epochNum >= s.epochLimit {
c.lostConn = true
s.sendDeadMsg(connId)
}
}
}
func (s *server) sendDeadMsg(connId int) {
dataMsg := NewData(connId, s.clients[connId].nextSN, nil, nil)
if s.flag {
s.readRequest.response <- dataMsg
s.flag = false
} else {
s.readList.PushBack(dataMsg)
}
}
func (s *server) readData(dataMsg *msgPackage) {
s.readFromClient(dataMsg)
}
func (s *server) handleEpochEvent() {
for connID, c := range s.clients {
/*
* if no data received from the client, then resend ack for connection requestRead
*/
if c.nextSmallestAck == 1 {
ack := NewAck(connID, 0)
go s.sendAck(ack)
}
/*
* sent but not acked, resend the data
*/
for key, received := range c.mapAck {
if !received {
msg := c.mapNeedAck[key]
fmt.Println("Server epoch: ", msg)
s.writeToClient(msg)
}
}
}
}
func (s *server) handleReadRequest() {
msg := s.getMessageFromReadList()
if msg != nil {
//////fmt.Println("101 no message")
/*elm := s.readList.Front()
msg := elm.Value.(*Message)*/
s.readRequest.response <- msg
} else {
s.flag = true
}
return
}
// find next msg that can be send
// return nil if no msg can be send
func (s *server) getMessageFromReadList() *Message {
for e := s.readList.Front(); e != nil; e = e.Next() {
msg := e.Value.(*Message)
if msg.SeqNum == s.clients[msg.ConnID].nextDataRead {
s.clients[msg.ConnID].nextDataRead += 1
s.readList.Remove(e)
return msg
}
}
return nil
}
func (s *server) writeToClient(msg *Message) error {
connId := msg.ConnID
addr := s.clients[connId].addr
mmsg, _ := json.Marshal(msg)
fmt.Println("364 Server write to client: ", msg)
_, err := s.conn.WriteToUDP(mmsg, addr)
if err != nil {
fmt.Println("364: ", err)
}
return err
}
func (s *server) readFromClient(msgPack *msgPackage) {
//////fmt.Println("129 read")
msg := msgPack.msg
addr := msgPack.addr
////fmt.Println("server read: ", msg)
if msg.Type == MsgConnect {
s.connectClient(msg, addr)
} else {
// set epoch number to 0
if s.clients[msg.ConnID] != nil {
s.clients[msg.ConnID].epochNum = 0
}
if msg.Type == MsgData {
if !s.waitToAckFinish {
//fmt.Println("Server read from client: ", msg)
s.receiveData(msg)
}
} else if msg.Type == MsgAck {
fmt.Println("389 Server: receive ack: ", msg)
s.receiveAck(msg)
}
if s.waitToAckFinish {
s.checkAckFinished()
}
}
return
}
func (s *server) alreadyHasClient(addr *lspnet.UDPAddr) bool {
////fmt.Println("server: in already has client")
for _, client := range s.clients {
if client.addr.String() == addr.String() {
return true
}
}
return false
}
func (s *server) receiveData(msg *Message) {
//fmt.Println("Server read: ", msg)
c := s.clients[msg.ConnID]
// check duplicated forst and within window size
/*if c.nextSmallestData+s.windowSize <= msg.SeqNum {
// check within window size, drop it if not
return
} else if c.mapReceived[msg.SeqNum] != nil {
// check duplicated outside of window size
return
}*/
//fmt.Println("pass window size")
// here is no duplicated
// decide should we push it to the readlist or chan
if s.flag && (s.clients[msg.ConnID].nextDataRead == msg.SeqNum) {
s.readRequest.response <- msg
c.nextDataRead += 1
s.flag = false
} else if msg.SeqNum >= c.nextDataRead {
s.readList.PushBack(msg)
}
// create ack message
connID := msg.ConnID
seqNum := msg.SeqNum
ack := NewAck(connID, seqNum)
// check the receive variables in client
// delete the message in mapreceived
if msg.SeqNum == c.nextSmallestData {
c.nextSmallestData += 1
for c.mapReceived[c.nextSmallestData] != nil {
delete(c.mapReceived, c.nextSmallestData)
c.nextSmallestData += 1
}
} else {
c.mapReceived[msg.SeqNum] = msg
}
go s.sendAck(ack)
return
}
func (s *server) sendAck(ack *Message) {
s.writeToClientChan <- ack
}
func (s *server) receiveAck(msg *Message) {
//fmt.Println("server: receiveAck")
c := s.clients[msg.ConnID]
if msg.SeqNum >= c.nextSmallestAck {
c.mapAck[msg.SeqNum] = true
}
if msg.SeqNum == c.nextSmallestAck {
// reset the next smallest ack numebr
value, exist := c.mapAck[c.nextSmallestAck]
for exist && value {
delete(c.mapAck, c.nextSmallestAck)
c.nextSmallestAck += 1
value, exist = c.mapAck[c.nextSmallestAck]
}
var l = *list.New()
for element := s.writeList.Front(); element != nil; element = element.Next() {
message := element.Value.(*Message)
if message.ConnID == msg.ConnID {
if (message.SeqNum - c.nextSmallestAck) < s.windowSize {
s.writeToClient(message)
l.PushBack(element)
}
}
}
////fmt.Println("loop finish go out of write")
for element := l.Front(); element != nil; element = element.Next() {
s.writeList.Remove(element)
}
////fmt.Println("go out of delte dfadfasdf")
if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
////fmt.Println("go asdfasdfe dfadfasdf")
}
return
}
func (s *server) connectClient(msg *Message, addr *lspnet.UDPAddr) {
////fmt.Println("Server: connect client ", msg)
// check duplication of connection
if s.alreadyHasClient(addr) {
////fmt.Println("already has client")
return
}
connID := s.nextConnectId
newClient := abstractClient{
addr: addr,
nextSN: 1,
nextDataRead: 1,
// server send msg to client
nextSmallestAck: 1,
mapAck: make(map[int]bool),
mapNeedAck: make(map[int]*Message),
// server receive msg from client
nextSmallestData: 1,
mapReceived: make(map[int]*Message),
// epoch
epochNum: 0,
lostConn: false,
// close
closeConn: false,
}
s.clients[connID] = &newClient
s.nextConnectId += 1
ack := NewAck(connID, 0)
////fmt.Println("Server: send ack to client")
go s.sendAck(ack)
return
}
func (s *server) readMessage() {
var b [1500]byte
//////fmt.Println("Read: read message")
for {
select {
case <-s.closeRead:
return
default:
var msg Message
// unmarshal
n, addr, err := s.conn.ReadFromUDP(b[0:])
err = json.Unmarshal(b[0:n], &msg)
if err == nil {
msgPck := &msgPackage{
msg: &msg,
addr: addr,
}
//fmt.Println("server read: ", &msg)
s.readFromClientChan <- msgPck
} else {
//fmt.Println(err)
return
}
}
}
}
func (s *server) Read() (int, []byte, error) {
s.readRequest.ask <- 1
msg := <-s.readRequest.response
////fmt.Println("Server called read: ", msg)
// check if the client still alive or not
if msg.Payload == nil {
////fmt.Println("client closed when read")
s.deleteClient <- msg.ConnID
//fmt.Println("read ")
return msg.ConnID, nil, errors.New("Client connection lost")
}
// TODO: client lost return connection id
return msg.ConnID, msg.Payload, nil
}
func (s *server) Write(connID int, payload []byte) error {
//fmt.Println("server called write")
/*client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
s.writeToClientChan <- dataMsg
client.nextSN += 1*/
//////fmt.Println(s.clients[connID])
if s.clients[connID] == nil {
return errors.New("Client connection lost")
}
s.writeRequest.ask <- payload
s.writeRequest.connId <- connID
response := <-s.writeRequest.response
return response
}
func (s *server) CloseConn(connID int) error {
//fmt.Println("server closeconn!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.closeConnRequest.ask <- connID
err := <-s.closeConnRequest.getError
//fmt.Println("server closed finish!!!!!!!!!!!!!!!!!!!!!!")
return err
}
func (s *server) Close() error {
return errors.New("text")
//fmt.Println("server: server close !!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.writeFinished <- 0
//fmt.Println("Server: finish wirte1")
<-s.writeFinished
//fmt.Println("Server: finish wirte")
/*s.ackFinished <- 0
<-s.ackFinished*/
//fmt.Println("Server: finish ack")
s.closeEpoch <- 1
s.closeRead <- 1
s.closeEvent <- 1
//fmt.Println("Server: close every thing")
err := s.conn.Close()
if err == nil {
return nil
} else {
return err
}
return errors.New("not yet implemented")
}
| {
//////fmt.Println(connID)
client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
go s.sendData(dataMsg)
client.nextSN += 1
return nil
} | identifier_body |
server_impl.go | // Contains the implementation of a LSP server.
package lsp
import "errors"
import "strconv"
import "github.com/cmu440/lspnet"
import "container/list"
import "encoding/json"
import "fmt"
import "time"
type msgPackage struct {
msg *Message
addr *lspnet.UDPAddr
}
type abstractClient struct {
addr *lspnet.UDPAddr
nextSN int
nextDataRead int
// window size variable
// server send msg to client
nextSmallestAck int
mapAck map[int]bool
mapNeedAck map[int]*Message
// server receive msg from client
nextSmallestData int
mapReceived map[int]*Message
// variables of epoch
epochNum int
lostConn bool
// close
closeConn bool
}
type requestRead struct {
ask chan int
response chan *Message
}
type requestWrite struct {
ask chan []byte
connId chan int
response chan error
}
type requestClose struct {
ask chan int
//response chan int
getError chan error
}
type requestCloseConn struct {
ask chan int
getError chan error
}
type server struct {
nextConnectId int
clients map[int]*abstractClient
conn *lspnet.UDPConn
readFromClientChan chan *msgPackage
writeToClientChan chan *Message
readRequest *requestRead
writeRequest *requestWrite
readList *list.List
writeList *list.List
flag bool
// variables for window size
windowSize int
mapNeedSend *list.List
// variables for epoch
epochChan chan int
epochMillis int
epochLimit int
// close
deleteClient chan int
closeConnRequest *requestCloseConn
// close every thing
waitToWriteFinish bool
writeFinished chan int
waitToAckFinish bool
ackFinished chan int
closeRead chan int
closeEpoch chan int
closeEvent chan int
// close connection
closeConn chan int
}
// NewServer creates, initiates, and returns a new server. This function should
// NOT block. Instead, it should spawn one or more goroutines (to handle things
// like accepting incoming client connections, triggering epoch events at
// fixed intervals, synchronizing events using a for-select loop like you saw in
// project 0, etc.) and immediately return. It should return a non-nil error if
// there was an error resolving or listening on the specified port number.
func NewServer(port int, params *Params) (Server, error) {
s := &server{
nextConnectId: 1,
clients: make(map[int]*abstractClient),
readFromClientChan: make(chan *msgPackage),
writeToClientChan: make(chan *Message),
readRequest: &requestRead{
ask: make(chan int),
response: make(chan *Message),
},
writeRequest: &requestWrite{
ask: make(chan []byte),
connId: make(chan int),
response: make(chan error),
},
readList: list.New(),
writeList: list.New(),
flag: false,
// variables for window size
windowSize: params.WindowSize,
mapNeedSend: list.New(),
// variables for epoch
epochChan: make(chan int),
epochMillis: params.EpochMillis,
epochLimit: params.EpochLimit,
// close
deleteClient: make(chan int),
closeConnRequest: &requestCloseConn{
ask: make(chan int),
getError: make(chan error),
},
waitToWriteFinish: false,
writeFinished: make(chan int),
waitToAckFinish: false,
ackFinished: make(chan int),
closeRead: make(chan int, 1),
closeEpoch: make(chan int, 1),
closeEvent: make(chan int, 1),
// close conn
closeConn: make(chan int, 1),
}
// start server
addr, err := lspnet.ResolveUDPAddr("udp", "localhost:"+strconv.Itoa(port))
if err != nil {
return nil, err
}
conn, err := lspnet.ListenUDP("udp", addr)
if err != nil {
return nil, err
}
s.conn = conn
go s.readMessage()
go s.handleMessage()
go s.epochFire()
fmt.Println("new server")
return s, nil
}
func (s *server) epochFire() {
for {
select {
case <-s.closeEpoch:
return
default:
////fmt.Println("Server EpochFire!!!!!!!!!!!!!!!!!!!!!!!!!!")
<-time.After(time.Duration(s.epochMillis) * time.Millisecond)
s.epochChan <- 1
}
}
}
func (s *server) handleMessage() {
for {
select {
case <-s.writeFinished:
if s.writeList.Len() == 0 {
s.writeFinished <- 1
} else {
s.waitToWriteFinish = true
}
case <-s.ackFinished:
//fmt.Println("Server 192: receive ack finished")
s.checkAckFinished()
case msg := <-s.readFromClientChan:
////fmt.Println("Server 162: read from client: ", msg)
s.readFromClient(msg)
case msg := <-s.writeToClientChan:
////fmt.Println("Server 165: write to client: ", msg)
if s.clients[msg.ConnID] == nil {
continue
}
if msg.Type == MsgData {
s.clients[msg.ConnID].mapAck[msg.SeqNum] = false
s.clients[msg.ConnID].mapNeedAck[msg.SeqNum] = msg
// data type: need to consider order
if (msg.SeqNum - s.clients[msg.ConnID].nextSmallestAck) < s.windowSize {
s.writeToClient(msg)
} else {
s.writeList.PushBack(msg)
}
} else {
// ack/conn type: don't need to consider order
s.writeToClient(msg)
}
if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
case <-s.readRequest.ask:
s.handleReadRequest()
case payload := <-s.writeRequest.ask:
connId := <-s.writeRequest.connId
var response = s.handleWriteRequest(connId, payload)
s.writeRequest.response <- response
case <-s.epochChan:
s.addEpochNum()
s.handleEpochEvent()
case id := <-s.deleteClient:
delete(s.clients, id)
case id := <-s.closeConnRequest.ask:
if s.clients[id] == nil {
s.closeConnRequest.getError <- errors.New("No client")
}
s.sendDeadMsg(id)
s.closeConnRequest.getError <- nil
}
}
}
func (s *server) checkAckFinished() {
for _, c := range s.clients {
if len(c.mapAck) != 0 {
s.waitToAckFinish = true
//fmt.Println("Server: need to wait ack finish")
return
}
}
s.ackFinished <- 1
}
func (s *server) handleWriteRequest(connID int, payload []byte) error {
//////fmt.Println(connID)
client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
go s.sendData(dataMsg)
client.nextSN += 1
return nil
}
func (s *server) sendData(dataMsg *Message) {
s.writeToClientChan <- dataMsg
}
func (s *server) addEpochNum() {
for connId, c := range s.clients {
c.epochNum += 1
if c.epochNum >= s.epochLimit {
c.lostConn = true
s.sendDeadMsg(connId)
}
}
}
func (s *server) sendDeadMsg(connId int) {
dataMsg := NewData(connId, s.clients[connId].nextSN, nil, nil)
if s.flag {
s.readRequest.response <- dataMsg
s.flag = false
} else {
s.readList.PushBack(dataMsg)
}
}
func (s *server) readData(dataMsg *msgPackage) {
s.readFromClient(dataMsg)
}
func (s *server) handleEpochEvent() {
for connID, c := range s.clients {
/*
* if no data received from the client, then resend ack for connection requestRead
*/
if c.nextSmallestAck == 1 {
ack := NewAck(connID, 0)
go s.sendAck(ack)
}
/*
* sent but not acked, resend the data
*/
for key, received := range c.mapAck {
if !received {
msg := c.mapNeedAck[key]
fmt.Println("Server epoch: ", msg)
s.writeToClient(msg)
}
}
}
}
func (s *server) handleReadRequest() {
msg := s.getMessageFromReadList()
if msg != nil {
//////fmt.Println("101 no message")
/*elm := s.readList.Front()
msg := elm.Value.(*Message)*/
s.readRequest.response <- msg
} else {
s.flag = true
}
return
}
// find next msg that can be send
// return nil if no msg can be send
func (s *server) getMessageFromReadList() *Message {
for e := s.readList.Front(); e != nil; e = e.Next() {
msg := e.Value.(*Message)
if msg.SeqNum == s.clients[msg.ConnID].nextDataRead {
s.clients[msg.ConnID].nextDataRead += 1
s.readList.Remove(e)
return msg
}
}
return nil
}
func (s *server) writeToClient(msg *Message) error {
connId := msg.ConnID
addr := s.clients[connId].addr
mmsg, _ := json.Marshal(msg)
fmt.Println("364 Server write to client: ", msg)
_, err := s.conn.WriteToUDP(mmsg, addr)
if err != nil {
fmt.Println("364: ", err)
}
return err
}
func (s *server) readFromClient(msgPack *msgPackage) {
//////fmt.Println("129 read")
msg := msgPack.msg
addr := msgPack.addr
////fmt.Println("server read: ", msg)
if msg.Type == MsgConnect {
s.connectClient(msg, addr)
} else {
// set epoch number to 0
if s.clients[msg.ConnID] != nil {
s.clients[msg.ConnID].epochNum = 0
}
if msg.Type == MsgData {
if !s.waitToAckFinish {
//fmt.Println("Server read from client: ", msg)
s.receiveData(msg)
}
} else if msg.Type == MsgAck {
fmt.Println("389 Server: receive ack: ", msg)
s.receiveAck(msg)
}
if s.waitToAckFinish {
s.checkAckFinished()
}
}
return
}
func (s *server) alreadyHasClient(addr *lspnet.UDPAddr) bool {
////fmt.Println("server: in already has client")
for _, client := range s.clients {
if client.addr.String() == addr.String() {
return true
}
}
return false
}
func (s *server) receiveData(msg *Message) {
//fmt.Println("Server read: ", msg)
c := s.clients[msg.ConnID]
// check duplicated forst and within window size
/*if c.nextSmallestData+s.windowSize <= msg.SeqNum {
// check within window size, drop it if not
return
} else if c.mapReceived[msg.SeqNum] != nil {
// check duplicated outside of window size
return
}*/
//fmt.Println("pass window size")
// here is no duplicated
// decide should we push it to the readlist or chan
if s.flag && (s.clients[msg.ConnID].nextDataRead == msg.SeqNum) {
s.readRequest.response <- msg
c.nextDataRead += 1
s.flag = false
} else if msg.SeqNum >= c.nextDataRead |
// create ack message
connID := msg.ConnID
seqNum := msg.SeqNum
ack := NewAck(connID, seqNum)
// check the receive variables in client
// delete the message in mapreceived
if msg.SeqNum == c.nextSmallestData {
c.nextSmallestData += 1
for c.mapReceived[c.nextSmallestData] != nil {
delete(c.mapReceived, c.nextSmallestData)
c.nextSmallestData += 1
}
} else {
c.mapReceived[msg.SeqNum] = msg
}
go s.sendAck(ack)
return
}
func (s *server) sendAck(ack *Message) {
s.writeToClientChan <- ack
}
func (s *server) receiveAck(msg *Message) {
//fmt.Println("server: receiveAck")
c := s.clients[msg.ConnID]
if msg.SeqNum >= c.nextSmallestAck {
c.mapAck[msg.SeqNum] = true
}
if msg.SeqNum == c.nextSmallestAck {
// reset the next smallest ack numebr
value, exist := c.mapAck[c.nextSmallestAck]
for exist && value {
delete(c.mapAck, c.nextSmallestAck)
c.nextSmallestAck += 1
value, exist = c.mapAck[c.nextSmallestAck]
}
var l = *list.New()
for element := s.writeList.Front(); element != nil; element = element.Next() {
message := element.Value.(*Message)
if message.ConnID == msg.ConnID {
if (message.SeqNum - c.nextSmallestAck) < s.windowSize {
s.writeToClient(message)
l.PushBack(element)
}
}
}
////fmt.Println("loop finish go out of write")
for element := l.Front(); element != nil; element = element.Next() {
s.writeList.Remove(element)
}
////fmt.Println("go out of delte dfadfasdf")
if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
////fmt.Println("go asdfasdfe dfadfasdf")
}
return
}
func (s *server) connectClient(msg *Message, addr *lspnet.UDPAddr) {
////fmt.Println("Server: connect client ", msg)
// check duplication of connection
if s.alreadyHasClient(addr) {
////fmt.Println("already has client")
return
}
connID := s.nextConnectId
newClient := abstractClient{
addr: addr,
nextSN: 1,
nextDataRead: 1,
// server send msg to client
nextSmallestAck: 1,
mapAck: make(map[int]bool),
mapNeedAck: make(map[int]*Message),
// server receive msg from client
nextSmallestData: 1,
mapReceived: make(map[int]*Message),
// epoch
epochNum: 0,
lostConn: false,
// close
closeConn: false,
}
s.clients[connID] = &newClient
s.nextConnectId += 1
ack := NewAck(connID, 0)
////fmt.Println("Server: send ack to client")
go s.sendAck(ack)
return
}
func (s *server) readMessage() {
var b [1500]byte
//////fmt.Println("Read: read message")
for {
select {
case <-s.closeRead:
return
default:
var msg Message
// unmarshal
n, addr, err := s.conn.ReadFromUDP(b[0:])
err = json.Unmarshal(b[0:n], &msg)
if err == nil {
msgPck := &msgPackage{
msg: &msg,
addr: addr,
}
//fmt.Println("server read: ", &msg)
s.readFromClientChan <- msgPck
} else {
//fmt.Println(err)
return
}
}
}
}
func (s *server) Read() (int, []byte, error) {
s.readRequest.ask <- 1
msg := <-s.readRequest.response
////fmt.Println("Server called read: ", msg)
// check if the client still alive or not
if msg.Payload == nil {
////fmt.Println("client closed when read")
s.deleteClient <- msg.ConnID
//fmt.Println("read ")
return msg.ConnID, nil, errors.New("Client connection lost")
}
// TODO: client lost return connection id
return msg.ConnID, msg.Payload, nil
}
func (s *server) Write(connID int, payload []byte) error {
//fmt.Println("server called write")
/*client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
s.writeToClientChan <- dataMsg
client.nextSN += 1*/
//////fmt.Println(s.clients[connID])
if s.clients[connID] == nil {
return errors.New("Client connection lost")
}
s.writeRequest.ask <- payload
s.writeRequest.connId <- connID
response := <-s.writeRequest.response
return response
}
func (s *server) CloseConn(connID int) error {
//fmt.Println("server closeconn!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.closeConnRequest.ask <- connID
err := <-s.closeConnRequest.getError
//fmt.Println("server closed finish!!!!!!!!!!!!!!!!!!!!!!")
return err
}
func (s *server) Close() error {
return errors.New("text")
//fmt.Println("server: server close !!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.writeFinished <- 0
//fmt.Println("Server: finish wirte1")
<-s.writeFinished
//fmt.Println("Server: finish wirte")
/*s.ackFinished <- 0
<-s.ackFinished*/
//fmt.Println("Server: finish ack")
s.closeEpoch <- 1
s.closeRead <- 1
s.closeEvent <- 1
//fmt.Println("Server: close every thing")
err := s.conn.Close()
if err == nil {
return nil
} else {
return err
}
return errors.New("not yet implemented")
}
| {
s.readList.PushBack(msg)
} | conditional_block |
server_impl.go | // Contains the implementation of a LSP server.
package lsp
import "errors"
import "strconv"
import "github.com/cmu440/lspnet"
import "container/list"
import "encoding/json"
import "fmt"
import "time"
type msgPackage struct {
msg *Message
addr *lspnet.UDPAddr
}
type abstractClient struct {
addr *lspnet.UDPAddr
nextSN int
nextDataRead int
// window size variable
// server send msg to client
nextSmallestAck int
mapAck map[int]bool
mapNeedAck map[int]*Message
// server receive msg from client
nextSmallestData int
mapReceived map[int]*Message
// variables of epoch
epochNum int
lostConn bool
// close
closeConn bool
}
type requestRead struct {
ask chan int
response chan *Message
}
type requestWrite struct {
ask chan []byte
connId chan int
response chan error
}
type requestClose struct {
ask chan int
//response chan int
getError chan error
}
type requestCloseConn struct {
ask chan int
getError chan error
}
type server struct {
nextConnectId int
clients map[int]*abstractClient
conn *lspnet.UDPConn
readFromClientChan chan *msgPackage
writeToClientChan chan *Message
readRequest *requestRead
writeRequest *requestWrite
readList *list.List
writeList *list.List
flag bool
// variables for window size
windowSize int
mapNeedSend *list.List
// variables for epoch
epochChan chan int
epochMillis int
epochLimit int
// close
deleteClient chan int
closeConnRequest *requestCloseConn
// close every thing
waitToWriteFinish bool
writeFinished chan int
waitToAckFinish bool
ackFinished chan int
closeRead chan int
closeEpoch chan int
closeEvent chan int
// close connection
closeConn chan int
}
// NewServer creates, initiates, and returns a new server. This function should
// NOT block. Instead, it should spawn one or more goroutines (to handle things
// like accepting incoming client connections, triggering epoch events at
// fixed intervals, synchronizing events using a for-select loop like you saw in
// project 0, etc.) and immediately return. It should return a non-nil error if
// there was an error resolving or listening on the specified port number.
func NewServer(port int, params *Params) (Server, error) {
s := &server{
nextConnectId: 1,
clients: make(map[int]*abstractClient),
readFromClientChan: make(chan *msgPackage),
writeToClientChan: make(chan *Message),
readRequest: &requestRead{
ask: make(chan int),
response: make(chan *Message),
},
writeRequest: &requestWrite{
ask: make(chan []byte),
connId: make(chan int),
response: make(chan error),
},
readList: list.New(),
writeList: list.New(),
flag: false,
// variables for window size
windowSize: params.WindowSize,
mapNeedSend: list.New(),
// variables for epoch
epochChan: make(chan int),
epochMillis: params.EpochMillis,
epochLimit: params.EpochLimit,
// close
deleteClient: make(chan int),
closeConnRequest: &requestCloseConn{
ask: make(chan int),
getError: make(chan error),
},
waitToWriteFinish: false,
writeFinished: make(chan int),
waitToAckFinish: false,
ackFinished: make(chan int),
closeRead: make(chan int, 1),
closeEpoch: make(chan int, 1),
closeEvent: make(chan int, 1),
// close conn
closeConn: make(chan int, 1),
}
// start server
addr, err := lspnet.ResolveUDPAddr("udp", "localhost:"+strconv.Itoa(port))
if err != nil {
return nil, err
}
conn, err := lspnet.ListenUDP("udp", addr)
if err != nil {
return nil, err
}
s.conn = conn
go s.readMessage()
go s.handleMessage()
go s.epochFire()
fmt.Println("new server")
return s, nil
}
func (s *server) epochFire() {
for {
select {
case <-s.closeEpoch:
return
default:
////fmt.Println("Server EpochFire!!!!!!!!!!!!!!!!!!!!!!!!!!")
<-time.After(time.Duration(s.epochMillis) * time.Millisecond)
s.epochChan <- 1
}
}
}
func (s *server) handleMessage() {
for {
select {
case <-s.writeFinished:
if s.writeList.Len() == 0 {
s.writeFinished <- 1
} else {
s.waitToWriteFinish = true
}
case <-s.ackFinished:
//fmt.Println("Server 192: receive ack finished")
s.checkAckFinished()
case msg := <-s.readFromClientChan:
////fmt.Println("Server 162: read from client: ", msg)
s.readFromClient(msg)
case msg := <-s.writeToClientChan:
////fmt.Println("Server 165: write to client: ", msg)
if s.clients[msg.ConnID] == nil {
continue
}
if msg.Type == MsgData {
s.clients[msg.ConnID].mapAck[msg.SeqNum] = false
s.clients[msg.ConnID].mapNeedAck[msg.SeqNum] = msg
// data type: need to consider order
if (msg.SeqNum - s.clients[msg.ConnID].nextSmallestAck) < s.windowSize {
s.writeToClient(msg)
} else {
s.writeList.PushBack(msg)
}
} else {
// ack/conn type: don't need to consider order
s.writeToClient(msg)
}
if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
case <-s.readRequest.ask:
s.handleReadRequest()
case payload := <-s.writeRequest.ask:
connId := <-s.writeRequest.connId
var response = s.handleWriteRequest(connId, payload)
s.writeRequest.response <- response
case <-s.epochChan:
s.addEpochNum()
s.handleEpochEvent()
case id := <-s.deleteClient:
delete(s.clients, id)
case id := <-s.closeConnRequest.ask:
if s.clients[id] == nil {
s.closeConnRequest.getError <- errors.New("No client")
}
s.sendDeadMsg(id)
s.closeConnRequest.getError <- nil
}
}
}
func (s *server) checkAckFinished() {
for _, c := range s.clients {
if len(c.mapAck) != 0 {
s.waitToAckFinish = true
//fmt.Println("Server: need to wait ack finish")
return
}
}
s.ackFinished <- 1
}
func (s *server) handleWriteRequest(connID int, payload []byte) error {
//////fmt.Println(connID)
client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
go s.sendData(dataMsg)
client.nextSN += 1
return nil
}
func (s *server) sendData(dataMsg *Message) {
s.writeToClientChan <- dataMsg
}
func (s *server) addEpochNum() {
for connId, c := range s.clients {
c.epochNum += 1
if c.epochNum >= s.epochLimit {
c.lostConn = true
s.sendDeadMsg(connId)
}
}
}
func (s *server) sendDeadMsg(connId int) {
dataMsg := NewData(connId, s.clients[connId].nextSN, nil, nil)
if s.flag {
s.readRequest.response <- dataMsg
s.flag = false
} else {
s.readList.PushBack(dataMsg)
}
}
func (s *server) readData(dataMsg *msgPackage) {
s.readFromClient(dataMsg)
}
func (s *server) handleEpochEvent() {
for connID, c := range s.clients {
/*
* if no data received from the client, then resend ack for connection requestRead
*/
if c.nextSmallestAck == 1 {
ack := NewAck(connID, 0)
go s.sendAck(ack)
}
/*
* sent but not acked, resend the data
*/
for key, received := range c.mapAck {
if !received {
msg := c.mapNeedAck[key]
fmt.Println("Server epoch: ", msg)
s.writeToClient(msg)
}
}
}
}
func (s *server) handleReadRequest() {
msg := s.getMessageFromReadList()
if msg != nil {
//////fmt.Println("101 no message")
/*elm := s.readList.Front()
msg := elm.Value.(*Message)*/
s.readRequest.response <- msg
} else {
s.flag = true
}
return
}
// find next msg that can be send
// return nil if no msg can be send
func (s *server) getMessageFromReadList() *Message {
for e := s.readList.Front(); e != nil; e = e.Next() {
msg := e.Value.(*Message)
if msg.SeqNum == s.clients[msg.ConnID].nextDataRead {
s.clients[msg.ConnID].nextDataRead += 1
s.readList.Remove(e)
return msg
}
}
return nil
}
func (s *server) writeToClient(msg *Message) error {
connId := msg.ConnID
addr := s.clients[connId].addr
mmsg, _ := json.Marshal(msg)
fmt.Println("364 Server write to client: ", msg)
_, err := s.conn.WriteToUDP(mmsg, addr)
if err != nil {
fmt.Println("364: ", err)
}
return err
}
func (s *server) readFromClient(msgPack *msgPackage) {
//////fmt.Println("129 read")
msg := msgPack.msg
addr := msgPack.addr
////fmt.Println("server read: ", msg)
if msg.Type == MsgConnect {
s.connectClient(msg, addr)
} else {
// set epoch number to 0
if s.clients[msg.ConnID] != nil {
s.clients[msg.ConnID].epochNum = 0
}
if msg.Type == MsgData {
if !s.waitToAckFinish {
//fmt.Println("Server read from client: ", msg)
s.receiveData(msg)
}
} else if msg.Type == MsgAck {
fmt.Println("389 Server: receive ack: ", msg)
s.receiveAck(msg)
}
if s.waitToAckFinish {
s.checkAckFinished()
}
}
return
}
func (s *server) alreadyHasClient(addr *lspnet.UDPAddr) bool {
////fmt.Println("server: in already has client")
for _, client := range s.clients {
if client.addr.String() == addr.String() {
return true
}
}
return false
}
func (s *server) receiveData(msg *Message) {
//fmt.Println("Server read: ", msg)
c := s.clients[msg.ConnID]
// check duplicated forst and within window size
/*if c.nextSmallestData+s.windowSize <= msg.SeqNum {
// check within window size, drop it if not
return
} else if c.mapReceived[msg.SeqNum] != nil {
// check duplicated outside of window size
return
}*/
//fmt.Println("pass window size")
// here is no duplicated
// decide should we push it to the readlist or chan
if s.flag && (s.clients[msg.ConnID].nextDataRead == msg.SeqNum) {
s.readRequest.response <- msg
c.nextDataRead += 1
s.flag = false
} else if msg.SeqNum >= c.nextDataRead {
s.readList.PushBack(msg)
}
// create ack message
connID := msg.ConnID
seqNum := msg.SeqNum
ack := NewAck(connID, seqNum)
// check the receive variables in client
// delete the message in mapreceived
if msg.SeqNum == c.nextSmallestData {
c.nextSmallestData += 1
for c.mapReceived[c.nextSmallestData] != nil {
delete(c.mapReceived, c.nextSmallestData)
c.nextSmallestData += 1
}
} else {
c.mapReceived[msg.SeqNum] = msg
}
go s.sendAck(ack)
return
}
func (s *server) sendAck(ack *Message) {
s.writeToClientChan <- ack
}
func (s *server) receiveAck(msg *Message) {
//fmt.Println("server: receiveAck")
c := s.clients[msg.ConnID]
if msg.SeqNum >= c.nextSmallestAck {
c.mapAck[msg.SeqNum] = true
}
if msg.SeqNum == c.nextSmallestAck {
// reset the next smallest ack numebr
value, exist := c.mapAck[c.nextSmallestAck]
for exist && value {
delete(c.mapAck, c.nextSmallestAck)
c.nextSmallestAck += 1
value, exist = c.mapAck[c.nextSmallestAck]
}
var l = *list.New()
for element := s.writeList.Front(); element != nil; element = element.Next() {
message := element.Value.(*Message)
if message.ConnID == msg.ConnID {
if (message.SeqNum - c.nextSmallestAck) < s.windowSize {
s.writeToClient(message)
l.PushBack(element)
}
}
}
////fmt.Println("loop finish go out of write")
for element := l.Front(); element != nil; element = element.Next() {
s.writeList.Remove(element)
}
////fmt.Println("go out of delte dfadfasdf")
if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
////fmt.Println("go asdfasdfe dfadfasdf")
}
return
}
func (s *server) connectClient(msg *Message, addr *lspnet.UDPAddr) {
////fmt.Println("Server: connect client ", msg)
// check duplication of connection
if s.alreadyHasClient(addr) {
////fmt.Println("already has client")
return
}
connID := s.nextConnectId
newClient := abstractClient{
addr: addr,
nextSN: 1,
nextDataRead: 1,
// server send msg to client
nextSmallestAck: 1,
mapAck: make(map[int]bool),
mapNeedAck: make(map[int]*Message),
// server receive msg from client
nextSmallestData: 1,
mapReceived: make(map[int]*Message),
// epoch
epochNum: 0,
lostConn: false,
// close
closeConn: false,
}
s.clients[connID] = &newClient
s.nextConnectId += 1
ack := NewAck(connID, 0)
////fmt.Println("Server: send ack to client")
go s.sendAck(ack)
return
}
func (s *server) readMessage() {
var b [1500]byte
//////fmt.Println("Read: read message")
for {
select {
case <-s.closeRead:
return
default:
var msg Message
// unmarshal
n, addr, err := s.conn.ReadFromUDP(b[0:])
err = json.Unmarshal(b[0:n], &msg)
if err == nil {
msgPck := &msgPackage{
msg: &msg,
addr: addr,
}
//fmt.Println("server read: ", &msg)
s.readFromClientChan <- msgPck
} else {
//fmt.Println(err)
return
}
}
}
}
func (s *server) Read() (int, []byte, error) {
s.readRequest.ask <- 1
msg := <-s.readRequest.response
////fmt.Println("Server called read: ", msg)
// check if the client still alive or not
if msg.Payload == nil {
////fmt.Println("client closed when read")
s.deleteClient <- msg.ConnID
//fmt.Println("read ")
return msg.ConnID, nil, errors.New("Client connection lost")
}
// TODO: client lost return connection id
return msg.ConnID, msg.Payload, nil
}
func (s *server) Write(connID int, payload []byte) error {
//fmt.Println("server called write")
/*client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
s.writeToClientChan <- dataMsg
client.nextSN += 1*/
//////fmt.Println(s.clients[connID])
if s.clients[connID] == nil {
return errors.New("Client connection lost")
}
s.writeRequest.ask <- payload
s.writeRequest.connId <- connID
response := <-s.writeRequest.response
return response
}
func (s *server) CloseConn(connID int) error {
//fmt.Println("server closeconn!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.closeConnRequest.ask <- connID
err := <-s.closeConnRequest.getError
//fmt.Println("server closed finish!!!!!!!!!!!!!!!!!!!!!!")
return err
}
func (s *server) | () error {
return errors.New("text")
//fmt.Println("server: server close !!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.writeFinished <- 0
//fmt.Println("Server: finish wirte1")
<-s.writeFinished
//fmt.Println("Server: finish wirte")
/*s.ackFinished <- 0
<-s.ackFinished*/
//fmt.Println("Server: finish ack")
s.closeEpoch <- 1
s.closeRead <- 1
s.closeEvent <- 1
//fmt.Println("Server: close every thing")
err := s.conn.Close()
if err == nil {
return nil
} else {
return err
}
return errors.New("not yet implemented")
}
| Close | identifier_name |
server_impl.go | // Contains the implementation of a LSP server.
package lsp
import "errors"
import "strconv"
import "github.com/cmu440/lspnet"
import "container/list"
import "encoding/json"
import "fmt"
import "time"
type msgPackage struct {
msg *Message
addr *lspnet.UDPAddr
}
type abstractClient struct {
addr *lspnet.UDPAddr
nextSN int
nextDataRead int
// window size variable
// server send msg to client
nextSmallestAck int
mapAck map[int]bool
mapNeedAck map[int]*Message
// server receive msg from client
nextSmallestData int
mapReceived map[int]*Message
// variables of epoch
epochNum int
lostConn bool
// close
closeConn bool
}
type requestRead struct {
ask chan int
response chan *Message
}
type requestWrite struct {
ask chan []byte
connId chan int
response chan error
}
type requestClose struct {
ask chan int
//response chan int
getError chan error
}
type requestCloseConn struct {
ask chan int
getError chan error
}
type server struct {
nextConnectId int
clients map[int]*abstractClient
conn *lspnet.UDPConn
readFromClientChan chan *msgPackage
writeToClientChan chan *Message
readRequest *requestRead
writeRequest *requestWrite
readList *list.List
writeList *list.List
flag bool
// variables for window size
windowSize int
mapNeedSend *list.List
// variables for epoch
epochChan chan int
epochMillis int
epochLimit int
// close
deleteClient chan int
closeConnRequest *requestCloseConn
// close every thing
waitToWriteFinish bool
writeFinished chan int
waitToAckFinish bool
ackFinished chan int
closeRead chan int
closeEpoch chan int
closeEvent chan int
// close connection
closeConn chan int
}
// NewServer creates, initiates, and returns a new server. This function should
// NOT block. Instead, it should spawn one or more goroutines (to handle things
// like accepting incoming client connections, triggering epoch events at
// fixed intervals, synchronizing events using a for-select loop like you saw in
// project 0, etc.) and immediately return. It should return a non-nil error if
// there was an error resolving or listening on the specified port number.
func NewServer(port int, params *Params) (Server, error) {
s := &server{
nextConnectId: 1,
clients: make(map[int]*abstractClient),
readFromClientChan: make(chan *msgPackage),
writeToClientChan: make(chan *Message),
readRequest: &requestRead{
ask: make(chan int),
response: make(chan *Message),
},
writeRequest: &requestWrite{
ask: make(chan []byte),
connId: make(chan int),
response: make(chan error),
},
readList: list.New(),
writeList: list.New(),
flag: false,
// variables for window size
windowSize: params.WindowSize,
mapNeedSend: list.New(),
// variables for epoch
epochChan: make(chan int),
epochMillis: params.EpochMillis,
epochLimit: params.EpochLimit,
// close
deleteClient: make(chan int),
closeConnRequest: &requestCloseConn{
ask: make(chan int),
getError: make(chan error),
},
waitToWriteFinish: false,
writeFinished: make(chan int),
waitToAckFinish: false,
ackFinished: make(chan int),
closeRead: make(chan int, 1),
closeEpoch: make(chan int, 1),
closeEvent: make(chan int, 1),
// close conn
closeConn: make(chan int, 1),
}
// start server
addr, err := lspnet.ResolveUDPAddr("udp", "localhost:"+strconv.Itoa(port))
if err != nil {
return nil, err
}
conn, err := lspnet.ListenUDP("udp", addr)
if err != nil {
return nil, err
}
s.conn = conn
go s.readMessage()
go s.handleMessage()
go s.epochFire()
fmt.Println("new server")
return s, nil
}
func (s *server) epochFire() {
for {
select {
case <-s.closeEpoch:
return
default:
////fmt.Println("Server EpochFire!!!!!!!!!!!!!!!!!!!!!!!!!!")
<-time.After(time.Duration(s.epochMillis) * time.Millisecond)
s.epochChan <- 1
}
}
}
func (s *server) handleMessage() {
for {
select {
case <-s.writeFinished:
if s.writeList.Len() == 0 {
s.writeFinished <- 1
} else {
s.waitToWriteFinish = true
}
case <-s.ackFinished:
//fmt.Println("Server 192: receive ack finished")
s.checkAckFinished()
case msg := <-s.readFromClientChan:
////fmt.Println("Server 162: read from client: ", msg)
s.readFromClient(msg)
case msg := <-s.writeToClientChan:
////fmt.Println("Server 165: write to client: ", msg)
if s.clients[msg.ConnID] == nil {
continue
}
if msg.Type == MsgData {
s.clients[msg.ConnID].mapAck[msg.SeqNum] = false
s.clients[msg.ConnID].mapNeedAck[msg.SeqNum] = msg
// data type: need to consider order
if (msg.SeqNum - s.clients[msg.ConnID].nextSmallestAck) < s.windowSize {
s.writeToClient(msg)
} else {
s.writeList.PushBack(msg)
}
} else { | if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
case <-s.readRequest.ask:
s.handleReadRequest()
case payload := <-s.writeRequest.ask:
connId := <-s.writeRequest.connId
var response = s.handleWriteRequest(connId, payload)
s.writeRequest.response <- response
case <-s.epochChan:
s.addEpochNum()
s.handleEpochEvent()
case id := <-s.deleteClient:
delete(s.clients, id)
case id := <-s.closeConnRequest.ask:
if s.clients[id] == nil {
s.closeConnRequest.getError <- errors.New("No client")
}
s.sendDeadMsg(id)
s.closeConnRequest.getError <- nil
}
}
}
func (s *server) checkAckFinished() {
for _, c := range s.clients {
if len(c.mapAck) != 0 {
s.waitToAckFinish = true
//fmt.Println("Server: need to wait ack finish")
return
}
}
s.ackFinished <- 1
}
func (s *server) handleWriteRequest(connID int, payload []byte) error {
//////fmt.Println(connID)
client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
go s.sendData(dataMsg)
client.nextSN += 1
return nil
}
func (s *server) sendData(dataMsg *Message) {
s.writeToClientChan <- dataMsg
}
func (s *server) addEpochNum() {
for connId, c := range s.clients {
c.epochNum += 1
if c.epochNum >= s.epochLimit {
c.lostConn = true
s.sendDeadMsg(connId)
}
}
}
func (s *server) sendDeadMsg(connId int) {
dataMsg := NewData(connId, s.clients[connId].nextSN, nil, nil)
if s.flag {
s.readRequest.response <- dataMsg
s.flag = false
} else {
s.readList.PushBack(dataMsg)
}
}
func (s *server) readData(dataMsg *msgPackage) {
s.readFromClient(dataMsg)
}
func (s *server) handleEpochEvent() {
for connID, c := range s.clients {
/*
* if no data received from the client, then resend ack for connection requestRead
*/
if c.nextSmallestAck == 1 {
ack := NewAck(connID, 0)
go s.sendAck(ack)
}
/*
* sent but not acked, resend the data
*/
for key, received := range c.mapAck {
if !received {
msg := c.mapNeedAck[key]
fmt.Println("Server epoch: ", msg)
s.writeToClient(msg)
}
}
}
}
func (s *server) handleReadRequest() {
msg := s.getMessageFromReadList()
if msg != nil {
//////fmt.Println("101 no message")
/*elm := s.readList.Front()
msg := elm.Value.(*Message)*/
s.readRequest.response <- msg
} else {
s.flag = true
}
return
}
// find next msg that can be send
// return nil if no msg can be send
func (s *server) getMessageFromReadList() *Message {
for e := s.readList.Front(); e != nil; e = e.Next() {
msg := e.Value.(*Message)
if msg.SeqNum == s.clients[msg.ConnID].nextDataRead {
s.clients[msg.ConnID].nextDataRead += 1
s.readList.Remove(e)
return msg
}
}
return nil
}
func (s *server) writeToClient(msg *Message) error {
connId := msg.ConnID
addr := s.clients[connId].addr
mmsg, _ := json.Marshal(msg)
fmt.Println("364 Server write to client: ", msg)
_, err := s.conn.WriteToUDP(mmsg, addr)
if err != nil {
fmt.Println("364: ", err)
}
return err
}
func (s *server) readFromClient(msgPack *msgPackage) {
//////fmt.Println("129 read")
msg := msgPack.msg
addr := msgPack.addr
////fmt.Println("server read: ", msg)
if msg.Type == MsgConnect {
s.connectClient(msg, addr)
} else {
// set epoch number to 0
if s.clients[msg.ConnID] != nil {
s.clients[msg.ConnID].epochNum = 0
}
if msg.Type == MsgData {
if !s.waitToAckFinish {
//fmt.Println("Server read from client: ", msg)
s.receiveData(msg)
}
} else if msg.Type == MsgAck {
fmt.Println("389 Server: receive ack: ", msg)
s.receiveAck(msg)
}
if s.waitToAckFinish {
s.checkAckFinished()
}
}
return
}
func (s *server) alreadyHasClient(addr *lspnet.UDPAddr) bool {
////fmt.Println("server: in already has client")
for _, client := range s.clients {
if client.addr.String() == addr.String() {
return true
}
}
return false
}
func (s *server) receiveData(msg *Message) {
//fmt.Println("Server read: ", msg)
c := s.clients[msg.ConnID]
// check duplicated forst and within window size
/*if c.nextSmallestData+s.windowSize <= msg.SeqNum {
// check within window size, drop it if not
return
} else if c.mapReceived[msg.SeqNum] != nil {
// check duplicated outside of window size
return
}*/
//fmt.Println("pass window size")
// here is no duplicated
// decide should we push it to the readlist or chan
if s.flag && (s.clients[msg.ConnID].nextDataRead == msg.SeqNum) {
s.readRequest.response <- msg
c.nextDataRead += 1
s.flag = false
} else if msg.SeqNum >= c.nextDataRead {
s.readList.PushBack(msg)
}
// create ack message
connID := msg.ConnID
seqNum := msg.SeqNum
ack := NewAck(connID, seqNum)
// check the receive variables in client
// delete the message in mapreceived
if msg.SeqNum == c.nextSmallestData {
c.nextSmallestData += 1
for c.mapReceived[c.nextSmallestData] != nil {
delete(c.mapReceived, c.nextSmallestData)
c.nextSmallestData += 1
}
} else {
c.mapReceived[msg.SeqNum] = msg
}
go s.sendAck(ack)
return
}
func (s *server) sendAck(ack *Message) {
s.writeToClientChan <- ack
}
func (s *server) receiveAck(msg *Message) {
//fmt.Println("server: receiveAck")
c := s.clients[msg.ConnID]
if msg.SeqNum >= c.nextSmallestAck {
c.mapAck[msg.SeqNum] = true
}
if msg.SeqNum == c.nextSmallestAck {
// reset the next smallest ack numebr
value, exist := c.mapAck[c.nextSmallestAck]
for exist && value {
delete(c.mapAck, c.nextSmallestAck)
c.nextSmallestAck += 1
value, exist = c.mapAck[c.nextSmallestAck]
}
var l = *list.New()
for element := s.writeList.Front(); element != nil; element = element.Next() {
message := element.Value.(*Message)
if message.ConnID == msg.ConnID {
if (message.SeqNum - c.nextSmallestAck) < s.windowSize {
s.writeToClient(message)
l.PushBack(element)
}
}
}
////fmt.Println("loop finish go out of write")
for element := l.Front(); element != nil; element = element.Next() {
s.writeList.Remove(element)
}
////fmt.Println("go out of delte dfadfasdf")
if s.waitToWriteFinish && s.writeList.Len() == 0 {
s.writeFinished <- 1
s.waitToWriteFinish = false
}
////fmt.Println("go asdfasdfe dfadfasdf")
}
return
}
func (s *server) connectClient(msg *Message, addr *lspnet.UDPAddr) {
////fmt.Println("Server: connect client ", msg)
// check duplication of connection
if s.alreadyHasClient(addr) {
////fmt.Println("already has client")
return
}
connID := s.nextConnectId
newClient := abstractClient{
addr: addr,
nextSN: 1,
nextDataRead: 1,
// server send msg to client
nextSmallestAck: 1,
mapAck: make(map[int]bool),
mapNeedAck: make(map[int]*Message),
// server receive msg from client
nextSmallestData: 1,
mapReceived: make(map[int]*Message),
// epoch
epochNum: 0,
lostConn: false,
// close
closeConn: false,
}
s.clients[connID] = &newClient
s.nextConnectId += 1
ack := NewAck(connID, 0)
////fmt.Println("Server: send ack to client")
go s.sendAck(ack)
return
}
func (s *server) readMessage() {
var b [1500]byte
//////fmt.Println("Read: read message")
for {
select {
case <-s.closeRead:
return
default:
var msg Message
// unmarshal
n, addr, err := s.conn.ReadFromUDP(b[0:])
err = json.Unmarshal(b[0:n], &msg)
if err == nil {
msgPck := &msgPackage{
msg: &msg,
addr: addr,
}
//fmt.Println("server read: ", &msg)
s.readFromClientChan <- msgPck
} else {
//fmt.Println(err)
return
}
}
}
}
func (s *server) Read() (int, []byte, error) {
s.readRequest.ask <- 1
msg := <-s.readRequest.response
////fmt.Println("Server called read: ", msg)
// check if the client still alive or not
if msg.Payload == nil {
////fmt.Println("client closed when read")
s.deleteClient <- msg.ConnID
//fmt.Println("read ")
return msg.ConnID, nil, errors.New("Client connection lost")
}
// TODO: client lost return connection id
return msg.ConnID, msg.Payload, nil
}
func (s *server) Write(connID int, payload []byte) error {
//fmt.Println("server called write")
/*client := s.clients[connID]
dataMsg := NewData(connID, client.nextSN, payload, payload)
////fmt.Println("Server: 197 write ", dataMsg)
s.writeToClientChan <- dataMsg
client.nextSN += 1*/
//////fmt.Println(s.clients[connID])
if s.clients[connID] == nil {
return errors.New("Client connection lost")
}
s.writeRequest.ask <- payload
s.writeRequest.connId <- connID
response := <-s.writeRequest.response
return response
}
func (s *server) CloseConn(connID int) error {
//fmt.Println("server closeconn!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.closeConnRequest.ask <- connID
err := <-s.closeConnRequest.getError
//fmt.Println("server closed finish!!!!!!!!!!!!!!!!!!!!!!")
return err
}
func (s *server) Close() error {
return errors.New("text")
//fmt.Println("server: server close !!!!!!!!!!!!!!!!!!!!!!!!!!!!")
s.writeFinished <- 0
//fmt.Println("Server: finish wirte1")
<-s.writeFinished
//fmt.Println("Server: finish wirte")
/*s.ackFinished <- 0
<-s.ackFinished*/
//fmt.Println("Server: finish ack")
s.closeEpoch <- 1
s.closeRead <- 1
s.closeEvent <- 1
//fmt.Println("Server: close every thing")
err := s.conn.Close()
if err == nil {
return nil
} else {
return err
}
return errors.New("not yet implemented")
} | // ack/conn type: don't need to consider order
s.writeToClient(msg)
}
| random_line_split |
simulation.py | """Assignment 1 - Simulation
=== CSC148 Fall 2017 ===
Diane Horton and David Liu
Department of Computer Science,
University of Toronto
=== Module Description ===
This file contains the Simulation class, which is the main class for your
bike-share simulation.
At the bottom of the file, there is a sample_simulation function that you
can use to try running the simulation at any time.
"""
import csv
from datetime import datetime, timedelta
import json
from typing import Dict, List, Tuple
from bikeshare import Ride, Station
from container import PriorityQueue
from visualizer import Visualizer
# Datetime format to parse the ride data
DATETIME_FORMAT = '%Y-%m-%d %H:%M'
#Function constants for reporting beggining and ending of rides
RIDE_BEGUN = True
RIDE_FINISHED = False
# Stock issues report
SUCCESSFUL_REPORT = 0
EMPTY_STATION_ISSUE = 1
FULL_STATION_ISSUE = 2
class Simulation:
"""Runs the core of the simulation through time.
=== Attributes ===
all_rides:
A list of all the rides in this simulation.
Note that not all rides might be used, depending on the timeframe
when the simulation is run.
all_stations:
A dictionary containing all the stations in this simulation.
visualizer:
A helper class for visualizing the simulation.
active_rides:
All rides that are active during simulation
"""
all_stations: Dict[str, Station]
all_rides: List[Ride]
visualizer: Visualizer
active_rides: List[Ride]
def __init__(self, station_file: str, ride_file: str) -> None:
"""Initialize this simulation with the given configuration settings.
"""
self.all_stations = create_stations(station_file)
self.all_rides = create_rides(ride_file, self.all_stations)
self.visualizer = Visualizer()
self.active_rides = []
def run(self, start: datetime, end: datetime) -> None:
"""Run the simulation from <start> to <end>.
"""
step = timedelta(minutes=1) # Each iteration spans one minute of time
current_time = start
while current_time <= end:
self._update_active_rides(current_time)
# Method that draws all stations and bikes on the map
self.visualizer.render_drawables(list(self.all_stations.values())\
+ self.active_rides, current_time)
current_time += step
# Leave this code at the very bottom of this method.
# It will keep the visualization window open until you close
# it by pressing the 'X'.
while True:
if self.visualizer.handle_window_events():
return # Stop the simulation
def _update_active_rides(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- Loop through `self.all_rides` and compare each Ride's start and
end times with <time>.
If <time> is between the ride's start and end times (inclusive),
then add the ride to self.active_rides if it isn't already in
that list.
Otherwise, remove the ride from self.active_rides if it is in
that list.
- This means that if a ride started before the simulation's time
period but ends during or after the simulation's time period,
it should still be added to self.active_rides.
"""
for ride in self.all_rides:
# Check whether the current time is between start and end
if time > ride.start_time and time < ride.end_time:
# Check whether it is already in the active rides list
if not ride in self.active_rides:
# Reports that this ride has just begun
# this method will also lead to an update on station's bike
# number.
if ride.report_status(RIDE_BEGUN, time) \
== SUCCESSFUL_REPORT:
# Adds to the active rides list if no anomaly was
# detected
self.active_rides.append(ride)
else:
if ride in self.active_rides:
# Reports that this ride has finished
# this method will also lead to an update on station's bike
# number
ride.report_status(RIDE_FINISHED, time)
# If a finished or not yet beginned ride is in this
# list, removes it.
self.active_rides.remove(ride)
def calculate_statistics(self) -> Dict[str, Tuple[str, float]]:
|
def _update_active_rides_fast(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- see Task 5 of the assignment handout
"""
pass
def create_stations(stations_file: str) -> Dict[str, 'Station']:
"""Return the stations described in the given JSON data file.
Each key in the returned dictionary is a station id,
and each value is the corresponding Station object.
Note that you need to call Station(...) to create these objects!
Precondition: stations_file matches the format specified in the
assignment handout.
This function should be called *before* _read_rides because the
rides CSV file refers to station ids.
"""
# Read in raw data using the json library.
with open(stations_file) as file:
raw_stations = json.load(file)
stations = {}
for s in raw_stations['stations']:
# Extract the relevant fields from the raw station JSON.
# s is a dictionary with the keys 'n', 's', 'la', 'lo', 'da', and 'ba'
# as described in the assignment handout.
# NOTE: all of the corresponding values are strings, and so you need
# to convert some of them to numbers explicitly using int() or float().
# Inputs for Station Class (with keys from dictionary) :
# -Lon/Lat (lo/la)
# -Capacity (da + ba)
# -Number of bikes (da)
# -Name (s)
station = Station((s['lo'], s['la']), s['da'] + s['ba'],\
s['da'], s['s'])
stations[s['n']] = station
return stations
def create_rides(rides_file: str,
stations: Dict[str, 'Station']) -> List['Ride']:
"""Return the rides described in the given CSV file.
Lookup the station ids contained in the rides file in <stations>
to access the corresponding Station objects.
Ignore any ride whose start or end station is not present in <stations>.
Precondition: rides_file matches the format specified in the
assignment handout.
"""
rides = []
with open(rides_file) as file:
for line in csv.reader(file):
# line is a list of strings, following the format described
# in the assignment handout.
#
# Convert between a string and a datetime object
# using the function datetime.strptime and the DATETIME_FORMAT
# constant we defined above. Example:
# >>> datetime.strptime('2017-06-01 8:00', DATETIME_FORMAT)
# datetime.datetime(2017, 6, 1, 8, 0)
# CSV Columns:
# -Start time [0]
# -Start station code [1]
# -End time [2]
# -End station code [3]
# -ignore the rest [4,5]
# Ride class inputs:
# -Start station (Station())
# -End station (Station())
# -Times (Tuple(start,end))
start_station_code = line[1]
end_station_code = line[3]
# First check whether the stations from the ride are present in the
# stations registry
if start_station_code and end_station_code in stations:
start_time = datetime.strptime(line[0], DATETIME_FORMAT)
end_time = datetime.strptime(line[2], DATETIME_FORMAT)
start_station = stations[start_station_code]
end_station = stations[end_station_code]
ride = Ride(start_station, end_station, (start_time, end_time))
rides.append(ride)
return rides
class Event:
"""An event in the bike share simulation.
Events are ordered by their timestamp.
"""
simulation: 'Simulation'
time: datetime
def __init__(self, simulation: 'Simulation', time: datetime) -> None:
"""Initialize a new event."""
self.simulation = simulation
self.time = time
def __lt__(self, other: 'Event') -> bool:
"""Return whether this event is less than <other>.
Events are ordered by their timestamp.
"""
return self.time < other.time
def process(self) -> List['Event']:
"""Process this event by updating the state of the simulation.
Return a list of new events spawned by this event.
"""
raise NotImplementedError
class RideStartEvent(Event):
"""An event corresponding to the start of a ride."""
pass
class RideEndEvent(Event):
"""An event corresponding to the start of a ride."""
pass
def sample_simulation() -> Dict[str, Tuple[str, float]]:
"""Run a sample simulation. For testing purposes only."""
sim = Simulation('stations.json', 'sample_rides.csv')
sim.run(datetime(2017, 6, 1, 8, 0, 0),
datetime(2017, 6, 1, 9, 0, 0))
return sim.calculate_statistics()
if __name__ == '__main__':
# Uncomment these lines when you want to check your work using python_ta!
# import python_ta
# python_ta.check_all(config={
# 'allowed-io': ['create_stations', 'create_rides'],
# 'allowed-import-modules': [
# 'doctest', 'python_ta', 'typing',
# 'csv', 'datetime', 'json',
# 'bikeshare', 'container', 'visualizer'
# ]
# })
print(sample_simulation())
| """Return a dictionary containing statistics for this simulation.
The returned dictionary has exactly four keys, corresponding
to the four statistics tracked for each station:
- 'max_start'
- 'max_end'
- 'max_time_low_availability'
- 'max_time_low_unoccupied'
The corresponding value of each key is a tuple of two elements,
where the first element is the name (NOT id) of the station that has
the maximum value of the quantity specified by that key,
and the second element is the value of that quantity.
For example, the value corresponding to key 'max_start' should be the
name of the station with the most number of rides started at that
station, and the number of rides that started at that station.
"""
tempDict = {
'max_start': ('', -1),
'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
}
return {
'max_start': ('', -1),
'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
} | identifier_body |
simulation.py | """Assignment 1 - Simulation
=== CSC148 Fall 2017 ===
Diane Horton and David Liu
Department of Computer Science,
University of Toronto
=== Module Description ===
This file contains the Simulation class, which is the main class for your
bike-share simulation.
At the bottom of the file, there is a sample_simulation function that you
can use to try running the simulation at any time.
"""
import csv
from datetime import datetime, timedelta
import json
from typing import Dict, List, Tuple
from bikeshare import Ride, Station
from container import PriorityQueue
from visualizer import Visualizer
# Datetime format to parse the ride data
DATETIME_FORMAT = '%Y-%m-%d %H:%M'
#Function constants for reporting beggining and ending of rides
RIDE_BEGUN = True
RIDE_FINISHED = False
# Stock issues report
SUCCESSFUL_REPORT = 0
EMPTY_STATION_ISSUE = 1
FULL_STATION_ISSUE = 2
class | :
"""Runs the core of the simulation through time.
=== Attributes ===
all_rides:
A list of all the rides in this simulation.
Note that not all rides might be used, depending on the timeframe
when the simulation is run.
all_stations:
A dictionary containing all the stations in this simulation.
visualizer:
A helper class for visualizing the simulation.
active_rides:
All rides that are active during simulation
"""
all_stations: Dict[str, Station]
all_rides: List[Ride]
visualizer: Visualizer
active_rides: List[Ride]
def __init__(self, station_file: str, ride_file: str) -> None:
"""Initialize this simulation with the given configuration settings.
"""
self.all_stations = create_stations(station_file)
self.all_rides = create_rides(ride_file, self.all_stations)
self.visualizer = Visualizer()
self.active_rides = []
def run(self, start: datetime, end: datetime) -> None:
"""Run the simulation from <start> to <end>.
"""
step = timedelta(minutes=1) # Each iteration spans one minute of time
current_time = start
while current_time <= end:
self._update_active_rides(current_time)
# Method that draws all stations and bikes on the map
self.visualizer.render_drawables(list(self.all_stations.values())\
+ self.active_rides, current_time)
current_time += step
# Leave this code at the very bottom of this method.
# It will keep the visualization window open until you close
# it by pressing the 'X'.
while True:
if self.visualizer.handle_window_events():
return # Stop the simulation
def _update_active_rides(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- Loop through `self.all_rides` and compare each Ride's start and
end times with <time>.
If <time> is between the ride's start and end times (inclusive),
then add the ride to self.active_rides if it isn't already in
that list.
Otherwise, remove the ride from self.active_rides if it is in
that list.
- This means that if a ride started before the simulation's time
period but ends during or after the simulation's time period,
it should still be added to self.active_rides.
"""
for ride in self.all_rides:
# Check whether the current time is between start and end
if time > ride.start_time and time < ride.end_time:
# Check whether it is already in the active rides list
if not ride in self.active_rides:
# Reports that this ride has just begun
# this method will also lead to an update on station's bike
# number.
if ride.report_status(RIDE_BEGUN, time) \
== SUCCESSFUL_REPORT:
# Adds to the active rides list if no anomaly was
# detected
self.active_rides.append(ride)
else:
if ride in self.active_rides:
# Reports that this ride has finished
# this method will also lead to an update on station's bike
# number
ride.report_status(RIDE_FINISHED, time)
# If a finished or not yet beginned ride is in this
# list, removes it.
self.active_rides.remove(ride)
def calculate_statistics(self) -> Dict[str, Tuple[str, float]]:
"""Return a dictionary containing statistics for this simulation.
The returned dictionary has exactly four keys, corresponding
to the four statistics tracked for each station:
- 'max_start'
- 'max_end'
- 'max_time_low_availability'
- 'max_time_low_unoccupied'
The corresponding value of each key is a tuple of two elements,
where the first element is the name (NOT id) of the station that has
the maximum value of the quantity specified by that key,
and the second element is the value of that quantity.
For example, the value corresponding to key 'max_start' should be the
name of the station with the most number of rides started at that
station, and the number of rides that started at that station.
"""
tempDict = {
'max_start': ('', -1),
'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
}
return {
'max_start': ('', -1),
'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
}
def _update_active_rides_fast(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- see Task 5 of the assignment handout
"""
pass
def create_stations(stations_file: str) -> Dict[str, 'Station']:
"""Return the stations described in the given JSON data file.
Each key in the returned dictionary is a station id,
and each value is the corresponding Station object.
Note that you need to call Station(...) to create these objects!
Precondition: stations_file matches the format specified in the
assignment handout.
This function should be called *before* _read_rides because the
rides CSV file refers to station ids.
"""
# Read in raw data using the json library.
with open(stations_file) as file:
raw_stations = json.load(file)
stations = {}
for s in raw_stations['stations']:
# Extract the relevant fields from the raw station JSON.
# s is a dictionary with the keys 'n', 's', 'la', 'lo', 'da', and 'ba'
# as described in the assignment handout.
# NOTE: all of the corresponding values are strings, and so you need
# to convert some of them to numbers explicitly using int() or float().
# Inputs for Station Class (with keys from dictionary) :
# -Lon/Lat (lo/la)
# -Capacity (da + ba)
# -Number of bikes (da)
# -Name (s)
station = Station((s['lo'], s['la']), s['da'] + s['ba'],\
s['da'], s['s'])
stations[s['n']] = station
return stations
def create_rides(rides_file: str,
stations: Dict[str, 'Station']) -> List['Ride']:
"""Return the rides described in the given CSV file.
Lookup the station ids contained in the rides file in <stations>
to access the corresponding Station objects.
Ignore any ride whose start or end station is not present in <stations>.
Precondition: rides_file matches the format specified in the
assignment handout.
"""
rides = []
with open(rides_file) as file:
for line in csv.reader(file):
# line is a list of strings, following the format described
# in the assignment handout.
#
# Convert between a string and a datetime object
# using the function datetime.strptime and the DATETIME_FORMAT
# constant we defined above. Example:
# >>> datetime.strptime('2017-06-01 8:00', DATETIME_FORMAT)
# datetime.datetime(2017, 6, 1, 8, 0)
# CSV Columns:
# -Start time [0]
# -Start station code [1]
# -End time [2]
# -End station code [3]
# -ignore the rest [4,5]
# Ride class inputs:
# -Start station (Station())
# -End station (Station())
# -Times (Tuple(start,end))
start_station_code = line[1]
end_station_code = line[3]
# First check whether the stations from the ride are present in the
# stations registry
if start_station_code and end_station_code in stations:
start_time = datetime.strptime(line[0], DATETIME_FORMAT)
end_time = datetime.strptime(line[2], DATETIME_FORMAT)
start_station = stations[start_station_code]
end_station = stations[end_station_code]
ride = Ride(start_station, end_station, (start_time, end_time))
rides.append(ride)
return rides
class Event:
"""An event in the bike share simulation.
Events are ordered by their timestamp.
"""
simulation: 'Simulation'
time: datetime
def __init__(self, simulation: 'Simulation', time: datetime) -> None:
"""Initialize a new event."""
self.simulation = simulation
self.time = time
def __lt__(self, other: 'Event') -> bool:
"""Return whether this event is less than <other>.
Events are ordered by their timestamp.
"""
return self.time < other.time
def process(self) -> List['Event']:
"""Process this event by updating the state of the simulation.
Return a list of new events spawned by this event.
"""
raise NotImplementedError
class RideStartEvent(Event):
"""An event corresponding to the start of a ride."""
pass
class RideEndEvent(Event):
"""An event corresponding to the start of a ride."""
pass
def sample_simulation() -> Dict[str, Tuple[str, float]]:
"""Run a sample simulation. For testing purposes only."""
sim = Simulation('stations.json', 'sample_rides.csv')
sim.run(datetime(2017, 6, 1, 8, 0, 0),
datetime(2017, 6, 1, 9, 0, 0))
return sim.calculate_statistics()
if __name__ == '__main__':
# Uncomment these lines when you want to check your work using python_ta!
# import python_ta
# python_ta.check_all(config={
# 'allowed-io': ['create_stations', 'create_rides'],
# 'allowed-import-modules': [
# 'doctest', 'python_ta', 'typing',
# 'csv', 'datetime', 'json',
# 'bikeshare', 'container', 'visualizer'
# ]
# })
print(sample_simulation())
| Simulation | identifier_name |
simulation.py | """Assignment 1 - Simulation
=== CSC148 Fall 2017 ===
Diane Horton and David Liu
Department of Computer Science,
University of Toronto
=== Module Description ===
This file contains the Simulation class, which is the main class for your
bike-share simulation.
At the bottom of the file, there is a sample_simulation function that you
can use to try running the simulation at any time.
"""
import csv
from datetime import datetime, timedelta
import json
from typing import Dict, List, Tuple
from bikeshare import Ride, Station
from container import PriorityQueue
from visualizer import Visualizer
# Datetime format to parse the ride data
DATETIME_FORMAT = '%Y-%m-%d %H:%M'
#Function constants for reporting beggining and ending of rides
RIDE_BEGUN = True
RIDE_FINISHED = False
# Stock issues report
SUCCESSFUL_REPORT = 0
EMPTY_STATION_ISSUE = 1
FULL_STATION_ISSUE = 2
class Simulation:
"""Runs the core of the simulation through time.
=== Attributes ===
all_rides:
A list of all the rides in this simulation.
Note that not all rides might be used, depending on the timeframe
when the simulation is run.
all_stations:
A dictionary containing all the stations in this simulation.
visualizer:
A helper class for visualizing the simulation.
active_rides:
All rides that are active during simulation
"""
all_stations: Dict[str, Station]
all_rides: List[Ride]
visualizer: Visualizer
active_rides: List[Ride]
def __init__(self, station_file: str, ride_file: str) -> None:
"""Initialize this simulation with the given configuration settings.
"""
self.all_stations = create_stations(station_file)
self.all_rides = create_rides(ride_file, self.all_stations)
self.visualizer = Visualizer()
self.active_rides = []
def run(self, start: datetime, end: datetime) -> None:
"""Run the simulation from <start> to <end>.
"""
step = timedelta(minutes=1) # Each iteration spans one minute of time
current_time = start
while current_time <= end:
self._update_active_rides(current_time)
# Method that draws all stations and bikes on the map
self.visualizer.render_drawables(list(self.all_stations.values())\
+ self.active_rides, current_time)
current_time += step
# Leave this code at the very bottom of this method.
# It will keep the visualization window open until you close
# it by pressing the 'X'.
while True:
if self.visualizer.handle_window_events():
return # Stop the simulation
def _update_active_rides(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- Loop through `self.all_rides` and compare each Ride's start and
end times with <time>.
If <time> is between the ride's start and end times (inclusive),
then add the ride to self.active_rides if it isn't already in
that list.
Otherwise, remove the ride from self.active_rides if it is in
that list.
- This means that if a ride started before the simulation's time
period but ends during or after the simulation's time period,
it should still be added to self.active_rides.
"""
for ride in self.all_rides:
# Check whether the current time is between start and end
if time > ride.start_time and time < ride.end_time:
# Check whether it is already in the active rides list
if not ride in self.active_rides:
# Reports that this ride has just begun
# this method will also lead to an update on station's bike
# number.
if ride.report_status(RIDE_BEGUN, time) \
== SUCCESSFUL_REPORT:
# Adds to the active rides list if no anomaly was
# detected
self.active_rides.append(ride)
else:
if ride in self.active_rides:
# Reports that this ride has finished
# this method will also lead to an update on station's bike
# number
ride.report_status(RIDE_FINISHED, time)
# If a finished or not yet beginned ride is in this
# list, removes it.
self.active_rides.remove(ride)
def calculate_statistics(self) -> Dict[str, Tuple[str, float]]:
"""Return a dictionary containing statistics for this simulation.
The returned dictionary has exactly four keys, corresponding
to the four statistics tracked for each station:
- 'max_start'
- 'max_end'
- 'max_time_low_availability'
- 'max_time_low_unoccupied'
The corresponding value of each key is a tuple of two elements,
where the first element is the name (NOT id) of the station that has
the maximum value of the quantity specified by that key,
and the second element is the value of that quantity.
For example, the value corresponding to key 'max_start' should be the
name of the station with the most number of rides started at that
station, and the number of rides that started at that station.
"""
tempDict = {
'max_start': ('', -1),
'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
}
return { | 'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
}
def _update_active_rides_fast(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- see Task 5 of the assignment handout
"""
pass
def create_stations(stations_file: str) -> Dict[str, 'Station']:
"""Return the stations described in the given JSON data file.
Each key in the returned dictionary is a station id,
and each value is the corresponding Station object.
Note that you need to call Station(...) to create these objects!
Precondition: stations_file matches the format specified in the
assignment handout.
This function should be called *before* _read_rides because the
rides CSV file refers to station ids.
"""
# Read in raw data using the json library.
with open(stations_file) as file:
raw_stations = json.load(file)
stations = {}
for s in raw_stations['stations']:
# Extract the relevant fields from the raw station JSON.
# s is a dictionary with the keys 'n', 's', 'la', 'lo', 'da', and 'ba'
# as described in the assignment handout.
# NOTE: all of the corresponding values are strings, and so you need
# to convert some of them to numbers explicitly using int() or float().
# Inputs for Station Class (with keys from dictionary) :
# -Lon/Lat (lo/la)
# -Capacity (da + ba)
# -Number of bikes (da)
# -Name (s)
station = Station((s['lo'], s['la']), s['da'] + s['ba'],\
s['da'], s['s'])
stations[s['n']] = station
return stations
def create_rides(rides_file: str,
stations: Dict[str, 'Station']) -> List['Ride']:
"""Return the rides described in the given CSV file.
Lookup the station ids contained in the rides file in <stations>
to access the corresponding Station objects.
Ignore any ride whose start or end station is not present in <stations>.
Precondition: rides_file matches the format specified in the
assignment handout.
"""
rides = []
with open(rides_file) as file:
for line in csv.reader(file):
# line is a list of strings, following the format described
# in the assignment handout.
#
# Convert between a string and a datetime object
# using the function datetime.strptime and the DATETIME_FORMAT
# constant we defined above. Example:
# >>> datetime.strptime('2017-06-01 8:00', DATETIME_FORMAT)
# datetime.datetime(2017, 6, 1, 8, 0)
# CSV Columns:
# -Start time [0]
# -Start station code [1]
# -End time [2]
# -End station code [3]
# -ignore the rest [4,5]
# Ride class inputs:
# -Start station (Station())
# -End station (Station())
# -Times (Tuple(start,end))
start_station_code = line[1]
end_station_code = line[3]
# First check whether the stations from the ride are present in the
# stations registry
if start_station_code and end_station_code in stations:
start_time = datetime.strptime(line[0], DATETIME_FORMAT)
end_time = datetime.strptime(line[2], DATETIME_FORMAT)
start_station = stations[start_station_code]
end_station = stations[end_station_code]
ride = Ride(start_station, end_station, (start_time, end_time))
rides.append(ride)
return rides
class Event:
"""An event in the bike share simulation.
Events are ordered by their timestamp.
"""
simulation: 'Simulation'
time: datetime
def __init__(self, simulation: 'Simulation', time: datetime) -> None:
"""Initialize a new event."""
self.simulation = simulation
self.time = time
def __lt__(self, other: 'Event') -> bool:
"""Return whether this event is less than <other>.
Events are ordered by their timestamp.
"""
return self.time < other.time
def process(self) -> List['Event']:
"""Process this event by updating the state of the simulation.
Return a list of new events spawned by this event.
"""
raise NotImplementedError
class RideStartEvent(Event):
"""An event corresponding to the start of a ride."""
pass
class RideEndEvent(Event):
"""An event corresponding to the start of a ride."""
pass
def sample_simulation() -> Dict[str, Tuple[str, float]]:
"""Run a sample simulation. For testing purposes only."""
sim = Simulation('stations.json', 'sample_rides.csv')
sim.run(datetime(2017, 6, 1, 8, 0, 0),
datetime(2017, 6, 1, 9, 0, 0))
return sim.calculate_statistics()
if __name__ == '__main__':
# Uncomment these lines when you want to check your work using python_ta!
# import python_ta
# python_ta.check_all(config={
# 'allowed-io': ['create_stations', 'create_rides'],
# 'allowed-import-modules': [
# 'doctest', 'python_ta', 'typing',
# 'csv', 'datetime', 'json',
# 'bikeshare', 'container', 'visualizer'
# ]
# })
print(sample_simulation()) | 'max_start': ('', -1), | random_line_split |
simulation.py | """Assignment 1 - Simulation
=== CSC148 Fall 2017 ===
Diane Horton and David Liu
Department of Computer Science,
University of Toronto
=== Module Description ===
This file contains the Simulation class, which is the main class for your
bike-share simulation.
At the bottom of the file, there is a sample_simulation function that you
can use to try running the simulation at any time.
"""
import csv
from datetime import datetime, timedelta
import json
from typing import Dict, List, Tuple
from bikeshare import Ride, Station
from container import PriorityQueue
from visualizer import Visualizer
# Datetime format to parse the ride data
DATETIME_FORMAT = '%Y-%m-%d %H:%M'
#Function constants for reporting beggining and ending of rides
RIDE_BEGUN = True
RIDE_FINISHED = False
# Stock issues report
SUCCESSFUL_REPORT = 0
EMPTY_STATION_ISSUE = 1
FULL_STATION_ISSUE = 2
class Simulation:
"""Runs the core of the simulation through time.
=== Attributes ===
all_rides:
A list of all the rides in this simulation.
Note that not all rides might be used, depending on the timeframe
when the simulation is run.
all_stations:
A dictionary containing all the stations in this simulation.
visualizer:
A helper class for visualizing the simulation.
active_rides:
All rides that are active during simulation
"""
all_stations: Dict[str, Station]
all_rides: List[Ride]
visualizer: Visualizer
active_rides: List[Ride]
def __init__(self, station_file: str, ride_file: str) -> None:
"""Initialize this simulation with the given configuration settings.
"""
self.all_stations = create_stations(station_file)
self.all_rides = create_rides(ride_file, self.all_stations)
self.visualizer = Visualizer()
self.active_rides = []
def run(self, start: datetime, end: datetime) -> None:
"""Run the simulation from <start> to <end>.
"""
step = timedelta(minutes=1) # Each iteration spans one minute of time
current_time = start
while current_time <= end:
self._update_active_rides(current_time)
# Method that draws all stations and bikes on the map
self.visualizer.render_drawables(list(self.all_stations.values())\
+ self.active_rides, current_time)
current_time += step
# Leave this code at the very bottom of this method.
# It will keep the visualization window open until you close
# it by pressing the 'X'.
while True:
|
def _update_active_rides(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- Loop through `self.all_rides` and compare each Ride's start and
end times with <time>.
If <time> is between the ride's start and end times (inclusive),
then add the ride to self.active_rides if it isn't already in
that list.
Otherwise, remove the ride from self.active_rides if it is in
that list.
- This means that if a ride started before the simulation's time
period but ends during or after the simulation's time period,
it should still be added to self.active_rides.
"""
for ride in self.all_rides:
# Check whether the current time is between start and end
if time > ride.start_time and time < ride.end_time:
# Check whether it is already in the active rides list
if not ride in self.active_rides:
# Reports that this ride has just begun
# this method will also lead to an update on station's bike
# number.
if ride.report_status(RIDE_BEGUN, time) \
== SUCCESSFUL_REPORT:
# Adds to the active rides list if no anomaly was
# detected
self.active_rides.append(ride)
else:
if ride in self.active_rides:
# Reports that this ride has finished
# this method will also lead to an update on station's bike
# number
ride.report_status(RIDE_FINISHED, time)
# If a finished or not yet beginned ride is in this
# list, removes it.
self.active_rides.remove(ride)
def calculate_statistics(self) -> Dict[str, Tuple[str, float]]:
"""Return a dictionary containing statistics for this simulation.
The returned dictionary has exactly four keys, corresponding
to the four statistics tracked for each station:
- 'max_start'
- 'max_end'
- 'max_time_low_availability'
- 'max_time_low_unoccupied'
The corresponding value of each key is a tuple of two elements,
where the first element is the name (NOT id) of the station that has
the maximum value of the quantity specified by that key,
and the second element is the value of that quantity.
For example, the value corresponding to key 'max_start' should be the
name of the station with the most number of rides started at that
station, and the number of rides that started at that station.
"""
tempDict = {
'max_start': ('', -1),
'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
}
return {
'max_start': ('', -1),
'max_end': ('', -1),
'max_time_low_availability': ('', -1),
'max_time_low_unoccupied': ('', -1)
}
def _update_active_rides_fast(self, time: datetime) -> None:
"""Update this simulation's list of active rides for the given time.
REQUIRED IMPLEMENTATION NOTES:
- see Task 5 of the assignment handout
"""
pass
def create_stations(stations_file: str) -> Dict[str, 'Station']:
"""Return the stations described in the given JSON data file.
Each key in the returned dictionary is a station id,
and each value is the corresponding Station object.
Note that you need to call Station(...) to create these objects!
Precondition: stations_file matches the format specified in the
assignment handout.
This function should be called *before* _read_rides because the
rides CSV file refers to station ids.
"""
# Read in raw data using the json library.
with open(stations_file) as file:
raw_stations = json.load(file)
stations = {}
for s in raw_stations['stations']:
# Extract the relevant fields from the raw station JSON.
# s is a dictionary with the keys 'n', 's', 'la', 'lo', 'da', and 'ba'
# as described in the assignment handout.
# NOTE: all of the corresponding values are strings, and so you need
# to convert some of them to numbers explicitly using int() or float().
# Inputs for Station Class (with keys from dictionary) :
# -Lon/Lat (lo/la)
# -Capacity (da + ba)
# -Number of bikes (da)
# -Name (s)
station = Station((s['lo'], s['la']), s['da'] + s['ba'],\
s['da'], s['s'])
stations[s['n']] = station
return stations
def create_rides(rides_file: str,
stations: Dict[str, 'Station']) -> List['Ride']:
"""Return the rides described in the given CSV file.
Lookup the station ids contained in the rides file in <stations>
to access the corresponding Station objects.
Ignore any ride whose start or end station is not present in <stations>.
Precondition: rides_file matches the format specified in the
assignment handout.
"""
rides = []
with open(rides_file) as file:
for line in csv.reader(file):
# line is a list of strings, following the format described
# in the assignment handout.
#
# Convert between a string and a datetime object
# using the function datetime.strptime and the DATETIME_FORMAT
# constant we defined above. Example:
# >>> datetime.strptime('2017-06-01 8:00', DATETIME_FORMAT)
# datetime.datetime(2017, 6, 1, 8, 0)
# CSV Columns:
# -Start time [0]
# -Start station code [1]
# -End time [2]
# -End station code [3]
# -ignore the rest [4,5]
# Ride class inputs:
# -Start station (Station())
# -End station (Station())
# -Times (Tuple(start,end))
start_station_code = line[1]
end_station_code = line[3]
# First check whether the stations from the ride are present in the
# stations registry
if start_station_code and end_station_code in stations:
start_time = datetime.strptime(line[0], DATETIME_FORMAT)
end_time = datetime.strptime(line[2], DATETIME_FORMAT)
start_station = stations[start_station_code]
end_station = stations[end_station_code]
ride = Ride(start_station, end_station, (start_time, end_time))
rides.append(ride)
return rides
class Event:
"""An event in the bike share simulation.
Events are ordered by their timestamp.
"""
simulation: 'Simulation'
time: datetime
def __init__(self, simulation: 'Simulation', time: datetime) -> None:
"""Initialize a new event."""
self.simulation = simulation
self.time = time
def __lt__(self, other: 'Event') -> bool:
"""Return whether this event is less than <other>.
Events are ordered by their timestamp.
"""
return self.time < other.time
def process(self) -> List['Event']:
"""Process this event by updating the state of the simulation.
Return a list of new events spawned by this event.
"""
raise NotImplementedError
class RideStartEvent(Event):
"""An event corresponding to the start of a ride."""
pass
class RideEndEvent(Event):
"""An event corresponding to the start of a ride."""
pass
def sample_simulation() -> Dict[str, Tuple[str, float]]:
"""Run a sample simulation. For testing purposes only."""
sim = Simulation('stations.json', 'sample_rides.csv')
sim.run(datetime(2017, 6, 1, 8, 0, 0),
datetime(2017, 6, 1, 9, 0, 0))
return sim.calculate_statistics()
if __name__ == '__main__':
# Uncomment these lines when you want to check your work using python_ta!
# import python_ta
# python_ta.check_all(config={
# 'allowed-io': ['create_stations', 'create_rides'],
# 'allowed-import-modules': [
# 'doctest', 'python_ta', 'typing',
# 'csv', 'datetime', 'json',
# 'bikeshare', 'container', 'visualizer'
# ]
# })
print(sample_simulation())
| if self.visualizer.handle_window_events():
return # Stop the simulation | conditional_block |
conn.rs | // Copyright 2016 The Grin Developers
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//! Provides a connection wrapper that handles the lower level tasks in sending
//! or receiving data from the TCP socket, as well as dealing with timeouts.
use std::iter;
use std::ops::Deref;
use std::sync::{Arc, Mutex};
use std::time::{Duration, Instant};
use futures;
use futures::{Future, Stream};
use futures::stream;
use futures::sync::mpsc::{Sender, UnboundedReceiver, UnboundedSender};
use tokio_core::net::TcpStream;
use tokio_io::{AsyncRead, AsyncWrite};
use tokio_io::io::{read_exact, write_all};
use tokio_timer::{Timer, TimerError};
use core::core::hash::Hash;
use core::ser;
use msg::*;
use types::Error;
use rate_limit::*;
use util::LOGGER;
/// Handler to provide to the connection, will be called back anytime a message
/// is received. The provided sender can be use to immediately send back
/// another message.
pub trait Handler: Sync + Send {
/// Handle function to implement to process incoming messages. A sender to
/// reply immediately as well as the message header and its unparsed body
/// are provided.
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error>;
}
impl<F> Handler for F
where
F: Fn(UnboundedSender<Vec<u8>>, MsgHeader, Vec<u8>)
-> Result<Option<Hash>, ser::Error>,
F: Sync + Send,
{
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error> {
self(sender, header, body)
}
}
/// A higher level connection wrapping the TcpStream. Maintains the amount of
/// data transmitted and deals with the low-level task of sending and
/// receiving data, parsing message headers and timeouts.
#[allow(dead_code)]
pub struct Connection {
// Channel to push bytes to the remote peer
outbound_chan: UnboundedSender<Vec<u8>>,
// Close the connection with the remote peer
close_chan: Sender<()>,
// Bytes we've sent.
sent_bytes: Arc<Mutex<u64>>,
// Bytes we've received.
received_bytes: Arc<Mutex<u64>>,
// Counter for read errors.
error_count: Mutex<u64>,
}
impl Connection {
/// Start listening on the provided connection and wraps it. Does not hang
/// the current thread, instead just returns a future and the Connection
/// itself.
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (Connection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let (reader, writer) = conn.split();
// Set Max Read to 12 Mb/s
let reader = ThrottledReader::new(reader, 12_000_000);
// Set Max Write to 12 Mb/s
let writer = ThrottledWriter::new(writer, 12_000_000);
// prepare the channel that will transmit data to the connection writer
let (tx, rx) = futures::sync::mpsc::unbounded();
// same for closing the connection
let (close_tx, close_rx) = futures::sync::mpsc::channel(1);
let close_conn = close_rx
.for_each(|_| Ok(()))
.map_err(|_| Error::ConnectionClose);
let me = Connection {
outbound_chan: tx.clone(),
close_chan: close_tx,
sent_bytes: Arc::new(Mutex::new(0)),
received_bytes: Arc::new(Mutex::new(0)),
error_count: Mutex::new(0),
};
// setup the reading future, getting messages from the peer and processing them
let read_msg = me.read_msg(tx, reader, handler).map(|_| ());
// setting the writing future, getting messages from our system and sending
// them out
let write_msg = me.write_msg(rx, writer).map(|_| ());
// select between our different futures and return them
let fut = Box::new(
close_conn
.select(read_msg.select(write_msg).map(|_| ()).map_err(|(e, _)| e))
.map(|_| ())
.map_err(|(e, _)| e),
);
(me, fut)
}
/// Prepares the future that gets message data produced by our system and
/// sends it to the peer connection
fn write_msg<W>(
&self,
rx: UnboundedReceiver<Vec<u8>>,
writer: W,
) -> Box<Future<Item = W, Error = Error>>
where
W: AsyncWrite + 'static,
{
let sent_bytes = self.sent_bytes.clone();
let send_data = rx
.map_err(|_| Error::ConnectionClose)
.map(move |data| {
// add the count of bytes sent
let mut sent_bytes = sent_bytes.lock().unwrap();
*sent_bytes += data.len() as u64;
data
})
// write the data and make sure the future returns the right types
.fold(writer, |writer, data| {
write_all(writer, data).map_err(|e| Error::Connection(e)).map(|(writer, _)| writer)
});
Box::new(send_data)
}
/// Prepares the future reading from the peer connection, parsing each
/// message and forwarding them appropriately based on their type
fn read_msg<F, R>(
&self,
sender: UnboundedSender<Vec<u8>>,
reader: R,
handler: F,
) -> Box<Future<Item = R, Error = Error>>
where
F: Handler + 'static,
R: AsyncRead + 'static,
{
// infinite iterator stream so we repeat the message reading logic until the
// peer is stopped
let iter = stream::iter_ok(iter::repeat(()).map(Ok::<(), Error>));
// setup the reading future, getting messages from the peer and processing them
let recv_bytes = self.received_bytes.clone();
let handler = Arc::new(handler);
let read_msg = iter.fold(reader, move |reader, _| {
let recv_bytes = recv_bytes.clone();
let handler = handler.clone();
let sender_inner = sender.clone();
// first read the message header
read_exact(reader, vec![0u8; HEADER_LEN as usize])
.from_err()
.and_then(move |(reader, buf)| {
let header = try!(ser::deserialize::<MsgHeader>(&mut &buf[..]));
Ok((reader, header))
})
.and_then(move |(reader, header)| {
// now that we have a size, proceed with the body
read_exact(reader, vec![0u8; header.msg_len as usize])
.map(|(reader, buf)| (reader, header, buf))
.from_err()
})
.and_then(move |(reader, header, buf)| {
// add the count of bytes received
let mut recv_bytes = recv_bytes.lock().unwrap();
*recv_bytes += header.serialized_len() + header.msg_len;
// and handle the different message types
let msg_type = header.msg_type;
if let Err(e) = handler.handle(sender_inner.clone(), header, buf) {
debug!(LOGGER, "Invalid {:?} message: {}", msg_type, e);
return Err(Error::Serialization(e));
}
Ok(reader)
})
});
Box::new(read_msg)
}
/// Utility function to send any Writeable. Handles adding the header and
/// serialization.
pub fn | <W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
let mut body_data = vec![];
try!(ser::serialize(&mut body_data, body));
let mut data = vec![];
try!(ser::serialize(
&mut data,
&MsgHeader::new(t, body_data.len() as u64),
));
data.append(&mut body_data);
self.outbound_chan
.unbounded_send(data)
.map_err(|_| Error::ConnectionClose)
}
/// Bytes sent and received by this peer to the remote peer.
pub fn transmitted_bytes(&self) -> (u64, u64) {
let sent = *self.sent_bytes.lock().unwrap();
let recv = *self.received_bytes.lock().unwrap();
(sent, recv)
}
}
/// Connection wrapper that handles a request/response oriented interaction with
/// a timeout.
pub struct TimeoutConnection {
underlying: Connection,
expected_responses: Arc<Mutex<Vec<(Type, Option<Hash>, Instant)>>>,
}
impl TimeoutConnection {
/// Same as Connection
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (TimeoutConnection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let expects = Arc::new(Mutex::new(vec![]));
// Decorates the handler to remove the "subscription" from the expected
// responses. We got our replies, so no timeout should occur.
let exp = expects.clone();
let (conn, fut) = Connection::listen(conn, move |sender, header: MsgHeader, data| {
let msg_type = header.msg_type;
let recv_h = try!(handler.handle(sender, header, data));
let mut expects = exp.lock().unwrap();
let filtered = expects
.iter()
.filter(|&&(typ, h, _): &&(Type, Option<Hash>, Instant)| {
msg_type != typ || h.is_some() && recv_h != h
})
.map(|&x| x)
.collect::<Vec<_>>();
*expects = filtered;
Ok(recv_h)
});
// Registers a timer with the event loop to regularly check for timeouts.
let exp = expects.clone();
let timer = Timer::default()
.interval(Duration::new(2, 0))
.fold((), move |_, _| {
let exp = exp.lock().unwrap();
for &(ty, h, t) in exp.deref() {
if Instant::now() - t > Duration::new(5, 0) {
trace!(LOGGER, "Too long: {:?} {:?}", ty, h);
return Err(TimerError::TooLong);
}
}
Ok(())
})
.from_err();
let me = TimeoutConnection {
underlying: conn,
expected_responses: expects,
};
(
me,
Box::new(fut.select(timer).map(|_| ()).map_err(|(e1, _)| e1)),
)
}
/// Sends a request and registers a timer on the provided message type and
/// optionally the hash of the sent data.
pub fn send_request<W: ser::Writeable>(
&self,
t: Type,
rt: Type,
body: &W,
expect_h: Option<(Hash)>,
) -> Result<(), Error> {
let _sent = try!(self.underlying.send_msg(t, body));
let mut expects = self.expected_responses.lock().unwrap();
expects.push((rt, expect_h, Instant::now()));
Ok(())
}
/// Same as Connection
pub fn send_msg<W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
self.underlying.send_msg(t, body)
}
/// Same as Connection
pub fn transmitted_bytes(&self) -> (u64, u64) {
self.underlying.transmitted_bytes()
}
}
| send_msg | identifier_name |
conn.rs | // Copyright 2016 The Grin Developers
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//! Provides a connection wrapper that handles the lower level tasks in sending
//! or receiving data from the TCP socket, as well as dealing with timeouts.
use std::iter;
use std::ops::Deref;
use std::sync::{Arc, Mutex};
use std::time::{Duration, Instant};
use futures;
use futures::{Future, Stream};
use futures::stream;
use futures::sync::mpsc::{Sender, UnboundedReceiver, UnboundedSender};
use tokio_core::net::TcpStream;
use tokio_io::{AsyncRead, AsyncWrite};
use tokio_io::io::{read_exact, write_all};
use tokio_timer::{Timer, TimerError};
use core::core::hash::Hash;
use core::ser;
use msg::*;
use types::Error;
use rate_limit::*;
use util::LOGGER;
/// Handler to provide to the connection, will be called back anytime a message
/// is received. The provided sender can be use to immediately send back
/// another message.
pub trait Handler: Sync + Send {
/// Handle function to implement to process incoming messages. A sender to
/// reply immediately as well as the message header and its unparsed body
/// are provided.
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error>;
}
impl<F> Handler for F
where
F: Fn(UnboundedSender<Vec<u8>>, MsgHeader, Vec<u8>)
-> Result<Option<Hash>, ser::Error>,
F: Sync + Send,
{
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error> {
self(sender, header, body)
}
}
/// A higher level connection wrapping the TcpStream. Maintains the amount of
/// data transmitted and deals with the low-level task of sending and
/// receiving data, parsing message headers and timeouts.
#[allow(dead_code)]
pub struct Connection {
// Channel to push bytes to the remote peer
outbound_chan: UnboundedSender<Vec<u8>>,
// Close the connection with the remote peer
close_chan: Sender<()>,
// Bytes we've sent.
sent_bytes: Arc<Mutex<u64>>,
// Bytes we've received.
received_bytes: Arc<Mutex<u64>>,
// Counter for read errors.
error_count: Mutex<u64>,
}
impl Connection {
/// Start listening on the provided connection and wraps it. Does not hang
/// the current thread, instead just returns a future and the Connection
/// itself.
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (Connection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let (reader, writer) = conn.split();
// Set Max Read to 12 Mb/s
let reader = ThrottledReader::new(reader, 12_000_000);
// Set Max Write to 12 Mb/s
let writer = ThrottledWriter::new(writer, 12_000_000);
// prepare the channel that will transmit data to the connection writer
let (tx, rx) = futures::sync::mpsc::unbounded();
// same for closing the connection
let (close_tx, close_rx) = futures::sync::mpsc::channel(1);
let close_conn = close_rx
.for_each(|_| Ok(()))
.map_err(|_| Error::ConnectionClose);
let me = Connection {
outbound_chan: tx.clone(),
close_chan: close_tx,
sent_bytes: Arc::new(Mutex::new(0)),
received_bytes: Arc::new(Mutex::new(0)),
error_count: Mutex::new(0),
};
// setup the reading future, getting messages from the peer and processing them
let read_msg = me.read_msg(tx, reader, handler).map(|_| ());
// setting the writing future, getting messages from our system and sending
// them out
let write_msg = me.write_msg(rx, writer).map(|_| ());
// select between our different futures and return them
let fut = Box::new(
close_conn
.select(read_msg.select(write_msg).map(|_| ()).map_err(|(e, _)| e))
.map(|_| ())
.map_err(|(e, _)| e),
);
(me, fut)
}
/// Prepares the future that gets message data produced by our system and
/// sends it to the peer connection
fn write_msg<W>(
&self,
rx: UnboundedReceiver<Vec<u8>>,
writer: W,
) -> Box<Future<Item = W, Error = Error>>
where
W: AsyncWrite + 'static,
{
let sent_bytes = self.sent_bytes.clone();
let send_data = rx
.map_err(|_| Error::ConnectionClose)
.map(move |data| {
// add the count of bytes sent
let mut sent_bytes = sent_bytes.lock().unwrap();
*sent_bytes += data.len() as u64;
data
})
// write the data and make sure the future returns the right types
.fold(writer, |writer, data| {
write_all(writer, data).map_err(|e| Error::Connection(e)).map(|(writer, _)| writer)
});
Box::new(send_data)
}
/// Prepares the future reading from the peer connection, parsing each
/// message and forwarding them appropriately based on their type
fn read_msg<F, R>(
&self,
sender: UnboundedSender<Vec<u8>>,
reader: R,
handler: F,
) -> Box<Future<Item = R, Error = Error>>
where
F: Handler + 'static,
R: AsyncRead + 'static,
{
// infinite iterator stream so we repeat the message reading logic until the
// peer is stopped
let iter = stream::iter_ok(iter::repeat(()).map(Ok::<(), Error>));
// setup the reading future, getting messages from the peer and processing them
let recv_bytes = self.received_bytes.clone();
let handler = Arc::new(handler);
let read_msg = iter.fold(reader, move |reader, _| {
let recv_bytes = recv_bytes.clone();
let handler = handler.clone();
let sender_inner = sender.clone();
// first read the message header
read_exact(reader, vec![0u8; HEADER_LEN as usize])
.from_err()
.and_then(move |(reader, buf)| {
let header = try!(ser::deserialize::<MsgHeader>(&mut &buf[..]));
Ok((reader, header))
})
.and_then(move |(reader, header)| {
// now that we have a size, proceed with the body
read_exact(reader, vec![0u8; header.msg_len as usize])
.map(|(reader, buf)| (reader, header, buf))
.from_err()
})
.and_then(move |(reader, header, buf)| {
// add the count of bytes received
let mut recv_bytes = recv_bytes.lock().unwrap();
*recv_bytes += header.serialized_len() + header.msg_len;
// and handle the different message types
let msg_type = header.msg_type;
if let Err(e) = handler.handle(sender_inner.clone(), header, buf) {
debug!(LOGGER, "Invalid {:?} message: {}", msg_type, e);
return Err(Error::Serialization(e));
}
Ok(reader)
})
});
Box::new(read_msg)
}
/// Utility function to send any Writeable. Handles adding the header and
/// serialization.
pub fn send_msg<W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
let mut body_data = vec![];
try!(ser::serialize(&mut body_data, body));
let mut data = vec![];
try!(ser::serialize(
&mut data,
&MsgHeader::new(t, body_data.len() as u64),
));
data.append(&mut body_data);
self.outbound_chan
.unbounded_send(data)
.map_err(|_| Error::ConnectionClose)
}
/// Bytes sent and received by this peer to the remote peer.
pub fn transmitted_bytes(&self) -> (u64, u64) {
let sent = *self.sent_bytes.lock().unwrap();
let recv = *self.received_bytes.lock().unwrap();
(sent, recv)
}
}
/// Connection wrapper that handles a request/response oriented interaction with
/// a timeout.
pub struct TimeoutConnection {
underlying: Connection,
expected_responses: Arc<Mutex<Vec<(Type, Option<Hash>, Instant)>>>,
}
impl TimeoutConnection {
/// Same as Connection
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (TimeoutConnection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let expects = Arc::new(Mutex::new(vec![]));
| let (conn, fut) = Connection::listen(conn, move |sender, header: MsgHeader, data| {
let msg_type = header.msg_type;
let recv_h = try!(handler.handle(sender, header, data));
let mut expects = exp.lock().unwrap();
let filtered = expects
.iter()
.filter(|&&(typ, h, _): &&(Type, Option<Hash>, Instant)| {
msg_type != typ || h.is_some() && recv_h != h
})
.map(|&x| x)
.collect::<Vec<_>>();
*expects = filtered;
Ok(recv_h)
});
// Registers a timer with the event loop to regularly check for timeouts.
let exp = expects.clone();
let timer = Timer::default()
.interval(Duration::new(2, 0))
.fold((), move |_, _| {
let exp = exp.lock().unwrap();
for &(ty, h, t) in exp.deref() {
if Instant::now() - t > Duration::new(5, 0) {
trace!(LOGGER, "Too long: {:?} {:?}", ty, h);
return Err(TimerError::TooLong);
}
}
Ok(())
})
.from_err();
let me = TimeoutConnection {
underlying: conn,
expected_responses: expects,
};
(
me,
Box::new(fut.select(timer).map(|_| ()).map_err(|(e1, _)| e1)),
)
}
/// Sends a request and registers a timer on the provided message type and
/// optionally the hash of the sent data.
pub fn send_request<W: ser::Writeable>(
&self,
t: Type,
rt: Type,
body: &W,
expect_h: Option<(Hash)>,
) -> Result<(), Error> {
let _sent = try!(self.underlying.send_msg(t, body));
let mut expects = self.expected_responses.lock().unwrap();
expects.push((rt, expect_h, Instant::now()));
Ok(())
}
/// Same as Connection
pub fn send_msg<W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
self.underlying.send_msg(t, body)
}
/// Same as Connection
pub fn transmitted_bytes(&self) -> (u64, u64) {
self.underlying.transmitted_bytes()
}
} | // Decorates the handler to remove the "subscription" from the expected
// responses. We got our replies, so no timeout should occur.
let exp = expects.clone(); | random_line_split |
conn.rs | // Copyright 2016 The Grin Developers
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//! Provides a connection wrapper that handles the lower level tasks in sending
//! or receiving data from the TCP socket, as well as dealing with timeouts.
use std::iter;
use std::ops::Deref;
use std::sync::{Arc, Mutex};
use std::time::{Duration, Instant};
use futures;
use futures::{Future, Stream};
use futures::stream;
use futures::sync::mpsc::{Sender, UnboundedReceiver, UnboundedSender};
use tokio_core::net::TcpStream;
use tokio_io::{AsyncRead, AsyncWrite};
use tokio_io::io::{read_exact, write_all};
use tokio_timer::{Timer, TimerError};
use core::core::hash::Hash;
use core::ser;
use msg::*;
use types::Error;
use rate_limit::*;
use util::LOGGER;
/// Handler to provide to the connection, will be called back anytime a message
/// is received. The provided sender can be use to immediately send back
/// another message.
pub trait Handler: Sync + Send {
/// Handle function to implement to process incoming messages. A sender to
/// reply immediately as well as the message header and its unparsed body
/// are provided.
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error>;
}
impl<F> Handler for F
where
F: Fn(UnboundedSender<Vec<u8>>, MsgHeader, Vec<u8>)
-> Result<Option<Hash>, ser::Error>,
F: Sync + Send,
{
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error> |
}
/// A higher level connection wrapping the TcpStream. Maintains the amount of
/// data transmitted and deals with the low-level task of sending and
/// receiving data, parsing message headers and timeouts.
#[allow(dead_code)]
pub struct Connection {
// Channel to push bytes to the remote peer
outbound_chan: UnboundedSender<Vec<u8>>,
// Close the connection with the remote peer
close_chan: Sender<()>,
// Bytes we've sent.
sent_bytes: Arc<Mutex<u64>>,
// Bytes we've received.
received_bytes: Arc<Mutex<u64>>,
// Counter for read errors.
error_count: Mutex<u64>,
}
impl Connection {
/// Start listening on the provided connection and wraps it. Does not hang
/// the current thread, instead just returns a future and the Connection
/// itself.
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (Connection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let (reader, writer) = conn.split();
// Set Max Read to 12 Mb/s
let reader = ThrottledReader::new(reader, 12_000_000);
// Set Max Write to 12 Mb/s
let writer = ThrottledWriter::new(writer, 12_000_000);
// prepare the channel that will transmit data to the connection writer
let (tx, rx) = futures::sync::mpsc::unbounded();
// same for closing the connection
let (close_tx, close_rx) = futures::sync::mpsc::channel(1);
let close_conn = close_rx
.for_each(|_| Ok(()))
.map_err(|_| Error::ConnectionClose);
let me = Connection {
outbound_chan: tx.clone(),
close_chan: close_tx,
sent_bytes: Arc::new(Mutex::new(0)),
received_bytes: Arc::new(Mutex::new(0)),
error_count: Mutex::new(0),
};
// setup the reading future, getting messages from the peer and processing them
let read_msg = me.read_msg(tx, reader, handler).map(|_| ());
// setting the writing future, getting messages from our system and sending
// them out
let write_msg = me.write_msg(rx, writer).map(|_| ());
// select between our different futures and return them
let fut = Box::new(
close_conn
.select(read_msg.select(write_msg).map(|_| ()).map_err(|(e, _)| e))
.map(|_| ())
.map_err(|(e, _)| e),
);
(me, fut)
}
/// Prepares the future that gets message data produced by our system and
/// sends it to the peer connection
fn write_msg<W>(
&self,
rx: UnboundedReceiver<Vec<u8>>,
writer: W,
) -> Box<Future<Item = W, Error = Error>>
where
W: AsyncWrite + 'static,
{
let sent_bytes = self.sent_bytes.clone();
let send_data = rx
.map_err(|_| Error::ConnectionClose)
.map(move |data| {
// add the count of bytes sent
let mut sent_bytes = sent_bytes.lock().unwrap();
*sent_bytes += data.len() as u64;
data
})
// write the data and make sure the future returns the right types
.fold(writer, |writer, data| {
write_all(writer, data).map_err(|e| Error::Connection(e)).map(|(writer, _)| writer)
});
Box::new(send_data)
}
/// Prepares the future reading from the peer connection, parsing each
/// message and forwarding them appropriately based on their type
fn read_msg<F, R>(
&self,
sender: UnboundedSender<Vec<u8>>,
reader: R,
handler: F,
) -> Box<Future<Item = R, Error = Error>>
where
F: Handler + 'static,
R: AsyncRead + 'static,
{
// infinite iterator stream so we repeat the message reading logic until the
// peer is stopped
let iter = stream::iter_ok(iter::repeat(()).map(Ok::<(), Error>));
// setup the reading future, getting messages from the peer and processing them
let recv_bytes = self.received_bytes.clone();
let handler = Arc::new(handler);
let read_msg = iter.fold(reader, move |reader, _| {
let recv_bytes = recv_bytes.clone();
let handler = handler.clone();
let sender_inner = sender.clone();
// first read the message header
read_exact(reader, vec![0u8; HEADER_LEN as usize])
.from_err()
.and_then(move |(reader, buf)| {
let header = try!(ser::deserialize::<MsgHeader>(&mut &buf[..]));
Ok((reader, header))
})
.and_then(move |(reader, header)| {
// now that we have a size, proceed with the body
read_exact(reader, vec![0u8; header.msg_len as usize])
.map(|(reader, buf)| (reader, header, buf))
.from_err()
})
.and_then(move |(reader, header, buf)| {
// add the count of bytes received
let mut recv_bytes = recv_bytes.lock().unwrap();
*recv_bytes += header.serialized_len() + header.msg_len;
// and handle the different message types
let msg_type = header.msg_type;
if let Err(e) = handler.handle(sender_inner.clone(), header, buf) {
debug!(LOGGER, "Invalid {:?} message: {}", msg_type, e);
return Err(Error::Serialization(e));
}
Ok(reader)
})
});
Box::new(read_msg)
}
/// Utility function to send any Writeable. Handles adding the header and
/// serialization.
pub fn send_msg<W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
let mut body_data = vec![];
try!(ser::serialize(&mut body_data, body));
let mut data = vec![];
try!(ser::serialize(
&mut data,
&MsgHeader::new(t, body_data.len() as u64),
));
data.append(&mut body_data);
self.outbound_chan
.unbounded_send(data)
.map_err(|_| Error::ConnectionClose)
}
/// Bytes sent and received by this peer to the remote peer.
pub fn transmitted_bytes(&self) -> (u64, u64) {
let sent = *self.sent_bytes.lock().unwrap();
let recv = *self.received_bytes.lock().unwrap();
(sent, recv)
}
}
/// Connection wrapper that handles a request/response oriented interaction with
/// a timeout.
pub struct TimeoutConnection {
underlying: Connection,
expected_responses: Arc<Mutex<Vec<(Type, Option<Hash>, Instant)>>>,
}
impl TimeoutConnection {
/// Same as Connection
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (TimeoutConnection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let expects = Arc::new(Mutex::new(vec![]));
// Decorates the handler to remove the "subscription" from the expected
// responses. We got our replies, so no timeout should occur.
let exp = expects.clone();
let (conn, fut) = Connection::listen(conn, move |sender, header: MsgHeader, data| {
let msg_type = header.msg_type;
let recv_h = try!(handler.handle(sender, header, data));
let mut expects = exp.lock().unwrap();
let filtered = expects
.iter()
.filter(|&&(typ, h, _): &&(Type, Option<Hash>, Instant)| {
msg_type != typ || h.is_some() && recv_h != h
})
.map(|&x| x)
.collect::<Vec<_>>();
*expects = filtered;
Ok(recv_h)
});
// Registers a timer with the event loop to regularly check for timeouts.
let exp = expects.clone();
let timer = Timer::default()
.interval(Duration::new(2, 0))
.fold((), move |_, _| {
let exp = exp.lock().unwrap();
for &(ty, h, t) in exp.deref() {
if Instant::now() - t > Duration::new(5, 0) {
trace!(LOGGER, "Too long: {:?} {:?}", ty, h);
return Err(TimerError::TooLong);
}
}
Ok(())
})
.from_err();
let me = TimeoutConnection {
underlying: conn,
expected_responses: expects,
};
(
me,
Box::new(fut.select(timer).map(|_| ()).map_err(|(e1, _)| e1)),
)
}
/// Sends a request and registers a timer on the provided message type and
/// optionally the hash of the sent data.
pub fn send_request<W: ser::Writeable>(
&self,
t: Type,
rt: Type,
body: &W,
expect_h: Option<(Hash)>,
) -> Result<(), Error> {
let _sent = try!(self.underlying.send_msg(t, body));
let mut expects = self.expected_responses.lock().unwrap();
expects.push((rt, expect_h, Instant::now()));
Ok(())
}
/// Same as Connection
pub fn send_msg<W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
self.underlying.send_msg(t, body)
}
/// Same as Connection
pub fn transmitted_bytes(&self) -> (u64, u64) {
self.underlying.transmitted_bytes()
}
}
| {
self(sender, header, body)
} | identifier_body |
conn.rs | // Copyright 2016 The Grin Developers
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//! Provides a connection wrapper that handles the lower level tasks in sending
//! or receiving data from the TCP socket, as well as dealing with timeouts.
use std::iter;
use std::ops::Deref;
use std::sync::{Arc, Mutex};
use std::time::{Duration, Instant};
use futures;
use futures::{Future, Stream};
use futures::stream;
use futures::sync::mpsc::{Sender, UnboundedReceiver, UnboundedSender};
use tokio_core::net::TcpStream;
use tokio_io::{AsyncRead, AsyncWrite};
use tokio_io::io::{read_exact, write_all};
use tokio_timer::{Timer, TimerError};
use core::core::hash::Hash;
use core::ser;
use msg::*;
use types::Error;
use rate_limit::*;
use util::LOGGER;
/// Handler to provide to the connection, will be called back anytime a message
/// is received. The provided sender can be use to immediately send back
/// another message.
pub trait Handler: Sync + Send {
/// Handle function to implement to process incoming messages. A sender to
/// reply immediately as well as the message header and its unparsed body
/// are provided.
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error>;
}
impl<F> Handler for F
where
F: Fn(UnboundedSender<Vec<u8>>, MsgHeader, Vec<u8>)
-> Result<Option<Hash>, ser::Error>,
F: Sync + Send,
{
fn handle(
&self,
sender: UnboundedSender<Vec<u8>>,
header: MsgHeader,
body: Vec<u8>,
) -> Result<Option<Hash>, ser::Error> {
self(sender, header, body)
}
}
/// A higher level connection wrapping the TcpStream. Maintains the amount of
/// data transmitted and deals with the low-level task of sending and
/// receiving data, parsing message headers and timeouts.
#[allow(dead_code)]
pub struct Connection {
// Channel to push bytes to the remote peer
outbound_chan: UnboundedSender<Vec<u8>>,
// Close the connection with the remote peer
close_chan: Sender<()>,
// Bytes we've sent.
sent_bytes: Arc<Mutex<u64>>,
// Bytes we've received.
received_bytes: Arc<Mutex<u64>>,
// Counter for read errors.
error_count: Mutex<u64>,
}
impl Connection {
/// Start listening on the provided connection and wraps it. Does not hang
/// the current thread, instead just returns a future and the Connection
/// itself.
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (Connection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let (reader, writer) = conn.split();
// Set Max Read to 12 Mb/s
let reader = ThrottledReader::new(reader, 12_000_000);
// Set Max Write to 12 Mb/s
let writer = ThrottledWriter::new(writer, 12_000_000);
// prepare the channel that will transmit data to the connection writer
let (tx, rx) = futures::sync::mpsc::unbounded();
// same for closing the connection
let (close_tx, close_rx) = futures::sync::mpsc::channel(1);
let close_conn = close_rx
.for_each(|_| Ok(()))
.map_err(|_| Error::ConnectionClose);
let me = Connection {
outbound_chan: tx.clone(),
close_chan: close_tx,
sent_bytes: Arc::new(Mutex::new(0)),
received_bytes: Arc::new(Mutex::new(0)),
error_count: Mutex::new(0),
};
// setup the reading future, getting messages from the peer and processing them
let read_msg = me.read_msg(tx, reader, handler).map(|_| ());
// setting the writing future, getting messages from our system and sending
// them out
let write_msg = me.write_msg(rx, writer).map(|_| ());
// select between our different futures and return them
let fut = Box::new(
close_conn
.select(read_msg.select(write_msg).map(|_| ()).map_err(|(e, _)| e))
.map(|_| ())
.map_err(|(e, _)| e),
);
(me, fut)
}
/// Prepares the future that gets message data produced by our system and
/// sends it to the peer connection
fn write_msg<W>(
&self,
rx: UnboundedReceiver<Vec<u8>>,
writer: W,
) -> Box<Future<Item = W, Error = Error>>
where
W: AsyncWrite + 'static,
{
let sent_bytes = self.sent_bytes.clone();
let send_data = rx
.map_err(|_| Error::ConnectionClose)
.map(move |data| {
// add the count of bytes sent
let mut sent_bytes = sent_bytes.lock().unwrap();
*sent_bytes += data.len() as u64;
data
})
// write the data and make sure the future returns the right types
.fold(writer, |writer, data| {
write_all(writer, data).map_err(|e| Error::Connection(e)).map(|(writer, _)| writer)
});
Box::new(send_data)
}
/// Prepares the future reading from the peer connection, parsing each
/// message and forwarding them appropriately based on their type
fn read_msg<F, R>(
&self,
sender: UnboundedSender<Vec<u8>>,
reader: R,
handler: F,
) -> Box<Future<Item = R, Error = Error>>
where
F: Handler + 'static,
R: AsyncRead + 'static,
{
// infinite iterator stream so we repeat the message reading logic until the
// peer is stopped
let iter = stream::iter_ok(iter::repeat(()).map(Ok::<(), Error>));
// setup the reading future, getting messages from the peer and processing them
let recv_bytes = self.received_bytes.clone();
let handler = Arc::new(handler);
let read_msg = iter.fold(reader, move |reader, _| {
let recv_bytes = recv_bytes.clone();
let handler = handler.clone();
let sender_inner = sender.clone();
// first read the message header
read_exact(reader, vec![0u8; HEADER_LEN as usize])
.from_err()
.and_then(move |(reader, buf)| {
let header = try!(ser::deserialize::<MsgHeader>(&mut &buf[..]));
Ok((reader, header))
})
.and_then(move |(reader, header)| {
// now that we have a size, proceed with the body
read_exact(reader, vec![0u8; header.msg_len as usize])
.map(|(reader, buf)| (reader, header, buf))
.from_err()
})
.and_then(move |(reader, header, buf)| {
// add the count of bytes received
let mut recv_bytes = recv_bytes.lock().unwrap();
*recv_bytes += header.serialized_len() + header.msg_len;
// and handle the different message types
let msg_type = header.msg_type;
if let Err(e) = handler.handle(sender_inner.clone(), header, buf) |
Ok(reader)
})
});
Box::new(read_msg)
}
/// Utility function to send any Writeable. Handles adding the header and
/// serialization.
pub fn send_msg<W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
let mut body_data = vec![];
try!(ser::serialize(&mut body_data, body));
let mut data = vec![];
try!(ser::serialize(
&mut data,
&MsgHeader::new(t, body_data.len() as u64),
));
data.append(&mut body_data);
self.outbound_chan
.unbounded_send(data)
.map_err(|_| Error::ConnectionClose)
}
/// Bytes sent and received by this peer to the remote peer.
pub fn transmitted_bytes(&self) -> (u64, u64) {
let sent = *self.sent_bytes.lock().unwrap();
let recv = *self.received_bytes.lock().unwrap();
(sent, recv)
}
}
/// Connection wrapper that handles a request/response oriented interaction with
/// a timeout.
pub struct TimeoutConnection {
underlying: Connection,
expected_responses: Arc<Mutex<Vec<(Type, Option<Hash>, Instant)>>>,
}
impl TimeoutConnection {
/// Same as Connection
pub fn listen<F>(
conn: TcpStream,
handler: F,
) -> (TimeoutConnection, Box<Future<Item = (), Error = Error>>)
where
F: Handler + 'static,
{
let expects = Arc::new(Mutex::new(vec![]));
// Decorates the handler to remove the "subscription" from the expected
// responses. We got our replies, so no timeout should occur.
let exp = expects.clone();
let (conn, fut) = Connection::listen(conn, move |sender, header: MsgHeader, data| {
let msg_type = header.msg_type;
let recv_h = try!(handler.handle(sender, header, data));
let mut expects = exp.lock().unwrap();
let filtered = expects
.iter()
.filter(|&&(typ, h, _): &&(Type, Option<Hash>, Instant)| {
msg_type != typ || h.is_some() && recv_h != h
})
.map(|&x| x)
.collect::<Vec<_>>();
*expects = filtered;
Ok(recv_h)
});
// Registers a timer with the event loop to regularly check for timeouts.
let exp = expects.clone();
let timer = Timer::default()
.interval(Duration::new(2, 0))
.fold((), move |_, _| {
let exp = exp.lock().unwrap();
for &(ty, h, t) in exp.deref() {
if Instant::now() - t > Duration::new(5, 0) {
trace!(LOGGER, "Too long: {:?} {:?}", ty, h);
return Err(TimerError::TooLong);
}
}
Ok(())
})
.from_err();
let me = TimeoutConnection {
underlying: conn,
expected_responses: expects,
};
(
me,
Box::new(fut.select(timer).map(|_| ()).map_err(|(e1, _)| e1)),
)
}
/// Sends a request and registers a timer on the provided message type and
/// optionally the hash of the sent data.
pub fn send_request<W: ser::Writeable>(
&self,
t: Type,
rt: Type,
body: &W,
expect_h: Option<(Hash)>,
) -> Result<(), Error> {
let _sent = try!(self.underlying.send_msg(t, body));
let mut expects = self.expected_responses.lock().unwrap();
expects.push((rt, expect_h, Instant::now()));
Ok(())
}
/// Same as Connection
pub fn send_msg<W: ser::Writeable>(&self, t: Type, body: &W) -> Result<(), Error> {
self.underlying.send_msg(t, body)
}
/// Same as Connection
pub fn transmitted_bytes(&self) -> (u64, u64) {
self.underlying.transmitted_bytes()
}
}
| {
debug!(LOGGER, "Invalid {:?} message: {}", msg_type, e);
return Err(Error::Serialization(e));
} | conditional_block |
dpkg_install.go | package main
import (
"archive/tar"
"bytes"
"fmt"
"io"
"log"
"os"
"os/exec"
"path/filepath"
"strings"
"github.com/blakesmith/ar"
"github.com/julien-sobczak/deb822"
)
func main() {
// This program expects one or more package files to install.
if len(os.Args) < 2 {
log.Fatalf("Missing package archive(s)")
}
// Read the DPKG database
db, _ := loadDatabase()
// Unpack and configure the archive(s)
for _, archivePath := range os.Args[1:] {
processArchive(db, archivePath)
}
// For simplicity reasons, we don't manage a queue to defer
// the configuration of packages like in the official code.
}
//
// Dpkg Database
//
type Database struct {
// File /var/lib/dpkg/status
Status deb822.Document
// Packages under /var/lib/dpkg/info/
Packages []*PackageInfo
}
type PackageInfo struct {
Paragraph deb822.Paragraph // Extracted section in /var/lib/dpkg/status
// info
Files []string // File <name>.list
Conffiles []string // File <name>.conffiles
MaintainerScripts map[string]string // File <name>.{preinst,prerm,...}
Status string // Current status (as present in `Paragraph`)
StatusDirty bool // True to ask for sync
}
func (p *PackageInfo) Name() string {
// Extract the package name from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Package")
}
func (p *PackageInfo) Version() string {
// Extract the package version from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Version")
}
// isConffile determines if a file must be processed as a conffile.
func (p *PackageInfo) isConffile(path string) bool {
for _, conffile := range p.Conffiles {
if path == conffile {
return true
}
}
return false
}
// InfoPath returns the path of a file under /var/lib/dpkg/info/.
// Ex: "list" => /var/lib/dpkg/info/hello.list
func (p *PackageInfo) | (filename string) string {
return filepath.Join("/var/lib/dpkg", p.Name()+"."+filename)
}
// We now add a method to change the package status
// and make sure the section in the status file is updated too.
// This method will be used several times at the different steps
// of the installation process.
func (p *PackageInfo) SetStatus(new string) {
p.Status = new
p.StatusDirty = true
// Override in DEB 822 document used to write the status file
old := p.Paragraph.Values["Status"]
parts := strings.Split(old, " ")
newStatus := fmt.Sprintf("%s %s %s", parts[0], parts[1], new)
p.Paragraph.Values["Status"] = newStatus
}
// Now, we are ready to read the database directory to initialize the structs.
func loadDatabase() (*Database, error) {
// Load the status file
f, _ := os.Open("/var/lib/dpkg/status")
parser, _ := deb822.NewParser(f)
status, _ := parser.Parse()
// Read the info directory
var packages []*PackageInfo
for _, statusParagraph := range status.Paragraphs {
statusField := statusParagraph.Value("Status") // install ok installed
statusValues := strings.Split(statusField, " ")
pkg := PackageInfo{
Paragraph: statusParagraph,
MaintainerScripts: make(map[string]string),
Status: statusValues[2],
StatusDirty: false,
}
// Read the configuration files
pkg.Files, _ = ReadLines(pkg.InfoPath("list"))
pkg.Conffiles, _ = ReadLines(pkg.InfoPath("conffiles"))
// Read the maintainer scripts
maintainerScripts := []string{"preinst", "postinst", "prerm", "postrm"}
for _, script := range maintainerScripts {
scriptPath := pkg.InfoPath(script)
if _, err := os.Stat(scriptPath); !os.IsNotExist(err) {
content, err := os.ReadFile(scriptPath)
if err != nil {
return nil, err
}
pkg.MaintainerScripts[script] = string(content)
}
}
packages = append(packages, &pkg)
}
// We have read everything that interest us and are ready
// to populate the Database struct.
return &Database{
Status: status,
Packages: packages,
}, nil
}
// Now we are ready to process an archive to install.
func processArchive(db *Database, archivePath string) error {
// Read the Debian archive file
f, err := os.Open(archivePath)
if err != nil {
return err
}
defer f.Close()
reader := ar.NewReader(f)
// Skip debian-binary
reader.Next()
// control.tar
reader.Next()
var bufControl bytes.Buffer
io.Copy(&bufControl, reader)
pkg, err := parseControl(db, bufControl)
if err != nil {
return err
}
// Add the new package in the database
db.Packages = append(db.Packages, pkg)
db.Sync()
// data.tar
reader.Next()
var bufData bytes.Buffer
io.Copy(&bufData, reader)
fmt.Printf("Preparing to unpack %s ...\n", filepath.Base(archivePath))
if err := pkg.Unpack(bufData); err != nil {
return err
}
if err := pkg.Configure(); err != nil {
return err
}
db.Sync()
return nil
}
// parseControl processes the control.tar archive.
func parseControl(db *Database, buf bytes.Buffer) (*PackageInfo, error) {
// The control.tar archive contains the most important files
// we need to install the package.
// We need to extract metadata from the control file, determine
// if the package contains conffiles and maintainer scripts.
pkg := PackageInfo{
MaintainerScripts: make(map[string]string),
Status: "not-installed",
StatusDirty: true,
}
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return nil, err
}
// Read the file content
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return nil, err
}
switch filepath.Base(hdr.Name) {
case "control":
parser, _ := deb822.NewParser(strings.NewReader(buf.String()))
document, _ := parser.Parse()
controlParagraph := document.Paragraphs[0]
// Copy control fields and add the Status field in second position
pkg.Paragraph = deb822.Paragraph{
Values: make(map[string]string),
}
// Make sure the field "Package' comes first, then "Status",
// then remaining fields.
pkg.Paragraph.Order = append(
pkg.Paragraph.Order, "Package", "Status")
pkg.Paragraph.Values["Package"] = controlParagraph.Value("Package")
pkg.Paragraph.Values["Status"] = "install ok non-installed"
for _, field := range controlParagraph.Order {
if field == "Package" {
continue
}
pkg.Paragraph.Order = append(pkg.Paragraph.Order, field)
pkg.Paragraph.Values[field] = controlParagraph.Value(field)
}
case "conffiles":
pkg.Conffiles = SplitLines(buf.String())
case "prerm":
fallthrough
case "preinst":
fallthrough
case "postinst":
fallthrough
case "postrm":
pkg.MaintainerScripts[filepath.Base(hdr.Name)] = buf.String()
}
}
return &pkg, nil
}
// Unpack processes the data.tar archive.
func (p *PackageInfo) Unpack(buf bytes.Buffer) error {
// The unpacking process consists in extracting all files
// in data.tar to their final destination, except for conffiles,
// which are copied with a special extension that will be removed
// in the configure step.
if err := p.runMaintainerScript("preinst"); err != nil {
return err
}
fmt.Printf("Unpacking %s (%s) ...\n", p.Name(), p.Version())
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return err
}
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return err
}
switch hdr.Typeflag {
case tar.TypeReg:
dest := hdr.Name
if strings.HasPrefix(dest, "./") {
// ./usr/bin/hello => /usr/bin/hello
dest = dest[1:]
}
if !strings.HasPrefix(dest, "/") {
// usr/bin/hello => /usr/bin/hello
dest = "/" + dest
}
tmpdest := dest
if p.isConffile(tmpdest) {
// Extract using the extension .dpkg-new
tmpdest += ".dpkg-new"
}
if err := os.MkdirAll(filepath.Dir(tmpdest), 0755); err != nil {
log.Fatalf("Failed to unpack directory %s: %v", tmpdest, err)
}
content := buf.Bytes()
if err := os.WriteFile(tmpdest, content, 0755); err != nil {
log.Fatalf("Failed to unpack file %s: %v", tmpdest, err)
}
p.Files = append(p.Files, dest)
}
}
p.SetStatus("unpacked")
p.Sync()
return nil
}
// Configure processes the conffiles.
func (p *PackageInfo) Configure() error {
// The configure process consists in renaming the conffiles
// unpacked at the previous step.
//
// We ignore some implementation concerns like checking if a conffile
// has been updated using the last known checksum.
fmt.Printf("Setting up %s (%s) ...\n", p.Name(), p.Version())
// Rename conffiles
for _, conffile := range p.Conffiles {
os.Rename(conffile+".dpkg-new", conffile)
}
p.SetStatus("half-configured")
p.Sync()
// Run maintainer script
if err := p.runMaintainerScript("postinst"); err != nil {
return err
}
p.SetStatus("installed")
p.Sync()
return nil
}
func (p *PackageInfo) runMaintainerScript(name string) error {
// The control.tar file can contains scripts to be run at
// specific moments. This function uses the standard Go library
// to run the `sh` command with a maintainer scrpit as an argument.
if _, ok := p.MaintainerScripts[name]; !ok {
// Nothing to run
return nil
}
out, err := exec.Command("/bin/sh", p.InfoPath(name)).Output()
if err != nil {
return err
}
fmt.Print(string(out))
return nil
}
// We have covered the different steps of the installation process.
// We still need to write the code to sync the database.
func (d *Database) Sync() error {
newStatus := deb822.Document{
Paragraphs: []deb822.Paragraph{},
}
// Sync the /var/lib/dpkg/info directory
for _, pkg := range d.Packages {
newStatus.Paragraphs = append(newStatus.Paragraphs, pkg.Paragraph)
if pkg.StatusDirty {
if err := pkg.Sync(); err != nil {
return err
}
}
}
// Make a new version of /var/lib/dpkg/status
os.Rename("/var/lib/dpkg/status", "/var/lib/dpkg/status-old")
formatter := deb822.NewFormatter()
formatter.SetFoldedFields("Description")
formatter.SetMultilineFields("Conffiles")
if err := os.WriteFile("/var/lib/dpkg/status",
[]byte(formatter.Format(newStatus)), 0644); err != nil {
return err
}
return nil
}
func (p *PackageInfo) Sync() error {
// This function synchronizes the files under /var/lib/dpkg/info
// for a single package.
// Write <package>.list
if err := os.WriteFile(p.InfoPath("list"),
[]byte(MergeLines(p.Files)), 0644); err != nil {
return err
}
// Write <package>.conffiles
if err := os.WriteFile(p.InfoPath("conffiles"),
[]byte(MergeLines(p.Conffiles)), 0644); err != nil {
return err
}
// Write <package>.{preinst,prerm,postinst,postrm}
for name, content := range p.MaintainerScripts {
err := os.WriteFile(p.InfoPath(name), []byte(content), 0755)
if err != nil {
return err
}
}
p.StatusDirty = false
return nil
}
/* Utility functions */
func ReadLines(path string) ([]string, error) {
if _, err := os.Stat(path); !os.IsNotExist(err) {
content, err := os.ReadFile(path)
if err != nil {
return nil, err
}
return SplitLines(string(content)), nil
}
return nil, nil
}
func SplitLines(content string) []string {
var lines []string
for _, line := range strings.Split(string(content), "\n") {
if strings.TrimSpace(line) == "" {
continue
}
lines = append(lines, line)
}
return lines
}
func MergeLines(lines []string) string {
return strings.Join(lines, "\n") + "\n"
}
| InfoPath | identifier_name |
dpkg_install.go | package main
import (
"archive/tar"
"bytes"
"fmt"
"io"
"log"
"os"
"os/exec"
"path/filepath"
"strings"
"github.com/blakesmith/ar"
"github.com/julien-sobczak/deb822"
)
func main() {
// This program expects one or more package files to install.
if len(os.Args) < 2 {
log.Fatalf("Missing package archive(s)")
}
// Read the DPKG database
db, _ := loadDatabase()
// Unpack and configure the archive(s)
for _, archivePath := range os.Args[1:] {
processArchive(db, archivePath)
}
// For simplicity reasons, we don't manage a queue to defer
// the configuration of packages like in the official code.
}
//
// Dpkg Database
//
type Database struct {
// File /var/lib/dpkg/status
Status deb822.Document
// Packages under /var/lib/dpkg/info/
Packages []*PackageInfo
}
type PackageInfo struct {
Paragraph deb822.Paragraph // Extracted section in /var/lib/dpkg/status
// info
Files []string // File <name>.list
Conffiles []string // File <name>.conffiles
MaintainerScripts map[string]string // File <name>.{preinst,prerm,...}
Status string // Current status (as present in `Paragraph`)
StatusDirty bool // True to ask for sync
}
func (p *PackageInfo) Name() string {
// Extract the package name from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Package")
}
func (p *PackageInfo) Version() string {
// Extract the package version from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Version")
}
// isConffile determines if a file must be processed as a conffile.
func (p *PackageInfo) isConffile(path string) bool {
for _, conffile := range p.Conffiles {
if path == conffile {
return true
}
}
return false
}
// InfoPath returns the path of a file under /var/lib/dpkg/info/.
// Ex: "list" => /var/lib/dpkg/info/hello.list
func (p *PackageInfo) InfoPath(filename string) string {
return filepath.Join("/var/lib/dpkg", p.Name()+"."+filename)
}
// We now add a method to change the package status
// and make sure the section in the status file is updated too.
// This method will be used several times at the different steps
// of the installation process.
func (p *PackageInfo) SetStatus(new string) {
p.Status = new
p.StatusDirty = true
// Override in DEB 822 document used to write the status file
old := p.Paragraph.Values["Status"]
parts := strings.Split(old, " ")
newStatus := fmt.Sprintf("%s %s %s", parts[0], parts[1], new)
p.Paragraph.Values["Status"] = newStatus
}
// Now, we are ready to read the database directory to initialize the structs.
func loadDatabase() (*Database, error) {
// Load the status file
f, _ := os.Open("/var/lib/dpkg/status")
parser, _ := deb822.NewParser(f)
status, _ := parser.Parse()
// Read the info directory
var packages []*PackageInfo
for _, statusParagraph := range status.Paragraphs {
statusField := statusParagraph.Value("Status") // install ok installed
statusValues := strings.Split(statusField, " ")
pkg := PackageInfo{
Paragraph: statusParagraph,
MaintainerScripts: make(map[string]string),
Status: statusValues[2],
StatusDirty: false,
}
// Read the configuration files
pkg.Files, _ = ReadLines(pkg.InfoPath("list"))
pkg.Conffiles, _ = ReadLines(pkg.InfoPath("conffiles"))
// Read the maintainer scripts
maintainerScripts := []string{"preinst", "postinst", "prerm", "postrm"}
for _, script := range maintainerScripts {
scriptPath := pkg.InfoPath(script)
if _, err := os.Stat(scriptPath); !os.IsNotExist(err) {
content, err := os.ReadFile(scriptPath)
if err != nil {
return nil, err
}
pkg.MaintainerScripts[script] = string(content)
}
}
packages = append(packages, &pkg)
}
// We have read everything that interest us and are ready
// to populate the Database struct.
return &Database{
Status: status,
Packages: packages,
}, nil
}
// Now we are ready to process an archive to install.
func processArchive(db *Database, archivePath string) error {
// Read the Debian archive file
f, err := os.Open(archivePath)
if err != nil {
return err
}
defer f.Close()
reader := ar.NewReader(f)
// Skip debian-binary
reader.Next()
// control.tar
reader.Next()
var bufControl bytes.Buffer
io.Copy(&bufControl, reader)
pkg, err := parseControl(db, bufControl)
if err != nil {
return err
}
// Add the new package in the database
db.Packages = append(db.Packages, pkg)
db.Sync()
// data.tar
reader.Next()
var bufData bytes.Buffer
io.Copy(&bufData, reader)
fmt.Printf("Preparing to unpack %s ...\n", filepath.Base(archivePath))
if err := pkg.Unpack(bufData); err != nil {
return err
}
if err := pkg.Configure(); err != nil {
return err
}
db.Sync()
return nil
}
// parseControl processes the control.tar archive.
func parseControl(db *Database, buf bytes.Buffer) (*PackageInfo, error) {
// The control.tar archive contains the most important files
// we need to install the package.
// We need to extract metadata from the control file, determine
// if the package contains conffiles and maintainer scripts.
pkg := PackageInfo{
MaintainerScripts: make(map[string]string),
Status: "not-installed",
StatusDirty: true,
}
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return nil, err
}
// Read the file content
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return nil, err
}
switch filepath.Base(hdr.Name) {
case "control":
parser, _ := deb822.NewParser(strings.NewReader(buf.String()))
document, _ := parser.Parse()
controlParagraph := document.Paragraphs[0]
// Copy control fields and add the Status field in second position
pkg.Paragraph = deb822.Paragraph{
Values: make(map[string]string),
}
// Make sure the field "Package' comes first, then "Status",
// then remaining fields.
pkg.Paragraph.Order = append(
pkg.Paragraph.Order, "Package", "Status")
pkg.Paragraph.Values["Package"] = controlParagraph.Value("Package")
pkg.Paragraph.Values["Status"] = "install ok non-installed"
for _, field := range controlParagraph.Order {
if field == "Package" {
continue
}
pkg.Paragraph.Order = append(pkg.Paragraph.Order, field)
pkg.Paragraph.Values[field] = controlParagraph.Value(field)
}
case "conffiles":
pkg.Conffiles = SplitLines(buf.String())
case "prerm":
fallthrough
case "preinst":
fallthrough
case "postinst":
fallthrough
case "postrm":
pkg.MaintainerScripts[filepath.Base(hdr.Name)] = buf.String()
}
}
return &pkg, nil
}
// Unpack processes the data.tar archive.
func (p *PackageInfo) Unpack(buf bytes.Buffer) error {
// The unpacking process consists in extracting all files
// in data.tar to their final destination, except for conffiles,
// which are copied with a special extension that will be removed
// in the configure step.
if err := p.runMaintainerScript("preinst"); err != nil {
return err
}
fmt.Printf("Unpacking %s (%s) ...\n", p.Name(), p.Version())
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return err
}
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return err
}
switch hdr.Typeflag {
case tar.TypeReg:
dest := hdr.Name
if strings.HasPrefix(dest, "./") {
// ./usr/bin/hello => /usr/bin/hello
dest = dest[1:]
}
if !strings.HasPrefix(dest, "/") {
// usr/bin/hello => /usr/bin/hello
dest = "/" + dest
}
tmpdest := dest
if p.isConffile(tmpdest) {
// Extract using the extension .dpkg-new
tmpdest += ".dpkg-new"
}
if err := os.MkdirAll(filepath.Dir(tmpdest), 0755); err != nil {
log.Fatalf("Failed to unpack directory %s: %v", tmpdest, err)
}
content := buf.Bytes()
if err := os.WriteFile(tmpdest, content, 0755); err != nil {
log.Fatalf("Failed to unpack file %s: %v", tmpdest, err)
}
p.Files = append(p.Files, dest)
}
}
p.SetStatus("unpacked")
p.Sync()
return nil
}
// Configure processes the conffiles.
func (p *PackageInfo) Configure() error {
// The configure process consists in renaming the conffiles
// unpacked at the previous step.
//
// We ignore some implementation concerns like checking if a conffile
// has been updated using the last known checksum.
fmt.Printf("Setting up %s (%s) ...\n", p.Name(), p.Version())
// Rename conffiles
for _, conffile := range p.Conffiles {
os.Rename(conffile+".dpkg-new", conffile)
}
p.SetStatus("half-configured")
p.Sync()
// Run maintainer script
if err := p.runMaintainerScript("postinst"); err != nil {
return err
}
p.SetStatus("installed")
p.Sync()
return nil
}
func (p *PackageInfo) runMaintainerScript(name string) error {
// The control.tar file can contains scripts to be run at
// specific moments. This function uses the standard Go library
// to run the `sh` command with a maintainer scrpit as an argument.
if _, ok := p.MaintainerScripts[name]; !ok {
// Nothing to run
return nil
}
out, err := exec.Command("/bin/sh", p.InfoPath(name)).Output()
if err != nil {
return err
}
fmt.Print(string(out))
return nil
}
// We have covered the different steps of the installation process.
// We still need to write the code to sync the database.
func (d *Database) Sync() error {
newStatus := deb822.Document{
Paragraphs: []deb822.Paragraph{},
}
// Sync the /var/lib/dpkg/info directory
for _, pkg := range d.Packages {
newStatus.Paragraphs = append(newStatus.Paragraphs, pkg.Paragraph)
if pkg.StatusDirty {
if err := pkg.Sync(); err != nil {
return err
}
}
}
// Make a new version of /var/lib/dpkg/status
os.Rename("/var/lib/dpkg/status", "/var/lib/dpkg/status-old")
formatter := deb822.NewFormatter()
formatter.SetFoldedFields("Description")
formatter.SetMultilineFields("Conffiles")
if err := os.WriteFile("/var/lib/dpkg/status",
[]byte(formatter.Format(newStatus)), 0644); err != nil {
return err
}
return nil
}
func (p *PackageInfo) Sync() error |
/* Utility functions */
func ReadLines(path string) ([]string, error) {
if _, err := os.Stat(path); !os.IsNotExist(err) {
content, err := os.ReadFile(path)
if err != nil {
return nil, err
}
return SplitLines(string(content)), nil
}
return nil, nil
}
func SplitLines(content string) []string {
var lines []string
for _, line := range strings.Split(string(content), "\n") {
if strings.TrimSpace(line) == "" {
continue
}
lines = append(lines, line)
}
return lines
}
func MergeLines(lines []string) string {
return strings.Join(lines, "\n") + "\n"
}
| {
// This function synchronizes the files under /var/lib/dpkg/info
// for a single package.
// Write <package>.list
if err := os.WriteFile(p.InfoPath("list"),
[]byte(MergeLines(p.Files)), 0644); err != nil {
return err
}
// Write <package>.conffiles
if err := os.WriteFile(p.InfoPath("conffiles"),
[]byte(MergeLines(p.Conffiles)), 0644); err != nil {
return err
}
// Write <package>.{preinst,prerm,postinst,postrm}
for name, content := range p.MaintainerScripts {
err := os.WriteFile(p.InfoPath(name), []byte(content), 0755)
if err != nil {
return err
}
}
p.StatusDirty = false
return nil
} | identifier_body |
dpkg_install.go | package main
import (
"archive/tar"
"bytes"
"fmt"
"io"
"log"
"os"
"os/exec"
"path/filepath"
"strings"
"github.com/blakesmith/ar"
"github.com/julien-sobczak/deb822"
)
func main() {
// This program expects one or more package files to install.
if len(os.Args) < 2 {
log.Fatalf("Missing package archive(s)")
}
// Read the DPKG database
db, _ := loadDatabase()
// Unpack and configure the archive(s)
for _, archivePath := range os.Args[1:] {
processArchive(db, archivePath)
}
// For simplicity reasons, we don't manage a queue to defer
// the configuration of packages like in the official code.
}
//
// Dpkg Database
//
type Database struct {
// File /var/lib/dpkg/status
Status deb822.Document
// Packages under /var/lib/dpkg/info/
Packages []*PackageInfo
}
type PackageInfo struct {
Paragraph deb822.Paragraph // Extracted section in /var/lib/dpkg/status
// info
Files []string // File <name>.list
Conffiles []string // File <name>.conffiles
MaintainerScripts map[string]string // File <name>.{preinst,prerm,...}
Status string // Current status (as present in `Paragraph`)
StatusDirty bool // True to ask for sync
}
func (p *PackageInfo) Name() string {
// Extract the package name from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Package")
}
func (p *PackageInfo) Version() string {
// Extract the package version from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Version")
}
// isConffile determines if a file must be processed as a conffile.
func (p *PackageInfo) isConffile(path string) bool {
for _, conffile := range p.Conffiles {
if path == conffile {
return true
}
}
return false
}
// InfoPath returns the path of a file under /var/lib/dpkg/info/.
// Ex: "list" => /var/lib/dpkg/info/hello.list
func (p *PackageInfo) InfoPath(filename string) string {
return filepath.Join("/var/lib/dpkg", p.Name()+"."+filename)
}
// We now add a method to change the package status
// and make sure the section in the status file is updated too.
// This method will be used several times at the different steps
// of the installation process.
func (p *PackageInfo) SetStatus(new string) {
p.Status = new
p.StatusDirty = true
// Override in DEB 822 document used to write the status file
old := p.Paragraph.Values["Status"]
parts := strings.Split(old, " ")
newStatus := fmt.Sprintf("%s %s %s", parts[0], parts[1], new)
p.Paragraph.Values["Status"] = newStatus
}
// Now, we are ready to read the database directory to initialize the structs.
func loadDatabase() (*Database, error) {
// Load the status file
f, _ := os.Open("/var/lib/dpkg/status")
parser, _ := deb822.NewParser(f)
status, _ := parser.Parse()
// Read the info directory
var packages []*PackageInfo
for _, statusParagraph := range status.Paragraphs {
statusField := statusParagraph.Value("Status") // install ok installed
statusValues := strings.Split(statusField, " ")
pkg := PackageInfo{
Paragraph: statusParagraph,
MaintainerScripts: make(map[string]string),
Status: statusValues[2],
StatusDirty: false,
}
// Read the configuration files
pkg.Files, _ = ReadLines(pkg.InfoPath("list"))
pkg.Conffiles, _ = ReadLines(pkg.InfoPath("conffiles"))
// Read the maintainer scripts
maintainerScripts := []string{"preinst", "postinst", "prerm", "postrm"}
for _, script := range maintainerScripts {
scriptPath := pkg.InfoPath(script)
if _, err := os.Stat(scriptPath); !os.IsNotExist(err) {
content, err := os.ReadFile(scriptPath)
if err != nil {
return nil, err
}
pkg.MaintainerScripts[script] = string(content)
}
}
packages = append(packages, &pkg)
}
// We have read everything that interest us and are ready
// to populate the Database struct.
return &Database{
Status: status,
Packages: packages,
}, nil
}
// Now we are ready to process an archive to install.
func processArchive(db *Database, archivePath string) error {
// Read the Debian archive file
f, err := os.Open(archivePath)
if err != nil {
return err
}
defer f.Close()
reader := ar.NewReader(f)
// Skip debian-binary
reader.Next() | var bufControl bytes.Buffer
io.Copy(&bufControl, reader)
pkg, err := parseControl(db, bufControl)
if err != nil {
return err
}
// Add the new package in the database
db.Packages = append(db.Packages, pkg)
db.Sync()
// data.tar
reader.Next()
var bufData bytes.Buffer
io.Copy(&bufData, reader)
fmt.Printf("Preparing to unpack %s ...\n", filepath.Base(archivePath))
if err := pkg.Unpack(bufData); err != nil {
return err
}
if err := pkg.Configure(); err != nil {
return err
}
db.Sync()
return nil
}
// parseControl processes the control.tar archive.
func parseControl(db *Database, buf bytes.Buffer) (*PackageInfo, error) {
// The control.tar archive contains the most important files
// we need to install the package.
// We need to extract metadata from the control file, determine
// if the package contains conffiles and maintainer scripts.
pkg := PackageInfo{
MaintainerScripts: make(map[string]string),
Status: "not-installed",
StatusDirty: true,
}
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return nil, err
}
// Read the file content
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return nil, err
}
switch filepath.Base(hdr.Name) {
case "control":
parser, _ := deb822.NewParser(strings.NewReader(buf.String()))
document, _ := parser.Parse()
controlParagraph := document.Paragraphs[0]
// Copy control fields and add the Status field in second position
pkg.Paragraph = deb822.Paragraph{
Values: make(map[string]string),
}
// Make sure the field "Package' comes first, then "Status",
// then remaining fields.
pkg.Paragraph.Order = append(
pkg.Paragraph.Order, "Package", "Status")
pkg.Paragraph.Values["Package"] = controlParagraph.Value("Package")
pkg.Paragraph.Values["Status"] = "install ok non-installed"
for _, field := range controlParagraph.Order {
if field == "Package" {
continue
}
pkg.Paragraph.Order = append(pkg.Paragraph.Order, field)
pkg.Paragraph.Values[field] = controlParagraph.Value(field)
}
case "conffiles":
pkg.Conffiles = SplitLines(buf.String())
case "prerm":
fallthrough
case "preinst":
fallthrough
case "postinst":
fallthrough
case "postrm":
pkg.MaintainerScripts[filepath.Base(hdr.Name)] = buf.String()
}
}
return &pkg, nil
}
// Unpack processes the data.tar archive.
func (p *PackageInfo) Unpack(buf bytes.Buffer) error {
// The unpacking process consists in extracting all files
// in data.tar to their final destination, except for conffiles,
// which are copied with a special extension that will be removed
// in the configure step.
if err := p.runMaintainerScript("preinst"); err != nil {
return err
}
fmt.Printf("Unpacking %s (%s) ...\n", p.Name(), p.Version())
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return err
}
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return err
}
switch hdr.Typeflag {
case tar.TypeReg:
dest := hdr.Name
if strings.HasPrefix(dest, "./") {
// ./usr/bin/hello => /usr/bin/hello
dest = dest[1:]
}
if !strings.HasPrefix(dest, "/") {
// usr/bin/hello => /usr/bin/hello
dest = "/" + dest
}
tmpdest := dest
if p.isConffile(tmpdest) {
// Extract using the extension .dpkg-new
tmpdest += ".dpkg-new"
}
if err := os.MkdirAll(filepath.Dir(tmpdest), 0755); err != nil {
log.Fatalf("Failed to unpack directory %s: %v", tmpdest, err)
}
content := buf.Bytes()
if err := os.WriteFile(tmpdest, content, 0755); err != nil {
log.Fatalf("Failed to unpack file %s: %v", tmpdest, err)
}
p.Files = append(p.Files, dest)
}
}
p.SetStatus("unpacked")
p.Sync()
return nil
}
// Configure processes the conffiles.
func (p *PackageInfo) Configure() error {
// The configure process consists in renaming the conffiles
// unpacked at the previous step.
//
// We ignore some implementation concerns like checking if a conffile
// has been updated using the last known checksum.
fmt.Printf("Setting up %s (%s) ...\n", p.Name(), p.Version())
// Rename conffiles
for _, conffile := range p.Conffiles {
os.Rename(conffile+".dpkg-new", conffile)
}
p.SetStatus("half-configured")
p.Sync()
// Run maintainer script
if err := p.runMaintainerScript("postinst"); err != nil {
return err
}
p.SetStatus("installed")
p.Sync()
return nil
}
func (p *PackageInfo) runMaintainerScript(name string) error {
// The control.tar file can contains scripts to be run at
// specific moments. This function uses the standard Go library
// to run the `sh` command with a maintainer scrpit as an argument.
if _, ok := p.MaintainerScripts[name]; !ok {
// Nothing to run
return nil
}
out, err := exec.Command("/bin/sh", p.InfoPath(name)).Output()
if err != nil {
return err
}
fmt.Print(string(out))
return nil
}
// We have covered the different steps of the installation process.
// We still need to write the code to sync the database.
func (d *Database) Sync() error {
newStatus := deb822.Document{
Paragraphs: []deb822.Paragraph{},
}
// Sync the /var/lib/dpkg/info directory
for _, pkg := range d.Packages {
newStatus.Paragraphs = append(newStatus.Paragraphs, pkg.Paragraph)
if pkg.StatusDirty {
if err := pkg.Sync(); err != nil {
return err
}
}
}
// Make a new version of /var/lib/dpkg/status
os.Rename("/var/lib/dpkg/status", "/var/lib/dpkg/status-old")
formatter := deb822.NewFormatter()
formatter.SetFoldedFields("Description")
formatter.SetMultilineFields("Conffiles")
if err := os.WriteFile("/var/lib/dpkg/status",
[]byte(formatter.Format(newStatus)), 0644); err != nil {
return err
}
return nil
}
func (p *PackageInfo) Sync() error {
// This function synchronizes the files under /var/lib/dpkg/info
// for a single package.
// Write <package>.list
if err := os.WriteFile(p.InfoPath("list"),
[]byte(MergeLines(p.Files)), 0644); err != nil {
return err
}
// Write <package>.conffiles
if err := os.WriteFile(p.InfoPath("conffiles"),
[]byte(MergeLines(p.Conffiles)), 0644); err != nil {
return err
}
// Write <package>.{preinst,prerm,postinst,postrm}
for name, content := range p.MaintainerScripts {
err := os.WriteFile(p.InfoPath(name), []byte(content), 0755)
if err != nil {
return err
}
}
p.StatusDirty = false
return nil
}
/* Utility functions */
func ReadLines(path string) ([]string, error) {
if _, err := os.Stat(path); !os.IsNotExist(err) {
content, err := os.ReadFile(path)
if err != nil {
return nil, err
}
return SplitLines(string(content)), nil
}
return nil, nil
}
func SplitLines(content string) []string {
var lines []string
for _, line := range strings.Split(string(content), "\n") {
if strings.TrimSpace(line) == "" {
continue
}
lines = append(lines, line)
}
return lines
}
func MergeLines(lines []string) string {
return strings.Join(lines, "\n") + "\n"
} |
// control.tar
reader.Next() | random_line_split |
dpkg_install.go | package main
import (
"archive/tar"
"bytes"
"fmt"
"io"
"log"
"os"
"os/exec"
"path/filepath"
"strings"
"github.com/blakesmith/ar"
"github.com/julien-sobczak/deb822"
)
func main() {
// This program expects one or more package files to install.
if len(os.Args) < 2 {
log.Fatalf("Missing package archive(s)")
}
// Read the DPKG database
db, _ := loadDatabase()
// Unpack and configure the archive(s)
for _, archivePath := range os.Args[1:] {
processArchive(db, archivePath)
}
// For simplicity reasons, we don't manage a queue to defer
// the configuration of packages like in the official code.
}
//
// Dpkg Database
//
type Database struct {
// File /var/lib/dpkg/status
Status deb822.Document
// Packages under /var/lib/dpkg/info/
Packages []*PackageInfo
}
type PackageInfo struct {
Paragraph deb822.Paragraph // Extracted section in /var/lib/dpkg/status
// info
Files []string // File <name>.list
Conffiles []string // File <name>.conffiles
MaintainerScripts map[string]string // File <name>.{preinst,prerm,...}
Status string // Current status (as present in `Paragraph`)
StatusDirty bool // True to ask for sync
}
func (p *PackageInfo) Name() string {
// Extract the package name from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Package")
}
func (p *PackageInfo) Version() string {
// Extract the package version from its section in /var/lib/dpkg/status
return p.Paragraph.Value("Version")
}
// isConffile determines if a file must be processed as a conffile.
func (p *PackageInfo) isConffile(path string) bool {
for _, conffile := range p.Conffiles {
if path == conffile {
return true
}
}
return false
}
// InfoPath returns the path of a file under /var/lib/dpkg/info/.
// Ex: "list" => /var/lib/dpkg/info/hello.list
func (p *PackageInfo) InfoPath(filename string) string {
return filepath.Join("/var/lib/dpkg", p.Name()+"."+filename)
}
// We now add a method to change the package status
// and make sure the section in the status file is updated too.
// This method will be used several times at the different steps
// of the installation process.
func (p *PackageInfo) SetStatus(new string) {
p.Status = new
p.StatusDirty = true
// Override in DEB 822 document used to write the status file
old := p.Paragraph.Values["Status"]
parts := strings.Split(old, " ")
newStatus := fmt.Sprintf("%s %s %s", parts[0], parts[1], new)
p.Paragraph.Values["Status"] = newStatus
}
// Now, we are ready to read the database directory to initialize the structs.
func loadDatabase() (*Database, error) {
// Load the status file
f, _ := os.Open("/var/lib/dpkg/status")
parser, _ := deb822.NewParser(f)
status, _ := parser.Parse()
// Read the info directory
var packages []*PackageInfo
for _, statusParagraph := range status.Paragraphs {
statusField := statusParagraph.Value("Status") // install ok installed
statusValues := strings.Split(statusField, " ")
pkg := PackageInfo{
Paragraph: statusParagraph,
MaintainerScripts: make(map[string]string),
Status: statusValues[2],
StatusDirty: false,
}
// Read the configuration files
pkg.Files, _ = ReadLines(pkg.InfoPath("list"))
pkg.Conffiles, _ = ReadLines(pkg.InfoPath("conffiles"))
// Read the maintainer scripts
maintainerScripts := []string{"preinst", "postinst", "prerm", "postrm"}
for _, script := range maintainerScripts {
scriptPath := pkg.InfoPath(script)
if _, err := os.Stat(scriptPath); !os.IsNotExist(err) {
content, err := os.ReadFile(scriptPath)
if err != nil {
return nil, err
}
pkg.MaintainerScripts[script] = string(content)
}
}
packages = append(packages, &pkg)
}
// We have read everything that interest us and are ready
// to populate the Database struct.
return &Database{
Status: status,
Packages: packages,
}, nil
}
// Now we are ready to process an archive to install.
func processArchive(db *Database, archivePath string) error {
// Read the Debian archive file
f, err := os.Open(archivePath)
if err != nil {
return err
}
defer f.Close()
reader := ar.NewReader(f)
// Skip debian-binary
reader.Next()
// control.tar
reader.Next()
var bufControl bytes.Buffer
io.Copy(&bufControl, reader)
pkg, err := parseControl(db, bufControl)
if err != nil |
// Add the new package in the database
db.Packages = append(db.Packages, pkg)
db.Sync()
// data.tar
reader.Next()
var bufData bytes.Buffer
io.Copy(&bufData, reader)
fmt.Printf("Preparing to unpack %s ...\n", filepath.Base(archivePath))
if err := pkg.Unpack(bufData); err != nil {
return err
}
if err := pkg.Configure(); err != nil {
return err
}
db.Sync()
return nil
}
// parseControl processes the control.tar archive.
func parseControl(db *Database, buf bytes.Buffer) (*PackageInfo, error) {
// The control.tar archive contains the most important files
// we need to install the package.
// We need to extract metadata from the control file, determine
// if the package contains conffiles and maintainer scripts.
pkg := PackageInfo{
MaintainerScripts: make(map[string]string),
Status: "not-installed",
StatusDirty: true,
}
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return nil, err
}
// Read the file content
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return nil, err
}
switch filepath.Base(hdr.Name) {
case "control":
parser, _ := deb822.NewParser(strings.NewReader(buf.String()))
document, _ := parser.Parse()
controlParagraph := document.Paragraphs[0]
// Copy control fields and add the Status field in second position
pkg.Paragraph = deb822.Paragraph{
Values: make(map[string]string),
}
// Make sure the field "Package' comes first, then "Status",
// then remaining fields.
pkg.Paragraph.Order = append(
pkg.Paragraph.Order, "Package", "Status")
pkg.Paragraph.Values["Package"] = controlParagraph.Value("Package")
pkg.Paragraph.Values["Status"] = "install ok non-installed"
for _, field := range controlParagraph.Order {
if field == "Package" {
continue
}
pkg.Paragraph.Order = append(pkg.Paragraph.Order, field)
pkg.Paragraph.Values[field] = controlParagraph.Value(field)
}
case "conffiles":
pkg.Conffiles = SplitLines(buf.String())
case "prerm":
fallthrough
case "preinst":
fallthrough
case "postinst":
fallthrough
case "postrm":
pkg.MaintainerScripts[filepath.Base(hdr.Name)] = buf.String()
}
}
return &pkg, nil
}
// Unpack processes the data.tar archive.
func (p *PackageInfo) Unpack(buf bytes.Buffer) error {
// The unpacking process consists in extracting all files
// in data.tar to their final destination, except for conffiles,
// which are copied with a special extension that will be removed
// in the configure step.
if err := p.runMaintainerScript("preinst"); err != nil {
return err
}
fmt.Printf("Unpacking %s (%s) ...\n", p.Name(), p.Version())
tr := tar.NewReader(&buf)
for {
hdr, err := tr.Next()
if err == io.EOF {
break // End of archive
}
if err != nil {
return err
}
var buf bytes.Buffer
if _, err := io.Copy(&buf, tr); err != nil {
return err
}
switch hdr.Typeflag {
case tar.TypeReg:
dest := hdr.Name
if strings.HasPrefix(dest, "./") {
// ./usr/bin/hello => /usr/bin/hello
dest = dest[1:]
}
if !strings.HasPrefix(dest, "/") {
// usr/bin/hello => /usr/bin/hello
dest = "/" + dest
}
tmpdest := dest
if p.isConffile(tmpdest) {
// Extract using the extension .dpkg-new
tmpdest += ".dpkg-new"
}
if err := os.MkdirAll(filepath.Dir(tmpdest), 0755); err != nil {
log.Fatalf("Failed to unpack directory %s: %v", tmpdest, err)
}
content := buf.Bytes()
if err := os.WriteFile(tmpdest, content, 0755); err != nil {
log.Fatalf("Failed to unpack file %s: %v", tmpdest, err)
}
p.Files = append(p.Files, dest)
}
}
p.SetStatus("unpacked")
p.Sync()
return nil
}
// Configure processes the conffiles.
func (p *PackageInfo) Configure() error {
// The configure process consists in renaming the conffiles
// unpacked at the previous step.
//
// We ignore some implementation concerns like checking if a conffile
// has been updated using the last known checksum.
fmt.Printf("Setting up %s (%s) ...\n", p.Name(), p.Version())
// Rename conffiles
for _, conffile := range p.Conffiles {
os.Rename(conffile+".dpkg-new", conffile)
}
p.SetStatus("half-configured")
p.Sync()
// Run maintainer script
if err := p.runMaintainerScript("postinst"); err != nil {
return err
}
p.SetStatus("installed")
p.Sync()
return nil
}
func (p *PackageInfo) runMaintainerScript(name string) error {
// The control.tar file can contains scripts to be run at
// specific moments. This function uses the standard Go library
// to run the `sh` command with a maintainer scrpit as an argument.
if _, ok := p.MaintainerScripts[name]; !ok {
// Nothing to run
return nil
}
out, err := exec.Command("/bin/sh", p.InfoPath(name)).Output()
if err != nil {
return err
}
fmt.Print(string(out))
return nil
}
// We have covered the different steps of the installation process.
// We still need to write the code to sync the database.
func (d *Database) Sync() error {
newStatus := deb822.Document{
Paragraphs: []deb822.Paragraph{},
}
// Sync the /var/lib/dpkg/info directory
for _, pkg := range d.Packages {
newStatus.Paragraphs = append(newStatus.Paragraphs, pkg.Paragraph)
if pkg.StatusDirty {
if err := pkg.Sync(); err != nil {
return err
}
}
}
// Make a new version of /var/lib/dpkg/status
os.Rename("/var/lib/dpkg/status", "/var/lib/dpkg/status-old")
formatter := deb822.NewFormatter()
formatter.SetFoldedFields("Description")
formatter.SetMultilineFields("Conffiles")
if err := os.WriteFile("/var/lib/dpkg/status",
[]byte(formatter.Format(newStatus)), 0644); err != nil {
return err
}
return nil
}
func (p *PackageInfo) Sync() error {
// This function synchronizes the files under /var/lib/dpkg/info
// for a single package.
// Write <package>.list
if err := os.WriteFile(p.InfoPath("list"),
[]byte(MergeLines(p.Files)), 0644); err != nil {
return err
}
// Write <package>.conffiles
if err := os.WriteFile(p.InfoPath("conffiles"),
[]byte(MergeLines(p.Conffiles)), 0644); err != nil {
return err
}
// Write <package>.{preinst,prerm,postinst,postrm}
for name, content := range p.MaintainerScripts {
err := os.WriteFile(p.InfoPath(name), []byte(content), 0755)
if err != nil {
return err
}
}
p.StatusDirty = false
return nil
}
/* Utility functions */
func ReadLines(path string) ([]string, error) {
if _, err := os.Stat(path); !os.IsNotExist(err) {
content, err := os.ReadFile(path)
if err != nil {
return nil, err
}
return SplitLines(string(content)), nil
}
return nil, nil
}
func SplitLines(content string) []string {
var lines []string
for _, line := range strings.Split(string(content), "\n") {
if strings.TrimSpace(line) == "" {
continue
}
lines = append(lines, line)
}
return lines
}
func MergeLines(lines []string) string {
return strings.Join(lines, "\n") + "\n"
}
| {
return err
} | conditional_block |
clarans.py | """!
@brief Cluster analysis algorithm: CLARANS.
@details Implementation based on paper @cite article::clarans::1.
@authors Andrei Novikov (pyclustering@yandex.ru)
@date 2014-2019
@copyright GNU Public License
@cond GNU_PUBLIC_LICENSE
PyClustering is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
PyClustering is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
@endcond
"""
import random
from pyclustering.cluster.encoder import type_encoding
from pyclustering.utils import euclidean_distance_square
def plot_pam(data, cl, equal_axis_scale = False):
import matplotlib.pyplot as plt
import numpy as np
fig,ax = plt.subplots(figsize=(14,6))
plt.scatter(np.array(data)[:,0], np.array(data)[:,1], s=300, color="white", edgecolor="black")
colors = { 0:"seagreen", 1:'beige', 2:'yellow', 3:'grey',
4:'pink', 5:'turquoise', 6:'orange', 7:'purple', 8:'yellowgreen', 9:'olive', 10:'brown',
11:'tan', 12: 'plum', 13:'rosybrown', 14:'lightblue', 15:"khaki", 16:"gainsboro", 17:"peachpuff"}
for i,el in enumerate(list(cl.values())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=300, color=colors[i%17], edgecolor="black")
for i,el in enumerate(list(cl.keys())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=500, color="red", marker="X", edgecolor="black")
xmin, xmax, ymin, ymax = plt.axis()
xwidth = xmax - xmin
ywidth = ymax - ymin
xw1 = xwidth*0.01
yw1 = ywidth*0.01
xw2 = xwidth*0.005
yw2 = ywidth*0.01
xw3 = xwidth*0.01
yw3 = ywidth*0.01
for i, txt in enumerate(range(len(data))):
if len(str(txt))==2:
ax.annotate(txt, (np.array(data)[:,0][i]-xw1, np.array(data)[:,1][i]-yw1), fontsize=12, size=12)
elif len(str(txt))==1:
ax.annotate(txt, (np.array(data)[:,0][i]-xw2, np.array(data)[:,1][i]-yw2), fontsize=12, size=12)
else:
ax.annotate(txt, (np.array(data)[:,0][i]-xw3, np.array(data)[:,1][i]-yw3), fontsize=9, size=9)
if equal_axis_scale == True:
plt.gca().set_aspect('equal', adjustable='box')
plt.show()
class clarans:
"""!
@brief Class represents clustering algorithm CLARANS (a method for clustering objects for spatial data mining).
"""
def __init__(self, data, number_clusters, numlocal, maxneighbor):
"""!
@brief Constructor of clustering algorithm CLARANS.
@details The higher the value of maxneighbor, the closer is CLARANS to K-Medoids, and the longer is each search of a local minima.
@param[in] data (list): Input data that is presented as list of points (objects), each point should be represented by list or tuple.
@param[in] number_clusters (uint): Amount of clusters that should be allocated.
@param[in] numlocal (uint): The number of local minima obtained (amount of iterations for solving the problem).
@param[in] maxneighbor (uint): The maximum number of neighbors examined.
"""
self.__pointer_data = data
self.__numlocal = numlocal
self.__maxneighbor = maxneighbor
self.__number_clusters = number_clusters
self.__clusters = []
self.__current = []
self.__belong = []
self.__optimal_medoids = []
self.__optimal_estimation = float('inf')
self.__verify_arguments()
def __verify_arguments(self):
"""!
@brief Verify input parameters for the algorithm and throw exception in case of incorrectness.
"""
if len(self.__pointer_data) == 0:
raise ValueError("Input data is empty (size: '%d')." % len(self.__pointer_data))
if self.__number_clusters <= 0:
raise ValueError("Amount of cluster (current value: '%d') for allocation should be greater than 0." %
self.__number_clusters)
if self.__numlocal < 0:
raise ValueError("Local minima (current value: '%d') should be greater or equal to 0." % self.__numlocal)
if self.__maxneighbor < 0:
raise ValueError("Maximum number of neighbors (current value: '%d') should be greater or "
"equal to 0." % self.__maxneighbor)
def process(self, plotting=False):
"""!
@brief Performs cluster analysis in line with rules of CLARANS algorithm.
@return (clarans) Returns itself (CLARANS instance).
@see get_clusters()
@see get_medoids()
"""
random.seed()
# loop for a numlocal number of times
for _ in range(0, self.__numlocal):
print("numlocal: ", _)
# set (current) random medoids
self.__current = random.sample(range(0, len(self.__pointer_data)), self.__number_clusters)
# update clusters in line with random allocated medoids
self.__update_clusters(self.__current)
# optimize configuration
self.__optimize_configuration()
# obtain cost of current cluster configuration and compare it with the best obtained
estimation = self.__calculate_estimation()
if estimation < self.__optimal_estimation:
print("Better configuration found with medoids: {0} and cost: {1}".format(self.__current[:], estimation))
self.__optimal_medoids = self.__current[:]
self.__optimal_estimation = estimation
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
else:
print("Configuration found does not improve current best one because its cost is {0}".format(estimation))
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
self.__update_clusters(self.__optimal_medoids)
return self
def get_clusters(self):
"""!
@brief Returns allocated clusters by the algorithm.
@remark Allocated clusters can be returned only after data processing (use method process()), otherwise empty list is returned.
@return (list) List of allocated clusters, each cluster contains indexes of objects in list of data.
@see process()
@see get_medoids()
"""
return self.__clusters
def get_medoids(self):
"""!
@brief Returns list of medoids of allocated clusters.
@see process()
@see get_clusters()
"""
return self.__optimal_medoids
def get_cluster_encoding(self):
"""!
@brief Returns clustering result representation type that indicate how clusters are encoded.
@return (type_encoding) Clustering result representation.
@see get_clusters()
"""
return type_encoding.CLUSTER_INDEX_LIST_SEPARATION
def __update_clusters(self, medoids):
"""!
@brief Forms cluster in line with specified medoids by calculation distance from each point to medoids.
"""
self.__belong = [0] * len(self.__pointer_data)
self.__clusters = [[] for i in range(len(medoids))]
for index_point in range(len(self.__pointer_data)):
index_optim = -1
dist_optim = 0.0
for index in range(len(medoids)):
dist = euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[medoids[index]])
if (dist < dist_optim) or (index is 0):
index_optim = index
dist_optim = dist
self.__clusters[index_optim].append(index_point)
self.__belong[index_point] = index_optim
# If cluster is not able to capture object it should be removed
self.__clusters = [cluster for cluster in self.__clusters if len(cluster) > 0]
def __optimize_configuration(self):
"""!
@brief Finds quasi-optimal medoids and updates in line with them clusters in line with algorithm's rules.
"""
index_neighbor = 0
counter = 0
while (index_neighbor < self.__maxneighbor):
# get random current medoid that is to be replaced
current_medoid_index = self.__current[random.randint(0, self.__number_clusters - 1)]
current_medoid_cluster_index = self.__belong[current_medoid_index]
# get new candidate to be medoid
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
while candidate_medoid_index in self.__current:
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
candidate_cost = 0.0
for point_index in range(0, len(self.__pointer_data)):
if point_index not in self.__current:
# get non-medoid point and its medoid
point_cluster_index = self.__belong[point_index]
point_medoid_index = self.__current[point_cluster_index]
# get other medoid that is nearest to the point (except current and candidate)
other_medoid_index = self.__find_another_nearest_medoid(point_index, current_medoid_index)
other_medoid_cluster_index = self.__belong[other_medoid_index]
# for optimization calculate all required distances
# from the point to current medoid
distance_current = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
# from the point to candidate median
distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[candidate_medoid_index])
# from the point to nearest (own) medoid
distance_nearest = float('inf')
if ( (point_medoid_index != candidate_medoid_index) and (point_medoid_index != current_medoid_cluster_index) ):
distance_nearest = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[point_medoid_index])
# apply rules for cost calculation
if (point_cluster_index == current_medoid_cluster_index):
# case 1:
if (distance_candidate >= distance_nearest):
candidate_cost += distance_nearest - distance_current
# case 2:
else:
candidate_cost += distance_candidate - distance_current
elif (point_cluster_index == other_medoid_cluster_index):
# case 3 ('nearest medoid' is the representative object of that cluster and object is more similar to 'nearest' than to 'candidate'):
if (distance_candidate > distance_nearest):
pass;
# case 4:
else:
candidate_cost += distance_candidate - distance_nearest
if (candidate_cost < 0):
counter+=1
# set candidate that has won
self.__current[current_medoid_cluster_index] = candidate_medoid_index
# recalculate clusters
self.__update_clusters(self.__current)
# reset iterations and starts investigation from the begining
index_neighbor = 0
else:
index_neighbor += 1
print("Medoid set changed {0} times".format(counter))
def __find_another_nearest_medoid(self, point_index, current_medoid_index):
|
def __calculate_estimation(self):
"""!
@brief Calculates estimation (cost) of the current clusters. The lower the estimation,
the more optimally configuration of clusters.
@return (double) estimation of current clusters.
"""
estimation = 0.0
for index_cluster in range(0, len(self.__clusters)):
cluster = self.__clusters[index_cluster]
index_medoid = self.__current[index_cluster]
for index_point in cluster:
estimation += euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[index_medoid])
return estimation
| """!
@brief Finds the another nearest medoid for the specified point that is different from the specified medoid.
@param[in] point_index: index of point in dataspace for that searching of medoid in current list of medoids is perfomed.
@param[in] current_medoid_index: index of medoid that shouldn't be considered as a nearest.
@return (uint) index of the another nearest medoid for the point.
"""
other_medoid_index = -1
other_distance_nearest = float('inf')
for index_medoid in self.__current:
if (index_medoid != current_medoid_index):
other_distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
if other_distance_candidate < other_distance_nearest:
other_distance_nearest = other_distance_candidate
other_medoid_index = index_medoid
return other_medoid_index | identifier_body |
clarans.py | """!
@brief Cluster analysis algorithm: CLARANS.
@details Implementation based on paper @cite article::clarans::1.
@authors Andrei Novikov (pyclustering@yandex.ru)
@date 2014-2019
@copyright GNU Public License
@cond GNU_PUBLIC_LICENSE
PyClustering is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
PyClustering is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
@endcond
"""
import random
from pyclustering.cluster.encoder import type_encoding
from pyclustering.utils import euclidean_distance_square
def plot_pam(data, cl, equal_axis_scale = False):
import matplotlib.pyplot as plt
import numpy as np
fig,ax = plt.subplots(figsize=(14,6))
plt.scatter(np.array(data)[:,0], np.array(data)[:,1], s=300, color="white", edgecolor="black")
colors = { 0:"seagreen", 1:'beige', 2:'yellow', 3:'grey',
4:'pink', 5:'turquoise', 6:'orange', 7:'purple', 8:'yellowgreen', 9:'olive', 10:'brown',
11:'tan', 12: 'plum', 13:'rosybrown', 14:'lightblue', 15:"khaki", 16:"gainsboro", 17:"peachpuff"}
for i,el in enumerate(list(cl.values())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=300, color=colors[i%17], edgecolor="black")
for i,el in enumerate(list(cl.keys())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=500, color="red", marker="X", edgecolor="black")
xmin, xmax, ymin, ymax = plt.axis()
xwidth = xmax - xmin
ywidth = ymax - ymin
xw1 = xwidth*0.01
yw1 = ywidth*0.01
xw2 = xwidth*0.005
yw2 = ywidth*0.01
xw3 = xwidth*0.01
yw3 = ywidth*0.01
for i, txt in enumerate(range(len(data))):
if len(str(txt))==2:
ax.annotate(txt, (np.array(data)[:,0][i]-xw1, np.array(data)[:,1][i]-yw1), fontsize=12, size=12)
elif len(str(txt))==1:
ax.annotate(txt, (np.array(data)[:,0][i]-xw2, np.array(data)[:,1][i]-yw2), fontsize=12, size=12)
else:
ax.annotate(txt, (np.array(data)[:,0][i]-xw3, np.array(data)[:,1][i]-yw3), fontsize=9, size=9)
if equal_axis_scale == True:
plt.gca().set_aspect('equal', adjustable='box')
plt.show()
class clarans:
"""!
@brief Class represents clustering algorithm CLARANS (a method for clustering objects for spatial data mining).
"""
def __init__(self, data, number_clusters, numlocal, maxneighbor):
"""!
@brief Constructor of clustering algorithm CLARANS.
@details The higher the value of maxneighbor, the closer is CLARANS to K-Medoids, and the longer is each search of a local minima.
@param[in] data (list): Input data that is presented as list of points (objects), each point should be represented by list or tuple.
@param[in] number_clusters (uint): Amount of clusters that should be allocated.
@param[in] numlocal (uint): The number of local minima obtained (amount of iterations for solving the problem).
@param[in] maxneighbor (uint): The maximum number of neighbors examined.
"""
self.__pointer_data = data
self.__numlocal = numlocal
self.__maxneighbor = maxneighbor
self.__number_clusters = number_clusters
self.__clusters = []
self.__current = []
self.__belong = []
self.__optimal_medoids = []
self.__optimal_estimation = float('inf')
self.__verify_arguments()
def __verify_arguments(self):
"""!
@brief Verify input parameters for the algorithm and throw exception in case of incorrectness.
"""
if len(self.__pointer_data) == 0:
raise ValueError("Input data is empty (size: '%d')." % len(self.__pointer_data))
if self.__number_clusters <= 0:
raise ValueError("Amount of cluster (current value: '%d') for allocation should be greater than 0." %
self.__number_clusters)
if self.__numlocal < 0:
raise ValueError("Local minima (current value: '%d') should be greater or equal to 0." % self.__numlocal)
if self.__maxneighbor < 0:
raise ValueError("Maximum number of neighbors (current value: '%d') should be greater or "
"equal to 0." % self.__maxneighbor)
def process(self, plotting=False):
"""!
@brief Performs cluster analysis in line with rules of CLARANS algorithm.
@return (clarans) Returns itself (CLARANS instance).
@see get_clusters()
@see get_medoids()
"""
random.seed()
# loop for a numlocal number of times
for _ in range(0, self.__numlocal):
print("numlocal: ", _)
# set (current) random medoids
self.__current = random.sample(range(0, len(self.__pointer_data)), self.__number_clusters)
# update clusters in line with random allocated medoids
self.__update_clusters(self.__current)
# optimize configuration
self.__optimize_configuration()
# obtain cost of current cluster configuration and compare it with the best obtained
estimation = self.__calculate_estimation()
if estimation < self.__optimal_estimation:
print("Better configuration found with medoids: {0} and cost: {1}".format(self.__current[:], estimation))
self.__optimal_medoids = self.__current[:]
self.__optimal_estimation = estimation
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
else:
print("Configuration found does not improve current best one because its cost is {0}".format(estimation))
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
self.__update_clusters(self.__optimal_medoids)
return self
def get_clusters(self):
"""!
@brief Returns allocated clusters by the algorithm.
@remark Allocated clusters can be returned only after data processing (use method process()), otherwise empty list is returned.
@return (list) List of allocated clusters, each cluster contains indexes of objects in list of data.
@see process()
@see get_medoids()
"""
return self.__clusters
def get_medoids(self):
"""!
@brief Returns list of medoids of allocated clusters.
@see process()
@see get_clusters()
"""
return self.__optimal_medoids
def get_cluster_encoding(self):
"""!
@brief Returns clustering result representation type that indicate how clusters are encoded.
@return (type_encoding) Clustering result representation.
@see get_clusters()
"""
return type_encoding.CLUSTER_INDEX_LIST_SEPARATION
def __update_clusters(self, medoids):
"""!
@brief Forms cluster in line with specified medoids by calculation distance from each point to medoids.
"""
self.__belong = [0] * len(self.__pointer_data)
self.__clusters = [[] for i in range(len(medoids))]
for index_point in range(len(self.__pointer_data)):
index_optim = -1
dist_optim = 0.0
for index in range(len(medoids)):
dist = euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[medoids[index]])
if (dist < dist_optim) or (index is 0):
index_optim = index
dist_optim = dist
self.__clusters[index_optim].append(index_point)
self.__belong[index_point] = index_optim
# If cluster is not able to capture object it should be removed
self.__clusters = [cluster for cluster in self.__clusters if len(cluster) > 0]
def __optimize_configuration(self):
"""!
@brief Finds quasi-optimal medoids and updates in line with them clusters in line with algorithm's rules.
"""
index_neighbor = 0
counter = 0
while (index_neighbor < self.__maxneighbor):
# get random current medoid that is to be replaced
current_medoid_index = self.__current[random.randint(0, self.__number_clusters - 1)]
current_medoid_cluster_index = self.__belong[current_medoid_index]
# get new candidate to be medoid
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
while candidate_medoid_index in self.__current:
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
candidate_cost = 0.0
for point_index in range(0, len(self.__pointer_data)):
if point_index not in self.__current:
# get non-medoid point and its medoid
point_cluster_index = self.__belong[point_index]
point_medoid_index = self.__current[point_cluster_index]
# get other medoid that is nearest to the point (except current and candidate)
other_medoid_index = self.__find_another_nearest_medoid(point_index, current_medoid_index)
other_medoid_cluster_index = self.__belong[other_medoid_index]
# for optimization calculate all required distances
# from the point to current medoid
distance_current = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
# from the point to candidate median
distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[candidate_medoid_index])
# from the point to nearest (own) medoid
distance_nearest = float('inf')
if ( (point_medoid_index != candidate_medoid_index) and (point_medoid_index != current_medoid_cluster_index) ):
distance_nearest = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[point_medoid_index])
# apply rules for cost calculation
if (point_cluster_index == current_medoid_cluster_index):
# case 1:
if (distance_candidate >= distance_nearest):
|
# case 2:
else:
candidate_cost += distance_candidate - distance_current
elif (point_cluster_index == other_medoid_cluster_index):
# case 3 ('nearest medoid' is the representative object of that cluster and object is more similar to 'nearest' than to 'candidate'):
if (distance_candidate > distance_nearest):
pass;
# case 4:
else:
candidate_cost += distance_candidate - distance_nearest
if (candidate_cost < 0):
counter+=1
# set candidate that has won
self.__current[current_medoid_cluster_index] = candidate_medoid_index
# recalculate clusters
self.__update_clusters(self.__current)
# reset iterations and starts investigation from the begining
index_neighbor = 0
else:
index_neighbor += 1
print("Medoid set changed {0} times".format(counter))
def __find_another_nearest_medoid(self, point_index, current_medoid_index):
"""!
@brief Finds the another nearest medoid for the specified point that is different from the specified medoid.
@param[in] point_index: index of point in dataspace for that searching of medoid in current list of medoids is perfomed.
@param[in] current_medoid_index: index of medoid that shouldn't be considered as a nearest.
@return (uint) index of the another nearest medoid for the point.
"""
other_medoid_index = -1
other_distance_nearest = float('inf')
for index_medoid in self.__current:
if (index_medoid != current_medoid_index):
other_distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
if other_distance_candidate < other_distance_nearest:
other_distance_nearest = other_distance_candidate
other_medoid_index = index_medoid
return other_medoid_index
def __calculate_estimation(self):
"""!
@brief Calculates estimation (cost) of the current clusters. The lower the estimation,
the more optimally configuration of clusters.
@return (double) estimation of current clusters.
"""
estimation = 0.0
for index_cluster in range(0, len(self.__clusters)):
cluster = self.__clusters[index_cluster]
index_medoid = self.__current[index_cluster]
for index_point in cluster:
estimation += euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[index_medoid])
return estimation
| candidate_cost += distance_nearest - distance_current | conditional_block |
clarans.py | """!
@brief Cluster analysis algorithm: CLARANS.
@details Implementation based on paper @cite article::clarans::1.
@authors Andrei Novikov (pyclustering@yandex.ru)
@date 2014-2019
@copyright GNU Public License
@cond GNU_PUBLIC_LICENSE
PyClustering is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
PyClustering is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
@endcond
"""
import random
from pyclustering.cluster.encoder import type_encoding
from pyclustering.utils import euclidean_distance_square
def plot_pam(data, cl, equal_axis_scale = False):
import matplotlib.pyplot as plt
import numpy as np
fig,ax = plt.subplots(figsize=(14,6))
plt.scatter(np.array(data)[:,0], np.array(data)[:,1], s=300, color="white", edgecolor="black")
colors = { 0:"seagreen", 1:'beige', 2:'yellow', 3:'grey',
4:'pink', 5:'turquoise', 6:'orange', 7:'purple', 8:'yellowgreen', 9:'olive', 10:'brown',
11:'tan', 12: 'plum', 13:'rosybrown', 14:'lightblue', 15:"khaki", 16:"gainsboro", 17:"peachpuff"}
for i,el in enumerate(list(cl.values())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=300, color=colors[i%17], edgecolor="black")
for i,el in enumerate(list(cl.keys())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=500, color="red", marker="X", edgecolor="black")
xmin, xmax, ymin, ymax = plt.axis()
xwidth = xmax - xmin
ywidth = ymax - ymin
xw1 = xwidth*0.01
yw1 = ywidth*0.01
xw2 = xwidth*0.005
yw2 = ywidth*0.01
xw3 = xwidth*0.01
yw3 = ywidth*0.01
for i, txt in enumerate(range(len(data))):
if len(str(txt))==2:
ax.annotate(txt, (np.array(data)[:,0][i]-xw1, np.array(data)[:,1][i]-yw1), fontsize=12, size=12)
elif len(str(txt))==1:
ax.annotate(txt, (np.array(data)[:,0][i]-xw2, np.array(data)[:,1][i]-yw2), fontsize=12, size=12)
else:
ax.annotate(txt, (np.array(data)[:,0][i]-xw3, np.array(data)[:,1][i]-yw3), fontsize=9, size=9)
if equal_axis_scale == True:
plt.gca().set_aspect('equal', adjustable='box')
plt.show()
class clarans:
"""!
@brief Class represents clustering algorithm CLARANS (a method for clustering objects for spatial data mining).
"""
def __init__(self, data, number_clusters, numlocal, maxneighbor):
"""!
@brief Constructor of clustering algorithm CLARANS.
@details The higher the value of maxneighbor, the closer is CLARANS to K-Medoids, and the longer is each search of a local minima.
@param[in] data (list): Input data that is presented as list of points (objects), each point should be represented by list or tuple.
@param[in] number_clusters (uint): Amount of clusters that should be allocated.
@param[in] numlocal (uint): The number of local minima obtained (amount of iterations for solving the problem).
@param[in] maxneighbor (uint): The maximum number of neighbors examined.
"""
self.__pointer_data = data
self.__numlocal = numlocal
self.__maxneighbor = maxneighbor
self.__number_clusters = number_clusters
self.__clusters = []
self.__current = []
self.__belong = []
self.__optimal_medoids = []
self.__optimal_estimation = float('inf')
self.__verify_arguments()
def __verify_arguments(self):
"""!
@brief Verify input parameters for the algorithm and throw exception in case of incorrectness.
"""
if len(self.__pointer_data) == 0:
raise ValueError("Input data is empty (size: '%d')." % len(self.__pointer_data))
if self.__number_clusters <= 0:
raise ValueError("Amount of cluster (current value: '%d') for allocation should be greater than 0." %
self.__number_clusters)
if self.__numlocal < 0:
raise ValueError("Local minima (current value: '%d') should be greater or equal to 0." % self.__numlocal)
if self.__maxneighbor < 0:
raise ValueError("Maximum number of neighbors (current value: '%d') should be greater or "
"equal to 0." % self.__maxneighbor)
def process(self, plotting=False):
"""!
@brief Performs cluster analysis in line with rules of CLARANS algorithm.
@return (clarans) Returns itself (CLARANS instance).
@see get_clusters()
@see get_medoids()
"""
random.seed()
# loop for a numlocal number of times
for _ in range(0, self.__numlocal):
print("numlocal: ", _)
# set (current) random medoids
self.__current = random.sample(range(0, len(self.__pointer_data)), self.__number_clusters)
# update clusters in line with random allocated medoids
self.__update_clusters(self.__current)
# optimize configuration
self.__optimize_configuration()
# obtain cost of current cluster configuration and compare it with the best obtained
estimation = self.__calculate_estimation()
if estimation < self.__optimal_estimation:
print("Better configuration found with medoids: {0} and cost: {1}".format(self.__current[:], estimation))
self.__optimal_medoids = self.__current[:]
self.__optimal_estimation = estimation
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
else:
print("Configuration found does not improve current best one because its cost is {0}".format(estimation))
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
self.__update_clusters(self.__optimal_medoids)
return self
def | (self):
"""!
@brief Returns allocated clusters by the algorithm.
@remark Allocated clusters can be returned only after data processing (use method process()), otherwise empty list is returned.
@return (list) List of allocated clusters, each cluster contains indexes of objects in list of data.
@see process()
@see get_medoids()
"""
return self.__clusters
def get_medoids(self):
"""!
@brief Returns list of medoids of allocated clusters.
@see process()
@see get_clusters()
"""
return self.__optimal_medoids
def get_cluster_encoding(self):
"""!
@brief Returns clustering result representation type that indicate how clusters are encoded.
@return (type_encoding) Clustering result representation.
@see get_clusters()
"""
return type_encoding.CLUSTER_INDEX_LIST_SEPARATION
def __update_clusters(self, medoids):
"""!
@brief Forms cluster in line with specified medoids by calculation distance from each point to medoids.
"""
self.__belong = [0] * len(self.__pointer_data)
self.__clusters = [[] for i in range(len(medoids))]
for index_point in range(len(self.__pointer_data)):
index_optim = -1
dist_optim = 0.0
for index in range(len(medoids)):
dist = euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[medoids[index]])
if (dist < dist_optim) or (index is 0):
index_optim = index
dist_optim = dist
self.__clusters[index_optim].append(index_point)
self.__belong[index_point] = index_optim
# If cluster is not able to capture object it should be removed
self.__clusters = [cluster for cluster in self.__clusters if len(cluster) > 0]
def __optimize_configuration(self):
"""!
@brief Finds quasi-optimal medoids and updates in line with them clusters in line with algorithm's rules.
"""
index_neighbor = 0
counter = 0
while (index_neighbor < self.__maxneighbor):
# get random current medoid that is to be replaced
current_medoid_index = self.__current[random.randint(0, self.__number_clusters - 1)]
current_medoid_cluster_index = self.__belong[current_medoid_index]
# get new candidate to be medoid
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
while candidate_medoid_index in self.__current:
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
candidate_cost = 0.0
for point_index in range(0, len(self.__pointer_data)):
if point_index not in self.__current:
# get non-medoid point and its medoid
point_cluster_index = self.__belong[point_index]
point_medoid_index = self.__current[point_cluster_index]
# get other medoid that is nearest to the point (except current and candidate)
other_medoid_index = self.__find_another_nearest_medoid(point_index, current_medoid_index)
other_medoid_cluster_index = self.__belong[other_medoid_index]
# for optimization calculate all required distances
# from the point to current medoid
distance_current = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
# from the point to candidate median
distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[candidate_medoid_index])
# from the point to nearest (own) medoid
distance_nearest = float('inf')
if ( (point_medoid_index != candidate_medoid_index) and (point_medoid_index != current_medoid_cluster_index) ):
distance_nearest = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[point_medoid_index])
# apply rules for cost calculation
if (point_cluster_index == current_medoid_cluster_index):
# case 1:
if (distance_candidate >= distance_nearest):
candidate_cost += distance_nearest - distance_current
# case 2:
else:
candidate_cost += distance_candidate - distance_current
elif (point_cluster_index == other_medoid_cluster_index):
# case 3 ('nearest medoid' is the representative object of that cluster and object is more similar to 'nearest' than to 'candidate'):
if (distance_candidate > distance_nearest):
pass;
# case 4:
else:
candidate_cost += distance_candidate - distance_nearest
if (candidate_cost < 0):
counter+=1
# set candidate that has won
self.__current[current_medoid_cluster_index] = candidate_medoid_index
# recalculate clusters
self.__update_clusters(self.__current)
# reset iterations and starts investigation from the begining
index_neighbor = 0
else:
index_neighbor += 1
print("Medoid set changed {0} times".format(counter))
def __find_another_nearest_medoid(self, point_index, current_medoid_index):
"""!
@brief Finds the another nearest medoid for the specified point that is different from the specified medoid.
@param[in] point_index: index of point in dataspace for that searching of medoid in current list of medoids is perfomed.
@param[in] current_medoid_index: index of medoid that shouldn't be considered as a nearest.
@return (uint) index of the another nearest medoid for the point.
"""
other_medoid_index = -1
other_distance_nearest = float('inf')
for index_medoid in self.__current:
if (index_medoid != current_medoid_index):
other_distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
if other_distance_candidate < other_distance_nearest:
other_distance_nearest = other_distance_candidate
other_medoid_index = index_medoid
return other_medoid_index
def __calculate_estimation(self):
"""!
@brief Calculates estimation (cost) of the current clusters. The lower the estimation,
the more optimally configuration of clusters.
@return (double) estimation of current clusters.
"""
estimation = 0.0
for index_cluster in range(0, len(self.__clusters)):
cluster = self.__clusters[index_cluster]
index_medoid = self.__current[index_cluster]
for index_point in cluster:
estimation += euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[index_medoid])
return estimation
| get_clusters | identifier_name |
clarans.py | """!
@brief Cluster analysis algorithm: CLARANS.
@details Implementation based on paper @cite article::clarans::1.
@authors Andrei Novikov (pyclustering@yandex.ru)
@date 2014-2019
@copyright GNU Public License
@cond GNU_PUBLIC_LICENSE
PyClustering is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
PyClustering is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
@endcond
"""
import random
from pyclustering.cluster.encoder import type_encoding
from pyclustering.utils import euclidean_distance_square
def plot_pam(data, cl, equal_axis_scale = False):
import matplotlib.pyplot as plt
import numpy as np
fig,ax = plt.subplots(figsize=(14,6))
plt.scatter(np.array(data)[:,0], np.array(data)[:,1], s=300, color="white", edgecolor="black")
colors = { 0:"seagreen", 1:'beige', 2:'yellow', 3:'grey',
4:'pink', 5:'turquoise', 6:'orange', 7:'purple', 8:'yellowgreen', 9:'olive', 10:'brown',
11:'tan', 12: 'plum', 13:'rosybrown', 14:'lightblue', 15:"khaki", 16:"gainsboro", 17:"peachpuff"}
for i,el in enumerate(list(cl.values())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=300, color=colors[i%17], edgecolor="black")
for i,el in enumerate(list(cl.keys())):
plt.scatter(np.array(data)[el,0], np.array(data)[el,1], s=500, color="red", marker="X", edgecolor="black")
xmin, xmax, ymin, ymax = plt.axis()
xwidth = xmax - xmin
ywidth = ymax - ymin
xw1 = xwidth*0.01
yw1 = ywidth*0.01
xw2 = xwidth*0.005
yw2 = ywidth*0.01
xw3 = xwidth*0.01
yw3 = ywidth*0.01
for i, txt in enumerate(range(len(data))):
if len(str(txt))==2:
ax.annotate(txt, (np.array(data)[:,0][i]-xw1, np.array(data)[:,1][i]-yw1), fontsize=12, size=12)
elif len(str(txt))==1:
ax.annotate(txt, (np.array(data)[:,0][i]-xw2, np.array(data)[:,1][i]-yw2), fontsize=12, size=12)
else:
ax.annotate(txt, (np.array(data)[:,0][i]-xw3, np.array(data)[:,1][i]-yw3), fontsize=9, size=9)
if equal_axis_scale == True:
plt.gca().set_aspect('equal', adjustable='box')
plt.show()
class clarans:
"""!
@brief Class represents clustering algorithm CLARANS (a method for clustering objects for spatial data mining).
"""
def __init__(self, data, number_clusters, numlocal, maxneighbor):
"""!
@brief Constructor of clustering algorithm CLARANS.
@details The higher the value of maxneighbor, the closer is CLARANS to K-Medoids, and the longer is each search of a local minima.
@param[in] data (list): Input data that is presented as list of points (objects), each point should be represented by list or tuple.
@param[in] number_clusters (uint): Amount of clusters that should be allocated.
@param[in] numlocal (uint): The number of local minima obtained (amount of iterations for solving the problem).
@param[in] maxneighbor (uint): The maximum number of neighbors examined.
"""
self.__pointer_data = data
self.__numlocal = numlocal
self.__maxneighbor = maxneighbor
self.__number_clusters = number_clusters
self.__clusters = []
self.__current = []
self.__belong = []
self.__optimal_medoids = []
self.__optimal_estimation = float('inf')
self.__verify_arguments()
def __verify_arguments(self):
"""!
@brief Verify input parameters for the algorithm and throw exception in case of incorrectness.
"""
if len(self.__pointer_data) == 0:
raise ValueError("Input data is empty (size: '%d')." % len(self.__pointer_data))
if self.__number_clusters <= 0:
raise ValueError("Amount of cluster (current value: '%d') for allocation should be greater than 0." %
self.__number_clusters)
if self.__numlocal < 0:
raise ValueError("Local minima (current value: '%d') should be greater or equal to 0." % self.__numlocal)
if self.__maxneighbor < 0:
raise ValueError("Maximum number of neighbors (current value: '%d') should be greater or "
"equal to 0." % self.__maxneighbor)
|
@see get_clusters()
@see get_medoids()
"""
random.seed()
# loop for a numlocal number of times
for _ in range(0, self.__numlocal):
print("numlocal: ", _)
# set (current) random medoids
self.__current = random.sample(range(0, len(self.__pointer_data)), self.__number_clusters)
# update clusters in line with random allocated medoids
self.__update_clusters(self.__current)
# optimize configuration
self.__optimize_configuration()
# obtain cost of current cluster configuration and compare it with the best obtained
estimation = self.__calculate_estimation()
if estimation < self.__optimal_estimation:
print("Better configuration found with medoids: {0} and cost: {1}".format(self.__current[:], estimation))
self.__optimal_medoids = self.__current[:]
self.__optimal_estimation = estimation
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
else:
print("Configuration found does not improve current best one because its cost is {0}".format(estimation))
if plotting == True:
self.__update_clusters(self.__optimal_medoids)
plot_pam(self.__pointer_data,
dict(zip(self.__optimal_medoids,self.__clusters)))
self.__update_clusters(self.__optimal_medoids)
return self
def get_clusters(self):
"""!
@brief Returns allocated clusters by the algorithm.
@remark Allocated clusters can be returned only after data processing (use method process()), otherwise empty list is returned.
@return (list) List of allocated clusters, each cluster contains indexes of objects in list of data.
@see process()
@see get_medoids()
"""
return self.__clusters
def get_medoids(self):
"""!
@brief Returns list of medoids of allocated clusters.
@see process()
@see get_clusters()
"""
return self.__optimal_medoids
def get_cluster_encoding(self):
"""!
@brief Returns clustering result representation type that indicate how clusters are encoded.
@return (type_encoding) Clustering result representation.
@see get_clusters()
"""
return type_encoding.CLUSTER_INDEX_LIST_SEPARATION
def __update_clusters(self, medoids):
"""!
@brief Forms cluster in line with specified medoids by calculation distance from each point to medoids.
"""
self.__belong = [0] * len(self.__pointer_data)
self.__clusters = [[] for i in range(len(medoids))]
for index_point in range(len(self.__pointer_data)):
index_optim = -1
dist_optim = 0.0
for index in range(len(medoids)):
dist = euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[medoids[index]])
if (dist < dist_optim) or (index is 0):
index_optim = index
dist_optim = dist
self.__clusters[index_optim].append(index_point)
self.__belong[index_point] = index_optim
# If cluster is not able to capture object it should be removed
self.__clusters = [cluster for cluster in self.__clusters if len(cluster) > 0]
def __optimize_configuration(self):
"""!
@brief Finds quasi-optimal medoids and updates in line with them clusters in line with algorithm's rules.
"""
index_neighbor = 0
counter = 0
while (index_neighbor < self.__maxneighbor):
# get random current medoid that is to be replaced
current_medoid_index = self.__current[random.randint(0, self.__number_clusters - 1)]
current_medoid_cluster_index = self.__belong[current_medoid_index]
# get new candidate to be medoid
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
while candidate_medoid_index in self.__current:
candidate_medoid_index = random.randint(0, len(self.__pointer_data) - 1)
candidate_cost = 0.0
for point_index in range(0, len(self.__pointer_data)):
if point_index not in self.__current:
# get non-medoid point and its medoid
point_cluster_index = self.__belong[point_index]
point_medoid_index = self.__current[point_cluster_index]
# get other medoid that is nearest to the point (except current and candidate)
other_medoid_index = self.__find_another_nearest_medoid(point_index, current_medoid_index)
other_medoid_cluster_index = self.__belong[other_medoid_index]
# for optimization calculate all required distances
# from the point to current medoid
distance_current = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
# from the point to candidate median
distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[candidate_medoid_index])
# from the point to nearest (own) medoid
distance_nearest = float('inf')
if ( (point_medoid_index != candidate_medoid_index) and (point_medoid_index != current_medoid_cluster_index) ):
distance_nearest = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[point_medoid_index])
# apply rules for cost calculation
if (point_cluster_index == current_medoid_cluster_index):
# case 1:
if (distance_candidate >= distance_nearest):
candidate_cost += distance_nearest - distance_current
# case 2:
else:
candidate_cost += distance_candidate - distance_current
elif (point_cluster_index == other_medoid_cluster_index):
# case 3 ('nearest medoid' is the representative object of that cluster and object is more similar to 'nearest' than to 'candidate'):
if (distance_candidate > distance_nearest):
pass;
# case 4:
else:
candidate_cost += distance_candidate - distance_nearest
if (candidate_cost < 0):
counter+=1
# set candidate that has won
self.__current[current_medoid_cluster_index] = candidate_medoid_index
# recalculate clusters
self.__update_clusters(self.__current)
# reset iterations and starts investigation from the begining
index_neighbor = 0
else:
index_neighbor += 1
print("Medoid set changed {0} times".format(counter))
def __find_another_nearest_medoid(self, point_index, current_medoid_index):
"""!
@brief Finds the another nearest medoid for the specified point that is different from the specified medoid.
@param[in] point_index: index of point in dataspace for that searching of medoid in current list of medoids is perfomed.
@param[in] current_medoid_index: index of medoid that shouldn't be considered as a nearest.
@return (uint) index of the another nearest medoid for the point.
"""
other_medoid_index = -1
other_distance_nearest = float('inf')
for index_medoid in self.__current:
if (index_medoid != current_medoid_index):
other_distance_candidate = euclidean_distance_square(self.__pointer_data[point_index], self.__pointer_data[current_medoid_index])
if other_distance_candidate < other_distance_nearest:
other_distance_nearest = other_distance_candidate
other_medoid_index = index_medoid
return other_medoid_index
def __calculate_estimation(self):
"""!
@brief Calculates estimation (cost) of the current clusters. The lower the estimation,
the more optimally configuration of clusters.
@return (double) estimation of current clusters.
"""
estimation = 0.0
for index_cluster in range(0, len(self.__clusters)):
cluster = self.__clusters[index_cluster]
index_medoid = self.__current[index_cluster]
for index_point in cluster:
estimation += euclidean_distance_square(self.__pointer_data[index_point], self.__pointer_data[index_medoid])
return estimation | def process(self, plotting=False):
"""!
@brief Performs cluster analysis in line with rules of CLARANS algorithm.
@return (clarans) Returns itself (CLARANS instance).
| random_line_split |
client.rs | use std::thread;
use std::net::SocketAddr;
use std::net::ToSocketAddrs;
use std::sync::mpsc;
use std::sync::Arc;
use futures;
use futures::Future;
use futures::stream::Stream;
use tokio_core;
use tokio_core::reactor;
use solicit::http::HttpScheme;
use solicit::http::HttpError;
use solicit::http::Header;
use solicit::http::StaticHeader;
use method::MethodDescriptor;
use error::*;
use result::*;
use futures_misc::*;
use futures_grpc::*;
use grpc::*;
use http_client::*;
use solicit_misc::*;
use assert_types::*;
trait GrpcResponseHandlerTrait : Send + 'static + HttpClientResponseHandler {
}
struct GrpcResponseHandlerTyped<Req : Send + 'static, Resp : Send + 'static> {
method: Arc<MethodDescriptor<Req, Resp>>,
complete: tokio_core::channel::Sender<ResultOrEof<Resp, GrpcError>>,
remaining_response: Vec<u8>,
}
impl<Req : Send + 'static, Resp : Send + 'static> GrpcResponseHandlerTrait for GrpcResponseHandlerTyped<Req, Resp> {
}
impl<Req : Send + 'static, Resp : Send + 'static> HttpClientResponseHandler for GrpcResponseHandlerTyped<Req, Resp> {
fn headers(&mut self, headers: Vec<StaticHeader>) -> bool {
println!("client: received headers");
if slice_get_header(&headers, ":status") != Some("200") {
if let Some(message) = slice_get_header(&headers, HEADER_GRPC_MESSAGE) {
self.complete.send(ResultOrEof::Error(GrpcError::GrpcMessage(GrpcMessageError { grpc_message: message.to_owned() }))).unwrap();
} else {
self.complete.send(ResultOrEof::Error(GrpcError::Other("not 200"))).unwrap();
}
false
} else {
true
}
}
fn data_frame(&mut self, chunk: Vec<u8>) -> bool {
self.remaining_response.extend(&chunk);
loop {
let len = match parse_grpc_frame(&self.remaining_response) {
Err(e) => {
self.complete.send(ResultOrEof::Error(e)).unwrap();
return false;
}
Ok(None) => break,
Ok(Some((message, len))) => {
let resp = self.method.resp_marshaller.read(&message);
self.complete.send(From::from(resp)).ok();
len
}
};
self.remaining_response.drain(..len);
}
true
}
fn trailers(&mut self, headers: Vec<StaticHeader>) -> bool {
let _status_200 = slice_get_header(&headers, ":status") == Some("200");
let grpc_status_0 = slice_get_header(&headers, HEADER_GRPC_STATUS) == Some("0");
if /* status_200 && */ grpc_status_0 {
true
} else {
if let Some(message) = slice_get_header(&headers, HEADER_GRPC_MESSAGE) {
self.complete.send(ResultOrEof::Error(GrpcError::GrpcMessage(GrpcMessageError { grpc_message: message.to_owned() }))).unwrap();
} else {
self.complete.send(ResultOrEof::Error(GrpcError::Other("not xxx"))).unwrap();
}
false
}
}
fn end(&mut self) {
self.complete.send(ResultOrEof::Eof).unwrap();
}
}
struct GrpcResponseHandler {
tr: Box<GrpcResponseHandlerTrait>,
}
impl HttpClientResponseHandler for GrpcResponseHandler {
fn headers(&mut self, headers: Vec<StaticHeader>) -> bool {
self.tr.headers(headers) |
fn data_frame(&mut self, chunk: Vec<u8>) -> bool {
self.tr.data_frame(chunk)
}
fn trailers(&mut self, headers: Vec<StaticHeader>) -> bool {
self.tr.trailers(headers)
}
fn end(&mut self) {
self.tr.end()
}
}
// Data sent from event loop to GrpcClient
struct LoopToClient {
// used only once to send shutdown signal
shutdown_tx: tokio_core::channel::Sender<()>,
loop_handle: reactor::Remote,
http_conn: Arc<HttpClientConnectionAsync<GrpcResponseHandler>>,
}
fn _assert_loop_to_client() {
assert_send::<reactor::Remote>();
assert_send::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_send::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_sync::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_send::<Arc<HttpClientConnectionAsync<GrpcResponseHandler>>>();
assert_send::<tokio_core::channel::Sender<()>>();
assert_send::<LoopToClient>();
}
/// gRPC client implementation.
/// Used by generated code.
pub struct GrpcClient {
loop_to_client: LoopToClient,
thread_join_handle: Option<thread::JoinHandle<()>>,
host: String,
http_scheme: HttpScheme,
}
impl GrpcClient {
/// Create a client connected to specified host and port.
pub fn new(host: &str, port: u16) -> GrpcResult<GrpcClient> {
// TODO: sync
// TODO: try connect to all addrs
let socket_addr = try!((host, port).to_socket_addrs()).next().unwrap();
// We need some data back from event loop.
// This channel is used to exchange that data
let (get_from_loop_tx, get_from_loop_rx) = mpsc::channel();
// Start event loop.
let join_handle = thread::spawn(move || {
run_client_event_loop(socket_addr, get_from_loop_tx);
});
// Get back call channel and shutdown channel.
let loop_to_client = try!(get_from_loop_rx.recv()
.map_err(|_| GrpcError::Other("get response from loop")));
Ok(GrpcClient {
loop_to_client: loop_to_client,
thread_join_handle: Some(join_handle),
host: host.to_owned(),
http_scheme: HttpScheme::Http,
})
}
pub fn new_resp_channel<Resp : Send + 'static>(&self)
-> futures::Oneshot<(tokio_core::channel::Sender<ResultOrEof<Resp, GrpcError>>, GrpcStreamSend<Resp>)>
{
let (one_sender, one_receiver) = futures::oneshot();
self.loop_to_client.loop_handle.spawn(move |handle| {
let (sender, receiver) = tokio_core::channel::channel(&handle).unwrap();
let receiver: GrpcStreamSend<ResultOrEof<Resp, GrpcError>> = Box::new(receiver.map_err(GrpcError::from));
let receiver: GrpcStreamSend<Resp> = Box::new(stream_with_eof_and_error(receiver));
one_sender.complete((sender, receiver));
futures::finished(())
});
one_receiver
}
pub fn call_impl<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
let host = self.host.clone();
let http_scheme = self.http_scheme.clone();
let http_conn = self.loop_to_client.http_conn.clone();
// A channel to send response back to caller
let future = self.new_resp_channel().map_err(GrpcError::from).and_then(move |(complete, receiver)| {
let headers = vec![
Header::new(":method", "POST"),
Header::new(":path", method.name.clone()),
Header::new(":authority", host.clone()),
Header::new(":scheme", http_scheme.as_bytes()),
];
let request_frames = {
let method = method.clone();
req
.and_then(move |req| {
let grpc_frame = try!(method.req_marshaller.write(&req));
Ok(write_grpc_frame_to_vec(&grpc_frame))
})
.map_err(|e| HttpError::Other(Box::new(e)))
};
let start_request = http_conn.start_request(
headers,
Box::new(request_frames),
GrpcResponseHandler {
tr: Box::new(GrpcResponseHandlerTyped {
method: method.clone(),
complete: complete,
remaining_response: Vec::new(),
}),
}
).map_err(GrpcError::from);
let receiver: GrpcStreamSend<Resp> = receiver;
start_request.map(move |()| receiver)
});
let s: GrpcStreamSend<Resp> = future_flatten_to_stream(future);
s
}
pub fn call_unary<Req : Send + 'static, Resp : Send + 'static>(&self, req: Req, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcFutureSend<Resp>
{
stream_single_send(self.call_impl(Box::new(stream_once_send(req)), method))
}
pub fn call_server_streaming<Req : Send + 'static, Resp : Send + 'static>(&self, req: Req, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
self.call_impl(stream_once_send(req).boxed(), method)
}
pub fn call_client_streaming<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcFutureSend<Resp>
{
stream_single_send(self.call_impl(req, method))
}
pub fn call_bidi<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
self.call_impl(req, method)
}
}
// We shutdown client in destructor.
impl Drop for GrpcClient {
fn drop(&mut self) {
// ignore error because even loop may be already dead
self.loop_to_client.shutdown_tx.send(()).ok();
// do not ignore errors because we own event loop thread
self.thread_join_handle.take().expect("handle.take")
.join().expect("join thread");
}
}
// Event loop entry point
fn run_client_event_loop(
socket_addr: SocketAddr,
send_to_back: mpsc::Sender<LoopToClient>)
{
// Create an event loop.
let mut lp = reactor::Core::new().unwrap();
// Create a channel to receive shutdown signal.
let (shutdown_tx, shutdown_rx) = tokio_core::channel::channel(&lp.handle()).unwrap();
let (http_conn, http_conn_future) = HttpClientConnectionAsync::new(lp.handle(), &socket_addr);
let http_conn_future: GrpcFuture<_> = Box::new(http_conn_future.map_err(GrpcError::from));
// Send channels back to GrpcClient
send_to_back
.send(LoopToClient {
shutdown_tx: shutdown_tx,
loop_handle: lp.remote(),
http_conn: Arc::new(http_conn),
})
.expect("send back");
let shutdown = shutdown_rx.into_future().map_err(|(e, _)| GrpcError::from(e)).and_then(move |_| {
// Must complete with error,
// so `join` with this future cancels another future.
futures::failed::<(), _>(GrpcError::Other("shutdown"))
});
// Wait for either completion of connection (i. e. error)
// or shutdown signal.
let done = http_conn_future.join(shutdown);
// TODO: do not ignore error
lp.run(done).ok();
} | } | random_line_split |
client.rs | use std::thread;
use std::net::SocketAddr;
use std::net::ToSocketAddrs;
use std::sync::mpsc;
use std::sync::Arc;
use futures;
use futures::Future;
use futures::stream::Stream;
use tokio_core;
use tokio_core::reactor;
use solicit::http::HttpScheme;
use solicit::http::HttpError;
use solicit::http::Header;
use solicit::http::StaticHeader;
use method::MethodDescriptor;
use error::*;
use result::*;
use futures_misc::*;
use futures_grpc::*;
use grpc::*;
use http_client::*;
use solicit_misc::*;
use assert_types::*;
trait GrpcResponseHandlerTrait : Send + 'static + HttpClientResponseHandler {
}
struct GrpcResponseHandlerTyped<Req : Send + 'static, Resp : Send + 'static> {
method: Arc<MethodDescriptor<Req, Resp>>,
complete: tokio_core::channel::Sender<ResultOrEof<Resp, GrpcError>>,
remaining_response: Vec<u8>,
}
impl<Req : Send + 'static, Resp : Send + 'static> GrpcResponseHandlerTrait for GrpcResponseHandlerTyped<Req, Resp> {
}
impl<Req : Send + 'static, Resp : Send + 'static> HttpClientResponseHandler for GrpcResponseHandlerTyped<Req, Resp> {
fn headers(&mut self, headers: Vec<StaticHeader>) -> bool {
println!("client: received headers");
if slice_get_header(&headers, ":status") != Some("200") {
if let Some(message) = slice_get_header(&headers, HEADER_GRPC_MESSAGE) {
self.complete.send(ResultOrEof::Error(GrpcError::GrpcMessage(GrpcMessageError { grpc_message: message.to_owned() }))).unwrap();
} else {
self.complete.send(ResultOrEof::Error(GrpcError::Other("not 200"))).unwrap();
}
false
} else {
true
}
}
fn data_frame(&mut self, chunk: Vec<u8>) -> bool {
self.remaining_response.extend(&chunk);
loop {
let len = match parse_grpc_frame(&self.remaining_response) {
Err(e) => {
self.complete.send(ResultOrEof::Error(e)).unwrap();
return false;
}
Ok(None) => break,
Ok(Some((message, len))) => {
let resp = self.method.resp_marshaller.read(&message);
self.complete.send(From::from(resp)).ok();
len
}
};
self.remaining_response.drain(..len);
}
true
}
fn trailers(&mut self, headers: Vec<StaticHeader>) -> bool {
let _status_200 = slice_get_header(&headers, ":status") == Some("200");
let grpc_status_0 = slice_get_header(&headers, HEADER_GRPC_STATUS) == Some("0");
if /* status_200 && */ grpc_status_0 {
true
} else {
if let Some(message) = slice_get_header(&headers, HEADER_GRPC_MESSAGE) {
self.complete.send(ResultOrEof::Error(GrpcError::GrpcMessage(GrpcMessageError { grpc_message: message.to_owned() }))).unwrap();
} else {
self.complete.send(ResultOrEof::Error(GrpcError::Other("not xxx"))).unwrap();
}
false
}
}
fn end(&mut self) {
self.complete.send(ResultOrEof::Eof).unwrap();
}
}
struct GrpcResponseHandler {
tr: Box<GrpcResponseHandlerTrait>,
}
impl HttpClientResponseHandler for GrpcResponseHandler {
fn headers(&mut self, headers: Vec<StaticHeader>) -> bool {
self.tr.headers(headers)
}
fn data_frame(&mut self, chunk: Vec<u8>) -> bool {
self.tr.data_frame(chunk)
}
fn trailers(&mut self, headers: Vec<StaticHeader>) -> bool {
self.tr.trailers(headers)
}
fn end(&mut self) {
self.tr.end()
}
}
// Data sent from event loop to GrpcClient
struct LoopToClient {
// used only once to send shutdown signal
shutdown_tx: tokio_core::channel::Sender<()>,
loop_handle: reactor::Remote,
http_conn: Arc<HttpClientConnectionAsync<GrpcResponseHandler>>,
}
fn _assert_loop_to_client() {
assert_send::<reactor::Remote>();
assert_send::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_send::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_sync::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_send::<Arc<HttpClientConnectionAsync<GrpcResponseHandler>>>();
assert_send::<tokio_core::channel::Sender<()>>();
assert_send::<LoopToClient>();
}
/// gRPC client implementation.
/// Used by generated code.
pub struct GrpcClient {
loop_to_client: LoopToClient,
thread_join_handle: Option<thread::JoinHandle<()>>,
host: String,
http_scheme: HttpScheme,
}
impl GrpcClient {
/// Create a client connected to specified host and port.
pub fn new(host: &str, port: u16) -> GrpcResult<GrpcClient> {
// TODO: sync
// TODO: try connect to all addrs
let socket_addr = try!((host, port).to_socket_addrs()).next().unwrap();
// We need some data back from event loop.
// This channel is used to exchange that data
let (get_from_loop_tx, get_from_loop_rx) = mpsc::channel();
// Start event loop.
let join_handle = thread::spawn(move || {
run_client_event_loop(socket_addr, get_from_loop_tx);
});
// Get back call channel and shutdown channel.
let loop_to_client = try!(get_from_loop_rx.recv()
.map_err(|_| GrpcError::Other("get response from loop")));
Ok(GrpcClient {
loop_to_client: loop_to_client,
thread_join_handle: Some(join_handle),
host: host.to_owned(),
http_scheme: HttpScheme::Http,
})
}
pub fn new_resp_channel<Resp : Send + 'static>(&self)
-> futures::Oneshot<(tokio_core::channel::Sender<ResultOrEof<Resp, GrpcError>>, GrpcStreamSend<Resp>)>
{
let (one_sender, one_receiver) = futures::oneshot();
self.loop_to_client.loop_handle.spawn(move |handle| {
let (sender, receiver) = tokio_core::channel::channel(&handle).unwrap();
let receiver: GrpcStreamSend<ResultOrEof<Resp, GrpcError>> = Box::new(receiver.map_err(GrpcError::from));
let receiver: GrpcStreamSend<Resp> = Box::new(stream_with_eof_and_error(receiver));
one_sender.complete((sender, receiver));
futures::finished(())
});
one_receiver
}
pub fn call_impl<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
let host = self.host.clone();
let http_scheme = self.http_scheme.clone();
let http_conn = self.loop_to_client.http_conn.clone();
// A channel to send response back to caller
let future = self.new_resp_channel().map_err(GrpcError::from).and_then(move |(complete, receiver)| {
let headers = vec![
Header::new(":method", "POST"),
Header::new(":path", method.name.clone()),
Header::new(":authority", host.clone()),
Header::new(":scheme", http_scheme.as_bytes()),
];
let request_frames = {
let method = method.clone();
req
.and_then(move |req| {
let grpc_frame = try!(method.req_marshaller.write(&req));
Ok(write_grpc_frame_to_vec(&grpc_frame))
})
.map_err(|e| HttpError::Other(Box::new(e)))
};
let start_request = http_conn.start_request(
headers,
Box::new(request_frames),
GrpcResponseHandler {
tr: Box::new(GrpcResponseHandlerTyped {
method: method.clone(),
complete: complete,
remaining_response: Vec::new(),
}),
}
).map_err(GrpcError::from);
let receiver: GrpcStreamSend<Resp> = receiver;
start_request.map(move |()| receiver)
});
let s: GrpcStreamSend<Resp> = future_flatten_to_stream(future);
s
}
pub fn call_unary<Req : Send + 'static, Resp : Send + 'static>(&self, req: Req, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcFutureSend<Resp>
{
stream_single_send(self.call_impl(Box::new(stream_once_send(req)), method))
}
pub fn call_server_streaming<Req : Send + 'static, Resp : Send + 'static>(&self, req: Req, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
self.call_impl(stream_once_send(req).boxed(), method)
}
pub fn | <Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcFutureSend<Resp>
{
stream_single_send(self.call_impl(req, method))
}
pub fn call_bidi<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
self.call_impl(req, method)
}
}
// We shutdown client in destructor.
impl Drop for GrpcClient {
fn drop(&mut self) {
// ignore error because even loop may be already dead
self.loop_to_client.shutdown_tx.send(()).ok();
// do not ignore errors because we own event loop thread
self.thread_join_handle.take().expect("handle.take")
.join().expect("join thread");
}
}
// Event loop entry point
fn run_client_event_loop(
socket_addr: SocketAddr,
send_to_back: mpsc::Sender<LoopToClient>)
{
// Create an event loop.
let mut lp = reactor::Core::new().unwrap();
// Create a channel to receive shutdown signal.
let (shutdown_tx, shutdown_rx) = tokio_core::channel::channel(&lp.handle()).unwrap();
let (http_conn, http_conn_future) = HttpClientConnectionAsync::new(lp.handle(), &socket_addr);
let http_conn_future: GrpcFuture<_> = Box::new(http_conn_future.map_err(GrpcError::from));
// Send channels back to GrpcClient
send_to_back
.send(LoopToClient {
shutdown_tx: shutdown_tx,
loop_handle: lp.remote(),
http_conn: Arc::new(http_conn),
})
.expect("send back");
let shutdown = shutdown_rx.into_future().map_err(|(e, _)| GrpcError::from(e)).and_then(move |_| {
// Must complete with error,
// so `join` with this future cancels another future.
futures::failed::<(), _>(GrpcError::Other("shutdown"))
});
// Wait for either completion of connection (i. e. error)
// or shutdown signal.
let done = http_conn_future.join(shutdown);
// TODO: do not ignore error
lp.run(done).ok();
}
| call_client_streaming | identifier_name |
client.rs | use std::thread;
use std::net::SocketAddr;
use std::net::ToSocketAddrs;
use std::sync::mpsc;
use std::sync::Arc;
use futures;
use futures::Future;
use futures::stream::Stream;
use tokio_core;
use tokio_core::reactor;
use solicit::http::HttpScheme;
use solicit::http::HttpError;
use solicit::http::Header;
use solicit::http::StaticHeader;
use method::MethodDescriptor;
use error::*;
use result::*;
use futures_misc::*;
use futures_grpc::*;
use grpc::*;
use http_client::*;
use solicit_misc::*;
use assert_types::*;
trait GrpcResponseHandlerTrait : Send + 'static + HttpClientResponseHandler {
}
struct GrpcResponseHandlerTyped<Req : Send + 'static, Resp : Send + 'static> {
method: Arc<MethodDescriptor<Req, Resp>>,
complete: tokio_core::channel::Sender<ResultOrEof<Resp, GrpcError>>,
remaining_response: Vec<u8>,
}
impl<Req : Send + 'static, Resp : Send + 'static> GrpcResponseHandlerTrait for GrpcResponseHandlerTyped<Req, Resp> {
}
impl<Req : Send + 'static, Resp : Send + 'static> HttpClientResponseHandler for GrpcResponseHandlerTyped<Req, Resp> {
fn headers(&mut self, headers: Vec<StaticHeader>) -> bool {
println!("client: received headers");
if slice_get_header(&headers, ":status") != Some("200") {
if let Some(message) = slice_get_header(&headers, HEADER_GRPC_MESSAGE) {
self.complete.send(ResultOrEof::Error(GrpcError::GrpcMessage(GrpcMessageError { grpc_message: message.to_owned() }))).unwrap();
} else {
self.complete.send(ResultOrEof::Error(GrpcError::Other("not 200"))).unwrap();
}
false
} else {
true
}
}
fn data_frame(&mut self, chunk: Vec<u8>) -> bool {
self.remaining_response.extend(&chunk);
loop {
let len = match parse_grpc_frame(&self.remaining_response) {
Err(e) => {
self.complete.send(ResultOrEof::Error(e)).unwrap();
return false;
}
Ok(None) => break,
Ok(Some((message, len))) => {
let resp = self.method.resp_marshaller.read(&message);
self.complete.send(From::from(resp)).ok();
len
}
};
self.remaining_response.drain(..len);
}
true
}
fn trailers(&mut self, headers: Vec<StaticHeader>) -> bool {
let _status_200 = slice_get_header(&headers, ":status") == Some("200");
let grpc_status_0 = slice_get_header(&headers, HEADER_GRPC_STATUS) == Some("0");
if /* status_200 && */ grpc_status_0 {
true
} else {
if let Some(message) = slice_get_header(&headers, HEADER_GRPC_MESSAGE) {
self.complete.send(ResultOrEof::Error(GrpcError::GrpcMessage(GrpcMessageError { grpc_message: message.to_owned() }))).unwrap();
} else {
self.complete.send(ResultOrEof::Error(GrpcError::Other("not xxx"))).unwrap();
}
false
}
}
fn end(&mut self) {
self.complete.send(ResultOrEof::Eof).unwrap();
}
}
struct GrpcResponseHandler {
tr: Box<GrpcResponseHandlerTrait>,
}
impl HttpClientResponseHandler for GrpcResponseHandler {
fn headers(&mut self, headers: Vec<StaticHeader>) -> bool {
self.tr.headers(headers)
}
fn data_frame(&mut self, chunk: Vec<u8>) -> bool {
self.tr.data_frame(chunk)
}
fn trailers(&mut self, headers: Vec<StaticHeader>) -> bool {
self.tr.trailers(headers)
}
fn end(&mut self) {
self.tr.end()
}
}
// Data sent from event loop to GrpcClient
struct LoopToClient {
// used only once to send shutdown signal
shutdown_tx: tokio_core::channel::Sender<()>,
loop_handle: reactor::Remote,
http_conn: Arc<HttpClientConnectionAsync<GrpcResponseHandler>>,
}
fn _assert_loop_to_client() {
assert_send::<reactor::Remote>();
assert_send::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_send::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_sync::<HttpClientConnectionAsync<GrpcResponseHandler>>();
assert_send::<Arc<HttpClientConnectionAsync<GrpcResponseHandler>>>();
assert_send::<tokio_core::channel::Sender<()>>();
assert_send::<LoopToClient>();
}
/// gRPC client implementation.
/// Used by generated code.
pub struct GrpcClient {
loop_to_client: LoopToClient,
thread_join_handle: Option<thread::JoinHandle<()>>,
host: String,
http_scheme: HttpScheme,
}
impl GrpcClient {
/// Create a client connected to specified host and port.
pub fn new(host: &str, port: u16) -> GrpcResult<GrpcClient> {
// TODO: sync
// TODO: try connect to all addrs
let socket_addr = try!((host, port).to_socket_addrs()).next().unwrap();
// We need some data back from event loop.
// This channel is used to exchange that data
let (get_from_loop_tx, get_from_loop_rx) = mpsc::channel();
// Start event loop.
let join_handle = thread::spawn(move || {
run_client_event_loop(socket_addr, get_from_loop_tx);
});
// Get back call channel and shutdown channel.
let loop_to_client = try!(get_from_loop_rx.recv()
.map_err(|_| GrpcError::Other("get response from loop")));
Ok(GrpcClient {
loop_to_client: loop_to_client,
thread_join_handle: Some(join_handle),
host: host.to_owned(),
http_scheme: HttpScheme::Http,
})
}
pub fn new_resp_channel<Resp : Send + 'static>(&self)
-> futures::Oneshot<(tokio_core::channel::Sender<ResultOrEof<Resp, GrpcError>>, GrpcStreamSend<Resp>)>
{
let (one_sender, one_receiver) = futures::oneshot();
self.loop_to_client.loop_handle.spawn(move |handle| {
let (sender, receiver) = tokio_core::channel::channel(&handle).unwrap();
let receiver: GrpcStreamSend<ResultOrEof<Resp, GrpcError>> = Box::new(receiver.map_err(GrpcError::from));
let receiver: GrpcStreamSend<Resp> = Box::new(stream_with_eof_and_error(receiver));
one_sender.complete((sender, receiver));
futures::finished(())
});
one_receiver
}
pub fn call_impl<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
let host = self.host.clone();
let http_scheme = self.http_scheme.clone();
let http_conn = self.loop_to_client.http_conn.clone();
// A channel to send response back to caller
let future = self.new_resp_channel().map_err(GrpcError::from).and_then(move |(complete, receiver)| {
let headers = vec![
Header::new(":method", "POST"),
Header::new(":path", method.name.clone()),
Header::new(":authority", host.clone()),
Header::new(":scheme", http_scheme.as_bytes()),
];
let request_frames = {
let method = method.clone();
req
.and_then(move |req| {
let grpc_frame = try!(method.req_marshaller.write(&req));
Ok(write_grpc_frame_to_vec(&grpc_frame))
})
.map_err(|e| HttpError::Other(Box::new(e)))
};
let start_request = http_conn.start_request(
headers,
Box::new(request_frames),
GrpcResponseHandler {
tr: Box::new(GrpcResponseHandlerTyped {
method: method.clone(),
complete: complete,
remaining_response: Vec::new(),
}),
}
).map_err(GrpcError::from);
let receiver: GrpcStreamSend<Resp> = receiver;
start_request.map(move |()| receiver)
});
let s: GrpcStreamSend<Resp> = future_flatten_to_stream(future);
s
}
pub fn call_unary<Req : Send + 'static, Resp : Send + 'static>(&self, req: Req, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcFutureSend<Resp>
{
stream_single_send(self.call_impl(Box::new(stream_once_send(req)), method))
}
pub fn call_server_streaming<Req : Send + 'static, Resp : Send + 'static>(&self, req: Req, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
self.call_impl(stream_once_send(req).boxed(), method)
}
pub fn call_client_streaming<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcFutureSend<Resp>
{
stream_single_send(self.call_impl(req, method))
}
pub fn call_bidi<Req : Send + 'static, Resp : Send + 'static>(&self, req: GrpcStreamSend<Req>, method: Arc<MethodDescriptor<Req, Resp>>)
-> GrpcStreamSend<Resp>
{
self.call_impl(req, method)
}
}
// We shutdown client in destructor.
impl Drop for GrpcClient {
fn drop(&mut self) |
}
// Event loop entry point
fn run_client_event_loop(
socket_addr: SocketAddr,
send_to_back: mpsc::Sender<LoopToClient>)
{
// Create an event loop.
let mut lp = reactor::Core::new().unwrap();
// Create a channel to receive shutdown signal.
let (shutdown_tx, shutdown_rx) = tokio_core::channel::channel(&lp.handle()).unwrap();
let (http_conn, http_conn_future) = HttpClientConnectionAsync::new(lp.handle(), &socket_addr);
let http_conn_future: GrpcFuture<_> = Box::new(http_conn_future.map_err(GrpcError::from));
// Send channels back to GrpcClient
send_to_back
.send(LoopToClient {
shutdown_tx: shutdown_tx,
loop_handle: lp.remote(),
http_conn: Arc::new(http_conn),
})
.expect("send back");
let shutdown = shutdown_rx.into_future().map_err(|(e, _)| GrpcError::from(e)).and_then(move |_| {
// Must complete with error,
// so `join` with this future cancels another future.
futures::failed::<(), _>(GrpcError::Other("shutdown"))
});
// Wait for either completion of connection (i. e. error)
// or shutdown signal.
let done = http_conn_future.join(shutdown);
// TODO: do not ignore error
lp.run(done).ok();
}
| {
// ignore error because even loop may be already dead
self.loop_to_client.shutdown_tx.send(()).ok();
// do not ignore errors because we own event loop thread
self.thread_join_handle.take().expect("handle.take")
.join().expect("join thread");
} | identifier_body |
fixed.rs | // Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
use crate::array::PrimitiveArray;
use crate::null_sentinel;
use arrow_array::builder::BufferBuilder;
use arrow_array::{ArrowPrimitiveType, BooleanArray, FixedSizeBinaryArray};
use arrow_buffer::{bit_util, i256, ArrowNativeType, Buffer, MutableBuffer};
use arrow_data::{ArrayData, ArrayDataBuilder};
use arrow_schema::{DataType, SortOptions};
use half::f16;
pub trait FromSlice {
fn from_slice(slice: &[u8], invert: bool) -> Self;
}
impl<const N: usize> FromSlice for [u8; N] {
#[inline]
fn from_slice(slice: &[u8], invert: bool) -> Self {
let mut t: Self = slice.try_into().unwrap();
if invert {
t.iter_mut().for_each(|o| *o = !*o);
}
t
}
}
/// Encodes a value of a particular fixed width type into bytes according to the rules
/// described on [`super::RowConverter`]
pub trait FixedLengthEncoding: Copy {
const ENCODED_LEN: usize = 1 + std::mem::size_of::<Self::Encoded>();
type Encoded: Sized + Copy + FromSlice + AsRef<[u8]> + AsMut<[u8]>;
fn encode(self) -> Self::Encoded;
fn decode(encoded: Self::Encoded) -> Self;
}
impl FixedLengthEncoding for bool {
type Encoded = [u8; 1];
fn encode(self) -> [u8; 1] {
[self as u8]
}
fn decode(encoded: Self::Encoded) -> Self {
encoded[0] != 0
}
}
macro_rules! encode_signed {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
let mut b = self.to_be_bytes();
// Toggle top "sign" bit to ensure consistent sort order
b[0] ^= 0x80;
b
}
fn decode(mut encoded: Self::Encoded) -> Self {
// Toggle top "sign" bit
encoded[0] ^= 0x80;
Self::from_be_bytes(encoded)
}
}
};
}
encode_signed!(1, i8);
encode_signed!(2, i16);
encode_signed!(4, i32);
encode_signed!(8, i64);
encode_signed!(16, i128);
encode_signed!(32, i256);
macro_rules! encode_unsigned {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
self.to_be_bytes()
}
fn decode(encoded: Self::Encoded) -> Self {
Self::from_be_bytes(encoded)
}
}
};
}
encode_unsigned!(1, u8);
encode_unsigned!(2, u16);
encode_unsigned!(4, u32);
encode_unsigned!(8, u64);
impl FixedLengthEncoding for f16 {
type Encoded = [u8; 2];
fn encode(self) -> [u8; 2] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i16;
let val = s ^ (((s >> 15) as u16) >> 1) as i16;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i16::decode(encoded);
let val = bits ^ (((bits >> 15) as u16) >> 1) as i16;
Self::from_bits(val as u16)
}
}
impl FixedLengthEncoding for f32 {
type Encoded = [u8; 4];
fn encode(self) -> [u8; 4] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i32;
let val = s ^ (((s >> 31) as u32) >> 1) as i32;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i32::decode(encoded);
let val = bits ^ (((bits >> 31) as u32) >> 1) as i32;
Self::from_bits(val as u32)
}
}
impl FixedLengthEncoding for f64 {
type Encoded = [u8; 8];
fn encode(self) -> [u8; 8] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i64;
let val = s ^ (((s >> 63) as u64) >> 1) as i64;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i64::decode(encoded);
let val = bits ^ (((bits >> 63) as u64) >> 1) as i64;
Self::from_bits(val as u64)
}
}
/// Returns the total encoded length (including null byte) for a value of type `T::Native`
pub const fn | <T>(_col: &PrimitiveArray<T>) -> usize
where
T: ArrowPrimitiveType,
T::Native: FixedLengthEncoding,
{
T::Native::ENCODED_LEN
}
/// Fixed width types are encoded as
///
/// - 1 byte `0` if null or `1` if valid
/// - bytes of [`FixedLengthEncoding`]
pub fn encode<T: FixedLengthEncoding, I: IntoIterator<Item = Option<T>>>(
data: &mut [u8],
offsets: &mut [usize],
i: I,
opts: SortOptions,
) {
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(i) {
let end_offset = *offset + T::ENCODED_LEN;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
let mut encoded = val.encode();
if opts.descending {
// Flip bits to reverse order
encoded.as_mut().iter_mut().for_each(|v| *v = !*v)
}
to_write[1..].copy_from_slice(encoded.as_ref())
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
pub fn encode_fixed_size_binary(
data: &mut [u8],
offsets: &mut [usize],
array: &FixedSizeBinaryArray,
opts: SortOptions,
) {
let len = array.value_length() as usize;
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(array.iter()) {
let end_offset = *offset + len + 1;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
to_write[1..].copy_from_slice(&val[..len]);
if opts.descending {
// Flip bits to reverse order
to_write[1..1 + len].iter_mut().for_each(|v| *v = !*v)
}
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
/// Splits `len` bytes from `src`
#[inline]
fn split_off<'a>(src: &mut &'a [u8], len: usize) -> &'a [u8] {
let v = &src[..len];
*src = &src[len..];
v
}
/// Decodes a `BooleanArray` from rows
pub fn decode_bool(rows: &mut [&[u8]], options: SortOptions) -> BooleanArray {
let true_val = match options.descending {
true => !1,
false => 1,
};
let len = rows.len();
let mut null_count = 0;
let mut nulls = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let mut values = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let chunks = len / 64;
let remainder = len % 64;
for chunk in 0..chunks {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..64 {
let i = split_off(&mut rows[bit_idx + chunk * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
if remainder != 0 {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..remainder {
let i = split_off(&mut rows[bit_idx + chunks * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
let builder = ArrayDataBuilder::new(DataType::Boolean)
.len(rows.len())
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls.into()));
// SAFETY:
// Buffers are the correct length
unsafe { BooleanArray::from(builder.build_unchecked()) }
}
/// Decodes a single byte from each row, interpreting `0x01` as a valid value
/// and all other values as a null
///
/// Returns the null count and null buffer
pub fn decode_nulls(rows: &[&[u8]]) -> (usize, Buffer) {
let mut null_count = 0;
let buffer = MutableBuffer::collect_bool(rows.len(), |idx| {
let valid = rows[idx][0] == 1;
null_count += !valid as usize;
valid
})
.into();
(null_count, buffer)
}
/// Decodes a `ArrayData` from rows based on the provided `FixedLengthEncoding` `T`
///
/// # Safety
///
/// `data_type` must be appropriate native type for `T`
unsafe fn decode_fixed<T: FixedLengthEncoding + ArrowNativeType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> ArrayData {
let len = rows.len();
let mut values = BufferBuilder::<T>::new(len);
let (null_count, nulls) = decode_nulls(rows);
for row in rows {
let i = split_off(row, T::ENCODED_LEN);
let value = T::Encoded::from_slice(&i[1..], options.descending);
values.append(T::decode(value));
}
let builder = ArrayDataBuilder::new(data_type)
.len(len)
.null_count(null_count)
.add_buffer(values.finish())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
builder.build_unchecked()
}
/// Decodes a `PrimitiveArray` from rows
pub fn decode_primitive<T: ArrowPrimitiveType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> PrimitiveArray<T>
where
T::Native: FixedLengthEncoding,
{
assert!(PrimitiveArray::<T>::is_compatible(&data_type));
// SAFETY:
// Validated data type above
unsafe { decode_fixed::<T::Native>(rows, data_type, options).into() }
}
/// Decodes a `FixedLengthBinary` from rows
pub fn decode_fixed_size_binary(
rows: &mut [&[u8]],
size: i32,
options: SortOptions,
) -> FixedSizeBinaryArray {
let len = rows.len();
let mut values = MutableBuffer::new(size as usize * rows.len());
let (null_count, nulls) = decode_nulls(rows);
let encoded_len = size as usize + 1;
for row in rows {
let i = split_off(row, encoded_len);
values.extend_from_slice(&i[1..]);
}
if options.descending {
for v in values.as_slice_mut() {
*v = !*v;
}
}
let builder = ArrayDataBuilder::new(DataType::FixedSizeBinary(size))
.len(len)
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
unsafe { builder.build_unchecked().into() }
}
| encoded_len | identifier_name |
fixed.rs | // Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
use crate::array::PrimitiveArray;
use crate::null_sentinel;
use arrow_array::builder::BufferBuilder;
use arrow_array::{ArrowPrimitiveType, BooleanArray, FixedSizeBinaryArray};
use arrow_buffer::{bit_util, i256, ArrowNativeType, Buffer, MutableBuffer};
use arrow_data::{ArrayData, ArrayDataBuilder};
use arrow_schema::{DataType, SortOptions};
use half::f16;
pub trait FromSlice {
fn from_slice(slice: &[u8], invert: bool) -> Self;
}
impl<const N: usize> FromSlice for [u8; N] {
#[inline]
fn from_slice(slice: &[u8], invert: bool) -> Self {
let mut t: Self = slice.try_into().unwrap();
if invert {
t.iter_mut().for_each(|o| *o = !*o);
}
t
}
}
/// Encodes a value of a particular fixed width type into bytes according to the rules
/// described on [`super::RowConverter`]
pub trait FixedLengthEncoding: Copy {
const ENCODED_LEN: usize = 1 + std::mem::size_of::<Self::Encoded>();
type Encoded: Sized + Copy + FromSlice + AsRef<[u8]> + AsMut<[u8]>;
fn encode(self) -> Self::Encoded;
fn decode(encoded: Self::Encoded) -> Self;
}
impl FixedLengthEncoding for bool {
type Encoded = [u8; 1];
fn encode(self) -> [u8; 1] {
[self as u8]
}
fn decode(encoded: Self::Encoded) -> Self {
encoded[0] != 0
}
}
macro_rules! encode_signed {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
let mut b = self.to_be_bytes();
// Toggle top "sign" bit to ensure consistent sort order
b[0] ^= 0x80;
b
}
fn decode(mut encoded: Self::Encoded) -> Self {
// Toggle top "sign" bit
encoded[0] ^= 0x80;
Self::from_be_bytes(encoded)
}
}
};
}
encode_signed!(1, i8);
encode_signed!(2, i16);
encode_signed!(4, i32);
encode_signed!(8, i64);
encode_signed!(16, i128);
encode_signed!(32, i256);
macro_rules! encode_unsigned {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
self.to_be_bytes()
}
fn decode(encoded: Self::Encoded) -> Self {
Self::from_be_bytes(encoded)
}
}
};
}
encode_unsigned!(1, u8);
encode_unsigned!(2, u16);
encode_unsigned!(4, u32);
encode_unsigned!(8, u64);
impl FixedLengthEncoding for f16 {
type Encoded = [u8; 2];
fn encode(self) -> [u8; 2] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i16;
let val = s ^ (((s >> 15) as u16) >> 1) as i16;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i16::decode(encoded);
let val = bits ^ (((bits >> 15) as u16) >> 1) as i16;
Self::from_bits(val as u16)
}
}
impl FixedLengthEncoding for f32 {
type Encoded = [u8; 4];
fn encode(self) -> [u8; 4] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i32;
let val = s ^ (((s >> 31) as u32) >> 1) as i32;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i32::decode(encoded);
let val = bits ^ (((bits >> 31) as u32) >> 1) as i32;
Self::from_bits(val as u32)
}
}
|
fn encode(self) -> [u8; 8] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i64;
let val = s ^ (((s >> 63) as u64) >> 1) as i64;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i64::decode(encoded);
let val = bits ^ (((bits >> 63) as u64) >> 1) as i64;
Self::from_bits(val as u64)
}
}
/// Returns the total encoded length (including null byte) for a value of type `T::Native`
pub const fn encoded_len<T>(_col: &PrimitiveArray<T>) -> usize
where
T: ArrowPrimitiveType,
T::Native: FixedLengthEncoding,
{
T::Native::ENCODED_LEN
}
/// Fixed width types are encoded as
///
/// - 1 byte `0` if null or `1` if valid
/// - bytes of [`FixedLengthEncoding`]
pub fn encode<T: FixedLengthEncoding, I: IntoIterator<Item = Option<T>>>(
data: &mut [u8],
offsets: &mut [usize],
i: I,
opts: SortOptions,
) {
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(i) {
let end_offset = *offset + T::ENCODED_LEN;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
let mut encoded = val.encode();
if opts.descending {
// Flip bits to reverse order
encoded.as_mut().iter_mut().for_each(|v| *v = !*v)
}
to_write[1..].copy_from_slice(encoded.as_ref())
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
pub fn encode_fixed_size_binary(
data: &mut [u8],
offsets: &mut [usize],
array: &FixedSizeBinaryArray,
opts: SortOptions,
) {
let len = array.value_length() as usize;
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(array.iter()) {
let end_offset = *offset + len + 1;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
to_write[1..].copy_from_slice(&val[..len]);
if opts.descending {
// Flip bits to reverse order
to_write[1..1 + len].iter_mut().for_each(|v| *v = !*v)
}
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
/// Splits `len` bytes from `src`
#[inline]
fn split_off<'a>(src: &mut &'a [u8], len: usize) -> &'a [u8] {
let v = &src[..len];
*src = &src[len..];
v
}
/// Decodes a `BooleanArray` from rows
pub fn decode_bool(rows: &mut [&[u8]], options: SortOptions) -> BooleanArray {
let true_val = match options.descending {
true => !1,
false => 1,
};
let len = rows.len();
let mut null_count = 0;
let mut nulls = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let mut values = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let chunks = len / 64;
let remainder = len % 64;
for chunk in 0..chunks {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..64 {
let i = split_off(&mut rows[bit_idx + chunk * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
if remainder != 0 {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..remainder {
let i = split_off(&mut rows[bit_idx + chunks * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
let builder = ArrayDataBuilder::new(DataType::Boolean)
.len(rows.len())
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls.into()));
// SAFETY:
// Buffers are the correct length
unsafe { BooleanArray::from(builder.build_unchecked()) }
}
/// Decodes a single byte from each row, interpreting `0x01` as a valid value
/// and all other values as a null
///
/// Returns the null count and null buffer
pub fn decode_nulls(rows: &[&[u8]]) -> (usize, Buffer) {
let mut null_count = 0;
let buffer = MutableBuffer::collect_bool(rows.len(), |idx| {
let valid = rows[idx][0] == 1;
null_count += !valid as usize;
valid
})
.into();
(null_count, buffer)
}
/// Decodes a `ArrayData` from rows based on the provided `FixedLengthEncoding` `T`
///
/// # Safety
///
/// `data_type` must be appropriate native type for `T`
unsafe fn decode_fixed<T: FixedLengthEncoding + ArrowNativeType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> ArrayData {
let len = rows.len();
let mut values = BufferBuilder::<T>::new(len);
let (null_count, nulls) = decode_nulls(rows);
for row in rows {
let i = split_off(row, T::ENCODED_LEN);
let value = T::Encoded::from_slice(&i[1..], options.descending);
values.append(T::decode(value));
}
let builder = ArrayDataBuilder::new(data_type)
.len(len)
.null_count(null_count)
.add_buffer(values.finish())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
builder.build_unchecked()
}
/// Decodes a `PrimitiveArray` from rows
pub fn decode_primitive<T: ArrowPrimitiveType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> PrimitiveArray<T>
where
T::Native: FixedLengthEncoding,
{
assert!(PrimitiveArray::<T>::is_compatible(&data_type));
// SAFETY:
// Validated data type above
unsafe { decode_fixed::<T::Native>(rows, data_type, options).into() }
}
/// Decodes a `FixedLengthBinary` from rows
pub fn decode_fixed_size_binary(
rows: &mut [&[u8]],
size: i32,
options: SortOptions,
) -> FixedSizeBinaryArray {
let len = rows.len();
let mut values = MutableBuffer::new(size as usize * rows.len());
let (null_count, nulls) = decode_nulls(rows);
let encoded_len = size as usize + 1;
for row in rows {
let i = split_off(row, encoded_len);
values.extend_from_slice(&i[1..]);
}
if options.descending {
for v in values.as_slice_mut() {
*v = !*v;
}
}
let builder = ArrayDataBuilder::new(DataType::FixedSizeBinary(size))
.len(len)
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
unsafe { builder.build_unchecked().into() }
} | impl FixedLengthEncoding for f64 {
type Encoded = [u8; 8]; | random_line_split |
fixed.rs | // Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
use crate::array::PrimitiveArray;
use crate::null_sentinel;
use arrow_array::builder::BufferBuilder;
use arrow_array::{ArrowPrimitiveType, BooleanArray, FixedSizeBinaryArray};
use arrow_buffer::{bit_util, i256, ArrowNativeType, Buffer, MutableBuffer};
use arrow_data::{ArrayData, ArrayDataBuilder};
use arrow_schema::{DataType, SortOptions};
use half::f16;
pub trait FromSlice {
fn from_slice(slice: &[u8], invert: bool) -> Self;
}
impl<const N: usize> FromSlice for [u8; N] {
#[inline]
fn from_slice(slice: &[u8], invert: bool) -> Self {
let mut t: Self = slice.try_into().unwrap();
if invert {
t.iter_mut().for_each(|o| *o = !*o);
}
t
}
}
/// Encodes a value of a particular fixed width type into bytes according to the rules
/// described on [`super::RowConverter`]
pub trait FixedLengthEncoding: Copy {
const ENCODED_LEN: usize = 1 + std::mem::size_of::<Self::Encoded>();
type Encoded: Sized + Copy + FromSlice + AsRef<[u8]> + AsMut<[u8]>;
fn encode(self) -> Self::Encoded;
fn decode(encoded: Self::Encoded) -> Self;
}
impl FixedLengthEncoding for bool {
type Encoded = [u8; 1];
fn encode(self) -> [u8; 1] {
[self as u8]
}
fn decode(encoded: Self::Encoded) -> Self {
encoded[0] != 0
}
}
macro_rules! encode_signed {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
let mut b = self.to_be_bytes();
// Toggle top "sign" bit to ensure consistent sort order
b[0] ^= 0x80;
b
}
fn decode(mut encoded: Self::Encoded) -> Self {
// Toggle top "sign" bit
encoded[0] ^= 0x80;
Self::from_be_bytes(encoded)
}
}
};
}
encode_signed!(1, i8);
encode_signed!(2, i16);
encode_signed!(4, i32);
encode_signed!(8, i64);
encode_signed!(16, i128);
encode_signed!(32, i256);
macro_rules! encode_unsigned {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
self.to_be_bytes()
}
fn decode(encoded: Self::Encoded) -> Self {
Self::from_be_bytes(encoded)
}
}
};
}
encode_unsigned!(1, u8);
encode_unsigned!(2, u16);
encode_unsigned!(4, u32);
encode_unsigned!(8, u64);
impl FixedLengthEncoding for f16 {
type Encoded = [u8; 2];
fn encode(self) -> [u8; 2] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i16;
let val = s ^ (((s >> 15) as u16) >> 1) as i16;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i16::decode(encoded);
let val = bits ^ (((bits >> 15) as u16) >> 1) as i16;
Self::from_bits(val as u16)
}
}
impl FixedLengthEncoding for f32 {
type Encoded = [u8; 4];
fn encode(self) -> [u8; 4] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i32;
let val = s ^ (((s >> 31) as u32) >> 1) as i32;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i32::decode(encoded);
let val = bits ^ (((bits >> 31) as u32) >> 1) as i32;
Self::from_bits(val as u32)
}
}
impl FixedLengthEncoding for f64 {
type Encoded = [u8; 8];
fn encode(self) -> [u8; 8] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i64;
let val = s ^ (((s >> 63) as u64) >> 1) as i64;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i64::decode(encoded);
let val = bits ^ (((bits >> 63) as u64) >> 1) as i64;
Self::from_bits(val as u64)
}
}
/// Returns the total encoded length (including null byte) for a value of type `T::Native`
pub const fn encoded_len<T>(_col: &PrimitiveArray<T>) -> usize
where
T: ArrowPrimitiveType,
T::Native: FixedLengthEncoding,
{
T::Native::ENCODED_LEN
}
/// Fixed width types are encoded as
///
/// - 1 byte `0` if null or `1` if valid
/// - bytes of [`FixedLengthEncoding`]
pub fn encode<T: FixedLengthEncoding, I: IntoIterator<Item = Option<T>>>(
data: &mut [u8],
offsets: &mut [usize],
i: I,
opts: SortOptions,
) {
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(i) {
let end_offset = *offset + T::ENCODED_LEN;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
let mut encoded = val.encode();
if opts.descending |
to_write[1..].copy_from_slice(encoded.as_ref())
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
pub fn encode_fixed_size_binary(
data: &mut [u8],
offsets: &mut [usize],
array: &FixedSizeBinaryArray,
opts: SortOptions,
) {
let len = array.value_length() as usize;
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(array.iter()) {
let end_offset = *offset + len + 1;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
to_write[1..].copy_from_slice(&val[..len]);
if opts.descending {
// Flip bits to reverse order
to_write[1..1 + len].iter_mut().for_each(|v| *v = !*v)
}
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
/// Splits `len` bytes from `src`
#[inline]
fn split_off<'a>(src: &mut &'a [u8], len: usize) -> &'a [u8] {
let v = &src[..len];
*src = &src[len..];
v
}
/// Decodes a `BooleanArray` from rows
pub fn decode_bool(rows: &mut [&[u8]], options: SortOptions) -> BooleanArray {
let true_val = match options.descending {
true => !1,
false => 1,
};
let len = rows.len();
let mut null_count = 0;
let mut nulls = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let mut values = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let chunks = len / 64;
let remainder = len % 64;
for chunk in 0..chunks {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..64 {
let i = split_off(&mut rows[bit_idx + chunk * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
if remainder != 0 {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..remainder {
let i = split_off(&mut rows[bit_idx + chunks * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
let builder = ArrayDataBuilder::new(DataType::Boolean)
.len(rows.len())
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls.into()));
// SAFETY:
// Buffers are the correct length
unsafe { BooleanArray::from(builder.build_unchecked()) }
}
/// Decodes a single byte from each row, interpreting `0x01` as a valid value
/// and all other values as a null
///
/// Returns the null count and null buffer
pub fn decode_nulls(rows: &[&[u8]]) -> (usize, Buffer) {
let mut null_count = 0;
let buffer = MutableBuffer::collect_bool(rows.len(), |idx| {
let valid = rows[idx][0] == 1;
null_count += !valid as usize;
valid
})
.into();
(null_count, buffer)
}
/// Decodes a `ArrayData` from rows based on the provided `FixedLengthEncoding` `T`
///
/// # Safety
///
/// `data_type` must be appropriate native type for `T`
unsafe fn decode_fixed<T: FixedLengthEncoding + ArrowNativeType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> ArrayData {
let len = rows.len();
let mut values = BufferBuilder::<T>::new(len);
let (null_count, nulls) = decode_nulls(rows);
for row in rows {
let i = split_off(row, T::ENCODED_LEN);
let value = T::Encoded::from_slice(&i[1..], options.descending);
values.append(T::decode(value));
}
let builder = ArrayDataBuilder::new(data_type)
.len(len)
.null_count(null_count)
.add_buffer(values.finish())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
builder.build_unchecked()
}
/// Decodes a `PrimitiveArray` from rows
pub fn decode_primitive<T: ArrowPrimitiveType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> PrimitiveArray<T>
where
T::Native: FixedLengthEncoding,
{
assert!(PrimitiveArray::<T>::is_compatible(&data_type));
// SAFETY:
// Validated data type above
unsafe { decode_fixed::<T::Native>(rows, data_type, options).into() }
}
/// Decodes a `FixedLengthBinary` from rows
pub fn decode_fixed_size_binary(
rows: &mut [&[u8]],
size: i32,
options: SortOptions,
) -> FixedSizeBinaryArray {
let len = rows.len();
let mut values = MutableBuffer::new(size as usize * rows.len());
let (null_count, nulls) = decode_nulls(rows);
let encoded_len = size as usize + 1;
for row in rows {
let i = split_off(row, encoded_len);
values.extend_from_slice(&i[1..]);
}
if options.descending {
for v in values.as_slice_mut() {
*v = !*v;
}
}
let builder = ArrayDataBuilder::new(DataType::FixedSizeBinary(size))
.len(len)
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
unsafe { builder.build_unchecked().into() }
}
| {
// Flip bits to reverse order
encoded.as_mut().iter_mut().for_each(|v| *v = !*v)
} | conditional_block |
fixed.rs | // Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
use crate::array::PrimitiveArray;
use crate::null_sentinel;
use arrow_array::builder::BufferBuilder;
use arrow_array::{ArrowPrimitiveType, BooleanArray, FixedSizeBinaryArray};
use arrow_buffer::{bit_util, i256, ArrowNativeType, Buffer, MutableBuffer};
use arrow_data::{ArrayData, ArrayDataBuilder};
use arrow_schema::{DataType, SortOptions};
use half::f16;
pub trait FromSlice {
fn from_slice(slice: &[u8], invert: bool) -> Self;
}
impl<const N: usize> FromSlice for [u8; N] {
#[inline]
fn from_slice(slice: &[u8], invert: bool) -> Self {
let mut t: Self = slice.try_into().unwrap();
if invert {
t.iter_mut().for_each(|o| *o = !*o);
}
t
}
}
/// Encodes a value of a particular fixed width type into bytes according to the rules
/// described on [`super::RowConverter`]
pub trait FixedLengthEncoding: Copy {
const ENCODED_LEN: usize = 1 + std::mem::size_of::<Self::Encoded>();
type Encoded: Sized + Copy + FromSlice + AsRef<[u8]> + AsMut<[u8]>;
fn encode(self) -> Self::Encoded;
fn decode(encoded: Self::Encoded) -> Self;
}
impl FixedLengthEncoding for bool {
type Encoded = [u8; 1];
fn encode(self) -> [u8; 1] |
fn decode(encoded: Self::Encoded) -> Self {
encoded[0] != 0
}
}
macro_rules! encode_signed {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
let mut b = self.to_be_bytes();
// Toggle top "sign" bit to ensure consistent sort order
b[0] ^= 0x80;
b
}
fn decode(mut encoded: Self::Encoded) -> Self {
// Toggle top "sign" bit
encoded[0] ^= 0x80;
Self::from_be_bytes(encoded)
}
}
};
}
encode_signed!(1, i8);
encode_signed!(2, i16);
encode_signed!(4, i32);
encode_signed!(8, i64);
encode_signed!(16, i128);
encode_signed!(32, i256);
macro_rules! encode_unsigned {
($n:expr, $t:ty) => {
impl FixedLengthEncoding for $t {
type Encoded = [u8; $n];
fn encode(self) -> [u8; $n] {
self.to_be_bytes()
}
fn decode(encoded: Self::Encoded) -> Self {
Self::from_be_bytes(encoded)
}
}
};
}
encode_unsigned!(1, u8);
encode_unsigned!(2, u16);
encode_unsigned!(4, u32);
encode_unsigned!(8, u64);
impl FixedLengthEncoding for f16 {
type Encoded = [u8; 2];
fn encode(self) -> [u8; 2] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i16;
let val = s ^ (((s >> 15) as u16) >> 1) as i16;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i16::decode(encoded);
let val = bits ^ (((bits >> 15) as u16) >> 1) as i16;
Self::from_bits(val as u16)
}
}
impl FixedLengthEncoding for f32 {
type Encoded = [u8; 4];
fn encode(self) -> [u8; 4] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i32;
let val = s ^ (((s >> 31) as u32) >> 1) as i32;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i32::decode(encoded);
let val = bits ^ (((bits >> 31) as u32) >> 1) as i32;
Self::from_bits(val as u32)
}
}
impl FixedLengthEncoding for f64 {
type Encoded = [u8; 8];
fn encode(self) -> [u8; 8] {
// https://github.com/rust-lang/rust/blob/9c20b2a8cc7588decb6de25ac6a7912dcef24d65/library/core/src/num/f32.rs#L1176-L1260
let s = self.to_bits() as i64;
let val = s ^ (((s >> 63) as u64) >> 1) as i64;
val.encode()
}
fn decode(encoded: Self::Encoded) -> Self {
let bits = i64::decode(encoded);
let val = bits ^ (((bits >> 63) as u64) >> 1) as i64;
Self::from_bits(val as u64)
}
}
/// Returns the total encoded length (including null byte) for a value of type `T::Native`
pub const fn encoded_len<T>(_col: &PrimitiveArray<T>) -> usize
where
T: ArrowPrimitiveType,
T::Native: FixedLengthEncoding,
{
T::Native::ENCODED_LEN
}
/// Fixed width types are encoded as
///
/// - 1 byte `0` if null or `1` if valid
/// - bytes of [`FixedLengthEncoding`]
pub fn encode<T: FixedLengthEncoding, I: IntoIterator<Item = Option<T>>>(
data: &mut [u8],
offsets: &mut [usize],
i: I,
opts: SortOptions,
) {
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(i) {
let end_offset = *offset + T::ENCODED_LEN;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
let mut encoded = val.encode();
if opts.descending {
// Flip bits to reverse order
encoded.as_mut().iter_mut().for_each(|v| *v = !*v)
}
to_write[1..].copy_from_slice(encoded.as_ref())
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
pub fn encode_fixed_size_binary(
data: &mut [u8],
offsets: &mut [usize],
array: &FixedSizeBinaryArray,
opts: SortOptions,
) {
let len = array.value_length() as usize;
for (offset, maybe_val) in offsets.iter_mut().skip(1).zip(array.iter()) {
let end_offset = *offset + len + 1;
if let Some(val) = maybe_val {
let to_write = &mut data[*offset..end_offset];
to_write[0] = 1;
to_write[1..].copy_from_slice(&val[..len]);
if opts.descending {
// Flip bits to reverse order
to_write[1..1 + len].iter_mut().for_each(|v| *v = !*v)
}
} else {
data[*offset] = null_sentinel(opts);
}
*offset = end_offset;
}
}
/// Splits `len` bytes from `src`
#[inline]
fn split_off<'a>(src: &mut &'a [u8], len: usize) -> &'a [u8] {
let v = &src[..len];
*src = &src[len..];
v
}
/// Decodes a `BooleanArray` from rows
pub fn decode_bool(rows: &mut [&[u8]], options: SortOptions) -> BooleanArray {
let true_val = match options.descending {
true => !1,
false => 1,
};
let len = rows.len();
let mut null_count = 0;
let mut nulls = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let mut values = MutableBuffer::new(bit_util::ceil(len, 64) * 8);
let chunks = len / 64;
let remainder = len % 64;
for chunk in 0..chunks {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..64 {
let i = split_off(&mut rows[bit_idx + chunk * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
if remainder != 0 {
let mut null_packed = 0;
let mut values_packed = 0;
for bit_idx in 0..remainder {
let i = split_off(&mut rows[bit_idx + chunks * 64], 2);
let (null, value) = (i[0] == 1, i[1] == true_val);
null_count += !null as usize;
null_packed |= (null as u64) << bit_idx;
values_packed |= (value as u64) << bit_idx;
}
nulls.push(null_packed);
values.push(values_packed);
}
let builder = ArrayDataBuilder::new(DataType::Boolean)
.len(rows.len())
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls.into()));
// SAFETY:
// Buffers are the correct length
unsafe { BooleanArray::from(builder.build_unchecked()) }
}
/// Decodes a single byte from each row, interpreting `0x01` as a valid value
/// and all other values as a null
///
/// Returns the null count and null buffer
pub fn decode_nulls(rows: &[&[u8]]) -> (usize, Buffer) {
let mut null_count = 0;
let buffer = MutableBuffer::collect_bool(rows.len(), |idx| {
let valid = rows[idx][0] == 1;
null_count += !valid as usize;
valid
})
.into();
(null_count, buffer)
}
/// Decodes a `ArrayData` from rows based on the provided `FixedLengthEncoding` `T`
///
/// # Safety
///
/// `data_type` must be appropriate native type for `T`
unsafe fn decode_fixed<T: FixedLengthEncoding + ArrowNativeType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> ArrayData {
let len = rows.len();
let mut values = BufferBuilder::<T>::new(len);
let (null_count, nulls) = decode_nulls(rows);
for row in rows {
let i = split_off(row, T::ENCODED_LEN);
let value = T::Encoded::from_slice(&i[1..], options.descending);
values.append(T::decode(value));
}
let builder = ArrayDataBuilder::new(data_type)
.len(len)
.null_count(null_count)
.add_buffer(values.finish())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
builder.build_unchecked()
}
/// Decodes a `PrimitiveArray` from rows
pub fn decode_primitive<T: ArrowPrimitiveType>(
rows: &mut [&[u8]],
data_type: DataType,
options: SortOptions,
) -> PrimitiveArray<T>
where
T::Native: FixedLengthEncoding,
{
assert!(PrimitiveArray::<T>::is_compatible(&data_type));
// SAFETY:
// Validated data type above
unsafe { decode_fixed::<T::Native>(rows, data_type, options).into() }
}
/// Decodes a `FixedLengthBinary` from rows
pub fn decode_fixed_size_binary(
rows: &mut [&[u8]],
size: i32,
options: SortOptions,
) -> FixedSizeBinaryArray {
let len = rows.len();
let mut values = MutableBuffer::new(size as usize * rows.len());
let (null_count, nulls) = decode_nulls(rows);
let encoded_len = size as usize + 1;
for row in rows {
let i = split_off(row, encoded_len);
values.extend_from_slice(&i[1..]);
}
if options.descending {
for v in values.as_slice_mut() {
*v = !*v;
}
}
let builder = ArrayDataBuilder::new(DataType::FixedSizeBinary(size))
.len(len)
.null_count(null_count)
.add_buffer(values.into())
.null_bit_buffer(Some(nulls));
// SAFETY: Buffers correct length
unsafe { builder.build_unchecked().into() }
}
| {
[self as u8]
} | identifier_body |
pop.py | import math
from historia.utils import unique_id, position_in_range
from historia.pops.models.inventory import Inventory
from historia.economy.enums.resource import Good, NaturalResource
from historia.economy.enums.order_type import OrderType
from historia.economy.models.price_range import PriceRange
from historia.economy.models.order import Order
from historia.pops.enums.pop_job import PopJob
DEBUG = False
class Pop(object):
"""
A simulated unit of population
"""
def | (self, province, pop_job, population):
"""
Creates a new Pop.
manager (Historia)
province (SecondaryDivision)
culture (Culture)
religion (Religion)
language (Language)
job (Job)
"""
self.bankrupt_times = 0
self.home = province
self.location = province
self.id = unique_id('po')
self.population = population
self.population_yesterday = 0
self.pop_job = pop_job
# ECONOMY
self.money = pop_job.start_money
self.money_yesterday = 0
self.bankrupt = False
# set inventory and ideal amounts
self.inventory = Inventory(pop_job.inventory_size)
self.give_start_inventory()
self.update_ideal_inventory()
# a dictionary of Goods to PriceRanges
# represents the price range the agent considers valid for each Good
self.price_belief = {}
# a dictionary of Goods to price list
# represents the prices of the good that the Pop has observed
# during the time they have been trading
self.observed_trading_range = {}
self.successful_trades = 0
self.failed_trades = 0
# make some fake initial data
for good in Good.all():
avg_price = self.market.avg_historial_price(good, 15)
# fake trades
self.observed_trading_range[good] = [
avg_price * 0.5,
avg_price * 1.5
]
# generate fake price belief
self.price_belief[good] = PriceRange(avg_price * 0.5, avg_price * 1.5)
# Merchant logic
self.trade_location = None # the province this Pop is traveling to
self.trade_good = None # what good we're trading in right now
self.trade_amount = 0 # amount of trade_good we should be trading
self.trading_days = 0 # number of days waiting to trade
# Generic Pop properties
@property
def social_class(self):
return self.pop_job.social_class
@property
def market(self):
"Get the market instance"
return self.location.market
@property
def profit(self):
"Determine today's profit"
return self.money - self.money_yesterday
@property
def total_trades(self):
"Total number of trades this Pop participated in"
return self.successful_trades + self.failed_trades
@property
def trade_success(self):
"Percent of trades that were successful"
if self.total_trades == 0:
return 0
return (self.successful_trades / self.total_trades) * 100
@property
def is_away(self):
"Is this Pop away from it's home?"
return self.home is not self.location
# Merchant specific logic
def go_to_province(self, province):
"Moves the Pop to another Province"
self.location = province
def decide_trade_plan(self):
"""
Decide what good to trade in and how much.
Look for the most in demand good, or the most expensive good at the home Province
Find a province near home province where its the cheapest and there's inventory
"""
self.trade_amount = 5
most_demanded_goods = self.home.market.goods_demand_ratio(day_range=1)
most_demanded_goods = sorted(most_demanded_goods.items(), key=lambda i: i[1], reverse=True)
# if we already had a trade good, refresh ideal inventory
if self.trade_good:
self.update_ideal_inventory()
if DEBUG: print("Finding a Good to trade:")
for good, demand in most_demanded_goods:
if demand > 0:
# find nearby provinces where this has inventory and the price is lower
price_at_home = self.home.market.mean_price(good)
if DEBUG: print("Good: {}, Demand: {}, Price: ${}".format(good.title, demand, price_at_home))
neighboring_markets = [p.market for p in self.location.owned_neighbors]
neighboring_markets = [m for m in neighboring_markets if m.supply_for(good) > self.trade_amount]
neighboring_markets.sort(key=lambda m: m.supply_for(good), reverse=True)
if len(neighboring_markets) > 0:
# we found places where this good is cheaper and in inventory
target = neighboring_markets[0].location
price_at_target = target.market.mean_price(good)
# only trade with prices where we can make money
if price_at_home > price_at_target:
offset = 0
if good is Good.bread:
offset = 1
self.inventory.set_ideal(good, self.trade_amount + offset)
self.trade_location = target
if DEBUG:
print("\tTarget: {}, Supply: {}, Price: ${}, Price at home: ${}".format(
self.trade_location.name,
self.trade_location.market.supply_for(good),
self.trade_location.market.mean_price(good),
price_at_home)
)
self.trade_good = good
return
else:
if DEBUG: print("\tPrice is higher at target (home: ${} target: ${})".format(price_at_home, price_at_target))
else:
if DEBUG: print("\tNo markets selling {} found".format(good))
# Generic economic logic
def update_ideal_inventory(self):
"Update ideal inventory"
# reset so that the Pop can sell the inventory it doesn't need
for good in Good.all():
self.inventory.set_ideal(good, 0)
# update ideal inventory for new Job
for item in self.pop_job.ideal_inventory:
self.inventory.set_ideal(item['good'], item['amount'])
def give_start_inventory(self):
"Give the Pop the inventory it needs to do its job"
for item in self.pop_job.start_inventory:
self.inventory.add(item['good'], item['amount'])
def change_population(self, trade_success):
"Change the population based off the trade"
self.population_yesterday = self.population
if trade_success:
self.population += round(self.population * 0.01)
else:
self.population -= round(self.population * 0.002)
def handle_bankruptcy(self, pop_job):
"Change job, create money out of thin air, update ideal inventory"
# TODO: stop creating money out of thin air
self.pop_job = pop_job
self.bankrupt_times += 1
self.money = 2
self.update_ideal_inventory()
self.give_start_inventory()
def perform_logic(self):
"Depending on PopJob, perform logic (including production)"
logic = self.pop_job.logic(self)
logic.perform()
def create_buy_order(self, good, limit):
"Create a buy order for a given Good at a determined quantity"
bid_price = self.determine_price_of(good)
ideal = self.determine_buy_quantity(good)
# can't buy more than limit
quantity_to_buy = limit if ideal > limit else ideal
if quantity_to_buy > 0:
return Order(self, OrderType.buy_order, quantity_to_buy, bid_price, good)
return False
def create_sell_order(self, good, limit):
"Create a sell order for a given Good at a determined quantity"
sell_price = self.determine_price_of(good)
ideal = self.determine_sell_quantity(good)
# can't buy more than limit
quantity_to_sell = limit if ideal < limit else ideal
if quantity_to_sell > 0:
return Order(self, OrderType.sell_order, quantity_to_sell, sell_price, good)
return False
def price_belief_for(self, good):
"Gets the price belief this agent has for a particular Good"
if good in self.price_belief:
return self.price_belief[good]
def determine_price_of(self, good):
"Determine the price of a particular good"
return self.price_belief_for(good).random()
def trading_range_extremes(self, good):
"Gets the lowest and highst price of a Good this agent has seen"
trading_range = self.observed_trading_range[good]
return PriceRange(min(trading_range), max(trading_range))
def determine_sell_quantity(self, good):
"Determine how much inventory goods to sell based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = position_in_range(mean, trading_range.low, trading_range.high)
amount_to_sell = round(favoribility * self.inventory.surplus(good))
if amount_to_sell < 1:
amount_to_sell = 1
return amount_to_sell
def determine_buy_quantity(self, good):
"Determine how much goods to buy based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = 1 - position_in_range(mean, trading_range.low, trading_range.high)
amount_to_buy = round(favoribility * self.inventory.shortage(good))
if amount_to_buy < 1:
amount_to_buy = 1
return amount_to_buy
def generate_orders(self, good):
"""
If the Pop needs a Good to perform production, buy it
If the Pop has surplus Resources, sell them
"""
surplus = self.inventory.surplus(good)
if surplus >= 1: # sell inventory
# the original only old one item here
sell_amount = surplus
order = self.create_sell_order(good, surplus)
if order:
# print('{} sells {} {}'.format(self.pop_job.title, sell_amount, good.name))
self.market.sell(order)
else: # buy more
shortage = self.inventory.shortage(good)
free_space = self.inventory.empty_space
if shortage > 0:
if shortage <= free_space:
# enough space for ideal order
limit = shortage
else:
# not enough space for ideal order
limit = math.floor(free_space / shortage)
if limit > 0:
order = self.create_buy_order(good, limit)
if order:
# print('{} buys {} {}'.format(self.pop_job.title, limit, good.name))
self.market.buy(order)
# else:
# print("{} has no shortage of {} (has shortage: {})".format(self.pop_job.title, good.title, shortage))
def update_price_model(self, good, order_type, is_successful, clearing_price=0):
"""
Update the Pop's price model for the given resource
good (Good) The Good which was orderd
order_type (OrderType) Which kind of Order this was
is_successful (bool) whether or not the Order was successful
clearing_price (float) The price per unit of the good that was ordered
as defined by the Pop which ordered it
"""
SIGNIFICANT = 0.25 # 25% more or less is "significant"
SIG_IMBALANCE = 0.33
LOW_INVENTORY = 0.1 # 10% of ideal inventory = "LOW"
HIGH_INVENTORY = 2.0 # 200% of ideal inventory = "HIGH"
MIN_PRICE = 0.01 # lowest allowed price of a Good
if is_successful:
# add this trade to the observed trading range
self.observed_trading_range[good].append(clearing_price)
public_mean_price = self.market.mean_price(good)
belief = self.price_belief[good]
mean = belief.mean()
wobble = 0.05 # the degree which the Pop should bid outside the belief
# how different the public mean price is from the price belief
delta_to_mean = mean - public_mean_price
if is_successful:
if order_type is OrderType.buy_order and delta_to_mean > SIGNIFICANT:
# this Pop overpaid, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
elif order_type is OrderType.sell_order and delta_to_mean < -SIGNIFICANT:
# this Pop underpaid!, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# increase the belief's certainty
belief.low += wobble * mean
belief.high -= wobble * mean
else:
# shift towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# check for inventory special cases
stocks = self.inventory.get_amount(good)
ideal = self.inventory.get_ideal(good)
# if we're buying and inventory is too low
# meaning we're desperate to buy
if order_type is OrderType.buy_order and stocks < LOW_INVENTORY * ideal:
wobble *= 2
# if we're selling and inventory is too high
# meaning we're desperate to sell
elif order_type is OrderType.sell_order and stocks > HIGH_INVENTORY * ideal:
wobble *= 2
# all other cases
else:
sells = self.market.history.sell_orders.average(good, 1)
buys = self.market.history.buy_orders.average(good, 1)
# TODO: figure out why this is sometimes 0
if sells + buys > 0:
supply_vs_demand = (sells - buys) / (sells + buys)
if supply_vs_demand > SIG_IMBALANCE or supply_vs_demand < -SIG_IMBALANCE:
# too much supply? lower bid lower to sell faster
# too much demand? raise price to buy faster
new_mean = public_mean_price * (1 - supply_vs_demand)
delta_to_mean = mean - new_mean
# shift the price belief to the new price mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# decrease belief's certainty since we've just changed it (we could be wrong)
belief.low -= wobble * mean
belief.high += wobble * mean
# make sure the price belief doesn't decrease below the minimum
if belief.low < MIN_PRICE:
belief.low = MIN_PRICE
elif belief.high < MIN_PRICE:
belief.high = MIN_PRICE
# Python utility methods
def __repr__(self):
return "<Pop: id={} type={}>".format(self.id, self.pop_job.title)
def __eq__(self, other):
return self.id == other.id
def __key__(self):
return self.id
def __hash__(self):
return hash(self.__key__())
def export(self):
model = {
'pop_job': self.pop_job.ref(),
'population': self.population,
'population_yesterday': self.population_yesterday,
'inventory': self.inventory.export(),
'money': self.money,
'money_yesterday': self.money_yesterday,
'successful_trades': self.successful_trades,
'failed_trades': self.failed_trades,
'bankrupt_times': self.bankrupt_times,
}
if self.pop_job is PopJob.merchant:
location_id = None
if self.trade_location:
location_id = self.trade_location.id
model.update({
'location': self.location.id,
'trade_location': location_id,
'trade_good': self.trade_good,
'trade_amount': self.trade_amount
})
return model
| __init__ | identifier_name |
pop.py | import math
from historia.utils import unique_id, position_in_range
from historia.pops.models.inventory import Inventory
from historia.economy.enums.resource import Good, NaturalResource
from historia.economy.enums.order_type import OrderType
from historia.economy.models.price_range import PriceRange
from historia.economy.models.order import Order
from historia.pops.enums.pop_job import PopJob
DEBUG = False
class Pop(object):
"""
A simulated unit of population
"""
def __init__(self, province, pop_job, population):
"""
Creates a new Pop.
manager (Historia)
province (SecondaryDivision)
culture (Culture)
religion (Religion)
language (Language)
job (Job)
"""
self.bankrupt_times = 0
self.home = province
self.location = province
self.id = unique_id('po')
self.population = population
self.population_yesterday = 0
self.pop_job = pop_job
# ECONOMY
self.money = pop_job.start_money
self.money_yesterday = 0
self.bankrupt = False
# set inventory and ideal amounts
self.inventory = Inventory(pop_job.inventory_size)
self.give_start_inventory()
self.update_ideal_inventory()
# a dictionary of Goods to PriceRanges
# represents the price range the agent considers valid for each Good
self.price_belief = {}
# a dictionary of Goods to price list
# represents the prices of the good that the Pop has observed
# during the time they have been trading
self.observed_trading_range = {}
self.successful_trades = 0
self.failed_trades = 0
# make some fake initial data
for good in Good.all():
avg_price = self.market.avg_historial_price(good, 15)
# fake trades
self.observed_trading_range[good] = [
avg_price * 0.5,
avg_price * 1.5
]
# generate fake price belief
self.price_belief[good] = PriceRange(avg_price * 0.5, avg_price * 1.5)
# Merchant logic
self.trade_location = None # the province this Pop is traveling to
self.trade_good = None # what good we're trading in right now
self.trade_amount = 0 # amount of trade_good we should be trading
self.trading_days = 0 # number of days waiting to trade
# Generic Pop properties
@property
def social_class(self):
return self.pop_job.social_class
@property
def market(self):
"Get the market instance"
return self.location.market
@property
def profit(self):
"Determine today's profit"
return self.money - self.money_yesterday
@property
def total_trades(self):
"Total number of trades this Pop participated in"
return self.successful_trades + self.failed_trades
@property
def trade_success(self):
"Percent of trades that were successful"
if self.total_trades == 0:
return 0
return (self.successful_trades / self.total_trades) * 100
@property
def is_away(self):
"Is this Pop away from it's home?"
return self.home is not self.location
# Merchant specific logic
def go_to_province(self, province):
"Moves the Pop to another Province"
self.location = province
def decide_trade_plan(self):
"""
Decide what good to trade in and how much.
Look for the most in demand good, or the most expensive good at the home Province
Find a province near home province where its the cheapest and there's inventory
"""
self.trade_amount = 5
most_demanded_goods = self.home.market.goods_demand_ratio(day_range=1)
most_demanded_goods = sorted(most_demanded_goods.items(), key=lambda i: i[1], reverse=True)
# if we already had a trade good, refresh ideal inventory
if self.trade_good:
self.update_ideal_inventory()
if DEBUG: print("Finding a Good to trade:")
for good, demand in most_demanded_goods:
if demand > 0:
# find nearby provinces where this has inventory and the price is lower
price_at_home = self.home.market.mean_price(good)
if DEBUG: print("Good: {}, Demand: {}, Price: ${}".format(good.title, demand, price_at_home))
neighboring_markets = [p.market for p in self.location.owned_neighbors]
neighboring_markets = [m for m in neighboring_markets if m.supply_for(good) > self.trade_amount]
neighboring_markets.sort(key=lambda m: m.supply_for(good), reverse=True)
if len(neighboring_markets) > 0:
# we found places where this good is cheaper and in inventory
target = neighboring_markets[0].location
price_at_target = target.market.mean_price(good)
# only trade with prices where we can make money
if price_at_home > price_at_target:
offset = 0
if good is Good.bread:
offset = 1
self.inventory.set_ideal(good, self.trade_amount + offset)
self.trade_location = target
if DEBUG:
print("\tTarget: {}, Supply: {}, Price: ${}, Price at home: ${}".format(
self.trade_location.name,
self.trade_location.market.supply_for(good),
self.trade_location.market.mean_price(good),
price_at_home)
)
self.trade_good = good
return
else:
if DEBUG: print("\tPrice is higher at target (home: ${} target: ${})".format(price_at_home, price_at_target))
else:
if DEBUG: print("\tNo markets selling {} found".format(good))
# Generic economic logic
def update_ideal_inventory(self):
"Update ideal inventory"
# reset so that the Pop can sell the inventory it doesn't need
for good in Good.all():
self.inventory.set_ideal(good, 0)
# update ideal inventory for new Job
for item in self.pop_job.ideal_inventory:
self.inventory.set_ideal(item['good'], item['amount'])
def give_start_inventory(self):
"Give the Pop the inventory it needs to do its job"
for item in self.pop_job.start_inventory:
self.inventory.add(item['good'], item['amount'])
def change_population(self, trade_success):
"Change the population based off the trade"
self.population_yesterday = self.population
if trade_success:
self.population += round(self.population * 0.01)
else:
self.population -= round(self.population * 0.002)
def handle_bankruptcy(self, pop_job):
"Change job, create money out of thin air, update ideal inventory"
# TODO: stop creating money out of thin air
self.pop_job = pop_job
self.bankrupt_times += 1
self.money = 2
self.update_ideal_inventory()
self.give_start_inventory()
def perform_logic(self):
"Depending on PopJob, perform logic (including production)"
logic = self.pop_job.logic(self)
logic.perform()
def create_buy_order(self, good, limit):
"Create a buy order for a given Good at a determined quantity"
bid_price = self.determine_price_of(good)
ideal = self.determine_buy_quantity(good)
# can't buy more than limit
quantity_to_buy = limit if ideal > limit else ideal
if quantity_to_buy > 0:
return Order(self, OrderType.buy_order, quantity_to_buy, bid_price, good)
return False
def create_sell_order(self, good, limit):
"Create a sell order for a given Good at a determined quantity"
sell_price = self.determine_price_of(good)
ideal = self.determine_sell_quantity(good)
# can't buy more than limit
quantity_to_sell = limit if ideal < limit else ideal
if quantity_to_sell > 0:
return Order(self, OrderType.sell_order, quantity_to_sell, sell_price, good)
return False
def price_belief_for(self, good):
"Gets the price belief this agent has for a particular Good"
if good in self.price_belief:
return self.price_belief[good]
def determine_price_of(self, good):
"Determine the price of a particular good"
return self.price_belief_for(good).random()
def trading_range_extremes(self, good):
"Gets the lowest and highst price of a Good this agent has seen"
trading_range = self.observed_trading_range[good]
return PriceRange(min(trading_range), max(trading_range))
def determine_sell_quantity(self, good):
"Determine how much inventory goods to sell based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = position_in_range(mean, trading_range.low, trading_range.high)
amount_to_sell = round(favoribility * self.inventory.surplus(good))
if amount_to_sell < 1:
amount_to_sell = 1
return amount_to_sell
def determine_buy_quantity(self, good):
|
def generate_orders(self, good):
"""
If the Pop needs a Good to perform production, buy it
If the Pop has surplus Resources, sell them
"""
surplus = self.inventory.surplus(good)
if surplus >= 1: # sell inventory
# the original only old one item here
sell_amount = surplus
order = self.create_sell_order(good, surplus)
if order:
# print('{} sells {} {}'.format(self.pop_job.title, sell_amount, good.name))
self.market.sell(order)
else: # buy more
shortage = self.inventory.shortage(good)
free_space = self.inventory.empty_space
if shortage > 0:
if shortage <= free_space:
# enough space for ideal order
limit = shortage
else:
# not enough space for ideal order
limit = math.floor(free_space / shortage)
if limit > 0:
order = self.create_buy_order(good, limit)
if order:
# print('{} buys {} {}'.format(self.pop_job.title, limit, good.name))
self.market.buy(order)
# else:
# print("{} has no shortage of {} (has shortage: {})".format(self.pop_job.title, good.title, shortage))
def update_price_model(self, good, order_type, is_successful, clearing_price=0):
"""
Update the Pop's price model for the given resource
good (Good) The Good which was orderd
order_type (OrderType) Which kind of Order this was
is_successful (bool) whether or not the Order was successful
clearing_price (float) The price per unit of the good that was ordered
as defined by the Pop which ordered it
"""
SIGNIFICANT = 0.25 # 25% more or less is "significant"
SIG_IMBALANCE = 0.33
LOW_INVENTORY = 0.1 # 10% of ideal inventory = "LOW"
HIGH_INVENTORY = 2.0 # 200% of ideal inventory = "HIGH"
MIN_PRICE = 0.01 # lowest allowed price of a Good
if is_successful:
# add this trade to the observed trading range
self.observed_trading_range[good].append(clearing_price)
public_mean_price = self.market.mean_price(good)
belief = self.price_belief[good]
mean = belief.mean()
wobble = 0.05 # the degree which the Pop should bid outside the belief
# how different the public mean price is from the price belief
delta_to_mean = mean - public_mean_price
if is_successful:
if order_type is OrderType.buy_order and delta_to_mean > SIGNIFICANT:
# this Pop overpaid, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
elif order_type is OrderType.sell_order and delta_to_mean < -SIGNIFICANT:
# this Pop underpaid!, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# increase the belief's certainty
belief.low += wobble * mean
belief.high -= wobble * mean
else:
# shift towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# check for inventory special cases
stocks = self.inventory.get_amount(good)
ideal = self.inventory.get_ideal(good)
# if we're buying and inventory is too low
# meaning we're desperate to buy
if order_type is OrderType.buy_order and stocks < LOW_INVENTORY * ideal:
wobble *= 2
# if we're selling and inventory is too high
# meaning we're desperate to sell
elif order_type is OrderType.sell_order and stocks > HIGH_INVENTORY * ideal:
wobble *= 2
# all other cases
else:
sells = self.market.history.sell_orders.average(good, 1)
buys = self.market.history.buy_orders.average(good, 1)
# TODO: figure out why this is sometimes 0
if sells + buys > 0:
supply_vs_demand = (sells - buys) / (sells + buys)
if supply_vs_demand > SIG_IMBALANCE or supply_vs_demand < -SIG_IMBALANCE:
# too much supply? lower bid lower to sell faster
# too much demand? raise price to buy faster
new_mean = public_mean_price * (1 - supply_vs_demand)
delta_to_mean = mean - new_mean
# shift the price belief to the new price mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# decrease belief's certainty since we've just changed it (we could be wrong)
belief.low -= wobble * mean
belief.high += wobble * mean
# make sure the price belief doesn't decrease below the minimum
if belief.low < MIN_PRICE:
belief.low = MIN_PRICE
elif belief.high < MIN_PRICE:
belief.high = MIN_PRICE
# Python utility methods
def __repr__(self):
return "<Pop: id={} type={}>".format(self.id, self.pop_job.title)
def __eq__(self, other):
return self.id == other.id
def __key__(self):
return self.id
def __hash__(self):
return hash(self.__key__())
def export(self):
model = {
'pop_job': self.pop_job.ref(),
'population': self.population,
'population_yesterday': self.population_yesterday,
'inventory': self.inventory.export(),
'money': self.money,
'money_yesterday': self.money_yesterday,
'successful_trades': self.successful_trades,
'failed_trades': self.failed_trades,
'bankrupt_times': self.bankrupt_times,
}
if self.pop_job is PopJob.merchant:
location_id = None
if self.trade_location:
location_id = self.trade_location.id
model.update({
'location': self.location.id,
'trade_location': location_id,
'trade_good': self.trade_good,
'trade_amount': self.trade_amount
})
return model
| "Determine how much goods to buy based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = 1 - position_in_range(mean, trading_range.low, trading_range.high)
amount_to_buy = round(favoribility * self.inventory.shortage(good))
if amount_to_buy < 1:
amount_to_buy = 1
return amount_to_buy | identifier_body |
pop.py | import math
from historia.utils import unique_id, position_in_range
from historia.pops.models.inventory import Inventory
from historia.economy.enums.resource import Good, NaturalResource
from historia.economy.enums.order_type import OrderType
from historia.economy.models.price_range import PriceRange
from historia.economy.models.order import Order
from historia.pops.enums.pop_job import PopJob
DEBUG = False
class Pop(object):
"""
A simulated unit of population
"""
def __init__(self, province, pop_job, population):
"""
Creates a new Pop.
manager (Historia)
province (SecondaryDivision)
culture (Culture)
religion (Religion)
language (Language)
job (Job)
"""
self.bankrupt_times = 0
self.home = province
self.location = province
self.id = unique_id('po')
self.population = population
self.population_yesterday = 0
self.pop_job = pop_job
# ECONOMY
self.money = pop_job.start_money
self.money_yesterday = 0
self.bankrupt = False
# set inventory and ideal amounts
self.inventory = Inventory(pop_job.inventory_size)
self.give_start_inventory()
self.update_ideal_inventory()
# a dictionary of Goods to PriceRanges
# represents the price range the agent considers valid for each Good
self.price_belief = {}
# a dictionary of Goods to price list
# represents the prices of the good that the Pop has observed
# during the time they have been trading
self.observed_trading_range = {}
self.successful_trades = 0
self.failed_trades = 0
# make some fake initial data
for good in Good.all():
avg_price = self.market.avg_historial_price(good, 15)
# fake trades
self.observed_trading_range[good] = [
avg_price * 0.5,
avg_price * 1.5
]
# generate fake price belief
self.price_belief[good] = PriceRange(avg_price * 0.5, avg_price * 1.5)
# Merchant logic
self.trade_location = None # the province this Pop is traveling to
self.trade_good = None # what good we're trading in right now
self.trade_amount = 0 # amount of trade_good we should be trading
self.trading_days = 0 # number of days waiting to trade
# Generic Pop properties
@property
def social_class(self):
return self.pop_job.social_class
@property
def market(self):
"Get the market instance"
return self.location.market
@property
def profit(self):
"Determine today's profit"
return self.money - self.money_yesterday
@property
def total_trades(self):
"Total number of trades this Pop participated in"
return self.successful_trades + self.failed_trades
@property
def trade_success(self):
"Percent of trades that were successful"
if self.total_trades == 0:
return 0
return (self.successful_trades / self.total_trades) * 100
@property
def is_away(self):
"Is this Pop away from it's home?"
return self.home is not self.location
# Merchant specific logic
def go_to_province(self, province):
"Moves the Pop to another Province"
self.location = province
def decide_trade_plan(self):
"""
Decide what good to trade in and how much.
Look for the most in demand good, or the most expensive good at the home Province
Find a province near home province where its the cheapest and there's inventory
"""
self.trade_amount = 5
most_demanded_goods = self.home.market.goods_demand_ratio(day_range=1)
most_demanded_goods = sorted(most_demanded_goods.items(), key=lambda i: i[1], reverse=True)
# if we already had a trade good, refresh ideal inventory
if self.trade_good:
self.update_ideal_inventory()
if DEBUG: print("Finding a Good to trade:")
for good, demand in most_demanded_goods:
if demand > 0:
# find nearby provinces where this has inventory and the price is lower
price_at_home = self.home.market.mean_price(good)
if DEBUG: |
neighboring_markets = [p.market for p in self.location.owned_neighbors]
neighboring_markets = [m for m in neighboring_markets if m.supply_for(good) > self.trade_amount]
neighboring_markets.sort(key=lambda m: m.supply_for(good), reverse=True)
if len(neighboring_markets) > 0:
# we found places where this good is cheaper and in inventory
target = neighboring_markets[0].location
price_at_target = target.market.mean_price(good)
# only trade with prices where we can make money
if price_at_home > price_at_target:
offset = 0
if good is Good.bread:
offset = 1
self.inventory.set_ideal(good, self.trade_amount + offset)
self.trade_location = target
if DEBUG:
print("\tTarget: {}, Supply: {}, Price: ${}, Price at home: ${}".format(
self.trade_location.name,
self.trade_location.market.supply_for(good),
self.trade_location.market.mean_price(good),
price_at_home)
)
self.trade_good = good
return
else:
if DEBUG: print("\tPrice is higher at target (home: ${} target: ${})".format(price_at_home, price_at_target))
else:
if DEBUG: print("\tNo markets selling {} found".format(good))
# Generic economic logic
def update_ideal_inventory(self):
"Update ideal inventory"
# reset so that the Pop can sell the inventory it doesn't need
for good in Good.all():
self.inventory.set_ideal(good, 0)
# update ideal inventory for new Job
for item in self.pop_job.ideal_inventory:
self.inventory.set_ideal(item['good'], item['amount'])
def give_start_inventory(self):
"Give the Pop the inventory it needs to do its job"
for item in self.pop_job.start_inventory:
self.inventory.add(item['good'], item['amount'])
def change_population(self, trade_success):
"Change the population based off the trade"
self.population_yesterday = self.population
if trade_success:
self.population += round(self.population * 0.01)
else:
self.population -= round(self.population * 0.002)
def handle_bankruptcy(self, pop_job):
"Change job, create money out of thin air, update ideal inventory"
# TODO: stop creating money out of thin air
self.pop_job = pop_job
self.bankrupt_times += 1
self.money = 2
self.update_ideal_inventory()
self.give_start_inventory()
def perform_logic(self):
"Depending on PopJob, perform logic (including production)"
logic = self.pop_job.logic(self)
logic.perform()
def create_buy_order(self, good, limit):
"Create a buy order for a given Good at a determined quantity"
bid_price = self.determine_price_of(good)
ideal = self.determine_buy_quantity(good)
# can't buy more than limit
quantity_to_buy = limit if ideal > limit else ideal
if quantity_to_buy > 0:
return Order(self, OrderType.buy_order, quantity_to_buy, bid_price, good)
return False
def create_sell_order(self, good, limit):
"Create a sell order for a given Good at a determined quantity"
sell_price = self.determine_price_of(good)
ideal = self.determine_sell_quantity(good)
# can't buy more than limit
quantity_to_sell = limit if ideal < limit else ideal
if quantity_to_sell > 0:
return Order(self, OrderType.sell_order, quantity_to_sell, sell_price, good)
return False
def price_belief_for(self, good):
"Gets the price belief this agent has for a particular Good"
if good in self.price_belief:
return self.price_belief[good]
def determine_price_of(self, good):
"Determine the price of a particular good"
return self.price_belief_for(good).random()
def trading_range_extremes(self, good):
"Gets the lowest and highst price of a Good this agent has seen"
trading_range = self.observed_trading_range[good]
return PriceRange(min(trading_range), max(trading_range))
def determine_sell_quantity(self, good):
"Determine how much inventory goods to sell based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = position_in_range(mean, trading_range.low, trading_range.high)
amount_to_sell = round(favoribility * self.inventory.surplus(good))
if amount_to_sell < 1:
amount_to_sell = 1
return amount_to_sell
def determine_buy_quantity(self, good):
"Determine how much goods to buy based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = 1 - position_in_range(mean, trading_range.low, trading_range.high)
amount_to_buy = round(favoribility * self.inventory.shortage(good))
if amount_to_buy < 1:
amount_to_buy = 1
return amount_to_buy
def generate_orders(self, good):
"""
If the Pop needs a Good to perform production, buy it
If the Pop has surplus Resources, sell them
"""
surplus = self.inventory.surplus(good)
if surplus >= 1: # sell inventory
# the original only old one item here
sell_amount = surplus
order = self.create_sell_order(good, surplus)
if order:
# print('{} sells {} {}'.format(self.pop_job.title, sell_amount, good.name))
self.market.sell(order)
else: # buy more
shortage = self.inventory.shortage(good)
free_space = self.inventory.empty_space
if shortage > 0:
if shortage <= free_space:
# enough space for ideal order
limit = shortage
else:
# not enough space for ideal order
limit = math.floor(free_space / shortage)
if limit > 0:
order = self.create_buy_order(good, limit)
if order:
# print('{} buys {} {}'.format(self.pop_job.title, limit, good.name))
self.market.buy(order)
# else:
# print("{} has no shortage of {} (has shortage: {})".format(self.pop_job.title, good.title, shortage))
def update_price_model(self, good, order_type, is_successful, clearing_price=0):
"""
Update the Pop's price model for the given resource
good (Good) The Good which was orderd
order_type (OrderType) Which kind of Order this was
is_successful (bool) whether or not the Order was successful
clearing_price (float) The price per unit of the good that was ordered
as defined by the Pop which ordered it
"""
SIGNIFICANT = 0.25 # 25% more or less is "significant"
SIG_IMBALANCE = 0.33
LOW_INVENTORY = 0.1 # 10% of ideal inventory = "LOW"
HIGH_INVENTORY = 2.0 # 200% of ideal inventory = "HIGH"
MIN_PRICE = 0.01 # lowest allowed price of a Good
if is_successful:
# add this trade to the observed trading range
self.observed_trading_range[good].append(clearing_price)
public_mean_price = self.market.mean_price(good)
belief = self.price_belief[good]
mean = belief.mean()
wobble = 0.05 # the degree which the Pop should bid outside the belief
# how different the public mean price is from the price belief
delta_to_mean = mean - public_mean_price
if is_successful:
if order_type is OrderType.buy_order and delta_to_mean > SIGNIFICANT:
# this Pop overpaid, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
elif order_type is OrderType.sell_order and delta_to_mean < -SIGNIFICANT:
# this Pop underpaid!, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# increase the belief's certainty
belief.low += wobble * mean
belief.high -= wobble * mean
else:
# shift towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# check for inventory special cases
stocks = self.inventory.get_amount(good)
ideal = self.inventory.get_ideal(good)
# if we're buying and inventory is too low
# meaning we're desperate to buy
if order_type is OrderType.buy_order and stocks < LOW_INVENTORY * ideal:
wobble *= 2
# if we're selling and inventory is too high
# meaning we're desperate to sell
elif order_type is OrderType.sell_order and stocks > HIGH_INVENTORY * ideal:
wobble *= 2
# all other cases
else:
sells = self.market.history.sell_orders.average(good, 1)
buys = self.market.history.buy_orders.average(good, 1)
# TODO: figure out why this is sometimes 0
if sells + buys > 0:
supply_vs_demand = (sells - buys) / (sells + buys)
if supply_vs_demand > SIG_IMBALANCE or supply_vs_demand < -SIG_IMBALANCE:
# too much supply? lower bid lower to sell faster
# too much demand? raise price to buy faster
new_mean = public_mean_price * (1 - supply_vs_demand)
delta_to_mean = mean - new_mean
# shift the price belief to the new price mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# decrease belief's certainty since we've just changed it (we could be wrong)
belief.low -= wobble * mean
belief.high += wobble * mean
# make sure the price belief doesn't decrease below the minimum
if belief.low < MIN_PRICE:
belief.low = MIN_PRICE
elif belief.high < MIN_PRICE:
belief.high = MIN_PRICE
# Python utility methods
def __repr__(self):
return "<Pop: id={} type={}>".format(self.id, self.pop_job.title)
def __eq__(self, other):
return self.id == other.id
def __key__(self):
return self.id
def __hash__(self):
return hash(self.__key__())
def export(self):
model = {
'pop_job': self.pop_job.ref(),
'population': self.population,
'population_yesterday': self.population_yesterday,
'inventory': self.inventory.export(),
'money': self.money,
'money_yesterday': self.money_yesterday,
'successful_trades': self.successful_trades,
'failed_trades': self.failed_trades,
'bankrupt_times': self.bankrupt_times,
}
if self.pop_job is PopJob.merchant:
location_id = None
if self.trade_location:
location_id = self.trade_location.id
model.update({
'location': self.location.id,
'trade_location': location_id,
'trade_good': self.trade_good,
'trade_amount': self.trade_amount
})
return model
| print("Good: {}, Demand: {}, Price: ${}".format(good.title, demand, price_at_home)) | conditional_block |
pop.py | import math
from historia.utils import unique_id, position_in_range
from historia.pops.models.inventory import Inventory
from historia.economy.enums.resource import Good, NaturalResource
from historia.economy.enums.order_type import OrderType
from historia.economy.models.price_range import PriceRange
from historia.economy.models.order import Order
from historia.pops.enums.pop_job import PopJob
DEBUG = False
class Pop(object):
"""
A simulated unit of population
"""
def __init__(self, province, pop_job, population):
"""
Creates a new Pop.
manager (Historia)
province (SecondaryDivision)
culture (Culture)
religion (Religion)
language (Language)
job (Job)
"""
self.bankrupt_times = 0
self.home = province
self.location = province
self.id = unique_id('po')
self.population = population
self.population_yesterday = 0
self.pop_job = pop_job
# ECONOMY
self.money = pop_job.start_money
self.money_yesterday = 0
self.bankrupt = False
# set inventory and ideal amounts
self.inventory = Inventory(pop_job.inventory_size)
self.give_start_inventory()
self.update_ideal_inventory()
# a dictionary of Goods to PriceRanges | self.price_belief = {}
# a dictionary of Goods to price list
# represents the prices of the good that the Pop has observed
# during the time they have been trading
self.observed_trading_range = {}
self.successful_trades = 0
self.failed_trades = 0
# make some fake initial data
for good in Good.all():
avg_price = self.market.avg_historial_price(good, 15)
# fake trades
self.observed_trading_range[good] = [
avg_price * 0.5,
avg_price * 1.5
]
# generate fake price belief
self.price_belief[good] = PriceRange(avg_price * 0.5, avg_price * 1.5)
# Merchant logic
self.trade_location = None # the province this Pop is traveling to
self.trade_good = None # what good we're trading in right now
self.trade_amount = 0 # amount of trade_good we should be trading
self.trading_days = 0 # number of days waiting to trade
# Generic Pop properties
@property
def social_class(self):
return self.pop_job.social_class
@property
def market(self):
"Get the market instance"
return self.location.market
@property
def profit(self):
"Determine today's profit"
return self.money - self.money_yesterday
@property
def total_trades(self):
"Total number of trades this Pop participated in"
return self.successful_trades + self.failed_trades
@property
def trade_success(self):
"Percent of trades that were successful"
if self.total_trades == 0:
return 0
return (self.successful_trades / self.total_trades) * 100
@property
def is_away(self):
"Is this Pop away from it's home?"
return self.home is not self.location
# Merchant specific logic
def go_to_province(self, province):
"Moves the Pop to another Province"
self.location = province
def decide_trade_plan(self):
"""
Decide what good to trade in and how much.
Look for the most in demand good, or the most expensive good at the home Province
Find a province near home province where its the cheapest and there's inventory
"""
self.trade_amount = 5
most_demanded_goods = self.home.market.goods_demand_ratio(day_range=1)
most_demanded_goods = sorted(most_demanded_goods.items(), key=lambda i: i[1], reverse=True)
# if we already had a trade good, refresh ideal inventory
if self.trade_good:
self.update_ideal_inventory()
if DEBUG: print("Finding a Good to trade:")
for good, demand in most_demanded_goods:
if demand > 0:
# find nearby provinces where this has inventory and the price is lower
price_at_home = self.home.market.mean_price(good)
if DEBUG: print("Good: {}, Demand: {}, Price: ${}".format(good.title, demand, price_at_home))
neighboring_markets = [p.market for p in self.location.owned_neighbors]
neighboring_markets = [m for m in neighboring_markets if m.supply_for(good) > self.trade_amount]
neighboring_markets.sort(key=lambda m: m.supply_for(good), reverse=True)
if len(neighboring_markets) > 0:
# we found places where this good is cheaper and in inventory
target = neighboring_markets[0].location
price_at_target = target.market.mean_price(good)
# only trade with prices where we can make money
if price_at_home > price_at_target:
offset = 0
if good is Good.bread:
offset = 1
self.inventory.set_ideal(good, self.trade_amount + offset)
self.trade_location = target
if DEBUG:
print("\tTarget: {}, Supply: {}, Price: ${}, Price at home: ${}".format(
self.trade_location.name,
self.trade_location.market.supply_for(good),
self.trade_location.market.mean_price(good),
price_at_home)
)
self.trade_good = good
return
else:
if DEBUG: print("\tPrice is higher at target (home: ${} target: ${})".format(price_at_home, price_at_target))
else:
if DEBUG: print("\tNo markets selling {} found".format(good))
# Generic economic logic
def update_ideal_inventory(self):
"Update ideal inventory"
# reset so that the Pop can sell the inventory it doesn't need
for good in Good.all():
self.inventory.set_ideal(good, 0)
# update ideal inventory for new Job
for item in self.pop_job.ideal_inventory:
self.inventory.set_ideal(item['good'], item['amount'])
def give_start_inventory(self):
"Give the Pop the inventory it needs to do its job"
for item in self.pop_job.start_inventory:
self.inventory.add(item['good'], item['amount'])
def change_population(self, trade_success):
"Change the population based off the trade"
self.population_yesterday = self.population
if trade_success:
self.population += round(self.population * 0.01)
else:
self.population -= round(self.population * 0.002)
def handle_bankruptcy(self, pop_job):
"Change job, create money out of thin air, update ideal inventory"
# TODO: stop creating money out of thin air
self.pop_job = pop_job
self.bankrupt_times += 1
self.money = 2
self.update_ideal_inventory()
self.give_start_inventory()
def perform_logic(self):
"Depending on PopJob, perform logic (including production)"
logic = self.pop_job.logic(self)
logic.perform()
def create_buy_order(self, good, limit):
"Create a buy order for a given Good at a determined quantity"
bid_price = self.determine_price_of(good)
ideal = self.determine_buy_quantity(good)
# can't buy more than limit
quantity_to_buy = limit if ideal > limit else ideal
if quantity_to_buy > 0:
return Order(self, OrderType.buy_order, quantity_to_buy, bid_price, good)
return False
def create_sell_order(self, good, limit):
"Create a sell order for a given Good at a determined quantity"
sell_price = self.determine_price_of(good)
ideal = self.determine_sell_quantity(good)
# can't buy more than limit
quantity_to_sell = limit if ideal < limit else ideal
if quantity_to_sell > 0:
return Order(self, OrderType.sell_order, quantity_to_sell, sell_price, good)
return False
def price_belief_for(self, good):
"Gets the price belief this agent has for a particular Good"
if good in self.price_belief:
return self.price_belief[good]
def determine_price_of(self, good):
"Determine the price of a particular good"
return self.price_belief_for(good).random()
def trading_range_extremes(self, good):
"Gets the lowest and highst price of a Good this agent has seen"
trading_range = self.observed_trading_range[good]
return PriceRange(min(trading_range), max(trading_range))
def determine_sell_quantity(self, good):
"Determine how much inventory goods to sell based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = position_in_range(mean, trading_range.low, trading_range.high)
amount_to_sell = round(favoribility * self.inventory.surplus(good))
if amount_to_sell < 1:
amount_to_sell = 1
return amount_to_sell
def determine_buy_quantity(self, good):
"Determine how much goods to buy based on market conditions"
mean = self.market.avg_historial_price(good, 15)
trading_range = self.trading_range_extremes(good)
favoribility = 1 - position_in_range(mean, trading_range.low, trading_range.high)
amount_to_buy = round(favoribility * self.inventory.shortage(good))
if amount_to_buy < 1:
amount_to_buy = 1
return amount_to_buy
def generate_orders(self, good):
"""
If the Pop needs a Good to perform production, buy it
If the Pop has surplus Resources, sell them
"""
surplus = self.inventory.surplus(good)
if surplus >= 1: # sell inventory
# the original only old one item here
sell_amount = surplus
order = self.create_sell_order(good, surplus)
if order:
# print('{} sells {} {}'.format(self.pop_job.title, sell_amount, good.name))
self.market.sell(order)
else: # buy more
shortage = self.inventory.shortage(good)
free_space = self.inventory.empty_space
if shortage > 0:
if shortage <= free_space:
# enough space for ideal order
limit = shortage
else:
# not enough space for ideal order
limit = math.floor(free_space / shortage)
if limit > 0:
order = self.create_buy_order(good, limit)
if order:
# print('{} buys {} {}'.format(self.pop_job.title, limit, good.name))
self.market.buy(order)
# else:
# print("{} has no shortage of {} (has shortage: {})".format(self.pop_job.title, good.title, shortage))
def update_price_model(self, good, order_type, is_successful, clearing_price=0):
"""
Update the Pop's price model for the given resource
good (Good) The Good which was orderd
order_type (OrderType) Which kind of Order this was
is_successful (bool) whether or not the Order was successful
clearing_price (float) The price per unit of the good that was ordered
as defined by the Pop which ordered it
"""
SIGNIFICANT = 0.25 # 25% more or less is "significant"
SIG_IMBALANCE = 0.33
LOW_INVENTORY = 0.1 # 10% of ideal inventory = "LOW"
HIGH_INVENTORY = 2.0 # 200% of ideal inventory = "HIGH"
MIN_PRICE = 0.01 # lowest allowed price of a Good
if is_successful:
# add this trade to the observed trading range
self.observed_trading_range[good].append(clearing_price)
public_mean_price = self.market.mean_price(good)
belief = self.price_belief[good]
mean = belief.mean()
wobble = 0.05 # the degree which the Pop should bid outside the belief
# how different the public mean price is from the price belief
delta_to_mean = mean - public_mean_price
if is_successful:
if order_type is OrderType.buy_order and delta_to_mean > SIGNIFICANT:
# this Pop overpaid, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
elif order_type is OrderType.sell_order and delta_to_mean < -SIGNIFICANT:
# this Pop underpaid!, shift belief towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# increase the belief's certainty
belief.low += wobble * mean
belief.high -= wobble * mean
else:
# shift towards mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# check for inventory special cases
stocks = self.inventory.get_amount(good)
ideal = self.inventory.get_ideal(good)
# if we're buying and inventory is too low
# meaning we're desperate to buy
if order_type is OrderType.buy_order and stocks < LOW_INVENTORY * ideal:
wobble *= 2
# if we're selling and inventory is too high
# meaning we're desperate to sell
elif order_type is OrderType.sell_order and stocks > HIGH_INVENTORY * ideal:
wobble *= 2
# all other cases
else:
sells = self.market.history.sell_orders.average(good, 1)
buys = self.market.history.buy_orders.average(good, 1)
# TODO: figure out why this is sometimes 0
if sells + buys > 0:
supply_vs_demand = (sells - buys) / (sells + buys)
if supply_vs_demand > SIG_IMBALANCE or supply_vs_demand < -SIG_IMBALANCE:
# too much supply? lower bid lower to sell faster
# too much demand? raise price to buy faster
new_mean = public_mean_price * (1 - supply_vs_demand)
delta_to_mean = mean - new_mean
# shift the price belief to the new price mean
belief.low -= delta_to_mean / 2
belief.high -= delta_to_mean / 2
# decrease belief's certainty since we've just changed it (we could be wrong)
belief.low -= wobble * mean
belief.high += wobble * mean
# make sure the price belief doesn't decrease below the minimum
if belief.low < MIN_PRICE:
belief.low = MIN_PRICE
elif belief.high < MIN_PRICE:
belief.high = MIN_PRICE
# Python utility methods
def __repr__(self):
return "<Pop: id={} type={}>".format(self.id, self.pop_job.title)
def __eq__(self, other):
return self.id == other.id
def __key__(self):
return self.id
def __hash__(self):
return hash(self.__key__())
def export(self):
model = {
'pop_job': self.pop_job.ref(),
'population': self.population,
'population_yesterday': self.population_yesterday,
'inventory': self.inventory.export(),
'money': self.money,
'money_yesterday': self.money_yesterday,
'successful_trades': self.successful_trades,
'failed_trades': self.failed_trades,
'bankrupt_times': self.bankrupt_times,
}
if self.pop_job is PopJob.merchant:
location_id = None
if self.trade_location:
location_id = self.trade_location.id
model.update({
'location': self.location.id,
'trade_location': location_id,
'trade_good': self.trade_good,
'trade_amount': self.trade_amount
})
return model | # represents the price range the agent considers valid for each Good | random_line_split |
server.go | package reflector
import (
"bufio"
"crypto/sha512"
"encoding/hex"
"encoding/json"
"io"
"net"
"time"
"github.com/lbryio/reflector.go/internal/metrics"
"github.com/lbryio/reflector.go/store"
"github.com/lbryio/lbry.go/v2/extras/errors"
"github.com/lbryio/lbry.go/v2/extras/stop"
"github.com/lbryio/lbry.go/v2/stream"
log "github.com/sirupsen/logrus"
)
const (
// DefaultPort is the port the reflector server listens on if not passed in.
DefaultPort = 5566
// DefaultTimeout is the default timeout to read or write the next message
DefaultTimeout = 5 * time.Second
network = "tcp4"
protocolVersion1 = 0
protocolVersion2 = 1
maxBlobSize = stream.MaxBlobSize
)
var ErrBlobTooBig = errors.Base("blob must be at most %d bytes", maxBlobSize)
// Server is and instance of the reflector server. It houses the blob store and listener.
type Server struct {
Timeout time.Duration // timeout to read or write next message
EnableBlocklist bool // if true, blocklist checking and blob deletion will be enabled
underlyingStore store.BlobStore
outerStore store.BlobStore
grp *stop.Group
}
// NewServer returns an initialized reflector server pointer.
func NewServer(underlying store.BlobStore, outer store.BlobStore) *Server {
return &Server{
Timeout: DefaultTimeout,
underlyingStore: underlying,
outerStore: outer,
grp: stop.New(),
}
}
// Shutdown shuts down the reflector server gracefully.
func (s *Server) Shutdown() {
log.Println("shutting down reflector server...")
s.grp.StopAndWait()
log.Println("reflector server stopped")
}
// Start starts the server to handle connections.
func (s *Server) Start(address string) error {
l, err := net.Listen(network, address)
if err != nil {
return errors.Err(err)
}
log.Println("reflector listening on " + address)
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Dec()
<-s.grp.Ch()
err := l.Close()
if err != nil {
log.Error(errors.Prefix("closing listener", err))
}
s.grp.Done()
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "start").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "start").Dec()
s.listenAndServe(l)
s.grp.Done()
}()
if s.EnableBlocklist {
if b, ok := s.underlyingStore.(store.Blocklister); ok {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Dec()
s.enableBlocklist(b)
s.grp.Done()
}()
} else {
//s.Shutdown()
return errors.Err("blocklist is enabled but blob store does not support blocklisting")
}
}
return nil
}
func (s *Server) listenAndServe(listener net.Listener) {
for {
conn, err := listener.Accept()
if err != nil {
if s.quitting() {
return
}
log.Error(err)
} else {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Dec()
s.handleConn(conn)
s.grp.Done()
}()
}
}
}
func (s *Server) handleConn(conn net.Conn) {
// all this stuff is to close the connections correctly when we're shutting down the server
connNeedsClosing := make(chan struct{})
defer func() {
close(connNeedsClosing)
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Dec()
defer s.grp.Done()
select {
case <-connNeedsClosing:
case <-s.grp.Ch():
}
err := conn.Close()
if err != nil {
log.Error(errors.Prefix("closing peer conn", err))
}
}()
err := s.doHandshake(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending handshake error", err))
}
return
}
for {
err = s.receiveBlob(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending blob receive error", err))
}
return
}
}
}
func (s *Server) doError(conn net.Conn, err error) error {
if err == nil {
return nil
}
shouldLog := metrics.TrackError(metrics.DirectionUpload, err)
if shouldLog {
log.Errorln(errors.FullTrace(err))
}
if e2, ok := err.(*json.SyntaxError); ok {
log.Errorf("syntax error at byte offset %d", e2.Offset)
}
//resp, err := json.Marshal(errorResponse{Error: err.Error()})
//if err != nil {
// return err
//}
//return s.write(conn, resp)
return nil
}
func (s *Server) receiveBlob(conn net.Conn) error {
blobSize, blobHash, isSdBlob, err := s.readBlobRequest(conn)
if err != nil {
return err
}
var wantsBlob bool
if bl, ok := s.underlyingStore.(store.Blocklister); ok {
wantsBlob, err = bl.Wants(blobHash)
if err != nil {
return err
}
} else {
blobExists, err := s.underlyingStore.Has(blobHash)
if err != nil {
return err
}
wantsBlob = !blobExists
}
var neededBlobs []string
if isSdBlob && !wantsBlob {
if nbc, ok := s.underlyingStore.(neededBlobChecker); ok {
neededBlobs, err = nbc.MissingBlobsForKnownStream(blobHash)
if err != nil {
return err
}
} else {
// if we can't check for blobs in a stream, we have to say that the sd blob is
// missing. if we say we have the sd blob, they wont try to send any content blobs
wantsBlob = true
}
}
err = s.sendBlobResponse(conn, wantsBlob, isSdBlob, neededBlobs)
if err != nil {
return err
}
if !wantsBlob {
return nil
}
blob, err := s.readRawBlob(conn, blobSize)
if err != nil {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Prefix("error reading blob "+blobHash[:8], err)
}
receivedBlobHash := BlobHash(blob)
if blobHash != receivedBlobHash {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Err("hash of received blob data does not match hash from send request")
// this can also happen if the blob size is wrong, because the server will read the wrong number of bytes from the stream
}
log.Debugln("Got blob " + blobHash[:8])
if isSdBlob {
err = s.outerStore.PutSD(blobHash, blob)
} else {
err = s.outerStore.Put(blobHash, blob)
}
if err != nil {
return err
}
metrics.MtrInBytesReflector.Add(float64(len(blob)))
metrics.BlobUploadCount.Inc()
if isSdBlob {
metrics.SDBlobUploadCount.Inc()
}
return s.sendTransferResponse(conn, true, isSdBlob)
}
func (s *Server) doHandshake(conn net.Conn) error {
var handshake handshakeRequestResponse
err := s.read(conn, &handshake)
if err != nil | else if handshake.Version == nil {
return errors.Err("handshake is missing protocol version")
} else if *handshake.Version != protocolVersion1 && *handshake.Version != protocolVersion2 {
return errors.Err("protocol version not supported")
}
resp, err := json.Marshal(handshakeRequestResponse{Version: handshake.Version})
if err != nil {
return err
}
return s.write(conn, resp)
}
func (s *Server) readBlobRequest(conn net.Conn) (int, string, bool, error) {
var sendRequest sendBlobRequest
err := s.read(conn, &sendRequest)
if err != nil {
return 0, "", false, err
}
var blobHash string
var blobSize int
isSdBlob := sendRequest.SdBlobHash != ""
if isSdBlob {
blobSize = sendRequest.SdBlobSize
blobHash = sendRequest.SdBlobHash
} else {
blobSize = sendRequest.BlobSize
blobHash = sendRequest.BlobHash
}
if blobHash == "" {
return blobSize, blobHash, isSdBlob, errors.Err("blob hash is empty")
}
if blobSize > maxBlobSize {
return blobSize, blobHash, isSdBlob, errors.Err(ErrBlobTooBig)
}
if blobSize == 0 {
return blobSize, blobHash, isSdBlob, errors.Err("0-byte blob received")
}
return blobSize, blobHash, isSdBlob, nil
}
func (s *Server) sendBlobResponse(conn net.Conn, shouldSendBlob, isSdBlob bool, neededBlobs []string) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sendSdBlobResponse{SendSdBlob: shouldSendBlob, NeededBlobs: neededBlobs})
} else {
response, err = json.Marshal(sendBlobResponse{SendBlob: shouldSendBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) sendTransferResponse(conn net.Conn, receivedBlob, isSdBlob bool) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sdBlobTransferResponse{ReceivedSdBlob: receivedBlob})
} else {
response, err = json.Marshal(blobTransferResponse{ReceivedBlob: receivedBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) read(conn net.Conn, v interface{}) error {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
dec := json.NewDecoder(conn)
err = dec.Decode(v)
if err != nil {
data, _ := io.ReadAll(dec.Buffered())
if len(data) > 0 {
return errors.Err("%s. Data: %s", err.Error(), hex.EncodeToString(data))
}
return errors.Err(err)
}
return nil
}
func (s *Server) readRawBlob(conn net.Conn, blobSize int) ([]byte, error) {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return nil, errors.Err(err)
}
blob := make([]byte, blobSize)
_, err = io.ReadFull(bufio.NewReader(conn), blob)
return blob, errors.Err(err)
}
func (s *Server) write(conn net.Conn, b []byte) error {
err := conn.SetWriteDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
n, err := conn.Write(b)
if err == nil && n != len(b) {
err = io.ErrShortWrite
}
return errors.Err(err)
}
func (s *Server) quitting() bool {
select {
case <-s.grp.Ch():
return true
default:
return false
}
}
// BlobHash returns the sha512 hash hex encoded string of the blob byte slice.
func BlobHash(blob []byte) string {
hashBytes := sha512.Sum384(blob)
return hex.EncodeToString(hashBytes[:])
}
func IsValidJSON(b []byte) bool {
var r json.RawMessage
return json.Unmarshal(b, &r) == nil
}
//type errorResponse struct {
// Error string `json:"error"`
//}
type handshakeRequestResponse struct {
Version *int `json:"version"`
}
type sendBlobRequest struct {
BlobHash string `json:"blob_hash,omitempty"`
BlobSize int `json:"blob_size,omitempty"`
SdBlobHash string `json:"sd_blob_hash,omitempty"`
SdBlobSize int `json:"sd_blob_size,omitempty"`
}
type sendBlobResponse struct {
SendBlob bool `json:"send_blob"`
}
type sendSdBlobResponse struct {
SendSdBlob bool `json:"send_sd_blob"`
NeededBlobs []string `json:"needed_blobs,omitempty"`
}
type blobTransferResponse struct {
ReceivedBlob bool `json:"received_blob"`
}
type sdBlobTransferResponse struct {
ReceivedSdBlob bool `json:"received_sd_blob"`
}
// neededBlobChecker can check which blobs from a known stream are not uploaded yet
type neededBlobChecker interface {
MissingBlobsForKnownStream(string) ([]string, error)
}
| {
return err
} | conditional_block |
server.go | package reflector
import (
"bufio"
"crypto/sha512"
"encoding/hex"
"encoding/json"
"io"
"net"
"time"
"github.com/lbryio/reflector.go/internal/metrics"
"github.com/lbryio/reflector.go/store"
"github.com/lbryio/lbry.go/v2/extras/errors"
"github.com/lbryio/lbry.go/v2/extras/stop"
"github.com/lbryio/lbry.go/v2/stream"
log "github.com/sirupsen/logrus"
)
const (
// DefaultPort is the port the reflector server listens on if not passed in.
DefaultPort = 5566
// DefaultTimeout is the default timeout to read or write the next message
DefaultTimeout = 5 * time.Second
network = "tcp4"
protocolVersion1 = 0
protocolVersion2 = 1
maxBlobSize = stream.MaxBlobSize
)
var ErrBlobTooBig = errors.Base("blob must be at most %d bytes", maxBlobSize)
// Server is and instance of the reflector server. It houses the blob store and listener.
type Server struct {
Timeout time.Duration // timeout to read or write next message
EnableBlocklist bool // if true, blocklist checking and blob deletion will be enabled
underlyingStore store.BlobStore
outerStore store.BlobStore
grp *stop.Group
}
// NewServer returns an initialized reflector server pointer.
func NewServer(underlying store.BlobStore, outer store.BlobStore) *Server {
return &Server{
Timeout: DefaultTimeout,
underlyingStore: underlying,
outerStore: outer,
grp: stop.New(),
}
}
// Shutdown shuts down the reflector server gracefully.
func (s *Server) Shutdown() {
log.Println("shutting down reflector server...")
s.grp.StopAndWait()
log.Println("reflector server stopped")
}
// Start starts the server to handle connections.
func (s *Server) Start(address string) error {
l, err := net.Listen(network, address)
if err != nil {
return errors.Err(err)
}
log.Println("reflector listening on " + address)
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Dec()
<-s.grp.Ch()
err := l.Close()
if err != nil {
log.Error(errors.Prefix("closing listener", err))
}
s.grp.Done()
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "start").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "start").Dec()
s.listenAndServe(l)
s.grp.Done()
}()
if s.EnableBlocklist {
if b, ok := s.underlyingStore.(store.Blocklister); ok {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Dec()
s.enableBlocklist(b)
s.grp.Done()
}()
} else {
//s.Shutdown()
return errors.Err("blocklist is enabled but blob store does not support blocklisting")
}
}
return nil
}
func (s *Server) listenAndServe(listener net.Listener) |
func (s *Server) handleConn(conn net.Conn) {
// all this stuff is to close the connections correctly when we're shutting down the server
connNeedsClosing := make(chan struct{})
defer func() {
close(connNeedsClosing)
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Dec()
defer s.grp.Done()
select {
case <-connNeedsClosing:
case <-s.grp.Ch():
}
err := conn.Close()
if err != nil {
log.Error(errors.Prefix("closing peer conn", err))
}
}()
err := s.doHandshake(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending handshake error", err))
}
return
}
for {
err = s.receiveBlob(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending blob receive error", err))
}
return
}
}
}
func (s *Server) doError(conn net.Conn, err error) error {
if err == nil {
return nil
}
shouldLog := metrics.TrackError(metrics.DirectionUpload, err)
if shouldLog {
log.Errorln(errors.FullTrace(err))
}
if e2, ok := err.(*json.SyntaxError); ok {
log.Errorf("syntax error at byte offset %d", e2.Offset)
}
//resp, err := json.Marshal(errorResponse{Error: err.Error()})
//if err != nil {
// return err
//}
//return s.write(conn, resp)
return nil
}
func (s *Server) receiveBlob(conn net.Conn) error {
blobSize, blobHash, isSdBlob, err := s.readBlobRequest(conn)
if err != nil {
return err
}
var wantsBlob bool
if bl, ok := s.underlyingStore.(store.Blocklister); ok {
wantsBlob, err = bl.Wants(blobHash)
if err != nil {
return err
}
} else {
blobExists, err := s.underlyingStore.Has(blobHash)
if err != nil {
return err
}
wantsBlob = !blobExists
}
var neededBlobs []string
if isSdBlob && !wantsBlob {
if nbc, ok := s.underlyingStore.(neededBlobChecker); ok {
neededBlobs, err = nbc.MissingBlobsForKnownStream(blobHash)
if err != nil {
return err
}
} else {
// if we can't check for blobs in a stream, we have to say that the sd blob is
// missing. if we say we have the sd blob, they wont try to send any content blobs
wantsBlob = true
}
}
err = s.sendBlobResponse(conn, wantsBlob, isSdBlob, neededBlobs)
if err != nil {
return err
}
if !wantsBlob {
return nil
}
blob, err := s.readRawBlob(conn, blobSize)
if err != nil {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Prefix("error reading blob "+blobHash[:8], err)
}
receivedBlobHash := BlobHash(blob)
if blobHash != receivedBlobHash {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Err("hash of received blob data does not match hash from send request")
// this can also happen if the blob size is wrong, because the server will read the wrong number of bytes from the stream
}
log.Debugln("Got blob " + blobHash[:8])
if isSdBlob {
err = s.outerStore.PutSD(blobHash, blob)
} else {
err = s.outerStore.Put(blobHash, blob)
}
if err != nil {
return err
}
metrics.MtrInBytesReflector.Add(float64(len(blob)))
metrics.BlobUploadCount.Inc()
if isSdBlob {
metrics.SDBlobUploadCount.Inc()
}
return s.sendTransferResponse(conn, true, isSdBlob)
}
func (s *Server) doHandshake(conn net.Conn) error {
var handshake handshakeRequestResponse
err := s.read(conn, &handshake)
if err != nil {
return err
} else if handshake.Version == nil {
return errors.Err("handshake is missing protocol version")
} else if *handshake.Version != protocolVersion1 && *handshake.Version != protocolVersion2 {
return errors.Err("protocol version not supported")
}
resp, err := json.Marshal(handshakeRequestResponse{Version: handshake.Version})
if err != nil {
return err
}
return s.write(conn, resp)
}
func (s *Server) readBlobRequest(conn net.Conn) (int, string, bool, error) {
var sendRequest sendBlobRequest
err := s.read(conn, &sendRequest)
if err != nil {
return 0, "", false, err
}
var blobHash string
var blobSize int
isSdBlob := sendRequest.SdBlobHash != ""
if isSdBlob {
blobSize = sendRequest.SdBlobSize
blobHash = sendRequest.SdBlobHash
} else {
blobSize = sendRequest.BlobSize
blobHash = sendRequest.BlobHash
}
if blobHash == "" {
return blobSize, blobHash, isSdBlob, errors.Err("blob hash is empty")
}
if blobSize > maxBlobSize {
return blobSize, blobHash, isSdBlob, errors.Err(ErrBlobTooBig)
}
if blobSize == 0 {
return blobSize, blobHash, isSdBlob, errors.Err("0-byte blob received")
}
return blobSize, blobHash, isSdBlob, nil
}
func (s *Server) sendBlobResponse(conn net.Conn, shouldSendBlob, isSdBlob bool, neededBlobs []string) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sendSdBlobResponse{SendSdBlob: shouldSendBlob, NeededBlobs: neededBlobs})
} else {
response, err = json.Marshal(sendBlobResponse{SendBlob: shouldSendBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) sendTransferResponse(conn net.Conn, receivedBlob, isSdBlob bool) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sdBlobTransferResponse{ReceivedSdBlob: receivedBlob})
} else {
response, err = json.Marshal(blobTransferResponse{ReceivedBlob: receivedBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) read(conn net.Conn, v interface{}) error {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
dec := json.NewDecoder(conn)
err = dec.Decode(v)
if err != nil {
data, _ := io.ReadAll(dec.Buffered())
if len(data) > 0 {
return errors.Err("%s. Data: %s", err.Error(), hex.EncodeToString(data))
}
return errors.Err(err)
}
return nil
}
func (s *Server) readRawBlob(conn net.Conn, blobSize int) ([]byte, error) {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return nil, errors.Err(err)
}
blob := make([]byte, blobSize)
_, err = io.ReadFull(bufio.NewReader(conn), blob)
return blob, errors.Err(err)
}
func (s *Server) write(conn net.Conn, b []byte) error {
err := conn.SetWriteDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
n, err := conn.Write(b)
if err == nil && n != len(b) {
err = io.ErrShortWrite
}
return errors.Err(err)
}
func (s *Server) quitting() bool {
select {
case <-s.grp.Ch():
return true
default:
return false
}
}
// BlobHash returns the sha512 hash hex encoded string of the blob byte slice.
func BlobHash(blob []byte) string {
hashBytes := sha512.Sum384(blob)
return hex.EncodeToString(hashBytes[:])
}
func IsValidJSON(b []byte) bool {
var r json.RawMessage
return json.Unmarshal(b, &r) == nil
}
//type errorResponse struct {
// Error string `json:"error"`
//}
type handshakeRequestResponse struct {
Version *int `json:"version"`
}
type sendBlobRequest struct {
BlobHash string `json:"blob_hash,omitempty"`
BlobSize int `json:"blob_size,omitempty"`
SdBlobHash string `json:"sd_blob_hash,omitempty"`
SdBlobSize int `json:"sd_blob_size,omitempty"`
}
type sendBlobResponse struct {
SendBlob bool `json:"send_blob"`
}
type sendSdBlobResponse struct {
SendSdBlob bool `json:"send_sd_blob"`
NeededBlobs []string `json:"needed_blobs,omitempty"`
}
type blobTransferResponse struct {
ReceivedBlob bool `json:"received_blob"`
}
type sdBlobTransferResponse struct {
ReceivedSdBlob bool `json:"received_sd_blob"`
}
// neededBlobChecker can check which blobs from a known stream are not uploaded yet
type neededBlobChecker interface {
MissingBlobsForKnownStream(string) ([]string, error)
}
| {
for {
conn, err := listener.Accept()
if err != nil {
if s.quitting() {
return
}
log.Error(err)
} else {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Dec()
s.handleConn(conn)
s.grp.Done()
}()
}
}
} | identifier_body |
server.go | package reflector
import (
"bufio"
"crypto/sha512"
"encoding/hex"
"encoding/json"
"io"
"net"
"time"
"github.com/lbryio/reflector.go/internal/metrics"
"github.com/lbryio/reflector.go/store"
"github.com/lbryio/lbry.go/v2/extras/errors"
"github.com/lbryio/lbry.go/v2/extras/stop"
"github.com/lbryio/lbry.go/v2/stream"
log "github.com/sirupsen/logrus"
)
const (
// DefaultPort is the port the reflector server listens on if not passed in.
DefaultPort = 5566
// DefaultTimeout is the default timeout to read or write the next message
DefaultTimeout = 5 * time.Second
network = "tcp4"
protocolVersion1 = 0
protocolVersion2 = 1
maxBlobSize = stream.MaxBlobSize
)
var ErrBlobTooBig = errors.Base("blob must be at most %d bytes", maxBlobSize)
// Server is and instance of the reflector server. It houses the blob store and listener.
type Server struct {
Timeout time.Duration // timeout to read or write next message
EnableBlocklist bool // if true, blocklist checking and blob deletion will be enabled
underlyingStore store.BlobStore
outerStore store.BlobStore
grp *stop.Group
}
// NewServer returns an initialized reflector server pointer.
func NewServer(underlying store.BlobStore, outer store.BlobStore) *Server {
return &Server{
Timeout: DefaultTimeout,
underlyingStore: underlying,
outerStore: outer,
grp: stop.New(),
}
}
// Shutdown shuts down the reflector server gracefully.
func (s *Server) Shutdown() {
log.Println("shutting down reflector server...")
s.grp.StopAndWait()
log.Println("reflector server stopped")
}
// Start starts the server to handle connections.
func (s *Server) Start(address string) error {
l, err := net.Listen(network, address)
if err != nil {
return errors.Err(err)
}
log.Println("reflector listening on " + address)
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Dec()
<-s.grp.Ch()
err := l.Close()
if err != nil {
log.Error(errors.Prefix("closing listener", err))
}
s.grp.Done()
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "start").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "start").Dec()
s.listenAndServe(l)
s.grp.Done()
}()
if s.EnableBlocklist {
if b, ok := s.underlyingStore.(store.Blocklister); ok {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Dec()
s.enableBlocklist(b)
s.grp.Done()
}()
} else {
//s.Shutdown()
return errors.Err("blocklist is enabled but blob store does not support blocklisting")
}
}
return nil
}
func (s *Server) listenAndServe(listener net.Listener) {
for {
conn, err := listener.Accept()
if err != nil {
if s.quitting() {
return
}
log.Error(err)
} else {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Dec()
s.handleConn(conn)
s.grp.Done()
}()
}
}
}
func (s *Server) handleConn(conn net.Conn) {
// all this stuff is to close the connections correctly when we're shutting down the server
connNeedsClosing := make(chan struct{})
defer func() {
close(connNeedsClosing)
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Dec()
defer s.grp.Done()
select {
case <-connNeedsClosing:
case <-s.grp.Ch():
}
err := conn.Close()
if err != nil {
log.Error(errors.Prefix("closing peer conn", err))
}
}()
err := s.doHandshake(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending handshake error", err))
}
return
}
for {
err = s.receiveBlob(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending blob receive error", err))
}
return
}
}
}
func (s *Server) doError(conn net.Conn, err error) error {
if err == nil {
return nil
}
shouldLog := metrics.TrackError(metrics.DirectionUpload, err)
if shouldLog {
log.Errorln(errors.FullTrace(err))
}
if e2, ok := err.(*json.SyntaxError); ok {
log.Errorf("syntax error at byte offset %d", e2.Offset)
}
//resp, err := json.Marshal(errorResponse{Error: err.Error()})
//if err != nil {
// return err
//}
//return s.write(conn, resp)
return nil
}
func (s *Server) receiveBlob(conn net.Conn) error {
blobSize, blobHash, isSdBlob, err := s.readBlobRequest(conn)
if err != nil {
return err
}
var wantsBlob bool
if bl, ok := s.underlyingStore.(store.Blocklister); ok {
wantsBlob, err = bl.Wants(blobHash)
if err != nil {
return err
}
} else {
blobExists, err := s.underlyingStore.Has(blobHash)
if err != nil {
return err
}
wantsBlob = !blobExists
}
var neededBlobs []string
if isSdBlob && !wantsBlob {
if nbc, ok := s.underlyingStore.(neededBlobChecker); ok {
neededBlobs, err = nbc.MissingBlobsForKnownStream(blobHash)
if err != nil {
return err
}
} else {
// if we can't check for blobs in a stream, we have to say that the sd blob is
// missing. if we say we have the sd blob, they wont try to send any content blobs
wantsBlob = true
}
}
err = s.sendBlobResponse(conn, wantsBlob, isSdBlob, neededBlobs)
if err != nil {
return err
}
if !wantsBlob {
return nil
}
blob, err := s.readRawBlob(conn, blobSize)
if err != nil {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Prefix("error reading blob "+blobHash[:8], err)
}
receivedBlobHash := BlobHash(blob)
if blobHash != receivedBlobHash {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Err("hash of received blob data does not match hash from send request")
// this can also happen if the blob size is wrong, because the server will read the wrong number of bytes from the stream
}
log.Debugln("Got blob " + blobHash[:8])
if isSdBlob {
err = s.outerStore.PutSD(blobHash, blob)
} else {
err = s.outerStore.Put(blobHash, blob)
}
if err != nil {
return err
}
metrics.MtrInBytesReflector.Add(float64(len(blob)))
metrics.BlobUploadCount.Inc()
if isSdBlob {
metrics.SDBlobUploadCount.Inc()
}
return s.sendTransferResponse(conn, true, isSdBlob)
}
func (s *Server) doHandshake(conn net.Conn) error {
var handshake handshakeRequestResponse
err := s.read(conn, &handshake)
if err != nil {
return err
} else if handshake.Version == nil {
return errors.Err("handshake is missing protocol version")
} else if *handshake.Version != protocolVersion1 && *handshake.Version != protocolVersion2 {
return errors.Err("protocol version not supported")
}
resp, err := json.Marshal(handshakeRequestResponse{Version: handshake.Version})
if err != nil {
return err
}
return s.write(conn, resp)
}
func (s *Server) readBlobRequest(conn net.Conn) (int, string, bool, error) {
var sendRequest sendBlobRequest
err := s.read(conn, &sendRequest)
if err != nil {
return 0, "", false, err
}
var blobHash string
var blobSize int
isSdBlob := sendRequest.SdBlobHash != ""
if isSdBlob {
blobSize = sendRequest.SdBlobSize
blobHash = sendRequest.SdBlobHash
} else {
blobSize = sendRequest.BlobSize
blobHash = sendRequest.BlobHash
}
if blobHash == "" {
return blobSize, blobHash, isSdBlob, errors.Err("blob hash is empty")
}
if blobSize > maxBlobSize {
return blobSize, blobHash, isSdBlob, errors.Err(ErrBlobTooBig)
}
if blobSize == 0 {
return blobSize, blobHash, isSdBlob, errors.Err("0-byte blob received")
}
return blobSize, blobHash, isSdBlob, nil
}
func (s *Server) sendBlobResponse(conn net.Conn, shouldSendBlob, isSdBlob bool, neededBlobs []string) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sendSdBlobResponse{SendSdBlob: shouldSendBlob, NeededBlobs: neededBlobs})
} else {
response, err = json.Marshal(sendBlobResponse{SendBlob: shouldSendBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) sendTransferResponse(conn net.Conn, receivedBlob, isSdBlob bool) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sdBlobTransferResponse{ReceivedSdBlob: receivedBlob})
} else {
response, err = json.Marshal(blobTransferResponse{ReceivedBlob: receivedBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) read(conn net.Conn, v interface{}) error {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
dec := json.NewDecoder(conn)
err = dec.Decode(v)
if err != nil {
data, _ := io.ReadAll(dec.Buffered())
if len(data) > 0 {
return errors.Err("%s. Data: %s", err.Error(), hex.EncodeToString(data))
}
return errors.Err(err)
}
return nil | func (s *Server) readRawBlob(conn net.Conn, blobSize int) ([]byte, error) {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return nil, errors.Err(err)
}
blob := make([]byte, blobSize)
_, err = io.ReadFull(bufio.NewReader(conn), blob)
return blob, errors.Err(err)
}
func (s *Server) write(conn net.Conn, b []byte) error {
err := conn.SetWriteDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
n, err := conn.Write(b)
if err == nil && n != len(b) {
err = io.ErrShortWrite
}
return errors.Err(err)
}
func (s *Server) quitting() bool {
select {
case <-s.grp.Ch():
return true
default:
return false
}
}
// BlobHash returns the sha512 hash hex encoded string of the blob byte slice.
func BlobHash(blob []byte) string {
hashBytes := sha512.Sum384(blob)
return hex.EncodeToString(hashBytes[:])
}
func IsValidJSON(b []byte) bool {
var r json.RawMessage
return json.Unmarshal(b, &r) == nil
}
//type errorResponse struct {
// Error string `json:"error"`
//}
type handshakeRequestResponse struct {
Version *int `json:"version"`
}
type sendBlobRequest struct {
BlobHash string `json:"blob_hash,omitempty"`
BlobSize int `json:"blob_size,omitempty"`
SdBlobHash string `json:"sd_blob_hash,omitempty"`
SdBlobSize int `json:"sd_blob_size,omitempty"`
}
type sendBlobResponse struct {
SendBlob bool `json:"send_blob"`
}
type sendSdBlobResponse struct {
SendSdBlob bool `json:"send_sd_blob"`
NeededBlobs []string `json:"needed_blobs,omitempty"`
}
type blobTransferResponse struct {
ReceivedBlob bool `json:"received_blob"`
}
type sdBlobTransferResponse struct {
ReceivedSdBlob bool `json:"received_sd_blob"`
}
// neededBlobChecker can check which blobs from a known stream are not uploaded yet
type neededBlobChecker interface {
MissingBlobsForKnownStream(string) ([]string, error)
} | }
| random_line_split |
server.go | package reflector
import (
"bufio"
"crypto/sha512"
"encoding/hex"
"encoding/json"
"io"
"net"
"time"
"github.com/lbryio/reflector.go/internal/metrics"
"github.com/lbryio/reflector.go/store"
"github.com/lbryio/lbry.go/v2/extras/errors"
"github.com/lbryio/lbry.go/v2/extras/stop"
"github.com/lbryio/lbry.go/v2/stream"
log "github.com/sirupsen/logrus"
)
const (
// DefaultPort is the port the reflector server listens on if not passed in.
DefaultPort = 5566
// DefaultTimeout is the default timeout to read or write the next message
DefaultTimeout = 5 * time.Second
network = "tcp4"
protocolVersion1 = 0
protocolVersion2 = 1
maxBlobSize = stream.MaxBlobSize
)
var ErrBlobTooBig = errors.Base("blob must be at most %d bytes", maxBlobSize)
// Server is and instance of the reflector server. It houses the blob store and listener.
type Server struct {
Timeout time.Duration // timeout to read or write next message
EnableBlocklist bool // if true, blocklist checking and blob deletion will be enabled
underlyingStore store.BlobStore
outerStore store.BlobStore
grp *stop.Group
}
// NewServer returns an initialized reflector server pointer.
func NewServer(underlying store.BlobStore, outer store.BlobStore) *Server {
return &Server{
Timeout: DefaultTimeout,
underlyingStore: underlying,
outerStore: outer,
grp: stop.New(),
}
}
// Shutdown shuts down the reflector server gracefully.
func (s *Server) Shutdown() {
log.Println("shutting down reflector server...")
s.grp.StopAndWait()
log.Println("reflector server stopped")
}
// Start starts the server to handle connections.
func (s *Server) Start(address string) error {
l, err := net.Listen(network, address)
if err != nil {
return errors.Err(err)
}
log.Println("reflector listening on " + address)
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "listener").Dec()
<-s.grp.Ch()
err := l.Close()
if err != nil {
log.Error(errors.Prefix("closing listener", err))
}
s.grp.Done()
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "start").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "start").Dec()
s.listenAndServe(l)
s.grp.Done()
}()
if s.EnableBlocklist {
if b, ok := s.underlyingStore.(store.Blocklister); ok {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "enableblocklist").Dec()
s.enableBlocklist(b)
s.grp.Done()
}()
} else {
//s.Shutdown()
return errors.Err("blocklist is enabled but blob store does not support blocklisting")
}
}
return nil
}
func (s *Server) listenAndServe(listener net.Listener) {
for {
conn, err := listener.Accept()
if err != nil {
if s.quitting() {
return
}
log.Error(err)
} else {
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-listenandserve").Dec()
s.handleConn(conn)
s.grp.Done()
}()
}
}
}
func (s *Server) handleConn(conn net.Conn) {
// all this stuff is to close the connections correctly when we're shutting down the server
connNeedsClosing := make(chan struct{})
defer func() {
close(connNeedsClosing)
}()
s.grp.Add(1)
metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Inc()
go func() {
defer metrics.RoutinesQueue.WithLabelValues("reflector", "server-handleconn").Dec()
defer s.grp.Done()
select {
case <-connNeedsClosing:
case <-s.grp.Ch():
}
err := conn.Close()
if err != nil {
log.Error(errors.Prefix("closing peer conn", err))
}
}()
err := s.doHandshake(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending handshake error", err))
}
return
}
for {
err = s.receiveBlob(conn)
if err != nil {
if errors.Is(err, io.EOF) || s.quitting() {
return
}
err := s.doError(conn, err)
if err != nil {
log.Error(errors.Prefix("sending blob receive error", err))
}
return
}
}
}
func (s *Server) doError(conn net.Conn, err error) error {
if err == nil {
return nil
}
shouldLog := metrics.TrackError(metrics.DirectionUpload, err)
if shouldLog {
log.Errorln(errors.FullTrace(err))
}
if e2, ok := err.(*json.SyntaxError); ok {
log.Errorf("syntax error at byte offset %d", e2.Offset)
}
//resp, err := json.Marshal(errorResponse{Error: err.Error()})
//if err != nil {
// return err
//}
//return s.write(conn, resp)
return nil
}
func (s *Server) receiveBlob(conn net.Conn) error {
blobSize, blobHash, isSdBlob, err := s.readBlobRequest(conn)
if err != nil {
return err
}
var wantsBlob bool
if bl, ok := s.underlyingStore.(store.Blocklister); ok {
wantsBlob, err = bl.Wants(blobHash)
if err != nil {
return err
}
} else {
blobExists, err := s.underlyingStore.Has(blobHash)
if err != nil {
return err
}
wantsBlob = !blobExists
}
var neededBlobs []string
if isSdBlob && !wantsBlob {
if nbc, ok := s.underlyingStore.(neededBlobChecker); ok {
neededBlobs, err = nbc.MissingBlobsForKnownStream(blobHash)
if err != nil {
return err
}
} else {
// if we can't check for blobs in a stream, we have to say that the sd blob is
// missing. if we say we have the sd blob, they wont try to send any content blobs
wantsBlob = true
}
}
err = s.sendBlobResponse(conn, wantsBlob, isSdBlob, neededBlobs)
if err != nil {
return err
}
if !wantsBlob {
return nil
}
blob, err := s.readRawBlob(conn, blobSize)
if err != nil {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Prefix("error reading blob "+blobHash[:8], err)
}
receivedBlobHash := BlobHash(blob)
if blobHash != receivedBlobHash {
sendErr := s.sendTransferResponse(conn, false, isSdBlob)
if sendErr != nil {
return sendErr
}
return errors.Err("hash of received blob data does not match hash from send request")
// this can also happen if the blob size is wrong, because the server will read the wrong number of bytes from the stream
}
log.Debugln("Got blob " + blobHash[:8])
if isSdBlob {
err = s.outerStore.PutSD(blobHash, blob)
} else {
err = s.outerStore.Put(blobHash, blob)
}
if err != nil {
return err
}
metrics.MtrInBytesReflector.Add(float64(len(blob)))
metrics.BlobUploadCount.Inc()
if isSdBlob {
metrics.SDBlobUploadCount.Inc()
}
return s.sendTransferResponse(conn, true, isSdBlob)
}
func (s *Server) | (conn net.Conn) error {
var handshake handshakeRequestResponse
err := s.read(conn, &handshake)
if err != nil {
return err
} else if handshake.Version == nil {
return errors.Err("handshake is missing protocol version")
} else if *handshake.Version != protocolVersion1 && *handshake.Version != protocolVersion2 {
return errors.Err("protocol version not supported")
}
resp, err := json.Marshal(handshakeRequestResponse{Version: handshake.Version})
if err != nil {
return err
}
return s.write(conn, resp)
}
func (s *Server) readBlobRequest(conn net.Conn) (int, string, bool, error) {
var sendRequest sendBlobRequest
err := s.read(conn, &sendRequest)
if err != nil {
return 0, "", false, err
}
var blobHash string
var blobSize int
isSdBlob := sendRequest.SdBlobHash != ""
if isSdBlob {
blobSize = sendRequest.SdBlobSize
blobHash = sendRequest.SdBlobHash
} else {
blobSize = sendRequest.BlobSize
blobHash = sendRequest.BlobHash
}
if blobHash == "" {
return blobSize, blobHash, isSdBlob, errors.Err("blob hash is empty")
}
if blobSize > maxBlobSize {
return blobSize, blobHash, isSdBlob, errors.Err(ErrBlobTooBig)
}
if blobSize == 0 {
return blobSize, blobHash, isSdBlob, errors.Err("0-byte blob received")
}
return blobSize, blobHash, isSdBlob, nil
}
func (s *Server) sendBlobResponse(conn net.Conn, shouldSendBlob, isSdBlob bool, neededBlobs []string) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sendSdBlobResponse{SendSdBlob: shouldSendBlob, NeededBlobs: neededBlobs})
} else {
response, err = json.Marshal(sendBlobResponse{SendBlob: shouldSendBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) sendTransferResponse(conn net.Conn, receivedBlob, isSdBlob bool) error {
var response []byte
var err error
if isSdBlob {
response, err = json.Marshal(sdBlobTransferResponse{ReceivedSdBlob: receivedBlob})
} else {
response, err = json.Marshal(blobTransferResponse{ReceivedBlob: receivedBlob})
}
if err != nil {
return err
}
return s.write(conn, response)
}
func (s *Server) read(conn net.Conn, v interface{}) error {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
dec := json.NewDecoder(conn)
err = dec.Decode(v)
if err != nil {
data, _ := io.ReadAll(dec.Buffered())
if len(data) > 0 {
return errors.Err("%s. Data: %s", err.Error(), hex.EncodeToString(data))
}
return errors.Err(err)
}
return nil
}
func (s *Server) readRawBlob(conn net.Conn, blobSize int) ([]byte, error) {
err := conn.SetReadDeadline(time.Now().Add(s.Timeout))
if err != nil {
return nil, errors.Err(err)
}
blob := make([]byte, blobSize)
_, err = io.ReadFull(bufio.NewReader(conn), blob)
return blob, errors.Err(err)
}
func (s *Server) write(conn net.Conn, b []byte) error {
err := conn.SetWriteDeadline(time.Now().Add(s.Timeout))
if err != nil {
return errors.Err(err)
}
n, err := conn.Write(b)
if err == nil && n != len(b) {
err = io.ErrShortWrite
}
return errors.Err(err)
}
func (s *Server) quitting() bool {
select {
case <-s.grp.Ch():
return true
default:
return false
}
}
// BlobHash returns the sha512 hash hex encoded string of the blob byte slice.
func BlobHash(blob []byte) string {
hashBytes := sha512.Sum384(blob)
return hex.EncodeToString(hashBytes[:])
}
func IsValidJSON(b []byte) bool {
var r json.RawMessage
return json.Unmarshal(b, &r) == nil
}
//type errorResponse struct {
// Error string `json:"error"`
//}
type handshakeRequestResponse struct {
Version *int `json:"version"`
}
type sendBlobRequest struct {
BlobHash string `json:"blob_hash,omitempty"`
BlobSize int `json:"blob_size,omitempty"`
SdBlobHash string `json:"sd_blob_hash,omitempty"`
SdBlobSize int `json:"sd_blob_size,omitempty"`
}
type sendBlobResponse struct {
SendBlob bool `json:"send_blob"`
}
type sendSdBlobResponse struct {
SendSdBlob bool `json:"send_sd_blob"`
NeededBlobs []string `json:"needed_blobs,omitempty"`
}
type blobTransferResponse struct {
ReceivedBlob bool `json:"received_blob"`
}
type sdBlobTransferResponse struct {
ReceivedSdBlob bool `json:"received_sd_blob"`
}
// neededBlobChecker can check which blobs from a known stream are not uploaded yet
type neededBlobChecker interface {
MissingBlobsForKnownStream(string) ([]string, error)
}
| doHandshake | identifier_name |
mod.rs | // Copyright (c) The Starcoin Core Contributors
// SPDX-License-Identifier: Apache-2.0
// Copyright (c) The Libra Core Contributors
// SPDX-License-Identifier: Apache-2.0
use crate::{
account_address::AccountAddress,
account_config::stc_type_tag,
block_metadata::BlockMetadata,
contract_event::ContractEvent,
vm_error::{StatusCode, StatusType, VMStatus},
write_set::WriteSet,
};
use anyhow::{format_err, Error, Result};
use serde::{de, ser, Deserialize, Serialize};
use starcoin_crypto::keygen::KeyGen;
use starcoin_crypto::{
ed25519::*,
hash::{CryptoHash, CryptoHasher, PlainCryptoHash},
traits::*,
HashValue,
};
use std::ops::Deref;
use std::{convert::TryFrom, fmt, time::Duration};
pub mod authenticator {
pub use libra_types::transaction::authenticator::{
AuthenticationKey, AuthenticationKeyPreimage, Scheme, TransactionAuthenticator,
};
}
mod error;
pub mod helpers;
mod pending_transaction;
mod transaction_argument;
mod upgrade;
use crate::transaction::authenticator::TransactionAuthenticator;
pub use error::CallError;
pub use error::Error as TransactionError;
pub use libra_types::transaction::{ChangeSet, Module, Script};
pub use pending_transaction::{Condition, PendingTransaction};
use starcoin_crypto::multi_ed25519::{MultiEd25519PublicKey, MultiEd25519Signature};
pub use transaction_argument::{
parse_transaction_argument, parse_transaction_arguments, TransactionArgument,
};
pub use upgrade::{InitScript, UpgradePackage};
pub type Version = u64; // Height - also used for MVCC in StateDB
pub const MAX_TRANSACTION_SIZE_IN_BYTES: usize = 4096;
/// RawUserTransaction is the portion of a transaction that a client signs
#[derive(Clone, Debug, Hash, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct RawUserTransaction {
/// Sender's address.
sender: AccountAddress,
// Sequence number of this transaction corresponding to sender's account.
sequence_number: u64,
// The transaction script to execute.
payload: TransactionPayload,
// Maximal total gas specified by wallet to spend for this transaction.
max_gas_amount: u64,
// Maximal price can be paid per gas.
gas_unit_price: u64,
// Expiration time for this transaction. If storage is queried and
// the time returned is greater than or equal to this time and this
// transaction has not been included, you can be certain that it will
// never be included.
// A transaction that doesn't expire is represented by a very large value like
// u64::max_value().
#[serde(serialize_with = "serialize_duration")]
#[serde(deserialize_with = "deserialize_duration")]
expiration_time: Duration,
}
// TODO(#1307)
fn serialize_duration<S>(d: &Duration, serializer: S) -> std::result::Result<S::Ok, S::Error>
where
S: ser::Serializer,
{
serializer.serialize_u64(d.as_secs())
}
fn deserialize_duration<'de, D>(deserializer: D) -> std::result::Result<Duration, D::Error>
where
D: de::Deserializer<'de>,
{
struct DurationVisitor;
impl<'de> de::Visitor<'de> for DurationVisitor {
type Value = Duration;
fn expecting(&self, formatter: &mut fmt::Formatter<'_>) -> fmt::Result {
formatter.write_str("Duration as u64")
}
fn visit_u64<E>(self, v: u64) -> std::result::Result<Self::Value, E>
where
E: de::Error,
{
Ok(Duration::from_secs(v))
}
}
deserializer.deserialize_u64(DurationVisitor)
}
impl RawUserTransaction {
/// Create a new `RawUserTransaction` with a payload.
///
/// It can be either to publish a module, to execute a script, or to issue a writeset
/// transaction.
pub fn new(
sender: AccountAddress,
sequence_number: u64,
payload: TransactionPayload,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload,
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Create a new `RawUserTransaction` with a script.
///
/// A script transaction contains only code to execute. No publishing is allowed in scripts.
pub fn | (
sender: AccountAddress,
sequence_number: u64,
script: Script,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload: TransactionPayload::Script(script),
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Create a new `RawUserTransaction` with a module to publish.
///
/// A module transaction is the only way to publish code. Only one module per transaction
/// can be published.
pub fn new_module(
sender: AccountAddress,
sequence_number: u64,
module: Module,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload: TransactionPayload::Module(module),
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Signs the given `RawUserTransaction`. Note that this consumes the `RawUserTransaction` and turns it
/// into a `SignatureCheckedTransaction`.
///
/// For a transaction that has just been signed, its signature is expected to be valid.
pub fn sign(
self,
private_key: &Ed25519PrivateKey,
public_key: Ed25519PublicKey,
) -> Result<SignatureCheckedTransaction> {
let signature = private_key.sign_message(&self.crypto_hash());
Ok(SignatureCheckedTransaction(SignedUserTransaction::new(
self, public_key, signature,
)))
}
pub fn into_payload(self) -> TransactionPayload {
self.payload
}
/// Return the sender of this transaction.
pub fn sender(&self) -> AccountAddress {
self.sender
}
pub fn mock() -> Self {
Self::mock_by_sender(AccountAddress::random())
}
pub fn mock_by_sender(sender: AccountAddress) -> Self {
Self::new(
sender,
0,
TransactionPayload::Script(Script::new(vec![], vec![], vec![])),
0,
0,
Duration::new(0, 0),
)
}
pub fn mock_from(compiled_script: Vec<u8>) -> Self {
Self::new(
AccountAddress::default(),
0,
TransactionPayload::Script(Script::new(compiled_script, vec![stc_type_tag()], vec![])),
600,
0,
Duration::new(0, 0),
)
}
}
#[derive(Clone, Debug, Hash, Eq, PartialEq, Serialize, Deserialize)]
pub enum TransactionPayload {
/// A transaction that executes code.
Script(Script),
/// A transaction that publishes code.
Module(Module),
/// A transaction that publish or update module code by a package.
Package(UpgradePackage),
}
/// A transaction that has been signed.
///
/// A `SignedUserTransaction` is a single transaction that can be atomically executed. Clients submit
/// these to validator nodes, and the validator and executor submits these to the VM.
///
/// **IMPORTANT:** The signature of a `SignedUserTransaction` is not guaranteed to be verified. For a
/// transaction whose signature is statically guaranteed to be verified, see
/// [`SignatureCheckedTransaction`].
#[derive(Clone, Eq, PartialEq, Hash, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct SignedUserTransaction {
/// The raw transaction
raw_txn: RawUserTransaction,
/// Public key and signature to authenticate
authenticator: TransactionAuthenticator,
}
/// A transaction for which the signature has been verified. Created by
/// [`SignedUserTransaction::check_signature`] and [`RawUserTransaction::sign`].
#[derive(Clone, Debug, Eq, PartialEq, Hash)]
pub struct SignatureCheckedTransaction(SignedUserTransaction);
impl SignatureCheckedTransaction {
/// Returns the `SignedUserTransaction` within.
pub fn into_inner(self) -> SignedUserTransaction {
self.0
}
/// Returns the `RawUserTransaction` within.
pub fn into_raw_transaction(self) -> RawUserTransaction {
self.0.into_raw_transaction()
}
}
impl Deref for SignatureCheckedTransaction {
type Target = SignedUserTransaction;
fn deref(&self) -> &Self::Target {
&self.0
}
}
impl fmt::Debug for SignedUserTransaction {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
write!(
f,
"SignedTransaction {{ \n \
{{ raw_txn: {:#?}, \n \
authenticator: {:#?}, \n \
}} \n \
}}",
self.raw_txn, self.authenticator
)
}
}
impl SignedUserTransaction {
pub fn new(
raw_txn: RawUserTransaction,
public_key: Ed25519PublicKey,
signature: Ed25519Signature,
) -> SignedUserTransaction {
let authenticator = TransactionAuthenticator::ed25519(public_key, signature);
SignedUserTransaction {
raw_txn,
authenticator,
}
}
pub fn multi_ed25519(
raw_txn: RawUserTransaction,
public_key: MultiEd25519PublicKey,
signature: MultiEd25519Signature,
) -> SignedUserTransaction {
let authenticator = TransactionAuthenticator::multi_ed25519(public_key, signature);
SignedUserTransaction {
raw_txn,
authenticator,
}
}
pub fn authenticator(&self) -> TransactionAuthenticator {
self.authenticator.clone()
}
pub fn raw_txn(&self) -> &RawUserTransaction {
&self.raw_txn
}
pub fn sender(&self) -> AccountAddress {
self.raw_txn.sender
}
pub fn into_raw_transaction(self) -> RawUserTransaction {
self.raw_txn
}
pub fn sequence_number(&self) -> u64 {
self.raw_txn.sequence_number
}
pub fn payload(&self) -> &TransactionPayload {
&self.raw_txn.payload
}
pub fn max_gas_amount(&self) -> u64 {
self.raw_txn.max_gas_amount
}
pub fn gas_unit_price(&self) -> u64 {
self.raw_txn.gas_unit_price
}
pub fn expiration_time(&self) -> Duration {
self.raw_txn.expiration_time
}
pub fn raw_txn_bytes_len(&self) -> usize {
scs::to_bytes(&self.raw_txn)
.expect("Unable to serialize RawUserTransaction")
.len()
}
/// Checks that the signature of given transaction. Returns `Ok(SignatureCheckedTransaction)` if
/// the signature is valid.
pub fn check_signature(self) -> Result<SignatureCheckedTransaction> {
self.authenticator
.verify_signature(&self.raw_txn.crypto_hash())?;
Ok(SignatureCheckedTransaction(self))
}
//TODO
pub fn mock() -> Self {
let mut gen = KeyGen::from_os_rng();
let (private_key, public_key) = gen.generate_keypair();
let raw_txn = RawUserTransaction::mock();
raw_txn.sign(&private_key, public_key).unwrap().into_inner()
}
pub fn mock_from(compiled_script: Vec<u8>) -> Self {
let mut gen = KeyGen::from_os_rng();
let (private_key, public_key) = gen.generate_keypair();
let raw_txn = RawUserTransaction::mock_from(compiled_script);
raw_txn.sign(&private_key, public_key).unwrap().into_inner()
}
}
/// The status of executing a transaction. The VM decides whether or not we should `Keep` the
/// transaction output or `Discard` it based upon the execution of the transaction. We wrap these
/// decisions around a `VMStatus` that provides more detail on the final execution state of the VM.
#[derive(Clone, Debug, Eq, PartialEq)]
pub enum TransactionStatus {
/// Discard the transaction output
Discard(VMStatus),
/// Keep the transaction output
Keep(VMStatus),
}
impl TransactionStatus {
pub fn vm_status(&self) -> &VMStatus {
match self {
TransactionStatus::Discard(vm_status) | TransactionStatus::Keep(vm_status) => vm_status,
}
}
}
impl From<VMStatus> for TransactionStatus {
fn from(vm_status: VMStatus) -> Self {
let should_discard = match vm_status.status_type() {
// Any unknown error should be discarded
StatusType::Unknown => true,
// Any error that is a validation status (i.e. an error arising from the prologue)
// causes the transaction to not be included.
StatusType::Validation => true,
// If the VM encountered an invalid internal state, we should discard the transaction.
StatusType::InvariantViolation => true,
// A transaction that publishes code that cannot be verified will be charged.
StatusType::Verification => false,
// Even if we are unable to decode the transaction, there should be a charge made to
// that user's account for the gas fees related to decoding, running the prologue etc.
StatusType::Deserialization => false,
// Any error encountered during the execution of the transaction will charge gas.
StatusType::Execution => false,
};
if should_discard {
TransactionStatus::Discard(vm_status)
} else {
TransactionStatus::Keep(vm_status)
}
}
}
/// The output of executing a transaction.
#[derive(Clone, Debug, Eq, PartialEq)]
pub struct TransactionOutput {
write_set: WriteSet,
/// The list of events emitted during this transaction.
events: Vec<ContractEvent>,
/// The amount of gas used during execution.
gas_used: u64,
/// The resource increment size
delta_size: i64,
/// The execution status.
status: TransactionStatus,
}
impl TransactionOutput {
pub fn new(
write_set: WriteSet,
events: Vec<ContractEvent>,
gas_used: u64,
delta_size: i64,
status: TransactionStatus,
) -> Self {
TransactionOutput {
write_set,
events,
gas_used,
delta_size,
status,
}
}
pub fn write_set(&self) -> &WriteSet {
&self.write_set
}
pub fn events(&self) -> &[ContractEvent] {
&self.events
}
pub fn gas_used(&self) -> u64 {
self.gas_used
}
pub fn status(&self) -> &TransactionStatus {
&self.status
}
pub fn delta_size(&self) -> i64 {
self.delta_size
}
pub fn into_inner(self) -> (WriteSet, Vec<ContractEvent>, u64, TransactionStatus) {
(self.write_set, self.events, self.gas_used, self.status)
}
}
/// `TransactionInfo` is the object we store in the transaction accumulator. It consists of the
/// transaction as well as the execution result of this transaction.
#[derive(Clone, Debug, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct TransactionInfo {
/// The hash of this transaction.
transaction_hash: HashValue,
/// The root hash of Sparse Merkle Tree describing the world state at the end of this
/// transaction.
state_root_hash: HashValue,
/// The root hash of Merkle Accumulator storing all events emitted during this transaction.
event_root_hash: HashValue,
events: Vec<ContractEvent>,
/// The amount of gas used.
gas_used: u64,
/// The major status. This will provide the general error class. Note that this is not
/// particularly high fidelity in the presence of sub statuses but, the major status does
/// determine whether or not the transaction is applied to the global state or not.
major_status: StatusCode,
}
impl TransactionInfo {
/// Constructs a new `TransactionInfo` object using transaction hash, state root hash and event
/// root hash.
pub fn new(
transaction_hash: HashValue,
state_root_hash: HashValue,
event_root_hash: HashValue,
events: Vec<ContractEvent>,
gas_used: u64,
major_status: StatusCode,
) -> TransactionInfo {
TransactionInfo {
transaction_hash,
state_root_hash,
event_root_hash,
events,
gas_used,
major_status,
}
}
pub fn id(&self) -> HashValue {
self.crypto_hash()
}
/// Returns the hash of this transaction.
pub fn transaction_hash(&self) -> HashValue {
self.transaction_hash
}
/// Returns root hash of Sparse Merkle Tree describing the world state at the end of this
/// transaction.
pub fn state_root_hash(&self) -> HashValue {
self.state_root_hash
}
/// Returns the root hash of Merkle Accumulator storing all events emitted during this
/// transaction.
pub fn event_root_hash(&self) -> HashValue {
self.event_root_hash
}
pub fn events(&self) -> &[ContractEvent] {
&self.events
}
/// Returns the amount of gas used by this transaction.
pub fn gas_used(&self) -> u64 {
self.gas_used
}
pub fn major_status(&self) -> StatusCode {
self.major_status
}
}
/// `Transaction` will be the transaction type used internally in the libra node to represent the
/// transaction to be processed and persisted.
///
/// We suppress the clippy warning here as we would expect most of the transaction to be user
/// transaction.
#[allow(clippy::large_enum_variant)]
#[derive(Clone, Debug, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub enum Transaction {
/// Transaction submitted by the user. e.g: P2P payment transaction, publishing module
/// transaction, etc.
UserTransaction(SignedUserTransaction),
/// Transaction that applies a ChangeSet to the current ChainState. This should be used for ONLY for
/// genesis right now.
ChangeSet(ChangeSet),
/// Transaction to update the block metadata resource at the beginning of a block.
BlockMetadata(BlockMetadata),
}
impl Transaction {
pub fn as_signed_user_txn(&self) -> Result<&SignedUserTransaction> {
match self {
Transaction::UserTransaction(txn) => Ok(txn),
_ => Err(format_err!("Not a user transaction.")),
}
}
pub fn id(&self) -> HashValue {
//TODO rethink txn id's represent.
self.crypto_hash()
}
}
impl TryFrom<Transaction> for SignedUserTransaction {
type Error = Error;
fn try_from(txn: Transaction) -> Result<Self> {
match txn {
Transaction::UserTransaction(txn) => Ok(txn),
_ => Err(format_err!("Not a user transaction.")),
}
}
}
/// Pool transactions status
#[derive(Debug, Clone, Copy, PartialEq, Eq, Serialize, Deserialize)]
#[serde(rename_all = "camelCase")]
pub enum TxStatus {
/// Added transaction
Added,
/// Rejected transaction
Rejected,
/// Dropped transaction
Dropped,
/// Invalid transaction
Invalid,
/// Canceled transaction
Canceled,
/// Culled transaction
Culled,
}
impl std::fmt::Display for TxStatus {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
let s = match self {
TxStatus::Added => "added",
TxStatus::Rejected => "rejected",
TxStatus::Dropped => "dropped",
TxStatus::Invalid => "invalid",
TxStatus::Canceled => "canceled",
TxStatus::Culled => "culled",
};
write!(f, "{}", s)
}
}
| new_script | identifier_name |
mod.rs | // Copyright (c) The Starcoin Core Contributors
// SPDX-License-Identifier: Apache-2.0
// Copyright (c) The Libra Core Contributors
// SPDX-License-Identifier: Apache-2.0
use crate::{
account_address::AccountAddress,
account_config::stc_type_tag,
block_metadata::BlockMetadata,
contract_event::ContractEvent,
vm_error::{StatusCode, StatusType, VMStatus},
write_set::WriteSet,
};
use anyhow::{format_err, Error, Result};
use serde::{de, ser, Deserialize, Serialize};
use starcoin_crypto::keygen::KeyGen;
use starcoin_crypto::{
ed25519::*,
hash::{CryptoHash, CryptoHasher, PlainCryptoHash},
traits::*,
HashValue,
};
use std::ops::Deref;
use std::{convert::TryFrom, fmt, time::Duration};
pub mod authenticator {
pub use libra_types::transaction::authenticator::{
AuthenticationKey, AuthenticationKeyPreimage, Scheme, TransactionAuthenticator,
};
}
mod error;
pub mod helpers;
mod pending_transaction;
mod transaction_argument;
mod upgrade;
use crate::transaction::authenticator::TransactionAuthenticator;
pub use error::CallError;
pub use error::Error as TransactionError;
pub use libra_types::transaction::{ChangeSet, Module, Script};
pub use pending_transaction::{Condition, PendingTransaction};
use starcoin_crypto::multi_ed25519::{MultiEd25519PublicKey, MultiEd25519Signature};
pub use transaction_argument::{
parse_transaction_argument, parse_transaction_arguments, TransactionArgument,
};
pub use upgrade::{InitScript, UpgradePackage};
pub type Version = u64; // Height - also used for MVCC in StateDB
pub const MAX_TRANSACTION_SIZE_IN_BYTES: usize = 4096;
/// RawUserTransaction is the portion of a transaction that a client signs
#[derive(Clone, Debug, Hash, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct RawUserTransaction {
/// Sender's address.
sender: AccountAddress,
// Sequence number of this transaction corresponding to sender's account.
sequence_number: u64,
// The transaction script to execute.
payload: TransactionPayload,
// Maximal total gas specified by wallet to spend for this transaction.
max_gas_amount: u64,
// Maximal price can be paid per gas.
gas_unit_price: u64,
// Expiration time for this transaction. If storage is queried and
// the time returned is greater than or equal to this time and this
// transaction has not been included, you can be certain that it will
// never be included. | expiration_time: Duration,
}
// TODO(#1307)
fn serialize_duration<S>(d: &Duration, serializer: S) -> std::result::Result<S::Ok, S::Error>
where
S: ser::Serializer,
{
serializer.serialize_u64(d.as_secs())
}
fn deserialize_duration<'de, D>(deserializer: D) -> std::result::Result<Duration, D::Error>
where
D: de::Deserializer<'de>,
{
struct DurationVisitor;
impl<'de> de::Visitor<'de> for DurationVisitor {
type Value = Duration;
fn expecting(&self, formatter: &mut fmt::Formatter<'_>) -> fmt::Result {
formatter.write_str("Duration as u64")
}
fn visit_u64<E>(self, v: u64) -> std::result::Result<Self::Value, E>
where
E: de::Error,
{
Ok(Duration::from_secs(v))
}
}
deserializer.deserialize_u64(DurationVisitor)
}
impl RawUserTransaction {
/// Create a new `RawUserTransaction` with a payload.
///
/// It can be either to publish a module, to execute a script, or to issue a writeset
/// transaction.
pub fn new(
sender: AccountAddress,
sequence_number: u64,
payload: TransactionPayload,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload,
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Create a new `RawUserTransaction` with a script.
///
/// A script transaction contains only code to execute. No publishing is allowed in scripts.
pub fn new_script(
sender: AccountAddress,
sequence_number: u64,
script: Script,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload: TransactionPayload::Script(script),
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Create a new `RawUserTransaction` with a module to publish.
///
/// A module transaction is the only way to publish code. Only one module per transaction
/// can be published.
pub fn new_module(
sender: AccountAddress,
sequence_number: u64,
module: Module,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload: TransactionPayload::Module(module),
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Signs the given `RawUserTransaction`. Note that this consumes the `RawUserTransaction` and turns it
/// into a `SignatureCheckedTransaction`.
///
/// For a transaction that has just been signed, its signature is expected to be valid.
pub fn sign(
self,
private_key: &Ed25519PrivateKey,
public_key: Ed25519PublicKey,
) -> Result<SignatureCheckedTransaction> {
let signature = private_key.sign_message(&self.crypto_hash());
Ok(SignatureCheckedTransaction(SignedUserTransaction::new(
self, public_key, signature,
)))
}
pub fn into_payload(self) -> TransactionPayload {
self.payload
}
/// Return the sender of this transaction.
pub fn sender(&self) -> AccountAddress {
self.sender
}
pub fn mock() -> Self {
Self::mock_by_sender(AccountAddress::random())
}
pub fn mock_by_sender(sender: AccountAddress) -> Self {
Self::new(
sender,
0,
TransactionPayload::Script(Script::new(vec![], vec![], vec![])),
0,
0,
Duration::new(0, 0),
)
}
pub fn mock_from(compiled_script: Vec<u8>) -> Self {
Self::new(
AccountAddress::default(),
0,
TransactionPayload::Script(Script::new(compiled_script, vec![stc_type_tag()], vec![])),
600,
0,
Duration::new(0, 0),
)
}
}
#[derive(Clone, Debug, Hash, Eq, PartialEq, Serialize, Deserialize)]
pub enum TransactionPayload {
/// A transaction that executes code.
Script(Script),
/// A transaction that publishes code.
Module(Module),
/// A transaction that publish or update module code by a package.
Package(UpgradePackage),
}
/// A transaction that has been signed.
///
/// A `SignedUserTransaction` is a single transaction that can be atomically executed. Clients submit
/// these to validator nodes, and the validator and executor submits these to the VM.
///
/// **IMPORTANT:** The signature of a `SignedUserTransaction` is not guaranteed to be verified. For a
/// transaction whose signature is statically guaranteed to be verified, see
/// [`SignatureCheckedTransaction`].
#[derive(Clone, Eq, PartialEq, Hash, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct SignedUserTransaction {
/// The raw transaction
raw_txn: RawUserTransaction,
/// Public key and signature to authenticate
authenticator: TransactionAuthenticator,
}
/// A transaction for which the signature has been verified. Created by
/// [`SignedUserTransaction::check_signature`] and [`RawUserTransaction::sign`].
#[derive(Clone, Debug, Eq, PartialEq, Hash)]
pub struct SignatureCheckedTransaction(SignedUserTransaction);
impl SignatureCheckedTransaction {
/// Returns the `SignedUserTransaction` within.
pub fn into_inner(self) -> SignedUserTransaction {
self.0
}
/// Returns the `RawUserTransaction` within.
pub fn into_raw_transaction(self) -> RawUserTransaction {
self.0.into_raw_transaction()
}
}
impl Deref for SignatureCheckedTransaction {
type Target = SignedUserTransaction;
fn deref(&self) -> &Self::Target {
&self.0
}
}
impl fmt::Debug for SignedUserTransaction {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
write!(
f,
"SignedTransaction {{ \n \
{{ raw_txn: {:#?}, \n \
authenticator: {:#?}, \n \
}} \n \
}}",
self.raw_txn, self.authenticator
)
}
}
impl SignedUserTransaction {
pub fn new(
raw_txn: RawUserTransaction,
public_key: Ed25519PublicKey,
signature: Ed25519Signature,
) -> SignedUserTransaction {
let authenticator = TransactionAuthenticator::ed25519(public_key, signature);
SignedUserTransaction {
raw_txn,
authenticator,
}
}
pub fn multi_ed25519(
raw_txn: RawUserTransaction,
public_key: MultiEd25519PublicKey,
signature: MultiEd25519Signature,
) -> SignedUserTransaction {
let authenticator = TransactionAuthenticator::multi_ed25519(public_key, signature);
SignedUserTransaction {
raw_txn,
authenticator,
}
}
pub fn authenticator(&self) -> TransactionAuthenticator {
self.authenticator.clone()
}
pub fn raw_txn(&self) -> &RawUserTransaction {
&self.raw_txn
}
pub fn sender(&self) -> AccountAddress {
self.raw_txn.sender
}
pub fn into_raw_transaction(self) -> RawUserTransaction {
self.raw_txn
}
pub fn sequence_number(&self) -> u64 {
self.raw_txn.sequence_number
}
pub fn payload(&self) -> &TransactionPayload {
&self.raw_txn.payload
}
pub fn max_gas_amount(&self) -> u64 {
self.raw_txn.max_gas_amount
}
pub fn gas_unit_price(&self) -> u64 {
self.raw_txn.gas_unit_price
}
pub fn expiration_time(&self) -> Duration {
self.raw_txn.expiration_time
}
pub fn raw_txn_bytes_len(&self) -> usize {
scs::to_bytes(&self.raw_txn)
.expect("Unable to serialize RawUserTransaction")
.len()
}
/// Checks that the signature of given transaction. Returns `Ok(SignatureCheckedTransaction)` if
/// the signature is valid.
pub fn check_signature(self) -> Result<SignatureCheckedTransaction> {
self.authenticator
.verify_signature(&self.raw_txn.crypto_hash())?;
Ok(SignatureCheckedTransaction(self))
}
//TODO
pub fn mock() -> Self {
let mut gen = KeyGen::from_os_rng();
let (private_key, public_key) = gen.generate_keypair();
let raw_txn = RawUserTransaction::mock();
raw_txn.sign(&private_key, public_key).unwrap().into_inner()
}
pub fn mock_from(compiled_script: Vec<u8>) -> Self {
let mut gen = KeyGen::from_os_rng();
let (private_key, public_key) = gen.generate_keypair();
let raw_txn = RawUserTransaction::mock_from(compiled_script);
raw_txn.sign(&private_key, public_key).unwrap().into_inner()
}
}
/// The status of executing a transaction. The VM decides whether or not we should `Keep` the
/// transaction output or `Discard` it based upon the execution of the transaction. We wrap these
/// decisions around a `VMStatus` that provides more detail on the final execution state of the VM.
#[derive(Clone, Debug, Eq, PartialEq)]
pub enum TransactionStatus {
/// Discard the transaction output
Discard(VMStatus),
/// Keep the transaction output
Keep(VMStatus),
}
impl TransactionStatus {
pub fn vm_status(&self) -> &VMStatus {
match self {
TransactionStatus::Discard(vm_status) | TransactionStatus::Keep(vm_status) => vm_status,
}
}
}
impl From<VMStatus> for TransactionStatus {
fn from(vm_status: VMStatus) -> Self {
let should_discard = match vm_status.status_type() {
// Any unknown error should be discarded
StatusType::Unknown => true,
// Any error that is a validation status (i.e. an error arising from the prologue)
// causes the transaction to not be included.
StatusType::Validation => true,
// If the VM encountered an invalid internal state, we should discard the transaction.
StatusType::InvariantViolation => true,
// A transaction that publishes code that cannot be verified will be charged.
StatusType::Verification => false,
// Even if we are unable to decode the transaction, there should be a charge made to
// that user's account for the gas fees related to decoding, running the prologue etc.
StatusType::Deserialization => false,
// Any error encountered during the execution of the transaction will charge gas.
StatusType::Execution => false,
};
if should_discard {
TransactionStatus::Discard(vm_status)
} else {
TransactionStatus::Keep(vm_status)
}
}
}
/// The output of executing a transaction.
#[derive(Clone, Debug, Eq, PartialEq)]
pub struct TransactionOutput {
write_set: WriteSet,
/// The list of events emitted during this transaction.
events: Vec<ContractEvent>,
/// The amount of gas used during execution.
gas_used: u64,
/// The resource increment size
delta_size: i64,
/// The execution status.
status: TransactionStatus,
}
impl TransactionOutput {
pub fn new(
write_set: WriteSet,
events: Vec<ContractEvent>,
gas_used: u64,
delta_size: i64,
status: TransactionStatus,
) -> Self {
TransactionOutput {
write_set,
events,
gas_used,
delta_size,
status,
}
}
pub fn write_set(&self) -> &WriteSet {
&self.write_set
}
pub fn events(&self) -> &[ContractEvent] {
&self.events
}
pub fn gas_used(&self) -> u64 {
self.gas_used
}
pub fn status(&self) -> &TransactionStatus {
&self.status
}
pub fn delta_size(&self) -> i64 {
self.delta_size
}
pub fn into_inner(self) -> (WriteSet, Vec<ContractEvent>, u64, TransactionStatus) {
(self.write_set, self.events, self.gas_used, self.status)
}
}
/// `TransactionInfo` is the object we store in the transaction accumulator. It consists of the
/// transaction as well as the execution result of this transaction.
#[derive(Clone, Debug, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct TransactionInfo {
/// The hash of this transaction.
transaction_hash: HashValue,
/// The root hash of Sparse Merkle Tree describing the world state at the end of this
/// transaction.
state_root_hash: HashValue,
/// The root hash of Merkle Accumulator storing all events emitted during this transaction.
event_root_hash: HashValue,
events: Vec<ContractEvent>,
/// The amount of gas used.
gas_used: u64,
/// The major status. This will provide the general error class. Note that this is not
/// particularly high fidelity in the presence of sub statuses but, the major status does
/// determine whether or not the transaction is applied to the global state or not.
major_status: StatusCode,
}
impl TransactionInfo {
/// Constructs a new `TransactionInfo` object using transaction hash, state root hash and event
/// root hash.
pub fn new(
transaction_hash: HashValue,
state_root_hash: HashValue,
event_root_hash: HashValue,
events: Vec<ContractEvent>,
gas_used: u64,
major_status: StatusCode,
) -> TransactionInfo {
TransactionInfo {
transaction_hash,
state_root_hash,
event_root_hash,
events,
gas_used,
major_status,
}
}
pub fn id(&self) -> HashValue {
self.crypto_hash()
}
/// Returns the hash of this transaction.
pub fn transaction_hash(&self) -> HashValue {
self.transaction_hash
}
/// Returns root hash of Sparse Merkle Tree describing the world state at the end of this
/// transaction.
pub fn state_root_hash(&self) -> HashValue {
self.state_root_hash
}
/// Returns the root hash of Merkle Accumulator storing all events emitted during this
/// transaction.
pub fn event_root_hash(&self) -> HashValue {
self.event_root_hash
}
pub fn events(&self) -> &[ContractEvent] {
&self.events
}
/// Returns the amount of gas used by this transaction.
pub fn gas_used(&self) -> u64 {
self.gas_used
}
pub fn major_status(&self) -> StatusCode {
self.major_status
}
}
/// `Transaction` will be the transaction type used internally in the libra node to represent the
/// transaction to be processed and persisted.
///
/// We suppress the clippy warning here as we would expect most of the transaction to be user
/// transaction.
#[allow(clippy::large_enum_variant)]
#[derive(Clone, Debug, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub enum Transaction {
/// Transaction submitted by the user. e.g: P2P payment transaction, publishing module
/// transaction, etc.
UserTransaction(SignedUserTransaction),
/// Transaction that applies a ChangeSet to the current ChainState. This should be used for ONLY for
/// genesis right now.
ChangeSet(ChangeSet),
/// Transaction to update the block metadata resource at the beginning of a block.
BlockMetadata(BlockMetadata),
}
impl Transaction {
pub fn as_signed_user_txn(&self) -> Result<&SignedUserTransaction> {
match self {
Transaction::UserTransaction(txn) => Ok(txn),
_ => Err(format_err!("Not a user transaction.")),
}
}
pub fn id(&self) -> HashValue {
//TODO rethink txn id's represent.
self.crypto_hash()
}
}
impl TryFrom<Transaction> for SignedUserTransaction {
type Error = Error;
fn try_from(txn: Transaction) -> Result<Self> {
match txn {
Transaction::UserTransaction(txn) => Ok(txn),
_ => Err(format_err!("Not a user transaction.")),
}
}
}
/// Pool transactions status
#[derive(Debug, Clone, Copy, PartialEq, Eq, Serialize, Deserialize)]
#[serde(rename_all = "camelCase")]
pub enum TxStatus {
/// Added transaction
Added,
/// Rejected transaction
Rejected,
/// Dropped transaction
Dropped,
/// Invalid transaction
Invalid,
/// Canceled transaction
Canceled,
/// Culled transaction
Culled,
}
impl std::fmt::Display for TxStatus {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
let s = match self {
TxStatus::Added => "added",
TxStatus::Rejected => "rejected",
TxStatus::Dropped => "dropped",
TxStatus::Invalid => "invalid",
TxStatus::Canceled => "canceled",
TxStatus::Culled => "culled",
};
write!(f, "{}", s)
}
} | // A transaction that doesn't expire is represented by a very large value like
// u64::max_value().
#[serde(serialize_with = "serialize_duration")]
#[serde(deserialize_with = "deserialize_duration")] | random_line_split |
mod.rs | // Copyright (c) The Starcoin Core Contributors
// SPDX-License-Identifier: Apache-2.0
// Copyright (c) The Libra Core Contributors
// SPDX-License-Identifier: Apache-2.0
use crate::{
account_address::AccountAddress,
account_config::stc_type_tag,
block_metadata::BlockMetadata,
contract_event::ContractEvent,
vm_error::{StatusCode, StatusType, VMStatus},
write_set::WriteSet,
};
use anyhow::{format_err, Error, Result};
use serde::{de, ser, Deserialize, Serialize};
use starcoin_crypto::keygen::KeyGen;
use starcoin_crypto::{
ed25519::*,
hash::{CryptoHash, CryptoHasher, PlainCryptoHash},
traits::*,
HashValue,
};
use std::ops::Deref;
use std::{convert::TryFrom, fmt, time::Duration};
pub mod authenticator {
pub use libra_types::transaction::authenticator::{
AuthenticationKey, AuthenticationKeyPreimage, Scheme, TransactionAuthenticator,
};
}
mod error;
pub mod helpers;
mod pending_transaction;
mod transaction_argument;
mod upgrade;
use crate::transaction::authenticator::TransactionAuthenticator;
pub use error::CallError;
pub use error::Error as TransactionError;
pub use libra_types::transaction::{ChangeSet, Module, Script};
pub use pending_transaction::{Condition, PendingTransaction};
use starcoin_crypto::multi_ed25519::{MultiEd25519PublicKey, MultiEd25519Signature};
pub use transaction_argument::{
parse_transaction_argument, parse_transaction_arguments, TransactionArgument,
};
pub use upgrade::{InitScript, UpgradePackage};
pub type Version = u64; // Height - also used for MVCC in StateDB
pub const MAX_TRANSACTION_SIZE_IN_BYTES: usize = 4096;
/// RawUserTransaction is the portion of a transaction that a client signs
#[derive(Clone, Debug, Hash, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct RawUserTransaction {
/// Sender's address.
sender: AccountAddress,
// Sequence number of this transaction corresponding to sender's account.
sequence_number: u64,
// The transaction script to execute.
payload: TransactionPayload,
// Maximal total gas specified by wallet to spend for this transaction.
max_gas_amount: u64,
// Maximal price can be paid per gas.
gas_unit_price: u64,
// Expiration time for this transaction. If storage is queried and
// the time returned is greater than or equal to this time and this
// transaction has not been included, you can be certain that it will
// never be included.
// A transaction that doesn't expire is represented by a very large value like
// u64::max_value().
#[serde(serialize_with = "serialize_duration")]
#[serde(deserialize_with = "deserialize_duration")]
expiration_time: Duration,
}
// TODO(#1307)
fn serialize_duration<S>(d: &Duration, serializer: S) -> std::result::Result<S::Ok, S::Error>
where
S: ser::Serializer,
{
serializer.serialize_u64(d.as_secs())
}
fn deserialize_duration<'de, D>(deserializer: D) -> std::result::Result<Duration, D::Error>
where
D: de::Deserializer<'de>,
{
struct DurationVisitor;
impl<'de> de::Visitor<'de> for DurationVisitor {
type Value = Duration;
fn expecting(&self, formatter: &mut fmt::Formatter<'_>) -> fmt::Result {
formatter.write_str("Duration as u64")
}
fn visit_u64<E>(self, v: u64) -> std::result::Result<Self::Value, E>
where
E: de::Error,
{
Ok(Duration::from_secs(v))
}
}
deserializer.deserialize_u64(DurationVisitor)
}
impl RawUserTransaction {
/// Create a new `RawUserTransaction` with a payload.
///
/// It can be either to publish a module, to execute a script, or to issue a writeset
/// transaction.
pub fn new(
sender: AccountAddress,
sequence_number: u64,
payload: TransactionPayload,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload,
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Create a new `RawUserTransaction` with a script.
///
/// A script transaction contains only code to execute. No publishing is allowed in scripts.
pub fn new_script(
sender: AccountAddress,
sequence_number: u64,
script: Script,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload: TransactionPayload::Script(script),
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Create a new `RawUserTransaction` with a module to publish.
///
/// A module transaction is the only way to publish code. Only one module per transaction
/// can be published.
pub fn new_module(
sender: AccountAddress,
sequence_number: u64,
module: Module,
max_gas_amount: u64,
gas_unit_price: u64,
expiration_time: Duration,
) -> Self {
RawUserTransaction {
sender,
sequence_number,
payload: TransactionPayload::Module(module),
max_gas_amount,
gas_unit_price,
expiration_time,
}
}
/// Signs the given `RawUserTransaction`. Note that this consumes the `RawUserTransaction` and turns it
/// into a `SignatureCheckedTransaction`.
///
/// For a transaction that has just been signed, its signature is expected to be valid.
pub fn sign(
self,
private_key: &Ed25519PrivateKey,
public_key: Ed25519PublicKey,
) -> Result<SignatureCheckedTransaction> {
let signature = private_key.sign_message(&self.crypto_hash());
Ok(SignatureCheckedTransaction(SignedUserTransaction::new(
self, public_key, signature,
)))
}
pub fn into_payload(self) -> TransactionPayload {
self.payload
}
/// Return the sender of this transaction.
pub fn sender(&self) -> AccountAddress {
self.sender
}
pub fn mock() -> Self {
Self::mock_by_sender(AccountAddress::random())
}
pub fn mock_by_sender(sender: AccountAddress) -> Self {
Self::new(
sender,
0,
TransactionPayload::Script(Script::new(vec![], vec![], vec![])),
0,
0,
Duration::new(0, 0),
)
}
pub fn mock_from(compiled_script: Vec<u8>) -> Self {
Self::new(
AccountAddress::default(),
0,
TransactionPayload::Script(Script::new(compiled_script, vec![stc_type_tag()], vec![])),
600,
0,
Duration::new(0, 0),
)
}
}
#[derive(Clone, Debug, Hash, Eq, PartialEq, Serialize, Deserialize)]
pub enum TransactionPayload {
/// A transaction that executes code.
Script(Script),
/// A transaction that publishes code.
Module(Module),
/// A transaction that publish or update module code by a package.
Package(UpgradePackage),
}
/// A transaction that has been signed.
///
/// A `SignedUserTransaction` is a single transaction that can be atomically executed. Clients submit
/// these to validator nodes, and the validator and executor submits these to the VM.
///
/// **IMPORTANT:** The signature of a `SignedUserTransaction` is not guaranteed to be verified. For a
/// transaction whose signature is statically guaranteed to be verified, see
/// [`SignatureCheckedTransaction`].
#[derive(Clone, Eq, PartialEq, Hash, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct SignedUserTransaction {
/// The raw transaction
raw_txn: RawUserTransaction,
/// Public key and signature to authenticate
authenticator: TransactionAuthenticator,
}
/// A transaction for which the signature has been verified. Created by
/// [`SignedUserTransaction::check_signature`] and [`RawUserTransaction::sign`].
#[derive(Clone, Debug, Eq, PartialEq, Hash)]
pub struct SignatureCheckedTransaction(SignedUserTransaction);
impl SignatureCheckedTransaction {
/// Returns the `SignedUserTransaction` within.
pub fn into_inner(self) -> SignedUserTransaction {
self.0
}
/// Returns the `RawUserTransaction` within.
pub fn into_raw_transaction(self) -> RawUserTransaction {
self.0.into_raw_transaction()
}
}
impl Deref for SignatureCheckedTransaction {
type Target = SignedUserTransaction;
fn deref(&self) -> &Self::Target {
&self.0
}
}
impl fmt::Debug for SignedUserTransaction {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
write!(
f,
"SignedTransaction {{ \n \
{{ raw_txn: {:#?}, \n \
authenticator: {:#?}, \n \
}} \n \
}}",
self.raw_txn, self.authenticator
)
}
}
impl SignedUserTransaction {
pub fn new(
raw_txn: RawUserTransaction,
public_key: Ed25519PublicKey,
signature: Ed25519Signature,
) -> SignedUserTransaction {
let authenticator = TransactionAuthenticator::ed25519(public_key, signature);
SignedUserTransaction {
raw_txn,
authenticator,
}
}
pub fn multi_ed25519(
raw_txn: RawUserTransaction,
public_key: MultiEd25519PublicKey,
signature: MultiEd25519Signature,
) -> SignedUserTransaction {
let authenticator = TransactionAuthenticator::multi_ed25519(public_key, signature);
SignedUserTransaction {
raw_txn,
authenticator,
}
}
pub fn authenticator(&self) -> TransactionAuthenticator {
self.authenticator.clone()
}
pub fn raw_txn(&self) -> &RawUserTransaction |
pub fn sender(&self) -> AccountAddress {
self.raw_txn.sender
}
pub fn into_raw_transaction(self) -> RawUserTransaction {
self.raw_txn
}
pub fn sequence_number(&self) -> u64 {
self.raw_txn.sequence_number
}
pub fn payload(&self) -> &TransactionPayload {
&self.raw_txn.payload
}
pub fn max_gas_amount(&self) -> u64 {
self.raw_txn.max_gas_amount
}
pub fn gas_unit_price(&self) -> u64 {
self.raw_txn.gas_unit_price
}
pub fn expiration_time(&self) -> Duration {
self.raw_txn.expiration_time
}
pub fn raw_txn_bytes_len(&self) -> usize {
scs::to_bytes(&self.raw_txn)
.expect("Unable to serialize RawUserTransaction")
.len()
}
/// Checks that the signature of given transaction. Returns `Ok(SignatureCheckedTransaction)` if
/// the signature is valid.
pub fn check_signature(self) -> Result<SignatureCheckedTransaction> {
self.authenticator
.verify_signature(&self.raw_txn.crypto_hash())?;
Ok(SignatureCheckedTransaction(self))
}
//TODO
pub fn mock() -> Self {
let mut gen = KeyGen::from_os_rng();
let (private_key, public_key) = gen.generate_keypair();
let raw_txn = RawUserTransaction::mock();
raw_txn.sign(&private_key, public_key).unwrap().into_inner()
}
pub fn mock_from(compiled_script: Vec<u8>) -> Self {
let mut gen = KeyGen::from_os_rng();
let (private_key, public_key) = gen.generate_keypair();
let raw_txn = RawUserTransaction::mock_from(compiled_script);
raw_txn.sign(&private_key, public_key).unwrap().into_inner()
}
}
/// The status of executing a transaction. The VM decides whether or not we should `Keep` the
/// transaction output or `Discard` it based upon the execution of the transaction. We wrap these
/// decisions around a `VMStatus` that provides more detail on the final execution state of the VM.
#[derive(Clone, Debug, Eq, PartialEq)]
pub enum TransactionStatus {
/// Discard the transaction output
Discard(VMStatus),
/// Keep the transaction output
Keep(VMStatus),
}
impl TransactionStatus {
pub fn vm_status(&self) -> &VMStatus {
match self {
TransactionStatus::Discard(vm_status) | TransactionStatus::Keep(vm_status) => vm_status,
}
}
}
impl From<VMStatus> for TransactionStatus {
fn from(vm_status: VMStatus) -> Self {
let should_discard = match vm_status.status_type() {
// Any unknown error should be discarded
StatusType::Unknown => true,
// Any error that is a validation status (i.e. an error arising from the prologue)
// causes the transaction to not be included.
StatusType::Validation => true,
// If the VM encountered an invalid internal state, we should discard the transaction.
StatusType::InvariantViolation => true,
// A transaction that publishes code that cannot be verified will be charged.
StatusType::Verification => false,
// Even if we are unable to decode the transaction, there should be a charge made to
// that user's account for the gas fees related to decoding, running the prologue etc.
StatusType::Deserialization => false,
// Any error encountered during the execution of the transaction will charge gas.
StatusType::Execution => false,
};
if should_discard {
TransactionStatus::Discard(vm_status)
} else {
TransactionStatus::Keep(vm_status)
}
}
}
/// The output of executing a transaction.
#[derive(Clone, Debug, Eq, PartialEq)]
pub struct TransactionOutput {
write_set: WriteSet,
/// The list of events emitted during this transaction.
events: Vec<ContractEvent>,
/// The amount of gas used during execution.
gas_used: u64,
/// The resource increment size
delta_size: i64,
/// The execution status.
status: TransactionStatus,
}
impl TransactionOutput {
pub fn new(
write_set: WriteSet,
events: Vec<ContractEvent>,
gas_used: u64,
delta_size: i64,
status: TransactionStatus,
) -> Self {
TransactionOutput {
write_set,
events,
gas_used,
delta_size,
status,
}
}
pub fn write_set(&self) -> &WriteSet {
&self.write_set
}
pub fn events(&self) -> &[ContractEvent] {
&self.events
}
pub fn gas_used(&self) -> u64 {
self.gas_used
}
pub fn status(&self) -> &TransactionStatus {
&self.status
}
pub fn delta_size(&self) -> i64 {
self.delta_size
}
pub fn into_inner(self) -> (WriteSet, Vec<ContractEvent>, u64, TransactionStatus) {
(self.write_set, self.events, self.gas_used, self.status)
}
}
/// `TransactionInfo` is the object we store in the transaction accumulator. It consists of the
/// transaction as well as the execution result of this transaction.
#[derive(Clone, Debug, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub struct TransactionInfo {
/// The hash of this transaction.
transaction_hash: HashValue,
/// The root hash of Sparse Merkle Tree describing the world state at the end of this
/// transaction.
state_root_hash: HashValue,
/// The root hash of Merkle Accumulator storing all events emitted during this transaction.
event_root_hash: HashValue,
events: Vec<ContractEvent>,
/// The amount of gas used.
gas_used: u64,
/// The major status. This will provide the general error class. Note that this is not
/// particularly high fidelity in the presence of sub statuses but, the major status does
/// determine whether or not the transaction is applied to the global state or not.
major_status: StatusCode,
}
impl TransactionInfo {
/// Constructs a new `TransactionInfo` object using transaction hash, state root hash and event
/// root hash.
pub fn new(
transaction_hash: HashValue,
state_root_hash: HashValue,
event_root_hash: HashValue,
events: Vec<ContractEvent>,
gas_used: u64,
major_status: StatusCode,
) -> TransactionInfo {
TransactionInfo {
transaction_hash,
state_root_hash,
event_root_hash,
events,
gas_used,
major_status,
}
}
pub fn id(&self) -> HashValue {
self.crypto_hash()
}
/// Returns the hash of this transaction.
pub fn transaction_hash(&self) -> HashValue {
self.transaction_hash
}
/// Returns root hash of Sparse Merkle Tree describing the world state at the end of this
/// transaction.
pub fn state_root_hash(&self) -> HashValue {
self.state_root_hash
}
/// Returns the root hash of Merkle Accumulator storing all events emitted during this
/// transaction.
pub fn event_root_hash(&self) -> HashValue {
self.event_root_hash
}
pub fn events(&self) -> &[ContractEvent] {
&self.events
}
/// Returns the amount of gas used by this transaction.
pub fn gas_used(&self) -> u64 {
self.gas_used
}
pub fn major_status(&self) -> StatusCode {
self.major_status
}
}
/// `Transaction` will be the transaction type used internally in the libra node to represent the
/// transaction to be processed and persisted.
///
/// We suppress the clippy warning here as we would expect most of the transaction to be user
/// transaction.
#[allow(clippy::large_enum_variant)]
#[derive(Clone, Debug, Eq, PartialEq, Serialize, Deserialize, CryptoHasher, CryptoHash)]
pub enum Transaction {
/// Transaction submitted by the user. e.g: P2P payment transaction, publishing module
/// transaction, etc.
UserTransaction(SignedUserTransaction),
/// Transaction that applies a ChangeSet to the current ChainState. This should be used for ONLY for
/// genesis right now.
ChangeSet(ChangeSet),
/// Transaction to update the block metadata resource at the beginning of a block.
BlockMetadata(BlockMetadata),
}
impl Transaction {
pub fn as_signed_user_txn(&self) -> Result<&SignedUserTransaction> {
match self {
Transaction::UserTransaction(txn) => Ok(txn),
_ => Err(format_err!("Not a user transaction.")),
}
}
pub fn id(&self) -> HashValue {
//TODO rethink txn id's represent.
self.crypto_hash()
}
}
impl TryFrom<Transaction> for SignedUserTransaction {
type Error = Error;
fn try_from(txn: Transaction) -> Result<Self> {
match txn {
Transaction::UserTransaction(txn) => Ok(txn),
_ => Err(format_err!("Not a user transaction.")),
}
}
}
/// Pool transactions status
#[derive(Debug, Clone, Copy, PartialEq, Eq, Serialize, Deserialize)]
#[serde(rename_all = "camelCase")]
pub enum TxStatus {
/// Added transaction
Added,
/// Rejected transaction
Rejected,
/// Dropped transaction
Dropped,
/// Invalid transaction
Invalid,
/// Canceled transaction
Canceled,
/// Culled transaction
Culled,
}
impl std::fmt::Display for TxStatus {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
let s = match self {
TxStatus::Added => "added",
TxStatus::Rejected => "rejected",
TxStatus::Dropped => "dropped",
TxStatus::Invalid => "invalid",
TxStatus::Canceled => "canceled",
TxStatus::Culled => "culled",
};
write!(f, "{}", s)
}
}
| {
&self.raw_txn
} | identifier_body |
sorting.go | package rivescript
// Data sorting functions
import (
"errors"
"sort"
"strconv"
"strings"
)
// Sort buffer data, for RiveScript.SortReplies()
type sortBuffer struct {
topics map[string][]sortedTriggerEntry // Topic name -> array of triggers
thats map[string][]sortedTriggerEntry
sub []string // Substitutions
person []string // Person substitutions
}
// Holds a sorted trigger and the pointer to that trigger's data
type sortedTriggerEntry struct {
trigger string
pointer *astTrigger
}
// Temporary categorization of triggers while sorting
type sortTrack struct {
atomic map[int][]sortedTriggerEntry // Sort by number of whole words
option map[int][]sortedTriggerEntry // Sort optionals by number of words
alpha map[int][]sortedTriggerEntry // Sort alpha wildcards by no. of words
number map[int][]sortedTriggerEntry // Sort numeric wildcards by no. of words
wild map[int][]sortedTriggerEntry // Sort wildcards by no. of words
pound []sortedTriggerEntry // Triggers of just '#'
under []sortedTriggerEntry // Triggers of just '_'
star []sortedTriggerEntry // Triggers of just '*'
}
/*
SortReplies sorts the reply structures in memory for optimal matching.
After you have finished loading your RiveScript code, call this method to
populate the various sort buffers. This is absolutely necessary for reply
matching to work efficiently!
If the bot has loaded no topics, or if it ends up with no sorted triggers at
the end, it will return an error saying such. This usually means the bot didn't
load any RiveScript code, for example because it looked in the wrong directory.
*/
func (rs *RiveScript) SortReplies() error {
// (Re)initialize the sort cache.
rs.sorted.topics = map[string][]sortedTriggerEntry{}
rs.sorted.thats = map[string][]sortedTriggerEntry{}
rs.say("Sorting triggers...")
// If there are no topics, give an error.
if len(rs.topics) == 0 {
return errors.New("SortReplies: no topics were found; did you load any RiveScript code?")
}
// Loop through all the topics.
for topic := range rs.topics {
rs.say("Analyzing topic %s", topic)
// Collect a list of all the triggers we're going to worry about. If this
// topic inherits another topic, we need to recursively add those to the
// list as well.
allTriggers := rs.getTopicTriggers(topic, false)
// Sort these triggers.
rs.sorted.topics[topic] = rs.sortTriggerSet(allTriggers, true)
// Get all of the %Previous triggers for this topic.
thatTriggers := rs.getTopicTriggers(topic, true)
// And sort them, too.
rs.sorted.thats[topic] = rs.sortTriggerSet(thatTriggers, false)
}
// Sort the substitution lists.
rs.sorted.sub = sortList(rs.sub)
rs.sorted.person = sortList(rs.person)
// Did we sort anything at all?
if len(rs.sorted.topics) == 0 && len(rs.sorted.thats) == 0 {
return errors.New("SortReplies: ended up with empty trigger lists; did you load any RiveScript code?")
}
return nil
}
/*
sortTriggerSet sorts a group of triggers in an optimal sorting order.
This function has two use cases:
1. Create a sort buffer for "normal" (matchable) triggers, which are triggers
that are NOT accompanied by a %Previous tag.
2. Create a sort buffer for triggers that had %Previous tags.
Use the `excludePrevious` parameter to control which one is being done. This
function will return a list of sortedTriggerEntry items, and it's intended to
have no duplicate trigger patterns (unless the source RiveScript code explicitly
uses the same duplicate pattern twice, which is a user error).
*/
func (rs *RiveScript) sortTriggerSet(triggers []sortedTriggerEntry, excludePrevious bool) []sortedTriggerEntry {
// Create a priority map, of priority numbers -> their triggers.
prior := map[int][]sortedTriggerEntry{}
// Go through and bucket each trigger by weight (priority).
for _, trig := range triggers {
if excludePrevious && trig.pointer.previous != "" {
continue
}
// Check the trigger text for any {weight} tags, default being 0
match := reWeight.FindStringSubmatch(trig.trigger)
weight := 0
if len(match) > 0 {
weight, _ = strconv.Atoi(match[1])
}
// First trigger of this priority? Initialize the weight map.
if _, ok := prior[weight]; !ok {
prior[weight] = []sortedTriggerEntry{}
}
prior[weight] = append(prior[weight], trig)
}
// Keep a running list of sorted triggers for this topic.
running := []sortedTriggerEntry{}
// Sort the priorities with the highest number first.
var sortedPriorities []int
for k := range prior {
sortedPriorities = append(sortedPriorities, k)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedPriorities)))
// Go through each priority set.
for _, p := range sortedPriorities {
rs.say("Sorting triggers with priority %d", p)
// So, some of these triggers may include an {inherits} tag, if they
// came from a topic which inherits another topic. Lower inherits values
// mean higher priority on the stack. Triggers that have NO inherits
// value at all (which will default to -1), will be moved to the END of
// the stack at the end (have the highest number/lowest priority).
inherits := -1 // -1 means no {inherits} tag
highestInherits := -1 // Highest number seen so far
// Loop through and categorize these triggers.
track := map[int]*sortTrack{}
track[inherits] = initSortTrack()
// Loop through all the triggers.
for _, trig := range prior[p] {
pattern := trig.trigger
rs.say("Looking at trigger: %s", pattern)
// See if the trigger has an {inherits} tag.
match := reInherits.FindStringSubmatch(pattern)
if len(match) > 0 {
inherits, _ = strconv.Atoi(match[1])
if inherits > highestInherits {
highestInherits = inherits
}
rs.say("Trigger belongs to a topic that inherits other topics. "+
"Level=%d", inherits)
pattern = reInherits.ReplaceAllString(pattern, "")
} else {
inherits = -1
}
// If this is the first time we've seen this inheritance level,
// initialize its sort track structure.
if _, ok := track[inherits]; !ok {
track[inherits] = initSortTrack()
}
// Start inspecting the trigger's contents.
if strings.Contains(pattern, "_") {
// Alphabetic wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a _ wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].alpha[cnt]; !ok {
track[inherits].alpha[cnt] = []sortedTriggerEntry{}
}
track[inherits].alpha[cnt] = append(track[inherits].alpha[cnt], trig)
} else {
track[inherits].under = append(track[inherits].under, trig)
}
} else if strings.Contains(pattern, "#") {
// Numeric wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a # wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].number[cnt]; !ok {
track[inherits].number[cnt] = []sortedTriggerEntry{}
}
track[inherits].number[cnt] = append(track[inherits].number[cnt], trig)
} else {
track[inherits].pound = append(track[inherits].pound, trig)
}
} else if strings.Contains(pattern, "*") {
// Wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a * wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].wild[cnt]; !ok {
track[inherits].wild[cnt] = []sortedTriggerEntry{}
}
track[inherits].wild[cnt] = append(track[inherits].wild[cnt], trig)
} else {
track[inherits].star = append(track[inherits].star, trig)
}
} else if strings.Contains(pattern, "[") {
// Optionals included.
cnt := wordCount(pattern, false)
rs.say("Has optionals with %d words", cnt)
if _, ok := track[inherits].option[cnt]; !ok {
track[inherits].option[cnt] = []sortedTriggerEntry{}
}
track[inherits].option[cnt] = append(track[inherits].option[cnt], trig)
} else {
// Totally atomic.
cnt := wordCount(pattern, false)
rs.say("Totally atomic trigger with %d words", cnt)
if _, ok := track[inherits].atomic[cnt]; !ok {
track[inherits].atomic[cnt] = []sortedTriggerEntry{}
}
track[inherits].atomic[cnt] = append(track[inherits].atomic[cnt], trig)
}
}
// Move the no-{inherits} triggers to the bottom of the stack.
track[highestInherits+1] = track[-1]
delete(track, -1)
// Sort the track from the lowest to the highest.
var trackSorted []int
for k := range track {
trackSorted = append(trackSorted, k)
}
sort.Ints(trackSorted)
// Go through each priority level from greatest to smallest.
for _, ip := range trackSorted {
rs.say("ip=%d", ip)
// Sort each of the main kinds of triggers by their word counts.
running = sortByWords(running, track[ip].atomic)
running = sortByWords(running, track[ip].option)
running = sortByWords(running, track[ip].alpha)
running = sortByWords(running, track[ip].number)
running = sortByWords(running, track[ip].wild)
// Add the single wildcard triggers, sorted by length.
running = sortByLength(running, track[ip].under)
running = sortByLength(running, track[ip].pound)
running = sortByLength(running, track[ip].star)
}
}
return running
}
// sortList sorts lists (like substitutions) from a string:string map.
func | (dict map[string]string) []string {
output := []string{}
// Track by number of words.
track := map[int][]string{}
// Loop through each item.
for item := range dict {
cnt := wordCount(item, true)
if _, ok := track[cnt]; !ok {
track[cnt] = []string{}
}
track[cnt] = append(track[cnt], item)
}
// Sort them by word count, descending.
sortedCounts := []int{}
for cnt := range track {
sortedCounts = append(sortedCounts, cnt)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedCounts)))
for _, cnt := range sortedCounts {
// Sort the strings of this word-count by their lengths.
sortedLengths := track[cnt]
sort.Sort(sort.Reverse(byLength(sortedLengths)))
output = append(output, sortedLengths...)
}
return output
}
/*
sortByWords sorts a set of triggers by word count and overall length.
This is a helper function for sorting the `atomic`, `option`, `alpha`, `number`
and `wild` attributes of the sortTrack and adding them to the running sort
buffer in that specific order. Since attribute lookup by reflection is expensive
in Go, this function is given the relevant sort buffer directly, and the current
running sort buffer to add the results to.
The `triggers` parameter is a map between word counts and the triggers that
fit that number of words.
*/
func sortByWords(running []sortedTriggerEntry, triggers map[int][]sortedTriggerEntry) []sortedTriggerEntry {
// Sort the triggers by their word counts from greatest to smallest.
var sortedWords []int
for wc := range triggers {
sortedWords = append(sortedWords, wc)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedWords)))
for _, wc := range sortedWords {
// Triggers with equal word lengths should be sorted by overall trigger length.
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers[wc] {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Add the triggers to the running triggers bucket.
for _, pattern := range sortedPatterns {
running = append(running, patternMap[pattern]...)
}
}
return running
}
/*
sortByLength sorts a set of triggers purely by character length.
This is like `sortByWords`, but it's intended for triggers that consist solely
of wildcard-like symbols with no real words. For example a trigger of `* * *`
qualifies for this, and it has no words, so we sort by length so it gets a
higher priority than simply `*`.
*/
func sortByLength(running []sortedTriggerEntry, triggers []sortedTriggerEntry) []sortedTriggerEntry {
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Only loop through unique patterns.
patternSet := map[string]bool{}
// Add them to the running triggers bucket.
for _, pattern := range sortedPatterns {
if _, ok := patternSet[pattern]; ok {
continue
}
patternSet[pattern] = true
running = append(running, patternMap[pattern]...)
}
return running
}
// initSortTrack initializes a new, empty sortTrack object.
func initSortTrack() *sortTrack {
return &sortTrack{
atomic: map[int][]sortedTriggerEntry{},
option: map[int][]sortedTriggerEntry{},
alpha: map[int][]sortedTriggerEntry{},
number: map[int][]sortedTriggerEntry{},
wild: map[int][]sortedTriggerEntry{},
pound: []sortedTriggerEntry{},
under: []sortedTriggerEntry{},
star: []sortedTriggerEntry{},
}
}
| sortList | identifier_name |
sorting.go | package rivescript
// Data sorting functions
import (
"errors"
"sort"
"strconv"
"strings"
)
// Sort buffer data, for RiveScript.SortReplies()
type sortBuffer struct {
topics map[string][]sortedTriggerEntry // Topic name -> array of triggers
thats map[string][]sortedTriggerEntry
sub []string // Substitutions
person []string // Person substitutions
}
// Holds a sorted trigger and the pointer to that trigger's data
type sortedTriggerEntry struct {
trigger string
pointer *astTrigger
}
// Temporary categorization of triggers while sorting
type sortTrack struct {
atomic map[int][]sortedTriggerEntry // Sort by number of whole words
option map[int][]sortedTriggerEntry // Sort optionals by number of words
alpha map[int][]sortedTriggerEntry // Sort alpha wildcards by no. of words
number map[int][]sortedTriggerEntry // Sort numeric wildcards by no. of words
wild map[int][]sortedTriggerEntry // Sort wildcards by no. of words
pound []sortedTriggerEntry // Triggers of just '#'
under []sortedTriggerEntry // Triggers of just '_'
star []sortedTriggerEntry // Triggers of just '*'
}
/*
SortReplies sorts the reply structures in memory for optimal matching.
After you have finished loading your RiveScript code, call this method to
populate the various sort buffers. This is absolutely necessary for reply
matching to work efficiently!
If the bot has loaded no topics, or if it ends up with no sorted triggers at
the end, it will return an error saying such. This usually means the bot didn't
load any RiveScript code, for example because it looked in the wrong directory.
*/
func (rs *RiveScript) SortReplies() error {
// (Re)initialize the sort cache.
rs.sorted.topics = map[string][]sortedTriggerEntry{}
rs.sorted.thats = map[string][]sortedTriggerEntry{}
rs.say("Sorting triggers...")
// If there are no topics, give an error.
if len(rs.topics) == 0 {
return errors.New("SortReplies: no topics were found; did you load any RiveScript code?")
}
// Loop through all the topics.
for topic := range rs.topics {
rs.say("Analyzing topic %s", topic)
// Collect a list of all the triggers we're going to worry about. If this
// topic inherits another topic, we need to recursively add those to the
// list as well.
allTriggers := rs.getTopicTriggers(topic, false)
// Sort these triggers.
rs.sorted.topics[topic] = rs.sortTriggerSet(allTriggers, true)
// Get all of the %Previous triggers for this topic.
thatTriggers := rs.getTopicTriggers(topic, true)
// And sort them, too.
rs.sorted.thats[topic] = rs.sortTriggerSet(thatTriggers, false)
}
// Sort the substitution lists.
rs.sorted.sub = sortList(rs.sub)
rs.sorted.person = sortList(rs.person)
// Did we sort anything at all?
if len(rs.sorted.topics) == 0 && len(rs.sorted.thats) == 0 {
return errors.New("SortReplies: ended up with empty trigger lists; did you load any RiveScript code?")
}
return nil
}
/*
sortTriggerSet sorts a group of triggers in an optimal sorting order.
This function has two use cases:
1. Create a sort buffer for "normal" (matchable) triggers, which are triggers
that are NOT accompanied by a %Previous tag.
2. Create a sort buffer for triggers that had %Previous tags.
Use the `excludePrevious` parameter to control which one is being done. This
function will return a list of sortedTriggerEntry items, and it's intended to
have no duplicate trigger patterns (unless the source RiveScript code explicitly
uses the same duplicate pattern twice, which is a user error).
*/
func (rs *RiveScript) sortTriggerSet(triggers []sortedTriggerEntry, excludePrevious bool) []sortedTriggerEntry {
// Create a priority map, of priority numbers -> their triggers.
prior := map[int][]sortedTriggerEntry{}
// Go through and bucket each trigger by weight (priority).
for _, trig := range triggers {
if excludePrevious && trig.pointer.previous != "" {
continue
}
// Check the trigger text for any {weight} tags, default being 0
match := reWeight.FindStringSubmatch(trig.trigger)
weight := 0
if len(match) > 0 {
weight, _ = strconv.Atoi(match[1])
}
// First trigger of this priority? Initialize the weight map.
if _, ok := prior[weight]; !ok {
prior[weight] = []sortedTriggerEntry{}
}
prior[weight] = append(prior[weight], trig)
}
// Keep a running list of sorted triggers for this topic.
running := []sortedTriggerEntry{}
// Sort the priorities with the highest number first.
var sortedPriorities []int
for k := range prior {
sortedPriorities = append(sortedPriorities, k)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedPriorities)))
// Go through each priority set.
for _, p := range sortedPriorities {
rs.say("Sorting triggers with priority %d", p)
// So, some of these triggers may include an {inherits} tag, if they
// came from a topic which inherits another topic. Lower inherits values
// mean higher priority on the stack. Triggers that have NO inherits
// value at all (which will default to -1), will be moved to the END of
// the stack at the end (have the highest number/lowest priority).
inherits := -1 // -1 means no {inherits} tag
highestInherits := -1 // Highest number seen so far
// Loop through and categorize these triggers.
track := map[int]*sortTrack{}
track[inherits] = initSortTrack()
// Loop through all the triggers.
for _, trig := range prior[p] {
pattern := trig.trigger
rs.say("Looking at trigger: %s", pattern)
// See if the trigger has an {inherits} tag.
match := reInherits.FindStringSubmatch(pattern)
if len(match) > 0 {
inherits, _ = strconv.Atoi(match[1])
if inherits > highestInherits {
highestInherits = inherits
}
rs.say("Trigger belongs to a topic that inherits other topics. "+
"Level=%d", inherits)
pattern = reInherits.ReplaceAllString(pattern, "")
} else {
inherits = -1
}
// If this is the first time we've seen this inheritance level,
// initialize its sort track structure.
if _, ok := track[inherits]; !ok {
track[inherits] = initSortTrack()
}
// Start inspecting the trigger's contents.
if strings.Contains(pattern, "_") {
// Alphabetic wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a _ wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].alpha[cnt]; !ok {
track[inherits].alpha[cnt] = []sortedTriggerEntry{}
}
track[inherits].alpha[cnt] = append(track[inherits].alpha[cnt], trig)
} else {
track[inherits].under = append(track[inherits].under, trig)
}
} else if strings.Contains(pattern, "#") {
// Numeric wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a # wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].number[cnt]; !ok {
track[inherits].number[cnt] = []sortedTriggerEntry{}
}
track[inherits].number[cnt] = append(track[inherits].number[cnt], trig)
} else {
track[inherits].pound = append(track[inherits].pound, trig)
}
} else if strings.Contains(pattern, "*") {
// Wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a * wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].wild[cnt]; !ok {
track[inherits].wild[cnt] = []sortedTriggerEntry{}
}
track[inherits].wild[cnt] = append(track[inherits].wild[cnt], trig)
} else {
track[inherits].star = append(track[inherits].star, trig)
}
} else if strings.Contains(pattern, "[") {
// Optionals included.
cnt := wordCount(pattern, false)
rs.say("Has optionals with %d words", cnt)
if _, ok := track[inherits].option[cnt]; !ok {
track[inherits].option[cnt] = []sortedTriggerEntry{}
}
track[inherits].option[cnt] = append(track[inherits].option[cnt], trig)
} else {
// Totally atomic.
cnt := wordCount(pattern, false)
rs.say("Totally atomic trigger with %d words", cnt)
if _, ok := track[inherits].atomic[cnt]; !ok {
track[inherits].atomic[cnt] = []sortedTriggerEntry{}
}
track[inherits].atomic[cnt] = append(track[inherits].atomic[cnt], trig)
}
}
// Move the no-{inherits} triggers to the bottom of the stack.
track[highestInherits+1] = track[-1]
delete(track, -1)
// Sort the track from the lowest to the highest.
var trackSorted []int
for k := range track {
trackSorted = append(trackSorted, k)
}
sort.Ints(trackSorted)
// Go through each priority level from greatest to smallest.
for _, ip := range trackSorted {
rs.say("ip=%d", ip)
// Sort each of the main kinds of triggers by their word counts.
running = sortByWords(running, track[ip].atomic)
running = sortByWords(running, track[ip].option)
running = sortByWords(running, track[ip].alpha)
running = sortByWords(running, track[ip].number)
running = sortByWords(running, track[ip].wild)
// Add the single wildcard triggers, sorted by length.
running = sortByLength(running, track[ip].under)
running = sortByLength(running, track[ip].pound)
running = sortByLength(running, track[ip].star)
}
}
return running
}
// sortList sorts lists (like substitutions) from a string:string map.
func sortList(dict map[string]string) []string {
output := []string{}
// Track by number of words.
track := map[int][]string{}
// Loop through each item.
for item := range dict {
cnt := wordCount(item, true)
if _, ok := track[cnt]; !ok {
track[cnt] = []string{}
}
track[cnt] = append(track[cnt], item)
}
// Sort them by word count, descending.
sortedCounts := []int{}
for cnt := range track {
sortedCounts = append(sortedCounts, cnt)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedCounts)))
for _, cnt := range sortedCounts {
// Sort the strings of this word-count by their lengths.
sortedLengths := track[cnt]
sort.Sort(sort.Reverse(byLength(sortedLengths)))
output = append(output, sortedLengths...)
}
return output
}
/*
sortByWords sorts a set of triggers by word count and overall length.
This is a helper function for sorting the `atomic`, `option`, `alpha`, `number`
and `wild` attributes of the sortTrack and adding them to the running sort
buffer in that specific order. Since attribute lookup by reflection is expensive
in Go, this function is given the relevant sort buffer directly, and the current
running sort buffer to add the results to.
The `triggers` parameter is a map between word counts and the triggers that
fit that number of words.
*/
func sortByWords(running []sortedTriggerEntry, triggers map[int][]sortedTriggerEntry) []sortedTriggerEntry {
// Sort the triggers by their word counts from greatest to smallest.
var sortedWords []int
for wc := range triggers {
sortedWords = append(sortedWords, wc)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedWords)))
for _, wc := range sortedWords {
// Triggers with equal word lengths should be sorted by overall trigger length.
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers[wc] {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Add the triggers to the running triggers bucket.
for _, pattern := range sortedPatterns {
running = append(running, patternMap[pattern]...)
}
}
return running
}
/*
sortByLength sorts a set of triggers purely by character length.
This is like `sortByWords`, but it's intended for triggers that consist solely
of wildcard-like symbols with no real words. For example a trigger of `* * *`
qualifies for this, and it has no words, so we sort by length so it gets a
higher priority than simply `*`.
*/
func sortByLength(running []sortedTriggerEntry, triggers []sortedTriggerEntry) []sortedTriggerEntry {
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Only loop through unique patterns.
patternSet := map[string]bool{}
// Add them to the running triggers bucket.
for _, pattern := range sortedPatterns {
if _, ok := patternSet[pattern]; ok {
continue
}
patternSet[pattern] = true
running = append(running, patternMap[pattern]...)
}
return running
}
// initSortTrack initializes a new, empty sortTrack object.
func initSortTrack() *sortTrack | {
return &sortTrack{
atomic: map[int][]sortedTriggerEntry{},
option: map[int][]sortedTriggerEntry{},
alpha: map[int][]sortedTriggerEntry{},
number: map[int][]sortedTriggerEntry{},
wild: map[int][]sortedTriggerEntry{},
pound: []sortedTriggerEntry{},
under: []sortedTriggerEntry{},
star: []sortedTriggerEntry{},
}
} | identifier_body | |
sorting.go | package rivescript
// Data sorting functions
import (
"errors"
"sort"
"strconv"
"strings"
)
// Sort buffer data, for RiveScript.SortReplies()
type sortBuffer struct {
topics map[string][]sortedTriggerEntry // Topic name -> array of triggers
thats map[string][]sortedTriggerEntry
sub []string // Substitutions
person []string // Person substitutions
}
// Holds a sorted trigger and the pointer to that trigger's data
type sortedTriggerEntry struct {
trigger string
pointer *astTrigger
}
// Temporary categorization of triggers while sorting
type sortTrack struct {
atomic map[int][]sortedTriggerEntry // Sort by number of whole words
option map[int][]sortedTriggerEntry // Sort optionals by number of words
alpha map[int][]sortedTriggerEntry // Sort alpha wildcards by no. of words
number map[int][]sortedTriggerEntry // Sort numeric wildcards by no. of words
wild map[int][]sortedTriggerEntry // Sort wildcards by no. of words
pound []sortedTriggerEntry // Triggers of just '#'
under []sortedTriggerEntry // Triggers of just '_'
star []sortedTriggerEntry // Triggers of just '*'
}
/*
SortReplies sorts the reply structures in memory for optimal matching.
After you have finished loading your RiveScript code, call this method to
populate the various sort buffers. This is absolutely necessary for reply
matching to work efficiently!
If the bot has loaded no topics, or if it ends up with no sorted triggers at
the end, it will return an error saying such. This usually means the bot didn't
load any RiveScript code, for example because it looked in the wrong directory.
*/
func (rs *RiveScript) SortReplies() error {
// (Re)initialize the sort cache.
rs.sorted.topics = map[string][]sortedTriggerEntry{}
rs.sorted.thats = map[string][]sortedTriggerEntry{}
rs.say("Sorting triggers...")
// If there are no topics, give an error.
if len(rs.topics) == 0 {
return errors.New("SortReplies: no topics were found; did you load any RiveScript code?")
}
// Loop through all the topics.
for topic := range rs.topics {
rs.say("Analyzing topic %s", topic)
// Collect a list of all the triggers we're going to worry about. If this
// topic inherits another topic, we need to recursively add those to the
// list as well.
allTriggers := rs.getTopicTriggers(topic, false)
// Sort these triggers.
rs.sorted.topics[topic] = rs.sortTriggerSet(allTriggers, true)
// Get all of the %Previous triggers for this topic.
thatTriggers := rs.getTopicTriggers(topic, true)
// And sort them, too.
rs.sorted.thats[topic] = rs.sortTriggerSet(thatTriggers, false)
}
// Sort the substitution lists.
rs.sorted.sub = sortList(rs.sub)
rs.sorted.person = sortList(rs.person)
// Did we sort anything at all?
if len(rs.sorted.topics) == 0 && len(rs.sorted.thats) == 0 {
return errors.New("SortReplies: ended up with empty trigger lists; did you load any RiveScript code?")
}
return nil
}
/*
sortTriggerSet sorts a group of triggers in an optimal sorting order.
This function has two use cases:
1. Create a sort buffer for "normal" (matchable) triggers, which are triggers
that are NOT accompanied by a %Previous tag.
2. Create a sort buffer for triggers that had %Previous tags.
Use the `excludePrevious` parameter to control which one is being done. This
function will return a list of sortedTriggerEntry items, and it's intended to
have no duplicate trigger patterns (unless the source RiveScript code explicitly
uses the same duplicate pattern twice, which is a user error).
*/
func (rs *RiveScript) sortTriggerSet(triggers []sortedTriggerEntry, excludePrevious bool) []sortedTriggerEntry {
// Create a priority map, of priority numbers -> their triggers.
prior := map[int][]sortedTriggerEntry{}
// Go through and bucket each trigger by weight (priority).
for _, trig := range triggers {
if excludePrevious && trig.pointer.previous != "" {
continue
}
// Check the trigger text for any {weight} tags, default being 0
match := reWeight.FindStringSubmatch(trig.trigger)
weight := 0
if len(match) > 0 {
weight, _ = strconv.Atoi(match[1])
}
// First trigger of this priority? Initialize the weight map.
if _, ok := prior[weight]; !ok {
prior[weight] = []sortedTriggerEntry{}
}
prior[weight] = append(prior[weight], trig)
}
// Keep a running list of sorted triggers for this topic.
running := []sortedTriggerEntry{}
// Sort the priorities with the highest number first.
var sortedPriorities []int
for k := range prior {
sortedPriorities = append(sortedPriorities, k)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedPriorities)))
// Go through each priority set.
for _, p := range sortedPriorities {
rs.say("Sorting triggers with priority %d", p)
// So, some of these triggers may include an {inherits} tag, if they
// came from a topic which inherits another topic. Lower inherits values
// mean higher priority on the stack. Triggers that have NO inherits
// value at all (which will default to -1), will be moved to the END of
// the stack at the end (have the highest number/lowest priority).
inherits := -1 // -1 means no {inherits} tag
highestInherits := -1 // Highest number seen so far
// Loop through and categorize these triggers.
track := map[int]*sortTrack{}
track[inherits] = initSortTrack()
// Loop through all the triggers.
for _, trig := range prior[p] {
pattern := trig.trigger
rs.say("Looking at trigger: %s", pattern)
// See if the trigger has an {inherits} tag.
match := reInherits.FindStringSubmatch(pattern)
if len(match) > 0 {
inherits, _ = strconv.Atoi(match[1])
if inherits > highestInherits {
highestInherits = inherits
}
rs.say("Trigger belongs to a topic that inherits other topics. "+
"Level=%d", inherits)
pattern = reInherits.ReplaceAllString(pattern, "")
} else {
inherits = -1
}
// If this is the first time we've seen this inheritance level,
// initialize its sort track structure.
if _, ok := track[inherits]; !ok {
track[inherits] = initSortTrack()
}
// Start inspecting the trigger's contents.
if strings.Contains(pattern, "_") {
// Alphabetic wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a _ wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].alpha[cnt]; !ok {
track[inherits].alpha[cnt] = []sortedTriggerEntry{}
}
track[inherits].alpha[cnt] = append(track[inherits].alpha[cnt], trig)
} else {
track[inherits].under = append(track[inherits].under, trig)
}
} else if strings.Contains(pattern, "#") {
// Numeric wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a # wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].number[cnt]; !ok {
track[inherits].number[cnt] = []sortedTriggerEntry{}
}
track[inherits].number[cnt] = append(track[inherits].number[cnt], trig)
} else {
track[inherits].pound = append(track[inherits].pound, trig)
}
} else if strings.Contains(pattern, "*") {
// Wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a * wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].wild[cnt]; !ok {
track[inherits].wild[cnt] = []sortedTriggerEntry{}
}
track[inherits].wild[cnt] = append(track[inherits].wild[cnt], trig)
} else {
track[inherits].star = append(track[inherits].star, trig)
}
} else if strings.Contains(pattern, "[") {
// Optionals included.
cnt := wordCount(pattern, false)
rs.say("Has optionals with %d words", cnt)
if _, ok := track[inherits].option[cnt]; !ok {
track[inherits].option[cnt] = []sortedTriggerEntry{}
}
track[inherits].option[cnt] = append(track[inherits].option[cnt], trig)
} else {
// Totally atomic.
cnt := wordCount(pattern, false) | track[inherits].atomic[cnt] = append(track[inherits].atomic[cnt], trig)
}
}
// Move the no-{inherits} triggers to the bottom of the stack.
track[highestInherits+1] = track[-1]
delete(track, -1)
// Sort the track from the lowest to the highest.
var trackSorted []int
for k := range track {
trackSorted = append(trackSorted, k)
}
sort.Ints(trackSorted)
// Go through each priority level from greatest to smallest.
for _, ip := range trackSorted {
rs.say("ip=%d", ip)
// Sort each of the main kinds of triggers by their word counts.
running = sortByWords(running, track[ip].atomic)
running = sortByWords(running, track[ip].option)
running = sortByWords(running, track[ip].alpha)
running = sortByWords(running, track[ip].number)
running = sortByWords(running, track[ip].wild)
// Add the single wildcard triggers, sorted by length.
running = sortByLength(running, track[ip].under)
running = sortByLength(running, track[ip].pound)
running = sortByLength(running, track[ip].star)
}
}
return running
}
// sortList sorts lists (like substitutions) from a string:string map.
func sortList(dict map[string]string) []string {
output := []string{}
// Track by number of words.
track := map[int][]string{}
// Loop through each item.
for item := range dict {
cnt := wordCount(item, true)
if _, ok := track[cnt]; !ok {
track[cnt] = []string{}
}
track[cnt] = append(track[cnt], item)
}
// Sort them by word count, descending.
sortedCounts := []int{}
for cnt := range track {
sortedCounts = append(sortedCounts, cnt)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedCounts)))
for _, cnt := range sortedCounts {
// Sort the strings of this word-count by their lengths.
sortedLengths := track[cnt]
sort.Sort(sort.Reverse(byLength(sortedLengths)))
output = append(output, sortedLengths...)
}
return output
}
/*
sortByWords sorts a set of triggers by word count and overall length.
This is a helper function for sorting the `atomic`, `option`, `alpha`, `number`
and `wild` attributes of the sortTrack and adding them to the running sort
buffer in that specific order. Since attribute lookup by reflection is expensive
in Go, this function is given the relevant sort buffer directly, and the current
running sort buffer to add the results to.
The `triggers` parameter is a map between word counts and the triggers that
fit that number of words.
*/
func sortByWords(running []sortedTriggerEntry, triggers map[int][]sortedTriggerEntry) []sortedTriggerEntry {
// Sort the triggers by their word counts from greatest to smallest.
var sortedWords []int
for wc := range triggers {
sortedWords = append(sortedWords, wc)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedWords)))
for _, wc := range sortedWords {
// Triggers with equal word lengths should be sorted by overall trigger length.
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers[wc] {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Add the triggers to the running triggers bucket.
for _, pattern := range sortedPatterns {
running = append(running, patternMap[pattern]...)
}
}
return running
}
/*
sortByLength sorts a set of triggers purely by character length.
This is like `sortByWords`, but it's intended for triggers that consist solely
of wildcard-like symbols with no real words. For example a trigger of `* * *`
qualifies for this, and it has no words, so we sort by length so it gets a
higher priority than simply `*`.
*/
func sortByLength(running []sortedTriggerEntry, triggers []sortedTriggerEntry) []sortedTriggerEntry {
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Only loop through unique patterns.
patternSet := map[string]bool{}
// Add them to the running triggers bucket.
for _, pattern := range sortedPatterns {
if _, ok := patternSet[pattern]; ok {
continue
}
patternSet[pattern] = true
running = append(running, patternMap[pattern]...)
}
return running
}
// initSortTrack initializes a new, empty sortTrack object.
func initSortTrack() *sortTrack {
return &sortTrack{
atomic: map[int][]sortedTriggerEntry{},
option: map[int][]sortedTriggerEntry{},
alpha: map[int][]sortedTriggerEntry{},
number: map[int][]sortedTriggerEntry{},
wild: map[int][]sortedTriggerEntry{},
pound: []sortedTriggerEntry{},
under: []sortedTriggerEntry{},
star: []sortedTriggerEntry{},
}
} | rs.say("Totally atomic trigger with %d words", cnt)
if _, ok := track[inherits].atomic[cnt]; !ok {
track[inherits].atomic[cnt] = []sortedTriggerEntry{}
} | random_line_split |
sorting.go | package rivescript
// Data sorting functions
import (
"errors"
"sort"
"strconv"
"strings"
)
// Sort buffer data, for RiveScript.SortReplies()
type sortBuffer struct {
topics map[string][]sortedTriggerEntry // Topic name -> array of triggers
thats map[string][]sortedTriggerEntry
sub []string // Substitutions
person []string // Person substitutions
}
// Holds a sorted trigger and the pointer to that trigger's data
type sortedTriggerEntry struct {
trigger string
pointer *astTrigger
}
// Temporary categorization of triggers while sorting
type sortTrack struct {
atomic map[int][]sortedTriggerEntry // Sort by number of whole words
option map[int][]sortedTriggerEntry // Sort optionals by number of words
alpha map[int][]sortedTriggerEntry // Sort alpha wildcards by no. of words
number map[int][]sortedTriggerEntry // Sort numeric wildcards by no. of words
wild map[int][]sortedTriggerEntry // Sort wildcards by no. of words
pound []sortedTriggerEntry // Triggers of just '#'
under []sortedTriggerEntry // Triggers of just '_'
star []sortedTriggerEntry // Triggers of just '*'
}
/*
SortReplies sorts the reply structures in memory for optimal matching.
After you have finished loading your RiveScript code, call this method to
populate the various sort buffers. This is absolutely necessary for reply
matching to work efficiently!
If the bot has loaded no topics, or if it ends up with no sorted triggers at
the end, it will return an error saying such. This usually means the bot didn't
load any RiveScript code, for example because it looked in the wrong directory.
*/
func (rs *RiveScript) SortReplies() error {
// (Re)initialize the sort cache.
rs.sorted.topics = map[string][]sortedTriggerEntry{}
rs.sorted.thats = map[string][]sortedTriggerEntry{}
rs.say("Sorting triggers...")
// If there are no topics, give an error.
if len(rs.topics) == 0 {
return errors.New("SortReplies: no topics were found; did you load any RiveScript code?")
}
// Loop through all the topics.
for topic := range rs.topics {
rs.say("Analyzing topic %s", topic)
// Collect a list of all the triggers we're going to worry about. If this
// topic inherits another topic, we need to recursively add those to the
// list as well.
allTriggers := rs.getTopicTriggers(topic, false)
// Sort these triggers.
rs.sorted.topics[topic] = rs.sortTriggerSet(allTriggers, true)
// Get all of the %Previous triggers for this topic.
thatTriggers := rs.getTopicTriggers(topic, true)
// And sort them, too.
rs.sorted.thats[topic] = rs.sortTriggerSet(thatTriggers, false)
}
// Sort the substitution lists.
rs.sorted.sub = sortList(rs.sub)
rs.sorted.person = sortList(rs.person)
// Did we sort anything at all?
if len(rs.sorted.topics) == 0 && len(rs.sorted.thats) == 0 {
return errors.New("SortReplies: ended up with empty trigger lists; did you load any RiveScript code?")
}
return nil
}
/*
sortTriggerSet sorts a group of triggers in an optimal sorting order.
This function has two use cases:
1. Create a sort buffer for "normal" (matchable) triggers, which are triggers
that are NOT accompanied by a %Previous tag.
2. Create a sort buffer for triggers that had %Previous tags.
Use the `excludePrevious` parameter to control which one is being done. This
function will return a list of sortedTriggerEntry items, and it's intended to
have no duplicate trigger patterns (unless the source RiveScript code explicitly
uses the same duplicate pattern twice, which is a user error).
*/
func (rs *RiveScript) sortTriggerSet(triggers []sortedTriggerEntry, excludePrevious bool) []sortedTriggerEntry {
// Create a priority map, of priority numbers -> their triggers.
prior := map[int][]sortedTriggerEntry{}
// Go through and bucket each trigger by weight (priority).
for _, trig := range triggers {
if excludePrevious && trig.pointer.previous != "" {
continue
}
// Check the trigger text for any {weight} tags, default being 0
match := reWeight.FindStringSubmatch(trig.trigger)
weight := 0
if len(match) > 0 {
weight, _ = strconv.Atoi(match[1])
}
// First trigger of this priority? Initialize the weight map.
if _, ok := prior[weight]; !ok {
prior[weight] = []sortedTriggerEntry{}
}
prior[weight] = append(prior[weight], trig)
}
// Keep a running list of sorted triggers for this topic.
running := []sortedTriggerEntry{}
// Sort the priorities with the highest number first.
var sortedPriorities []int
for k := range prior {
sortedPriorities = append(sortedPriorities, k)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedPriorities)))
// Go through each priority set.
for _, p := range sortedPriorities {
rs.say("Sorting triggers with priority %d", p)
// So, some of these triggers may include an {inherits} tag, if they
// came from a topic which inherits another topic. Lower inherits values
// mean higher priority on the stack. Triggers that have NO inherits
// value at all (which will default to -1), will be moved to the END of
// the stack at the end (have the highest number/lowest priority).
inherits := -1 // -1 means no {inherits} tag
highestInherits := -1 // Highest number seen so far
// Loop through and categorize these triggers.
track := map[int]*sortTrack{}
track[inherits] = initSortTrack()
// Loop through all the triggers.
for _, trig := range prior[p] {
pattern := trig.trigger
rs.say("Looking at trigger: %s", pattern)
// See if the trigger has an {inherits} tag.
match := reInherits.FindStringSubmatch(pattern)
if len(match) > 0 {
inherits, _ = strconv.Atoi(match[1])
if inherits > highestInherits {
highestInherits = inherits
}
rs.say("Trigger belongs to a topic that inherits other topics. "+
"Level=%d", inherits)
pattern = reInherits.ReplaceAllString(pattern, "")
} else {
inherits = -1
}
// If this is the first time we've seen this inheritance level,
// initialize its sort track structure.
if _, ok := track[inherits]; !ok {
track[inherits] = initSortTrack()
}
// Start inspecting the trigger's contents.
if strings.Contains(pattern, "_") {
// Alphabetic wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a _ wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].alpha[cnt]; !ok {
track[inherits].alpha[cnt] = []sortedTriggerEntry{}
}
track[inherits].alpha[cnt] = append(track[inherits].alpha[cnt], trig)
} else {
track[inherits].under = append(track[inherits].under, trig)
}
} else if strings.Contains(pattern, "#") {
// Numeric wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a # wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].number[cnt]; !ok {
track[inherits].number[cnt] = []sortedTriggerEntry{}
}
track[inherits].number[cnt] = append(track[inherits].number[cnt], trig)
} else {
track[inherits].pound = append(track[inherits].pound, trig)
}
} else if strings.Contains(pattern, "*") {
// Wildcard included.
cnt := wordCount(pattern, false)
rs.say("Has a * wildcard with %d words", cnt)
if cnt > 0 {
if _, ok := track[inherits].wild[cnt]; !ok {
track[inherits].wild[cnt] = []sortedTriggerEntry{}
}
track[inherits].wild[cnt] = append(track[inherits].wild[cnt], trig)
} else {
track[inherits].star = append(track[inherits].star, trig)
}
} else if strings.Contains(pattern, "[") {
// Optionals included.
cnt := wordCount(pattern, false)
rs.say("Has optionals with %d words", cnt)
if _, ok := track[inherits].option[cnt]; !ok {
track[inherits].option[cnt] = []sortedTriggerEntry{}
}
track[inherits].option[cnt] = append(track[inherits].option[cnt], trig)
} else {
// Totally atomic.
cnt := wordCount(pattern, false)
rs.say("Totally atomic trigger with %d words", cnt)
if _, ok := track[inherits].atomic[cnt]; !ok {
track[inherits].atomic[cnt] = []sortedTriggerEntry{}
}
track[inherits].atomic[cnt] = append(track[inherits].atomic[cnt], trig)
}
}
// Move the no-{inherits} triggers to the bottom of the stack.
track[highestInherits+1] = track[-1]
delete(track, -1)
// Sort the track from the lowest to the highest.
var trackSorted []int
for k := range track {
trackSorted = append(trackSorted, k)
}
sort.Ints(trackSorted)
// Go through each priority level from greatest to smallest.
for _, ip := range trackSorted {
rs.say("ip=%d", ip)
// Sort each of the main kinds of triggers by their word counts.
running = sortByWords(running, track[ip].atomic)
running = sortByWords(running, track[ip].option)
running = sortByWords(running, track[ip].alpha)
running = sortByWords(running, track[ip].number)
running = sortByWords(running, track[ip].wild)
// Add the single wildcard triggers, sorted by length.
running = sortByLength(running, track[ip].under)
running = sortByLength(running, track[ip].pound)
running = sortByLength(running, track[ip].star)
}
}
return running
}
// sortList sorts lists (like substitutions) from a string:string map.
func sortList(dict map[string]string) []string {
output := []string{}
// Track by number of words.
track := map[int][]string{}
// Loop through each item.
for item := range dict {
cnt := wordCount(item, true)
if _, ok := track[cnt]; !ok {
track[cnt] = []string{}
}
track[cnt] = append(track[cnt], item)
}
// Sort them by word count, descending.
sortedCounts := []int{}
for cnt := range track {
sortedCounts = append(sortedCounts, cnt)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedCounts)))
for _, cnt := range sortedCounts {
// Sort the strings of this word-count by their lengths.
sortedLengths := track[cnt]
sort.Sort(sort.Reverse(byLength(sortedLengths)))
output = append(output, sortedLengths...)
}
return output
}
/*
sortByWords sorts a set of triggers by word count and overall length.
This is a helper function for sorting the `atomic`, `option`, `alpha`, `number`
and `wild` attributes of the sortTrack and adding them to the running sort
buffer in that specific order. Since attribute lookup by reflection is expensive
in Go, this function is given the relevant sort buffer directly, and the current
running sort buffer to add the results to.
The `triggers` parameter is a map between word counts and the triggers that
fit that number of words.
*/
func sortByWords(running []sortedTriggerEntry, triggers map[int][]sortedTriggerEntry) []sortedTriggerEntry {
// Sort the triggers by their word counts from greatest to smallest.
var sortedWords []int
for wc := range triggers {
sortedWords = append(sortedWords, wc)
}
sort.Sort(sort.Reverse(sort.IntSlice(sortedWords)))
for _, wc := range sortedWords {
// Triggers with equal word lengths should be sorted by overall trigger length.
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers[wc] {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Add the triggers to the running triggers bucket.
for _, pattern := range sortedPatterns {
running = append(running, patternMap[pattern]...)
}
}
return running
}
/*
sortByLength sorts a set of triggers purely by character length.
This is like `sortByWords`, but it's intended for triggers that consist solely
of wildcard-like symbols with no real words. For example a trigger of `* * *`
qualifies for this, and it has no words, so we sort by length so it gets a
higher priority than simply `*`.
*/
func sortByLength(running []sortedTriggerEntry, triggers []sortedTriggerEntry) []sortedTriggerEntry {
var sortedPatterns []string
patternMap := map[string][]sortedTriggerEntry{}
for _, trig := range triggers {
sortedPatterns = append(sortedPatterns, trig.trigger)
if _, ok := patternMap[trig.trigger]; !ok {
patternMap[trig.trigger] = []sortedTriggerEntry{}
}
patternMap[trig.trigger] = append(patternMap[trig.trigger], trig)
}
sort.Sort(sort.Reverse(byLength(sortedPatterns)))
// Only loop through unique patterns.
patternSet := map[string]bool{}
// Add them to the running triggers bucket.
for _, pattern := range sortedPatterns {
if _, ok := patternSet[pattern]; ok |
patternSet[pattern] = true
running = append(running, patternMap[pattern]...)
}
return running
}
// initSortTrack initializes a new, empty sortTrack object.
func initSortTrack() *sortTrack {
return &sortTrack{
atomic: map[int][]sortedTriggerEntry{},
option: map[int][]sortedTriggerEntry{},
alpha: map[int][]sortedTriggerEntry{},
number: map[int][]sortedTriggerEntry{},
wild: map[int][]sortedTriggerEntry{},
pound: []sortedTriggerEntry{},
under: []sortedTriggerEntry{},
star: []sortedTriggerEntry{},
}
}
| {
continue
} | conditional_block |
app.js | var SECONDS = 59;
var MINUTES = 3;
var SCORE = [0,0];
var VALUE1;
var VALUE2;
var sliceCount = 0;
var p1data = [];
var p2data = [];
var p1trend = "0";
var p2trend = "0";
var multiplierA;
var multiplierB;
var makePointsFlag = true;
var P1NAME;
var P2NAME;
var P1EMAIL;
var P2EMAIL;
var COMBONAME = null;
function add(a, b) {
return a + b;
}
function aggregateAverage () {//calculates the average SCORE of all players
var average = 0; //holds a cumulative list of all of the player scores
var tempString = localStorage.getItem("cumulativescores");
var allScores = tempString.split(",") ? tempString.split(",") :"0,0";
average = allScores.reduce(function(a, b){a= parseFloat(a);b= parseFloat(b); return a+b;})
// for (var i = 0; i < allScores.length; i++){
average = Math.round(average/allScores.length);
document.getElementById("average").innerHTML = (average);
}
function average (data) {
var sum = data.reduce(function(sum, value){
return sum + value;
}, 0);
var avg = sum / data.length;
return avg;
}
function contactNameInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[0].split("=")[1];
nameTwo = chopped[1].split("=")[1];
var re = new RegExp("\s*([A-Za-z\d]+)\s*");
nameOne = re.exec(nameOne)[0];
nameTwo = re.exec(nameTwo)[0];
P1NAME = nameOne;
P2NAME = nameTwo;
coupleName = this.nameMash (nameOne, nameTwo);
COMBONAME = coupleName;
document.getElementById("player-1").innerHTML = (nameOne + "'s Score is:");
document.getElementById("player-2").innerHTML = (nameTwo + "'s Score is:");
document.getElementById("couple-name").innerHTML = ("Couple Name: " + coupleName);
}
function contactEmailInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
var atRegex = /%40/g;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[2].split("=")[1];
nameTwo = chopped[3].split("=")[1];
P1EMAIL = nameOne.replace(atRegex, '@');
P2EMAIL = nameTwo.replace(atRegex,'@');
document.getElementById("email-message").innerHTML = ("Thanks! We will be getting back to you sometime before the next burn!");
}
function datesTimes () {//grabs the time... no network or persistant time... Need a GPS module for the Raspberry PI
var cleanDate;
cleanDate = new Date().toLocaleString()
localStorage.setItem("timestamp", cleanDate);
}
function dataLoader(){//pulls the data from a local JSON file
$.getJSON("http://localhost:8000/Love-Staring-Machine/result.json", function(data) {
VALUE1 = data["player1"];
VALUE2 = data["player2"];
});
}
function finalMessage () {//Delivers the final message in the form of a modal overlay based on compadability of scores
//derive a compatability index based on distance of final scores
var p1Score = SCORE[0];
var p2Score = SCORE[1];
var higherScore = p1Score > p2Score ? p1Score : p2Score;
var distance = Math.abs(p1Score - p2Score);
var relDistance = distance / higherScore;
var message = [
"Hell No!!!",
"Y'all Fuck",
"True Love!"
];
var messenger;
if (relDistance > .1) {
messenger = 0;
}
else if (relDistance > .05) {
messenger = 1;
}
else {
messenger = 2;
}
//render the modal overlay
document.getElementById("timer-readout").innerHTML = (message[messenger]);
}
function finalData () {
//Stpres all he data as an object
var finalObject = {};
var totalFinalObjects = JSON.parse(localStorage.getItem("finalobjects"));
var participants = localStorage.getItem("totalparticipants");
var p1Name = P1NAME ? P1NAME : null;
p2Name = P2NAME ? P2NAME : null;
var coupleName = COMBONAME ? COMBONAME : null;
p1Email = P1EMAIL ? P1EMAIL : null;
p2Email = P2EMAIL ? P2EMAIL : null;
//grabs from localstorage the entire
totalFinalObjects[participants] = {
'p1': {
'n': p1Name,
'e': p1Email,
'p': SCORE[0],
's': p1data
},
'p2' : {
'n': p2Name,
'e': p1Email,
'p': SCORE[1],
's': p2data
},
'n': COMBONAME
};
localStorage.setItem("finalobjects", JSON.stringify(totalFinalObjects));
}
function highScore () {//checks if any of the present SCOREs are higher than the all time high SCORE
var tempSCORE = parseInt(localStorage.getItem("highScore"));
if (SCORE[0] > tempSCORE) {
localStorage.setItem("highScore", SCORE[0]);
}
if (SCORE[1] > tempSCORE) {
localStorage.setItem("highScore", SCORE[1]);
}
}
function initialization () {//initializes the persistant dashboard metrics for the first time
//add all of the local storage variables
//localStorage.setItem("highScore", 0);
//localStorage.setItem("totalparticipants", 0);
//localStorage.setItem("average", 0);
//localStorage.setItem("cumulativescores", 0)
localStorage.setItem("finalobjects", "{}");
}
function myTimer() {//Simple, non-accurate clock funtion
var secs = "" + SECONDS;
var min = MINUTES;
if ( MINUTES >= 0){
SECONDS --;
if (SECONDS < 9) {
secs = "0" + secs;
}
if (SECONDS === 0){
SECONDS = 59;
secs = "00";//TODO: fix this where the at 1:00 it displays :00
MINUTES --;
}
if (MINUTES > 0){
document.getElementById("timer-readout").innerHTML = (min + ":" + secs);
}
else {
document.getElementById("timer-readout").innerHTML = (":" + secs);
}
}
else {
makePointsFlag = false;
storeScoresForAverages();
finalMessage();
finalData();
clearTimeout(loop);
clearTimeout(slice);
clearTimeout(countdown);
clearTimeout(points);
clearTimeout(leader);
}
}
function nameMash (p1,p2) {//celebrity name generator splits on the vowels or if all else fails just smash them together
if (p1.length > 5 || p2.length > 5) {//TODO: Clean up this logic and make it more compact
//use the p1 and p2 arguments and loop through to grab the regex vowels
var re = /[aeiou]/gi;
var p1Array = p1.match(re);
var p2Array = p2.match(re);
//if there are more than 2 split on the second
if (p1Array.length >= 2 && p2Array.length >= 2) {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == p1Array.length){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == p2Array.length){
var player2 = p2.substring(match.index - 2, p2.length);
}
}
return player1 + player2;
}
else {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == 1){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == 1){
var player2 = p2.substring(match.index - 1, p2.length);
}
}
return player1 + player2;
}
}
else {
if (p1.length > 2 && p2.length > 2){
return p1.substring(0,1).toUpperCase() + p1.substring(1,Math.floor(p1.length/2)).toLowerCase() + p2.substring(Math.floor(p2.length/2),p2.length).toLowerCase();
}
else {
return p1.substring(0,1).toUpperCase() + p1.substring(1, p1.length).toLowerCase() + p2.substring(2,p2.length).toLowerCase();
}
}
}
function onRefreshInitialization () {//initializes the persistant dashboard metrics
document.getElementById('p1').focus();
$('#form')[0].reset();
}
function overlay() {
el = document.getElementById("overlay");
el.style.visibility = (el.style.visibility == "visible") ? "hidden" : "visible";
}
function pointGenerator () {//this function computes the points
//TODO: Make a funtion to check directional trends and integrate the result into this function
if (makePointsFlag) {
var trendValue1,
trendValue2;
trendOfReadings();
if (p1trend === '0'){
trendValue1 = 9;
}
else if (p1trend === '+'){
trendValue1 = 30;
}
else if (p1trend === '-'){
trendValue1 = 5;
}
else {
trendValue1 = 1;
}
if (p2trend === '0'){
trendValue2 = 9;
}
else if (p2trend === '+'){
trendValue2 = 30;
}
else if (p2trend === '-') |
else {
trendValue2 = 1;
}
trendValue1 ? SCORE[0] += Math.round(trendValue1) : null;
trendValue2 ? SCORE[1] += Math.round(trendValue2) : null;
scoreRender();
}
}
function scoreRender () {
document.getElementById("player-1-score").innerHTML = SCORE[0] ? SCORE[0] : "Loading...";
document.getElementById("player-2-score").innerHTML = SCORE[1] ? SCORE[1] : "Loading...";
//sets the high scoreRender
document.getElementById("high-score").innerHTML = (localStorage.getItem("highScore"));
}
function standardDeviation(values){
var avg = average(values);
var squareDiffs = values.map(function(value){
var diff = value - avg;
var sqrDiff = diff * diff;
return sqrDiff;
});
var avgSquareDiff = average(squareDiffs);
var stdDev = Math.sqrt(avgSquareDiff);
return stdDev;
}
function storeScoresForAverages() {
var tempArray = [];
tempArray.push(localStorage.getItem("cumulativescores"));
tempArray.push(SCORE[0]);
tempArray.push(SCORE[1]);
localStorage.setItem("cumulativescores", tempArray);
}
function timeSlice () {
p1data[sliceCount] = VALUE1;
p2data[sliceCount] = VALUE2;
sliceCount ++;
//Store 240 time slices (one per second) for both players
multiplierA = standardDeviation(p1data);
multiplierB = standardDeviation(p2data);
//this data is used for modeling the results
//creates a baseline to determine if disconnet has occured
//writes the list of time slices into localstorage
localStorage.setItem("player1slices", p1data);
localStorage.setItem("player2slices", p2data);
}
function totalParticipants () {//keeps a rolling total on how many times the test has been ran
var newTotal = parseInt(localStorage.getItem("totalparticipants")) + 1;
document.getElementById("participant-total").innerHTML = (newTotal);
localStorage.setItem("totalparticipants", newTotal);
}
function totalPoints () {//create a rolling total
var newTotal;
localStorage.setItem("averagepoints", rollingAverage);
localStorage.setItem("totalparticipants", newTotal);
}
function trendOfReadings () {//Takes in the last n readings and makes comparison to determine directional trending
var driftDebounce = 0; //tune based off of how much it wobbles
if (sliceCount > 5){//slice count is always ahead by one
//simple and needs refactoring
//TODO: take the last five averages and do a simple comparison for smoothing
if (p1data[sliceCount - 1] > p1data[sliceCount - 2] && p1data[sliceCount - 2] > p1data[sliceCount - 3]){
p1trend = "+";
}
else if (p1data[sliceCount - 1] < p1data[sliceCount - 2] && p1data[sliceCount - 2] < p1data[sliceCount - 3]){
p1trend = "-";
}
else {
p1trend = "0";
}
if (p2data[sliceCount - 1] > p2data[sliceCount - 2] && p2data[sliceCount - 2] > p2data[sliceCount - 3]){
p2trend = "+";
}
else if (p2data[sliceCount - 1] < p2data[sliceCount - 2] && p2data[sliceCount - 2] < p2data[sliceCount - 3]){
p2trend = "-";
}
else {
p2trend = "0";
}
}
}
///////////////////////////////////////////////////////////////////////////////
function render() {//keep all of the execution of the app within this function
//initialization();
onRefreshInitialization();
aggregateAverage();
storeScoresForAverages();
totalParticipants();
datesTimes();
pointGenerator();
dataLoader();
loop = setInterval(dataLoader, 1000);
slice = setInterval(timeSlice, 1000);
countdown = setInterval(myTimer,1000);
points = setInterval(pointGenerator, 633);
leader = setInterval(highScore,1000);
//TODO: store everything in local storage better for extraction at a later date
}
$(document).ready(function() {
render();
});
| {
trendValue2 = 5;
} | conditional_block |
app.js | var SECONDS = 59;
var MINUTES = 3;
var SCORE = [0,0];
var VALUE1;
var VALUE2;
var sliceCount = 0;
var p1data = [];
var p2data = [];
var p1trend = "0";
var p2trend = "0";
var multiplierA;
var multiplierB;
var makePointsFlag = true;
var P1NAME;
var P2NAME;
var P1EMAIL;
var P2EMAIL;
var COMBONAME = null;
function add(a, b) {
return a + b;
}
function aggregateAverage () {//calculates the average SCORE of all players
var average = 0; //holds a cumulative list of all of the player scores
var tempString = localStorage.getItem("cumulativescores");
var allScores = tempString.split(",") ? tempString.split(",") :"0,0";
average = allScores.reduce(function(a, b){a= parseFloat(a);b= parseFloat(b); return a+b;})
// for (var i = 0; i < allScores.length; i++){
average = Math.round(average/allScores.length);
document.getElementById("average").innerHTML = (average);
}
function average (data) {
var sum = data.reduce(function(sum, value){
return sum + value;
}, 0);
var avg = sum / data.length;
return avg;
}
function contactNameInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[0].split("=")[1];
nameTwo = chopped[1].split("=")[1];
var re = new RegExp("\s*([A-Za-z\d]+)\s*");
nameOne = re.exec(nameOne)[0];
nameTwo = re.exec(nameTwo)[0];
P1NAME = nameOne;
P2NAME = nameTwo;
coupleName = this.nameMash (nameOne, nameTwo);
COMBONAME = coupleName;
document.getElementById("player-1").innerHTML = (nameOne + "'s Score is:");
document.getElementById("player-2").innerHTML = (nameTwo + "'s Score is:");
document.getElementById("couple-name").innerHTML = ("Couple Name: " + coupleName);
}
function contactEmailInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
var atRegex = /%40/g;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[2].split("=")[1];
nameTwo = chopped[3].split("=")[1];
P1EMAIL = nameOne.replace(atRegex, '@');
P2EMAIL = nameTwo.replace(atRegex,'@');
document.getElementById("email-message").innerHTML = ("Thanks! We will be getting back to you sometime before the next burn!");
}
function datesTimes () {//grabs the time... no network or persistant time... Need a GPS module for the Raspberry PI
var cleanDate;
cleanDate = new Date().toLocaleString()
localStorage.setItem("timestamp", cleanDate);
}
function dataLoader(){//pulls the data from a local JSON file
$.getJSON("http://localhost:8000/Love-Staring-Machine/result.json", function(data) {
VALUE1 = data["player1"];
VALUE2 = data["player2"];
});
}
function finalMessage () {//Delivers the final message in the form of a modal overlay based on compadability of scores
//derive a compatability index based on distance of final scores
var p1Score = SCORE[0];
var p2Score = SCORE[1];
var higherScore = p1Score > p2Score ? p1Score : p2Score;
var distance = Math.abs(p1Score - p2Score);
var relDistance = distance / higherScore;
var message = [
"Hell No!!!",
"Y'all Fuck",
"True Love!"
];
var messenger;
if (relDistance > .1) {
messenger = 0;
}
else if (relDistance > .05) {
messenger = 1;
}
else {
messenger = 2;
}
//render the modal overlay
document.getElementById("timer-readout").innerHTML = (message[messenger]);
}
function finalData () {
//Stpres all he data as an object
var finalObject = {};
var totalFinalObjects = JSON.parse(localStorage.getItem("finalobjects"));
var participants = localStorage.getItem("totalparticipants");
var p1Name = P1NAME ? P1NAME : null;
p2Name = P2NAME ? P2NAME : null;
var coupleName = COMBONAME ? COMBONAME : null;
p1Email = P1EMAIL ? P1EMAIL : null;
p2Email = P2EMAIL ? P2EMAIL : null;
//grabs from localstorage the entire
totalFinalObjects[participants] = {
'p1': {
'n': p1Name,
'e': p1Email,
'p': SCORE[0],
's': p1data
},
'p2' : {
'n': p2Name,
'e': p1Email,
'p': SCORE[1],
's': p2data
},
'n': COMBONAME
};
localStorage.setItem("finalobjects", JSON.stringify(totalFinalObjects));
}
function highScore () |
function initialization () {//initializes the persistant dashboard metrics for the first time
//add all of the local storage variables
//localStorage.setItem("highScore", 0);
//localStorage.setItem("totalparticipants", 0);
//localStorage.setItem("average", 0);
//localStorage.setItem("cumulativescores", 0)
localStorage.setItem("finalobjects", "{}");
}
function myTimer() {//Simple, non-accurate clock funtion
var secs = "" + SECONDS;
var min = MINUTES;
if ( MINUTES >= 0){
SECONDS --;
if (SECONDS < 9) {
secs = "0" + secs;
}
if (SECONDS === 0){
SECONDS = 59;
secs = "00";//TODO: fix this where the at 1:00 it displays :00
MINUTES --;
}
if (MINUTES > 0){
document.getElementById("timer-readout").innerHTML = (min + ":" + secs);
}
else {
document.getElementById("timer-readout").innerHTML = (":" + secs);
}
}
else {
makePointsFlag = false;
storeScoresForAverages();
finalMessage();
finalData();
clearTimeout(loop);
clearTimeout(slice);
clearTimeout(countdown);
clearTimeout(points);
clearTimeout(leader);
}
}
function nameMash (p1,p2) {//celebrity name generator splits on the vowels or if all else fails just smash them together
if (p1.length > 5 || p2.length > 5) {//TODO: Clean up this logic and make it more compact
//use the p1 and p2 arguments and loop through to grab the regex vowels
var re = /[aeiou]/gi;
var p1Array = p1.match(re);
var p2Array = p2.match(re);
//if there are more than 2 split on the second
if (p1Array.length >= 2 && p2Array.length >= 2) {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == p1Array.length){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == p2Array.length){
var player2 = p2.substring(match.index - 2, p2.length);
}
}
return player1 + player2;
}
else {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == 1){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == 1){
var player2 = p2.substring(match.index - 1, p2.length);
}
}
return player1 + player2;
}
}
else {
if (p1.length > 2 && p2.length > 2){
return p1.substring(0,1).toUpperCase() + p1.substring(1,Math.floor(p1.length/2)).toLowerCase() + p2.substring(Math.floor(p2.length/2),p2.length).toLowerCase();
}
else {
return p1.substring(0,1).toUpperCase() + p1.substring(1, p1.length).toLowerCase() + p2.substring(2,p2.length).toLowerCase();
}
}
}
function onRefreshInitialization () {//initializes the persistant dashboard metrics
document.getElementById('p1').focus();
$('#form')[0].reset();
}
function overlay() {
el = document.getElementById("overlay");
el.style.visibility = (el.style.visibility == "visible") ? "hidden" : "visible";
}
function pointGenerator () {//this function computes the points
//TODO: Make a funtion to check directional trends and integrate the result into this function
if (makePointsFlag) {
var trendValue1,
trendValue2;
trendOfReadings();
if (p1trend === '0'){
trendValue1 = 9;
}
else if (p1trend === '+'){
trendValue1 = 30;
}
else if (p1trend === '-'){
trendValue1 = 5;
}
else {
trendValue1 = 1;
}
if (p2trend === '0'){
trendValue2 = 9;
}
else if (p2trend === '+'){
trendValue2 = 30;
}
else if (p2trend === '-'){
trendValue2 = 5;
}
else {
trendValue2 = 1;
}
trendValue1 ? SCORE[0] += Math.round(trendValue1) : null;
trendValue2 ? SCORE[1] += Math.round(trendValue2) : null;
scoreRender();
}
}
function scoreRender () {
document.getElementById("player-1-score").innerHTML = SCORE[0] ? SCORE[0] : "Loading...";
document.getElementById("player-2-score").innerHTML = SCORE[1] ? SCORE[1] : "Loading...";
//sets the high scoreRender
document.getElementById("high-score").innerHTML = (localStorage.getItem("highScore"));
}
function standardDeviation(values){
var avg = average(values);
var squareDiffs = values.map(function(value){
var diff = value - avg;
var sqrDiff = diff * diff;
return sqrDiff;
});
var avgSquareDiff = average(squareDiffs);
var stdDev = Math.sqrt(avgSquareDiff);
return stdDev;
}
function storeScoresForAverages() {
var tempArray = [];
tempArray.push(localStorage.getItem("cumulativescores"));
tempArray.push(SCORE[0]);
tempArray.push(SCORE[1]);
localStorage.setItem("cumulativescores", tempArray);
}
function timeSlice () {
p1data[sliceCount] = VALUE1;
p2data[sliceCount] = VALUE2;
sliceCount ++;
//Store 240 time slices (one per second) for both players
multiplierA = standardDeviation(p1data);
multiplierB = standardDeviation(p2data);
//this data is used for modeling the results
//creates a baseline to determine if disconnet has occured
//writes the list of time slices into localstorage
localStorage.setItem("player1slices", p1data);
localStorage.setItem("player2slices", p2data);
}
function totalParticipants () {//keeps a rolling total on how many times the test has been ran
var newTotal = parseInt(localStorage.getItem("totalparticipants")) + 1;
document.getElementById("participant-total").innerHTML = (newTotal);
localStorage.setItem("totalparticipants", newTotal);
}
function totalPoints () {//create a rolling total
var newTotal;
localStorage.setItem("averagepoints", rollingAverage);
localStorage.setItem("totalparticipants", newTotal);
}
function trendOfReadings () {//Takes in the last n readings and makes comparison to determine directional trending
var driftDebounce = 0; //tune based off of how much it wobbles
if (sliceCount > 5){//slice count is always ahead by one
//simple and needs refactoring
//TODO: take the last five averages and do a simple comparison for smoothing
if (p1data[sliceCount - 1] > p1data[sliceCount - 2] && p1data[sliceCount - 2] > p1data[sliceCount - 3]){
p1trend = "+";
}
else if (p1data[sliceCount - 1] < p1data[sliceCount - 2] && p1data[sliceCount - 2] < p1data[sliceCount - 3]){
p1trend = "-";
}
else {
p1trend = "0";
}
if (p2data[sliceCount - 1] > p2data[sliceCount - 2] && p2data[sliceCount - 2] > p2data[sliceCount - 3]){
p2trend = "+";
}
else if (p2data[sliceCount - 1] < p2data[sliceCount - 2] && p2data[sliceCount - 2] < p2data[sliceCount - 3]){
p2trend = "-";
}
else {
p2trend = "0";
}
}
}
///////////////////////////////////////////////////////////////////////////////
function render() {//keep all of the execution of the app within this function
//initialization();
onRefreshInitialization();
aggregateAverage();
storeScoresForAverages();
totalParticipants();
datesTimes();
pointGenerator();
dataLoader();
loop = setInterval(dataLoader, 1000);
slice = setInterval(timeSlice, 1000);
countdown = setInterval(myTimer,1000);
points = setInterval(pointGenerator, 633);
leader = setInterval(highScore,1000);
//TODO: store everything in local storage better for extraction at a later date
}
$(document).ready(function() {
render();
});
| {//checks if any of the present SCOREs are higher than the all time high SCORE
var tempSCORE = parseInt(localStorage.getItem("highScore"));
if (SCORE[0] > tempSCORE) {
localStorage.setItem("highScore", SCORE[0]);
}
if (SCORE[1] > tempSCORE) {
localStorage.setItem("highScore", SCORE[1]);
}
} | identifier_body |
app.js | var SECONDS = 59;
var MINUTES = 3;
var SCORE = [0,0];
var VALUE1;
var VALUE2;
var sliceCount = 0;
var p1data = [];
var p2data = [];
var p1trend = "0";
var p2trend = "0";
var multiplierA;
var multiplierB;
var makePointsFlag = true;
var P1NAME;
var P2NAME;
var P1EMAIL;
var P2EMAIL;
var COMBONAME = null;
function add(a, b) {
return a + b;
}
function aggregateAverage () {//calculates the average SCORE of all players
var average = 0; //holds a cumulative list of all of the player scores
var tempString = localStorage.getItem("cumulativescores");
var allScores = tempString.split(",") ? tempString.split(",") :"0,0";
average = allScores.reduce(function(a, b){a= parseFloat(a);b= parseFloat(b); return a+b;})
// for (var i = 0; i < allScores.length; i++){
average = Math.round(average/allScores.length);
document.getElementById("average").innerHTML = (average);
}
function average (data) {
var sum = data.reduce(function(sum, value){
return sum + value;
}, 0);
var avg = sum / data.length;
return avg;
}
function contactNameInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[0].split("=")[1];
nameTwo = chopped[1].split("=")[1];
var re = new RegExp("\s*([A-Za-z\d]+)\s*");
nameOne = re.exec(nameOne)[0];
nameTwo = re.exec(nameTwo)[0];
P1NAME = nameOne;
P2NAME = nameTwo;
coupleName = this.nameMash (nameOne, nameTwo);
COMBONAME = coupleName;
document.getElementById("player-1").innerHTML = (nameOne + "'s Score is:");
document.getElementById("player-2").innerHTML = (nameTwo + "'s Score is:");
document.getElementById("couple-name").innerHTML = ("Couple Name: " + coupleName);
}
function contactEmailInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
var atRegex = /%40/g;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[2].split("=")[1];
nameTwo = chopped[3].split("=")[1];
P1EMAIL = nameOne.replace(atRegex, '@');
P2EMAIL = nameTwo.replace(atRegex,'@');
document.getElementById("email-message").innerHTML = ("Thanks! We will be getting back to you sometime before the next burn!");
}
function datesTimes () {//grabs the time... no network or persistant time... Need a GPS module for the Raspberry PI
var cleanDate;
cleanDate = new Date().toLocaleString()
localStorage.setItem("timestamp", cleanDate);
}
function dataLoader(){//pulls the data from a local JSON file
$.getJSON("http://localhost:8000/Love-Staring-Machine/result.json", function(data) {
VALUE1 = data["player1"];
VALUE2 = data["player2"];
});
}
function finalMessage () {//Delivers the final message in the form of a modal overlay based on compadability of scores
//derive a compatability index based on distance of final scores
var p1Score = SCORE[0];
var p2Score = SCORE[1];
var higherScore = p1Score > p2Score ? p1Score : p2Score;
var distance = Math.abs(p1Score - p2Score);
var relDistance = distance / higherScore;
var message = [
"Hell No!!!",
"Y'all Fuck",
"True Love!"
];
var messenger;
if (relDistance > .1) {
messenger = 0;
}
else if (relDistance > .05) {
messenger = 1;
}
else {
messenger = 2;
}
//render the modal overlay
document.getElementById("timer-readout").innerHTML = (message[messenger]);
}
function finalData () {
//Stpres all he data as an object
var finalObject = {};
var totalFinalObjects = JSON.parse(localStorage.getItem("finalobjects"));
var participants = localStorage.getItem("totalparticipants");
var p1Name = P1NAME ? P1NAME : null;
p2Name = P2NAME ? P2NAME : null;
var coupleName = COMBONAME ? COMBONAME : null;
p1Email = P1EMAIL ? P1EMAIL : null;
p2Email = P2EMAIL ? P2EMAIL : null;
//grabs from localstorage the entire
totalFinalObjects[participants] = {
'p1': {
'n': p1Name,
'e': p1Email,
'p': SCORE[0],
's': p1data
},
'p2' : {
'n': p2Name,
'e': p1Email,
'p': SCORE[1],
's': p2data
},
'n': COMBONAME
|
function highScore () {//checks if any of the present SCOREs are higher than the all time high SCORE
var tempSCORE = parseInt(localStorage.getItem("highScore"));
if (SCORE[0] > tempSCORE) {
localStorage.setItem("highScore", SCORE[0]);
}
if (SCORE[1] > tempSCORE) {
localStorage.setItem("highScore", SCORE[1]);
}
}
function initialization () {//initializes the persistant dashboard metrics for the first time
//add all of the local storage variables
//localStorage.setItem("highScore", 0);
//localStorage.setItem("totalparticipants", 0);
//localStorage.setItem("average", 0);
//localStorage.setItem("cumulativescores", 0)
localStorage.setItem("finalobjects", "{}");
}
function myTimer() {//Simple, non-accurate clock funtion
var secs = "" + SECONDS;
var min = MINUTES;
if ( MINUTES >= 0){
SECONDS --;
if (SECONDS < 9) {
secs = "0" + secs;
}
if (SECONDS === 0){
SECONDS = 59;
secs = "00";//TODO: fix this where the at 1:00 it displays :00
MINUTES --;
}
if (MINUTES > 0){
document.getElementById("timer-readout").innerHTML = (min + ":" + secs);
}
else {
document.getElementById("timer-readout").innerHTML = (":" + secs);
}
}
else {
makePointsFlag = false;
storeScoresForAverages();
finalMessage();
finalData();
clearTimeout(loop);
clearTimeout(slice);
clearTimeout(countdown);
clearTimeout(points);
clearTimeout(leader);
}
}
function nameMash (p1,p2) {//celebrity name generator splits on the vowels or if all else fails just smash them together
if (p1.length > 5 || p2.length > 5) {//TODO: Clean up this logic and make it more compact
//use the p1 and p2 arguments and loop through to grab the regex vowels
var re = /[aeiou]/gi;
var p1Array = p1.match(re);
var p2Array = p2.match(re);
//if there are more than 2 split on the second
if (p1Array.length >= 2 && p2Array.length >= 2) {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == p1Array.length){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == p2Array.length){
var player2 = p2.substring(match.index - 2, p2.length);
}
}
return player1 + player2;
}
else {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == 1){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == 1){
var player2 = p2.substring(match.index - 1, p2.length);
}
}
return player1 + player2;
}
}
else {
if (p1.length > 2 && p2.length > 2){
return p1.substring(0,1).toUpperCase() + p1.substring(1,Math.floor(p1.length/2)).toLowerCase() + p2.substring(Math.floor(p2.length/2),p2.length).toLowerCase();
}
else {
return p1.substring(0,1).toUpperCase() + p1.substring(1, p1.length).toLowerCase() + p2.substring(2,p2.length).toLowerCase();
}
}
}
function onRefreshInitialization () {//initializes the persistant dashboard metrics
document.getElementById('p1').focus();
$('#form')[0].reset();
}
function overlay() {
el = document.getElementById("overlay");
el.style.visibility = (el.style.visibility == "visible") ? "hidden" : "visible";
}
function pointGenerator () {//this function computes the points
//TODO: Make a funtion to check directional trends and integrate the result into this function
if (makePointsFlag) {
var trendValue1,
trendValue2;
trendOfReadings();
if (p1trend === '0'){
trendValue1 = 9;
}
else if (p1trend === '+'){
trendValue1 = 30;
}
else if (p1trend === '-'){
trendValue1 = 5;
}
else {
trendValue1 = 1;
}
if (p2trend === '0'){
trendValue2 = 9;
}
else if (p2trend === '+'){
trendValue2 = 30;
}
else if (p2trend === '-'){
trendValue2 = 5;
}
else {
trendValue2 = 1;
}
trendValue1 ? SCORE[0] += Math.round(trendValue1) : null;
trendValue2 ? SCORE[1] += Math.round(trendValue2) : null;
scoreRender();
}
}
function scoreRender () {
document.getElementById("player-1-score").innerHTML = SCORE[0] ? SCORE[0] : "Loading...";
document.getElementById("player-2-score").innerHTML = SCORE[1] ? SCORE[1] : "Loading...";
//sets the high scoreRender
document.getElementById("high-score").innerHTML = (localStorage.getItem("highScore"));
}
function standardDeviation(values){
var avg = average(values);
var squareDiffs = values.map(function(value){
var diff = value - avg;
var sqrDiff = diff * diff;
return sqrDiff;
});
var avgSquareDiff = average(squareDiffs);
var stdDev = Math.sqrt(avgSquareDiff);
return stdDev;
}
function storeScoresForAverages() {
var tempArray = [];
tempArray.push(localStorage.getItem("cumulativescores"));
tempArray.push(SCORE[0]);
tempArray.push(SCORE[1]);
localStorage.setItem("cumulativescores", tempArray);
}
function timeSlice () {
p1data[sliceCount] = VALUE1;
p2data[sliceCount] = VALUE2;
sliceCount ++;
//Store 240 time slices (one per second) for both players
multiplierA = standardDeviation(p1data);
multiplierB = standardDeviation(p2data);
//this data is used for modeling the results
//creates a baseline to determine if disconnet has occured
//writes the list of time slices into localstorage
localStorage.setItem("player1slices", p1data);
localStorage.setItem("player2slices", p2data);
}
function totalParticipants () {//keeps a rolling total on how many times the test has been ran
var newTotal = parseInt(localStorage.getItem("totalparticipants")) + 1;
document.getElementById("participant-total").innerHTML = (newTotal);
localStorage.setItem("totalparticipants", newTotal);
}
function totalPoints () {//create a rolling total
var newTotal;
localStorage.setItem("averagepoints", rollingAverage);
localStorage.setItem("totalparticipants", newTotal);
}
function trendOfReadings () {//Takes in the last n readings and makes comparison to determine directional trending
var driftDebounce = 0; //tune based off of how much it wobbles
if (sliceCount > 5){//slice count is always ahead by one
//simple and needs refactoring
//TODO: take the last five averages and do a simple comparison for smoothing
if (p1data[sliceCount - 1] > p1data[sliceCount - 2] && p1data[sliceCount - 2] > p1data[sliceCount - 3]){
p1trend = "+";
}
else if (p1data[sliceCount - 1] < p1data[sliceCount - 2] && p1data[sliceCount - 2] < p1data[sliceCount - 3]){
p1trend = "-";
}
else {
p1trend = "0";
}
if (p2data[sliceCount - 1] > p2data[sliceCount - 2] && p2data[sliceCount - 2] > p2data[sliceCount - 3]){
p2trend = "+";
}
else if (p2data[sliceCount - 1] < p2data[sliceCount - 2] && p2data[sliceCount - 2] < p2data[sliceCount - 3]){
p2trend = "-";
}
else {
p2trend = "0";
}
}
}
///////////////////////////////////////////////////////////////////////////////
function render() {//keep all of the execution of the app within this function
//initialization();
onRefreshInitialization();
aggregateAverage();
storeScoresForAverages();
totalParticipants();
datesTimes();
pointGenerator();
dataLoader();
loop = setInterval(dataLoader, 1000);
slice = setInterval(timeSlice, 1000);
countdown = setInterval(myTimer,1000);
points = setInterval(pointGenerator, 633);
leader = setInterval(highScore,1000);
//TODO: store everything in local storage better for extraction at a later date
}
$(document).ready(function() {
render();
}); | };
localStorage.setItem("finalobjects", JSON.stringify(totalFinalObjects));
} | random_line_split |
app.js | var SECONDS = 59;
var MINUTES = 3;
var SCORE = [0,0];
var VALUE1;
var VALUE2;
var sliceCount = 0;
var p1data = [];
var p2data = [];
var p1trend = "0";
var p2trend = "0";
var multiplierA;
var multiplierB;
var makePointsFlag = true;
var P1NAME;
var P2NAME;
var P1EMAIL;
var P2EMAIL;
var COMBONAME = null;
function add(a, b) {
return a + b;
}
function aggregateAverage () {//calculates the average SCORE of all players
var average = 0; //holds a cumulative list of all of the player scores
var tempString = localStorage.getItem("cumulativescores");
var allScores = tempString.split(",") ? tempString.split(",") :"0,0";
average = allScores.reduce(function(a, b){a= parseFloat(a);b= parseFloat(b); return a+b;})
// for (var i = 0; i < allScores.length; i++){
average = Math.round(average/allScores.length);
document.getElementById("average").innerHTML = (average);
}
function average (data) {
var sum = data.reduce(function(sum, value){
return sum + value;
}, 0);
var avg = sum / data.length;
return avg;
}
function contactNameInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[0].split("=")[1];
nameTwo = chopped[1].split("=")[1];
var re = new RegExp("\s*([A-Za-z\d]+)\s*");
nameOne = re.exec(nameOne)[0];
nameTwo = re.exec(nameTwo)[0];
P1NAME = nameOne;
P2NAME = nameTwo;
coupleName = this.nameMash (nameOne, nameTwo);
COMBONAME = coupleName;
document.getElementById("player-1").innerHTML = (nameOne + "'s Score is:");
document.getElementById("player-2").innerHTML = (nameTwo + "'s Score is:");
document.getElementById("couple-name").innerHTML = ("Couple Name: " + coupleName);
}
function contactEmailInfo () { //grabs the contact details and transforms them into a combination name and stores everything to local storage
var coupleName, name, email;
var atRegex = /%40/g;
nameTemp = $('#form').serialize();
chopped = nameTemp.split("&");
nameOne = chopped[2].split("=")[1];
nameTwo = chopped[3].split("=")[1];
P1EMAIL = nameOne.replace(atRegex, '@');
P2EMAIL = nameTwo.replace(atRegex,'@');
document.getElementById("email-message").innerHTML = ("Thanks! We will be getting back to you sometime before the next burn!");
}
function datesTimes () {//grabs the time... no network or persistant time... Need a GPS module for the Raspberry PI
var cleanDate;
cleanDate = new Date().toLocaleString()
localStorage.setItem("timestamp", cleanDate);
}
function dataLoader(){//pulls the data from a local JSON file
$.getJSON("http://localhost:8000/Love-Staring-Machine/result.json", function(data) {
VALUE1 = data["player1"];
VALUE2 = data["player2"];
});
}
function finalMessage () {//Delivers the final message in the form of a modal overlay based on compadability of scores
//derive a compatability index based on distance of final scores
var p1Score = SCORE[0];
var p2Score = SCORE[1];
var higherScore = p1Score > p2Score ? p1Score : p2Score;
var distance = Math.abs(p1Score - p2Score);
var relDistance = distance / higherScore;
var message = [
"Hell No!!!",
"Y'all Fuck",
"True Love!"
];
var messenger;
if (relDistance > .1) {
messenger = 0;
}
else if (relDistance > .05) {
messenger = 1;
}
else {
messenger = 2;
}
//render the modal overlay
document.getElementById("timer-readout").innerHTML = (message[messenger]);
}
function finalData () {
//Stpres all he data as an object
var finalObject = {};
var totalFinalObjects = JSON.parse(localStorage.getItem("finalobjects"));
var participants = localStorage.getItem("totalparticipants");
var p1Name = P1NAME ? P1NAME : null;
p2Name = P2NAME ? P2NAME : null;
var coupleName = COMBONAME ? COMBONAME : null;
p1Email = P1EMAIL ? P1EMAIL : null;
p2Email = P2EMAIL ? P2EMAIL : null;
//grabs from localstorage the entire
totalFinalObjects[participants] = {
'p1': {
'n': p1Name,
'e': p1Email,
'p': SCORE[0],
's': p1data
},
'p2' : {
'n': p2Name,
'e': p1Email,
'p': SCORE[1],
's': p2data
},
'n': COMBONAME
};
localStorage.setItem("finalobjects", JSON.stringify(totalFinalObjects));
}
function highScore () {//checks if any of the present SCOREs are higher than the all time high SCORE
var tempSCORE = parseInt(localStorage.getItem("highScore"));
if (SCORE[0] > tempSCORE) {
localStorage.setItem("highScore", SCORE[0]);
}
if (SCORE[1] > tempSCORE) {
localStorage.setItem("highScore", SCORE[1]);
}
}
function initialization () {//initializes the persistant dashboard metrics for the first time
//add all of the local storage variables
//localStorage.setItem("highScore", 0);
//localStorage.setItem("totalparticipants", 0);
//localStorage.setItem("average", 0);
//localStorage.setItem("cumulativescores", 0)
localStorage.setItem("finalobjects", "{}");
}
function myTimer() {//Simple, non-accurate clock funtion
var secs = "" + SECONDS;
var min = MINUTES;
if ( MINUTES >= 0){
SECONDS --;
if (SECONDS < 9) {
secs = "0" + secs;
}
if (SECONDS === 0){
SECONDS = 59;
secs = "00";//TODO: fix this where the at 1:00 it displays :00
MINUTES --;
}
if (MINUTES > 0){
document.getElementById("timer-readout").innerHTML = (min + ":" + secs);
}
else {
document.getElementById("timer-readout").innerHTML = (":" + secs);
}
}
else {
makePointsFlag = false;
storeScoresForAverages();
finalMessage();
finalData();
clearTimeout(loop);
clearTimeout(slice);
clearTimeout(countdown);
clearTimeout(points);
clearTimeout(leader);
}
}
function nameMash (p1,p2) {//celebrity name generator splits on the vowels or if all else fails just smash them together
if (p1.length > 5 || p2.length > 5) {//TODO: Clean up this logic and make it more compact
//use the p1 and p2 arguments and loop through to grab the regex vowels
var re = /[aeiou]/gi;
var p1Array = p1.match(re);
var p2Array = p2.match(re);
//if there are more than 2 split on the second
if (p1Array.length >= 2 && p2Array.length >= 2) {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == p1Array.length){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == p2Array.length){
var player2 = p2.substring(match.index - 2, p2.length);
}
}
return player1 + player2;
}
else {
var i = 0
while ((match = re.exec(p1)) != null) {
i ++;
if (i == 1){
var player1 = p1.substring(0,match.index + 1);
}
}
var z = 0
while ((match = re.exec(p2)) != null) {
z ++;
if (z == 1){
var player2 = p2.substring(match.index - 1, p2.length);
}
}
return player1 + player2;
}
}
else {
if (p1.length > 2 && p2.length > 2){
return p1.substring(0,1).toUpperCase() + p1.substring(1,Math.floor(p1.length/2)).toLowerCase() + p2.substring(Math.floor(p2.length/2),p2.length).toLowerCase();
}
else {
return p1.substring(0,1).toUpperCase() + p1.substring(1, p1.length).toLowerCase() + p2.substring(2,p2.length).toLowerCase();
}
}
}
function onRefreshInitialization () {//initializes the persistant dashboard metrics
document.getElementById('p1').focus();
$('#form')[0].reset();
}
function | () {
el = document.getElementById("overlay");
el.style.visibility = (el.style.visibility == "visible") ? "hidden" : "visible";
}
function pointGenerator () {//this function computes the points
//TODO: Make a funtion to check directional trends and integrate the result into this function
if (makePointsFlag) {
var trendValue1,
trendValue2;
trendOfReadings();
if (p1trend === '0'){
trendValue1 = 9;
}
else if (p1trend === '+'){
trendValue1 = 30;
}
else if (p1trend === '-'){
trendValue1 = 5;
}
else {
trendValue1 = 1;
}
if (p2trend === '0'){
trendValue2 = 9;
}
else if (p2trend === '+'){
trendValue2 = 30;
}
else if (p2trend === '-'){
trendValue2 = 5;
}
else {
trendValue2 = 1;
}
trendValue1 ? SCORE[0] += Math.round(trendValue1) : null;
trendValue2 ? SCORE[1] += Math.round(trendValue2) : null;
scoreRender();
}
}
function scoreRender () {
document.getElementById("player-1-score").innerHTML = SCORE[0] ? SCORE[0] : "Loading...";
document.getElementById("player-2-score").innerHTML = SCORE[1] ? SCORE[1] : "Loading...";
//sets the high scoreRender
document.getElementById("high-score").innerHTML = (localStorage.getItem("highScore"));
}
function standardDeviation(values){
var avg = average(values);
var squareDiffs = values.map(function(value){
var diff = value - avg;
var sqrDiff = diff * diff;
return sqrDiff;
});
var avgSquareDiff = average(squareDiffs);
var stdDev = Math.sqrt(avgSquareDiff);
return stdDev;
}
function storeScoresForAverages() {
var tempArray = [];
tempArray.push(localStorage.getItem("cumulativescores"));
tempArray.push(SCORE[0]);
tempArray.push(SCORE[1]);
localStorage.setItem("cumulativescores", tempArray);
}
function timeSlice () {
p1data[sliceCount] = VALUE1;
p2data[sliceCount] = VALUE2;
sliceCount ++;
//Store 240 time slices (one per second) for both players
multiplierA = standardDeviation(p1data);
multiplierB = standardDeviation(p2data);
//this data is used for modeling the results
//creates a baseline to determine if disconnet has occured
//writes the list of time slices into localstorage
localStorage.setItem("player1slices", p1data);
localStorage.setItem("player2slices", p2data);
}
function totalParticipants () {//keeps a rolling total on how many times the test has been ran
var newTotal = parseInt(localStorage.getItem("totalparticipants")) + 1;
document.getElementById("participant-total").innerHTML = (newTotal);
localStorage.setItem("totalparticipants", newTotal);
}
function totalPoints () {//create a rolling total
var newTotal;
localStorage.setItem("averagepoints", rollingAverage);
localStorage.setItem("totalparticipants", newTotal);
}
function trendOfReadings () {//Takes in the last n readings and makes comparison to determine directional trending
var driftDebounce = 0; //tune based off of how much it wobbles
if (sliceCount > 5){//slice count is always ahead by one
//simple and needs refactoring
//TODO: take the last five averages and do a simple comparison for smoothing
if (p1data[sliceCount - 1] > p1data[sliceCount - 2] && p1data[sliceCount - 2] > p1data[sliceCount - 3]){
p1trend = "+";
}
else if (p1data[sliceCount - 1] < p1data[sliceCount - 2] && p1data[sliceCount - 2] < p1data[sliceCount - 3]){
p1trend = "-";
}
else {
p1trend = "0";
}
if (p2data[sliceCount - 1] > p2data[sliceCount - 2] && p2data[sliceCount - 2] > p2data[sliceCount - 3]){
p2trend = "+";
}
else if (p2data[sliceCount - 1] < p2data[sliceCount - 2] && p2data[sliceCount - 2] < p2data[sliceCount - 3]){
p2trend = "-";
}
else {
p2trend = "0";
}
}
}
///////////////////////////////////////////////////////////////////////////////
function render() {//keep all of the execution of the app within this function
//initialization();
onRefreshInitialization();
aggregateAverage();
storeScoresForAverages();
totalParticipants();
datesTimes();
pointGenerator();
dataLoader();
loop = setInterval(dataLoader, 1000);
slice = setInterval(timeSlice, 1000);
countdown = setInterval(myTimer,1000);
points = setInterval(pointGenerator, 633);
leader = setInterval(highScore,1000);
//TODO: store everything in local storage better for extraction at a later date
}
$(document).ready(function() {
render();
});
| overlay | identifier_name |
XYRepresentation.py | # -*- coding: utf-8 -*-
"""
Created on Mon Nov 5 15:06:34 2018
@author: lawrence
"""
import numpy as np
import random as rd
import matplotlib.pyplot as plt
import matplotlib.path as mplpath
import matplotlib.patches as patches
from matplotlib.collections import PatchCollection
from matplotlib import collections as mc
import shapely.geometry as shapely
from XYaStarInterpreter import interpret
from Astar import find_path
import PRM
import copy
#from descartes.patch import PolygonPatch
class Graph():
def __init__(self,mp):
self.edges=[]
self.vertices=[]
self.mp=mp
def addVertex(self,point):
self.vertices.append(point)
def copy(self):
return copy.deepcopy(self)
def ngd(self, a):
neighbors=[]
distances=[]
for pt in self.vertices:
if pt[0]!=a[0] or pt[1]!=a[1]:
dist=np.sqrt(np.square(pt[0]-a[0])+np.square(pt[1]-a[1]))
if not distances:
distances.append(dist)
neighbors.append(pt)
else:
counter=0
while counter<len(distances) and dist>distances[counter]:
counter+=1
distances.insert(counter,dist)
neighbors.insert(counter,pt)
return neighbors
def can_connect(self,a,q):
return not self.mp.collide(a,q)
def connect(self,a,q):
|
def sample(self):
samp=self.mp.sample()
self.addVertex(samp)
return samp
class Map():
points=[] #PRM points
edges=[] #PRM edges
lines=[]
buff_lines=[]
obstacles=[]
buff=0
vis_graph=None
visibility=False
def __init__(self,lx=0,hx=0,ly=0,hy=0):
self.lx=lx
self.hx=hx
self.ly=ly
self.hy=hy
"""Returns a pt that is collision free"""
def sample(self):
x=None
y=None
while x==None or self.collide((x,y)):
xwidth=self.hx-self.lx
ywidth=self.hy-self.ly
x=xwidth*rd.random()+self.lx
y=ywidth*rd.random()+self.ly
return (x,y)
def collide(self,point1,point2=None):
if point2==None:
point=shapely.Point(point1[0],point1[1])
for obs in self.obstacles:
if point.within(obs[1]) and not point.touches(obs[1]):
return True
else:
path =shapely.LineString([point1,point2])
if self.buff>0:
path=path.buffer(self.buff)
for obs in self.obstacles:
if obs[1].intersects(path) and not path.touches(obs[1]):
return True
return False
def clear(self):
self.lines=[]
self.buff_lines=[]
self.obstacles=[]
"""vertices are of format [(a,b),(c,d)]"""
def add_Poly(self,vertices=[]):
if self.visibility:
init_poly=shapely.Polygon(vertices)
buff_poly=init_poly.buffer(self.buff,cap_style=2,join_style =2)
new_vert=list(buff_poly.exterior.coords)
self.obstacles.append((mplpath.Path(vertices),init_poly,new_vert,mplpath.Path(new_vert),buff_poly))
else:
self.obstacles.append((mplpath.Path(vertices),shapely.Polygon(vertices),vertices))
def display(self):
"""start grid"""
fig, ax = plt.subplots()
"""Add Obstacles"""
if self.buff>0 and self.visibility:
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[3], facecolor='orange',alpha=0.5))
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[0], facecolor='purple'))
"""Set boundaries"""
ax.set_xlim([self.lx,self.hx])
ax.set_ylim([self.ly,self.hy])
"""add edges"""
c=[]
for edge in self.edges:
c.append((1,0,0,0.5))
# vis_lines.append([(edge[0][0],edge[0][1]),(edge[1][0],edge[1][1])])
ax.add_collection(mc.LineCollection(self.edges,colors=c,linewidths = 1.0))
"""Set size of plot"""
fig_size = plt.rcParams["figure.figsize"]
fig_size[0] = 12
fig_size[1] = 9
plt.rcParams["figure.figsize"] = fig_size
"""draw points"""
for pt in self.points:
plt.plot(pt[0],pt[1],marker='o', markersize=5, color="black")
"""draw lines to the plot"""
if not self.visibility and self.buff>0:
path=shapely.LineString(self.buff_lines)
poly=path.buffer(self.buff)
poly_points=list(poly.exterior.coords)
ax.add_patch(patches.PathPatch(mplpath.Path(poly_points), facecolor='yellow',alpha=0.5))
lc = mc.LineCollection(self.lines,linewidths = 2.5)
ax.add_collection(lc)
"""Set up Axis """
ax.minorticks_on()
ax.xaxis.set_major_locator(plt.MultipleLocator(1))
ax.yaxis.set_major_locator(plt.MultipleLocator(1))
ax.xaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.yaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.grid(which='major', linestyle='-', linewidth='0.5', color='black')
ax.grid(which='minor', linestyle='-', linewidth='0.2', color='black')
plt.show()
def average(intlist):
total=0
for val in intlist:
total+=val
if len(intlist)>0:
return total/float(len(intlist))
else:
return -1
def sd(intlist):
avg= average(intlist)
variance=0
for val in intlist:
variance+=np.square(val-avg)
if len(intlist)>0:
variance= variance/float(len(intlist))
return np.sqrt(variance)
else:
return -1
def stats():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-8,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(9,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-7,-5) ,(-7,-4) ])
mp.add_Poly([(7,6), (0,4) ,(-8,7) ,(0,6), (4,8)])
mp.add_Poly([(-9,-9), (8,-5) ,(9,-8)])
starts=[]
goals=[]
for i in range(25):
starts.append(mp.sample())
goals.append(mp.sample())
prmcc_maps=[]
prmk_maps=[]
prmstar_maps=[]
conn=[]
for i in range(10):
g1=Graph(mp)
g2=Graph(mp)
g3=Graph(mp)
conn.append(PRM.prm_cc(g1,25*i+25))
PRM.prm_k(g2,25*i+25,5)
PRM.prm_star(g3,25*i+25)
prmcc_maps.append(g1)
prmk_maps.append(g2)
prmstar_maps.append(g3)
print i
cc_stats=[]
k_stats=[]
star_stats=[]
c=0
for prm_map in prmcc_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_cc(temp,0,starts[i],goals[i],copy.deepcopy(conn[c]))
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
cc_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmk_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_k(temp,0,5,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
k_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmstar_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_star(temp,0,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
star_stats.append(average(raw_dat))
c+=1
print c
print cc_stats
print k_stats
print star_stats
def test():
# mp = Map(-5,5,-5,5)
mp = Map(-7, 8, -7, 8)
start=(-6,7)
goal=(6,-6)
"""Add obstacles"""
# mp.add_Poly([(-1,-1),(-1,1),(1,1),(1,-1)])
mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
# mp.add_Poly([(-2,0),(-2,-1),(2,-1),(2,0)])
graph=Graph(mp)
"""Create graph for PRM"""
# PRM.prm_cc(graph,100,start,goal)
# PRM.prm_k(graph,100,3,start,goal)
PRM.prm_star(graph,5)
"""Add graph and tree to display and print out what it looks"""
mp.points=graph.vertices
mp.edges=graph.edges
mp.display()
"""Run A-Star to find a path"""
# ag=interpret(graph.vertices,graph.edges)
# end = ag.nodes.pop()#end is always last of list
# beg = ag.nodes.pop()#beginning is second to last
# find_path(beg,end)
# iterator = end
#
# """Get line segments to draw a-star path"""
# lines=[]
# while iterator.parent != None:
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
# iterator=iterator.parent
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
#
# """Input all of data to map display and display it"""
# mp.lines=lines
# mp.points=graph.vertices
# mp.edges=graph.edges
# mp.display()
def main():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
g1=Graph(mp)
PRM.prm_star(g1,20)
start=(-6,7)
goal=(6,-6)
PRM.prm_star(g1,0,start,goal)
if __name__ == '__main__':
stats()
| if self.can_connect(a,q):
if (a,q) not in self.edges and (q,a) not in self.edges:
self.edges.append((a,q)) | identifier_body |
XYRepresentation.py | # -*- coding: utf-8 -*-
"""
Created on Mon Nov 5 15:06:34 2018
@author: lawrence
"""
import numpy as np
import random as rd
import matplotlib.pyplot as plt
import matplotlib.path as mplpath
import matplotlib.patches as patches
from matplotlib.collections import PatchCollection
from matplotlib import collections as mc
import shapely.geometry as shapely
from XYaStarInterpreter import interpret
from Astar import find_path
import PRM
import copy
#from descartes.patch import PolygonPatch
class Graph():
def __init__(self,mp):
self.edges=[]
self.vertices=[]
self.mp=mp
def addVertex(self,point):
self.vertices.append(point)
def copy(self):
return copy.deepcopy(self)
def ngd(self, a):
neighbors=[]
distances=[]
for pt in self.vertices:
if pt[0]!=a[0] or pt[1]!=a[1]:
dist=np.sqrt(np.square(pt[0]-a[0])+np.square(pt[1]-a[1]))
if not distances:
distances.append(dist)
neighbors.append(pt)
else:
counter=0
while counter<len(distances) and dist>distances[counter]:
counter+=1
distances.insert(counter,dist)
neighbors.insert(counter,pt)
return neighbors
def can_connect(self,a,q):
return not self.mp.collide(a,q)
def connect(self,a,q):
if self.can_connect(a,q):
if (a,q) not in self.edges and (q,a) not in self.edges:
self.edges.append((a,q))
def sample(self):
samp=self.mp.sample()
self.addVertex(samp)
return samp
class Map():
points=[] #PRM points
edges=[] #PRM edges
lines=[]
buff_lines=[]
obstacles=[]
buff=0
vis_graph=None
visibility=False
def __init__(self,lx=0,hx=0,ly=0,hy=0):
self.lx=lx
self.hx=hx
self.ly=ly
self.hy=hy
"""Returns a pt that is collision free"""
def sample(self):
x=None
y=None
while x==None or self.collide((x,y)):
xwidth=self.hx-self.lx
ywidth=self.hy-self.ly
x=xwidth*rd.random()+self.lx
y=ywidth*rd.random()+self.ly
return (x,y)
def collide(self,point1,point2=None):
if point2==None:
point=shapely.Point(point1[0],point1[1])
for obs in self.obstacles:
if point.within(obs[1]) and not point.touches(obs[1]):
return True
else:
path =shapely.LineString([point1,point2])
if self.buff>0:
path=path.buffer(self.buff)
for obs in self.obstacles:
|
return False
def clear(self):
self.lines=[]
self.buff_lines=[]
self.obstacles=[]
"""vertices are of format [(a,b),(c,d)]"""
def add_Poly(self,vertices=[]):
if self.visibility:
init_poly=shapely.Polygon(vertices)
buff_poly=init_poly.buffer(self.buff,cap_style=2,join_style =2)
new_vert=list(buff_poly.exterior.coords)
self.obstacles.append((mplpath.Path(vertices),init_poly,new_vert,mplpath.Path(new_vert),buff_poly))
else:
self.obstacles.append((mplpath.Path(vertices),shapely.Polygon(vertices),vertices))
def display(self):
"""start grid"""
fig, ax = plt.subplots()
"""Add Obstacles"""
if self.buff>0 and self.visibility:
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[3], facecolor='orange',alpha=0.5))
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[0], facecolor='purple'))
"""Set boundaries"""
ax.set_xlim([self.lx,self.hx])
ax.set_ylim([self.ly,self.hy])
"""add edges"""
c=[]
for edge in self.edges:
c.append((1,0,0,0.5))
# vis_lines.append([(edge[0][0],edge[0][1]),(edge[1][0],edge[1][1])])
ax.add_collection(mc.LineCollection(self.edges,colors=c,linewidths = 1.0))
"""Set size of plot"""
fig_size = plt.rcParams["figure.figsize"]
fig_size[0] = 12
fig_size[1] = 9
plt.rcParams["figure.figsize"] = fig_size
"""draw points"""
for pt in self.points:
plt.plot(pt[0],pt[1],marker='o', markersize=5, color="black")
"""draw lines to the plot"""
if not self.visibility and self.buff>0:
path=shapely.LineString(self.buff_lines)
poly=path.buffer(self.buff)
poly_points=list(poly.exterior.coords)
ax.add_patch(patches.PathPatch(mplpath.Path(poly_points), facecolor='yellow',alpha=0.5))
lc = mc.LineCollection(self.lines,linewidths = 2.5)
ax.add_collection(lc)
"""Set up Axis """
ax.minorticks_on()
ax.xaxis.set_major_locator(plt.MultipleLocator(1))
ax.yaxis.set_major_locator(plt.MultipleLocator(1))
ax.xaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.yaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.grid(which='major', linestyle='-', linewidth='0.5', color='black')
ax.grid(which='minor', linestyle='-', linewidth='0.2', color='black')
plt.show()
def average(intlist):
total=0
for val in intlist:
total+=val
if len(intlist)>0:
return total/float(len(intlist))
else:
return -1
def sd(intlist):
avg= average(intlist)
variance=0
for val in intlist:
variance+=np.square(val-avg)
if len(intlist)>0:
variance= variance/float(len(intlist))
return np.sqrt(variance)
else:
return -1
def stats():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-8,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(9,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-7,-5) ,(-7,-4) ])
mp.add_Poly([(7,6), (0,4) ,(-8,7) ,(0,6), (4,8)])
mp.add_Poly([(-9,-9), (8,-5) ,(9,-8)])
starts=[]
goals=[]
for i in range(25):
starts.append(mp.sample())
goals.append(mp.sample())
prmcc_maps=[]
prmk_maps=[]
prmstar_maps=[]
conn=[]
for i in range(10):
g1=Graph(mp)
g2=Graph(mp)
g3=Graph(mp)
conn.append(PRM.prm_cc(g1,25*i+25))
PRM.prm_k(g2,25*i+25,5)
PRM.prm_star(g3,25*i+25)
prmcc_maps.append(g1)
prmk_maps.append(g2)
prmstar_maps.append(g3)
print i
cc_stats=[]
k_stats=[]
star_stats=[]
c=0
for prm_map in prmcc_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_cc(temp,0,starts[i],goals[i],copy.deepcopy(conn[c]))
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
cc_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmk_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_k(temp,0,5,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
k_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmstar_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_star(temp,0,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
star_stats.append(average(raw_dat))
c+=1
print c
print cc_stats
print k_stats
print star_stats
def test():
# mp = Map(-5,5,-5,5)
mp = Map(-7, 8, -7, 8)
start=(-6,7)
goal=(6,-6)
"""Add obstacles"""
# mp.add_Poly([(-1,-1),(-1,1),(1,1),(1,-1)])
mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
# mp.add_Poly([(-2,0),(-2,-1),(2,-1),(2,0)])
graph=Graph(mp)
"""Create graph for PRM"""
# PRM.prm_cc(graph,100,start,goal)
# PRM.prm_k(graph,100,3,start,goal)
PRM.prm_star(graph,5)
"""Add graph and tree to display and print out what it looks"""
mp.points=graph.vertices
mp.edges=graph.edges
mp.display()
"""Run A-Star to find a path"""
# ag=interpret(graph.vertices,graph.edges)
# end = ag.nodes.pop()#end is always last of list
# beg = ag.nodes.pop()#beginning is second to last
# find_path(beg,end)
# iterator = end
#
# """Get line segments to draw a-star path"""
# lines=[]
# while iterator.parent != None:
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
# iterator=iterator.parent
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
#
# """Input all of data to map display and display it"""
# mp.lines=lines
# mp.points=graph.vertices
# mp.edges=graph.edges
# mp.display()
def main():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
g1=Graph(mp)
PRM.prm_star(g1,20)
start=(-6,7)
goal=(6,-6)
PRM.prm_star(g1,0,start,goal)
if __name__ == '__main__':
stats()
| if obs[1].intersects(path) and not path.touches(obs[1]):
return True | conditional_block |
XYRepresentation.py | # -*- coding: utf-8 -*-
"""
Created on Mon Nov 5 15:06:34 2018
@author: lawrence
"""
import numpy as np
import random as rd
import matplotlib.pyplot as plt
import matplotlib.path as mplpath
import matplotlib.patches as patches
from matplotlib.collections import PatchCollection
from matplotlib import collections as mc
import shapely.geometry as shapely
from XYaStarInterpreter import interpret
from Astar import find_path
import PRM
import copy
#from descartes.patch import PolygonPatch
class Graph():
def __init__(self,mp):
self.edges=[]
self.vertices=[]
self.mp=mp
def addVertex(self,point):
self.vertices.append(point)
def copy(self):
return copy.deepcopy(self)
def ngd(self, a):
neighbors=[]
distances=[]
for pt in self.vertices:
if pt[0]!=a[0] or pt[1]!=a[1]:
dist=np.sqrt(np.square(pt[0]-a[0])+np.square(pt[1]-a[1]))
if not distances:
distances.append(dist)
neighbors.append(pt)
else:
counter=0
while counter<len(distances) and dist>distances[counter]:
counter+=1
distances.insert(counter,dist)
neighbors.insert(counter,pt)
return neighbors
def can_connect(self,a,q):
return not self.mp.collide(a,q)
def | (self,a,q):
if self.can_connect(a,q):
if (a,q) not in self.edges and (q,a) not in self.edges:
self.edges.append((a,q))
def sample(self):
samp=self.mp.sample()
self.addVertex(samp)
return samp
class Map():
points=[] #PRM points
edges=[] #PRM edges
lines=[]
buff_lines=[]
obstacles=[]
buff=0
vis_graph=None
visibility=False
def __init__(self,lx=0,hx=0,ly=0,hy=0):
self.lx=lx
self.hx=hx
self.ly=ly
self.hy=hy
"""Returns a pt that is collision free"""
def sample(self):
x=None
y=None
while x==None or self.collide((x,y)):
xwidth=self.hx-self.lx
ywidth=self.hy-self.ly
x=xwidth*rd.random()+self.lx
y=ywidth*rd.random()+self.ly
return (x,y)
def collide(self,point1,point2=None):
if point2==None:
point=shapely.Point(point1[0],point1[1])
for obs in self.obstacles:
if point.within(obs[1]) and not point.touches(obs[1]):
return True
else:
path =shapely.LineString([point1,point2])
if self.buff>0:
path=path.buffer(self.buff)
for obs in self.obstacles:
if obs[1].intersects(path) and not path.touches(obs[1]):
return True
return False
def clear(self):
self.lines=[]
self.buff_lines=[]
self.obstacles=[]
"""vertices are of format [(a,b),(c,d)]"""
def add_Poly(self,vertices=[]):
if self.visibility:
init_poly=shapely.Polygon(vertices)
buff_poly=init_poly.buffer(self.buff,cap_style=2,join_style =2)
new_vert=list(buff_poly.exterior.coords)
self.obstacles.append((mplpath.Path(vertices),init_poly,new_vert,mplpath.Path(new_vert),buff_poly))
else:
self.obstacles.append((mplpath.Path(vertices),shapely.Polygon(vertices),vertices))
def display(self):
"""start grid"""
fig, ax = plt.subplots()
"""Add Obstacles"""
if self.buff>0 and self.visibility:
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[3], facecolor='orange',alpha=0.5))
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[0], facecolor='purple'))
"""Set boundaries"""
ax.set_xlim([self.lx,self.hx])
ax.set_ylim([self.ly,self.hy])
"""add edges"""
c=[]
for edge in self.edges:
c.append((1,0,0,0.5))
# vis_lines.append([(edge[0][0],edge[0][1]),(edge[1][0],edge[1][1])])
ax.add_collection(mc.LineCollection(self.edges,colors=c,linewidths = 1.0))
"""Set size of plot"""
fig_size = plt.rcParams["figure.figsize"]
fig_size[0] = 12
fig_size[1] = 9
plt.rcParams["figure.figsize"] = fig_size
"""draw points"""
for pt in self.points:
plt.plot(pt[0],pt[1],marker='o', markersize=5, color="black")
"""draw lines to the plot"""
if not self.visibility and self.buff>0:
path=shapely.LineString(self.buff_lines)
poly=path.buffer(self.buff)
poly_points=list(poly.exterior.coords)
ax.add_patch(patches.PathPatch(mplpath.Path(poly_points), facecolor='yellow',alpha=0.5))
lc = mc.LineCollection(self.lines,linewidths = 2.5)
ax.add_collection(lc)
"""Set up Axis """
ax.minorticks_on()
ax.xaxis.set_major_locator(plt.MultipleLocator(1))
ax.yaxis.set_major_locator(plt.MultipleLocator(1))
ax.xaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.yaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.grid(which='major', linestyle='-', linewidth='0.5', color='black')
ax.grid(which='minor', linestyle='-', linewidth='0.2', color='black')
plt.show()
def average(intlist):
total=0
for val in intlist:
total+=val
if len(intlist)>0:
return total/float(len(intlist))
else:
return -1
def sd(intlist):
avg= average(intlist)
variance=0
for val in intlist:
variance+=np.square(val-avg)
if len(intlist)>0:
variance= variance/float(len(intlist))
return np.sqrt(variance)
else:
return -1
def stats():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-8,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(9,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-7,-5) ,(-7,-4) ])
mp.add_Poly([(7,6), (0,4) ,(-8,7) ,(0,6), (4,8)])
mp.add_Poly([(-9,-9), (8,-5) ,(9,-8)])
starts=[]
goals=[]
for i in range(25):
starts.append(mp.sample())
goals.append(mp.sample())
prmcc_maps=[]
prmk_maps=[]
prmstar_maps=[]
conn=[]
for i in range(10):
g1=Graph(mp)
g2=Graph(mp)
g3=Graph(mp)
conn.append(PRM.prm_cc(g1,25*i+25))
PRM.prm_k(g2,25*i+25,5)
PRM.prm_star(g3,25*i+25)
prmcc_maps.append(g1)
prmk_maps.append(g2)
prmstar_maps.append(g3)
print i
cc_stats=[]
k_stats=[]
star_stats=[]
c=0
for prm_map in prmcc_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_cc(temp,0,starts[i],goals[i],copy.deepcopy(conn[c]))
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
cc_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmk_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_k(temp,0,5,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
k_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmstar_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_star(temp,0,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
star_stats.append(average(raw_dat))
c+=1
print c
print cc_stats
print k_stats
print star_stats
def test():
# mp = Map(-5,5,-5,5)
mp = Map(-7, 8, -7, 8)
start=(-6,7)
goal=(6,-6)
"""Add obstacles"""
# mp.add_Poly([(-1,-1),(-1,1),(1,1),(1,-1)])
mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
# mp.add_Poly([(-2,0),(-2,-1),(2,-1),(2,0)])
graph=Graph(mp)
"""Create graph for PRM"""
# PRM.prm_cc(graph,100,start,goal)
# PRM.prm_k(graph,100,3,start,goal)
PRM.prm_star(graph,5)
"""Add graph and tree to display and print out what it looks"""
mp.points=graph.vertices
mp.edges=graph.edges
mp.display()
"""Run A-Star to find a path"""
# ag=interpret(graph.vertices,graph.edges)
# end = ag.nodes.pop()#end is always last of list
# beg = ag.nodes.pop()#beginning is second to last
# find_path(beg,end)
# iterator = end
#
# """Get line segments to draw a-star path"""
# lines=[]
# while iterator.parent != None:
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
# iterator=iterator.parent
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
#
# """Input all of data to map display and display it"""
# mp.lines=lines
# mp.points=graph.vertices
# mp.edges=graph.edges
# mp.display()
def main():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
g1=Graph(mp)
PRM.prm_star(g1,20)
start=(-6,7)
goal=(6,-6)
PRM.prm_star(g1,0,start,goal)
if __name__ == '__main__':
stats()
| connect | identifier_name |
XYRepresentation.py | # -*- coding: utf-8 -*-
"""
Created on Mon Nov 5 15:06:34 2018
@author: lawrence
"""
import numpy as np
import random as rd
import matplotlib.pyplot as plt
import matplotlib.path as mplpath
import matplotlib.patches as patches
from matplotlib.collections import PatchCollection
from matplotlib import collections as mc
import shapely.geometry as shapely
from XYaStarInterpreter import interpret
from Astar import find_path
import PRM
import copy
#from descartes.patch import PolygonPatch
class Graph():
def __init__(self,mp):
self.edges=[]
self.vertices=[]
self.mp=mp
def addVertex(self,point):
self.vertices.append(point)
def copy(self):
return copy.deepcopy(self)
def ngd(self, a):
neighbors=[]
distances=[]
for pt in self.vertices:
if pt[0]!=a[0] or pt[1]!=a[1]:
dist=np.sqrt(np.square(pt[0]-a[0])+np.square(pt[1]-a[1]))
if not distances:
distances.append(dist)
neighbors.append(pt)
else:
counter=0
while counter<len(distances) and dist>distances[counter]:
counter+=1
distances.insert(counter,dist)
neighbors.insert(counter,pt)
return neighbors
def can_connect(self,a,q):
return not self.mp.collide(a,q)
def connect(self,a,q):
if self.can_connect(a,q):
if (a,q) not in self.edges and (q,a) not in self.edges:
self.edges.append((a,q))
def sample(self):
samp=self.mp.sample()
self.addVertex(samp)
return samp
class Map():
points=[] #PRM points
edges=[] #PRM edges
lines=[]
buff_lines=[]
obstacles=[]
buff=0
vis_graph=None
visibility=False
def __init__(self,lx=0,hx=0,ly=0,hy=0):
self.lx=lx
self.hx=hx
self.ly=ly
self.hy=hy
"""Returns a pt that is collision free"""
def sample(self):
x=None
y=None
while x==None or self.collide((x,y)):
xwidth=self.hx-self.lx
ywidth=self.hy-self.ly
x=xwidth*rd.random()+self.lx
y=ywidth*rd.random()+self.ly
return (x,y)
def collide(self,point1,point2=None):
if point2==None:
point=shapely.Point(point1[0],point1[1])
for obs in self.obstacles:
if point.within(obs[1]) and not point.touches(obs[1]):
return True
else:
path =shapely.LineString([point1,point2])
if self.buff>0:
path=path.buffer(self.buff)
for obs in self.obstacles:
if obs[1].intersects(path) and not path.touches(obs[1]):
return True
return False
def clear(self):
self.lines=[]
self.buff_lines=[]
self.obstacles=[]
"""vertices are of format [(a,b),(c,d)]"""
def add_Poly(self,vertices=[]):
if self.visibility:
init_poly=shapely.Polygon(vertices)
buff_poly=init_poly.buffer(self.buff,cap_style=2,join_style =2)
new_vert=list(buff_poly.exterior.coords)
self.obstacles.append((mplpath.Path(vertices),init_poly,new_vert,mplpath.Path(new_vert),buff_poly))
else:
self.obstacles.append((mplpath.Path(vertices),shapely.Polygon(vertices),vertices))
def display(self):
"""start grid"""
fig, ax = plt.subplots()
"""Add Obstacles"""
if self.buff>0 and self.visibility:
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[3], facecolor='orange',alpha=0.5))
for poly in self.obstacles:
ax.add_patch(patches.PathPatch(poly[0], facecolor='purple'))
"""Set boundaries"""
ax.set_xlim([self.lx,self.hx])
ax.set_ylim([self.ly,self.hy])
"""add edges"""
c=[]
for edge in self.edges:
c.append((1,0,0,0.5))
# vis_lines.append([(edge[0][0],edge[0][1]),(edge[1][0],edge[1][1])])
ax.add_collection(mc.LineCollection(self.edges,colors=c,linewidths = 1.0))
"""Set size of plot"""
fig_size = plt.rcParams["figure.figsize"]
fig_size[0] = 12
fig_size[1] = 9
plt.rcParams["figure.figsize"] = fig_size
"""draw points"""
for pt in self.points:
plt.plot(pt[0],pt[1],marker='o', markersize=5, color="black")
"""draw lines to the plot"""
if not self.visibility and self.buff>0:
path=shapely.LineString(self.buff_lines)
poly=path.buffer(self.buff)
poly_points=list(poly.exterior.coords)
ax.add_patch(patches.PathPatch(mplpath.Path(poly_points), facecolor='yellow',alpha=0.5))
lc = mc.LineCollection(self.lines,linewidths = 2.5)
ax.add_collection(lc)
"""Set up Axis """
ax.minorticks_on()
ax.xaxis.set_major_locator(plt.MultipleLocator(1))
ax.yaxis.set_major_locator(plt.MultipleLocator(1))
ax.xaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.yaxis.set_minor_locator(plt.MultipleLocator(0.25))
ax.grid(which='major', linestyle='-', linewidth='0.5', color='black')
ax.grid(which='minor', linestyle='-', linewidth='0.2', color='black')
plt.show()
def average(intlist):
total=0
for val in intlist:
total+=val
if len(intlist)>0:
return total/float(len(intlist))
else:
return -1
def sd(intlist):
avg= average(intlist)
variance=0
for val in intlist:
variance+=np.square(val-avg)
if len(intlist)>0:
variance= variance/float(len(intlist))
return np.sqrt(variance)
else:
return -1
def stats():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-8,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(9,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-7,-5) ,(-7,-4) ])
mp.add_Poly([(7,6), (0,4) ,(-8,7) ,(0,6), (4,8)])
mp.add_Poly([(-9,-9), (8,-5) ,(9,-8)])
starts=[]
goals=[]
for i in range(25):
starts.append(mp.sample())
goals.append(mp.sample())
prmcc_maps=[]
prmk_maps=[]
prmstar_maps=[]
conn=[]
for i in range(10):
g1=Graph(mp)
g2=Graph(mp)
g3=Graph(mp)
conn.append(PRM.prm_cc(g1,25*i+25))
PRM.prm_k(g2,25*i+25,5)
PRM.prm_star(g3,25*i+25)
prmcc_maps.append(g1)
prmk_maps.append(g2)
prmstar_maps.append(g3)
print i
cc_stats=[]
k_stats=[]
star_stats=[]
c=0
for prm_map in prmcc_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_cc(temp,0,starts[i],goals[i],copy.deepcopy(conn[c]))
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
cc_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmk_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_k(temp,0,5,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
k_stats.append(average(raw_dat))
c+=1
print c
c=0
for prm_map in prmstar_maps:
raw_dat=[]
for i in range(len(starts)):
temp=prm_map.copy()
PRM.prm_star(temp,0,starts[i],goals[i])
ag=interpret(temp.vertices,temp.edges)
end = ag.nodes.pop()
beg = ag.nodes.pop()
if find_path(beg,end):
raw_dat.append(end.g)
star_stats.append(average(raw_dat))
c+=1
print c
print cc_stats
print k_stats
print star_stats
def test():
# mp = Map(-5,5,-5,5)
mp = Map(-7, 8, -7, 8)
start=(-6,7)
goal=(6,-6)
"""Add obstacles"""
# mp.add_Poly([(-1,-1),(-1,1),(1,1),(1,-1)]) | graph=Graph(mp)
"""Create graph for PRM"""
# PRM.prm_cc(graph,100,start,goal)
# PRM.prm_k(graph,100,3,start,goal)
PRM.prm_star(graph,5)
"""Add graph and tree to display and print out what it looks"""
mp.points=graph.vertices
mp.edges=graph.edges
mp.display()
"""Run A-Star to find a path"""
# ag=interpret(graph.vertices,graph.edges)
# end = ag.nodes.pop()#end is always last of list
# beg = ag.nodes.pop()#beginning is second to last
# find_path(beg,end)
# iterator = end
#
# """Get line segments to draw a-star path"""
# lines=[]
# while iterator.parent != None:
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
# iterator=iterator.parent
# lines.append([(iterator.x,iterator.y),(iterator.parent.x,iterator.parent.y)])
#
# """Input all of data to map display and display it"""
# mp.lines=lines
# mp.points=graph.vertices
# mp.edges=graph.edges
# mp.display()
def main():
mp = Map(-10,10,-10,10)
mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
g1=Graph(mp)
PRM.prm_star(g1,20)
start=(-6,7)
goal=(6,-6)
PRM.prm_star(g1,0,start,goal)
if __name__ == '__main__':
stats() | mp.add_Poly([(-1,-1), (-6,-2), (-5,2), (-3,2), (-4,0)])
mp.add_Poly([(6,5), (4,1), (5,-2), (2,-4), (1,2)])
mp.add_Poly([(0,-3) ,(0,-4) ,(1,-5) ,(-5,-5) ,(-5,-4) ])
mp.add_Poly([(6,6), (0,4) ,(-5,6) ,(0,6), (4,7)])
# mp.add_Poly([(-2,0),(-2,-1),(2,-1),(2,0)]) | random_line_split |
loadout-builder-reducer.ts | import {
defaultLoadoutParameters,
LoadoutParameters,
UpgradeSpendTier,
} from '@destinyitemmanager/dim-api-types';
import { D2ManifestDefinitions } from 'app/destiny2/d2-definitions';
import { t } from 'app/i18next-t';
import { DimItem, PluggableInventoryItemDefinition } from 'app/inventory/item-types';
import { DimStore } from 'app/inventory/store-types';
import { getCurrentStore, getItemAcrossStores } from 'app/inventory/stores-helpers';
import { Loadout } from 'app/loadout-drawer/loadout-types';
import { showNotification } from 'app/notifications/notifications';
import { armor2PlugCategoryHashesByName } from 'app/search/d2-known-values';
import { DestinyClass } from 'bungie-api-ts/destiny2';
import _ from 'lodash';
import { useReducer } from 'react';
import { isLoadoutBuilderItem } from '../loadout/item-utils';
import {
lockedModsFromLoadoutParameters,
statFiltersFromLoadoutParamaters,
statOrderFromLoadoutParameters,
} from './loadout-params';
import { ArmorSet, ArmorStatHashes, ExcludedItems, PinnedItems, StatFilters } from './types';
export interface LoadoutBuilderState {
statOrder: ArmorStatHashes[]; // stat hashes, including disabled stats
upgradeSpendTier: UpgradeSpendTier;
lockItemEnergyType: boolean;
pinnedItems: PinnedItems;
excludedItems: ExcludedItems;
lockedMods: PluggableInventoryItemDefinition[];
lockedExoticHash?: number;
selectedStoreId?: string;
statFilters: Readonly<StatFilters>;
modPicker: {
open: boolean;
initialQuery?: string;
};
compareSet?: ArmorSet;
}
function warnMissingClass(classType: DestinyClass, defs: D2ManifestDefinitions) {
const missingClassName = Object.values(defs.Class).find((c) => c.classType === classType)!
.displayProperties.name; | body: t('LoadoutBuilder.MissingClassDescription'),
});
}
const lbStateInit = ({
stores,
preloadedLoadout,
initialLoadoutParameters,
classType,
defs,
}: {
stores: DimStore[];
preloadedLoadout?: Loadout;
initialLoadoutParameters: LoadoutParameters;
classType: DestinyClass | undefined;
defs: D2ManifestDefinitions;
}): LoadoutBuilderState => {
const pinnedItems: PinnedItems = {};
const matchingClass =
classType !== undefined ? stores.find((store) => store.classType === classType) : undefined;
if (classType !== undefined && !matchingClass) {
warnMissingClass(classType, defs);
// Take out the exotic
initialLoadoutParameters = { ...initialLoadoutParameters, exoticArmorHash: undefined };
}
let selectedStoreId = (matchingClass ?? getCurrentStore(stores)!).id;
let loadoutParams = initialLoadoutParameters;
if (stores.length && preloadedLoadout) {
const loadoutStore = stores.find((store) => store.classType === preloadedLoadout.classType);
if (!loadoutStore) {
warnMissingClass(preloadedLoadout.classType, defs);
} else {
selectedStoreId = loadoutStore.id;
// TODO: instead of locking items, show the loadout fixed at the top to compare against and leave all items free
for (const loadoutItem of preloadedLoadout.items) {
if (loadoutItem.equipped) {
const item = getItemAcrossStores(stores, loadoutItem);
if (item && isLoadoutBuilderItem(item)) {
pinnedItems[item.bucket.hash] = item;
}
}
}
// Load all parameters from the loadout if we can
if (preloadedLoadout.parameters) {
loadoutParams = { ...defaultLoadoutParameters, ...preloadedLoadout.parameters };
}
}
}
const statOrder = statOrderFromLoadoutParameters(loadoutParams);
const statFilters = statFiltersFromLoadoutParamaters(loadoutParams);
const lockedMods = lockedModsFromLoadoutParameters(loadoutParams, defs);
const lockItemEnergyType = Boolean(loadoutParams?.lockItemEnergyType);
// We need to handle the deprecated case
const upgradeSpendTier =
loadoutParams.upgradeSpendTier === UpgradeSpendTier.AscendantShardsLockEnergyType
? UpgradeSpendTier.Nothing
: loadoutParams.upgradeSpendTier!;
const lockedExoticHash = loadoutParams.exoticArmorHash;
return {
lockItemEnergyType,
upgradeSpendTier,
statOrder,
pinnedItems,
excludedItems: [],
statFilters,
lockedMods,
lockedExoticHash,
selectedStoreId,
modPicker: {
open: false,
},
};
};
export type LoadoutBuilderAction =
| { type: 'changeCharacter'; storeId: string }
| { type: 'statFiltersChanged'; statFilters: LoadoutBuilderState['statFilters'] }
| { type: 'sortOrderChanged'; sortOrder: LoadoutBuilderState['statOrder'] }
| {
type: 'lockItemEnergyTypeChanged';
lockItemEnergyType: LoadoutBuilderState['lockItemEnergyType'];
}
| { type: 'upgradeSpendTierChanged'; upgradeSpendTier: LoadoutBuilderState['upgradeSpendTier'] }
| { type: 'pinItem'; item: DimItem }
| { type: 'setPinnedItems'; items: DimItem[] }
| { type: 'unpinItem'; item: DimItem }
| { type: 'excludeItem'; item: DimItem }
| { type: 'unexcludeItem'; item: DimItem }
| {
type: 'lockedModsChanged';
lockedMods: PluggableInventoryItemDefinition[];
}
| { type: 'removeLockedMod'; mod: PluggableInventoryItemDefinition }
| { type: 'addGeneralMods'; mods: PluggableInventoryItemDefinition[] }
| { type: 'lockExotic'; lockedExoticHash: number }
| { type: 'removeLockedExotic' }
| { type: 'openModPicker'; initialQuery?: string }
| { type: 'closeModPicker' }
| { type: 'openCompareDrawer'; set: ArmorSet }
| { type: 'closeCompareDrawer' };
// TODO: Move more logic inside the reducer
function lbStateReducer(
state: LoadoutBuilderState,
action: LoadoutBuilderAction
): LoadoutBuilderState {
switch (action.type) {
case 'changeCharacter':
return {
...state,
selectedStoreId: action.storeId,
pinnedItems: {},
excludedItems: {},
lockedExoticHash: undefined,
};
case 'statFiltersChanged':
return { ...state, statFilters: action.statFilters };
case 'pinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
// Remove any previously locked item in that bucket and add this one
pinnedItems: {
...state.pinnedItems,
[bucketHash]: item,
},
// Locking an item clears excluded items in this bucket
excludedItems: {
...state.excludedItems,
[bucketHash]: undefined,
},
};
}
case 'setPinnedItems': {
const { items } = action;
return {
...state,
pinnedItems: _.keyBy(items, (i) => i.bucket.hash),
excludedItems: {},
};
}
case 'unpinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
pinnedItems: {
...state.pinnedItems,
[bucketHash]: undefined,
},
};
}
case 'excludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
if (state.excludedItems[bucketHash]?.some((i) => i.id === item.id)) {
return state; // item's already there
}
const existingExcluded = state.excludedItems[bucketHash] ?? [];
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: [...existingExcluded, item],
},
};
}
case 'unexcludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
const newExcluded = (state.excludedItems[bucketHash] ?? []).filter((i) => i.id !== item.id);
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: newExcluded.length > 0 ? newExcluded : undefined,
},
};
}
case 'lockedModsChanged': {
return {
...state,
lockedMods: action.lockedMods,
};
}
case 'sortOrderChanged': {
return {
...state,
statOrder: action.sortOrder,
};
}
case 'lockItemEnergyTypeChanged': {
return {
...state,
lockItemEnergyType: action.lockItemEnergyType,
};
}
case 'upgradeSpendTierChanged': {
return {
...state,
upgradeSpendTier: action.upgradeSpendTier,
};
}
case 'addGeneralMods': {
let currentGeneralModsCount = state.lockedMods.filter(
(mod) => mod.plug.plugCategoryHash === armor2PlugCategoryHashesByName.general
).length;
const newMods = [...state.lockedMods];
const failures: string[] = [];
for (const mod of action.mods) {
if (currentGeneralModsCount < 5) {
newMods.push(mod);
currentGeneralModsCount++;
} else {
failures.push(mod.displayProperties.name);
}
}
if (failures.length) {
showNotification({
title: t('LoadoutBuilder.UnableToAddAllMods'),
body: t('LoadoutBuilder.UnableToAddAllModsBody', { mods: failures.join(', ') }),
type: 'warning',
});
}
return {
...state,
lockedMods: newMods,
};
}
case 'removeLockedMod': {
const indexToRemove = state.lockedMods.findIndex((mod) => mod.hash === action.mod.hash);
const newMods = [...state.lockedMods];
newMods.splice(indexToRemove, 1);
return {
...state,
lockedMods: newMods,
};
}
case 'lockExotic': {
const { lockedExoticHash } = action;
return { ...state, lockedExoticHash };
}
case 'removeLockedExotic': {
return { ...state, lockedExoticHash: undefined };
}
case 'openModPicker':
return {
...state,
modPicker: {
open: true,
initialQuery: action.initialQuery,
},
};
case 'closeModPicker':
return { ...state, modPicker: { open: false } };
case 'openCompareDrawer':
return { ...state, compareSet: action.set };
case 'closeCompareDrawer':
return { ...state, compareSet: undefined };
}
}
export function useLbState(
stores: DimStore[],
preloadedLoadout: Loadout | undefined,
classType: DestinyClass | undefined,
initialLoadoutParameters: LoadoutParameters,
defs: D2ManifestDefinitions
) {
return useReducer(
lbStateReducer,
{ stores, preloadedLoadout, initialLoadoutParameters, defs, classType },
lbStateInit
);
} |
showNotification({
type: 'error',
title: t('LoadoutBuilder.MissingClass', { className: missingClassName }), | random_line_split |
loadout-builder-reducer.ts | import {
defaultLoadoutParameters,
LoadoutParameters,
UpgradeSpendTier,
} from '@destinyitemmanager/dim-api-types';
import { D2ManifestDefinitions } from 'app/destiny2/d2-definitions';
import { t } from 'app/i18next-t';
import { DimItem, PluggableInventoryItemDefinition } from 'app/inventory/item-types';
import { DimStore } from 'app/inventory/store-types';
import { getCurrentStore, getItemAcrossStores } from 'app/inventory/stores-helpers';
import { Loadout } from 'app/loadout-drawer/loadout-types';
import { showNotification } from 'app/notifications/notifications';
import { armor2PlugCategoryHashesByName } from 'app/search/d2-known-values';
import { DestinyClass } from 'bungie-api-ts/destiny2';
import _ from 'lodash';
import { useReducer } from 'react';
import { isLoadoutBuilderItem } from '../loadout/item-utils';
import {
lockedModsFromLoadoutParameters,
statFiltersFromLoadoutParamaters,
statOrderFromLoadoutParameters,
} from './loadout-params';
import { ArmorSet, ArmorStatHashes, ExcludedItems, PinnedItems, StatFilters } from './types';
export interface LoadoutBuilderState {
statOrder: ArmorStatHashes[]; // stat hashes, including disabled stats
upgradeSpendTier: UpgradeSpendTier;
lockItemEnergyType: boolean;
pinnedItems: PinnedItems;
excludedItems: ExcludedItems;
lockedMods: PluggableInventoryItemDefinition[];
lockedExoticHash?: number;
selectedStoreId?: string;
statFilters: Readonly<StatFilters>;
modPicker: {
open: boolean;
initialQuery?: string;
};
compareSet?: ArmorSet;
}
function | (classType: DestinyClass, defs: D2ManifestDefinitions) {
const missingClassName = Object.values(defs.Class).find((c) => c.classType === classType)!
.displayProperties.name;
showNotification({
type: 'error',
title: t('LoadoutBuilder.MissingClass', { className: missingClassName }),
body: t('LoadoutBuilder.MissingClassDescription'),
});
}
const lbStateInit = ({
stores,
preloadedLoadout,
initialLoadoutParameters,
classType,
defs,
}: {
stores: DimStore[];
preloadedLoadout?: Loadout;
initialLoadoutParameters: LoadoutParameters;
classType: DestinyClass | undefined;
defs: D2ManifestDefinitions;
}): LoadoutBuilderState => {
const pinnedItems: PinnedItems = {};
const matchingClass =
classType !== undefined ? stores.find((store) => store.classType === classType) : undefined;
if (classType !== undefined && !matchingClass) {
warnMissingClass(classType, defs);
// Take out the exotic
initialLoadoutParameters = { ...initialLoadoutParameters, exoticArmorHash: undefined };
}
let selectedStoreId = (matchingClass ?? getCurrentStore(stores)!).id;
let loadoutParams = initialLoadoutParameters;
if (stores.length && preloadedLoadout) {
const loadoutStore = stores.find((store) => store.classType === preloadedLoadout.classType);
if (!loadoutStore) {
warnMissingClass(preloadedLoadout.classType, defs);
} else {
selectedStoreId = loadoutStore.id;
// TODO: instead of locking items, show the loadout fixed at the top to compare against and leave all items free
for (const loadoutItem of preloadedLoadout.items) {
if (loadoutItem.equipped) {
const item = getItemAcrossStores(stores, loadoutItem);
if (item && isLoadoutBuilderItem(item)) {
pinnedItems[item.bucket.hash] = item;
}
}
}
// Load all parameters from the loadout if we can
if (preloadedLoadout.parameters) {
loadoutParams = { ...defaultLoadoutParameters, ...preloadedLoadout.parameters };
}
}
}
const statOrder = statOrderFromLoadoutParameters(loadoutParams);
const statFilters = statFiltersFromLoadoutParamaters(loadoutParams);
const lockedMods = lockedModsFromLoadoutParameters(loadoutParams, defs);
const lockItemEnergyType = Boolean(loadoutParams?.lockItemEnergyType);
// We need to handle the deprecated case
const upgradeSpendTier =
loadoutParams.upgradeSpendTier === UpgradeSpendTier.AscendantShardsLockEnergyType
? UpgradeSpendTier.Nothing
: loadoutParams.upgradeSpendTier!;
const lockedExoticHash = loadoutParams.exoticArmorHash;
return {
lockItemEnergyType,
upgradeSpendTier,
statOrder,
pinnedItems,
excludedItems: [],
statFilters,
lockedMods,
lockedExoticHash,
selectedStoreId,
modPicker: {
open: false,
},
};
};
export type LoadoutBuilderAction =
| { type: 'changeCharacter'; storeId: string }
| { type: 'statFiltersChanged'; statFilters: LoadoutBuilderState['statFilters'] }
| { type: 'sortOrderChanged'; sortOrder: LoadoutBuilderState['statOrder'] }
| {
type: 'lockItemEnergyTypeChanged';
lockItemEnergyType: LoadoutBuilderState['lockItemEnergyType'];
}
| { type: 'upgradeSpendTierChanged'; upgradeSpendTier: LoadoutBuilderState['upgradeSpendTier'] }
| { type: 'pinItem'; item: DimItem }
| { type: 'setPinnedItems'; items: DimItem[] }
| { type: 'unpinItem'; item: DimItem }
| { type: 'excludeItem'; item: DimItem }
| { type: 'unexcludeItem'; item: DimItem }
| {
type: 'lockedModsChanged';
lockedMods: PluggableInventoryItemDefinition[];
}
| { type: 'removeLockedMod'; mod: PluggableInventoryItemDefinition }
| { type: 'addGeneralMods'; mods: PluggableInventoryItemDefinition[] }
| { type: 'lockExotic'; lockedExoticHash: number }
| { type: 'removeLockedExotic' }
| { type: 'openModPicker'; initialQuery?: string }
| { type: 'closeModPicker' }
| { type: 'openCompareDrawer'; set: ArmorSet }
| { type: 'closeCompareDrawer' };
// TODO: Move more logic inside the reducer
function lbStateReducer(
state: LoadoutBuilderState,
action: LoadoutBuilderAction
): LoadoutBuilderState {
switch (action.type) {
case 'changeCharacter':
return {
...state,
selectedStoreId: action.storeId,
pinnedItems: {},
excludedItems: {},
lockedExoticHash: undefined,
};
case 'statFiltersChanged':
return { ...state, statFilters: action.statFilters };
case 'pinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
// Remove any previously locked item in that bucket and add this one
pinnedItems: {
...state.pinnedItems,
[bucketHash]: item,
},
// Locking an item clears excluded items in this bucket
excludedItems: {
...state.excludedItems,
[bucketHash]: undefined,
},
};
}
case 'setPinnedItems': {
const { items } = action;
return {
...state,
pinnedItems: _.keyBy(items, (i) => i.bucket.hash),
excludedItems: {},
};
}
case 'unpinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
pinnedItems: {
...state.pinnedItems,
[bucketHash]: undefined,
},
};
}
case 'excludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
if (state.excludedItems[bucketHash]?.some((i) => i.id === item.id)) {
return state; // item's already there
}
const existingExcluded = state.excludedItems[bucketHash] ?? [];
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: [...existingExcluded, item],
},
};
}
case 'unexcludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
const newExcluded = (state.excludedItems[bucketHash] ?? []).filter((i) => i.id !== item.id);
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: newExcluded.length > 0 ? newExcluded : undefined,
},
};
}
case 'lockedModsChanged': {
return {
...state,
lockedMods: action.lockedMods,
};
}
case 'sortOrderChanged': {
return {
...state,
statOrder: action.sortOrder,
};
}
case 'lockItemEnergyTypeChanged': {
return {
...state,
lockItemEnergyType: action.lockItemEnergyType,
};
}
case 'upgradeSpendTierChanged': {
return {
...state,
upgradeSpendTier: action.upgradeSpendTier,
};
}
case 'addGeneralMods': {
let currentGeneralModsCount = state.lockedMods.filter(
(mod) => mod.plug.plugCategoryHash === armor2PlugCategoryHashesByName.general
).length;
const newMods = [...state.lockedMods];
const failures: string[] = [];
for (const mod of action.mods) {
if (currentGeneralModsCount < 5) {
newMods.push(mod);
currentGeneralModsCount++;
} else {
failures.push(mod.displayProperties.name);
}
}
if (failures.length) {
showNotification({
title: t('LoadoutBuilder.UnableToAddAllMods'),
body: t('LoadoutBuilder.UnableToAddAllModsBody', { mods: failures.join(', ') }),
type: 'warning',
});
}
return {
...state,
lockedMods: newMods,
};
}
case 'removeLockedMod': {
const indexToRemove = state.lockedMods.findIndex((mod) => mod.hash === action.mod.hash);
const newMods = [...state.lockedMods];
newMods.splice(indexToRemove, 1);
return {
...state,
lockedMods: newMods,
};
}
case 'lockExotic': {
const { lockedExoticHash } = action;
return { ...state, lockedExoticHash };
}
case 'removeLockedExotic': {
return { ...state, lockedExoticHash: undefined };
}
case 'openModPicker':
return {
...state,
modPicker: {
open: true,
initialQuery: action.initialQuery,
},
};
case 'closeModPicker':
return { ...state, modPicker: { open: false } };
case 'openCompareDrawer':
return { ...state, compareSet: action.set };
case 'closeCompareDrawer':
return { ...state, compareSet: undefined };
}
}
export function useLbState(
stores: DimStore[],
preloadedLoadout: Loadout | undefined,
classType: DestinyClass | undefined,
initialLoadoutParameters: LoadoutParameters,
defs: D2ManifestDefinitions
) {
return useReducer(
lbStateReducer,
{ stores, preloadedLoadout, initialLoadoutParameters, defs, classType },
lbStateInit
);
}
| warnMissingClass | identifier_name |
loadout-builder-reducer.ts | import {
defaultLoadoutParameters,
LoadoutParameters,
UpgradeSpendTier,
} from '@destinyitemmanager/dim-api-types';
import { D2ManifestDefinitions } from 'app/destiny2/d2-definitions';
import { t } from 'app/i18next-t';
import { DimItem, PluggableInventoryItemDefinition } from 'app/inventory/item-types';
import { DimStore } from 'app/inventory/store-types';
import { getCurrentStore, getItemAcrossStores } from 'app/inventory/stores-helpers';
import { Loadout } from 'app/loadout-drawer/loadout-types';
import { showNotification } from 'app/notifications/notifications';
import { armor2PlugCategoryHashesByName } from 'app/search/d2-known-values';
import { DestinyClass } from 'bungie-api-ts/destiny2';
import _ from 'lodash';
import { useReducer } from 'react';
import { isLoadoutBuilderItem } from '../loadout/item-utils';
import {
lockedModsFromLoadoutParameters,
statFiltersFromLoadoutParamaters,
statOrderFromLoadoutParameters,
} from './loadout-params';
import { ArmorSet, ArmorStatHashes, ExcludedItems, PinnedItems, StatFilters } from './types';
export interface LoadoutBuilderState {
statOrder: ArmorStatHashes[]; // stat hashes, including disabled stats
upgradeSpendTier: UpgradeSpendTier;
lockItemEnergyType: boolean;
pinnedItems: PinnedItems;
excludedItems: ExcludedItems;
lockedMods: PluggableInventoryItemDefinition[];
lockedExoticHash?: number;
selectedStoreId?: string;
statFilters: Readonly<StatFilters>;
modPicker: {
open: boolean;
initialQuery?: string;
};
compareSet?: ArmorSet;
}
function warnMissingClass(classType: DestinyClass, defs: D2ManifestDefinitions) {
const missingClassName = Object.values(defs.Class).find((c) => c.classType === classType)!
.displayProperties.name;
showNotification({
type: 'error',
title: t('LoadoutBuilder.MissingClass', { className: missingClassName }),
body: t('LoadoutBuilder.MissingClassDescription'),
});
}
const lbStateInit = ({
stores,
preloadedLoadout,
initialLoadoutParameters,
classType,
defs,
}: {
stores: DimStore[];
preloadedLoadout?: Loadout;
initialLoadoutParameters: LoadoutParameters;
classType: DestinyClass | undefined;
defs: D2ManifestDefinitions;
}): LoadoutBuilderState => {
const pinnedItems: PinnedItems = {};
const matchingClass =
classType !== undefined ? stores.find((store) => store.classType === classType) : undefined;
if (classType !== undefined && !matchingClass) {
warnMissingClass(classType, defs);
// Take out the exotic
initialLoadoutParameters = { ...initialLoadoutParameters, exoticArmorHash: undefined };
}
let selectedStoreId = (matchingClass ?? getCurrentStore(stores)!).id;
let loadoutParams = initialLoadoutParameters;
if (stores.length && preloadedLoadout) {
const loadoutStore = stores.find((store) => store.classType === preloadedLoadout.classType);
if (!loadoutStore) {
warnMissingClass(preloadedLoadout.classType, defs);
} else {
selectedStoreId = loadoutStore.id;
// TODO: instead of locking items, show the loadout fixed at the top to compare against and leave all items free
for (const loadoutItem of preloadedLoadout.items) {
if (loadoutItem.equipped) {
const item = getItemAcrossStores(stores, loadoutItem);
if (item && isLoadoutBuilderItem(item)) {
pinnedItems[item.bucket.hash] = item;
}
}
}
// Load all parameters from the loadout if we can
if (preloadedLoadout.parameters) {
loadoutParams = { ...defaultLoadoutParameters, ...preloadedLoadout.parameters };
}
}
}
const statOrder = statOrderFromLoadoutParameters(loadoutParams);
const statFilters = statFiltersFromLoadoutParamaters(loadoutParams);
const lockedMods = lockedModsFromLoadoutParameters(loadoutParams, defs);
const lockItemEnergyType = Boolean(loadoutParams?.lockItemEnergyType);
// We need to handle the deprecated case
const upgradeSpendTier =
loadoutParams.upgradeSpendTier === UpgradeSpendTier.AscendantShardsLockEnergyType
? UpgradeSpendTier.Nothing
: loadoutParams.upgradeSpendTier!;
const lockedExoticHash = loadoutParams.exoticArmorHash;
return {
lockItemEnergyType,
upgradeSpendTier,
statOrder,
pinnedItems,
excludedItems: [],
statFilters,
lockedMods,
lockedExoticHash,
selectedStoreId,
modPicker: {
open: false,
},
};
};
export type LoadoutBuilderAction =
| { type: 'changeCharacter'; storeId: string }
| { type: 'statFiltersChanged'; statFilters: LoadoutBuilderState['statFilters'] }
| { type: 'sortOrderChanged'; sortOrder: LoadoutBuilderState['statOrder'] }
| {
type: 'lockItemEnergyTypeChanged';
lockItemEnergyType: LoadoutBuilderState['lockItemEnergyType'];
}
| { type: 'upgradeSpendTierChanged'; upgradeSpendTier: LoadoutBuilderState['upgradeSpendTier'] }
| { type: 'pinItem'; item: DimItem }
| { type: 'setPinnedItems'; items: DimItem[] }
| { type: 'unpinItem'; item: DimItem }
| { type: 'excludeItem'; item: DimItem }
| { type: 'unexcludeItem'; item: DimItem }
| {
type: 'lockedModsChanged';
lockedMods: PluggableInventoryItemDefinition[];
}
| { type: 'removeLockedMod'; mod: PluggableInventoryItemDefinition }
| { type: 'addGeneralMods'; mods: PluggableInventoryItemDefinition[] }
| { type: 'lockExotic'; lockedExoticHash: number }
| { type: 'removeLockedExotic' }
| { type: 'openModPicker'; initialQuery?: string }
| { type: 'closeModPicker' }
| { type: 'openCompareDrawer'; set: ArmorSet }
| { type: 'closeCompareDrawer' };
// TODO: Move more logic inside the reducer
function lbStateReducer(
state: LoadoutBuilderState,
action: LoadoutBuilderAction
): LoadoutBuilderState |
export function useLbState(
stores: DimStore[],
preloadedLoadout: Loadout | undefined,
classType: DestinyClass | undefined,
initialLoadoutParameters: LoadoutParameters,
defs: D2ManifestDefinitions
) {
return useReducer(
lbStateReducer,
{ stores, preloadedLoadout, initialLoadoutParameters, defs, classType },
lbStateInit
);
}
| {
switch (action.type) {
case 'changeCharacter':
return {
...state,
selectedStoreId: action.storeId,
pinnedItems: {},
excludedItems: {},
lockedExoticHash: undefined,
};
case 'statFiltersChanged':
return { ...state, statFilters: action.statFilters };
case 'pinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
// Remove any previously locked item in that bucket and add this one
pinnedItems: {
...state.pinnedItems,
[bucketHash]: item,
},
// Locking an item clears excluded items in this bucket
excludedItems: {
...state.excludedItems,
[bucketHash]: undefined,
},
};
}
case 'setPinnedItems': {
const { items } = action;
return {
...state,
pinnedItems: _.keyBy(items, (i) => i.bucket.hash),
excludedItems: {},
};
}
case 'unpinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
pinnedItems: {
...state.pinnedItems,
[bucketHash]: undefined,
},
};
}
case 'excludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
if (state.excludedItems[bucketHash]?.some((i) => i.id === item.id)) {
return state; // item's already there
}
const existingExcluded = state.excludedItems[bucketHash] ?? [];
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: [...existingExcluded, item],
},
};
}
case 'unexcludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
const newExcluded = (state.excludedItems[bucketHash] ?? []).filter((i) => i.id !== item.id);
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: newExcluded.length > 0 ? newExcluded : undefined,
},
};
}
case 'lockedModsChanged': {
return {
...state,
lockedMods: action.lockedMods,
};
}
case 'sortOrderChanged': {
return {
...state,
statOrder: action.sortOrder,
};
}
case 'lockItemEnergyTypeChanged': {
return {
...state,
lockItemEnergyType: action.lockItemEnergyType,
};
}
case 'upgradeSpendTierChanged': {
return {
...state,
upgradeSpendTier: action.upgradeSpendTier,
};
}
case 'addGeneralMods': {
let currentGeneralModsCount = state.lockedMods.filter(
(mod) => mod.plug.plugCategoryHash === armor2PlugCategoryHashesByName.general
).length;
const newMods = [...state.lockedMods];
const failures: string[] = [];
for (const mod of action.mods) {
if (currentGeneralModsCount < 5) {
newMods.push(mod);
currentGeneralModsCount++;
} else {
failures.push(mod.displayProperties.name);
}
}
if (failures.length) {
showNotification({
title: t('LoadoutBuilder.UnableToAddAllMods'),
body: t('LoadoutBuilder.UnableToAddAllModsBody', { mods: failures.join(', ') }),
type: 'warning',
});
}
return {
...state,
lockedMods: newMods,
};
}
case 'removeLockedMod': {
const indexToRemove = state.lockedMods.findIndex((mod) => mod.hash === action.mod.hash);
const newMods = [...state.lockedMods];
newMods.splice(indexToRemove, 1);
return {
...state,
lockedMods: newMods,
};
}
case 'lockExotic': {
const { lockedExoticHash } = action;
return { ...state, lockedExoticHash };
}
case 'removeLockedExotic': {
return { ...state, lockedExoticHash: undefined };
}
case 'openModPicker':
return {
...state,
modPicker: {
open: true,
initialQuery: action.initialQuery,
},
};
case 'closeModPicker':
return { ...state, modPicker: { open: false } };
case 'openCompareDrawer':
return { ...state, compareSet: action.set };
case 'closeCompareDrawer':
return { ...state, compareSet: undefined };
}
} | identifier_body |
loadout-builder-reducer.ts | import {
defaultLoadoutParameters,
LoadoutParameters,
UpgradeSpendTier,
} from '@destinyitemmanager/dim-api-types';
import { D2ManifestDefinitions } from 'app/destiny2/d2-definitions';
import { t } from 'app/i18next-t';
import { DimItem, PluggableInventoryItemDefinition } from 'app/inventory/item-types';
import { DimStore } from 'app/inventory/store-types';
import { getCurrentStore, getItemAcrossStores } from 'app/inventory/stores-helpers';
import { Loadout } from 'app/loadout-drawer/loadout-types';
import { showNotification } from 'app/notifications/notifications';
import { armor2PlugCategoryHashesByName } from 'app/search/d2-known-values';
import { DestinyClass } from 'bungie-api-ts/destiny2';
import _ from 'lodash';
import { useReducer } from 'react';
import { isLoadoutBuilderItem } from '../loadout/item-utils';
import {
lockedModsFromLoadoutParameters,
statFiltersFromLoadoutParamaters,
statOrderFromLoadoutParameters,
} from './loadout-params';
import { ArmorSet, ArmorStatHashes, ExcludedItems, PinnedItems, StatFilters } from './types';
export interface LoadoutBuilderState {
statOrder: ArmorStatHashes[]; // stat hashes, including disabled stats
upgradeSpendTier: UpgradeSpendTier;
lockItemEnergyType: boolean;
pinnedItems: PinnedItems;
excludedItems: ExcludedItems;
lockedMods: PluggableInventoryItemDefinition[];
lockedExoticHash?: number;
selectedStoreId?: string;
statFilters: Readonly<StatFilters>;
modPicker: {
open: boolean;
initialQuery?: string;
};
compareSet?: ArmorSet;
}
function warnMissingClass(classType: DestinyClass, defs: D2ManifestDefinitions) {
const missingClassName = Object.values(defs.Class).find((c) => c.classType === classType)!
.displayProperties.name;
showNotification({
type: 'error',
title: t('LoadoutBuilder.MissingClass', { className: missingClassName }),
body: t('LoadoutBuilder.MissingClassDescription'),
});
}
const lbStateInit = ({
stores,
preloadedLoadout,
initialLoadoutParameters,
classType,
defs,
}: {
stores: DimStore[];
preloadedLoadout?: Loadout;
initialLoadoutParameters: LoadoutParameters;
classType: DestinyClass | undefined;
defs: D2ManifestDefinitions;
}): LoadoutBuilderState => {
const pinnedItems: PinnedItems = {};
const matchingClass =
classType !== undefined ? stores.find((store) => store.classType === classType) : undefined;
if (classType !== undefined && !matchingClass) {
warnMissingClass(classType, defs);
// Take out the exotic
initialLoadoutParameters = { ...initialLoadoutParameters, exoticArmorHash: undefined };
}
let selectedStoreId = (matchingClass ?? getCurrentStore(stores)!).id;
let loadoutParams = initialLoadoutParameters;
if (stores.length && preloadedLoadout) |
const statOrder = statOrderFromLoadoutParameters(loadoutParams);
const statFilters = statFiltersFromLoadoutParamaters(loadoutParams);
const lockedMods = lockedModsFromLoadoutParameters(loadoutParams, defs);
const lockItemEnergyType = Boolean(loadoutParams?.lockItemEnergyType);
// We need to handle the deprecated case
const upgradeSpendTier =
loadoutParams.upgradeSpendTier === UpgradeSpendTier.AscendantShardsLockEnergyType
? UpgradeSpendTier.Nothing
: loadoutParams.upgradeSpendTier!;
const lockedExoticHash = loadoutParams.exoticArmorHash;
return {
lockItemEnergyType,
upgradeSpendTier,
statOrder,
pinnedItems,
excludedItems: [],
statFilters,
lockedMods,
lockedExoticHash,
selectedStoreId,
modPicker: {
open: false,
},
};
};
export type LoadoutBuilderAction =
| { type: 'changeCharacter'; storeId: string }
| { type: 'statFiltersChanged'; statFilters: LoadoutBuilderState['statFilters'] }
| { type: 'sortOrderChanged'; sortOrder: LoadoutBuilderState['statOrder'] }
| {
type: 'lockItemEnergyTypeChanged';
lockItemEnergyType: LoadoutBuilderState['lockItemEnergyType'];
}
| { type: 'upgradeSpendTierChanged'; upgradeSpendTier: LoadoutBuilderState['upgradeSpendTier'] }
| { type: 'pinItem'; item: DimItem }
| { type: 'setPinnedItems'; items: DimItem[] }
| { type: 'unpinItem'; item: DimItem }
| { type: 'excludeItem'; item: DimItem }
| { type: 'unexcludeItem'; item: DimItem }
| {
type: 'lockedModsChanged';
lockedMods: PluggableInventoryItemDefinition[];
}
| { type: 'removeLockedMod'; mod: PluggableInventoryItemDefinition }
| { type: 'addGeneralMods'; mods: PluggableInventoryItemDefinition[] }
| { type: 'lockExotic'; lockedExoticHash: number }
| { type: 'removeLockedExotic' }
| { type: 'openModPicker'; initialQuery?: string }
| { type: 'closeModPicker' }
| { type: 'openCompareDrawer'; set: ArmorSet }
| { type: 'closeCompareDrawer' };
// TODO: Move more logic inside the reducer
function lbStateReducer(
state: LoadoutBuilderState,
action: LoadoutBuilderAction
): LoadoutBuilderState {
switch (action.type) {
case 'changeCharacter':
return {
...state,
selectedStoreId: action.storeId,
pinnedItems: {},
excludedItems: {},
lockedExoticHash: undefined,
};
case 'statFiltersChanged':
return { ...state, statFilters: action.statFilters };
case 'pinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
// Remove any previously locked item in that bucket and add this one
pinnedItems: {
...state.pinnedItems,
[bucketHash]: item,
},
// Locking an item clears excluded items in this bucket
excludedItems: {
...state.excludedItems,
[bucketHash]: undefined,
},
};
}
case 'setPinnedItems': {
const { items } = action;
return {
...state,
pinnedItems: _.keyBy(items, (i) => i.bucket.hash),
excludedItems: {},
};
}
case 'unpinItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
return {
...state,
pinnedItems: {
...state.pinnedItems,
[bucketHash]: undefined,
},
};
}
case 'excludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
if (state.excludedItems[bucketHash]?.some((i) => i.id === item.id)) {
return state; // item's already there
}
const existingExcluded = state.excludedItems[bucketHash] ?? [];
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: [...existingExcluded, item],
},
};
}
case 'unexcludeItem': {
const { item } = action;
const bucketHash = item.bucket.hash;
const newExcluded = (state.excludedItems[bucketHash] ?? []).filter((i) => i.id !== item.id);
return {
...state,
excludedItems: {
...state.excludedItems,
[bucketHash]: newExcluded.length > 0 ? newExcluded : undefined,
},
};
}
case 'lockedModsChanged': {
return {
...state,
lockedMods: action.lockedMods,
};
}
case 'sortOrderChanged': {
return {
...state,
statOrder: action.sortOrder,
};
}
case 'lockItemEnergyTypeChanged': {
return {
...state,
lockItemEnergyType: action.lockItemEnergyType,
};
}
case 'upgradeSpendTierChanged': {
return {
...state,
upgradeSpendTier: action.upgradeSpendTier,
};
}
case 'addGeneralMods': {
let currentGeneralModsCount = state.lockedMods.filter(
(mod) => mod.plug.plugCategoryHash === armor2PlugCategoryHashesByName.general
).length;
const newMods = [...state.lockedMods];
const failures: string[] = [];
for (const mod of action.mods) {
if (currentGeneralModsCount < 5) {
newMods.push(mod);
currentGeneralModsCount++;
} else {
failures.push(mod.displayProperties.name);
}
}
if (failures.length) {
showNotification({
title: t('LoadoutBuilder.UnableToAddAllMods'),
body: t('LoadoutBuilder.UnableToAddAllModsBody', { mods: failures.join(', ') }),
type: 'warning',
});
}
return {
...state,
lockedMods: newMods,
};
}
case 'removeLockedMod': {
const indexToRemove = state.lockedMods.findIndex((mod) => mod.hash === action.mod.hash);
const newMods = [...state.lockedMods];
newMods.splice(indexToRemove, 1);
return {
...state,
lockedMods: newMods,
};
}
case 'lockExotic': {
const { lockedExoticHash } = action;
return { ...state, lockedExoticHash };
}
case 'removeLockedExotic': {
return { ...state, lockedExoticHash: undefined };
}
case 'openModPicker':
return {
...state,
modPicker: {
open: true,
initialQuery: action.initialQuery,
},
};
case 'closeModPicker':
return { ...state, modPicker: { open: false } };
case 'openCompareDrawer':
return { ...state, compareSet: action.set };
case 'closeCompareDrawer':
return { ...state, compareSet: undefined };
}
}
export function useLbState(
stores: DimStore[],
preloadedLoadout: Loadout | undefined,
classType: DestinyClass | undefined,
initialLoadoutParameters: LoadoutParameters,
defs: D2ManifestDefinitions
) {
return useReducer(
lbStateReducer,
{ stores, preloadedLoadout, initialLoadoutParameters, defs, classType },
lbStateInit
);
}
| {
const loadoutStore = stores.find((store) => store.classType === preloadedLoadout.classType);
if (!loadoutStore) {
warnMissingClass(preloadedLoadout.classType, defs);
} else {
selectedStoreId = loadoutStore.id;
// TODO: instead of locking items, show the loadout fixed at the top to compare against and leave all items free
for (const loadoutItem of preloadedLoadout.items) {
if (loadoutItem.equipped) {
const item = getItemAcrossStores(stores, loadoutItem);
if (item && isLoadoutBuilderItem(item)) {
pinnedItems[item.bucket.hash] = item;
}
}
}
// Load all parameters from the loadout if we can
if (preloadedLoadout.parameters) {
loadoutParams = { ...defaultLoadoutParameters, ...preloadedLoadout.parameters };
}
}
} | conditional_block |
config.rs | // Copyright 2019 Twitter, Inc.
// Licensed under the Apache License, Version 2.0
// http://www.apache.org/licenses/LICENSE-2.0
use crate::common::*;
use clap::{App, Arg, ArgMatches};
use logger::*;
use std::collections::{HashMap, HashSet};
use std::fs::File;
use std::io::{BufRead, BufReader};
use std::net::SocketAddr;
use std::process;
use std::time::Duration;
pub const VERSION: &str = env!("CARGO_PKG_VERSION");
pub const NAME: &str = env!("CARGO_PKG_NAME");
pub const DEFAULT_SAMPLE_RATE_HZ: f64 = 1.0;
pub const DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS: usize = 50;
pub const DEFAULT_MAX_SAMPLER_TIMEOUTS: usize = 5;
pub const DEFAULT_INTERVAL_SECONDS: usize = 60;
/// This struct contains the configuration of the agent.
#[derive(Clone)]
pub struct Config {
/// the latching interval for stats
interval: u64,
/// sample rate for counters in Hz
sample_rate: f64,
/// the sampler timeout
sampler_timeout: Duration,
/// maximum consecutive sampler timeouts
max_sampler_timeouts: usize,
/// the listen address for the stats port
listen: SocketAddr,
/// the logging level
loglevel: Level,
/// memcache instance to instrument
memcache: Option<SocketAddr>,
/// flags for enabled statistics subsystems
flags: Flags,
/// the number of cores on the host
cores: usize,
/// an optional file to log stats to
stats_log: Option<String>,
/// flag to indicate Mesos sidecar mode
sidecar: bool,
}
#[derive(Clone)]
/// `Flags` is a simple wrapper for a doubly-keyed `HashSet`
pub struct Flags {
data: HashMap<String, HashSet<String>>,
}
impl Flags {
/// Creates a new empty set of `Flags`
pub fn new() -> Self {
Self {
data: HashMap::new(),
}
}
/// Insert a `pkey`+`lkey` into the set
pub fn insert(&mut self, pkey: &str, lkey: &str) {
let mut entry = self.data.remove(pkey).unwrap_or_default();
entry.insert(lkey.to_owned());
self.data.insert(pkey.to_owned(), entry);
}
/// True if the set contains `pkey`+`lkey`
pub fn contains(&self, pkey: &str, lkey: &str) -> bool {
if let Some(entry) = self.data.get(pkey) {
entry.get(lkey).is_some()
} else {
false
}
}
/// True if the set contains the `pkey`
pub fn | (&self, pkey: &str) -> bool {
self.data.get(pkey).is_some()
}
/// Remove a `pkey`+`lkey`
pub fn remove(&mut self, pkey: &str, lkey: &str) {
if let Some(entry) = self.data.get_mut(pkey) {
entry.remove(lkey);
}
}
/// Remove the `pkey` and all `lkey`s under it
pub fn remove_pkey(&mut self, pkey: &str) {
self.data.remove(pkey);
}
}
impl Config {
/// parse command line options and return `Config`
pub fn new() -> Config {
let matches = App::new(NAME)
.version(VERSION)
.author("Brian Martin <bmartin@twitter.com>")
.about("high-resolution systems performance telemetry agent")
.arg(
Arg::with_name("listen")
.short("l")
.long("listen")
.required(true)
.takes_value(true)
.value_name("IP:PORT")
.help("Sets the listen address for metrics"),
)
.arg(
Arg::with_name("verbose")
.short("v")
.long("verbose")
.multiple(true)
.help("Increase verbosity by one level. Can be used more than once"),
)
.arg(
Arg::with_name("interval")
.long("interval")
.value_name("Seconds")
.help("Integration window duration and stats endpoint refresh time")
.takes_value(true),
)
.arg(
Arg::with_name("sample-rate")
.long("sample-rate")
.value_name("Hertz")
.help("Sets the sampling frequency for the counters")
.takes_value(true),
)
.arg(
Arg::with_name("sampler-timeout")
.long("sampler-timeout")
.value_name("MS")
.help("Sets the timeout for per-sampler execution")
.takes_value(true),
)
.arg(
Arg::with_name("max-sampler-timeouts")
.long("max-sampler-timeouts")
.value_name("MS")
.help("Sets the maximum number of consecutive sampler timeouts")
.takes_value(true),
)
.arg(
Arg::with_name("cpu")
.long("cpu")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from CPU subsystem"),
)
.arg(
Arg::with_name("disk")
.long("disk")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Disk subsystem"),
)
.arg(
Arg::with_name("ebpf")
.long("ebpf")
.takes_value(true)
.multiple(true)
.possible_value("all")
.possible_value("block")
.possible_value("ext4")
.possible_value("scheduler")
.possible_value("xfs")
.help("Enable statistics from eBPF"),
)
.arg(
Arg::with_name("network")
.long("network")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Network subsystem"),
)
.arg(
Arg::with_name("perf")
.long("perf")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.possible_value("per-cgroup")
.help("Enable statistics from Perf Events subsystem"),
)
.arg(
Arg::with_name("memcache")
.long("memcache")
.required(false)
.takes_value(true)
.value_name("IP:PORT")
.help("Connect to the given memcache server and produce stats"),
)
.arg(
Arg::with_name("stats-log")
.long("stats-log")
.required(false)
.takes_value(true)
.value_name("LOG FILE")
.help("Enable logging of stats to file"),
)
.arg(
Arg::with_name("sidecar")
.long("sidecar")
.required(false)
.help("Enables Mesos sidecar mode, instrumenting the container"),
)
.get_matches();
let listen = matches
.value_of("listen")
.unwrap()
.parse()
.unwrap_or_else(|_| {
println!("ERROR: listen address is malformed");
process::exit(1);
});
let memcache = if let Some(sock) = matches.value_of("memcache") {
let socket = sock.parse().unwrap_or_else(|_| {
println!("ERROR: memcache address is malformed");
process::exit(1);
});
Some(socket)
} else {
None
};
let sample_rate =
parse_float_arg(&matches, "sample-rate").unwrap_or(DEFAULT_SAMPLE_RATE_HZ);
let sampler_timeout = Duration::from_millis(
parse_numeric_arg(&matches, "sampler-timeout")
.unwrap_or(DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS) as u64,
);
let max_sampler_timeouts = parse_numeric_arg(&matches, "max-sampler-timeouts")
.unwrap_or(DEFAULT_MAX_SAMPLER_TIMEOUTS);
let interval = parse_numeric_arg(&matches, "interval").unwrap_or(DEFAULT_INTERVAL_SECONDS)
as u64
* SECOND;
let cores = hardware_threads().unwrap_or(1);
let mut stats_enabled = Flags::new();
for subsystem in &["cpu", "disk", "ebpf", "network", "perf"] {
if let Some(values) = matches.values_of(subsystem) {
let flags: Vec<&str> = values.collect();
for flag in flags {
stats_enabled.insert(subsystem, flag);
}
}
}
let loglevel = match matches.occurrences_of("verbose") {
0 => Level::Info,
1 => Level::Debug,
_ => Level::Trace,
};
let stats_log = matches
.value_of("stats-log")
.map(std::string::ToString::to_string);
let sidecar = matches.is_present("sidecar");
Config {
cores,
flags: stats_enabled,
sample_rate,
sampler_timeout,
max_sampler_timeouts,
interval,
listen,
loglevel,
memcache,
stats_log,
sidecar,
}
}
/// what interval should the stats library latch on
pub fn interval(&self) -> u64 {
self.interval
}
/// what frequency the stats should be sampled on
pub fn sample_rate(&self) -> f64 {
self.sample_rate
}
/// the timeout for sampler execution
pub fn sampler_timeout(&self) -> Duration {
self.sampler_timeout
}
/// maximum consecutive sampler timeouts
pub fn max_sampler_timeouts(&self) -> usize {
self.max_sampler_timeouts
}
/// get listen address
pub fn listen(&self) -> SocketAddr {
self.listen
}
/// get log level
pub fn loglevel(&self) -> Level {
self.loglevel
}
/// how many cores on the host?
pub fn cores(&self) -> usize {
self.cores
}
pub fn memcache(&self) -> Option<SocketAddr> {
self.memcache
}
/// is a flag enabled for a subsystem?
pub fn is_enabled(&self, subsystem: &str, flag: &str) -> bool {
self.flags.contains(subsystem, flag)
}
pub fn is_subsystem_enabled(&self, subsystem: &str) -> bool {
self.flags.contains_pkey(subsystem)
}
pub fn stats_log(&self) -> Option<String> {
self.stats_log.clone()
}
}
/// a helper function to parse a numeric argument by name from `ArgMatches`
fn parse_numeric_arg(matches: &ArgMatches, key: &str) -> Option<usize> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1);
})
})
}
/// a helper function to parse a floating point argument by name from `ArgMatches`
fn parse_float_arg(matches: &ArgMatches, key: &str) -> Option<f64> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1);
})
})
}
/// helper function to discover the number of hardware threads
pub fn hardware_threads() -> Result<usize, ()> {
let path = "/sys/devices/system/cpu/present";
let f = File::open(path).map_err(|e| error!("failed to open file ({:?}): {}", path, e))?;
let mut f = BufReader::new(f);
let mut line = String::new();
f.read_line(&mut line)
.map_err(|_| error!("failed to read line"))?;
let line = line.trim();
let a: Vec<&str> = line.split('-').collect();
a.last()
.unwrap_or(&"0")
.parse::<usize>()
.map_err(|e| error!("could not parse num cpus from file ({:?}): {}", path, e))
.map(|i| i + 1)
}
| contains_pkey | identifier_name |
config.rs | // Copyright 2019 Twitter, Inc.
// Licensed under the Apache License, Version 2.0
// http://www.apache.org/licenses/LICENSE-2.0
use crate::common::*;
use clap::{App, Arg, ArgMatches};
use logger::*;
use std::collections::{HashMap, HashSet};
use std::fs::File;
use std::io::{BufRead, BufReader};
use std::net::SocketAddr;
use std::process;
use std::time::Duration;
pub const VERSION: &str = env!("CARGO_PKG_VERSION");
pub const NAME: &str = env!("CARGO_PKG_NAME");
pub const DEFAULT_SAMPLE_RATE_HZ: f64 = 1.0;
pub const DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS: usize = 50;
pub const DEFAULT_MAX_SAMPLER_TIMEOUTS: usize = 5;
pub const DEFAULT_INTERVAL_SECONDS: usize = 60;
/// This struct contains the configuration of the agent.
#[derive(Clone)]
pub struct Config {
/// the latching interval for stats
interval: u64,
/// sample rate for counters in Hz
sample_rate: f64,
/// the sampler timeout
sampler_timeout: Duration,
/// maximum consecutive sampler timeouts
max_sampler_timeouts: usize,
/// the listen address for the stats port
listen: SocketAddr,
/// the logging level
loglevel: Level,
/// memcache instance to instrument
memcache: Option<SocketAddr>,
/// flags for enabled statistics subsystems
flags: Flags,
/// the number of cores on the host
cores: usize,
/// an optional file to log stats to
stats_log: Option<String>,
/// flag to indicate Mesos sidecar mode
sidecar: bool,
}
#[derive(Clone)]
/// `Flags` is a simple wrapper for a doubly-keyed `HashSet`
pub struct Flags {
data: HashMap<String, HashSet<String>>,
}
impl Flags {
/// Creates a new empty set of `Flags`
pub fn new() -> Self {
Self {
data: HashMap::new(),
}
}
/// Insert a `pkey`+`lkey` into the set
pub fn insert(&mut self, pkey: &str, lkey: &str) {
let mut entry = self.data.remove(pkey).unwrap_or_default();
entry.insert(lkey.to_owned());
self.data.insert(pkey.to_owned(), entry);
}
/// True if the set contains `pkey`+`lkey`
pub fn contains(&self, pkey: &str, lkey: &str) -> bool {
if let Some(entry) = self.data.get(pkey) {
entry.get(lkey).is_some()
} else {
false
}
}
/// True if the set contains the `pkey`
pub fn contains_pkey(&self, pkey: &str) -> bool {
self.data.get(pkey).is_some()
}
/// Remove a `pkey`+`lkey`
pub fn remove(&mut self, pkey: &str, lkey: &str) {
if let Some(entry) = self.data.get_mut(pkey) {
entry.remove(lkey);
}
}
/// Remove the `pkey` and all `lkey`s under it
pub fn remove_pkey(&mut self, pkey: &str) {
self.data.remove(pkey);
}
}
impl Config {
/// parse command line options and return `Config`
pub fn new() -> Config {
let matches = App::new(NAME)
.version(VERSION)
.author("Brian Martin <bmartin@twitter.com>")
.about("high-resolution systems performance telemetry agent")
.arg(
Arg::with_name("listen")
.short("l")
.long("listen")
.required(true)
.takes_value(true)
.value_name("IP:PORT")
.help("Sets the listen address for metrics"),
)
.arg(
Arg::with_name("verbose")
.short("v")
.long("verbose")
.multiple(true)
.help("Increase verbosity by one level. Can be used more than once"),
)
.arg(
Arg::with_name("interval")
.long("interval")
.value_name("Seconds")
.help("Integration window duration and stats endpoint refresh time")
.takes_value(true),
)
.arg(
Arg::with_name("sample-rate")
.long("sample-rate")
.value_name("Hertz")
.help("Sets the sampling frequency for the counters")
.takes_value(true),
)
.arg(
Arg::with_name("sampler-timeout")
.long("sampler-timeout")
.value_name("MS")
.help("Sets the timeout for per-sampler execution")
.takes_value(true),
)
.arg(
Arg::with_name("max-sampler-timeouts")
.long("max-sampler-timeouts")
.value_name("MS")
.help("Sets the maximum number of consecutive sampler timeouts")
.takes_value(true),
)
.arg(
Arg::with_name("cpu")
.long("cpu")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from CPU subsystem"),
)
.arg(
Arg::with_name("disk")
.long("disk")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Disk subsystem"),
)
.arg(
Arg::with_name("ebpf")
.long("ebpf")
.takes_value(true)
.multiple(true)
.possible_value("all")
.possible_value("block")
.possible_value("ext4")
.possible_value("scheduler")
.possible_value("xfs")
.help("Enable statistics from eBPF"),
)
.arg(
Arg::with_name("network")
.long("network")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Network subsystem"),
)
.arg(
Arg::with_name("perf")
.long("perf")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.possible_value("per-cgroup")
.help("Enable statistics from Perf Events subsystem"),
)
.arg(
Arg::with_name("memcache")
.long("memcache")
.required(false)
.takes_value(true)
.value_name("IP:PORT")
.help("Connect to the given memcache server and produce stats"),
)
.arg(
Arg::with_name("stats-log")
.long("stats-log")
.required(false)
.takes_value(true)
.value_name("LOG FILE")
.help("Enable logging of stats to file"),
)
.arg(
Arg::with_name("sidecar")
.long("sidecar")
.required(false)
.help("Enables Mesos sidecar mode, instrumenting the container"),
)
.get_matches();
let listen = matches
.value_of("listen")
.unwrap()
.parse()
.unwrap_or_else(|_| {
println!("ERROR: listen address is malformed");
process::exit(1);
});
let memcache = if let Some(sock) = matches.value_of("memcache") {
let socket = sock.parse().unwrap_or_else(|_| {
println!("ERROR: memcache address is malformed");
process::exit(1);
});
Some(socket)
} else {
None
};
let sample_rate =
parse_float_arg(&matches, "sample-rate").unwrap_or(DEFAULT_SAMPLE_RATE_HZ);
let sampler_timeout = Duration::from_millis(
parse_numeric_arg(&matches, "sampler-timeout")
.unwrap_or(DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS) as u64,
);
let max_sampler_timeouts = parse_numeric_arg(&matches, "max-sampler-timeouts")
.unwrap_or(DEFAULT_MAX_SAMPLER_TIMEOUTS);
let interval = parse_numeric_arg(&matches, "interval").unwrap_or(DEFAULT_INTERVAL_SECONDS)
as u64
* SECOND;
let cores = hardware_threads().unwrap_or(1);
let mut stats_enabled = Flags::new();
for subsystem in &["cpu", "disk", "ebpf", "network", "perf"] {
if let Some(values) = matches.values_of(subsystem) {
let flags: Vec<&str> = values.collect();
for flag in flags {
stats_enabled.insert(subsystem, flag);
}
}
}
let loglevel = match matches.occurrences_of("verbose") {
0 => Level::Info,
1 => Level::Debug,
_ => Level::Trace,
};
let stats_log = matches
.value_of("stats-log")
.map(std::string::ToString::to_string);
let sidecar = matches.is_present("sidecar");
Config {
cores,
flags: stats_enabled,
sample_rate,
sampler_timeout,
max_sampler_timeouts,
interval,
listen,
loglevel,
memcache,
stats_log,
sidecar,
}
}
/// what interval should the stats library latch on
pub fn interval(&self) -> u64 {
self.interval
}
/// what frequency the stats should be sampled on
pub fn sample_rate(&self) -> f64 |
/// the timeout for sampler execution
pub fn sampler_timeout(&self) -> Duration {
self.sampler_timeout
}
/// maximum consecutive sampler timeouts
pub fn max_sampler_timeouts(&self) -> usize {
self.max_sampler_timeouts
}
/// get listen address
pub fn listen(&self) -> SocketAddr {
self.listen
}
/// get log level
pub fn loglevel(&self) -> Level {
self.loglevel
}
/// how many cores on the host?
pub fn cores(&self) -> usize {
self.cores
}
pub fn memcache(&self) -> Option<SocketAddr> {
self.memcache
}
/// is a flag enabled for a subsystem?
pub fn is_enabled(&self, subsystem: &str, flag: &str) -> bool {
self.flags.contains(subsystem, flag)
}
pub fn is_subsystem_enabled(&self, subsystem: &str) -> bool {
self.flags.contains_pkey(subsystem)
}
pub fn stats_log(&self) -> Option<String> {
self.stats_log.clone()
}
}
/// a helper function to parse a numeric argument by name from `ArgMatches`
fn parse_numeric_arg(matches: &ArgMatches, key: &str) -> Option<usize> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1);
})
})
}
/// a helper function to parse a floating point argument by name from `ArgMatches`
fn parse_float_arg(matches: &ArgMatches, key: &str) -> Option<f64> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1);
})
})
}
/// helper function to discover the number of hardware threads
pub fn hardware_threads() -> Result<usize, ()> {
let path = "/sys/devices/system/cpu/present";
let f = File::open(path).map_err(|e| error!("failed to open file ({:?}): {}", path, e))?;
let mut f = BufReader::new(f);
let mut line = String::new();
f.read_line(&mut line)
.map_err(|_| error!("failed to read line"))?;
let line = line.trim();
let a: Vec<&str> = line.split('-').collect();
a.last()
.unwrap_or(&"0")
.parse::<usize>()
.map_err(|e| error!("could not parse num cpus from file ({:?}): {}", path, e))
.map(|i| i + 1)
}
| {
self.sample_rate
} | identifier_body |
config.rs | // Copyright 2019 Twitter, Inc.
// Licensed under the Apache License, Version 2.0
// http://www.apache.org/licenses/LICENSE-2.0
use crate::common::*;
use clap::{App, Arg, ArgMatches};
use logger::*;
use std::collections::{HashMap, HashSet};
use std::fs::File;
use std::io::{BufRead, BufReader};
use std::net::SocketAddr;
use std::process;
use std::time::Duration;
pub const VERSION: &str = env!("CARGO_PKG_VERSION");
pub const NAME: &str = env!("CARGO_PKG_NAME");
pub const DEFAULT_SAMPLE_RATE_HZ: f64 = 1.0;
pub const DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS: usize = 50;
pub const DEFAULT_MAX_SAMPLER_TIMEOUTS: usize = 5;
pub const DEFAULT_INTERVAL_SECONDS: usize = 60;
/// This struct contains the configuration of the agent.
#[derive(Clone)]
pub struct Config {
/// the latching interval for stats
interval: u64,
/// sample rate for counters in Hz
sample_rate: f64,
/// the sampler timeout
sampler_timeout: Duration,
/// maximum consecutive sampler timeouts
max_sampler_timeouts: usize,
/// the listen address for the stats port
listen: SocketAddr,
/// the logging level
loglevel: Level,
/// memcache instance to instrument
memcache: Option<SocketAddr>,
/// flags for enabled statistics subsystems
flags: Flags,
/// the number of cores on the host
cores: usize,
/// an optional file to log stats to
stats_log: Option<String>,
/// flag to indicate Mesos sidecar mode
sidecar: bool,
}
#[derive(Clone)]
/// `Flags` is a simple wrapper for a doubly-keyed `HashSet`
pub struct Flags {
data: HashMap<String, HashSet<String>>,
}
impl Flags {
/// Creates a new empty set of `Flags`
pub fn new() -> Self {
Self {
data: HashMap::new(),
}
}
/// Insert a `pkey`+`lkey` into the set
pub fn insert(&mut self, pkey: &str, lkey: &str) {
let mut entry = self.data.remove(pkey).unwrap_or_default();
entry.insert(lkey.to_owned());
self.data.insert(pkey.to_owned(), entry);
}
/// True if the set contains `pkey`+`lkey`
pub fn contains(&self, pkey: &str, lkey: &str) -> bool {
if let Some(entry) = self.data.get(pkey) {
entry.get(lkey).is_some()
} else {
false
}
}
/// True if the set contains the `pkey`
pub fn contains_pkey(&self, pkey: &str) -> bool {
self.data.get(pkey).is_some()
}
/// Remove a `pkey`+`lkey`
pub fn remove(&mut self, pkey: &str, lkey: &str) {
if let Some(entry) = self.data.get_mut(pkey) {
entry.remove(lkey);
}
}
/// Remove the `pkey` and all `lkey`s under it
pub fn remove_pkey(&mut self, pkey: &str) {
self.data.remove(pkey);
}
}
impl Config {
/// parse command line options and return `Config`
pub fn new() -> Config {
let matches = App::new(NAME)
.version(VERSION)
.author("Brian Martin <bmartin@twitter.com>")
.about("high-resolution systems performance telemetry agent")
.arg(
Arg::with_name("listen")
.short("l")
.long("listen")
.required(true)
.takes_value(true)
.value_name("IP:PORT")
.help("Sets the listen address for metrics"),
)
.arg(
Arg::with_name("verbose")
.short("v")
.long("verbose")
.multiple(true)
.help("Increase verbosity by one level. Can be used more than once"),
)
.arg(
Arg::with_name("interval")
.long("interval")
.value_name("Seconds")
.help("Integration window duration and stats endpoint refresh time")
.takes_value(true),
)
.arg(
Arg::with_name("sample-rate")
.long("sample-rate")
.value_name("Hertz")
.help("Sets the sampling frequency for the counters")
.takes_value(true),
)
.arg(
Arg::with_name("sampler-timeout")
.long("sampler-timeout")
.value_name("MS")
.help("Sets the timeout for per-sampler execution")
.takes_value(true),
)
.arg(
Arg::with_name("max-sampler-timeouts")
.long("max-sampler-timeouts")
.value_name("MS")
.help("Sets the maximum number of consecutive sampler timeouts")
.takes_value(true),
)
.arg(
Arg::with_name("cpu")
.long("cpu")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from CPU subsystem"),
)
.arg(
Arg::with_name("disk")
.long("disk")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Disk subsystem"),
)
.arg(
Arg::with_name("ebpf")
.long("ebpf")
.takes_value(true)
.multiple(true)
.possible_value("all")
.possible_value("block")
.possible_value("ext4")
.possible_value("scheduler")
.possible_value("xfs")
.help("Enable statistics from eBPF"),
)
.arg(
Arg::with_name("network")
.long("network")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Network subsystem"),
)
.arg(
Arg::with_name("perf")
.long("perf")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.possible_value("per-cgroup")
.help("Enable statistics from Perf Events subsystem"),
)
.arg(
Arg::with_name("memcache")
.long("memcache")
.required(false)
.takes_value(true)
.value_name("IP:PORT")
.help("Connect to the given memcache server and produce stats"),
)
.arg(
Arg::with_name("stats-log")
.long("stats-log")
.required(false)
.takes_value(true)
.value_name("LOG FILE")
.help("Enable logging of stats to file"),
)
.arg(
Arg::with_name("sidecar")
.long("sidecar")
.required(false)
.help("Enables Mesos sidecar mode, instrumenting the container"),
)
.get_matches();
let listen = matches
.value_of("listen")
.unwrap()
.parse()
.unwrap_or_else(|_| {
println!("ERROR: listen address is malformed");
process::exit(1);
});
let memcache = if let Some(sock) = matches.value_of("memcache") | else {
None
};
let sample_rate =
parse_float_arg(&matches, "sample-rate").unwrap_or(DEFAULT_SAMPLE_RATE_HZ);
let sampler_timeout = Duration::from_millis(
parse_numeric_arg(&matches, "sampler-timeout")
.unwrap_or(DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS) as u64,
);
let max_sampler_timeouts = parse_numeric_arg(&matches, "max-sampler-timeouts")
.unwrap_or(DEFAULT_MAX_SAMPLER_TIMEOUTS);
let interval = parse_numeric_arg(&matches, "interval").unwrap_or(DEFAULT_INTERVAL_SECONDS)
as u64
* SECOND;
let cores = hardware_threads().unwrap_or(1);
let mut stats_enabled = Flags::new();
for subsystem in &["cpu", "disk", "ebpf", "network", "perf"] {
if let Some(values) = matches.values_of(subsystem) {
let flags: Vec<&str> = values.collect();
for flag in flags {
stats_enabled.insert(subsystem, flag);
}
}
}
let loglevel = match matches.occurrences_of("verbose") {
0 => Level::Info,
1 => Level::Debug,
_ => Level::Trace,
};
let stats_log = matches
.value_of("stats-log")
.map(std::string::ToString::to_string);
let sidecar = matches.is_present("sidecar");
Config {
cores,
flags: stats_enabled,
sample_rate,
sampler_timeout,
max_sampler_timeouts,
interval,
listen,
loglevel,
memcache,
stats_log,
sidecar,
}
}
/// what interval should the stats library latch on
pub fn interval(&self) -> u64 {
self.interval
}
/// what frequency the stats should be sampled on
pub fn sample_rate(&self) -> f64 {
self.sample_rate
}
/// the timeout for sampler execution
pub fn sampler_timeout(&self) -> Duration {
self.sampler_timeout
}
/// maximum consecutive sampler timeouts
pub fn max_sampler_timeouts(&self) -> usize {
self.max_sampler_timeouts
}
/// get listen address
pub fn listen(&self) -> SocketAddr {
self.listen
}
/// get log level
pub fn loglevel(&self) -> Level {
self.loglevel
}
/// how many cores on the host?
pub fn cores(&self) -> usize {
self.cores
}
pub fn memcache(&self) -> Option<SocketAddr> {
self.memcache
}
/// is a flag enabled for a subsystem?
pub fn is_enabled(&self, subsystem: &str, flag: &str) -> bool {
self.flags.contains(subsystem, flag)
}
pub fn is_subsystem_enabled(&self, subsystem: &str) -> bool {
self.flags.contains_pkey(subsystem)
}
pub fn stats_log(&self) -> Option<String> {
self.stats_log.clone()
}
}
/// a helper function to parse a numeric argument by name from `ArgMatches`
fn parse_numeric_arg(matches: &ArgMatches, key: &str) -> Option<usize> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1);
})
})
}
/// a helper function to parse a floating point argument by name from `ArgMatches`
fn parse_float_arg(matches: &ArgMatches, key: &str) -> Option<f64> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1);
})
})
}
/// helper function to discover the number of hardware threads
pub fn hardware_threads() -> Result<usize, ()> {
let path = "/sys/devices/system/cpu/present";
let f = File::open(path).map_err(|e| error!("failed to open file ({:?}): {}", path, e))?;
let mut f = BufReader::new(f);
let mut line = String::new();
f.read_line(&mut line)
.map_err(|_| error!("failed to read line"))?;
let line = line.trim();
let a: Vec<&str> = line.split('-').collect();
a.last()
.unwrap_or(&"0")
.parse::<usize>()
.map_err(|e| error!("could not parse num cpus from file ({:?}): {}", path, e))
.map(|i| i + 1)
}
| {
let socket = sock.parse().unwrap_or_else(|_| {
println!("ERROR: memcache address is malformed");
process::exit(1);
});
Some(socket)
} | conditional_block |
config.rs | // Copyright 2019 Twitter, Inc.
// Licensed under the Apache License, Version 2.0
// http://www.apache.org/licenses/LICENSE-2.0
use crate::common::*;
use clap::{App, Arg, ArgMatches};
use logger::*;
use std::collections::{HashMap, HashSet};
use std::fs::File;
use std::io::{BufRead, BufReader};
use std::net::SocketAddr;
use std::process;
use std::time::Duration;
pub const VERSION: &str = env!("CARGO_PKG_VERSION");
pub const NAME: &str = env!("CARGO_PKG_NAME");
pub const DEFAULT_SAMPLE_RATE_HZ: f64 = 1.0;
pub const DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS: usize = 50;
pub const DEFAULT_MAX_SAMPLER_TIMEOUTS: usize = 5;
pub const DEFAULT_INTERVAL_SECONDS: usize = 60;
/// This struct contains the configuration of the agent.
#[derive(Clone)]
pub struct Config {
/// the latching interval for stats
interval: u64,
/// sample rate for counters in Hz
sample_rate: f64,
/// the sampler timeout
sampler_timeout: Duration,
/// maximum consecutive sampler timeouts
max_sampler_timeouts: usize,
/// the listen address for the stats port
listen: SocketAddr,
/// the logging level
loglevel: Level,
/// memcache instance to instrument
memcache: Option<SocketAddr>,
/// flags for enabled statistics subsystems
flags: Flags,
/// the number of cores on the host
cores: usize,
/// an optional file to log stats to
stats_log: Option<String>,
/// flag to indicate Mesos sidecar mode
sidecar: bool,
}
#[derive(Clone)]
/// `Flags` is a simple wrapper for a doubly-keyed `HashSet`
pub struct Flags {
data: HashMap<String, HashSet<String>>,
}
impl Flags {
/// Creates a new empty set of `Flags`
pub fn new() -> Self {
Self {
data: HashMap::new(),
}
}
/// Insert a `pkey`+`lkey` into the set
pub fn insert(&mut self, pkey: &str, lkey: &str) {
let mut entry = self.data.remove(pkey).unwrap_or_default();
entry.insert(lkey.to_owned());
self.data.insert(pkey.to_owned(), entry);
}
/// True if the set contains `pkey`+`lkey`
pub fn contains(&self, pkey: &str, lkey: &str) -> bool {
if let Some(entry) = self.data.get(pkey) {
entry.get(lkey).is_some()
} else {
false
}
}
/// True if the set contains the `pkey`
pub fn contains_pkey(&self, pkey: &str) -> bool {
self.data.get(pkey).is_some()
}
/// Remove a `pkey`+`lkey`
pub fn remove(&mut self, pkey: &str, lkey: &str) {
if let Some(entry) = self.data.get_mut(pkey) {
entry.remove(lkey);
}
}
/// Remove the `pkey` and all `lkey`s under it
pub fn remove_pkey(&mut self, pkey: &str) {
self.data.remove(pkey);
}
}
impl Config {
/// parse command line options and return `Config`
pub fn new() -> Config {
let matches = App::new(NAME)
.version(VERSION)
.author("Brian Martin <bmartin@twitter.com>")
.about("high-resolution systems performance telemetry agent")
.arg(
Arg::with_name("listen")
.short("l")
.long("listen")
.required(true)
.takes_value(true)
.value_name("IP:PORT")
.help("Sets the listen address for metrics"),
)
.arg(
Arg::with_name("verbose")
.short("v")
.long("verbose")
.multiple(true)
.help("Increase verbosity by one level. Can be used more than once"),
)
.arg(
Arg::with_name("interval")
.long("interval")
.value_name("Seconds")
.help("Integration window duration and stats endpoint refresh time")
.takes_value(true),
)
.arg(
Arg::with_name("sample-rate")
.long("sample-rate")
.value_name("Hertz")
.help("Sets the sampling frequency for the counters")
.takes_value(true),
)
.arg(
Arg::with_name("sampler-timeout")
.long("sampler-timeout")
.value_name("MS")
.help("Sets the timeout for per-sampler execution")
.takes_value(true),
)
.arg(
Arg::with_name("max-sampler-timeouts")
.long("max-sampler-timeouts")
.value_name("MS")
.help("Sets the maximum number of consecutive sampler timeouts")
.takes_value(true),
)
.arg(
Arg::with_name("cpu")
.long("cpu")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from CPU subsystem"),
)
.arg(
Arg::with_name("disk")
.long("disk")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Disk subsystem"),
)
.arg(
Arg::with_name("ebpf")
.long("ebpf")
.takes_value(true)
.multiple(true)
.possible_value("all")
.possible_value("block")
.possible_value("ext4")
.possible_value("scheduler")
.possible_value("xfs")
.help("Enable statistics from eBPF"),
)
.arg(
Arg::with_name("network")
.long("network")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.help("Enable statistics from Network subsystem"),
)
.arg(
Arg::with_name("perf")
.long("perf")
.takes_value(true)
.multiple(true)
.possible_value("totals")
.possible_value("per-cgroup")
.help("Enable statistics from Perf Events subsystem"),
)
.arg(
Arg::with_name("memcache")
.long("memcache")
.required(false)
.takes_value(true)
.value_name("IP:PORT")
.help("Connect to the given memcache server and produce stats"),
)
.arg(
Arg::with_name("stats-log")
.long("stats-log")
.required(false)
.takes_value(true)
.value_name("LOG FILE")
.help("Enable logging of stats to file"),
)
.arg(
Arg::with_name("sidecar")
.long("sidecar")
.required(false)
.help("Enables Mesos sidecar mode, instrumenting the container"),
)
.get_matches();
let listen = matches
.value_of("listen")
.unwrap()
.parse()
.unwrap_or_else(|_| {
println!("ERROR: listen address is malformed");
process::exit(1);
});
let memcache = if let Some(sock) = matches.value_of("memcache") {
let socket = sock.parse().unwrap_or_else(|_| {
println!("ERROR: memcache address is malformed");
process::exit(1);
});
Some(socket)
} else {
None
};
let sample_rate =
parse_float_arg(&matches, "sample-rate").unwrap_or(DEFAULT_SAMPLE_RATE_HZ);
let sampler_timeout = Duration::from_millis(
parse_numeric_arg(&matches, "sampler-timeout")
.unwrap_or(DEFAULT_SAMPLER_TIMEOUT_MILLISECONDS) as u64,
);
let max_sampler_timeouts = parse_numeric_arg(&matches, "max-sampler-timeouts")
.unwrap_or(DEFAULT_MAX_SAMPLER_TIMEOUTS);
let interval = parse_numeric_arg(&matches, "interval").unwrap_or(DEFAULT_INTERVAL_SECONDS)
as u64
* SECOND;
let cores = hardware_threads().unwrap_or(1);
let mut stats_enabled = Flags::new();
for subsystem in &["cpu", "disk", "ebpf", "network", "perf"] {
if let Some(values) = matches.values_of(subsystem) {
let flags: Vec<&str> = values.collect();
for flag in flags {
stats_enabled.insert(subsystem, flag);
}
}
}
let loglevel = match matches.occurrences_of("verbose") {
0 => Level::Info,
1 => Level::Debug,
_ => Level::Trace,
};
let stats_log = matches
.value_of("stats-log")
.map(std::string::ToString::to_string);
let sidecar = matches.is_present("sidecar");
Config {
cores,
flags: stats_enabled,
sample_rate,
sampler_timeout,
max_sampler_timeouts,
interval,
listen,
loglevel,
memcache,
stats_log,
sidecar,
}
}
/// what interval should the stats library latch on
pub fn interval(&self) -> u64 {
self.interval
}
/// what frequency the stats should be sampled on
pub fn sample_rate(&self) -> f64 {
self.sample_rate
}
/// the timeout for sampler execution
pub fn sampler_timeout(&self) -> Duration {
self.sampler_timeout
}
/// maximum consecutive sampler timeouts
pub fn max_sampler_timeouts(&self) -> usize {
self.max_sampler_timeouts
}
/// get listen address
pub fn listen(&self) -> SocketAddr {
self.listen
}
/// get log level
pub fn loglevel(&self) -> Level {
self.loglevel
}
/// how many cores on the host?
pub fn cores(&self) -> usize {
self.cores
}
pub fn memcache(&self) -> Option<SocketAddr> {
self.memcache
}
/// is a flag enabled for a subsystem?
pub fn is_enabled(&self, subsystem: &str, flag: &str) -> bool {
self.flags.contains(subsystem, flag)
}
pub fn is_subsystem_enabled(&self, subsystem: &str) -> bool {
self.flags.contains_pkey(subsystem)
}
pub fn stats_log(&self) -> Option<String> {
self.stats_log.clone()
}
}
/// a helper function to parse a numeric argument by name from `ArgMatches`
fn parse_numeric_arg(matches: &ArgMatches, key: &str) -> Option<usize> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1);
})
})
}
/// a helper function to parse a floating point argument by name from `ArgMatches`
fn parse_float_arg(matches: &ArgMatches, key: &str) -> Option<f64> {
matches.value_of(key).map(|f| {
f.parse().unwrap_or_else(|_| {
println!("ERROR: could not parse {}", key);
process::exit(1); | }
/// helper function to discover the number of hardware threads
pub fn hardware_threads() -> Result<usize, ()> {
let path = "/sys/devices/system/cpu/present";
let f = File::open(path).map_err(|e| error!("failed to open file ({:?}): {}", path, e))?;
let mut f = BufReader::new(f);
let mut line = String::new();
f.read_line(&mut line)
.map_err(|_| error!("failed to read line"))?;
let line = line.trim();
let a: Vec<&str> = line.split('-').collect();
a.last()
.unwrap_or(&"0")
.parse::<usize>()
.map_err(|e| error!("could not parse num cpus from file ({:?}): {}", path, e))
.map(|i| i + 1)
} | })
}) | random_line_split |
tag.rs | //! Utilities and constants for OpenType tags.
//!
//! See also the [`tag!`](../macro.tag.html) macro for creating tags from a byte string.
use crate::error::ParseError;
use std::{fmt, str};
/// Generate a 4-byte OpenType tag from byte string
///
/// Example:
///
/// ```
/// use allsorts::tag;
/// assert_eq!(tag!(b"glyf"), 0x676C7966);
/// ```
#[macro_export]
macro_rules! tag {
($w:expr) => {
u32::from_be_bytes(*$w)
};
}
/// Wrapper type for a tag that implements `Display`
///
/// Example:
///
/// ```
/// use allsorts::tag::{self, DisplayTag};
///
/// // ASCII tag comes out as a string
/// assert_eq!(&DisplayTag(tag::NAME).to_string(), "name");
/// // Non-ASCII tag comes out as hex
/// assert_eq!(&DisplayTag(0x12345678).to_string(), "0x12345678");
///
/// println!("DisplayTag is handy for printing a tag: '{}'", DisplayTag(tag::CFF));
/// ```
#[derive(PartialEq, Eq, Clone, Copy)]
pub struct DisplayTag(pub u32);
pub fn from_string(s: &str) -> Result<u32, ParseError> {
if s.len() > 4 {
return Err(ParseError::BadValue);
}
let mut tag: u32 = 0;
let mut count = 0;
for c in s.chars() {
if !c.is_ascii() || c.is_ascii_control() {
return Err(ParseError::BadValue);
}
tag = (tag << 8) | (c as u32);
count += 1;
}
while count < 4 {
tag = (tag << 8) | (' ' as u32);
count += 1;
}
Ok(tag)
}
impl fmt::Display for DisplayTag {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
let tag = self.0;
let bytes = tag.to_be_bytes();
if bytes.iter().all(|c| c.is_ascii() && !c.is_ascii_control()) {
let s = str::from_utf8(&bytes).unwrap(); // unwrap safe due to above check
s.fmt(f)
} else {
write!(f, "0x{:08x}", tag)
}
}
}
impl fmt::Debug for DisplayTag {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
fmt::Display::fmt(self, f)
}
}
/// `abvf`
pub const ABVF: u32 = tag!(b"abvf");
/// `abvm`
pub const ABVM: u32 = tag!(b"abvm");
/// `abvs`
pub const ABVS: u32 = tag!(b"abvs");
/// `acnt`
pub const ACNT: u32 = tag!(b"acnt");
/// `afrc`
pub const AFRC: u32 = tag!(b"afrc");
/// `akhn`
pub const AKHN: u32 = tag!(b"akhn");
/// `arab`
pub const ARAB: u32 = tag!(b"arab");
/// `avar`
pub const AVAR: u32 = tag!(b"avar");
/// `BASE`
pub const BASE: u32 = tag!(b"BASE");
/// `bdat`
pub const BDAT: u32 = tag!(b"bdat");
/// `beng`
pub const BENG: u32 = tag!(b"beng");
/// `bloc`
pub const BLOC: u32 = tag!(b"bloc");
/// `blwf`
pub const BLWF: u32 = tag!(b"blwf");
/// `blwm`
pub const BLWM: u32 = tag!(b"blwm");
/// `blws`
pub const BLWS: u32 = tag!(b"blws");
/// `bng2`
pub const BNG2: u32 = tag!(b"bng2");
/// `bsln`
pub const BSLN: u32 = tag!(b"bsln");
/// `c2sc`
pub const C2SC: u32 = tag!(b"c2sc");
/// `calt`
pub const CALT: u32 = tag!(b"calt");
/// `CBDT`
pub const CBDT: u32 = tag!(b"CBDT");
/// `CBLC`
pub const CBLC: u32 = tag!(b"CBLC");
/// `ccmp`
pub const CCMP: u32 = tag!(b"ccmp");
/// `cfar`
pub const CFAR: u32 = tag!(b"cfar");
/// `CFF `
pub const CFF: u32 = tag!(b"CFF ");
/// `cjct`
pub const CJCT: u32 = tag!(b"cjct");
/// `clig`
pub const CLIG: u32 = tag!(b"clig");
/// `cmap`
pub const CMAP: u32 = tag!(b"cmap");
/// `COLR`
pub const COLR: u32 = tag!(b"COLR");
/// `CPAL`
pub const CPAL: u32 = tag!(b"CPAL");
/// `curs`
pub const CURS: u32 = tag!(b"curs");
/// `cvar`
pub const CVAR: u32 = tag!(b"cvar");
/// `cvt `
pub const CVT: u32 = tag!(b"cvt ");
/// `cyrl`
pub const CYRL: u32 = tag!(b"cyrl");
/// `dev2`
pub const DEV2: u32 = tag!(b"dev2");
/// `deva`
pub const DEVA: u32 = tag!(b"deva");
/// `DFLT`
pub const DFLT: u32 = tag!(b"DFLT");
/// `dist`
pub const DIST: u32 = tag!(b"dist");
/// `dlig`
pub const DLIG: u32 = tag!(b"dlig");
/// `dupe`
pub const DUPE: u32 = tag!(b"dupe");
/// `EBDT`
pub const EBDT: u32 = tag!(b"EBDT");
/// `EBLC`
pub const EBLC: u32 = tag!(b"EBLC");
/// `EBSC`
pub const EBSC: u32 = tag!(b"EBSC");
/// `FAR `
pub const FAR: u32 = tag!(b"FAR ");
/// `fdsc`
pub const FDSC: u32 = tag!(b"fdsc");
/// `Feat`
pub const FEAT2: u32 = tag!(b"Feat");
/// `feat`
pub const FEAT: u32 = tag!(b"feat");
/// `fin2`
pub const FIN2: u32 = tag!(b"fin2");
/// `fin3`
pub const FIN3: u32 = tag!(b"fin3");
/// `fina`
pub const FINA: u32 = tag!(b"fina");
/// `fmtx`
pub const FMTX: u32 = tag!(b"fmtx");
/// `fpgm`
pub const FPGM: u32 = tag!(b"fpgm");
/// `frac`
pub const FRAC: u32 = tag!(b"frac");
/// `fvar`
pub const FVAR: u32 = tag!(b"fvar");
/// `gasp`
pub const GASP: u32 = tag!(b"gasp");
/// `GDEF`
pub const GDEF: u32 = tag!(b"GDEF");
/// `gjr2`
pub const GJR2: u32 = tag!(b"gjr2");
/// `Glat`
pub const GLAT: u32 = tag!(b"Glat");
/// `Gloc`
pub const GLOC: u32 = tag!(b"Gloc");
/// `glyf`
pub const GLYF: u32 = tag!(b"glyf");
/// `GPOS`
pub const GPOS: u32 = tag!(b"GPOS");
/// `grek`
pub const GREK: u32 = tag!(b"grek");
/// `GSUB`
pub const GSUB: u32 = tag!(b"GSUB");
/// `gujr`
pub const GUJR: u32 = tag!(b"gujr");
/// `gur2`
pub const GUR2: u32 = tag!(b"gur2");
/// `guru`
pub const GURU: u32 = tag!(b"guru");
/// `gvar`
pub const GVAR: u32 = tag!(b"gvar");
/// `half`
pub const HALF: u32 = tag!(b"half");
/// `haln`
pub const HALN: u32 = tag!(b"haln");
/// `hdmx`
pub const HDMX: u32 = tag!(b"hdmx");
/// `head`
pub const HEAD: u32 = tag!(b"head");
/// `hhea`
pub const HHEA: u32 = tag!(b"hhea");
/// `hlig`
pub const HLIG: u32 = tag!(b"hlig");
/// `hmtx`
pub const HMTX: u32 = tag!(b"hmtx");
/// `hsty`
pub const HSTY: u32 = tag!(b"hsty");
/// `init`
pub const INIT: u32 = tag!(b"init");
/// `isol`
pub const ISOL: u32 = tag!(b"isol");
/// `jpg `
pub const JPG: u32 = tag!(b"jpg ");
/// `JSTF`
pub const JSTF: u32 = tag!(b"JSTF");
/// `just`
pub const JUST: u32 = tag!(b"just");
/// `kern`
pub const KERN: u32 = tag!(b"kern");
/// `khmr`
pub const KHMR: u32 = tag!(b"khmr");
/// `knd2`
pub const KND2: u32 = tag!(b"knd2");
/// `knda`
pub const KNDA: u32 = tag!(b"knda");
/// `lao `
pub const LAO: u32 = tag!(b"lao ");
/// `latn`
pub const LATN: u32 = tag!(b"latn");
/// `lcar`
pub const LCAR: u32 = tag!(b"lcar");
/// `liga`
pub const LIGA: u32 = tag!(b"liga");
/// `lnum`
pub const LNUM: u32 = tag!(b"lnum");
/// `loca`
pub const LOCA: u32 = tag!(b"loca");
/// `locl`
pub const LOCL: u32 = tag!(b"locl");
/// `LTSH`
pub const LTSH: u32 = tag!(b"LTSH");
/// `mark`
pub const MARK: u32 = tag!(b"mark");
/// `MATH`
pub const MATH: u32 = tag!(b"MATH");
/// `maxp`
pub const MAXP: u32 = tag!(b"maxp");
/// `med2`
pub const MED2: u32 = tag!(b"med2");
/// `medi`
pub const MEDI: u32 = tag!(b"medi");
/// `mkmk`
pub const MKMK: u32 = tag!(b"mkmk");
/// `mlm2`
pub const MLM2: u32 = tag!(b"mlm2");
/// `mlym`
pub const MLYM: u32 = tag!(b"mlym");
/// `mort`
pub const MORT: u32 = tag!(b"mort");
/// `morx`
pub const MORX: u32 = tag!(b"morx");
/// `mset`
pub const MSET: u32 = tag!(b"mset");
/// `name`
pub const NAME: u32 = tag!(b"name");
/// `nukt`
pub const NUKT: u32 = tag!(b"nukt");
/// `onum`
pub const ONUM: u32 = tag!(b"onum");
/// `opbd`
pub const OPBD: u32 = tag!(b"opbd");
/// `ordn`
pub const ORDN: u32 = tag!(b"ordn");
/// `ory2`
pub const ORY2: u32 = tag!(b"ory2");
/// `orya`
pub const ORYA: u32 = tag!(b"orya");
/// `OS/2`
pub const OS_2: u32 = tag!(b"OS/2");
/// `OTTO`
pub const OTTO: u32 = tag!(b"OTTO");
/// `PCLT`
pub const PCLT: u32 = tag!(b"PCLT");
/// `pnum`
pub const PNUM: u32 = tag!(b"pnum");
/// `png `
pub const PNG: u32 = tag!(b"png ");
/// `post`
pub const POST: u32 = tag!(b"post");
/// `pref`
pub const PREF: u32 = tag!(b"pref");
/// `prep`
pub const PREP: u32 = tag!(b"prep");
/// `pres`
pub const PRES: u32 = tag!(b"pres");
/// `prop`
pub const PROP: u32 = tag!(b"prop");
/// `pstf`
pub const PSTF: u32 = tag!(b"pstf");
/// `psts`
pub const PSTS: u32 = tag!(b"psts");
/// `rclt`
pub const RCLT: u32 = tag!(b"rclt");
/// `rkrf`
pub const RKRF: u32 = tag!(b"rkrf");
/// `rlig`
pub const RLIG: u32 = tag!(b"rlig");
/// `rphf`
pub const RPHF: u32 = tag!(b"rphf");
/// `sbix`
pub const SBIX: u32 = tag!(b"sbix");
/// `Silf`
pub const SILF: u32 = tag!(b"Silf");
/// `Sill`
pub const SILL: u32 = tag!(b"Sill");
/// `sinh`
pub const SINH: u32 = tag!(b"sinh");
/// `smcp`
pub const SMCP: u32 = tag!(b"smcp");
/// `SND `
pub const SND: u32 = tag!(b"SND ");
/// `SVG `
pub const SVG: u32 = tag!(b"SVG ");
/// `syrc`
pub const SYRC: u32 = tag!(b"syrc");
/// `taml`
pub const TAML: u32 = tag!(b"taml");
/// `tel2`
pub const TEL2: u32 = tag!(b"tel2");
/// `telu`
pub const TELU: u32 = tag!(b"telu");
/// `thai`
pub const THAI: u32 = tag!(b"thai");
/// `tiff`
pub const TIFF: u32 = tag!(b"tiff");
/// `tml2`
pub const TML2: u32 = tag!(b"tml2");
/// `tnum`
pub const TNUM: u32 = tag!(b"tnum");
/// `trak`
pub const TRAK: u32 = tag!(b"trak");
/// `ttcf`
pub const TTCF: u32 = tag!(b"ttcf");
/// `URD `
pub const URD: u32 = tag!(b"URD ");
/// `vatu`
pub const VATU: u32 = tag!(b"vatu");
/// `VDMX`
pub const VDMX: u32 = tag!(b"VDMX");
/// `vert`
pub const VERT: u32 = tag!(b"vert");
/// `vhea`
pub const VHEA: u32 = tag!(b"vhea");
/// `vmtx`
pub const VMTX: u32 = tag!(b"vmtx");
/// `VORG`
pub const VORG: u32 = tag!(b"VORG");
/// `vrt2`
pub const VRT2: u32 = tag!(b"vrt2");
/// `Zapf`
pub const ZAPF: u32 = tag!(b"Zapf");
/// `zero`
pub const ZERO: u32 = tag!(b"zero");
#[cfg(test)]
mod tests {
use super::*;
mod from_string {
use super::*;
#[test]
fn test_four_chars() |
#[test]
fn test_three_chars() {
let tag = from_string("BEN").expect("invalid tag");
assert_eq!(tag, 1111838240);
}
}
mod display_tag {
use crate::tag::{DisplayTag, NAME};
#[test]
fn test_simple_ascii() {
assert_eq!(DisplayTag(NAME).to_string(), "name".to_string());
}
#[test]
fn test_non_ascii() {
assert_eq!(DisplayTag(0x12345678).to_string(), "0x12345678".to_string());
}
#[test]
fn test_debug() {
assert_eq!(format!("{:?}", DisplayTag(NAME)), "name".to_string());
}
}
}
| {
let tag = from_string("beng").expect("invalid tag");
assert_eq!(tag, 1650814567);
} | identifier_body |
tag.rs | //! Utilities and constants for OpenType tags.
//!
//! See also the [`tag!`](../macro.tag.html) macro for creating tags from a byte string.
use crate::error::ParseError;
use std::{fmt, str};
/// Generate a 4-byte OpenType tag from byte string
///
/// Example:
///
/// ```
/// use allsorts::tag;
/// assert_eq!(tag!(b"glyf"), 0x676C7966);
/// ```
#[macro_export]
macro_rules! tag {
($w:expr) => {
u32::from_be_bytes(*$w)
};
}
/// Wrapper type for a tag that implements `Display`
///
/// Example:
///
/// ```
/// use allsorts::tag::{self, DisplayTag};
///
/// // ASCII tag comes out as a string
/// assert_eq!(&DisplayTag(tag::NAME).to_string(), "name");
/// // Non-ASCII tag comes out as hex
/// assert_eq!(&DisplayTag(0x12345678).to_string(), "0x12345678");
///
/// println!("DisplayTag is handy for printing a tag: '{}'", DisplayTag(tag::CFF));
/// ```
#[derive(PartialEq, Eq, Clone, Copy)]
pub struct DisplayTag(pub u32);
pub fn from_string(s: &str) -> Result<u32, ParseError> {
if s.len() > 4 {
return Err(ParseError::BadValue);
}
let mut tag: u32 = 0;
let mut count = 0;
for c in s.chars() {
if !c.is_ascii() || c.is_ascii_control() {
return Err(ParseError::BadValue);
}
tag = (tag << 8) | (c as u32);
count += 1;
}
while count < 4 {
tag = (tag << 8) | (' ' as u32);
count += 1;
}
Ok(tag)
}
impl fmt::Display for DisplayTag {
fn | (&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
let tag = self.0;
let bytes = tag.to_be_bytes();
if bytes.iter().all(|c| c.is_ascii() && !c.is_ascii_control()) {
let s = str::from_utf8(&bytes).unwrap(); // unwrap safe due to above check
s.fmt(f)
} else {
write!(f, "0x{:08x}", tag)
}
}
}
impl fmt::Debug for DisplayTag {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
fmt::Display::fmt(self, f)
}
}
/// `abvf`
pub const ABVF: u32 = tag!(b"abvf");
/// `abvm`
pub const ABVM: u32 = tag!(b"abvm");
/// `abvs`
pub const ABVS: u32 = tag!(b"abvs");
/// `acnt`
pub const ACNT: u32 = tag!(b"acnt");
/// `afrc`
pub const AFRC: u32 = tag!(b"afrc");
/// `akhn`
pub const AKHN: u32 = tag!(b"akhn");
/// `arab`
pub const ARAB: u32 = tag!(b"arab");
/// `avar`
pub const AVAR: u32 = tag!(b"avar");
/// `BASE`
pub const BASE: u32 = tag!(b"BASE");
/// `bdat`
pub const BDAT: u32 = tag!(b"bdat");
/// `beng`
pub const BENG: u32 = tag!(b"beng");
/// `bloc`
pub const BLOC: u32 = tag!(b"bloc");
/// `blwf`
pub const BLWF: u32 = tag!(b"blwf");
/// `blwm`
pub const BLWM: u32 = tag!(b"blwm");
/// `blws`
pub const BLWS: u32 = tag!(b"blws");
/// `bng2`
pub const BNG2: u32 = tag!(b"bng2");
/// `bsln`
pub const BSLN: u32 = tag!(b"bsln");
/// `c2sc`
pub const C2SC: u32 = tag!(b"c2sc");
/// `calt`
pub const CALT: u32 = tag!(b"calt");
/// `CBDT`
pub const CBDT: u32 = tag!(b"CBDT");
/// `CBLC`
pub const CBLC: u32 = tag!(b"CBLC");
/// `ccmp`
pub const CCMP: u32 = tag!(b"ccmp");
/// `cfar`
pub const CFAR: u32 = tag!(b"cfar");
/// `CFF `
pub const CFF: u32 = tag!(b"CFF ");
/// `cjct`
pub const CJCT: u32 = tag!(b"cjct");
/// `clig`
pub const CLIG: u32 = tag!(b"clig");
/// `cmap`
pub const CMAP: u32 = tag!(b"cmap");
/// `COLR`
pub const COLR: u32 = tag!(b"COLR");
/// `CPAL`
pub const CPAL: u32 = tag!(b"CPAL");
/// `curs`
pub const CURS: u32 = tag!(b"curs");
/// `cvar`
pub const CVAR: u32 = tag!(b"cvar");
/// `cvt `
pub const CVT: u32 = tag!(b"cvt ");
/// `cyrl`
pub const CYRL: u32 = tag!(b"cyrl");
/// `dev2`
pub const DEV2: u32 = tag!(b"dev2");
/// `deva`
pub const DEVA: u32 = tag!(b"deva");
/// `DFLT`
pub const DFLT: u32 = tag!(b"DFLT");
/// `dist`
pub const DIST: u32 = tag!(b"dist");
/// `dlig`
pub const DLIG: u32 = tag!(b"dlig");
/// `dupe`
pub const DUPE: u32 = tag!(b"dupe");
/// `EBDT`
pub const EBDT: u32 = tag!(b"EBDT");
/// `EBLC`
pub const EBLC: u32 = tag!(b"EBLC");
/// `EBSC`
pub const EBSC: u32 = tag!(b"EBSC");
/// `FAR `
pub const FAR: u32 = tag!(b"FAR ");
/// `fdsc`
pub const FDSC: u32 = tag!(b"fdsc");
/// `Feat`
pub const FEAT2: u32 = tag!(b"Feat");
/// `feat`
pub const FEAT: u32 = tag!(b"feat");
/// `fin2`
pub const FIN2: u32 = tag!(b"fin2");
/// `fin3`
pub const FIN3: u32 = tag!(b"fin3");
/// `fina`
pub const FINA: u32 = tag!(b"fina");
/// `fmtx`
pub const FMTX: u32 = tag!(b"fmtx");
/// `fpgm`
pub const FPGM: u32 = tag!(b"fpgm");
/// `frac`
pub const FRAC: u32 = tag!(b"frac");
/// `fvar`
pub const FVAR: u32 = tag!(b"fvar");
/// `gasp`
pub const GASP: u32 = tag!(b"gasp");
/// `GDEF`
pub const GDEF: u32 = tag!(b"GDEF");
/// `gjr2`
pub const GJR2: u32 = tag!(b"gjr2");
/// `Glat`
pub const GLAT: u32 = tag!(b"Glat");
/// `Gloc`
pub const GLOC: u32 = tag!(b"Gloc");
/// `glyf`
pub const GLYF: u32 = tag!(b"glyf");
/// `GPOS`
pub const GPOS: u32 = tag!(b"GPOS");
/// `grek`
pub const GREK: u32 = tag!(b"grek");
/// `GSUB`
pub const GSUB: u32 = tag!(b"GSUB");
/// `gujr`
pub const GUJR: u32 = tag!(b"gujr");
/// `gur2`
pub const GUR2: u32 = tag!(b"gur2");
/// `guru`
pub const GURU: u32 = tag!(b"guru");
/// `gvar`
pub const GVAR: u32 = tag!(b"gvar");
/// `half`
pub const HALF: u32 = tag!(b"half");
/// `haln`
pub const HALN: u32 = tag!(b"haln");
/// `hdmx`
pub const HDMX: u32 = tag!(b"hdmx");
/// `head`
pub const HEAD: u32 = tag!(b"head");
/// `hhea`
pub const HHEA: u32 = tag!(b"hhea");
/// `hlig`
pub const HLIG: u32 = tag!(b"hlig");
/// `hmtx`
pub const HMTX: u32 = tag!(b"hmtx");
/// `hsty`
pub const HSTY: u32 = tag!(b"hsty");
/// `init`
pub const INIT: u32 = tag!(b"init");
/// `isol`
pub const ISOL: u32 = tag!(b"isol");
/// `jpg `
pub const JPG: u32 = tag!(b"jpg ");
/// `JSTF`
pub const JSTF: u32 = tag!(b"JSTF");
/// `just`
pub const JUST: u32 = tag!(b"just");
/// `kern`
pub const KERN: u32 = tag!(b"kern");
/// `khmr`
pub const KHMR: u32 = tag!(b"khmr");
/// `knd2`
pub const KND2: u32 = tag!(b"knd2");
/// `knda`
pub const KNDA: u32 = tag!(b"knda");
/// `lao `
pub const LAO: u32 = tag!(b"lao ");
/// `latn`
pub const LATN: u32 = tag!(b"latn");
/// `lcar`
pub const LCAR: u32 = tag!(b"lcar");
/// `liga`
pub const LIGA: u32 = tag!(b"liga");
/// `lnum`
pub const LNUM: u32 = tag!(b"lnum");
/// `loca`
pub const LOCA: u32 = tag!(b"loca");
/// `locl`
pub const LOCL: u32 = tag!(b"locl");
/// `LTSH`
pub const LTSH: u32 = tag!(b"LTSH");
/// `mark`
pub const MARK: u32 = tag!(b"mark");
/// `MATH`
pub const MATH: u32 = tag!(b"MATH");
/// `maxp`
pub const MAXP: u32 = tag!(b"maxp");
/// `med2`
pub const MED2: u32 = tag!(b"med2");
/// `medi`
pub const MEDI: u32 = tag!(b"medi");
/// `mkmk`
pub const MKMK: u32 = tag!(b"mkmk");
/// `mlm2`
pub const MLM2: u32 = tag!(b"mlm2");
/// `mlym`
pub const MLYM: u32 = tag!(b"mlym");
/// `mort`
pub const MORT: u32 = tag!(b"mort");
/// `morx`
pub const MORX: u32 = tag!(b"morx");
/// `mset`
pub const MSET: u32 = tag!(b"mset");
/// `name`
pub const NAME: u32 = tag!(b"name");
/// `nukt`
pub const NUKT: u32 = tag!(b"nukt");
/// `onum`
pub const ONUM: u32 = tag!(b"onum");
/// `opbd`
pub const OPBD: u32 = tag!(b"opbd");
/// `ordn`
pub const ORDN: u32 = tag!(b"ordn");
/// `ory2`
pub const ORY2: u32 = tag!(b"ory2");
/// `orya`
pub const ORYA: u32 = tag!(b"orya");
/// `OS/2`
pub const OS_2: u32 = tag!(b"OS/2");
/// `OTTO`
pub const OTTO: u32 = tag!(b"OTTO");
/// `PCLT`
pub const PCLT: u32 = tag!(b"PCLT");
/// `pnum`
pub const PNUM: u32 = tag!(b"pnum");
/// `png `
pub const PNG: u32 = tag!(b"png ");
/// `post`
pub const POST: u32 = tag!(b"post");
/// `pref`
pub const PREF: u32 = tag!(b"pref");
/// `prep`
pub const PREP: u32 = tag!(b"prep");
/// `pres`
pub const PRES: u32 = tag!(b"pres");
/// `prop`
pub const PROP: u32 = tag!(b"prop");
/// `pstf`
pub const PSTF: u32 = tag!(b"pstf");
/// `psts`
pub const PSTS: u32 = tag!(b"psts");
/// `rclt`
pub const RCLT: u32 = tag!(b"rclt");
/// `rkrf`
pub const RKRF: u32 = tag!(b"rkrf");
/// `rlig`
pub const RLIG: u32 = tag!(b"rlig");
/// `rphf`
pub const RPHF: u32 = tag!(b"rphf");
/// `sbix`
pub const SBIX: u32 = tag!(b"sbix");
/// `Silf`
pub const SILF: u32 = tag!(b"Silf");
/// `Sill`
pub const SILL: u32 = tag!(b"Sill");
/// `sinh`
pub const SINH: u32 = tag!(b"sinh");
/// `smcp`
pub const SMCP: u32 = tag!(b"smcp");
/// `SND `
pub const SND: u32 = tag!(b"SND ");
/// `SVG `
pub const SVG: u32 = tag!(b"SVG ");
/// `syrc`
pub const SYRC: u32 = tag!(b"syrc");
/// `taml`
pub const TAML: u32 = tag!(b"taml");
/// `tel2`
pub const TEL2: u32 = tag!(b"tel2");
/// `telu`
pub const TELU: u32 = tag!(b"telu");
/// `thai`
pub const THAI: u32 = tag!(b"thai");
/// `tiff`
pub const TIFF: u32 = tag!(b"tiff");
/// `tml2`
pub const TML2: u32 = tag!(b"tml2");
/// `tnum`
pub const TNUM: u32 = tag!(b"tnum");
/// `trak`
pub const TRAK: u32 = tag!(b"trak");
/// `ttcf`
pub const TTCF: u32 = tag!(b"ttcf");
/// `URD `
pub const URD: u32 = tag!(b"URD ");
/// `vatu`
pub const VATU: u32 = tag!(b"vatu");
/// `VDMX`
pub const VDMX: u32 = tag!(b"VDMX");
/// `vert`
pub const VERT: u32 = tag!(b"vert");
/// `vhea`
pub const VHEA: u32 = tag!(b"vhea");
/// `vmtx`
pub const VMTX: u32 = tag!(b"vmtx");
/// `VORG`
pub const VORG: u32 = tag!(b"VORG");
/// `vrt2`
pub const VRT2: u32 = tag!(b"vrt2");
/// `Zapf`
pub const ZAPF: u32 = tag!(b"Zapf");
/// `zero`
pub const ZERO: u32 = tag!(b"zero");
#[cfg(test)]
mod tests {
use super::*;
mod from_string {
use super::*;
#[test]
fn test_four_chars() {
let tag = from_string("beng").expect("invalid tag");
assert_eq!(tag, 1650814567);
}
#[test]
fn test_three_chars() {
let tag = from_string("BEN").expect("invalid tag");
assert_eq!(tag, 1111838240);
}
}
mod display_tag {
use crate::tag::{DisplayTag, NAME};
#[test]
fn test_simple_ascii() {
assert_eq!(DisplayTag(NAME).to_string(), "name".to_string());
}
#[test]
fn test_non_ascii() {
assert_eq!(DisplayTag(0x12345678).to_string(), "0x12345678".to_string());
}
#[test]
fn test_debug() {
assert_eq!(format!("{:?}", DisplayTag(NAME)), "name".to_string());
}
}
}
| fmt | identifier_name |
tag.rs | //! Utilities and constants for OpenType tags.
//!
//! See also the [`tag!`](../macro.tag.html) macro for creating tags from a byte string.
use crate::error::ParseError;
use std::{fmt, str};
/// Generate a 4-byte OpenType tag from byte string
///
/// Example:
///
/// ```
/// use allsorts::tag;
/// assert_eq!(tag!(b"glyf"), 0x676C7966);
/// ```
#[macro_export]
macro_rules! tag {
($w:expr) => {
u32::from_be_bytes(*$w)
};
}
/// Wrapper type for a tag that implements `Display`
///
/// Example:
///
/// ```
/// use allsorts::tag::{self, DisplayTag};
///
/// // ASCII tag comes out as a string
/// assert_eq!(&DisplayTag(tag::NAME).to_string(), "name");
/// // Non-ASCII tag comes out as hex
/// assert_eq!(&DisplayTag(0x12345678).to_string(), "0x12345678");
///
/// println!("DisplayTag is handy for printing a tag: '{}'", DisplayTag(tag::CFF));
/// ```
#[derive(PartialEq, Eq, Clone, Copy)]
pub struct DisplayTag(pub u32);
pub fn from_string(s: &str) -> Result<u32, ParseError> {
if s.len() > 4 {
return Err(ParseError::BadValue);
}
let mut tag: u32 = 0;
let mut count = 0;
for c in s.chars() {
if !c.is_ascii() || c.is_ascii_control() {
return Err(ParseError::BadValue);
}
tag = (tag << 8) | (c as u32);
count += 1;
}
while count < 4 {
tag = (tag << 8) | (' ' as u32);
count += 1;
}
Ok(tag)
}
impl fmt::Display for DisplayTag {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
let tag = self.0;
let bytes = tag.to_be_bytes();
if bytes.iter().all(|c| c.is_ascii() && !c.is_ascii_control()) {
let s = str::from_utf8(&bytes).unwrap(); // unwrap safe due to above check
s.fmt(f)
} else {
write!(f, "0x{:08x}", tag)
}
}
}
impl fmt::Debug for DisplayTag {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
fmt::Display::fmt(self, f)
}
}
/// `abvf`
pub const ABVF: u32 = tag!(b"abvf");
/// `abvm`
pub const ABVM: u32 = tag!(b"abvm");
/// `abvs`
pub const ABVS: u32 = tag!(b"abvs");
/// `acnt`
pub const ACNT: u32 = tag!(b"acnt");
/// `afrc`
pub const AFRC: u32 = tag!(b"afrc");
/// `akhn`
pub const AKHN: u32 = tag!(b"akhn");
/// `arab`
pub const ARAB: u32 = tag!(b"arab");
/// `avar`
pub const AVAR: u32 = tag!(b"avar");
/// `BASE`
pub const BASE: u32 = tag!(b"BASE");
/// `bdat`
pub const BDAT: u32 = tag!(b"bdat");
/// `beng`
pub const BENG: u32 = tag!(b"beng");
/// `bloc`
pub const BLOC: u32 = tag!(b"bloc");
/// `blwf`
pub const BLWF: u32 = tag!(b"blwf");
/// `blwm`
pub const BLWM: u32 = tag!(b"blwm");
/// `blws`
pub const BLWS: u32 = tag!(b"blws");
/// `bng2`
pub const BNG2: u32 = tag!(b"bng2");
/// `bsln`
pub const BSLN: u32 = tag!(b"bsln");
/// `c2sc`
pub const C2SC: u32 = tag!(b"c2sc");
/// `calt`
pub const CALT: u32 = tag!(b"calt");
/// `CBDT`
pub const CBDT: u32 = tag!(b"CBDT");
/// `CBLC`
pub const CBLC: u32 = tag!(b"CBLC");
/// `ccmp`
pub const CCMP: u32 = tag!(b"ccmp");
/// `cfar`
pub const CFAR: u32 = tag!(b"cfar");
/// `CFF `
pub const CFF: u32 = tag!(b"CFF ");
/// `cjct`
pub const CJCT: u32 = tag!(b"cjct");
/// `clig`
pub const CLIG: u32 = tag!(b"clig");
/// `cmap`
pub const CMAP: u32 = tag!(b"cmap");
/// `COLR`
pub const COLR: u32 = tag!(b"COLR");
/// `CPAL`
pub const CPAL: u32 = tag!(b"CPAL");
/// `curs`
pub const CURS: u32 = tag!(b"curs");
/// `cvar`
pub const CVAR: u32 = tag!(b"cvar");
/// `cvt `
pub const CVT: u32 = tag!(b"cvt ");
/// `cyrl`
pub const CYRL: u32 = tag!(b"cyrl");
/// `dev2`
pub const DEV2: u32 = tag!(b"dev2");
/// `deva`
pub const DEVA: u32 = tag!(b"deva");
/// `DFLT`
pub const DFLT: u32 = tag!(b"DFLT");
/// `dist`
pub const DIST: u32 = tag!(b"dist");
/// `dlig`
pub const DLIG: u32 = tag!(b"dlig");
/// `dupe`
pub const DUPE: u32 = tag!(b"dupe");
/// `EBDT`
pub const EBDT: u32 = tag!(b"EBDT");
/// `EBLC`
pub const EBLC: u32 = tag!(b"EBLC");
/// `EBSC`
pub const EBSC: u32 = tag!(b"EBSC");
/// `FAR `
pub const FAR: u32 = tag!(b"FAR ");
/// `fdsc`
pub const FDSC: u32 = tag!(b"fdsc");
/// `Feat`
pub const FEAT2: u32 = tag!(b"Feat");
/// `feat`
pub const FEAT: u32 = tag!(b"feat");
/// `fin2`
pub const FIN2: u32 = tag!(b"fin2");
/// `fin3`
pub const FIN3: u32 = tag!(b"fin3");
/// `fina`
pub const FINA: u32 = tag!(b"fina");
/// `fmtx`
pub const FMTX: u32 = tag!(b"fmtx");
/// `fpgm`
pub const FPGM: u32 = tag!(b"fpgm");
/// `frac`
pub const FRAC: u32 = tag!(b"frac");
/// `fvar`
pub const FVAR: u32 = tag!(b"fvar");
/// `gasp`
pub const GASP: u32 = tag!(b"gasp");
/// `GDEF`
pub const GDEF: u32 = tag!(b"GDEF");
/// `gjr2`
pub const GJR2: u32 = tag!(b"gjr2");
/// `Glat`
pub const GLAT: u32 = tag!(b"Glat");
/// `Gloc`
pub const GLOC: u32 = tag!(b"Gloc");
/// `glyf`
pub const GLYF: u32 = tag!(b"glyf");
/// `GPOS`
pub const GPOS: u32 = tag!(b"GPOS");
/// `grek`
pub const GREK: u32 = tag!(b"grek");
/// `GSUB`
pub const GSUB: u32 = tag!(b"GSUB");
/// `gujr`
pub const GUJR: u32 = tag!(b"gujr");
/// `gur2`
pub const GUR2: u32 = tag!(b"gur2");
/// `guru`
pub const GURU: u32 = tag!(b"guru");
/// `gvar`
pub const GVAR: u32 = tag!(b"gvar");
/// `half`
pub const HALF: u32 = tag!(b"half");
/// `haln`
pub const HALN: u32 = tag!(b"haln");
/// `hdmx`
pub const HDMX: u32 = tag!(b"hdmx");
/// `head`
pub const HEAD: u32 = tag!(b"head");
/// `hhea`
pub const HHEA: u32 = tag!(b"hhea");
/// `hlig`
pub const HLIG: u32 = tag!(b"hlig");
/// `hmtx`
pub const HMTX: u32 = tag!(b"hmtx");
/// `hsty`
pub const HSTY: u32 = tag!(b"hsty");
/// `init`
pub const INIT: u32 = tag!(b"init");
/// `isol`
pub const ISOL: u32 = tag!(b"isol");
/// `jpg `
pub const JPG: u32 = tag!(b"jpg ");
/// `JSTF`
pub const JSTF: u32 = tag!(b"JSTF");
/// `just`
pub const JUST: u32 = tag!(b"just");
/// `kern`
pub const KERN: u32 = tag!(b"kern");
/// `khmr`
pub const KHMR: u32 = tag!(b"khmr");
/// `knd2`
pub const KND2: u32 = tag!(b"knd2");
/// `knda`
pub const KNDA: u32 = tag!(b"knda");
/// `lao `
pub const LAO: u32 = tag!(b"lao ");
/// `latn`
pub const LATN: u32 = tag!(b"latn");
/// `lcar`
pub const LCAR: u32 = tag!(b"lcar");
/// `liga`
pub const LIGA: u32 = tag!(b"liga");
/// `lnum`
pub const LNUM: u32 = tag!(b"lnum");
/// `loca`
pub const LOCA: u32 = tag!(b"loca");
/// `locl`
pub const LOCL: u32 = tag!(b"locl");
/// `LTSH`
pub const LTSH: u32 = tag!(b"LTSH");
/// `mark`
pub const MARK: u32 = tag!(b"mark");
/// `MATH`
pub const MATH: u32 = tag!(b"MATH");
/// `maxp`
pub const MAXP: u32 = tag!(b"maxp");
/// `med2`
pub const MED2: u32 = tag!(b"med2");
/// `medi` | pub const MKMK: u32 = tag!(b"mkmk");
/// `mlm2`
pub const MLM2: u32 = tag!(b"mlm2");
/// `mlym`
pub const MLYM: u32 = tag!(b"mlym");
/// `mort`
pub const MORT: u32 = tag!(b"mort");
/// `morx`
pub const MORX: u32 = tag!(b"morx");
/// `mset`
pub const MSET: u32 = tag!(b"mset");
/// `name`
pub const NAME: u32 = tag!(b"name");
/// `nukt`
pub const NUKT: u32 = tag!(b"nukt");
/// `onum`
pub const ONUM: u32 = tag!(b"onum");
/// `opbd`
pub const OPBD: u32 = tag!(b"opbd");
/// `ordn`
pub const ORDN: u32 = tag!(b"ordn");
/// `ory2`
pub const ORY2: u32 = tag!(b"ory2");
/// `orya`
pub const ORYA: u32 = tag!(b"orya");
/// `OS/2`
pub const OS_2: u32 = tag!(b"OS/2");
/// `OTTO`
pub const OTTO: u32 = tag!(b"OTTO");
/// `PCLT`
pub const PCLT: u32 = tag!(b"PCLT");
/// `pnum`
pub const PNUM: u32 = tag!(b"pnum");
/// `png `
pub const PNG: u32 = tag!(b"png ");
/// `post`
pub const POST: u32 = tag!(b"post");
/// `pref`
pub const PREF: u32 = tag!(b"pref");
/// `prep`
pub const PREP: u32 = tag!(b"prep");
/// `pres`
pub const PRES: u32 = tag!(b"pres");
/// `prop`
pub const PROP: u32 = tag!(b"prop");
/// `pstf`
pub const PSTF: u32 = tag!(b"pstf");
/// `psts`
pub const PSTS: u32 = tag!(b"psts");
/// `rclt`
pub const RCLT: u32 = tag!(b"rclt");
/// `rkrf`
pub const RKRF: u32 = tag!(b"rkrf");
/// `rlig`
pub const RLIG: u32 = tag!(b"rlig");
/// `rphf`
pub const RPHF: u32 = tag!(b"rphf");
/// `sbix`
pub const SBIX: u32 = tag!(b"sbix");
/// `Silf`
pub const SILF: u32 = tag!(b"Silf");
/// `Sill`
pub const SILL: u32 = tag!(b"Sill");
/// `sinh`
pub const SINH: u32 = tag!(b"sinh");
/// `smcp`
pub const SMCP: u32 = tag!(b"smcp");
/// `SND `
pub const SND: u32 = tag!(b"SND ");
/// `SVG `
pub const SVG: u32 = tag!(b"SVG ");
/// `syrc`
pub const SYRC: u32 = tag!(b"syrc");
/// `taml`
pub const TAML: u32 = tag!(b"taml");
/// `tel2`
pub const TEL2: u32 = tag!(b"tel2");
/// `telu`
pub const TELU: u32 = tag!(b"telu");
/// `thai`
pub const THAI: u32 = tag!(b"thai");
/// `tiff`
pub const TIFF: u32 = tag!(b"tiff");
/// `tml2`
pub const TML2: u32 = tag!(b"tml2");
/// `tnum`
pub const TNUM: u32 = tag!(b"tnum");
/// `trak`
pub const TRAK: u32 = tag!(b"trak");
/// `ttcf`
pub const TTCF: u32 = tag!(b"ttcf");
/// `URD `
pub const URD: u32 = tag!(b"URD ");
/// `vatu`
pub const VATU: u32 = tag!(b"vatu");
/// `VDMX`
pub const VDMX: u32 = tag!(b"VDMX");
/// `vert`
pub const VERT: u32 = tag!(b"vert");
/// `vhea`
pub const VHEA: u32 = tag!(b"vhea");
/// `vmtx`
pub const VMTX: u32 = tag!(b"vmtx");
/// `VORG`
pub const VORG: u32 = tag!(b"VORG");
/// `vrt2`
pub const VRT2: u32 = tag!(b"vrt2");
/// `Zapf`
pub const ZAPF: u32 = tag!(b"Zapf");
/// `zero`
pub const ZERO: u32 = tag!(b"zero");
#[cfg(test)]
mod tests {
use super::*;
mod from_string {
use super::*;
#[test]
fn test_four_chars() {
let tag = from_string("beng").expect("invalid tag");
assert_eq!(tag, 1650814567);
}
#[test]
fn test_three_chars() {
let tag = from_string("BEN").expect("invalid tag");
assert_eq!(tag, 1111838240);
}
}
mod display_tag {
use crate::tag::{DisplayTag, NAME};
#[test]
fn test_simple_ascii() {
assert_eq!(DisplayTag(NAME).to_string(), "name".to_string());
}
#[test]
fn test_non_ascii() {
assert_eq!(DisplayTag(0x12345678).to_string(), "0x12345678".to_string());
}
#[test]
fn test_debug() {
assert_eq!(format!("{:?}", DisplayTag(NAME)), "name".to_string());
}
}
} | pub const MEDI: u32 = tag!(b"medi");
/// `mkmk` | random_line_split |
tag.rs | //! Utilities and constants for OpenType tags.
//!
//! See also the [`tag!`](../macro.tag.html) macro for creating tags from a byte string.
use crate::error::ParseError;
use std::{fmt, str};
/// Generate a 4-byte OpenType tag from byte string
///
/// Example:
///
/// ```
/// use allsorts::tag;
/// assert_eq!(tag!(b"glyf"), 0x676C7966);
/// ```
#[macro_export]
macro_rules! tag {
($w:expr) => {
u32::from_be_bytes(*$w)
};
}
/// Wrapper type for a tag that implements `Display`
///
/// Example:
///
/// ```
/// use allsorts::tag::{self, DisplayTag};
///
/// // ASCII tag comes out as a string
/// assert_eq!(&DisplayTag(tag::NAME).to_string(), "name");
/// // Non-ASCII tag comes out as hex
/// assert_eq!(&DisplayTag(0x12345678).to_string(), "0x12345678");
///
/// println!("DisplayTag is handy for printing a tag: '{}'", DisplayTag(tag::CFF));
/// ```
#[derive(PartialEq, Eq, Clone, Copy)]
pub struct DisplayTag(pub u32);
pub fn from_string(s: &str) -> Result<u32, ParseError> {
if s.len() > 4 {
return Err(ParseError::BadValue);
}
let mut tag: u32 = 0;
let mut count = 0;
for c in s.chars() {
if !c.is_ascii() || c.is_ascii_control() |
tag = (tag << 8) | (c as u32);
count += 1;
}
while count < 4 {
tag = (tag << 8) | (' ' as u32);
count += 1;
}
Ok(tag)
}
impl fmt::Display for DisplayTag {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
let tag = self.0;
let bytes = tag.to_be_bytes();
if bytes.iter().all(|c| c.is_ascii() && !c.is_ascii_control()) {
let s = str::from_utf8(&bytes).unwrap(); // unwrap safe due to above check
s.fmt(f)
} else {
write!(f, "0x{:08x}", tag)
}
}
}
impl fmt::Debug for DisplayTag {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
fmt::Display::fmt(self, f)
}
}
/// `abvf`
pub const ABVF: u32 = tag!(b"abvf");
/// `abvm`
pub const ABVM: u32 = tag!(b"abvm");
/// `abvs`
pub const ABVS: u32 = tag!(b"abvs");
/// `acnt`
pub const ACNT: u32 = tag!(b"acnt");
/// `afrc`
pub const AFRC: u32 = tag!(b"afrc");
/// `akhn`
pub const AKHN: u32 = tag!(b"akhn");
/// `arab`
pub const ARAB: u32 = tag!(b"arab");
/// `avar`
pub const AVAR: u32 = tag!(b"avar");
/// `BASE`
pub const BASE: u32 = tag!(b"BASE");
/// `bdat`
pub const BDAT: u32 = tag!(b"bdat");
/// `beng`
pub const BENG: u32 = tag!(b"beng");
/// `bloc`
pub const BLOC: u32 = tag!(b"bloc");
/// `blwf`
pub const BLWF: u32 = tag!(b"blwf");
/// `blwm`
pub const BLWM: u32 = tag!(b"blwm");
/// `blws`
pub const BLWS: u32 = tag!(b"blws");
/// `bng2`
pub const BNG2: u32 = tag!(b"bng2");
/// `bsln`
pub const BSLN: u32 = tag!(b"bsln");
/// `c2sc`
pub const C2SC: u32 = tag!(b"c2sc");
/// `calt`
pub const CALT: u32 = tag!(b"calt");
/// `CBDT`
pub const CBDT: u32 = tag!(b"CBDT");
/// `CBLC`
pub const CBLC: u32 = tag!(b"CBLC");
/// `ccmp`
pub const CCMP: u32 = tag!(b"ccmp");
/// `cfar`
pub const CFAR: u32 = tag!(b"cfar");
/// `CFF `
pub const CFF: u32 = tag!(b"CFF ");
/// `cjct`
pub const CJCT: u32 = tag!(b"cjct");
/// `clig`
pub const CLIG: u32 = tag!(b"clig");
/// `cmap`
pub const CMAP: u32 = tag!(b"cmap");
/// `COLR`
pub const COLR: u32 = tag!(b"COLR");
/// `CPAL`
pub const CPAL: u32 = tag!(b"CPAL");
/// `curs`
pub const CURS: u32 = tag!(b"curs");
/// `cvar`
pub const CVAR: u32 = tag!(b"cvar");
/// `cvt `
pub const CVT: u32 = tag!(b"cvt ");
/// `cyrl`
pub const CYRL: u32 = tag!(b"cyrl");
/// `dev2`
pub const DEV2: u32 = tag!(b"dev2");
/// `deva`
pub const DEVA: u32 = tag!(b"deva");
/// `DFLT`
pub const DFLT: u32 = tag!(b"DFLT");
/// `dist`
pub const DIST: u32 = tag!(b"dist");
/// `dlig`
pub const DLIG: u32 = tag!(b"dlig");
/// `dupe`
pub const DUPE: u32 = tag!(b"dupe");
/// `EBDT`
pub const EBDT: u32 = tag!(b"EBDT");
/// `EBLC`
pub const EBLC: u32 = tag!(b"EBLC");
/// `EBSC`
pub const EBSC: u32 = tag!(b"EBSC");
/// `FAR `
pub const FAR: u32 = tag!(b"FAR ");
/// `fdsc`
pub const FDSC: u32 = tag!(b"fdsc");
/// `Feat`
pub const FEAT2: u32 = tag!(b"Feat");
/// `feat`
pub const FEAT: u32 = tag!(b"feat");
/// `fin2`
pub const FIN2: u32 = tag!(b"fin2");
/// `fin3`
pub const FIN3: u32 = tag!(b"fin3");
/// `fina`
pub const FINA: u32 = tag!(b"fina");
/// `fmtx`
pub const FMTX: u32 = tag!(b"fmtx");
/// `fpgm`
pub const FPGM: u32 = tag!(b"fpgm");
/// `frac`
pub const FRAC: u32 = tag!(b"frac");
/// `fvar`
pub const FVAR: u32 = tag!(b"fvar");
/// `gasp`
pub const GASP: u32 = tag!(b"gasp");
/// `GDEF`
pub const GDEF: u32 = tag!(b"GDEF");
/// `gjr2`
pub const GJR2: u32 = tag!(b"gjr2");
/// `Glat`
pub const GLAT: u32 = tag!(b"Glat");
/// `Gloc`
pub const GLOC: u32 = tag!(b"Gloc");
/// `glyf`
pub const GLYF: u32 = tag!(b"glyf");
/// `GPOS`
pub const GPOS: u32 = tag!(b"GPOS");
/// `grek`
pub const GREK: u32 = tag!(b"grek");
/// `GSUB`
pub const GSUB: u32 = tag!(b"GSUB");
/// `gujr`
pub const GUJR: u32 = tag!(b"gujr");
/// `gur2`
pub const GUR2: u32 = tag!(b"gur2");
/// `guru`
pub const GURU: u32 = tag!(b"guru");
/// `gvar`
pub const GVAR: u32 = tag!(b"gvar");
/// `half`
pub const HALF: u32 = tag!(b"half");
/// `haln`
pub const HALN: u32 = tag!(b"haln");
/// `hdmx`
pub const HDMX: u32 = tag!(b"hdmx");
/// `head`
pub const HEAD: u32 = tag!(b"head");
/// `hhea`
pub const HHEA: u32 = tag!(b"hhea");
/// `hlig`
pub const HLIG: u32 = tag!(b"hlig");
/// `hmtx`
pub const HMTX: u32 = tag!(b"hmtx");
/// `hsty`
pub const HSTY: u32 = tag!(b"hsty");
/// `init`
pub const INIT: u32 = tag!(b"init");
/// `isol`
pub const ISOL: u32 = tag!(b"isol");
/// `jpg `
pub const JPG: u32 = tag!(b"jpg ");
/// `JSTF`
pub const JSTF: u32 = tag!(b"JSTF");
/// `just`
pub const JUST: u32 = tag!(b"just");
/// `kern`
pub const KERN: u32 = tag!(b"kern");
/// `khmr`
pub const KHMR: u32 = tag!(b"khmr");
/// `knd2`
pub const KND2: u32 = tag!(b"knd2");
/// `knda`
pub const KNDA: u32 = tag!(b"knda");
/// `lao `
pub const LAO: u32 = tag!(b"lao ");
/// `latn`
pub const LATN: u32 = tag!(b"latn");
/// `lcar`
pub const LCAR: u32 = tag!(b"lcar");
/// `liga`
pub const LIGA: u32 = tag!(b"liga");
/// `lnum`
pub const LNUM: u32 = tag!(b"lnum");
/// `loca`
pub const LOCA: u32 = tag!(b"loca");
/// `locl`
pub const LOCL: u32 = tag!(b"locl");
/// `LTSH`
pub const LTSH: u32 = tag!(b"LTSH");
/// `mark`
pub const MARK: u32 = tag!(b"mark");
/// `MATH`
pub const MATH: u32 = tag!(b"MATH");
/// `maxp`
pub const MAXP: u32 = tag!(b"maxp");
/// `med2`
pub const MED2: u32 = tag!(b"med2");
/// `medi`
pub const MEDI: u32 = tag!(b"medi");
/// `mkmk`
pub const MKMK: u32 = tag!(b"mkmk");
/// `mlm2`
pub const MLM2: u32 = tag!(b"mlm2");
/// `mlym`
pub const MLYM: u32 = tag!(b"mlym");
/// `mort`
pub const MORT: u32 = tag!(b"mort");
/// `morx`
pub const MORX: u32 = tag!(b"morx");
/// `mset`
pub const MSET: u32 = tag!(b"mset");
/// `name`
pub const NAME: u32 = tag!(b"name");
/// `nukt`
pub const NUKT: u32 = tag!(b"nukt");
/// `onum`
pub const ONUM: u32 = tag!(b"onum");
/// `opbd`
pub const OPBD: u32 = tag!(b"opbd");
/// `ordn`
pub const ORDN: u32 = tag!(b"ordn");
/// `ory2`
pub const ORY2: u32 = tag!(b"ory2");
/// `orya`
pub const ORYA: u32 = tag!(b"orya");
/// `OS/2`
pub const OS_2: u32 = tag!(b"OS/2");
/// `OTTO`
pub const OTTO: u32 = tag!(b"OTTO");
/// `PCLT`
pub const PCLT: u32 = tag!(b"PCLT");
/// `pnum`
pub const PNUM: u32 = tag!(b"pnum");
/// `png `
pub const PNG: u32 = tag!(b"png ");
/// `post`
pub const POST: u32 = tag!(b"post");
/// `pref`
pub const PREF: u32 = tag!(b"pref");
/// `prep`
pub const PREP: u32 = tag!(b"prep");
/// `pres`
pub const PRES: u32 = tag!(b"pres");
/// `prop`
pub const PROP: u32 = tag!(b"prop");
/// `pstf`
pub const PSTF: u32 = tag!(b"pstf");
/// `psts`
pub const PSTS: u32 = tag!(b"psts");
/// `rclt`
pub const RCLT: u32 = tag!(b"rclt");
/// `rkrf`
pub const RKRF: u32 = tag!(b"rkrf");
/// `rlig`
pub const RLIG: u32 = tag!(b"rlig");
/// `rphf`
pub const RPHF: u32 = tag!(b"rphf");
/// `sbix`
pub const SBIX: u32 = tag!(b"sbix");
/// `Silf`
pub const SILF: u32 = tag!(b"Silf");
/// `Sill`
pub const SILL: u32 = tag!(b"Sill");
/// `sinh`
pub const SINH: u32 = tag!(b"sinh");
/// `smcp`
pub const SMCP: u32 = tag!(b"smcp");
/// `SND `
pub const SND: u32 = tag!(b"SND ");
/// `SVG `
pub const SVG: u32 = tag!(b"SVG ");
/// `syrc`
pub const SYRC: u32 = tag!(b"syrc");
/// `taml`
pub const TAML: u32 = tag!(b"taml");
/// `tel2`
pub const TEL2: u32 = tag!(b"tel2");
/// `telu`
pub const TELU: u32 = tag!(b"telu");
/// `thai`
pub const THAI: u32 = tag!(b"thai");
/// `tiff`
pub const TIFF: u32 = tag!(b"tiff");
/// `tml2`
pub const TML2: u32 = tag!(b"tml2");
/// `tnum`
pub const TNUM: u32 = tag!(b"tnum");
/// `trak`
pub const TRAK: u32 = tag!(b"trak");
/// `ttcf`
pub const TTCF: u32 = tag!(b"ttcf");
/// `URD `
pub const URD: u32 = tag!(b"URD ");
/// `vatu`
pub const VATU: u32 = tag!(b"vatu");
/// `VDMX`
pub const VDMX: u32 = tag!(b"VDMX");
/// `vert`
pub const VERT: u32 = tag!(b"vert");
/// `vhea`
pub const VHEA: u32 = tag!(b"vhea");
/// `vmtx`
pub const VMTX: u32 = tag!(b"vmtx");
/// `VORG`
pub const VORG: u32 = tag!(b"VORG");
/// `vrt2`
pub const VRT2: u32 = tag!(b"vrt2");
/// `Zapf`
pub const ZAPF: u32 = tag!(b"Zapf");
/// `zero`
pub const ZERO: u32 = tag!(b"zero");
#[cfg(test)]
mod tests {
use super::*;
mod from_string {
use super::*;
#[test]
fn test_four_chars() {
let tag = from_string("beng").expect("invalid tag");
assert_eq!(tag, 1650814567);
}
#[test]
fn test_three_chars() {
let tag = from_string("BEN").expect("invalid tag");
assert_eq!(tag, 1111838240);
}
}
mod display_tag {
use crate::tag::{DisplayTag, NAME};
#[test]
fn test_simple_ascii() {
assert_eq!(DisplayTag(NAME).to_string(), "name".to_string());
}
#[test]
fn test_non_ascii() {
assert_eq!(DisplayTag(0x12345678).to_string(), "0x12345678".to_string());
}
#[test]
fn test_debug() {
assert_eq!(format!("{:?}", DisplayTag(NAME)), "name".to_string());
}
}
}
| {
return Err(ParseError::BadValue);
} | conditional_block |
lib.rs | /*!
* This library provides an API client for Diffbot.
*
* Making API requests
* -------------------
*
* There are a handful of different ways to make API calls:
*
* 1. The most basic way to make a request is with the ``call()`` function.
* Everything must be specified for each request.
*
* 2. Use the ``Diffbot`` struct to keep track of your token and API version
* and then use its ``.call()`` method to make API calls. This has the
* advantage that you can specify those things just once and they'll be
* retained.
*
* 3. Instead of making a request in one step, you can make it two steps with
* the ``prepare_request()`` function. This allows you to specify to Diffbot
* certain details of how *it* should make the request. That gives you a
* ``Request`` object.
*
* 4. In the same manner, if you have a ``Diffbot`` struct you can call the
* ``.prepare_request()`` method on it.
*
* Prepared requests
* -----------------
*
* If you use the ``prepare_request()`` function or method, you can tweak the
* request that will be sent to Diffbot. You can alter the User-Agent, Referer
* or Cookie headers that it will send and then call ``.call()`` to make the
* request, or you can call ``.post_body()`` to send the HTML yourself, if it
* is not publicly available to the wider Internet.
*
* Getting data out of the result
* ------------------------------
*
* At present, the successful return value of a request is simply a JSON object,
* a tree map. This *will* make it moderately difficult to work with, but if
* you're determined, it's possible. You'll end up with results like these:
*
* // First of all, you must, of course, have a response to work on.
* let mut response: TreeMap<~str, Json>
* = diffbot::call(..., "article", ...).unwrap();
*
* // Get the title of the article
* let title = match response.pop(&~"title").unwrap() {
* json::String(s) => s,
* _ => unreachable!(),
* };
*
* // Get the URL of each image
* let image_urls: ~[Url] = match response.pop(&~"images").unwrap() {
* json::List(images) => images.move_iter().map(|image| match image {
* json::Object(~mut o) => {
* match o.pop(&~"url").unwrap() {
* json::String(ref s) => from_str(s),
* _ => unreachable!(),
* }
* },
* _ => unreachable!(),
* }),
* _ => unreachable!(),
* }.collect();
*
* (Yep, I'll freely admit that these are clumsier than they might be in another
* language, which might allow something like this:
*
* let response = ...;
*
* let title = response.title;
* let image_urls = [from_str(image.url) for image in response.images];
*
* In time we may get strongly typed interfaces which would be much nicer, but
* for now, you'd need to do that yourself. It can be done with the tools in
* ``extra::serialize``, by the way.)
*/
#[crate_id = "diffbot#1.0"];
#[crate_type = "dylib"];
#[crate_type = "rlib"];
#[doc(html_logo_url = "diffy-d.png",
html_favicon_url = "http://www.diffbot.com/favicon.ico")];
extern mod extra = "extra#0.10-pre";
extern mod http = "http#0.1-pre";
use std::io::net::tcp::TcpStream;
use extra::json;
use extra::url::Url;
use http::client::RequestWriter;
use http::method::{Get, Post};
use http::headers::content_type::MediaType;
/// A convenience type which simply keeps track of a developer token and version
/// number.
///
/// There is no necessity to use this type; you can call ``call()`` directly
/// should you so desire.
#[deriving(Eq, Clone)]
pub struct Diffbot {
/// The developer's token
token: ~str,
/// The API version number
version: uint,
}
// Basic methods
impl Diffbot {
/// Construct a new ``Diffbot`` instance from the passed parameters.
pub fn new(token: ~str, version: uint) -> Diffbot {
Diffbot {
token: token,
version: version,
}
}
/// Make a call to any Diffbot API with the stored token and API version.
///
/// See the ``call()`` function for an explanation of the parameters.
pub fn | (&self, url: &Url, api: &str, fields: &[&str])
-> Result<json::Object, Error> {
call(url, self.token, api, fields, self.version)
}
/// Prepare a request to any Diffbot API with the stored token and API version.
///
/// See the ``call()`` function for an explanation of the parameters.
pub fn prepare_request(&self, url: &Url, api: &str, fields: &[&str])
-> Request {
prepare_request(url, self.token, api, fields, self.version)
}
}
/// An in-progress Diffbot API call.
pub struct Request {
priv request: RequestWriter<TcpStream>,
}
impl Request {
/// Set the value for Diffbot to send as the ``User-Agent`` header when
/// making your request.
pub fn user_agent(&mut self, user_agent: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-User-Agent",
user_agent);
}
/// Set the value for Diffbot to send as the ``Referer`` header when
/// making your request.
pub fn referer(&mut self, referer: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-Referer",
referer);
}
/// Set the value for Diffbot to send as the ``Cookie`` header when
/// making your request.
pub fn cookie(&mut self, cookie: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-Cookie",
cookie);
}
/// Set Diffbot's timeout, in milliseconds. The default is five seconds.
pub fn timeout(&mut self, milliseconds: u64) {
self.request.url.query.push((~"timeout", milliseconds.to_str()));
}
/// Execute the request and get the results.
pub fn call(self) -> Result<json::Object, Error> {
let mut response = match self.request.read_response() {
Ok(response) => response,
Err(_request) => return Err(IoError), // Request failed
};
let json = match json::from_reader(&mut response as &mut Reader) {
Ok(json) => json,
Err(error) => return Err(JsonError(error)), // It... wasn't JSON!?
};
// Now let's see if this is an API error or not.
// API errors are of the form {"error":"Invalid API.","errorCode":500}
match json {
json::Object(~mut o) => {
match o.pop(&~"errorCode") {
Some(json::Number(num)) => {
let num = num as uint;
let msg = match o.pop(&~"error")
.expect("JSON had errorCode but not error") {
json::String(s) => s,
uh_oh => fail!("error was {} instead of a string", uh_oh.to_str()),
};
Err(ApiError(msg, num))
},
Some(uh_oh) => fail!("errorCode was {} instead of a number", uh_oh.to_str()),
None => Ok(o),
}
},
// All API responses must be objects.
// If it's not, there's something screwy going on.
_ => fail!("API return value wasn't a JSON object"),
}
}
/// Execute the request as a POST request, sending it through with the given
/// text/html entity body.
///
/// This has the effect that Diffbot will skip requesting the URL and will
/// instead take the passed body as the HTML it is to check. This is mainly
/// useful for non-public websites.
pub fn post_body(mut self, body: &[u8]) -> Result<json::Object, Error> {
self.request.method = Post;
self.request.headers.content_type = Some(MediaType(~"text", ~"html", ~[]));
self.request.headers.content_length = Some(body.len());
// Calling write_headers is an extra and unnecessary safety guard which
// will cause the task to fail if the request has already started to be
// sent (which would render the three statements above ineffectual)
self.request.write_headers();
self.request.write(body);
self.call()
}
}
/// Error code: "unauthorized token"
pub static UNAUTHORIZED_TOKEN: uint = 401;
/// Error code: "requested page not found"
pub static REQUESTED_PAGE_NOT_FOUND: uint = 404;
/// Error code: "your token has exceeded the allowed number of calls, or has
/// otherwise been throttled for API abuse."
pub static TOKEN_EXCEEDED_OR_THROTTLED: uint = 429;
/// Error code: "error processing the page. Specific information will be
/// returned in the JSON response."
pub static ERROR_PROCESSING: uint = 500;
/// Something went wrong with the Diffbot API call.
#[deriving(Eq)]
pub enum Error {
/// An error code returned by the Diffbot API, with message and code.
/// Refer to http://www.diffbot.com/dev/docs/error/ for an explanation of
/// the error codes.
///
/// When comparing the error code, you should use these constants:
///
/// - ``UNAUTHORIZED_TOKEN``: "unauthorized token"
/// - ``REQUESTED_PAGE_NOT_FOUND``: "requested page not found"
/// - ``TOKEN_EXCEEDED_OR_THROTTLED``: "your token has exceeded the allowed
/// number of calls, or has otherwise been throttled for API abuse."
/// - ``ERROR_PROCESSING``: "error processing the page. Specific information
/// will be returned in the JSON response."
ApiError(~str, uint),
/// The JSON was not valid. This is one of those ones that *should* never
/// happen; you know...
///
/// Actually, I can percieve that it might happen if a document returned
/// included invalid UTF-8, but this case has not been tested.
JsonError(json::Error),
/// An I/O error occurred and the condition was trapped somewhere (by you).
IoError,
}
impl ToStr for Error {
fn to_str(&self) -> ~str {
match *self {
ApiError(ref msg, code) => format!("API error {}: {}", code, *msg),
JsonError(ref error) => format!("JSON error: {}", error.to_str()),
IoError => format!("I/O error (already handled)"),
}
}
}
/// Make a simple Diffbot API call.
///
/// For more complex requests, use ``Diffbot`` or ``prepare_request()``.
///
/// Arguments
/// =========
///
/// - ``url`` is the URL that you wish Diffbot to operate upon. If this is a
/// publicly-inaccessible URL, you should use ``post_body()`` on a prepared
/// request instead of ``call()``.
///
/// - ``token`` is the developer's token.
///
/// - ``api`` is the name of the API endpoint, e.g. "article", "product".
///
/// - ``fields`` is the set of fields you want the API call to return; it
/// follows the form specified by the Diffbot API and so should have values
/// like "*", "meta", "querystring", "images(*)".
///
/// - ``version`` is the Diffbot API version number.
pub fn call(url: &Url, token: &str, api: &str, fields: &[&str], version: uint)
-> Result<json::Object, Error> {
prepare_request(url, token, api, fields, version).call()
}
/// Prepare, but do not send, a request.
///
/// This allows you to use some of the more advanced features of the API like
/// setting certain headers for Diffbot to use, or uploading a private document
/// for it.
pub fn prepare_request(url: &Url, token: &str, api: &str, fields: &[&str],
version: uint)
-> Request {
// First of all we must calculate the GET parameters.
let mut query = ~[(~"token", token.to_owned()),
(~"url", url.to_str())];
if fields.len() > 0 {
query.push((~"fields", fields.connect(",")));
}
// Now that we've got that, we can figure out the complete URL.
let url = Url::new(~"http", // scheme
None, // user
~"api.diffbot.com", // host
None, // port
format!("/v{}/{}", version, api), // path
query, // query
None); // fragment
// And with that, we can now make the request. Whee!
Request {
request: RequestWriter::new(Get, url)
}
}
| call | identifier_name |
lib.rs | /*!
* This library provides an API client for Diffbot.
*
* Making API requests
* -------------------
*
* There are a handful of different ways to make API calls:
*
* 1. The most basic way to make a request is with the ``call()`` function.
* Everything must be specified for each request.
*
* 2. Use the ``Diffbot`` struct to keep track of your token and API version
* and then use its ``.call()`` method to make API calls. This has the
* advantage that you can specify those things just once and they'll be
* retained.
*
* 3. Instead of making a request in one step, you can make it two steps with
* the ``prepare_request()`` function. This allows you to specify to Diffbot
* certain details of how *it* should make the request. That gives you a
* ``Request`` object.
*
* 4. In the same manner, if you have a ``Diffbot`` struct you can call the
* ``.prepare_request()`` method on it.
*
* Prepared requests
* -----------------
*
* If you use the ``prepare_request()`` function or method, you can tweak the
* request that will be sent to Diffbot. You can alter the User-Agent, Referer
* or Cookie headers that it will send and then call ``.call()`` to make the
* request, or you can call ``.post_body()`` to send the HTML yourself, if it
* is not publicly available to the wider Internet.
*
* Getting data out of the result
* ------------------------------
*
* At present, the successful return value of a request is simply a JSON object,
* a tree map. This *will* make it moderately difficult to work with, but if
* you're determined, it's possible. You'll end up with results like these:
*
* // First of all, you must, of course, have a response to work on.
* let mut response: TreeMap<~str, Json>
* = diffbot::call(..., "article", ...).unwrap();
*
* // Get the title of the article
* let title = match response.pop(&~"title").unwrap() {
* json::String(s) => s,
* _ => unreachable!(),
* };
*
* // Get the URL of each image
* let image_urls: ~[Url] = match response.pop(&~"images").unwrap() {
* json::List(images) => images.move_iter().map(|image| match image {
* json::Object(~mut o) => {
* match o.pop(&~"url").unwrap() {
* json::String(ref s) => from_str(s),
* _ => unreachable!(),
* }
* },
* _ => unreachable!(),
* }),
* _ => unreachable!(),
* }.collect();
*
* (Yep, I'll freely admit that these are clumsier than they might be in another
* language, which might allow something like this:
*
* let response = ...;
*
* let title = response.title;
* let image_urls = [from_str(image.url) for image in response.images];
*
* In time we may get strongly typed interfaces which would be much nicer, but
* for now, you'd need to do that yourself. It can be done with the tools in
* ``extra::serialize``, by the way.)
*/
#[crate_id = "diffbot#1.0"];
#[crate_type = "dylib"];
#[crate_type = "rlib"];
#[doc(html_logo_url = "diffy-d.png",
html_favicon_url = "http://www.diffbot.com/favicon.ico")];
extern mod extra = "extra#0.10-pre";
extern mod http = "http#0.1-pre";
use std::io::net::tcp::TcpStream;
use extra::json;
use extra::url::Url;
use http::client::RequestWriter;
use http::method::{Get, Post};
use http::headers::content_type::MediaType;
/// A convenience type which simply keeps track of a developer token and version
/// number.
///
/// There is no necessity to use this type; you can call ``call()`` directly
/// should you so desire.
#[deriving(Eq, Clone)]
pub struct Diffbot {
/// The developer's token
token: ~str,
/// The API version number
version: uint,
}
// Basic methods
impl Diffbot {
/// Construct a new ``Diffbot`` instance from the passed parameters.
pub fn new(token: ~str, version: uint) -> Diffbot {
Diffbot {
token: token,
version: version,
}
}
/// Make a call to any Diffbot API with the stored token and API version.
///
/// See the ``call()`` function for an explanation of the parameters.
pub fn call(&self, url: &Url, api: &str, fields: &[&str])
-> Result<json::Object, Error> {
call(url, self.token, api, fields, self.version)
}
/// Prepare a request to any Diffbot API with the stored token and API version.
///
/// See the ``call()`` function for an explanation of the parameters.
pub fn prepare_request(&self, url: &Url, api: &str, fields: &[&str])
-> Request {
prepare_request(url, self.token, api, fields, self.version)
}
}
/// An in-progress Diffbot API call.
pub struct Request {
priv request: RequestWriter<TcpStream>,
}
impl Request {
/// Set the value for Diffbot to send as the ``User-Agent`` header when
/// making your request.
pub fn user_agent(&mut self, user_agent: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-User-Agent",
user_agent);
}
/// Set the value for Diffbot to send as the ``Referer`` header when
/// making your request.
pub fn referer(&mut self, referer: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-Referer",
referer);
}
/// Set the value for Diffbot to send as the ``Cookie`` header when
/// making your request.
pub fn cookie(&mut self, cookie: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-Cookie",
cookie);
}
/// Set Diffbot's timeout, in milliseconds. The default is five seconds.
pub fn timeout(&mut self, milliseconds: u64) {
self.request.url.query.push((~"timeout", milliseconds.to_str()));
}
/// Execute the request and get the results.
pub fn call(self) -> Result<json::Object, Error> {
let mut response = match self.request.read_response() {
Ok(response) => response,
Err(_request) => return Err(IoError), // Request failed
};
let json = match json::from_reader(&mut response as &mut Reader) {
Ok(json) => json,
Err(error) => return Err(JsonError(error)), // It... wasn't JSON!?
};
// Now let's see if this is an API error or not.
// API errors are of the form {"error":"Invalid API.","errorCode":500} | let num = num as uint;
let msg = match o.pop(&~"error")
.expect("JSON had errorCode but not error") {
json::String(s) => s,
uh_oh => fail!("error was {} instead of a string", uh_oh.to_str()),
};
Err(ApiError(msg, num))
},
Some(uh_oh) => fail!("errorCode was {} instead of a number", uh_oh.to_str()),
None => Ok(o),
}
},
// All API responses must be objects.
// If it's not, there's something screwy going on.
_ => fail!("API return value wasn't a JSON object"),
}
}
/// Execute the request as a POST request, sending it through with the given
/// text/html entity body.
///
/// This has the effect that Diffbot will skip requesting the URL and will
/// instead take the passed body as the HTML it is to check. This is mainly
/// useful for non-public websites.
pub fn post_body(mut self, body: &[u8]) -> Result<json::Object, Error> {
self.request.method = Post;
self.request.headers.content_type = Some(MediaType(~"text", ~"html", ~[]));
self.request.headers.content_length = Some(body.len());
// Calling write_headers is an extra and unnecessary safety guard which
// will cause the task to fail if the request has already started to be
// sent (which would render the three statements above ineffectual)
self.request.write_headers();
self.request.write(body);
self.call()
}
}
/// Error code: "unauthorized token"
pub static UNAUTHORIZED_TOKEN: uint = 401;
/// Error code: "requested page not found"
pub static REQUESTED_PAGE_NOT_FOUND: uint = 404;
/// Error code: "your token has exceeded the allowed number of calls, or has
/// otherwise been throttled for API abuse."
pub static TOKEN_EXCEEDED_OR_THROTTLED: uint = 429;
/// Error code: "error processing the page. Specific information will be
/// returned in the JSON response."
pub static ERROR_PROCESSING: uint = 500;
/// Something went wrong with the Diffbot API call.
#[deriving(Eq)]
pub enum Error {
/// An error code returned by the Diffbot API, with message and code.
/// Refer to http://www.diffbot.com/dev/docs/error/ for an explanation of
/// the error codes.
///
/// When comparing the error code, you should use these constants:
///
/// - ``UNAUTHORIZED_TOKEN``: "unauthorized token"
/// - ``REQUESTED_PAGE_NOT_FOUND``: "requested page not found"
/// - ``TOKEN_EXCEEDED_OR_THROTTLED``: "your token has exceeded the allowed
/// number of calls, or has otherwise been throttled for API abuse."
/// - ``ERROR_PROCESSING``: "error processing the page. Specific information
/// will be returned in the JSON response."
ApiError(~str, uint),
/// The JSON was not valid. This is one of those ones that *should* never
/// happen; you know...
///
/// Actually, I can percieve that it might happen if a document returned
/// included invalid UTF-8, but this case has not been tested.
JsonError(json::Error),
/// An I/O error occurred and the condition was trapped somewhere (by you).
IoError,
}
impl ToStr for Error {
fn to_str(&self) -> ~str {
match *self {
ApiError(ref msg, code) => format!("API error {}: {}", code, *msg),
JsonError(ref error) => format!("JSON error: {}", error.to_str()),
IoError => format!("I/O error (already handled)"),
}
}
}
/// Make a simple Diffbot API call.
///
/// For more complex requests, use ``Diffbot`` or ``prepare_request()``.
///
/// Arguments
/// =========
///
/// - ``url`` is the URL that you wish Diffbot to operate upon. If this is a
/// publicly-inaccessible URL, you should use ``post_body()`` on a prepared
/// request instead of ``call()``.
///
/// - ``token`` is the developer's token.
///
/// - ``api`` is the name of the API endpoint, e.g. "article", "product".
///
/// - ``fields`` is the set of fields you want the API call to return; it
/// follows the form specified by the Diffbot API and so should have values
/// like "*", "meta", "querystring", "images(*)".
///
/// - ``version`` is the Diffbot API version number.
pub fn call(url: &Url, token: &str, api: &str, fields: &[&str], version: uint)
-> Result<json::Object, Error> {
prepare_request(url, token, api, fields, version).call()
}
/// Prepare, but do not send, a request.
///
/// This allows you to use some of the more advanced features of the API like
/// setting certain headers for Diffbot to use, or uploading a private document
/// for it.
pub fn prepare_request(url: &Url, token: &str, api: &str, fields: &[&str],
version: uint)
-> Request {
// First of all we must calculate the GET parameters.
let mut query = ~[(~"token", token.to_owned()),
(~"url", url.to_str())];
if fields.len() > 0 {
query.push((~"fields", fields.connect(",")));
}
// Now that we've got that, we can figure out the complete URL.
let url = Url::new(~"http", // scheme
None, // user
~"api.diffbot.com", // host
None, // port
format!("/v{}/{}", version, api), // path
query, // query
None); // fragment
// And with that, we can now make the request. Whee!
Request {
request: RequestWriter::new(Get, url)
}
} | match json {
json::Object(~mut o) => {
match o.pop(&~"errorCode") {
Some(json::Number(num)) => { | random_line_split |
lib.rs | /*!
* This library provides an API client for Diffbot.
*
* Making API requests
* -------------------
*
* There are a handful of different ways to make API calls:
*
* 1. The most basic way to make a request is with the ``call()`` function.
* Everything must be specified for each request.
*
* 2. Use the ``Diffbot`` struct to keep track of your token and API version
* and then use its ``.call()`` method to make API calls. This has the
* advantage that you can specify those things just once and they'll be
* retained.
*
* 3. Instead of making a request in one step, you can make it two steps with
* the ``prepare_request()`` function. This allows you to specify to Diffbot
* certain details of how *it* should make the request. That gives you a
* ``Request`` object.
*
* 4. In the same manner, if you have a ``Diffbot`` struct you can call the
* ``.prepare_request()`` method on it.
*
* Prepared requests
* -----------------
*
* If you use the ``prepare_request()`` function or method, you can tweak the
* request that will be sent to Diffbot. You can alter the User-Agent, Referer
* or Cookie headers that it will send and then call ``.call()`` to make the
* request, or you can call ``.post_body()`` to send the HTML yourself, if it
* is not publicly available to the wider Internet.
*
* Getting data out of the result
* ------------------------------
*
* At present, the successful return value of a request is simply a JSON object,
* a tree map. This *will* make it moderately difficult to work with, but if
* you're determined, it's possible. You'll end up with results like these:
*
* // First of all, you must, of course, have a response to work on.
* let mut response: TreeMap<~str, Json>
* = diffbot::call(..., "article", ...).unwrap();
*
* // Get the title of the article
* let title = match response.pop(&~"title").unwrap() {
* json::String(s) => s,
* _ => unreachable!(),
* };
*
* // Get the URL of each image
* let image_urls: ~[Url] = match response.pop(&~"images").unwrap() {
* json::List(images) => images.move_iter().map(|image| match image {
* json::Object(~mut o) => {
* match o.pop(&~"url").unwrap() {
* json::String(ref s) => from_str(s),
* _ => unreachable!(),
* }
* },
* _ => unreachable!(),
* }),
* _ => unreachable!(),
* }.collect();
*
* (Yep, I'll freely admit that these are clumsier than they might be in another
* language, which might allow something like this:
*
* let response = ...;
*
* let title = response.title;
* let image_urls = [from_str(image.url) for image in response.images];
*
* In time we may get strongly typed interfaces which would be much nicer, but
* for now, you'd need to do that yourself. It can be done with the tools in
* ``extra::serialize``, by the way.)
*/
#[crate_id = "diffbot#1.0"];
#[crate_type = "dylib"];
#[crate_type = "rlib"];
#[doc(html_logo_url = "diffy-d.png",
html_favicon_url = "http://www.diffbot.com/favicon.ico")];
extern mod extra = "extra#0.10-pre";
extern mod http = "http#0.1-pre";
use std::io::net::tcp::TcpStream;
use extra::json;
use extra::url::Url;
use http::client::RequestWriter;
use http::method::{Get, Post};
use http::headers::content_type::MediaType;
/// A convenience type which simply keeps track of a developer token and version
/// number.
///
/// There is no necessity to use this type; you can call ``call()`` directly
/// should you so desire.
#[deriving(Eq, Clone)]
pub struct Diffbot {
/// The developer's token
token: ~str,
/// The API version number
version: uint,
}
// Basic methods
impl Diffbot {
/// Construct a new ``Diffbot`` instance from the passed parameters.
pub fn new(token: ~str, version: uint) -> Diffbot {
Diffbot {
token: token,
version: version,
}
}
/// Make a call to any Diffbot API with the stored token and API version.
///
/// See the ``call()`` function for an explanation of the parameters.
pub fn call(&self, url: &Url, api: &str, fields: &[&str])
-> Result<json::Object, Error> {
call(url, self.token, api, fields, self.version)
}
/// Prepare a request to any Diffbot API with the stored token and API version.
///
/// See the ``call()`` function for an explanation of the parameters.
pub fn prepare_request(&self, url: &Url, api: &str, fields: &[&str])
-> Request |
}
/// An in-progress Diffbot API call.
pub struct Request {
priv request: RequestWriter<TcpStream>,
}
impl Request {
/// Set the value for Diffbot to send as the ``User-Agent`` header when
/// making your request.
pub fn user_agent(&mut self, user_agent: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-User-Agent",
user_agent);
}
/// Set the value for Diffbot to send as the ``Referer`` header when
/// making your request.
pub fn referer(&mut self, referer: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-Referer",
referer);
}
/// Set the value for Diffbot to send as the ``Cookie`` header when
/// making your request.
pub fn cookie(&mut self, cookie: ~str) {
self.request.headers.extensions.insert(~"X-Forwarded-Cookie",
cookie);
}
/// Set Diffbot's timeout, in milliseconds. The default is five seconds.
pub fn timeout(&mut self, milliseconds: u64) {
self.request.url.query.push((~"timeout", milliseconds.to_str()));
}
/// Execute the request and get the results.
pub fn call(self) -> Result<json::Object, Error> {
let mut response = match self.request.read_response() {
Ok(response) => response,
Err(_request) => return Err(IoError), // Request failed
};
let json = match json::from_reader(&mut response as &mut Reader) {
Ok(json) => json,
Err(error) => return Err(JsonError(error)), // It... wasn't JSON!?
};
// Now let's see if this is an API error or not.
// API errors are of the form {"error":"Invalid API.","errorCode":500}
match json {
json::Object(~mut o) => {
match o.pop(&~"errorCode") {
Some(json::Number(num)) => {
let num = num as uint;
let msg = match o.pop(&~"error")
.expect("JSON had errorCode but not error") {
json::String(s) => s,
uh_oh => fail!("error was {} instead of a string", uh_oh.to_str()),
};
Err(ApiError(msg, num))
},
Some(uh_oh) => fail!("errorCode was {} instead of a number", uh_oh.to_str()),
None => Ok(o),
}
},
// All API responses must be objects.
// If it's not, there's something screwy going on.
_ => fail!("API return value wasn't a JSON object"),
}
}
/// Execute the request as a POST request, sending it through with the given
/// text/html entity body.
///
/// This has the effect that Diffbot will skip requesting the URL and will
/// instead take the passed body as the HTML it is to check. This is mainly
/// useful for non-public websites.
pub fn post_body(mut self, body: &[u8]) -> Result<json::Object, Error> {
self.request.method = Post;
self.request.headers.content_type = Some(MediaType(~"text", ~"html", ~[]));
self.request.headers.content_length = Some(body.len());
// Calling write_headers is an extra and unnecessary safety guard which
// will cause the task to fail if the request has already started to be
// sent (which would render the three statements above ineffectual)
self.request.write_headers();
self.request.write(body);
self.call()
}
}
/// Error code: "unauthorized token"
pub static UNAUTHORIZED_TOKEN: uint = 401;
/// Error code: "requested page not found"
pub static REQUESTED_PAGE_NOT_FOUND: uint = 404;
/// Error code: "your token has exceeded the allowed number of calls, or has
/// otherwise been throttled for API abuse."
pub static TOKEN_EXCEEDED_OR_THROTTLED: uint = 429;
/// Error code: "error processing the page. Specific information will be
/// returned in the JSON response."
pub static ERROR_PROCESSING: uint = 500;
/// Something went wrong with the Diffbot API call.
#[deriving(Eq)]
pub enum Error {
/// An error code returned by the Diffbot API, with message and code.
/// Refer to http://www.diffbot.com/dev/docs/error/ for an explanation of
/// the error codes.
///
/// When comparing the error code, you should use these constants:
///
/// - ``UNAUTHORIZED_TOKEN``: "unauthorized token"
/// - ``REQUESTED_PAGE_NOT_FOUND``: "requested page not found"
/// - ``TOKEN_EXCEEDED_OR_THROTTLED``: "your token has exceeded the allowed
/// number of calls, or has otherwise been throttled for API abuse."
/// - ``ERROR_PROCESSING``: "error processing the page. Specific information
/// will be returned in the JSON response."
ApiError(~str, uint),
/// The JSON was not valid. This is one of those ones that *should* never
/// happen; you know...
///
/// Actually, I can percieve that it might happen if a document returned
/// included invalid UTF-8, but this case has not been tested.
JsonError(json::Error),
/// An I/O error occurred and the condition was trapped somewhere (by you).
IoError,
}
impl ToStr for Error {
fn to_str(&self) -> ~str {
match *self {
ApiError(ref msg, code) => format!("API error {}: {}", code, *msg),
JsonError(ref error) => format!("JSON error: {}", error.to_str()),
IoError => format!("I/O error (already handled)"),
}
}
}
/// Make a simple Diffbot API call.
///
/// For more complex requests, use ``Diffbot`` or ``prepare_request()``.
///
/// Arguments
/// =========
///
/// - ``url`` is the URL that you wish Diffbot to operate upon. If this is a
/// publicly-inaccessible URL, you should use ``post_body()`` on a prepared
/// request instead of ``call()``.
///
/// - ``token`` is the developer's token.
///
/// - ``api`` is the name of the API endpoint, e.g. "article", "product".
///
/// - ``fields`` is the set of fields you want the API call to return; it
/// follows the form specified by the Diffbot API and so should have values
/// like "*", "meta", "querystring", "images(*)".
///
/// - ``version`` is the Diffbot API version number.
pub fn call(url: &Url, token: &str, api: &str, fields: &[&str], version: uint)
-> Result<json::Object, Error> {
prepare_request(url, token, api, fields, version).call()
}
/// Prepare, but do not send, a request.
///
/// This allows you to use some of the more advanced features of the API like
/// setting certain headers for Diffbot to use, or uploading a private document
/// for it.
pub fn prepare_request(url: &Url, token: &str, api: &str, fields: &[&str],
version: uint)
-> Request {
// First of all we must calculate the GET parameters.
let mut query = ~[(~"token", token.to_owned()),
(~"url", url.to_str())];
if fields.len() > 0 {
query.push((~"fields", fields.connect(",")));
}
// Now that we've got that, we can figure out the complete URL.
let url = Url::new(~"http", // scheme
None, // user
~"api.diffbot.com", // host
None, // port
format!("/v{}/{}", version, api), // path
query, // query
None); // fragment
// And with that, we can now make the request. Whee!
Request {
request: RequestWriter::new(Get, url)
}
}
| {
prepare_request(url, self.token, api, fields, self.version)
} | identifier_body |
main.rs | use std::time;
use vulkano::instance::Instance;
use vulkano::instance::PhysicalDevice;
use vulkano::pipeline::GraphicsPipeline;
use vulkano::pipeline::viewport::Viewport;
use vulkano::device::Device;
use vulkano::device::Features;
use vulkano::device::RawDeviceExtensions;
use vulkano::framebuffer::{
Framebuffer,
FramebufferAbstract,
Subpass,
RenderPassAbstract
};
use vulkano::image::SwapchainImage;
use vulkano::sampler::{
Sampler,
SamplerAddressMode,
Filter,
MipmapMode
};
use vulkano::buffer::BufferUsage;
use vulkano::buffer::CpuAccessibleBuffer;
use vulkano::command_buffer::{
AutoCommandBufferBuilder,
DynamicState
};
use vulkano::descriptor::descriptor_set::PersistentDescriptorSet;
use std::sync::Arc;
// use vulkano::pipeline::ComputePipeline;
use vulkano::sync;
use vulkano::sync::{
GpuFuture,
FlushError
};
use vulkano::format::Format;
use vulkano::image::{
Dimensions,
ImmutableImage,
};
// use rand::{
// thread_rng,
// Rng
// };
use image::{
GenericImageView,
};
use vulkano_win::VkSurfaceBuild;
use winit::{
EventsLoop,
WindowBuilder,
Window,
Event,
WindowEvent
};
use vulkano::swapchain;
use vulkano::swapchain::{
AcquireError,
Swapchain,
SurfaceTransform,
PresentMode,
SwapchainCreationError
};
mod cs {
vulkano_shaders::shader! {
ty: "compute",
path: "./src/shader/particle_update.cp.glsl"
}
}
#[derive(Copy, Clone, Debug)]
struct Particle {
pos: [f32; 2],
speed: [f32; 2],
tail: [f32; 2],
prev_pos: [f32; 2],
prev_tail: [f32; 2],
}
// struct ImageDataIterator<P> {
// originalIterator: image::buffer::Pie,
// length: usize
// }
//
// impl<T> ImageDataIterator<T> {
// fn from_dynamic_image(img: &DynamicImage) -> ImageDataIterator<T> {
// let dimensions = img.dimensions();
//
// ImageDataIterator {
// originalIterator: img.to_rgba().pixels(),
// length: ( dimensions.0 * dimensions.1 ) as usize
// }
// }
// }
//
// impl<'a, T> Iterator for ImageDataIterator<'a, T> {
// type Item = [u8; 4];
// fn next(&mut self) -> Option<[u8; 4]> {
// return match self.originalIterator.next() {
// Some(pixel) => {
// let rgba = pixel.2;
// let data: [u8; 4] = [ rgba[0], rgba[1], rgba[2], rgba[3] ];
// return Some(data);
// },
// None => None
// }
// }
// }
//
// impl<'a, T> ExactSizeIterator for ImageDataIterator<'a, T> {
// fn len(&self) -> usize {
// return self.length;
// }
// }
fn | () {
let img = match image::open("./media/autumn.png") {
Ok(image) => image,
Err(err) => panic!("{:?}", err)
};
{ // stdout image info
println!("color {:?}", img.color());
println!("dimensions {:?}", img.dimensions());
// println!("first pixel {:?}", img.pixels().next().unwrap());
// println!("first pixel {:?}", img.pixels().next().map(|item| item.2).unwrap());
}
let instance = {
let inst_exts = vulkano_win::required_extensions();
Instance::new(None, &inst_exts, None).expect("failed to create instance")
};
//TODO: list devices, choose based on user input
for p in PhysicalDevice::enumerate(&instance) {
print!("{}", p.name());
println!(", driver version: {}", p.driver_version());
}
let physical = PhysicalDevice::enumerate(&instance)
.next()
.expect("no device available");
let queue_family = physical
.queue_families()
.find(|&q| q.supports_graphics())
.expect("couldn't find a graphical queue family");
let (device, mut queues) = {
let unraw_dev_exts = vulkano::device::DeviceExtensions {
khr_swapchain: true,
.. vulkano::device::DeviceExtensions::none()
};
let mut all_dev_exts = RawDeviceExtensions::from(&unraw_dev_exts);
all_dev_exts.insert(std::ffi::CString::new("VK_KHR_storage_buffer_storage_class").unwrap());
Device::new(
physical,
&Features::none(),
all_dev_exts,
[(queue_family, 0.5)].iter().cloned(),
)
.expect("failed to create device")
};
let queue = queues.next().unwrap();
// let particles = init_particles_buffer();
// let particles_buffer =
// CpuAccessibleBuffer::from_data(device.clone(), BufferUsage::all(), particles)
// .expect("failed to create buffer");
// let shader = cs::Shader::load(device.clone()).expect("failed to create shader module");
// let compute_pipeline = Arc::new(
// ComputePipeline::new(device.clone(), &shader.main_entry_point(), &())
// .expect("failed to create compute pipeline"),
// );
// let set = Arc::new(
// PersistentDescriptorSet::start(compute_pipeline.clone(), 0)
// .add_buffer(particles_buffer.clone())
// .unwrap()
// .build()
// .unwrap(),
// );
// let command_buffer = AutoCommandBufferBuilder::new(device.clone(), queue.family())
// .unwrap()
// .dispatch([PARTICLE_COUNT as u32 / 32, 1, 1], compute_pipeline.clone(), set.clone(), ())
// .unwrap()
// .build()
// .unwrap();
let mut events_loop = EventsLoop::new();
let surface = WindowBuilder::new().build_vk_surface(&events_loop, instance.clone()).unwrap();
let window = surface.window();
let (mut swapchain, images) = {
let caps = surface.capabilities(physical)
.expect("failed to get surface capabilities");
let usage = caps.supported_usage_flags;
let alpha = caps.supported_composite_alpha.iter().next().unwrap();
let format = caps.supported_formats[0].0;
caps.supported_formats.iter().for_each(|sth| println!("{:?}", sth));
let initial_dimensions = if let Some(dimensions) = window.get_inner_size() {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} else {
return;
};
Swapchain::new(device.clone(), surface.clone(),
caps.min_image_count, format, initial_dimensions, 1, usage, &queue,
SurfaceTransform::Identity, alpha, PresentMode::Fifo, true, None)
.expect("failed to create swapchain")
};
#[derive(Default, Debug, Clone)]
struct Vertex {
position: [f32; 2]
}
let vertex_buffer = {
vulkano::impl_vertex!(Vertex, position);
CpuAccessibleBuffer::from_iter(device.clone(), BufferUsage::all(), [
Vertex { position: [-0.5, -0.5] },
Vertex { position: [ 0.5, -0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, 0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, -0.5] },
].iter().cloned()).unwrap()
};
// texture
let img_dim = img.dimensions();
let (autumn_texture, autumn_texture_future) = match ImmutableImage::from_iter(
img.as_rgba8().unwrap().pixels().map(|rgba| {
let bytes : [u8; 4] = [rgba[0], rgba[1], rgba[2], rgba[3]];
bytes
}),
Dimensions::Dim2d { width: img_dim.0, height: img_dim.1 },
Format::R8G8B8A8Unorm,
queue.clone()
) {
Ok(i) => i,
Err(err) => panic!("{:?}", err)
};
let sampler = Sampler::new(device.clone(), Filter::Linear, Filter::Linear,
MipmapMode::Nearest, SamplerAddressMode::Repeat, SamplerAddressMode::Repeat,
SamplerAddressMode::Repeat, 0.0, 1.0, 0.0, 0.0).unwrap();
mod square_vs {
vulkano_shaders::shader!{
ty: "vertex",
path: "./src/shader/square.vs.glsl"
}
}
mod square_fs {
vulkano_shaders::shader!{
ty: "fragment",
path: "./src/shader/square.fs.glsl"
}
}
let square_vs = square_vs::Shader::load(device.clone()).unwrap();
let square_fs = square_fs::Shader::load(device.clone()).unwrap();
let render_pass = Arc::new(vulkano::single_pass_renderpass!(
device.clone(),
attachments: {
color: {
load: Clear,
store: Store,
format: swapchain.format(),
samples: 1,
}
},
pass: {
color: [color],
depth_stencil: {}
}
).unwrap());
let pipeline = Arc::new(GraphicsPipeline::start()
.vertex_input_single_buffer::<Vertex>()
.vertex_shader(square_vs.main_entry_point(), ())
.triangle_list()
.viewports_dynamic_scissors_irrelevant(1)
.fragment_shader(square_fs.main_entry_point(), ())
.blend_alpha_blending()
.render_pass(Subpass::from(render_pass.clone(), 0).unwrap())
.build(device.clone())
.unwrap());
let set = Arc::new(PersistentDescriptorSet::start(pipeline.clone(), 0)
.add_sampled_image(autumn_texture.clone(), sampler.clone()).unwrap()
.build().unwrap()
);
let mut dynamic_state = DynamicState {
line_width: None,
viewports: None,
scissors: None,
compare_mask: None,
write_mask: None,
reference: None
};
let mut framebuffers = window_size_dependent_setup(&images, render_pass.clone(), &mut dynamic_state);
let mut recreate_swapchain = false;
let mut previous_frame_end = Box::new(sync::now(device.clone()).join(autumn_texture_future)) as Box<dyn GpuFuture>;
let t0 = time::SystemTime::now();
let mut now = t0;
let mut then;
loop {
previous_frame_end.cleanup_finished();
if recreate_swapchain {
let dimensions = if let Some(dimensions) = window.get_inner_size() {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} else {
return ;
};
let (new_swapchain, new_images) = match swapchain.recreate_with_dimension(dimensions) {
Ok(r) => r,
Err(SwapchainCreationError::UnsupportedDimensions) => continue,
Err(err) => panic!("{:?}", err)
};
swapchain = new_swapchain;
framebuffers = window_size_dependent_setup(&new_images, render_pass.clone(), &mut dynamic_state);
recreate_swapchain = false;
}
let (image_num, acquire_future) = match swapchain::acquire_next_image(swapchain.clone(), None){
Ok(r) => r,
Err(AcquireError::OutOfDate) => {
recreate_swapchain = true;
continue;
},
Err(err) => panic!("{:?}", err)
};
if image_num > 2 {
recreate_swapchain = true;
continue;
} // ugly workaround for a situation when image_num is out of bounds
then = now;
now = time::SystemTime::now();
let time = now.duration_since(t0).unwrap().as_millis() as i32;
let dtime = now.duration_since(then).unwrap().as_millis() as i32;
let push_constants = square_fs::ty::PushConstantData {
time,
dtime
};
let clear_values = vec!([0.0, 0.0, 0.0, 1.0].into());
let command_buffer = AutoCommandBufferBuilder::primary_one_time_submit(
device.clone(),
queue.family()
)
.unwrap()
.begin_render_pass(framebuffers[image_num].clone(), false, clear_values)
.unwrap()
.draw(
pipeline.clone(),
&dynamic_state,
vertex_buffer.clone(),
set.clone(),
push_constants)
.unwrap()
.end_render_pass()
.unwrap()
.build()
.unwrap();
let future = previous_frame_end.join(acquire_future)
.then_execute(queue.clone(), command_buffer).unwrap()
.then_swapchain_present(queue.clone(), swapchain.clone(), image_num)
.then_signal_fence_and_flush();
match future {
Ok(future) => {
future.wait(None).unwrap();
previous_frame_end = Box::new(future) as Box<_>;
}
Err(FlushError::OutOfDate) => {
recreate_swapchain = true;
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
Err(e) => {
println!("{:?}", e);
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
}
let mut done = false;
events_loop.poll_events(|event| {
match event {
Event::WindowEvent { event: WindowEvent::CloseRequested, .. } => done = true,
Event::WindowEvent { event: WindowEvent::Resized(_), .. } => recreate_swapchain = true,
_ => ()
}
});
if done { return; }
}
}
// fn init_particles_buffer() -> [Particle; PARTICLE_COUNT] {
// let mut rng = thread_rng();
// let mut particles = [Particle {
// pos: [0.0, 0.0],
// tail: [0.0, 0.0],
// speed: [0.0, 0.0],
// prev_pos: [0.0, 0.0],
// prev_tail: [0.0, 0.0],
// }; PARTICLE_COUNT];
// for i in 0..particles.len() {
// particles[i].pos = [rng.gen_range(-1.0, 1.0), rng.gen_range(-1.0, 1.0)];
// particles[i].tail = particles[i].pos.clone();
// particles[i].speed = [rng.gen_range(-0.1, 0.1), rng.gen_range(-0.1, 0.1)];
// }
// return particles;
// }
fn window_size_dependent_setup(
images: &[Arc<SwapchainImage<Window>>],
render_pass: Arc<dyn RenderPassAbstract + Send + Sync>,
dynamic_state: &mut DynamicState
) -> Vec<Arc<dyn FramebufferAbstract + Send + Sync>> {
let dimensions = images[0].dimensions();
let viewport = Viewport {
origin: [0.0, 0.0],
dimensions: [dimensions[0] as f32, dimensions[1] as f32],
depth_range: 0.0 .. 1.0,
};
dynamic_state.viewports = Some(vec!(viewport));
images.iter().map(|image| {
Arc::new(
Framebuffer::start(render_pass.clone())
.add(image.clone()).unwrap()
.build().unwrap()
) as Arc<dyn FramebufferAbstract + Send + Sync>
}).collect::<Vec<_>>()
}
| main | identifier_name |
main.rs | use std::time;
use vulkano::instance::Instance;
use vulkano::instance::PhysicalDevice;
use vulkano::pipeline::GraphicsPipeline;
use vulkano::pipeline::viewport::Viewport;
use vulkano::device::Device;
use vulkano::device::Features;
use vulkano::device::RawDeviceExtensions;
use vulkano::framebuffer::{
Framebuffer,
FramebufferAbstract,
Subpass,
RenderPassAbstract
};
use vulkano::image::SwapchainImage;
use vulkano::sampler::{
Sampler,
SamplerAddressMode,
Filter,
MipmapMode
};
use vulkano::buffer::BufferUsage;
use vulkano::buffer::CpuAccessibleBuffer;
use vulkano::command_buffer::{
AutoCommandBufferBuilder,
DynamicState
};
use vulkano::descriptor::descriptor_set::PersistentDescriptorSet;
use std::sync::Arc;
// use vulkano::pipeline::ComputePipeline;
use vulkano::sync;
use vulkano::sync::{
GpuFuture,
FlushError
};
use vulkano::format::Format;
use vulkano::image::{
Dimensions,
ImmutableImage,
};
// use rand::{
// thread_rng,
// Rng
// };
use image::{
GenericImageView,
};
use vulkano_win::VkSurfaceBuild;
use winit::{
EventsLoop,
WindowBuilder,
Window,
Event,
WindowEvent
};
use vulkano::swapchain;
use vulkano::swapchain::{
AcquireError,
Swapchain,
SurfaceTransform,
PresentMode,
SwapchainCreationError
};
mod cs {
vulkano_shaders::shader! {
ty: "compute",
path: "./src/shader/particle_update.cp.glsl"
}
}
#[derive(Copy, Clone, Debug)]
struct Particle {
pos: [f32; 2],
speed: [f32; 2],
tail: [f32; 2],
prev_pos: [f32; 2],
prev_tail: [f32; 2],
}
// struct ImageDataIterator<P> {
// originalIterator: image::buffer::Pie,
// length: usize
// }
//
// impl<T> ImageDataIterator<T> {
// fn from_dynamic_image(img: &DynamicImage) -> ImageDataIterator<T> {
// let dimensions = img.dimensions();
//
// ImageDataIterator {
// originalIterator: img.to_rgba().pixels(),
// length: ( dimensions.0 * dimensions.1 ) as usize
// }
// }
// }
//
// impl<'a, T> Iterator for ImageDataIterator<'a, T> {
// type Item = [u8; 4];
// fn next(&mut self) -> Option<[u8; 4]> {
// return match self.originalIterator.next() {
// Some(pixel) => {
// let rgba = pixel.2;
// let data: [u8; 4] = [ rgba[0], rgba[1], rgba[2], rgba[3] ];
// return Some(data);
// },
// None => None
// }
// }
// }
//
// impl<'a, T> ExactSizeIterator for ImageDataIterator<'a, T> {
// fn len(&self) -> usize {
// return self.length;
// }
// }
fn main() {
let img = match image::open("./media/autumn.png") {
Ok(image) => image,
Err(err) => panic!("{:?}", err)
};
{ // stdout image info
println!("color {:?}", img.color());
println!("dimensions {:?}", img.dimensions());
// println!("first pixel {:?}", img.pixels().next().unwrap());
// println!("first pixel {:?}", img.pixels().next().map(|item| item.2).unwrap());
}
let instance = {
let inst_exts = vulkano_win::required_extensions();
Instance::new(None, &inst_exts, None).expect("failed to create instance")
};
//TODO: list devices, choose based on user input
for p in PhysicalDevice::enumerate(&instance) {
print!("{}", p.name());
println!(", driver version: {}", p.driver_version());
}
let physical = PhysicalDevice::enumerate(&instance)
.next()
.expect("no device available");
let queue_family = physical
.queue_families()
.find(|&q| q.supports_graphics())
.expect("couldn't find a graphical queue family");
let (device, mut queues) = {
let unraw_dev_exts = vulkano::device::DeviceExtensions {
khr_swapchain: true,
.. vulkano::device::DeviceExtensions::none()
};
let mut all_dev_exts = RawDeviceExtensions::from(&unraw_dev_exts);
all_dev_exts.insert(std::ffi::CString::new("VK_KHR_storage_buffer_storage_class").unwrap());
Device::new(
physical,
&Features::none(),
all_dev_exts,
[(queue_family, 0.5)].iter().cloned(),
)
.expect("failed to create device")
};
let queue = queues.next().unwrap();
// let particles = init_particles_buffer();
// let particles_buffer =
// CpuAccessibleBuffer::from_data(device.clone(), BufferUsage::all(), particles)
// .expect("failed to create buffer");
// let shader = cs::Shader::load(device.clone()).expect("failed to create shader module");
// let compute_pipeline = Arc::new(
// ComputePipeline::new(device.clone(), &shader.main_entry_point(), &())
// .expect("failed to create compute pipeline"),
// );
// let set = Arc::new(
// PersistentDescriptorSet::start(compute_pipeline.clone(), 0)
// .add_buffer(particles_buffer.clone())
// .unwrap()
// .build()
// .unwrap(),
// );
// let command_buffer = AutoCommandBufferBuilder::new(device.clone(), queue.family())
// .unwrap()
// .dispatch([PARTICLE_COUNT as u32 / 32, 1, 1], compute_pipeline.clone(), set.clone(), ())
// .unwrap()
// .build()
// .unwrap();
let mut events_loop = EventsLoop::new();
let surface = WindowBuilder::new().build_vk_surface(&events_loop, instance.clone()).unwrap();
let window = surface.window();
let (mut swapchain, images) = {
let caps = surface.capabilities(physical)
.expect("failed to get surface capabilities");
let usage = caps.supported_usage_flags;
let alpha = caps.supported_composite_alpha.iter().next().unwrap();
let format = caps.supported_formats[0].0;
caps.supported_formats.iter().for_each(|sth| println!("{:?}", sth));
let initial_dimensions = if let Some(dimensions) = window.get_inner_size() {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} else {
return;
};
Swapchain::new(device.clone(), surface.clone(),
caps.min_image_count, format, initial_dimensions, 1, usage, &queue,
SurfaceTransform::Identity, alpha, PresentMode::Fifo, true, None)
.expect("failed to create swapchain")
};
#[derive(Default, Debug, Clone)]
struct Vertex {
position: [f32; 2]
}
let vertex_buffer = {
vulkano::impl_vertex!(Vertex, position);
CpuAccessibleBuffer::from_iter(device.clone(), BufferUsage::all(), [
Vertex { position: [-0.5, -0.5] },
Vertex { position: [ 0.5, -0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, 0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, -0.5] },
].iter().cloned()).unwrap()
};
// texture
let img_dim = img.dimensions();
let (autumn_texture, autumn_texture_future) = match ImmutableImage::from_iter(
img.as_rgba8().unwrap().pixels().map(|rgba| {
let bytes : [u8; 4] = [rgba[0], rgba[1], rgba[2], rgba[3]];
bytes
}),
Dimensions::Dim2d { width: img_dim.0, height: img_dim.1 },
Format::R8G8B8A8Unorm,
queue.clone()
) {
Ok(i) => i,
Err(err) => panic!("{:?}", err)
};
let sampler = Sampler::new(device.clone(), Filter::Linear, Filter::Linear,
MipmapMode::Nearest, SamplerAddressMode::Repeat, SamplerAddressMode::Repeat,
SamplerAddressMode::Repeat, 0.0, 1.0, 0.0, 0.0).unwrap();
mod square_vs {
vulkano_shaders::shader!{
ty: "vertex",
path: "./src/shader/square.vs.glsl"
}
}
mod square_fs {
vulkano_shaders::shader!{
ty: "fragment",
path: "./src/shader/square.fs.glsl"
}
}
let square_vs = square_vs::Shader::load(device.clone()).unwrap();
let square_fs = square_fs::Shader::load(device.clone()).unwrap();
let render_pass = Arc::new(vulkano::single_pass_renderpass!(
device.clone(),
attachments: {
color: {
load: Clear,
store: Store,
format: swapchain.format(),
samples: 1,
}
},
pass: {
color: [color],
depth_stencil: {}
}
).unwrap());
let pipeline = Arc::new(GraphicsPipeline::start()
.vertex_input_single_buffer::<Vertex>()
.vertex_shader(square_vs.main_entry_point(), ())
.triangle_list()
.viewports_dynamic_scissors_irrelevant(1)
.fragment_shader(square_fs.main_entry_point(), ())
.blend_alpha_blending()
.render_pass(Subpass::from(render_pass.clone(), 0).unwrap())
.build(device.clone())
.unwrap());
let set = Arc::new(PersistentDescriptorSet::start(pipeline.clone(), 0)
.add_sampled_image(autumn_texture.clone(), sampler.clone()).unwrap()
.build().unwrap()
);
let mut dynamic_state = DynamicState {
line_width: None,
viewports: None,
scissors: None,
compare_mask: None,
write_mask: None,
reference: None
};
let mut framebuffers = window_size_dependent_setup(&images, render_pass.clone(), &mut dynamic_state);
let mut recreate_swapchain = false;
let mut previous_frame_end = Box::new(sync::now(device.clone()).join(autumn_texture_future)) as Box<dyn GpuFuture>;
let t0 = time::SystemTime::now();
let mut now = t0;
let mut then;
loop {
previous_frame_end.cleanup_finished();
if recreate_swapchain {
let dimensions = if let Some(dimensions) = window.get_inner_size() {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} else {
return ;
};
let (new_swapchain, new_images) = match swapchain.recreate_with_dimension(dimensions) {
Ok(r) => r,
Err(SwapchainCreationError::UnsupportedDimensions) => continue, |
swapchain = new_swapchain;
framebuffers = window_size_dependent_setup(&new_images, render_pass.clone(), &mut dynamic_state);
recreate_swapchain = false;
}
let (image_num, acquire_future) = match swapchain::acquire_next_image(swapchain.clone(), None){
Ok(r) => r,
Err(AcquireError::OutOfDate) => {
recreate_swapchain = true;
continue;
},
Err(err) => panic!("{:?}", err)
};
if image_num > 2 {
recreate_swapchain = true;
continue;
} // ugly workaround for a situation when image_num is out of bounds
then = now;
now = time::SystemTime::now();
let time = now.duration_since(t0).unwrap().as_millis() as i32;
let dtime = now.duration_since(then).unwrap().as_millis() as i32;
let push_constants = square_fs::ty::PushConstantData {
time,
dtime
};
let clear_values = vec!([0.0, 0.0, 0.0, 1.0].into());
let command_buffer = AutoCommandBufferBuilder::primary_one_time_submit(
device.clone(),
queue.family()
)
.unwrap()
.begin_render_pass(framebuffers[image_num].clone(), false, clear_values)
.unwrap()
.draw(
pipeline.clone(),
&dynamic_state,
vertex_buffer.clone(),
set.clone(),
push_constants)
.unwrap()
.end_render_pass()
.unwrap()
.build()
.unwrap();
let future = previous_frame_end.join(acquire_future)
.then_execute(queue.clone(), command_buffer).unwrap()
.then_swapchain_present(queue.clone(), swapchain.clone(), image_num)
.then_signal_fence_and_flush();
match future {
Ok(future) => {
future.wait(None).unwrap();
previous_frame_end = Box::new(future) as Box<_>;
}
Err(FlushError::OutOfDate) => {
recreate_swapchain = true;
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
Err(e) => {
println!("{:?}", e);
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
}
let mut done = false;
events_loop.poll_events(|event| {
match event {
Event::WindowEvent { event: WindowEvent::CloseRequested, .. } => done = true,
Event::WindowEvent { event: WindowEvent::Resized(_), .. } => recreate_swapchain = true,
_ => ()
}
});
if done { return; }
}
}
// fn init_particles_buffer() -> [Particle; PARTICLE_COUNT] {
// let mut rng = thread_rng();
// let mut particles = [Particle {
// pos: [0.0, 0.0],
// tail: [0.0, 0.0],
// speed: [0.0, 0.0],
// prev_pos: [0.0, 0.0],
// prev_tail: [0.0, 0.0],
// }; PARTICLE_COUNT];
// for i in 0..particles.len() {
// particles[i].pos = [rng.gen_range(-1.0, 1.0), rng.gen_range(-1.0, 1.0)];
// particles[i].tail = particles[i].pos.clone();
// particles[i].speed = [rng.gen_range(-0.1, 0.1), rng.gen_range(-0.1, 0.1)];
// }
// return particles;
// }
fn window_size_dependent_setup(
images: &[Arc<SwapchainImage<Window>>],
render_pass: Arc<dyn RenderPassAbstract + Send + Sync>,
dynamic_state: &mut DynamicState
) -> Vec<Arc<dyn FramebufferAbstract + Send + Sync>> {
let dimensions = images[0].dimensions();
let viewport = Viewport {
origin: [0.0, 0.0],
dimensions: [dimensions[0] as f32, dimensions[1] as f32],
depth_range: 0.0 .. 1.0,
};
dynamic_state.viewports = Some(vec!(viewport));
images.iter().map(|image| {
Arc::new(
Framebuffer::start(render_pass.clone())
.add(image.clone()).unwrap()
.build().unwrap()
) as Arc<dyn FramebufferAbstract + Send + Sync>
}).collect::<Vec<_>>()
} | Err(err) => panic!("{:?}", err)
}; | random_line_split |
main.rs | use std::time;
use vulkano::instance::Instance;
use vulkano::instance::PhysicalDevice;
use vulkano::pipeline::GraphicsPipeline;
use vulkano::pipeline::viewport::Viewport;
use vulkano::device::Device;
use vulkano::device::Features;
use vulkano::device::RawDeviceExtensions;
use vulkano::framebuffer::{
Framebuffer,
FramebufferAbstract,
Subpass,
RenderPassAbstract
};
use vulkano::image::SwapchainImage;
use vulkano::sampler::{
Sampler,
SamplerAddressMode,
Filter,
MipmapMode
};
use vulkano::buffer::BufferUsage;
use vulkano::buffer::CpuAccessibleBuffer;
use vulkano::command_buffer::{
AutoCommandBufferBuilder,
DynamicState
};
use vulkano::descriptor::descriptor_set::PersistentDescriptorSet;
use std::sync::Arc;
// use vulkano::pipeline::ComputePipeline;
use vulkano::sync;
use vulkano::sync::{
GpuFuture,
FlushError
};
use vulkano::format::Format;
use vulkano::image::{
Dimensions,
ImmutableImage,
};
// use rand::{
// thread_rng,
// Rng
// };
use image::{
GenericImageView,
};
use vulkano_win::VkSurfaceBuild;
use winit::{
EventsLoop,
WindowBuilder,
Window,
Event,
WindowEvent
};
use vulkano::swapchain;
use vulkano::swapchain::{
AcquireError,
Swapchain,
SurfaceTransform,
PresentMode,
SwapchainCreationError
};
mod cs {
vulkano_shaders::shader! {
ty: "compute",
path: "./src/shader/particle_update.cp.glsl"
}
}
#[derive(Copy, Clone, Debug)]
struct Particle {
pos: [f32; 2],
speed: [f32; 2],
tail: [f32; 2],
prev_pos: [f32; 2],
prev_tail: [f32; 2],
}
// struct ImageDataIterator<P> {
// originalIterator: image::buffer::Pie,
// length: usize
// }
//
// impl<T> ImageDataIterator<T> {
// fn from_dynamic_image(img: &DynamicImage) -> ImageDataIterator<T> {
// let dimensions = img.dimensions();
//
// ImageDataIterator {
// originalIterator: img.to_rgba().pixels(),
// length: ( dimensions.0 * dimensions.1 ) as usize
// }
// }
// }
//
// impl<'a, T> Iterator for ImageDataIterator<'a, T> {
// type Item = [u8; 4];
// fn next(&mut self) -> Option<[u8; 4]> {
// return match self.originalIterator.next() {
// Some(pixel) => {
// let rgba = pixel.2;
// let data: [u8; 4] = [ rgba[0], rgba[1], rgba[2], rgba[3] ];
// return Some(data);
// },
// None => None
// }
// }
// }
//
// impl<'a, T> ExactSizeIterator for ImageDataIterator<'a, T> {
// fn len(&self) -> usize {
// return self.length;
// }
// }
fn main() |
// fn init_particles_buffer() -> [Particle; PARTICLE_COUNT] {
// let mut rng = thread_rng();
// let mut particles = [Particle {
// pos: [0.0, 0.0],
// tail: [0.0, 0.0],
// speed: [0.0, 0.0],
// prev_pos: [0.0, 0.0],
// prev_tail: [0.0, 0.0],
// }; PARTICLE_COUNT];
// for i in 0..particles.len() {
// particles[i].pos = [rng.gen_range(-1.0, 1.0), rng.gen_range(-1.0, 1.0)];
// particles[i].tail = particles[i].pos.clone();
// particles[i].speed = [rng.gen_range(-0.1, 0.1), rng.gen_range(-0.1, 0.1)];
// }
// return particles;
// }
fn window_size_dependent_setup(
images: &[Arc<SwapchainImage<Window>>],
render_pass: Arc<dyn RenderPassAbstract + Send + Sync>,
dynamic_state: &mut DynamicState
) -> Vec<Arc<dyn FramebufferAbstract + Send + Sync>> {
let dimensions = images[0].dimensions();
let viewport = Viewport {
origin: [0.0, 0.0],
dimensions: [dimensions[0] as f32, dimensions[1] as f32],
depth_range: 0.0 .. 1.0,
};
dynamic_state.viewports = Some(vec!(viewport));
images.iter().map(|image| {
Arc::new(
Framebuffer::start(render_pass.clone())
.add(image.clone()).unwrap()
.build().unwrap()
) as Arc<dyn FramebufferAbstract + Send + Sync>
}).collect::<Vec<_>>()
}
| {
let img = match image::open("./media/autumn.png") {
Ok(image) => image,
Err(err) => panic!("{:?}", err)
};
{ // stdout image info
println!("color {:?}", img.color());
println!("dimensions {:?}", img.dimensions());
// println!("first pixel {:?}", img.pixels().next().unwrap());
// println!("first pixel {:?}", img.pixels().next().map(|item| item.2).unwrap());
}
let instance = {
let inst_exts = vulkano_win::required_extensions();
Instance::new(None, &inst_exts, None).expect("failed to create instance")
};
//TODO: list devices, choose based on user input
for p in PhysicalDevice::enumerate(&instance) {
print!("{}", p.name());
println!(", driver version: {}", p.driver_version());
}
let physical = PhysicalDevice::enumerate(&instance)
.next()
.expect("no device available");
let queue_family = physical
.queue_families()
.find(|&q| q.supports_graphics())
.expect("couldn't find a graphical queue family");
let (device, mut queues) = {
let unraw_dev_exts = vulkano::device::DeviceExtensions {
khr_swapchain: true,
.. vulkano::device::DeviceExtensions::none()
};
let mut all_dev_exts = RawDeviceExtensions::from(&unraw_dev_exts);
all_dev_exts.insert(std::ffi::CString::new("VK_KHR_storage_buffer_storage_class").unwrap());
Device::new(
physical,
&Features::none(),
all_dev_exts,
[(queue_family, 0.5)].iter().cloned(),
)
.expect("failed to create device")
};
let queue = queues.next().unwrap();
// let particles = init_particles_buffer();
// let particles_buffer =
// CpuAccessibleBuffer::from_data(device.clone(), BufferUsage::all(), particles)
// .expect("failed to create buffer");
// let shader = cs::Shader::load(device.clone()).expect("failed to create shader module");
// let compute_pipeline = Arc::new(
// ComputePipeline::new(device.clone(), &shader.main_entry_point(), &())
// .expect("failed to create compute pipeline"),
// );
// let set = Arc::new(
// PersistentDescriptorSet::start(compute_pipeline.clone(), 0)
// .add_buffer(particles_buffer.clone())
// .unwrap()
// .build()
// .unwrap(),
// );
// let command_buffer = AutoCommandBufferBuilder::new(device.clone(), queue.family())
// .unwrap()
// .dispatch([PARTICLE_COUNT as u32 / 32, 1, 1], compute_pipeline.clone(), set.clone(), ())
// .unwrap()
// .build()
// .unwrap();
let mut events_loop = EventsLoop::new();
let surface = WindowBuilder::new().build_vk_surface(&events_loop, instance.clone()).unwrap();
let window = surface.window();
let (mut swapchain, images) = {
let caps = surface.capabilities(physical)
.expect("failed to get surface capabilities");
let usage = caps.supported_usage_flags;
let alpha = caps.supported_composite_alpha.iter().next().unwrap();
let format = caps.supported_formats[0].0;
caps.supported_formats.iter().for_each(|sth| println!("{:?}", sth));
let initial_dimensions = if let Some(dimensions) = window.get_inner_size() {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} else {
return;
};
Swapchain::new(device.clone(), surface.clone(),
caps.min_image_count, format, initial_dimensions, 1, usage, &queue,
SurfaceTransform::Identity, alpha, PresentMode::Fifo, true, None)
.expect("failed to create swapchain")
};
#[derive(Default, Debug, Clone)]
struct Vertex {
position: [f32; 2]
}
let vertex_buffer = {
vulkano::impl_vertex!(Vertex, position);
CpuAccessibleBuffer::from_iter(device.clone(), BufferUsage::all(), [
Vertex { position: [-0.5, -0.5] },
Vertex { position: [ 0.5, -0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, 0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, -0.5] },
].iter().cloned()).unwrap()
};
// texture
let img_dim = img.dimensions();
let (autumn_texture, autumn_texture_future) = match ImmutableImage::from_iter(
img.as_rgba8().unwrap().pixels().map(|rgba| {
let bytes : [u8; 4] = [rgba[0], rgba[1], rgba[2], rgba[3]];
bytes
}),
Dimensions::Dim2d { width: img_dim.0, height: img_dim.1 },
Format::R8G8B8A8Unorm,
queue.clone()
) {
Ok(i) => i,
Err(err) => panic!("{:?}", err)
};
let sampler = Sampler::new(device.clone(), Filter::Linear, Filter::Linear,
MipmapMode::Nearest, SamplerAddressMode::Repeat, SamplerAddressMode::Repeat,
SamplerAddressMode::Repeat, 0.0, 1.0, 0.0, 0.0).unwrap();
mod square_vs {
vulkano_shaders::shader!{
ty: "vertex",
path: "./src/shader/square.vs.glsl"
}
}
mod square_fs {
vulkano_shaders::shader!{
ty: "fragment",
path: "./src/shader/square.fs.glsl"
}
}
let square_vs = square_vs::Shader::load(device.clone()).unwrap();
let square_fs = square_fs::Shader::load(device.clone()).unwrap();
let render_pass = Arc::new(vulkano::single_pass_renderpass!(
device.clone(),
attachments: {
color: {
load: Clear,
store: Store,
format: swapchain.format(),
samples: 1,
}
},
pass: {
color: [color],
depth_stencil: {}
}
).unwrap());
let pipeline = Arc::new(GraphicsPipeline::start()
.vertex_input_single_buffer::<Vertex>()
.vertex_shader(square_vs.main_entry_point(), ())
.triangle_list()
.viewports_dynamic_scissors_irrelevant(1)
.fragment_shader(square_fs.main_entry_point(), ())
.blend_alpha_blending()
.render_pass(Subpass::from(render_pass.clone(), 0).unwrap())
.build(device.clone())
.unwrap());
let set = Arc::new(PersistentDescriptorSet::start(pipeline.clone(), 0)
.add_sampled_image(autumn_texture.clone(), sampler.clone()).unwrap()
.build().unwrap()
);
let mut dynamic_state = DynamicState {
line_width: None,
viewports: None,
scissors: None,
compare_mask: None,
write_mask: None,
reference: None
};
let mut framebuffers = window_size_dependent_setup(&images, render_pass.clone(), &mut dynamic_state);
let mut recreate_swapchain = false;
let mut previous_frame_end = Box::new(sync::now(device.clone()).join(autumn_texture_future)) as Box<dyn GpuFuture>;
let t0 = time::SystemTime::now();
let mut now = t0;
let mut then;
loop {
previous_frame_end.cleanup_finished();
if recreate_swapchain {
let dimensions = if let Some(dimensions) = window.get_inner_size() {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} else {
return ;
};
let (new_swapchain, new_images) = match swapchain.recreate_with_dimension(dimensions) {
Ok(r) => r,
Err(SwapchainCreationError::UnsupportedDimensions) => continue,
Err(err) => panic!("{:?}", err)
};
swapchain = new_swapchain;
framebuffers = window_size_dependent_setup(&new_images, render_pass.clone(), &mut dynamic_state);
recreate_swapchain = false;
}
let (image_num, acquire_future) = match swapchain::acquire_next_image(swapchain.clone(), None){
Ok(r) => r,
Err(AcquireError::OutOfDate) => {
recreate_swapchain = true;
continue;
},
Err(err) => panic!("{:?}", err)
};
if image_num > 2 {
recreate_swapchain = true;
continue;
} // ugly workaround for a situation when image_num is out of bounds
then = now;
now = time::SystemTime::now();
let time = now.duration_since(t0).unwrap().as_millis() as i32;
let dtime = now.duration_since(then).unwrap().as_millis() as i32;
let push_constants = square_fs::ty::PushConstantData {
time,
dtime
};
let clear_values = vec!([0.0, 0.0, 0.0, 1.0].into());
let command_buffer = AutoCommandBufferBuilder::primary_one_time_submit(
device.clone(),
queue.family()
)
.unwrap()
.begin_render_pass(framebuffers[image_num].clone(), false, clear_values)
.unwrap()
.draw(
pipeline.clone(),
&dynamic_state,
vertex_buffer.clone(),
set.clone(),
push_constants)
.unwrap()
.end_render_pass()
.unwrap()
.build()
.unwrap();
let future = previous_frame_end.join(acquire_future)
.then_execute(queue.clone(), command_buffer).unwrap()
.then_swapchain_present(queue.clone(), swapchain.clone(), image_num)
.then_signal_fence_and_flush();
match future {
Ok(future) => {
future.wait(None).unwrap();
previous_frame_end = Box::new(future) as Box<_>;
}
Err(FlushError::OutOfDate) => {
recreate_swapchain = true;
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
Err(e) => {
println!("{:?}", e);
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
}
let mut done = false;
events_loop.poll_events(|event| {
match event {
Event::WindowEvent { event: WindowEvent::CloseRequested, .. } => done = true,
Event::WindowEvent { event: WindowEvent::Resized(_), .. } => recreate_swapchain = true,
_ => ()
}
});
if done { return; }
}
} | identifier_body |
main.rs | use std::time;
use vulkano::instance::Instance;
use vulkano::instance::PhysicalDevice;
use vulkano::pipeline::GraphicsPipeline;
use vulkano::pipeline::viewport::Viewport;
use vulkano::device::Device;
use vulkano::device::Features;
use vulkano::device::RawDeviceExtensions;
use vulkano::framebuffer::{
Framebuffer,
FramebufferAbstract,
Subpass,
RenderPassAbstract
};
use vulkano::image::SwapchainImage;
use vulkano::sampler::{
Sampler,
SamplerAddressMode,
Filter,
MipmapMode
};
use vulkano::buffer::BufferUsage;
use vulkano::buffer::CpuAccessibleBuffer;
use vulkano::command_buffer::{
AutoCommandBufferBuilder,
DynamicState
};
use vulkano::descriptor::descriptor_set::PersistentDescriptorSet;
use std::sync::Arc;
// use vulkano::pipeline::ComputePipeline;
use vulkano::sync;
use vulkano::sync::{
GpuFuture,
FlushError
};
use vulkano::format::Format;
use vulkano::image::{
Dimensions,
ImmutableImage,
};
// use rand::{
// thread_rng,
// Rng
// };
use image::{
GenericImageView,
};
use vulkano_win::VkSurfaceBuild;
use winit::{
EventsLoop,
WindowBuilder,
Window,
Event,
WindowEvent
};
use vulkano::swapchain;
use vulkano::swapchain::{
AcquireError,
Swapchain,
SurfaceTransform,
PresentMode,
SwapchainCreationError
};
mod cs {
vulkano_shaders::shader! {
ty: "compute",
path: "./src/shader/particle_update.cp.glsl"
}
}
#[derive(Copy, Clone, Debug)]
struct Particle {
pos: [f32; 2],
speed: [f32; 2],
tail: [f32; 2],
prev_pos: [f32; 2],
prev_tail: [f32; 2],
}
// struct ImageDataIterator<P> {
// originalIterator: image::buffer::Pie,
// length: usize
// }
//
// impl<T> ImageDataIterator<T> {
// fn from_dynamic_image(img: &DynamicImage) -> ImageDataIterator<T> {
// let dimensions = img.dimensions();
//
// ImageDataIterator {
// originalIterator: img.to_rgba().pixels(),
// length: ( dimensions.0 * dimensions.1 ) as usize
// }
// }
// }
//
// impl<'a, T> Iterator for ImageDataIterator<'a, T> {
// type Item = [u8; 4];
// fn next(&mut self) -> Option<[u8; 4]> {
// return match self.originalIterator.next() {
// Some(pixel) => {
// let rgba = pixel.2;
// let data: [u8; 4] = [ rgba[0], rgba[1], rgba[2], rgba[3] ];
// return Some(data);
// },
// None => None
// }
// }
// }
//
// impl<'a, T> ExactSizeIterator for ImageDataIterator<'a, T> {
// fn len(&self) -> usize {
// return self.length;
// }
// }
fn main() {
let img = match image::open("./media/autumn.png") {
Ok(image) => image,
Err(err) => panic!("{:?}", err)
};
{ // stdout image info
println!("color {:?}", img.color());
println!("dimensions {:?}", img.dimensions());
// println!("first pixel {:?}", img.pixels().next().unwrap());
// println!("first pixel {:?}", img.pixels().next().map(|item| item.2).unwrap());
}
let instance = {
let inst_exts = vulkano_win::required_extensions();
Instance::new(None, &inst_exts, None).expect("failed to create instance")
};
//TODO: list devices, choose based on user input
for p in PhysicalDevice::enumerate(&instance) {
print!("{}", p.name());
println!(", driver version: {}", p.driver_version());
}
let physical = PhysicalDevice::enumerate(&instance)
.next()
.expect("no device available");
let queue_family = physical
.queue_families()
.find(|&q| q.supports_graphics())
.expect("couldn't find a graphical queue family");
let (device, mut queues) = {
let unraw_dev_exts = vulkano::device::DeviceExtensions {
khr_swapchain: true,
.. vulkano::device::DeviceExtensions::none()
};
let mut all_dev_exts = RawDeviceExtensions::from(&unraw_dev_exts);
all_dev_exts.insert(std::ffi::CString::new("VK_KHR_storage_buffer_storage_class").unwrap());
Device::new(
physical,
&Features::none(),
all_dev_exts,
[(queue_family, 0.5)].iter().cloned(),
)
.expect("failed to create device")
};
let queue = queues.next().unwrap();
// let particles = init_particles_buffer();
// let particles_buffer =
// CpuAccessibleBuffer::from_data(device.clone(), BufferUsage::all(), particles)
// .expect("failed to create buffer");
// let shader = cs::Shader::load(device.clone()).expect("failed to create shader module");
// let compute_pipeline = Arc::new(
// ComputePipeline::new(device.clone(), &shader.main_entry_point(), &())
// .expect("failed to create compute pipeline"),
// );
// let set = Arc::new(
// PersistentDescriptorSet::start(compute_pipeline.clone(), 0)
// .add_buffer(particles_buffer.clone())
// .unwrap()
// .build()
// .unwrap(),
// );
// let command_buffer = AutoCommandBufferBuilder::new(device.clone(), queue.family())
// .unwrap()
// .dispatch([PARTICLE_COUNT as u32 / 32, 1, 1], compute_pipeline.clone(), set.clone(), ())
// .unwrap()
// .build()
// .unwrap();
let mut events_loop = EventsLoop::new();
let surface = WindowBuilder::new().build_vk_surface(&events_loop, instance.clone()).unwrap();
let window = surface.window();
let (mut swapchain, images) = {
let caps = surface.capabilities(physical)
.expect("failed to get surface capabilities");
let usage = caps.supported_usage_flags;
let alpha = caps.supported_composite_alpha.iter().next().unwrap();
let format = caps.supported_formats[0].0;
caps.supported_formats.iter().for_each(|sth| println!("{:?}", sth));
let initial_dimensions = if let Some(dimensions) = window.get_inner_size() {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} else {
return;
};
Swapchain::new(device.clone(), surface.clone(),
caps.min_image_count, format, initial_dimensions, 1, usage, &queue,
SurfaceTransform::Identity, alpha, PresentMode::Fifo, true, None)
.expect("failed to create swapchain")
};
#[derive(Default, Debug, Clone)]
struct Vertex {
position: [f32; 2]
}
let vertex_buffer = {
vulkano::impl_vertex!(Vertex, position);
CpuAccessibleBuffer::from_iter(device.clone(), BufferUsage::all(), [
Vertex { position: [-0.5, -0.5] },
Vertex { position: [ 0.5, -0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, 0.5] },
Vertex { position: [-0.5, 0.5] },
Vertex { position: [ 0.5, -0.5] },
].iter().cloned()).unwrap()
};
// texture
let img_dim = img.dimensions();
let (autumn_texture, autumn_texture_future) = match ImmutableImage::from_iter(
img.as_rgba8().unwrap().pixels().map(|rgba| {
let bytes : [u8; 4] = [rgba[0], rgba[1], rgba[2], rgba[3]];
bytes
}),
Dimensions::Dim2d { width: img_dim.0, height: img_dim.1 },
Format::R8G8B8A8Unorm,
queue.clone()
) {
Ok(i) => i,
Err(err) => panic!("{:?}", err)
};
let sampler = Sampler::new(device.clone(), Filter::Linear, Filter::Linear,
MipmapMode::Nearest, SamplerAddressMode::Repeat, SamplerAddressMode::Repeat,
SamplerAddressMode::Repeat, 0.0, 1.0, 0.0, 0.0).unwrap();
mod square_vs {
vulkano_shaders::shader!{
ty: "vertex",
path: "./src/shader/square.vs.glsl"
}
}
mod square_fs {
vulkano_shaders::shader!{
ty: "fragment",
path: "./src/shader/square.fs.glsl"
}
}
let square_vs = square_vs::Shader::load(device.clone()).unwrap();
let square_fs = square_fs::Shader::load(device.clone()).unwrap();
let render_pass = Arc::new(vulkano::single_pass_renderpass!(
device.clone(),
attachments: {
color: {
load: Clear,
store: Store,
format: swapchain.format(),
samples: 1,
}
},
pass: {
color: [color],
depth_stencil: {}
}
).unwrap());
let pipeline = Arc::new(GraphicsPipeline::start()
.vertex_input_single_buffer::<Vertex>()
.vertex_shader(square_vs.main_entry_point(), ())
.triangle_list()
.viewports_dynamic_scissors_irrelevant(1)
.fragment_shader(square_fs.main_entry_point(), ())
.blend_alpha_blending()
.render_pass(Subpass::from(render_pass.clone(), 0).unwrap())
.build(device.clone())
.unwrap());
let set = Arc::new(PersistentDescriptorSet::start(pipeline.clone(), 0)
.add_sampled_image(autumn_texture.clone(), sampler.clone()).unwrap()
.build().unwrap()
);
let mut dynamic_state = DynamicState {
line_width: None,
viewports: None,
scissors: None,
compare_mask: None,
write_mask: None,
reference: None
};
let mut framebuffers = window_size_dependent_setup(&images, render_pass.clone(), &mut dynamic_state);
let mut recreate_swapchain = false;
let mut previous_frame_end = Box::new(sync::now(device.clone()).join(autumn_texture_future)) as Box<dyn GpuFuture>;
let t0 = time::SystemTime::now();
let mut now = t0;
let mut then;
loop {
previous_frame_end.cleanup_finished();
if recreate_swapchain {
let dimensions = if let Some(dimensions) = window.get_inner_size() | else {
return ;
};
let (new_swapchain, new_images) = match swapchain.recreate_with_dimension(dimensions) {
Ok(r) => r,
Err(SwapchainCreationError::UnsupportedDimensions) => continue,
Err(err) => panic!("{:?}", err)
};
swapchain = new_swapchain;
framebuffers = window_size_dependent_setup(&new_images, render_pass.clone(), &mut dynamic_state);
recreate_swapchain = false;
}
let (image_num, acquire_future) = match swapchain::acquire_next_image(swapchain.clone(), None){
Ok(r) => r,
Err(AcquireError::OutOfDate) => {
recreate_swapchain = true;
continue;
},
Err(err) => panic!("{:?}", err)
};
if image_num > 2 {
recreate_swapchain = true;
continue;
} // ugly workaround for a situation when image_num is out of bounds
then = now;
now = time::SystemTime::now();
let time = now.duration_since(t0).unwrap().as_millis() as i32;
let dtime = now.duration_since(then).unwrap().as_millis() as i32;
let push_constants = square_fs::ty::PushConstantData {
time,
dtime
};
let clear_values = vec!([0.0, 0.0, 0.0, 1.0].into());
let command_buffer = AutoCommandBufferBuilder::primary_one_time_submit(
device.clone(),
queue.family()
)
.unwrap()
.begin_render_pass(framebuffers[image_num].clone(), false, clear_values)
.unwrap()
.draw(
pipeline.clone(),
&dynamic_state,
vertex_buffer.clone(),
set.clone(),
push_constants)
.unwrap()
.end_render_pass()
.unwrap()
.build()
.unwrap();
let future = previous_frame_end.join(acquire_future)
.then_execute(queue.clone(), command_buffer).unwrap()
.then_swapchain_present(queue.clone(), swapchain.clone(), image_num)
.then_signal_fence_and_flush();
match future {
Ok(future) => {
future.wait(None).unwrap();
previous_frame_end = Box::new(future) as Box<_>;
}
Err(FlushError::OutOfDate) => {
recreate_swapchain = true;
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
Err(e) => {
println!("{:?}", e);
previous_frame_end = Box::new(sync::now(device.clone())) as Box<_>;
}
}
let mut done = false;
events_loop.poll_events(|event| {
match event {
Event::WindowEvent { event: WindowEvent::CloseRequested, .. } => done = true,
Event::WindowEvent { event: WindowEvent::Resized(_), .. } => recreate_swapchain = true,
_ => ()
}
});
if done { return; }
}
}
// fn init_particles_buffer() -> [Particle; PARTICLE_COUNT] {
// let mut rng = thread_rng();
// let mut particles = [Particle {
// pos: [0.0, 0.0],
// tail: [0.0, 0.0],
// speed: [0.0, 0.0],
// prev_pos: [0.0, 0.0],
// prev_tail: [0.0, 0.0],
// }; PARTICLE_COUNT];
// for i in 0..particles.len() {
// particles[i].pos = [rng.gen_range(-1.0, 1.0), rng.gen_range(-1.0, 1.0)];
// particles[i].tail = particles[i].pos.clone();
// particles[i].speed = [rng.gen_range(-0.1, 0.1), rng.gen_range(-0.1, 0.1)];
// }
// return particles;
// }
fn window_size_dependent_setup(
images: &[Arc<SwapchainImage<Window>>],
render_pass: Arc<dyn RenderPassAbstract + Send + Sync>,
dynamic_state: &mut DynamicState
) -> Vec<Arc<dyn FramebufferAbstract + Send + Sync>> {
let dimensions = images[0].dimensions();
let viewport = Viewport {
origin: [0.0, 0.0],
dimensions: [dimensions[0] as f32, dimensions[1] as f32],
depth_range: 0.0 .. 1.0,
};
dynamic_state.viewports = Some(vec!(viewport));
images.iter().map(|image| {
Arc::new(
Framebuffer::start(render_pass.clone())
.add(image.clone()).unwrap()
.build().unwrap()
) as Arc<dyn FramebufferAbstract + Send + Sync>
}).collect::<Vec<_>>()
}
| {
let dimensions: (u32, u32) = dimensions.to_physical(window.get_hidpi_factor()).into();
[dimensions.0, dimensions.1]
} | conditional_block |
body.rs | #![allow(dead_code)]
#![allow(unused_doc_comments)]
/**
* body.rs contains the Body struct and implements methods for it. A body struct contains only the
* position and velocity vectors of the body, other parameters are calculated using methods. A body
* is instantiated using using the Body::new() method, which also determines the type of orbit the
* body has at the same time. orbit_type does not have to be given manually.
*/
use nalgebra::{Matrix3, Vector3};
use std::f64::consts::PI;
use colored::*;
const DAYTOSEC: f64 = 24.0 * 3600.0;
const SOLARGM: f64 = 1.328905188132376e11;
const PI2: f64 = 2.0 * PI;
/**
* OrbitType is used to abstract away some of the functions that depend on
* the type of orbit the body is in, like kepler's equation. That way, you
* can call one function and it will return the correct value
*/
#[derive(Debug)]
pub enum OrbitType {
Circular,
Elliptic,
Parabolic,
Hyperbolic,
}
impl OrbitType {
/// Return the orbit type given the eccentricity, Body::new
/// uses this function to set the orbit type when an instance
/// is constructed
pub fn new(eccentricity: f64) -> OrbitType {
if eccentricity == 0.0 {
return OrbitType::Circular;
} else if eccentricity < 1.0 && eccentricity > 0.0 {
return OrbitType::Elliptic;
} else if eccentricity == 1.0 {
return OrbitType::Parabolic;
} else {
return OrbitType::Hyperbolic;
}
}
}
#[derive(Debug)]
pub struct Body {
pub position: Vector3<f64>,
pub velocity: Vector3<f64>,
pub orbit_type: OrbitType,
}
/* Adds methods to Body struct */
impl Body {
pub fn new(position: Vector3<f64>, velocity: Vector3<f64>) -> Body {
// h and e are used for determining what kind of orbit the body is currently in
let h = position.cross(&velocity);
let e = ((velocity.cross(&h) / SOLARGM) - position.normalize()).norm();
Body {
position: position,
velocity: velocity,
orbit_type: OrbitType::new(e),
}
}
pub fn radial_velocity(&self) -> Vector3<f64> {
(self.velocity.dot(&self.position) / self.position.norm_squared()) * self.position
}
pub fn tangential_velocity(&self) -> Vector3<f64> {
self.omega().cross(&self.position)
}
pub fn true_anomaly(&self) -> f64 {
let e_vec = self.eccentricity_vector();
let posit = self.position.normalize();
let val = e_vec.dot(&posit) / (e_vec.norm() * posit.norm());
if posit.dot(&self.velocity.normalize()) < 0.0 {
return 2.0 * PI - val.acos();
} else {
return val.acos();
}
}
/* points from focus to perigee if I'm not mistaken */
pub fn eccentricity_vector(&self) -> Vector3<f64> {
let veloc = self.velocity;
let posit = self.position;
let h = self.angular_momentum();
(veloc.cross(&h) / SOLARGM) - posit.normalize()
}
pub fn angular_momentum(&self) -> Vector3<f64> {
self.position.cross(&self.velocity)
}
pub fn total_energy(&self) -> f64 {
let posit = self.position.norm();
let veloc = self.velocity.norm();
0.5 * veloc.powi(2) - (SOLARGM / posit)
}
pub fn omega(&self) -> Vector3<f64> {
self.angular_momentum() / self.position.norm_squared()
}
pub fn frame_rotation_rate(&self) -> f64 {
self.omega().norm()
}
pub fn position_at_angle(&self, angle: f64) -> Vector3<f64> {
let e = self.eccentricity();
let numer = self.angular_momentum().norm_squared() / SOLARGM;
let denom = 1_f64 + (e * (angle).cos());
let radius = numer / denom;
Vector3::new(radius, 0.0, 0.0)
}
pub fn velocity_at_angle(&self, angle: f64) -> Vector3<f64> {
let p = self.orbital_parameter();
let e = self.eccentricity();
let h = self.angular_momentum().norm_squared();
Vector3::new(
(h / p) * e * angle.sin(),
(h / p) * (1_f64 + e * angle.cos()),
0.0,
)
}
pub fn | (&self, angle: f64) -> (Vector3<f64>, Vector3<f64>) {
let r = self.position_at_angle(angle);
let v = self.velocity_at_angle(angle);
let tht = angle - self.true_anomaly();
let trans = Matrix3::from_rows(&[
Vector3::new(tht.cos(), -tht.sin(), 0.0).transpose(),
Vector3::new(tht.sin(), tht.cos(), 0.0).transpose(),
Vector3::new(0.0, 0.0, 1.0).transpose(),
]);
(trans * r, trans * v)
}
// Angle to other body, keep getting the wrong thing anyway, tried everything
pub fn angle_to(&self, other: &Body) -> f64 {
(self.position.dot(&other.position) / (self.position.norm() * other.position.norm())).acos()
}
/* Return a transformation matrix constructed from body's orbit in inertial frame */
pub fn make_frame(&self) -> Matrix3<f64> {
let e_r = self.position.normalize();
let e_h = self.angular_momentum().normalize();
let e_tht = e_h.cross(&e_r);
Matrix3::from_rows(&[e_r.transpose(), e_tht.transpose(), e_h.transpose()])
}
pub fn semi_major_axis(&self) -> f64 {
let ang_moment = self.angular_momentum().norm();
let e = self.eccentricity();
ang_moment.powi(2) / (SOLARGM * (1_f64 - e.powi(2)))
}
pub fn orbital_period(&self) -> f64 {
2_f64 * PI * (self.semi_major_axis().powi(3) / SOLARGM).sqrt()
}
pub fn orbital_parameter(&self) -> f64 {
let e = self.eccentricity();
self.semi_major_axis() * (1.0 - e.powi(2))
}
pub fn eccentric_anomaly(&self) -> f64 {
let e = self.eccentricity();
let theta = self.true_anomaly();
2.0 * ((theta / 2.0).tan() / ((1.0 + e) / (1.0 - e)).sqrt()).atan()
}
pub fn time_since_periapsis(&self) -> f64 {
let t_anom = self.true_anomaly();
let e_anom = self.true_to_eccentric(t_anom);
let a = self.semi_major_axis();
let e = self.eccentricity();
(a.powi(3) / SOLARGM).sqrt() * (e_anom - e * e_anom.sin())
}
pub fn eccentricity(&self) -> f64 {
self.eccentricity_vector().norm()
}
pub fn inclination(&self) -> f64 {
let h = self.angular_momentum();
(h[2] / h.norm()).acos() // h[2] is the z component of the vector
}
pub fn ascending_node(&self) -> Vector3<f64> {
let k = Vector3::new(0.0, 0.0, 1.0);
k.cross(&self.angular_momentum())
}
pub fn argument_of_periapsis(&self) -> f64 {
let n = self.ascending_node();
let e = self.eccentricity_vector();
let omega = (n.dot(&e) / (n.norm() * e.norm())).acos();
if e[2] < 0.0 {
PI2 - omega
} else {
omega
}
}
pub fn argument_of_ascending_node(&self) -> f64 {
let n = self.ascending_node();
let n_x = n[0];
let n_y = n[1];
if n_y >= 0.0 {
(n_x / n.norm()).acos()
} else {
PI2 - (n_x / n.norm()).acos()
}
}
pub fn true_to_eccentric(&self, t_anom: f64) -> f64 {
let a = self.semi_major_axis();
let e = self.eccentricity();
let b = a * (1.0 - e.powi(2)).sqrt();
let p = self.orbital_parameter();
let r = p / (1.0 + e * t_anom.cos());
let c = (a * e + r * t_anom.cos()) / a;
let s = (r / b) * t_anom.sin();
return s.atan2(c);
}
pub fn true_anomaly_at_time(&self, time: f64) -> f64 {
let t_peri = self.time_since_periapsis();
let m_anom = self.mean_anomaly((time * DAYTOSEC) + t_peri);
let angle = self.eccentric_from_mean(m_anom);
return PI2 - self.eccentric_to_true_anomaly(angle);
}
/// The eccentric anomaly at a certain time
pub fn eccentric_from_mean(&self, m_anom: f64) -> f64 {
match self.kepler(m_anom) {
Ok(num) => num,
Err(e) => {
eprintln!("{}: {}\n", "Invalid Orbit".red(), e);
return std::f64::NAN;
}
}
}
/// Return the eccentric anomaly using the appropriate Kepler equation
pub fn kepler(&self, m_anom: f64) -> Result<f64, &str> {
let e = self.eccentricity();
match &self.orbit_type {
OrbitType::Elliptic => Ok(elliptic_kepler(m_anom, e)),
OrbitType::Hyperbolic => Ok(hyper_kepler(m_anom, e)),
OrbitType::Circular => Err("cannot use Keler's equation with a circular orbit."),
OrbitType::Parabolic => Err("cannot use Kepler's equation with a parabolic orbit."),
}
}
pub fn eccentric_to_true_anomaly(&self, e_anom: f64) -> f64 {
let e = self.eccentricity();
// let sqrt_val = ((1.0 + e) / (1.0 - e)).sqrt();
// 2.0 * (sqrt_val * (e_anom / 2.0).tan()).atan() + PI2
((e_anom.cos() - e) / (1.0 - e * e_anom.cos())).acos()
}
/// Return the mean anomaly at a certain time from current position
pub fn mean_anomaly(&self, t: f64) -> f64 {
let n = (SOLARGM / self.semi_major_axis().powi(3)).sqrt();
n * t
}
}
/**
* Some of the kepler functions below. Body matches on its orbit type
* and uses the correct function to return the correct eccentric anomaly
*/
fn elliptic_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-200;
let kep = |e: f64| e - eccen * e.sin() - nt;
let kep_d = |e: f64| 1.0 - eccen * e.cos();
let mut e_0 = 0.0;
let mut e = e_0 - (kep(e_0) / kep_d(e_0));
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - (kep(e_0) / kep_d(e_0));
}
return e;
}
fn hyper_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-100;
let kep = |e: f64| eccen * e.sinh() - nt - e;
let kep_d = |e: f64| eccen * e.cosh() - 1.0;
let mut e_0 = nt;
let mut e = e_0 - kep(e_0) / kep_d(e_0);
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - kep(e_0) / kep_d(e_0);
}
return e;
}
pub fn three_one_three_transform(
arg_of_peri: f64,
inclination: f64,
arg_of_AN: f64,
) -> Matrix3<f64> {
let omega = arg_of_peri;
let inc = inclination;
let tht = arg_of_AN;
let m_c = Matrix3::new(
omega.cos(),
omega.sin(),
0.0,
-omega.sin(),
omega.cos(),
0.0,
0.0,
0.0,
1.0,
);
let m_b = Matrix3::new(
1.0,
0.0,
0.0,
0.0,
inc.cos(),
inc.sin(),
0.0,
-inc.sin(),
inc.cos(),
);
let m_a = Matrix3::new(
tht.cos(),
tht.sin(),
0.0,
-tht.sin(),
tht.cos(),
0.0,
0.0,
0.0,
1.0,
);
return m_c * m_b * m_a;
}
| position_and_velocity | identifier_name |
body.rs | #![allow(dead_code)]
#![allow(unused_doc_comments)]
/**
* body.rs contains the Body struct and implements methods for it. A body struct contains only the
* position and velocity vectors of the body, other parameters are calculated using methods. A body
* is instantiated using using the Body::new() method, which also determines the type of orbit the
* body has at the same time. orbit_type does not have to be given manually.
*/
use nalgebra::{Matrix3, Vector3};
use std::f64::consts::PI;
use colored::*;
const DAYTOSEC: f64 = 24.0 * 3600.0;
const SOLARGM: f64 = 1.328905188132376e11;
const PI2: f64 = 2.0 * PI;
/**
* OrbitType is used to abstract away some of the functions that depend on
* the type of orbit the body is in, like kepler's equation. That way, you
* can call one function and it will return the correct value
*/
#[derive(Debug)]
pub enum OrbitType {
Circular,
Elliptic,
Parabolic,
Hyperbolic,
}
impl OrbitType {
/// Return the orbit type given the eccentricity, Body::new
/// uses this function to set the orbit type when an instance
/// is constructed
pub fn new(eccentricity: f64) -> OrbitType {
if eccentricity == 0.0 {
return OrbitType::Circular;
} else if eccentricity < 1.0 && eccentricity > 0.0 {
return OrbitType::Elliptic;
} else if eccentricity == 1.0 {
return OrbitType::Parabolic;
} else {
return OrbitType::Hyperbolic;
}
}
}
#[derive(Debug)]
pub struct Body {
pub position: Vector3<f64>,
pub velocity: Vector3<f64>,
pub orbit_type: OrbitType,
}
/* Adds methods to Body struct */
impl Body {
pub fn new(position: Vector3<f64>, velocity: Vector3<f64>) -> Body {
// h and e are used for determining what kind of orbit the body is currently in
let h = position.cross(&velocity);
let e = ((velocity.cross(&h) / SOLARGM) - position.normalize()).norm();
Body {
position: position,
velocity: velocity,
orbit_type: OrbitType::new(e),
}
}
pub fn radial_velocity(&self) -> Vector3<f64> {
(self.velocity.dot(&self.position) / self.position.norm_squared()) * self.position
}
pub fn tangential_velocity(&self) -> Vector3<f64> {
self.omega().cross(&self.position)
}
pub fn true_anomaly(&self) -> f64 {
let e_vec = self.eccentricity_vector();
let posit = self.position.normalize();
let val = e_vec.dot(&posit) / (e_vec.norm() * posit.norm());
if posit.dot(&self.velocity.normalize()) < 0.0 {
return 2.0 * PI - val.acos();
} else |
}
/* points from focus to perigee if I'm not mistaken */
pub fn eccentricity_vector(&self) -> Vector3<f64> {
let veloc = self.velocity;
let posit = self.position;
let h = self.angular_momentum();
(veloc.cross(&h) / SOLARGM) - posit.normalize()
}
pub fn angular_momentum(&self) -> Vector3<f64> {
self.position.cross(&self.velocity)
}
pub fn total_energy(&self) -> f64 {
let posit = self.position.norm();
let veloc = self.velocity.norm();
0.5 * veloc.powi(2) - (SOLARGM / posit)
}
pub fn omega(&self) -> Vector3<f64> {
self.angular_momentum() / self.position.norm_squared()
}
pub fn frame_rotation_rate(&self) -> f64 {
self.omega().norm()
}
pub fn position_at_angle(&self, angle: f64) -> Vector3<f64> {
let e = self.eccentricity();
let numer = self.angular_momentum().norm_squared() / SOLARGM;
let denom = 1_f64 + (e * (angle).cos());
let radius = numer / denom;
Vector3::new(radius, 0.0, 0.0)
}
pub fn velocity_at_angle(&self, angle: f64) -> Vector3<f64> {
let p = self.orbital_parameter();
let e = self.eccentricity();
let h = self.angular_momentum().norm_squared();
Vector3::new(
(h / p) * e * angle.sin(),
(h / p) * (1_f64 + e * angle.cos()),
0.0,
)
}
pub fn position_and_velocity(&self, angle: f64) -> (Vector3<f64>, Vector3<f64>) {
let r = self.position_at_angle(angle);
let v = self.velocity_at_angle(angle);
let tht = angle - self.true_anomaly();
let trans = Matrix3::from_rows(&[
Vector3::new(tht.cos(), -tht.sin(), 0.0).transpose(),
Vector3::new(tht.sin(), tht.cos(), 0.0).transpose(),
Vector3::new(0.0, 0.0, 1.0).transpose(),
]);
(trans * r, trans * v)
}
// Angle to other body, keep getting the wrong thing anyway, tried everything
pub fn angle_to(&self, other: &Body) -> f64 {
(self.position.dot(&other.position) / (self.position.norm() * other.position.norm())).acos()
}
/* Return a transformation matrix constructed from body's orbit in inertial frame */
pub fn make_frame(&self) -> Matrix3<f64> {
let e_r = self.position.normalize();
let e_h = self.angular_momentum().normalize();
let e_tht = e_h.cross(&e_r);
Matrix3::from_rows(&[e_r.transpose(), e_tht.transpose(), e_h.transpose()])
}
pub fn semi_major_axis(&self) -> f64 {
let ang_moment = self.angular_momentum().norm();
let e = self.eccentricity();
ang_moment.powi(2) / (SOLARGM * (1_f64 - e.powi(2)))
}
pub fn orbital_period(&self) -> f64 {
2_f64 * PI * (self.semi_major_axis().powi(3) / SOLARGM).sqrt()
}
pub fn orbital_parameter(&self) -> f64 {
let e = self.eccentricity();
self.semi_major_axis() * (1.0 - e.powi(2))
}
pub fn eccentric_anomaly(&self) -> f64 {
let e = self.eccentricity();
let theta = self.true_anomaly();
2.0 * ((theta / 2.0).tan() / ((1.0 + e) / (1.0 - e)).sqrt()).atan()
}
pub fn time_since_periapsis(&self) -> f64 {
let t_anom = self.true_anomaly();
let e_anom = self.true_to_eccentric(t_anom);
let a = self.semi_major_axis();
let e = self.eccentricity();
(a.powi(3) / SOLARGM).sqrt() * (e_anom - e * e_anom.sin())
}
pub fn eccentricity(&self) -> f64 {
self.eccentricity_vector().norm()
}
pub fn inclination(&self) -> f64 {
let h = self.angular_momentum();
(h[2] / h.norm()).acos() // h[2] is the z component of the vector
}
pub fn ascending_node(&self) -> Vector3<f64> {
let k = Vector3::new(0.0, 0.0, 1.0);
k.cross(&self.angular_momentum())
}
pub fn argument_of_periapsis(&self) -> f64 {
let n = self.ascending_node();
let e = self.eccentricity_vector();
let omega = (n.dot(&e) / (n.norm() * e.norm())).acos();
if e[2] < 0.0 {
PI2 - omega
} else {
omega
}
}
pub fn argument_of_ascending_node(&self) -> f64 {
let n = self.ascending_node();
let n_x = n[0];
let n_y = n[1];
if n_y >= 0.0 {
(n_x / n.norm()).acos()
} else {
PI2 - (n_x / n.norm()).acos()
}
}
pub fn true_to_eccentric(&self, t_anom: f64) -> f64 {
let a = self.semi_major_axis();
let e = self.eccentricity();
let b = a * (1.0 - e.powi(2)).sqrt();
let p = self.orbital_parameter();
let r = p / (1.0 + e * t_anom.cos());
let c = (a * e + r * t_anom.cos()) / a;
let s = (r / b) * t_anom.sin();
return s.atan2(c);
}
pub fn true_anomaly_at_time(&self, time: f64) -> f64 {
let t_peri = self.time_since_periapsis();
let m_anom = self.mean_anomaly((time * DAYTOSEC) + t_peri);
let angle = self.eccentric_from_mean(m_anom);
return PI2 - self.eccentric_to_true_anomaly(angle);
}
/// The eccentric anomaly at a certain time
pub fn eccentric_from_mean(&self, m_anom: f64) -> f64 {
match self.kepler(m_anom) {
Ok(num) => num,
Err(e) => {
eprintln!("{}: {}\n", "Invalid Orbit".red(), e);
return std::f64::NAN;
}
}
}
/// Return the eccentric anomaly using the appropriate Kepler equation
pub fn kepler(&self, m_anom: f64) -> Result<f64, &str> {
let e = self.eccentricity();
match &self.orbit_type {
OrbitType::Elliptic => Ok(elliptic_kepler(m_anom, e)),
OrbitType::Hyperbolic => Ok(hyper_kepler(m_anom, e)),
OrbitType::Circular => Err("cannot use Keler's equation with a circular orbit."),
OrbitType::Parabolic => Err("cannot use Kepler's equation with a parabolic orbit."),
}
}
pub fn eccentric_to_true_anomaly(&self, e_anom: f64) -> f64 {
let e = self.eccentricity();
// let sqrt_val = ((1.0 + e) / (1.0 - e)).sqrt();
// 2.0 * (sqrt_val * (e_anom / 2.0).tan()).atan() + PI2
((e_anom.cos() - e) / (1.0 - e * e_anom.cos())).acos()
}
/// Return the mean anomaly at a certain time from current position
pub fn mean_anomaly(&self, t: f64) -> f64 {
let n = (SOLARGM / self.semi_major_axis().powi(3)).sqrt();
n * t
}
}
/**
* Some of the kepler functions below. Body matches on its orbit type
* and uses the correct function to return the correct eccentric anomaly
*/
fn elliptic_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-200;
let kep = |e: f64| e - eccen * e.sin() - nt;
let kep_d = |e: f64| 1.0 - eccen * e.cos();
let mut e_0 = 0.0;
let mut e = e_0 - (kep(e_0) / kep_d(e_0));
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - (kep(e_0) / kep_d(e_0));
}
return e;
}
fn hyper_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-100;
let kep = |e: f64| eccen * e.sinh() - nt - e;
let kep_d = |e: f64| eccen * e.cosh() - 1.0;
let mut e_0 = nt;
let mut e = e_0 - kep(e_0) / kep_d(e_0);
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - kep(e_0) / kep_d(e_0);
}
return e;
}
pub fn three_one_three_transform(
arg_of_peri: f64,
inclination: f64,
arg_of_AN: f64,
) -> Matrix3<f64> {
let omega = arg_of_peri;
let inc = inclination;
let tht = arg_of_AN;
let m_c = Matrix3::new(
omega.cos(),
omega.sin(),
0.0,
-omega.sin(),
omega.cos(),
0.0,
0.0,
0.0,
1.0,
);
let m_b = Matrix3::new(
1.0,
0.0,
0.0,
0.0,
inc.cos(),
inc.sin(),
0.0,
-inc.sin(),
inc.cos(),
);
let m_a = Matrix3::new(
tht.cos(),
tht.sin(),
0.0,
-tht.sin(),
tht.cos(),
0.0,
0.0,
0.0,
1.0,
);
return m_c * m_b * m_a;
}
| {
return val.acos();
} | conditional_block |
body.rs | #![allow(dead_code)]
#![allow(unused_doc_comments)]
/**
* body.rs contains the Body struct and implements methods for it. A body struct contains only the
* position and velocity vectors of the body, other parameters are calculated using methods. A body
* is instantiated using using the Body::new() method, which also determines the type of orbit the
* body has at the same time. orbit_type does not have to be given manually.
*/
use nalgebra::{Matrix3, Vector3};
use std::f64::consts::PI;
use colored::*;
const DAYTOSEC: f64 = 24.0 * 3600.0;
const SOLARGM: f64 = 1.328905188132376e11;
const PI2: f64 = 2.0 * PI;
/**
* OrbitType is used to abstract away some of the functions that depend on
* the type of orbit the body is in, like kepler's equation. That way, you
* can call one function and it will return the correct value
*/
#[derive(Debug)]
pub enum OrbitType {
Circular,
Elliptic,
Parabolic,
Hyperbolic,
}
impl OrbitType {
/// Return the orbit type given the eccentricity, Body::new
/// uses this function to set the orbit type when an instance
/// is constructed
pub fn new(eccentricity: f64) -> OrbitType {
if eccentricity == 0.0 {
return OrbitType::Circular;
} else if eccentricity < 1.0 && eccentricity > 0.0 {
return OrbitType::Elliptic;
} else if eccentricity == 1.0 {
return OrbitType::Parabolic;
} else {
return OrbitType::Hyperbolic;
}
}
}
#[derive(Debug)]
pub struct Body {
pub position: Vector3<f64>, | impl Body {
pub fn new(position: Vector3<f64>, velocity: Vector3<f64>) -> Body {
// h and e are used for determining what kind of orbit the body is currently in
let h = position.cross(&velocity);
let e = ((velocity.cross(&h) / SOLARGM) - position.normalize()).norm();
Body {
position: position,
velocity: velocity,
orbit_type: OrbitType::new(e),
}
}
pub fn radial_velocity(&self) -> Vector3<f64> {
(self.velocity.dot(&self.position) / self.position.norm_squared()) * self.position
}
pub fn tangential_velocity(&self) -> Vector3<f64> {
self.omega().cross(&self.position)
}
pub fn true_anomaly(&self) -> f64 {
let e_vec = self.eccentricity_vector();
let posit = self.position.normalize();
let val = e_vec.dot(&posit) / (e_vec.norm() * posit.norm());
if posit.dot(&self.velocity.normalize()) < 0.0 {
return 2.0 * PI - val.acos();
} else {
return val.acos();
}
}
/* points from focus to perigee if I'm not mistaken */
pub fn eccentricity_vector(&self) -> Vector3<f64> {
let veloc = self.velocity;
let posit = self.position;
let h = self.angular_momentum();
(veloc.cross(&h) / SOLARGM) - posit.normalize()
}
pub fn angular_momentum(&self) -> Vector3<f64> {
self.position.cross(&self.velocity)
}
pub fn total_energy(&self) -> f64 {
let posit = self.position.norm();
let veloc = self.velocity.norm();
0.5 * veloc.powi(2) - (SOLARGM / posit)
}
pub fn omega(&self) -> Vector3<f64> {
self.angular_momentum() / self.position.norm_squared()
}
pub fn frame_rotation_rate(&self) -> f64 {
self.omega().norm()
}
pub fn position_at_angle(&self, angle: f64) -> Vector3<f64> {
let e = self.eccentricity();
let numer = self.angular_momentum().norm_squared() / SOLARGM;
let denom = 1_f64 + (e * (angle).cos());
let radius = numer / denom;
Vector3::new(radius, 0.0, 0.0)
}
pub fn velocity_at_angle(&self, angle: f64) -> Vector3<f64> {
let p = self.orbital_parameter();
let e = self.eccentricity();
let h = self.angular_momentum().norm_squared();
Vector3::new(
(h / p) * e * angle.sin(),
(h / p) * (1_f64 + e * angle.cos()),
0.0,
)
}
pub fn position_and_velocity(&self, angle: f64) -> (Vector3<f64>, Vector3<f64>) {
let r = self.position_at_angle(angle);
let v = self.velocity_at_angle(angle);
let tht = angle - self.true_anomaly();
let trans = Matrix3::from_rows(&[
Vector3::new(tht.cos(), -tht.sin(), 0.0).transpose(),
Vector3::new(tht.sin(), tht.cos(), 0.0).transpose(),
Vector3::new(0.0, 0.0, 1.0).transpose(),
]);
(trans * r, trans * v)
}
// Angle to other body, keep getting the wrong thing anyway, tried everything
pub fn angle_to(&self, other: &Body) -> f64 {
(self.position.dot(&other.position) / (self.position.norm() * other.position.norm())).acos()
}
/* Return a transformation matrix constructed from body's orbit in inertial frame */
pub fn make_frame(&self) -> Matrix3<f64> {
let e_r = self.position.normalize();
let e_h = self.angular_momentum().normalize();
let e_tht = e_h.cross(&e_r);
Matrix3::from_rows(&[e_r.transpose(), e_tht.transpose(), e_h.transpose()])
}
pub fn semi_major_axis(&self) -> f64 {
let ang_moment = self.angular_momentum().norm();
let e = self.eccentricity();
ang_moment.powi(2) / (SOLARGM * (1_f64 - e.powi(2)))
}
pub fn orbital_period(&self) -> f64 {
2_f64 * PI * (self.semi_major_axis().powi(3) / SOLARGM).sqrt()
}
pub fn orbital_parameter(&self) -> f64 {
let e = self.eccentricity();
self.semi_major_axis() * (1.0 - e.powi(2))
}
pub fn eccentric_anomaly(&self) -> f64 {
let e = self.eccentricity();
let theta = self.true_anomaly();
2.0 * ((theta / 2.0).tan() / ((1.0 + e) / (1.0 - e)).sqrt()).atan()
}
pub fn time_since_periapsis(&self) -> f64 {
let t_anom = self.true_anomaly();
let e_anom = self.true_to_eccentric(t_anom);
let a = self.semi_major_axis();
let e = self.eccentricity();
(a.powi(3) / SOLARGM).sqrt() * (e_anom - e * e_anom.sin())
}
pub fn eccentricity(&self) -> f64 {
self.eccentricity_vector().norm()
}
pub fn inclination(&self) -> f64 {
let h = self.angular_momentum();
(h[2] / h.norm()).acos() // h[2] is the z component of the vector
}
pub fn ascending_node(&self) -> Vector3<f64> {
let k = Vector3::new(0.0, 0.0, 1.0);
k.cross(&self.angular_momentum())
}
pub fn argument_of_periapsis(&self) -> f64 {
let n = self.ascending_node();
let e = self.eccentricity_vector();
let omega = (n.dot(&e) / (n.norm() * e.norm())).acos();
if e[2] < 0.0 {
PI2 - omega
} else {
omega
}
}
pub fn argument_of_ascending_node(&self) -> f64 {
let n = self.ascending_node();
let n_x = n[0];
let n_y = n[1];
if n_y >= 0.0 {
(n_x / n.norm()).acos()
} else {
PI2 - (n_x / n.norm()).acos()
}
}
pub fn true_to_eccentric(&self, t_anom: f64) -> f64 {
let a = self.semi_major_axis();
let e = self.eccentricity();
let b = a * (1.0 - e.powi(2)).sqrt();
let p = self.orbital_parameter();
let r = p / (1.0 + e * t_anom.cos());
let c = (a * e + r * t_anom.cos()) / a;
let s = (r / b) * t_anom.sin();
return s.atan2(c);
}
pub fn true_anomaly_at_time(&self, time: f64) -> f64 {
let t_peri = self.time_since_periapsis();
let m_anom = self.mean_anomaly((time * DAYTOSEC) + t_peri);
let angle = self.eccentric_from_mean(m_anom);
return PI2 - self.eccentric_to_true_anomaly(angle);
}
/// The eccentric anomaly at a certain time
pub fn eccentric_from_mean(&self, m_anom: f64) -> f64 {
match self.kepler(m_anom) {
Ok(num) => num,
Err(e) => {
eprintln!("{}: {}\n", "Invalid Orbit".red(), e);
return std::f64::NAN;
}
}
}
/// Return the eccentric anomaly using the appropriate Kepler equation
pub fn kepler(&self, m_anom: f64) -> Result<f64, &str> {
let e = self.eccentricity();
match &self.orbit_type {
OrbitType::Elliptic => Ok(elliptic_kepler(m_anom, e)),
OrbitType::Hyperbolic => Ok(hyper_kepler(m_anom, e)),
OrbitType::Circular => Err("cannot use Keler's equation with a circular orbit."),
OrbitType::Parabolic => Err("cannot use Kepler's equation with a parabolic orbit."),
}
}
pub fn eccentric_to_true_anomaly(&self, e_anom: f64) -> f64 {
let e = self.eccentricity();
// let sqrt_val = ((1.0 + e) / (1.0 - e)).sqrt();
// 2.0 * (sqrt_val * (e_anom / 2.0).tan()).atan() + PI2
((e_anom.cos() - e) / (1.0 - e * e_anom.cos())).acos()
}
/// Return the mean anomaly at a certain time from current position
pub fn mean_anomaly(&self, t: f64) -> f64 {
let n = (SOLARGM / self.semi_major_axis().powi(3)).sqrt();
n * t
}
}
/**
* Some of the kepler functions below. Body matches on its orbit type
* and uses the correct function to return the correct eccentric anomaly
*/
fn elliptic_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-200;
let kep = |e: f64| e - eccen * e.sin() - nt;
let kep_d = |e: f64| 1.0 - eccen * e.cos();
let mut e_0 = 0.0;
let mut e = e_0 - (kep(e_0) / kep_d(e_0));
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - (kep(e_0) / kep_d(e_0));
}
return e;
}
fn hyper_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-100;
let kep = |e: f64| eccen * e.sinh() - nt - e;
let kep_d = |e: f64| eccen * e.cosh() - 1.0;
let mut e_0 = nt;
let mut e = e_0 - kep(e_0) / kep_d(e_0);
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - kep(e_0) / kep_d(e_0);
}
return e;
}
pub fn three_one_three_transform(
arg_of_peri: f64,
inclination: f64,
arg_of_AN: f64,
) -> Matrix3<f64> {
let omega = arg_of_peri;
let inc = inclination;
let tht = arg_of_AN;
let m_c = Matrix3::new(
omega.cos(),
omega.sin(),
0.0,
-omega.sin(),
omega.cos(),
0.0,
0.0,
0.0,
1.0,
);
let m_b = Matrix3::new(
1.0,
0.0,
0.0,
0.0,
inc.cos(),
inc.sin(),
0.0,
-inc.sin(),
inc.cos(),
);
let m_a = Matrix3::new(
tht.cos(),
tht.sin(),
0.0,
-tht.sin(),
tht.cos(),
0.0,
0.0,
0.0,
1.0,
);
return m_c * m_b * m_a;
} | pub velocity: Vector3<f64>,
pub orbit_type: OrbitType,
}
/* Adds methods to Body struct */ | random_line_split |
body.rs | #![allow(dead_code)]
#![allow(unused_doc_comments)]
/**
* body.rs contains the Body struct and implements methods for it. A body struct contains only the
* position and velocity vectors of the body, other parameters are calculated using methods. A body
* is instantiated using using the Body::new() method, which also determines the type of orbit the
* body has at the same time. orbit_type does not have to be given manually.
*/
use nalgebra::{Matrix3, Vector3};
use std::f64::consts::PI;
use colored::*;
const DAYTOSEC: f64 = 24.0 * 3600.0;
const SOLARGM: f64 = 1.328905188132376e11;
const PI2: f64 = 2.0 * PI;
/**
* OrbitType is used to abstract away some of the functions that depend on
* the type of orbit the body is in, like kepler's equation. That way, you
* can call one function and it will return the correct value
*/
#[derive(Debug)]
pub enum OrbitType {
Circular,
Elliptic,
Parabolic,
Hyperbolic,
}
impl OrbitType {
/// Return the orbit type given the eccentricity, Body::new
/// uses this function to set the orbit type when an instance
/// is constructed
pub fn new(eccentricity: f64) -> OrbitType {
if eccentricity == 0.0 {
return OrbitType::Circular;
} else if eccentricity < 1.0 && eccentricity > 0.0 {
return OrbitType::Elliptic;
} else if eccentricity == 1.0 {
return OrbitType::Parabolic;
} else {
return OrbitType::Hyperbolic;
}
}
}
#[derive(Debug)]
pub struct Body {
pub position: Vector3<f64>,
pub velocity: Vector3<f64>,
pub orbit_type: OrbitType,
}
/* Adds methods to Body struct */
impl Body {
pub fn new(position: Vector3<f64>, velocity: Vector3<f64>) -> Body {
// h and e are used for determining what kind of orbit the body is currently in
let h = position.cross(&velocity);
let e = ((velocity.cross(&h) / SOLARGM) - position.normalize()).norm();
Body {
position: position,
velocity: velocity,
orbit_type: OrbitType::new(e),
}
}
pub fn radial_velocity(&self) -> Vector3<f64> {
(self.velocity.dot(&self.position) / self.position.norm_squared()) * self.position
}
pub fn tangential_velocity(&self) -> Vector3<f64> |
pub fn true_anomaly(&self) -> f64 {
let e_vec = self.eccentricity_vector();
let posit = self.position.normalize();
let val = e_vec.dot(&posit) / (e_vec.norm() * posit.norm());
if posit.dot(&self.velocity.normalize()) < 0.0 {
return 2.0 * PI - val.acos();
} else {
return val.acos();
}
}
/* points from focus to perigee if I'm not mistaken */
pub fn eccentricity_vector(&self) -> Vector3<f64> {
let veloc = self.velocity;
let posit = self.position;
let h = self.angular_momentum();
(veloc.cross(&h) / SOLARGM) - posit.normalize()
}
pub fn angular_momentum(&self) -> Vector3<f64> {
self.position.cross(&self.velocity)
}
pub fn total_energy(&self) -> f64 {
let posit = self.position.norm();
let veloc = self.velocity.norm();
0.5 * veloc.powi(2) - (SOLARGM / posit)
}
pub fn omega(&self) -> Vector3<f64> {
self.angular_momentum() / self.position.norm_squared()
}
pub fn frame_rotation_rate(&self) -> f64 {
self.omega().norm()
}
pub fn position_at_angle(&self, angle: f64) -> Vector3<f64> {
let e = self.eccentricity();
let numer = self.angular_momentum().norm_squared() / SOLARGM;
let denom = 1_f64 + (e * (angle).cos());
let radius = numer / denom;
Vector3::new(radius, 0.0, 0.0)
}
pub fn velocity_at_angle(&self, angle: f64) -> Vector3<f64> {
let p = self.orbital_parameter();
let e = self.eccentricity();
let h = self.angular_momentum().norm_squared();
Vector3::new(
(h / p) * e * angle.sin(),
(h / p) * (1_f64 + e * angle.cos()),
0.0,
)
}
pub fn position_and_velocity(&self, angle: f64) -> (Vector3<f64>, Vector3<f64>) {
let r = self.position_at_angle(angle);
let v = self.velocity_at_angle(angle);
let tht = angle - self.true_anomaly();
let trans = Matrix3::from_rows(&[
Vector3::new(tht.cos(), -tht.sin(), 0.0).transpose(),
Vector3::new(tht.sin(), tht.cos(), 0.0).transpose(),
Vector3::new(0.0, 0.0, 1.0).transpose(),
]);
(trans * r, trans * v)
}
// Angle to other body, keep getting the wrong thing anyway, tried everything
pub fn angle_to(&self, other: &Body) -> f64 {
(self.position.dot(&other.position) / (self.position.norm() * other.position.norm())).acos()
}
/* Return a transformation matrix constructed from body's orbit in inertial frame */
pub fn make_frame(&self) -> Matrix3<f64> {
let e_r = self.position.normalize();
let e_h = self.angular_momentum().normalize();
let e_tht = e_h.cross(&e_r);
Matrix3::from_rows(&[e_r.transpose(), e_tht.transpose(), e_h.transpose()])
}
pub fn semi_major_axis(&self) -> f64 {
let ang_moment = self.angular_momentum().norm();
let e = self.eccentricity();
ang_moment.powi(2) / (SOLARGM * (1_f64 - e.powi(2)))
}
pub fn orbital_period(&self) -> f64 {
2_f64 * PI * (self.semi_major_axis().powi(3) / SOLARGM).sqrt()
}
pub fn orbital_parameter(&self) -> f64 {
let e = self.eccentricity();
self.semi_major_axis() * (1.0 - e.powi(2))
}
pub fn eccentric_anomaly(&self) -> f64 {
let e = self.eccentricity();
let theta = self.true_anomaly();
2.0 * ((theta / 2.0).tan() / ((1.0 + e) / (1.0 - e)).sqrt()).atan()
}
pub fn time_since_periapsis(&self) -> f64 {
let t_anom = self.true_anomaly();
let e_anom = self.true_to_eccentric(t_anom);
let a = self.semi_major_axis();
let e = self.eccentricity();
(a.powi(3) / SOLARGM).sqrt() * (e_anom - e * e_anom.sin())
}
pub fn eccentricity(&self) -> f64 {
self.eccentricity_vector().norm()
}
pub fn inclination(&self) -> f64 {
let h = self.angular_momentum();
(h[2] / h.norm()).acos() // h[2] is the z component of the vector
}
pub fn ascending_node(&self) -> Vector3<f64> {
let k = Vector3::new(0.0, 0.0, 1.0);
k.cross(&self.angular_momentum())
}
pub fn argument_of_periapsis(&self) -> f64 {
let n = self.ascending_node();
let e = self.eccentricity_vector();
let omega = (n.dot(&e) / (n.norm() * e.norm())).acos();
if e[2] < 0.0 {
PI2 - omega
} else {
omega
}
}
pub fn argument_of_ascending_node(&self) -> f64 {
let n = self.ascending_node();
let n_x = n[0];
let n_y = n[1];
if n_y >= 0.0 {
(n_x / n.norm()).acos()
} else {
PI2 - (n_x / n.norm()).acos()
}
}
pub fn true_to_eccentric(&self, t_anom: f64) -> f64 {
let a = self.semi_major_axis();
let e = self.eccentricity();
let b = a * (1.0 - e.powi(2)).sqrt();
let p = self.orbital_parameter();
let r = p / (1.0 + e * t_anom.cos());
let c = (a * e + r * t_anom.cos()) / a;
let s = (r / b) * t_anom.sin();
return s.atan2(c);
}
pub fn true_anomaly_at_time(&self, time: f64) -> f64 {
let t_peri = self.time_since_periapsis();
let m_anom = self.mean_anomaly((time * DAYTOSEC) + t_peri);
let angle = self.eccentric_from_mean(m_anom);
return PI2 - self.eccentric_to_true_anomaly(angle);
}
/// The eccentric anomaly at a certain time
pub fn eccentric_from_mean(&self, m_anom: f64) -> f64 {
match self.kepler(m_anom) {
Ok(num) => num,
Err(e) => {
eprintln!("{}: {}\n", "Invalid Orbit".red(), e);
return std::f64::NAN;
}
}
}
/// Return the eccentric anomaly using the appropriate Kepler equation
pub fn kepler(&self, m_anom: f64) -> Result<f64, &str> {
let e = self.eccentricity();
match &self.orbit_type {
OrbitType::Elliptic => Ok(elliptic_kepler(m_anom, e)),
OrbitType::Hyperbolic => Ok(hyper_kepler(m_anom, e)),
OrbitType::Circular => Err("cannot use Keler's equation with a circular orbit."),
OrbitType::Parabolic => Err("cannot use Kepler's equation with a parabolic orbit."),
}
}
pub fn eccentric_to_true_anomaly(&self, e_anom: f64) -> f64 {
let e = self.eccentricity();
// let sqrt_val = ((1.0 + e) / (1.0 - e)).sqrt();
// 2.0 * (sqrt_val * (e_anom / 2.0).tan()).atan() + PI2
((e_anom.cos() - e) / (1.0 - e * e_anom.cos())).acos()
}
/// Return the mean anomaly at a certain time from current position
pub fn mean_anomaly(&self, t: f64) -> f64 {
let n = (SOLARGM / self.semi_major_axis().powi(3)).sqrt();
n * t
}
}
/**
* Some of the kepler functions below. Body matches on its orbit type
* and uses the correct function to return the correct eccentric anomaly
*/
fn elliptic_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-200;
let kep = |e: f64| e - eccen * e.sin() - nt;
let kep_d = |e: f64| 1.0 - eccen * e.cos();
let mut e_0 = 0.0;
let mut e = e_0 - (kep(e_0) / kep_d(e_0));
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - (kep(e_0) / kep_d(e_0));
}
return e;
}
fn hyper_kepler(nt: f64, eccen: f64) -> f64 {
let tolerance = 1e-100;
let kep = |e: f64| eccen * e.sinh() - nt - e;
let kep_d = |e: f64| eccen * e.cosh() - 1.0;
let mut e_0 = nt;
let mut e = e_0 - kep(e_0) / kep_d(e_0);
while (e - e_0).abs() > tolerance {
e_0 = e;
e = e_0 - kep(e_0) / kep_d(e_0);
}
return e;
}
pub fn three_one_three_transform(
arg_of_peri: f64,
inclination: f64,
arg_of_AN: f64,
) -> Matrix3<f64> {
let omega = arg_of_peri;
let inc = inclination;
let tht = arg_of_AN;
let m_c = Matrix3::new(
omega.cos(),
omega.sin(),
0.0,
-omega.sin(),
omega.cos(),
0.0,
0.0,
0.0,
1.0,
);
let m_b = Matrix3::new(
1.0,
0.0,
0.0,
0.0,
inc.cos(),
inc.sin(),
0.0,
-inc.sin(),
inc.cos(),
);
let m_a = Matrix3::new(
tht.cos(),
tht.sin(),
0.0,
-tht.sin(),
tht.cos(),
0.0,
0.0,
0.0,
1.0,
);
return m_c * m_b * m_a;
}
| {
self.omega().cross(&self.position)
} | identifier_body |
thermald.py | #!/usr/bin/env python3.7
import os
import json
import copy
import datetime
from smbus2 import SMBus
from cereal import log
from common.basedir import BASEDIR
from common.params import Params
from common.realtime import sec_since_boot, DT_TRML
from common.numpy_fast import clip
from common.filter_simple import FirstOrderFilter
from selfdrive.version import terms_version, training_version
from selfdrive.swaglog import cloudlog
import selfdrive.messaging as messaging
from selfdrive.services import service_list
from selfdrive.loggerd.config import get_available_percent
ThermalStatus = log.ThermalData.ThermalStatus
CURRENT_TAU = 15. # 15s time constant
DAYS_NO_CONNECTIVITY_MAX = 7 # do not allow to engage after a week without internet
DAYS_NO_CONNECTIVITY_PROMPT = 4 # send an offroad prompt after 4 days with no internet
with open(BASEDIR + "/selfdrive/controls/lib/alerts_offroad.json") as json_file:
OFFROAD_ALERTS = json.load(json_file)
def read_tz(x):
with open("/sys/devices/virtual/thermal/thermal_zone%d/temp" % x) as f:
ret = max(0, int(f.read()))
return ret
def read_thermal():
dat = messaging.new_message()
dat.init('thermal')
dat.thermal.cpu0 = read_tz(5)
dat.thermal.cpu1 = read_tz(7)
dat.thermal.cpu2 = read_tz(10)
dat.thermal.cpu3 = read_tz(12)
dat.thermal.mem = read_tz(2)
dat.thermal.gpu = read_tz(16)
dat.thermal.bat = read_tz(29)
return dat
LEON = False
def setup_eon_fan():
global LEON
os.system("echo 2 > /sys/module/dwc3_msm/parameters/otg_switch")
bus = SMBus(7, force=True)
try:
bus.write_byte_data(0x21, 0x10, 0xf) # mask all interrupts
bus.write_byte_data(0x21, 0x03, 0x1) # set drive current and global interrupt disable
bus.write_byte_data(0x21, 0x02, 0x2) # needed?
bus.write_byte_data(0x21, 0x04, 0x4) # manual override source
except IOError:
print("LEON detected")
#os.system("echo 1 > /sys/devices/soc/6a00000.ssusb/power_supply/usb/usb_otg")
LEON = True
bus.close()
last_eon_fan_val = None
def set_eon_fan(val):
|
# temp thresholds to control fan speed - high hysteresis
_TEMP_THRS_H = [50., 65., 80., 10000]
# temp thresholds to control fan speed - low hysteresis
_TEMP_THRS_L = [42.5, 57.5, 72.5, 10000]
# fan speed options
_FAN_SPEEDS = [0, 16384, 32768, 65535]
# max fan speed only allowed if battery is hot
_BAT_TEMP_THERSHOLD = 45.
def handle_fan(max_cpu_temp, bat_temp, fan_speed):
new_speed_h = next(speed for speed, temp_h in zip(_FAN_SPEEDS, _TEMP_THRS_H) if temp_h > max_cpu_temp)
new_speed_l = next(speed for speed, temp_l in zip(_FAN_SPEEDS, _TEMP_THRS_L) if temp_l > max_cpu_temp)
if new_speed_h > fan_speed:
# update speed if using the high thresholds results in fan speed increment
fan_speed = new_speed_h
elif new_speed_l < fan_speed:
# update speed if using the low thresholds results in fan speed decrement
fan_speed = new_speed_l
if bat_temp < _BAT_TEMP_THERSHOLD:
# no max fan speed unless battery is hot
fan_speed = min(fan_speed, _FAN_SPEEDS[-2])
set_eon_fan(fan_speed//16384)
return fan_speed
def thermald_thread():
setup_eon_fan()
# prevent LEECO from undervoltage
BATT_PERC_OFF = 10 if LEON else 3
# now loop
thermal_sock = messaging.pub_sock(service_list['thermal'].port)
health_sock = messaging.sub_sock(service_list['health'].port)
location_sock = messaging.sub_sock(service_list['gpsLocation'].port)
fan_speed = 0
count = 0
off_ts = None
started_ts = None
started_seen = False
thermal_status = ThermalStatus.green
thermal_status_prev = ThermalStatus.green
usb_power = True
usb_power_prev = True
health_sock.RCVTIMEO = int(1000 * 2 * DT_TRML) # 2x the expected health frequency
current_filter = FirstOrderFilter(0., CURRENT_TAU, DT_TRML)
health_prev = None
current_connectivity_alert = None
params = Params()
while 1:
health = messaging.recv_sock(health_sock, wait=True)
location = messaging.recv_sock(location_sock)
location = location.gpsLocation if location else None
msg = read_thermal()
# clear car params when panda gets disconnected
if health is None and health_prev is not None:
params.panda_disconnect()
health_prev = health
if health is not None:
usb_power = health.health.usbPowerMode != log.HealthData.UsbPowerMode.client
# loggerd is gated based on free space
avail = get_available_percent() / 100.0
# thermal message now also includes free space
msg.thermal.freeSpace = avail
with open("/sys/class/power_supply/battery/capacity") as f:
msg.thermal.batteryPercent = int(f.read())
with open("/sys/class/power_supply/battery/status") as f:
msg.thermal.batteryStatus = f.read().strip()
with open("/sys/class/power_supply/battery/current_now") as f:
msg.thermal.batteryCurrent = int(f.read())
with open("/sys/class/power_supply/battery/voltage_now") as f:
msg.thermal.batteryVoltage = int(f.read())
with open("/sys/class/power_supply/usb/present") as f:
msg.thermal.usbOnline = bool(int(f.read()))
current_filter.update(msg.thermal.batteryCurrent / 1e6)
# TODO: add car battery voltage check
max_cpu_temp = max(msg.thermal.cpu0, msg.thermal.cpu1,
msg.thermal.cpu2, msg.thermal.cpu3) / 10.0
max_comp_temp = max(max_cpu_temp, msg.thermal.mem / 10., msg.thermal.gpu / 10.)
bat_temp = msg.thermal.bat/1000.
fan_speed = handle_fan(max_cpu_temp, bat_temp, fan_speed)
msg.thermal.fanSpeed = fan_speed
# thermal logic with hysterisis
if max_cpu_temp > 107. or bat_temp >= 63.:
# onroad not allowed
thermal_status = ThermalStatus.danger
elif max_comp_temp > 92.5 or bat_temp > 60.: # CPU throttling starts around ~90C
# hysteresis between onroad not allowed and engage not allowed
thermal_status = clip(thermal_status, ThermalStatus.red, ThermalStatus.danger)
elif max_cpu_temp > 87.5:
# hysteresis between engage not allowed and uploader not allowed
thermal_status = clip(thermal_status, ThermalStatus.yellow, ThermalStatus.red)
elif max_cpu_temp > 80.0:
# uploader not allowed
thermal_status = ThermalStatus.yellow
elif max_cpu_temp > 75.0:
# hysteresis between uploader not allowed and all good
thermal_status = clip(thermal_status, ThermalStatus.green, ThermalStatus.yellow)
else:
# all good
thermal_status = ThermalStatus.green
# **** starting logic ****
# Check for last update time and display alerts if needed
now = datetime.datetime.now()
try:
last_update = datetime.datetime.fromisoformat(params.get("LastUpdateTime", encoding='utf8'))
except (TypeError, ValueError):
last_update = now
dt = now - last_update
if dt.days > DAYS_NO_CONNECTIVITY_MAX:
if current_connectivity_alert != "expired":
current_connectivity_alert = "expired"
params.delete("Offroad_ConnectivityNeededPrompt")
params.put("Offroad_ConnectivityNeeded", json.dumps(OFFROAD_ALERTS["Offroad_ConnectivityNeeded"]))
elif dt.days > DAYS_NO_CONNECTIVITY_PROMPT:
remaining_time = str(DAYS_NO_CONNECTIVITY_MAX - dt.days)
if current_connectivity_alert != "prompt" + remaining_time:
current_connectivity_alert = "prompt" + remaining_time
alert_connectivity_prompt = copy.copy(OFFROAD_ALERTS["Offroad_ConnectivityNeededPrompt"])
alert_connectivity_prompt["text"] += remaining_time + " days."
params.delete("Offroad_ConnectivityNeeded")
params.put("Offroad_ConnectivityNeededPrompt", json.dumps(alert_connectivity_prompt))
elif current_connectivity_alert is not None:
current_connectivity_alert = None
params.delete("Offroad_ConnectivityNeeded")
params.delete("Offroad_ConnectivityNeededPrompt")
# start constellation of processes when the car starts
ignition = health is not None and health.health.started
do_uninstall = params.get("DoUninstall") == b"1"
accepted_terms = params.get("HasAcceptedTerms") == terms_version
completed_training = params.get("CompletedTrainingVersion") == training_version
should_start = ignition
# have we seen a panda?
passive = (params.get("Passive") == "1")
# with 2% left, we killall, otherwise the phone will take a long time to boot
should_start = should_start and msg.thermal.freeSpace > 0.02
# confirm we have completed training and aren't uninstalling
should_start = should_start and accepted_terms and (passive or completed_training) and (not do_uninstall)
# if any CPU gets above 107 or the battery gets above 63, kill all processes
# controls will warn with CPU above 95 or battery above 60
if thermal_status >= ThermalStatus.danger:
should_start = False
if thermal_status_prev < ThermalStatus.danger:
params.put("Offroad_TemperatureTooHigh", json.dumps(OFFROAD_ALERTS["Offroad_TemperatureTooHigh"]))
else:
if thermal_status_prev >= ThermalStatus.danger:
params.delete("Offroad_TemperatureTooHigh")
if should_start:
off_ts = None
if started_ts is None:
started_ts = sec_since_boot()
started_seen = True
os.system('echo performance > /sys/class/devfreq/soc:qcom,cpubw/governor')
else:
started_ts = None
if off_ts is None:
off_ts = sec_since_boot()
os.system('echo powersave > /sys/class/devfreq/soc:qcom,cpubw/governor')
# shutdown if the battery gets lower than 3%, it's discharging, we aren't running for
# more than a minute but we were running
if msg.thermal.batteryPercent < BATT_PERC_OFF and msg.thermal.batteryStatus == "Discharging" and \
started_seen and (sec_since_boot() - off_ts) > 60:
os.system('LD_LIBRARY_PATH="" svc power shutdown')
msg.thermal.chargingError = current_filter.x > 0. and msg.thermal.batteryPercent < 90 # if current is positive, then battery is being discharged
msg.thermal.started = started_ts is not None
msg.thermal.startedTs = int(1e9*(started_ts or 0))
msg.thermal.thermalStatus = thermal_status
thermal_sock.send(msg.to_bytes())
if usb_power_prev and not usb_power:
params.put("Offroad_ChargeDisabled", json.dumps(OFFROAD_ALERTS["Offroad_ChargeDisabled"]))
elif usb_power and not usb_power_prev:
params.delete("Offroad_ChargeDisabled")
thermal_status_prev = thermal_status
usb_power_prev = usb_power
print(msg)
# report to server once per minute
if (count % int(60. / DT_TRML)) == 0:
cloudlog.event("STATUS_PACKET",
count=count,
health=(health.to_dict() if health else None),
location=(location.to_dict() if location else None),
thermal=msg.to_dict())
count += 1
def main(gctx=None):
thermald_thread()
if __name__ == "__main__":
main()
| global LEON, last_eon_fan_val
if last_eon_fan_val is None or last_eon_fan_val != val:
bus = SMBus(7, force=True)
if LEON:
try:
i = [0x1, 0x3 | 0, 0x3 | 0x08, 0x3 | 0x10][val]
bus.write_i2c_block_data(0x3d, 0, [i])
except IOError:
# tusb320
if val == 0:
bus.write_i2c_block_data(0x67, 0xa, [0])
else:
bus.write_i2c_block_data(0x67, 0xa, [0x20])
bus.write_i2c_block_data(0x67, 0x8, [(val-1)<<6])
else:
bus.write_byte_data(0x21, 0x04, 0x2)
bus.write_byte_data(0x21, 0x03, (val*2)+1)
bus.write_byte_data(0x21, 0x04, 0x4)
bus.close()
last_eon_fan_val = val | identifier_body |
thermald.py | #!/usr/bin/env python3.7
import os
import json
import copy
import datetime
from smbus2 import SMBus
from cereal import log
from common.basedir import BASEDIR
from common.params import Params
from common.realtime import sec_since_boot, DT_TRML
from common.numpy_fast import clip
from common.filter_simple import FirstOrderFilter
from selfdrive.version import terms_version, training_version
from selfdrive.swaglog import cloudlog
import selfdrive.messaging as messaging
from selfdrive.services import service_list
from selfdrive.loggerd.config import get_available_percent
ThermalStatus = log.ThermalData.ThermalStatus
CURRENT_TAU = 15. # 15s time constant
DAYS_NO_CONNECTIVITY_MAX = 7 # do not allow to engage after a week without internet
DAYS_NO_CONNECTIVITY_PROMPT = 4 # send an offroad prompt after 4 days with no internet
with open(BASEDIR + "/selfdrive/controls/lib/alerts_offroad.json") as json_file:
OFFROAD_ALERTS = json.load(json_file)
def read_tz(x):
with open("/sys/devices/virtual/thermal/thermal_zone%d/temp" % x) as f:
ret = max(0, int(f.read()))
return ret
def read_thermal():
dat = messaging.new_message()
dat.init('thermal')
dat.thermal.cpu0 = read_tz(5)
dat.thermal.cpu1 = read_tz(7)
dat.thermal.cpu2 = read_tz(10)
dat.thermal.cpu3 = read_tz(12)
dat.thermal.mem = read_tz(2)
dat.thermal.gpu = read_tz(16)
dat.thermal.bat = read_tz(29)
return dat
LEON = False
def setup_eon_fan():
global LEON
os.system("echo 2 > /sys/module/dwc3_msm/parameters/otg_switch")
bus = SMBus(7, force=True)
try:
bus.write_byte_data(0x21, 0x10, 0xf) # mask all interrupts
bus.write_byte_data(0x21, 0x03, 0x1) # set drive current and global interrupt disable
bus.write_byte_data(0x21, 0x02, 0x2) # needed?
bus.write_byte_data(0x21, 0x04, 0x4) # manual override source
except IOError:
print("LEON detected")
#os.system("echo 1 > /sys/devices/soc/6a00000.ssusb/power_supply/usb/usb_otg")
LEON = True
bus.close()
last_eon_fan_val = None
def set_eon_fan(val):
global LEON, last_eon_fan_val
if last_eon_fan_val is None or last_eon_fan_val != val:
bus = SMBus(7, force=True)
if LEON:
try:
i = [0x1, 0x3 | 0, 0x3 | 0x08, 0x3 | 0x10][val]
bus.write_i2c_block_data(0x3d, 0, [i])
except IOError:
# tusb320
if val == 0:
bus.write_i2c_block_data(0x67, 0xa, [0])
else:
bus.write_i2c_block_data(0x67, 0xa, [0x20])
bus.write_i2c_block_data(0x67, 0x8, [(val-1)<<6])
else:
bus.write_byte_data(0x21, 0x04, 0x2)
bus.write_byte_data(0x21, 0x03, (val*2)+1)
bus.write_byte_data(0x21, 0x04, 0x4)
bus.close()
last_eon_fan_val = val
# temp thresholds to control fan speed - high hysteresis
_TEMP_THRS_H = [50., 65., 80., 10000]
# temp thresholds to control fan speed - low hysteresis
_TEMP_THRS_L = [42.5, 57.5, 72.5, 10000]
# fan speed options
_FAN_SPEEDS = [0, 16384, 32768, 65535]
# max fan speed only allowed if battery is hot
_BAT_TEMP_THERSHOLD = 45.
def handle_fan(max_cpu_temp, bat_temp, fan_speed):
new_speed_h = next(speed for speed, temp_h in zip(_FAN_SPEEDS, _TEMP_THRS_H) if temp_h > max_cpu_temp)
new_speed_l = next(speed for speed, temp_l in zip(_FAN_SPEEDS, _TEMP_THRS_L) if temp_l > max_cpu_temp)
if new_speed_h > fan_speed:
# update speed if using the high thresholds results in fan speed increment
fan_speed = new_speed_h
elif new_speed_l < fan_speed:
# update speed if using the low thresholds results in fan speed decrement
fan_speed = new_speed_l
if bat_temp < _BAT_TEMP_THERSHOLD:
# no max fan speed unless battery is hot
fan_speed = min(fan_speed, _FAN_SPEEDS[-2])
set_eon_fan(fan_speed//16384)
return fan_speed
def thermald_thread():
setup_eon_fan()
# prevent LEECO from undervoltage
BATT_PERC_OFF = 10 if LEON else 3
# now loop
thermal_sock = messaging.pub_sock(service_list['thermal'].port)
health_sock = messaging.sub_sock(service_list['health'].port)
location_sock = messaging.sub_sock(service_list['gpsLocation'].port)
fan_speed = 0
count = 0
off_ts = None
started_ts = None
started_seen = False
thermal_status = ThermalStatus.green
thermal_status_prev = ThermalStatus.green
usb_power = True
usb_power_prev = True
health_sock.RCVTIMEO = int(1000 * 2 * DT_TRML) # 2x the expected health frequency
current_filter = FirstOrderFilter(0., CURRENT_TAU, DT_TRML)
health_prev = None
current_connectivity_alert = None
params = Params()
while 1:
health = messaging.recv_sock(health_sock, wait=True)
location = messaging.recv_sock(location_sock)
location = location.gpsLocation if location else None
msg = read_thermal()
# clear car params when panda gets disconnected
if health is None and health_prev is not None:
params.panda_disconnect()
health_prev = health
if health is not None:
usb_power = health.health.usbPowerMode != log.HealthData.UsbPowerMode.client
# loggerd is gated based on free space
avail = get_available_percent() / 100.0
# thermal message now also includes free space
msg.thermal.freeSpace = avail
with open("/sys/class/power_supply/battery/capacity") as f:
msg.thermal.batteryPercent = int(f.read())
with open("/sys/class/power_supply/battery/status") as f:
msg.thermal.batteryStatus = f.read().strip()
with open("/sys/class/power_supply/battery/current_now") as f:
msg.thermal.batteryCurrent = int(f.read())
with open("/sys/class/power_supply/battery/voltage_now") as f:
msg.thermal.batteryVoltage = int(f.read())
with open("/sys/class/power_supply/usb/present") as f:
msg.thermal.usbOnline = bool(int(f.read()))
current_filter.update(msg.thermal.batteryCurrent / 1e6)
# TODO: add car battery voltage check
max_cpu_temp = max(msg.thermal.cpu0, msg.thermal.cpu1,
msg.thermal.cpu2, msg.thermal.cpu3) / 10.0
max_comp_temp = max(max_cpu_temp, msg.thermal.mem / 10., msg.thermal.gpu / 10.)
bat_temp = msg.thermal.bat/1000.
fan_speed = handle_fan(max_cpu_temp, bat_temp, fan_speed)
msg.thermal.fanSpeed = fan_speed
# thermal logic with hysterisis
if max_cpu_temp > 107. or bat_temp >= 63.:
# onroad not allowed
thermal_status = ThermalStatus.danger
elif max_comp_temp > 92.5 or bat_temp > 60.: # CPU throttling starts around ~90C
# hysteresis between onroad not allowed and engage not allowed
thermal_status = clip(thermal_status, ThermalStatus.red, ThermalStatus.danger)
elif max_cpu_temp > 87.5:
# hysteresis between engage not allowed and uploader not allowed
thermal_status = clip(thermal_status, ThermalStatus.yellow, ThermalStatus.red)
elif max_cpu_temp > 80.0:
# uploader not allowed
thermal_status = ThermalStatus.yellow
elif max_cpu_temp > 75.0:
# hysteresis between uploader not allowed and all good
thermal_status = clip(thermal_status, ThermalStatus.green, ThermalStatus.yellow)
else:
# all good
thermal_status = ThermalStatus.green
# **** starting logic ****
# Check for last update time and display alerts if needed
now = datetime.datetime.now()
try:
last_update = datetime.datetime.fromisoformat(params.get("LastUpdateTime", encoding='utf8'))
except (TypeError, ValueError):
last_update = now
dt = now - last_update
if dt.days > DAYS_NO_CONNECTIVITY_MAX:
if current_connectivity_alert != "expired":
current_connectivity_alert = "expired"
params.delete("Offroad_ConnectivityNeededPrompt")
params.put("Offroad_ConnectivityNeeded", json.dumps(OFFROAD_ALERTS["Offroad_ConnectivityNeeded"]))
elif dt.days > DAYS_NO_CONNECTIVITY_PROMPT:
remaining_time = str(DAYS_NO_CONNECTIVITY_MAX - dt.days)
if current_connectivity_alert != "prompt" + remaining_time:
current_connectivity_alert = "prompt" + remaining_time
alert_connectivity_prompt = copy.copy(OFFROAD_ALERTS["Offroad_ConnectivityNeededPrompt"])
alert_connectivity_prompt["text"] += remaining_time + " days."
params.delete("Offroad_ConnectivityNeeded")
params.put("Offroad_ConnectivityNeededPrompt", json.dumps(alert_connectivity_prompt))
elif current_connectivity_alert is not None:
current_connectivity_alert = None
params.delete("Offroad_ConnectivityNeeded") | ignition = health is not None and health.health.started
do_uninstall = params.get("DoUninstall") == b"1"
accepted_terms = params.get("HasAcceptedTerms") == terms_version
completed_training = params.get("CompletedTrainingVersion") == training_version
should_start = ignition
# have we seen a panda?
passive = (params.get("Passive") == "1")
# with 2% left, we killall, otherwise the phone will take a long time to boot
should_start = should_start and msg.thermal.freeSpace > 0.02
# confirm we have completed training and aren't uninstalling
should_start = should_start and accepted_terms and (passive or completed_training) and (not do_uninstall)
# if any CPU gets above 107 or the battery gets above 63, kill all processes
# controls will warn with CPU above 95 or battery above 60
if thermal_status >= ThermalStatus.danger:
should_start = False
if thermal_status_prev < ThermalStatus.danger:
params.put("Offroad_TemperatureTooHigh", json.dumps(OFFROAD_ALERTS["Offroad_TemperatureTooHigh"]))
else:
if thermal_status_prev >= ThermalStatus.danger:
params.delete("Offroad_TemperatureTooHigh")
if should_start:
off_ts = None
if started_ts is None:
started_ts = sec_since_boot()
started_seen = True
os.system('echo performance > /sys/class/devfreq/soc:qcom,cpubw/governor')
else:
started_ts = None
if off_ts is None:
off_ts = sec_since_boot()
os.system('echo powersave > /sys/class/devfreq/soc:qcom,cpubw/governor')
# shutdown if the battery gets lower than 3%, it's discharging, we aren't running for
# more than a minute but we were running
if msg.thermal.batteryPercent < BATT_PERC_OFF and msg.thermal.batteryStatus == "Discharging" and \
started_seen and (sec_since_boot() - off_ts) > 60:
os.system('LD_LIBRARY_PATH="" svc power shutdown')
msg.thermal.chargingError = current_filter.x > 0. and msg.thermal.batteryPercent < 90 # if current is positive, then battery is being discharged
msg.thermal.started = started_ts is not None
msg.thermal.startedTs = int(1e9*(started_ts or 0))
msg.thermal.thermalStatus = thermal_status
thermal_sock.send(msg.to_bytes())
if usb_power_prev and not usb_power:
params.put("Offroad_ChargeDisabled", json.dumps(OFFROAD_ALERTS["Offroad_ChargeDisabled"]))
elif usb_power and not usb_power_prev:
params.delete("Offroad_ChargeDisabled")
thermal_status_prev = thermal_status
usb_power_prev = usb_power
print(msg)
# report to server once per minute
if (count % int(60. / DT_TRML)) == 0:
cloudlog.event("STATUS_PACKET",
count=count,
health=(health.to_dict() if health else None),
location=(location.to_dict() if location else None),
thermal=msg.to_dict())
count += 1
def main(gctx=None):
thermald_thread()
if __name__ == "__main__":
main() | params.delete("Offroad_ConnectivityNeededPrompt")
# start constellation of processes when the car starts | random_line_split |
thermald.py | #!/usr/bin/env python3.7
import os
import json
import copy
import datetime
from smbus2 import SMBus
from cereal import log
from common.basedir import BASEDIR
from common.params import Params
from common.realtime import sec_since_boot, DT_TRML
from common.numpy_fast import clip
from common.filter_simple import FirstOrderFilter
from selfdrive.version import terms_version, training_version
from selfdrive.swaglog import cloudlog
import selfdrive.messaging as messaging
from selfdrive.services import service_list
from selfdrive.loggerd.config import get_available_percent
ThermalStatus = log.ThermalData.ThermalStatus
CURRENT_TAU = 15. # 15s time constant
DAYS_NO_CONNECTIVITY_MAX = 7 # do not allow to engage after a week without internet
DAYS_NO_CONNECTIVITY_PROMPT = 4 # send an offroad prompt after 4 days with no internet
with open(BASEDIR + "/selfdrive/controls/lib/alerts_offroad.json") as json_file:
OFFROAD_ALERTS = json.load(json_file)
def read_tz(x):
with open("/sys/devices/virtual/thermal/thermal_zone%d/temp" % x) as f:
ret = max(0, int(f.read()))
return ret
def read_thermal():
dat = messaging.new_message()
dat.init('thermal')
dat.thermal.cpu0 = read_tz(5)
dat.thermal.cpu1 = read_tz(7)
dat.thermal.cpu2 = read_tz(10)
dat.thermal.cpu3 = read_tz(12)
dat.thermal.mem = read_tz(2)
dat.thermal.gpu = read_tz(16)
dat.thermal.bat = read_tz(29)
return dat
LEON = False
def setup_eon_fan():
global LEON
os.system("echo 2 > /sys/module/dwc3_msm/parameters/otg_switch")
bus = SMBus(7, force=True)
try:
bus.write_byte_data(0x21, 0x10, 0xf) # mask all interrupts
bus.write_byte_data(0x21, 0x03, 0x1) # set drive current and global interrupt disable
bus.write_byte_data(0x21, 0x02, 0x2) # needed?
bus.write_byte_data(0x21, 0x04, 0x4) # manual override source
except IOError:
print("LEON detected")
#os.system("echo 1 > /sys/devices/soc/6a00000.ssusb/power_supply/usb/usb_otg")
LEON = True
bus.close()
last_eon_fan_val = None
def | (val):
global LEON, last_eon_fan_val
if last_eon_fan_val is None or last_eon_fan_val != val:
bus = SMBus(7, force=True)
if LEON:
try:
i = [0x1, 0x3 | 0, 0x3 | 0x08, 0x3 | 0x10][val]
bus.write_i2c_block_data(0x3d, 0, [i])
except IOError:
# tusb320
if val == 0:
bus.write_i2c_block_data(0x67, 0xa, [0])
else:
bus.write_i2c_block_data(0x67, 0xa, [0x20])
bus.write_i2c_block_data(0x67, 0x8, [(val-1)<<6])
else:
bus.write_byte_data(0x21, 0x04, 0x2)
bus.write_byte_data(0x21, 0x03, (val*2)+1)
bus.write_byte_data(0x21, 0x04, 0x4)
bus.close()
last_eon_fan_val = val
# temp thresholds to control fan speed - high hysteresis
_TEMP_THRS_H = [50., 65., 80., 10000]
# temp thresholds to control fan speed - low hysteresis
_TEMP_THRS_L = [42.5, 57.5, 72.5, 10000]
# fan speed options
_FAN_SPEEDS = [0, 16384, 32768, 65535]
# max fan speed only allowed if battery is hot
_BAT_TEMP_THERSHOLD = 45.
def handle_fan(max_cpu_temp, bat_temp, fan_speed):
new_speed_h = next(speed for speed, temp_h in zip(_FAN_SPEEDS, _TEMP_THRS_H) if temp_h > max_cpu_temp)
new_speed_l = next(speed for speed, temp_l in zip(_FAN_SPEEDS, _TEMP_THRS_L) if temp_l > max_cpu_temp)
if new_speed_h > fan_speed:
# update speed if using the high thresholds results in fan speed increment
fan_speed = new_speed_h
elif new_speed_l < fan_speed:
# update speed if using the low thresholds results in fan speed decrement
fan_speed = new_speed_l
if bat_temp < _BAT_TEMP_THERSHOLD:
# no max fan speed unless battery is hot
fan_speed = min(fan_speed, _FAN_SPEEDS[-2])
set_eon_fan(fan_speed//16384)
return fan_speed
def thermald_thread():
setup_eon_fan()
# prevent LEECO from undervoltage
BATT_PERC_OFF = 10 if LEON else 3
# now loop
thermal_sock = messaging.pub_sock(service_list['thermal'].port)
health_sock = messaging.sub_sock(service_list['health'].port)
location_sock = messaging.sub_sock(service_list['gpsLocation'].port)
fan_speed = 0
count = 0
off_ts = None
started_ts = None
started_seen = False
thermal_status = ThermalStatus.green
thermal_status_prev = ThermalStatus.green
usb_power = True
usb_power_prev = True
health_sock.RCVTIMEO = int(1000 * 2 * DT_TRML) # 2x the expected health frequency
current_filter = FirstOrderFilter(0., CURRENT_TAU, DT_TRML)
health_prev = None
current_connectivity_alert = None
params = Params()
while 1:
health = messaging.recv_sock(health_sock, wait=True)
location = messaging.recv_sock(location_sock)
location = location.gpsLocation if location else None
msg = read_thermal()
# clear car params when panda gets disconnected
if health is None and health_prev is not None:
params.panda_disconnect()
health_prev = health
if health is not None:
usb_power = health.health.usbPowerMode != log.HealthData.UsbPowerMode.client
# loggerd is gated based on free space
avail = get_available_percent() / 100.0
# thermal message now also includes free space
msg.thermal.freeSpace = avail
with open("/sys/class/power_supply/battery/capacity") as f:
msg.thermal.batteryPercent = int(f.read())
with open("/sys/class/power_supply/battery/status") as f:
msg.thermal.batteryStatus = f.read().strip()
with open("/sys/class/power_supply/battery/current_now") as f:
msg.thermal.batteryCurrent = int(f.read())
with open("/sys/class/power_supply/battery/voltage_now") as f:
msg.thermal.batteryVoltage = int(f.read())
with open("/sys/class/power_supply/usb/present") as f:
msg.thermal.usbOnline = bool(int(f.read()))
current_filter.update(msg.thermal.batteryCurrent / 1e6)
# TODO: add car battery voltage check
max_cpu_temp = max(msg.thermal.cpu0, msg.thermal.cpu1,
msg.thermal.cpu2, msg.thermal.cpu3) / 10.0
max_comp_temp = max(max_cpu_temp, msg.thermal.mem / 10., msg.thermal.gpu / 10.)
bat_temp = msg.thermal.bat/1000.
fan_speed = handle_fan(max_cpu_temp, bat_temp, fan_speed)
msg.thermal.fanSpeed = fan_speed
# thermal logic with hysterisis
if max_cpu_temp > 107. or bat_temp >= 63.:
# onroad not allowed
thermal_status = ThermalStatus.danger
elif max_comp_temp > 92.5 or bat_temp > 60.: # CPU throttling starts around ~90C
# hysteresis between onroad not allowed and engage not allowed
thermal_status = clip(thermal_status, ThermalStatus.red, ThermalStatus.danger)
elif max_cpu_temp > 87.5:
# hysteresis between engage not allowed and uploader not allowed
thermal_status = clip(thermal_status, ThermalStatus.yellow, ThermalStatus.red)
elif max_cpu_temp > 80.0:
# uploader not allowed
thermal_status = ThermalStatus.yellow
elif max_cpu_temp > 75.0:
# hysteresis between uploader not allowed and all good
thermal_status = clip(thermal_status, ThermalStatus.green, ThermalStatus.yellow)
else:
# all good
thermal_status = ThermalStatus.green
# **** starting logic ****
# Check for last update time and display alerts if needed
now = datetime.datetime.now()
try:
last_update = datetime.datetime.fromisoformat(params.get("LastUpdateTime", encoding='utf8'))
except (TypeError, ValueError):
last_update = now
dt = now - last_update
if dt.days > DAYS_NO_CONNECTIVITY_MAX:
if current_connectivity_alert != "expired":
current_connectivity_alert = "expired"
params.delete("Offroad_ConnectivityNeededPrompt")
params.put("Offroad_ConnectivityNeeded", json.dumps(OFFROAD_ALERTS["Offroad_ConnectivityNeeded"]))
elif dt.days > DAYS_NO_CONNECTIVITY_PROMPT:
remaining_time = str(DAYS_NO_CONNECTIVITY_MAX - dt.days)
if current_connectivity_alert != "prompt" + remaining_time:
current_connectivity_alert = "prompt" + remaining_time
alert_connectivity_prompt = copy.copy(OFFROAD_ALERTS["Offroad_ConnectivityNeededPrompt"])
alert_connectivity_prompt["text"] += remaining_time + " days."
params.delete("Offroad_ConnectivityNeeded")
params.put("Offroad_ConnectivityNeededPrompt", json.dumps(alert_connectivity_prompt))
elif current_connectivity_alert is not None:
current_connectivity_alert = None
params.delete("Offroad_ConnectivityNeeded")
params.delete("Offroad_ConnectivityNeededPrompt")
# start constellation of processes when the car starts
ignition = health is not None and health.health.started
do_uninstall = params.get("DoUninstall") == b"1"
accepted_terms = params.get("HasAcceptedTerms") == terms_version
completed_training = params.get("CompletedTrainingVersion") == training_version
should_start = ignition
# have we seen a panda?
passive = (params.get("Passive") == "1")
# with 2% left, we killall, otherwise the phone will take a long time to boot
should_start = should_start and msg.thermal.freeSpace > 0.02
# confirm we have completed training and aren't uninstalling
should_start = should_start and accepted_terms and (passive or completed_training) and (not do_uninstall)
# if any CPU gets above 107 or the battery gets above 63, kill all processes
# controls will warn with CPU above 95 or battery above 60
if thermal_status >= ThermalStatus.danger:
should_start = False
if thermal_status_prev < ThermalStatus.danger:
params.put("Offroad_TemperatureTooHigh", json.dumps(OFFROAD_ALERTS["Offroad_TemperatureTooHigh"]))
else:
if thermal_status_prev >= ThermalStatus.danger:
params.delete("Offroad_TemperatureTooHigh")
if should_start:
off_ts = None
if started_ts is None:
started_ts = sec_since_boot()
started_seen = True
os.system('echo performance > /sys/class/devfreq/soc:qcom,cpubw/governor')
else:
started_ts = None
if off_ts is None:
off_ts = sec_since_boot()
os.system('echo powersave > /sys/class/devfreq/soc:qcom,cpubw/governor')
# shutdown if the battery gets lower than 3%, it's discharging, we aren't running for
# more than a minute but we were running
if msg.thermal.batteryPercent < BATT_PERC_OFF and msg.thermal.batteryStatus == "Discharging" and \
started_seen and (sec_since_boot() - off_ts) > 60:
os.system('LD_LIBRARY_PATH="" svc power shutdown')
msg.thermal.chargingError = current_filter.x > 0. and msg.thermal.batteryPercent < 90 # if current is positive, then battery is being discharged
msg.thermal.started = started_ts is not None
msg.thermal.startedTs = int(1e9*(started_ts or 0))
msg.thermal.thermalStatus = thermal_status
thermal_sock.send(msg.to_bytes())
if usb_power_prev and not usb_power:
params.put("Offroad_ChargeDisabled", json.dumps(OFFROAD_ALERTS["Offroad_ChargeDisabled"]))
elif usb_power and not usb_power_prev:
params.delete("Offroad_ChargeDisabled")
thermal_status_prev = thermal_status
usb_power_prev = usb_power
print(msg)
# report to server once per minute
if (count % int(60. / DT_TRML)) == 0:
cloudlog.event("STATUS_PACKET",
count=count,
health=(health.to_dict() if health else None),
location=(location.to_dict() if location else None),
thermal=msg.to_dict())
count += 1
def main(gctx=None):
thermald_thread()
if __name__ == "__main__":
main()
| set_eon_fan | identifier_name |
thermald.py | #!/usr/bin/env python3.7
import os
import json
import copy
import datetime
from smbus2 import SMBus
from cereal import log
from common.basedir import BASEDIR
from common.params import Params
from common.realtime import sec_since_boot, DT_TRML
from common.numpy_fast import clip
from common.filter_simple import FirstOrderFilter
from selfdrive.version import terms_version, training_version
from selfdrive.swaglog import cloudlog
import selfdrive.messaging as messaging
from selfdrive.services import service_list
from selfdrive.loggerd.config import get_available_percent
ThermalStatus = log.ThermalData.ThermalStatus
CURRENT_TAU = 15. # 15s time constant
DAYS_NO_CONNECTIVITY_MAX = 7 # do not allow to engage after a week without internet
DAYS_NO_CONNECTIVITY_PROMPT = 4 # send an offroad prompt after 4 days with no internet
with open(BASEDIR + "/selfdrive/controls/lib/alerts_offroad.json") as json_file:
OFFROAD_ALERTS = json.load(json_file)
def read_tz(x):
with open("/sys/devices/virtual/thermal/thermal_zone%d/temp" % x) as f:
ret = max(0, int(f.read()))
return ret
def read_thermal():
dat = messaging.new_message()
dat.init('thermal')
dat.thermal.cpu0 = read_tz(5)
dat.thermal.cpu1 = read_tz(7)
dat.thermal.cpu2 = read_tz(10)
dat.thermal.cpu3 = read_tz(12)
dat.thermal.mem = read_tz(2)
dat.thermal.gpu = read_tz(16)
dat.thermal.bat = read_tz(29)
return dat
LEON = False
def setup_eon_fan():
global LEON
os.system("echo 2 > /sys/module/dwc3_msm/parameters/otg_switch")
bus = SMBus(7, force=True)
try:
bus.write_byte_data(0x21, 0x10, 0xf) # mask all interrupts
bus.write_byte_data(0x21, 0x03, 0x1) # set drive current and global interrupt disable
bus.write_byte_data(0x21, 0x02, 0x2) # needed?
bus.write_byte_data(0x21, 0x04, 0x4) # manual override source
except IOError:
print("LEON detected")
#os.system("echo 1 > /sys/devices/soc/6a00000.ssusb/power_supply/usb/usb_otg")
LEON = True
bus.close()
last_eon_fan_val = None
def set_eon_fan(val):
global LEON, last_eon_fan_val
if last_eon_fan_val is None or last_eon_fan_val != val:
bus = SMBus(7, force=True)
if LEON:
try:
i = [0x1, 0x3 | 0, 0x3 | 0x08, 0x3 | 0x10][val]
bus.write_i2c_block_data(0x3d, 0, [i])
except IOError:
# tusb320
if val == 0:
bus.write_i2c_block_data(0x67, 0xa, [0])
else:
bus.write_i2c_block_data(0x67, 0xa, [0x20])
bus.write_i2c_block_data(0x67, 0x8, [(val-1)<<6])
else:
bus.write_byte_data(0x21, 0x04, 0x2)
bus.write_byte_data(0x21, 0x03, (val*2)+1)
bus.write_byte_data(0x21, 0x04, 0x4)
bus.close()
last_eon_fan_val = val
# temp thresholds to control fan speed - high hysteresis
_TEMP_THRS_H = [50., 65., 80., 10000]
# temp thresholds to control fan speed - low hysteresis
_TEMP_THRS_L = [42.5, 57.5, 72.5, 10000]
# fan speed options
_FAN_SPEEDS = [0, 16384, 32768, 65535]
# max fan speed only allowed if battery is hot
_BAT_TEMP_THERSHOLD = 45.
def handle_fan(max_cpu_temp, bat_temp, fan_speed):
new_speed_h = next(speed for speed, temp_h in zip(_FAN_SPEEDS, _TEMP_THRS_H) if temp_h > max_cpu_temp)
new_speed_l = next(speed for speed, temp_l in zip(_FAN_SPEEDS, _TEMP_THRS_L) if temp_l > max_cpu_temp)
if new_speed_h > fan_speed:
# update speed if using the high thresholds results in fan speed increment
fan_speed = new_speed_h
elif new_speed_l < fan_speed:
# update speed if using the low thresholds results in fan speed decrement
fan_speed = new_speed_l
if bat_temp < _BAT_TEMP_THERSHOLD:
# no max fan speed unless battery is hot
fan_speed = min(fan_speed, _FAN_SPEEDS[-2])
set_eon_fan(fan_speed//16384)
return fan_speed
def thermald_thread():
setup_eon_fan()
# prevent LEECO from undervoltage
BATT_PERC_OFF = 10 if LEON else 3
# now loop
thermal_sock = messaging.pub_sock(service_list['thermal'].port)
health_sock = messaging.sub_sock(service_list['health'].port)
location_sock = messaging.sub_sock(service_list['gpsLocation'].port)
fan_speed = 0
count = 0
off_ts = None
started_ts = None
started_seen = False
thermal_status = ThermalStatus.green
thermal_status_prev = ThermalStatus.green
usb_power = True
usb_power_prev = True
health_sock.RCVTIMEO = int(1000 * 2 * DT_TRML) # 2x the expected health frequency
current_filter = FirstOrderFilter(0., CURRENT_TAU, DT_TRML)
health_prev = None
current_connectivity_alert = None
params = Params()
while 1:
|
def main(gctx=None):
thermald_thread()
if __name__ == "__main__":
main()
| health = messaging.recv_sock(health_sock, wait=True)
location = messaging.recv_sock(location_sock)
location = location.gpsLocation if location else None
msg = read_thermal()
# clear car params when panda gets disconnected
if health is None and health_prev is not None:
params.panda_disconnect()
health_prev = health
if health is not None:
usb_power = health.health.usbPowerMode != log.HealthData.UsbPowerMode.client
# loggerd is gated based on free space
avail = get_available_percent() / 100.0
# thermal message now also includes free space
msg.thermal.freeSpace = avail
with open("/sys/class/power_supply/battery/capacity") as f:
msg.thermal.batteryPercent = int(f.read())
with open("/sys/class/power_supply/battery/status") as f:
msg.thermal.batteryStatus = f.read().strip()
with open("/sys/class/power_supply/battery/current_now") as f:
msg.thermal.batteryCurrent = int(f.read())
with open("/sys/class/power_supply/battery/voltage_now") as f:
msg.thermal.batteryVoltage = int(f.read())
with open("/sys/class/power_supply/usb/present") as f:
msg.thermal.usbOnline = bool(int(f.read()))
current_filter.update(msg.thermal.batteryCurrent / 1e6)
# TODO: add car battery voltage check
max_cpu_temp = max(msg.thermal.cpu0, msg.thermal.cpu1,
msg.thermal.cpu2, msg.thermal.cpu3) / 10.0
max_comp_temp = max(max_cpu_temp, msg.thermal.mem / 10., msg.thermal.gpu / 10.)
bat_temp = msg.thermal.bat/1000.
fan_speed = handle_fan(max_cpu_temp, bat_temp, fan_speed)
msg.thermal.fanSpeed = fan_speed
# thermal logic with hysterisis
if max_cpu_temp > 107. or bat_temp >= 63.:
# onroad not allowed
thermal_status = ThermalStatus.danger
elif max_comp_temp > 92.5 or bat_temp > 60.: # CPU throttling starts around ~90C
# hysteresis between onroad not allowed and engage not allowed
thermal_status = clip(thermal_status, ThermalStatus.red, ThermalStatus.danger)
elif max_cpu_temp > 87.5:
# hysteresis between engage not allowed and uploader not allowed
thermal_status = clip(thermal_status, ThermalStatus.yellow, ThermalStatus.red)
elif max_cpu_temp > 80.0:
# uploader not allowed
thermal_status = ThermalStatus.yellow
elif max_cpu_temp > 75.0:
# hysteresis between uploader not allowed and all good
thermal_status = clip(thermal_status, ThermalStatus.green, ThermalStatus.yellow)
else:
# all good
thermal_status = ThermalStatus.green
# **** starting logic ****
# Check for last update time and display alerts if needed
now = datetime.datetime.now()
try:
last_update = datetime.datetime.fromisoformat(params.get("LastUpdateTime", encoding='utf8'))
except (TypeError, ValueError):
last_update = now
dt = now - last_update
if dt.days > DAYS_NO_CONNECTIVITY_MAX:
if current_connectivity_alert != "expired":
current_connectivity_alert = "expired"
params.delete("Offroad_ConnectivityNeededPrompt")
params.put("Offroad_ConnectivityNeeded", json.dumps(OFFROAD_ALERTS["Offroad_ConnectivityNeeded"]))
elif dt.days > DAYS_NO_CONNECTIVITY_PROMPT:
remaining_time = str(DAYS_NO_CONNECTIVITY_MAX - dt.days)
if current_connectivity_alert != "prompt" + remaining_time:
current_connectivity_alert = "prompt" + remaining_time
alert_connectivity_prompt = copy.copy(OFFROAD_ALERTS["Offroad_ConnectivityNeededPrompt"])
alert_connectivity_prompt["text"] += remaining_time + " days."
params.delete("Offroad_ConnectivityNeeded")
params.put("Offroad_ConnectivityNeededPrompt", json.dumps(alert_connectivity_prompt))
elif current_connectivity_alert is not None:
current_connectivity_alert = None
params.delete("Offroad_ConnectivityNeeded")
params.delete("Offroad_ConnectivityNeededPrompt")
# start constellation of processes when the car starts
ignition = health is not None and health.health.started
do_uninstall = params.get("DoUninstall") == b"1"
accepted_terms = params.get("HasAcceptedTerms") == terms_version
completed_training = params.get("CompletedTrainingVersion") == training_version
should_start = ignition
# have we seen a panda?
passive = (params.get("Passive") == "1")
# with 2% left, we killall, otherwise the phone will take a long time to boot
should_start = should_start and msg.thermal.freeSpace > 0.02
# confirm we have completed training and aren't uninstalling
should_start = should_start and accepted_terms and (passive or completed_training) and (not do_uninstall)
# if any CPU gets above 107 or the battery gets above 63, kill all processes
# controls will warn with CPU above 95 or battery above 60
if thermal_status >= ThermalStatus.danger:
should_start = False
if thermal_status_prev < ThermalStatus.danger:
params.put("Offroad_TemperatureTooHigh", json.dumps(OFFROAD_ALERTS["Offroad_TemperatureTooHigh"]))
else:
if thermal_status_prev >= ThermalStatus.danger:
params.delete("Offroad_TemperatureTooHigh")
if should_start:
off_ts = None
if started_ts is None:
started_ts = sec_since_boot()
started_seen = True
os.system('echo performance > /sys/class/devfreq/soc:qcom,cpubw/governor')
else:
started_ts = None
if off_ts is None:
off_ts = sec_since_boot()
os.system('echo powersave > /sys/class/devfreq/soc:qcom,cpubw/governor')
# shutdown if the battery gets lower than 3%, it's discharging, we aren't running for
# more than a minute but we were running
if msg.thermal.batteryPercent < BATT_PERC_OFF and msg.thermal.batteryStatus == "Discharging" and \
started_seen and (sec_since_boot() - off_ts) > 60:
os.system('LD_LIBRARY_PATH="" svc power shutdown')
msg.thermal.chargingError = current_filter.x > 0. and msg.thermal.batteryPercent < 90 # if current is positive, then battery is being discharged
msg.thermal.started = started_ts is not None
msg.thermal.startedTs = int(1e9*(started_ts or 0))
msg.thermal.thermalStatus = thermal_status
thermal_sock.send(msg.to_bytes())
if usb_power_prev and not usb_power:
params.put("Offroad_ChargeDisabled", json.dumps(OFFROAD_ALERTS["Offroad_ChargeDisabled"]))
elif usb_power and not usb_power_prev:
params.delete("Offroad_ChargeDisabled")
thermal_status_prev = thermal_status
usb_power_prev = usb_power
print(msg)
# report to server once per minute
if (count % int(60. / DT_TRML)) == 0:
cloudlog.event("STATUS_PACKET",
count=count,
health=(health.to_dict() if health else None),
location=(location.to_dict() if location else None),
thermal=msg.to_dict())
count += 1 | conditional_block |
falcon.py | """
Python implementation of Falcon:
https://falcon-sign.info/.
"""
from .common import q
from numpy import set_printoptions
from math import sqrt
from .fft import fft, ifft, sub, neg, add_fft, mul_fft
from .ntt import sub_zq, mul_zq, div_zq
from .ffsampling import gram, ffldl_fft, ffsampling_fft
from .ntrugen import ntru_gen
from .encoding import compress, decompress
# https://pycryptodome.readthedocs.io/en/latest/src/hash/shake256.html
from Crypto.Hash import SHAKE256
# Randomness
from os import urandom
from .rng import ChaCha20
# For debugging purposes
import sys
if sys.version_info >= (3, 4):
from importlib import reload # Python 3.4+ only.
set_printoptions(linewidth=200, precision=5, suppress=True)
logn = {
2: 1,
4: 2,
8: 3,
16: 4,
32: 5,
64: 6,
128: 7,
256: 8,
512: 9,
1024: 10
}
# Bytelength of the signing salt and header
HEAD_LEN = 1
SALT_LEN = 40
SEED_LEN = 56
# Parameter sets for Falcon:
# - n is the dimension/degree of the cyclotomic ring
# - sigma is the std. dev. of signatures (Gaussians over a lattice)
# - sigmin is a lower bounds on the std. dev. of each Gaussian over Z
# - sigbound is the upper bound on ||s0||^2 + ||s1||^2
# - sig_bytelen is the bytelength of signatures
Params = {
# FalconParam(2, 2)
2: {
"n": 2,
"sigma": 144.81253976308423,
"sigmin": 1.1165085072329104,
"sig_bound": 101498,
"sig_bytelen": 44,
},
# FalconParam(4, 2)
4: {
"n": 4,
"sigma": 146.83798833523608,
"sigmin": 1.1321247692325274,
"sig_bound": 208714,
"sig_bytelen": 47,
},
# FalconParam(8, 2)
8: {
"n": 8,
"sigma": 148.83587593064718,
"sigmin": 1.147528535373367,
"sig_bound": 428865,
"sig_bytelen": 52,
},
# FalconParam(16, 4)
16: {
"n": 16,
"sigma": 151.78340713845503,
"sigmin": 1.170254078853483,
"sig_bound": 892039,
"sig_bytelen": 63,
},
# FalconParam(32, 8)
32: {
"n": 32,
"sigma": 154.6747794602761,
"sigmin": 1.1925466358390344,
"sig_bound": 1852696,
"sig_bytelen": 82,
},
# FalconParam(64, 16)
64: {
"n": 64,
"sigma": 157.51308555044122,
"sigmin": 1.2144300507766141,
"sig_bound": 3842630,
"sig_bytelen": 122,
},
# FalconParam(128, 32)
128: {
"n": 128,
"sigma": 160.30114421975344,
"sigmin": 1.235926056771981,
"sig_bound": 7959734,
"sig_bytelen": 200,
},
# FalconParam(256, 64)
256: {
"n": 256,
"sigma": 163.04153322607107,
"sigmin": 1.2570545284063217,
"sig_bound": 16468416,
"sig_bytelen": 356,
},
# FalconParam(512, 128)
512: {
"n": 512,
"sigma": 165.7366171829776,
"sigmin": 1.2778336969128337,
"sig_bound": 34034726,
"sig_bytelen": 666,
},
# FalconParam(1024, 256)
1024: {
"n": 1024,
"sigma": 168.38857144654395,
"sigmin": 1.298280334344292,
"sig_bound": 70265242,
"sig_bytelen": 1280,
},
}
def print_tree(tree, pref=""):
"""
Display a LDL tree in a readable form.
Args:
T: a LDL tree
Format: coefficient or fft
"""
leaf = "|_____> "
top = "|_______"
son1 = "| "
son2 = " "
width = len(top)
a = ""
if len(tree) == 3:
if (pref == ""):
a += pref + str(tree[0]) + "\n"
else:
a += pref[:-width] + top + str(tree[0]) + "\n"
a += print_tree(tree[1], pref + son1)
a += print_tree(tree[2], pref + son2)
return a
else:
return (pref[:-width] + leaf + str(tree) + "\n")
def normalize_tree(tree, sigma):
"""
Normalize leaves of a LDL tree (from values ||b_i||**2 to sigma/||b_i||).
Args:
T: a LDL tree
sigma: a standard deviation
Format: coefficient or fft
"""
if len(tree) == 3:
normalize_tree(tree[1], sigma)
normalize_tree(tree[2], sigma)
else:
tree[0] = sigma / sqrt(tree[0].real)
tree[1] = 0
class PublicKey:
"""
This class contains methods for performing public key operations in Falcon.
"""
def __init__(self, sk=None, n=None, h=None):
"""Initialize a public key."""
if sk:
self.n = sk.n
self.h = sk.h
elif n and h:
self.n = n
self.h = h
else:
raise Exception("Public Key construction failed: insufficient/wrong arguments")
self.signature_bound = Params[self.n]["sig_bound"]
self.sig_bytelen = Params[self.n]["sig_bytelen"]
def __repr__(self):
"""Print the object in readable form."""
rep = "Public for n = {n}:\n\n".format(n=self.n)
rep += "h = {h}\n".format(h=self.h)
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def verify(self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True
class SecretKey:
"""
This class contains methods for performing
secret key operations (and also public key operations) in Falcon.
One can:
- initialize a secret key for:
- n = 128, 256, 512, 1024,
- phi = x ** n + 1,
- q = 12 * 1024 + 1
- find a preimage t of a point c (both in ( Z[x] mod (Phi,q) )**2 ) such that t*B0 = c
- hash a message to a point of Z[x] mod (Phi,q)
- sign a message
- verify the signature of a message
"""
def __init__(self, n, polys=None):
"""Initialize a secret key."""
# Public parameters
self.n = n
self.sigma = Params[n]["sigma"]
self.sigmin = Params[n]["sigmin"]
self.signature_bound = Params[n]["sig_bound"]
self.sig_bytelen = Params[n]["sig_bytelen"]
# Compute NTRU polynomials f, g, F, G verifying fG - gF = q mod Phi
if polys is None:
self.f, self.g, self.F, self.G = ntru_gen(n)
else:
[f, g, F, G] = polys
assert all((len(poly) == n) for poly in [f, g, F, G])
self.f = f[:]
self.g = g[:]
self.F = F[:]
self.G = G[:]
# From f, g, F, G, compute the basis B0 of a NTRU lattice
# as well as its Gram matrix and their fft's.
B0 = [[self.g, neg(self.f)], [self.G, neg(self.F)]]
G0 = gram(B0)
self.B0_fft = [[fft(elt) for elt in row] for row in B0]
G0_fft = [[fft(elt) for elt in row] for row in G0]
self.T_fft = ffldl_fft(G0_fft)
# Normalize Falcon tree
normalize_tree(self.T_fft, self.sigma)
# The public key is a polynomial such that h*f = g mod (Phi,q)
self.h = div_zq(self.g, self.f)
def __repr__(self, verbose=False):
"""Print the object in readable form."""
rep = "Private key for n = {n}:\n\n".format(n=self.n)
rep += "f = {f}\n\n".format(f=self.f)
rep += "g = {g}\n\n".format(g=self.g)
rep += "F = {F}\n\n".format(F=self.F)
rep += "G = {G}\n\n".format(G=self.G)
if verbose:
rep += "\nFFT tree\n"
rep += print_tree(self.T_fft, pref="")
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def sample_preimage(self, point, seed=None):
"""
Sample a short vector s such that s[0] + s[1] * h = point.
"""
[[a, b], [c, d]] = self.B0_fft
# We compute a vector t_fft such that:
# (fft(point), fft(0)) * B0_fft = t_fft
# Because fft(0) = 0 and the inverse of B has a very specific form,
# we can do several optimizations.
'''
print("---------Inside sample_preimage----------")
print("point: ", point)
'''
point_fft = fft(point)
t0_fft = [(point_fft[i] * d[i]) / q for i in range(self.n)]
t1_fft = [(-point_fft[i] * b[i]) / q for i in range(self.n)]
t_fft = [t0_fft, t1_fft]
# We now compute v such that:
# v = z * B0 for an integral vector z
# v is close to (point, 0)
if seed is None:
# If no seed is defined, use urandom as the pseudo-random source.
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin, urandom)
else:
# If a seed is defined, initialize a ChaCha20 PRG
# that is used to generate pseudo-randomness.
chacha_prng = ChaCha20(seed)
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin,
chacha_prng.randombytes)
v0_fft = add_fft(mul_fft(z_fft[0], a), mul_fft(z_fft[1], c))
v1_fft = add_fft(mul_fft(z_fft[0], b), mul_fft(z_fft[1], d))
v0 = [int(round(elt)) for elt in ifft(v0_fft)]
v1 = [int(round(elt)) for elt in ifft(v1_fft)]
# The difference s = (point, 0) - v is such that:
# s is short
# s[0] + s[1] * h = point
s = [sub(point, v0), neg(v1)]
return s
def sign(self, message, randombytes=urandom):
"""
Sign a message. The message MUST be a byte string or byte array.
Optionally, one can select the source of (pseudo-)randomness used
(default: urandom).
"""
int_header = 0x30 + logn[self.n]
header = int_header.to_bytes(1, "little")
salt = randombytes(SALT_LEN)
hashed = self.hash_to_point(message, salt)
# We repeat the signing procedure until we find a signature that is
# short enough (both the Euclidean norm and the bytelength)
'''
print("---------Inside sign----------")
'''
while(1):
if (randombytes == urandom):
s = self.sample_preimage(hashed)
'''
print("s: ", s)
'''
else:
seed = randombytes(SEED_LEN)
s = self.sample_preimage(hashed, seed=seed)
norm_sign = sum(coef ** 2 for coef in s[0])
norm_sign += sum(coef ** 2 for coef in s[1])
# Check the Euclidean norm
if norm_sign <= self.signature_bound:
enc_s = compress(s[1], self.sig_bytelen - HEAD_LEN - SALT_LEN)
# Check that the encoding is valid (sometimes it fails)
if (enc_s is not False):
return header + salt + enc_s
'''
else:
print("-------------INVALID encoding---------------")
else:
print("-------------NOT within signature bound---------------")
'''
def verify(self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
| s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True | random_line_split | |
falcon.py | """
Python implementation of Falcon:
https://falcon-sign.info/.
"""
from .common import q
from numpy import set_printoptions
from math import sqrt
from .fft import fft, ifft, sub, neg, add_fft, mul_fft
from .ntt import sub_zq, mul_zq, div_zq
from .ffsampling import gram, ffldl_fft, ffsampling_fft
from .ntrugen import ntru_gen
from .encoding import compress, decompress
# https://pycryptodome.readthedocs.io/en/latest/src/hash/shake256.html
from Crypto.Hash import SHAKE256
# Randomness
from os import urandom
from .rng import ChaCha20
# For debugging purposes
import sys
if sys.version_info >= (3, 4):
from importlib import reload # Python 3.4+ only.
set_printoptions(linewidth=200, precision=5, suppress=True)
logn = {
2: 1,
4: 2,
8: 3,
16: 4,
32: 5,
64: 6,
128: 7,
256: 8,
512: 9,
1024: 10
}
# Bytelength of the signing salt and header
HEAD_LEN = 1
SALT_LEN = 40
SEED_LEN = 56
# Parameter sets for Falcon:
# - n is the dimension/degree of the cyclotomic ring
# - sigma is the std. dev. of signatures (Gaussians over a lattice)
# - sigmin is a lower bounds on the std. dev. of each Gaussian over Z
# - sigbound is the upper bound on ||s0||^2 + ||s1||^2
# - sig_bytelen is the bytelength of signatures
Params = {
# FalconParam(2, 2)
2: {
"n": 2,
"sigma": 144.81253976308423,
"sigmin": 1.1165085072329104,
"sig_bound": 101498,
"sig_bytelen": 44,
},
# FalconParam(4, 2)
4: {
"n": 4,
"sigma": 146.83798833523608,
"sigmin": 1.1321247692325274,
"sig_bound": 208714,
"sig_bytelen": 47,
},
# FalconParam(8, 2)
8: {
"n": 8,
"sigma": 148.83587593064718,
"sigmin": 1.147528535373367,
"sig_bound": 428865,
"sig_bytelen": 52,
},
# FalconParam(16, 4)
16: {
"n": 16,
"sigma": 151.78340713845503,
"sigmin": 1.170254078853483,
"sig_bound": 892039,
"sig_bytelen": 63,
},
# FalconParam(32, 8)
32: {
"n": 32,
"sigma": 154.6747794602761,
"sigmin": 1.1925466358390344,
"sig_bound": 1852696,
"sig_bytelen": 82,
},
# FalconParam(64, 16)
64: {
"n": 64,
"sigma": 157.51308555044122,
"sigmin": 1.2144300507766141,
"sig_bound": 3842630,
"sig_bytelen": 122,
},
# FalconParam(128, 32)
128: {
"n": 128,
"sigma": 160.30114421975344,
"sigmin": 1.235926056771981,
"sig_bound": 7959734,
"sig_bytelen": 200,
},
# FalconParam(256, 64)
256: {
"n": 256,
"sigma": 163.04153322607107,
"sigmin": 1.2570545284063217,
"sig_bound": 16468416,
"sig_bytelen": 356,
},
# FalconParam(512, 128)
512: {
"n": 512,
"sigma": 165.7366171829776,
"sigmin": 1.2778336969128337,
"sig_bound": 34034726,
"sig_bytelen": 666,
},
# FalconParam(1024, 256)
1024: {
"n": 1024,
"sigma": 168.38857144654395,
"sigmin": 1.298280334344292,
"sig_bound": 70265242,
"sig_bytelen": 1280,
},
}
def print_tree(tree, pref=""):
"""
Display a LDL tree in a readable form.
Args:
T: a LDL tree
Format: coefficient or fft
"""
leaf = "|_____> "
top = "|_______"
son1 = "| "
son2 = " "
width = len(top)
a = ""
if len(tree) == 3:
if (pref == ""):
a += pref + str(tree[0]) + "\n"
else:
a += pref[:-width] + top + str(tree[0]) + "\n"
a += print_tree(tree[1], pref + son1)
a += print_tree(tree[2], pref + son2)
return a
else:
return (pref[:-width] + leaf + str(tree) + "\n")
def normalize_tree(tree, sigma):
"""
Normalize leaves of a LDL tree (from values ||b_i||**2 to sigma/||b_i||).
Args:
T: a LDL tree
sigma: a standard deviation
Format: coefficient or fft
"""
if len(tree) == 3:
normalize_tree(tree[1], sigma)
normalize_tree(tree[2], sigma)
else:
tree[0] = sigma / sqrt(tree[0].real)
tree[1] = 0
class PublicKey:
"""
This class contains methods for performing public key operations in Falcon.
"""
def __init__(self, sk=None, n=None, h=None):
"""Initialize a public key."""
if sk:
self.n = sk.n
self.h = sk.h
elif n and h:
self.n = n
self.h = h
else:
raise Exception("Public Key construction failed: insufficient/wrong arguments")
self.signature_bound = Params[self.n]["sig_bound"]
self.sig_bytelen = Params[self.n]["sig_bytelen"]
def __repr__(self):
"""Print the object in readable form."""
rep = "Public for n = {n}:\n\n".format(n=self.n)
rep += "h = {h}\n".format(h=self.h)
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def verify(self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True
class SecretKey:
"""
This class contains methods for performing
secret key operations (and also public key operations) in Falcon.
One can:
- initialize a secret key for:
- n = 128, 256, 512, 1024,
- phi = x ** n + 1,
- q = 12 * 1024 + 1
- find a preimage t of a point c (both in ( Z[x] mod (Phi,q) )**2 ) such that t*B0 = c
- hash a message to a point of Z[x] mod (Phi,q)
- sign a message
- verify the signature of a message
"""
def __init__(self, n, polys=None):
"""Initialize a secret key."""
# Public parameters
self.n = n
self.sigma = Params[n]["sigma"]
self.sigmin = Params[n]["sigmin"]
self.signature_bound = Params[n]["sig_bound"]
self.sig_bytelen = Params[n]["sig_bytelen"]
# Compute NTRU polynomials f, g, F, G verifying fG - gF = q mod Phi
if polys is None:
self.f, self.g, self.F, self.G = ntru_gen(n)
else:
[f, g, F, G] = polys
assert all((len(poly) == n) for poly in [f, g, F, G])
self.f = f[:]
self.g = g[:]
self.F = F[:]
self.G = G[:]
# From f, g, F, G, compute the basis B0 of a NTRU lattice
# as well as its Gram matrix and their fft's.
B0 = [[self.g, neg(self.f)], [self.G, neg(self.F)]]
G0 = gram(B0)
self.B0_fft = [[fft(elt) for elt in row] for row in B0]
G0_fft = [[fft(elt) for elt in row] for row in G0]
self.T_fft = ffldl_fft(G0_fft)
# Normalize Falcon tree
normalize_tree(self.T_fft, self.sigma)
# The public key is a polynomial such that h*f = g mod (Phi,q)
self.h = div_zq(self.g, self.f)
def __repr__(self, verbose=False):
"""Print the object in readable form."""
rep = "Private key for n = {n}:\n\n".format(n=self.n)
rep += "f = {f}\n\n".format(f=self.f)
rep += "g = {g}\n\n".format(g=self.g)
rep += "F = {F}\n\n".format(F=self.F)
rep += "G = {G}\n\n".format(G=self.G)
if verbose:
rep += "\nFFT tree\n"
rep += print_tree(self.T_fft, pref="")
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def sample_preimage(self, point, seed=None):
"""
Sample a short vector s such that s[0] + s[1] * h = point.
"""
[[a, b], [c, d]] = self.B0_fft
# We compute a vector t_fft such that:
# (fft(point), fft(0)) * B0_fft = t_fft
# Because fft(0) = 0 and the inverse of B has a very specific form,
# we can do several optimizations.
'''
print("---------Inside sample_preimage----------")
print("point: ", point)
'''
point_fft = fft(point)
t0_fft = [(point_fft[i] * d[i]) / q for i in range(self.n)]
t1_fft = [(-point_fft[i] * b[i]) / q for i in range(self.n)]
t_fft = [t0_fft, t1_fft]
# We now compute v such that:
# v = z * B0 for an integral vector z
# v is close to (point, 0)
if seed is None:
# If no seed is defined, use urandom as the pseudo-random source.
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin, urandom)
else:
# If a seed is defined, initialize a ChaCha20 PRG
# that is used to generate pseudo-randomness.
chacha_prng = ChaCha20(seed)
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin,
chacha_prng.randombytes)
v0_fft = add_fft(mul_fft(z_fft[0], a), mul_fft(z_fft[1], c))
v1_fft = add_fft(mul_fft(z_fft[0], b), mul_fft(z_fft[1], d))
v0 = [int(round(elt)) for elt in ifft(v0_fft)]
v1 = [int(round(elt)) for elt in ifft(v1_fft)]
# The difference s = (point, 0) - v is such that:
# s is short
# s[0] + s[1] * h = point
s = [sub(point, v0), neg(v1)]
return s
def sign(self, message, randombytes=urandom):
"""
Sign a message. The message MUST be a byte string or byte array.
Optionally, one can select the source of (pseudo-)randomness used
(default: urandom).
"""
int_header = 0x30 + logn[self.n]
header = int_header.to_bytes(1, "little")
salt = randombytes(SALT_LEN)
hashed = self.hash_to_point(message, salt)
# We repeat the signing procedure until we find a signature that is
# short enough (both the Euclidean norm and the bytelength)
'''
print("---------Inside sign----------")
'''
while(1):
if (randombytes == urandom):
s = self.sample_preimage(hashed)
'''
print("s: ", s)
'''
else:
seed = randombytes(SEED_LEN)
s = self.sample_preimage(hashed, seed=seed)
norm_sign = sum(coef ** 2 for coef in s[0])
norm_sign += sum(coef ** 2 for coef in s[1])
# Check the Euclidean norm
if norm_sign <= self.signature_bound:
enc_s = compress(s[1], self.sig_bytelen - HEAD_LEN - SALT_LEN)
# Check that the encoding is valid (sometimes it fails)
if (enc_s is not False):
|
'''
else:
print("-------------INVALID encoding---------------")
else:
print("-------------NOT within signature bound---------------")
'''
def verify(self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True
| return header + salt + enc_s | conditional_block |
falcon.py | """
Python implementation of Falcon:
https://falcon-sign.info/.
"""
from .common import q
from numpy import set_printoptions
from math import sqrt
from .fft import fft, ifft, sub, neg, add_fft, mul_fft
from .ntt import sub_zq, mul_zq, div_zq
from .ffsampling import gram, ffldl_fft, ffsampling_fft
from .ntrugen import ntru_gen
from .encoding import compress, decompress
# https://pycryptodome.readthedocs.io/en/latest/src/hash/shake256.html
from Crypto.Hash import SHAKE256
# Randomness
from os import urandom
from .rng import ChaCha20
# For debugging purposes
import sys
if sys.version_info >= (3, 4):
from importlib import reload # Python 3.4+ only.
set_printoptions(linewidth=200, precision=5, suppress=True)
logn = {
2: 1,
4: 2,
8: 3,
16: 4,
32: 5,
64: 6,
128: 7,
256: 8,
512: 9,
1024: 10
}
# Bytelength of the signing salt and header
HEAD_LEN = 1
SALT_LEN = 40
SEED_LEN = 56
# Parameter sets for Falcon:
# - n is the dimension/degree of the cyclotomic ring
# - sigma is the std. dev. of signatures (Gaussians over a lattice)
# - sigmin is a lower bounds on the std. dev. of each Gaussian over Z
# - sigbound is the upper bound on ||s0||^2 + ||s1||^2
# - sig_bytelen is the bytelength of signatures
Params = {
# FalconParam(2, 2)
2: {
"n": 2,
"sigma": 144.81253976308423,
"sigmin": 1.1165085072329104,
"sig_bound": 101498,
"sig_bytelen": 44,
},
# FalconParam(4, 2)
4: {
"n": 4,
"sigma": 146.83798833523608,
"sigmin": 1.1321247692325274,
"sig_bound": 208714,
"sig_bytelen": 47,
},
# FalconParam(8, 2)
8: {
"n": 8,
"sigma": 148.83587593064718,
"sigmin": 1.147528535373367,
"sig_bound": 428865,
"sig_bytelen": 52,
},
# FalconParam(16, 4)
16: {
"n": 16,
"sigma": 151.78340713845503,
"sigmin": 1.170254078853483,
"sig_bound": 892039,
"sig_bytelen": 63,
},
# FalconParam(32, 8)
32: {
"n": 32,
"sigma": 154.6747794602761,
"sigmin": 1.1925466358390344,
"sig_bound": 1852696,
"sig_bytelen": 82,
},
# FalconParam(64, 16)
64: {
"n": 64,
"sigma": 157.51308555044122,
"sigmin": 1.2144300507766141,
"sig_bound": 3842630,
"sig_bytelen": 122,
},
# FalconParam(128, 32)
128: {
"n": 128,
"sigma": 160.30114421975344,
"sigmin": 1.235926056771981,
"sig_bound": 7959734,
"sig_bytelen": 200,
},
# FalconParam(256, 64)
256: {
"n": 256,
"sigma": 163.04153322607107,
"sigmin": 1.2570545284063217,
"sig_bound": 16468416,
"sig_bytelen": 356,
},
# FalconParam(512, 128)
512: {
"n": 512,
"sigma": 165.7366171829776,
"sigmin": 1.2778336969128337,
"sig_bound": 34034726,
"sig_bytelen": 666,
},
# FalconParam(1024, 256)
1024: {
"n": 1024,
"sigma": 168.38857144654395,
"sigmin": 1.298280334344292,
"sig_bound": 70265242,
"sig_bytelen": 1280,
},
}
def print_tree(tree, pref=""):
"""
Display a LDL tree in a readable form.
Args:
T: a LDL tree
Format: coefficient or fft
"""
leaf = "|_____> "
top = "|_______"
son1 = "| "
son2 = " "
width = len(top)
a = ""
if len(tree) == 3:
if (pref == ""):
a += pref + str(tree[0]) + "\n"
else:
a += pref[:-width] + top + str(tree[0]) + "\n"
a += print_tree(tree[1], pref + son1)
a += print_tree(tree[2], pref + son2)
return a
else:
return (pref[:-width] + leaf + str(tree) + "\n")
def normalize_tree(tree, sigma):
"""
Normalize leaves of a LDL tree (from values ||b_i||**2 to sigma/||b_i||).
Args:
T: a LDL tree
sigma: a standard deviation
Format: coefficient or fft
"""
if len(tree) == 3:
normalize_tree(tree[1], sigma)
normalize_tree(tree[2], sigma)
else:
tree[0] = sigma / sqrt(tree[0].real)
tree[1] = 0
class PublicKey:
"""
This class contains methods for performing public key operations in Falcon.
"""
def __init__(self, sk=None, n=None, h=None):
"""Initialize a public key."""
if sk:
self.n = sk.n
self.h = sk.h
elif n and h:
self.n = n
self.h = h
else:
raise Exception("Public Key construction failed: insufficient/wrong arguments")
self.signature_bound = Params[self.n]["sig_bound"]
self.sig_bytelen = Params[self.n]["sig_bytelen"]
def __repr__(self):
"""Print the object in readable form."""
rep = "Public for n = {n}:\n\n".format(n=self.n)
rep += "h = {h}\n".format(h=self.h)
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def | (self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True
class SecretKey:
"""
This class contains methods for performing
secret key operations (and also public key operations) in Falcon.
One can:
- initialize a secret key for:
- n = 128, 256, 512, 1024,
- phi = x ** n + 1,
- q = 12 * 1024 + 1
- find a preimage t of a point c (both in ( Z[x] mod (Phi,q) )**2 ) such that t*B0 = c
- hash a message to a point of Z[x] mod (Phi,q)
- sign a message
- verify the signature of a message
"""
def __init__(self, n, polys=None):
"""Initialize a secret key."""
# Public parameters
self.n = n
self.sigma = Params[n]["sigma"]
self.sigmin = Params[n]["sigmin"]
self.signature_bound = Params[n]["sig_bound"]
self.sig_bytelen = Params[n]["sig_bytelen"]
# Compute NTRU polynomials f, g, F, G verifying fG - gF = q mod Phi
if polys is None:
self.f, self.g, self.F, self.G = ntru_gen(n)
else:
[f, g, F, G] = polys
assert all((len(poly) == n) for poly in [f, g, F, G])
self.f = f[:]
self.g = g[:]
self.F = F[:]
self.G = G[:]
# From f, g, F, G, compute the basis B0 of a NTRU lattice
# as well as its Gram matrix and their fft's.
B0 = [[self.g, neg(self.f)], [self.G, neg(self.F)]]
G0 = gram(B0)
self.B0_fft = [[fft(elt) for elt in row] for row in B0]
G0_fft = [[fft(elt) for elt in row] for row in G0]
self.T_fft = ffldl_fft(G0_fft)
# Normalize Falcon tree
normalize_tree(self.T_fft, self.sigma)
# The public key is a polynomial such that h*f = g mod (Phi,q)
self.h = div_zq(self.g, self.f)
def __repr__(self, verbose=False):
"""Print the object in readable form."""
rep = "Private key for n = {n}:\n\n".format(n=self.n)
rep += "f = {f}\n\n".format(f=self.f)
rep += "g = {g}\n\n".format(g=self.g)
rep += "F = {F}\n\n".format(F=self.F)
rep += "G = {G}\n\n".format(G=self.G)
if verbose:
rep += "\nFFT tree\n"
rep += print_tree(self.T_fft, pref="")
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def sample_preimage(self, point, seed=None):
"""
Sample a short vector s such that s[0] + s[1] * h = point.
"""
[[a, b], [c, d]] = self.B0_fft
# We compute a vector t_fft such that:
# (fft(point), fft(0)) * B0_fft = t_fft
# Because fft(0) = 0 and the inverse of B has a very specific form,
# we can do several optimizations.
'''
print("---------Inside sample_preimage----------")
print("point: ", point)
'''
point_fft = fft(point)
t0_fft = [(point_fft[i] * d[i]) / q for i in range(self.n)]
t1_fft = [(-point_fft[i] * b[i]) / q for i in range(self.n)]
t_fft = [t0_fft, t1_fft]
# We now compute v such that:
# v = z * B0 for an integral vector z
# v is close to (point, 0)
if seed is None:
# If no seed is defined, use urandom as the pseudo-random source.
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin, urandom)
else:
# If a seed is defined, initialize a ChaCha20 PRG
# that is used to generate pseudo-randomness.
chacha_prng = ChaCha20(seed)
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin,
chacha_prng.randombytes)
v0_fft = add_fft(mul_fft(z_fft[0], a), mul_fft(z_fft[1], c))
v1_fft = add_fft(mul_fft(z_fft[0], b), mul_fft(z_fft[1], d))
v0 = [int(round(elt)) for elt in ifft(v0_fft)]
v1 = [int(round(elt)) for elt in ifft(v1_fft)]
# The difference s = (point, 0) - v is such that:
# s is short
# s[0] + s[1] * h = point
s = [sub(point, v0), neg(v1)]
return s
def sign(self, message, randombytes=urandom):
"""
Sign a message. The message MUST be a byte string or byte array.
Optionally, one can select the source of (pseudo-)randomness used
(default: urandom).
"""
int_header = 0x30 + logn[self.n]
header = int_header.to_bytes(1, "little")
salt = randombytes(SALT_LEN)
hashed = self.hash_to_point(message, salt)
# We repeat the signing procedure until we find a signature that is
# short enough (both the Euclidean norm and the bytelength)
'''
print("---------Inside sign----------")
'''
while(1):
if (randombytes == urandom):
s = self.sample_preimage(hashed)
'''
print("s: ", s)
'''
else:
seed = randombytes(SEED_LEN)
s = self.sample_preimage(hashed, seed=seed)
norm_sign = sum(coef ** 2 for coef in s[0])
norm_sign += sum(coef ** 2 for coef in s[1])
# Check the Euclidean norm
if norm_sign <= self.signature_bound:
enc_s = compress(s[1], self.sig_bytelen - HEAD_LEN - SALT_LEN)
# Check that the encoding is valid (sometimes it fails)
if (enc_s is not False):
return header + salt + enc_s
'''
else:
print("-------------INVALID encoding---------------")
else:
print("-------------NOT within signature bound---------------")
'''
def verify(self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True
| verify | identifier_name |
falcon.py | """
Python implementation of Falcon:
https://falcon-sign.info/.
"""
from .common import q
from numpy import set_printoptions
from math import sqrt
from .fft import fft, ifft, sub, neg, add_fft, mul_fft
from .ntt import sub_zq, mul_zq, div_zq
from .ffsampling import gram, ffldl_fft, ffsampling_fft
from .ntrugen import ntru_gen
from .encoding import compress, decompress
# https://pycryptodome.readthedocs.io/en/latest/src/hash/shake256.html
from Crypto.Hash import SHAKE256
# Randomness
from os import urandom
from .rng import ChaCha20
# For debugging purposes
import sys
if sys.version_info >= (3, 4):
from importlib import reload # Python 3.4+ only.
set_printoptions(linewidth=200, precision=5, suppress=True)
logn = {
2: 1,
4: 2,
8: 3,
16: 4,
32: 5,
64: 6,
128: 7,
256: 8,
512: 9,
1024: 10
}
# Bytelength of the signing salt and header
HEAD_LEN = 1
SALT_LEN = 40
SEED_LEN = 56
# Parameter sets for Falcon:
# - n is the dimension/degree of the cyclotomic ring
# - sigma is the std. dev. of signatures (Gaussians over a lattice)
# - sigmin is a lower bounds on the std. dev. of each Gaussian over Z
# - sigbound is the upper bound on ||s0||^2 + ||s1||^2
# - sig_bytelen is the bytelength of signatures
Params = {
# FalconParam(2, 2)
2: {
"n": 2,
"sigma": 144.81253976308423,
"sigmin": 1.1165085072329104,
"sig_bound": 101498,
"sig_bytelen": 44,
},
# FalconParam(4, 2)
4: {
"n": 4,
"sigma": 146.83798833523608,
"sigmin": 1.1321247692325274,
"sig_bound": 208714,
"sig_bytelen": 47,
},
# FalconParam(8, 2)
8: {
"n": 8,
"sigma": 148.83587593064718,
"sigmin": 1.147528535373367,
"sig_bound": 428865,
"sig_bytelen": 52,
},
# FalconParam(16, 4)
16: {
"n": 16,
"sigma": 151.78340713845503,
"sigmin": 1.170254078853483,
"sig_bound": 892039,
"sig_bytelen": 63,
},
# FalconParam(32, 8)
32: {
"n": 32,
"sigma": 154.6747794602761,
"sigmin": 1.1925466358390344,
"sig_bound": 1852696,
"sig_bytelen": 82,
},
# FalconParam(64, 16)
64: {
"n": 64,
"sigma": 157.51308555044122,
"sigmin": 1.2144300507766141,
"sig_bound": 3842630,
"sig_bytelen": 122,
},
# FalconParam(128, 32)
128: {
"n": 128,
"sigma": 160.30114421975344,
"sigmin": 1.235926056771981,
"sig_bound": 7959734,
"sig_bytelen": 200,
},
# FalconParam(256, 64)
256: {
"n": 256,
"sigma": 163.04153322607107,
"sigmin": 1.2570545284063217,
"sig_bound": 16468416,
"sig_bytelen": 356,
},
# FalconParam(512, 128)
512: {
"n": 512,
"sigma": 165.7366171829776,
"sigmin": 1.2778336969128337,
"sig_bound": 34034726,
"sig_bytelen": 666,
},
# FalconParam(1024, 256)
1024: {
"n": 1024,
"sigma": 168.38857144654395,
"sigmin": 1.298280334344292,
"sig_bound": 70265242,
"sig_bytelen": 1280,
},
}
def print_tree(tree, pref=""):
"""
Display a LDL tree in a readable form.
Args:
T: a LDL tree
Format: coefficient or fft
"""
leaf = "|_____> "
top = "|_______"
son1 = "| "
son2 = " "
width = len(top)
a = ""
if len(tree) == 3:
if (pref == ""):
a += pref + str(tree[0]) + "\n"
else:
a += pref[:-width] + top + str(tree[0]) + "\n"
a += print_tree(tree[1], pref + son1)
a += print_tree(tree[2], pref + son2)
return a
else:
return (pref[:-width] + leaf + str(tree) + "\n")
def normalize_tree(tree, sigma):
"""
Normalize leaves of a LDL tree (from values ||b_i||**2 to sigma/||b_i||).
Args:
T: a LDL tree
sigma: a standard deviation
Format: coefficient or fft
"""
if len(tree) == 3:
normalize_tree(tree[1], sigma)
normalize_tree(tree[2], sigma)
else:
tree[0] = sigma / sqrt(tree[0].real)
tree[1] = 0
class PublicKey:
|
class SecretKey:
"""
This class contains methods for performing
secret key operations (and also public key operations) in Falcon.
One can:
- initialize a secret key for:
- n = 128, 256, 512, 1024,
- phi = x ** n + 1,
- q = 12 * 1024 + 1
- find a preimage t of a point c (both in ( Z[x] mod (Phi,q) )**2 ) such that t*B0 = c
- hash a message to a point of Z[x] mod (Phi,q)
- sign a message
- verify the signature of a message
"""
def __init__(self, n, polys=None):
"""Initialize a secret key."""
# Public parameters
self.n = n
self.sigma = Params[n]["sigma"]
self.sigmin = Params[n]["sigmin"]
self.signature_bound = Params[n]["sig_bound"]
self.sig_bytelen = Params[n]["sig_bytelen"]
# Compute NTRU polynomials f, g, F, G verifying fG - gF = q mod Phi
if polys is None:
self.f, self.g, self.F, self.G = ntru_gen(n)
else:
[f, g, F, G] = polys
assert all((len(poly) == n) for poly in [f, g, F, G])
self.f = f[:]
self.g = g[:]
self.F = F[:]
self.G = G[:]
# From f, g, F, G, compute the basis B0 of a NTRU lattice
# as well as its Gram matrix and their fft's.
B0 = [[self.g, neg(self.f)], [self.G, neg(self.F)]]
G0 = gram(B0)
self.B0_fft = [[fft(elt) for elt in row] for row in B0]
G0_fft = [[fft(elt) for elt in row] for row in G0]
self.T_fft = ffldl_fft(G0_fft)
# Normalize Falcon tree
normalize_tree(self.T_fft, self.sigma)
# The public key is a polynomial such that h*f = g mod (Phi,q)
self.h = div_zq(self.g, self.f)
def __repr__(self, verbose=False):
"""Print the object in readable form."""
rep = "Private key for n = {n}:\n\n".format(n=self.n)
rep += "f = {f}\n\n".format(f=self.f)
rep += "g = {g}\n\n".format(g=self.g)
rep += "F = {F}\n\n".format(F=self.F)
rep += "G = {G}\n\n".format(G=self.G)
if verbose:
rep += "\nFFT tree\n"
rep += print_tree(self.T_fft, pref="")
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def sample_preimage(self, point, seed=None):
"""
Sample a short vector s such that s[0] + s[1] * h = point.
"""
[[a, b], [c, d]] = self.B0_fft
# We compute a vector t_fft such that:
# (fft(point), fft(0)) * B0_fft = t_fft
# Because fft(0) = 0 and the inverse of B has a very specific form,
# we can do several optimizations.
'''
print("---------Inside sample_preimage----------")
print("point: ", point)
'''
point_fft = fft(point)
t0_fft = [(point_fft[i] * d[i]) / q for i in range(self.n)]
t1_fft = [(-point_fft[i] * b[i]) / q for i in range(self.n)]
t_fft = [t0_fft, t1_fft]
# We now compute v such that:
# v = z * B0 for an integral vector z
# v is close to (point, 0)
if seed is None:
# If no seed is defined, use urandom as the pseudo-random source.
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin, urandom)
else:
# If a seed is defined, initialize a ChaCha20 PRG
# that is used to generate pseudo-randomness.
chacha_prng = ChaCha20(seed)
z_fft = ffsampling_fft(t_fft, self.T_fft, self.sigmin,
chacha_prng.randombytes)
v0_fft = add_fft(mul_fft(z_fft[0], a), mul_fft(z_fft[1], c))
v1_fft = add_fft(mul_fft(z_fft[0], b), mul_fft(z_fft[1], d))
v0 = [int(round(elt)) for elt in ifft(v0_fft)]
v1 = [int(round(elt)) for elt in ifft(v1_fft)]
# The difference s = (point, 0) - v is such that:
# s is short
# s[0] + s[1] * h = point
s = [sub(point, v0), neg(v1)]
return s
def sign(self, message, randombytes=urandom):
"""
Sign a message. The message MUST be a byte string or byte array.
Optionally, one can select the source of (pseudo-)randomness used
(default: urandom).
"""
int_header = 0x30 + logn[self.n]
header = int_header.to_bytes(1, "little")
salt = randombytes(SALT_LEN)
hashed = self.hash_to_point(message, salt)
# We repeat the signing procedure until we find a signature that is
# short enough (both the Euclidean norm and the bytelength)
'''
print("---------Inside sign----------")
'''
while(1):
if (randombytes == urandom):
s = self.sample_preimage(hashed)
'''
print("s: ", s)
'''
else:
seed = randombytes(SEED_LEN)
s = self.sample_preimage(hashed, seed=seed)
norm_sign = sum(coef ** 2 for coef in s[0])
norm_sign += sum(coef ** 2 for coef in s[1])
# Check the Euclidean norm
if norm_sign <= self.signature_bound:
enc_s = compress(s[1], self.sig_bytelen - HEAD_LEN - SALT_LEN)
# Check that the encoding is valid (sometimes it fails)
if (enc_s is not False):
return header + salt + enc_s
'''
else:
print("-------------INVALID encoding---------------")
else:
print("-------------NOT within signature bound---------------")
'''
def verify(self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True
| """
This class contains methods for performing public key operations in Falcon.
"""
def __init__(self, sk=None, n=None, h=None):
"""Initialize a public key."""
if sk:
self.n = sk.n
self.h = sk.h
elif n and h:
self.n = n
self.h = h
else:
raise Exception("Public Key construction failed: insufficient/wrong arguments")
self.signature_bound = Params[self.n]["sig_bound"]
self.sig_bytelen = Params[self.n]["sig_bytelen"]
def __repr__(self):
"""Print the object in readable form."""
rep = "Public for n = {n}:\n\n".format(n=self.n)
rep += "h = {h}\n".format(h=self.h)
return rep
def hash_to_point(self, message, salt):
"""
Hash a message to a point in Z[x] mod(Phi, q).
Inspired by the Parse function from NewHope.
"""
n = self.n
if q > (1 << 16):
raise ValueError("The modulus is too large")
k = (1 << 16) // q
# Create a SHAKE object and hash the salt and message.
shake = SHAKE256.new()
shake.update(salt)
shake.update(message)
# Output pseudorandom bytes and map them to coefficients.
hashed = [0 for i in range(n)]
i = 0
j = 0
while i < n:
# Takes 2 bytes, transform them in a 16 bits integer
twobytes = shake.read(2)
elt = (twobytes[0] << 8) + twobytes[1] # This breaks in Python 2.x
# Implicit rejection sampling
if elt < k * q:
hashed[i] = elt % q
i += 1
j += 1
return hashed
def verify(self, message, signature):
"""
Verify a signature.
"""
# Unpack the salt and the short polynomial s1
salt = signature[HEAD_LEN:HEAD_LEN + SALT_LEN]
enc_s = signature[HEAD_LEN + SALT_LEN:]
s1 = decompress(enc_s, self.sig_bytelen - HEAD_LEN - SALT_LEN, self.n)
# Check that the encoding is valid
if (s1 is False):
print("Invalid encoding")
return False
# Compute s0 and normalize its coefficients in (-q/2, q/2]
hashed = self.hash_to_point(message, salt)
s0 = sub_zq(hashed, mul_zq(s1, self.h))
s0 = [(coef + (q >> 1)) % q - (q >> 1) for coef in s0]
# Check that the (s0, s1) is short
norm_sign = sum(coef ** 2 for coef in s0)
norm_sign += sum(coef ** 2 for coef in s1)
if norm_sign > self.signature_bound:
print("Squared norm of signature is too large:", norm_sign)
return False
# If all checks are passed, accept
return True | identifier_body |
cluster_feeder.go | /*
Copyright 2018 The Kubernetes Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package input
import (
"context"
"fmt"
"time"
apiv1 "k8s.io/api/core/v1"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/fields"
"k8s.io/apimachinery/pkg/labels"
"k8s.io/apimachinery/pkg/util/wait"
"k8s.io/apimachinery/pkg/watch"
kube_client "k8s.io/client-go/kubernetes"
corev1 "k8s.io/client-go/kubernetes/typed/core/v1"
v1lister "k8s.io/client-go/listers/core/v1"
"k8s.io/client-go/tools/cache"
"k8s.io/klog/v2"
vpa_types "k8s.io/autoscaler/vertical-pod-autoscaler/pkg/apis/autoscaling.k8s.io/v1"
vpa_api "k8s.io/autoscaler/vertical-pod-autoscaler/pkg/client/clientset/versioned/typed/autoscaling.k8s.io/v1"
vpa_lister "k8s.io/autoscaler/vertical-pod-autoscaler/pkg/client/listers/autoscaling.k8s.io/v1"
controllerfetcher "k8s.io/autoscaler/vertical-pod-autoscaler/pkg/recommender/input/controller_fetcher"
"k8s.io/autoscaler/vertical-pod-autoscaler/pkg/recommender/input/history"
"k8s.io/autoscaler/vertical-pod-autoscaler/pkg/recommender/input/metrics"
"k8s.io/autoscaler/vertical-pod-autoscaler/pkg/recommender/input/oom"
"k8s.io/autoscaler/vertical-pod-autoscaler/pkg/recommender/input/spec"
"k8s.io/autoscaler/vertical-pod-autoscaler/pkg/recommender/model"
"k8s.io/autoscaler/vertical-pod-autoscaler/pkg/target"
metrics_recommender "k8s.io/autoscaler/vertical-pod-autoscaler/pkg/utils/metrics/recommender"
)
const (
evictionWatchRetryWait = 10 * time.Second
evictionWatchJitterFactor = 0.5
// DefaultRecommenderName recommender name explicitly (and so implicitly specify that the default recommender should handle them)
DefaultRecommenderName = "default"
)
// ClusterStateFeeder can update state of ClusterState object.
type ClusterStateFeeder interface {
// InitFromHistoryProvider loads historical pod spec into clusterState.
InitFromHistoryProvider(historyProvider history.HistoryProvider)
// InitFromCheckpoints loads historical checkpoints into clusterState.
InitFromCheckpoints()
// LoadVPAs updates clusterState with current state of VPAs.
LoadVPAs()
// LoadPods updates clusterState with current specification of Pods and their Containers.
LoadPods()
// LoadRealTimeMetrics updates clusterState with current usage metrics of containers.
LoadRealTimeMetrics()
// GarbageCollectCheckpoints removes historical checkpoints that don't have a matching VPA.
GarbageCollectCheckpoints()
}
// ClusterStateFeederFactory makes instances of ClusterStateFeeder.
type ClusterStateFeederFactory struct {
ClusterState *model.ClusterState
KubeClient kube_client.Interface
MetricsClient metrics.MetricsClient
VpaCheckpointClient vpa_api.VerticalPodAutoscalerCheckpointsGetter
VpaLister vpa_lister.VerticalPodAutoscalerLister
PodLister v1lister.PodLister
OOMObserver oom.Observer
SelectorFetcher target.VpaTargetSelectorFetcher
MemorySaveMode bool
ControllerFetcher controllerfetcher.ControllerFetcher
RecommenderName string
}
// Make creates new ClusterStateFeeder with internal data providers, based on kube client.
func (m ClusterStateFeederFactory) Make() *clusterStateFeeder {
return &clusterStateFeeder{
coreClient: m.KubeClient.CoreV1(),
metricsClient: m.MetricsClient,
oomChan: m.OOMObserver.GetObservedOomsChannel(),
vpaCheckpointClient: m.VpaCheckpointClient,
vpaLister: m.VpaLister,
clusterState: m.ClusterState,
specClient: spec.NewSpecClient(m.PodLister),
selectorFetcher: m.SelectorFetcher,
memorySaveMode: m.MemorySaveMode,
controllerFetcher: m.ControllerFetcher,
recommenderName: m.RecommenderName,
}
}
// WatchEvictionEventsWithRetries watches new Events with reason=Evicted and passes them to the observer.
func WatchEvictionEventsWithRetries(kubeClient kube_client.Interface, observer oom.Observer, namespace string) {
go func() {
options := metav1.ListOptions{
FieldSelector: "reason=Evicted",
}
watchEvictionEventsOnce := func() {
watchInterface, err := kubeClient.CoreV1().Events(namespace).Watch(context.TODO(), options)
if err != nil {
klog.Errorf("Cannot initialize watching events. Reason %v", err)
return
}
watchEvictionEvents(watchInterface.ResultChan(), observer)
}
for {
watchEvictionEventsOnce()
// Wait between attempts, retrying too often breaks API server.
waitTime := wait.Jitter(evictionWatchRetryWait, evictionWatchJitterFactor)
klog.V(1).Infof("An attempt to watch eviction events finished. Waiting %v before the next one.", waitTime)
time.Sleep(waitTime)
}
}()
}
func watchEvictionEvents(evictedEventChan <-chan watch.Event, observer oom.Observer) {
for {
evictedEvent, ok := <-evictedEventChan
if !ok {
klog.V(3).Infof("Eviction event chan closed")
return
}
if evictedEvent.Type == watch.Added {
evictedEvent, ok := evictedEvent.Object.(*apiv1.Event)
if !ok {
continue
}
observer.OnEvent(evictedEvent)
}
}
}
// Creates clients watching pods: PodLister (listing only not terminated pods).
func newPodClients(kubeClient kube_client.Interface, resourceEventHandler cache.ResourceEventHandler, namespace string) v1lister.PodLister {
// We are interested in pods which are Running or Unknown (in case the pod is
// running but there are some transient errors we don't want to delete it from
// our model).
// We don't want to watch Pending pods because they didn't generate any usage
// yet.
// Succeeded and Failed failed pods don't generate any usage anymore but we
// don't necessarily want to immediately delete them.
selector := fields.ParseSelectorOrDie("status.phase!=" + string(apiv1.PodPending))
podListWatch := cache.NewListWatchFromClient(kubeClient.CoreV1().RESTClient(), "pods", namespace, selector)
indexer, controller := cache.NewIndexerInformer(
podListWatch,
&apiv1.Pod{},
time.Hour,
resourceEventHandler,
cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc},
)
podLister := v1lister.NewPodLister(indexer)
stopCh := make(chan struct{})
go controller.Run(stopCh)
return podLister
}
// NewPodListerAndOOMObserver creates pair of pod lister and OOM observer.
func NewPodListerAndOOMObserver(kubeClient kube_client.Interface, namespace string) (v1lister.PodLister, oom.Observer) {
oomObserver := oom.NewObserver()
podLister := newPodClients(kubeClient, oomObserver, namespace)
WatchEvictionEventsWithRetries(kubeClient, oomObserver, namespace)
return podLister, oomObserver
}
type clusterStateFeeder struct {
coreClient corev1.CoreV1Interface
specClient spec.SpecClient
metricsClient metrics.MetricsClient
oomChan <-chan oom.OomInfo
vpaCheckpointClient vpa_api.VerticalPodAutoscalerCheckpointsGetter
vpaLister vpa_lister.VerticalPodAutoscalerLister
clusterState *model.ClusterState
selectorFetcher target.VpaTargetSelectorFetcher
memorySaveMode bool
controllerFetcher controllerfetcher.ControllerFetcher
recommenderName string
}
func (feeder *clusterStateFeeder) InitFromHistoryProvider(historyProvider history.HistoryProvider) {
klog.V(3).Info("Initializing VPA from history provider")
clusterHistory, err := historyProvider.GetClusterHistory()
if err != nil {
klog.Errorf("Cannot get cluster history: %v", err)
}
for podID, podHistory := range clusterHistory {
klog.V(4).Infof("Adding pod %v with labels %v", podID, podHistory.LastLabels)
feeder.clusterState.AddOrUpdatePod(podID, podHistory.LastLabels, apiv1.PodUnknown)
for containerName, sampleList := range podHistory.Samples {
containerID := model.ContainerID{
PodID: podID,
ContainerName: containerName,
}
if err = feeder.clusterState.AddOrUpdateContainer(containerID, nil); err != nil {
klog.Warningf("Failed to add container %+v. Reason: %+v", containerID, err)
}
klog.V(4).Infof("Adding %d samples for container %v", len(sampleList), containerID)
for _, sample := range sampleList {
if err := feeder.clusterState.AddSample(
&model.ContainerUsageSampleWithKey{
ContainerUsageSample: sample,
Container: containerID,
}); err != nil {
klog.Warningf("Error adding metric sample for container %v: %v", containerID, err)
}
}
}
}
}
func (feeder *clusterStateFeeder) setVpaCheckpoint(checkpoint *vpa_types.VerticalPodAutoscalerCheckpoint) error {
vpaID := model.VpaID{Namespace: checkpoint.Namespace, VpaName: checkpoint.Spec.VPAObjectName}
vpa, exists := feeder.clusterState.Vpas[vpaID]
if !exists {
return fmt.Errorf("cannot load checkpoint to missing VPA object %+v", vpaID)
}
cs := model.NewAggregateContainerState()
err := cs.LoadFromCheckpoint(&checkpoint.Status)
if err != nil {
return fmt.Errorf("cannot load checkpoint for VPA %+v. Reason: %v", vpa.ID, err)
}
vpa.ContainersInitialAggregateState[checkpoint.Spec.ContainerName] = cs
return nil
}
func (feeder *clusterStateFeeder) InitFromCheckpoints() {
klog.V(3).Info("Initializing VPA from checkpoints")
feeder.LoadVPAs()
namespaces := make(map[string]bool)
for _, v := range feeder.clusterState.Vpas {
namespaces[v.ID.Namespace] = true
}
for namespace := range namespaces {
klog.V(3).Infof("Fetching checkpoints from namespace %s", namespace)
checkpointList, err := feeder.vpaCheckpointClient.VerticalPodAutoscalerCheckpoints(namespace).List(context.TODO(), metav1.ListOptions{})
if err != nil |
for _, checkpoint := range checkpointList.Items {
klog.V(3).Infof("Loading VPA %s/%s checkpoint for %s", checkpoint.ObjectMeta.Namespace, checkpoint.Spec.VPAObjectName, checkpoint.Spec.ContainerName)
err = feeder.setVpaCheckpoint(&checkpoint)
if err != nil {
klog.Errorf("Error while loading checkpoint. Reason: %+v", err)
}
}
}
}
func (feeder *clusterStateFeeder) GarbageCollectCheckpoints() {
klog.V(3).Info("Starting garbage collection of checkpoints")
feeder.LoadVPAs()
namespaceList, err := feeder.coreClient.Namespaces().List(context.TODO(), metav1.ListOptions{})
if err != nil {
klog.Errorf("Cannot list namespaces. Reason: %+v", err)
return
}
for _, namespaceItem := range namespaceList.Items {
namespace := namespaceItem.Name
checkpointList, err := feeder.vpaCheckpointClient.VerticalPodAutoscalerCheckpoints(namespace).List(context.TODO(), metav1.ListOptions{})
if err != nil {
klog.Errorf("Cannot list VPA checkpoints from namespace %v. Reason: %+v", namespace, err)
}
for _, checkpoint := range checkpointList.Items {
vpaID := model.VpaID{Namespace: checkpoint.Namespace, VpaName: checkpoint.Spec.VPAObjectName}
_, exists := feeder.clusterState.Vpas[vpaID]
if !exists {
err = feeder.vpaCheckpointClient.VerticalPodAutoscalerCheckpoints(namespace).Delete(context.TODO(), checkpoint.Name, metav1.DeleteOptions{})
if err == nil {
klog.V(3).Infof("Orphaned VPA checkpoint cleanup - deleting %v/%v.", namespace, checkpoint.Name)
} else {
klog.Errorf("Cannot delete VPA checkpoint %v/%v. Reason: %+v", namespace, checkpoint.Name, err)
}
}
}
}
}
func implicitDefaultRecommender(selectors []*vpa_types.VerticalPodAutoscalerRecommenderSelector) bool {
return len(selectors) == 0
}
func selectsRecommender(selectors []*vpa_types.VerticalPodAutoscalerRecommenderSelector, name *string) bool {
for _, s := range selectors {
if s.Name == *name {
return true
}
}
return false
}
// Filter VPA objects whose specified recommender names are not default
func filterVPAs(feeder *clusterStateFeeder, allVpaCRDs []*vpa_types.VerticalPodAutoscaler) []*vpa_types.VerticalPodAutoscaler {
klog.V(3).Infof("Start selecting the vpaCRDs.")
var vpaCRDs []*vpa_types.VerticalPodAutoscaler
for _, vpaCRD := range allVpaCRDs {
if feeder.recommenderName == DefaultRecommenderName {
if !implicitDefaultRecommender(vpaCRD.Spec.Recommenders) && !selectsRecommender(vpaCRD.Spec.Recommenders, &feeder.recommenderName) {
klog.V(6).Infof("Ignoring vpaCRD %s in namespace %s as current recommender's name %v doesn't appear among its recommenders", vpaCRD.Name, vpaCRD.Namespace, feeder.recommenderName)
continue
}
} else {
if implicitDefaultRecommender(vpaCRD.Spec.Recommenders) {
klog.V(6).Infof("Ignoring vpaCRD %s in namespace %s as %v recommender doesn't process CRDs implicitly destined to %v recommender", vpaCRD.Name, vpaCRD.Namespace, feeder.recommenderName, DefaultRecommenderName)
continue
}
if !selectsRecommender(vpaCRD.Spec.Recommenders, &feeder.recommenderName) {
klog.V(6).Infof("Ignoring vpaCRD %s in namespace %s as current recommender's name %v doesn't appear among its recommenders", vpaCRD.Name, vpaCRD.Namespace, feeder.recommenderName)
continue
}
}
vpaCRDs = append(vpaCRDs, vpaCRD)
}
return vpaCRDs
}
// LoadVPAs fetches VPA objects and loads them into the cluster state.
func (feeder *clusterStateFeeder) LoadVPAs() {
// List VPA API objects.
allVpaCRDs, err := feeder.vpaLister.List(labels.Everything())
if err != nil {
klog.Errorf("Cannot list VPAs. Reason: %+v", err)
return
}
// Filter out VPAs that specified recommenders with names not equal to "default"
vpaCRDs := filterVPAs(feeder, allVpaCRDs)
klog.V(3).Infof("Fetched %d VPAs.", len(vpaCRDs))
// Add or update existing VPAs in the model.
vpaKeys := make(map[model.VpaID]bool)
for _, vpaCRD := range vpaCRDs {
vpaID := model.VpaID{
Namespace: vpaCRD.Namespace,
VpaName: vpaCRD.Name,
}
selector, conditions := feeder.getSelector(vpaCRD)
klog.V(4).Infof("Using selector %s for VPA %s/%s", selector.String(), vpaCRD.Namespace, vpaCRD.Name)
if feeder.clusterState.AddOrUpdateVpa(vpaCRD, selector) == nil {
// Successfully added VPA to the model.
vpaKeys[vpaID] = true
for _, condition := range conditions {
if condition.delete {
delete(feeder.clusterState.Vpas[vpaID].Conditions, condition.conditionType)
} else {
feeder.clusterState.Vpas[vpaID].Conditions.Set(condition.conditionType, true, "", condition.message)
}
}
}
}
// Delete non-existent VPAs from the model.
for vpaID := range feeder.clusterState.Vpas {
if _, exists := vpaKeys[vpaID]; !exists {
klog.V(3).Infof("Deleting VPA %v", vpaID)
if err := feeder.clusterState.DeleteVpa(vpaID); err != nil {
klog.Errorf("Deleting VPA %v failed: %v", vpaID, err)
}
}
}
feeder.clusterState.ObservedVpas = vpaCRDs
}
// LoadPods loads pod into the cluster state.
func (feeder *clusterStateFeeder) LoadPods() {
podSpecs, err := feeder.specClient.GetPodSpecs()
if err != nil {
klog.Errorf("Cannot get SimplePodSpecs. Reason: %+v", err)
}
pods := make(map[model.PodID]*spec.BasicPodSpec)
for _, spec := range podSpecs {
pods[spec.ID] = spec
}
for key := range feeder.clusterState.Pods {
if _, exists := pods[key]; !exists {
klog.V(3).Infof("Deleting Pod %v", key)
feeder.clusterState.DeletePod(key)
}
}
for _, pod := range pods {
if feeder.memorySaveMode && !feeder.matchesVPA(pod) {
continue
}
feeder.clusterState.AddOrUpdatePod(pod.ID, pod.PodLabels, pod.Phase)
for _, container := range pod.Containers {
if err = feeder.clusterState.AddOrUpdateContainer(container.ID, container.Request); err != nil {
klog.Warningf("Failed to add container %+v. Reason: %+v", container.ID, err)
}
}
}
}
func (feeder *clusterStateFeeder) LoadRealTimeMetrics() {
containersMetrics, err := feeder.metricsClient.GetContainersMetrics()
if err != nil {
klog.Errorf("Cannot get ContainerMetricsSnapshot from MetricsClient. Reason: %+v", err)
}
sampleCount := 0
droppedSampleCount := 0
for _, containerMetrics := range containersMetrics {
for _, sample := range newContainerUsageSamplesWithKey(containerMetrics) {
if err := feeder.clusterState.AddSample(sample); err != nil {
// Not all pod states are tracked in memory saver mode
if _, isKeyError := err.(model.KeyError); isKeyError && feeder.memorySaveMode {
continue
}
klog.Warningf("Error adding metric sample for container %v: %v", sample.Container, err)
droppedSampleCount++
} else {
sampleCount++
}
}
}
klog.V(3).Infof("ClusterSpec fed with #%v ContainerUsageSamples for #%v containers. Dropped #%v samples.", sampleCount, len(containersMetrics), droppedSampleCount)
Loop:
for {
select {
case oomInfo := <-feeder.oomChan:
klog.V(3).Infof("OOM detected %+v", oomInfo)
if err = feeder.clusterState.RecordOOM(oomInfo.ContainerID, oomInfo.Timestamp, oomInfo.Memory); err != nil {
klog.Warningf("Failed to record OOM %+v. Reason: %+v", oomInfo, err)
}
default:
break Loop
}
}
metrics_recommender.RecordAggregateContainerStatesCount(feeder.clusterState.StateMapSize())
}
func (feeder *clusterStateFeeder) matchesVPA(pod *spec.BasicPodSpec) bool {
for vpaKey, vpa := range feeder.clusterState.Vpas {
podLabels := labels.Set(pod.PodLabels)
if vpaKey.Namespace == pod.ID.Namespace && vpa.PodSelector.Matches(podLabels) {
return true
}
}
return false
}
func newContainerUsageSamplesWithKey(metrics *metrics.ContainerMetricsSnapshot) []*model.ContainerUsageSampleWithKey {
var samples []*model.ContainerUsageSampleWithKey
for metricName, resourceAmount := range metrics.Usage {
sample := &model.ContainerUsageSampleWithKey{
Container: metrics.ID,
ContainerUsageSample: model.ContainerUsageSample{
MeasureStart: metrics.SnapshotTime,
Resource: metricName,
Usage: resourceAmount,
},
}
samples = append(samples, sample)
}
return samples
}
type condition struct {
conditionType vpa_types.VerticalPodAutoscalerConditionType
delete bool
message string
}
func (feeder *clusterStateFeeder) validateTargetRef(vpa *vpa_types.VerticalPodAutoscaler) (bool, condition) {
//
if vpa.Spec.TargetRef == nil {
return false, condition{}
}
k := controllerfetcher.ControllerKeyWithAPIVersion{
ControllerKey: controllerfetcher.ControllerKey{
Namespace: vpa.Namespace,
Kind: vpa.Spec.TargetRef.Kind,
Name: vpa.Spec.TargetRef.Name,
},
ApiVersion: vpa.Spec.TargetRef.APIVersion,
}
top, err := feeder.controllerFetcher.FindTopMostWellKnownOrScalable(&k)
if err != nil {
return false, condition{conditionType: vpa_types.ConfigUnsupported, delete: false, message: fmt.Sprintf("Error checking if target is a topmost well-known or scalable controller: %s", err)}
}
if top == nil {
return false, condition{conditionType: vpa_types.ConfigUnsupported, delete: false, message: fmt.Sprintf("Unknown error during checking if target is a topmost well-known or scalable controller: %s", err)}
}
if *top != k {
return false, condition{conditionType: vpa_types.ConfigUnsupported, delete: false, message: "The targetRef controller has a parent but it should point to a topmost well-known or scalable controller"}
}
return true, condition{}
}
func (feeder *clusterStateFeeder) getSelector(vpa *vpa_types.VerticalPodAutoscaler) (labels.Selector, []condition) {
selector, fetchErr := feeder.selectorFetcher.Fetch(vpa)
if selector != nil {
validTargetRef, unsupportedCondition := feeder.validateTargetRef(vpa)
if !validTargetRef {
return labels.Nothing(), []condition{
unsupportedCondition,
{conditionType: vpa_types.ConfigDeprecated, delete: true},
}
}
return selector, []condition{
{conditionType: vpa_types.ConfigUnsupported, delete: true},
{conditionType: vpa_types.ConfigDeprecated, delete: true},
}
}
msg := "Cannot read targetRef"
if fetchErr != nil {
klog.Errorf("Cannot get target selector from VPA's targetRef. Reason: %+v", fetchErr)
msg = fmt.Sprintf("Cannot read targetRef. Reason: %s", fetchErr.Error())
}
return labels.Nothing(), []condition{
{conditionType: vpa_types.ConfigUnsupported, delete: false, message: msg},
{conditionType: vpa_types.ConfigDeprecated, delete: true},
}
}
| {
klog.Errorf("Cannot list VPA checkpoints from namespace %v. Reason: %+v", namespace, err)
} | conditional_block |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.