code stringlengths 3 6.57k |
|---|
override_native_generic_func("argwhere_strategy") |
argwhere_strategy(attrs, inputs, out_type, target) |
_op.OpStrategy() |
wrap_compute_argwhere(topi.argwhere) |
wrap_topi_schedule(topi.generic.schedule_argwhere) |
logging.getLogger() |
logging.StreamHandler() |
handler.setFormatter(formatter) |
logger.addHandler(handler) |
logger.setLevel(logging.INFO) |
flip_coords(xy_list) |
for (x, y) |
make_ids(n, prefix='row_') |
int(math.log10(n) |
format(num=i, pad=k) |
range(n) |
to_df(distance_matrix_response, origin_ids=None, destination_ids=None) |
a (decoded) |
pd.DataFrame([], columns=columns) |
product(r['origin_addresses'], r['destination_addresses']) |
make_ids(len(r['origin_addresses']) |
make_ids(len(r['destination_addresses']) |
product(origin_ids, destination_ids) |
durs.append(e[dur_key]['value']) |
dists.append(e['distance']['value']) |
durs.append(np.nan) |
dists.append(np.nan) |
point_df_to_gdf(f, x_col='lon', y_col='lat', from_crs=WGS84) |
f.copy() |
apply(lambda p: sg.Point(p) |
f.drop([x_col, y_col], axis=1) |
gpd.GeoDataFrame(f) |
point_gdf_to_df(f, x_col='lon', y_col='lat', to_crs=WGS84) |
f.copy() |
ValueError('GeoDataFrame needs a crs attribute') |
f.to_crs(to_crs) |
zip(*f['geometry'].map(lambda p: p.coords[0]) |
pd.DataFrame(f) |
origins_gdf.copy() |
destinations_gdf.copy() |
o_gdf.to_crs(WGS84) |
make_ids(o_gdf.shape[0]) |
d_gdf.to_crs(WGS84) |
make_ids(d_gdf.shape[0]) |
client.distance_matrix(flip_coords(o_locs) |
flip_coords(d_locs) |
to_df(r, o_ids, d_ids) |
except (googlemaps.exceptions.HTTPError, googlemaps.exceptions.Timeout) |
origins_gdf.copy() |
destinations_gdf.copy() |
make_ids(n_o, 'orig_row_') |
make_ids(n_d, 'dest_row_') |
distance_matrix_kwargs.get('mode', 'driving') |
Path(out_dir) |
out_dir.exists() |
out_dir.mkdir(parents=True) |
itertuples() |
logger.info('Working on origin {} of {} (id {}) |
attempt (e.g. through API usage over limit) |
range(math.ceil(n_d/max_elements) |
min(max_elements*(j + 1) |
Path(out_dir) |
f.to_csv(path, index=False) |
logger.info('* Failed to get data for ' + path.stem) |
OrderedDict() |
format(daily_limit) |
max(0, n - num_freebies) |
pd.Series(d) |
mocked_rpc_server() |
MockedRpcServer(object) |
Queue() |
__init__(self, host, port) |
mocked_send(cls, message) |
cls.queue.put(message.serialize() |
sleep(0) |
recv(self) |
self.queue.get() |
Message.unserialize(results) |
send(self, message) |
self.outbox.append(message.serialize() |
TestMasterRunner(LocustTestCase) |
setUp(self) |
global_stats.reset_all() |
parse_options() |
parser.parse_args(args) |
tearDown(self) |
test_slave_connect(self) |
MyTestLocust(Locust) |
mock.patch("locust.rpc.rpc.Server", mocked_rpc_server() |
MasterLocustRunner(MyTestLocust, self.options) |
server.mocked_send(Message("client_ready", None, "zeh_fake_client1") |
self.assertEqual(1, len(master.clients) |
self.assertTrue("zeh_fake_client1" in master.clients, "Could not find fake client in master instance's clients dict") |
server.mocked_send(Message("client_ready", None, "zeh_fake_client2") |
server.mocked_send(Message("client_ready", None, "zeh_fake_client3") |
server.mocked_send(Message("client_ready", None, "zeh_fake_client4") |
self.assertEqual(4, len(master.clients) |
server.mocked_send(Message("quit", None, "zeh_fake_client3") |
self.assertEqual(3, len(master.clients) |
test_slave_stats_report_median(self) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.