file_name large_stringlengths 4 140 | prefix large_stringlengths 0 12.1k | suffix large_stringlengths 0 12k | middle large_stringlengths 0 7.51k | fim_type large_stringclasses 4
values |
|---|---|---|---|---|
main.rs | r: init::Resources) -> init::LateResources {
let mut rcc = p.device.RCC.constrain();
let mut flash = p.device.FLASH.constrain();
let mut gpioa = p.device.GPIOA.split(&mut rcc.ahb);
let mut gpiob = p.device.GPIOB.split(&mut rcc.ahb);
let mut gpioc = p.device.GPIOC.split(&mut rcc.ahb);
// Enable the syscfg
rcc.apb2.enr().modify(|_, w| w.syscfgen().enabled());
rcc.apb2.rstr().modify(|_, w| w.syscfgrst().set_bit());
rcc.apb2.rstr().modify(|_, w| w.syscfgrst().clear_bit());
// Enable systick
p.core.SYST.set_clock_source(SystClkSource::Core);
p.core.SYST.set_reload(16_000_000);
p.core.SYST.enable_interrupt();
p.core.SYST.enable_counter();
// Set up our clocks & timer & delay
let clocks = rcc.cfgr.freeze(&mut flash.acr);
let mut timer = Timer::tim7(p.device.TIM7, 1.hz(), clocks, &mut rcc.apb1);
//timer.listen(timer::Event::TimeOut);
let mut delay = Delay::new(p.core.SYST, clocks);
// Set up our GPIO pins
let disp_en = gpiob.pb2.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let extcomin = gpiob.pb1.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let cs = gpiob.pb0.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let mut v5_en = gpioa.pa3.into_push_pull_output(
&mut gpioa.moder,
&mut gpioa.otyper,
);
let reset_ble = gpiob.pb5.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let sck = gpioa.pa5.into_af5(&mut gpioa.moder, &mut gpioa.afrl);
let miso = gpioa.pa6.into_af5(&mut gpioa.moder, &mut gpioa.afrl);
let mosi = gpioa.pa7.into_af5(&mut gpioa.moder, &mut gpioa.afrl);
let tx = gpiob.pb6.into_af7(&mut gpiob.moder, &mut gpiob.afrl);
let rx = gpiob.pb7.into_af7(&mut gpiob.moder, &mut gpiob.afrl);
let button_1 = gpiob.pb8.into_pull_up_input(
&mut gpiob.moder,
&mut gpiob.pupdr,
);
let button_2 = gpiob.pb9.into_pull_up_input(
&mut gpiob.moder,
&mut gpiob.pupdr,
);
let button_3 = gpioc.pc13.into_pull_up_input(
&mut gpioc.moder,
&mut gpioc.pupdr,
);
// Set up our display
let mode = ls010b7dh01::MODE;
let spi = Spi::spi1(
p.device.SPI1,
(sck, miso, mosi),
mode,
1.mhz(),
clocks,
&mut rcc.apb2,
);
let mut display = Ls010b7dh01::new(spi, cs, disp_en);
// Set up our BLE
let mut serial = Serial::usart1(
p.device.USART1,
(tx, rx),
115_200.bps(),
clocks,
&mut rcc.apb2,
);
serial.listen(serial::Event::Rxne); // TODO: Serial interrupts?
let mut ble = rn4870::Rn4870::new(serial, reset_ble);
// Set the default values
v5_en.set_high();
display.enable();
// Set up syscfg to link GPIO to EXTI
p.device.SYSCFG.exticr3.modify(|_, w| unsafe {
w.bits(0x11)
/* This does not work
w.exti8().bits(0b001) // Port b
.exti9().bits(0b001) // Port b
*/
});
p.device.SYSCFG.exticr4.modify(|_, w| unsafe {
w.exti13().bits(0b010) // Port c
});
p.device.EXTI.imr1.modify(|_, w| {
w.mr8().set_bit().mr9().set_bit().mr13().set_bit()
});
p.device.EXTI.ftsr1.modify(|_, w| {
w.tr8().set_bit().tr9().set_bit().tr13().set_bit()
});
init::LateResources {
DISPLAY: display,
EXTCOMIN: extcomin,
BLE: ble,
EXTI: p.device.EXTI,
}
}
fn idle() -> ! {
loop {
rtfm::wfi();
}
}
fn ble_message(_t: &mut Threshold, mut r: USART1_EXTI25::Resources) {
let res = r.BLE.read_raw();
match res {
Ok(n) => {
if n < 32 {
return
}
(*r.DRAW_BUFFER)[*r.BUFFER_POS as usize] = n;
*r.BUFFER_POS += 1;
if *r.BUFFER_POS == 16 |
}
Err(nb::Error::Other(_)) => {
r.BLE.handle_error(|uart| { uart.clear_overflow_error(); } );
}
Err(nb::Error::WouldBlock) => {}
}
}
fn exti9_5(_t: &mut Threshold, mut r: EXTI9_5::Resources) {
if r.EXTI.pr1.read().pr8().bit_is_set() {
r.EXTI.pr1.modify(|_, w| w.pr8().set_bit());
*r.STATE = State::Ble;
}
if r.EXTI.pr1.read().pr9().bit_is_set() {
r.EXTI.pr1.modify(|_, w| w.pr9().set_bit());
*r.STATE = State::Time;
}
}
fn exti15_10(_t: &mut Threshold, mut r: EXTI15_10::Resources) {
if r.EXTI.pr1.read().pr13().bit_is_set() {
r.EXTI.pr1.modify(|_, w| w.pr13().set_bit());
*r.STATE = State::Face;
}
}
fn tick(_t: &mut Threshold, mut r: TIM7::Resources) {
}
fn sys_tick(_t: &mut Threshold, mut r: SYS_TICK::Resources) {
let toggle = *r.TOGGLE;
let extcomin = &mut *r.EXTCOMIN;
if *r.RESET_BLE {
r.BLE.hard_reset_on();
*r.RESET_BLE = false;
} else {
r.BLE.hard_reset_off();
}
match *r.STATE {
State::Ble => {
r.DISPLAY.clear();
//let s = String::from_utf8_lossy(&*r.DRAW_BUFFER);
unsafe {
let s = &*(&*r.DRAW_BUFFER as *const [u8] as *const str);
r.DISPLAY.draw(Font6x8::render_str(s).translate((5, 50)).into_iter());
r.DISPLAY.flush_buffer();
}
}
State::Time => {
*r.REDRAW = true;
draw_time(&mut *r.DISPLAY, *r.TIME);
*r.TIME += 1;
if *r.TIME == 60 {
*r.TIME = 0;
}
}
State::Face => {
if *r.REDRAW {
draw_face(&mut *r.DISPLAY);
*r.REDRAW = false;
}
}
}
// Toggle extcomin manually
if toggle {
(*extcomin).set_high();
} else {
(*extcomin).set_low();
}
*r.TOGGLE = !toggle;
}
fn draw_face(mut display: &mut display::Display) {
display.clear();
let bpp = Image1BPP::new(include_bytes!("../data/face_1bpp_neg.raw"), 120, 120)
.translate((0, 0));
display.draw(bpp.into_iter());
display.flush_buffer();
}
fn draw_time(mut display: &mut display::Display, time: u8) {
display.clear();
/*
let values = [
(125, 65), (124, 71), (123, 77 | {
*r.BUFFER_POS = 0;
} | conditional_block |
main.rs | r: init::Resources) -> init::LateResources {
let mut rcc = p.device.RCC.constrain();
let mut flash = p.device.FLASH.constrain();
let mut gpioa = p.device.GPIOA.split(&mut rcc.ahb);
let mut gpiob = p.device.GPIOB.split(&mut rcc.ahb);
let mut gpioc = p.device.GPIOC.split(&mut rcc.ahb);
// Enable the syscfg
rcc.apb2.enr().modify(|_, w| w.syscfgen().enabled());
rcc.apb2.rstr().modify(|_, w| w.syscfgrst().set_bit());
rcc.apb2.rstr().modify(|_, w| w.syscfgrst().clear_bit());
// Enable systick
p.core.SYST.set_clock_source(SystClkSource::Core);
p.core.SYST.set_reload(16_000_000);
p.core.SYST.enable_interrupt();
p.core.SYST.enable_counter();
// Set up our clocks & timer & delay
let clocks = rcc.cfgr.freeze(&mut flash.acr);
let mut timer = Timer::tim7(p.device.TIM7, 1.hz(), clocks, &mut rcc.apb1);
//timer.listen(timer::Event::TimeOut);
let mut delay = Delay::new(p.core.SYST, clocks);
// Set up our GPIO pins
let disp_en = gpiob.pb2.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let extcomin = gpiob.pb1.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let cs = gpiob.pb0.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let mut v5_en = gpioa.pa3.into_push_pull_output(
&mut gpioa.moder,
&mut gpioa.otyper,
);
let reset_ble = gpiob.pb5.into_push_pull_output(
&mut gpiob.moder,
&mut gpiob.otyper,
);
let sck = gpioa.pa5.into_af5(&mut gpioa.moder, &mut gpioa.afrl);
let miso = gpioa.pa6.into_af5(&mut gpioa.moder, &mut gpioa.afrl);
let mosi = gpioa.pa7.into_af5(&mut gpioa.moder, &mut gpioa.afrl);
let tx = gpiob.pb6.into_af7(&mut gpiob.moder, &mut gpiob.afrl);
let rx = gpiob.pb7.into_af7(&mut gpiob.moder, &mut gpiob.afrl);
let button_1 = gpiob.pb8.into_pull_up_input(
&mut gpiob.moder,
&mut gpiob.pupdr,
);
let button_2 = gpiob.pb9.into_pull_up_input(
&mut gpiob.moder,
&mut gpiob.pupdr,
);
let button_3 = gpioc.pc13.into_pull_up_input(
&mut gpioc.moder,
&mut gpioc.pupdr,
);
// Set up our display
let mode = ls010b7dh01::MODE;
let spi = Spi::spi1(
p.device.SPI1,
(sck, miso, mosi),
mode,
1.mhz(),
clocks,
&mut rcc.apb2,
);
let mut display = Ls010b7dh01::new(spi, cs, disp_en);
// Set up our BLE
let mut serial = Serial::usart1(
p.device.USART1,
(tx, rx),
115_200.bps(),
clocks,
&mut rcc.apb2,
);
serial.listen(serial::Event::Rxne); // TODO: Serial interrupts?
let mut ble = rn4870::Rn4870::new(serial, reset_ble);
// Set the default values
v5_en.set_high();
display.enable();
// Set up syscfg to link GPIO to EXTI
p.device.SYSCFG.exticr3.modify(|_, w| unsafe {
w.bits(0x11)
/* This does not work
w.exti8().bits(0b001) // Port b
.exti9().bits(0b001) // Port b
*/
});
p.device.SYSCFG.exticr4.modify(|_, w| unsafe {
w.exti13().bits(0b010) // Port c
});
p.device.EXTI.imr1.modify(|_, w| {
w.mr8().set_bit().mr9().set_bit().mr13().set_bit()
});
p.device.EXTI.ftsr1.modify(|_, w| {
w.tr8().set_bit().tr9().set_bit().tr13().set_bit()
});
init::LateResources {
DISPLAY: display,
EXTCOMIN: extcomin,
BLE: ble,
EXTI: p.device.EXTI,
}
}
fn idle() -> ! {
loop {
rtfm::wfi();
}
}
fn ble_message(_t: &mut Threshold, mut r: USART1_EXTI25::Resources) {
let res = r.BLE.read_raw();
match res {
Ok(n) => {
if n < 32 {
return
}
(*r.DRAW_BUFFER)[*r.BUFFER_POS as usize] = n;
*r.BUFFER_POS += 1;
if *r.BUFFER_POS == 16 {
*r.BUFFER_POS = 0;
}
}
Err(nb::Error::Other(_)) => {
r.BLE.handle_error(|uart| { uart.clear_overflow_error(); } );
}
Err(nb::Error::WouldBlock) => {}
}
}
fn exti9_5(_t: &mut Threshold, mut r: EXTI9_5::Resources) {
if r.EXTI.pr1.read().pr8().bit_is_set() {
r.EXTI.pr1.modify(|_, w| w.pr8().set_bit());
*r.STATE = State::Ble;
}
if r.EXTI.pr1.read().pr9().bit_is_set() {
r.EXTI.pr1.modify(|_, w| w.pr9().set_bit());
*r.STATE = State::Time;
}
}
fn exti15_10(_t: &mut Threshold, mut r: EXTI15_10::Resources) {
if r.EXTI.pr1.read().pr13().bit_is_set() {
r.EXTI.pr1.modify(|_, w| w.pr13().set_bit());
*r.STATE = State::Face;
}
}
fn | (_t: &mut Threshold, mut r: TIM7::Resources) {
}
fn sys_tick(_t: &mut Threshold, mut r: SYS_TICK::Resources) {
let toggle = *r.TOGGLE;
let extcomin = &mut *r.EXTCOMIN;
if *r.RESET_BLE {
r.BLE.hard_reset_on();
*r.RESET_BLE = false;
} else {
r.BLE.hard_reset_off();
}
match *r.STATE {
State::Ble => {
r.DISPLAY.clear();
//let s = String::from_utf8_lossy(&*r.DRAW_BUFFER);
unsafe {
let s = &*(&*r.DRAW_BUFFER as *const [u8] as *const str);
r.DISPLAY.draw(Font6x8::render_str(s).translate((5, 50)).into_iter());
r.DISPLAY.flush_buffer();
}
}
State::Time => {
*r.REDRAW = true;
draw_time(&mut *r.DISPLAY, *r.TIME);
*r.TIME += 1;
if *r.TIME == 60 {
*r.TIME = 0;
}
}
State::Face => {
if *r.REDRAW {
draw_face(&mut *r.DISPLAY);
*r.REDRAW = false;
}
}
}
// Toggle extcomin manually
if toggle {
(*extcomin).set_high();
} else {
(*extcomin).set_low();
}
*r.TOGGLE = !toggle;
}
fn draw_face(mut display: &mut display::Display) {
display.clear();
let bpp = Image1BPP::new(include_bytes!("../data/face_1bpp_neg.raw"), 120, 120)
.translate((0, 0));
display.draw(bpp.into_iter());
display.flush_buffer();
}
fn draw_time(mut display: &mut display::Display, time: u8) {
display.clear();
/*
let values = [
(125, 65), (124, 71), (123, 77), | tick | identifier_name |
model.py | 1)
# Scalar Encoder
self.scalar_encoder = nn.Linear(num_players, 128)
# transformer
# self.max_entities = 10
self.action_map = [None, "NORTH", "EAST", "SOUTH", "WEST", "CONVERT", "SPAWN"]
self.SHIP_TYPE = 2
self.SHIPYARD_TYPE = 1
num_actions = len(self.action_map)
self.transformer = nn.TransformerEncoder(nn.TransformerEncoderLayer(d_model=128, nhead=8, dim_feedforward=100), 2, norm=None)
self.policy = nn.Sequential(
nn.Linear(128, 128),
nn.ReLU(),
nn.Linear(128, num_actions)
)
self.value = nn.Sequential(
nn.Linear(128 * self.MAX_ENTITIES, 400),
nn.ReLU(),
nn.Linear(400, 100),
torch.nn.ReLU(),
nn.Linear(100, 1)
)
self.softmax = nn.Softmax(-1)
def device(self):
return next(self.parameters()).device
def | (self, state, mask = False):
# Scalar encoding
state = self.parse_state(state)
scalar = state['scalar'].to(self.device())
scalar_encoding = F.relu(self.scalar_encoder(scalar)).unsqueeze(1)
# Spatial Encoding
game_map = state['map'].to(self.device())
map_encoding = self.map(game_map).unsqueeze(1)
# Entity Encoding
entity_typ = state['entity_typ'].to(self.device())
entity_pos = state['entity_pos'].to(self.device())
entity_scalar = state['entity_scalar'].to(self.device())
entity_encodings = self.entity(entity_typ, entity_pos, entity_scalar)
embeddings = map_encoding + entity_encodings + scalar_encoding
set_embedding = self.transformer(embeddings)
out = self.policy(set_embedding)
if mask == True:
lens = []
for eid in state['entity_id']:
n_entities = len(eid)
lens.append(torch.tensor([1] * n_entities + [0] * (self.MAX_ENTITIES - n_entities)))
m = torch.stack(lens).to(self.device())
return self.softmax(out), m
return self.softmax(out)
def action(self, states):
if not isinstance(states, list):
states = [states]
t_states = self.parse_state(states)
out = self.forward(states)
actions_iter = []
raw_actions_iter = []
for i, state in enumerate(states):
raw_actions = Categorical(probs=out[i]).sample()
actions = {}
n_entities = len(t_states['entity_id'][i])
# TODO: Migrate this code to env helper
for e, eid in enumerate(t_states['entity_id'][i]):
act = self.action_map[raw_actions[e]]
typ = t_states['entity_typ'][i][e]
if typ == self.SHIP_TYPE and act == "SPAWN":
act = None
elif typ == self.SHIPYARD_TYPE and (act != "SPAWN" and act != None):
act = None
elif typ == 0:
continue
if act == "SPAWN":
if n_entities < self.MAX_ENTITIES:
n_entities += 1
else:
act = None
if act is not None:
actions[eid] = act
actions_iter.append(actions)
raw_actions_iter.append(raw_actions)
return actions_iter, raw_actions_iter
class ParseState(object):
"""Rescale the image in a sample to a given size.
Args:
output_size (tuple or int): Desired output size. If tuple, output is
matched to output_size. If int, smaller of image edges is matched
to output_size keeping aspect ratio the same.
"""
def __init__(self, config, max_entities):
self.map_size = config['size']
self.max_halite = config['maxCellHalite']
self.starting_halite = config['startingHalite']
self.max_entities = max_entities
def __call__(self, states):
if not isinstance(states, list):
states = [states]
spat_map_iter = []
entity_typ_iter = []
entity_pos_iter = []
entity_id_iter = []
entity_scalar_iter = []
scalar_iter = []
for s in states:
step = s['step']
halite = torch.tensor(s['halite']).float()
halite = halite.reshape(self.map_size, self.map_size, 1) / self.max_halite
obstruction = torch.zeros(self.map_size**2).float()
me = s['players'][s['player']]
my_halite, my_shipyards, my_ships = tuple(me)
scalar = torch.zeros(len(s['players']))
scalar[0] = my_halite
entity_typ = []
entity_pos = []
entity_scalar = []
entity_id = []
for shipyard_id, shipyard_pos in my_shipyards.items():
obstruction[shipyard_pos] = 1.0
x = int(shipyard_pos % self.map_size)
y = int(shipyard_pos / self.map_size)
entity_typ.append(1)
entity_pos.append([x,y])
entity_scalar.append([0])
entity_id.append(shipyard_id)
for ship_id, ship_pos in my_ships.items():
obstruction[ship_pos[0]] = 1.0
x = int(ship_pos[0] % self.map_size)
y = int(ship_pos[0] / self.map_size)
entity_typ.append(2)
entity_pos.append([x,y])
entity_scalar.append([ship_pos[1]])
entity_id.append(ship_id)
opponents = s['players']
scalar_loc = 1
for i, opponent in enumerate(opponents):
if i != s['player']:
opp_halite, opp_shipyards, opp_ships = tuple(opponent)
scalar[scalar_loc] = opp_halite
for shipyard_pos in opp_shipyards.values():
obstruction[shipyard_pos] = 1.0
for ship_pos in opp_ships.values():
obstruction[ship_pos[0]] = 1.0
scalar_loc += 1
obstruction = obstruction.reshape(self.map_size, self.map_size, 1)
spat_map = torch.cat((halite, obstruction), 2).unsqueeze(0).permute(0,3,1,2)
n_entities = len(entity_id)
diff = self.max_entities - n_entities
entity_typ = F.pad(torch.tensor(entity_typ).long().unsqueeze(0), (0, diff), "constant", 0)
entity_pos = F.pad(torch.tensor(entity_pos).long().unsqueeze(0), (0, 0, 0, diff), "constant", 0)
entity_scalar = F.pad(torch.tensor(entity_scalar).float().unsqueeze(0), (0, 0, 0, diff), "constant", 0)
scalar = scalar.unsqueeze(0) / self.starting_halite
spat_map_iter.append(spat_map)
entity_typ_iter.append(entity_typ)
entity_pos_iter.append(entity_pos)
entity_id_iter.append(entity_id)
entity_scalar_iter.append(entity_scalar)
scalar_iter.append(scalar)
return {
'map': torch.cat(spat_map_iter),
'entity_typ': torch.cat(entity_typ_iter),
'entity_pos': torch.cat(entity_pos_iter),
'entity_scalar': torch.cat(entity_scalar_iter),
'entity_id': entity_id_iter,
'scalar': torch.cat(scalar_iter)
}
class MapEmbedding(nn.Module):
def __init__(self, embed_size=256, depth=2, maps=2):
super(MapEmbedding, self).__init__()
blocks = []
c_b = 64
while c_b < embed_size:
blocks.append(c_b)
c_b *= 2
blocks.append(embed_size)
deepths = [depth] * len(blocks)
self.resnet = ResNetEncoder(in_channels=maps, blocks_sizes=blocks, deepths=deepths)
def forward(self, multi_layer_map):
return self.resnet(multi_layer_map)
class EntityEmbedding(nn.Module):
def __init__(self, d_model, map_size, n_scalars):
super(EntityEmbedding, self).__init__()
# self.lut = pre_trained.embeddings.word_embeddings
self.EntityType = nn.Embedding(2 + 1, d_model)
self.EntityPosition = PositionalEncoding2D(d_model, map_size, map_size)
self.fc = nn.Linear(n_scalars, d_model)
self.EntityType.weight.data.uniform_(-0.1, .1)
def forward(self, typ, pos, scalar):
return self.EntityType(typ) + self.EntityPosition(pos) + self.fc(scalar)
# Retrieved from pytorch website
class PositionalEncoding2D(nn.Module):
def __init__(self, d_model, height, width):
super(PositionalEncoding2D, self).__init__()
if d_model % 4 != 0:
raise Error()
pe = torch.zeros(d_model, height, width)
d_model = int(d_model / 2)
div_term = torch.exp(torch.arange(0., d_model | forward | identifier_name |
model.py | 1)
# Scalar Encoder
self.scalar_encoder = nn.Linear(num_players, 128)
# transformer
# self.max_entities = 10
self.action_map = [None, "NORTH", "EAST", "SOUTH", "WEST", "CONVERT", "SPAWN"]
self.SHIP_TYPE = 2
self.SHIPYARD_TYPE = 1
num_actions = len(self.action_map)
self.transformer = nn.TransformerEncoder(nn.TransformerEncoderLayer(d_model=128, nhead=8, dim_feedforward=100), 2, norm=None)
self.policy = nn.Sequential(
nn.Linear(128, 128),
nn.ReLU(),
nn.Linear(128, num_actions)
)
self.value = nn.Sequential(
nn.Linear(128 * self.MAX_ENTITIES, 400),
nn.ReLU(),
nn.Linear(400, 100),
torch.nn.ReLU(),
nn.Linear(100, 1)
)
self.softmax = nn.Softmax(-1)
def device(self):
return next(self.parameters()).device
def forward(self, state, mask = False):
# Scalar encoding
state = self.parse_state(state)
scalar = state['scalar'].to(self.device())
scalar_encoding = F.relu(self.scalar_encoder(scalar)).unsqueeze(1)
# Spatial Encoding
game_map = state['map'].to(self.device())
map_encoding = self.map(game_map).unsqueeze(1)
# Entity Encoding
entity_typ = state['entity_typ'].to(self.device())
entity_pos = state['entity_pos'].to(self.device())
entity_scalar = state['entity_scalar'].to(self.device())
entity_encodings = self.entity(entity_typ, entity_pos, entity_scalar)
embeddings = map_encoding + entity_encodings + scalar_encoding
set_embedding = self.transformer(embeddings)
out = self.policy(set_embedding)
if mask == True:
lens = []
for eid in state['entity_id']:
n_entities = len(eid)
lens.append(torch.tensor([1] * n_entities + [0] * (self.MAX_ENTITIES - n_entities)))
m = torch.stack(lens).to(self.device())
return self.softmax(out), m
return self.softmax(out)
def action(self, states):
if not isinstance(states, list):
states = [states]
t_states = self.parse_state(states)
out = self.forward(states)
actions_iter = []
raw_actions_iter = []
for i, state in enumerate(states):
raw_actions = Categorical(probs=out[i]).sample()
actions = {}
n_entities = len(t_states['entity_id'][i])
# TODO: Migrate this code to env helper
for e, eid in enumerate(t_states['entity_id'][i]):
act = self.action_map[raw_actions[e]]
typ = t_states['entity_typ'][i][e]
if typ == self.SHIP_TYPE and act == "SPAWN":
act = None
elif typ == self.SHIPYARD_TYPE and (act != "SPAWN" and act != None):
act = None
elif typ == 0:
continue
if act == "SPAWN":
if n_entities < self.MAX_ENTITIES:
n_entities += 1
else:
act = None
if act is not None:
actions[eid] = act
actions_iter.append(actions)
raw_actions_iter.append(raw_actions)
return actions_iter, raw_actions_iter
class ParseState(object):
"""Rescale the image in a sample to a given size.
Args:
output_size (tuple or int): Desired output size. If tuple, output is
matched to output_size. If int, smaller of image edges is matched
to output_size keeping aspect ratio the same.
"""
def __init__(self, config, max_entities):
self.map_size = config['size']
self.max_halite = config['maxCellHalite']
self.starting_halite = config['startingHalite']
self.max_entities = max_entities
def __call__(self, states):
if not isinstance(states, list):
states = [states]
spat_map_iter = []
entity_typ_iter = []
entity_pos_iter = []
entity_id_iter = []
entity_scalar_iter = []
scalar_iter = []
for s in states:
step = s['step']
halite = torch.tensor(s['halite']).float()
halite = halite.reshape(self.map_size, self.map_size, 1) / self.max_halite
obstruction = torch.zeros(self.map_size**2).float()
me = s['players'][s['player']]
my_halite, my_shipyards, my_ships = tuple(me)
scalar = torch.zeros(len(s['players']))
scalar[0] = my_halite
entity_typ = []
entity_pos = []
entity_scalar = []
entity_id = []
for shipyard_id, shipyard_pos in my_shipyards.items():
obstruction[shipyard_pos] = 1.0
x = int(shipyard_pos % self.map_size)
y = int(shipyard_pos / self.map_size)
entity_typ.append(1)
entity_pos.append([x,y])
entity_scalar.append([0])
entity_id.append(shipyard_id)
for ship_id, ship_pos in my_ships.items():
obstruction[ship_pos[0]] = 1.0
x = int(ship_pos[0] % self.map_size)
y = int(ship_pos[0] / self.map_size)
entity_typ.append(2)
entity_pos.append([x,y])
entity_scalar.append([ship_pos[1]])
entity_id.append(ship_id)
opponents = s['players']
scalar_loc = 1
for i, opponent in enumerate(opponents):
if i != s['player']:
opp_halite, opp_shipyards, opp_ships = tuple(opponent)
scalar[scalar_loc] = opp_halite
for shipyard_pos in opp_shipyards.values():
obstruction[shipyard_pos] = 1.0
for ship_pos in opp_ships.values():
obstruction[ship_pos[0]] = 1.0
scalar_loc += 1
obstruction = obstruction.reshape(self.map_size, self.map_size, 1)
spat_map = torch.cat((halite, obstruction), 2).unsqueeze(0).permute(0,3,1,2)
n_entities = len(entity_id)
diff = self.max_entities - n_entities
entity_typ = F.pad(torch.tensor(entity_typ).long().unsqueeze(0), (0, diff), "constant", 0)
entity_pos = F.pad(torch.tensor(entity_pos).long().unsqueeze(0), (0, 0, 0, diff), "constant", 0)
entity_scalar = F.pad(torch.tensor(entity_scalar).float().unsqueeze(0), (0, 0, 0, diff), "constant", 0)
scalar = scalar.unsqueeze(0) / self.starting_halite
spat_map_iter.append(spat_map)
entity_typ_iter.append(entity_typ)
entity_pos_iter.append(entity_pos)
entity_id_iter.append(entity_id)
entity_scalar_iter.append(entity_scalar)
scalar_iter.append(scalar)
return {
'map': torch.cat(spat_map_iter),
'entity_typ': torch.cat(entity_typ_iter),
'entity_pos': torch.cat(entity_pos_iter),
'entity_scalar': torch.cat(entity_scalar_iter),
'entity_id': entity_id_iter,
'scalar': torch.cat(scalar_iter)
}
class MapEmbedding(nn.Module):
def __init__(self, embed_size=256, depth=2, maps=2):
super(MapEmbedding, self).__init__()
blocks = []
c_b = 64
while c_b < embed_size:
blocks.append(c_b)
c_b *= 2
blocks.append(embed_size)
deepths = [depth] * len(blocks)
self.resnet = ResNetEncoder(in_channels=maps, blocks_sizes=blocks, deepths=deepths)
def forward(self, multi_layer_map):
return self.resnet(multi_layer_map)
class EntityEmbedding(nn.Module):
def __init__(self, d_model, map_size, n_scalars):
super(EntityEmbedding, self).__init__()
# self.lut = pre_trained.embeddings.word_embeddings
self.EntityType = nn.Embedding(2 + 1, d_model)
self.EntityPosition = PositionalEncoding2D(d_model, map_size, map_size)
self.fc = nn.Linear(n_scalars, d_model)
self.EntityType.weight.data.uniform_(-0.1, .1)
def forward(self, typ, pos, scalar):
|
# Retrieved from pytorch website
class PositionalEncoding2D(nn.Module):
def __init__(self, d_model, height, width):
super(PositionalEncoding2D, self).__init__()
if d_model % 4 != 0:
raise Error()
pe = torch.zeros(d_model, height, width)
d_model = int(d_model / 2)
div_term = torch.exp(torch.arange(0., d_model | return self.EntityType(typ) + self.EntityPosition(pos) + self.fc(scalar) | identifier_body |
model.py | 1)
# Scalar Encoder
self.scalar_encoder = nn.Linear(num_players, 128)
# transformer
# self.max_entities = 10
self.action_map = [None, "NORTH", "EAST", "SOUTH", "WEST", "CONVERT", "SPAWN"]
self.SHIP_TYPE = 2
self.SHIPYARD_TYPE = 1
num_actions = len(self.action_map)
self.transformer = nn.TransformerEncoder(nn.TransformerEncoderLayer(d_model=128, nhead=8, dim_feedforward=100), 2, norm=None)
self.policy = nn.Sequential(
nn.Linear(128, 128),
nn.ReLU(),
nn.Linear(128, num_actions)
)
self.value = nn.Sequential(
nn.Linear(128 * self.MAX_ENTITIES, 400),
nn.ReLU(),
nn.Linear(400, 100),
torch.nn.ReLU(),
nn.Linear(100, 1)
)
self.softmax = nn.Softmax(-1)
def device(self):
return next(self.parameters()).device
def forward(self, state, mask = False):
# Scalar encoding
state = self.parse_state(state)
scalar = state['scalar'].to(self.device())
scalar_encoding = F.relu(self.scalar_encoder(scalar)).unsqueeze(1)
# Spatial Encoding
game_map = state['map'].to(self.device())
map_encoding = self.map(game_map).unsqueeze(1)
# Entity Encoding
entity_typ = state['entity_typ'].to(self.device())
entity_pos = state['entity_pos'].to(self.device())
entity_scalar = state['entity_scalar'].to(self.device())
entity_encodings = self.entity(entity_typ, entity_pos, entity_scalar)
embeddings = map_encoding + entity_encodings + scalar_encoding
set_embedding = self.transformer(embeddings)
out = self.policy(set_embedding)
if mask == True:
lens = []
for eid in state['entity_id']:
n_entities = len(eid)
lens.append(torch.tensor([1] * n_entities + [0] * (self.MAX_ENTITIES - n_entities)))
m = torch.stack(lens).to(self.device())
return self.softmax(out), m
return self.softmax(out)
def action(self, states):
if not isinstance(states, list):
states = [states]
t_states = self.parse_state(states)
out = self.forward(states)
actions_iter = []
raw_actions_iter = []
for i, state in enumerate(states):
raw_actions = Categorical(probs=out[i]).sample()
actions = {}
n_entities = len(t_states['entity_id'][i])
# TODO: Migrate this code to env helper
for e, eid in enumerate(t_states['entity_id'][i]):
act = self.action_map[raw_actions[e]]
typ = t_states['entity_typ'][i][e]
if typ == self.SHIP_TYPE and act == "SPAWN":
act = None
elif typ == self.SHIPYARD_TYPE and (act != "SPAWN" and act != None):
act = None
elif typ == 0:
continue
if act == "SPAWN":
if n_entities < self.MAX_ENTITIES:
n_entities += 1
else:
act = None
if act is not None:
actions[eid] = act
actions_iter.append(actions)
raw_actions_iter.append(raw_actions)
return actions_iter, raw_actions_iter
class ParseState(object):
"""Rescale the image in a sample to a given size.
Args:
output_size (tuple or int): Desired output size. If tuple, output is
matched to output_size. If int, smaller of image edges is matched
to output_size keeping aspect ratio the same.
"""
def __init__(self, config, max_entities):
self.map_size = config['size']
self.max_halite = config['maxCellHalite']
self.starting_halite = config['startingHalite']
self.max_entities = max_entities
def __call__(self, states):
if not isinstance(states, list):
states = [states]
spat_map_iter = []
entity_typ_iter = []
entity_pos_iter = []
entity_id_iter = []
entity_scalar_iter = []
scalar_iter = []
for s in states:
step = s['step']
halite = torch.tensor(s['halite']).float()
halite = halite.reshape(self.map_size, self.map_size, 1) / self.max_halite
obstruction = torch.zeros(self.map_size**2).float()
me = s['players'][s['player']]
my_halite, my_shipyards, my_ships = tuple(me)
scalar = torch.zeros(len(s['players']))
scalar[0] = my_halite
entity_typ = []
entity_pos = []
entity_scalar = []
entity_id = []
for shipyard_id, shipyard_pos in my_shipyards.items():
obstruction[shipyard_pos] = 1.0
x = int(shipyard_pos % self.map_size)
y = int(shipyard_pos / self.map_size)
entity_typ.append(1)
entity_pos.append([x,y])
entity_scalar.append([0])
entity_id.append(shipyard_id)
for ship_id, ship_pos in my_ships.items():
obstruction[ship_pos[0]] = 1.0
x = int(ship_pos[0] % self.map_size)
y = int(ship_pos[0] / self.map_size)
entity_typ.append(2)
entity_pos.append([x,y])
entity_scalar.append([ship_pos[1]])
entity_id.append(ship_id)
opponents = s['players']
scalar_loc = 1
for i, opponent in enumerate(opponents):
if i != s['player']:
opp_halite, opp_shipyards, opp_ships = tuple(opponent)
scalar[scalar_loc] = opp_halite
for shipyard_pos in opp_shipyards.values():
obstruction[shipyard_pos] = 1.0
for ship_pos in opp_ships.values():
obstruction[ship_pos[0]] = 1.0
scalar_loc += 1
obstruction = obstruction.reshape(self.map_size, self.map_size, 1)
spat_map = torch.cat((halite, obstruction), 2).unsqueeze(0).permute(0,3,1,2)
n_entities = len(entity_id)
diff = self.max_entities - n_entities
entity_typ = F.pad(torch.tensor(entity_typ).long().unsqueeze(0), (0, diff), "constant", 0)
entity_pos = F.pad(torch.tensor(entity_pos).long().unsqueeze(0), (0, 0, 0, diff), "constant", 0) | entity_typ_iter.append(entity_typ)
entity_pos_iter.append(entity_pos)
entity_id_iter.append(entity_id)
entity_scalar_iter.append(entity_scalar)
scalar_iter.append(scalar)
return {
'map': torch.cat(spat_map_iter),
'entity_typ': torch.cat(entity_typ_iter),
'entity_pos': torch.cat(entity_pos_iter),
'entity_scalar': torch.cat(entity_scalar_iter),
'entity_id': entity_id_iter,
'scalar': torch.cat(scalar_iter)
}
class MapEmbedding(nn.Module):
def __init__(self, embed_size=256, depth=2, maps=2):
super(MapEmbedding, self).__init__()
blocks = []
c_b = 64
while c_b < embed_size:
blocks.append(c_b)
c_b *= 2
blocks.append(embed_size)
deepths = [depth] * len(blocks)
self.resnet = ResNetEncoder(in_channels=maps, blocks_sizes=blocks, deepths=deepths)
def forward(self, multi_layer_map):
return self.resnet(multi_layer_map)
class EntityEmbedding(nn.Module):
def __init__(self, d_model, map_size, n_scalars):
super(EntityEmbedding, self).__init__()
# self.lut = pre_trained.embeddings.word_embeddings
self.EntityType = nn.Embedding(2 + 1, d_model)
self.EntityPosition = PositionalEncoding2D(d_model, map_size, map_size)
self.fc = nn.Linear(n_scalars, d_model)
self.EntityType.weight.data.uniform_(-0.1, .1)
def forward(self, typ, pos, scalar):
return self.EntityType(typ) + self.EntityPosition(pos) + self.fc(scalar)
# Retrieved from pytorch website
class PositionalEncoding2D(nn.Module):
def __init__(self, d_model, height, width):
super(PositionalEncoding2D, self).__init__()
if d_model % 4 != 0:
raise Error()
pe = torch.zeros(d_model, height, width)
d_model = int(d_model / 2)
div_term = torch.exp(torch.arange(0., d_model, | entity_scalar = F.pad(torch.tensor(entity_scalar).float().unsqueeze(0), (0, 0, 0, diff), "constant", 0)
scalar = scalar.unsqueeze(0) / self.starting_halite
spat_map_iter.append(spat_map) | random_line_split |
model.py | 1)
# Scalar Encoder
self.scalar_encoder = nn.Linear(num_players, 128)
# transformer
# self.max_entities = 10
self.action_map = [None, "NORTH", "EAST", "SOUTH", "WEST", "CONVERT", "SPAWN"]
self.SHIP_TYPE = 2
self.SHIPYARD_TYPE = 1
num_actions = len(self.action_map)
self.transformer = nn.TransformerEncoder(nn.TransformerEncoderLayer(d_model=128, nhead=8, dim_feedforward=100), 2, norm=None)
self.policy = nn.Sequential(
nn.Linear(128, 128),
nn.ReLU(),
nn.Linear(128, num_actions)
)
self.value = nn.Sequential(
nn.Linear(128 * self.MAX_ENTITIES, 400),
nn.ReLU(),
nn.Linear(400, 100),
torch.nn.ReLU(),
nn.Linear(100, 1)
)
self.softmax = nn.Softmax(-1)
def device(self):
return next(self.parameters()).device
def forward(self, state, mask = False):
# Scalar encoding
state = self.parse_state(state)
scalar = state['scalar'].to(self.device())
scalar_encoding = F.relu(self.scalar_encoder(scalar)).unsqueeze(1)
# Spatial Encoding
game_map = state['map'].to(self.device())
map_encoding = self.map(game_map).unsqueeze(1)
# Entity Encoding
entity_typ = state['entity_typ'].to(self.device())
entity_pos = state['entity_pos'].to(self.device())
entity_scalar = state['entity_scalar'].to(self.device())
entity_encodings = self.entity(entity_typ, entity_pos, entity_scalar)
embeddings = map_encoding + entity_encodings + scalar_encoding
set_embedding = self.transformer(embeddings)
out = self.policy(set_embedding)
if mask == True:
lens = []
for eid in state['entity_id']:
n_entities = len(eid)
lens.append(torch.tensor([1] * n_entities + [0] * (self.MAX_ENTITIES - n_entities)))
m = torch.stack(lens).to(self.device())
return self.softmax(out), m
return self.softmax(out)
def action(self, states):
if not isinstance(states, list):
states = [states]
t_states = self.parse_state(states)
out = self.forward(states)
actions_iter = []
raw_actions_iter = []
for i, state in enumerate(states):
raw_actions = Categorical(probs=out[i]).sample()
actions = {}
n_entities = len(t_states['entity_id'][i])
# TODO: Migrate this code to env helper
for e, eid in enumerate(t_states['entity_id'][i]):
act = self.action_map[raw_actions[e]]
typ = t_states['entity_typ'][i][e]
if typ == self.SHIP_TYPE and act == "SPAWN":
act = None
elif typ == self.SHIPYARD_TYPE and (act != "SPAWN" and act != None):
act = None
elif typ == 0:
continue
if act == "SPAWN":
if n_entities < self.MAX_ENTITIES:
|
else:
act = None
if act is not None:
actions[eid] = act
actions_iter.append(actions)
raw_actions_iter.append(raw_actions)
return actions_iter, raw_actions_iter
class ParseState(object):
"""Rescale the image in a sample to a given size.
Args:
output_size (tuple or int): Desired output size. If tuple, output is
matched to output_size. If int, smaller of image edges is matched
to output_size keeping aspect ratio the same.
"""
def __init__(self, config, max_entities):
self.map_size = config['size']
self.max_halite = config['maxCellHalite']
self.starting_halite = config['startingHalite']
self.max_entities = max_entities
def __call__(self, states):
if not isinstance(states, list):
states = [states]
spat_map_iter = []
entity_typ_iter = []
entity_pos_iter = []
entity_id_iter = []
entity_scalar_iter = []
scalar_iter = []
for s in states:
step = s['step']
halite = torch.tensor(s['halite']).float()
halite = halite.reshape(self.map_size, self.map_size, 1) / self.max_halite
obstruction = torch.zeros(self.map_size**2).float()
me = s['players'][s['player']]
my_halite, my_shipyards, my_ships = tuple(me)
scalar = torch.zeros(len(s['players']))
scalar[0] = my_halite
entity_typ = []
entity_pos = []
entity_scalar = []
entity_id = []
for shipyard_id, shipyard_pos in my_shipyards.items():
obstruction[shipyard_pos] = 1.0
x = int(shipyard_pos % self.map_size)
y = int(shipyard_pos / self.map_size)
entity_typ.append(1)
entity_pos.append([x,y])
entity_scalar.append([0])
entity_id.append(shipyard_id)
for ship_id, ship_pos in my_ships.items():
obstruction[ship_pos[0]] = 1.0
x = int(ship_pos[0] % self.map_size)
y = int(ship_pos[0] / self.map_size)
entity_typ.append(2)
entity_pos.append([x,y])
entity_scalar.append([ship_pos[1]])
entity_id.append(ship_id)
opponents = s['players']
scalar_loc = 1
for i, opponent in enumerate(opponents):
if i != s['player']:
opp_halite, opp_shipyards, opp_ships = tuple(opponent)
scalar[scalar_loc] = opp_halite
for shipyard_pos in opp_shipyards.values():
obstruction[shipyard_pos] = 1.0
for ship_pos in opp_ships.values():
obstruction[ship_pos[0]] = 1.0
scalar_loc += 1
obstruction = obstruction.reshape(self.map_size, self.map_size, 1)
spat_map = torch.cat((halite, obstruction), 2).unsqueeze(0).permute(0,3,1,2)
n_entities = len(entity_id)
diff = self.max_entities - n_entities
entity_typ = F.pad(torch.tensor(entity_typ).long().unsqueeze(0), (0, diff), "constant", 0)
entity_pos = F.pad(torch.tensor(entity_pos).long().unsqueeze(0), (0, 0, 0, diff), "constant", 0)
entity_scalar = F.pad(torch.tensor(entity_scalar).float().unsqueeze(0), (0, 0, 0, diff), "constant", 0)
scalar = scalar.unsqueeze(0) / self.starting_halite
spat_map_iter.append(spat_map)
entity_typ_iter.append(entity_typ)
entity_pos_iter.append(entity_pos)
entity_id_iter.append(entity_id)
entity_scalar_iter.append(entity_scalar)
scalar_iter.append(scalar)
return {
'map': torch.cat(spat_map_iter),
'entity_typ': torch.cat(entity_typ_iter),
'entity_pos': torch.cat(entity_pos_iter),
'entity_scalar': torch.cat(entity_scalar_iter),
'entity_id': entity_id_iter,
'scalar': torch.cat(scalar_iter)
}
class MapEmbedding(nn.Module):
def __init__(self, embed_size=256, depth=2, maps=2):
super(MapEmbedding, self).__init__()
blocks = []
c_b = 64
while c_b < embed_size:
blocks.append(c_b)
c_b *= 2
blocks.append(embed_size)
deepths = [depth] * len(blocks)
self.resnet = ResNetEncoder(in_channels=maps, blocks_sizes=blocks, deepths=deepths)
def forward(self, multi_layer_map):
return self.resnet(multi_layer_map)
class EntityEmbedding(nn.Module):
def __init__(self, d_model, map_size, n_scalars):
super(EntityEmbedding, self).__init__()
# self.lut = pre_trained.embeddings.word_embeddings
self.EntityType = nn.Embedding(2 + 1, d_model)
self.EntityPosition = PositionalEncoding2D(d_model, map_size, map_size)
self.fc = nn.Linear(n_scalars, d_model)
self.EntityType.weight.data.uniform_(-0.1, .1)
def forward(self, typ, pos, scalar):
return self.EntityType(typ) + self.EntityPosition(pos) + self.fc(scalar)
# Retrieved from pytorch website
class PositionalEncoding2D(nn.Module):
def __init__(self, d_model, height, width):
super(PositionalEncoding2D, self).__init__()
if d_model % 4 != 0:
raise Error()
pe = torch.zeros(d_model, height, width)
d_model = int(d_model / 2)
div_term = torch.exp(torch.arange(0., d_model, | n_entities += 1 | conditional_block |
match.go | the default tag to return in case there is no match
// - desired: list of desired tags, ordered by preference, starting with
// the most-preferred.
//
// Algorithm:
// 1) Set the best match to the lowest confidence level
// 2) For each tag in "desired":
// a) For each tag in "supported":
// 1) compute the match between the two tags.
// 2) if the match is better than the previous best match, replace it
// with the new match. (see next section)
// b) if the current best match is Exact and pin is true the result will be
// frozen to the language found thusfar, although better matches may
// still be found for the same language.
// 3) If the best match so far is below a certain threshold, return "default".
//
// Ranking:
// We use two phases to determine whether one pair of tags are a better match
// than another pair of tags. First, we determine a rough confidence level. If the
// levels are different, the one with the highest confidence wins.
// Second, if the rough confidence levels are identical, we use a set of tie-breaker
// rules.
//
// The confidence level of matching a pair of tags is determined by finding the
// lowest confidence level of any matches of the corresponding subtags (the
// result is deemed as good as its weakest link).
// We define the following levels:
// Exact - An exact match of a subtag, before adding likely subtags.
// MaxExact - An exact match of a subtag, after adding likely subtags.
// [See Note 2].
// High - High level of mutual intelligibility between different subtag
// variants.
// Low - Low level of mutual intelligibility between different subtag
// variants.
// No - No mutual intelligibility.
//
// The following levels can occur for each type of subtag:
// Base: Exact, MaxExact, High, Low, No
// Script: Exact, MaxExact [see Note 3], Low, No
// Region: Exact, MaxExact, High
// Variant: Exact, High
// Private: Exact, No
//
// Any result with a confidence level of Low or higher is deemed a possible match.
// Once a desired tag matches any of the supported tags with a level of MaxExact
// or higher, the next desired tag is not considered (see Step 2.b).
// Note that CLDR provides languageMatching data that defines close equivalence
// classes for base languages, scripts and regions.
//
// Tie-breaking
// If we get the same confidence level for two matches, we apply a sequence of
// tie-breaking rules. The first that succeeds defines the result. The rules are
// applied in the following order.
// 1) Original language was defined and was identical.
// 2) Original region was defined and was identical.
// 3) Distance between two maximized regions was the smallest.
// 4) Original script was defined and was identical.
// 5) Distance from want tag to have tag using the parent relation [see Note 5.]
// If there is still no winner after these rules are applied, the first match
// found wins.
//
// Notes:
// [2] In practice, as matching of Exact is done in a separate phase from
// matching the other levels, we reuse the Exact level to mean MaxExact in
// the second phase. As a consequence, we only need the levels defined by
// the Confidence type. The MaxExact confidence level is mapped to High in
// the public API.
// [3] We do not differentiate between maximized script values that were derived
// from suppressScript versus most likely tag data. We determined that in
// ranking the two, one ranks just after the other. Moreover, the two cannot
// occur concurrently. As a consequence, they are identical for practical
// purposes.
// [4] In case of deprecated, macro-equivalents and legacy mappings, we assign
// the MaxExact level to allow iw vs he to still be a closer match than
// en-AU vs en-US, for example.
// [5] In CLDR a locale inherits fields that are unspecified for this locale
// from its parent. Therefore, if a locale is a parent of another locale,
// it is a strong measure for closeness, especially when no other tie
// breaker rule applies. One could also argue it is inconsistent, for
// example, when pt-AO matches pt (which CLDR equates with pt-BR), even
// though its parent is pt-PT according to the inheritance rules.
//
// Implementation Details:
// There are several performance considerations worth pointing out. Most notably,
// we preprocess as much as possible (within reason) at the time of creation of a
// matcher. This includes:
// - creating a per-language map, which includes data for the raw base language
// and its canonicalized variant (if applicable),
// - expanding entries for the equivalence classes defined in CLDR's
// languageMatch data.
// The per-language map ensures that typically only a very small number of tags
// need to be considered. The pre-expansion of canonicalized subtags and
// equivalence classes reduces the amount of map lookups that need to be done at
// runtime.
// matcher keeps a set of supported language tags, indexed by language.
type matcher struct {
default_ *haveTag
supported []*haveTag
index map[language.Language]*matchHeader
passSettings bool
preferSameScript bool
}
// matchHeader has the lists of tags for exact matches and matches based on
// maximized and canonicalized tags for a given language.
type matchHeader struct {
haveTags []*haveTag
original bool
}
// haveTag holds a supported Tag and its maximized script and region. The maximized
// or canonicalized language is not stored as it is not needed during matching.
type haveTag struct {
tag language.Tag
// index of this tag in the original list of supported tags.
index int
// conf is the maximum confidence that can result from matching this haveTag.
// When conf < Exact this means it was inserted after applying a CLDR equivalence rule.
conf Confidence
// Maximized region and script.
maxRegion language.Region
maxScript language.Script
// altScript may be checked as an alternative match to maxScript. If altScript
// matches, the confidence level for this match is Low. Theoretically there
// could be multiple alternative scripts. This does not occur in practice.
altScript language.Script
// nextMax is the index of the next haveTag with the same maximized tags.
nextMax uint16
}
func makeHaveTag(tag language.Tag, index int) (haveTag, language.Language) {
max := tag
if tag.LangID != 0 || tag.RegionID != 0 || tag.ScriptID != 0 {
max, _ = canonicalize(All, max)
max, _ = max.Maximize()
max.RemakeString()
}
return haveTag{tag, index, Exact, max.RegionID, max.ScriptID, altScript(max.LangID, max.ScriptID), 0}, max.LangID
}
// altScript returns an alternative script that may match the given script with
// a low confidence. At the moment, the langMatch data allows for at most one
// script to map to another and we rely on this to keep the code simple.
func | (l language.Language, s language.Script) language.Script {
for _, alt := range matchScript {
// TODO: also match cases where language is not the same.
if (language.Language(alt.wantLang) == l || language.Language(alt.haveLang) == l) &&
language.Script(alt.haveScript) == s {
return language.Script(alt.wantScript)
}
}
return 0
}
// addIfNew adds a haveTag to the list of tags only if it is a unique tag.
// Tags that have the same maximized values are linked by index.
func (h *matchHeader) addIfNew(n haveTag, exact bool) {
h.original = h.original || exact
// Don't add new exact matches.
for _, v := range h.haveTags {
if equalsRest(v.tag, n.tag) {
return
}
}
// Allow duplicate maximized tags, but create a linked list to allow quickly
// comparing the equivalents and bail out.
for i, v := range h.haveTags {
if v.maxScript == n.maxScript &&
v.maxRegion == n.maxRegion &&
v.tag.VariantOrPrivateUseTags() == n.tag.VariantOrPrivateUseTags() {
for h.haveTags[i].nextMax != 0 {
i = int(h.haveTags[i].nextMax)
}
h.haveTags[i].nextMax = uint16(len(h.haveTags))
break
}
}
h.haveTags = append(h.haveTags, &n)
}
// header returns the matchHeader for the given language. It creates one if
// it doesn't already exist.
func (m *matcher) header(l language.Language) *matchHeader {
if h := m.index[l]; h != nil {
return h
}
h := &matchHeader{}
m.index[l] = h
return h
}
func toConf(d uint8) Confidence {
if d <= | altScript | identifier_name |
match.go | n.tag.VariantOrPrivateUseTags() {
for h.haveTags[i].nextMax != 0 {
i = int(h.haveTags[i].nextMax)
}
h.haveTags[i].nextMax = uint16(len(h.haveTags))
break
}
}
h.haveTags = append(h.haveTags, &n)
}
// header returns the matchHeader for the given language. It creates one if
// it doesn't already exist.
func (m *matcher) header(l language.Language) *matchHeader {
if h := m.index[l]; h != nil {
return h
}
h := &matchHeader{}
m.index[l] = h
return h
}
func toConf(d uint8) Confidence {
if d <= 10 {
return High
}
if d < 30 {
return Low
}
return No
}
// newMatcher builds an index for the given supported tags and returns it as
// a matcher. It also expands the index by considering various equivalence classes
// for a given tag.
func newMatcher(supported []Tag, options []MatchOption) *matcher {
m := &matcher{
index: make(map[language.Language]*matchHeader),
preferSameScript: true,
}
for _, o := range options {
o(m)
}
if len(supported) == 0 {
m.default_ = &haveTag{}
return m
}
// Add supported languages to the index. Add exact matches first to give
// them precedence.
for i, tag := range supported {
tt := tag.tag()
pair, _ := makeHaveTag(tt, i)
m.header(tt.LangID).addIfNew(pair, true)
m.supported = append(m.supported, &pair)
}
m.default_ = m.header(supported[0].lang()).haveTags[0]
// Keep these in two different loops to support the case that two equivalent
// languages are distinguished, such as iw and he.
for i, tag := range supported {
tt := tag.tag()
pair, max := makeHaveTag(tt, i)
if max != tt.LangID {
m.header(max).addIfNew(pair, true)
}
}
// update is used to add indexes in the map for equivalent languages.
// update will only add entries to original indexes, thus not computing any
// transitive relations.
update := func(want, have uint16, conf Confidence) {
if hh := m.index[language.Language(have)]; hh != nil {
if !hh.original {
return
}
hw := m.header(language.Language(want))
for _, ht := range hh.haveTags {
v := *ht
if conf < v.conf {
v.conf = conf
}
v.nextMax = 0 // this value needs to be recomputed
if v.altScript != 0 {
v.altScript = altScript(language.Language(want), v.maxScript)
}
hw.addIfNew(v, conf == Exact && hh.original)
}
}
}
// Add entries for languages with mutual intelligibility as defined by CLDR's
// languageMatch data.
for _, ml := range matchLang {
update(ml.want, ml.have, toConf(ml.distance))
if !ml.oneway {
update(ml.have, ml.want, toConf(ml.distance))
}
}
// Add entries for possible canonicalizations. This is an optimization to
// ensure that only one map lookup needs to be done at runtime per desired tag.
// First we match deprecated equivalents. If they are perfect equivalents
// (their canonicalization simply substitutes a different language code, but
// nothing else), the match confidence is Exact, otherwise it is High.
for i, lm := range language.AliasMap {
// If deprecated codes match and there is no fiddling with the script or
// or region, we consider it an exact match.
conf := Exact
if language.AliasTypes[i] != language.Macro {
if !isExactEquivalent(language.Language(lm.From)) {
conf = High
}
update(lm.To, lm.From, conf)
}
update(lm.From, lm.To, conf)
}
return m
}
// getBest gets the best matching tag in m for any of the given tags, taking into
// account the order of preference of the given tags.
func (m *matcher) getBest(want ...Tag) (got *haveTag, orig language.Tag, c Confidence) {
best := bestMatch{}
for i, ww := range want {
w := ww.tag()
var max language.Tag
// Check for exact match first.
h := m.index[w.LangID]
if w.LangID != 0 {
if h == nil {
continue
}
// Base language is defined.
max, _ = canonicalize(Legacy|Deprecated|Macro, w)
// A region that is added through canonicalization is stronger than
// a maximized region: set it in the original (e.g. mo -> ro-MD).
if w.RegionID != max.RegionID {
w.RegionID = max.RegionID
}
// TODO: should we do the same for scripts?
// See test case: en, sr, nl ; sh ; sr
max, _ = max.Maximize()
} else {
// Base language is not defined.
if h != nil {
for i := range h.haveTags {
have := h.haveTags[i]
if equalsRest(have.tag, w) {
return have, w, Exact
}
}
}
if w.ScriptID == 0 && w.RegionID == 0 {
// We skip all tags matching und for approximate matching, including
// private tags.
continue
}
max, _ = w.Maximize()
if h = m.index[max.LangID]; h == nil {
continue
}
}
pin := true
for _, t := range want[i+1:] {
if w.LangID == t.lang() {
pin = false
break
}
}
// Check for match based on maximized tag.
for i := range h.haveTags {
have := h.haveTags[i]
best.update(have, w, max.ScriptID, max.RegionID, pin)
if best.conf == Exact {
for have.nextMax != 0 {
have = h.haveTags[have.nextMax]
best.update(have, w, max.ScriptID, max.RegionID, pin)
}
return best.have, best.want, best.conf
}
}
}
if best.conf <= No {
if len(want) != 0 {
return nil, want[0].tag(), No
}
return nil, language.Tag{}, No
}
return best.have, best.want, best.conf
}
// bestMatch accumulates the best match so far.
type bestMatch struct {
have *haveTag
want language.Tag
conf Confidence
pinnedRegion language.Region
pinLanguage bool
sameRegionGroup bool
// Cached results from applying tie-breaking rules.
origLang bool
origReg bool
paradigmReg bool
regGroupDist uint8
origScript bool
}
// update updates the existing best match if the new pair is considered to be a
// better match. To determine if the given pair is a better match, it first
// computes the rough confidence level. If this surpasses the current match, it
// will replace it and update the tie-breaker rule cache. If there is a tie, it
// proceeds with applying a series of tie-breaker rules. If there is no
// conclusive winner after applying the tie-breaker rules, it leaves the current
// match as the preferred match.
//
// If pin is true and have and tag are a strong match, it will henceforth only
// consider matches for this language. This corresponds to the idea that most
// users have a strong preference for the first defined language. A user can
// still prefer a second language over a dialect of the preferred language by
// explicitly specifying dialects, e.g. "en, nl, en-GB". In this case pin should
// be false.
func (m *bestMatch) update(have *haveTag, tag language.Tag, maxScript language.Script, maxRegion language.Region, pin bool) | {
// Bail if the maximum attainable confidence is below that of the current best match.
c := have.conf
if c < m.conf {
return
}
// Don't change the language once we already have found an exact match.
if m.pinLanguage && tag.LangID != m.want.LangID {
return
}
// Pin the region group if we are comparing tags for the same language.
if tag.LangID == m.want.LangID && m.sameRegionGroup {
_, sameGroup := regionGroupDist(m.pinnedRegion, have.maxRegion, have.maxScript, m.want.LangID)
if !sameGroup {
return
}
}
if c == Exact && have.maxScript == maxScript {
// If there is another language and then another entry of this language,
// don't pin anything, otherwise pin the language. | identifier_body | |
match.go |
// A MatchOption configures a Matcher.
type MatchOption func(*matcher)
// PreferSameScript will, in the absence of a match, result in the first
// preferred tag with the same script as a supported tag to match this supported
// tag. The default is currently true, but this may change in the future.
func PreferSameScript(preferSame bool) MatchOption {
return func(m *matcher) { m.preferSameScript = preferSame }
}
// TODO(v1.0.0): consider making Matcher a concrete type, instead of interface.
// There doesn't seem to be too much need for multiple types.
// Making it a concrete type allows MatchStrings to be a method, which will
// improve its discoverability.
// MatchStrings parses and matches the given strings until one of them matches
// the language in the Matcher. A string may be an Accept-Language header as
// handled by ParseAcceptLanguage. The default language is returned if no
// other language matched.
func MatchStrings(m Matcher, lang ...string) (tag Tag, index int) {
for _, accept := range lang {
desired, _, err := ParseAcceptLanguage(accept)
if err != nil {
continue
}
if tag, index, conf := m.Match(desired...); conf != No {
return tag, index
}
}
tag, index, _ = m.Match()
return
}
// Matcher is the interface that wraps the Match method.
//
// Match returns the best match for any of the given tags, along with
// a unique index associated with the returned tag and a confidence
// score.
type Matcher interface {
Match(t ...Tag) (tag Tag, index int, c Confidence)
}
// Comprehends reports the confidence score for a speaker of a given language
// to being able to comprehend the written form of an alternative language.
func Comprehends(speaker, alternative Tag) Confidence {
_, _, c := NewMatcher([]Tag{alternative}).Match(speaker)
return c
}
// NewMatcher returns a Matcher that matches an ordered list of preferred tags
// against a list of supported tags based on written intelligibility, closeness
// of dialect, equivalence of subtags and various other rules. It is initialized
// with the list of supported tags. The first element is used as the default
// value in case no match is found.
//
// Its Match method matches the first of the given Tags to reach a certain
// confidence threshold. The tags passed to Match should therefore be specified
// in order of preference. Extensions are ignored for matching.
//
// The index returned by the Match method corresponds to the index of the
// matched tag in t, but is augmented with the Unicode extension ('u')of the
// corresponding preferred tag. This allows user locale options to be passed
// transparently.
func NewMatcher(t []Tag, options ...MatchOption) Matcher {
return newMatcher(t, options)
}
func (m *matcher) Match(want ...Tag) (t Tag, index int, c Confidence) {
var tt language.Tag
match, w, c := m.getBest(want...)
if match != nil {
tt, index = match.tag, match.index
} else {
// TODO: this should be an option
tt = m.default_.tag
if m.preferSameScript {
outer:
for _, w := range want {
script, _ := w.Script()
if script.scriptID == 0 {
// Don't do anything if there is no script, such as with
// private subtags.
continue
}
for i, h := range m.supported {
if script.scriptID == h.maxScript {
tt, index = h.tag, i
break outer
}
}
}
}
// TODO: select first language tag based on script.
}
if w.RegionID != tt.RegionID && w.RegionID != 0 {
if w.RegionID != 0 && tt.RegionID != 0 && tt.RegionID.Contains(w.RegionID) {
tt.RegionID = w.RegionID
tt.RemakeString()
} else if r := w.RegionID.String(); len(r) == 2 {
// TODO: also filter macro and deprecated.
tt, _ = tt.SetTypeForKey("rg", strings.ToLower(r)+"zzzz")
}
}
// Copy options from the user-provided tag into the result tag. This is hard
// to do after the fact, so we do it here.
// TODO: add in alternative variants to -u-va-.
// TODO: add preferred region to -u-rg-.
if e := w.Extensions(); len(e) > 0 {
b := language.Builder{}
b.SetTag(tt)
for _, e := range e {
b.AddExt(e)
}
tt = b.Make()
}
return makeTag(tt), index, c
}
// ErrMissingLikelyTagsData indicates no information was available
// to compute likely values of missing tags.
var ErrMissingLikelyTagsData = errors.New("missing likely tags data")
// func (t *Tag) setTagsFrom(id Tag) {
// t.LangID = id.LangID
// t.ScriptID = id.ScriptID
// t.RegionID = id.RegionID
// }
// Tag Matching
// CLDR defines an algorithm for finding the best match between two sets of language
// tags. The basic algorithm defines how to score a possible match and then find
// the match with the best score
// (see https://www.unicode.org/reports/tr35/#LanguageMatching).
// Using scoring has several disadvantages. The scoring obfuscates the importance of
// the various factors considered, making the algorithm harder to understand. Using
// scoring also requires the full score to be computed for each pair of tags.
//
// We will use a different algorithm which aims to have the following properties:
// - clarity on the precedence of the various selection factors, and
// - improved performance by allowing early termination of a comparison.
//
// Matching algorithm (overview)
// Input:
// - supported: a set of supported tags
// - default: the default tag to return in case there is no match
// - desired: list of desired tags, ordered by preference, starting with
// the most-preferred.
//
// Algorithm:
// 1) Set the best match to the lowest confidence level
// 2) For each tag in "desired":
// a) For each tag in "supported":
// 1) compute the match between the two tags.
// 2) if the match is better than the previous best match, replace it
// with the new match. (see next section)
// b) if the current best match is Exact and pin is true the result will be
// frozen to the language found thusfar, although better matches may
// still be found for the same language.
// 3) If the best match so far is below a certain threshold, return "default".
//
// Ranking:
// We use two phases to determine whether one pair of tags are a better match
// than another pair of tags. First, we determine a rough confidence level. If the
// levels are different, the one with the highest confidence wins.
// Second, if the rough confidence levels are identical, we use a set of tie-breaker
// rules.
//
// The confidence level of matching a pair of tags is determined by finding the
// lowest confidence level of any matches of the corresponding subtags (the
// result is deemed as good as its weakest link).
// We define the following levels:
// Exact - An exact match of a subtag, before adding likely subtags.
// MaxExact - An exact match of a subtag, after adding likely subtags.
// [See Note 2].
// High - High level of mutual intelligibility between different subtag
// variants.
// Low - Low level of mutual intelligibility between different subtag
// variants.
// No - No mutual intelligibility.
//
// The following levels can occur for each type of subtag:
// Base: Exact, MaxExact, High, Low, No
// Script: Exact, MaxExact [see Note 3], Low, No
// Region: Exact, MaxExact, High
// Variant: Exact, High
// Private: Exact, No
//
// Any result with a confidence level of Low or higher is deemed a possible match.
// Once a desired tag matches any of the supported tags with a level of MaxExact
// or higher, the next desired tag is not considered (see Step 2.b).
// Note that CLDR provides languageMatching data that defines close equivalence
// classes for base languages, scripts and regions.
//
// Tie-breaking
// If we get the same confidence level for two matches, we apply a sequence of
// tie-breaking rules. The first that succeeds defines the result. The rules are
// applied in the following order.
// 1) Original language was defined and was identical.
// 2) Original region was defined and was identical.
// 3) Distance between two maximized regions was the smallest.
// 4) Original script was defined and was identical.
// 5) Distance from want tag to have tag using the parent relation [see Note 5.]
// If there is still no winner after these rules are applied, the first match
// found wins.
| "golang.org/x/text/internal/language"
) | random_line_split | |
match.go | m *matcher) header(l language.Language) *matchHeader {
if h := m.index[l]; h != nil {
return h
}
h := &matchHeader{}
m.index[l] = h
return h
}
func toConf(d uint8) Confidence {
if d <= 10 {
return High
}
if d < 30 {
return Low
}
return No
}
// newMatcher builds an index for the given supported tags and returns it as
// a matcher. It also expands the index by considering various equivalence classes
// for a given tag.
func newMatcher(supported []Tag, options []MatchOption) *matcher {
m := &matcher{
index: make(map[language.Language]*matchHeader),
preferSameScript: true,
}
for _, o := range options {
o(m)
}
if len(supported) == 0 {
m.default_ = &haveTag{}
return m
}
// Add supported languages to the index. Add exact matches first to give
// them precedence.
for i, tag := range supported {
tt := tag.tag()
pair, _ := makeHaveTag(tt, i)
m.header(tt.LangID).addIfNew(pair, true)
m.supported = append(m.supported, &pair)
}
m.default_ = m.header(supported[0].lang()).haveTags[0]
// Keep these in two different loops to support the case that two equivalent
// languages are distinguished, such as iw and he.
for i, tag := range supported {
tt := tag.tag()
pair, max := makeHaveTag(tt, i)
if max != tt.LangID {
m.header(max).addIfNew(pair, true)
}
}
// update is used to add indexes in the map for equivalent languages.
// update will only add entries to original indexes, thus not computing any
// transitive relations.
update := func(want, have uint16, conf Confidence) {
if hh := m.index[language.Language(have)]; hh != nil {
if !hh.original {
return
}
hw := m.header(language.Language(want))
for _, ht := range hh.haveTags {
v := *ht
if conf < v.conf {
v.conf = conf
}
v.nextMax = 0 // this value needs to be recomputed
if v.altScript != 0 {
v.altScript = altScript(language.Language(want), v.maxScript)
}
hw.addIfNew(v, conf == Exact && hh.original)
}
}
}
// Add entries for languages with mutual intelligibility as defined by CLDR's
// languageMatch data.
for _, ml := range matchLang {
update(ml.want, ml.have, toConf(ml.distance))
if !ml.oneway {
update(ml.have, ml.want, toConf(ml.distance))
}
}
// Add entries for possible canonicalizations. This is an optimization to
// ensure that only one map lookup needs to be done at runtime per desired tag.
// First we match deprecated equivalents. If they are perfect equivalents
// (their canonicalization simply substitutes a different language code, but
// nothing else), the match confidence is Exact, otherwise it is High.
for i, lm := range language.AliasMap {
// If deprecated codes match and there is no fiddling with the script or
// or region, we consider it an exact match.
conf := Exact
if language.AliasTypes[i] != language.Macro {
if !isExactEquivalent(language.Language(lm.From)) {
conf = High
}
update(lm.To, lm.From, conf)
}
update(lm.From, lm.To, conf)
}
return m
}
// getBest gets the best matching tag in m for any of the given tags, taking into
// account the order of preference of the given tags.
func (m *matcher) getBest(want ...Tag) (got *haveTag, orig language.Tag, c Confidence) {
best := bestMatch{}
for i, ww := range want {
w := ww.tag()
var max language.Tag
// Check for exact match first.
h := m.index[w.LangID]
if w.LangID != 0 {
if h == nil {
continue
}
// Base language is defined.
max, _ = canonicalize(Legacy|Deprecated|Macro, w)
// A region that is added through canonicalization is stronger than
// a maximized region: set it in the original (e.g. mo -> ro-MD).
if w.RegionID != max.RegionID {
w.RegionID = max.RegionID
}
// TODO: should we do the same for scripts?
// See test case: en, sr, nl ; sh ; sr
max, _ = max.Maximize()
} else {
// Base language is not defined.
if h != nil {
for i := range h.haveTags {
have := h.haveTags[i]
if equalsRest(have.tag, w) {
return have, w, Exact
}
}
}
if w.ScriptID == 0 && w.RegionID == 0 {
// We skip all tags matching und for approximate matching, including
// private tags.
continue
}
max, _ = w.Maximize()
if h = m.index[max.LangID]; h == nil {
continue
}
}
pin := true
for _, t := range want[i+1:] {
if w.LangID == t.lang() {
pin = false
break
}
}
// Check for match based on maximized tag.
for i := range h.haveTags {
have := h.haveTags[i]
best.update(have, w, max.ScriptID, max.RegionID, pin)
if best.conf == Exact {
for have.nextMax != 0 {
have = h.haveTags[have.nextMax]
best.update(have, w, max.ScriptID, max.RegionID, pin)
}
return best.have, best.want, best.conf
}
}
}
if best.conf <= No {
if len(want) != 0 {
return nil, want[0].tag(), No
}
return nil, language.Tag{}, No
}
return best.have, best.want, best.conf
}
// bestMatch accumulates the best match so far.
type bestMatch struct {
have *haveTag
want language.Tag
conf Confidence
pinnedRegion language.Region
pinLanguage bool
sameRegionGroup bool
// Cached results from applying tie-breaking rules.
origLang bool
origReg bool
paradigmReg bool
regGroupDist uint8
origScript bool
}
// update updates the existing best match if the new pair is considered to be a
// better match. To determine if the given pair is a better match, it first
// computes the rough confidence level. If this surpasses the current match, it
// will replace it and update the tie-breaker rule cache. If there is a tie, it
// proceeds with applying a series of tie-breaker rules. If there is no
// conclusive winner after applying the tie-breaker rules, it leaves the current
// match as the preferred match.
//
// If pin is true and have and tag are a strong match, it will henceforth only
// consider matches for this language. This corresponds to the idea that most
// users have a strong preference for the first defined language. A user can
// still prefer a second language over a dialect of the preferred language by
// explicitly specifying dialects, e.g. "en, nl, en-GB". In this case pin should
// be false.
func (m *bestMatch) update(have *haveTag, tag language.Tag, maxScript language.Script, maxRegion language.Region, pin bool) {
// Bail if the maximum attainable confidence is below that of the current best match.
c := have.conf
if c < m.conf {
return
}
// Don't change the language once we already have found an exact match.
if m.pinLanguage && tag.LangID != m.want.LangID {
return
}
// Pin the region group if we are comparing tags for the same language.
if tag.LangID == m.want.LangID && m.sameRegionGroup {
_, sameGroup := regionGroupDist(m.pinnedRegion, have.maxRegion, have.maxScript, m.want.LangID)
if !sameGroup {
return
}
}
if c == Exact && have.maxScript == maxScript {
// If there is another language and then another entry of this language,
// don't pin anything, otherwise pin the language.
m.pinLanguage = pin
}
if equalsRest(have.tag, tag) {
} else if have.maxScript != maxScript | {
// There is usually very little comprehension between different scripts.
// In a few cases there may still be Low comprehension. This possibility
// is pre-computed and stored in have.altScript.
if Low < m.conf || have.altScript != maxScript {
return
}
c = Low
} | conditional_block | |
lane_detector.py | (self, img):
''' Apply polynomial fit to the given image, returning fit for left/right lanes
Called when one frame of image has previously found left_fit/right_fit.
This method attempts to find lane fits in the vicinity of previous fits
:param img -- input image with lane lines
:return left_fit, right_fit
'''
if self.left_fit is None or self.right_fit is None:
return self.sliding_window_fit(img)
# from the next frame of video (also called "binary_warped")
# It's now much easier to find line pixels!
nonzero = img.nonzero()
nonzeroy = np.array(nonzero[0])
nonzerox = np.array(nonzero[1])
margin = 100
left_lane_inds = ((nonzerox > (self.left_fit[0]*(nonzeroy**2) + self.left_fit[1]*nonzeroy + self.left_fit[2] - margin)) &
(nonzerox < (self.left_fit[0]*(nonzeroy**2) + self.left_fit[1]*nonzeroy + self.left_fit[2] + margin)))
right_lane_inds = ((nonzerox > (self.right_fit[0]*(nonzeroy**2) + self.right_fit[1]*nonzeroy + self.right_fit[2] - margin)) &
(nonzerox < (self.right_fit[0]*(nonzeroy**2) + self.right_fit[1]*nonzeroy + self.right_fit[2] + margin)))
# Again, extract left and right line pixel positions
self.leftx = nonzerox[left_lane_inds]
lefty = nonzeroy[left_lane_inds]
self.rightx = nonzerox[right_lane_inds]
righty = nonzeroy[right_lane_inds]
# Fit a second order polynomial to each
self.left_fit = np.polyfit(lefty, self.leftx, 2)
self.right_fit = np.polyfit(righty, self.rightx, 2)
return self.left_fit, self.right_fit
def sliding_window_fit(self, img):
''' Apply sliding windows search to the given image to find polynomial to find lane lines
Code based largely on Udacity lessons
:param img - given image
:return left_fit, right_fit - polynomials fitting the left/right lane lines
'''
y_half = int(img.shape[0]/2)
# take histogram of bottom half of img
histogram = np.sum(img[y_half:, :], axis=0)
# Create an output image to draw on and visualize the result
out_img = np.dstack((img, img, img))*255
# Find the peak of the left and right halves of the histogram
# These will be the starting point for the left and right lines
midpoint = np.int(histogram.shape[0]/2)
leftx_base = np.argmax(histogram[:midpoint])
rightx_base = np.argmax(histogram[midpoint:]) + midpoint
# Choose the number of sliding windows
nwindows = 9
# Set height of windows
window_height = np.int(img.shape[0]/nwindows)
# Identify the x and y positions of all nonzero pixels in the image
nonzero = img.nonzero()
nonzeroy = np.array(nonzero[0])
nonzerox = np.array(nonzero[1])
# Current positions to be updated for each window
leftx_current = leftx_base
rightx_current = rightx_base
# Set the width of the windows +/- margin
margin = 100
# Set minimum number of pixels found to recenter window
minpix = 50
# Create empty lists to receive left and right lane pixel indices
left_lane_inds = []
right_lane_inds = []
for window in range(nwindows):
# Identify window boundaries in x and y (and right and left)
win_y_low = img.shape[0] - (window+1) * window_height
win_y_high = img.shape[0] - window * window_height
win_xleft_low = leftx_current - margin
win_xleft_high = leftx_current + margin
win_xright_low = rightx_current - margin
win_xright_high = rightx_current + margin
# Draw the windows on the visualization image
cv2.rectangle(out_img,(win_xleft_low,win_y_low), (win_xleft_high,win_y_high), (0,255,0), 2)
cv2.rectangle(out_img,(win_xright_low,win_y_low),(win_xright_high,win_y_high),(0,255,0), 2)
# Identify the nonzero pixels in x and y within the window
good_left_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xleft_low) & (nonzerox < win_xleft_high)).nonzero()[0]
good_right_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xright_low) & (nonzerox < win_xright_high)).nonzero()[0]
# Append these indices to the lists
left_lane_inds.append(good_left_inds)
right_lane_inds.append(good_right_inds)
# If you found > minpix pixels, recenter next window on their mean position
if len(good_left_inds) > minpix:
leftx_current = np.int( np.mean(nonzerox[good_left_inds]) )
if len(good_right_inds) > minpix:
rightx_current = np.int( np.mean(nonzerox[good_right_inds]) )
# Concatenate the arrays of indices
left_lane_inds = np.concatenate(left_lane_inds)
right_lane_inds = np.concatenate(right_lane_inds)
# Extract left and right line pixel positions
self.leftx = nonzerox[left_lane_inds]
lefty = nonzeroy[left_lane_inds]
righty = nonzeroy[right_lane_inds]
self.rightx = nonzerox[right_lane_inds]
# Fit a second order polynomial to each
self.left_fit = np.polyfit(lefty, self.leftx, 2)
self.right_fit = np.polyfit(righty,self.rightx, 2)
return self.left_fit, self.right_fit
def find_lane_curvature(self, img):
''' Find lane curvature for the given img
:param img - the input image
:return lane curvature
'''
# Generate some fake data to represent lane-line pixels
ploty = np.linspace(0, 719, num=720) # to cover same y-range as image
quadratic_coeff = 3e-4 # arbitrary quadratic coefficient
# For each y position generate random x position within +/-50 pix
# of the line base position in each case (x=200 for left, and x=900 for right)
leftx = np.array([200 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
rightx = np.array([900 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
leftx = leftx[::-1] # Reverse to match top-to-bottom in y
rightx = rightx[::-1] # Reverse to match top-to-bottom in y
# Fit a second order polynomial to pixel positions in each fake lane line
# left_fit = np.polyfit(ploty, leftx, 2)
# left_fitx = left_fit[0]*ploty**2 + left_fit[1]*ploty + left_fit[2]
# right_fit = np.polyfit(ploty, rightx, 2)
# right_fitx = right_fit[0]*ploty**2 + right_fit[1]*ploty + right_fit[2]
# Define y-value where we want radius of curvature
# I'll choose the maximum y-value, corresponding to the bottom of the image
y_eval = np.max(ploty)
# left_curverad = ((1 + (2*left_fit[0]*y_eval + left_fit[1])**2)**1.5) / np.absolute(2*left_fit[0])
# right_curverad = ((1 + (2*right_fit[0]*y_eval + right_fit[1])**2)**1.5) / np.absolute(2*right_fit[0])
# print(left_curverad, right_curverad)
# Example values: 1926.74 1908.48
# Define conversions in x and y from pixels space to meters
ym_per_pix = 30/720 # meters per pixel in y dimension
xm_per_pix = 3.7/700 # meters per pixel in x dimension
# Fit new polynomials to x,y in world space
left_fit_cr = np.polyfit(ploty * ym_per_pix, leftx * xm_per_pix, | window_fit | identifier_name | |
lane_detector.py |
# from the next frame of video (also called "binary_warped")
# It's now much easier to find line pixels!
nonzero = img.nonzero()
nonzeroy = np.array(nonzero[0])
nonzerox = np.array(nonzero[1])
margin = 100
left_lane_inds = ((nonzerox > (self.left_fit[0]*(nonzeroy**2) + self.left_fit[1]*nonzeroy + self.left_fit[2] - margin)) &
(nonzerox < (self.left_fit[0]*(nonzeroy**2) + self.left_fit[1]*nonzeroy + self.left_fit[2] + margin)))
right_lane_inds = ((nonzerox > (self.right_fit[0]*(nonzeroy**2) + self.right_fit[1]*nonzeroy + self.right_fit[2] - margin)) &
(nonzerox < (self.right_fit[0]*(nonzeroy**2) + self.right_fit[1]*nonzeroy + self.right_fit[2] + margin)))
# Again, extract left and right line pixel positions
self.leftx = nonzerox[left_lane_inds]
lefty = nonzeroy[left_lane_inds]
self.rightx = nonzerox[right_lane_inds]
righty = nonzeroy[right_lane_inds]
# Fit a second order polynomial to each
self.left_fit = np.polyfit(lefty, self.leftx, 2)
self.right_fit = np.polyfit(righty, self.rightx, 2)
return self.left_fit, self.right_fit
def sliding_window_fit(self, img):
''' Apply sliding windows search to the given image to find polynomial to find lane lines
Code based largely on Udacity lessons
:param img - given image
:return left_fit, right_fit - polynomials fitting the left/right lane lines
'''
y_half = int(img.shape[0]/2)
# take histogram of bottom half of img
histogram = np.sum(img[y_half:, :], axis=0)
# Create an output image to draw on and visualize the result
out_img = np.dstack((img, img, img))*255
# Find the peak of the left and right halves of the histogram
# These will be the starting point for the left and right lines
midpoint = np.int(histogram.shape[0]/2)
leftx_base = np.argmax(histogram[:midpoint])
rightx_base = np.argmax(histogram[midpoint:]) + midpoint
# Choose the number of sliding windows
nwindows = 9
# Set height of windows
window_height = np.int(img.shape[0]/nwindows)
# Identify the x and y positions of all nonzero pixels in the image
nonzero = img.nonzero()
nonzeroy = np.array(nonzero[0])
nonzerox = np.array(nonzero[1])
# Current positions to be updated for each window
leftx_current = leftx_base
rightx_current = rightx_base
# Set the width of the windows +/- margin
margin = 100
# Set minimum number of pixels found to recenter window
minpix = 50
# Create empty lists to receive left and right lane pixel indices
left_lane_inds = []
right_lane_inds = []
for window in range(nwindows):
# Identify window boundaries in x and y (and right and left)
win_y_low = img.shape[0] - (window+1) * window_height
win_y_high = img.shape[0] - window * window_height
win_xleft_low = leftx_current - margin
win_xleft_high = leftx_current + margin
win_xright_low = rightx_current - margin
win_xright_high = rightx_current + margin
# Draw the windows on the visualization image
cv2.rectangle(out_img,(win_xleft_low,win_y_low), (win_xleft_high,win_y_high), (0,255,0), 2)
cv2.rectangle(out_img,(win_xright_low,win_y_low),(win_xright_high,win_y_high),(0,255,0), 2)
# Identify the nonzero pixels in x and y within the window
good_left_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xleft_low) & (nonzerox < win_xleft_high)).nonzero()[0]
good_right_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xright_low) & (nonzerox < win_xright_high)).nonzero()[0]
# Append these indices to the lists
left_lane_inds.append(good_left_inds)
right_lane_inds.append(good_right_inds)
# If you found > minpix pixels, recenter next window on their mean position
if len(good_left_inds) > minpix:
leftx_current = np.int( np.mean(nonzerox[good_left_inds]) )
if len(good_right_inds) > minpix:
rightx_current = np.int( np.mean(nonzerox[good_right_inds]) )
# Concatenate the arrays of indices
left_lane_inds = np.concatenate(left_lane_inds)
right_lane_inds = np.concatenate(right_lane_inds)
# Extract left and right line pixel positions
self.leftx = nonzerox[left_lane_inds]
lefty = nonzeroy[left_lane_inds]
righty = nonzeroy[right_lane_inds]
self.rightx = nonzerox[right_lane_inds]
# Fit a second order polynomial to each
self.left_fit = np.polyfit(lefty, self.leftx, 2)
self.right_fit = np.polyfit(righty,self.rightx, 2)
return self.left_fit, self.right_fit
def find_lane_curvature(self, img):
''' Find lane curvature for the given img
:param img - the input image
:return lane curvature
'''
# Generate some fake data to represent lane-line pixels
ploty = np.linspace(0, 719, num=720) # to cover same y-range as image
quadratic_coeff = 3e-4 # arbitrary quadratic coefficient
# For each y position generate random x position within +/-50 pix
# of the line base position in each case (x=200 for left, and x=900 for right)
leftx = np.array([200 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
rightx = np.array([900 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
leftx = leftx[::-1] # Reverse to match top-to-bottom in y
rightx = rightx[::-1] # Reverse to match top-to-bottom in y
# Fit a second order polynomial to pixel positions in each fake lane line
# left_fit = np.polyfit(ploty, leftx, 2)
# left_fitx = left_fit[0]*ploty**2 + left_fit[1]*ploty + left_fit[2]
# right_fit = np.polyfit(ploty, rightx, 2)
# right_fitx = right_fit[0]*ploty**2 + right_fit[1]*ploty + right_fit[2]
# Define y-value where we want radius of curvature
# I'll choose the maximum y-value, corresponding to the bottom of the image
y_eval = np.max(ploty)
# left_curverad = ((1 + (2*left_fit[0]*y_eval + left_fit[1])**2)**1.5) / np.absolute(2*left_fit[0])
# right_curverad = ((1 + (2*right_fit[0]*y_eval + right_fit[1])**2)**1.5) / np.absolute(2*right_fit[0])
# print(left_curverad, right_curverad)
# Example values: 1926.74 1908.48
# Define conversions in x and y from pixels space to meters
ym_per_pix = 30/720 # meters per pixel in y dimension
xm_per_pix = 3.7/700 # meters per pixel in x dimension
# Fit new polynomials to x,y in world space
left_fit_cr = np.polyfit(ploty * ym_per_pix, leftx * xm_per_pix, 2)
right_fit_cr = np.polyfit(ploty * ym_per_pix, rightx * xm_per_pix, 2)
# Calculate the new radii of curvature
left_curverad = ((1 + (2 * left_fit_cr[0] * y_eval*ym_per_pix + \
left_fit_cr[1]) ** 2) ** 1.5) / np.absolute(2 * left_fit_cr[ | return self.sliding_window_fit(img) | conditional_block | |
lane_detector.py |
# Again, extract left and right line pixel positions
self.leftx = nonzerox[left_lane_inds]
lefty = nonzeroy[left_lane_inds]
self.rightx = nonzerox[right_lane_inds]
righty = nonzeroy[right_lane_inds]
# Fit a second order polynomial to each
self.left_fit = np.polyfit(lefty, self.leftx, 2)
self.right_fit = np.polyfit(righty, self.rightx, 2)
return self.left_fit, self.right_fit
def sliding_window_fit(self, img):
''' Apply sliding windows search to the given image to find polynomial to find lane lines
Code based largely on Udacity lessons
:param img - given image
:return left_fit, right_fit - polynomials fitting the left/right lane lines
'''
y_half = int(img.shape[0]/2)
# take histogram of bottom half of img
histogram = np.sum(img[y_half:, :], axis=0)
# Create an output image to draw on and visualize the result
out_img = np.dstack((img, img, img))*255
# Find the peak of the left and right halves of the histogram
# These will be the starting point for the left and right lines
midpoint = np.int(histogram.shape[0]/2)
leftx_base = np.argmax(histogram[:midpoint])
rightx_base = np.argmax(histogram[midpoint:]) + midpoint
# Choose the number of sliding windows
nwindows = 9
# Set height of windows
window_height = np.int(img.shape[0]/nwindows)
# Identify the x and y positions of all nonzero pixels in the image
nonzero = img.nonzero()
nonzeroy = np.array(nonzero[0])
nonzerox = np.array(nonzero[1])
# Current positions to be updated for each window
leftx_current = leftx_base
rightx_current = rightx_base
# Set the width of the windows +/- margin
margin = 100
# Set minimum number of pixels found to recenter window
minpix = 50
# Create empty lists to receive left and right lane pixel indices
left_lane_inds = []
right_lane_inds = []
for window in range(nwindows):
# Identify window boundaries in x and y (and right and left)
win_y_low = img.shape[0] - (window+1) * window_height
win_y_high = img.shape[0] - window * window_height
win_xleft_low = leftx_current - margin
win_xleft_high = leftx_current + margin
win_xright_low = rightx_current - margin
win_xright_high = rightx_current + margin
# Draw the windows on the visualization image
cv2.rectangle(out_img,(win_xleft_low,win_y_low), (win_xleft_high,win_y_high), (0,255,0), 2)
cv2.rectangle(out_img,(win_xright_low,win_y_low),(win_xright_high,win_y_high),(0,255,0), 2)
# Identify the nonzero pixels in x and y within the window
good_left_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xleft_low) & (nonzerox < win_xleft_high)).nonzero()[0]
good_right_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xright_low) & (nonzerox < win_xright_high)).nonzero()[0]
# Append these indices to the lists
left_lane_inds.append(good_left_inds)
right_lane_inds.append(good_right_inds)
# If you found > minpix pixels, recenter next window on their mean position
if len(good_left_inds) > minpix:
leftx_current = np.int( np.mean(nonzerox[good_left_inds]) )
if len(good_right_inds) > minpix:
rightx_current = np.int( np.mean(nonzerox[good_right_inds]) )
# Concatenate the arrays of indices
left_lane_inds = np.concatenate(left_lane_inds)
right_lane_inds = np.concatenate(right_lane_inds)
# Extract left and right line pixel positions
self.leftx = nonzerox[left_lane_inds]
lefty = nonzeroy[left_lane_inds]
righty = nonzeroy[right_lane_inds]
self.rightx = nonzerox[right_lane_inds]
# Fit a second order polynomial to each
self.left_fit = np.polyfit(lefty, self.leftx, 2)
self.right_fit = np.polyfit(righty,self.rightx, 2)
return self.left_fit, self.right_fit
def find_lane_curvature(self, img):
''' Find lane curvature for the given img
:param img - the input image
:return lane curvature
'''
# Generate some fake data to represent lane-line pixels
ploty = np.linspace(0, 719, num=720) # to cover same y-range as image
quadratic_coeff = 3e-4 # arbitrary quadratic coefficient
# For each y position generate random x position within +/-50 pix
# of the line base position in each case (x=200 for left, and x=900 for right)
leftx = np.array([200 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
rightx = np.array([900 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
leftx = leftx[::-1] # Reverse to match top-to-bottom in y
rightx = rightx[::-1] # Reverse to match top-to-bottom in y
# Fit a second order polynomial to pixel positions in each fake lane line
# left_fit = np.polyfit(ploty, leftx, 2)
# left_fitx = left_fit[0]*ploty**2 + left_fit[1]*ploty + left_fit[2]
# right_fit = np.polyfit(ploty, rightx, 2)
# right_fitx = right_fit[0]*ploty**2 + right_fit[1]*ploty + right_fit[2]
# Define y-value where we want radius of curvature
# I'll choose the maximum y-value, corresponding to the bottom of the image
y_eval = np.max(ploty)
# left_curverad = ((1 + (2*left_fit[0]*y_eval + left_fit[1])**2)**1.5) / np.absolute(2*left_fit[0])
# right_curverad = ((1 + (2*right_fit[0]*y_eval + right_fit[1])**2)**1.5) / np.absolute(2*right_fit[0])
# print(left_curverad, right_curverad)
# Example values: 1926.74 1908.48
# Define conversions in x and y from pixels space to meters
ym_per_pix = 30/720 # meters per pixel in y dimension
xm_per_pix = 3.7/700 # meters per pixel in x dimension
# Fit new polynomials to x,y in world space
left_fit_cr = np.polyfit(ploty * ym_per_pix, leftx * xm_per_pix, 2)
right_fit_cr | ''' Apply polynomial fit to the given image, returning fit for left/right lanes
Called when one frame of image has previously found left_fit/right_fit.
This method attempts to find lane fits in the vicinity of previous fits
:param img -- input image with lane lines
:return left_fit, right_fit
'''
if self.left_fit is None or self.right_fit is None:
return self.sliding_window_fit(img)
# from the next frame of video (also called "binary_warped")
# It's now much easier to find line pixels!
nonzero = img.nonzero()
nonzeroy = np.array(nonzero[0])
nonzerox = np.array(nonzero[1])
margin = 100
left_lane_inds = ((nonzerox > (self.left_fit[0]*(nonzeroy**2) + self.left_fit[1]*nonzeroy + self.left_fit[2] - margin)) &
(nonzerox < (self.left_fit[0]*(nonzeroy**2) + self.left_fit[1]*nonzeroy + self.left_fit[2] + margin)))
right_lane_inds = ((nonzerox > (self.right_fit[0]*(nonzeroy**2) + self.right_fit[1]*nonzeroy + self.right_fit[2] - margin)) &
(nonzerox < (self.right_fit[0]*(nonzeroy**2) + self.right_fit[1]*nonzeroy + self.right_fit[2] + margin))) | identifier_body | |
lane_detector.py | , 2)
self.right_fit = np.polyfit(righty, self.rightx, 2)
return self.left_fit, self.right_fit
def sliding_window_fit(self, img):
''' Apply sliding windows search to the given image to find polynomial to find lane lines
Code based largely on Udacity lessons
:param img - given image
:return left_fit, right_fit - polynomials fitting the left/right lane lines
'''
y_half = int(img.shape[0]/2)
# take histogram of bottom half of img
histogram = np.sum(img[y_half:, :], axis=0)
# Create an output image to draw on and visualize the result
out_img = np.dstack((img, img, img))*255
# Find the peak of the left and right halves of the histogram
# These will be the starting point for the left and right lines
midpoint = np.int(histogram.shape[0]/2)
leftx_base = np.argmax(histogram[:midpoint])
rightx_base = np.argmax(histogram[midpoint:]) + midpoint
# Choose the number of sliding windows
nwindows = 9
# Set height of windows
window_height = np.int(img.shape[0]/nwindows)
# Identify the x and y positions of all nonzero pixels in the image
nonzero = img.nonzero()
nonzeroy = np.array(nonzero[0])
nonzerox = np.array(nonzero[1])
# Current positions to be updated for each window
leftx_current = leftx_base
rightx_current = rightx_base
# Set the width of the windows +/- margin
margin = 100
# Set minimum number of pixels found to recenter window
minpix = 50
# Create empty lists to receive left and right lane pixel indices
left_lane_inds = []
right_lane_inds = []
for window in range(nwindows):
# Identify window boundaries in x and y (and right and left)
win_y_low = img.shape[0] - (window+1) * window_height
win_y_high = img.shape[0] - window * window_height
win_xleft_low = leftx_current - margin
win_xleft_high = leftx_current + margin
win_xright_low = rightx_current - margin
win_xright_high = rightx_current + margin
# Draw the windows on the visualization image
cv2.rectangle(out_img,(win_xleft_low,win_y_low), (win_xleft_high,win_y_high), (0,255,0), 2)
cv2.rectangle(out_img,(win_xright_low,win_y_low),(win_xright_high,win_y_high),(0,255,0), 2)
# Identify the nonzero pixels in x and y within the window
good_left_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xleft_low) & (nonzerox < win_xleft_high)).nonzero()[0]
good_right_inds = ((nonzeroy >= win_y_low) & (nonzeroy < win_y_high) &
(nonzerox >= win_xright_low) & (nonzerox < win_xright_high)).nonzero()[0]
# Append these indices to the lists
left_lane_inds.append(good_left_inds)
right_lane_inds.append(good_right_inds)
# If you found > minpix pixels, recenter next window on their mean position
if len(good_left_inds) > minpix:
leftx_current = np.int( np.mean(nonzerox[good_left_inds]) )
if len(good_right_inds) > minpix:
rightx_current = np.int( np.mean(nonzerox[good_right_inds]) )
# Concatenate the arrays of indices
left_lane_inds = np.concatenate(left_lane_inds)
right_lane_inds = np.concatenate(right_lane_inds)
# Extract left and right line pixel positions
self.leftx = nonzerox[left_lane_inds]
lefty = nonzeroy[left_lane_inds]
righty = nonzeroy[right_lane_inds]
self.rightx = nonzerox[right_lane_inds]
# Fit a second order polynomial to each
self.left_fit = np.polyfit(lefty, self.leftx, 2)
self.right_fit = np.polyfit(righty,self.rightx, 2)
return self.left_fit, self.right_fit
def find_lane_curvature(self, img):
''' Find lane curvature for the given img
:param img - the input image
:return lane curvature
'''
# Generate some fake data to represent lane-line pixels
ploty = np.linspace(0, 719, num=720) # to cover same y-range as image
quadratic_coeff = 3e-4 # arbitrary quadratic coefficient
# For each y position generate random x position within +/-50 pix
# of the line base position in each case (x=200 for left, and x=900 for right)
leftx = np.array([200 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
rightx = np.array([900 + (y**2)*quadratic_coeff + np.random.randint(-50, high=51)
for y in ploty])
leftx = leftx[::-1] # Reverse to match top-to-bottom in y
rightx = rightx[::-1] # Reverse to match top-to-bottom in y
# Fit a second order polynomial to pixel positions in each fake lane line
# left_fit = np.polyfit(ploty, leftx, 2)
# left_fitx = left_fit[0]*ploty**2 + left_fit[1]*ploty + left_fit[2]
# right_fit = np.polyfit(ploty, rightx, 2)
# right_fitx = right_fit[0]*ploty**2 + right_fit[1]*ploty + right_fit[2]
# Define y-value where we want radius of curvature
# I'll choose the maximum y-value, corresponding to the bottom of the image
y_eval = np.max(ploty)
# left_curverad = ((1 + (2*left_fit[0]*y_eval + left_fit[1])**2)**1.5) / np.absolute(2*left_fit[0])
# right_curverad = ((1 + (2*right_fit[0]*y_eval + right_fit[1])**2)**1.5) / np.absolute(2*right_fit[0])
# print(left_curverad, right_curverad)
# Example values: 1926.74 1908.48
# Define conversions in x and y from pixels space to meters
ym_per_pix = 30/720 # meters per pixel in y dimension
xm_per_pix = 3.7/700 # meters per pixel in x dimension
# Fit new polynomials to x,y in world space
left_fit_cr = np.polyfit(ploty * ym_per_pix, leftx * xm_per_pix, 2)
right_fit_cr = np.polyfit(ploty * ym_per_pix, rightx * xm_per_pix, 2)
# Calculate the new radii of curvature
left_curverad = ((1 + (2 * left_fit_cr[0] * y_eval*ym_per_pix + \
left_fit_cr[1]) ** 2) ** 1.5) / np.absolute(2 * left_fit_cr[0])
right_curverad = ((1 + (2 * right_fit_cr[0] * y_eval * ym_per_pix + \
right_fit_cr[1]) ** 2) ** 1.5) / np.absolute(2 * right_fit_cr[0])
# Now our radius of curvature is in meters
# print(left_curverad, 'm', right_curverad, 'm')
# Example values: 632.1 m 626.2 m
lx = self.left_fit[0] * (img.shape[0] - 1)**2 + \
self.left_fit[1] * (img.shape[0] - 1) + \
self.left_fit[2]
rx = self.right_fit[0] * (img.shape[0] - 1)**2 + \
self.right_fit[1] * (img.shape[0] - 1) + \
self.right_fit[2]
# calc car's position in the lane w.r.to center
position = ((img.shape[1] / 2) - ((lx + rx)/2)) * xm_per_pix
# calc mean curvature
mean_curverad = (left_curverad + right_curverad) / 2
# save the car's position
self.car_position = position.round(2)
return mean_curverad
| def draw_polygon(self, img, left_fit, right_fit, M_inverse):
''' Draw shaded polygon on the lane between left_fit and right_fit | random_line_split | |
quiz.js | : 0,
quizNameArray: [],
aniArray: [],
aniId: null,
init: function () {
var quiz = QSAll('.quiz'),
popupQuiz = QSAll('.popupPageContainer .quiz');
this.objCount = new Array(quiz.length);
for (var i = 0; i < quiz.length; i++) {
var quizNameArray = quiz[i].getAttribute('quiz'),
quizNum = i + 1;
this.objCount[i] = 0;
quiz[i].classList.add('quiz_' + quizNum);
quiz[i].setAttribute('idx', quizNum);
if (quizNameArray !== null) {
quizNameArray = (new String(quizNameArray).indexOf(',') > -1) ? quizNameArray.split(',') : [quizNameArray];
this.start(quizNum, quizNameArray);
}
else {
console.log('noQuiz');
}
}
},
start: function (quizNum, quizName) {
this.quizNameArray.push(quizName);
for (var i = 0; i < quizName.length; i++) this[quizName[i]]['init'](quizNum);
}
}
return quizObj;
})();
// 퀴즈 : dragLine 이벤트
QUIZ.dragLine = {
name: 'dragLine',
dragLineObj: null,
dropArea: null,
path: null,
objSize: {width: null, height: null},
dropSize: {width: null, height: null},
dropPosition: [],
objPosition: [],
// 초기 설정
init: function (quizNum) { console.log('>>>>>> dragLine');
var svgContainer = CESVG(QS('.quiz_'+ quizNum), 'svg');
svgContainer.setAttribute('class', 'svgContainer');
// $('.svgContainer').insertBefore($('.questionDotWrap'));
this.append(quizNum);
QUIZ.objCount[quizNum-1] += this.dragLineObj.length;
for (var i = 0; i < this.dragLineObj.length; i++) {
this.dragLineObj[i].style.cursor = 'pointer';
this.dragLineObj[i].setAttribute('value', i + 1);
new DragDrop({
quizNum: quizNum,
quizName: this.name,
element: this.dragLineObj[i],
top: this.objPosition[i].top,
left: this.objPosition[i].left,
width: this.objSize.width,
height: this.objSize.height,
callBack: function (e, param) {
var eventMaster = eventSelector('eventUp', e),
dropArea = QSAll('.quiz_'+ param.quizNum +' .lineDropArea'),
answerCount = 0;
if (eventMaster !== undefined && QUIZ.dragLine.dropCompare(param.quizNum, this, dropArea, eventMaster.clientX, eventMaster.clientY)) {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, true);
} else {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, false);
}
}
});
}
this.path = QSAll('.quiz_'+ quizNum +' .svgContainer > path');
for (var i = 0; i < this.path.length; i++) {
this.path[i].setAttribute('class', 'answerLine');
this.path[i].setAttribute('value', this.dragLineObj[i].getAttribute('value'));
}
},
// 각 요소 위치 저장
append: function (quizNum) {
var svgContainer = QS('.quiz_'+ quizNum + ' .svgContainer');
this.dragLineObj = QSAll('.quiz_'+ quizNum +' .dragLineObj');
this.dropArea = QSAll('.quiz_'+ quizNum +' .lineDropArea');
//this.objSize.width = QS('.quiz_'+ quizNum +' .dragLineObj').offsetWidth / 2;
//this.objSize.height = QS('.quiz_'+ quizNum +' .dragLineObj').offsetHeight / 2;
this.objSize.width = dlo[quizNum-1].width[0] / 2;
this.objSize.height = dlo[quizNum-1].height[0] / 2;
this.dropSize.width = QS('.quiz_'+ quizNum +' .lineDropArea').offsetWidth / 2;
this.dropSize.height = QS('.quiz_'+ quizNum +' .lineDropArea').offsetHeight / 2;
this.dropPosition = [];
this.objPosition = [];
for (var i = 0; i < this.dropArea.length; i++) {
this.dropPosition.push({top: this.dropArea[i].offsetTop, left: this.dropArea[i].offsetLeft});
}
for (var i = 0; i < this.dragLineObj.length; i++) {
this.objPosition.push({top: dlo[quizNum-1].top[i], left: dlo[quizNum-1].left[i]});
//this.objPosition.push({top: this.dragLineObj[i].offsetTop, left: this.dragLineObj[i].offsetLeft});
}
},
// 드랍 영역 체크
dropCompare: function (quizNum, dragObj, dropArea, x, y) {
var dragObjValue = dragObj.element !== undefined ? dragObj.element.getAttribute('value') : dragObj.getAttribute('value'),
allDap = false,
result;
//변수처리함 .
var dotLeft = dragObj.element.parentNode.parentNode.getElementsByClassName('dLeft');
var dotRight = dragObj.element.parentNode.parentNode.getElementsByClassName('dRight');
for (var i = 0; i < dropArea.length; i++) {
var dropValue = dropArea[i].getAttribute('value').indexOf(',') > -1 ? dropArea[i].getAttribute('value').split(',') : [dropArea[i].getAttribute('value')],
dropAreaCss = dropArea[i].getBoundingClientRect();
if (x === undefined && y === undefined) allDap = true;
var comparePosition = x >= dropAreaCss.left &&
x <= (dropAreaCss.left + dropAreaCss.width) &&
y >= dropAreaCss.top &&
y <= dropAreaCss.top + dropAreaCss.height;
if (comparePosition || allDap) {
for (var j = 0; j < dropValue.length; j++) {
if (dragObjValue == dropValue[j]) {
//var dLeft = QSAll('.dLeft');
//var dRight = QSAll('.dRight');
var dLeft = dotLeft;//신규추가
var dRight = dotRight;//신규추가
dLeft[dragObjValue-1].childNodes[0].style.backgroundColor = '#000';
dRight[dragObjValue-1].childNodes[0].style.backgroundColor = '#000';
result = true;
}
}
if (result === undefined) result = false;
}
}
return result;
},
setDragObjPosition: function (quizNum, dragObj, param, type) {
console.log('setDragObjPosition')
var obj = dragObj.element !== undefined ? dragObj.element : dragObj,
idx = obj.getAttribute('value') - 1,
top, left, targetPath, value, dropTop, dropLeft;
QUIZ.dragLine.append(quizNum);
this.path = QSAll('.quiz_' + quizNum +' .svgContainer > path');
for (var i = 0; i < this.path.length; i++) {
if (obj.getAttribute('value') == this.path[i].getAttribute('value')) {
targetPath = this.path[i];
}
}
value = targetPath.getAttribute('value');
for (var i = 0; i < this.dropArea.length; i++) {
if (obj.getAttribute('value') == this.dropArea[i].getAttribute('value')) {
dropTop = this.dropArea[i].offsetTop + this.dropSize.width;
dropLeft = this.dropArea[i].offsetLeft + this.dropSize.height;
}
}
obj.style.left = param.left + 'px';
obj.style.top = param.top + 'px';
left = param.left + param.width;
top = param.top + param.height;
if (type) {
obj.style.pointerEvents = 'none';
obj.classList.add(this.name + 'Complete');
targetPath.setAttribute('d', 'M '+ left +' '+ top + ' L '+ dropLeft +' '+ dropTop);
} else {
targetPath.setAttribute('d', 'M '+ 0 +' '+ 0 + ' L '+ 0 +' '+ 0);
}
//console.error(obj.parentNode.parentNode.parentNode.parentNode.getAttribute('data-qid'));
var dataQid = obj.parentNode.parentNode.parentNode.parentNode.getAttribute('data-qid');
DTCaliperSensor.fire({
correct: null, // 정답 여부입력 [true, false] 중에서 택일
itemObject: document.querySelector('[data-qid='+dataQid+']'), // 해당 문항 객체
value: '' // 실제 정답 데이터 입력 <correctResponse>에 입력된 값이랑 동일 | random_line_split | ||
quiz.js | on DragDrop (param) {
this.element = param.element;
this.parentElment = window;
this.createDragDrop(param);
}
// 드래그&드랍 : 위치 이동
DragDrop.prototype.createDragDrop = function (param) {
var dragObj = this,
left = param.left + param.width,
top = param.top + param.height,
answerLine = (param.quizName !== 'dragLine') ? null : CESVG(QS('.quiz_'+ param.quizNum +' .svgContainer'), 'path'),
startDrag = function (e) {
e.preventDefault();
var eventMaster = eventSelector('eventDown', e);
dragObj.element.style.zIndex = 3;
dragObj.offY = eventMaster.clientY - (dragObj.element.offsetTop * GameManager.event.zoomRate);
dragObj.offX = eventMaster.clientX - (dragObj.element.offsetLeft * GameManager.event.zoomRate);
dragObj.element.addEventListener(GameManager.event.eventSelector('eventMove'), drag, true);
dragObj.element.addEventListener(GameManager.event.eventSelector('eventOut'), endDrag, true);
},
drag = function (e) {
e.preventDefault();
var eventMaster = eventSelector('eventMove', e);
dragObj.newY = eventMaster.clientY - dragObj.offY;
dragObj.newX = eventMaster.clientX - dragObj.offX;
dragObj.element.style.left = (dragObj.newX / GameManager.event.zoomRate) + 'px';
dragObj.element.style.top = (dragObj.newY / GameManager.event.zoomRate) + 'px';
var newLeft = (dragObj.newX + param.width * GameManager.event.zoomRate) / GameManager.event.zoomRate;
var newTop = (dragObj.newY + param.height * GameManager.event.zoomRate) / GameManager.event.zoomRate;
if (answerLine !== null) {
answerLine.setAttribute('d', 'M '+ left +' '+ top + ' L '+ newLeft +' '+ newTop);
}
},
endDrag = function (e) {
e.preventDefault();
var eventMaster = eventSelector('eventUp', e);
dragObj.element.removeEventListener(GameManager.event.eventSelector('eventMove'), drag, true);
dragObj.element.style.zIndex = 3;
param.callBack(e, param);
}
dragObj.element.addEventListener(GameManager.event.eventSelector('eventDown'), startDrag, true);
dragObj.element.addEventListener(GameManager.event.eventSelector('eventUp'), endDrag, true);
};
// ********************************************************************************
// 퀴즈 : 기본 설정
var QUIZ = QUIZ || {};
QUIZ = (function(){
var quizObj = {
objCount: null,
totalCount: 0,
quizNameArray: [],
aniArray: [],
aniId: null,
init: function () {
var quiz = QSAll('.quiz'),
popupQuiz = QSAll('.popupPageContainer .quiz');
this.objCount = new Array(quiz.length);
for (var i = 0; i < quiz.length; i++) {
var quizNameArray = quiz[i].getAttribute('quiz'),
quizNum = i + 1;
this.objCount[i] = 0;
quiz[i].classList.add('quiz_' + quizNum);
quiz[i].setAttribute('idx', quizNum);
if (quizNameArray !== null) {
quizNameArray = (new String(quizNameArray).indexOf(',') > -1) ? quizNameArray.split(',') : [quizNameArray];
this.start(quizNum, quizNameArray);
}
else {
console.log('noQuiz');
}
}
},
start: function (quizNum, quizName) {
this.quizNameArray.push(quizName);
for (var i = 0; i < quizName.length; i++) this[quizName[i]]['init'](quizNum);
}
}
return quizObj;
})();
// 퀴즈 : dragLine 이벤트
QUIZ.dragLine = {
name: 'dragLine',
dragLineObj: null,
dropArea: null,
path: null,
objSize: {width: null, height: null},
dropSize: {width: null, height: null},
dropPosition: [],
objPosition: [],
// 초기 설정
init: function (quizNum) { console.log('>>>>>> dragLine');
var svgContainer = CESVG(QS('.quiz_'+ quizNum), 'svg');
svgContainer.setAttribute('class', 'svgContainer');
// $('.svgContainer').insertBefore($('.questionDotWrap'));
this.append(quizNum);
QUIZ.objCount[quizNum-1] += this.dragLineObj.length;
for (var i = 0; i < this.dragLineObj.length; i++) {
this.dragLineObj[i].style.cursor = 'pointer';
this.dragLineObj[i].setAttribute('value', i + 1);
new DragDrop({
quizNum: quizNum,
quizName: this.name,
element: this.dragLineObj[i],
top: this.objPosition[i].top,
left: this.objPosition[i].left,
width: this.objSize.width,
height: this.objSize.height,
callBack: function (e, param) {
var eventMaster = eventSelector('eventUp', e),
dropArea = QSAll('.quiz_'+ param.quizNum +' .lineDropArea'),
answerCount = 0;
if (eventMaster !== undefined && QUIZ.dragLine.dropCompare(param.quizNum, this, dropArea, eventMaster.clientX, eventMaster.clientY)) {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, true);
} else {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, false);
}
}
});
}
this.path = QSAll('.quiz_'+ quizNum +' .svgContainer > path');
for (var i = 0; i < this.path.length; i++) {
this.path[i].setAttribute('class', 'answerLine');
this.path[i].setAttribute('value', this.dragLineObj[i].getAttribute('value'));
}
},
// 각 요소 위치 저장
append: function (quizNum) {
var svgContainer = QS('.quiz_'+ quizNum + ' .svgContainer');
this.dragLineObj = QSAll('.quiz_'+ quizNum +' .dragLineObj');
this.dropArea = QSAll('.quiz_'+ quizNum +' .lineDropArea');
//this.objSize.width = QS('.quiz_'+ quizNum +' .dragLineObj').offsetWidth / 2;
//this.objSize.height = QS('.quiz_'+ quizNum +' .dragLineObj').offsetHeight / 2;
this.objSize.width = dlo[quizNum-1].width[0] / 2;
this.objSize.height = dlo[quizNum-1].height[0] / 2;
this.dropSize.width = QS('.quiz_'+ quizNum +' .lineDropArea').offsetWidth / 2;
this.dropSize.height = QS('.quiz_'+ quizNum +' .lineDropArea').offsetHeight / 2;
this.dropPosition = [];
this.objPosition = [];
for (var i = 0; i < this.dropArea.length; i++) {
this.dropPosition.push({top: this.dropArea[i].offsetTop, left: this.dropArea[i].offsetLeft});
}
for (var i = 0; i < this.dragLineObj.length; i++) {
this.objPosition.push({top: dlo[quizNum-1].top[i], left: dlo[quizNum-1].left[i]});
//this.objPosition.push({top: this.dragLineObj[i].offsetTop, left: this.dragLineObj[i].offsetLeft});
}
},
// 드랍 영역 체크
dropCompare: function (quizNum, dragObj, dropArea, x, y) {
var dragObjValue = dragObj.element !== undefined ? dragObj.element.getAttribute('value') : dragObj.getAttribute('value'),
allDap = false,
result;
//변수처리함 .
var dotLeft = dragObj.element.parentNode.parentNode.getElementsByClassName('dLeft');
var dotRight = dragObj.element.parentNode.parentNode.getElementsByClassName('dRight');
for (var i = 0; i < dropArea.length; i++) {
var dropValue = dropArea[i].getAttribute('value').indexOf(',') > -1 ? dropArea[i].getAttribute('value').split(',') : [dropArea[i].getAttribute('value')],
dropAreaCss = dropArea[i].getBoundingClientRect();
if (x === undefined && y === undefined) allDap = true;
var comparePosition = x >= dropAreaCss.left &&
x <= (dropAreaCss.left + dropAreaCss.width) &&
y >= dropAreaCss.top &&
y <= dropAreaCss.top + dropAreaCss.height;
if (comparePosition || allDap) {
for (var j | Audio;
var efPlay = function () {
efAudio.removeEventListener('loadeddata', efPlay);
efAudio.play();
};
efAudio.src = src;
efAudio.addEventListener('loadeddata', efPlay);
efAudio.load();
}
// 드래그&드랍 : 설정
functi | identifier_body | |
quiz.js | type) {
var svgContainer = document.createElementNS('http://www.w3.org/2000/svg', type);
target.appendChild(svgContainer);
return svgContainer;
}
// 효과음 기본 설정
function efSound (src) {
var efAudio = new Audio;
var efPlay = function () {
efAudio.removeEventListener('loadeddata', efPlay);
efAudio.play();
};
efAudio.src = src;
efAudio.addEventListener('loadeddata', efPlay);
efAudio.load();
}
// 드래그&드랍 : 설정
function DragDrop (param) {
this.element = param.element;
this.parentElment = window;
this.createDragDrop(param);
}
// 드래그&드랍 : 위치 이동
DragDrop.prototype.createDragDrop = function (param) {
var dragObj = this,
left = param.left + param.width,
top = param.top + param.height,
answerLine = (param.quizName !== 'dragLine') ? null : CESVG(QS('.quiz_'+ param.quizNum +' .svgContainer'), 'path'),
startDrag = function (e) {
e.preventDefault();
var eventMaster = eventSelector('eventDown', e);
dragObj.element.style.zIndex = 3;
dragObj.offY = eventMaster.clientY - (dragObj.element.offsetTop * GameManager.event.zoomRate);
dragObj.offX = eventMaster.clientX - (dragObj.element.offsetLeft * GameManager.event.zoomRate);
dragObj.element.addEventListener(GameManager.event.eventSelector('eventMove'), drag, true);
dragObj.element.addEventListener(GameManager.event.eventSelector('eventOut'), endDrag, true);
},
drag = function (e) {
e.preventDefault();
var eventMaster = eventSelector('eventMove', e);
dragObj.newY = eventMaster.clientY - dragObj.offY;
dragObj.newX = eventMaster.clientX - dragObj.offX;
dragObj.element.style.left = (dragObj.newX / GameManager.event.zoomRate) + 'px';
dragObj.element.style.top = (dragObj.newY / GameManager.event.zoomRate) + 'px';
var newLeft = (dragObj.newX + param.width * GameManager.event.zoomRate) / GameManager.event.zoomRate;
var newTop = (dragObj.newY + param.height * GameManager.event.zoomRate) / GameManager.event.zoomRate;
if (answerLine !== null) {
answerLine.setAttribute('d', 'M '+ left +' '+ top + ' L '+ newLeft +' '+ newTop);
}
},
endDrag = function (e) {
e.preventDefault();
var eventMaster = eventSelector('eventUp', e);
dragObj.element.removeEventListener(GameManager.event.eventSelector('eventMove'), drag, true);
dragObj.element.style.zIndex = 3;
param.callBack(e, param);
}
dragObj.element.addEventListener(GameManager.event.eventSelector('eventDown'), startDrag, true);
dragObj.element.addEventListener(GameManager.event.eventSelector('eventUp'), endDrag, true);
};
// ********************************************************************************
// 퀴즈 : 기본 설정
var QUIZ = QUIZ || {};
QUIZ = (function(){
var quizObj = {
objCount: null,
totalCount: 0,
quizNameArray: [],
aniArray: [],
aniId: null,
init: function () {
var quiz = QSAll('.quiz'),
popupQuiz = QSAll('.popupPageContainer .quiz');
this.objCount = new Array(quiz.length);
for (var i = 0; i < quiz.length; i++) {
var quizNameArray = quiz[i].getAttribute('quiz'),
quizNum = i + 1;
this.objCount[i] = 0;
quiz[i].classList.add('quiz_' + quizNum);
quiz[i].setAttribute('idx', quizNum);
if (quizNameArray !== null) {
quizNameArray = (new String(quizNameArray).indexOf(',') > -1) ? quizNameArray.split(',') : [quizNameArray];
this.start(quizNum, quizNameArray);
}
else {
console.log('noQuiz');
}
}
},
start: function (quizNum, quizName) {
this.quizNameArray.push(quizName);
for (var i = 0; i < quizName.length; i++) this[quizName[i]]['init'](quizNum);
}
}
return quizObj;
})();
// 퀴즈 : dragLine 이벤트
QUIZ.dragLine = {
name: 'dragLine',
dragLineObj: null,
dropArea: null,
path: null,
objSize: {width: null, height: null},
dropSize: {width: null, height: null},
dropPosition: [],
objPosition: [],
// 초기 설정
init: function (quizNum) { console.log('>>>>>> dragLine');
var svgContainer = CESVG(QS('.quiz_'+ quizNum), 'svg');
svgContainer.setAttribute('class', 'svgContainer');
// $('.svgContainer').insertBefore($('.questionDotWrap'));
this.append(quizNum);
QUIZ.objCount[quizNum-1] += this.dragLineObj.length;
for (var i = 0; i < this.dragLineObj.length; i++) {
this.dragLineObj[i].style.cursor = 'pointer';
this.dragLineObj[i].setAttribute('value', i + 1);
new DragDrop({
quizNum: quizNum,
quizName: this.name,
element: this.dragLineObj[i],
top: this.objPosition[i].top,
left: this.objPosition[i].left,
width: this.objSize.width,
height: this.objSize.height,
callBack: function (e, param) {
var eventMaster = eventSelector('eventUp', e),
dropArea = QSAll('.quiz_'+ param.quizNum +' .lineDropArea'),
answerCount = 0;
if (eventMaster !== undefined && QUIZ.dragLine.dropCompare(param.quizNum, this, dropArea, eventMaster.clientX, eventMaster.clientY)) {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, true);
} else {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, false);
}
}
});
}
this.path = QSAll('.quiz_'+ quizNum +' .svgContainer > path');
for (var i = 0; i < this.path.length; i++) {
this.path[i].setAttribute('class', 'answerLine');
this.path[i].setAttribute('value', this.dragLineObj[i].getAttribute('value'));
}
},
// 각 요소 위치 저장
append: function (quizNum) {
var svgContainer = QS('.quiz_'+ quizNum + ' .svgContainer');
this.dragLineObj = QSAll('.quiz_'+ quizNum +' .dragLineObj');
this.dropArea = QSAll('.quiz_'+ quizNum +' .lineDropArea');
//this.objSize.width = QS('.quiz_'+ quizNum +' .dragLineObj').offsetWidth / 2;
//this.objSize.height = QS('.quiz_'+ quizNum +' .dragLineObj').offsetHeight / 2;
this.objSize.width = dlo[quizNum-1].width[0] / 2;
this.objSize.height = dlo[quizNum-1].height[0] / 2;
this.dropSize.width = QS('.quiz_'+ quizNum +' .lineDropArea').offsetWidth / 2;
this.dropSize.height = QS('.quiz_'+ quizNum +' .lineDropArea').offsetHeight / 2;
this.dropPosition = [];
this.objPosition = [];
for (var i = 0; i < this.dropArea.length; i++) {
this.dropPosition.push({top: this.dropArea[i].offsetTop, left: this.dropArea[i].offsetLeft});
}
for (var i = 0; i < this.dragLineObj.length; i++) {
this.objPosition.push({top: dlo[quizNum-1].top[i], left: dlo[quizNum-1].left[i]});
//this.objPosition.push({top: this.dragLineObj[i].offsetTop, left: this.dragLineObj[i].offsetLeft});
}
},
// 드랍 영역 체크
dropCompare: function (quizNum, dragObj, dropArea, x, y) {
var dragObjValue = dragObj.element !== undefined ? dragObj.element.getAttribute('value') : dragObj.getAttribute('value'),
allDap = false,
result;
//변수처리함 .
var dotLeft = dragObj.element.parentNode.parentNode.getElementsByClassName('dLeft');
var dotRight = dragObj.element.parentNode.parentNode.getElementsByClassName('dRight');
for (var i = 0; i < dropArea.length; i++) {
var dropValue = dropArea[i].getAttribute('value').indexOf(',') > -1 ? dropArea[i].getAttribute('value').split(',') : [dropArea[i].getAttribute('value')],
dropAreaCss = dropArea[i].getBoundingClientRect();
if (x === undefined && y === undefined) allDap = true;
var comparePosition = x | get, | identifier_name | |
quiz.js | / GameManager.event.zoomRate;
var newTop = (dragObj.newY + param.height * GameManager.event.zoomRate) / GameManager.event.zoomRate;
if (answerLine !== null) {
answerLine.setAttribute('d', 'M '+ left +' '+ top + ' L '+ newLeft +' '+ newTop);
}
},
endDrag = function (e) {
e.preventDefault();
var eventMaster = eventSelector('eventUp', e);
dragObj.element.removeEventListener(GameManager.event.eventSelector('eventMove'), drag, true);
dragObj.element.style.zIndex = 3;
param.callBack(e, param);
}
dragObj.element.addEventListener(GameManager.event.eventSelector('eventDown'), startDrag, true);
dragObj.element.addEventListener(GameManager.event.eventSelector('eventUp'), endDrag, true);
};
// ********************************************************************************
// 퀴즈 : 기본 설정
var QUIZ = QUIZ || {};
QUIZ = (function(){
var quizObj = {
objCount: null,
totalCount: 0,
quizNameArray: [],
aniArray: [],
aniId: null,
init: function () {
var quiz = QSAll('.quiz'),
popupQuiz = QSAll('.popupPageContainer .quiz');
this.objCount = new Array(quiz.length);
for (var i = 0; i < quiz.length; i++) {
var quizNameArray = quiz[i].getAttribute('quiz'),
quizNum = i + 1;
this.objCount[i] = 0;
quiz[i].classList.add('quiz_' + quizNum);
quiz[i].setAttribute('idx', quizNum);
if (quizNameArray !== null) {
quizNameArray = (new String(quizNameArray).indexOf(',') > -1) | function (quizNum, quizName) {
this.quizNameArray.push(quizName);
for (var i = 0; i < quizName.length; i++) this[quizName[i]]['init'](quizNum);
}
}
return quizObj;
})();
// 퀴즈 : dragLine 이벤트
QUIZ.dragLine = {
name: 'dragLine',
dragLineObj: null,
dropArea: null,
path: null,
objSize: {width: null, height: null},
dropSize: {width: null, height: null},
dropPosition: [],
objPosition: [],
// 초기 설정
init: function (quizNum) { console.log('>>>>>> dragLine');
var svgContainer = CESVG(QS('.quiz_'+ quizNum), 'svg');
svgContainer.setAttribute('class', 'svgContainer');
// $('.svgContainer').insertBefore($('.questionDotWrap'));
this.append(quizNum);
QUIZ.objCount[quizNum-1] += this.dragLineObj.length;
for (var i = 0; i < this.dragLineObj.length; i++) {
this.dragLineObj[i].style.cursor = 'pointer';
this.dragLineObj[i].setAttribute('value', i + 1);
new DragDrop({
quizNum: quizNum,
quizName: this.name,
element: this.dragLineObj[i],
top: this.objPosition[i].top,
left: this.objPosition[i].left,
width: this.objSize.width,
height: this.objSize.height,
callBack: function (e, param) {
var eventMaster = eventSelector('eventUp', e),
dropArea = QSAll('.quiz_'+ param.quizNum +' .lineDropArea'),
answerCount = 0;
if (eventMaster !== undefined && QUIZ.dragLine.dropCompare(param.quizNum, this, dropArea, eventMaster.clientX, eventMaster.clientY)) {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, true);
} else {
QUIZ.dragLine.setDragObjPosition(param.quizNum, this, param, false);
}
}
});
}
this.path = QSAll('.quiz_'+ quizNum +' .svgContainer > path');
for (var i = 0; i < this.path.length; i++) {
this.path[i].setAttribute('class', 'answerLine');
this.path[i].setAttribute('value', this.dragLineObj[i].getAttribute('value'));
}
},
// 각 요소 위치 저장
append: function (quizNum) {
var svgContainer = QS('.quiz_'+ quizNum + ' .svgContainer');
this.dragLineObj = QSAll('.quiz_'+ quizNum +' .dragLineObj');
this.dropArea = QSAll('.quiz_'+ quizNum +' .lineDropArea');
//this.objSize.width = QS('.quiz_'+ quizNum +' .dragLineObj').offsetWidth / 2;
//this.objSize.height = QS('.quiz_'+ quizNum +' .dragLineObj').offsetHeight / 2;
this.objSize.width = dlo[quizNum-1].width[0] / 2;
this.objSize.height = dlo[quizNum-1].height[0] / 2;
this.dropSize.width = QS('.quiz_'+ quizNum +' .lineDropArea').offsetWidth / 2;
this.dropSize.height = QS('.quiz_'+ quizNum +' .lineDropArea').offsetHeight / 2;
this.dropPosition = [];
this.objPosition = [];
for (var i = 0; i < this.dropArea.length; i++) {
this.dropPosition.push({top: this.dropArea[i].offsetTop, left: this.dropArea[i].offsetLeft});
}
for (var i = 0; i < this.dragLineObj.length; i++) {
this.objPosition.push({top: dlo[quizNum-1].top[i], left: dlo[quizNum-1].left[i]});
//this.objPosition.push({top: this.dragLineObj[i].offsetTop, left: this.dragLineObj[i].offsetLeft});
}
},
// 드랍 영역 체크
dropCompare: function (quizNum, dragObj, dropArea, x, y) {
var dragObjValue = dragObj.element !== undefined ? dragObj.element.getAttribute('value') : dragObj.getAttribute('value'),
allDap = false,
result;
//변수처리함 .
var dotLeft = dragObj.element.parentNode.parentNode.getElementsByClassName('dLeft');
var dotRight = dragObj.element.parentNode.parentNode.getElementsByClassName('dRight');
for (var i = 0; i < dropArea.length; i++) {
var dropValue = dropArea[i].getAttribute('value').indexOf(',') > -1 ? dropArea[i].getAttribute('value').split(',') : [dropArea[i].getAttribute('value')],
dropAreaCss = dropArea[i].getBoundingClientRect();
if (x === undefined && y === undefined) allDap = true;
var comparePosition = x >= dropAreaCss.left &&
x <= (dropAreaCss.left + dropAreaCss.width) &&
y >= dropAreaCss.top &&
y <= dropAreaCss.top + dropAreaCss.height;
if (comparePosition || allDap) {
for (var j = 0; j < dropValue.length; j++) {
if (dragObjValue == dropValue[j]) {
//var dLeft = QSAll('.dLeft');
//var dRight = QSAll('.dRight');
var dLeft = dotLeft;//신규추가
var dRight = dotRight;//신규추가
dLeft[dragObjValue-1].childNodes[0].style.backgroundColor = '#000';
dRight[dragObjValue-1].childNodes[0].style.backgroundColor = '#000';
result = true;
}
}
if (result === undefined) result = false;
}
}
return result;
},
setDragObjPosition: function (quizNum, dragObj, param, type) {
console.log('setDragObjPosition')
var obj = dragObj.element !== undefined ? dragObj.element : dragObj,
idx = obj.getAttribute('value') - 1,
top, left, targetPath, value, dropTop, dropLeft;
QUIZ.dragLine.append(quizNum);
this.path = QSAll('.quiz_' + quizNum +' .svgContainer > path');
for (var i = 0; i < this.path.length; i++) {
if (obj.getAttribute('value') == this.path[i].getAttribute('value')) {
targetPath = this.path[i];
}
}
value = targetPath.getAttribute('value');
for (var i = 0; i < this.dropArea.length; i++) {
if (obj.getAttribute('value') == this.dropArea[i].getAttribute('value')) {
dropTop = this.dropArea[i].offsetTop + this.dropSize.width;
dropLeft = this.dropArea[i].offsetLeft + this.dropSize.height;
}
}
obj.style.left = param.left + 'px';
obj.style.top = param.top | ? quizNameArray.split(',') : [quizNameArray];
this.start(quizNum, quizNameArray);
}
else {
console.log('noQuiz');
}
}
},
start: | conditional_block |
intraday.py |
self._state = {}
self._syms = [
"BINANCE:UNIUSD", "BINANCE:ETHUSD", "BINANCE:DOTUSD", "SGX:ES3",
"SGX:CLR"
]
self._t = None
self._timeframe = 240 # Default to 4 hours chart
self._ws_url = "wss://data.tradingview.com/socket.io/websocket"
def get(self, type: str, **kwargs):
"""
Type is either quote (live) or chart (historical + live)
Support kwargs:
syms: list of symbols, e.g. [BINANCE:ETHUSD]
indicators: list of indicators, e.g. [rsi]
timeframe: int of minutes of chart time frame, e.g. 240 -> 4 hours chart
histbars: int of number of historical data points, e.g. 300
"""
websocket.enableTrace(True)
ws = websocket.WebSocketApp(
self._ws_url,
on_open=lambda ws: self.on_open(ws, type, **kwargs),
on_close=self.on_close,
on_message=lambda ws, message: self.on_message(ws, message),
on_error=self.on_error)
ws.run_forever()
def send_slack(self):
"""
Periodic slack alerts - Indicators
"""
while True:
indicators = self._alerts.get("indicators")
if indicators:
res = pd.DataFrame(indicators).transpose().reset_index()
res.rename(columns={"index": "sym"}, inplace=True)
send_alert(self._slackchannel, [("Indicators", res)])
time.sleep(self._slackfreq)
def on_message(self, ws, message):
pattern = re.compile(r'~m~\d+~m~~h~\d+$')
if pattern.match(message):
ws.send(message)
else:
msg_body = re.compile(r'~m~\d+~m~')
messages = msg_body.split(message)
for msg in messages:
if msg:
parsed_msg = json.loads(msg)
params = parsed_msg.get("p")
if parsed_msg.get("m") == "timescale_update":
# timescale_update -> initial historical data
# TODO: handling of these data for plotting on UI
continue
if parsed_msg.get("m") == "du":
# du -> data update
sym = self._state.get(params[0]).get("sym")
now = datetime.datetime.now().strftime(
'%Y-%m-%d %H:%M:%S')
for k, v in params[1].items():
if v.get("st"):
# study
indicator = k.split("_")[0]
vals = v.get("st")[0].get("v")
val = vals[1]
val_dict = {"dtime": now, indicator: val}
# print({sym: val_dict})
if not self._alerts["indicators"].get(sym):
|
self._alerts["indicators"][sym][
indicator] = val
elif v.get("s"):
# series
vals = v.get("s")[0].get("v")
val_dict = dict(
zip([
"dtime", "open", "high", "low", "last",
"vol"
], vals))
val_dict["dtime"] = now
# print({sym: val_dict})
if not self._alerts["price"].get(sym):
self._alerts["price"][sym] = {}
self._alerts["price"][sym]["last"] = val_dict[
"last"]
@staticmethod
def on_error(ws, error):
print(error)
@staticmethod
def on_close(ws):
print("### closed ###")
def on_open(self, ws, type: str, **kwargs):
def run(*args, **kwargs):
# ~m~52~m~{"m":"quote_create_session","p":["qs_3bDnffZvz5ur"]}
# ~m~395~m~{"m":"quote_set_fields","p":["qs_3bDnffZvz5ur","ch","chp","lp"]}
# ~m~89~m~{"m":"quote_add_symbols","p":["qs_3bDnffZvz5ur","SP:SPX",{"flags":["force_permission"]}]}
# ~m~315~m~{"m":"quote_fast_symbols","p":["qs_3bDnffZvz5ur","SP:SPX","TVC:NDX","CBOE:VIX","TVC:DXY","SGX:ES3","NASDAQ:AAPL","NASDAQ:MSFT","NASDAQ:TSLA","TVC:USOIL","TVC:GOLD","TVC:SILVER","FX:AUDUSD","FX:EURUSD","FX:GBPUSD","FX:USDJPY","BITSTAMP:BTCUSD","BITSTAMP:ETHUSD","COINBASE:UNIUSD","BINANCE:DOGEUSD","BINANCE:DOTUSD"]}
syms = kwargs.get("syms") or self._syms
timeframe = f'{kwargs.get("timeframe") or self._timeframe}'
indicators = kwargs.get("indicators") or self._indicators
histbars = kwargs.get("histbars") or self._histbars
send = self._send
send(ws, "set_auth_token", ["unauthorized_user_token"])
# Quote session
if not args or (args and args[0] == "quote"):
session = self._gen_session() # Quote session ID
send(ws, "quote_create_session", [session])
send(ws, "quote_set_fields", [session, "lp", "volume"])
[ws.send(self._add_symbol(session, s)) for s in syms]
send(ws, "quote_fast_symbols", [session, *syms])
send(ws, "quote_hibernate_all", [session])
# Chart session - Prefer to use this over quote sessions since it has a historical series
else:
for i, sym in enumerate(syms):
# Each ticker warrants a separate chart session ID
c_session = self._gen_session(type="chart")
self._state[c_session] = {
"sym": sym,
"indicators": [],
"series": [],
"timeframe": timeframe
}
# Users are allowed to select specific tickers
send(ws, "chart_create_session", [c_session, ""])
send(ws, "switch_timezone", [c_session, "Asia/Singapore"])
send(ws, "resolve_symbol", [
c_session, f"symbol_{i}",
self._add_chart_symbol(sym)
])
# s (in resp) -> series
self._state[c_session].get("series").append(f"s_{i}")
send(ws, "create_series", [
c_session, f"s_{i}", f"s_{i}", f"symbol_{i}",
timeframe, histbars
])
for indicator in indicators:
# Users are allowed to select specific indicators
# st (in resp) -> study
self._state[c_session].get("indicators").append(
f"{indicator}_{i}")
send(ws, "create_study", [
c_session, f"{indicator}_{i}", f"{indicator}_{i}",
f"s_{i}", "Script@tv-scripting-101!",
self._indicator_mapper(indicator)
])
self._t = threading.Thread(target=run, args=(type, ), kwargs=kwargs)
self._t.setDaemon(True)
self._t.start()
def _send(self, ws, func, params):
""" Client sends msg to websockets server """
ws.send(self._create_msg(func, params))
def _indicator_mapper(self, indicator: str) -> dict:
""" Indicator params that are accepted by the tv server """
return {
"rsi": {
"text":
"1f0fkZ72S0de2geyaUhXXw==_xwY73vljRXeew69Rl27RumLDs6aJ9NLsTYN9Xrht254BTb8uSOgccpLDt/cdRWopwJPNZx40m19yEFwJFswkSi62X4guNJYpXe4A6S9iq2n+OXM6mqWeWzDbjTl0lYmEf1ujbg7i3FvUdV/zCSrqd+iwnvvZSV+O2acpfNLpUlDdB6PZX4Y9y8tlQLWA2PiF8CVJng7DF1LPeecWC4fv+lNg+s5OXU46AjIhc+TFu8DOwiuKjNh7wWz6EZ7gpQS3",
"pineId": "STD;RSI",
"pineVersion": "12.0",
"in_2": {
"v": "",
"f": True,
"t": "resolution"
},
"in_0": {
"v": 14,
"f": True,
"t": " | self._alerts["indicators"][sym] = {} | conditional_block |
intraday.py | mins
self._state = {}
self._syms = [
"BINANCE:UNIUSD", "BINANCE:ETHUSD", "BINANCE:DOTUSD", "SGX:ES3",
"SGX:CLR"
]
self._t = None
self._timeframe = 240 # Default to 4 hours chart
self._ws_url = "wss://data.tradingview.com/socket.io/websocket"
def get(self, type: str, **kwargs):
"""
Type is either quote (live) or chart (historical + live)
Support kwargs:
syms: list of symbols, e.g. [BINANCE:ETHUSD]
indicators: list of indicators, e.g. [rsi]
timeframe: int of minutes of chart time frame, e.g. 240 -> 4 hours chart
histbars: int of number of historical data points, e.g. 300
"""
websocket.enableTrace(True)
ws = websocket.WebSocketApp(
self._ws_url,
on_open=lambda ws: self.on_open(ws, type, **kwargs),
on_close=self.on_close,
on_message=lambda ws, message: self.on_message(ws, message),
on_error=self.on_error)
ws.run_forever()
def send_slack(self):
"""
Periodic slack alerts - Indicators
"""
while True:
indicators = self._alerts.get("indicators")
if indicators:
res = pd.DataFrame(indicators).transpose().reset_index()
res.rename(columns={"index": "sym"}, inplace=True)
send_alert(self._slackchannel, [("Indicators", res)])
time.sleep(self._slackfreq)
def on_message(self, ws, message):
pattern = re.compile(r'~m~\d+~m~~h~\d+$')
if pattern.match(message):
ws.send(message)
else:
msg_body = re.compile(r'~m~\d+~m~')
messages = msg_body.split(message)
for msg in messages:
if msg:
parsed_msg = json.loads(msg)
params = parsed_msg.get("p")
if parsed_msg.get("m") == "timescale_update":
# timescale_update -> initial historical data
# TODO: handling of these data for plotting on UI
continue
if parsed_msg.get("m") == "du":
# du -> data update
sym = self._state.get(params[0]).get("sym")
now = datetime.datetime.now().strftime(
'%Y-%m-%d %H:%M:%S')
for k, v in params[1].items():
if v.get("st"):
# study
indicator = k.split("_")[0]
vals = v.get("st")[0].get("v")
val = vals[1]
val_dict = {"dtime": now, indicator: val}
# print({sym: val_dict})
if not self._alerts["indicators"].get(sym):
self._alerts["indicators"][sym] = {}
self._alerts["indicators"][sym][
indicator] = val
elif v.get("s"):
# series
vals = v.get("s")[0].get("v")
val_dict = dict(
zip([
"dtime", "open", "high", "low", "last",
"vol"
], vals))
val_dict["dtime"] = now
# print({sym: val_dict})
if not self._alerts["price"].get(sym):
self._alerts["price"][sym] = {}
self._alerts["price"][sym]["last"] = val_dict[
"last"]
@staticmethod
def on_error(ws, error):
print(error)
@staticmethod
def on_close(ws):
print("### closed ###")
def on_open(self, ws, type: str, **kwargs):
def run(*args, **kwargs):
# ~m~52~m~{"m":"quote_create_session","p":["qs_3bDnffZvz5ur"]}
# ~m~395~m~{"m":"quote_set_fields","p":["qs_3bDnffZvz5ur","ch","chp","lp"]}
# ~m~89~m~{"m":"quote_add_symbols","p":["qs_3bDnffZvz5ur","SP:SPX",{"flags":["force_permission"]}]}
# ~m~315~m~{"m":"quote_fast_symbols","p":["qs_3bDnffZvz5ur","SP:SPX","TVC:NDX","CBOE:VIX","TVC:DXY","SGX:ES3","NASDAQ:AAPL","NASDAQ:MSFT","NASDAQ:TSLA","TVC:USOIL","TVC:GOLD","TVC:SILVER","FX:AUDUSD","FX:EURUSD","FX:GBPUSD","FX:USDJPY","BITSTAMP:BTCUSD","BITSTAMP:ETHUSD","COINBASE:UNIUSD","BINANCE:DOGEUSD","BINANCE:DOTUSD"]}
syms = kwargs.get("syms") or self._syms
timeframe = f'{kwargs.get("timeframe") or self._timeframe}'
indicators = kwargs.get("indicators") or self._indicators
histbars = kwargs.get("histbars") or self._histbars
send = self._send
send(ws, "set_auth_token", ["unauthorized_user_token"])
# Quote session
if not args or (args and args[0] == "quote"):
session = self._gen_session() # Quote session ID
send(ws, "quote_create_session", [session])
send(ws, "quote_set_fields", [session, "lp", "volume"])
[ws.send(self._add_symbol(session, s)) for s in syms]
send(ws, "quote_fast_symbols", [session, *syms])
send(ws, "quote_hibernate_all", [session])
# Chart session - Prefer to use this over quote sessions since it has a historical series
else:
for i, sym in enumerate(syms):
# Each ticker warrants a separate chart session ID
c_session = self._gen_session(type="chart")
self._state[c_session] = {
"sym": sym,
"indicators": [],
"series": [],
"timeframe": timeframe
}
# Users are allowed to select specific tickers | send(ws, "switch_timezone", [c_session, "Asia/Singapore"])
send(ws, "resolve_symbol", [
c_session, f"symbol_{i}",
self._add_chart_symbol(sym)
])
# s (in resp) -> series
self._state[c_session].get("series").append(f"s_{i}")
send(ws, "create_series", [
c_session, f"s_{i}", f"s_{i}", f"symbol_{i}",
timeframe, histbars
])
for indicator in indicators:
# Users are allowed to select specific indicators
# st (in resp) -> study
self._state[c_session].get("indicators").append(
f"{indicator}_{i}")
send(ws, "create_study", [
c_session, f"{indicator}_{i}", f"{indicator}_{i}",
f"s_{i}", "Script@tv-scripting-101!",
self._indicator_mapper(indicator)
])
self._t = threading.Thread(target=run, args=(type, ), kwargs=kwargs)
self._t.setDaemon(True)
self._t.start()
def _send(self, ws, func, params):
""" Client sends msg to websockets server """
ws.send(self._create_msg(func, params))
def _indicator_mapper(self, indicator: str) -> dict:
""" Indicator params that are accepted by the tv server """
return {
"rsi": {
"text":
"1f0fkZ72S0de2geyaUhXXw==_xwY73vljRXeew69Rl27RumLDs6aJ9NLsTYN9Xrht254BTb8uSOgccpLDt/cdRWopwJPNZx40m19yEFwJFswkSi62X4guNJYpXe4A6S9iq2n+OXM6mqWeWzDbjTl0lYmEf1ujbg7i3FvUdV/zCSrqd+iwnvvZSV+O2acpfNLpUlDdB6PZX4Y9y8tlQLWA2PiF8CVJng7DF1LPeecWC4fv+lNg+s5OXU46AjIhc+TFu8DOwiuKjNh7wWz6EZ7gpQS3",
"pineId": "STD;RSI",
"pineVersion": "12.0",
"in_2": {
"v": "",
"f": True,
"t": "resolution"
},
"in_0": {
"v": 14,
"f": True,
"t": "integer | send(ws, "chart_create_session", [c_session, ""]) | random_line_split |
intraday.py |
self._state = {}
self._syms = [
"BINANCE:UNIUSD", "BINANCE:ETHUSD", "BINANCE:DOTUSD", "SGX:ES3",
"SGX:CLR"
]
self._t = None
self._timeframe = 240 # Default to 4 hours chart
self._ws_url = "wss://data.tradingview.com/socket.io/websocket"
def get(self, type: str, **kwargs):
"""
Type is either quote (live) or chart (historical + live)
Support kwargs:
syms: list of symbols, e.g. [BINANCE:ETHUSD]
indicators: list of indicators, e.g. [rsi]
timeframe: int of minutes of chart time frame, e.g. 240 -> 4 hours chart
histbars: int of number of historical data points, e.g. 300
"""
websocket.enableTrace(True)
ws = websocket.WebSocketApp(
self._ws_url,
on_open=lambda ws: self.on_open(ws, type, **kwargs),
on_close=self.on_close,
on_message=lambda ws, message: self.on_message(ws, message),
on_error=self.on_error)
ws.run_forever()
def send_slack(self):
"""
Periodic slack alerts - Indicators
"""
while True:
indicators = self._alerts.get("indicators")
if indicators:
res = pd.DataFrame(indicators).transpose().reset_index()
res.rename(columns={"index": "sym"}, inplace=True)
send_alert(self._slackchannel, [("Indicators", res)])
time.sleep(self._slackfreq)
def on_message(self, ws, message):
pattern = re.compile(r'~m~\d+~m~~h~\d+$')
if pattern.match(message):
ws.send(message)
else:
msg_body = re.compile(r'~m~\d+~m~')
messages = msg_body.split(message)
for msg in messages:
if msg:
parsed_msg = json.loads(msg)
params = parsed_msg.get("p")
if parsed_msg.get("m") == "timescale_update":
# timescale_update -> initial historical data
# TODO: handling of these data for plotting on UI
continue
if parsed_msg.get("m") == "du":
# du -> data update
sym = self._state.get(params[0]).get("sym")
now = datetime.datetime.now().strftime(
'%Y-%m-%d %H:%M:%S')
for k, v in params[1].items():
if v.get("st"):
# study
indicator = k.split("_")[0]
vals = v.get("st")[0].get("v")
val = vals[1]
val_dict = {"dtime": now, indicator: val}
# print({sym: val_dict})
if not self._alerts["indicators"].get(sym):
self._alerts["indicators"][sym] = {}
self._alerts["indicators"][sym][
indicator] = val
elif v.get("s"):
# series
vals = v.get("s")[0].get("v")
val_dict = dict(
zip([
"dtime", "open", "high", "low", "last",
"vol"
], vals))
val_dict["dtime"] = now
# print({sym: val_dict})
if not self._alerts["price"].get(sym):
self._alerts["price"][sym] = {}
self._alerts["price"][sym]["last"] = val_dict[
"last"]
@staticmethod
def on_error(ws, error):
print(error)
@staticmethod
def on_close(ws):
print("### closed ###")
def on_open(self, ws, type: str, **kwargs):
def | (*args, **kwargs):
# ~m~52~m~{"m":"quote_create_session","p":["qs_3bDnffZvz5ur"]}
# ~m~395~m~{"m":"quote_set_fields","p":["qs_3bDnffZvz5ur","ch","chp","lp"]}
# ~m~89~m~{"m":"quote_add_symbols","p":["qs_3bDnffZvz5ur","SP:SPX",{"flags":["force_permission"]}]}
# ~m~315~m~{"m":"quote_fast_symbols","p":["qs_3bDnffZvz5ur","SP:SPX","TVC:NDX","CBOE:VIX","TVC:DXY","SGX:ES3","NASDAQ:AAPL","NASDAQ:MSFT","NASDAQ:TSLA","TVC:USOIL","TVC:GOLD","TVC:SILVER","FX:AUDUSD","FX:EURUSD","FX:GBPUSD","FX:USDJPY","BITSTAMP:BTCUSD","BITSTAMP:ETHUSD","COINBASE:UNIUSD","BINANCE:DOGEUSD","BINANCE:DOTUSD"]}
syms = kwargs.get("syms") or self._syms
timeframe = f'{kwargs.get("timeframe") or self._timeframe}'
indicators = kwargs.get("indicators") or self._indicators
histbars = kwargs.get("histbars") or self._histbars
send = self._send
send(ws, "set_auth_token", ["unauthorized_user_token"])
# Quote session
if not args or (args and args[0] == "quote"):
session = self._gen_session() # Quote session ID
send(ws, "quote_create_session", [session])
send(ws, "quote_set_fields", [session, "lp", "volume"])
[ws.send(self._add_symbol(session, s)) for s in syms]
send(ws, "quote_fast_symbols", [session, *syms])
send(ws, "quote_hibernate_all", [session])
# Chart session - Prefer to use this over quote sessions since it has a historical series
else:
for i, sym in enumerate(syms):
# Each ticker warrants a separate chart session ID
c_session = self._gen_session(type="chart")
self._state[c_session] = {
"sym": sym,
"indicators": [],
"series": [],
"timeframe": timeframe
}
# Users are allowed to select specific tickers
send(ws, "chart_create_session", [c_session, ""])
send(ws, "switch_timezone", [c_session, "Asia/Singapore"])
send(ws, "resolve_symbol", [
c_session, f"symbol_{i}",
self._add_chart_symbol(sym)
])
# s (in resp) -> series
self._state[c_session].get("series").append(f"s_{i}")
send(ws, "create_series", [
c_session, f"s_{i}", f"s_{i}", f"symbol_{i}",
timeframe, histbars
])
for indicator in indicators:
# Users are allowed to select specific indicators
# st (in resp) -> study
self._state[c_session].get("indicators").append(
f"{indicator}_{i}")
send(ws, "create_study", [
c_session, f"{indicator}_{i}", f"{indicator}_{i}",
f"s_{i}", "Script@tv-scripting-101!",
self._indicator_mapper(indicator)
])
self._t = threading.Thread(target=run, args=(type, ), kwargs=kwargs)
self._t.setDaemon(True)
self._t.start()
def _send(self, ws, func, params):
""" Client sends msg to websockets server """
ws.send(self._create_msg(func, params))
def _indicator_mapper(self, indicator: str) -> dict:
""" Indicator params that are accepted by the tv server """
return {
"rsi": {
"text":
"1f0fkZ72S0de2geyaUhXXw==_xwY73vljRXeew69Rl27RumLDs6aJ9NLsTYN9Xrht254BTb8uSOgccpLDt/cdRWopwJPNZx40m19yEFwJFswkSi62X4guNJYpXe4A6S9iq2n+OXM6mqWeWzDbjTl0lYmEf1ujbg7i3FvUdV/zCSrqd+iwnvvZSV+O2acpfNLpUlDdB6PZX4Y9y8tlQLWA2PiF8CVJng7DF1LPeecWC4fv+lNg+s5OXU46AjIhc+TFu8DOwiuKjNh7wWz6EZ7gpQS3",
"pineId": "STD;RSI",
"pineVersion": "12.0",
"in_2": {
"v": "",
"f": True,
"t": "resolution"
},
"in_0": {
"v": 14,
"f": True,
"t": " | run | identifier_name |
intraday.py | "BINANCE:ETHUSD", "BINANCE:DOTUSD", "SGX:ES3",
"SGX:CLR"
]
self._t = None
self._timeframe = 240 # Default to 4 hours chart
self._ws_url = "wss://data.tradingview.com/socket.io/websocket"
def get(self, type: str, **kwargs):
"""
Type is either quote (live) or chart (historical + live)
Support kwargs:
syms: list of symbols, e.g. [BINANCE:ETHUSD]
indicators: list of indicators, e.g. [rsi]
timeframe: int of minutes of chart time frame, e.g. 240 -> 4 hours chart
histbars: int of number of historical data points, e.g. 300
"""
websocket.enableTrace(True)
ws = websocket.WebSocketApp(
self._ws_url,
on_open=lambda ws: self.on_open(ws, type, **kwargs),
on_close=self.on_close,
on_message=lambda ws, message: self.on_message(ws, message),
on_error=self.on_error)
ws.run_forever()
def send_slack(self):
"""
Periodic slack alerts - Indicators
"""
while True:
indicators = self._alerts.get("indicators")
if indicators:
res = pd.DataFrame(indicators).transpose().reset_index()
res.rename(columns={"index": "sym"}, inplace=True)
send_alert(self._slackchannel, [("Indicators", res)])
time.sleep(self._slackfreq)
def on_message(self, ws, message):
pattern = re.compile(r'~m~\d+~m~~h~\d+$')
if pattern.match(message):
ws.send(message)
else:
msg_body = re.compile(r'~m~\d+~m~')
messages = msg_body.split(message)
for msg in messages:
if msg:
parsed_msg = json.loads(msg)
params = parsed_msg.get("p")
if parsed_msg.get("m") == "timescale_update":
# timescale_update -> initial historical data
# TODO: handling of these data for plotting on UI
continue
if parsed_msg.get("m") == "du":
# du -> data update
sym = self._state.get(params[0]).get("sym")
now = datetime.datetime.now().strftime(
'%Y-%m-%d %H:%M:%S')
for k, v in params[1].items():
if v.get("st"):
# study
indicator = k.split("_")[0]
vals = v.get("st")[0].get("v")
val = vals[1]
val_dict = {"dtime": now, indicator: val}
# print({sym: val_dict})
if not self._alerts["indicators"].get(sym):
self._alerts["indicators"][sym] = {}
self._alerts["indicators"][sym][
indicator] = val
elif v.get("s"):
# series
vals = v.get("s")[0].get("v")
val_dict = dict(
zip([
"dtime", "open", "high", "low", "last",
"vol"
], vals))
val_dict["dtime"] = now
# print({sym: val_dict})
if not self._alerts["price"].get(sym):
self._alerts["price"][sym] = {}
self._alerts["price"][sym]["last"] = val_dict[
"last"]
@staticmethod
def on_error(ws, error):
print(error)
@staticmethod
def on_close(ws):
print("### closed ###")
def on_open(self, ws, type: str, **kwargs):
def run(*args, **kwargs):
# ~m~52~m~{"m":"quote_create_session","p":["qs_3bDnffZvz5ur"]}
# ~m~395~m~{"m":"quote_set_fields","p":["qs_3bDnffZvz5ur","ch","chp","lp"]}
# ~m~89~m~{"m":"quote_add_symbols","p":["qs_3bDnffZvz5ur","SP:SPX",{"flags":["force_permission"]}]}
# ~m~315~m~{"m":"quote_fast_symbols","p":["qs_3bDnffZvz5ur","SP:SPX","TVC:NDX","CBOE:VIX","TVC:DXY","SGX:ES3","NASDAQ:AAPL","NASDAQ:MSFT","NASDAQ:TSLA","TVC:USOIL","TVC:GOLD","TVC:SILVER","FX:AUDUSD","FX:EURUSD","FX:GBPUSD","FX:USDJPY","BITSTAMP:BTCUSD","BITSTAMP:ETHUSD","COINBASE:UNIUSD","BINANCE:DOGEUSD","BINANCE:DOTUSD"]}
syms = kwargs.get("syms") or self._syms
timeframe = f'{kwargs.get("timeframe") or self._timeframe}'
indicators = kwargs.get("indicators") or self._indicators
histbars = kwargs.get("histbars") or self._histbars
send = self._send
send(ws, "set_auth_token", ["unauthorized_user_token"])
# Quote session
if not args or (args and args[0] == "quote"):
session = self._gen_session() # Quote session ID
send(ws, "quote_create_session", [session])
send(ws, "quote_set_fields", [session, "lp", "volume"])
[ws.send(self._add_symbol(session, s)) for s in syms]
send(ws, "quote_fast_symbols", [session, *syms])
send(ws, "quote_hibernate_all", [session])
# Chart session - Prefer to use this over quote sessions since it has a historical series
else:
for i, sym in enumerate(syms):
# Each ticker warrants a separate chart session ID
c_session = self._gen_session(type="chart")
self._state[c_session] = {
"sym": sym,
"indicators": [],
"series": [],
"timeframe": timeframe
}
# Users are allowed to select specific tickers
send(ws, "chart_create_session", [c_session, ""])
send(ws, "switch_timezone", [c_session, "Asia/Singapore"])
send(ws, "resolve_symbol", [
c_session, f"symbol_{i}",
self._add_chart_symbol(sym)
])
# s (in resp) -> series
self._state[c_session].get("series").append(f"s_{i}")
send(ws, "create_series", [
c_session, f"s_{i}", f"s_{i}", f"symbol_{i}",
timeframe, histbars
])
for indicator in indicators:
# Users are allowed to select specific indicators
# st (in resp) -> study
self._state[c_session].get("indicators").append(
f"{indicator}_{i}")
send(ws, "create_study", [
c_session, f"{indicator}_{i}", f"{indicator}_{i}",
f"s_{i}", "Script@tv-scripting-101!",
self._indicator_mapper(indicator)
])
self._t = threading.Thread(target=run, args=(type, ), kwargs=kwargs)
self._t.setDaemon(True)
self._t.start()
def _send(self, ws, func, params):
""" Client sends msg to websockets server """
ws.send(self._create_msg(func, params))
def _indicator_mapper(self, indicator: str) -> dict:
| """ Indicator params that are accepted by the tv server """
return {
"rsi": {
"text":
"1f0fkZ72S0de2geyaUhXXw==_xwY73vljRXeew69Rl27RumLDs6aJ9NLsTYN9Xrht254BTb8uSOgccpLDt/cdRWopwJPNZx40m19yEFwJFswkSi62X4guNJYpXe4A6S9iq2n+OXM6mqWeWzDbjTl0lYmEf1ujbg7i3FvUdV/zCSrqd+iwnvvZSV+O2acpfNLpUlDdB6PZX4Y9y8tlQLWA2PiF8CVJng7DF1LPeecWC4fv+lNg+s5OXU46AjIhc+TFu8DOwiuKjNh7wWz6EZ7gpQS3",
"pineId": "STD;RSI",
"pineVersion": "12.0",
"in_2": {
"v": "",
"f": True,
"t": "resolution"
},
"in_0": {
"v": 14,
"f": True,
"t": "integer"
},
"in_1": {
"v": "close",
"f": True, | identifier_body | |
conv2d_compact.rs | = Vec::new();
ext.resize((packed_size + k_col - 1) as usize, T::zero());
for k in 0..(packed_size + k_col - 1) * bit_length {
ext[(k / bit_length) as usize] += T::from_i32(((val[(k/8) as usize] >> (k % 8)) & 1) as i32) * power_of_two(k % bit_length);
}
for k in 0..packed_size + k_col - 1 {
let idx = c * n_packed + k;
if idx >= row_dim {
break
}
var_dict[mem[extracted].at_idx(&[layer_out,r,idx]) as usize] = ext[k as usize] - offset;
}
}
}
}
}
pub fn conv2d_compact(&mut self, input: TensorAddress, output: TensorAddress, weight_rev: TensorAddress, bias: Option<(TensorAddress, u32)>, bit_length: u8, act: ActivationFunction) {
// packing weight
let dim = &self.mem[weight_rev].dim;
let (fout, fin, k_row, k_col) = (dim[0], dim[1], dim[2], dim[3]);
let packed_weight = self.mem.alloc(&[fout, fin, k_row]);
assert!(k_col * (bit_length as u32) <= SCALAR_SIZE);
self.packing_tensor(weight_rev, packed_weight, bit_length, k_col as u8,1, BigScalar::one(), true);
let (row, col) = (self.mem[input].dim[1], self.mem[input].dim[2]);
let packed_size = min((SCALAR_SIZE / (bit_length as u32)).checked_sub(k_col).unwrap(),col);
let col_packed = (col-1)/packed_size + 1;
let packed_layer = self.mem.alloc(&[fin, row, col_packed]);
// packing row of inputs
self.packing_tensor_by_dim(input,&[-1], packed_layer, bit_length, packed_size as u8,1,BigScalar::one(), true);
// splicing output by row
let mut mul_input = Vec::new();
for r in 0..row - k_row + 1 { | }
mul_input.push(mul_input_row);
}
//packing bias
let mut packed_bias: Vec<Vec<TensorAddress>> = Vec::with_capacity(fout as usize);
let mut bias_dim = 0;
let mut bias_scale = 0;
if let Some((b, scale)) = bias {
bias_dim = (col - k_col)/packed_size + 1;
bias_scale = scale;
for layer_out in 0..fout {
let mut packed_bias_row: Vec<TensorAddress> = Vec::with_capacity(((row - k_row)/scale + 1) as usize);
for r in 0..(row - k_row)/scale + 1 {
let packed_bias = self.mem.alloc(&[bias_dim]);
let bias_row = self.mem.save(self.mem[b].at_(&[layer_out, r]));
self.packing_tensor(bias_row, packed_bias, bit_length, packed_size as u8, scale,power_of_two(bit_length as u32 * (k_col - 1)), true);
packed_bias_row.push(packed_bias);
}
packed_bias.push(packed_bias_row);
}
}
let mul_result = self.mem.alloc(&[fout, row - k_row + 1, col_packed]);
for layer_out in 0..fout {
let packed_weight = self.mem.save(self.mem[packed_weight].at_(&[layer_out]));
for r in 0..row - k_row + 1 {
for c in 0..col_packed {
let cur_bias = if c < bias_dim {Some(self.mem[packed_bias[layer_out as usize][(r/bias_scale) as usize]].at_idx(&[c]))} else {None};
self.dot(mul_input[r as usize][c as usize], packed_weight, self.mem[mul_result].at_idx(&[layer_out, r, c]), cur_bias);
}
}
}
// sign extraction
let n_packed = packed_size + k_col - 1;
let extracted_length = (col_packed - 1) * n_packed + ((col-1) % packed_size) + k_col;
let extracted = self.mem.alloc(&[fout, row - k_row + 1, extracted_length]);
self.packing_tensor_by_dim(extracted,&[-1], mul_result, bit_length, n_packed as u8,1,BigScalar::one(), false);
let params = vec![mul_result, k_col, packed_size, bit_length as u32, extracted];
self.compute.push((params.into_boxed_slice(), Functions::ConvCompact));
fn split_tensor<const N:usize>(mem: &mut MemoryManager,tensor: TensorAddress, length: u32, pos: [u32; N]) -> [(Option<TensorAddress>, Option<TensorAddress>); N] {
let fully_packed = mem[tensor].dim[2]/length;
let remainder = mem[tensor].dim[2] % length;
// should not save this
let tmp=mem[tensor].partition(2, length);
let mut res: [(Option<TensorAddress>, Option<TensorAddress>); N] = [(None, None); N];
for i in 0..N - 1 {
if pos[i] == pos[i+1] {
res[i] = (None, None);
continue;
}
let n= fully_packed + if remainder >= pos[i+1] {1} else {0};
let full = if n > 0 {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), RangeTo(..n), Range(pos[i]..pos[i+1])])))
} else {
None
};
let rem = if pos[i] < remainder && remainder < pos[i+1] {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), Id(n), Range(pos[i]..remainder)])))
} else {
None
};
res[i] = (full, rem);
}
res
}
fn extract_sign_part(c: &mut ConstraintSystem, extracted: TensorAddress, bit_length: u8) {
let output = c.mem.alloc(&c.mem[extracted].dim.to_owned());
c.sign(extracted, output, bit_length - 1);
}
let reduced_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeTo(..extracted_length - k_col + 1)]));
if k_col != 1 {
let rem_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeFrom(extracted_length - k_col + 1..)]));
extract_sign_part(self, rem_extract, bit_length);
}
let [(output_full, output_full_rem), (output_part, output_part_rem), (_,_)]= split_tensor(&mut self.mem, output, packed_size, [0, packed_size-(k_col-1), packed_size]);
let [(ext_left, ext_left_rem), (ext_full, ext_full_rem), (ext_right,ext_right_rem), (_,_)]= split_tensor(&mut self.mem, reduced_extract, n_packed, [0, k_col-1, packed_size, n_packed]);
// extract the fully correct part
if let Some(e) = ext_full {
self.activation(e, output_full.unwrap(), bit_length - 1, act);
}
if let Some(e) = ext_full_rem {
self.activation(e, output_full_rem.unwrap(), bit_length - 1, act);
}
//extract left and right sign part
if let Some(e) = ext_left {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_left_rem {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_right {
extract_sign_part(self,e, bit_length);
}
assert_eq!(ext_right_rem, None);
if let Some(left_rem) = ext_left_rem {
if let Some(right) = ext_right {
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[right].dim[2] - 1, k_col - 1]);
let left = self.mem.save(self.mem[ext_left.unwrap()].at(&[RangeFull(), RangeFrom(1..)]));
self.sum_two(right, left, sum_res);
self.activation(sum_res, output_part.unwrap(), bit_length - 1, act);
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[left_rem].dim[2]]);
let right_rem = self.mem.save(self.mem[right].at(&[RangeFull(), Id(self.mem[right].dim[2] - 1), RangeTo(..self.mem[left_rem].dim[2])]));
self.sum_two(right_rem, | let mut mul_input_row = Vec::new();
for c in 0..col_packed {
mul_input_row.push(self.mem.save(self.mem[packed_layer].at(&[RangeFull(), Range(r..r + k_row), Id(c)]))); | random_line_split |
conv2d_compact.rs | = Vec::new();
ext.resize((packed_size + k_col - 1) as usize, T::zero());
for k in 0..(packed_size + k_col - 1) * bit_length {
ext[(k / bit_length) as usize] += T::from_i32(((val[(k/8) as usize] >> (k % 8)) & 1) as i32) * power_of_two(k % bit_length);
}
for k in 0..packed_size + k_col - 1 {
let idx = c * n_packed + k;
if idx >= row_dim {
break
}
var_dict[mem[extracted].at_idx(&[layer_out,r,idx]) as usize] = ext[k as usize] - offset;
}
}
}
}
}
pub fn conv2d_compact(&mut self, input: TensorAddress, output: TensorAddress, weight_rev: TensorAddress, bias: Option<(TensorAddress, u32)>, bit_length: u8, act: ActivationFunction) | for r in 0..row - k_row + 1 {
let mut mul_input_row = Vec::new();
for c in 0..col_packed {
mul_input_row.push(self.mem.save(self.mem[packed_layer].at(&[RangeFull(), Range(r..r + k_row), Id(c)])));
}
mul_input.push(mul_input_row);
}
//packing bias
let mut packed_bias: Vec<Vec<TensorAddress>> = Vec::with_capacity(fout as usize);
let mut bias_dim = 0;
let mut bias_scale = 0;
if let Some((b, scale)) = bias {
bias_dim = (col - k_col)/packed_size + 1;
bias_scale = scale;
for layer_out in 0..fout {
let mut packed_bias_row: Vec<TensorAddress> = Vec::with_capacity(((row - k_row)/scale + 1) as usize);
for r in 0..(row - k_row)/scale + 1 {
let packed_bias = self.mem.alloc(&[bias_dim]);
let bias_row = self.mem.save(self.mem[b].at_(&[layer_out, r]));
self.packing_tensor(bias_row, packed_bias, bit_length, packed_size as u8, scale,power_of_two(bit_length as u32 * (k_col - 1)), true);
packed_bias_row.push(packed_bias);
}
packed_bias.push(packed_bias_row);
}
}
let mul_result = self.mem.alloc(&[fout, row - k_row + 1, col_packed]);
for layer_out in 0..fout {
let packed_weight = self.mem.save(self.mem[packed_weight].at_(&[layer_out]));
for r in 0..row - k_row + 1 {
for c in 0..col_packed {
let cur_bias = if c < bias_dim {Some(self.mem[packed_bias[layer_out as usize][(r/bias_scale) as usize]].at_idx(&[c]))} else {None};
self.dot(mul_input[r as usize][c as usize], packed_weight, self.mem[mul_result].at_idx(&[layer_out, r, c]), cur_bias);
}
}
}
// sign extraction
let n_packed = packed_size + k_col - 1;
let extracted_length = (col_packed - 1) * n_packed + ((col-1) % packed_size) + k_col;
let extracted = self.mem.alloc(&[fout, row - k_row + 1, extracted_length]);
self.packing_tensor_by_dim(extracted,&[-1], mul_result, bit_length, n_packed as u8,1,BigScalar::one(), false);
let params = vec![mul_result, k_col, packed_size, bit_length as u32, extracted];
self.compute.push((params.into_boxed_slice(), Functions::ConvCompact));
fn split_tensor<const N:usize>(mem: &mut MemoryManager,tensor: TensorAddress, length: u32, pos: [u32; N]) -> [(Option<TensorAddress>, Option<TensorAddress>); N] {
let fully_packed = mem[tensor].dim[2]/length;
let remainder = mem[tensor].dim[2] % length;
// should not save this
let tmp=mem[tensor].partition(2, length);
let mut res: [(Option<TensorAddress>, Option<TensorAddress>); N] = [(None, None); N];
for i in 0..N - 1 {
if pos[i] == pos[i+1] {
res[i] = (None, None);
continue;
}
let n= fully_packed + if remainder >= pos[i+1] {1} else {0};
let full = if n > 0 {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), RangeTo(..n), Range(pos[i]..pos[i+1])])))
} else {
None
};
let rem = if pos[i] < remainder && remainder < pos[i+1] {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), Id(n), Range(pos[i]..remainder)])))
} else {
None
};
res[i] = (full, rem);
}
res
}
fn extract_sign_part(c: &mut ConstraintSystem, extracted: TensorAddress, bit_length: u8) {
let output = c.mem.alloc(&c.mem[extracted].dim.to_owned());
c.sign(extracted, output, bit_length - 1);
}
let reduced_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeTo(..extracted_length - k_col + 1)]));
if k_col != 1 {
let rem_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeFrom(extracted_length - k_col + 1..)]));
extract_sign_part(self, rem_extract, bit_length);
}
let [(output_full, output_full_rem), (output_part, output_part_rem), (_,_)]= split_tensor(&mut self.mem, output, packed_size, [0, packed_size-(k_col-1), packed_size]);
let [(ext_left, ext_left_rem), (ext_full, ext_full_rem), (ext_right,ext_right_rem), (_,_)]= split_tensor(&mut self.mem, reduced_extract, n_packed, [0, k_col-1, packed_size, n_packed]);
// extract the fully correct part
if let Some(e) = ext_full {
self.activation(e, output_full.unwrap(), bit_length - 1, act);
}
if let Some(e) = ext_full_rem {
self.activation(e, output_full_rem.unwrap(), bit_length - 1, act);
}
//extract left and right sign part
if let Some(e) = ext_left {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_left_rem {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_right {
extract_sign_part(self,e, bit_length);
}
assert_eq!(ext_right_rem, None);
if let Some(left_rem) = ext_left_rem {
if let Some(right) = ext_right {
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[right].dim[2] - 1, k_col - 1]);
let left = self.mem.save(self.mem[ext_left.unwrap()].at(&[RangeFull(), RangeFrom(1..)]));
self.sum_two(right, left, sum_res);
self.activation(sum_res, output_part.unwrap(), bit_length - 1, act);
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[left_rem].dim[2]]);
let right_rem = self.mem.save(self.mem[right].at(&[RangeFull(), Id(self.mem[right].dim[2] - 1), RangeTo(..self.mem[left_rem].dim[2])]));
self.sum_two(right_rem | {
// packing weight
let dim = &self.mem[weight_rev].dim;
let (fout, fin, k_row, k_col) = (dim[0], dim[1], dim[2], dim[3]);
let packed_weight = self.mem.alloc(&[fout, fin, k_row]);
assert!(k_col * (bit_length as u32) <= SCALAR_SIZE);
self.packing_tensor(weight_rev, packed_weight, bit_length, k_col as u8,1, BigScalar::one(), true);
let (row, col) = (self.mem[input].dim[1], self.mem[input].dim[2]);
let packed_size = min((SCALAR_SIZE / (bit_length as u32)).checked_sub(k_col).unwrap(),col);
let col_packed = (col-1)/packed_size + 1;
let packed_layer = self.mem.alloc(&[fin, row, col_packed]);
// packing row of inputs
self.packing_tensor_by_dim(input,&[-1], packed_layer, bit_length, packed_size as u8,1,BigScalar::one(), true);
// splicing output by row
let mut mul_input = Vec::new(); | identifier_body |
conv2d_compact.rs | + k_row), Id(c)])));
}
mul_input.push(mul_input_row);
}
//packing bias
let mut packed_bias: Vec<Vec<TensorAddress>> = Vec::with_capacity(fout as usize);
let mut bias_dim = 0;
let mut bias_scale = 0;
if let Some((b, scale)) = bias {
bias_dim = (col - k_col)/packed_size + 1;
bias_scale = scale;
for layer_out in 0..fout {
let mut packed_bias_row: Vec<TensorAddress> = Vec::with_capacity(((row - k_row)/scale + 1) as usize);
for r in 0..(row - k_row)/scale + 1 {
let packed_bias = self.mem.alloc(&[bias_dim]);
let bias_row = self.mem.save(self.mem[b].at_(&[layer_out, r]));
self.packing_tensor(bias_row, packed_bias, bit_length, packed_size as u8, scale,power_of_two(bit_length as u32 * (k_col - 1)), true);
packed_bias_row.push(packed_bias);
}
packed_bias.push(packed_bias_row);
}
}
let mul_result = self.mem.alloc(&[fout, row - k_row + 1, col_packed]);
for layer_out in 0..fout {
let packed_weight = self.mem.save(self.mem[packed_weight].at_(&[layer_out]));
for r in 0..row - k_row + 1 {
for c in 0..col_packed {
let cur_bias = if c < bias_dim {Some(self.mem[packed_bias[layer_out as usize][(r/bias_scale) as usize]].at_idx(&[c]))} else {None};
self.dot(mul_input[r as usize][c as usize], packed_weight, self.mem[mul_result].at_idx(&[layer_out, r, c]), cur_bias);
}
}
}
// sign extraction
let n_packed = packed_size + k_col - 1;
let extracted_length = (col_packed - 1) * n_packed + ((col-1) % packed_size) + k_col;
let extracted = self.mem.alloc(&[fout, row - k_row + 1, extracted_length]);
self.packing_tensor_by_dim(extracted,&[-1], mul_result, bit_length, n_packed as u8,1,BigScalar::one(), false);
let params = vec![mul_result, k_col, packed_size, bit_length as u32, extracted];
self.compute.push((params.into_boxed_slice(), Functions::ConvCompact));
fn split_tensor<const N:usize>(mem: &mut MemoryManager,tensor: TensorAddress, length: u32, pos: [u32; N]) -> [(Option<TensorAddress>, Option<TensorAddress>); N] {
let fully_packed = mem[tensor].dim[2]/length;
let remainder = mem[tensor].dim[2] % length;
// should not save this
let tmp=mem[tensor].partition(2, length);
let mut res: [(Option<TensorAddress>, Option<TensorAddress>); N] = [(None, None); N];
for i in 0..N - 1 {
if pos[i] == pos[i+1] {
res[i] = (None, None);
continue;
}
let n= fully_packed + if remainder >= pos[i+1] {1} else {0};
let full = if n > 0 {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), RangeTo(..n), Range(pos[i]..pos[i+1])])))
} else {
None
};
let rem = if pos[i] < remainder && remainder < pos[i+1] {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), Id(n), Range(pos[i]..remainder)])))
} else {
None
};
res[i] = (full, rem);
}
res
}
fn extract_sign_part(c: &mut ConstraintSystem, extracted: TensorAddress, bit_length: u8) {
let output = c.mem.alloc(&c.mem[extracted].dim.to_owned());
c.sign(extracted, output, bit_length - 1);
}
let reduced_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeTo(..extracted_length - k_col + 1)]));
if k_col != 1 {
let rem_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeFrom(extracted_length - k_col + 1..)]));
extract_sign_part(self, rem_extract, bit_length);
}
let [(output_full, output_full_rem), (output_part, output_part_rem), (_,_)]= split_tensor(&mut self.mem, output, packed_size, [0, packed_size-(k_col-1), packed_size]);
let [(ext_left, ext_left_rem), (ext_full, ext_full_rem), (ext_right,ext_right_rem), (_,_)]= split_tensor(&mut self.mem, reduced_extract, n_packed, [0, k_col-1, packed_size, n_packed]);
// extract the fully correct part
if let Some(e) = ext_full {
self.activation(e, output_full.unwrap(), bit_length - 1, act);
}
if let Some(e) = ext_full_rem {
self.activation(e, output_full_rem.unwrap(), bit_length - 1, act);
}
//extract left and right sign part
if let Some(e) = ext_left {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_left_rem {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_right {
extract_sign_part(self,e, bit_length);
}
assert_eq!(ext_right_rem, None);
if let Some(left_rem) = ext_left_rem {
if let Some(right) = ext_right {
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[right].dim[2] - 1, k_col - 1]);
let left = self.mem.save(self.mem[ext_left.unwrap()].at(&[RangeFull(), RangeFrom(1..)]));
self.sum_two(right, left, sum_res);
self.activation(sum_res, output_part.unwrap(), bit_length - 1, act);
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[left_rem].dim[2]]);
let right_rem = self.mem.save(self.mem[right].at(&[RangeFull(), Id(self.mem[right].dim[2] - 1), RangeTo(..self.mem[left_rem].dim[2])]));
self.sum_two(right_rem, left_rem, sum_res);
self.activation(sum_res, output_part_rem.unwrap(), bit_length - 1, act);
}
}
}
}
#[cfg(test)]
mod tests {
use super::*;
use crate::scalar::slice_to_scalar;
#[test]
fn conv2d_compact_test() {
let mut x = ConstraintSystem::new();
let input = x.mem.alloc(&[2,5,5]);
let weight = x.mem.alloc(&[2,2,3,3]);
let output = x.mem.alloc(&[2,3,3]);
let bias = x.mem.alloc(&[2,3,3]);
let weight_rev = x.mem.save(x.mem[weight].reverse(3));
x.conv2d_compact(input, output, weight_rev, Some((bias, 1)), 7, ActivationFunction::Sign);
let mut mem: Vec<BigScalar> = slice_to_scalar(&[1,0,1,-1,0,0,0,-2,4,-1,-4,0,3,-4,0,0,0,1,-1,1,-4,2,3,-1,0,-4,2,2,-3,-1,-1,1,2,-1,1,4,4,2,3,-3,0,3,-2,3,0,2,3,3,-2,2,4,3,3,-4,-4,-1,3,1,4,-2,-2,0,-2,4,-3,0,0,0,-2,0,0,0,0,3,4,-3,-4,-1,-1,-4,3,1,-2,0,0,0,-1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,-3,0,-3,0,1,-4,-1,2,0,0,-4,2,1,3,2,-3,4,-3]);
mem.resize(x.mem.n_var as usize, Scalar::zero());
x.compute(&mut mem);
assert_eq!(mem[87..87+18], slice_to_scalar(&[1,1,-1,-1,-1,1,1,-1,-1,-1,1,-1,-1,-1,1,-1,-1,-1]));
x.sort_cons();
assert!(x.verify(&mem));
}
#[test]
fn | conv2d_compact_test_small | identifier_name | |
conv2d_compact.rs | = Vec::new();
ext.resize((packed_size + k_col - 1) as usize, T::zero());
for k in 0..(packed_size + k_col - 1) * bit_length {
ext[(k / bit_length) as usize] += T::from_i32(((val[(k/8) as usize] >> (k % 8)) & 1) as i32) * power_of_two(k % bit_length);
}
for k in 0..packed_size + k_col - 1 {
let idx = c * n_packed + k;
if idx >= row_dim {
break
}
var_dict[mem[extracted].at_idx(&[layer_out,r,idx]) as usize] = ext[k as usize] - offset;
}
}
}
}
}
pub fn conv2d_compact(&mut self, input: TensorAddress, output: TensorAddress, weight_rev: TensorAddress, bias: Option<(TensorAddress, u32)>, bit_length: u8, act: ActivationFunction) {
// packing weight
let dim = &self.mem[weight_rev].dim;
let (fout, fin, k_row, k_col) = (dim[0], dim[1], dim[2], dim[3]);
let packed_weight = self.mem.alloc(&[fout, fin, k_row]);
assert!(k_col * (bit_length as u32) <= SCALAR_SIZE);
self.packing_tensor(weight_rev, packed_weight, bit_length, k_col as u8,1, BigScalar::one(), true);
let (row, col) = (self.mem[input].dim[1], self.mem[input].dim[2]);
let packed_size = min((SCALAR_SIZE / (bit_length as u32)).checked_sub(k_col).unwrap(),col);
let col_packed = (col-1)/packed_size + 1;
let packed_layer = self.mem.alloc(&[fin, row, col_packed]);
// packing row of inputs
self.packing_tensor_by_dim(input,&[-1], packed_layer, bit_length, packed_size as u8,1,BigScalar::one(), true);
// splicing output by row
let mut mul_input = Vec::new();
for r in 0..row - k_row + 1 {
let mut mul_input_row = Vec::new();
for c in 0..col_packed {
mul_input_row.push(self.mem.save(self.mem[packed_layer].at(&[RangeFull(), Range(r..r + k_row), Id(c)])));
}
mul_input.push(mul_input_row);
}
//packing bias
let mut packed_bias: Vec<Vec<TensorAddress>> = Vec::with_capacity(fout as usize);
let mut bias_dim = 0;
let mut bias_scale = 0;
if let Some((b, scale)) = bias {
bias_dim = (col - k_col)/packed_size + 1;
bias_scale = scale;
for layer_out in 0..fout {
let mut packed_bias_row: Vec<TensorAddress> = Vec::with_capacity(((row - k_row)/scale + 1) as usize);
for r in 0..(row - k_row)/scale + 1 {
let packed_bias = self.mem.alloc(&[bias_dim]);
let bias_row = self.mem.save(self.mem[b].at_(&[layer_out, r]));
self.packing_tensor(bias_row, packed_bias, bit_length, packed_size as u8, scale,power_of_two(bit_length as u32 * (k_col - 1)), true);
packed_bias_row.push(packed_bias);
}
packed_bias.push(packed_bias_row);
}
}
let mul_result = self.mem.alloc(&[fout, row - k_row + 1, col_packed]);
for layer_out in 0..fout {
let packed_weight = self.mem.save(self.mem[packed_weight].at_(&[layer_out]));
for r in 0..row - k_row + 1 {
for c in 0..col_packed {
let cur_bias = if c < bias_dim {Some(self.mem[packed_bias[layer_out as usize][(r/bias_scale) as usize]].at_idx(&[c]))} else {None};
self.dot(mul_input[r as usize][c as usize], packed_weight, self.mem[mul_result].at_idx(&[layer_out, r, c]), cur_bias);
}
}
}
// sign extraction
let n_packed = packed_size + k_col - 1;
let extracted_length = (col_packed - 1) * n_packed + ((col-1) % packed_size) + k_col;
let extracted = self.mem.alloc(&[fout, row - k_row + 1, extracted_length]);
self.packing_tensor_by_dim(extracted,&[-1], mul_result, bit_length, n_packed as u8,1,BigScalar::one(), false);
let params = vec![mul_result, k_col, packed_size, bit_length as u32, extracted];
self.compute.push((params.into_boxed_slice(), Functions::ConvCompact));
fn split_tensor<const N:usize>(mem: &mut MemoryManager,tensor: TensorAddress, length: u32, pos: [u32; N]) -> [(Option<TensorAddress>, Option<TensorAddress>); N] {
let fully_packed = mem[tensor].dim[2]/length;
let remainder = mem[tensor].dim[2] % length;
// should not save this
let tmp=mem[tensor].partition(2, length);
let mut res: [(Option<TensorAddress>, Option<TensorAddress>); N] = [(None, None); N];
for i in 0..N - 1 {
if pos[i] == pos[i+1] {
res[i] = (None, None);
continue;
}
let n= fully_packed + if remainder >= pos[i+1] {1} else {0};
let full = if n > 0 {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), RangeTo(..n), Range(pos[i]..pos[i+1])])))
} else {
None
};
let rem = if pos[i] < remainder && remainder < pos[i+1] | else {
None
};
res[i] = (full, rem);
}
res
}
fn extract_sign_part(c: &mut ConstraintSystem, extracted: TensorAddress, bit_length: u8) {
let output = c.mem.alloc(&c.mem[extracted].dim.to_owned());
c.sign(extracted, output, bit_length - 1);
}
let reduced_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeTo(..extracted_length - k_col + 1)]));
if k_col != 1 {
let rem_extract = self.mem.save(self.mem[extracted].at(&[RangeFull(), RangeFull(), RangeFrom(extracted_length - k_col + 1..)]));
extract_sign_part(self, rem_extract, bit_length);
}
let [(output_full, output_full_rem), (output_part, output_part_rem), (_,_)]= split_tensor(&mut self.mem, output, packed_size, [0, packed_size-(k_col-1), packed_size]);
let [(ext_left, ext_left_rem), (ext_full, ext_full_rem), (ext_right,ext_right_rem), (_,_)]= split_tensor(&mut self.mem, reduced_extract, n_packed, [0, k_col-1, packed_size, n_packed]);
// extract the fully correct part
if let Some(e) = ext_full {
self.activation(e, output_full.unwrap(), bit_length - 1, act);
}
if let Some(e) = ext_full_rem {
self.activation(e, output_full_rem.unwrap(), bit_length - 1, act);
}
//extract left and right sign part
if let Some(e) = ext_left {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_left_rem {
extract_sign_part(self,e, bit_length);
}
if let Some(e) = ext_right {
extract_sign_part(self,e, bit_length);
}
assert_eq!(ext_right_rem, None);
if let Some(left_rem) = ext_left_rem {
if let Some(right) = ext_right {
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[right].dim[2] - 1, k_col - 1]);
let left = self.mem.save(self.mem[ext_left.unwrap()].at(&[RangeFull(), RangeFrom(1..)]));
self.sum_two(right, left, sum_res);
self.activation(sum_res, output_part.unwrap(), bit_length - 1, act);
let sum_res = self.mem.alloc(&[fout, row - k_row + 1, self.mem[left_rem].dim[2]]);
let right_rem = self.mem.save(self.mem[right].at(&[RangeFull(), Id(self.mem[right].dim[2] - 1), RangeTo(..self.mem[left_rem].dim[2])]));
self.sum_two(right_rem | {
Some(mem.save(tmp.at(&[RangeFull(), RangeFull(), Id(n), Range(pos[i]..remainder)])))
} | conditional_block |
director_test.go | //
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package receiver
import (
"fmt"
"log"
"math"
"os"
"strings"
"sync"
"testing"
"time"
"github.com/hashicorp/memberlist"
"github.com/tgres/tgres/cluster"
"github.com/tgres/tgres/rrd"
"github.com/tgres/tgres/serde"
)
type fakeLogger struct {
last []byte
}
func (f *fakeLogger) Write(p []byte) (n int, err error) {
f.last = p
return len(p), nil
}
func Test_directorIncomingDPMessages(t *testing.T) {
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
fl := &fakeLogger{}
log.SetOutput(fl)
rcv := make(chan *cluster.Msg)
dpCh := make(chan *IncomingDP)
count := 0
go func() {
for {
if _, ok := <-dpCh; !ok {
break
}
count++
}
}()
go directorIncomingDPMessages(rcv, dpCh)
// Sending a bogus message should not cause anything be written to dpCh
rcv <- &cluster.Msg{}
rcv <- &cluster.Msg{} // second send ensures the loop has gone full circle
if count > 0 {
t.Errorf("Malformed messages should not cause data points, count: %d", count)
}
if !strings.Contains(string(fl.last), "decoding FAILED") {
t.Errorf("Malformed messages should log 'decoding FAILED'")
}
// now we need a real message
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
m, _ := cluster.NewMsg(&cluster.Node{}, dp)
rcv <- m
rcv <- m
if count < 1 {
t.Errorf("At least 1 data point should have been sent to dpCh")
}
dp.Hops = 1000 // exceed maxhops (which in fakeCluster is 0?)
m, _ = cluster.NewMsg(&cluster.Node{}, dp)
rcv <- m // "clear" the loop
count = 0
rcv <- m
rcv <- m
if count > 0 {
t.Errorf("Hops exceeded should not cause data points, count: %d", count)
}
if !strings.Contains(string(fl.last), "max hops") {
t.Errorf("Hops exceeded messages should log 'max hops'")
}
// Closing the dpCh should cause the recover() to happen
// The test here is that it doesn't panic
close(dpCh)
dp.Hops = 0
m, _ = cluster.NewMsg(&cluster.Node{}, dp)
rcv <- m
// Closing the channel exists (not sure how to really test for that)
go directorIncomingDPMessages(rcv, dpCh)
close(rcv)
}
func Test_directorForwardDPToNode(t *testing.T) {
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
md := make([]byte, 20)
md[0] = 1 // Ready
node := &cluster.Node{Node: &memberlist.Node{Meta: md}}
snd := make(chan *cluster.Msg)
count := 0
go func() {
for {
if _, ok := <-snd; !ok {
break
}
count++
}
}()
// if hops is > 0, nothing happens
dp.Hops = 1
directorForwardDPToNode(dp, node, snd)
directorForwardDPToNode(dp, node, snd)
if count > 0 {
t.Errorf("directorForwardDPToNode: Data points with hops > 0 should not be forwarded")
}
// otherwise it should work
dp.Hops = 0
directorForwardDPToNode(dp, node, snd)
dp.Hops = 0 // because it just got incremented
directorForwardDPToNode(dp, node, snd)
if count < 1 {
t.Errorf("Data point not sent to channel?")
}
// mark node not Ready
md[0] = 0
dp.Hops = 0 // because it just got incremented
if err := directorForwardDPToNode(dp, node, snd); err == nil {
t.Errorf("not ready node should cause an error")
}
}
func Test_directorProcessOrForward(t *testing.T) {
saveFn := directorForwardDPToNode
forward, fwErr := 0, error(nil)
directorForwardDPToNode = func(dp *IncomingDP, node *cluster.Node, snd chan *cluster.Msg) error {
forward++
return fwErr
}
// dsc
db := &fakeSerde{}
df := &SimpleDSFinder{DftDSSPec}
sr := &fakeSr{}
dsf := &dsFlusher{db: db, sr: sr}
dsc := newDsCache(db, df, dsf)
// rds
foo := serde.Ident{"name": "foo"}
ds := serde.NewDbDataSource(0, foo, rrd.NewDataSource(*DftDSSPec))
rds := &cachedDs{DbDataSourcer: ds}
// cluster
clstr := &fakeCluster{}
md := make([]byte, 20)
md[0] = 1 // Ready
node := &cluster.Node{Node: &memberlist.Node{Meta: md, Name: "local"}}
clstr.nodesForDd = []*cluster.Node{node}
clstr.ln = node
// workerChs
workerChs := make([]chan *incomingDpWithDs, 1)
workerChs[0] = make(chan *incomingDpWithDs)
sent := 0
go func() {
for {
<-workerChs[0]
sent++
}
}()
// Test if we are LocalNode
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if sent < 1 {
t.Errorf("directorProcessOrForward: Nothing sent to workerChs")
}
// Now test we are NOT LN, forward
remote := &cluster.Node{Node: &memberlist.Node{Meta: md, Name: "remote"}}
clstr.nodesForDd = []*cluster.Node{remote}
n := directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if forward != 1 {
t.Errorf("directorProcessOrForward: directorForwardDPToNode not called")
}
if n != 1 {
t.Errorf("directorProcessOrForward: return value != 1")
}
fl := &fakeLogger{}
log.SetOutput(fl)
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
fwErr = fmt.Errorf("some error")
n = directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if n != 0 {
t.Errorf("directorProcessOrForward: return value != 0")
}
if !strings.Contains(string(fl.last), "some error") {
t.Errorf("directorProcessOrForward: directorForwardDPToNode not logged")
}
fwErr = nil
// make an rds with points
foo = serde.Ident{"name": "foo"}
ds = serde.NewDbDataSource(0, foo, rrd.NewDataSource(rrd.DSSpec{
Step: 10 * time.Second,
RRAs: []rrd.RRASpec{
rrd.RRASpec{Function: rrd.WMEAN,
Step: 10 * time.Second,
Span: 30 * time.Second,
Latest: time.Unix(1000, 0),
},
},
}))
ds.ProcessDataPoint(123, time.Unix(2000, 0))
ds.ProcessDataPoint(123, time.Unix(3000, 0))
rds = &cachedDs{DbDataSourcer: ds}
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if !strings.Contains(string(fl.last), "PointCount") {
t.Errorf("directorProcessOrForward: Missing the PointCount warning log")
}
if rds.PointCount() != 0 {
t.Errorf("directorProcessOrForward: ClearRRAs(true) not called")
}
// restore directorForwardDPToNode
directorForwardDPToNode = saveFn
}
func Test_directorProcessIncomingDP(t *testing.T) {
saveFn := directorProcessOrForward
dpofCalled := 0
directorProcess | // Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0 | random_line_split | |
director_test.go | }()
go directorIncomingDPMessages(rcv, dpCh)
// Sending a bogus message should not cause anything be written to dpCh
rcv <- &cluster.Msg{}
rcv <- &cluster.Msg{} // second send ensures the loop has gone full circle
if count > 0 {
t.Errorf("Malformed messages should not cause data points, count: %d", count)
}
if !strings.Contains(string(fl.last), "decoding FAILED") {
t.Errorf("Malformed messages should log 'decoding FAILED'")
}
// now we need a real message
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
m, _ := cluster.NewMsg(&cluster.Node{}, dp)
rcv <- m
rcv <- m
if count < 1 {
t.Errorf("At least 1 data point should have been sent to dpCh")
}
dp.Hops = 1000 // exceed maxhops (which in fakeCluster is 0?)
m, _ = cluster.NewMsg(&cluster.Node{}, dp)
rcv <- m // "clear" the loop
count = 0
rcv <- m
rcv <- m
if count > 0 {
t.Errorf("Hops exceeded should not cause data points, count: %d", count)
}
if !strings.Contains(string(fl.last), "max hops") {
t.Errorf("Hops exceeded messages should log 'max hops'")
}
// Closing the dpCh should cause the recover() to happen
// The test here is that it doesn't panic
close(dpCh)
dp.Hops = 0
m, _ = cluster.NewMsg(&cluster.Node{}, dp)
rcv <- m
// Closing the channel exists (not sure how to really test for that)
go directorIncomingDPMessages(rcv, dpCh)
close(rcv)
}
func Test_directorForwardDPToNode(t *testing.T) {
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
md := make([]byte, 20)
md[0] = 1 // Ready
node := &cluster.Node{Node: &memberlist.Node{Meta: md}}
snd := make(chan *cluster.Msg)
count := 0
go func() {
for {
if _, ok := <-snd; !ok {
break
}
count++
}
}()
// if hops is > 0, nothing happens
dp.Hops = 1
directorForwardDPToNode(dp, node, snd)
directorForwardDPToNode(dp, node, snd)
if count > 0 {
t.Errorf("directorForwardDPToNode: Data points with hops > 0 should not be forwarded")
}
// otherwise it should work
dp.Hops = 0
directorForwardDPToNode(dp, node, snd)
dp.Hops = 0 // because it just got incremented
directorForwardDPToNode(dp, node, snd)
if count < 1 {
t.Errorf("Data point not sent to channel?")
}
// mark node not Ready
md[0] = 0
dp.Hops = 0 // because it just got incremented
if err := directorForwardDPToNode(dp, node, snd); err == nil {
t.Errorf("not ready node should cause an error")
}
}
func Test_directorProcessOrForward(t *testing.T) {
saveFn := directorForwardDPToNode
forward, fwErr := 0, error(nil)
directorForwardDPToNode = func(dp *IncomingDP, node *cluster.Node, snd chan *cluster.Msg) error {
forward++
return fwErr
}
// dsc
db := &fakeSerde{}
df := &SimpleDSFinder{DftDSSPec}
sr := &fakeSr{}
dsf := &dsFlusher{db: db, sr: sr}
dsc := newDsCache(db, df, dsf)
// rds
foo := serde.Ident{"name": "foo"}
ds := serde.NewDbDataSource(0, foo, rrd.NewDataSource(*DftDSSPec))
rds := &cachedDs{DbDataSourcer: ds}
// cluster
clstr := &fakeCluster{}
md := make([]byte, 20)
md[0] = 1 // Ready
node := &cluster.Node{Node: &memberlist.Node{Meta: md, Name: "local"}}
clstr.nodesForDd = []*cluster.Node{node}
clstr.ln = node
// workerChs
workerChs := make([]chan *incomingDpWithDs, 1)
workerChs[0] = make(chan *incomingDpWithDs)
sent := 0
go func() {
for {
<-workerChs[0]
sent++
}
}()
// Test if we are LocalNode
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if sent < 1 {
t.Errorf("directorProcessOrForward: Nothing sent to workerChs")
}
// Now test we are NOT LN, forward
remote := &cluster.Node{Node: &memberlist.Node{Meta: md, Name: "remote"}}
clstr.nodesForDd = []*cluster.Node{remote}
n := directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if forward != 1 {
t.Errorf("directorProcessOrForward: directorForwardDPToNode not called")
}
if n != 1 {
t.Errorf("directorProcessOrForward: return value != 1")
}
fl := &fakeLogger{}
log.SetOutput(fl)
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
fwErr = fmt.Errorf("some error")
n = directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if n != 0 {
t.Errorf("directorProcessOrForward: return value != 0")
}
if !strings.Contains(string(fl.last), "some error") {
t.Errorf("directorProcessOrForward: directorForwardDPToNode not logged")
}
fwErr = nil
// make an rds with points
foo = serde.Ident{"name": "foo"}
ds = serde.NewDbDataSource(0, foo, rrd.NewDataSource(rrd.DSSpec{
Step: 10 * time.Second,
RRAs: []rrd.RRASpec{
rrd.RRASpec{Function: rrd.WMEAN,
Step: 10 * time.Second,
Span: 30 * time.Second,
Latest: time.Unix(1000, 0),
},
},
}))
ds.ProcessDataPoint(123, time.Unix(2000, 0))
ds.ProcessDataPoint(123, time.Unix(3000, 0))
rds = &cachedDs{DbDataSourcer: ds}
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if !strings.Contains(string(fl.last), "PointCount") {
t.Errorf("directorProcessOrForward: Missing the PointCount warning log")
}
if rds.PointCount() != 0 {
t.Errorf("directorProcessOrForward: ClearRRAs(true) not called")
}
// restore directorForwardDPToNode
directorForwardDPToNode = saveFn
}
func Test_directorProcessIncomingDP(t *testing.T) {
saveFn := directorProcessOrForward
dpofCalled := 0
directorProcessOrForward = func(dsc *dsCache, cds *cachedDs, clstr clusterer, workerChs workerChannels, dp *IncomingDP, snd chan *cluster.Msg) (forwarded int) {
dpofCalled++
return 0
}
fl := &fakeLogger{}
log.SetOutput(fl)
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
// dp
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
// dsc
db := &fakeSerde{}
df := &SimpleDSFinder{DftDSSPec}
scr := &fakeSr{}
dsf := &dsFlusher{db: db, sr: scr}
dsc := newDsCache(db, df, dsf)
// cluster
clstr := &fakeCluster{cChange: make(chan bool)}
// workerChs
workerChs := make([]chan *incomingDpWithDs, 1)
workerChs[0] = make(chan *incomingDpWithDs | {
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
fl := &fakeLogger{}
log.SetOutput(fl)
rcv := make(chan *cluster.Msg)
dpCh := make(chan *IncomingDP)
count := 0
go func() {
for {
if _, ok := <-dpCh; !ok {
break
}
count++
} | identifier_body | |
director_test.go | }
if rds.PointCount() != 0 {
t.Errorf("directorProcessOrForward: ClearRRAs(true) not called")
}
// restore directorForwardDPToNode
directorForwardDPToNode = saveFn
}
func Test_directorProcessIncomingDP(t *testing.T) {
saveFn := directorProcessOrForward
dpofCalled := 0
directorProcessOrForward = func(dsc *dsCache, cds *cachedDs, clstr clusterer, workerChs workerChannels, dp *IncomingDP, snd chan *cluster.Msg) (forwarded int) {
dpofCalled++
return 0
}
fl := &fakeLogger{}
log.SetOutput(fl)
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
// dp
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
// dsc
db := &fakeSerde{}
df := &SimpleDSFinder{DftDSSPec}
scr := &fakeSr{}
dsf := &dsFlusher{db: db, sr: scr}
dsc := newDsCache(db, df, dsf)
// cluster
clstr := &fakeCluster{cChange: make(chan bool)}
// workerChs
workerChs := make([]chan *incomingDpWithDs, 1)
workerChs[0] = make(chan *incomingDpWithDs)
sent := 0
go func() {
for {
<-workerChs[0]
sent++
}
}()
// NaN
dp.Value = math.NaN()
directorProcessIncomingDP(dp, scr, dsc, nil, nil, nil)
if scr.called != 1 {
t.Errorf("directorProcessIncomingDP: With a NaN, reportStatCount() should only be called once")
}
if dpofCalled > 0 {
t.Errorf("directorProcessIncomingDP: With a NaN, directorProcessOrForward should not be called")
}
// A value
dp.Value = 1234
scr.called, dpofCalled = 0, 0
directorProcessIncomingDP(dp, scr, dsc, workerChs, clstr, nil)
if scr.called != 2 {
t.Errorf("directorProcessIncomingDP: With a value, reportStatCount() should be called twice: %v", scr.called)
}
if dpofCalled != 1 {
t.Errorf("directorProcessIncomingDP: With a value, directorProcessOrForward should be called once: %v", dpofCalled)
}
// A blank name should cause a nil rds
dp.Name = ""
scr.called, dpofCalled = 0, 0
directorProcessIncomingDP(dp, scr, dsc, nil, nil, nil)
if scr.called != 1 {
t.Errorf("directorProcessIncomingDP: With a blank name, reportStatCount() should be called once")
}
if dpofCalled > 0 {
t.Errorf("directorProcessIncomingDP: With a blank name, directorProcessOrForward should not be called")
}
if !strings.Contains(string(fl.last), "No spec matched") {
t.Errorf("should log 'No spec matched'")
}
// fake a db error
dp.Name = "blah"
db.fakeErr = true
scr.called, dpofCalled = 0, 0
directorProcessIncomingDP(dp, scr, dsc, nil, nil, nil)
if scr.called != 1 {
t.Errorf("directorProcessIncomingDP: With a db error, reportStatCount() should be called once")
}
if dpofCalled > 0 {
t.Errorf("directorProcessIncomingDP: With a db error, directorProcessOrForward should not be called")
}
if !strings.Contains(string(fl.last), "error") {
t.Errorf("should log 'error'")
}
// nil cluster
dp.Value = 1234
db.fakeErr = false
scr.called = 0
directorProcessIncomingDP(dp, scr, dsc, workerChs, nil, nil)
if scr.called != 1 {
t.Errorf("directorProcessIncomingDP: With a value, reportStatCount() should be called once: %v", scr.called)
}
if dpofCalled != 0 {
t.Errorf("directorProcessIncomingDP: With a value and no cluster, directorProcessOrForward should not be called: %v", dpofCalled)
}
directorProcessOrForward = saveFn
}
func Test_the_director(t *testing.T) {
saveFn1 := directorIncomingDPMessages
saveFn2 := directorProcessIncomingDP
dimCalled := 0
directorIncomingDPMessages = func(rcv chan *cluster.Msg, dpCh chan *IncomingDP) { dimCalled++ }
dpidpCalled := 0
directorProcessIncomingDP = func(dp *IncomingDP, scr statReporter, dsc *dsCache, workerChs workerChannels, clstr clusterer, snd chan *cluster.Msg) {
dpidpCalled++
}
fl := &fakeLogger{}
log.SetOutput(fl)
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
wc := &wrkCtl{wg: &sync.WaitGroup{}, startWg: &sync.WaitGroup{}, id: "FOO"}
clstr := &fakeCluster{cChange: make(chan bool)}
dpCh := make(chan *IncomingDP)
// dsc
db := &fakeSerde{}
df := &SimpleDSFinder{DftDSSPec}
sr := &fakeSr{}
dsf := &dsFlusher{db: db, sr: sr}
dsc := newDsCache(db, df, dsf)
wc.startWg.Add(1)
go director(wc, dpCh, clstr, sr, dsc, nil)
wc.startWg.Wait()
if clstr.nReady == 0 {
t.Errorf("director: Ready(true) not called on cluster")
}
if clstr.nReg == 0 {
t.Errorf("director: cluster.RegisterMsgType() not called")
}
// This sometimes can fail because we don't wait for that goroutine in this test...
time.Sleep(5 * time.Millisecond)
if dimCalled == 0 {
t.Errorf("director: directorIncomingDPMessages not started")
}
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
dpCh <- dp
dpCh <- dp
if dpidpCalled == 0 {
t.Errorf("director: directorProcessIncomingDP not called")
}
// Trigger a transition
clstr.cChange <- true
dpCh <- dp
if clstr.nTrans == 0 {
t.Errorf("director: on cluster change, Transition() not called")
}
// Transition with error
clstr.tErr = true
clstr.cChange <- true
dpCh <- dp
if !strings.Contains(string(fl.last), "some error") {
t.Errorf("director: on transition error, 'some error' missing from logs")
}
dpidpCalled = 0
close(dpCh)
time.Sleep(1 * time.Second) // so that nil dp goroutine panics/recovers
if dpidpCalled > 0 {
t.Errorf("director: directorProcessIncomingDP must not be called on channel close")
}
if !strings.Contains(string(fl.last), "shutting down") {
t.Errorf("director: on channel close, missing 'shutting down' log entry")
}
// overrun
dpCh = make(chan *IncomingDP, 5)
dpCh <- dp
dpCh <- dp
dpCh <- dp
dpCh <- dp
wc.startWg.Add(1)
go director(wc, dpCh, clstr, sr, dsc, nil)
wc.startWg.Wait()
time.Sleep(100 * time.Millisecond)
close(dpCh)
directorIncomingDPMessages = saveFn1
directorProcessIncomingDP = saveFn2
}
func Test_director_reportDirectorChannelFillPercent(t *testing.T) {
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
fl := &fakeLogger{}
log.SetOutput(fl)
ch := make(chan *IncomingDP, 10)
sr := &fakeSr{}
for i := 0; i < 9; i++ {
ch <- &IncomingDP{}
}
queue := &dpQueue{}
queue.push(&IncomingDP{})
go reportDirectorChannelFillPercent(ch, queue, sr, time.Millisecond)
time.Sleep(50 * time.Millisecond)
if sr.called == 0 {
t.Errorf("reportDirectorChannelFillPercent: statReporter should have been called a bunch of times")
}
if !strings.Contains(string(fl.last), "WARNING") {
t.Errorf("reportDirectorChannelFillPercent: there should be a warning about director channel nearly full")
}
}
func Test_director_queue(t *testing.T) {
queue := &dpQueue{}
dp := &IncomingDP{}
queue.push(dp)
if queue.pop() != dp {
t.Errorf("queue: pop returned wrong dp")
}
if queue.size() != 0 {
t.Errorf("queue: should be empty")
}
queue.push(&IncomingDP{})
if queue.size() != 1 {
t.Errorf("queue: size != 1")
}
}
func | Test_director_checkSetAside | identifier_name | |
director_test.go | t.Errorf("Hops exceeded messages should log 'max hops'")
}
// Closing the dpCh should cause the recover() to happen
// The test here is that it doesn't panic
close(dpCh)
dp.Hops = 0
m, _ = cluster.NewMsg(&cluster.Node{}, dp)
rcv <- m
// Closing the channel exists (not sure how to really test for that)
go directorIncomingDPMessages(rcv, dpCh)
close(rcv)
}
func Test_directorForwardDPToNode(t *testing.T) {
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
md := make([]byte, 20)
md[0] = 1 // Ready
node := &cluster.Node{Node: &memberlist.Node{Meta: md}}
snd := make(chan *cluster.Msg)
count := 0
go func() {
for {
if _, ok := <-snd; !ok {
break
}
count++
}
}()
// if hops is > 0, nothing happens
dp.Hops = 1
directorForwardDPToNode(dp, node, snd)
directorForwardDPToNode(dp, node, snd)
if count > 0 {
t.Errorf("directorForwardDPToNode: Data points with hops > 0 should not be forwarded")
}
// otherwise it should work
dp.Hops = 0
directorForwardDPToNode(dp, node, snd)
dp.Hops = 0 // because it just got incremented
directorForwardDPToNode(dp, node, snd)
if count < 1 {
t.Errorf("Data point not sent to channel?")
}
// mark node not Ready
md[0] = 0
dp.Hops = 0 // because it just got incremented
if err := directorForwardDPToNode(dp, node, snd); err == nil {
t.Errorf("not ready node should cause an error")
}
}
func Test_directorProcessOrForward(t *testing.T) {
saveFn := directorForwardDPToNode
forward, fwErr := 0, error(nil)
directorForwardDPToNode = func(dp *IncomingDP, node *cluster.Node, snd chan *cluster.Msg) error {
forward++
return fwErr
}
// dsc
db := &fakeSerde{}
df := &SimpleDSFinder{DftDSSPec}
sr := &fakeSr{}
dsf := &dsFlusher{db: db, sr: sr}
dsc := newDsCache(db, df, dsf)
// rds
foo := serde.Ident{"name": "foo"}
ds := serde.NewDbDataSource(0, foo, rrd.NewDataSource(*DftDSSPec))
rds := &cachedDs{DbDataSourcer: ds}
// cluster
clstr := &fakeCluster{}
md := make([]byte, 20)
md[0] = 1 // Ready
node := &cluster.Node{Node: &memberlist.Node{Meta: md, Name: "local"}}
clstr.nodesForDd = []*cluster.Node{node}
clstr.ln = node
// workerChs
workerChs := make([]chan *incomingDpWithDs, 1)
workerChs[0] = make(chan *incomingDpWithDs)
sent := 0
go func() {
for {
<-workerChs[0]
sent++
}
}()
// Test if we are LocalNode
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if sent < 1 {
t.Errorf("directorProcessOrForward: Nothing sent to workerChs")
}
// Now test we are NOT LN, forward
remote := &cluster.Node{Node: &memberlist.Node{Meta: md, Name: "remote"}}
clstr.nodesForDd = []*cluster.Node{remote}
n := directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if forward != 1 {
t.Errorf("directorProcessOrForward: directorForwardDPToNode not called")
}
if n != 1 {
t.Errorf("directorProcessOrForward: return value != 1")
}
fl := &fakeLogger{}
log.SetOutput(fl)
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
fwErr = fmt.Errorf("some error")
n = directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if n != 0 {
t.Errorf("directorProcessOrForward: return value != 0")
}
if !strings.Contains(string(fl.last), "some error") {
t.Errorf("directorProcessOrForward: directorForwardDPToNode not logged")
}
fwErr = nil
// make an rds with points
foo = serde.Ident{"name": "foo"}
ds = serde.NewDbDataSource(0, foo, rrd.NewDataSource(rrd.DSSpec{
Step: 10 * time.Second,
RRAs: []rrd.RRASpec{
rrd.RRASpec{Function: rrd.WMEAN,
Step: 10 * time.Second,
Span: 30 * time.Second,
Latest: time.Unix(1000, 0),
},
},
}))
ds.ProcessDataPoint(123, time.Unix(2000, 0))
ds.ProcessDataPoint(123, time.Unix(3000, 0))
rds = &cachedDs{DbDataSourcer: ds}
directorProcessOrForward(dsc, rds, clstr, workerChs, nil, nil)
if !strings.Contains(string(fl.last), "PointCount") {
t.Errorf("directorProcessOrForward: Missing the PointCount warning log")
}
if rds.PointCount() != 0 {
t.Errorf("directorProcessOrForward: ClearRRAs(true) not called")
}
// restore directorForwardDPToNode
directorForwardDPToNode = saveFn
}
func Test_directorProcessIncomingDP(t *testing.T) {
saveFn := directorProcessOrForward
dpofCalled := 0
directorProcessOrForward = func(dsc *dsCache, cds *cachedDs, clstr clusterer, workerChs workerChannels, dp *IncomingDP, snd chan *cluster.Msg) (forwarded int) {
dpofCalled++
return 0
}
fl := &fakeLogger{}
log.SetOutput(fl)
defer func() {
// restore default output
log.SetOutput(os.Stderr)
}()
// dp
dp := &IncomingDP{Name: "foo", TimeStamp: time.Unix(1000, 0), Value: 123}
// dsc
db := &fakeSerde{}
df := &SimpleDSFinder{DftDSSPec}
scr := &fakeSr{}
dsf := &dsFlusher{db: db, sr: scr}
dsc := newDsCache(db, df, dsf)
// cluster
clstr := &fakeCluster{cChange: make(chan bool)}
// workerChs
workerChs := make([]chan *incomingDpWithDs, 1)
workerChs[0] = make(chan *incomingDpWithDs)
sent := 0
go func() {
for {
<-workerChs[0]
sent++
}
}()
// NaN
dp.Value = math.NaN()
directorProcessIncomingDP(dp, scr, dsc, nil, nil, nil)
if scr.called != 1 {
t.Errorf("directorProcessIncomingDP: With a NaN, reportStatCount() should only be called once")
}
if dpofCalled > 0 {
t.Errorf("directorProcessIncomingDP: With a NaN, directorProcessOrForward should not be called")
}
// A value
dp.Value = 1234
scr.called, dpofCalled = 0, 0
directorProcessIncomingDP(dp, scr, dsc, workerChs, clstr, nil)
if scr.called != 2 |
if dpofCalled != 1 {
t.Errorf("directorProcessIncomingDP: With a value, directorProcessOrForward should be called once: %v", dpofCalled)
}
// A blank name should cause a nil rds
dp.Name = ""
scr.called, dpofCalled = 0, 0
directorProcessIncomingDP(dp, scr, dsc, nil, nil, nil)
if scr.called != 1 {
t.Errorf("directorProcessIncomingDP: With a blank name, reportStatCount() should be called once")
}
if dpofCalled > 0 {
t.Errorf("directorProcessIncomingDP: With a blank name, directorProcessOrForward should not be called")
}
if !strings.Contains(string(fl.last), "No spec matched") {
t.Errorf("should log 'No spec matched'")
}
// fake | {
t.Errorf("directorProcessIncomingDP: With a value, reportStatCount() should be called twice: %v", scr.called)
} | conditional_block |
search.rs | > {
pub fn new(cl: C) -> Self |
}
impl<C: IntoIterator> Iterator for TaskIter<C> {
type Item = <C::IntoIter as Iterator>::Item;
fn next(&mut self) -> Option<Self::Item> {
self.0.next()
}
}
pub type Task<'t, P, C> = TaskIter<super::clause::OClause<'t, Lit<P, C, usize>>>;
pub type Context<'t, P, C> = context::Context<Vec<OLit<'t, P, C>>>;
#[derive(Clone, Debug)]
pub enum Action<'t, P, C> {
Prove,
Reduce(OLit<'t, P, C>, Index),
Extend(OLit<'t, P, C>, Contras<'t, P, C>, Index),
}
impl<'t, P, C> Action<'t, P, C> {
pub fn max_children(&self) -> usize {
use Action::*;
match self {
Prove | Reduce(_, _) => 0,
Extend(_, cs, skip) => cs[*skip].rest.len(),
}
}
}
type Index = usize;
type Contras<'t, P, C> = &'t [Contrapositive<P, C, usize>];
struct Alternative<'t, P, C> {
task: Task<'t, P, C>,
// when we do *not* use cut, then we may need to backtrack to
// contexts that are larger than the current context,
// so we save the whole context here
ctx: Option<Context<'t, P, C>>,
// when we use cut, then we always backtrack to contexts that are
// prefixes of the current context, so in that case,
// storing just a pointer to the context suffices
ctx_ptr: context::Ptr,
promises: Option<Vec<Promise<Task<'t, P, C>>>>,
promises_len: usize,
sub: SubPtr,
proof_len: usize,
}
#[derive(Clone)]
struct Promise<T> {
task: T,
ctx_ptr: context::Ptr,
alt_len: usize,
}
pub struct Opt {
pub lim: usize,
pub cuts: Cuts,
}
impl<'t, P, C> Search<'t, P, C> {
pub fn new(task: Task<'t, P, C>, db: &'t Db<P, C, usize>, opt: Opt) -> Self {
Self {
task,
ctx: Context::default(),
promises: Vec::new(),
sub: Sub::default(),
proof: Steps::new(),
alternatives: Vec::new(),
inferences: 0,
literals: 0,
db,
opt,
}
}
}
type State<'t, P, C> = Result<Action<'t, P, C>, bool>;
impl<'t, P, C> Search<'t, P, C>
where
P: Clone + Display + Eq + Hash + Neg<Output = P>,
C: Clone + Display + Eq,
{
pub fn prove(&mut self) -> Option<&Steps<'t, P, C>> {
let mut action: Action<'t, P, C> = Action::Prove;
loop {
let result = match action {
Action::Prove => match self.task.clone().next() {
Some(lit) => self.chk(lit),
None => self.fulfill_promise(),
},
Action::Reduce(lit, skip) => self.red(lit, skip),
Action::Extend(lit, contras, skip) => self.ext(lit, contras, skip),
};
match result {
Ok(next) => action = next,
Err(true) => return Some(&self.proof),
Err(false) => return None,
}
}
}
pub fn inferences(&self) -> usize {
self.inferences
}
fn chk(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("checks: {}", lit);
debug!("{} {}", self.literals, lit.head());
debug!("lemmas: {}", self.ctx.lemmas.len());
debug!("path: {}", self.ctx.path.len());
self.literals += 1;
let mut lits = self.task.clone();
let mut path = self.ctx.path.iter();
let mut lemmas = self.ctx.lemmas.iter();
if lits.any(|cl| path.any(|pl| pl.eq_mod(&self.sub, &cl))) {
debug!("regularity");
self.try_alternative()
} else if lemmas.any(|lem| lem.eq_mod(&self.sub, &lit)) {
debug!("lemma");
self.proof.push(Action::Prove);
// do not add lit to lemmas, unlike original leanCoP
// furthermore, do not try red/ext steps if we found a lemma,
// because it does not add anything to substitution
// note that Jens said that this might sometimes be counterproductive,
// because adding to the substitution is also beneficial to cut down search space
self.task.next();
Ok(Action::Prove)
} else {
Ok(Action::Reduce(lit, 0))
}
}
fn red(&mut self, lit: OLit<'t, P, C>, skip: usize) -> State<'t, P, C> {
debug!("reduce: {}", lit);
let alternative = Alternative::from(&*self);
for (pidx, pat) in self.ctx.path.iter().rev().enumerate().skip(skip) {
debug!("try reduce: {}", pat);
let sub_dom_len = self.sub.get_dom_len();
if pat.head() != &-lit.head().clone() {
continue;
}
if pat.args().unify(&mut self.sub, lit.args()) {
debug!("reduce succeeded");
self.proof.push(Action::Reduce(lit, pidx));
if !self.opt.cuts.reduction {
let action = Action::Reduce(lit, pidx + 1);
self.alternatives.push((alternative, action));
}
self.ctx.lemmas.push(lit);
self.task.next();
return Ok(Action::Prove);
} else {
self.sub.set_dom_len(sub_dom_len)
}
}
self.ext0(lit)
}
fn ext0(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("extend: {}", lit);
let neg = -lit.head().clone();
match self.db.get(&neg) {
Some(entries) => self.ext(lit, entries, 0),
None => self.try_alternative(),
}
}
fn ext(&mut self, lit: OLit<'t, P, C>, cs: Contras<'t, P, C>, skip: usize) -> State<'t, P, C> {
let alt = Alternative::from(&*self);
let prm = Promise::from(&*self);
let sub = SubPtr::from(&self.sub);
for (eidx, entry) in cs.iter().enumerate().skip(skip) {
debug!(
"try extend {}{} (lit = {}, |path| = {})",
lit.head(),
entry,
lit,
self.ctx.path.len()
);
if self.ctx.path.len() >= self.opt.lim && entry.vars.is_some() {
debug!("path limit reached");
continue;
};
let eargs = Offset::new(sub.dom_max(), &entry.args);
if let Some(vars) = entry.vars {
// we have to add 1 here because the lowest variable is 0
self.sub.set_dom_max(sub.dom_max() + vars + 1)
};
debug!("unify {} ~? {}, sub = {}", eargs, lit.args(), self.sub);
if eargs.unify(&mut self.sub, lit.args()) {
debug!("unify succeeded with {}, sub = {}", entry.rest, self.sub);
self.inferences += 1;
// promise to fulfill the current task
// (if the promise is kept and cut is enabled,
// then all alternatives that came after will be discarded)
self.promises.push(prm);
self.proof.push(Action::Extend(lit, cs, eidx));
let action = Action::Extend(lit, cs, eidx + 1);
// register an alternative (that will be discarded
// if the above promise is kept and cut is enabled)
self.alternatives.push((alt, action));
self.task = Task::new(Offset::new(sub.dom_max(), &entry.rest));
self.ctx.path.push(lit);
return Ok(Action::Prove);
} else {
debug!("unify failed");
self.sub.rewind(&sub)
}
}
self.try_alternative()
}
fn fulfill_promise(&mut self) -> State<'t, P, C> {
debug!("fulfill promise ({} left)", self.promises.len());
let prm = self.promises.pop().ok_or(true)?;
self.task = prm.task;
self.ctx.rewind(prm.ctx_ptr);
if let Some(prev) = self.task.next() {
self.ctx.lemmas.push(prev)
};
if let Some(cut) = self.opt.cuts.extension {
use super::cuts::Cut::*;
let alt_len = match cut | {
Self(cl.into_iter().skip(0))
} | identifier_body |
search.rs | pub type Task<'t, P, C> = TaskIter<super::clause::OClause<'t, Lit<P, C, usize>>>;
pub type Context<'t, P, C> = context::Context<Vec<OLit<'t, P, C>>>;
#[derive(Clone, Debug)]
pub enum Action<'t, P, C> {
Prove,
Reduce(OLit<'t, P, C>, Index),
Extend(OLit<'t, P, C>, Contras<'t, P, C>, Index),
}
impl<'t, P, C> Action<'t, P, C> {
pub fn max_children(&self) -> usize {
use Action::*;
match self {
Prove | Reduce(_, _) => 0,
Extend(_, cs, skip) => cs[*skip].rest.len(),
}
}
}
type Index = usize;
type Contras<'t, P, C> = &'t [Contrapositive<P, C, usize>];
struct Alternative<'t, P, C> {
task: Task<'t, P, C>,
// when we do *not* use cut, then we may need to backtrack to
// contexts that are larger than the current context,
// so we save the whole context here
ctx: Option<Context<'t, P, C>>,
// when we use cut, then we always backtrack to contexts that are
// prefixes of the current context, so in that case,
// storing just a pointer to the context suffices
ctx_ptr: context::Ptr,
promises: Option<Vec<Promise<Task<'t, P, C>>>>,
promises_len: usize,
sub: SubPtr,
proof_len: usize,
}
#[derive(Clone)]
struct Promise<T> {
task: T,
ctx_ptr: context::Ptr,
alt_len: usize,
}
pub struct Opt {
pub lim: usize,
pub cuts: Cuts,
}
impl<'t, P, C> Search<'t, P, C> {
pub fn new(task: Task<'t, P, C>, db: &'t Db<P, C, usize>, opt: Opt) -> Self {
Self {
task,
ctx: Context::default(),
promises: Vec::new(),
sub: Sub::default(),
proof: Steps::new(),
alternatives: Vec::new(),
inferences: 0,
literals: 0,
db,
opt,
}
}
}
type State<'t, P, C> = Result<Action<'t, P, C>, bool>;
impl<'t, P, C> Search<'t, P, C>
where
P: Clone + Display + Eq + Hash + Neg<Output = P>,
C: Clone + Display + Eq,
{
pub fn prove(&mut self) -> Option<&Steps<'t, P, C>> {
let mut action: Action<'t, P, C> = Action::Prove;
loop {
let result = match action {
Action::Prove => match self.task.clone().next() {
Some(lit) => self.chk(lit),
None => self.fulfill_promise(),
},
Action::Reduce(lit, skip) => self.red(lit, skip),
Action::Extend(lit, contras, skip) => self.ext(lit, contras, skip),
};
match result {
Ok(next) => action = next,
Err(true) => return Some(&self.proof),
Err(false) => return None,
}
}
}
pub fn inferences(&self) -> usize {
self.inferences
}
fn chk(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("checks: {}", lit);
debug!("{} {}", self.literals, lit.head());
debug!("lemmas: {}", self.ctx.lemmas.len());
debug!("path: {}", self.ctx.path.len());
self.literals += 1;
let mut lits = self.task.clone();
let mut path = self.ctx.path.iter();
let mut lemmas = self.ctx.lemmas.iter();
if lits.any(|cl| path.any(|pl| pl.eq_mod(&self.sub, &cl))) {
debug!("regularity");
self.try_alternative()
} else if lemmas.any(|lem| lem.eq_mod(&self.sub, &lit)) {
debug!("lemma");
self.proof.push(Action::Prove);
// do not add lit to lemmas, unlike original leanCoP
// furthermore, do not try red/ext steps if we found a lemma,
// because it does not add anything to substitution
// note that Jens said that this might sometimes be counterproductive,
// because adding to the substitution is also beneficial to cut down search space
self.task.next();
Ok(Action::Prove)
} else {
Ok(Action::Reduce(lit, 0))
}
}
fn red(&mut self, lit: OLit<'t, P, C>, skip: usize) -> State<'t, P, C> {
debug!("reduce: {}", lit);
let alternative = Alternative::from(&*self);
for (pidx, pat) in self.ctx.path.iter().rev().enumerate().skip(skip) {
debug!("try reduce: {}", pat);
let sub_dom_len = self.sub.get_dom_len();
if pat.head() != &-lit.head().clone() {
continue;
}
if pat.args().unify(&mut self.sub, lit.args()) {
debug!("reduce succeeded");
self.proof.push(Action::Reduce(lit, pidx));
if !self.opt.cuts.reduction {
let action = Action::Reduce(lit, pidx + 1);
self.alternatives.push((alternative, action));
}
self.ctx.lemmas.push(lit);
self.task.next();
return Ok(Action::Prove);
} else {
self.sub.set_dom_len(sub_dom_len)
}
}
self.ext0(lit)
}
fn ext0(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("extend: {}", lit);
let neg = -lit.head().clone();
match self.db.get(&neg) {
Some(entries) => self.ext(lit, entries, 0),
None => self.try_alternative(),
}
}
fn ext(&mut self, lit: OLit<'t, P, C>, cs: Contras<'t, P, C>, skip: usize) -> State<'t, P, C> {
let alt = Alternative::from(&*self);
let prm = Promise::from(&*self);
let sub = SubPtr::from(&self.sub);
for (eidx, entry) in cs.iter().enumerate().skip(skip) {
debug!(
"try extend {}{} (lit = {}, |path| = {})",
lit.head(),
entry,
lit,
self.ctx.path.len()
);
if self.ctx.path.len() >= self.opt.lim && entry.vars.is_some() {
debug!("path limit reached");
continue;
};
let eargs = Offset::new(sub.dom_max(), &entry.args);
if let Some(vars) = entry.vars {
// we have to add 1 here because the lowest variable is 0
self.sub.set_dom_max(sub.dom_max() + vars + 1)
};
debug!("unify {} ~? {}, sub = {}", eargs, lit.args(), self.sub);
if eargs.unify(&mut self.sub, lit.args()) {
debug!("unify succeeded with {}, sub = {}", entry.rest, self.sub);
self.inferences += 1;
// promise to fulfill the current task
// (if the promise is kept and cut is enabled,
// then all alternatives that came after will be discarded)
self.promises.push(prm);
self.proof.push(Action::Extend(lit, cs, eidx));
let action = Action::Extend(lit, cs, eidx + 1);
// register an alternative (that will be discarded
// if the above promise is kept and cut is enabled)
self.alternatives.push((alt, action));
self.task = Task::new(Offset::new(sub.dom_max(), &entry.rest));
self.ctx.path.push(lit);
return Ok(Action::Prove);
} else {
debug!("unify failed");
self.sub.rewind(&sub)
}
}
self.try_alternative()
}
fn fulfill_promise(&mut self) -> State<'t, P, C> {
debug!("fulfill promise ({} left)", self.promises.len());
let prm = self.promises.pop().ok_or(true)?;
self.task = prm.task;
self.ctx.rewind(prm.ctx_ptr);
if let Some(prev) = self.task.next() {
self.ctx.lemmas.push(prev)
};
if let Some(cut) = self.opt.cuts.extension {
use super::cuts::Cut::*;
let alt_len = match cut {
Exclusive => prm.alt_len + 1,
Inclusive => prm.alt_len,
};
debug!("cut {} alternatives", self.alternatives.len() - alt_len);
assert!(alt_len <= self.alternatives.len());
self.alternatives.truncate(alt_len);
}
Ok(Action::Prove)
}
fn | try_alternative | identifier_name | |
search.rs | {
task: T,
ctx_ptr: context::Ptr,
alt_len: usize,
}
pub struct Opt {
pub lim: usize,
pub cuts: Cuts,
}
impl<'t, P, C> Search<'t, P, C> {
pub fn new(task: Task<'t, P, C>, db: &'t Db<P, C, usize>, opt: Opt) -> Self {
Self {
task,
ctx: Context::default(),
promises: Vec::new(),
sub: Sub::default(),
proof: Steps::new(),
alternatives: Vec::new(),
inferences: 0,
literals: 0,
db,
opt,
}
}
}
type State<'t, P, C> = Result<Action<'t, P, C>, bool>;
impl<'t, P, C> Search<'t, P, C>
where
P: Clone + Display + Eq + Hash + Neg<Output = P>,
C: Clone + Display + Eq,
{
pub fn prove(&mut self) -> Option<&Steps<'t, P, C>> {
let mut action: Action<'t, P, C> = Action::Prove;
loop {
let result = match action {
Action::Prove => match self.task.clone().next() {
Some(lit) => self.chk(lit),
None => self.fulfill_promise(),
},
Action::Reduce(lit, skip) => self.red(lit, skip),
Action::Extend(lit, contras, skip) => self.ext(lit, contras, skip),
};
match result {
Ok(next) => action = next,
Err(true) => return Some(&self.proof),
Err(false) => return None,
}
}
}
pub fn inferences(&self) -> usize {
self.inferences
}
fn chk(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("checks: {}", lit);
debug!("{} {}", self.literals, lit.head());
debug!("lemmas: {}", self.ctx.lemmas.len());
debug!("path: {}", self.ctx.path.len());
self.literals += 1;
let mut lits = self.task.clone();
let mut path = self.ctx.path.iter();
let mut lemmas = self.ctx.lemmas.iter();
if lits.any(|cl| path.any(|pl| pl.eq_mod(&self.sub, &cl))) {
debug!("regularity");
self.try_alternative()
} else if lemmas.any(|lem| lem.eq_mod(&self.sub, &lit)) {
debug!("lemma");
self.proof.push(Action::Prove);
// do not add lit to lemmas, unlike original leanCoP
// furthermore, do not try red/ext steps if we found a lemma,
// because it does not add anything to substitution
// note that Jens said that this might sometimes be counterproductive,
// because adding to the substitution is also beneficial to cut down search space
self.task.next();
Ok(Action::Prove)
} else {
Ok(Action::Reduce(lit, 0))
}
}
fn red(&mut self, lit: OLit<'t, P, C>, skip: usize) -> State<'t, P, C> {
debug!("reduce: {}", lit);
let alternative = Alternative::from(&*self);
for (pidx, pat) in self.ctx.path.iter().rev().enumerate().skip(skip) {
debug!("try reduce: {}", pat);
let sub_dom_len = self.sub.get_dom_len();
if pat.head() != &-lit.head().clone() {
continue;
}
if pat.args().unify(&mut self.sub, lit.args()) {
debug!("reduce succeeded");
self.proof.push(Action::Reduce(lit, pidx));
if !self.opt.cuts.reduction {
let action = Action::Reduce(lit, pidx + 1);
self.alternatives.push((alternative, action));
}
self.ctx.lemmas.push(lit);
self.task.next();
return Ok(Action::Prove);
} else {
self.sub.set_dom_len(sub_dom_len)
}
}
self.ext0(lit)
}
fn ext0(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("extend: {}", lit);
let neg = -lit.head().clone();
match self.db.get(&neg) {
Some(entries) => self.ext(lit, entries, 0),
None => self.try_alternative(),
}
}
fn ext(&mut self, lit: OLit<'t, P, C>, cs: Contras<'t, P, C>, skip: usize) -> State<'t, P, C> {
let alt = Alternative::from(&*self);
let prm = Promise::from(&*self);
let sub = SubPtr::from(&self.sub);
for (eidx, entry) in cs.iter().enumerate().skip(skip) {
debug!(
"try extend {}{} (lit = {}, |path| = {})",
lit.head(),
entry,
lit,
self.ctx.path.len()
);
if self.ctx.path.len() >= self.opt.lim && entry.vars.is_some() {
debug!("path limit reached");
continue;
};
let eargs = Offset::new(sub.dom_max(), &entry.args);
if let Some(vars) = entry.vars {
// we have to add 1 here because the lowest variable is 0
self.sub.set_dom_max(sub.dom_max() + vars + 1)
};
debug!("unify {} ~? {}, sub = {}", eargs, lit.args(), self.sub);
if eargs.unify(&mut self.sub, lit.args()) {
debug!("unify succeeded with {}, sub = {}", entry.rest, self.sub);
self.inferences += 1;
// promise to fulfill the current task
// (if the promise is kept and cut is enabled,
// then all alternatives that came after will be discarded)
self.promises.push(prm);
self.proof.push(Action::Extend(lit, cs, eidx));
let action = Action::Extend(lit, cs, eidx + 1);
// register an alternative (that will be discarded
// if the above promise is kept and cut is enabled)
self.alternatives.push((alt, action));
self.task = Task::new(Offset::new(sub.dom_max(), &entry.rest));
self.ctx.path.push(lit);
return Ok(Action::Prove);
} else {
debug!("unify failed");
self.sub.rewind(&sub)
}
}
self.try_alternative()
}
fn fulfill_promise(&mut self) -> State<'t, P, C> {
debug!("fulfill promise ({} left)", self.promises.len());
let prm = self.promises.pop().ok_or(true)?;
self.task = prm.task;
self.ctx.rewind(prm.ctx_ptr);
if let Some(prev) = self.task.next() {
self.ctx.lemmas.push(prev)
};
if let Some(cut) = self.opt.cuts.extension {
use super::cuts::Cut::*;
let alt_len = match cut {
Exclusive => prm.alt_len + 1,
Inclusive => prm.alt_len,
};
debug!("cut {} alternatives", self.alternatives.len() - alt_len);
assert!(alt_len <= self.alternatives.len());
self.alternatives.truncate(alt_len);
}
Ok(Action::Prove)
}
fn try_alternative(&mut self) -> State<'t, P, C> {
debug!("try alternative ({} left)", self.alternatives.len());
self.alternatives.pop().ok_or(false).map(|(alt, action)| {
self.rewind(alt);
action
})
}
}
impl<'t, P, C> From<&Search<'t, P, C>> for Alternative<'t, P, C> {
fn from(st: &Search<'t, P, C>) -> Self {
Self {
task: st.task.clone(),
ctx: if st.opt.cuts.extension.is_none() {
Some(st.ctx.clone())
} else {
None
},
ctx_ptr: context::Ptr::from(&st.ctx),
promises: if st.opt.cuts.extension.is_none() {
Some(st.promises.clone())
} else {
None
},
promises_len: st.promises.len(),
sub: SubPtr::from(&st.sub),
proof_len: st.proof.len(),
}
}
}
impl<'t, P, C> From<&Search<'t, P, C>> for Promise<Task<'t, P, C>> {
fn from(st: &Search<'t, P, C>) -> Self {
Self {
task: st.task.clone(),
ctx_ptr: context::Ptr::from(&st.ctx),
alt_len: st.alternatives.len(),
}
}
}
impl<'t, P, C> Rewind<Alternative<'t, P, C>> for Search<'t, P, C> {
fn rewind(&mut self, alt: Alternative<'t, P, C>) {
self.task = alt.task;
if let Some(ctx) = alt.ctx | {
self.ctx = ctx;
} | conditional_block | |
search.rs | > {
pub fn new(cl: C) -> Self {
Self(cl.into_iter().skip(0))
}
}
impl<C: IntoIterator> Iterator for TaskIter<C> {
type Item = <C::IntoIter as Iterator>::Item;
fn next(&mut self) -> Option<Self::Item> {
self.0.next()
}
}
pub type Task<'t, P, C> = TaskIter<super::clause::OClause<'t, Lit<P, C, usize>>>;
pub type Context<'t, P, C> = context::Context<Vec<OLit<'t, P, C>>>;
#[derive(Clone, Debug)]
pub enum Action<'t, P, C> {
Prove,
Reduce(OLit<'t, P, C>, Index),
Extend(OLit<'t, P, C>, Contras<'t, P, C>, Index),
}
impl<'t, P, C> Action<'t, P, C> {
pub fn max_children(&self) -> usize {
use Action::*;
match self {
Prove | Reduce(_, _) => 0,
Extend(_, cs, skip) => cs[*skip].rest.len(),
}
}
}
type Index = usize;
type Contras<'t, P, C> = &'t [Contrapositive<P, C, usize>];
struct Alternative<'t, P, C> {
task: Task<'t, P, C>,
// when we do *not* use cut, then we may need to backtrack to
// contexts that are larger than the current context,
// so we save the whole context here
ctx: Option<Context<'t, P, C>>,
// when we use cut, then we always backtrack to contexts that are
// prefixes of the current context, so in that case,
// storing just a pointer to the context suffices
ctx_ptr: context::Ptr,
promises: Option<Vec<Promise<Task<'t, P, C>>>>,
promises_len: usize,
sub: SubPtr,
proof_len: usize,
}
#[derive(Clone)]
struct Promise<T> {
task: T,
ctx_ptr: context::Ptr,
alt_len: usize,
}
pub struct Opt {
pub lim: usize,
pub cuts: Cuts,
}
impl<'t, P, C> Search<'t, P, C> {
pub fn new(task: Task<'t, P, C>, db: &'t Db<P, C, usize>, opt: Opt) -> Self {
Self {
task,
ctx: Context::default(),
promises: Vec::new(),
sub: Sub::default(),
proof: Steps::new(),
alternatives: Vec::new(),
inferences: 0,
literals: 0,
db,
opt,
}
}
}
type State<'t, P, C> = Result<Action<'t, P, C>, bool>;
impl<'t, P, C> Search<'t, P, C>
where
P: Clone + Display + Eq + Hash + Neg<Output = P>,
C: Clone + Display + Eq,
{
pub fn prove(&mut self) -> Option<&Steps<'t, P, C>> {
let mut action: Action<'t, P, C> = Action::Prove;
loop {
let result = match action {
Action::Prove => match self.task.clone().next() {
Some(lit) => self.chk(lit),
None => self.fulfill_promise(),
},
Action::Reduce(lit, skip) => self.red(lit, skip),
Action::Extend(lit, contras, skip) => self.ext(lit, contras, skip),
};
match result {
Ok(next) => action = next,
Err(true) => return Some(&self.proof),
Err(false) => return None,
}
}
}
pub fn inferences(&self) -> usize {
self.inferences
}
fn chk(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("checks: {}", lit);
debug!("{} {}", self.literals, lit.head());
debug!("lemmas: {}", self.ctx.lemmas.len());
debug!("path: {}", self.ctx.path.len());
self.literals += 1;
let mut lits = self.task.clone();
let mut path = self.ctx.path.iter();
let mut lemmas = self.ctx.lemmas.iter();
if lits.any(|cl| path.any(|pl| pl.eq_mod(&self.sub, &cl))) {
debug!("regularity");
self.try_alternative()
} else if lemmas.any(|lem| lem.eq_mod(&self.sub, &lit)) {
debug!("lemma");
self.proof.push(Action::Prove);
// do not add lit to lemmas, unlike original leanCoP
// furthermore, do not try red/ext steps if we found a lemma,
// because it does not add anything to substitution
// note that Jens said that this might sometimes be counterproductive,
// because adding to the substitution is also beneficial to cut down search space
self.task.next();
Ok(Action::Prove)
} else {
Ok(Action::Reduce(lit, 0))
}
}
fn red(&mut self, lit: OLit<'t, P, C>, skip: usize) -> State<'t, P, C> {
debug!("reduce: {}", lit);
let alternative = Alternative::from(&*self);
for (pidx, pat) in self.ctx.path.iter().rev().enumerate().skip(skip) {
debug!("try reduce: {}", pat);
let sub_dom_len = self.sub.get_dom_len();
if pat.head() != &-lit.head().clone() {
continue;
}
if pat.args().unify(&mut self.sub, lit.args()) {
debug!("reduce succeeded");
self.proof.push(Action::Reduce(lit, pidx));
if !self.opt.cuts.reduction {
let action = Action::Reduce(lit, pidx + 1);
self.alternatives.push((alternative, action));
}
self.ctx.lemmas.push(lit);
self.task.next();
return Ok(Action::Prove);
} else {
self.sub.set_dom_len(sub_dom_len)
}
}
self.ext0(lit)
}
fn ext0(&mut self, lit: OLit<'t, P, C>) -> State<'t, P, C> {
debug!("extend: {}", lit); | }
}
fn ext(&mut self, lit: OLit<'t, P, C>, cs: Contras<'t, P, C>, skip: usize) -> State<'t, P, C> {
let alt = Alternative::from(&*self);
let prm = Promise::from(&*self);
let sub = SubPtr::from(&self.sub);
for (eidx, entry) in cs.iter().enumerate().skip(skip) {
debug!(
"try extend {}{} (lit = {}, |path| = {})",
lit.head(),
entry,
lit,
self.ctx.path.len()
);
if self.ctx.path.len() >= self.opt.lim && entry.vars.is_some() {
debug!("path limit reached");
continue;
};
let eargs = Offset::new(sub.dom_max(), &entry.args);
if let Some(vars) = entry.vars {
// we have to add 1 here because the lowest variable is 0
self.sub.set_dom_max(sub.dom_max() + vars + 1)
};
debug!("unify {} ~? {}, sub = {}", eargs, lit.args(), self.sub);
if eargs.unify(&mut self.sub, lit.args()) {
debug!("unify succeeded with {}, sub = {}", entry.rest, self.sub);
self.inferences += 1;
// promise to fulfill the current task
// (if the promise is kept and cut is enabled,
// then all alternatives that came after will be discarded)
self.promises.push(prm);
self.proof.push(Action::Extend(lit, cs, eidx));
let action = Action::Extend(lit, cs, eidx + 1);
// register an alternative (that will be discarded
// if the above promise is kept and cut is enabled)
self.alternatives.push((alt, action));
self.task = Task::new(Offset::new(sub.dom_max(), &entry.rest));
self.ctx.path.push(lit);
return Ok(Action::Prove);
} else {
debug!("unify failed");
self.sub.rewind(&sub)
}
}
self.try_alternative()
}
fn fulfill_promise(&mut self) -> State<'t, P, C> {
debug!("fulfill promise ({} left)", self.promises.len());
let prm = self.promises.pop().ok_or(true)?;
self.task = prm.task;
self.ctx.rewind(prm.ctx_ptr);
if let Some(prev) = self.task.next() {
self.ctx.lemmas.push(prev)
};
if let Some(cut) = self.opt.cuts.extension {
use super::cuts::Cut::*;
let alt_len = match cut {
| let neg = -lit.head().clone();
match self.db.get(&neg) {
Some(entries) => self.ext(lit, entries, 0),
None => self.try_alternative(), | random_line_split |
ESealedRadioactiveSourceDetail.js | $("#sourceType").val("0");
$("#sourceType1").html("");
var sourceType = sourceCode.substr(sourceCode.length-1,1);
return;
}
//根据最后一个字符生成类别
var sourceType = sourceCode.substr(sourceCode.length-1,1);
if(sourceType == "N"){
$("#sourceType").val("0");
$("#sourceType1").html("");
}
if(sourceType == 1){
$("#sourceType").val("1");
$("#sourceType1").html("I");
}
if(sourceType == 2){
$("#sourceType").val("2");
$("#sourceType1").html("II");
}
if(sourceType == 3){
$("#sourceType").val("3");
$("#sourceType1").html("III");
}
if(sourceType == 4){
$("#sourceType").val("4");
$("#sourceType1").html("IV");
}
if(sourceType == 5){
$("#sourceType").val("5");
$("#sourceType1").html("V");
}
});
ESealedRadioactiveSourceDetail.initValidate();
},
initValidate : function(){
//提交
$("#ESealedRadioactiveSourceForm").find(".a-submit").click(function(){
if ($("#ESealedRadioactiveSourceForm").valid()) {
var data = {
id : $("#ESealedRadioactiveSourceForm").find("#id").val(),
nuclideId : $("#ESealedRadioactiveSourceForm").find("#nuclideId").val(),// 核素表id
staffId : $("#ESealedRadioactiveSourceForm").find("#staffId").val(),// 存储用户人的id
areaId : $("#ESealedRadioactiveSourceForm").find("#areaId").val(),//区域表的id
nuclideManufacturers : $("#ESealedRadioactiveSourceForm").find("#nuclideManufacturers").val(),//核素生产厂家id
activityExponent : $("#ESealedRadioactiveSourceForm").find("#activityExponent").val(),//活度指数
activity : $("#ESealedRadioactiveSourceForm").find("#activity").val(),//活度正数
measureTime : $("#ESealedRadioactiveSourceForm").find("#measureTime").val(),//测量日期
sourceCode : $("#ESealedRadioactiveSourceForm").find("#sourceCode").val(),//放射编号
purpose : $("#ESealedRadioactiveSourceForm").find("#purpose").val(),//用途
sourceType : $("#ESealedRadioactiveSourceForm").find("#sourceType").val(),// 放射源类别
activityType : $("#ESealedRadioactiveSourceForm").find("#activityType").val(),// 活动种类
nuclideName : $("#ESealedRadioactiveSourceForm").find("#nuclideName").val()//核素名称
};
$.post(home_url + "/EDevice/sealRadiate/submit",data,
function (response) {
if (response.code == 0){//提示 确认后跳转到查看
gridAddCallback("ESealedRadioactiveSourceTable");//刷新父页面table
location.href = home_url + "/EDevice/sealRadiate/lookOver?id=" + response.data;
}
}, 'json');
}
});
//验证编号一二位 生产单位代码
jQuery.validator.addMethod("verifySourceCodeUnit", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
var state = new Array(
"AR", "GR", "PL", "PT", "RU" ,"SI", "SK" ,"RO" ,"ZA", "SE" ,"ZZ", "TR" ,"GB",
"UA", "US" ,"UZ" ,"ES" ,"HU", "IN" ,"IE", "ID" ,"IS" ,"IL" ,"JP",
"NO","KZ","KG","KR","MX","NL","IT","AU","BE","BG","BR","BY","CA","CZ","DE","DK",
"EE","EG","FI","FR","HR"
);//国家名称代码
var str = SourceCode.substring(0,2);//获取1 2位编号
var b = false;//判断1 2位编号是否合格 true为不合格
if($.inArray(str, state) == -1){//不包含
b = true;
}
if(str == "NN" || !b || !isNaN(str)){
return true;
}
return false;
});
//验证编号三四位 出厂年份
jQuery.validator.addMethod("verifySourceCodeYear", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var time = SourceCode.substring(2,4);//获取年份
if(/^[0-9N]{2}$/.test(time) ){//不是数字或NN
return true;
}
return false;
});
//验证编号五六位 核素代码
jQuery.validator.addMethod("verifySourceCodeNuclide", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var str = SourceCode.substring(4,6);//获取5 6位编号()核素代码
if(str == "NN"){
return true;
}
var nuclide = new Array(//现在能用的核素 英文名
"H3","FE","C7","CO","NI","GE","SE","KR","CD","CS","PM",
"GD","IR","PO","RA","P8","AM","CM","CF"
);
if($.inArray(str, nuclide) == -1){//不包含
return false;
}
var nuclide2 = new Array(//现在能用的核素 中文名
"氢-3","铁-55","钴-57","钴-60","镍-63","锗-68","硒-75","氪-85","镉-109","铯-137","钷-147",
"钆-153","铱-192","钋-210","镭-226","钚-238","镅-241","锔-244","锎-252",
"磷-32","钼-99","钯-103","碘-125","碘-131","金-198"
);
//获取选取的核素名称
var element = $("#element span").html();
if($.inArray(element, nuclide2) == -1){//不包含
return false;
}
for(i = 0; i < nuclide.length; i++){//看元素是否对应 代码 和名字
if(element == nuclide2[i]){
if(str != nuclide[i]){//编号5 6位是否和核素英文代码相同
return false;
|
if( /^[A-Z0-9\d]{2}$/.test(str) ){
return true;
}
return false;
});
//验证编号是否正确
jQuery.validator.addMethod("verifySourceCode", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
if( /^[A-Z\d]{2}[0-9N]{2}[A-Z0-9\d]{2}[\d]{5}[1-5N]$/.test(SourceCode) ){
return true;
}
return false;
});
//所在活度正数不为空时指数 没有值时 指数默认为1
$("#activity").change(function(){
var str = $(this).val();
var activityExponent = $("#activityExponent").val();//指数
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
if(str == ""){
$("#activityExponent").val("")
}
});
//当活度正数有值时候 指数不能为 空
$("#activityExponent").change(function(){
var activityExponent = $(this).val();
var str = $("#activity").val();
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
| }
break;
}
}
| conditional_block |
ESealedRadioactiveSourceDetail.js | 2}$/.test(time) ){//不是数字或NN
return true;
}
return false;
});
//验证编号五六位 核素代码
jQuery.validator.addMethod("verifySourceCodeNuclide", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var str = SourceCode.substring(4,6);//获取5 6位编号()核素代码
if(str == "NN"){
return true;
}
var nuclide = new Array(//现在能用的核素 英文名
"H3","FE","C7","CO","NI","GE","SE","KR","CD","CS","PM",
"GD","IR","PO","RA","P8","AM","CM","CF"
);
if($.inArray(str, nuclide) == -1){//不包含
return false;
}
var nuclide2 = new Array(//现在能用的核素 中文名
"氢-3","铁-55","钴-57","钴-60","镍-63","锗-68","硒-75","氪-85","镉-109","铯-137","钷-147",
"钆-153","铱-192","钋-210","镭-226","钚-238","镅-241","锔-244","锎-252",
"磷-32","钼-99","钯-103","碘-125","碘-131","金-198"
);
//获取选取的核素名称
var element = $("#element span").html();
if($.inArray(element, nuclide2) == -1){//不包含
return false;
}
for(i = 0; i < nuclide.length; i++){//看元素是否对应 代码 和名字
if(element == nuclide2[i]){
if(str != nuclide[i]){//编号5 6位是否和核素英文代码相同
return false;
}
break;
}
}
if( /^[A-Z0-9\d]{2}$/.test(str) ){
return true;
}
return false;
});
//验证编号是否正确
jQuery.validator.addMethod("verifySourceCode", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
if( /^[A-Z\d]{2}[0-9N]{2}[A-Z0-9\d]{2}[\d]{5}[1-5N]$/.test(SourceCode) ){
return true;
}
return false;
});
//所在活度正数不为空时指数 没有值时 指数默认为1
$("#activity").change(function(){
var str = $(this).val();
var activityExponent = $("#activityExponent").val();//指数
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
if(str == ""){
$("#activityExponent").val("")
}
});
//当活度正数有值时候 指数不能为 空
$("#activityExponent").change(function(){
var activityExponent = $(this).val();
var str = $("#activity").val();
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
});
$("#ESealedRadioactiveSourceForm").validate({
rules: {
nuclideId: {//核素名称
required: true
},
purpose : {//用途
required:true
},
areaId: {//场所
required: true
},
sourceCode : {//放射源编号
rangelength : [12,12],
verifySourceCodeUnit : true,
verifySourceCodeYear : true,
verifySourceCodeNuclide : true,
verifySourceCode : true
},
activityExponent : {//活度指数
digits:true,
range:[0,20]
},
activity : {//活度 正数
number : true,
verifyfigure : true
}
},
messages: {
sourceCode : {//放射源编号
rangelength :$.validator.format("<div class='tisyz'><i class='tisbefore'></i>请输入{0}位的放射源编号</div>"),
verifySourceCodeUnit : "<div class='tisyz'><i class='tisbefore'></i>一二位是生产单位代码</div>",
verifySourceCodeYear : "<div class='tisyz'><i class='tisbefore'></i>三四位是出厂年份后两位或NN</div>",
verifySourceCodeNuclide : "<div class='tisyz'><i class='tisbefore'></i>五六位是核素代码编号或NN</div>",
verifySourceCode : "<div class='tisyz'><i class='tisbefore'></i>放射编号后六位不正确</div>"
},
activityExponent : {
digits:"<div class='tisyz'><i class='tisbefore'></i>请输入0到20之间的整数</div>",
range:"<div class='tisyz'><i class='tisbefore'></i>请输入0到20之间的整数</div>"
},
activity : {
number : "<div class='tisyz'><i class='tisbefore'></i>请输入正数且4位有效数字</div>",
verifyfigure : "<div class='tisyz'><i class='tisbefore'></i>请输入正数且4位有效数字</div>"
}
}
});
}
}
})();
jQuery(document).ready(function() {
ESealedRadioactiveSourceDetail.init();//初始化
});
//核素生产厂家 单选
function getNuclideManufacturers(){
Company_Single(function(data){
$("#nuclideManufacturersList1").html("");//清空数据
//保存生产厂家的id
$("#nuclideManufacturers").val($(data).val());
$("#nuclideManufacturersList1").append(data);
$("#nuclideManufacturers").focus().blur()//光标切入移除
},5);//1.个人剂量监测单位;2.职业健康检查单位;3.评价与检测机构; 4.供货商;5.生产厂家;6.设计单位;7.施工单位;
}
//区域单选
function selectArea1() {
SelectRoom_Single(function(data) {
$("#areaList1").children().remove();//单选 删除里面多余的数据
$("#hospitalArea").html("");//清空 所在院区
//所在机房或区域 id
$("#areaId").val("");
if(data.id == undefined || data.id == ""){
return;
}
//本级区域
$("#areaList1").append(dataNodeSelected(data.id, data.name));
//生成所在院区
$("#hospitalArea").html(data.areaName);
//所在机房或区域 id
$("#areaId").val(data.id);
$("#areaId").focus().blur()//光标切入移除
});
}
//人员单选
function selectStaff1() {
SelectStaff_Single(function(data) {
$("#staffList1 li").remove();
$("#staffId").val("");
if(data.id == undefined || data.id == ""){
return;
}
$("#staffList1").append(dataNodeSelected(data.id, data.name));
//保存 人员id
$("#staffId").val(data.id);
$("#staffId").focus().blur()//光标切入移除
});
}
//元素周期表
function getElement(){
element(function(data){
$("#element").html("");
if( $(data).attr("name") == 'undefined' || $(data).attr("name") == undefined){//判断数据是否为null
$("#nuclideName").val("");//元素名称
$("#nuclideId").val("");//id
return;
}
$("#element").append(data);
//保存数据
$("#nuclideName").val($(data).attr("name"));//元素名称
$("#nuclideId").val($(data).val());//id
$("#nuclideId").focus().blur()//光标切入移除
});
}
//数组去重
Array.prototype.unique3 = function(){
var res = [];
var json = {};
for(var i = 0; i < this.length; i++){
if(!json[this[i]]){
res.push(this[i]);
json[this[i]] = 1;
}
}
return res;
} | identifier_body | ||
ESealedRadioactiveSourceDetail.js | $("#sourceType").val("3");
$("#sourceType1").html("III");
}
if(sourceType == 4){
$("#sourceType").val("4");
$("#sourceType1").html("IV");
}
if(sourceType == 5){
$("#sourceType").val("5");
$("#sourceType1").html("V");
}
});
ESealedRadioactiveSourceDetail.initValidate();
},
initValidate : function(){
//提交
$("#ESealedRadioactiveSourceForm").find(".a-submit").click(function(){
if ($("#ESealedRadioactiveSourceForm").valid()) {
var data = {
id : $("#ESealedRadioactiveSourceForm").find("#id").val(),
nuclideId : $("#ESealedRadioactiveSourceForm").find("#nuclideId").val(),// 核素表id
staffId : $("#ESealedRadioactiveSourceForm").find("#staffId").val(),// 存储用户人的id
areaId : $("#ESealedRadioactiveSourceForm").find("#areaId").val(),//区域表的id
nuclideManufacturers : $("#ESealedRadioactiveSourceForm").find("#nuclideManufacturers").val(),//核素生产厂家id
activityExponent : $("#ESealedRadioactiveSourceForm").find("#activityExponent").val(),//活度指数
activity : $("#ESealedRadioactiveSourceForm").find("#activity").val(),//活度正数
measureTime : $("#ESealedRadioactiveSourceForm").find("#measureTime").val(),//测量日期
sourceCode : $("#ESealedRadioactiveSourceForm").find("#sourceCode").val(),//放射编号
purpose : $("#ESealedRadioactiveSourceForm").find("#purpose").val(),//用途
sourceType : $("#ESealedRadioactiveSourceForm").find("#sourceType").val(),// 放射源类别
activityType : $("#ESealedRadioactiveSourceForm").find("#activityType").val(),// 活动种类
nuclideName : $("#ESealedRadioactiveSourceForm").find("#nuclideName").val()//核素名称
};
$.post(home_url + "/EDevice/sealRadiate/submit",data,
function (response) {
if (response.code == 0){//提示 确认后跳转到查看
gridAddCallback("ESealedRadioactiveSourceTable");//刷新父页面table
location.href = home_url + "/EDevice/sealRadiate/lookOver?id=" + response.data;
}
}, 'json');
}
});
//验证编号一二位 生产单位代码
jQuery.validator.addMethod("verifySourceCodeUnit", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
var state = new Array(
"AR", "GR", "PL", "PT", "RU" ,"SI", "SK" ,"RO" ,"ZA", "SE" ,"ZZ", "TR" ,"GB",
"UA", "US" ,"UZ" ,"ES" ,"HU", "IN" ,"IE", "ID" ,"IS" ,"IL" ,"JP",
"NO","KZ","KG","KR","MX","NL","IT","AU","BE","BG","BR","BY","CA","CZ","DE","DK",
"EE","EG","FI","FR","HR"
);//国家名称代码
var str = SourceCode.substring(0,2);//获取1 2位编号
var b = false;//判断1 2位编号是否合格 true为不合格
if($.inArray(str, state) == -1){//不包含
b = true;
}
if(str == "NN" || !b || !isNaN(str)){
return true;
}
return false;
});
//验证编号三四位 出厂年份
jQuery.validator.addMethod("verifySourceCodeYear", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var time = SourceCode.substring(2,4);//获取年份
if(/^[0-9N]{2}$/.test(time) ){//不是数字或NN
return true;
}
return false;
});
//验证编号五六位 核素代码
jQuery.validator.addMethod("verifySourceCodeNuclide", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var str = SourceCode.substring(4,6);//获取5 6位编号()核素代码
if(str == "NN"){
return true;
}
var nuclide = new Array(//现在能用的核素 英文名
"H3","FE","C7","CO","NI","GE","SE","KR","CD","CS","PM",
"GD","IR","PO","RA","P8","AM","CM","CF"
);
if($.inArray(str, nuclide) == -1){//不包含
return false;
}
var nuclide2 = new Array(//现在能用的核素 中文名
"氢-3","铁-55","钴-57","钴-60","镍-63","锗-68","硒-75","氪-85","镉-109","铯-137","钷-147",
"钆-153","铱-192","钋-210","镭-226","钚-238","镅-241","锔-244","锎-252",
"磷-32","钼-99","钯-103","碘-125","碘-131","金-198"
);
//获取选取的核素名称
var element = $("#element span").html();
if($.inArray(element, nuclide2) == -1){//不包含
return false;
}
for(i = 0; i < nuclide.length; i++){//看元素是否对应 代码 和名字
if(element == nuclide2[i]){
if(str != nuclide[i]){//编号5 6位是否和核素英文代码相同
return false;
}
break;
}
}
if( /^[A-Z0-9\d]{2}$/.test(str) ){
return true;
}
return false;
});
//验证编号是否正确
jQuery.validator.addMethod("verifySourceCode", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
if( /^[A-Z\d]{2}[0-9N]{2}[A-Z0-9\d]{2}[\d]{5}[1-5N]$/.test(SourceCode) ){
return true;
}
return false;
});
//所在活度正数不为空时指数 没有值时 指数默认为1
$("#activity").change(function(){
var str = $(this).val();
var activityExponent = $("#activityExponent").val();//指数
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
if(str == ""){
$("#activityExponent").val("")
}
});
//当活度正数有值时候 指数不能为 空
$("#activityExponent").change(function(){
var activityExponent = $(this).val();
var str = $("#activity").val();
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
});
$("#ESealedRadioactiveSourceForm").validate({
rules: {
nuclideId: {//核素名称
required: true
},
purpose : {//用途
required:true
},
areaId: {//场所
required: true
},
sourceCode : {//放射源编号
rangelength : [12,12],
verifySourceCodeUnit : true,
verifySourceCodeYear : true,
verifySourceCodeNuclide : true,
verifySourceCode : true
},
activityExponent : {//活度指数
digits:true,
range:[0,20]
},
activity : {//活度 正数
number : true,
verifyfigure : true
| }
| random_line_split | |
ESealedRadioactiveSourceDetail.js | jQuery.validator.addMethod("verifySourceCodeYear", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var time = SourceCode.substring(2,4);//获取年份
if(/^[0-9N]{2}$/.test(time) ){//不是数字或NN
return true;
}
return false;
});
//验证编号五六位 核素代码
jQuery.validator.addMethod("verifySourceCodeNuclide", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
var str = SourceCode.substring(4,6);//获取5 6位编号()核素代码
if(str == "NN"){
return true;
}
var nuclide = new Array(//现在能用的核素 英文名
"H3","FE","C7","CO","NI","GE","SE","KR","CD","CS","PM",
"GD","IR","PO","RA","P8","AM","CM","CF"
);
if($.inArray(str, nuclide) == -1){//不包含
return false;
}
var nuclide2 = new Array(//现在能用的核素 中文名
"氢-3","铁-55","钴-57","钴-60","镍-63","锗-68","硒-75","氪-85","镉-109","铯-137","钷-147",
"钆-153","铱-192","钋-210","镭-226","钚-238","镅-241","锔-244","锎-252",
"磷-32","钼-99","钯-103","碘-125","碘-131","金-198"
);
//获取选取的核素名称
var element = $("#element span").html();
if($.inArray(element, nuclide2) == -1){//不包含
return false;
}
for(i = 0; i < nuclide.length; i++){//看元素是否对应 代码 和名字
if(element == nuclide2[i]){
if(str != nuclide[i]){//编号5 6位是否和核素英文代码相同
return false;
}
break;
}
}
if( /^[A-Z0-9\d]{2}$/.test(str) ){
return true;
}
return false;
});
//验证编号是否正确
jQuery.validator.addMethod("verifySourceCode", function(value, element) {
var SourceCode = $.trim($("#ESealedRadioactiveSourceForm").find("#sourceCode").val());
if(SourceCode == ''){
return true;
}
if( /^[A-Z\d]{2}[0-9N]{2}[A-Z0-9\d]{2}[\d]{5}[1-5N]$/.test(SourceCode) ){
return true;
}
return false;
});
//所在活度正数不为空时指数 没有值时 指数默认为1
$("#activity").change(function(){
var str = $(this).val();
var activityExponent = $("#activityExponent").val();//指数
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
if(str == ""){
$("#activityExponent").val("")
}
});
//当活度正数有值时候 指数不能为 空
$("#activityExponent").change(function(){
var activityExponent = $(this).val();
var str = $("#activity").val();
if(str != "" && activityExponent == ""){
$("#activityExponent").val("0")
}
});
$("#ESealedRadioactiveSourceForm").validate({
rules: {
nuclideId: {//核素名称
required: true
},
purpose : {//用途
required:true
},
areaId: {//场所
required: true
},
sourceCode : {//放射源编号
rangelength : [12,12],
verifySourceCodeUnit : true,
verifySourceCodeYear : true,
verifySourceCodeNuclide : true,
verifySourceCode : true
},
activityExponent : {//活度指数
digits:true,
range:[0,20]
},
activity : {//活度 正数
number : true,
verifyfigure : true
}
},
messages: {
sourceCode : {//放射源编号
rangelength :$.validator.format("<div class='tisyz'><i class='tisbefore'></i>请输入{0}位的放射源编号</div>"),
verifySourceCodeUnit : "<div class='tisyz'><i class='tisbefore'></i>一二位是生产单位代码</div>",
verifySourceCodeYear : "<div class='tisyz'><i class='tisbefore'></i>三四位是出厂年份后两位或NN</div>",
verifySourceCodeNuclide : "<div class='tisyz'><i class='tisbefore'></i>五六位是核素代码编号或NN</div>",
verifySourceCode : "<div class='tisyz'><i class='tisbefore'></i>放射编号后六位不正确</div>"
},
activityExponent : {
digits:"<div class='tisyz'><i class='tisbefore'></i>请输入0到20之间的整数</div>",
range:"<div class='tisyz'><i class='tisbefore'></i>请输入0到20之间的整数</div>"
},
activity : {
number : "<div class='tisyz'><i class='tisbefore'></i>请输入正数且4位有效数字</div>",
verifyfigure : "<div class='tisyz'><i class='tisbefore'></i>请输入正数且4位有效数字</div>"
}
}
});
}
}
})();
jQuery(document).ready(function() {
ESealedRadioactiveSourceDetail.init();//初始化
});
//核素生产厂家 单选
function getNuclideManufacturers(){
Company_Single(function(data){
$("#nuclideManufacturersList1").html("");//清空数据
//保存生产厂家的id
$("#nuclideManufacturers").val($(data).val());
$("#nuclideManufacturersList1").append(data);
$("#nuclideManufacturers").focus().blur()//光标切入移除
},5);//1.个人剂量监测单位;2.职业健康检查单位;3.评价与检测机构; 4.供货商;5.生产厂家;6.设计单位;7.施工单位;
}
//区域单选
function selectArea1() {
SelectRoom_Single(function(data) {
$("#areaList1").children().remove();//单选 删除里面多余的数据
$("#hospitalArea").html("");//清空 所在院区
//所在机房或区域 id
$("#areaId").val("");
if(data.id == undefined || data.id == ""){
return;
}
//本级区域
$("#areaList1").append(dataNodeSelected(data.id, data.name));
//生成所在院区
$("#hospitalArea").html(data.areaName);
//所在机房或区域 id
$("#areaId").val(data.id);
$("#areaId").focus().blur()//光标切入移除
});
}
//人员单选
function selectStaff1() {
SelectStaff_Single(function(data) {
$("#staffList1 li").remove();
$("#staffId").val("");
if(data.id == undefined || data.id == ""){
return;
}
$("#staffList1").append(dataNodeSelected(data.id, data.name));
//保存 人员id
$("#staffId").val(data.id);
$("#staffId").focus().blur()//光标切入移除
});
}
//元素周期表
function getElement(){
element(function(data){
$("#element").html("");
if( $(data).attr("name") == 'undefined' || $(data).attr("name") == undefined){//判断数据是否为null
$("#nuclideName").val("");//元素名称
$("#nuclideId").val("");//id
return;
}
$("#element").append(data);
//保存数据
$("#nuclideName").val($(data).attr("name"));//元素名称
$("#nuclideId").val($(data).val());//id
$("#nuclideId").fo | cus().blur( | identifier_name | |
array.ts | < inputs.length; index++)
inputs[index] = getNewValue(inputs[index], index)
}
/**
Delete elements of an array not satisfying `predicate`.
If `predicate` throws, the array will be left in a bad state.
*/
export function filterMutate<T>(inputs: T[], predicate: Predicate<T>): void {
let writeIndex = 0
for (const input of inputs)
if (predicate(input)) {
inputs[writeIndex] = input
writeIndex++
}
inputs.length = writeIndex
}
/**
Replace elements with the result of [[tryGetOutput]] or delete them if that returns `undefined`.
If [[tryGetOutput] throws, the array will be left in a bad state.
*/
export function mapDefinedMutate<T>(inputs: T[], tryGetOutput: (input: T) => Option<T>): void {
let writeIndex = 0
for (const input of inputs) {
const output = tryGetOutput(input)
if (output !== undefined) {
inputs[writeIndex] = output
writeIndex++
}
}
inputs.length = writeIndex
}
/** Change the value at a single index in an array by applying a function to it. */
export function mutate<T>(inputs: T[], index: Nat, transform: (t: T) => T): void {
checkIndex(inputs, index)
inputs[index] = transform(inputs[index])
}
/**
Remove an element from an array and do not preserve the array's order.
Useful for arrays used to represent small sets.
Returns whether the value was successfully removed.
*/
export function removeUnordered<T>(inputs: T[], value: T, equal?: Comparer<T>): boolean {
for (let i = 0; i < inputs.length; i++)
if (exists(equal) ? equal(inputs[i], value) : inputs[i] === value) {
inputs[i] = last(inputs)!
inputs.length--
return true
}
return false
}
/**
Mutate [[inputs]] by combining them with each in [[other]].
If [[other]] is shorter than [[inputs]], this will reduce [[inputs]] in length.
If [[other]] is longer, the extra entries are ignored.
*/
export function zipMutate<T, U>(inputs: T[], other: Iterable<U>, zipper: (input: T, other: U) => T): void {
const iter = other[Symbol.iterator]()
for (let index = 0; index < inputs.length; index++) {
const { value, done } = iter.next()
if (done) {
inputs.length = index
break
}
inputs[index] = zipper(inputs[index], value)
}
}
/** Provides async utilities for an array. */
export function asyncArray<T>(inputs: T[]): AsyncArrayOps<T> {
return new AsyncArrayOps(inputs)
}
/**
Wrapper class for utilities that mutate arrays asynchronously.
For non-mutating utilities use [[AsyncSeq]].
*/
export class AsyncArrayOps<T> {
constructor(private inputs: T[]) {}
/** Asynchronous [[mapMutate]]. */
async map(getNewValue: (element: T, index: number) => Promise<T>): Promise<void> {
const { inputs } = this
for (let index = 0; index < inputs.length; index++)
inputs[index] = await getNewValue(inputs[index], index)
}
/** Asynchronous [[filterMutate]]. */
async filter(predicate: (element: T) => Promise<boolean>): Promise<void> {
const { inputs } = this
let writeIndex = 0
for (let readIndex = 0; readIndex < inputs.length; readIndex++)
if (await predicate(inputs[readIndex])) {
inputs[writeIndex] = inputs[readIndex]
writeIndex++
}
inputs.length = writeIndex
}
/** Asynchronous [[mapDefinedMutate]]. Performs `tryGetOutput` one element at a time. */
async mapDefined(tryGetOutput: (input: T) => Promise<Option<T>>): Promise<void> {
const { inputs } = this
let writeIndex = 0
for (let readIndex = 0; readIndex < inputs.length; readIndex++) {
const output = await tryGetOutput(inputs[readIndex])
if (output !== undefined) {
inputs[writeIndex] = output
writeIndex++
}
}
inputs.length = writeIndex
}
/** Asynchronous [[mutate]]. */
async mutate(index: Nat, transform: (t: T) => Promise<T>): Promise<void> {
const { inputs } = this
checkIndex(inputs, index)
inputs[index] = await transform(inputs[index])
}
}
/** Provides parallel utilities for an array. */
export function parallelArray<T>(inputs: T[], maxNumberOfThreads?: number): ParallelArrayOps<T> {
return new ParallelArrayOps(inputs, maxNumberOfThreads)
}
/**
Wrapper class for utilities that mutate arrays in parallel.
For non-mutating utilities use [[ParallelSeq]].
*/
export class ParallelArrayOps<T> {
/** Use [[parallelArray]] rather than calling this directly. */
constructor(readonly inputs: T[], readonly maxNumberOfThreads: number = Number.POSITIVE_INFINITY) {
if (maxNumberOfThreads !== Number.POSITIVE_INFINITY)
checkNat(maxNumberOfThreads)
}
/** Parallel [[mapMutate]]. */ |
let writeIndex = 0
let readIndex = 0
while (readIndex < maxNumberOfThreads && readIndex < inputs.length)
startOne()
while (readIndex < inputs.length) {
await awaitOne()
startOne()
}
while (writeIndex < inputs.length)
await awaitOne()
async function awaitOne(): Promise<void> {
inputs[writeIndex] = await (inputs as any as Array<Promise<T>>)[writeIndex]
writeIndex++
}
function startOne(): void {
(inputs as any as Array<Promise<T>>)[readIndex] = mapper(inputs[readIndex], readIndex)
readIndex++
}
}
/** Parallel [[filterMutate]]. */
filter(predicate: (element: T, index: number) => Promise<boolean>): Promise<void> {
return this.mapDefined(async (input, index) =>
optional(await predicate(input, index), () => input))
}
/** Parallel [[mapDefinedMutate]]. */
async mapDefined(tryGetOutput: (input: T, index: number) => Promise<Option<T>>): Promise<void> {
const { inputs, maxNumberOfThreads } = this
/** Next index to write a (defined) result to. */
let writeOutputIndex = 0
/** Next index to await a thread at. */
let readPromiseIndex = 0
/** Next index to read an input value from; the thread for that input will be written to the same index. */
let readValueIndex = 0
// Start initial threads.
while (readValueIndex < maxNumberOfThreads && readValueIndex < inputs.length)
startOne()
// Keep awaiting threads and starting new ones.
// Invariants: writeIndex <= readPromiseIndex, readPromiseIndex = readValueIndex - numberOfThreads
while (readValueIndex < inputs.length) {
await awaitOne()
startOne()
}
// Await remaining threads.
while (readPromiseIndex < inputs.length)
await awaitOne()
// Shorten array to new length.
inputs.length = writeOutputIndex
async function awaitOne(): Promise<void> {
const output = await (inputs as any as Array<Promise<Option<T>>>)[readPromiseIndex]
readPromiseIndex++
if (output !== undefined) {
inputs[writeOutputIndex] = output
writeOutputIndex++
}
}
function startOne(): void {
(inputs as any as Array<Promise<Option<T>>>)[readValueIndex] = tryGetOutput(inputs[readValueIndex], readValueIndex)
readValueIndex++
}
}
}
/**
Whether a number is an integer between 0 and array.length.
Does *not* check for whether there is a "hole" at the index.
*/
export function isValidIndex(inputs: Array<{}>, index: Nat): boolean {
return isNat(index) && index < inputs.length
}
/** Throws an error if [[index]] is not a valid index. */
export function checkIndex(inputs: Array<{}>, index: Nat): void {
if (!isValidIndex(inputs, index))
throw new Error(`Expected an array index < ${inputs.length}, got ${index}`)
}
/** Swap two values in an array. */
export function swap(inputs: Array<{}>, firstIndex: Nat, secondIndex: Nat): void {
checkIndex(inputs, firstIndex)
checkIndex(inputs, secondIndex)
const tmp = inputs[firstIndex]
inputs[firstIndex] = inputs[secondIndex]
inputs[secondIndex] = tmp
}
/** Initialize a new array by calling [[makeElement]] [[length]] times. */
export function initArray<T>(length: number, makeElement: (index: number) => T): T[] {
const arr = new Array(length)
for (let i = 0; i < length; i++)
arr[i] = makeElement(i)
return arr
}
/** Asynchronous [[initArray]]. */
export function initArrayAsync<T>(length: number, | async map(mapper: (element: T, index: number) => Promise<T>): Promise<void> {
const { inputs, maxNumberOfThreads } = this | random_line_split |
array.ts | < inputs.length; index++)
inputs[index] = getNewValue(inputs[index], index)
}
/**
Delete elements of an array not satisfying `predicate`.
If `predicate` throws, the array will be left in a bad state.
*/
export function filterMutate<T>(inputs: T[], predicate: Predicate<T>): void {
let writeIndex = 0
for (const input of inputs)
if (predicate(input)) {
inputs[writeIndex] = input
writeIndex++
}
inputs.length = writeIndex
}
/**
Replace elements with the result of [[tryGetOutput]] or delete them if that returns `undefined`.
If [[tryGetOutput] throws, the array will be left in a bad state.
*/
export function mapDefinedMutate<T>(inputs: T[], tryGetOutput: (input: T) => Option<T>): void {
let writeIndex = 0
for (const input of inputs) {
const output = tryGetOutput(input)
if (output !== undefined) {
inputs[writeIndex] = output
writeIndex++
}
}
inputs.length = writeIndex
}
/** Change the value at a single index in an array by applying a function to it. */
export function mutate<T>(inputs: T[], index: Nat, transform: (t: T) => T): void {
checkIndex(inputs, index)
inputs[index] = transform(inputs[index])
}
/**
Remove an element from an array and do not preserve the array's order.
Useful for arrays used to represent small sets.
Returns whether the value was successfully removed.
*/
export function removeUnordered<T>(inputs: T[], value: T, equal?: Comparer<T>): boolean {
for (let i = 0; i < inputs.length; i++)
if (exists(equal) ? equal(inputs[i], value) : inputs[i] === value) {
inputs[i] = last(inputs)!
inputs.length--
return true
}
return false
}
/**
Mutate [[inputs]] by combining them with each in [[other]].
If [[other]] is shorter than [[inputs]], this will reduce [[inputs]] in length.
If [[other]] is longer, the extra entries are ignored.
*/
export function zipMutate<T, U>(inputs: T[], other: Iterable<U>, zipper: (input: T, other: U) => T): void {
const iter = other[Symbol.iterator]()
for (let index = 0; index < inputs.length; index++) {
const { value, done } = iter.next()
if (done) {
inputs.length = index
break
}
inputs[index] = zipper(inputs[index], value)
}
}
/** Provides async utilities for an array. */
export function asyncArray<T>(inputs: T[]): AsyncArrayOps<T> {
return new AsyncArrayOps(inputs)
}
/**
Wrapper class for utilities that mutate arrays asynchronously.
For non-mutating utilities use [[AsyncSeq]].
*/
export class AsyncArrayOps<T> {
constructor(private inputs: T[]) {}
/** Asynchronous [[mapMutate]]. */
async map(getNewValue: (element: T, index: number) => Promise<T>): Promise<void> {
const { inputs } = this
for (let index = 0; index < inputs.length; index++)
inputs[index] = await getNewValue(inputs[index], index)
}
/** Asynchronous [[filterMutate]]. */
async filter(predicate: (element: T) => Promise<boolean>): Promise<void> {
const { inputs } = this
let writeIndex = 0
for (let readIndex = 0; readIndex < inputs.length; readIndex++)
if (await predicate(inputs[readIndex])) {
inputs[writeIndex] = inputs[readIndex]
writeIndex++
}
inputs.length = writeIndex
}
/** Asynchronous [[mapDefinedMutate]]. Performs `tryGetOutput` one element at a time. */
async mapDefined(tryGetOutput: (input: T) => Promise<Option<T>>): Promise<void> {
const { inputs } = this
let writeIndex = 0
for (let readIndex = 0; readIndex < inputs.length; readIndex++) {
const output = await tryGetOutput(inputs[readIndex])
if (output !== undefined) {
inputs[writeIndex] = output
writeIndex++
}
}
inputs.length = writeIndex
}
/** Asynchronous [[mutate]]. */
async mutate(index: Nat, transform: (t: T) => Promise<T>): Promise<void> {
const { inputs } = this
checkIndex(inputs, index)
inputs[index] = await transform(inputs[index])
}
}
/** Provides parallel utilities for an array. */
export function parallelArray<T>(inputs: T[], maxNumberOfThreads?: number): ParallelArrayOps<T> {
return new ParallelArrayOps(inputs, maxNumberOfThreads)
}
/**
Wrapper class for utilities that mutate arrays in parallel.
For non-mutating utilities use [[ParallelSeq]].
*/
export class ParallelArrayOps<T> {
/** Use [[parallelArray]] rather than calling this directly. */
| (readonly inputs: T[], readonly maxNumberOfThreads: number = Number.POSITIVE_INFINITY) {
if (maxNumberOfThreads !== Number.POSITIVE_INFINITY)
checkNat(maxNumberOfThreads)
}
/** Parallel [[mapMutate]]. */
async map(mapper: (element: T, index: number) => Promise<T>): Promise<void> {
const { inputs, maxNumberOfThreads } = this
let writeIndex = 0
let readIndex = 0
while (readIndex < maxNumberOfThreads && readIndex < inputs.length)
startOne()
while (readIndex < inputs.length) {
await awaitOne()
startOne()
}
while (writeIndex < inputs.length)
await awaitOne()
async function awaitOne(): Promise<void> {
inputs[writeIndex] = await (inputs as any as Array<Promise<T>>)[writeIndex]
writeIndex++
}
function startOne(): void {
(inputs as any as Array<Promise<T>>)[readIndex] = mapper(inputs[readIndex], readIndex)
readIndex++
}
}
/** Parallel [[filterMutate]]. */
filter(predicate: (element: T, index: number) => Promise<boolean>): Promise<void> {
return this.mapDefined(async (input, index) =>
optional(await predicate(input, index), () => input))
}
/** Parallel [[mapDefinedMutate]]. */
async mapDefined(tryGetOutput: (input: T, index: number) => Promise<Option<T>>): Promise<void> {
const { inputs, maxNumberOfThreads } = this
/** Next index to write a (defined) result to. */
let writeOutputIndex = 0
/** Next index to await a thread at. */
let readPromiseIndex = 0
/** Next index to read an input value from; the thread for that input will be written to the same index. */
let readValueIndex = 0
// Start initial threads.
while (readValueIndex < maxNumberOfThreads && readValueIndex < inputs.length)
startOne()
// Keep awaiting threads and starting new ones.
// Invariants: writeIndex <= readPromiseIndex, readPromiseIndex = readValueIndex - numberOfThreads
while (readValueIndex < inputs.length) {
await awaitOne()
startOne()
}
// Await remaining threads.
while (readPromiseIndex < inputs.length)
await awaitOne()
// Shorten array to new length.
inputs.length = writeOutputIndex
async function awaitOne(): Promise<void> {
const output = await (inputs as any as Array<Promise<Option<T>>>)[readPromiseIndex]
readPromiseIndex++
if (output !== undefined) {
inputs[writeOutputIndex] = output
writeOutputIndex++
}
}
function startOne(): void {
(inputs as any as Array<Promise<Option<T>>>)[readValueIndex] = tryGetOutput(inputs[readValueIndex], readValueIndex)
readValueIndex++
}
}
}
/**
Whether a number is an integer between 0 and array.length.
Does *not* check for whether there is a "hole" at the index.
*/
export function isValidIndex(inputs: Array<{}>, index: Nat): boolean {
return isNat(index) && index < inputs.length
}
/** Throws an error if [[index]] is not a valid index. */
export function checkIndex(inputs: Array<{}>, index: Nat): void {
if (!isValidIndex(inputs, index))
throw new Error(`Expected an array index < ${inputs.length}, got ${index}`)
}
/** Swap two values in an array. */
export function swap(inputs: Array<{}>, firstIndex: Nat, secondIndex: Nat): void {
checkIndex(inputs, firstIndex)
checkIndex(inputs, secondIndex)
const tmp = inputs[firstIndex]
inputs[firstIndex] = inputs[secondIndex]
inputs[secondIndex] = tmp
}
/** Initialize a new array by calling [[makeElement]] [[length]] times. */
export function initArray<T>(length: number, makeElement: (index: number) => T): T[] {
const arr = new Array(length)
for (let i = 0; i < length; i++)
arr[i] = makeElement(i)
return arr
}
/** Asynchronous [[initArray]]. */
export function initArrayAsync<T>(length: number | constructor | identifier_name |
array.ts | inputs.length; index++)
inputs[index] = getNewValue(inputs[index], index)
}
/**
Delete elements of an array not satisfying `predicate`.
If `predicate` throws, the array will be left in a bad state.
*/
export function filterMutate<T>(inputs: T[], predicate: Predicate<T>): void {
let writeIndex = 0
for (const input of inputs)
if (predicate(input)) {
inputs[writeIndex] = input
writeIndex++
}
inputs.length = writeIndex
}
/**
Replace elements with the result of [[tryGetOutput]] or delete them if that returns `undefined`.
If [[tryGetOutput] throws, the array will be left in a bad state.
*/
export function mapDefinedMutate<T>(inputs: T[], tryGetOutput: (input: T) => Option<T>): void {
let writeIndex = 0
for (const input of inputs) {
const output = tryGetOutput(input)
if (output !== undefined) {
inputs[writeIndex] = output
writeIndex++
}
}
inputs.length = writeIndex
}
/** Change the value at a single index in an array by applying a function to it. */
export function mutate<T>(inputs: T[], index: Nat, transform: (t: T) => T): void {
checkIndex(inputs, index)
inputs[index] = transform(inputs[index])
}
/**
Remove an element from an array and do not preserve the array's order.
Useful for arrays used to represent small sets.
Returns whether the value was successfully removed.
*/
export function removeUnordered<T>(inputs: T[], value: T, equal?: Comparer<T>): boolean {
for (let i = 0; i < inputs.length; i++)
if (exists(equal) ? equal(inputs[i], value) : inputs[i] === value) {
inputs[i] = last(inputs)!
inputs.length--
return true
}
return false
}
/**
Mutate [[inputs]] by combining them with each in [[other]].
If [[other]] is shorter than [[inputs]], this will reduce [[inputs]] in length.
If [[other]] is longer, the extra entries are ignored.
*/
export function zipMutate<T, U>(inputs: T[], other: Iterable<U>, zipper: (input: T, other: U) => T): void {
const iter = other[Symbol.iterator]()
for (let index = 0; index < inputs.length; index++) {
const { value, done } = iter.next()
if (done) {
inputs.length = index
break
}
inputs[index] = zipper(inputs[index], value)
}
}
/** Provides async utilities for an array. */
export function asyncArray<T>(inputs: T[]): AsyncArrayOps<T> {
return new AsyncArrayOps(inputs)
}
/**
Wrapper class for utilities that mutate arrays asynchronously.
For non-mutating utilities use [[AsyncSeq]].
*/
export class AsyncArrayOps<T> {
constructor(private inputs: T[]) {}
/** Asynchronous [[mapMutate]]. */
async map(getNewValue: (element: T, index: number) => Promise<T>): Promise<void> {
const { inputs } = this
for (let index = 0; index < inputs.length; index++)
inputs[index] = await getNewValue(inputs[index], index)
}
/** Asynchronous [[filterMutate]]. */
async filter(predicate: (element: T) => Promise<boolean>): Promise<void> {
const { inputs } = this
let writeIndex = 0
for (let readIndex = 0; readIndex < inputs.length; readIndex++)
if (await predicate(inputs[readIndex])) {
inputs[writeIndex] = inputs[readIndex]
writeIndex++
}
inputs.length = writeIndex
}
/** Asynchronous [[mapDefinedMutate]]. Performs `tryGetOutput` one element at a time. */
async mapDefined(tryGetOutput: (input: T) => Promise<Option<T>>): Promise<void> {
const { inputs } = this
let writeIndex = 0
for (let readIndex = 0; readIndex < inputs.length; readIndex++) {
const output = await tryGetOutput(inputs[readIndex])
if (output !== undefined) {
inputs[writeIndex] = output
writeIndex++
}
}
inputs.length = writeIndex
}
/** Asynchronous [[mutate]]. */
async mutate(index: Nat, transform: (t: T) => Promise<T>): Promise<void> {
const { inputs } = this
checkIndex(inputs, index)
inputs[index] = await transform(inputs[index])
}
}
/** Provides parallel utilities for an array. */
export function parallelArray<T>(inputs: T[], maxNumberOfThreads?: number): ParallelArrayOps<T> {
return new ParallelArrayOps(inputs, maxNumberOfThreads)
}
/**
Wrapper class for utilities that mutate arrays in parallel.
For non-mutating utilities use [[ParallelSeq]].
*/
export class ParallelArrayOps<T> {
/** Use [[parallelArray]] rather than calling this directly. */
constructor(readonly inputs: T[], readonly maxNumberOfThreads: number = Number.POSITIVE_INFINITY) |
/** Parallel [[mapMutate]]. */
async map(mapper: (element: T, index: number) => Promise<T>): Promise<void> {
const { inputs, maxNumberOfThreads } = this
let writeIndex = 0
let readIndex = 0
while (readIndex < maxNumberOfThreads && readIndex < inputs.length)
startOne()
while (readIndex < inputs.length) {
await awaitOne()
startOne()
}
while (writeIndex < inputs.length)
await awaitOne()
async function awaitOne(): Promise<void> {
inputs[writeIndex] = await (inputs as any as Array<Promise<T>>)[writeIndex]
writeIndex++
}
function startOne(): void {
(inputs as any as Array<Promise<T>>)[readIndex] = mapper(inputs[readIndex], readIndex)
readIndex++
}
}
/** Parallel [[filterMutate]]. */
filter(predicate: (element: T, index: number) => Promise<boolean>): Promise<void> {
return this.mapDefined(async (input, index) =>
optional(await predicate(input, index), () => input))
}
/** Parallel [[mapDefinedMutate]]. */
async mapDefined(tryGetOutput: (input: T, index: number) => Promise<Option<T>>): Promise<void> {
const { inputs, maxNumberOfThreads } = this
/** Next index to write a (defined) result to. */
let writeOutputIndex = 0
/** Next index to await a thread at. */
let readPromiseIndex = 0
/** Next index to read an input value from; the thread for that input will be written to the same index. */
let readValueIndex = 0
// Start initial threads.
while (readValueIndex < maxNumberOfThreads && readValueIndex < inputs.length)
startOne()
// Keep awaiting threads and starting new ones.
// Invariants: writeIndex <= readPromiseIndex, readPromiseIndex = readValueIndex - numberOfThreads
while (readValueIndex < inputs.length) {
await awaitOne()
startOne()
}
// Await remaining threads.
while (readPromiseIndex < inputs.length)
await awaitOne()
// Shorten array to new length.
inputs.length = writeOutputIndex
async function awaitOne(): Promise<void> {
const output = await (inputs as any as Array<Promise<Option<T>>>)[readPromiseIndex]
readPromiseIndex++
if (output !== undefined) {
inputs[writeOutputIndex] = output
writeOutputIndex++
}
}
function startOne(): void {
(inputs as any as Array<Promise<Option<T>>>)[readValueIndex] = tryGetOutput(inputs[readValueIndex], readValueIndex)
readValueIndex++
}
}
}
/**
Whether a number is an integer between 0 and array.length.
Does *not* check for whether there is a "hole" at the index.
*/
export function isValidIndex(inputs: Array<{}>, index: Nat): boolean {
return isNat(index) && index < inputs.length
}
/** Throws an error if [[index]] is not a valid index. */
export function checkIndex(inputs: Array<{}>, index: Nat): void {
if (!isValidIndex(inputs, index))
throw new Error(`Expected an array index < ${inputs.length}, got ${index}`)
}
/** Swap two values in an array. */
export function swap(inputs: Array<{}>, firstIndex: Nat, secondIndex: Nat): void {
checkIndex(inputs, firstIndex)
checkIndex(inputs, secondIndex)
const tmp = inputs[firstIndex]
inputs[firstIndex] = inputs[secondIndex]
inputs[secondIndex] = tmp
}
/** Initialize a new array by calling [[makeElement]] [[length]] times. */
export function initArray<T>(length: number, makeElement: (index: number) => T): T[] {
const arr = new Array(length)
for (let i = 0; i < length; i++)
arr[i] = makeElement(i)
return arr
}
/** Asynchronous [[initArray]]. */
export function initArrayAsync<T>(length: number | {
if (maxNumberOfThreads !== Number.POSITIVE_INFINITY)
checkNat(maxNumberOfThreads)
} | identifier_body |
main.rs | Error> {
fern::Dispatch::new()
.format(|out, message, record| unsafe {
out.finish(format_args!(
"{}[{}] {}",
chrono::Local::now().format("[%Y-%m-%d][%H:%M:%S]"),
libc::pthread_self(),
message
))
})
.level(log::LevelFilter::Info)
.chain(fern::log_file("output.log")?)
.apply()?;
Ok(())
}
/*==============================================================================
* Packet
*------------------------------------------------------------------------------
*
*/
struct Packet {
from: VirtualLink,
workload: u64,
origin: u32,
timestamp: SystemTime,
}
impl Packet {
fn from_iface(iface: &NetworkInterface, workload: u64, origin: u32) -> Packet {
Packet {
from: (*iface).get_virtual_link(),
workload: workload,
origin: origin,
timestamp: SystemTime::now(),
}
}
fn answer_me_at(tx: &mpsc::Sender<Packet>, workload: u64, origin: u32, timestamp: SystemTime) -> Packet {
Packet {
from: VirtualLink::linked_to(tx),
workload: workload,
origin: origin,
timestamp: timestamp,
}
}
}
/*==============================================================================
* VirtualLink
*------------------------------------------------------------------------------
*
*/
struct VirtualLink {
s: mpsc::Sender<Packet>
}
impl VirtualLink {
fn to_iface(interface: &NetworkInterface) -> VirtualLink {
VirtualLink {
s: (*interface).s.clone()
}
}
fn linked_to(tx: &mpsc::Sender<Packet>) -> VirtualLink {
VirtualLink {
s: (*tx).clone()
}
}
fn send_through(&self, packet: Packet) {
self.s.send(packet).unwrap()
}
}
/*==============================================================================
* Network Interface
*------------------------------------------------------------------------------
*
*/
struct NetworkInterface {
s: mpsc::Sender<Packet>,
r: mpsc::Receiver<Packet>
}
impl NetworkInterface {
fn new() -> NetworkInterface {
let (tx, rx) = mpsc::channel();
NetworkInterface {
s: tx,
r: rx
}
}
fn read(&self) -> Packet |
fn get_virtual_link(&self) -> VirtualLink {
VirtualLink::to_iface(self)
}
}
/*==============================================================================
* Host
*
*/
struct Host {
nic: NetworkInterface,
}
impl Host {
fn new() -> Host {
Host {
nic: NetworkInterface::new(),
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.nic.get_virtual_link()
}
}
/*==============================================================================
* Stats
*
*/
struct Stats {
samples: u64,
total: u64,
}
impl Stats {
fn new() -> Stats {
Stats {
samples: 0,
total: 0,
}
}
fn update_stats(&mut self, new_sample_time: u64) {
self.samples += 1;
self.total += new_sample_time;
}
fn get_average(&self) -> f64 {
if self.samples == 0 {
return 0.0;
}
(self.total as f64) / (self.samples as f64)
}
}
/*==============================================================================
* Server
*
*/
struct Server {
id: u32,
host: Host,
processing_power: u64
}
impl Server {
fn new(id: u32, server_data: ServerData) -> Server {
Server {
id: id,
host: Host::new(),
processing_power: server_data.get_processing_power()
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.host.get_virtual_link()
}
fn run(self) {
info!("[S{}] Ejecutando servidor {}", self.id, self.id);
let rx = self.host.nic.r;
let tx = self.host.nic.s;
for message in rx {
// Obtenemos la cantidad de cuadrantes a procesar.
let workload = message.workload;
info!("[S{}] Recibidas {} unidades de trabajo desde observatorio {}", self.id, workload, message.origin);
/*
* Procesamos los cuadrantes.
*
* El workload tiene unidades de trabajo. El poder de procesamiento
* tiene unidades de trabajo por segundo. El sleep time tiene unidades
* de milisegundos.
*
* Por ejemplo, un servidor recibe 5 unidades de trabajo desde el
* cliente. El servidor puede procesar dos unidades de trabajo por
* segundo. El hilo dormirá entonces 2500 milisegundos simulando
* el procesamiento de la carga. Para acelerar o relentizar
* la simulación, podemos ajustar el factor global de velocidad;
* por ejemplo, si el factor global es 2.0, en vez de dormir los 2500
* milisegundos dormiría 1250.
*
*/
let sleep_time = (1000*workload)/self.processing_power;
let sleep_time_scaled = ((sleep_time as f64)/GLOBAL_SPEED) as u64;
info!("[S{}] Tiempo estimado: {}ms (s: {}ms)", self.id, sleep_time, sleep_time_scaled);
thread::sleep(time::Duration::from_millis(sleep_time_scaled));
info!("[S{}] Procesamiento terminado; devolviendo ACK a observatorio {}", self.id, message.origin);
// Devolvemos el ACK.
let response = Packet::answer_me_at(&tx, 0, self.id, message.timestamp);
message.from.send_through(response);
}
}
}
/*==============================================================================
* Client
*
*/
struct Target {
virtual_link: VirtualLink,
weight: f64
}
struct Client {
id: u32,
host: Host,
distribution_scheme: Vec<Target>,
work_generation_rate: u64
}
impl Client {
fn new(id: u32, servers: &Vec<Server>, client_data: ClientData) -> Client {
let workshare: &Vec<f64> = client_data.get_workshare();
let mut distribution = Vec::new();
for i in 0..servers.len() {
distribution.push(Target {
virtual_link: servers[i].get_virtual_link(),
weight: workshare[i]
});
}
Client {
id: id,
host: Host::new(),
distribution_scheme: distribution,
work_generation_rate: client_data.get_work_generation_rate()
}
}
fn run(self) {
info!("[C{}] Ejecutando cliente {}", self.id, self.id);
/*
* Cada cierta cantidad de tiempo, el observatorio genera x cuadrantes.
* A partir de ahí itera por la lista de servidores distribuyendo los
* cuadrantes según los factores de distribución (e.g., si debe enviar
* una fracción p_k de los cuadrantes al servidor k, enviará p_k*x
* cuadrantes al servidor k).
*
* Habiendo enviado los mensajes, simplemente espera las respuestas.
* Suponiendo alternativamente que hay que seguir generando cuadrantes
* mientras se toman fotos, se pueden tener internamente dos threads,
* uno acumulando cuadrantes y otro tomando cuadrantes y distribuyendolos.
*
* Para medir el tiempo de respuesta del observatorio se puede ir
* calculando una media móvil, tomando el tiempo que tarda en responder
* cada servidor.
*/
let targets = &self.distribution_scheme;
let mut stats : Stats = Stats::new();
loop {
let x = self.work_generation_rate;
info!("[C{}] Generando {} unidades de trabajo", self.id, x);
// Distribuimos los x cuadrantes generados.
let mut sid = 0;
for target in targets {
sid += 1;
let workload = ((x as f64)*(target.weight)) as u64;
let packet = Packet::from_iface(&self.host.nic, workload, self.id);
info!("[C{}] Enviando {} unidades al servidor {}", self.id, workload, sid);
target.virtual_link.send_through(packet);
}
// Esperamos la respuesta de cada servidor.
info!("[C{}] Esperando respuestas", self.id);
for _d in targets {
let _response = self.host.nic.read();
// Cálculo de tiempo de respuesta
let response_time_duration = _response.timestamp.elapsed().unwrap();
let response_time_ms = response_time_duration.as_secs() + ((response_time_duration.subsec_millis() * 1000) as u64);
stats.update_stats(response_time_ms);
}
// Impresión de estadística hasta el momento
info!("[C{}] Promedio de respuesta parcial: {} ms", self.id, format!("{:.*}", 2, stats.get_average()));
info!("[C{}] Todos los servidores terminaron de procesar el bache", self.id);
let sleep_time = (3000.0/GLOBAL_SPEED) as u64;
thread::sleep(time::Duration::from_millis(sleep_time));
}
}
}
/*==============================================================================
| {
self.r.recv().unwrap()
} | identifier_body |
main.rs | Error> {
fern::Dispatch::new()
.format(|out, message, record| unsafe {
out.finish(format_args!(
"{}[{}] {}",
chrono::Local::now().format("[%Y-%m-%d][%H:%M:%S]"),
libc::pthread_self(),
message
))
})
.level(log::LevelFilter::Info)
.chain(fern::log_file("output.log")?)
.apply()?;
Ok(())
}
/*==============================================================================
* Packet
*------------------------------------------------------------------------------
*
*/
struct Packet {
from: VirtualLink,
workload: u64,
origin: u32,
timestamp: SystemTime,
}
impl Packet {
fn from_iface(iface: &NetworkInterface, workload: u64, origin: u32) -> Packet {
Packet {
from: (*iface).get_virtual_link(),
workload: workload,
origin: origin,
timestamp: SystemTime::now(),
}
}
fn answer_me_at(tx: &mpsc::Sender<Packet>, workload: u64, origin: u32, timestamp: SystemTime) -> Packet {
Packet {
from: VirtualLink::linked_to(tx),
workload: workload,
origin: origin,
timestamp: timestamp,
}
}
}
/*==============================================================================
* VirtualLink
*------------------------------------------------------------------------------
*
*/
struct VirtualLink {
s: mpsc::Sender<Packet>
}
impl VirtualLink {
fn to_iface(interface: &NetworkInterface) -> VirtualLink {
VirtualLink {
s: (*interface).s.clone()
}
}
fn linked_to(tx: &mpsc::Sender<Packet>) -> VirtualLink {
VirtualLink {
s: (*tx).clone()
}
}
fn send_through(&self, packet: Packet) {
self.s.send(packet).unwrap()
}
}
/*==============================================================================
* Network Interface
*------------------------------------------------------------------------------
*
*/
struct NetworkInterface {
s: mpsc::Sender<Packet>,
r: mpsc::Receiver<Packet>
}
impl NetworkInterface {
fn new() -> NetworkInterface {
let (tx, rx) = mpsc::channel();
NetworkInterface {
s: tx,
r: rx
}
}
fn read(&self) -> Packet {
self.r.recv().unwrap()
}
fn get_virtual_link(&self) -> VirtualLink {
VirtualLink::to_iface(self)
}
}
/*==============================================================================
* Host
*
*/
struct Host {
nic: NetworkInterface,
}
impl Host {
fn new() -> Host {
Host {
nic: NetworkInterface::new(),
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.nic.get_virtual_link()
}
}
/*==============================================================================
* Stats
*
*/
struct Stats {
samples: u64,
total: u64,
}
impl Stats {
fn new() -> Stats {
Stats {
samples: 0,
total: 0,
}
}
fn update_stats(&mut self, new_sample_time: u64) {
self.samples += 1;
self.total += new_sample_time;
}
fn get_average(&self) -> f64 {
if self.samples == 0 |
(self.total as f64) / (self.samples as f64)
}
}
/*==============================================================================
* Server
*
*/
struct Server {
id: u32,
host: Host,
processing_power: u64
}
impl Server {
fn new(id: u32, server_data: ServerData) -> Server {
Server {
id: id,
host: Host::new(),
processing_power: server_data.get_processing_power()
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.host.get_virtual_link()
}
fn run(self) {
info!("[S{}] Ejecutando servidor {}", self.id, self.id);
let rx = self.host.nic.r;
let tx = self.host.nic.s;
for message in rx {
// Obtenemos la cantidad de cuadrantes a procesar.
let workload = message.workload;
info!("[S{}] Recibidas {} unidades de trabajo desde observatorio {}", self.id, workload, message.origin);
/*
* Procesamos los cuadrantes.
*
* El workload tiene unidades de trabajo. El poder de procesamiento
* tiene unidades de trabajo por segundo. El sleep time tiene unidades
* de milisegundos.
*
* Por ejemplo, un servidor recibe 5 unidades de trabajo desde el
* cliente. El servidor puede procesar dos unidades de trabajo por
* segundo. El hilo dormirá entonces 2500 milisegundos simulando
* el procesamiento de la carga. Para acelerar o relentizar
* la simulación, podemos ajustar el factor global de velocidad;
* por ejemplo, si el factor global es 2.0, en vez de dormir los 2500
* milisegundos dormiría 1250.
*
*/
let sleep_time = (1000*workload)/self.processing_power;
let sleep_time_scaled = ((sleep_time as f64)/GLOBAL_SPEED) as u64;
info!("[S{}] Tiempo estimado: {}ms (s: {}ms)", self.id, sleep_time, sleep_time_scaled);
thread::sleep(time::Duration::from_millis(sleep_time_scaled));
info!("[S{}] Procesamiento terminado; devolviendo ACK a observatorio {}", self.id, message.origin);
// Devolvemos el ACK.
let response = Packet::answer_me_at(&tx, 0, self.id, message.timestamp);
message.from.send_through(response);
}
}
}
/*==============================================================================
* Client
*
*/
struct Target {
virtual_link: VirtualLink,
weight: f64
}
struct Client {
id: u32,
host: Host,
distribution_scheme: Vec<Target>,
work_generation_rate: u64
}
impl Client {
fn new(id: u32, servers: &Vec<Server>, client_data: ClientData) -> Client {
let workshare: &Vec<f64> = client_data.get_workshare();
let mut distribution = Vec::new();
for i in 0..servers.len() {
distribution.push(Target {
virtual_link: servers[i].get_virtual_link(),
weight: workshare[i]
});
}
Client {
id: id,
host: Host::new(),
distribution_scheme: distribution,
work_generation_rate: client_data.get_work_generation_rate()
}
}
fn run(self) {
info!("[C{}] Ejecutando cliente {}", self.id, self.id);
/*
* Cada cierta cantidad de tiempo, el observatorio genera x cuadrantes.
* A partir de ahí itera por la lista de servidores distribuyendo los
* cuadrantes según los factores de distribución (e.g., si debe enviar
* una fracción p_k de los cuadrantes al servidor k, enviará p_k*x
* cuadrantes al servidor k).
*
* Habiendo enviado los mensajes, simplemente espera las respuestas.
* Suponiendo alternativamente que hay que seguir generando cuadrantes
* mientras se toman fotos, se pueden tener internamente dos threads,
* uno acumulando cuadrantes y otro tomando cuadrantes y distribuyendolos.
*
* Para medir el tiempo de respuesta del observatorio se puede ir
* calculando una media móvil, tomando el tiempo que tarda en responder
* cada servidor.
*/
let targets = &self.distribution_scheme;
let mut stats : Stats = Stats::new();
loop {
let x = self.work_generation_rate;
info!("[C{}] Generando {} unidades de trabajo", self.id, x);
// Distribuimos los x cuadrantes generados.
let mut sid = 0;
for target in targets {
sid += 1;
let workload = ((x as f64)*(target.weight)) as u64;
let packet = Packet::from_iface(&self.host.nic, workload, self.id);
info!("[C{}] Enviando {} unidades al servidor {}", self.id, workload, sid);
target.virtual_link.send_through(packet);
}
// Esperamos la respuesta de cada servidor.
info!("[C{}] Esperando respuestas", self.id);
for _d in targets {
let _response = self.host.nic.read();
// Cálculo de tiempo de respuesta
let response_time_duration = _response.timestamp.elapsed().unwrap();
let response_time_ms = response_time_duration.as_secs() + ((response_time_duration.subsec_millis() * 1000) as u64);
stats.update_stats(response_time_ms);
}
// Impresión de estadística hasta el momento
info!("[C{}] Promedio de respuesta parcial: {} ms", self.id, format!("{:.*}", 2, stats.get_average()));
info!("[C{}] Todos los servidores terminaron de procesar el bache", self.id);
let sleep_time = (3000.0/GLOBAL_SPEED) as u64;
thread::sleep(time::Duration::from_millis(sleep_time));
}
}
}
/*==============================================================================
| {
return 0.0;
} | conditional_block |
main.rs | from: (*iface).get_virtual_link(),
workload: workload,
origin: origin,
timestamp: SystemTime::now(),
}
}
fn answer_me_at(tx: &mpsc::Sender<Packet>, workload: u64, origin: u32, timestamp: SystemTime) -> Packet {
Packet {
from: VirtualLink::linked_to(tx),
workload: workload,
origin: origin,
timestamp: timestamp,
}
}
}
/*==============================================================================
* VirtualLink
*------------------------------------------------------------------------------
*
*/
struct VirtualLink {
s: mpsc::Sender<Packet>
}
impl VirtualLink {
fn to_iface(interface: &NetworkInterface) -> VirtualLink {
VirtualLink {
s: (*interface).s.clone()
}
}
fn linked_to(tx: &mpsc::Sender<Packet>) -> VirtualLink {
VirtualLink {
s: (*tx).clone()
}
}
fn send_through(&self, packet: Packet) {
self.s.send(packet).unwrap()
}
}
/*==============================================================================
* Network Interface
*------------------------------------------------------------------------------
*
*/
struct NetworkInterface {
s: mpsc::Sender<Packet>,
r: mpsc::Receiver<Packet>
}
impl NetworkInterface {
fn new() -> NetworkInterface {
let (tx, rx) = mpsc::channel();
NetworkInterface {
s: tx,
r: rx
}
}
fn read(&self) -> Packet {
self.r.recv().unwrap()
}
fn get_virtual_link(&self) -> VirtualLink {
VirtualLink::to_iface(self)
}
}
/*==============================================================================
* Host
*
*/
struct Host {
nic: NetworkInterface,
}
impl Host {
fn new() -> Host {
Host {
nic: NetworkInterface::new(),
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.nic.get_virtual_link()
}
}
/*==============================================================================
* Stats
*
*/
struct Stats {
samples: u64,
total: u64,
}
impl Stats {
fn new() -> Stats {
Stats {
samples: 0,
total: 0,
}
}
fn update_stats(&mut self, new_sample_time: u64) {
self.samples += 1;
self.total += new_sample_time;
}
fn get_average(&self) -> f64 {
if self.samples == 0 {
return 0.0;
}
(self.total as f64) / (self.samples as f64)
}
}
/*==============================================================================
* Server
*
*/
struct Server {
id: u32,
host: Host,
processing_power: u64
}
impl Server {
fn new(id: u32, server_data: ServerData) -> Server {
Server {
id: id,
host: Host::new(),
processing_power: server_data.get_processing_power()
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.host.get_virtual_link()
}
fn run(self) {
info!("[S{}] Ejecutando servidor {}", self.id, self.id);
let rx = self.host.nic.r;
let tx = self.host.nic.s;
for message in rx {
// Obtenemos la cantidad de cuadrantes a procesar.
let workload = message.workload;
info!("[S{}] Recibidas {} unidades de trabajo desde observatorio {}", self.id, workload, message.origin);
/*
* Procesamos los cuadrantes.
*
* El workload tiene unidades de trabajo. El poder de procesamiento
* tiene unidades de trabajo por segundo. El sleep time tiene unidades
* de milisegundos.
*
* Por ejemplo, un servidor recibe 5 unidades de trabajo desde el
* cliente. El servidor puede procesar dos unidades de trabajo por
* segundo. El hilo dormirá entonces 2500 milisegundos simulando
* el procesamiento de la carga. Para acelerar o relentizar
* la simulación, podemos ajustar el factor global de velocidad;
* por ejemplo, si el factor global es 2.0, en vez de dormir los 2500
* milisegundos dormiría 1250.
*
*/
let sleep_time = (1000*workload)/self.processing_power;
let sleep_time_scaled = ((sleep_time as f64)/GLOBAL_SPEED) as u64;
info!("[S{}] Tiempo estimado: {}ms (s: {}ms)", self.id, sleep_time, sleep_time_scaled);
thread::sleep(time::Duration::from_millis(sleep_time_scaled));
info!("[S{}] Procesamiento terminado; devolviendo ACK a observatorio {}", self.id, message.origin);
// Devolvemos el ACK.
let response = Packet::answer_me_at(&tx, 0, self.id, message.timestamp);
message.from.send_through(response);
}
}
}
/*==============================================================================
* Client
*
*/
struct Target {
virtual_link: VirtualLink,
weight: f64
}
struct Client {
id: u32,
host: Host,
distribution_scheme: Vec<Target>,
work_generation_rate: u64
}
impl Client {
fn new(id: u32, servers: &Vec<Server>, client_data: ClientData) -> Client {
let workshare: &Vec<f64> = client_data.get_workshare();
let mut distribution = Vec::new();
for i in 0..servers.len() {
distribution.push(Target {
virtual_link: servers[i].get_virtual_link(),
weight: workshare[i]
});
}
Client {
id: id,
host: Host::new(),
distribution_scheme: distribution,
work_generation_rate: client_data.get_work_generation_rate()
}
}
fn run(self) {
info!("[C{}] Ejecutando cliente {}", self.id, self.id);
/*
* Cada cierta cantidad de tiempo, el observatorio genera x cuadrantes.
* A partir de ahí itera por la lista de servidores distribuyendo los
* cuadrantes según los factores de distribución (e.g., si debe enviar
* una fracción p_k de los cuadrantes al servidor k, enviará p_k*x
* cuadrantes al servidor k).
*
* Habiendo enviado los mensajes, simplemente espera las respuestas.
* Suponiendo alternativamente que hay que seguir generando cuadrantes
* mientras se toman fotos, se pueden tener internamente dos threads,
* uno acumulando cuadrantes y otro tomando cuadrantes y distribuyendolos.
*
* Para medir el tiempo de respuesta del observatorio se puede ir
* calculando una media móvil, tomando el tiempo que tarda en responder
* cada servidor.
*/
let targets = &self.distribution_scheme;
let mut stats : Stats = Stats::new();
loop {
let x = self.work_generation_rate;
info!("[C{}] Generando {} unidades de trabajo", self.id, x);
// Distribuimos los x cuadrantes generados.
let mut sid = 0;
for target in targets {
sid += 1;
let workload = ((x as f64)*(target.weight)) as u64;
let packet = Packet::from_iface(&self.host.nic, workload, self.id);
info!("[C{}] Enviando {} unidades al servidor {}", self.id, workload, sid);
target.virtual_link.send_through(packet);
}
// Esperamos la respuesta de cada servidor.
info!("[C{}] Esperando respuestas", self.id);
for _d in targets {
let _response = self.host.nic.read();
// Cálculo de tiempo de respuesta
let response_time_duration = _response.timestamp.elapsed().unwrap();
let response_time_ms = response_time_duration.as_secs() + ((response_time_duration.subsec_millis() * 1000) as u64);
stats.update_stats(response_time_ms);
}
// Impresión de estadística hasta el momento
info!("[C{}] Promedio de respuesta parcial: {} ms", self.id, format!("{:.*}", 2, stats.get_average()));
info!("[C{}] Todos los servidores terminaron de procesar el bache", self.id);
let sleep_time = (3000.0/GLOBAL_SPEED) as u64;
thread::sleep(time::Duration::from_millis(sleep_time));
}
}
}
/*==============================================================================
* Main
*
*/
const GLOBAL_SPEED: f64 = 1.0;
fn main() {
let args : Vec<String> = env::args().collect();
if args.len() > 1 && args[1] == "--debug" {
setup_file_logging().expect("Couldn't set up logger");
} else {
setup_terminal_logging().expect("Couldn't set up logger");
}
/*
* Cargamos la configuración. La configuración es un archivo de texto con
* pares clave-valor. El objeto de configuración puede usarse como
*
* configuration.get("clave") // retorna el valor asociado a "clave".
*/
info!("[T0] Cargando configuración");
let mut configuration = configuration::Configuration::new();
configuration.load(); | random_line_split | ||
main.rs | Error> {
fern::Dispatch::new()
.format(|out, message, record| unsafe {
out.finish(format_args!(
"{}[{}] {}",
chrono::Local::now().format("[%Y-%m-%d][%H:%M:%S]"),
libc::pthread_self(),
message
))
})
.level(log::LevelFilter::Info)
.chain(fern::log_file("output.log")?)
.apply()?;
Ok(())
}
/*==============================================================================
* Packet
*------------------------------------------------------------------------------
*
*/
struct Packet {
from: VirtualLink,
workload: u64,
origin: u32,
timestamp: SystemTime,
}
impl Packet {
fn from_iface(iface: &NetworkInterface, workload: u64, origin: u32) -> Packet {
Packet {
from: (*iface).get_virtual_link(),
workload: workload,
origin: origin,
timestamp: SystemTime::now(),
}
}
fn answer_me_at(tx: &mpsc::Sender<Packet>, workload: u64, origin: u32, timestamp: SystemTime) -> Packet {
Packet {
from: VirtualLink::linked_to(tx),
workload: workload,
origin: origin,
timestamp: timestamp,
}
}
}
/*==============================================================================
* VirtualLink
*------------------------------------------------------------------------------
*
*/
struct VirtualLink {
s: mpsc::Sender<Packet>
}
impl VirtualLink {
fn to_iface(interface: &NetworkInterface) -> VirtualLink {
VirtualLink {
s: (*interface).s.clone()
}
}
fn linked_to(tx: &mpsc::Sender<Packet>) -> VirtualLink {
VirtualLink {
s: (*tx).clone()
}
}
fn send_through(&self, packet: Packet) {
self.s.send(packet).unwrap()
}
}
/*==============================================================================
* Network Interface
*------------------------------------------------------------------------------
*
*/
struct NetworkInterface {
s: mpsc::Sender<Packet>,
r: mpsc::Receiver<Packet>
}
impl NetworkInterface {
fn | () -> NetworkInterface {
let (tx, rx) = mpsc::channel();
NetworkInterface {
s: tx,
r: rx
}
}
fn read(&self) -> Packet {
self.r.recv().unwrap()
}
fn get_virtual_link(&self) -> VirtualLink {
VirtualLink::to_iface(self)
}
}
/*==============================================================================
* Host
*
*/
struct Host {
nic: NetworkInterface,
}
impl Host {
fn new() -> Host {
Host {
nic: NetworkInterface::new(),
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.nic.get_virtual_link()
}
}
/*==============================================================================
* Stats
*
*/
struct Stats {
samples: u64,
total: u64,
}
impl Stats {
fn new() -> Stats {
Stats {
samples: 0,
total: 0,
}
}
fn update_stats(&mut self, new_sample_time: u64) {
self.samples += 1;
self.total += new_sample_time;
}
fn get_average(&self) -> f64 {
if self.samples == 0 {
return 0.0;
}
(self.total as f64) / (self.samples as f64)
}
}
/*==============================================================================
* Server
*
*/
struct Server {
id: u32,
host: Host,
processing_power: u64
}
impl Server {
fn new(id: u32, server_data: ServerData) -> Server {
Server {
id: id,
host: Host::new(),
processing_power: server_data.get_processing_power()
}
}
fn get_virtual_link(&self) -> VirtualLink {
self.host.get_virtual_link()
}
fn run(self) {
info!("[S{}] Ejecutando servidor {}", self.id, self.id);
let rx = self.host.nic.r;
let tx = self.host.nic.s;
for message in rx {
// Obtenemos la cantidad de cuadrantes a procesar.
let workload = message.workload;
info!("[S{}] Recibidas {} unidades de trabajo desde observatorio {}", self.id, workload, message.origin);
/*
* Procesamos los cuadrantes.
*
* El workload tiene unidades de trabajo. El poder de procesamiento
* tiene unidades de trabajo por segundo. El sleep time tiene unidades
* de milisegundos.
*
* Por ejemplo, un servidor recibe 5 unidades de trabajo desde el
* cliente. El servidor puede procesar dos unidades de trabajo por
* segundo. El hilo dormirá entonces 2500 milisegundos simulando
* el procesamiento de la carga. Para acelerar o relentizar
* la simulación, podemos ajustar el factor global de velocidad;
* por ejemplo, si el factor global es 2.0, en vez de dormir los 2500
* milisegundos dormiría 1250.
*
*/
let sleep_time = (1000*workload)/self.processing_power;
let sleep_time_scaled = ((sleep_time as f64)/GLOBAL_SPEED) as u64;
info!("[S{}] Tiempo estimado: {}ms (s: {}ms)", self.id, sleep_time, sleep_time_scaled);
thread::sleep(time::Duration::from_millis(sleep_time_scaled));
info!("[S{}] Procesamiento terminado; devolviendo ACK a observatorio {}", self.id, message.origin);
// Devolvemos el ACK.
let response = Packet::answer_me_at(&tx, 0, self.id, message.timestamp);
message.from.send_through(response);
}
}
}
/*==============================================================================
* Client
*
*/
struct Target {
virtual_link: VirtualLink,
weight: f64
}
struct Client {
id: u32,
host: Host,
distribution_scheme: Vec<Target>,
work_generation_rate: u64
}
impl Client {
fn new(id: u32, servers: &Vec<Server>, client_data: ClientData) -> Client {
let workshare: &Vec<f64> = client_data.get_workshare();
let mut distribution = Vec::new();
for i in 0..servers.len() {
distribution.push(Target {
virtual_link: servers[i].get_virtual_link(),
weight: workshare[i]
});
}
Client {
id: id,
host: Host::new(),
distribution_scheme: distribution,
work_generation_rate: client_data.get_work_generation_rate()
}
}
fn run(self) {
info!("[C{}] Ejecutando cliente {}", self.id, self.id);
/*
* Cada cierta cantidad de tiempo, el observatorio genera x cuadrantes.
* A partir de ahí itera por la lista de servidores distribuyendo los
* cuadrantes según los factores de distribución (e.g., si debe enviar
* una fracción p_k de los cuadrantes al servidor k, enviará p_k*x
* cuadrantes al servidor k).
*
* Habiendo enviado los mensajes, simplemente espera las respuestas.
* Suponiendo alternativamente que hay que seguir generando cuadrantes
* mientras se toman fotos, se pueden tener internamente dos threads,
* uno acumulando cuadrantes y otro tomando cuadrantes y distribuyendolos.
*
* Para medir el tiempo de respuesta del observatorio se puede ir
* calculando una media móvil, tomando el tiempo que tarda en responder
* cada servidor.
*/
let targets = &self.distribution_scheme;
let mut stats : Stats = Stats::new();
loop {
let x = self.work_generation_rate;
info!("[C{}] Generando {} unidades de trabajo", self.id, x);
// Distribuimos los x cuadrantes generados.
let mut sid = 0;
for target in targets {
sid += 1;
let workload = ((x as f64)*(target.weight)) as u64;
let packet = Packet::from_iface(&self.host.nic, workload, self.id);
info!("[C{}] Enviando {} unidades al servidor {}", self.id, workload, sid);
target.virtual_link.send_through(packet);
}
// Esperamos la respuesta de cada servidor.
info!("[C{}] Esperando respuestas", self.id);
for _d in targets {
let _response = self.host.nic.read();
// Cálculo de tiempo de respuesta
let response_time_duration = _response.timestamp.elapsed().unwrap();
let response_time_ms = response_time_duration.as_secs() + ((response_time_duration.subsec_millis() * 1000) as u64);
stats.update_stats(response_time_ms);
}
// Impresión de estadística hasta el momento
info!("[C{}] Promedio de respuesta parcial: {} ms", self.id, format!("{:.*}", 2, stats.get_average()));
info!("[C{}] Todos los servidores terminaron de procesar el bache", self.id);
let sleep_time = (3000.0/GLOBAL_SPEED) as u64;
thread::sleep(time::Duration::from_millis(sleep_time));
}
}
}
/*==============================================================================
* | new | identifier_name |
get_company_test.py | ;q=0.9",
# "Connection": "keep-alive",
# "Content-Length": "18",
"Content-Type": "application/x-www-form-urlencoded; charset=UTF-8",
"Cookie": "__root_domain_v=.52wmb.com; _qddaz=QD.oxnn0z.8hqf3r.kjru3sro; _QUY=wqXCh8KywpbCmMKVwplrwpzCk8KeZ8KbwpJnZnDCk2htaQ==; _DP=2; company_search_tips=1; _QP=1; promote=proname=auto; _qdda=3-1.1; _qddab=3-cx3zj9.klpw0b65; access_token=13609ab52b8b529a; 52BY_TOKEN=8ed9a124-7a31-11eb-a26e-00155d391e0d; _MNG=1; vip_expired_tips=none; _qddamta_2885855166=3-0",
"DNT": "1",
"Host": "www.52wmb.com",
"Origin": "https://www.52wmb.com",
"Referer": "https://www.52wmb.com/buyer/" + str(data1['company_id']),
"Sec-Fetch-Dest": "empty",
"Sec-Fetch-Mode": "cors",
"Sec-Fetch-Site": "same-origin",
"User-Agent": "Mozilla/5.0(Windows NT 10.0; Win64; x64)AppleWebKit/537.36(KHTML,like Gecko)Chrome/88.0.4324.182 Safari/537.36 Edg/88.0.705.81",
# "X-Requested-With": "XMLHttpRequest",
}
data = {
'company_id': data1['company_id']
}
try:
html = requests.post(url=url, data=data, headers=headers) # data=data,
time.sleep(random.randint(3, 5))
html.raise_for_status()
html.encoding = html.apparent_encoding
# print(html.url)
html = html.text
# print('8'*20,'\n',html)
parser_email(html, company_info)
except Exception as e:
return ''
def parser_email(html, company_info):
html = etree.HTML(html)
try:
tbod = html.xpath('//*[@id="contact-detail"]')
# print("tbod:ss",tbod)
# titlle = html.xpath('//*[@id="companies-detail"]/div/div/div/h1/text()')//*[@id="contact-detail"]/table/tbody
if len(tbod) >= 1:
pass
else:
automation()
except Exception as e:
pass
try:
html = html.xpath('//tbody')[0]
if len(html) >= 1:
print('有数据')
strt = int(time.time())
company_info['time'] = strt
company_info['tableid'] = 0
company_info['spider'] = 1
try:
# 数据来源
region = html.xpath('tr[4]/td[1]/text()')[1].strip()
except:
region = '-'
company_info['region'] = region
try:
# 联系电话
c_tel = html.xpath('tr[3]/td[1]/text()')[1].strip()
except:
c_tel = '-'
company_info['tel'] = c_tel
try:
# 联 系 人
lianxiren = html.xpath('tr[2]/td[1]/text()')[1].strip()
except:
lianxiren = '-'
company_info['lianxiren'] = lianxiren
try:
# 邮箱地址
c_email = html.xpath('tr[2]/td[2]/text()')[1].strip()
except:
c_email = '-'
company_info['company_email'] = c_email
try:
# 公司地址
c_adress = html.xpath('tr[3]/td[2]/text()')[1].strip()
except:
c_adress = '-'
company_info['company_address'] = c_adress
# try:
# # 数据来源
# c_source = html.xpath('tr[4]/td[1]/text()')[1].strip()
# except:
# c_source = '-'
# company_info['company_source'] = c_source
try:
# 官方网站
c_site = html.xpath('tr[4]/td[2]/a/text()')[0]
except:
c_site = '-'
company_info['company_site'] = c_site
except Exception as e:
print('没有数据' + e)
if len(company_info) > 10:
information_list.append(company_info)
print(json.dumps(company_info, ensure_ascii=False))
else:
print('数据不完整')
def Verification(companys_list):
"""
# 判断页面是否请求成功是否拿到信息
# :param companys_list:
# :return:
"""
all_dict = collections.OrderedDict()
if len(companys_list) > 0:
all_dict['code'] = 200
all_dict['msg'] = "sucess"
else:
all_dict['code'] = 404
all_dict['msg'] = "没有找到该信息,请更换关键词再试试"
test_dict = collections.OrderedDict()
test_dict['number'] = '1111111'
test_dict['name'] = 'name'
test_dict['update_time'] = time.strftime('%Y-%m-%d')
test_dict['other_trade'] = 'None'
# companys_list[-1] = test_dict
all_dict['data'] = companys_list
print(json.dumps(companys_list, ensure_ascii=False))
insert_company(companys_list)
# 插入数据库
if all_dict['code'] == 404:
Verification_code()
def insert_company(mysql_db):
"""
# 把信息插入数据库
# :param data:
# :return:
"""
strt = int(time.time())
try:
conn = pymysql.connect(host=mysql_db['host'],
user=mysql_db['user'],
password=mysql_db['password'],
database=mysql_db['user'],
charset='utf8')
cursor = conn.cursor()
print("数据库连接成功")
except Exception as e:
print(e)
return
if len(information_list) >= 0:
# print("准备插入数据库->",information_list)
# print("数据库信息->",mysql_db['host'],mysql_db['user'],mysql_db['password'],mysql_db['user'])
list1 = [tuple([i["update_time"]] +
[i['name']] +
[i["trade_number"]] +
[i["region"]] +
[i["company_email"]] +
[i["lianxiren"]] +
[i["tel"]] +
[i['company_address']] +
[i["company_site"]] +
[i["time"]] +
[i["tableid"]] +
[i["spider"]]
) for i in information_list]
for data in list1:
insert_sql = 'insert into wp_haiguan_data (day,company_name,number,country,email,lianxiren,phone,address,' \
'website,update_time,tableid,is_spider) values {};'.format(data)
try:
# print("插入数据库的sql语句",insert_sql)
cursor.execute(insert_sql)
conn.commit()
except Exception as e:
print(e)
cursor.close()
conn.close()
else:
print('没有数据')
def automation():
"""
解决验证码,验证码处出来截图并连接打码平台
:return:
"""
driver = webdriver.Chrome(options=chrome_options)
"""
处理selenium被检测这个问题
"""
driver.maximize_window()
driver.get('https://www.52wmb.com/buyer/35206685?SRID=acKVwpdnwpdrbA%3D%3D&key=mask&st=2')
with open('w_cookies.txt', 'r', encoding='utf8') as f:
listCookies = json.loads(f.read())
# 往br | owser里添加cookies
for cookie in listCookies:
cookie_dict = {
'name': cookie.get('name'),
'value': cookie.get('value'),
"expires": '',
'path': '/',
'httpOnly': False,
'HostOnly': False,
'Secure': False
}
# 添加cookie进行免登陆
driver.add_cookie(cookie_dict)
# 刷新当前页面
driver.refresh()
time.sleep(2)
# 切换iframe
driver.switch_to.frame('mainFrame')
# 定位验证码图片位置 | identifier_body | |
get_company_test.py | # item = item.rstrip("\n")
# with open('adjunct.json') as f:
# con_dict = json.loads(f.read())
# # cookie_path = con_dict['cookie_filepath']
# cookie_path = 'w_cookies.txt'
# mysql_db = con_dict[item]['datebase']
# data_hs = con_dict[item]['hs']
# try:
# data_hs = data_hs.split(',')
# except Exception as e:
# print('无需分割' + e)
class Chaojiying_Client(object):
def __init__(self, username, password, soft_id):
self.username = username
password = password.encode('utf8')
self.password = md5(password).hexdigest()
self.soft_id = soft_id
self.base_params = {
'user': self.username,
'pass2': self.password,
'softid': self.soft_id,
}
self.headers = {
'Connection': 'Keep-Alive',
'User-Agent': 'Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 5.1; Trident/4.0)',
}
def PostPic(self, im, codetype):
"""
im: 图片字节
codetype: 题目类型 参考 http://www.chaojiying.com/price.html
"""
params = {
'codetype': codetype,
}
params.update(self.base_params)
files = {'userfile': ('ccc.jpg', im)}
r = requests.post('http://upload.chaojiying.net/Upload/Processing.php', data=params, files=files,
headers=self.headers)
return r.json()
def ReportError(self, im_id):
"""
im_id:报错题目的图片ID
"""
params = {
'id': im_id,
}
params.update(self.base_params)
r = requests.post('http://upload.chaojiying.net/Upload/ReportError.php', data=params, headers=self.headers)
return r.json()
def readCookies(cookie_path):
"""
读取本地的cookie并进行遍历
:param cookie_path:
:return:
"""
with open(cookie_path, 'r') as f: # ,encoding='utf-8'
listCookies = json.loads(f.read())
cookie = [item["name"] + "=" + item["value"] for item in listCookies]
cookiestr = '; '.join(item for item in cookie)
return cookiestr
def getHTMLText(data, cookiestr):
"""
# requests请求获取到页面的html
# :param data:
# :param cookiestr:
# :return:
"""
pageNum = 1
while pageNum < 5:
url2 = 'https://www.52wmb.com/async/company?key={}&old_key={}&country=&country_id=0&type=0&sort' \
'=default&click_search=0&st=3&old_st=3&main=0&extend_search=false&fw%5B%5D=email&filterweight=email&_' \
'=1604476115171&page={}'.format(data['key'], data['key'], pageNum)
url1 = 'https://www.52wmb.com/async/company?key={}&old_key={' \
'}&country=&country_id=0&type=0&sort=default&click_search=0&st=2&old_st=2&main=0&extend_search=false' \
'&fw%5B%5D=email&filterweight=email&_=1603852119428&page={} '.format(data['key'], data['key'], pageNum)
url = 'https://www.52wmb.com/async/company?key={}&old_key={' \
'}&country=&country_id=0&type=0&sort=default&click_search=0&st=3&old_st=3&main=0&extend_search=false&fw' \
'%5B%5D=email&filterweight=email&_=1603434620022&page={}'.format(data['key'], data['key'], pageNum)
url3 = 'https://www.52wmb.com/async/company?country=*&key={}&type=0&click_search=1&fw%5B%5D=email&filterweight=email' \
'&sort=default&country_data=&search_type=3&is_label=1&st=3&page={}&_=1614564485810'.format(data['key'], pageNum)
headers = {
'cookie': cookiestr,
'user-agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36',
'referer': 'https://www.52wmb.com/buy-3926909090?st=3'
}
try:
# html = requests.get(url=url2, headers=headers) #miao zhu shi 2021/3/1
html = requests.get(url=url3, headers=headers) # miao xin zeng 2021/3/1
time.sleep(random.randint(4, 8))
html.raise_for_status()
html.encoding = html.apparent_encoding
# print(html.url)
html = html.text
information(html, cookiestr)
except Exception as e:
print(e)
pageNum += 1
def information(html, cookiestr):
"""
# 获取公司的信息
# :param html:
# :return:
"""
companys_list = etree.HTML(html).xpath('//li[@class="ssList-li company-fun-li"]')
for i in companys_list[:]:
company_info = collections.OrderedDict()
try:
# number 公司id
c_id = i.xpath('@data-id')[0]
company_info['number'] = c_id
# 公司名称
company_info['name'] = i.xpath('div[1]/div[@class="ssContent"]/a/text()')[0]
# 地区
# company_info['trade_country'] = i.xpath('div[1]/div[@class="ssContent"]/p[@class="ss-Ctag"]/text()')[0]
# 货运次数
company_info['trade_number'] = i.xpath('div[1]/div[@class="ssContent"]/p[@class="ss-Ctag"]/text()')[1].replace(
'总货运', '')
# 更新时间
update_time = i.xpath('div[1]/div[@class="ssContent"]/div[@class="ss-Cjl"]/text()')[0].strip()[-10:]
company_info['update_time'] = update_time
# print("company_info------",company_info)
get_next_level(company_info, cookiestr)
except Exception as e:
print(e)
def get_next_level(company_info, cookiestr):
url = 'https://www.52wmb.com/async/contact'
# url = 'https://www.52wmb.com/buyer/35274878?SRID=Z8KWwphnwpZsag%3D%3D&key=crane&st=2'
data1 = collections.OrderedDict()
data1['company_id'] = company_info['number']
# headers = {
# 'Referer': 'https://www.52wmb.com/buyer/{}'.format(data1['company_id']),
# 'cookie': cookiestr,
# 'user-agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/81.0.4044.122 Safari/537.36'
# } miao zhu shi 2021/3/1
headers = {
"Accept": "*/*",
# "Accept-Encoding": "gzip,deflate,br",
# "Accept-Language": "zh-CN,zh;q=0.9",
# "Connection": "keep-alive",
# "Content-Length": "18",
"Content-Type": "application/x-www-form-urlencoded; charset=UTF-8",
"Cookie": "__root_domain_v=.52wmb.com; _qddaz=QD.oxnn0z.8hqf3r.kjru3sro; _QUY=wqXCh8KywpbCmMKVwplrwpzCk8KeZ8KbwpJnZnDCk2htaQ==; _DP=2; company_search_tips=1; _QP=1; promote=proname=auto; _qdda=3-1.1; _qddab=3-cx3zj9.klpw0b65; access_token=13609ab52b8b529a; 52BY_TOKEN=8ed9a124-7a31-11eb-a26 | random_line_split | ||
get_company_test.py | _st=3&main=0&extend_search=false&fw%5B%5D=email&filterweight=email&_' \
'=1604476115171&page={}'.format(data['key'], data['key'], pageNum)
url1 = 'https://www.52wmb.com/async/company?key={}&old_key={' \
'}&country=&country_id=0&type=0&sort=default&click_search=0&st=2&old_st=2&main=0&extend_search=false' \
'&fw%5B%5D=email&filterweight=email&_=1603852119428&page={} '.format(data['key'], data['key'], pageNum)
url = 'https://www.52wmb.com/async/company?key={}&old_key={' \
'}&country=&country_id=0&type=0&sort=default&click_search=0&st=3&old_st=3&main=0&extend_search=false&fw' \
'%5B%5D=email&filterweight=email&_=1603434620022&page={}'.format(data['key'], data['key'], pageNum)
url3 = 'https://www.52wmb.com/async/company?country=*&key={}&type=0&click_search=1&fw%5B%5D=email&filterweight=email' \
'&sort=default&country_data=&search_type=3&is_label=1&st=3&page={}&_=1614564485810'.format(data['key'], pageNum)
headers = {
'cookie': cookiestr,
'user-agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36',
'referer': 'https://www.52wmb.com/buy-3926909090?st=3'
}
try:
# html = requests.get(url=url2, headers=headers) #miao zhu shi 2021/3/1
html = requests.get(url=url3, headers=headers) # miao xin zeng 2021/3/1
time.sleep(random.randint(4, 8))
html.raise_for_status()
html.encoding = html.apparent_encoding
# print(html.url)
html = html.text
information(html, cookiestr)
except Exception as e:
print(e)
pageNum += 1
def information(html, cookiestr):
"""
# 获取公司的信息
# :param html:
# :return:
"""
companys_list = etree.HTML(html).xpath('//li[@class="ssList-li company-fun-li"]')
for i in companys_list[:]:
company_info = collections.OrderedDict()
try:
# number 公司id
c_id = i.xpath('@data-id')[0]
company_info['number'] = c_id
# 公司名称
company_info['name'] = i.xpath('div[1]/div[@class="ssContent"]/a/text()')[0]
# 地区
# company_info['trade_country'] = i.xpath('div[1]/div[@class="ssContent"]/p[@class="ss-Ctag"]/text()')[0]
# 货运次数
company_info['trade_number'] = i.xpath('div[1]/div[@class="ssContent"]/p[@class="ss-Ctag"]/text()')[1].replace(
'总货运', '')
# 更新时间
update_time = i.xpath('div[1]/div[@class="ssContent"]/div[@class="ss-Cjl"]/text()')[0].strip()[-10:]
company_info['update_time'] = update_time
# print("company_info------",company_info)
get_next_level(company_info, cookiestr)
except Exception as e:
print(e)
def get_next_level(company_info, cookiestr):
url = 'https://www.52wmb.com/async/contact'
# url = 'https://www.52wmb.com/buyer/35274878?SRID=Z8KW | 3D&key=crane&st=2'
data1 = collections.OrderedDict()
data1['company_id'] = company_info['number']
# headers = {
# 'Referer': 'https://www.52wmb.com/buyer/{}'.format(data1['company_id']),
# 'cookie': cookiestr,
# 'user-agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/81.0.4044.122 Safari/537.36'
# } miao zhu shi 2021/3/1
headers = {
"Accept": "*/*",
# "Accept-Encoding": "gzip,deflate,br",
# "Accept-Language": "zh-CN,zh;q=0.9",
# "Connection": "keep-alive",
# "Content-Length": "18",
"Content-Type": "application/x-www-form-urlencoded; charset=UTF-8",
"Cookie": "__root_domain_v=.52wmb.com; _qddaz=QD.oxnn0z.8hqf3r.kjru3sro; _QUY=wqXCh8KywpbCmMKVwplrwpzCk8KeZ8KbwpJnZnDCk2htaQ==; _DP=2; company_search_tips=1; _QP=1; promote=proname=auto; _qdda=3-1.1; _qddab=3-cx3zj9.klpw0b65; access_token=13609ab52b8b529a; 52BY_TOKEN=8ed9a124-7a31-11eb-a26e-00155d391e0d; _MNG=1; vip_expired_tips=none; _qddamta_2885855166=3-0",
"DNT": "1",
"Host": "www.52wmb.com",
"Origin": "https://www.52wmb.com",
"Referer": "https://www.52wmb.com/buyer/" + str(data1['company_id']),
"Sec-Fetch-Dest": "empty",
"Sec-Fetch-Mode": "cors",
"Sec-Fetch-Site": "same-origin",
"User-Agent": "Mozilla/5.0(Windows NT 10.0; Win64; x64)AppleWebKit/537.36(KHTML,like Gecko)Chrome/88.0.4324.182 Safari/537.36 Edg/88.0.705.81",
# "X-Requested-With": "XMLHttpRequest",
}
data = {
'company_id': data1['company_id']
}
try:
html = requests.post(url=url, data=data, headers=headers) # data=data,
time.sleep(random.randint(3, 5))
html.raise_for_status()
html.encoding = html.apparent_encoding
# print(html.url)
html = html.text
# print('8'*20,'\n',html)
parser_email(html, company_info)
except Exception as e:
return ''
def parser_email(html, company_info):
html = etree.HTML(html)
try:
tbod = html.xpath('//*[@id="contact-detail"]')
# print("tbod:ss",tbod)
# titlle = html.xpath('//*[@id="companies-detail"]/div/div/div/h1/text()')//*[@id="contact-detail"]/table/tbody
if len(tbod) >= 1:
pass
else:
automation()
except Exception as e:
pass
try:
html = html.xpath('//tbody')[0]
if len(html) >= 1:
print('有数据')
strt = int(time.time())
company_info['time'] = strt
company_info['tableid'] = 0
company_info['spider'] = 1
try:
# 数据来源
region = html.xpath('tr[4]/td[1]/text()')[1].strip()
except:
region = '-'
company_info['region'] = region
try:
# 联系电话
c_tel = html.xpath('tr[3]/td[1]/text()')[1].strip()
except:
c_tel = '-'
company_info['tel'] = c_tel
try:
# 联 系 人
lianxiren = html.xpath('tr[2]/td[1]/text()')[ | wphnwpZsag%3D% | identifier_name |
get_company_test.py | _st=3&main=0&extend_search=false&fw%5B%5D=email&filterweight=email&_' \
'=1604476115171&page={}'.format(data['key'], data['key'], pageNum)
url1 = 'https://www.52wmb.com/async/company?key={}&old_key={' \
'}&country=&country_id=0&type=0&sort=default&click_search=0&st=2&old_st=2&main=0&extend_search=false' \
'&fw%5B%5D=email&filterweight=email&_=1603852119428&page={} '.format(data['key'], data['key'], pageNum)
url = 'https://www.52wmb.com/async/company?key={}&old_key={' \
'}&country=&country_id=0&type=0&sort=default&click_search=0&st=3&old_st=3&main=0&extend_search=false&fw' \
'%5B%5D=email&filterweight=email&_=1603434620022&page={}'.format(data['key'], data['key'], pageNum)
url3 = 'https://www.52wmb.com/async/company?country=*&key={}&type=0&click_search=1&fw%5B%5D=email&filterweight=email' \
'&sort=default&country_data=&search_type=3&is_label=1&st=3&page={}&_=1614564485810'.format(data['key'], pageNum)
headers = {
'cookie': cookiestr,
'user-agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36',
'referer': 'https://www.52wmb.com/buy-3926909090?st=3'
}
try:
# html = requests.get(url=url2, headers=headers) #miao zhu shi 2021/3/1
html = requests.get(url=url3, headers=headers) # miao xin zeng 2021/3/1
time.sleep(random.randint(4, 8))
html.raise_for_status()
html.encoding = html.apparent_encoding
# print(html.url)
html = html.text
information(html, cookiestr)
except Exception as e:
print(e)
pageNum += 1
def information(html, cookiestr):
"""
# 获取公司的信息
# :param html:
# :return:
"""
companys_list = etree.HTML(html).xpath('//li[@class="ssList-li company-fun-li"]')
for i in companys_list[:]:
company_info = collections.OrderedDict()
try:
# number 公司id
c_id = i.xpath('@data-id')[0]
company_info['number'] = c_id
# 公司名称
company_info['name'] = i.xpath('div[1]/div[@class="ssContent"]/a/text()')[0]
# 地区
# company_info['trade_country'] = i.xpath('div[1]/div[@class="ssContent"]/p[@class="ss-Ctag"]/text()')[0]
# 货运次数
company_info['trade_number'] = i.xpath('div[1]/div[@class="ssContent"]/p[@class="ss-Ctag"]/text()')[1].replace(
'总货运', '')
# 更新时间
update_time = i.xpath('div[1]/div[@class="ssContent"]/div[@class="ss-Cjl"]/text()')[0].strip()[-10:]
company_info['update_time'] = update_time
# print("company_info------",company_info)
get_next_level(company_info, cookiestr)
except Exception as e:
print(e)
def get_next_level(company_info, cookiestr):
url = 'https://www.52wmb.com/async/contact'
# url = 'https://www.52wmb.com/buyer/35274878?SRID=Z8KWwphnwpZsag%3D%3D&key=crane&st=2'
data1 = collections.OrderedDict()
data1['company_id'] = company_info['number']
# headers = {
# 'Referer': 'https://www.52wmb.com/buyer/{}'.format(data1['company_id']),
# 'cookie': cookiestr,
# 'user-agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/81.0.4044.122 Safari/537.36'
# } miao zhu shi 2021/3/1
headers = {
"Accept": "*/*",
# "Accept-Encoding": "gzip,deflate,br",
# "Accept-Language": "zh-CN,zh;q=0.9",
# "Connection": "keep-alive",
# "Content-Length": "18",
"Content-Type": "application/x-www-form-urlencoded; charset=UTF-8",
"Cookie": "__root_domain_v=.52wmb.com; _qddaz=QD.oxnn0z.8hqf3r.kjru3sro; _QUY=wqXCh8KywpbCmMKVwplrwpzCk8KeZ8KbwpJnZnDCk2htaQ==; _DP=2; company_search_tips=1; _QP=1; promote=proname=auto; _qdda=3-1.1; _qddab=3-cx3zj9.klpw0b65; access_token=13609ab52b8b529a; 52BY_TOKEN=8ed9a124-7a31-11eb-a26e-00155d391e0d; _MNG=1; vip_expired_tips=none; _qddamta_2885855166=3-0",
"DNT": "1",
"Host": "www.52wmb.com",
"Origin": "https://www.52wmb.com",
"Referer": "https://www.52wmb.com/buyer/" + str(data1['company_id']),
"Sec-Fetch-Dest": "empty",
"Sec-Fetch-Mode": "cors",
"Sec-Fetch-Site": "same-origin",
"User-Agent": "Mozilla/5.0(Windows NT 10.0; Win64; x64)AppleWebKit/537.36(KHTML,like Gecko)Chrome/88.0.4324.182 Safari/537.36 Edg/88.0.705.81",
# "X-Requested-With": "XMLHttpRequest",
}
data = {
'company_id': data1['company_id']
}
try:
html = requests.post(url=url, data=data, headers=headers) # data=data,
time.sleep(random.randint(3, 5))
html.raise_for_status()
html.encoding = html.apparent_encoding
# print(html.url)
html = html.text
# print('8'*20,'\n',html)
parser_email(html, company_info)
except Exception as e:
return ''
def parser_email(html, company_info):
html = etree.HTML(html)
try:
tbod = html.xpath('//*[@id="contact-detail"]')
# print("tbod:ss",tbod)
# titlle = html.xpath('//*[@id="companies-detail"]/div/div/div/h1/text()')//*[@id="contact-detail"]/table/tbody
if len(tbod) >= 1:
pass
else:
automation()
except Exception as e:
pass
try:
html = html.xpath('//tbody')[0]
if len( | ) >= 1:
print('有数据')
strt = int(time.time())
company_info['time'] = strt
company_info['tableid'] = 0
company_info['spider'] = 1
try:
# 数据来源
region = html.xpath('tr[4]/td[1]/text()')[1].strip()
except:
region = '-'
company_info['region'] = region
try:
# 联系电话
c_tel = html.xpath('tr[3]/td[1]/text()')[1].strip()
except:
c_tel = '-'
company_info['tel'] = c_tel
try:
# 联 系 人
lianxiren = html.xpath('tr[2]/td[1]/text() | html | conditional_block |
main.rs |
fn vectors() {
let v: Vec<i32> = Vec::new();
let mut v = vec![1, 2, 3];
match v.binary_search(&16) {
Ok(pos) => v.insert(pos, 16),
Err(_) => v.push(16)
}
match v.binary_search(&12) {
Ok(pos) => v.insert(pos, 12),
Err(pos) => v.insert(pos, 12)
}
println!("Binary Search -> {:?}", v);
let mut v = Vec::new();
v.push(5);
v.push(6);
v.push(7);
v.push(8);
let v = vec![1, 2, 3, 4, 5];
let third: &i32 = &v[2];
println!("The third element is {}", third);
match v.get(2) {
Some(third) => println!("The third element is {}", third),
None => println!("There is no third element."),
}
// When the program has a valid reference, the borrow checker enforces the ownership and
// borrowing rules (covered in Chapter 4) to ensure this reference and any other references to
// the contents of the vector remain valid. Recall the rule that states you can’t have mutable
// and immutable references in the same scope. That rule applies in Listing 8-7, where we hold
// an immutable reference to the first element in a vector and try to add an element to the end,
// which won’t work.
let mut v = vec![1, 2, 3, 4, 5];
let first = &v[0];
v.push(6);
//Below line causes Compilation Error
//println!("The first element is: {}", first);
// This error is due to the way vectors work: adding a new element onto the end of the vector
// might require allocating new memory and copying the old elements to the new space, if there
// isn’t enough room to put all the elements next to each other where the vector currently is.
// In that case, the reference to the first element would be pointing to deallocated memory.
// The borrowing rules prevent programs from ending up in that situation.
let v = vec![100, 32, 57];
for i in &v {
println!("{}", i);
}
// To change the value that the mutable reference refers to, we have to use the dereference
// operator (*) to get to the value in i before we can use the += operator.
let mut v = vec![100, 32, 57];
for i in &mut v {
*i += 50;
}
for i in &v {
println!("{}", i);
}
enum SpreadsheetCell {
Int(i32),
Float(f64),
Text(String),
}
let row = vec![
SpreadsheetCell::Int(3),
SpreadsheetCell::Text(String::from("blue")),
SpreadsheetCell::Float(10.12),
];
}
fn strings() {
let mut s = String::new();
let m = String::from("sdfsdf");
let data = "initial contents";
let s = data.to_string();
// the method also works on a literal directly:
let s = "initial contents".to_string();
let hello = String::from("السلام عليكم");
let hello = String::from("Dobrý den");
let hello = String::from("Hello");
let hello = String::from("שָׁלוֹם");
let hello = String::from("नमस्ते");
let hello = String::from("こんにちは");
let hello = String::from("안녕하세요");
let hello = String::from("你好");
let hello = String::from("Olá");
let hello = String::from("Здравствуйте");
let hello = String::from("Hola");
let mut s1 = String::from("foo");
let s2 = "bar";
s1.push_str(s2);
println!("s2 is {}", s2);
let mut s = String::from("lo");
s.push('l');
use std::ops::Add;
let s1 = String::from("Hello, ");
let s2 = String::from("world!");
// The reason we’re able to use &s2 in the call to add is that the compiler can coerce the
// &String argument into a &str. When we call the add method, Rust uses a deref coercion, which
// here turns &s2 into &s2[..]. We’ll discuss deref coercion in more depth in Chapter 15.
// Because add does not take ownership of the s parameter, s2 will still be a valid String after
// this operation.
// looks like it will copy both strings and create a new one, this statement actually takes
// ownership of s1, appends a copy of the contents of s2, and then returns ownership of the
// result. In other words, it looks like it’s making a lot of copies but isn’t; the
// implementation is more efficient than copying.
//let s3 = s1.add(&s2);
let s3 = s1 + &s2;
println!("{}", s3);
let s1 = String::from("tic");
let s2 = String::from("tac");
let s3 = String::from("toe");
//let s = s1 + "-" + &s2 + "-" + &s3;
let s = format!("{}-{}-{}", s1, s2, s3);
println!("{}", s);
// The version of the code using format! is much easier to read and doesn’t take ownership of
// any of its parameters.
println!("{}", s1);
// A String is a wrapper over a Vec<u8>
let len = String::from("Hola").len();
// In this case, len will be 4, which means the vector storing the string “Hola” is 4 bytes long.
// Each of these letters takes 1 byte when encoded in UTF-8
println!("{}", len);
let len = String::from("Здравствуйте").len();
println!("{}", len);
// It takes 24 bytes to encode “Здравствуйте” in UTF-8, because each Unicode scalar value in that string
// takes 2 bytes of storage. Therefore, an index into the string’s bytes will not always
// correlate to a valid Unicode scalar value. To demonstrate, consider this invalid Rust code:
// let hello = "Здравствуйте";
// let answer = &hello[0];
// println!("{}", answer);
// error[E0277]: the type `str` cannot be indexed by `{integer}`
// Another point about UTF-8 is that there are actually three relevant ways to look at strings
// from Rust’s perspective: as bytes, scalar values, and grapheme clusters (the closest thing to
// what we would call letters).
// “नमस्ते”
// Bytes: [224, 164, 168, 224, 164, 174, 224, 164, 184, 224, 165, 141, 224, 164, 164, 224, 165, 135]
// Unicode scalar values (Rust's char type): ['न', 'म', 'स', '्', 'त', 'े']
// There are six char values here, but the fourth and sixth are not letters: they’re diacritics
// that don’t make sense on their own
// Grapheme clusters: ["न", "म", "स्", "ते"]
let namaste = "नमस्ते";
println!("{}", &namaste[0..12]);
let hello = "Здравствуйте";
let s = &hello[0..4];
println!("{}", s);
for c in "नमस्ते".chars() {
println!("{}", c);
}
for b in "नमस्ते".bytes() {
print!("{},", b);
}
// But be sure to remember that valid Unicode scalar values may be made up of more than 1 byte.
// Getting grapheme clusters from strings is complex, so this functionality is not provided by
// the standard library. Crates are available on crates.io if this is the functionality you need.
}
fn hashmaps() {
use std::collections::HashMap;
let mut scores = HashMap::new();
scores.insert(String::from("Blue"), 10);
scores.insert(String::from("Yellow"), 50);
// Note that we need to first use the HashMap from the collections portion of the standard
// library. Of our three common collections, this one is the least often used, so it’s not
// included in the features brought into scope automatically in the prelude.
// The type | {
vectors();
strings();
hashmaps();
} | identifier_body | |
main.rs | can’t have mutable
// and immutable references in the same scope. That rule applies in Listing 8-7, where we hold
// an immutable reference to the first element in a vector and try to add an element to the end,
// which won’t work.
let mut v = vec![1, 2, 3, 4, 5];
let first = &v[0];
v.push(6);
//Below line causes Compilation Error
//println!("The first element is: {}", first);
// This error is due to the way vectors work: adding a new element onto the end of the vector
// might require allocating new memory and copying the old elements to the new space, if there
// isn’t enough room to put all the elements next to each other where the vector currently is.
// In that case, the reference to the first element would be pointing to deallocated memory.
// The borrowing rules prevent programs from ending up in that situation.
let v = vec![100, 32, 57];
for i in &v {
println!("{}", i);
}
// To change the value that the mutable reference refers to, we have to use the dereference
// operator (*) to get to the value in i before we can use the += operator.
let mut v = vec![100, 32, 57];
for i in &mut v {
*i += 50;
}
for i in &v {
println!("{}", i);
}
enum SpreadsheetCell {
Int(i32),
Float(f64),
Text(String),
}
let row = vec![
SpreadsheetCell::Int(3),
SpreadsheetCell::Text(String::from("blue")),
SpreadsheetCell::Float(10.12),
];
}
fn strings() {
let mut s = String::new();
let m = String::from("sdfsdf");
let data = "initial contents";
let s = data.to_string();
// the method also works on a literal directly:
let s = "initial contents".to_string();
let hello = String::from("السلام عليكم"); | let hello = String::from("שָׁלוֹם");
let hello = String::from("नमस्ते");
let hello = String::from("こんにちは");
let hello = String::from("안녕하세요");
let hello = String::from("你好");
let hello = String::from("Olá");
let hello = String::from("Здравствуйте");
let hello = String::from("Hola");
let mut s1 = String::from("foo");
let s2 = "bar";
s1.push_str(s2);
println!("s2 is {}", s2);
let mut s = String::from("lo");
s.push('l');
use std::ops::Add;
let s1 = String::from("Hello, ");
let s2 = String::from("world!");
// The reason we’re able to use &s2 in the call to add is that the compiler can coerce the
// &String argument into a &str. When we call the add method, Rust uses a deref coercion, which
// here turns &s2 into &s2[..]. We’ll discuss deref coercion in more depth in Chapter 15.
// Because add does not take ownership of the s parameter, s2 will still be a valid String after
// this operation.
// looks like it will copy both strings and create a new one, this statement actually takes
// ownership of s1, appends a copy of the contents of s2, and then returns ownership of the
// result. In other words, it looks like it’s making a lot of copies but isn’t; the
// implementation is more efficient than copying.
//let s3 = s1.add(&s2);
let s3 = s1 + &s2;
println!("{}", s3);
let s1 = String::from("tic");
let s2 = String::from("tac");
let s3 = String::from("toe");
//let s = s1 + "-" + &s2 + "-" + &s3;
let s = format!("{}-{}-{}", s1, s2, s3);
println!("{}", s);
// The version of the code using format! is much easier to read and doesn’t take ownership of
// any of its parameters.
println!("{}", s1);
// A String is a wrapper over a Vec<u8>
let len = String::from("Hola").len();
// In this case, len will be 4, which means the vector storing the string “Hola” is 4 bytes long.
// Each of these letters takes 1 byte when encoded in UTF-8
println!("{}", len);
let len = String::from("Здравствуйте").len();
println!("{}", len);
// It takes 24 bytes to encode “Здравствуйте” in UTF-8, because each Unicode scalar value in that string
// takes 2 bytes of storage. Therefore, an index into the string’s bytes will not always
// correlate to a valid Unicode scalar value. To demonstrate, consider this invalid Rust code:
// let hello = "Здравствуйте";
// let answer = &hello[0];
// println!("{}", answer);
// error[E0277]: the type `str` cannot be indexed by `{integer}`
// Another point about UTF-8 is that there are actually three relevant ways to look at strings
// from Rust’s perspective: as bytes, scalar values, and grapheme clusters (the closest thing to
// what we would call letters).
// “नमस्ते”
// Bytes: [224, 164, 168, 224, 164, 174, 224, 164, 184, 224, 165, 141, 224, 164, 164, 224, 165, 135]
// Unicode scalar values (Rust's char type): ['न', 'म', 'स', '्', 'त', 'े']
// There are six char values here, but the fourth and sixth are not letters: they’re diacritics
// that don’t make sense on their own
// Grapheme clusters: ["न", "म", "स्", "ते"]
let namaste = "नमस्ते";
println!("{}", &namaste[0..12]);
let hello = "Здравствуйте";
let s = &hello[0..4];
println!("{}", s);
for c in "नमस्ते".chars() {
println!("{}", c);
}
for b in "नमस्ते".bytes() {
print!("{},", b);
}
// But be sure to remember that valid Unicode scalar values may be made up of more than 1 byte.
// Getting grapheme clusters from strings is complex, so this functionality is not provided by
// the standard library. Crates are available on crates.io if this is the functionality you need.
}
fn hashmaps() {
use std::collections::HashMap;
let mut scores = HashMap::new();
scores.insert(String::from("Blue"), 10);
scores.insert(String::from("Yellow"), 50);
// Note that we need to first use the HashMap from the collections portion of the standard
// library. Of our three common collections, this one is the least often used, so it’s not
// included in the features brought into scope automatically in the prelude.
// The type annotation HashMap<_, _> is needed here because it’s possible to collect into many
// different data structures and Rust doesn’t know which you want unless you specify. For the
// parameters for the key and value types, however, we use underscores, and Rust can infer the
// types that the hash map contains based on the types of the data in the vectors.
let teams = vec![String::from("Blue"), String::from("Yellow")];
let initial_scores = vec![10, 50];
println!("");
let scores: HashMap<_, _> = teams.iter().zip(initial_scores.iter()).collect();
for (k, v) in &scores {
println!("{},{}", k, v);
}
let score = scores.get(&String::from("Blue"));
match score {
Some(s) => println!("{}", s),
None => ()
}
// For types that implement the Copy trait, like i32, the values are copied into the hash map.
// For owned values like String, the values will be moved and the hash map will be the owner of
// those values
let field_name = String::from("Favorite color");
let field_value = String::from("Blue");
let mut map = HashMap::new();
map.insert(field_name, field_value);
//error[E0382]: borrow of moved value: | let hello = String::from("Dobrý den");
let hello = String::from("Hello"); | random_line_split |
main.rs | can’t have mutable
// and immutable references in the same scope. That rule applies in Listing 8-7, where we hold
// an immutable reference to the first element in a vector and try to add an element to the end,
// which won’t work.
let mut v = vec![1, 2, 3, 4, 5];
let first = &v[0];
v.push(6);
//Below line causes Compilation Error
//println!("The first element is: {}", first);
// This error is due to the way vectors work: adding a new element onto the end of the vector
// might require allocating new memory and copying the old elements to the new space, if there
// isn’t enough room to put all the elements next to each other where the vector currently is.
// In that case, the reference to the first element would be pointing to deallocated memory.
// The borrowing rules prevent programs from ending up in that situation.
let v = vec![100, 32, 57];
for i in &v {
println!("{}", i);
}
// To change the value that the mutable reference refers to, we have to use the dereference
// operator (*) to get to the value in i before we can use the += operator.
let mut v = vec![100, 32, 57];
for i in &mut v {
*i += 50;
}
for i in &v {
println!("{}", i);
}
enum SpreadsheetCell {
Int(i32),
Float(f64),
Text(String),
}
let row = vec![
SpreadsheetCell::Int(3),
SpreadsheetCell::Text(String::from("blue")),
SpreadsheetCell::Float(10.12),
];
}
fn string | let mut s = String::new();
let m = String::from("sdfsdf");
let data = "initial contents";
let s = data.to_string();
// the method also works on a literal directly:
let s = "initial contents".to_string();
let hello = String::from("السلام عليكم");
let hello = String::from("Dobrý den");
let hello = String::from("Hello");
let hello = String::from("שָׁלוֹם");
let hello = String::from("नमस्ते");
let hello = String::from("こんにちは");
let hello = String::from("안녕하세요");
let hello = String::from("你好");
let hello = String::from("Olá");
let hello = String::from("Здравствуйте");
let hello = String::from("Hola");
let mut s1 = String::from("foo");
let s2 = "bar";
s1.push_str(s2);
println!("s2 is {}", s2);
let mut s = String::from("lo");
s.push('l');
use std::ops::Add;
let s1 = String::from("Hello, ");
let s2 = String::from("world!");
// The reason we’re able to use &s2 in the call to add is that the compiler can coerce the
// &String argument into a &str. When we call the add method, Rust uses a deref coercion, which
// here turns &s2 into &s2[..]. We’ll discuss deref coercion in more depth in Chapter 15.
// Because add does not take ownership of the s parameter, s2 will still be a valid String after
// this operation.
// looks like it will copy both strings and create a new one, this statement actually takes
// ownership of s1, appends a copy of the contents of s2, and then returns ownership of the
// result. In other words, it looks like it’s making a lot of copies but isn’t; the
// implementation is more efficient than copying.
//let s3 = s1.add(&s2);
let s3 = s1 + &s2;
println!("{}", s3);
let s1 = String::from("tic");
let s2 = String::from("tac");
let s3 = String::from("toe");
//let s = s1 + "-" + &s2 + "-" + &s3;
let s = format!("{}-{}-{}", s1, s2, s3);
println!("{}", s);
// The version of the code using format! is much easier to read and doesn’t take ownership of
// any of its parameters.
println!("{}", s1);
// A String is a wrapper over a Vec<u8>
let len = String::from("Hola").len();
// In this case, len will be 4, which means the vector storing the string “Hola” is 4 bytes long.
// Each of these letters takes 1 byte when encoded in UTF-8
println!("{}", len);
let len = String::from("Здравствуйте").len();
println!("{}", len);
// It takes 24 bytes to encode “Здравствуйте” in UTF-8, because each Unicode scalar value in that string
// takes 2 bytes of storage. Therefore, an index into the string’s bytes will not always
// correlate to a valid Unicode scalar value. To demonstrate, consider this invalid Rust code:
// let hello = "Здравствуйте";
// let answer = &hello[0];
// println!("{}", answer);
// error[E0277]: the type `str` cannot be indexed by `{integer}`
// Another point about UTF-8 is that there are actually three relevant ways to look at strings
// from Rust’s perspective: as bytes, scalar values, and grapheme clusters (the closest thing to
// what we would call letters).
// “नमस्ते”
// Bytes: [224, 164, 168, 224, 164, 174, 224, 164, 184, 224, 165, 141, 224, 164, 164, 224, 165, 135]
// Unicode scalar values (Rust's char type): ['न', 'म', 'स', '्', 'त', 'े']
// There are six char values here, but the fourth and sixth are not letters: they’re diacritics
// that don’t make sense on their own
// Grapheme clusters: ["न", "म", "स्", "ते"]
let namaste = "नमस्ते";
println!("{}", &namaste[0..12]);
let hello = "Здравствуйте";
let s = &hello[0..4];
println!("{}", s);
for c in "नमस्ते".chars() {
println!("{}", c);
}
for b in "नमस्ते".bytes() {
print!("{},", b);
}
// But be sure to remember that valid Unicode scalar values may be made up of more than 1 byte.
// Getting grapheme clusters from strings is complex, so this functionality is not provided by
// the standard library. Crates are available on crates.io if this is the functionality you need.
}
fn hashmaps() {
use std::collections::HashMap;
let mut scores = HashMap::new();
scores.insert(String::from("Blue"), 10);
scores.insert(String::from("Yellow"), 50);
// Note that we need to first use the HashMap from the collections portion of the standard
// library. Of our three common collections, this one is the least often used, so it’s not
// included in the features brought into scope automatically in the prelude.
// The type annotation HashMap<_, _> is needed here because it’s possible to collect into many
// different data structures and Rust doesn’t know which you want unless you specify. For the
// parameters for the key and value types, however, we use underscores, and Rust can infer the
// types that the hash map contains based on the types of the data in the vectors.
let teams = vec![String::from("Blue"), String::from("Yellow")];
let initial_scores = vec![10, 50];
println!("");
let scores: HashMap<_, _> = teams.iter().zip(initial_scores.iter()).collect();
for (k, v) in &scores {
println!("{},{}", k, v);
}
let score = scores.get(&String::from("Blue"));
match score {
Some(s) => println!("{}", s),
None => ()
}
// For types that implement the Copy trait, like i32, the values are copied into the hash map.
// For owned values like String, the values will be moved and the hash map will be the owner of
// those values
let field_name = String::from("Favorite color");
let field_value = String::from("Blue");
let mut map = HashMap::new();
map.insert(field_name, field_value);
//error[E0382]: borrow of moved | s() {
| identifier_name |
protocol.py | 0000000000000ccff41007200690061006c000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"
self.msg_blob = self.msg_blob.decode('hex')
print "UDP start"
self.poll_messages()
self.hello_everybody_lc = task.LoopingCall(self.hello_everybody).start(30, False)
self.tries = {}
self.resenders = {}
self.received = []
def poll_messages(self):
self.factory.bot.messages.get().addCallback(self.send_message)
def hello_everybody(self):
# print "UDP hello => all"
for id in [id for id, online in self.factory.bot.online.items() if online]:
self.say_hello(id)
def say_hello(self, id):
addr = self.factory.bot.addr(id)
if addr:
hello_packet = struct.pack("< I I 8x", 2, self.factory.account.id)
self.transport.write(hello_packet, addr)
def datagramReceived(self, data, host_port):
host, port = host_port
packet_type = ord(data[0])
if packet_type == 2:
self.handle_hello(data)
if packet_type == 15:
pass
# this is {HELLO REPLY} packet, we don't really need it, so -> ignore
#print "UDP hello reply <= ", host_port
#
if packet_type == 51:
self.handle_message(data)
if packet_type == 57:
self.invalidate_resender(data)
def handle_message(self, data):
data = data[1:]
# print len(data)
# print data.encode('hex')
format = "< I I I 96x I"
unpacked = struct.unpack(format, data[:112])
seq, from_id, to_id, length = unpacked
msg = data[112:].decode('utf_16_le', 'ignore')
# print self.factory.account.login + " => " + msg
# player = self.tcp.players.get(from_id)
# me = self.tcp.players.get(to_id)
addr = self.factory.bot.addr(from_id)
# print addr
key = "%s#%s" % (from_id, seq)
if addr and not key in self.received:
self.received.append(key)
reactor.callLater(10, lambda: self.received.remove(key))
# print "{MESSAGE #%s from %s of length %s(bytes)}" % (seq, login, length)
# print "{MSG BODY => %s}" % msg
reply = struct.pack("< B I I 8x", 57, seq, self.factory.account.id)
self.transport.write(reply, addr)
reactor.callLater(0, self.factory.bot.message_received, from_id, msg, True)
# self.send_message(me, player, u"you said => " + msg)
def send_message(self, player_and_msg):
self.poll_messages()
to_player, msg = player_and_msg
addr = self.factory.bot.addr(to_player.id)
if addr:
self.msg_seq += 1
seq = self.msg_seq
from_id = self.factory.account.id
length = len(msg) * 2
header = struct.pack("< B I I I", 51, seq, from_id, to_player.id)
packet = header + self.msg_blob + struct.pack("< I", length) + msg.encode('utf_16_le', 'ignore')
#self.transport.write(packet, addr)
self.resenders[seq] = task.LoopingCall(self.resend_message, seq, packet, addr)
self.tries[seq] = 0
self.resenders[seq].start(0.4)
# print "{MESSAGE to %s}" % to_player.login
# print "{MSG BODY => %s}" % msg
def | (self, data):
seq = struct.unpack("<I", data[1:5])[0]
# print "remote => i got #%s" % seq
lc = self.resenders.get(seq)
if lc:
lc.stop()
del self.resenders[seq]
del self.tries[seq]
def resend_message(self, seq, packet, addr):
lc = self.resenders.get(seq)
if lc:
self.tries[seq] += 1
self.transport.write(packet, addr)
# print "sending #%s, tries: %s" % (seq, self.tries[seq])
if self.tries[seq] > 6:
lc.stop()
del self.resenders[seq]
del self.tries[seq]
def handle_hello(self, data):
id = struct.unpack("<I", data[4:8])[0]
addr = self.factory.bot.addr(id)
if addr:
reply = struct.pack("< I I 4x I", 15, self.factory.account.id, id)
self.transport.write(reply, addr)
class GarenaRSProtocol(Protocol):
def __init__(self):
self.buffer = ''
def write(self, data): self.transport.write(data)
def write_hex(self, data): self.write(data.decode('hex'))
def connectionMade(self):
self.log = logging.getLogger("GRSP[%s]" % self.factory.account.login)
self.log.info(u"connection made, sending auth packet")
self.write_hex(self.factory.packet)
self.log.info(u"issuing disconnect in 45 seconds if Garena did not respond with WELCOME")
self.timeout_deferred = reactor.callLater(45, self.timeout)
def timeout(self):
self.log.error(u"Garena did not send WELCOME packet in 45 seconds, dropping connection now")
self.transport.loseConnection()
def dataReceived(self, data):
self.buffer += data
self.decodeHeader()
def decodeHeader(self):
if len(self.buffer) >= 5:
header = struct.unpack("< I B", self.buffer[:5])
if len(self.buffer) >= header[0]+4:
packet = self.buffer[5:header[0]+4]
self.buffer = self.buffer[header[0]+4:]
if len(self.buffer) >= 5:
reactor.callLater(0, self.decodeHeader)
self.decodePacket(header[1], packet)
def decodePacket(self, packet_type, data):
if self.factory.write_only and packet_type != 48: return
getattr(self, 'handle_' + {
34: 'player_came',
35: 'player_left',
37: 'message',
44: 'userlist',
48: 'welcome'
}.get(packet_type, 'non_existing'), lambda data: None)(data)
def handle_non_existing(self, data):
self.log.info(u"??? -> %s", data.encode('hex'))
def handle_player_left(self, data):
id = struct.unpack("< I", data)[0]
self.factory.bot.player_left(id)
def handle_player_came(self, data):
format = "< I 15s 6x 1B 2x 4B 32x"
unpacked = struct.unpack(format, data)
id = unpacked[0]
login = unicode(unpacked[1].rstrip(chr(0)), 'ascii', 'ignore')
ip = "%s.%s.%s.%s" % unpacked[3:]
lvl = unpacked[2]
port = struct.unpack(">H", data[40:42])[0]
if not Account.get_or(pk = id):
self.factory.bot.player_came(id, login, ip, port, lvl)
else:
self.log.info(u"%s is bot's account -> do nothing", login)
#if hasattr(self.factory, 'udp_protocol'):
# self.factory.udp_protocol.say_hello(id)
def handle_userlist(self, data):
self.log.info(u"cancelling TIMEOUT")
self.factory.connection = self
timeout_deferred = getattr(self, 'timeout_deferred', None)
if timeout_deferred and timeout_deferred.active:
timeout_deferred.cancel()
del self.timeout_deferred
self.log.info(u"got userlist")
for user_data in [ud for ud in split_by(data[8:], 64) if len(ud) == 64]:
self.handle_player_came(user_data)
def handle_message(self, data):
id = struct.unpack("<I", data[4:8])[0]
message = unicode(data[12:], 'utf_16_le', 'ignore').strip()
reactor.callLater(0, self.factory.bot.message_received, id, message)
def handle_welcome(self, data):
self.log.info(u"got WELCOME")
self.log.info | invalidate_resender | identifier_name |
protocol.py | action = msg.get("action")
print data
if action == "start":
for bot in self.bots:
for login in [bot.logins.get(id) for id, online in bot.online.items() if online]:
self.player_came(bot.name, login)
def message_received(self, room, by, body):
self.send_json({
"action": "message",
"room": room,
"by": by,
"body": body
})
def player_came(self, room, login):
self.send_json({
"action": "player_came",
"room": room,
"login": login
})
def player_left(self, room, login):
self.send_json({
"action": "player_left",
"room": room,
"login": login
})
#udp_info = LocalUDPInfo()
#reactor.listenUDP(8125, udp_info)
class GarenaRSUDPProtocol(DatagramProtocol):
def __init__(self, factory):
self.factory = factory
self.msg_seq = int(time.time()) # because of how large unsigned int is, it is ok to do this
self.msg_blob = "000000005c0000003f0000f800000040b40000000000000000000000ccff41007200690061006c000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"
self.msg_blob = self.msg_blob.decode('hex')
print "UDP start"
self.poll_messages()
self.hello_everybody_lc = task.LoopingCall(self.hello_everybody).start(30, False)
self.tries = {}
self.resenders = {}
self.received = []
def poll_messages(self):
self.factory.bot.messages.get().addCallback(self.send_message)
def hello_everybody(self):
# print "UDP hello => all"
for id in [id for id, online in self.factory.bot.online.items() if online]:
self.say_hello(id)
def say_hello(self, id):
addr = self.factory.bot.addr(id)
if addr:
hello_packet = struct.pack("< I I 8x", 2, self.factory.account.id)
self.transport.write(hello_packet, addr)
def datagramReceived(self, data, host_port):
host, port = host_port
packet_type = ord(data[0])
if packet_type == 2:
self.handle_hello(data)
if packet_type == 15:
pass
# this is {HELLO REPLY} packet, we don't really need it, so -> ignore
#print "UDP hello reply <= ", host_port
#
if packet_type == 51:
self.handle_message(data)
if packet_type == 57:
self.invalidate_resender(data)
def handle_message(self, data):
data = data[1:]
# print len(data)
# print data.encode('hex')
format = "< I I I 96x I"
unpacked = struct.unpack(format, data[:112])
seq, from_id, to_id, length = unpacked
msg = data[112:].decode('utf_16_le', 'ignore')
# print self.factory.account.login + " => " + msg
# player = self.tcp.players.get(from_id)
# me = self.tcp.players.get(to_id)
addr = self.factory.bot.addr(from_id)
# print addr
key = "%s#%s" % (from_id, seq)
if addr and not key in self.received:
self.received.append(key)
reactor.callLater(10, lambda: self.received.remove(key))
# print "{MESSAGE #%s from %s of length %s(bytes)}" % (seq, login, length)
# print "{MSG BODY => %s}" % msg
reply = struct.pack("< B I I 8x", 57, seq, self.factory.account.id)
self.transport.write(reply, addr)
reactor.callLater(0, self.factory.bot.message_received, from_id, msg, True)
# self.send_message(me, player, u"you said => " + msg)
def send_message(self, player_and_msg):
self.poll_messages()
to_player, msg = player_and_msg
addr = self.factory.bot.addr(to_player.id)
if addr:
self.msg_seq += 1
seq = self.msg_seq
from_id = self.factory.account.id
length = len(msg) * 2
header = struct.pack("< B I I I", 51, seq, from_id, to_player.id)
packet = header + self.msg_blob + struct.pack("< I", length) + msg.encode('utf_16_le', 'ignore')
#self.transport.write(packet, addr)
self.resenders[seq] = task.LoopingCall(self.resend_message, seq, packet, addr)
self.tries[seq] = 0
self.resenders[seq].start(0.4)
# print "{MESSAGE to %s}" % to_player.login
# print "{MSG BODY => %s}" % msg
def invalidate_resender(self, data):
seq = struct.unpack("<I", data[1:5])[0]
# print "remote => i got #%s" % seq
lc = self.resenders.get(seq)
if lc:
lc.stop()
del self.resenders[seq]
del self.tries[seq]
def resend_message(self, seq, packet, addr):
lc = self.resenders.get(seq)
if lc:
self.tries[seq] += 1
self.transport.write(packet, addr)
# print "sending #%s, tries: %s" % (seq, self.tries[seq])
if self.tries[seq] > 6:
lc.stop()
del self.resenders[seq]
del self.tries[seq]
def handle_hello(self, data):
id = struct.unpack("<I", data[4:8])[0]
addr = self.factory.bot.addr(id)
if addr:
reply = struct.pack("< I I 4x I", 15, self.factory.account.id, id)
self.transport.write(reply, addr)
class GarenaRSProtocol(Protocol):
def __init__(self):
self.buffer = ''
def write(self, data): self.transport.write(data)
def write_hex(self, data): self.write(data.decode('hex'))
def connectionMade(self):
self.log = logging.getLogger("GRSP[%s]" % self.factory.account.login)
self.log.info(u"connection made, sending auth packet")
self.write_hex(self.factory.packet)
self.log.info(u"issuing disconnect in 45 seconds if Garena did not respond with WELCOME")
self.timeout_deferred = reactor.callLater(45, self.timeout)
def timeout(self):
self.log.error(u"Garena did not send WELCOME packet in 45 seconds, dropping connection now")
self.transport.loseConnection()
def dataReceived(self, data):
self.buffer += data
self.decodeHeader()
def decodeHeader(self):
if len(self.buffer) >= 5:
header = struct.unpack("< I B", self.buffer[:5])
if len(self.buffer) >= header[0]+4:
packet = self.buffer[5:header[0]+4]
self.buffer = self.buffer[header[0]+4:]
if len(self.buffer) >= 5:
reactor.callLater(0, self.decodeHeader)
self.decodePacket(header[1], packet)
def decodePacket(self, packet_type, data):
if self.factory.write_only and packet_type != 48: return
getattr(self, 'handle_' + {
34: 'player_came',
35: 'player_left',
37: 'message',
44: 'userlist',
48: 'welcome'
}.get(packet_type, 'non_existing'), lambda data: None)(data)
def handle_non_existing(self, data):
self.log.info(u"??? -> %s", data.encode('hex'))
def handle_player_left(self, data):
id = struct.unpack("< I", data)[0]
self.factory.bot.player_left(id)
def handle_player_came(self, data):
format = "< I 15s 6x 1B 2x 4B 32x"
unpacked = struct.unpack(format, | def send_json(self, obj): self.transport.write(json.dumps(obj), self.node_io_addr)
def datagramReceived(self, data, addr):
msg = json.loads(data)
| random_line_split | |
protocol.py | 0000000000000ccff41007200690061006c000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"
self.msg_blob = self.msg_blob.decode('hex')
print "UDP start"
self.poll_messages()
self.hello_everybody_lc = task.LoopingCall(self.hello_everybody).start(30, False)
self.tries = {}
self.resenders = {}
self.received = []
def poll_messages(self):
self.factory.bot.messages.get().addCallback(self.send_message)
def hello_everybody(self):
# print "UDP hello => all"
for id in [id for id, online in self.factory.bot.online.items() if online]:
self.say_hello(id)
def say_hello(self, id):
addr = self.factory.bot.addr(id)
if addr:
hello_packet = struct.pack("< I I 8x", 2, self.factory.account.id)
self.transport.write(hello_packet, addr)
def datagramReceived(self, data, host_port):
host, port = host_port
packet_type = ord(data[0])
if packet_type == 2:
self.handle_hello(data)
if packet_type == 15:
pass
# this is {HELLO REPLY} packet, we don't really need it, so -> ignore
#print "UDP hello reply <= ", host_port
#
if packet_type == 51:
self.handle_message(data)
if packet_type == 57:
self.invalidate_resender(data)
def handle_message(self, data):
data = data[1:]
# print len(data)
# print data.encode('hex')
format = "< I I I 96x I"
unpacked = struct.unpack(format, data[:112])
seq, from_id, to_id, length = unpacked
msg = data[112:].decode('utf_16_le', 'ignore')
# print self.factory.account.login + " => " + msg
# player = self.tcp.players.get(from_id)
# me = self.tcp.players.get(to_id)
addr = self.factory.bot.addr(from_id)
# print addr
key = "%s#%s" % (from_id, seq)
if addr and not key in self.received:
self.received.append(key)
reactor.callLater(10, lambda: self.received.remove(key))
# print "{MESSAGE #%s from %s of length %s(bytes)}" % (seq, login, length)
# print "{MSG BODY => %s}" % msg
reply = struct.pack("< B I I 8x", 57, seq, self.factory.account.id)
self.transport.write(reply, addr)
reactor.callLater(0, self.factory.bot.message_received, from_id, msg, True)
# self.send_message(me, player, u"you said => " + msg)
def send_message(self, player_and_msg):
self.poll_messages()
to_player, msg = player_and_msg
addr = self.factory.bot.addr(to_player.id)
if addr:
self.msg_seq += 1
seq = self.msg_seq
from_id = self.factory.account.id
length = len(msg) * 2
header = struct.pack("< B I I I", 51, seq, from_id, to_player.id)
packet = header + self.msg_blob + struct.pack("< I", length) + msg.encode('utf_16_le', 'ignore')
#self.transport.write(packet, addr)
self.resenders[seq] = task.LoopingCall(self.resend_message, seq, packet, addr)
self.tries[seq] = 0
self.resenders[seq].start(0.4)
# print "{MESSAGE to %s}" % to_player.login
# print "{MSG BODY => %s}" % msg
def invalidate_resender(self, data):
seq = struct.unpack("<I", data[1:5])[0]
# print "remote => i got #%s" % seq
lc = self.resenders.get(seq)
if lc:
lc.stop()
del self.resenders[seq]
del self.tries[seq]
def resend_message(self, seq, packet, addr):
lc = self.resenders.get(seq)
if lc:
self.tries[seq] += 1
self.transport.write(packet, addr)
# print "sending #%s, tries: %s" % (seq, self.tries[seq])
if self.tries[seq] > 6:
lc.stop()
del self.resenders[seq]
del self.tries[seq]
def handle_hello(self, data):
id = struct.unpack("<I", data[4:8])[0]
addr = self.factory.bot.addr(id)
if addr:
reply = struct.pack("< I I 4x I", 15, self.factory.account.id, id)
self.transport.write(reply, addr)
class GarenaRSProtocol(Protocol):
def __init__(self):
self.buffer = ''
def write(self, data): self.transport.write(data)
def write_hex(self, data): self.write(data.decode('hex'))
def connectionMade(self):
|
def timeout(self):
self.log.error(u"Garena did not send WELCOME packet in 45 seconds, dropping connection now")
self.transport.loseConnection()
def dataReceived(self, data):
self.buffer += data
self.decodeHeader()
def decodeHeader(self):
if len(self.buffer) >= 5:
header = struct.unpack("< I B", self.buffer[:5])
if len(self.buffer) >= header[0]+4:
packet = self.buffer[5:header[0]+4]
self.buffer = self.buffer[header[0]+4:]
if len(self.buffer) >= 5:
reactor.callLater(0, self.decodeHeader)
self.decodePacket(header[1], packet)
def decodePacket(self, packet_type, data):
if self.factory.write_only and packet_type != 48: return
getattr(self, 'handle_' + {
34: 'player_came',
35: 'player_left',
37: 'message',
44: 'userlist',
48: 'welcome'
}.get(packet_type, 'non_existing'), lambda data: None)(data)
def handle_non_existing(self, data):
self.log.info(u"??? -> %s", data.encode('hex'))
def handle_player_left(self, data):
id = struct.unpack("< I", data)[0]
self.factory.bot.player_left(id)
def handle_player_came(self, data):
format = "< I 15s 6x 1B 2x 4B 32x"
unpacked = struct.unpack(format, data)
id = unpacked[0]
login = unicode(unpacked[1].rstrip(chr(0)), 'ascii', 'ignore')
ip = "%s.%s.%s.%s" % unpacked[3:]
lvl = unpacked[2]
port = struct.unpack(">H", data[40:42])[0]
if not Account.get_or(pk = id):
self.factory.bot.player_came(id, login, ip, port, lvl)
else:
self.log.info(u"%s is bot's account -> do nothing", login)
#if hasattr(self.factory, 'udp_protocol'):
# self.factory.udp_protocol.say_hello(id)
def handle_userlist(self, data):
self.log.info(u"cancelling TIMEOUT")
self.factory.connection = self
timeout_deferred = getattr(self, 'timeout_deferred', None)
if timeout_deferred and timeout_deferred.active:
timeout_deferred.cancel()
del self.timeout_deferred
self.log.info(u"got userlist")
for user_data in [ud for ud in split_by(data[8:], 64) if len(ud) == 64]:
self.handle_player_came(user_data)
def handle_message(self, data):
id = struct.unpack("<I", data[4:8])[0]
message = unicode(data[12:], 'utf_16_le', 'ignore').strip()
reactor.callLater(0, self.factory.bot.message_received, id, message)
def handle_welcome(self, data):
self.log.info(u"got WELCOME")
self.log.info(u | self.log = logging.getLogger("GRSP[%s]" % self.factory.account.login)
self.log.info(u"connection made, sending auth packet")
self.write_hex(self.factory.packet)
self.log.info(u"issuing disconnect in 45 seconds if Garena did not respond with WELCOME")
self.timeout_deferred = reactor.callLater(45, self.timeout) | identifier_body |
protocol.py | 0000000000000ccff41007200690061006c000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"
self.msg_blob = self.msg_blob.decode('hex')
print "UDP start"
self.poll_messages()
self.hello_everybody_lc = task.LoopingCall(self.hello_everybody).start(30, False)
self.tries = {}
self.resenders = {}
self.received = []
def poll_messages(self):
self.factory.bot.messages.get().addCallback(self.send_message)
def hello_everybody(self):
# print "UDP hello => all"
for id in [id for id, online in self.factory.bot.online.items() if online]:
self.say_hello(id)
def say_hello(self, id):
addr = self.factory.bot.addr(id)
if addr:
hello_packet = struct.pack("< I I 8x", 2, self.factory.account.id)
self.transport.write(hello_packet, addr)
def datagramReceived(self, data, host_port):
host, port = host_port
packet_type = ord(data[0])
if packet_type == 2:
self.handle_hello(data)
if packet_type == 15:
pass
# this is {HELLO REPLY} packet, we don't really need it, so -> ignore
#print "UDP hello reply <= ", host_port
#
if packet_type == 51:
self.handle_message(data)
if packet_type == 57:
self.invalidate_resender(data)
def handle_message(self, data):
data = data[1:]
# print len(data)
# print data.encode('hex')
format = "< I I I 96x I"
unpacked = struct.unpack(format, data[:112])
seq, from_id, to_id, length = unpacked
msg = data[112:].decode('utf_16_le', 'ignore')
# print self.factory.account.login + " => " + msg
# player = self.tcp.players.get(from_id)
# me = self.tcp.players.get(to_id)
addr = self.factory.bot.addr(from_id)
# print addr
key = "%s#%s" % (from_id, seq)
if addr and not key in self.received:
self.received.append(key)
reactor.callLater(10, lambda: self.received.remove(key))
# print "{MESSAGE #%s from %s of length %s(bytes)}" % (seq, login, length)
# print "{MSG BODY => %s}" % msg
reply = struct.pack("< B I I 8x", 57, seq, self.factory.account.id)
self.transport.write(reply, addr)
reactor.callLater(0, self.factory.bot.message_received, from_id, msg, True)
# self.send_message(me, player, u"you said => " + msg)
def send_message(self, player_and_msg):
self.poll_messages()
to_player, msg = player_and_msg
addr = self.factory.bot.addr(to_player.id)
if addr:
self.msg_seq += 1
seq = self.msg_seq
from_id = self.factory.account.id
length = len(msg) * 2
header = struct.pack("< B I I I", 51, seq, from_id, to_player.id)
packet = header + self.msg_blob + struct.pack("< I", length) + msg.encode('utf_16_le', 'ignore')
#self.transport.write(packet, addr)
self.resenders[seq] = task.LoopingCall(self.resend_message, seq, packet, addr)
self.tries[seq] = 0
self.resenders[seq].start(0.4)
# print "{MESSAGE to %s}" % to_player.login
# print "{MSG BODY => %s}" % msg
def invalidate_resender(self, data):
seq = struct.unpack("<I", data[1:5])[0]
# print "remote => i got #%s" % seq
lc = self.resenders.get(seq)
if lc:
lc.stop()
del self.resenders[seq]
del self.tries[seq]
def resend_message(self, seq, packet, addr):
lc = self.resenders.get(seq)
if lc:
|
def handle_hello(self, data):
id = struct.unpack("<I", data[4:8])[0]
addr = self.factory.bot.addr(id)
if addr:
reply = struct.pack("< I I 4x I", 15, self.factory.account.id, id)
self.transport.write(reply, addr)
class GarenaRSProtocol(Protocol):
def __init__(self):
self.buffer = ''
def write(self, data): self.transport.write(data)
def write_hex(self, data): self.write(data.decode('hex'))
def connectionMade(self):
self.log = logging.getLogger("GRSP[%s]" % self.factory.account.login)
self.log.info(u"connection made, sending auth packet")
self.write_hex(self.factory.packet)
self.log.info(u"issuing disconnect in 45 seconds if Garena did not respond with WELCOME")
self.timeout_deferred = reactor.callLater(45, self.timeout)
def timeout(self):
self.log.error(u"Garena did not send WELCOME packet in 45 seconds, dropping connection now")
self.transport.loseConnection()
def dataReceived(self, data):
self.buffer += data
self.decodeHeader()
def decodeHeader(self):
if len(self.buffer) >= 5:
header = struct.unpack("< I B", self.buffer[:5])
if len(self.buffer) >= header[0]+4:
packet = self.buffer[5:header[0]+4]
self.buffer = self.buffer[header[0]+4:]
if len(self.buffer) >= 5:
reactor.callLater(0, self.decodeHeader)
self.decodePacket(header[1], packet)
def decodePacket(self, packet_type, data):
if self.factory.write_only and packet_type != 48: return
getattr(self, 'handle_' + {
34: 'player_came',
35: 'player_left',
37: 'message',
44: 'userlist',
48: 'welcome'
}.get(packet_type, 'non_existing'), lambda data: None)(data)
def handle_non_existing(self, data):
self.log.info(u"??? -> %s", data.encode('hex'))
def handle_player_left(self, data):
id = struct.unpack("< I", data)[0]
self.factory.bot.player_left(id)
def handle_player_came(self, data):
format = "< I 15s 6x 1B 2x 4B 32x"
unpacked = struct.unpack(format, data)
id = unpacked[0]
login = unicode(unpacked[1].rstrip(chr(0)), 'ascii', 'ignore')
ip = "%s.%s.%s.%s" % unpacked[3:]
lvl = unpacked[2]
port = struct.unpack(">H", data[40:42])[0]
if not Account.get_or(pk = id):
self.factory.bot.player_came(id, login, ip, port, lvl)
else:
self.log.info(u"%s is bot's account -> do nothing", login)
#if hasattr(self.factory, 'udp_protocol'):
# self.factory.udp_protocol.say_hello(id)
def handle_userlist(self, data):
self.log.info(u"cancelling TIMEOUT")
self.factory.connection = self
timeout_deferred = getattr(self, 'timeout_deferred', None)
if timeout_deferred and timeout_deferred.active:
timeout_deferred.cancel()
del self.timeout_deferred
self.log.info(u"got userlist")
for user_data in [ud for ud in split_by(data[8:], 64) if len(ud) == 64]:
self.handle_player_came(user_data)
def handle_message(self, data):
id = struct.unpack("<I", data[4:8])[0]
message = unicode(data[12:], 'utf_16_le', 'ignore').strip()
reactor.callLater(0, self.factory.bot.message_received, id, message)
def handle_welcome(self, data):
self.log.info(u"got WELCOME")
self.log.info(u | self.tries[seq] += 1
self.transport.write(packet, addr)
# print "sending #%s, tries: %s" % (seq, self.tries[seq])
if self.tries[seq] > 6:
lc.stop()
del self.resenders[seq]
del self.tries[seq] | conditional_block |
WeightedSum.go | // The basic idea of WeightedSum is to substitute the WeightedSum equation to
// many Ci*Xi=Hi, so that a Sum constraint results with H1+H2+...+Hn=Z. This
// Sum constraint is substituted as well with X+Y=Z equations (see Sum
// constraint for more information).
type WeightedSum struct {
vars []core.VarId
hvars []core.VarId //helper-variables
cs []int
resultVar core.VarId
outCh chan<- *core.ChangeEvent
inCh <-chan *core.ChangeEntry
varidToDomainMap map[core.VarId]*core.IvDomain
id core.PropId
store *core.Store
pseudoPropsXCY []*PseudoXmultCeqY
pseudoPropsXYZ []*PseudoXplusYeqZ
}
func (this *WeightedSum) Start() {
// initial check
evt := core.CreateChangeEvent()
this.ivweightSumInitialCheck(evt)
core.SendChangesToStore(evt, this)
for changeEntry := range this.inCh {
core.LogIncomingChange(this, this.store, changeEntry)
evt = core.CreateChangeEvent()
varidChanged := changeEntry.GetID()
changedDom := this.varidToDomainMap[varidChanged]
changedDom.Removes(changeEntry.GetValues())
this.ivweightedSumPropagate(varidChanged, evt)
core.SendChangesToStore(evt, this)
}
}
// propagate check for changes. First look for X*C=Y propagators, then
// X+Y=Z and finally for the final propagator X=Y. Collect changes
func (this *WeightedSum) ivweightedSumPropagate(varid core.VarId, evt *core.ChangeEvent) {
this.checkXmultCeqY(varid, evt)
this.ivsumPropagate(varid, evt)
}
func (this *WeightedSum) ivsumPropagate(varid core.VarId, evt *core.ChangeEvent) {
hvarXCY := make([]core.VarId, len(this.pseudoPropsXCY))
for i, hxcy := range this.pseudoPropsXCY {
hvarXCY[i] = hxcy.y
}
ivsumBoundsPropagate(varid, this.varidToDomainMap, this.pseudoPropsXYZ, evt)
ivsumArcPropagate(varid, this.varidToDomainMap, this.pseudoPropsXYZ, evt)
}
// initialCheck check for changes. First look for X*C=Y propagators, then
// X+Y=Z and finally for the final propagator X=Y. Collect changes
func (this *WeightedSum) ivweightSumInitialCheck(evt *core.ChangeEvent) {
this.checkXmultCeqY(-1, evt)
hvarXCY := make([]core.VarId, len(this.pseudoPropsXCY))
for i, hxcy := range this.pseudoPropsXCY {
hvarXCY[i] = hxcy.y
}
ivsumBoundsInitialCheck(this.varidToDomainMap, this.pseudoPropsXYZ, evt)
ivsumArcInitialCheck(this.varidToDomainMap, this.pseudoPropsXYZ, evt)
}
func (this *WeightedSum) checkXmultCeqY(varid core.VarId, evt *core.ChangeEvent) {
for i, prop := range this.pseudoPropsXCY {
xDom := this.varidToDomainMap[prop.x]
yDom := this.varidToDomainMap[prop.y]
c := this.cs[i]
if varid == -1 {
xDom := this.varidToDomainMap[prop.x]
yDom := this.varidToDomainMap[prop.y]
c := this.cs[i]
ivfirstInMultSecondOutBOUNDS(xDom, c, yDom, prop.y, evt)
ivsecondInMultFirstOutBOUNDS(yDom, c, xDom, prop.x, evt)
ivfirstInMultSecondOutARC(xDom, c, yDom, prop.y, evt)
ivsecondInMultFirstOutARC(yDom, c, xDom, prop.x, evt)
} else if prop.x == varid {
ivfirstInMultSecondOutBOUNDS(xDom, c, yDom, prop.y, evt)
ivfirstInMultSecondOutARC(xDom, c, yDom, prop.y, evt)
} else if prop.y == varid {
ivsecondInMultFirstOutBOUNDS(yDom, c, xDom, prop.x, evt)
ivsecondInMultFirstOutARC(yDom, c, xDom, prop.x, evt)
}
}
}
// Register generates auxiliary variables and makes pseudo structs
// and all vars will be registered at store and get domains and channels
func (this *WeightedSum) Register(store *core.Store) {
allvars := make([]core.VarId, len(this.vars)+len(this.hvars)+1)
i := 0
for _, v := range this.vars {
allvars[i] = v
i++
}
for _, v := range this.hvars {
allvars[i] = v
i++
}
allvars[i] = this.resultVar
var domains map[core.VarId]core.Domain
this.inCh, domains, this.outCh =
store.RegisterPropagatorMap(allvars, this.id)
this.varidToDomainMap = core.GetVaridToIntervalDomains(domains)
this.store = store
}
// SetID is used by the store to set the propagator's ID, don't use it
// yourself or bad things will happen
func (this *WeightedSum) SetID(propID core.PropId) {
this.id = propID
}
func (this *WeightedSum) GetID() core.PropId {
return this.id
}
func CreateWeightedSum(store *core.Store, resultVar core.VarId, cs []int,
intVars ...core.VarId) *WeightedSum {
if core.GetLogger().DoDebug() {
core.GetLogger().Dln("CreateWeightedSum-propagator")
}
prop := new(WeightedSum)
prop.vars = intVars
prop.resultVar = resultVar
prop.cs = cs
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(prop.vars))
prop.hvars = make([]core.VarId, 0)
for i, X := range prop.vars {
H := core.CreateAuxIntVarIvValues(store,
core.ScalarSlice(prop.cs[i], store.GetDomain(X).Values_asSlice()))
prop.pseudoPropsXCY[i] = CreatePseudoXmultCeqY(X, prop.cs[i], H)
prop.hvars = append(prop.hvars, H)
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(prop.pseudoPropsXCY)-1)
H := prop.pseudoPropsXCY[0].y
newHVars := make([]core.VarId, 0)
for i, p := range prop.pseudoPropsXCY[1 : len(prop.vars)-1] {
NewH := core.CreateAuxIntVarIvFromTo(store,
store.GetDomain(H).GetMin()+store.GetDomain(p.y).GetMin(),
store.GetDomain(H).GetMax()+store.GetDomain(p.y).GetMax())
prop.pseudoPropsXYZ[i] = CreatePseudoXplusYeqZ(H, p.y, NewH)
H = NewH
newHVars = append(newHVars, NewH)
}
X := prop.hvars[len(prop.hvars)-1]
prop.hvars = append(prop.hvars, newHVars...)
prop.pseudoPropsXYZ[len(prop.pseudoPropsXYZ)-1] = CreatePseudoXplusYeqZ(H, X, prop.resultVar)
return prop
}
func (this *WeightedSum) String() string {
vars_str := make([]string, len(this.vars))
for i, var_id := range this.vars {
vars_str[i] = fmt.Sprintf("%v*%s",
this.cs[i], this.store.GetName(var_id))
}
return fmt.Sprintf("PROP_%d %s = %s",
this.id, strings.Join(vars_str, "+"),
this.store.GetName(this.resultVar))
}
func (this *WeightedSum) Clone() core.Constraint {
prop := new(WeightedSum)
prop.vars = make([]core.VarId, len(this.vars))
for i, single_var := range this.vars {
prop.vars[i] = single_var
}
prop.resultVar = this.resultVar
prop.cs = make([]int, len(this.cs))
for i, c := range this.cs {
prop.cs[i] = c
}
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(this.pseudoPropsXCY))
for i, p := range this.pseudoPropsXCY {
prop.pseudoPropsXCY[i] = p.Clone()
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(this.pseudoPropsXYZ))
for i, p := range this.pseudoPropsXYZ {
prop.pseudoPropsXYZ[i] = p.Clone()
}
prop.hvars = make([]core.VarId, len(this.hvars))
for i, single_var := range this.hvars {
prop.hvars[i] = single_var
}
return prop
}
// INFO: the following two propagation-functions do not work on special
// interval-operations. Reason is: the | random_line_split | ||
WeightedSum.go | , this.outCh =
store.RegisterPropagatorMap(allvars, this.id)
this.varidToDomainMap = core.GetVaridToIntervalDomains(domains)
this.store = store
}
// SetID is used by the store to set the propagator's ID, don't use it
// yourself or bad things will happen
func (this *WeightedSum) SetID(propID core.PropId) {
this.id = propID
}
func (this *WeightedSum) GetID() core.PropId {
return this.id
}
func CreateWeightedSum(store *core.Store, resultVar core.VarId, cs []int,
intVars ...core.VarId) *WeightedSum {
if core.GetLogger().DoDebug() {
core.GetLogger().Dln("CreateWeightedSum-propagator")
}
prop := new(WeightedSum)
prop.vars = intVars
prop.resultVar = resultVar
prop.cs = cs
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(prop.vars))
prop.hvars = make([]core.VarId, 0)
for i, X := range prop.vars {
H := core.CreateAuxIntVarIvValues(store,
core.ScalarSlice(prop.cs[i], store.GetDomain(X).Values_asSlice()))
prop.pseudoPropsXCY[i] = CreatePseudoXmultCeqY(X, prop.cs[i], H)
prop.hvars = append(prop.hvars, H)
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(prop.pseudoPropsXCY)-1)
H := prop.pseudoPropsXCY[0].y
newHVars := make([]core.VarId, 0)
for i, p := range prop.pseudoPropsXCY[1 : len(prop.vars)-1] {
NewH := core.CreateAuxIntVarIvFromTo(store,
store.GetDomain(H).GetMin()+store.GetDomain(p.y).GetMin(),
store.GetDomain(H).GetMax()+store.GetDomain(p.y).GetMax())
prop.pseudoPropsXYZ[i] = CreatePseudoXplusYeqZ(H, p.y, NewH)
H = NewH
newHVars = append(newHVars, NewH)
}
X := prop.hvars[len(prop.hvars)-1]
prop.hvars = append(prop.hvars, newHVars...)
prop.pseudoPropsXYZ[len(prop.pseudoPropsXYZ)-1] = CreatePseudoXplusYeqZ(H, X, prop.resultVar)
return prop
}
func (this *WeightedSum) String() string {
vars_str := make([]string, len(this.vars))
for i, var_id := range this.vars {
vars_str[i] = fmt.Sprintf("%v*%s",
this.cs[i], this.store.GetName(var_id))
}
return fmt.Sprintf("PROP_%d %s = %s",
this.id, strings.Join(vars_str, "+"),
this.store.GetName(this.resultVar))
}
func (this *WeightedSum) Clone() core.Constraint {
prop := new(WeightedSum)
prop.vars = make([]core.VarId, len(this.vars))
for i, single_var := range this.vars {
prop.vars[i] = single_var
}
prop.resultVar = this.resultVar
prop.cs = make([]int, len(this.cs))
for i, c := range this.cs {
prop.cs[i] = c
}
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(this.pseudoPropsXCY))
for i, p := range this.pseudoPropsXCY {
prop.pseudoPropsXCY[i] = p.Clone()
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(this.pseudoPropsXYZ))
for i, p := range this.pseudoPropsXYZ {
prop.pseudoPropsXYZ[i] = p.Clone()
}
prop.hvars = make([]core.VarId, len(this.hvars))
for i, single_var := range this.hvars {
prop.hvars[i] = single_var
}
return prop
}
// INFO: the following two propagation-functions do not work on special
// interval-operations. Reason is: the multiplication is always worst case!
// example-CSP:
// X*C=Y
// X:[1,11]
// C:2
// Y:[1,20]
// RESULT:
// X:[1,10]
// Y:[2] [4] [6] ... [20]
// X*C=Y
// ivfirstInMultSecondOutARC collect changes, when first variable has changed
// e.g. X=Y/C, then X is first variable
func ivfirstInMultSecondOutBOUNDS(firstInDomain *core.IvDomain, c int,
secondOutDomain *core.IvDomain, secondOutVarId core.VarId,
evt *core.ChangeEvent) {
if firstInDomain.IsEmpty() {
return
}
removeParts := make([]*core.IvDomPart, 0)
minX := firstInDomain.GetMin() * c
maxX := firstInDomain.GetMax() * c
for i, yPart := range secondOutDomain.GetParts() {
removeParts = append(removeParts, yPart.DIFFERENCE_MIN_MAX(minX, maxX)...)
if yPart.GT(maxX) {
removeParts = append(removeParts, secondOutDomain.GetParts()[i:]...)
break
}
}
if len(removeParts) != 0 {
chEntry := core.CreateChangeEntryWithValues(secondOutVarId, core.CreateIvDomainDomParts(removeParts))
evt.AddChangeEntry(chEntry)
}
}
// ivsecondInMultFirstOutARC collect changes, when second variable has changed
// e.g. X*C=Y, then Y is second variable
func ivsecondInMultFirstOutBOUNDS(secondInDomain *core.IvDomain, c int,
firstOutDomain *core.IvDomain, firstOutVarId core.VarId,
evt *core.ChangeEvent) {
removeParts := make([]*core.IvDomPart, 0)
if secondInDomain.IsEmpty() {
return
}
minY := secondInDomain.GetMin() / c
maxY := secondInDomain.GetMax() / c
for i, xPart := range firstOutDomain.GetParts() {
removeParts = append(removeParts, xPart.DIFFERENCE_MIN_MAX(minY, maxY)...)
if xPart.GT(maxY) {
removeParts = append(removeParts, firstOutDomain.GetParts()[i:]...)
break
}
}
if len(removeParts) != 0 {
chEntry := core.CreateChangeEntryWithValues(firstOutVarId, core.CreateIvDomainDomParts(removeParts))
evt.AddChangeEntry(chEntry)
}
}
// X*C=Y
// ivfirstInMultSecondOutARC collect changes, when first variable has changed
// e.g. Y=X/C, then X is first variable
func ivfirstInMultSecondOutARC(firstInDomain *core.IvDomain, c int,
secondOutDomain *core.IvDomain, secondOutVarId core.VarId,
evt *core.ChangeEvent) {
if firstInDomain.IsEmpty() {
return
}
vals := make([]int, 0)
for _, y_val := range secondOutDomain.GetValues() {
if c == 0 || !(y_val%c == 0) {
if y_val != 0 {
vals = append(vals, y_val)
}
} else {
if !firstInDomain.Contains(y_val / c) {
vals = append(vals, y_val)
}
}
}
if len(vals) != 0 {
chEntry := core.CreateChangeEntryWithIntValues(secondOutVarId, vals)
evt.AddChangeEntry(chEntry)
}
}
// ivsecondInMultFirstOutARC collect changes, when second variable has changed
// e.g. X*C=Y, then Y is second variable
func ivsecondInMultFirstOutARC(secondInDomain *core.IvDomain, c int,
firstOutDomain *core.IvDomain, firstOutVarId core.VarId,
evt *core.ChangeEvent) {
if secondInDomain.IsEmpty() {
return
}
vals := make([]int, 0)
for _, x_val := range firstOutDomain.GetValues() {
if !secondInDomain.Contains(x_val * c) {
vals = append(vals, x_val)
}
}
if len(vals) != 0 {
chEntry := core.CreateChangeEntryWithIntValues(firstOutVarId, vals)
evt.AddChangeEntry(chEntry)
}
}
func (this *WeightedSum) GetAllVars() []core.VarId {
allvars := make([]core.VarId, len(this.vars)+len(this.hvars)+1)
i := 0
for _, v := range this.vars {
allvars[i] = v
i++
}
for _, v := range this.hvars {
allvars[i] = v
i++
}
allvars[i] = this.resultVar
return allvars
}
func (this *WeightedSum) GetVarIds() []core.VarId {
return this.GetAllVars()
}
func (this *WeightedSum) GetDomains() []core.Domain {
return core.ValuesOfMapVarIdToIvDomain(this.GetAllVars(), this.varidToDomainMap)
}
func (this *WeightedSum) GetInCh() <-chan *core.ChangeEntry | {
return this.inCh
} | identifier_body | |
WeightedSum.go | .varidToDomainMap, this.pseudoPropsXYZ, evt)
}
// initialCheck check for changes. First look for X*C=Y propagators, then
// X+Y=Z and finally for the final propagator X=Y. Collect changes
func (this *WeightedSum) ivweightSumInitialCheck(evt *core.ChangeEvent) {
this.checkXmultCeqY(-1, evt)
hvarXCY := make([]core.VarId, len(this.pseudoPropsXCY))
for i, hxcy := range this.pseudoPropsXCY {
hvarXCY[i] = hxcy.y
}
ivsumBoundsInitialCheck(this.varidToDomainMap, this.pseudoPropsXYZ, evt)
ivsumArcInitialCheck(this.varidToDomainMap, this.pseudoPropsXYZ, evt)
}
func (this *WeightedSum) checkXmultCeqY(varid core.VarId, evt *core.ChangeEvent) {
for i, prop := range this.pseudoPropsXCY {
xDom := this.varidToDomainMap[prop.x]
yDom := this.varidToDomainMap[prop.y]
c := this.cs[i]
if varid == -1 {
xDom := this.varidToDomainMap[prop.x]
yDom := this.varidToDomainMap[prop.y]
c := this.cs[i]
ivfirstInMultSecondOutBOUNDS(xDom, c, yDom, prop.y, evt)
ivsecondInMultFirstOutBOUNDS(yDom, c, xDom, prop.x, evt)
ivfirstInMultSecondOutARC(xDom, c, yDom, prop.y, evt)
ivsecondInMultFirstOutARC(yDom, c, xDom, prop.x, evt)
} else if prop.x == varid {
ivfirstInMultSecondOutBOUNDS(xDom, c, yDom, prop.y, evt)
ivfirstInMultSecondOutARC(xDom, c, yDom, prop.y, evt)
} else if prop.y == varid {
ivsecondInMultFirstOutBOUNDS(yDom, c, xDom, prop.x, evt)
ivsecondInMultFirstOutARC(yDom, c, xDom, prop.x, evt)
}
}
}
// Register generates auxiliary variables and makes pseudo structs
// and all vars will be registered at store and get domains and channels
func (this *WeightedSum) Register(store *core.Store) {
allvars := make([]core.VarId, len(this.vars)+len(this.hvars)+1)
i := 0
for _, v := range this.vars {
allvars[i] = v
i++
}
for _, v := range this.hvars {
allvars[i] = v
i++
}
allvars[i] = this.resultVar
var domains map[core.VarId]core.Domain
this.inCh, domains, this.outCh =
store.RegisterPropagatorMap(allvars, this.id)
this.varidToDomainMap = core.GetVaridToIntervalDomains(domains)
this.store = store
}
// SetID is used by the store to set the propagator's ID, don't use it
// yourself or bad things will happen
func (this *WeightedSum) SetID(propID core.PropId) {
this.id = propID
}
func (this *WeightedSum) GetID() core.PropId {
return this.id
}
func CreateWeightedSum(store *core.Store, resultVar core.VarId, cs []int,
intVars ...core.VarId) *WeightedSum {
if core.GetLogger().DoDebug() {
core.GetLogger().Dln("CreateWeightedSum-propagator")
}
prop := new(WeightedSum)
prop.vars = intVars
prop.resultVar = resultVar
prop.cs = cs
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(prop.vars))
prop.hvars = make([]core.VarId, 0)
for i, X := range prop.vars {
H := core.CreateAuxIntVarIvValues(store,
core.ScalarSlice(prop.cs[i], store.GetDomain(X).Values_asSlice()))
prop.pseudoPropsXCY[i] = CreatePseudoXmultCeqY(X, prop.cs[i], H)
prop.hvars = append(prop.hvars, H)
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(prop.pseudoPropsXCY)-1)
H := prop.pseudoPropsXCY[0].y
newHVars := make([]core.VarId, 0)
for i, p := range prop.pseudoPropsXCY[1 : len(prop.vars)-1] {
NewH := core.CreateAuxIntVarIvFromTo(store,
store.GetDomain(H).GetMin()+store.GetDomain(p.y).GetMin(),
store.GetDomain(H).GetMax()+store.GetDomain(p.y).GetMax())
prop.pseudoPropsXYZ[i] = CreatePseudoXplusYeqZ(H, p.y, NewH)
H = NewH
newHVars = append(newHVars, NewH)
}
X := prop.hvars[len(prop.hvars)-1]
prop.hvars = append(prop.hvars, newHVars...)
prop.pseudoPropsXYZ[len(prop.pseudoPropsXYZ)-1] = CreatePseudoXplusYeqZ(H, X, prop.resultVar)
return prop
}
func (this *WeightedSum) String() string {
vars_str := make([]string, len(this.vars))
for i, var_id := range this.vars {
vars_str[i] = fmt.Sprintf("%v*%s",
this.cs[i], this.store.GetName(var_id))
}
return fmt.Sprintf("PROP_%d %s = %s",
this.id, strings.Join(vars_str, "+"),
this.store.GetName(this.resultVar))
}
func (this *WeightedSum) Clone() core.Constraint {
prop := new(WeightedSum)
prop.vars = make([]core.VarId, len(this.vars))
for i, single_var := range this.vars {
prop.vars[i] = single_var
}
prop.resultVar = this.resultVar
prop.cs = make([]int, len(this.cs))
for i, c := range this.cs {
prop.cs[i] = c
}
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(this.pseudoPropsXCY))
for i, p := range this.pseudoPropsXCY {
prop.pseudoPropsXCY[i] = p.Clone()
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(this.pseudoPropsXYZ))
for i, p := range this.pseudoPropsXYZ {
prop.pseudoPropsXYZ[i] = p.Clone()
}
prop.hvars = make([]core.VarId, len(this.hvars))
for i, single_var := range this.hvars {
prop.hvars[i] = single_var
}
return prop
}
// INFO: the following two propagation-functions do not work on special
// interval-operations. Reason is: the multiplication is always worst case!
// example-CSP:
// X*C=Y
// X:[1,11]
// C:2
// Y:[1,20]
// RESULT:
// X:[1,10]
// Y:[2] [4] [6] ... [20]
// X*C=Y
// ivfirstInMultSecondOutARC collect changes, when first variable has changed
// e.g. X=Y/C, then X is first variable
func ivfirstInMultSecondOutBOUNDS(firstInDomain *core.IvDomain, c int,
secondOutDomain *core.IvDomain, secondOutVarId core.VarId,
evt *core.ChangeEvent) {
if firstInDomain.IsEmpty() {
return
}
removeParts := make([]*core.IvDomPart, 0)
minX := firstInDomain.GetMin() * c
maxX := firstInDomain.GetMax() * c
for i, yPart := range secondOutDomain.GetParts() {
removeParts = append(removeParts, yPart.DIFFERENCE_MIN_MAX(minX, maxX)...)
if yPart.GT(maxX) {
removeParts = append(removeParts, secondOutDomain.GetParts()[i:]...)
break
}
}
if len(removeParts) != 0 {
chEntry := core.CreateChangeEntryWithValues(secondOutVarId, core.CreateIvDomainDomParts(removeParts))
evt.AddChangeEntry(chEntry)
}
}
// ivsecondInMultFirstOutARC collect changes, when second variable has changed
// e.g. X*C=Y, then Y is second variable
func ivsecondInMultFirstOutBOUNDS(secondInDomain *core.IvDomain, c int,
firstOutDomain *core.IvDomain, firstOutVarId core.VarId,
evt *core.ChangeEvent) {
removeParts := make([]*core.IvDomPart, 0)
if secondInDomain.IsEmpty() {
return
}
minY := secondInDomain.GetMin() / c
maxY := secondInDomain.GetMax() / c
for i, xPart := range firstOutDomain.GetParts() | {
removeParts = append(removeParts, xPart.DIFFERENCE_MIN_MAX(minY, maxY)...)
if xPart.GT(maxY) {
removeParts = append(removeParts, firstOutDomain.GetParts()[i:]...)
break
}
} | conditional_block | |
WeightedSum.go |
i++
}
for _, v := range this.hvars {
allvars[i] = v
i++
}
allvars[i] = this.resultVar
var domains map[core.VarId]core.Domain
this.inCh, domains, this.outCh =
store.RegisterPropagatorMap(allvars, this.id)
this.varidToDomainMap = core.GetVaridToIntervalDomains(domains)
this.store = store
}
// SetID is used by the store to set the propagator's ID, don't use it
// yourself or bad things will happen
func (this *WeightedSum) SetID(propID core.PropId) {
this.id = propID
}
func (this *WeightedSum) GetID() core.PropId {
return this.id
}
func CreateWeightedSum(store *core.Store, resultVar core.VarId, cs []int,
intVars ...core.VarId) *WeightedSum {
if core.GetLogger().DoDebug() {
core.GetLogger().Dln("CreateWeightedSum-propagator")
}
prop := new(WeightedSum)
prop.vars = intVars
prop.resultVar = resultVar
prop.cs = cs
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(prop.vars))
prop.hvars = make([]core.VarId, 0)
for i, X := range prop.vars {
H := core.CreateAuxIntVarIvValues(store,
core.ScalarSlice(prop.cs[i], store.GetDomain(X).Values_asSlice()))
prop.pseudoPropsXCY[i] = CreatePseudoXmultCeqY(X, prop.cs[i], H)
prop.hvars = append(prop.hvars, H)
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(prop.pseudoPropsXCY)-1)
H := prop.pseudoPropsXCY[0].y
newHVars := make([]core.VarId, 0)
for i, p := range prop.pseudoPropsXCY[1 : len(prop.vars)-1] {
NewH := core.CreateAuxIntVarIvFromTo(store,
store.GetDomain(H).GetMin()+store.GetDomain(p.y).GetMin(),
store.GetDomain(H).GetMax()+store.GetDomain(p.y).GetMax())
prop.pseudoPropsXYZ[i] = CreatePseudoXplusYeqZ(H, p.y, NewH)
H = NewH
newHVars = append(newHVars, NewH)
}
X := prop.hvars[len(prop.hvars)-1]
prop.hvars = append(prop.hvars, newHVars...)
prop.pseudoPropsXYZ[len(prop.pseudoPropsXYZ)-1] = CreatePseudoXplusYeqZ(H, X, prop.resultVar)
return prop
}
func (this *WeightedSum) String() string {
vars_str := make([]string, len(this.vars))
for i, var_id := range this.vars {
vars_str[i] = fmt.Sprintf("%v*%s",
this.cs[i], this.store.GetName(var_id))
}
return fmt.Sprintf("PROP_%d %s = %s",
this.id, strings.Join(vars_str, "+"),
this.store.GetName(this.resultVar))
}
func (this *WeightedSum) Clone() core.Constraint {
prop := new(WeightedSum)
prop.vars = make([]core.VarId, len(this.vars))
for i, single_var := range this.vars {
prop.vars[i] = single_var
}
prop.resultVar = this.resultVar
prop.cs = make([]int, len(this.cs))
for i, c := range this.cs {
prop.cs[i] = c
}
prop.pseudoPropsXCY = make([]*PseudoXmultCeqY, len(this.pseudoPropsXCY))
for i, p := range this.pseudoPropsXCY {
prop.pseudoPropsXCY[i] = p.Clone()
}
prop.pseudoPropsXYZ = make([]*PseudoXplusYeqZ, len(this.pseudoPropsXYZ))
for i, p := range this.pseudoPropsXYZ {
prop.pseudoPropsXYZ[i] = p.Clone()
}
prop.hvars = make([]core.VarId, len(this.hvars))
for i, single_var := range this.hvars {
prop.hvars[i] = single_var
}
return prop
}
// INFO: the following two propagation-functions do not work on special
// interval-operations. Reason is: the multiplication is always worst case!
// example-CSP:
// X*C=Y
// X:[1,11]
// C:2
// Y:[1,20]
// RESULT:
// X:[1,10]
// Y:[2] [4] [6] ... [20]
// X*C=Y
// ivfirstInMultSecondOutARC collect changes, when first variable has changed
// e.g. X=Y/C, then X is first variable
func ivfirstInMultSecondOutBOUNDS(firstInDomain *core.IvDomain, c int,
secondOutDomain *core.IvDomain, secondOutVarId core.VarId,
evt *core.ChangeEvent) {
if firstInDomain.IsEmpty() {
return
}
removeParts := make([]*core.IvDomPart, 0)
minX := firstInDomain.GetMin() * c
maxX := firstInDomain.GetMax() * c
for i, yPart := range secondOutDomain.GetParts() {
removeParts = append(removeParts, yPart.DIFFERENCE_MIN_MAX(minX, maxX)...)
if yPart.GT(maxX) {
removeParts = append(removeParts, secondOutDomain.GetParts()[i:]...)
break
}
}
if len(removeParts) != 0 {
chEntry := core.CreateChangeEntryWithValues(secondOutVarId, core.CreateIvDomainDomParts(removeParts))
evt.AddChangeEntry(chEntry)
}
}
// ivsecondInMultFirstOutARC collect changes, when second variable has changed
// e.g. X*C=Y, then Y is second variable
func ivsecondInMultFirstOutBOUNDS(secondInDomain *core.IvDomain, c int,
firstOutDomain *core.IvDomain, firstOutVarId core.VarId,
evt *core.ChangeEvent) {
removeParts := make([]*core.IvDomPart, 0)
if secondInDomain.IsEmpty() {
return
}
minY := secondInDomain.GetMin() / c
maxY := secondInDomain.GetMax() / c
for i, xPart := range firstOutDomain.GetParts() {
removeParts = append(removeParts, xPart.DIFFERENCE_MIN_MAX(minY, maxY)...)
if xPart.GT(maxY) {
removeParts = append(removeParts, firstOutDomain.GetParts()[i:]...)
break
}
}
if len(removeParts) != 0 {
chEntry := core.CreateChangeEntryWithValues(firstOutVarId, core.CreateIvDomainDomParts(removeParts))
evt.AddChangeEntry(chEntry)
}
}
// X*C=Y
// ivfirstInMultSecondOutARC collect changes, when first variable has changed
// e.g. Y=X/C, then X is first variable
func ivfirstInMultSecondOutARC(firstInDomain *core.IvDomain, c int,
secondOutDomain *core.IvDomain, secondOutVarId core.VarId,
evt *core.ChangeEvent) {
if firstInDomain.IsEmpty() {
return
}
vals := make([]int, 0)
for _, y_val := range secondOutDomain.GetValues() {
if c == 0 || !(y_val%c == 0) {
if y_val != 0 {
vals = append(vals, y_val)
}
} else {
if !firstInDomain.Contains(y_val / c) {
vals = append(vals, y_val)
}
}
}
if len(vals) != 0 {
chEntry := core.CreateChangeEntryWithIntValues(secondOutVarId, vals)
evt.AddChangeEntry(chEntry)
}
}
// ivsecondInMultFirstOutARC collect changes, when second variable has changed
// e.g. X*C=Y, then Y is second variable
func ivsecondInMultFirstOutARC(secondInDomain *core.IvDomain, c int,
firstOutDomain *core.IvDomain, firstOutVarId core.VarId,
evt *core.ChangeEvent) {
if secondInDomain.IsEmpty() {
return
}
vals := make([]int, 0)
for _, x_val := range firstOutDomain.GetValues() {
if !secondInDomain.Contains(x_val * c) {
vals = append(vals, x_val)
}
}
if len(vals) != 0 {
chEntry := core.CreateChangeEntryWithIntValues(firstOutVarId, vals)
evt.AddChangeEntry(chEntry)
}
}
func (this *WeightedSum) GetAllVars() []core.VarId {
allvars := make([]core.VarId, len(this.vars)+len(this.hvars)+1)
i := 0
for _, v := range this.vars {
allvars[i] = v
i++
}
for _, v := range this.hvars {
allvars[i] = v
i++
}
allvars[i] = this.resultVar
return allvars
}
func (this *WeightedSum) GetVarIds() []core.VarId {
return this.GetAllVars()
}
func (this *WeightedSum) | GetDomains | identifier_name | |
1505.js | --------------
var CITY = $.cookie('city');
var urlTemp = dxlHttp.m + "index/jsonpnew/index?act=hunShaList&callback=?&"; //瀑布流接口
var groupSubmit = $(".group.submit"); //筛选 确认按钮
var loadWrap = $(".loadWrap"); //加载提示
var allSellerSpan = $(".allSeller .selectShow span"); //全部商区 span
var allStyleSpan = $(".allStyle .selectShow span"); //全部商户 span
var smartSortSpan = $(".smartSort .selectShow span"); //智能排序 span
var par = {
region: allSellerSpan.attr("url"), //商区
feature: allStyleSpan.attr("url"), //酒店商户类型
city: $.cookie("city") //智能排序
}
var minPriceInput = $(".priceRange .dSpanInput").eq(0).find("input"); //最低价格输入框
var maxPriceInput = $(".priceRange .dSpanInput").eq(1).find("input"); //最高价格输入框
var priceRangeSpan = $(".priceRange span"); //价格tag元素
var jsonPrice = {}; //价格tag json
//记录cookie
$.cookie("weddingListUrl", window.location.href, {
expires: 7,
domain: "daoxila." + s4Com,
path: "/"
});
//点击指定区域 input失焦
$("#sellerSort").touchClick(function (e) {
if (e.target.tagName !== "INPUT") {
$("input").blur();
}
})
//把页面内的数据清除。
$("#sellerList ul").html("").css("display", 'block');
//关闭广告条
$(".banner .close").touchClick(function (e) {
e.stopPropagation();
e.preventDefault();
$(".banner").slideUp(500);
})
//底部app下载调用
$.wFootAppDown();
//返回顶部
$.mTopCall();
//搜索
$.mSearch(function () {
$.dxlGaPageTracker("/VPTracker/WapHunShaSheYing/Search");
});
//动态取值 排序区域宽高
sortUlWidth()
$(window).on("resize", sortUlWidth);
//全部商户 婚宴/婚纱/婚庆业务切换
$(".selectTag").touchClick(function () {
$(".allStyle").find(".active").removeClass("active");
$(this).parent().addClass("active");
});
//点击显示筛选条目
$(".selectShow").touchClick(function (e) {
//添加虚拟url
var parentIndex = $(this).parent().index(); //父级的索引
var gaUrl = "";
switch (parentIndex) {
case 0:
gaUrl = "/VPTracker/WapList/ShangQu";
break;
case 1:
gaUrl = "/VPTracker/WapList/YeWu";
break;
case 2:
gaUrl = "/VPTracker/WapList/PaiXu";
break;
case 3:
gaUrl = "/VPTracker/WapList/ShaiXuan";
break;
}
$.dxlGaPageTracker(gaUrl);
$(this).next(".popup").toggleClass("show")
.parent().siblings().find(".popup").removeClass("show");
$(this).find("span").toggleClass("cur")
.parents('.sortNav').siblings().find("span").removeClass("cur");
if ($(this).next(".popup").hasClass("show")) {
$("#sellerList,footer").css("display", "none");
$.dxlGaPageTracker('/VPTracker/WapHunShaSheYing/' + $.cookie("city") + '/ShangHu-filter?filter=' + $(this).data("type"));
if ($(this).data("type") == "4") {
groupSubmit.show();
} else {
groupSubmit.hide();
}
} else {
groupSubmit.hide();
$("input").blur();
$("#sellerList,footer").css("display", "block");
};
});
//折扣和优惠
$(".defaultSort .discounts span").touchClick(function (e) {
$(this).toggleClass("selected");
});
//价格区间(元)
$(".defaultSort .priceRange span").touchClick(function (e) {
$(this).toggleClass("selected").siblings().removeClass("selected");
});
//将价格tag存入json
priceRangeSpan.each(function (i, n) {
if ($(this).hasClass("dSpanInput")) return;
var _thisText = $(this).text();
jsonPrice[i] = _thisText.match(/下|上/) ? (_thisText.indexOf("下") != -1 ? "-" + parseInt(_thisText) : parseInt(_thisText) + "-") : _thisText;
})
// 滑动获取数据
$.dxlInclud(["fall"], function () {
$.isUrlPar("q") ? par.q = $.isUrlPar("q") : "";
$("#sellerSort").data("help") ? par.helpTag = $("#sellerSort").data("help") : "";
var option = {
"setting": {
url: urlTemp + $.param(par), //获取数据网址
tpl: tpl,
dom: "ul", //竖列容器元素标签(可省略)
selector: "#sellerList", //瀑布流容器
preLoad: false, //无图片或无需预加载时设为false, 默认为true(可省略)
imgUrl: "path",
initNum: 15, //初始化获取数量
newNum: 15, //每次新获取数量
watchHeight: 100 //页面离底部多远拉取数据,单位px(可省略)
},
"haddle": {
"onLoading": onLoading,
"onLoadingOk": loadingOk,
"onComplete": onComplete,
"onNoData": onNoData
}
};
$.dxlWaterFall(option);
//input聚焦 失焦
$(".priceRange input").focus(function () {
$(this).parent().addClass("selected").siblings(".dSpanInput").addClass("selected");
groupSubmit.css({
position: "absolute",
top: $(".popup.show").height() - 20 + "px"
});
}).blur(function () {
if ($(this).val() == "") {
$(this).parent().removeClass("selected").siblings(".dSpanInput").removeClass("selected");
};
groupSubmit.css({
position: "fixed",
top: ""
});
});
//价格筛选后赋值输入框
priceRangeSpan.touchClick(function () {
if ($(this).hasClass("dSpanInput")) return;
var minPrice,
maxPrice;
priceStr = $(this).text();
if (!priceStr) return;
if (priceStr.indexOf("-") != -1) {
minPrice = parseInt(priceStr.substring(0, priceStr.indexOf("-")));
maxPrice = parseInt(priceStr.substr(priceStr.indexOf("-") + 1));
}
if (priceStr.indexOf("以下") != -1) {
maxPrice = parseInt(priceStr);
}
if (priceStr.indexOf("以上") != -1) {
minPrice = parseInt(priceStr);
}
minPriceInput.val(minPrice);
maxPriceInput.val(maxPrice);
priceRangeSpan.find("input").removeClass("cur");
})
//价格区间输入框,匹配价格tag
$(".priceRange input").on("input", function () {
var minPriceInputVal = parseInt(minPriceInput.val()) || "";
var maxPriceInputVal = parseInt(maxPriceInput.val()) || "";
var priceStr = minPriceInputVal + "-" + maxPriceInputVal;
$(this).parent().parent().siblings().find("span").removeClass("selected");
$.each(jsonPrice, function (i, v) {
v == priceStr && priceRangeSpan.eq(i).addClass("selected");
})
})
//提交
$(".defaultSort .submit").touchClick(function (e) {
var _price = priceJudge();
//价格为空或价格左小右大符合条件
if (!_price || (!_price.unfill && _price[1] < _price[0])) {
$.mAlert("价格区间格式不正确,请重新输入");
return false;
}
//符合条件:
if ($.dxlWaterFall && $.isFunction($.dxlWaterFall)) {
$("#sellerList ul").empty();
$.dxlWaterFall("emptyData");
var group0 = $(".defaultSort .group").eq(0).find("span"),
regionVal = $.trim(allSellerSpan.attr("url")), //商区
featureVal = $.trim(allStyleSpan.attr("url")); //商户
var obj = {
city: $.cookie("city"),
min_price: _price[0],
max_price: _price[1],
region: region | $("footer").remove();
};
// --------------------- | conditional_block | |
1505.js | 框,匹配价格tag
$(".priceRange input").on("input", function () {
var minPriceInputVal = parseInt(minPriceInput.val()) || "";
var maxPriceInputVal = parseInt(maxPriceInput.val()) || "";
var priceStr = minPriceInputVal + "-" + maxPriceInputVal;
$(this).parent().parent().siblings().find("span").removeClass("selected");
$.each(jsonPrice, function (i, v) {
v == priceStr && priceRangeSpan.eq(i).addClass("selected");
})
})
//提交
$(".defaultSort .submit").touchClick(function (e) {
var _price = priceJudge();
//价格为空或价格左小右大符合条件
if (!_price || (!_price.unfill && _price[1] < _price[0])) {
$.mAlert("价格区间格式不正确,请重新输入");
return false;
}
//符合条件:
if ($.dxlWaterFall && $.isFunction($.dxlWaterFall)) {
$("#sellerList ul").empty();
$.dxlWaterFall("emptyData");
var group0 = $(".defaultSort .group").eq(0).find("span"),
regionVal = $.trim(allSellerSpan.attr("url")), //商区
featureVal = $.trim(allStyleSpan.attr("url")); //商户
var obj = {
city: $.cookie("city"),
min_price: _price[0],
max_price: _price[1],
region: regionVal == "" ? "" : regionVal, //商区
feature: featureVal == "" ? "" : featureVal, //商户
sort: smartSortSpan.attr("url"), //智能排序
libao: $(group0[0]).hasClass("selected") ? 1 : 0,
youhui: $(group0[1]).hasClass("selected") ? 1 : 0,
};
option.setting.url = urlTemp + $.param(obj);
$(".dataEmpty").remove();
$.dxlWaterFall(option);
sortItemHide();
}
});
//全城商区
$(".allSeller .popup li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
$(".allSeller .selectShow span").html(_thisText + '<i class="sortIconI"></i>');
$(".allSeller .selectShow span").attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "region");
});
//全部商户
$(".allStyle .popup div:nth-child(2)").find("li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
$(".allStyle .selectShow span").html(_thisText + '<i class="sortIconI"></i>');
$(".allStyle .selectShow span").attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "feature");
});
//智能排序
$(".smartSort .popup li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
smartSortSpan.html(_thisText + '<i class="sortIconI"></i>');
smartSortSpan.attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "sort");
})
//全城商区/全部商户请求接口时传参
function searchAjax(elem, pa) {
if ($.dxlWaterFall && $.isFunction($.dxlWaterFall)) {
var liHuiDiv = $(".defaultSort .group").eq(0);
$("#sellerList ul").empty();
$.dxlWaterFall("emptyData");
par[pa] = elem.attr("url") != "" ? elem.attr("url") : "";
par.min_price = minPriceInput.val(); //最低价格
par.max_price = maxPriceInput.val(); //最高价格
par.libao = liHuiDiv.find("span:nth-child(1)").hasClass("selected") ? 1 : 0; //礼包
par.youhui = liHuiDiv.find("span:nth-child(2)").hasClass("selected") ? 1 : 0; //优惠
par.sort = smartSortSpan.attr("url") //智能排序
option.setting.url = urlTemp + $.param(par);
$(".dataEmpty").remove();
$.dxlWaterFall(option);
}
sortItemHide();
}
});
//动态赋值 banner高/弹层宽高
function sortUlWidth() {
$("#sellerSort .sortNav .popup").css({
"width": $(window).width(),
"height": $(window).height() - 86
});
$(".banner img").height($(window).width() * 50 / 320);
}
//关闭排序条件总区域
function sortItemHide() {
setTimeout(function () {
$("input").blur();
groupSubmit.hide().css("position", "fixed");
$("#sellerSort .sortNav .popup").removeClass("show");
$("#sellerSort .sortNav span").removeClass("cur");
$("#sellerList,footer").css("display", "block");
}, 300);
}
//筛选弹层 价格区间判断
function priceJudge() {
var price = [];
$(".priceRange input").each(function (i) {
if (is_money_withzero($(this).val())) {
price[i] = Number($(this).val());
} else if ($(this).val() == "") {
price.unfill = true;
} else {
price = false;
return false;
}
});
return price;
}
//瀑布流设置对象生成器
function tpl(data, obj) {
var html = '<li>' +
'<a href="' + dxlHttp.m + CITY + '/HunShaSheYing/' + data.url + '-Info" class="sellerListItem">' +
'<dl class="clearfix">' +
'<dt><img src="' + data.image + '" ></dt>' +
'<dd>' +
'<div class="title">' +
'<h3>' +
'<i>' + data.name + '</i>' +
(data.xixuntong_status == 1 ? (data.coupon_daodianli_id ? '<span class="gift"></span>' : "") +
/*(data.price_back ? '<span class="fan"></span>':"") +*/
(data.coupon_putong_id ? '<span class="sale"></span>' : "") : '') +
(data.fu_flag ? '<span class="fu"></span>' : '') +
'</h3>' +
'</div>' +
'<div class="returnCash">' +
'<p class="row1">' +
'<span class="big">¥';
if (data.price_min == data.price_max) { //如果最低价格 和 最高价格一样,取其中一个
html += '<i>' + data.price_min + '</i>/套';
} else {
html += '<i>' + data.price_min + '-' + data.price_max + '</i>/套';
}
html += '</span>' +
/*(data.xixuntong_status==1 && data.price_back ? '<span class="spetial">'+ data.price_back +'</span>' : '') + */
//(data.xixuntong_status==1 ? '<span class="spetial">最高返'+ data.price_back +'</span>' : '') +
'</p>' +
'<p class="row2">' +
'<span class="district">' + data.region + '</span>' +
'<span class="style">' + data.features + '</span>' +
'</p>' +
'</div>' +
'</dd>' +
'</dl>' +
'</a>' +
'</li>';
obj.append(html);
}
//加载数据中处理
function onLoading() {
loadWrap.show();
}
//本次拉取成功处理
function loadingOk() {
loadWrap.hide();
}
//加载数据完成
function onComplete() {
loadWrap.html("没有更多了哦").show();
}
//首次获取数据量为0时
function onNoData() {
loadWrap.hide();
var str = '<div class="dataEmpty"><h3>没有找到合适的商户</h3><p>换个条件再试试吧</p></div>';
$("#sellerList").append(str);
}
//判断金额是否为正整数和0==为正整数和0返回true,不为正整数和0返回false
function is_money_withzero(value) {
var reg = /^[1-9]\d*$|^0$/;
return reg.test(value);
};
}) | identifier_body | ||
1505.js | ("display", 'block');
//关闭广告条
$(".banner .close").touchClick(function (e) {
e.stopPropagation();
e.preventDefault();
$(".banner").slideUp(500);
})
//底部app下载调用
$.wFootAppDown();
//返回顶部
$.mTopCall();
//搜索
$.mSearch(function () {
$.dxlGaPageTracker("/VPTracker/WapHunShaSheYing/Search");
});
|
//动态取值 排序区域宽高
sortUlWidth()
$(window).on("resize", sortUlWidth);
//全部商户 婚宴/婚纱/婚庆业务切换
$(".selectTag").touchClick(function () {
$(".allStyle").find(".active").removeClass("active");
$(this).parent().addClass("active");
});
//点击显示筛选条目
$(".selectShow").touchClick(function (e) {
//添加虚拟url
var parentIndex = $(this).parent().index(); //父级的索引
var gaUrl = "";
switch (parentIndex) {
case 0:
gaUrl = "/VPTracker/WapList/ShangQu";
break;
case 1:
gaUrl = "/VPTracker/WapList/YeWu";
break;
case 2:
gaUrl = "/VPTracker/WapList/PaiXu";
break;
case 3:
gaUrl = "/VPTracker/WapList/ShaiXuan";
break;
}
$.dxlGaPageTracker(gaUrl);
$(this).next(".popup").toggleClass("show")
.parent().siblings().find(".popup").removeClass("show");
$(this).find("span").toggleClass("cur")
.parents('.sortNav').siblings().find("span").removeClass("cur");
if ($(this).next(".popup").hasClass("show")) {
$("#sellerList,footer").css("display", "none");
$.dxlGaPageTracker('/VPTracker/WapHunShaSheYing/' + $.cookie("city") + '/ShangHu-filter?filter=' + $(this).data("type"));
if ($(this).data("type") == "4") {
groupSubmit.show();
} else {
groupSubmit.hide();
}
} else {
groupSubmit.hide();
$("input").blur();
$("#sellerList,footer").css("display", "block");
};
});
//折扣和优惠
$(".defaultSort .discounts span").touchClick(function (e) {
$(this).toggleClass("selected");
});
//价格区间(元)
$(".defaultSort .priceRange span").touchClick(function (e) {
$(this).toggleClass("selected").siblings().removeClass("selected");
});
//将价格tag存入json
priceRangeSpan.each(function (i, n) {
if ($(this).hasClass("dSpanInput")) return;
var _thisText = $(this).text();
jsonPrice[i] = _thisText.match(/下|上/) ? (_thisText.indexOf("下") != -1 ? "-" + parseInt(_thisText) : parseInt(_thisText) + "-") : _thisText;
})
// 滑动获取数据
$.dxlInclud(["fall"], function () {
$.isUrlPar("q") ? par.q = $.isUrlPar("q") : "";
$("#sellerSort").data("help") ? par.helpTag = $("#sellerSort").data("help") : "";
var option = {
"setting": {
url: urlTemp + $.param(par), //获取数据网址
tpl: tpl,
dom: "ul", //竖列容器元素标签(可省略)
selector: "#sellerList", //瀑布流容器
preLoad: false, //无图片或无需预加载时设为false, 默认为true(可省略)
imgUrl: "path",
initNum: 15, //初始化获取数量
newNum: 15, //每次新获取数量
watchHeight: 100 //页面离底部多远拉取数据,单位px(可省略)
},
"haddle": {
"onLoading": onLoading,
"onLoadingOk": loadingOk,
"onComplete": onComplete,
"onNoData": onNoData
}
};
$.dxlWaterFall(option);
//input聚焦 失焦
$(".priceRange input").focus(function () {
$(this).parent().addClass("selected").siblings(".dSpanInput").addClass("selected");
groupSubmit.css({
position: "absolute",
top: $(".popup.show").height() - 20 + "px"
});
}).blur(function () {
if ($(this).val() == "") {
$(this).parent().removeClass("selected").siblings(".dSpanInput").removeClass("selected");
};
groupSubmit.css({
position: "fixed",
top: ""
});
});
//价格筛选后赋值输入框
priceRangeSpan.touchClick(function () {
if ($(this).hasClass("dSpanInput")) return;
var minPrice,
maxPrice;
priceStr = $(this).text();
if (!priceStr) return;
if (priceStr.indexOf("-") != -1) {
minPrice = parseInt(priceStr.substring(0, priceStr.indexOf("-")));
maxPrice = parseInt(priceStr.substr(priceStr.indexOf("-") + 1));
}
if (priceStr.indexOf("以下") != -1) {
maxPrice = parseInt(priceStr);
}
if (priceStr.indexOf("以上") != -1) {
minPrice = parseInt(priceStr);
}
minPriceInput.val(minPrice);
maxPriceInput.val(maxPrice);
priceRangeSpan.find("input").removeClass("cur");
})
//价格区间输入框,匹配价格tag
$(".priceRange input").on("input", function () {
var minPriceInputVal = parseInt(minPriceInput.val()) || "";
var maxPriceInputVal = parseInt(maxPriceInput.val()) || "";
var priceStr = minPriceInputVal + "-" + maxPriceInputVal;
$(this).parent().parent().siblings().find("span").removeClass("selected");
$.each(jsonPrice, function (i, v) {
v == priceStr && priceRangeSpan.eq(i).addClass("selected");
})
})
//提交
$(".defaultSort .submit").touchClick(function (e) {
var _price = priceJudge();
//价格为空或价格左小右大符合条件
if (!_price || (!_price.unfill && _price[1] < _price[0])) {
$.mAlert("价格区间格式不正确,请重新输入");
return false;
}
//符合条件:
if ($.dxlWaterFall && $.isFunction($.dxlWaterFall)) {
$("#sellerList ul").empty();
$.dxlWaterFall("emptyData");
var group0 = $(".defaultSort .group").eq(0).find("span"),
regionVal = $.trim(allSellerSpan.attr("url")), //商区
featureVal = $.trim(allStyleSpan.attr("url")); //商户
var obj = {
city: $.cookie("city"),
min_price: _price[0],
max_price: _price[1],
region: regionVal == "" ? "" : regionVal, //商区
feature: featureVal == "" ? "" : featureVal, //商户
sort: smartSortSpan.attr("url"), //智能排序
libao: $(group0[0]).hasClass("selected") ? 1 : 0,
youhui: $(group0[1]).hasClass("selected") ? 1 : 0,
};
option.setting.url = urlTemp + $.param(obj);
$(".dataEmpty").remove();
$.dxlWaterFall(option);
sortItemHide();
}
});
//全城商区
$(".allSeller .popup li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
$(".allSeller .selectShow span").html(_thisText + '<i class="sortIconI"></i>');
$(".allSeller .selectShow span").attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "region");
});
//全部商户
$(".allStyle .popup div:nth-child(2)").find("li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
$(".allStyle .selectShow span").html(_thisText + '<i class="sortIconI"></i>');
$(".allStyle .selectShow span").attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "feature");
| random_line_split | |
1505.js | pl: tpl,
dom: "ul", //竖列容器元素标签(可省略)
selector: "#sellerList", //瀑布流容器
preLoad: false, //无图片或无需预加载时设为false, 默认为true(可省略)
imgUrl: "path",
initNum: 15, //初始化获取数量
newNum: 15, //每次新获取数量
watchHeight: 100 //页面离底部多远拉取数据,单位px(可省略)
},
"haddle": {
"onLoading": onLoading,
"onLoadingOk": loadingOk,
"onComplete": onComplete,
"onNoData": onNoData
}
};
$.dxlWaterFall(option);
//input聚焦 失焦
$(".priceRange input").focus(function () {
$(this).parent().addClass("selected").siblings(".dSpanInput").addClass("selected");
groupSubmit.css({
position: "absolute",
top: $(".popup.show").height() - 20 + "px"
});
}).blur(function () {
if ($(this).val() == "") {
$(this).parent().removeClass("selected").siblings(".dSpanInput").removeClass("selected");
};
groupSubmit.css({
position: "fixed",
top: ""
});
});
//价格筛选后赋值输入框
priceRangeSpan.touchClick(function () {
if ($(this).hasClass("dSpanInput")) return;
var minPrice,
maxPrice;
priceStr = $(this).text();
if (!priceStr) return;
if (priceStr.indexOf("-") != -1) {
minPrice = parseInt(priceStr.substring(0, priceStr.indexOf("-")));
maxPrice = parseInt(priceStr.substr(priceStr.indexOf("-") + 1));
}
if (priceStr.indexOf("以下") != -1) {
maxPrice = parseInt(priceStr);
}
if (priceStr.indexOf("以上") != -1) {
minPrice = parseInt(priceStr);
}
minPriceInput.val(minPrice);
maxPriceInput.val(maxPrice);
priceRangeSpan.find("input").removeClass("cur");
})
//价格区间输入框,匹配价格tag
$(".priceRange input").on("input", function () {
var minPriceInputVal = parseInt(minPriceInput.val()) || "";
var maxPriceInputVal = parseInt(maxPriceInput.val()) || "";
var priceStr = minPriceInputVal + "-" + maxPriceInputVal;
$(this).parent().parent().siblings().find("span").removeClass("selected");
$.each(jsonPrice, function (i, v) {
v == priceStr && priceRangeSpan.eq(i).addClass("selected");
})
})
//提交
$(".defaultSort .submit").touchClick(function (e) {
var _price = priceJudge();
//价格为空或价格左小右大符合条件
if (!_price || (!_price.unfill && _price[1] < _price[0])) {
$.mAlert("价格区间格式不正确,请重新输入");
return false;
}
//符合条件:
if ($.dxlWaterFall && $.isFunction($.dxlWaterFall)) {
$("#sellerList ul").empty();
$.dxlWaterFall("emptyData");
var group0 = $(".defaultSort .group").eq(0).find("span"),
regionVal = $.trim(allSellerSpan.attr("url")), //商区
featureVal = $.trim(allStyleSpan.attr("url")); //商户
var obj = {
city: $.cookie("city"),
min_price: _price[0],
max_price: _price[1],
region: regionVal == "" ? "" : regionVal, //商区
feature: featureVal == "" ? "" : featureVal, //商户
sort: smartSortSpan.attr("url"), //智能排序
libao: $(group0[0]).hasClass("selected") ? 1 : 0,
youhui: $(group0[1]).hasClass("selected") ? 1 : 0,
};
option.setting.url = urlTemp + $.param(obj);
$(".dataEmpty").remove();
$.dxlWaterFall(option);
sortItemHide();
}
});
//全城商区
$(".allSeller .popup li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
$(".allSeller .selectShow span").html(_thisText + '<i class="sortIconI"></i>');
$(".allSeller .selectShow span").attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "region");
});
//全部商户
$(".allStyle .popup div:nth-child(2)").find("li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
$(".allStyle .selectShow span").html(_thisText + '<i class="sortIconI"></i>');
$(".allStyle .selectShow span").attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "feature");
});
//智能排序
$(".smartSort .popup li").touchClick(function () {
var _thisText = $.trim($(this).text()),
_thisUrl = $.trim($(this).attr("url"));
smartSortSpan.html(_thisText + '<i class="sortIconI"></i>');
smartSortSpan.attr("url", _thisUrl);
$(this).addClass("selected").siblings().removeClass("selected");
searchAjax($(this), "sort");
})
//全城商区/全部商户请求接口时传参
function searchAjax(elem, pa) {
if ($.dxlWaterFall && $.isFunction($.dxlWaterFall)) {
var liHuiDiv = $(".defaultSort .group").eq(0);
$("#sellerList ul").empty();
$.dxlWaterFall("emptyData");
par[pa] = elem.attr("url") != "" ? elem.attr("url") : "";
par.min_price = minPriceInput.val(); //最低价格
par.max_price = maxPriceInput.val(); //最高价格
par.libao = liHuiDiv.find("span:nth-child(1)").hasClass("selected") ? 1 : 0; //礼包
par.youhui = liHuiDiv.find("span:nth-child(2)").hasClass("selected") ? 1 : 0; //优惠
par.sort = smartSortSpan.attr("url") //智能排序
option.setting.url = urlTemp + $.param(par);
$(".dataEmpty").remove();
$.dxlWaterFall(option);
}
sortItemHide();
}
});
//动态赋值 banner高/弹层宽高
function sortUlWidth() {
$("#sellerSort .sortNav .popup").css({
"width": $(window).width(),
"height": $(window).height() - 86
});
$(".banner img").height($(window).width() * 50 / 320);
}
//关闭排序条件总区域
function sortItemHide() {
setTimeout(function () {
$("input").blur();
groupSubmit.hide().css("position", "fixed");
$("#sellerSort .sortNav .popup").removeClass("show");
$("#sellerSort .sortNav span").removeClass("cur");
$("#sellerList,footer").css("display", "block");
}, 300);
}
//筛选弹层 价格区间判断
function priceJudge() {
var price = [];
$(".priceRange input").each(function (i) {
if (is_money_withzero($(this).val())) {
price[i] = Number($(this).val());
} else if ($(this).val() == "") {
price.unfill = true;
} else {
price = false;
return false;
}
});
return price;
}
//瀑布流设置对象生成器
function tpl(data, obj) {
var html = '<li>' +
'<a href="' + dxlHttp.m + CITY + '/HunShaSheYing/' + data.url + '-Info" class="sellerListItem">' +
'<dl class="clearfix">' +
'<dt><img src="' + data.image + '" ></dt>' +
'<dd>' +
'<div class="title">' +
'<h3>' +
'<i>' + data.name + '</i>' +
(data.xixuntong_status == 1 ? (data.coupon_daodianli_id ? '<span class="gift"></span>' : "") +
/*(data.price_back ? '<span class="fan"></span>':"") + | */
(d | identifier_name | |
udpopnet_random.py | range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k
link = copy.deepcopy(linkcopy)
#print link
for j in range(0, npop):
if(link[j]>maxdiameter):
maxdiameter=link[j]
#print "diameter: ", maxdiameter
# alternative could return diameter by uncommenting below
#return maxdiameter
#print "linearness: ", (float(maxdiameter)-1)/(float(npop)-2)
return ((float(maxdiameter)-1)/(float(npop)-2))
def eff(net, npop):
# special case for a network of 2 populations
if(npop == 2):
return 1
else:
# determine the distance between all popualtion pairs
invert=[]
for i in range(1, npop):
link = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
linkcopy = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
for k in range(1, npop):
for j in range(0, npop):
if (i != j):
if(link[j]==0):
if(net[i][j]==1):
linkcopy[j]=k
else:
for l in range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k
link = copy.deepcopy(linkcopy)
for m in range(len(link)):
if link[m]!=0:
invert.append(1/float(link[m]))
return float(sum(invert))/(float(npop)*(npop-1))
# check if the network is fully interconnected
def fullconnectcheck(net):
connected = [0 for x in range(nodes)]
connected[0]=1
current=0
for k in range(0, nodes):
for j in range(0, nodes):
current = j
if(connected[j] == 1):
for i in range(0, nodes): # walks through and finds connections to first node
if(i != current):
if(net[current][i]==1):
connected[i]=1
connectcount=0
for j in range(0, nodes):
if(connected[j]==1):
connectcount=connectcount+1
if(connectcount == nodes):
# fully connected
return 1
else:
# not connected
return 0
#Function for calculating dendrisity value.
def dend(net):
#Makes a list of the number of edges each node in the network has.
edgesum=[]
for i in net:
edgesum.append(sum(i))
#print "The number of edges at each node: %s" % edgesum
#Makes a list of the number of nodes that have three edges.
threecount=0
for i in edgesum:
if i==3:
threecount=threecount+1
#print "The number of nodes with three edges: %i" % threecount
edges=[] #Place-holder for list of all edges.
bridges=[] #Place-holder for list of all bridge edges.
nontermnodebridges=[] #Place-holder list for non-terminal bridge edges.
nontermbnodes=[] #Place-holder list for non-terminal nodes connected tobridges.
#Finds bridges in network by breaking connections and then checking connectivity of network, then checks if nodes and edges are non-terminal.
for i in range(0,nodes):
for j in range(0,nodes):
netp=copy.deepcopy(net)
if i!=j and netp[i][j]==netp[j][i]==1:
edges.append(1)
netp[i][j]=netp[j][i]=0
check=fullconnectcheck(netp)
if check==0:
bridges.append(1)
if sum(net[i])>=2:
nontermnodebridges.append(str(i)+str(j))
nontermbnodes.append(i)
else:
bridges.append(0)
#Gives the total number of non-terminal bridge edges.
#nontermbridge=len(nontermbridge)/2
def unique(seq):
checked=[]
for e in seq:
if e not in checked:
checked.append(e)
return checked
nontermbnodes=unique(nontermbnodes)
#Makes a list of non-terminal edges and checks if they are connected to each other.
bridgeconnect=[]
threebridgeconnect=[]
for i in nontermbnodes:
for j in nontermbnodes:
if i>=0 and j>=0 and i!=j and j>i:
if net[i][j]==net[j][i]==1:
bridgeconnect.append(str(i)+str(j))
if sum(net[i])==3:
threebridgeconnect.append(i)
totedge=(sum(edgesum)/2)
#print "Then total number of edges in the matrix: %i" % totedge
totbridge=sum(bridges)/2
#print "The total number of bridges in the network: %i" % totbridge
#Sums the total number of non-terminal edges.
#nontbedge=float(len(unique(bridgeconnect)))
#Checks if the bridges involving non-terminal nodes are to another non-terminal node.
nontermbridgeconnect=0
for i in range(0,len(bridgeconnect)):
if bridgeconnect[i] in nontermnodebridges:
nontermbridgeconnect+=1
#print "The number of non-terminal bridge edges in the network: %i" % nontermbridgeconnect
Totnontbridgeflankthree=float(len(unique(threebridgeconnect)))
#print "The number of non-terminal nodes in the network that flank nodes with three edges: %i" % Totnontbridgeflankthree
if nontermbridgeconnect!=float(0):
dend = Totnontbridgeflankthree/float(len(nontermbnodes))
else:
dend="0"
return dend
def pnextcalc(w,m,freq,popnet,immi):
# calculate avereage allele frequency after migration
pave = [0 for x in range(nodes)]
for i in range(0, nodes):
for j in range(0, nodes):
if i == j:
# normalization step if m is too large
if(immi[i]*m <= 1):
# p = global freq[j]
pave[i]=pave[i]+(1-immi[i]*m) * freq[j]
else:
pave[i]=0
elif popnet[i][j] == 1:
if(immi[i]*m <= 1):
pave[i] = pave[i] + m * freq[j]
else:
pave[i] = pave[i] + m * freq[j] / immi[i]*m
# calculate average fitness
wbar = [0 for x in range(nodes)]
for i in range(0, nodes):
wbar[i] = pave[i]*pave[i] + 2*w*pave[i]*(1-pave[i]) + (1-pave[i])*(1-pave[i])
# update frequencies with selection
pnext = [0 for x in range(nodes)]
for i in range(0, nodes):
pnext[i] = (pave[i]*pave[i] + w*pave[i]*(1-pave[i]))/wbar[i]
# boundry at zero and one
for i in range(0, nodes):
if(pnext[i]>1):
pnext[i]=1
if(pnext[i]<0):
pnext[i]=0
# store new values
for i in range(0, nodes):
freq[i]=pnext[i]
return freq
def diff(freq, pnext):
# calculate change
diff=0
for i in range(0, nodes):
diff=diff+abs(pnext[i]-freq[i])
return diff
# calculate distance in allele frequency between popualtions
def dist(freq):
distance=0
for i in range(0, nodes):
for j in range(0, nodes):
distance=distance+abs(freq[i]-freq[j])
return distance
def | (net,npop):
# what fraction of the nodes are terminal nodes
# these only have one connection in the rows
term = 0
for i in range(0, npop):
rowsum = 0
for j in range(0, npop):
rowsum = rowsum + net[i][j]
if(rowsum == 1):
term = term + 1
#print rowsum
#print term
return float(term)/float(npop)
for r in range(0, replicates):
if(noderange):
nodes = random.randrange(minnode,maxnode+1)
if(edgerange):
minedge = nodes-1 # minimum # edges possible
maxedge = nodes*minedge/2 | terminalness | identifier_name |
udpopnet_random.py | link = copy.deepcopy(linkcopy)
#print link
for j in range(0, npop):
if(link[j]>maxdiameter):
maxdiameter=link[j]
#print "diameter: ", maxdiameter
# alternative could return diameter by uncommenting below
#return maxdiameter
#print "linearness: ", (float(maxdiameter)-1)/(float(npop)-2)
return ((float(maxdiameter)-1)/(float(npop)-2))
def eff(net, npop):
# special case for a network of 2 populations
if(npop == 2):
return 1
else:
# determine the distance between all popualtion pairs
invert=[]
for i in range(1, npop):
link = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
linkcopy = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
for k in range(1, npop):
for j in range(0, npop):
if (i != j):
if(link[j]==0):
if(net[i][j]==1):
linkcopy[j]=k
else:
for l in range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k
link = copy.deepcopy(linkcopy)
for m in range(len(link)):
if link[m]!=0:
invert.append(1/float(link[m]))
return float(sum(invert))/(float(npop)*(npop-1))
# check if the network is fully interconnected
def fullconnectcheck(net):
connected = [0 for x in range(nodes)]
connected[0]=1
current=0
for k in range(0, nodes):
for j in range(0, nodes):
current = j
if(connected[j] == 1):
for i in range(0, nodes): # walks through and finds connections to first node
if(i != current):
if(net[current][i]==1):
connected[i]=1
connectcount=0
for j in range(0, nodes):
if(connected[j]==1):
connectcount=connectcount+1
if(connectcount == nodes):
# fully connected
return 1
else:
# not connected
return 0
#Function for calculating dendrisity value.
def dend(net):
#Makes a list of the number of edges each node in the network has.
edgesum=[]
for i in net:
edgesum.append(sum(i))
#print "The number of edges at each node: %s" % edgesum
#Makes a list of the number of nodes that have three edges.
threecount=0
for i in edgesum:
if i==3:
threecount=threecount+1
#print "The number of nodes with three edges: %i" % threecount
edges=[] #Place-holder for list of all edges.
bridges=[] #Place-holder for list of all bridge edges.
nontermnodebridges=[] #Place-holder list for non-terminal bridge edges.
nontermbnodes=[] #Place-holder list for non-terminal nodes connected tobridges.
#Finds bridges in network by breaking connections and then checking connectivity of network, then checks if nodes and edges are non-terminal.
for i in range(0,nodes):
for j in range(0,nodes):
netp=copy.deepcopy(net)
if i!=j and netp[i][j]==netp[j][i]==1:
edges.append(1)
netp[i][j]=netp[j][i]=0
check=fullconnectcheck(netp)
if check==0:
bridges.append(1)
if sum(net[i])>=2:
nontermnodebridges.append(str(i)+str(j))
nontermbnodes.append(i)
else:
bridges.append(0)
#Gives the total number of non-terminal bridge edges.
#nontermbridge=len(nontermbridge)/2
def unique(seq):
checked=[]
for e in seq:
if e not in checked:
checked.append(e)
return checked
nontermbnodes=unique(nontermbnodes)
#Makes a list of non-terminal edges and checks if they are connected to each other.
bridgeconnect=[]
threebridgeconnect=[]
for i in nontermbnodes:
for j in nontermbnodes:
if i>=0 and j>=0 and i!=j and j>i:
if net[i][j]==net[j][i]==1:
bridgeconnect.append(str(i)+str(j))
if sum(net[i])==3:
threebridgeconnect.append(i)
totedge=(sum(edgesum)/2)
#print "Then total number of edges in the matrix: %i" % totedge
totbridge=sum(bridges)/2
#print "The total number of bridges in the network: %i" % totbridge
#Sums the total number of non-terminal edges.
#nontbedge=float(len(unique(bridgeconnect)))
#Checks if the bridges involving non-terminal nodes are to another non-terminal node.
nontermbridgeconnect=0
for i in range(0,len(bridgeconnect)):
if bridgeconnect[i] in nontermnodebridges:
nontermbridgeconnect+=1
#print "The number of non-terminal bridge edges in the network: %i" % nontermbridgeconnect
Totnontbridgeflankthree=float(len(unique(threebridgeconnect)))
#print "The number of non-terminal nodes in the network that flank nodes with three edges: %i" % Totnontbridgeflankthree
if nontermbridgeconnect!=float(0):
dend = Totnontbridgeflankthree/float(len(nontermbnodes))
else:
dend="0"
return dend
def pnextcalc(w,m,freq,popnet,immi):
# calculate avereage allele frequency after migration
pave = [0 for x in range(nodes)]
for i in range(0, nodes):
for j in range(0, nodes):
if i == j:
# normalization step if m is too large
if(immi[i]*m <= 1):
# p = global freq[j]
pave[i]=pave[i]+(1-immi[i]*m) * freq[j]
else:
pave[i]=0
elif popnet[i][j] == 1:
if(immi[i]*m <= 1):
pave[i] = pave[i] + m * freq[j]
else:
pave[i] = pave[i] + m * freq[j] / immi[i]*m
# calculate average fitness
wbar = [0 for x in range(nodes)]
for i in range(0, nodes):
wbar[i] = pave[i]*pave[i] + 2*w*pave[i]*(1-pave[i]) + (1-pave[i])*(1-pave[i])
# update frequencies with selection
pnext = [0 for x in range(nodes)]
for i in range(0, nodes):
pnext[i] = (pave[i]*pave[i] + w*pave[i]*(1-pave[i]))/wbar[i]
# boundry at zero and one
for i in range(0, nodes):
if(pnext[i]>1):
pnext[i]=1
if(pnext[i]<0):
pnext[i]=0
# store new values
for i in range(0, nodes):
freq[i]=pnext[i]
return freq
def diff(freq, pnext):
# calculate change
diff=0
for i in range(0, nodes):
diff=diff+abs(pnext[i]-freq[i])
return diff
# calculate distance in allele frequency between popualtions
def dist(freq):
distance=0
for i in range(0, nodes | if(npop == 2):
return 1
else:
# determine the distance between all popualtion pairs
maxdiameter=0
for i in range(1, npop):
link = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
linkcopy = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
for k in range(1, npop):
for j in range(0, npop):
if (i != j):
if(link[j]==0):
if(net[i][j]==1):
linkcopy[j]=k
else:
for l in range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k | identifier_body | |
udpopnet_random.py | in range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k
link = copy.deepcopy(linkcopy)
#print link
for j in range(0, npop):
if(link[j]>maxdiameter):
maxdiameter=link[j]
#print "diameter: ", maxdiameter
# alternative could return diameter by uncommenting below
#return maxdiameter
#print "linearness: ", (float(maxdiameter)-1)/(float(npop)-2)
return ((float(maxdiameter)-1)/(float(npop)-2))
def eff(net, npop):
# special case for a network of 2 populations
if(npop == 2):
return 1
else:
# determine the distance between all popualtion pairs
invert=[]
for i in range(1, npop):
link = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
linkcopy = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
for k in range(1, npop):
for j in range(0, npop):
if (i != j):
if(link[j]==0):
if(net[i][j]==1):
linkcopy[j]=k
else:
for l in range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k
link = copy.deepcopy(linkcopy)
for m in range(len(link)):
if link[m]!=0:
invert.append(1/float(link[m]))
return float(sum(invert))/(float(npop)*(npop-1))
# check if the network is fully interconnected
def fullconnectcheck(net):
connected = [0 for x in range(nodes)]
connected[0]=1
current=0
for k in range(0, nodes):
for j in range(0, nodes):
current = j
if(connected[j] == 1):
for i in range(0, nodes): # walks through and finds connections to first node
if(i != current):
if(net[current][i]==1):
connected[i]=1
connectcount=0
for j in range(0, nodes):
if(connected[j]==1):
connectcount=connectcount+1
if(connectcount == nodes):
# fully connected
return 1
else:
# not connected
return 0
#Function for calculating dendrisity value.
def dend(net):
#Makes a list of the number of edges each node in the network has.
edgesum=[]
for i in net:
edgesum.append(sum(i))
#print "The number of edges at each node: %s" % edgesum
#Makes a list of the number of nodes that have three edges.
threecount=0
for i in edgesum:
if i==3:
threecount=threecount+1
#print "The number of nodes with three edges: %i" % threecount
edges=[] #Place-holder for list of all edges.
bridges=[] #Place-holder for list of all bridge edges.
nontermnodebridges=[] #Place-holder list for non-terminal bridge edges.
nontermbnodes=[] #Place-holder list for non-terminal nodes connected tobridges.
#Finds bridges in network by breaking connections and then checking connectivity of network, then checks if nodes and edges are non-terminal.
for i in range(0,nodes):
for j in range(0,nodes):
netp=copy.deepcopy(net)
if i!=j and netp[i][j]==netp[j][i]==1:
edges.append(1)
netp[i][j]=netp[j][i]=0
check=fullconnectcheck(netp)
if check==0:
bridges.append(1)
if sum(net[i])>=2:
nontermnodebridges.append(str(i)+str(j))
nontermbnodes.append(i)
else:
bridges.append(0)
#Gives the total number of non-terminal bridge edges.
#nontermbridge=len(nontermbridge)/2
def unique(seq):
checked=[]
for e in seq:
if e not in checked:
checked.append(e)
return checked
nontermbnodes=unique(nontermbnodes)
#Makes a list of non-terminal edges and checks if they are connected to each other.
bridgeconnect=[]
threebridgeconnect=[]
for i in nontermbnodes:
for j in nontermbnodes:
if i>=0 and j>=0 and i!=j and j>i:
if net[i][j]==net[j][i]==1:
bridgeconnect.append(str(i)+str(j))
if sum(net[i])==3:
threebridgeconnect.append(i)
totedge=(sum(edgesum)/2)
#print "Then total number of edges in the matrix: %i" % totedge
totbridge=sum(bridges)/2
#print "The total number of bridges in the network: %i" % totbridge
#Sums the total number of non-terminal edges.
#nontbedge=float(len(unique(bridgeconnect)))
#Checks if the bridges involving non-terminal nodes are to another non-terminal node.
nontermbridgeconnect=0
for i in range(0,len(bridgeconnect)):
if bridgeconnect[i] in nontermnodebridges:
nontermbridgeconnect+=1
#print "The number of non-terminal bridge edges in the network: %i" % nontermbridgeconnect
Totnontbridgeflankthree=float(len(unique(threebridgeconnect)))
#print "The number of non-terminal nodes in the network that flank nodes with three edges: %i" % Totnontbridgeflankthree
if nontermbridgeconnect!=float(0):
dend = Totnontbridgeflankthree/float(len(nontermbnodes))
else:
dend="0"
return dend
def pnextcalc(w,m,freq,popnet,immi):
# calculate avereage allele frequency after migration
pave = [0 for x in range(nodes)]
for i in range(0, nodes):
for j in range(0, nodes):
if i == j:
# normalization step if m is too large
if(immi[i]*m <= 1):
# p = global freq[j]
pave[i]=pave[i]+(1-immi[i]*m) * freq[j]
else:
pave[i]=0
elif popnet[i][j] == 1:
if(immi[i]*m <= 1):
pave[i] = pave[i] + m * freq[j]
else:
pave[i] = pave[i] + m * freq[j] / immi[i]*m
# calculate average fitness
wbar = [0 for x in range(nodes)]
for i in range(0, nodes):
wbar[i] = pave[i]*pave[i] + 2*w*pave[i]*(1-pave[i]) + (1-pave[i])*(1-pave[i])
# update frequencies with selection
pnext = [0 for x in range(nodes)]
for i in range(0, nodes):
pnext[i] = (pave[i]*pave[i] + w*pave[i]*(1-pave[i]))/wbar[i]
# boundry at zero and one
for i in range(0, nodes):
if(pnext[i]>1):
pnext[i]=1
if(pnext[i]<0):
pnext[i]=0
# store new values
for i in range(0, nodes):
freq[i]=pnext[i]
return freq
def diff(freq, pnext):
# calculate change
diff=0
for i in range(0, nodes):
diff=diff+abs(pnext[i]-freq[i]) | # calculate distance in allele frequency between popualtions
def dist(freq):
distance=0
for i in range(0, nodes):
for j in range(0, nodes):
distance=distance+abs(freq[i]-freq[j])
return distance
def terminalness(net,npop):
# what fraction of the nodes are terminal nodes
# these only have one connection in the rows
term = 0
for i in range(0, npop):
rowsum = 0
for j in range(0, npop):
rowsum = rowsum + net[i][j]
if(rowsum == 1):
term = term + 1
#print rowsum
#print term
return float(term)/float(npop)
for r in range(0, replicates):
if(noderange):
nodes = random.randrange(minnode,maxnode+1)
if(edgerange):
minedge = nodes-1 # minimum # edges possible
maxedge = nodes*minedge/2 | return diff
| random_line_split |
udpopnet_random.py | range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k
link = copy.deepcopy(linkcopy)
#print link
for j in range(0, npop):
if(link[j]>maxdiameter):
maxdiameter=link[j]
#print "diameter: ", maxdiameter
# alternative could return diameter by uncommenting below
#return maxdiameter
#print "linearness: ", (float(maxdiameter)-1)/(float(npop)-2)
return ((float(maxdiameter)-1)/(float(npop)-2))
def eff(net, npop):
# special case for a network of 2 populations
if(npop == 2):
return 1
else:
# determine the distance between all popualtion pairs
invert=[]
for i in range(1, npop):
link = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
linkcopy = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
for k in range(1, npop):
for j in range(0, npop):
if (i != j):
if(link[j]==0):
if(net[i][j]==1):
linkcopy[j]=k
else:
for l in range(0, npop):
if (i != l):
if(link[l]==0):
if(net[j][l]==1):
linkcopy[l]=k
link = copy.deepcopy(linkcopy)
for m in range(len(link)):
if link[m]!=0:
invert.append(1/float(link[m]))
return float(sum(invert))/(float(npop)*(npop-1))
# check if the network is fully interconnected
def fullconnectcheck(net):
connected = [0 for x in range(nodes)]
connected[0]=1
current=0
for k in range(0, nodes):
for j in range(0, nodes):
current = j
if(connected[j] == 1):
for i in range(0, nodes): # walks through and finds connections to first node
if(i != current):
if(net[current][i]==1):
connected[i]=1
connectcount=0
for j in range(0, nodes):
if(connected[j]==1):
connectcount=connectcount+1
if(connectcount == nodes):
# fully connected
return 1
else:
# not connected
return 0
#Function for calculating dendrisity value.
def dend(net):
#Makes a list of the number of edges each node in the network has.
edgesum=[]
for i in net:
edgesum.append(sum(i))
#print "The number of edges at each node: %s" % edgesum
#Makes a list of the number of nodes that have three edges.
threecount=0
for i in edgesum:
if i==3:
threecount=threecount+1
#print "The number of nodes with three edges: %i" % threecount
edges=[] #Place-holder for list of all edges.
bridges=[] #Place-holder for list of all bridge edges.
nontermnodebridges=[] #Place-holder list for non-terminal bridge edges.
nontermbnodes=[] #Place-holder list for non-terminal nodes connected tobridges.
#Finds bridges in network by breaking connections and then checking connectivity of network, then checks if nodes and edges are non-terminal.
for i in range(0,nodes):
for j in range(0,nodes):
netp=copy.deepcopy(net)
if i!=j and netp[i][j]==netp[j][i]==1:
edges.append(1)
netp[i][j]=netp[j][i]=0
check=fullconnectcheck(netp)
if check==0:
bridges.append(1)
if sum(net[i])>=2:
nontermnodebridges.append(str(i)+str(j))
nontermbnodes.append(i)
else:
bridges.append(0)
#Gives the total number of non-terminal bridge edges.
#nontermbridge=len(nontermbridge)/2
def unique(seq):
checked=[]
for e in seq:
if e not in checked:
checked.append(e)
return checked
nontermbnodes=unique(nontermbnodes)
#Makes a list of non-terminal edges and checks if they are connected to each other.
bridgeconnect=[]
threebridgeconnect=[]
for i in nontermbnodes:
for j in nontermbnodes:
if i>=0 and j>=0 and i!=j and j>i:
|
totedge=(sum(edgesum)/2)
#print "Then total number of edges in the matrix: %i" % totedge
totbridge=sum(bridges)/2
#print "The total number of bridges in the network: %i" % totbridge
#Sums the total number of non-terminal edges.
#nontbedge=float(len(unique(bridgeconnect)))
#Checks if the bridges involving non-terminal nodes are to another non-terminal node.
nontermbridgeconnect=0
for i in range(0,len(bridgeconnect)):
if bridgeconnect[i] in nontermnodebridges:
nontermbridgeconnect+=1
#print "The number of non-terminal bridge edges in the network: %i" % nontermbridgeconnect
Totnontbridgeflankthree=float(len(unique(threebridgeconnect)))
#print "The number of non-terminal nodes in the network that flank nodes with three edges: %i" % Totnontbridgeflankthree
if nontermbridgeconnect!=float(0):
dend = Totnontbridgeflankthree/float(len(nontermbnodes))
else:
dend="0"
return dend
def pnextcalc(w,m,freq,popnet,immi):
# calculate avereage allele frequency after migration
pave = [0 for x in range(nodes)]
for i in range(0, nodes):
for j in range(0, nodes):
if i == j:
# normalization step if m is too large
if(immi[i]*m <= 1):
# p = global freq[j]
pave[i]=pave[i]+(1-immi[i]*m) * freq[j]
else:
pave[i]=0
elif popnet[i][j] == 1:
if(immi[i]*m <= 1):
pave[i] = pave[i] + m * freq[j]
else:
pave[i] = pave[i] + m * freq[j] / immi[i]*m
# calculate average fitness
wbar = [0 for x in range(nodes)]
for i in range(0, nodes):
wbar[i] = pave[i]*pave[i] + 2*w*pave[i]*(1-pave[i]) + (1-pave[i])*(1-pave[i])
# update frequencies with selection
pnext = [0 for x in range(nodes)]
for i in range(0, nodes):
pnext[i] = (pave[i]*pave[i] + w*pave[i]*(1-pave[i]))/wbar[i]
# boundry at zero and one
for i in range(0, nodes):
if(pnext[i]>1):
pnext[i]=1
if(pnext[i]<0):
pnext[i]=0
# store new values
for i in range(0, nodes):
freq[i]=pnext[i]
return freq
def diff(freq, pnext):
# calculate change
diff=0
for i in range(0, nodes):
diff=diff+abs(pnext[i]-freq[i])
return diff
# calculate distance in allele frequency between popualtions
def dist(freq):
distance=0
for i in range(0, nodes):
for j in range(0, nodes):
distance=distance+abs(freq[i]-freq[j])
return distance
def terminalness(net,npop):
# what fraction of the nodes are terminal nodes
# these only have one connection in the rows
term = 0
for i in range(0, npop):
rowsum = 0
for j in range(0, npop):
rowsum = rowsum + net[i][j]
if(rowsum == 1):
term = term + 1
#print rowsum
#print term
return float(term)/float(npop)
for r in range(0, replicates):
if(noderange):
nodes = random.randrange(minnode,maxnode+1)
if(edgerange):
minedge = nodes-1 # minimum # edges possible
maxedge = nodes*minedge/2 | if net[i][j]==net[j][i]==1:
bridgeconnect.append(str(i)+str(j))
if sum(net[i])==3:
threebridgeconnect.append(i) | conditional_block |
transcribe.py | if not line.endswith(BR):
print "SKIP", line,
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
except ValueError:
if PRINT_PRE:
print >>o, line, BR,
else:
if PRINT_PRE:
print >>o, line, BR,
assert False
corrections = [
(' star war ', ' Star Wars '),
(' yzaca ', ' ycaza '),
]
icons = {
'Mr. Richardson': 'richardson.jpg',
'The Respondent': 'leif.png',
'The Witness': 'banks.jpg',
'A': 'banks.jpg',
'by mr. richardson:': 'richardson.jpg',
'by the court:': 'avatar.png',
'by the respondent:': 'leif.png',
}
examiner = None
class Dialogue:
def __init__(self, pageNo, lastLineNo, speaker):
self.pageNo = pageNo
self.lineNo = lastLineNo
speaker = speaker.lower().title()
self.speaker = speaker
self.lines = []
self.words = []
if speaker == 'Q':
self.icon = icons[examiner]
else:
self.icon = icons.get(speaker, 'avatar.png')
def addLine(self, line):
l = line.lower()
if l.startswith('by '):
global examiner
examiner = l
elif l == 'constantine evans,':
icons['The Witness'] = 'costi.png'
icons['A'] = 'costi.png'
for r in corrections:
l = l.replace(r[0], r[1])
self.lines.append(l)
self.words.extend(l.split())
def | (o, dialogue, pageNo, lastLineNo, speaker):
if dialogue:
printDialogue(o, dialogue)
dialogue = Dialogue(pageNo, lastLineNo, speaker)
return dialogue
def parsePageAnchor(anchor):
# e.g.,
# <A name=38></a>35<br>
# The first number is produced by the pdftohtml conversion;
# the second is the one printed in the top-right of
# the original PDF.
return int(anchor.split('>')[2].split('<')[0])
def parsePage(o, i, p, pageAnchor, dialogue, lastLineNo = None):
global examiner
print "page", p
prevLineNo = None
pageNo = parsePageAnchor(pageAnchor)
print >>o, '<a name=p%02d></a>' % (pageNo, )
for line in i:
# these are eaten by the outer loop
assert not line.startswith('<A ')
if line == HR:
break # end of page
# 16<br>
# THE COURT:<br>
# AS YOU CAN SEE, WE HAVE A BUSY<br>
# 17<br>
# CALENDAR THIS MORNING.<br>
# I HAVE A PRIORITY FOR MATTER<br>
# 18<br>
# NUMBER 13, WHICH IS CASTLE VERSUS STRAND.<br>
if not line.endswith(BR):
# HTML tags at begin/end
print "SKIP", line,
assert line[0] == '<'
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
if prevLineNo and lastLineNo == prevLineNo+1:
if dialogue and dialogue.speaker == 'Narrator':
printDialogue(o, dialogue)
dialogue = None
else:
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
prevLineNo = lastLineNo
except ValueError:
dialogue.addLine(line)
prevLineNo = None
elif line.endswith(':'):
if line.count(' ') < 2:
if dialogue is None or dialogue.speaker != 'Narrator':
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
else:
if line != 'THE COURT:':
dialogue.addLine(line)
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
prevLineNo = None
else:
if dialogue.speaker == 'Narrator':
dialogue.addLine(line)
printDialogue(o, dialogue)
dialogue = None
else:
dialogue.addLine(line)
prevLineNo = None
elif line in ('Q','A'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line)
elif line.startswith('(WHEREUPON,'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
else:
if not dialogue or (line == '///' and dialogue.speaker != 'Narrator'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
return dialogue
# from Pygments:
# <span class="lineno"> 6 </span><span class="p">{</span>
LINENO = '<span class="lineno">%2d </span>'
printLineNumbers = False # XXX: anchor
def printDialogue(o, dialogue):
if printLineNumbers:
print >>o, (LINENO % dialogue.lineNo),
print >>o, '<a name=p%02dl%02d></a>' % (dialogue.pageNo, dialogue.lineNo)
if dialogue.speaker == 'Narrator':
print >>o, '<p class="narration">'
else:
if dialogue.icon:
print >>o, ('<p class="spk"><img src="%s" width=48 height=48 align=bottom> %s</p><p class="dlg">' % (dialogue.icon, dialogue.speaker))
else:
print >>o, ('<p class="spk">%s</p><p class="dlg">' % dialogue.speaker)
if False:
for line in dialogue.lines:
print >>o, line,
else:
printWords(o, dialogue.words)
print >>o, '</p>'
return
properNames = set([
'god',
'castle', 'green', 'homeowners', 'association', "association's",
'leif', 'strand',
'dianne', 'patrizzi', "dianne's",
'randy', 'banks',
'kelly', 'richardson',
'constantine', 'evans',
'richard', 'ycaza',
'cathy', 'brown',
'i', "i'm", "i'll", "i've",
'mr.',
'association',
'treasurer', 'president', 'chairman',
'pasadena',
'south', 'el', 'molino', 'avenue',
'twitter',
'slack',
'wednesday', 'september', 'october',
])
from collections import deque
renter = deque()
renter.append('/baby/vm-2016-07-21.mov')
renter.append('/baby/vm-2016-07-22-1.mov')
renter.append('/baby/vm-2016-07-22-2.mov')
renter.append('/baby/vm-2016-08-04.mov')
links = {
'"slack."': '''"<a href="https://slack.com/">Slack</a>."''',
'newsletters': '<a href="/havisham/v2i1.pdf">newsletters</a>',
}
def printWords(o, words):
# "I was afraid of worms, Roxanne!"
newSentence = True
lastWord = None
for word in words:
if newSentence:
word = word.capitalize()
newSentence = False
if word in properNames or (word[-1] in '.,?!' and word[:-1] in properNames):
word = word.capitalize()
if 'pdro' in word:
# case number
word = word.upper()
if max(word.count('-'), word.count('.')) > 1:
# e.g., H.O.A. or B-A-N-K-S
# BUG: capitalizes 'MAN-TO-MAN', which is funny, so I left it in
word = word.upper()
if word[:6] == 'renter':
url = renter.popleft()
word = '<a href="%s">renter</a>%s' % (url, word[6:])
elif word in links:
word = links[word]
elif word == '///' and lastWord != '///':
print >>o, '<br>'
print >>o, word,
lastWord = word
if word == '///':
print >>o, '<br>'
if word[-1] in '.?!':
newSentence = True
return
def main():
i = open("input.html", "r")
o = open("index.html", "w");
print >>o, HEAD
print >>o, '<div class=" | switchDialogue | identifier_name |
transcribe.py |
if not line.endswith(BR):
print "SKIP", line,
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
except ValueError:
if PRINT_PRE:
print >>o, line, BR,
else:
if PRINT_PRE:
print >>o, line, BR,
assert False
corrections = [
(' star war ', ' Star Wars '),
(' yzaca ', ' ycaza '),
]
icons = {
'Mr. Richardson': 'richardson.jpg',
'The Respondent': 'leif.png',
'The Witness': 'banks.jpg',
'A': 'banks.jpg',
'by mr. richardson:': 'richardson.jpg',
'by the court:': 'avatar.png',
'by the respondent:': 'leif.png',
}
examiner = None
class Dialogue:
def __init__(self, pageNo, lastLineNo, speaker):
self.pageNo = pageNo
self.lineNo = lastLineNo
speaker = speaker.lower().title()
self.speaker = speaker
self.lines = []
self.words = []
if speaker == 'Q':
self.icon = icons[examiner]
else:
self.icon = icons.get(speaker, 'avatar.png')
def addLine(self, line):
l = line.lower()
if l.startswith('by '):
global examiner
examiner = l
elif l == 'constantine evans,':
icons['The Witness'] = 'costi.png'
icons['A'] = 'costi.png'
for r in corrections:
l = l.replace(r[0], r[1])
self.lines.append(l)
self.words.extend(l.split())
def switchDialogue(o, dialogue, pageNo, lastLineNo, speaker):
if dialogue:
printDialogue(o, dialogue)
dialogue = Dialogue(pageNo, lastLineNo, speaker)
return dialogue
def parsePageAnchor(anchor):
# e.g.,
# <A name=38></a>35<br>
# The first number is produced by the pdftohtml conversion;
# the second is the one printed in the top-right of
# the original PDF.
return int(anchor.split('>')[2].split('<')[0])
def parsePage(o, i, p, pageAnchor, dialogue, lastLineNo = None):
global examiner
print "page", p
prevLineNo = None
pageNo = parsePageAnchor(pageAnchor)
print >>o, '<a name=p%02d></a>' % (pageNo, )
for line in i:
# these are eaten by the outer loop
assert not line.startswith('<A ')
if line == HR:
break # end of page
# 16<br>
# THE COURT:<br>
# AS YOU CAN SEE, WE HAVE A BUSY<br>
# 17<br>
# CALENDAR THIS MORNING.<br>
# I HAVE A PRIORITY FOR MATTER<br>
# 18<br>
# NUMBER 13, WHICH IS CASTLE VERSUS STRAND.<br>
if not line.endswith(BR):
# HTML tags at begin/end
print "SKIP", line,
assert line[0] == '<'
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
if prevLineNo and lastLineNo == prevLineNo+1:
if dialogue and dialogue.speaker == 'Narrator':
printDialogue(o, dialogue)
dialogue = None
else:
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
prevLineNo = lastLineNo
except ValueError:
dialogue.addLine(line)
prevLineNo = None
elif line.endswith(':'):
if line.count(' ') < 2:
if dialogue is None or dialogue.speaker != 'Narrator':
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
else:
if line != 'THE COURT:':
dialogue.addLine(line)
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
prevLineNo = None
else:
if dialogue.speaker == 'Narrator':
dialogue.addLine(line)
printDialogue(o, dialogue)
dialogue = None
else:
dialogue.addLine(line)
prevLineNo = None
elif line in ('Q','A'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line)
elif line.startswith('(WHEREUPON,'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
else:
if not dialogue or (line == '///' and dialogue.speaker != 'Narrator'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
return dialogue
# from Pygments:
# <span class="lineno"> 6 </span><span class="p">{</span>
LINENO = '<span class="lineno">%2d </span>'
printLineNumbers = False # XXX: anchor
def printDialogue(o, dialogue):
if printLineNumbers:
print >>o, (LINENO % dialogue.lineNo),
print >>o, '<a name=p%02dl%02d></a>' % (dialogue.pageNo, dialogue.lineNo)
if dialogue.speaker == 'Narrator':
print >>o, '<p class="narration">'
else:
if dialogue.icon:
print >>o, ('<p class="spk"><img src="%s" width=48 height=48 align=bottom> %s</p><p class="dlg">' % (dialogue.icon, dialogue.speaker))
else:
print >>o, ('<p class="spk">%s</p><p class="dlg">' % dialogue.speaker)
if False:
for line in dialogue.lines:
print >>o, line,
else:
printWords(o, dialogue.words)
print >>o, '</p>'
return
properNames = set([
'god',
'castle', 'green', 'homeowners', 'association', "association's",
'leif', 'strand',
'dianne', 'patrizzi', "dianne's",
'randy', 'banks',
'kelly', 'richardson',
'constantine', 'evans',
'richard', 'ycaza',
'cathy', 'brown',
'i', "i'm", "i'll", "i've",
'mr.',
'association',
'treasurer', 'president', 'chairman',
'pasadena',
'south', 'el', 'molino', 'avenue',
'twitter',
'slack',
'wednesday', 'september', 'october',
])
from collections import deque
renter = deque()
renter.append('/baby/vm-2016-07-21.mov')
renter.append('/baby/vm-2016-07-22-1.mov')
renter.append('/baby/vm-2016-07-22-2.mov')
renter.append('/baby/vm-2016-08-04.mov')
links = {
'"slack."': '''"<a href="https://slack.com/">Slack</a>."''',
'newsletters': '<a href="/havisham/v2i1.pdf">newsletters</a>',
}
def printWords(o, words):
# "I was afraid of worms, Roxanne!"
newSentence = True
lastWord = None
for word in words:
if newSentence:
word = word.capitalize()
newSentence = False
if word in properNames or (word[-1] in '.,?!' and word[:-1] in properNames):
word = word.capitalize()
if 'pdro' in word:
# case number
word = word.upper()
if max(word.count('-'), word.count('.')) > 1:
# e.g., H.O.A. or B-A-N-K-S
# BUG: capitalizes 'MAN-TO-MAN', which is funny, so I left it in
word = word.upper()
if word[:6] == 'renter':
url = renter.popleft()
word = '<a href="%s">renter</a>%s' % (url, word[6:])
elif word in links:
word = links[word]
elif word == '///' and lastWord != '///':
print >>o, '<br>'
print >>o, word,
lastWord = word
if word == '///':
print >>o, '<br>'
if word[-1] in '.?!':
newSentence = True
return
def main():
i = open("input.html", "r")
o = open("index.html | pageAnchor = line
if PRINT_PRE:
print >>o, line
continue | conditional_block | |
transcribe.py | if not line.endswith(BR):
print "SKIP", line,
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
except ValueError:
if PRINT_PRE:
print >>o, line, BR,
else:
if PRINT_PRE:
print >>o, line, BR,
assert False
corrections = [
(' star war ', ' Star Wars '),
(' yzaca ', ' ycaza '),
]
icons = {
'Mr. Richardson': 'richardson.jpg',
'The Respondent': 'leif.png',
'The Witness': 'banks.jpg',
'A': 'banks.jpg',
'by mr. richardson:': 'richardson.jpg',
'by the court:': 'avatar.png',
'by the respondent:': 'leif.png',
}
examiner = None
class Dialogue:
def __init__(self, pageNo, lastLineNo, speaker):
self.pageNo = pageNo
self.lineNo = lastLineNo
speaker = speaker.lower().title()
self.speaker = speaker
self.lines = []
self.words = []
if speaker == 'Q':
self.icon = icons[examiner]
else:
self.icon = icons.get(speaker, 'avatar.png')
def addLine(self, line):
l = line.lower()
if l.startswith('by '):
global examiner
examiner = l
elif l == 'constantine evans,':
icons['The Witness'] = 'costi.png'
icons['A'] = 'costi.png'
for r in corrections:
l = l.replace(r[0], r[1])
self.lines.append(l)
self.words.extend(l.split())
def switchDialogue(o, dialogue, pageNo, lastLineNo, speaker):
if dialogue:
printDialogue(o, dialogue)
dialogue = Dialogue(pageNo, lastLineNo, speaker)
return dialogue
def parsePageAnchor(anchor):
# e.g.,
# <A name=38></a>35<br>
# The first number is produced by the pdftohtml conversion;
# the second is the one printed in the top-right of
# the original PDF.
|
def parsePage(o, i, p, pageAnchor, dialogue, lastLineNo = None):
global examiner
print "page", p
prevLineNo = None
pageNo = parsePageAnchor(pageAnchor)
print >>o, '<a name=p%02d></a>' % (pageNo, )
for line in i:
# these are eaten by the outer loop
assert not line.startswith('<A ')
if line == HR:
break # end of page
# 16<br>
# THE COURT:<br>
# AS YOU CAN SEE, WE HAVE A BUSY<br>
# 17<br>
# CALENDAR THIS MORNING.<br>
# I HAVE A PRIORITY FOR MATTER<br>
# 18<br>
# NUMBER 13, WHICH IS CASTLE VERSUS STRAND.<br>
if not line.endswith(BR):
# HTML tags at begin/end
print "SKIP", line,
assert line[0] == '<'
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
if prevLineNo and lastLineNo == prevLineNo+1:
if dialogue and dialogue.speaker == 'Narrator':
printDialogue(o, dialogue)
dialogue = None
else:
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
prevLineNo = lastLineNo
except ValueError:
dialogue.addLine(line)
prevLineNo = None
elif line.endswith(':'):
if line.count(' ') < 2:
if dialogue is None or dialogue.speaker != 'Narrator':
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
else:
if line != 'THE COURT:':
dialogue.addLine(line)
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
prevLineNo = None
else:
if dialogue.speaker == 'Narrator':
dialogue.addLine(line)
printDialogue(o, dialogue)
dialogue = None
else:
dialogue.addLine(line)
prevLineNo = None
elif line in ('Q','A'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line)
elif line.startswith('(WHEREUPON,'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
else:
if not dialogue or (line == '///' and dialogue.speaker != 'Narrator'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
return dialogue
# from Pygments:
# <span class="lineno"> 6 </span><span class="p">{</span>
LINENO = '<span class="lineno">%2d </span>'
printLineNumbers = False # XXX: anchor
def printDialogue(o, dialogue):
if printLineNumbers:
print >>o, (LINENO % dialogue.lineNo),
print >>o, '<a name=p%02dl%02d></a>' % (dialogue.pageNo, dialogue.lineNo)
if dialogue.speaker == 'Narrator':
print >>o, '<p class="narration">'
else:
if dialogue.icon:
print >>o, ('<p class="spk"><img src="%s" width=48 height=48 align=bottom> %s</p><p class="dlg">' % (dialogue.icon, dialogue.speaker))
else:
print >>o, ('<p class="spk">%s</p><p class="dlg">' % dialogue.speaker)
if False:
for line in dialogue.lines:
print >>o, line,
else:
printWords(o, dialogue.words)
print >>o, '</p>'
return
properNames = set([
'god',
'castle', 'green', 'homeowners', 'association', "association's",
'leif', 'strand',
'dianne', 'patrizzi', "dianne's",
'randy', 'banks',
'kelly', 'richardson',
'constantine', 'evans',
'richard', 'ycaza',
'cathy', 'brown',
'i', "i'm", "i'll", "i've",
'mr.',
'association',
'treasurer', 'president', 'chairman',
'pasadena',
'south', 'el', 'molino', 'avenue',
'twitter',
'slack',
'wednesday', 'september', 'october',
])
from collections import deque
renter = deque()
renter.append('/baby/vm-2016-07-21.mov')
renter.append('/baby/vm-2016-07-22-1.mov')
renter.append('/baby/vm-2016-07-22-2.mov')
renter.append('/baby/vm-2016-08-04.mov')
links = {
'"slack."': '''"<a href="https://slack.com/">Slack</a>."''',
'newsletters': '<a href="/havisham/v2i1.pdf">newsletters</a>',
}
def printWords(o, words):
# "I was afraid of worms, Roxanne!"
newSentence = True
lastWord = None
for word in words:
if newSentence:
word = word.capitalize()
newSentence = False
if word in properNames or (word[-1] in '.,?!' and word[:-1] in properNames):
word = word.capitalize()
if 'pdro' in word:
# case number
word = word.upper()
if max(word.count('-'), word.count('.')) > 1:
# e.g., H.O.A. or B-A-N-K-S
# BUG: capitalizes 'MAN-TO-MAN', which is funny, so I left it in
word = word.upper()
if word[:6] == 'renter':
url = renter.popleft()
word = '<a href="%s">renter</a>%s' % (url, word[6:])
elif word in links:
word = links[word]
elif word == '///' and lastWord != '///':
print >>o, '<br>'
print >>o, word,
lastWord = word
if word == '///':
print >>o, '<br>'
if word[-1] in '.?!':
newSentence = True
return
def main():
i = open("input.html", "r")
o = open("index.html", "w");
print >>o, HEAD
print >>o, '<div class="trans | return int(anchor.split('>')[2].split('<')[0]) | identifier_body |
transcribe.py | if not line.endswith(BR):
print "SKIP", line,
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
except ValueError:
if PRINT_PRE:
print >>o, line, BR,
else:
if PRINT_PRE:
print >>o, line, BR,
assert False
corrections = [
(' star war ', ' Star Wars '),
(' yzaca ', ' ycaza '),
]
icons = {
'Mr. Richardson': 'richardson.jpg',
'The Respondent': 'leif.png',
'The Witness': 'banks.jpg',
'A': 'banks.jpg',
'by mr. richardson:': 'richardson.jpg',
'by the court:': 'avatar.png',
'by the respondent:': 'leif.png',
}
examiner = None
class Dialogue:
def __init__(self, pageNo, lastLineNo, speaker):
self.pageNo = pageNo
self.lineNo = lastLineNo
speaker = speaker.lower().title()
self.speaker = speaker
self.lines = []
self.words = []
if speaker == 'Q':
self.icon = icons[examiner]
else:
self.icon = icons.get(speaker, 'avatar.png')
def addLine(self, line):
l = line.lower()
if l.startswith('by '):
global examiner
examiner = l
elif l == 'constantine evans,':
icons['The Witness'] = 'costi.png'
icons['A'] = 'costi.png'
for r in corrections:
l = l.replace(r[0], r[1])
self.lines.append(l)
self.words.extend(l.split())
def switchDialogue(o, dialogue, pageNo, lastLineNo, speaker):
if dialogue:
printDialogue(o, dialogue)
dialogue = Dialogue(pageNo, lastLineNo, speaker)
return dialogue
def parsePageAnchor(anchor):
# e.g.,
# <A name=38></a>35<br>
# The first number is produced by the pdftohtml conversion;
# the second is the one printed in the top-right of
# the original PDF.
return int(anchor.split('>')[2].split('<')[0])
def parsePage(o, i, p, pageAnchor, dialogue, lastLineNo = None):
global examiner
print "page", p
prevLineNo = None
pageNo = parsePageAnchor(pageAnchor)
print >>o, '<a name=p%02d></a>' % (pageNo, )
for line in i:
# these are eaten by the outer loop
assert not line.startswith('<A ')
if line == HR:
break # end of page
# 16<br>
# THE COURT:<br>
# AS YOU CAN SEE, WE HAVE A BUSY<br>
# 17<br>
# CALENDAR THIS MORNING.<br>
# I HAVE A PRIORITY FOR MATTER<br>
# 18<br>
# NUMBER 13, WHICH IS CASTLE VERSUS STRAND.<br>
if not line.endswith(BR):
# HTML tags at begin/end
print "SKIP", line,
assert line[0] == '<'
continue
line = line[0:-len(BR)]
if line[0].isdigit():
try:
lastLineNo = int(line)
if prevLineNo and lastLineNo == prevLineNo+1:
if dialogue and dialogue.speaker == 'Narrator':
printDialogue(o, dialogue)
dialogue = None
else:
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
prevLineNo = lastLineNo
except ValueError:
dialogue.addLine(line)
prevLineNo = None | if line.count(' ') < 2:
if dialogue is None or dialogue.speaker != 'Narrator':
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
else:
if line != 'THE COURT:':
dialogue.addLine(line)
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line[0:-1])
prevLineNo = None
else:
if dialogue.speaker == 'Narrator':
dialogue.addLine(line)
printDialogue(o, dialogue)
dialogue = None
else:
dialogue.addLine(line)
prevLineNo = None
elif line in ('Q','A'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker=line)
elif line.startswith('(WHEREUPON,'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
else:
if not dialogue or (line == '///' and dialogue.speaker != 'Narrator'):
dialogue = switchDialogue(o, dialogue, pageNo, lastLineNo, speaker='Narrator')
dialogue.addLine(line)
prevLineNo = None
return dialogue
# from Pygments:
# <span class="lineno"> 6 </span><span class="p">{</span>
LINENO = '<span class="lineno">%2d </span>'
printLineNumbers = False # XXX: anchor
def printDialogue(o, dialogue):
if printLineNumbers:
print >>o, (LINENO % dialogue.lineNo),
print >>o, '<a name=p%02dl%02d></a>' % (dialogue.pageNo, dialogue.lineNo)
if dialogue.speaker == 'Narrator':
print >>o, '<p class="narration">'
else:
if dialogue.icon:
print >>o, ('<p class="spk"><img src="%s" width=48 height=48 align=bottom> %s</p><p class="dlg">' % (dialogue.icon, dialogue.speaker))
else:
print >>o, ('<p class="spk">%s</p><p class="dlg">' % dialogue.speaker)
if False:
for line in dialogue.lines:
print >>o, line,
else:
printWords(o, dialogue.words)
print >>o, '</p>'
return
properNames = set([
'god',
'castle', 'green', 'homeowners', 'association', "association's",
'leif', 'strand',
'dianne', 'patrizzi', "dianne's",
'randy', 'banks',
'kelly', 'richardson',
'constantine', 'evans',
'richard', 'ycaza',
'cathy', 'brown',
'i', "i'm", "i'll", "i've",
'mr.',
'association',
'treasurer', 'president', 'chairman',
'pasadena',
'south', 'el', 'molino', 'avenue',
'twitter',
'slack',
'wednesday', 'september', 'october',
])
from collections import deque
renter = deque()
renter.append('/baby/vm-2016-07-21.mov')
renter.append('/baby/vm-2016-07-22-1.mov')
renter.append('/baby/vm-2016-07-22-2.mov')
renter.append('/baby/vm-2016-08-04.mov')
links = {
'"slack."': '''"<a href="https://slack.com/">Slack</a>."''',
'newsletters': '<a href="/havisham/v2i1.pdf">newsletters</a>',
}
def printWords(o, words):
# "I was afraid of worms, Roxanne!"
newSentence = True
lastWord = None
for word in words:
if newSentence:
word = word.capitalize()
newSentence = False
if word in properNames or (word[-1] in '.,?!' and word[:-1] in properNames):
word = word.capitalize()
if 'pdro' in word:
# case number
word = word.upper()
if max(word.count('-'), word.count('.')) > 1:
# e.g., H.O.A. or B-A-N-K-S
# BUG: capitalizes 'MAN-TO-MAN', which is funny, so I left it in
word = word.upper()
if word[:6] == 'renter':
url = renter.popleft()
word = '<a href="%s">renter</a>%s' % (url, word[6:])
elif word in links:
word = links[word]
elif word == '///' and lastWord != '///':
print >>o, '<br>'
print >>o, word,
lastWord = word
if word == '///':
print >>o, '<br>'
if word[-1] in '.?!':
newSentence = True
return
def main():
i = open("input.html", "r")
o = open("index.html", "w");
print >>o, HEAD
print >>o, '<div class="trans |
elif line.endswith(':'): | random_line_split |
order.go | NoStorageException, _ := logic.VerifyGoodsNum(goodsIdMap, goodsList)
// 获取 goodsId 信息
shipTimeList := []string{"XX", "XX"}
// shipTimeList := logic.GetShipTime()
// 获取用户所有地址
myAddressList, err := daoSql.GetAddressListByUid(uid, false)
if nil != err && RecordEmpty != err {
// log
return util.Fail(ctx, 10, err.Error())
}
var myAddress *daoSql.Address
for idx, addressItem := range myAddressList {
// 默认取第一个
if 0 == idx {
myAddress = addressItem
}
// 取默认地址
if uint8(1) == addressItem.IsDefault {
myAddress = addressItem
}
}
if nil == myAddress {
myAddress = &daoSql.Address{}
} else {
myAddress.IsDefault = 1
}
// 读入配置信息
orderConf, err := daoConf.OrderConf()
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
// 生成预处理订单
orderMap := logic.OrderMap{
Address: myAddress,
GoodsIdMap: goodsIdMap,
GoodsList: goodsList,
}
orderInfo, orderGoodsList, err := logic.GenOrder(uid, orderMap)
if nil != err {
// log
}
// 过滤订单参数
orderInfo.Filter()
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(orderGoodsList))
for idx, item := range orderGoodsList {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
// 拼装接口数据
orderData := apiIndex.Order{
Address: (*apiIndex.AddressType)(myAddress),
ShipTimeList: shipTimeList,
OrderInfo: apiIndex.OrderInfo{
GoodsList: arrApiOrderGoods,
Order: &apiIndex.OrderBase{Order: orderInfo},
},
}
if 0 < len(goodsException) {
orderData.Alert = fmt.Sprintf(orderConf.Alert, goodsException)
} else if 0 < len(goodsNoStorageException) {
orderData.Alert = fmt.Sprintf(orderConf.StorageAlert, goodsNoStorageException)
} else {
orderData.Alert = ""
}
// 格式化地址列表
for _, addressItem := range myAddressList {
orderData.AddressList = append(orderData.AddressList, (*apiIndex.AddressType)(addressItem))
}
orderData.Format()
return util.Success(ctx, orderData) | }
// 确认信息页 form-> goods_list:[{"goods_id":"3","selected":"1","goods_num":"2"}]
func (OrderController) DoOrder(ctx echo.Context) error {
// 设置 redis key
uid := ctx.Get("uid").(uint64)
// 避免同一个订单重复提交
// data, _ := json.Marshal(ctx.Request())
// curOrderMd5 = fmt.Printf("%x", md5.Sum(data))
// preOrderMd5 := daoRedis.NewRedisClient().Key(daoRedis.KeyOrder, util.Itoa(uid)).GET("")
// if preOrder == preOrderMd5 {
// // log
// return util.Fail(ctx, 10, RepeatDoOrder)
// }
// daoRedis.NewRedisClient().SET("", curOrderMd5, 30)
// 获取购物车商品列表
goodsList, err := getCartGoodsList(ctx)
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
// 收集 goodsId
goodsIdList := []uint64{}
for _, goodsInfo := range goodsList {
if 1 == goodsInfo.Selected {
if 0 < goodsInfo.GoodsNum {
goodsIdList = append(goodsIdList, goodsInfo.GoodsId)
}
}
}
// 获取 商品详情
goodsException, goodsIdMap, _ := logic.GetCartInfo(goodsIdList)
// 验证并修正库存信息 如果只有3个,购买5个,会强制改为3个
goodsNoStorageException, _ := logic.VerifyGoodsNum(goodsIdMap, goodsList)
// 获取地址信息
address, err := fetchAddress(ctx)
if nil != err {
return util.Fail(ctx, 10, "地址信息无效")
}
// 读入配置信息
orderConf, err := daoConf.OrderConf()
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
if 0 < len(goodsException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.Alert, goodsException))
}
if 0 < len(goodsNoStorageException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.StorageAlert, goodsNoStorageException))
}
// 提交订单
orderMap := logic.OrderMap{
Address: address,
GoodsIdMap: goodsIdMap,
GoodsList: goodsList,
ExceptTime: time.Now().Unix(),
OrderMessage: ctx.FormValue("order_message"),
}
orderInfo, orderGoodsList, err := logic.SubmitOrder(uid, orderMap)
if nil != err {
// log
}
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(orderGoodsList))
for idx, item := range orderGoodsList {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
// 拼装接口数据
orderData := apiIndex.Order{
Address: (*apiIndex.AddressType)(address),
ShipTimeList: []string{},
OrderInfo: apiIndex.OrderInfo{
GoodsList: arrApiOrderGoods,
Order: &apiIndex.OrderBase{Order: orderInfo},
},
Cancel: genApiCancel(orderInfo),
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) MyOrderList(ctx echo.Context) error {
// 获取订单列表信息
// uid, base_id, rn
myOrderMapList, hasMore, err := logic.GetMyOrderList(10, 0, 20)
if nil != err {
return err
}
// 拼装接口数据
orderData := &apiIndex.OrderList{
HasMore: hasMore,
}
// orderData
for _, v := range myOrderMapList {
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(v["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range v["goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData.List = append(orderData.List, &apiIndex.Order{
Address: (*apiIndex.AddressType)(v["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order: &apiIndex.OrderBase{Order: v["order"].(*daoSql.Order)},
GoodsList: arrApiOrderGoods,
},
})
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) CancelOrder(ctx echo.Context) error {
// 获取订单列表信息
uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
cancelFlag := util.Atoi(ctx.FormValue("cancel_flag"), 16, false).(uint16)
err := logic.CancelOrder(uid, orderSn, cancelFlag)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) EvalOrder(ctx echo.Context) error {
// 获取订单列表信息
uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
stars := util.Atoi(ctx.FormValue("stars"), 8, false).(uint8)
feedback := ctx.FormValue("feedback")
err := logic.EvalOrder(uid, orderSn, stars, feedback)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) Detail(ctx echo.Context) error {
ordeSn := ctx.QueryParam("order_sn")
// 获取订单列表信息
// uid, orderSn
myOrderMap, err := logic.GetOrderDetail(10, ordeSn)
if nil != err {
return err
}
// 拼装接口数据
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(myOrderMap["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range myOrderMap["goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData := &apiIndex.Order{
Alert: "",
Address: (*apiIndex.AddressType)(myOrderMap["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order: | random_line_split | |
order.go | StorageException, _ := logic.VerifyGoodsNum(goodsIdMap, goodsList)
// 获取 goodsId 信息
shipTimeList := []string{"XX", "XX"}
// shipTimeList := logic.GetShipTime()
// 获取用户所有地址
myAddressList, err := daoSql.GetAddressListByUid(uid, false)
if nil != err && RecordEmpty != err {
// log
return util.Fail(ctx, 10, err.Error())
}
var myAddress *daoSql.Address
for idx, addressItem := range myAddressList {
// 默认取第一个
if 0 == idx {
myAddress = addressItem
}
// 取默认地址
if uint8(1) == addressItem.IsDefault {
myAddress = addressItem
}
}
if nil == myAddress {
myAddress = &daoSql.Address{}
} else {
myAddress.IsDefault = 1
}
// 读入配置信息
orderConf, err := daoConf.OrderConf()
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
// 生成预处理订单
orderMap := logic.OrderMap{
Address: myAddress,
GoodsIdMap: goodsIdMap,
GoodsList: goodsList,
}
orderInfo, orderGoodsList, err := logic.GenOrder(uid, orderMap)
if nil != err {
// log
}
// 过滤订单参数
orderInfo.Filter()
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(orderGoodsList))
for idx, item := range orderGoodsList {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
// 拼装接口数据
orderData := apiIndex.Order{
Address: (*apiIndex.AddressType)(myAddress),
ShipTimeList: shipTimeList,
OrderInfo: apiIndex.OrderInfo{
GoodsList: arrApiOrderGoods,
Order: &apiIndex.OrderBase{Order: orderInfo},
},
}
if 0 < len(goodsException) {
orderData.Alert = fmt.Sprintf(orderConf.Alert, goodsException)
} else if 0 < len(goodsNoStorageException) {
orderData.Alert = fmt.Sprintf(orderConf.StorageAlert, goodsNoStorageException)
} else {
orderData.Alert = ""
}
// 格式化地址列表
for _, addressItem := range myAddressList {
orderData.AddressList = append(orderData.AddressList, (*apiIndex.AddressType)(addressItem))
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 确认信息页 form-> goods_list:[{"goods_id":"3","selected":"1","goods_num":"2"}]
func (OrderController) DoOrder(ctx echo.Context) error {
// 设置 redis key
uid := ctx.Get("uid").(uint64)
// 避免同一个订单重复提交
// data, _ := json.Marshal(ctx.Request())
// curOrderMd5 = fmt.Printf("%x", md5.Sum(data))
// preOrderMd5 := daoRedis.NewRedisClient().Key(daoRedis.KeyOrder, util.Itoa(uid)).GET("")
// if preOrder == preOrderMd5 {
// // log
// return util.Fail(ctx, 10, RepeatDoOrder)
// }
// daoRedis.NewRedisClient().SET("", curOrderMd5, 30)
// 获取购物车商品列表
goodsList, err := getCartGoodsList(ctx)
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
// 收集 goodsId
goodsIdList := []uint64{}
for _, goodsInfo := range goodsList {
if 1 == goodsInfo.Selected {
if 0 < goodsInfo.GoodsNum {
goodsIdList = append(goodsIdList, goodsInfo.GoodsId)
}
}
}
// 获取 商品详情
goodsException, goodsIdMap, _ := logic.GetCartInfo(goodsIdList)
// 验证并修正库存信息 如果只有3个,购买5个,会强制改为3个
goodsNoStorageException, _ := logic.VerifyGoodsNum(goodsIdMap, goodsList)
// 获取地址信息
address, err := fetchAddress(ctx)
if nil != err {
return util.Fail(ctx, 10, "地址信息无效")
}
// 读入配置信息
orderConf, err := daoConf.OrderConf()
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
if 0 < len(goodsException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.Alert, goodsException))
}
if 0 < len(goodsNoStorageException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.StorageAlert, goodsNoStorageException))
}
// 提交订单
orderMap := logic.OrderMap{
Address: address,
GoodsIdMap: goodsIdMap,
GoodsList: goodsList,
ExceptTime: time.Now().Unix(),
OrderMessage: ctx.FormValue("order_message"),
}
orderInfo, orderGoodsList, err := logic.SubmitOrder(uid, orderMap)
if nil != err {
// log
}
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(orderGoodsList))
for idx, item := range orderGoodsList {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
// 拼装接口数据
orderData := apiIndex.Order{
Address: (*apiIndex.AddressType)(address),
ShipTimeList: []string{},
OrderInfo: apiIndex.OrderInfo{
GoodsList: arrApiOrderGoods,
Order: &apiIndex.OrderBase{Order: orderInfo},
},
Cancel: genApiCancel(orderInfo),
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) MyOrderList(ctx echo.Context) error {
// 获取订单列表信息
// uid, base_id, rn
myOrderMapList, hasMore, err := logic.GetMyOrderList(10, 0, 20)
if nil != err {
return err
}
// 拼装接口数据
orderData := &apiIndex.OrderList{
HasMore: hasMore,
}
// orderData
for _, v := range myOrderMapList {
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(v["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range v[ | uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
cancelFlag := util.Atoi(ctx.FormValue("cancel_flag"), 16, false).(uint16)
err := logic.CancelOrder(uid, orderSn, cancelFlag)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) EvalOrd
er(ctx echo.Context) error {
// 获取订单列表信息
uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
stars := util.Atoi(ctx.FormValue("stars"), 8, false).(uint8)
feedback := ctx.FormValue("feedback")
err := logic.EvalOrder(uid, orderSn, stars, feedback)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) Detail(ctx echo.Context) error {
ordeSn := ctx.QueryParam("order_sn")
// 获取订单列表信息
// uid, orderSn
myOrderMap, err := logic.GetOrderDetail(10, ordeSn)
if nil != err {
return err
}
// 拼装接口数据
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(myOrderMap["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range myOrderMap["goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData := &apiIndex.Order{
Alert: "",
Address: (*apiIndex.AddressType)(myOrderMap["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order | "goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData.List = append(orderData.List, &apiIndex.Order{
Address: (*apiIndex.AddressType)(v["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order: &apiIndex.OrderBase{Order: v["order"].(*daoSql.Order)},
GoodsList: arrApiOrderGoods,
},
})
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) CancelOrder(ctx echo.Context) error {
// 获取订单列表信息 | identifier_body |
order.go | = append(goodsIdList, goodsInfo.GoodsId)
}
}
}
// 获取 商品详情
goodsException, goodsIdMap, _ := logic.GetCartInfo(goodsIdList)
// 验证并修正库存信息 如果只有3个,购买5个,会强制改为3个
goodsNoStorageException, _ := logic.VerifyGoodsNum(goodsIdMap, goodsList)
// 获取地址信息
address, err := fetchAddress(ctx)
if nil != err {
return util.Fail(ctx, 10, "地址信息无效")
}
// 读入配置信息
orderConf, err := daoConf.OrderConf()
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
if 0 < len(goodsException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.Alert, goodsException))
}
if 0 < len(goodsNoStorageException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.StorageAlert, goodsNoStorageException))
}
// 提交订单
orderMap := logic.OrderMap{
Address: address,
GoodsIdMap: goodsIdMap,
GoodsList: goodsList,
ExceptTime: time.Now().Unix(),
OrderMessage: ctx.FormValue("order_message"),
}
orderInfo, orderGoodsList, err := logic.SubmitOrder(uid, orderMap)
if nil != err {
// log
}
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(orderGoodsList))
for idx, item := range orderGoodsList {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
// 拼装接口数据
orderData := apiIndex.Order{
Address: (*apiIndex.AddressType)(address),
ShipTimeList: []string{},
OrderInfo: apiIndex.OrderInfo{
GoodsList: arrApiOrderGoods,
Order: &apiIndex.OrderBase{Order: orderInfo},
},
Cancel: genApiCancel(orderInfo),
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) MyOrderList(ctx echo.Context) error {
// 获取订单列表信息
// uid, base_id, rn
myOrderMapList, hasMore, err := logic.GetMyOrderList(10, 0, 20)
if nil != err {
return err
}
// 拼装接口数据
orderData := &apiIndex.OrderList{
HasMore: hasMore,
}
// orderData
for _, v := range myOrderMapList {
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(v["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range v["goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData.List = append(orderData.List, &apiIndex.Order{
Address: (*apiIndex.AddressType)(v["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order: &apiIndex.OrderBase{Order: v["order"].(*daoSql.Order)},
GoodsList: arrApiOrderGoods,
},
})
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) CancelOrder(ctx echo.Context) error {
// 获取订单列表信息
uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
cancelFlag := util.Atoi(ctx.FormValue("cancel_flag"), 16, false).(uint16)
err := logic.CancelOrder(uid, orderSn, cancelFlag)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) EvalOrder(ctx echo.Context) error {
// 获取订单列表信息
uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
stars := util.Atoi(ctx.FormValue("stars"), 8, false).(uint8)
feedback := ctx.FormValue("feedback")
err := logic.EvalOrder(uid, orderSn, stars, feedback)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) Detail(ctx echo.Context) error {
ordeSn := ctx.QueryParam("order_sn")
// 获取订单列表信息
// uid, orderSn
myOrderMap, err := logic.GetOrderDetail(10, ordeSn)
if nil != err {
return err
}
// 拼装接口数据
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(myOrderMap["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range myOrderMap["goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData := &apiIndex.Order{
Alert: "",
Address: (*apiIndex.AddressType)(myOrderMap["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order: &apiIndex.OrderBase{Order: myOrderMap["order"].(*daoSql.Order)},
GoodsList: arrApiOrderGoods,
},
Cancel: genApiCancel(myOrderMap["order"].(*daoSql.Order)),
}
orderData.Format()
return util.Render(ctx, "order/info", "订单详情", orderData)
}
// 生成订单列表页页的 html 不提供外部接口
func (OrderController) GenOrderListHtml(ctx echo.Context) error {
return util.Render(ctx, "order/list", "订单列表", map[string]interface{}{})
}
// 获取购物车中商品列表
func getCartGoodsList(ctx echo.Context) (goodsList []*logic.CartInfo, err error) {
goodsList = []*logic.CartInfo{}
goodsListStr := ctx.FormValue("goods_list")
goodsListMap := []map[string]interface{}{}
err = json.Unmarshal([]byte(goodsListStr), &goodsListMap)
for _, item := range goodsListMap {
tmpInfo := &logic.CartInfo{
GoodsId: util.MustNum(item["goods_id"], 64, false).(uint64),
Selected: util.MustNum(item["selected"], 8, false).(uint8),
GoodsNum: util.MustNum(item["goods_num"], 16, false).(uint16),
}
goodsList = append(goodsList, tmpInfo)
}
if err != nil {
// log
return
}
if 0 >= len(goodsList) {
return goodsList, CartEmpty
}
return
}
// 提交订单时新地址(address_id<=0) 会先插入地址表
func fetchAddress(ctx echo.Context) (*daoSql.Address, error) {
uid := ctx.Get("uid").(uint64)
// 获取提交订单时指定的地址
addressId := util.Atoi(ctx.FormValue("address_id"), 64, false).(uint64)
if 0 < addressId {
myAddressMap, err := daoSql.GetAddressListById([]uint64{addressId})
if nil != err {
// log
return &daoSql.Address{}, err
}
myAddress, ok := myAddressMap[addressId]
if !ok || uid != myAddress.MemberId {
// log
return &daoSql.Address{}, RecordEmpty
}
return myAddress, nil
}
// 插入新的地址信息
trueName := ctx.FormValue("true_name")
liveArea := ctx.FormValue("live_area")
address := ctx.FormValue("address")
mobile := ctx.FormValue("mobile")
// 显式提取地址信息
addressInfo := daoSql.UserAddressInfo{
TrueName: trueName,
LiveArea: liveArea,
Address: address,
Mobile: mobile,
}
myAddress, err := daoSql.SaveMyAddress(uid, &addressInfo)
if nil != err {
// log
return &daoSql.Address{}, err
}
return myAddress, err
}
// 根据订单
func genApiCancel(orderInfo *daoSql.Order) (cancel *apiIndex.Cancel) {
// 读入配置信息
envConf, _ := daoConf.EnvConf()
// 订单取消信息
cancelInfo := logic.GetCancelInfo(orderInfo)
cancel = &apiIndex.Cancel{
CanCancel: cancelInfo.CanCancel,
}
if !cancelInfo.CanCancel {
if 0 < len(envConf.ServiceTel) {
cancel.CancelTip.Tel = envConf.ServiceTel
} else {
cancel.CancelTip.Tel = cancelInfo.CancelTip.Tel
}
cancel.CancelTip.Tip = cancelInfo.CancelTip.Tip
} else {
cancelReasonList := []*apiIndex.CancelReasonType{}
for k, v := range cancelInfo.CancelReason {
tmp := &apiIndex.CancelReasonType{
Flag: util.Itoa(k),
Context: v,
}
| cancelRea | identifier_name | |
order.go | StorageException, _ := logic.VerifyGoodsNum(goodsIdMap, goodsList)
// 获取 goodsId 信息
shipTimeList := []string{"XX", "XX"}
// shipTimeList := logic.GetShipTime()
// 获取用户所有地址
myAddressList, err := daoSql.GetAddressListByUid(uid, false)
if nil != err && RecordEmpty != err {
// log
return util.Fail(ctx, 10, err.Error())
}
var myAddress *daoSql.Address
for idx, addressItem := range myAddressList {
// 默认取第一个
if 0 == idx {
myAddress = addressItem
}
// 取默认地址
if uint8(1) == addressItem.IsDefault {
myAddress = addressItem
}
}
if nil == myAddress {
myAddress = &daoSql.Address{}
} else {
myAddress.IsDefault = 1
}
// 读入配置信息
orderConf, err := daoConf.OrderConf()
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
// 生成预处理订单
orderMap := logic.OrderMap{
Address: myAddress,
GoodsIdMap: goodsIdMap,
GoodsList: goodsList,
}
orderInfo, orderGoodsList, err := logic.GenOrder(uid, orderMap)
if nil != err {
// log
}
// 过滤订单参数
orderInfo.Filter()
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(orderGoodsList))
for idx, item := range orderGoodsList {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
// 拼装接口数据
orderData := apiIndex.Order{
Address: (*apiIndex.AddressType)(myAddress),
ShipTimeList: shipTimeList,
OrderInfo: apiIndex.OrderInfo{
GoodsList: arrApiOrderGoods,
Order: &apiIndex.OrderBase{Order: orderInfo},
},
}
if 0 < len(goodsException) {
orderData.Alert = fmt.Sprintf(orderConf.Alert, goodsException)
} else if 0 < len(goodsNoStorageException) {
orderData.Alert = fmt.Sprintf(orderConf.StorageAlert, goodsNoStorageException)
} else {
orderData.Alert = ""
}
// 格式化地址列表
for _, addressItem := range myAddressList {
orderData.AddressList = append(orderData.AddressList, (*apiIndex.AddressType)(addressItem))
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 确认信息页 form-> goods_list:[{"goods_id":"3","selected":"1","goods_num":"2"}]
func (OrderController) DoOrder(ctx echo.Context) error {
// 设置 redis key
uid := ctx.Get("uid").(uint64)
// 避免同一个订单重复提交
// data, _ := json.Marshal(ctx.Request())
// curOrderMd5 = fmt.Printf("%x", md5.Sum(data))
// preOrderMd5 := daoRedis.NewRedisClient().Key(daoRedis.KeyOrder, util.Itoa(uid)).GET("")
// if preOrder == preOrderMd5 {
// // log
// return util.Fail(ctx, 10, RepeatDoOrder)
// }
// daoRedis.NewRedisClient().SET("", curOrderMd5, 30)
// 获取购物车商品列表
goodsList, err := getCartGoodsList(ctx)
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
// 收集 goodsId
goodsIdList := []uint64{}
for _, goodsInfo := range goodsList {
if 1 == goodsInfo.Selected {
if 0 < goodsInfo.GoodsNum {
goodsIdList = append(goodsIdList, goodsIn | goodsIdMap, _ := logic.GetCartInfo(goodsIdList)
// 验证并修正库存信息 如果只有3个,购买5个,会强制改为3个
goodsNoStorageException, _ := logic.VerifyGoodsNum(goodsIdMap, goodsList)
// 获取地址信息
address, err := fetchAddress(ctx)
if nil != err {
return util.Fail(ctx, 10, "地址信息无效")
}
// 读入配置信息
orderConf, err := daoConf.OrderConf()
if nil != err {
// log
return util.Fail(ctx, 10, err.Error())
}
if 0 < len(goodsException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.Alert, goodsException))
}
if 0 < len(goodsNoStorageException) {
return util.Fail(ctx, 10, fmt.Sprintf(orderConf.StorageAlert, goodsNoStorageException))
}
// 提交订单
orderMap := logic.OrderMap{
Address: address,
GoodsIdMap: goodsIdMap,
GoodsList: goodsList,
ExceptTime: time.Now().Unix(),
OrderMessage: ctx.FormValue("order_message"),
}
orderInfo, orderGoodsList, err := logic.SubmitOrder(uid, orderMap)
if nil != err {
// log
}
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(orderGoodsList))
for idx, item := range orderGoodsList {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
// 拼装接口数据
orderData := apiIndex.Order{
Address: (*apiIndex.AddressType)(address),
ShipTimeList: []string{},
OrderInfo: apiIndex.OrderInfo{
GoodsList: arrApiOrderGoods,
Order: &apiIndex.OrderBase{Order: orderInfo},
},
Cancel: genApiCancel(orderInfo),
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) MyOrderList(ctx echo.Context) error {
// 获取订单列表信息
// uid, base_id, rn
myOrderMapList, hasMore, err := logic.GetMyOrderList(10, 0, 20)
if nil != err {
return err
}
// 拼装接口数据
orderData := &apiIndex.OrderList{
HasMore: hasMore,
}
// orderData
for _, v := range myOrderMapList {
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(v["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range v["goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData.List = append(orderData.List, &apiIndex.Order{
Address: (*apiIndex.AddressType)(v["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order: &apiIndex.OrderBase{Order: v["order"].(*daoSql.Order)},
GoodsList: arrApiOrderGoods,
},
})
}
orderData.Format()
return util.Success(ctx, orderData)
}
// 订单列表页
func (OrderController) CancelOrder(ctx echo.Context) error {
// 获取订单列表信息
uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
cancelFlag := util.Atoi(ctx.FormValue("cancel_flag"), 16, false).(uint16)
err := logic.CancelOrder(uid, orderSn, cancelFlag)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) EvalOrder(ctx echo.Context) error {
// 获取订单列表信息
uid := uint64(10)
orderSn := ctx.FormValue("order_sn")
stars := util.Atoi(ctx.FormValue("stars"), 8, false).(uint8)
feedback := ctx.FormValue("feedback")
err := logic.EvalOrder(uid, orderSn, stars, feedback)
if nil != err {
return util.Fail(ctx, 10, err.Error())
}
return util.Success(ctx, nil)
}
// 订单列表页
func (OrderController) Detail(ctx echo.Context) error {
ordeSn := ctx.QueryParam("order_sn")
// 获取订单列表信息
// uid, orderSn
myOrderMap, err := logic.GetOrderDetail(10, ordeSn)
if nil != err {
return err
}
// 拼装接口数据
arrApiOrderGoods := make([]*apiIndex.OrderGoods, len(myOrderMap["goodsList"].([]*daoSql.OrderGoods)))
for idx, item := range myOrderMap["goodsList"].([]*daoSql.OrderGoods) {
arrApiOrderGoods[idx] = &apiIndex.OrderGoods{OrderGoods: item}
}
orderData := &apiIndex.Order{
Alert: "",
Address: (*apiIndex.AddressType)(myOrderMap["addressInfo"].(*daoSql.Address)),
OrderInfo: apiIndex.OrderInfo{
Order | fo.GoodsId)
}
}
}
// 获取 商品详情
goodsException, | conditional_block |
symbol.ts | ';
import { SymbolOptionMixin, ZRColor } from './types';
import { parsePercent } from './number';
export type ECSymbol = graphic.Path & {
__isEmptyBrush?: boolean
setColor: (color: ZRColor, innerColor?: ZRColor) => void
getColor: () => ZRColor
};
type SymbolCtor = { new(): ECSymbol };
type SymbolShapeMaker = (x: number, y: number, w: number, h: number, shape: Dictionary<any>) => void;
/**
* Triangle shape
* @inner
*/
const Triangle = graphic.Path.extend({
type: 'triangle',
shape: {
cx: 0,
cy: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const cx = shape.cx;
const cy = shape.cy;
const width = shape.width / 2;
const height = shape.height / 2;
path.moveTo(cx, cy - height);
path.lineTo(cx + width, cy + height);
path.lineTo(cx - width, cy + height);
path.closePath();
}
});
/**
* Diamond shape
* @inner
*/
const Diamond = graphic.Path.extend({
type: 'diamond',
shape: {
cx: 0,
cy: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const cx = shape.cx;
const cy = shape.cy;
const width = shape.width / 2;
const height = shape.height / 2;
path.moveTo(cx, cy - height);
path.lineTo(cx + width, cy);
path.lineTo(cx, cy + height);
path.lineTo(cx - width, cy);
path.closePath();
}
});
/**
* Pin shape
* @inner
*/
const Pin = graphic.Path.extend({
type: 'pin',
shape: {
// x, y on the cusp
x: 0,
y: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const x = shape.x;
const y = shape.y;
const w = shape.width / 5 * 3;
// Height must be larger than width
const h = Math.max(w, shape.height);
const r = w / 2;
// Dist on y with tangent point and circle center
const dy = r * r / (h - r);
const cy = y - h + r + dy;
const angle = Math.asin(dy / r);
// Dist on x with tangent point and circle center
const dx = Math.cos(angle) * r;
const tanX = Math.sin(angle);
const tanY = Math.cos(angle);
const cpLen = r * 0.6;
const cpLen2 = r * 0.7;
path.moveTo(x - dx, cy + dy);
path.arc(
x, cy, r,
Math.PI - angle,
Math.PI * 2 + angle
);
path.bezierCurveTo(
x + dx - tanX * cpLen, cy + dy + tanY * cpLen,
x, y - cpLen2,
x, y
);
path.bezierCurveTo(
x, y - cpLen2,
x - dx + tanX * cpLen, cy + dy + tanY * cpLen,
x - dx, cy + dy
);
path.closePath();
}
});
/**
* Arrow shape
* @inner
*/
const Arrow = graphic.Path.extend({
type: 'arrow',
shape: {
x: 0,
y: 0,
width: 0,
height: 0
},
buildPath: function (ctx, shape) {
const height = shape.height;
const width = shape.width;
const x = shape.x;
const y = shape.y;
const dx = width / 3 * 2;
ctx.moveTo(x, y);
ctx.lineTo(x + dx, y + height);
ctx.lineTo(x, y + height / 4 * 3);
ctx.lineTo(x - dx, y + height);
ctx.lineTo(x, y);
ctx.closePath();
}
});
/**
* Map of path constructors | rect: graphic.Rect as unknown as SymbolCtor,
roundRect: graphic.Rect as unknown as SymbolCtor,
square: graphic.Rect as unknown as SymbolCtor,
circle: graphic.Circle as unknown as SymbolCtor,
diamond: Diamond as unknown as SymbolCtor,
pin: Pin as unknown as SymbolCtor,
arrow: Arrow as unknown as SymbolCtor,
triangle: Triangle as unknown as SymbolCtor
};
const symbolShapeMakers: Dictionary<SymbolShapeMaker> = {
line: function (x, y, w, h, shape: graphic.Line['shape']) {
shape.x1 = x;
shape.y1 = y + h / 2;
shape.x2 = x + w;
shape.y2 = y + h / 2;
},
rect: function (x, y, w, h, shape: graphic.Rect['shape']) {
shape.x = x;
shape.y = y;
shape.width = w;
shape.height = h;
},
roundRect: function (x, y, w, h, shape: graphic.Rect['shape']) {
shape.x = x;
shape.y = y;
shape.width = w;
shape.height = h;
shape.r = Math.min(w, h) / 4;
},
square: function (x, y, w, h, shape: graphic.Rect['shape']) {
const size = Math.min(w, h);
shape.x = x;
shape.y = y;
shape.width = size;
shape.height = size;
},
circle: function (x, y, w, h, shape: graphic.Circle['shape']) {
// Put circle in the center of square
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.r = Math.min(w, h) / 2;
},
diamond: function (x, y, w, h, shape: InstanceType<typeof Diamond>['shape']) {
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.width = w;
shape.height = h;
},
pin: function (x, y, w, h, shape: InstanceType<typeof Pin>['shape']) {
shape.x = x + w / 2;
shape.y = y + h / 2;
shape.width = w;
shape.height = h;
},
arrow: function (x, y, w, h, shape: InstanceType<typeof Arrow>['shape']) {
shape.x = x + w / 2;
shape.y = y + h / 2;
shape.width = w;
shape.height = h;
},
triangle: function (x, y, w, h, shape: InstanceType<typeof Triangle>['shape']) {
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.width = w;
shape.height = h;
}
};
export const symbolBuildProxies: Dictionary<ECSymbol> = {};
each(symbolCtors, function (Ctor, name) {
symbolBuildProxies[name] = new Ctor();
});
const SymbolClz = graphic.Path.extend({
type: 'symbol',
shape: {
symbolType: '',
x: 0,
y: 0,
width: 0,
height: 0
},
calculateTextPosition(out, config, rect) {
const res = calculateTextPosition(out, config, rect);
const shape = this.shape;
if (shape && shape.symbolType === 'pin' && config.position === 'inside') {
res.y = rect.y + rect.height * 0.4;
}
return res;
},
buildPath(ctx, shape, inBundle) {
let symbolType = shape.symbolType;
if (symbolType !== 'none') {
let proxySymbol = symbolBuildProxies[symbolType];
if (!proxySymbol) {
// Default rect
symbolType = 'rect';
proxySymbol = symbolBuildProxies[symbolType];
}
symbolShapeMakers[symbolType](
shape.x, shape.y, shape.width, shape.height, proxySymbol.shape
);
proxySymbol.buildPath(ctx, proxySymbol.shape, inBundle);
}
}
});
// Provide setColor helper method to avoid determine if set the fill or stroke outside
function symbolPathSetColor(this: ECSymbol, color: ZRColor, innerColor?: ZRColor) {
if (this.type !== 'image') {
const symbolStyle = this.style;
if (this.__isEmptyBrush) {
symbolStyle.stroke = color;
symbolStyle.fill = innerColor || '#fff';
// TODO Same width with lineStyle in LineView
symbolStyle.lineWidth = 2;
}
else if (this.shape.symbolType === 'line') {
symbolStyle.stroke = color;
}
else {
symbolStyle.fill = color;
}
| */
// TODO Use function to build symbol path.
const symbolCtors: Dictionary<SymbolCtor> = {
line: graphic.Line as unknown as SymbolCtor,
| random_line_split |
symbol.ts | boolean
setColor: (color: ZRColor, innerColor?: ZRColor) => void
getColor: () => ZRColor
};
type SymbolCtor = { new(): ECSymbol };
type SymbolShapeMaker = (x: number, y: number, w: number, h: number, shape: Dictionary<any>) => void;
/**
* Triangle shape
* @inner
*/
const Triangle = graphic.Path.extend({
type: 'triangle',
shape: {
cx: 0,
cy: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const cx = shape.cx;
const cy = shape.cy;
const width = shape.width / 2;
const height = shape.height / 2;
path.moveTo(cx, cy - height);
path.lineTo(cx + width, cy + height);
path.lineTo(cx - width, cy + height);
path.closePath();
}
});
/**
* Diamond shape
* @inner
*/
const Diamond = graphic.Path.extend({
type: 'diamond',
shape: {
cx: 0,
cy: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const cx = shape.cx;
const cy = shape.cy;
const width = shape.width / 2;
const height = shape.height / 2;
path.moveTo(cx, cy - height);
path.lineTo(cx + width, cy);
path.lineTo(cx, cy + height);
path.lineTo(cx - width, cy);
path.closePath();
}
});
/**
* Pin shape
* @inner
*/
const Pin = graphic.Path.extend({
type: 'pin',
shape: {
// x, y on the cusp
x: 0,
y: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const x = shape.x;
const y = shape.y;
const w = shape.width / 5 * 3;
// Height must be larger than width
const h = Math.max(w, shape.height);
const r = w / 2;
// Dist on y with tangent point and circle center
const dy = r * r / (h - r);
const cy = y - h + r + dy;
const angle = Math.asin(dy / r);
// Dist on x with tangent point and circle center
const dx = Math.cos(angle) * r;
const tanX = Math.sin(angle);
const tanY = Math.cos(angle);
const cpLen = r * 0.6;
const cpLen2 = r * 0.7;
path.moveTo(x - dx, cy + dy);
path.arc(
x, cy, r,
Math.PI - angle,
Math.PI * 2 + angle
);
path.bezierCurveTo(
x + dx - tanX * cpLen, cy + dy + tanY * cpLen,
x, y - cpLen2,
x, y
);
path.bezierCurveTo(
x, y - cpLen2,
x - dx + tanX * cpLen, cy + dy + tanY * cpLen,
x - dx, cy + dy
);
path.closePath();
}
});
/**
* Arrow shape
* @inner
*/
const Arrow = graphic.Path.extend({
type: 'arrow',
shape: {
x: 0,
y: 0,
width: 0,
height: 0
},
buildPath: function (ctx, shape) {
const height = shape.height;
const width = shape.width;
const x = shape.x;
const y = shape.y;
const dx = width / 3 * 2;
ctx.moveTo(x, y);
ctx.lineTo(x + dx, y + height);
ctx.lineTo(x, y + height / 4 * 3);
ctx.lineTo(x - dx, y + height);
ctx.lineTo(x, y);
ctx.closePath();
}
});
/**
* Map of path constructors
*/
// TODO Use function to build symbol path.
const symbolCtors: Dictionary<SymbolCtor> = {
line: graphic.Line as unknown as SymbolCtor,
rect: graphic.Rect as unknown as SymbolCtor,
roundRect: graphic.Rect as unknown as SymbolCtor,
square: graphic.Rect as unknown as SymbolCtor,
circle: graphic.Circle as unknown as SymbolCtor,
diamond: Diamond as unknown as SymbolCtor,
pin: Pin as unknown as SymbolCtor,
arrow: Arrow as unknown as SymbolCtor,
triangle: Triangle as unknown as SymbolCtor
};
const symbolShapeMakers: Dictionary<SymbolShapeMaker> = {
line: function (x, y, w, h, shape: graphic.Line['shape']) {
shape.x1 = x;
shape.y1 = y + h / 2;
shape.x2 = x + w;
shape.y2 = y + h / 2;
},
rect: function (x, y, w, h, shape: graphic.Rect['shape']) {
shape.x = x;
shape.y = y;
shape.width = w;
shape.height = h;
},
roundRect: function (x, y, w, h, shape: graphic.Rect['shape']) {
shape.x = x;
shape.y = y;
shape.width = w;
shape.height = h;
shape.r = Math.min(w, h) / 4;
},
square: function (x, y, w, h, shape: graphic.Rect['shape']) {
const size = Math.min(w, h);
shape.x = x;
shape.y = y;
shape.width = size;
shape.height = size;
},
circle: function (x, y, w, h, shape: graphic.Circle['shape']) {
// Put circle in the center of square
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.r = Math.min(w, h) / 2;
},
diamond: function (x, y, w, h, shape: InstanceType<typeof Diamond>['shape']) {
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.width = w;
shape.height = h;
},
pin: function (x, y, w, h, shape: InstanceType<typeof Pin>['shape']) {
shape.x = x + w / 2;
shape.y = y + h / 2;
shape.width = w;
shape.height = h;
},
arrow: function (x, y, w, h, shape: InstanceType<typeof Arrow>['shape']) {
shape.x = x + w / 2;
shape.y = y + h / 2;
shape.width = w;
shape.height = h;
},
triangle: function (x, y, w, h, shape: InstanceType<typeof Triangle>['shape']) {
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.width = w;
shape.height = h;
}
};
export const symbolBuildProxies: Dictionary<ECSymbol> = {};
each(symbolCtors, function (Ctor, name) {
symbolBuildProxies[name] = new Ctor();
});
const SymbolClz = graphic.Path.extend({
type: 'symbol',
shape: {
symbolType: '',
x: 0,
y: 0,
width: 0,
height: 0
},
calculateTextPosition(out, config, rect) {
const res = calculateTextPosition(out, config, rect);
const shape = this.shape;
if (shape && shape.symbolType === 'pin' && config.position === 'inside') {
res.y = rect.y + rect.height * 0.4;
}
return res;
},
buildPath(ctx, shape, inBundle) {
let symbolType = shape.symbolType;
if (symbolType !== 'none') {
let proxySymbol = symbolBuildProxies[symbolType];
if (!proxySymbol) {
// Default rect
symbolType = 'rect';
proxySymbol = symbolBuildProxies[symbolType];
}
symbolShapeMakers[symbolType](
shape.x, shape.y, shape.width, shape.height, proxySymbol.shape
);
proxySymbol.buildPath(ctx, proxySymbol.shape, inBundle);
}
}
});
// Provide setColor helper method to avoid determine if set the fill or stroke outside
function symbolPathSetColor(this: ECSymbol, color: ZRColor, innerColor?: ZRColor) {
if (this.type !== 'image') {
const symbolStyle = this.style;
if (this.__isEmptyBrush) {
symbolStyle.stroke = color;
symbolStyle.fill = innerColor || '#fff';
// TODO Same width with lineStyle in LineView
symbolStyle.lineWidth = 2;
}
else if (this.shape.symbolType === 'line') {
symbolStyle.stroke = color;
}
else {
symbolStyle.fill = color;
}
this.markRedraw();
}
}
/**
* Create a symbol element with given symbol configuration: shape, x, y, width, height, color
*/
export function | createSymbol | identifier_name | |
symbol.ts | import { SymbolOptionMixin, ZRColor } from './types';
import { parsePercent } from './number';
export type ECSymbol = graphic.Path & {
__isEmptyBrush?: boolean
setColor: (color: ZRColor, innerColor?: ZRColor) => void
getColor: () => ZRColor
};
type SymbolCtor = { new(): ECSymbol };
type SymbolShapeMaker = (x: number, y: number, w: number, h: number, shape: Dictionary<any>) => void;
/**
* Triangle shape
* @inner
*/
const Triangle = graphic.Path.extend({
type: 'triangle',
shape: {
cx: 0,
cy: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const cx = shape.cx;
const cy = shape.cy;
const width = shape.width / 2;
const height = shape.height / 2;
path.moveTo(cx, cy - height);
path.lineTo(cx + width, cy + height);
path.lineTo(cx - width, cy + height);
path.closePath();
}
});
/**
* Diamond shape
* @inner
*/
const Diamond = graphic.Path.extend({
type: 'diamond',
shape: {
cx: 0,
cy: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const cx = shape.cx;
const cy = shape.cy;
const width = shape.width / 2;
const height = shape.height / 2;
path.moveTo(cx, cy - height);
path.lineTo(cx + width, cy);
path.lineTo(cx, cy + height);
path.lineTo(cx - width, cy);
path.closePath();
}
});
/**
* Pin shape
* @inner
*/
const Pin = graphic.Path.extend({
type: 'pin',
shape: {
// x, y on the cusp
x: 0,
y: 0,
width: 0,
height: 0
},
buildPath: function (path, shape) {
const x = shape.x;
const y = shape.y;
const w = shape.width / 5 * 3;
// Height must be larger than width
const h = Math.max(w, shape.height);
const r = w / 2;
// Dist on y with tangent point and circle center
const dy = r * r / (h - r);
const cy = y - h + r + dy;
const angle = Math.asin(dy / r);
// Dist on x with tangent point and circle center
const dx = Math.cos(angle) * r;
const tanX = Math.sin(angle);
const tanY = Math.cos(angle);
const cpLen = r * 0.6;
const cpLen2 = r * 0.7;
path.moveTo(x - dx, cy + dy);
path.arc(
x, cy, r,
Math.PI - angle,
Math.PI * 2 + angle
);
path.bezierCurveTo(
x + dx - tanX * cpLen, cy + dy + tanY * cpLen,
x, y - cpLen2,
x, y
);
path.bezierCurveTo(
x, y - cpLen2,
x - dx + tanX * cpLen, cy + dy + tanY * cpLen,
x - dx, cy + dy
);
path.closePath();
}
});
/**
* Arrow shape
* @inner
*/
const Arrow = graphic.Path.extend({
type: 'arrow',
shape: {
x: 0,
y: 0,
width: 0,
height: 0
},
buildPath: function (ctx, shape) {
const height = shape.height;
const width = shape.width;
const x = shape.x;
const y = shape.y;
const dx = width / 3 * 2;
ctx.moveTo(x, y);
ctx.lineTo(x + dx, y + height);
ctx.lineTo(x, y + height / 4 * 3);
ctx.lineTo(x - dx, y + height);
ctx.lineTo(x, y);
ctx.closePath();
}
});
/**
* Map of path constructors
*/
// TODO Use function to build symbol path.
const symbolCtors: Dictionary<SymbolCtor> = {
line: graphic.Line as unknown as SymbolCtor,
rect: graphic.Rect as unknown as SymbolCtor,
roundRect: graphic.Rect as unknown as SymbolCtor,
square: graphic.Rect as unknown as SymbolCtor,
circle: graphic.Circle as unknown as SymbolCtor,
diamond: Diamond as unknown as SymbolCtor,
pin: Pin as unknown as SymbolCtor,
arrow: Arrow as unknown as SymbolCtor,
triangle: Triangle as unknown as SymbolCtor
};
const symbolShapeMakers: Dictionary<SymbolShapeMaker> = {
line: function (x, y, w, h, shape: graphic.Line['shape']) {
shape.x1 = x;
shape.y1 = y + h / 2;
shape.x2 = x + w;
shape.y2 = y + h / 2;
},
rect: function (x, y, w, h, shape: graphic.Rect['shape']) {
shape.x = x;
shape.y = y;
shape.width = w;
shape.height = h;
},
roundRect: function (x, y, w, h, shape: graphic.Rect['shape']) {
shape.x = x;
shape.y = y;
shape.width = w;
shape.height = h;
shape.r = Math.min(w, h) / 4;
},
square: function (x, y, w, h, shape: graphic.Rect['shape']) {
const size = Math.min(w, h);
shape.x = x;
shape.y = y;
shape.width = size;
shape.height = size;
},
circle: function (x, y, w, h, shape: graphic.Circle['shape']) {
// Put circle in the center of square
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.r = Math.min(w, h) / 2;
},
diamond: function (x, y, w, h, shape: InstanceType<typeof Diamond>['shape']) {
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.width = w;
shape.height = h;
},
pin: function (x, y, w, h, shape: InstanceType<typeof Pin>['shape']) {
shape.x = x + w / 2;
shape.y = y + h / 2;
shape.width = w;
shape.height = h;
},
arrow: function (x, y, w, h, shape: InstanceType<typeof Arrow>['shape']) {
shape.x = x + w / 2;
shape.y = y + h / 2;
shape.width = w;
shape.height = h;
},
triangle: function (x, y, w, h, shape: InstanceType<typeof Triangle>['shape']) {
shape.cx = x + w / 2;
shape.cy = y + h / 2;
shape.width = w;
shape.height = h;
}
};
export const symbolBuildProxies: Dictionary<ECSymbol> = {};
each(symbolCtors, function (Ctor, name) {
symbolBuildProxies[name] = new Ctor();
});
const SymbolClz = graphic.Path.extend({
type: 'symbol',
shape: {
symbolType: '',
x: 0,
y: 0,
width: 0,
height: 0
},
calculateTextPosition(out, config, rect) {
const res = calculateTextPosition(out, config, rect);
const shape = this.shape;
if (shape && shape.symbolType === 'pin' && config.position === 'inside') {
res.y = rect.y + rect.height * 0.4;
}
return res;
},
buildPath(ctx, shape, inBundle) |
});
// Provide setColor helper method to avoid determine if set the fill or stroke outside
function symbolPathSetColor(this: ECSymbol, color: ZRColor, innerColor?: ZRColor) {
if (this.type !== 'image') {
const symbolStyle = this.style;
if (this.__isEmptyBrush) {
symbolStyle.stroke = color;
symbolStyle.fill = innerColor || '#fff';
// TODO Same width with lineStyle in LineView
symbolStyle.lineWidth = 2;
}
else if (this.shape.symbolType === 'line') {
symbolStyle.stroke = color;
}
else {
symbolStyle.fill = color;
| {
let symbolType = shape.symbolType;
if (symbolType !== 'none') {
let proxySymbol = symbolBuildProxies[symbolType];
if (!proxySymbol) {
// Default rect
symbolType = 'rect';
proxySymbol = symbolBuildProxies[symbolType];
}
symbolShapeMakers[symbolType](
shape.x, shape.y, shape.width, shape.height, proxySymbol.shape
);
proxySymbol.buildPath(ctx, proxySymbol.shape, inBundle);
}
} | identifier_body |
main.rs | root.step_in_direction(&d)))
.filter_map(|(d, p)| match gd.get(&p) {
Some(Tile::Unexplored) | None => Some((d, p)),
Some(_) => None,
})
.collect::<Vec<_>>();
for (d, target) in unexplored {
let (ended_up, tile) = robot.try_move(d);
if tile == Tile::Oxygen {
assert!(oxygen_at.is_none());
oxygen_at = Some(ended_up);
}
if target == ended_up {
gd.insert(&target, tile);
// push to the same side as we are popping will decrease the amount of running
// around on the map so maybe depth first?
work.push_front(target);
let (back_at, _) = robot.try_move(d.reverse());
assert_eq!(back_at, root);
} else {
gd.insert(&target, tile);
}
}
}
}
println!("oxygen at: {:?}", oxygen_at);
println!("robot moves: {}", robot.moves);
println!("stage1: {}", path_to(&gd, &( 0, 0), oxygen_at.as_ref().unwrap()).unwrap().len());
{
// stage2 is probably just a dfs from the oxygen, mark the coordinates and ... push all new
// marked ones to the queue?
let mut frontier = VecDeque::new();
let mut oxygen = HashSet::new();
oxygen.insert(oxygen_at.unwrap());
frontier.push_back((oxygen_at.unwrap(), 0));
let mut prev_time = 0;
let mut minutes = 0;
while let Some((p1, time)) = frontier.pop_front() {
oxygen.insert(p1);
if prev_time != time {
assert!(prev_time < time, "{} should be less than {}", prev_time, time);
prev_time = time;
minutes += 1;
println!("{:>3} minutes ... {} slots oxygenated", minutes, oxygen.len());
}
let unoxinated = Direction::all()
.map(|d| p1.step_in_direction(&d))
.filter_map(|p| match gd.get(&p) {
Some(Tile::Empty) => Some(p),
Some(_) | None => None,
})
.filter(|p| !oxygen.contains(&p))
.collect::<Vec<_>>();
for p2 in unoxinated {
frontier.push_back((p2, time + 1));
}
}
println!("stage2: {}", minutes);
}
}
/// Wasteful dijkstra ... could share the hashmaps across queries maybe?
fn | (gd: &GameDisplay<Tile>, pos: &(Word, Word), target: &(Word, Word)) -> Option<Vec<Direction>> {
//println!("path_to: {:?} to {:?}", pos, target);
use std::collections::{HashMap, BinaryHeap};
use std::collections::hash_map::Entry;
use std::cmp;
let mut ret = Vec::new();
let mut work = BinaryHeap::new();
let mut dist = HashMap::new();
let mut prev = HashMap::new();
work.push(cmp::Reverse((0, *pos)));
while let Some(cmp::Reverse((steps_here, p))) = work.pop() {
//println!("path_to: popped {:?}", (p, steps_here));
if p == *target {
//println!("path_to: found target {:?}", p);
let mut backwards = p;
ret.push(p);
while backwards != *pos {
let previous = prev.remove(&backwards).unwrap();
ret.push(previous);
backwards = previous;
}
ret.reverse();
let dirs = ret.windows(2)
.map(|slice| {
let a = slice[0];
let b = slice[1];
let d = (b.0 - a.0, b.1 - a.1);
match d {
( 0,-1) => Direction::Down,
( 0, 1) => Direction::Up,
(-1, 0) => Direction::Left,
( 1, 0) => Direction::Right,
x => unreachable!("cannot have this {:?} between {:?} and {:?}", x, a, b),
}
}).collect();
return Some(dirs);
}
match dist.entry(p) {
Entry::Vacant(vcnt) => {
vcnt.insert(steps_here);
},
Entry::Occupied(mut o) => {
if *o.get() >= steps_here {
*o.get_mut() = steps_here;
} else {
println!("already visited {:?} with lower dist {} than {} from {:?}", p, o.get(), steps_here, prev[&p]);
continue;
}
}
}
for (p2, dir) in adjacent(gd, &p) {
let alt = steps_here + 1;
if alt < *dist.get(&p2).unwrap_or(&usize::max_value()) {
//println!(" {:?} --{:?}--> {:?}", p, dir, p2);
dist.insert(p2, alt);
prev.insert(p2, p);
work.push(cmp::Reverse((alt, p2)));
}
}
}
None
}
#[test]
fn test_path_to() {
use Direction::*;
let mut gd: GameDisplay<Tile> = GameDisplay::default();
gd.insert(&(-1, 0), Tile::Wall);
gd.insert(&(-1,-1), Tile::Wall);
gd.insert(&( 0,-1), Tile::Wall);
gd.insert(&( 2, 0), Tile::Wall);
gd.insert(&( 2,-1), Tile::Wall);
gd.insert(&( 0, 0), Tile::Empty); // right
gd.insert(&( 1, 0), Tile::Empty); // down
gd.insert(&( 1, 1), Tile::Empty); // down
gd.insert(&( 1, 2), Tile::Empty); // down
gd.insert(&( 1, 3), Tile::Empty); // down
gd.insert(&( 1, 4), Tile::Empty); // down
gd.insert(&( 2, 4), Tile::Empty); // down
gd.insert(&( 3, 4), Tile::Empty); // down
gd.insert(&( 4, 4), Tile::Empty); // down
gd.insert(&( 4, 3), Tile::Empty); // down
gd.insert(&( 4, 2), Tile::Empty); // down
gd.insert(&( 4, 1), Tile::Empty); // down
gd.insert(&( 1,-1), Tile::Empty); // down
gd.insert(&( 1,-2), Tile::Empty); // down
gd.insert(&( 2,-2), Tile::Empty); // right
gd.insert(&( 3,-2), Tile::Empty); // right
gd.insert(&( 3,-1), Tile::Empty);
gd.insert(&( 3, 0), Tile::Empty);
gd.insert(&( 4, 0), Tile::Empty);
println!("{}", gd);
assert_eq!(vec![Right, Down, Down, Right, Right, Up, Up, Right], path_to(&gd, &( 0, 0), &( 4, 0)).unwrap());
}
fn adjacent<'a>(gd: &'a GameDisplay<Tile>, pos: &'a (Word, Word)) -> impl Iterator<Item = ((Word, Word), Direction)> + 'a {
Direction::all()
.into_iter()
.map(move |d| (pos.step_in_direction(&d), d))
.filter_map(move |(p2, d)| gd.get(&p2).map(|t| (p2, d, t)))
//.inspect(|x| println!(" c: {:?}", x))
.filter_map(|(p2, d, t)| match t {
&Tile::Empty | &Tile::Robot | &Tile::Oxygen => Some((p2, d)),
_ => None,
})
//.inspect(|x| println!(" d: {:?}", x))
}
struct Robot {
program: Program<'static>,
regs: Option<Registers>,
pos: (Word, Word),
moves: usize,
}
impl Robot {
fn new(data: Vec<Word>) -> Self {
let mem = intcode::Memory::from(data).with_memory_expansion();
let program = Program::from(mem);
Robot {
program,
regs: Some(Registers::default()),
pos: (0, 0),
moves: 0,
}
}
fn position(&self) -> &(Word, Word) {
&self.pos
}
fn try_move(&mut self, dir: Direction) -> ((Word, Word), Tile) {
loop {
let mut ret = None;
self.regs = Some(match self.program.eval_from_instruction(self.regs.take().unwrap()).unwrap() {
ExecutionState::HaltedAt(regs) => unreachable!("Halted at: {:?}", regs),
ExecutionState::Paused(regs) => unreachable!("Paused? {:?}", regs),
ExecutionState::InputIO(io) => {
let val: i64 = dir.into();
//println!("robot <-- {}", val);
self.program.handle_input_completion(io, val).unwrap()
},
ExecutionState::OutputIO(io, value) => {
//println!("robot --> {}", value);
let moved = value != 0;
let | path_to | identifier_name |
main.rs | root.step_in_direction(&d)))
.filter_map(|(d, p)| match gd.get(&p) {
Some(Tile::Unexplored) | None => Some((d, p)),
Some(_) => None,
})
.collect::<Vec<_>>();
for (d, target) in unexplored {
let (ended_up, tile) = robot.try_move(d);
if tile == Tile::Oxygen {
assert!(oxygen_at.is_none());
oxygen_at = Some(ended_up);
}
if target == ended_up {
gd.insert(&target, tile);
// push to the same side as we are popping will decrease the amount of running
// around on the map so maybe depth first?
work.push_front(target);
let (back_at, _) = robot.try_move(d.reverse());
assert_eq!(back_at, root);
} else {
gd.insert(&target, tile);
}
}
}
}
println!("oxygen at: {:?}", oxygen_at);
println!("robot moves: {}", robot.moves);
println!("stage1: {}", path_to(&gd, &( 0, 0), oxygen_at.as_ref().unwrap()).unwrap().len());
{
// stage2 is probably just a dfs from the oxygen, mark the coordinates and ... push all new
// marked ones to the queue?
let mut frontier = VecDeque::new();
let mut oxygen = HashSet::new();
oxygen.insert(oxygen_at.unwrap());
frontier.push_back((oxygen_at.unwrap(), 0));
let mut prev_time = 0;
let mut minutes = 0;
while let Some((p1, time)) = frontier.pop_front() {
oxygen.insert(p1);
if prev_time != time {
assert!(prev_time < time, "{} should be less than {}", prev_time, time);
prev_time = time;
minutes += 1;
println!("{:>3} minutes ... {} slots oxygenated", minutes, oxygen.len());
}
let unoxinated = Direction::all()
.map(|d| p1.step_in_direction(&d))
.filter_map(|p| match gd.get(&p) {
Some(Tile::Empty) => Some(p),
Some(_) | None => None,
})
.filter(|p| !oxygen.contains(&p))
.collect::<Vec<_>>();
for p2 in unoxinated {
frontier.push_back((p2, time + 1));
}
}
println!("stage2: {}", minutes);
}
}
/// Wasteful dijkstra ... could share the hashmaps across queries maybe?
fn path_to(gd: &GameDisplay<Tile>, pos: &(Word, Word), target: &(Word, Word)) -> Option<Vec<Direction>> {
//println!("path_to: {:?} to {:?}", pos, target);
use std::collections::{HashMap, BinaryHeap};
use std::collections::hash_map::Entry;
use std::cmp;
let mut ret = Vec::new();
let mut work = BinaryHeap::new();
let mut dist = HashMap::new();
let mut prev = HashMap::new();
work.push(cmp::Reverse((0, *pos)));
while let Some(cmp::Reverse((steps_here, p))) = work.pop() {
//println!("path_to: popped {:?}", (p, steps_here));
if p == *target {
//println!("path_to: found target {:?}", p);
let mut backwards = p;
ret.push(p);
while backwards != *pos {
let previous = prev.remove(&backwards).unwrap();
ret.push(previous);
backwards = previous;
}
ret.reverse();
let dirs = ret.windows(2)
.map(|slice| {
let a = slice[0];
let b = slice[1];
let d = (b.0 - a.0, b.1 - a.1);
match d {
( 0,-1) => Direction::Down,
( 0, 1) => Direction::Up,
(-1, 0) => Direction::Left,
( 1, 0) => Direction::Right,
x => unreachable!("cannot have this {:?} between {:?} and {:?}", x, a, b),
}
}).collect();
return Some(dirs);
}
match dist.entry(p) {
Entry::Vacant(vcnt) => {
vcnt.insert(steps_here);
},
Entry::Occupied(mut o) => {
if *o.get() >= steps_here {
*o.get_mut() = steps_here;
} else {
println!("already visited {:?} with lower dist {} than {} from {:?}", p, o.get(), steps_here, prev[&p]);
continue;
} |
if alt < *dist.get(&p2).unwrap_or(&usize::max_value()) {
//println!(" {:?} --{:?}--> {:?}", p, dir, p2);
dist.insert(p2, alt);
prev.insert(p2, p);
work.push(cmp::Reverse((alt, p2)));
}
}
}
None
}
#[test]
fn test_path_to() {
use Direction::*;
let mut gd: GameDisplay<Tile> = GameDisplay::default();
gd.insert(&(-1, 0), Tile::Wall);
gd.insert(&(-1,-1), Tile::Wall);
gd.insert(&( 0,-1), Tile::Wall);
gd.insert(&( 2, 0), Tile::Wall);
gd.insert(&( 2,-1), Tile::Wall);
gd.insert(&( 0, 0), Tile::Empty); // right
gd.insert(&( 1, 0), Tile::Empty); // down
gd.insert(&( 1, 1), Tile::Empty); // down
gd.insert(&( 1, 2), Tile::Empty); // down
gd.insert(&( 1, 3), Tile::Empty); // down
gd.insert(&( 1, 4), Tile::Empty); // down
gd.insert(&( 2, 4), Tile::Empty); // down
gd.insert(&( 3, 4), Tile::Empty); // down
gd.insert(&( 4, 4), Tile::Empty); // down
gd.insert(&( 4, 3), Tile::Empty); // down
gd.insert(&( 4, 2), Tile::Empty); // down
gd.insert(&( 4, 1), Tile::Empty); // down
gd.insert(&( 1,-1), Tile::Empty); // down
gd.insert(&( 1,-2), Tile::Empty); // down
gd.insert(&( 2,-2), Tile::Empty); // right
gd.insert(&( 3,-2), Tile::Empty); // right
gd.insert(&( 3,-1), Tile::Empty);
gd.insert(&( 3, 0), Tile::Empty);
gd.insert(&( 4, 0), Tile::Empty);
println!("{}", gd);
assert_eq!(vec![Right, Down, Down, Right, Right, Up, Up, Right], path_to(&gd, &( 0, 0), &( 4, 0)).unwrap());
}
fn adjacent<'a>(gd: &'a GameDisplay<Tile>, pos: &'a (Word, Word)) -> impl Iterator<Item = ((Word, Word), Direction)> + 'a {
Direction::all()
.into_iter()
.map(move |d| (pos.step_in_direction(&d), d))
.filter_map(move |(p2, d)| gd.get(&p2).map(|t| (p2, d, t)))
//.inspect(|x| println!(" c: {:?}", x))
.filter_map(|(p2, d, t)| match t {
&Tile::Empty | &Tile::Robot | &Tile::Oxygen => Some((p2, d)),
_ => None,
})
//.inspect(|x| println!(" d: {:?}", x))
}
struct Robot {
program: Program<'static>,
regs: Option<Registers>,
pos: (Word, Word),
moves: usize,
}
impl Robot {
fn new(data: Vec<Word>) -> Self {
let mem = intcode::Memory::from(data).with_memory_expansion();
let program = Program::from(mem);
Robot {
program,
regs: Some(Registers::default()),
pos: (0, 0),
moves: 0,
}
}
fn position(&self) -> &(Word, Word) {
&self.pos
}
fn try_move(&mut self, dir: Direction) -> ((Word, Word), Tile) {
loop {
let mut ret = None;
self.regs = Some(match self.program.eval_from_instruction(self.regs.take().unwrap()).unwrap() {
ExecutionState::HaltedAt(regs) => unreachable!("Halted at: {:?}", regs),
ExecutionState::Paused(regs) => unreachable!("Paused? {:?}", regs),
ExecutionState::InputIO(io) => {
let val: i64 = dir.into();
//println!("robot <-- {}", val);
self.program.handle_input_completion(io, val).unwrap()
},
ExecutionState::OutputIO(io, value) => {
//println!("robot --> {}", value);
let moved = value != 0;
let | }
}
for (p2, dir) in adjacent(gd, &p) {
let alt = steps_here + 1; | random_line_split |
booting.py | (msg)
# should we run configure for all the wan devices? or just wan?
for x in devices:
# if isinstance(x, DebianWAN): # does not work for mitm
if hasattr(x, "name") and "wan" in x.name:
logger.info(f"Configuring {x.name}")
x.configure(config=config)
# if more than 1 tftp server should we start them all?
# currently starting the 1 being used
logger.info(f"Starting TFTP server on {tftp_device.name}")
tftp_device.start_tftp_server()
devices.board.tftp_device = tftp_device
def pre_boot_lan_clients(config, env_helper, devices):
for x in devices.lan_clients:
if isinstance(x, DebianLAN):
logger.info(f"Configuring {x.name}")
x.configure()
def pre_boot_wlan_clients(config, env_helper, devices):
for x in getattr(devices, "wlan_clients", []):
logger.info(f"Configuring {x.name}")
x.configure()
def pre_boot_board(config, env_helper, devices):
env_cwmp_v = env_helper.get_cwmp_version()
if env_cwmp_v:
assert Version(env_cwmp_v) == Version(
devices.board.cwmp_version()
), f"CWMP version mismatch, Expected version {env_cwmp_v}"
def pre_boot_env(config, env_helper, devices):
# this should take care of provisioner/tr069/voice/etc
# depending on what the env_helperd has configured
if env_helper.mitm_enabled() and not hasattr(devices, "mitm"):
raise DeviceDoesNotExistError("No mitm device (requested by environment)")
cm_boot_file = None
mta_boot_file = None
if env_helper.has_board_boot_file():
cm_boot_file = env_helper.get_board_boot_file()
if env_helper.has_board_boot_file_mta():
mta_boot_file = env_helper.get_board_boot_file_mta()
devices.board.env_config(cm_boot_file, mta_boot_file, devices.board.mibs_path)
if env_helper.voice_enabled():
dev_list = [
devices.sipcenter,
devices.softphone,
]
if env_helper.get_external_voip():
dev_list.append(devices.softphone2)
dev_list.append(getattr(devices, "FXS", [devices.lan, devices.lan2]))
boardfarm.lib.voice.voice_configure(
dev_list,
devices.sipcenter,
config,
)
prov = getattr(config, "provisioner", None)
if prov:
if env_helper.vendor_encap_opts(ip_proto="ipv4"):
devices.provisioner.vendor_opts_acsv4_url = True
if env_helper.vendor_encap_opts(ip_proto="ipv6"):
devices.provisioner.vendor_opts_acsv6_url = True
devices.provisioner.valid_route_gateway = env_helper.is_route_gateway_valid()
logger.info("Provisioning board")
provision_board()
else:
# should this be an error?
logger.error(
colored(
"No provisioner found! Board provisioned skipped",
color="yellow",
attrs=["bold"],
)
)
if hasattr(devices.board, "pre_boot_env"):
devices.board.pre_boot_env(env_helper)
pre_boot_actions = {
"wan_clients_pre_boot": pre_boot_wan_clients,
"lan_clients_pre_boot": pre_boot_lan_clients,
"wlan_clients_pre_boot": pre_boot_wlan_clients,
"board_pre_boot": pre_boot_board,
"environment_pre_boot": pre_boot_env,
}
def boot_board(config, env_helper, devices):
try:
devices.board.reset()
if env_helper.get_software():
devices.board.flash(env_helper)
# store the timestamp, for uptime check later (in case the board
# crashes on boot)
devices.board.__reset__timestamp = time.time()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
except Exception as e:
logger.critical(colored("\n\nFailed to Boot", color="red", attrs=["bold"]))
logger.error(e)
raise BootFail
boot_actions = {"board_boot": boot_board}
def _wait_for_cm_online(config, env_helper, devices):
for _ in range(180):
if (
devices.cmts.is_cm_online(
ignore_bpi=is_bpi_privacy_disabled(), ignore_partial=True
)
is False
):
# show the arm prompt as it is a log in itself
devices.board.touch()
time.sleep(15)
continue
if devices.board.finalize_boot():
break
logger.info("######Rebooting######")
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
else:
msg = "\n\nFailed to Boot: board not online on CMTS"
logger.critical(msg)
raise BootFail(msg)
def post_boot_board(config, env_helper, devices):
try:
_wait_for_cm_online(config, env_helper, devices)
except BootFail:
logger.warning(
colored(
"Board not online on CMTS, resetting and retrying",
color="yellow",
attrs=["bold"],
)
)
devices.board.reset()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
_wait_for_cm_online(config, env_helper, devices)
devices.board.post_boot_init()
board_uptime = devices.board.get_seconds_uptime()
logger.info(f"Time up: {board_uptime}")
if hasattr(devices.board, "__reset__timestamp"):
time_elapsed = time.time() - devices.board.__reset__timestamp
logger.info(f"Time since reboot: {time_elapsed}")
if time_elapsed < board_uptime:
# TODO: the following should be an exception and not
# just a print!!!!
logger.warning("Error: possibly the board did not reset!")
if (time_elapsed - board_uptime) > 60:
logger.warning(
colored(
"Board may have rebooted multiple times after flashing process",
color="yellow",
attrs=["bold"],
)
)
if isinstance(devices.board, DocsisCPE):
pass # maybe new method to be added
else:
# the old way for legacy
devices.board.check_valid_docsis_ip_networking(strict=False)
def post_boot_wan_clients(config, env_helper, devices):
pass
def post_boot_lan_clients(config, env_helper, devices):
for i, v in enumerate(devices.board.dev.lan_clients):
if getattr(env_helper, "has_lan_advertise_identity", None):
for option in ["125", "17"]:
if env_helper.has_lan_advertise_identity(i):
v.configure_dhclient(([option, True],))
else:
v.configure_dhclient(([option, False],))
if devices.board.routing and config.setup_device_networking:
for idx, x in enumerate(devices.board.dev.lan_clients):
if isinstance(x, DebianLAN): # should this use devices.lan_clients?
logger.info(f"Starting LAN client on {x.name}")
for n in range(3):
try:
x.configure_docker_iface()
if env_helper.get_prov_mode() == "ipv6":
x.start_ipv6_lan_client(wan_gw=devices.wan.gw)
if (
devices.board.cm_cfg.dslite
and env_helper.is_dhcpv4_enabled_on_lan()
):
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
elif env_helper.get_prov_mode() == "dual":
x.start_ipv6_lan_client(wan_gw=devices.wan.gw)
if env_helper.is_dhcpv4_enabled_on_lan():
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
elif env_helper.is_dhcpv4_enabled_on_lan():
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
x.configure_proxy_pkgs()
if env_helper.get_prov_mode() in [
"ipv4",
"dual",
] and env_helper.is_set_static_ipv4(idx):
# set static ip address
set_static_ip_and_default_gw(
client=x,
)
break
except Exception as e:
logger.warning(e)
logger.error(
colored(
f"Failed to start lan client on '{x.name}' device, attempt #{n}",
color="red",
attrs=["bold"],
)
)
time.sleep(10)
else:
msg = f"Failed to start lan client on {x.name}"
logger.warning(colored(msg, color="yellow", attrs=["bold"]))
# do not fail the boot with raise BootFail(msg)
# reason: the board config may be such that the
# clients are not getting an ip (see LLCs)
def post_boot_wlan_clients(config, env_helper, devices):
wifi_clients = env_helper.wifi_clients()
if wifi_clients:
# Register all wifi clients in wifi manager
for client in wifi_clients:
|
# Start to connect all clients after registartions done:
for client in wifi_clients:
check_and_connect_to_wifi(devices, client)
logger.info(colored("\nWlan clients:", color="green"))
devices.wlan_clients.registered_clients_summary()
| devices.wlan_clients.register(client) | conditional_block |
booting.py | devices.sipcenter,
devices.softphone,
]
if env_helper.get_external_voip():
dev_list.append(devices.softphone2)
dev_list.append(getattr(devices, "FXS", [devices.lan, devices.lan2]))
boardfarm.lib.voice.voice_configure(
dev_list,
devices.sipcenter,
config,
)
prov = getattr(config, "provisioner", None)
if prov:
if env_helper.vendor_encap_opts(ip_proto="ipv4"):
devices.provisioner.vendor_opts_acsv4_url = True
if env_helper.vendor_encap_opts(ip_proto="ipv6"):
devices.provisioner.vendor_opts_acsv6_url = True
devices.provisioner.valid_route_gateway = env_helper.is_route_gateway_valid()
logger.info("Provisioning board")
provision_board()
else:
# should this be an error?
logger.error(
colored(
"No provisioner found! Board provisioned skipped",
color="yellow",
attrs=["bold"],
)
)
if hasattr(devices.board, "pre_boot_env"):
devices.board.pre_boot_env(env_helper)
pre_boot_actions = {
"wan_clients_pre_boot": pre_boot_wan_clients,
"lan_clients_pre_boot": pre_boot_lan_clients,
"wlan_clients_pre_boot": pre_boot_wlan_clients,
"board_pre_boot": pre_boot_board,
"environment_pre_boot": pre_boot_env,
}
def boot_board(config, env_helper, devices):
try:
devices.board.reset()
if env_helper.get_software():
devices.board.flash(env_helper)
# store the timestamp, for uptime check later (in case the board
# crashes on boot)
devices.board.__reset__timestamp = time.time()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
except Exception as e:
logger.critical(colored("\n\nFailed to Boot", color="red", attrs=["bold"]))
logger.error(e)
raise BootFail
boot_actions = {"board_boot": boot_board}
def _wait_for_cm_online(config, env_helper, devices):
for _ in range(180):
if (
devices.cmts.is_cm_online(
ignore_bpi=is_bpi_privacy_disabled(), ignore_partial=True
)
is False
):
# show the arm prompt as it is a log in itself
devices.board.touch()
time.sleep(15)
continue
if devices.board.finalize_boot():
break
logger.info("######Rebooting######")
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
else:
msg = "\n\nFailed to Boot: board not online on CMTS"
logger.critical(msg)
raise BootFail(msg)
def post_boot_board(config, env_helper, devices):
try:
_wait_for_cm_online(config, env_helper, devices)
except BootFail:
logger.warning(
colored(
"Board not online on CMTS, resetting and retrying",
color="yellow",
attrs=["bold"],
)
)
devices.board.reset()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
_wait_for_cm_online(config, env_helper, devices)
devices.board.post_boot_init()
board_uptime = devices.board.get_seconds_uptime()
logger.info(f"Time up: {board_uptime}")
if hasattr(devices.board, "__reset__timestamp"):
time_elapsed = time.time() - devices.board.__reset__timestamp
logger.info(f"Time since reboot: {time_elapsed}")
if time_elapsed < board_uptime:
# TODO: the following should be an exception and not
# just a print!!!!
logger.warning("Error: possibly the board did not reset!")
if (time_elapsed - board_uptime) > 60:
logger.warning(
colored(
"Board may have rebooted multiple times after flashing process",
color="yellow",
attrs=["bold"],
)
)
if isinstance(devices.board, DocsisCPE):
pass # maybe new method to be added
else:
# the old way for legacy
devices.board.check_valid_docsis_ip_networking(strict=False)
def post_boot_wan_clients(config, env_helper, devices):
pass
def post_boot_lan_clients(config, env_helper, devices):
for i, v in enumerate(devices.board.dev.lan_clients):
if getattr(env_helper, "has_lan_advertise_identity", None):
for option in ["125", "17"]:
if env_helper.has_lan_advertise_identity(i):
v.configure_dhclient(([option, True],))
else:
v.configure_dhclient(([option, False],))
if devices.board.routing and config.setup_device_networking:
for idx, x in enumerate(devices.board.dev.lan_clients):
if isinstance(x, DebianLAN): # should this use devices.lan_clients?
logger.info(f"Starting LAN client on {x.name}")
for n in range(3):
try:
x.configure_docker_iface()
if env_helper.get_prov_mode() == "ipv6":
x.start_ipv6_lan_client(wan_gw=devices.wan.gw)
if (
devices.board.cm_cfg.dslite
and env_helper.is_dhcpv4_enabled_on_lan()
):
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
elif env_helper.get_prov_mode() == "dual":
x.start_ipv6_lan_client(wan_gw=devices.wan.gw)
if env_helper.is_dhcpv4_enabled_on_lan():
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
elif env_helper.is_dhcpv4_enabled_on_lan():
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
x.configure_proxy_pkgs()
if env_helper.get_prov_mode() in [
"ipv4",
"dual",
] and env_helper.is_set_static_ipv4(idx):
# set static ip address
set_static_ip_and_default_gw(
client=x,
)
break
except Exception as e:
logger.warning(e)
logger.error(
colored(
f"Failed to start lan client on '{x.name}' device, attempt #{n}",
color="red",
attrs=["bold"],
)
)
time.sleep(10)
else:
msg = f"Failed to start lan client on {x.name}"
logger.warning(colored(msg, color="yellow", attrs=["bold"]))
# do not fail the boot with raise BootFail(msg)
# reason: the board config may be such that the
# clients are not getting an ip (see LLCs)
def post_boot_wlan_clients(config, env_helper, devices):
wifi_clients = env_helper.wifi_clients()
if wifi_clients:
# Register all wifi clients in wifi manager
for client in wifi_clients:
devices.wlan_clients.register(client)
# Start to connect all clients after registartions done:
for client in wifi_clients:
check_and_connect_to_wifi(devices, client)
logger.info(colored("\nWlan clients:", color="green"))
devices.wlan_clients.registered_clients_summary()
def post_boot_env(config, env_helper, devices):
if env_helper.mitm_enabled():
activate_mitm(devices, env_helper.get_mitm_devices())
eMTA_iface_status = env_helper.get_emta_interface_status()
if eMTA_iface_status:
devices.board.set_eMTA_interface(devices.board.mta_iface, eMTA_iface_status)
if env_helper.voice_enabled() and eMTA_iface_status != "down":
devices.board.wait_for_mta_provisioning()
register_fxs_details(getattr(devices, "FXS", []), devices.board)
cfg = env_helper.get_prov_mode()
tr069check = cfg not in ["disabled", "bridge", "none"]
tr069provision = env_helper.get_tr069_provisioning()
if tr069check:
for _ in range(20):
try:
devices.board.get_cpeid()
break
except Exception as e:
logger.error(e)
warnings.warn("Failed to connect to ACS, retrying")
time.sleep(10)
else:
raise BootFail("Failed to connect to ACS")
if tr069provision:
reset_val = any(
x in env_helper.get_software()
for x in [
"factory_reset",
"pre_flash_factory_reset",
]
)
if reset_val:
for i in tr069provision:
for acs_api in i:
API_func = getattr(devices.acs_server, acs_api)
for param in i[acs_api]:
API_func(param)
else:
raise BootFail(
"Factory reset has to performed for tr069 provisioning. Env json with factory reset true should be used."
)
if hasattr(devices.board, "post_boot_env"):
devices.board.post_boot_env()
post_boot_actions = {
"board_post_boot": post_boot_board,
"wan_clients_post_boot": post_boot_wan_clients,
"lan_clients_post_boot": post_boot_lan_clients,
"environment_post_boot": post_boot_env,
"wlan_clients_connection": post_boot_wlan_clients,
}
def | run_actions | identifier_name | |
booting.py | import boardfarm.lib.booting
import boardfarm.lib.voice
from boardfarm.devices.debian_lan import DebianLAN
from boardfarm.exceptions import (
BootFail,
CodeError,
DeviceDoesNotExistError,
NoTFTPServer,
)
from boardfarm.lib.booting_utils import check_and_connect_to_wifi
from boardfarm.library import check_devices
from packaging.version import Version
from termcolor import colored
from boardfarm_docsis.devices.base_devices.board import DocsisCPE
from boardfarm_docsis.lib.booting_utils import (
activate_mitm,
register_fxs_details,
set_static_ip_and_default_gw,
)
from boardfarm_docsis.lib.dns_helper import dns_acs_config
from boardfarm_docsis.use_cases.cmts_interactions import is_bpi_privacy_disabled
from boardfarm_docsis.use_cases.provision_helper import provision_board
logger = logging.getLogger("bft")
def pre_boot_wan_clients(config, env_helper, devices):
if env_helper.get_dns_dict():
# to get reachable and unreachable ips for ACS DNS
devices.wan.auth_dns = True
dns_acs_config(devices, env_helper.get_dns_dict())
tftp_device, tftp_servers = boardfarm.lib.booting.get_tftp(config)
if not tftp_servers:
logger.error(colored("No tftp server found", color="red", attrs=["bold"]))
# currently we must have at least 1 tftp server configured
raise NoTFTPServer
if len(tftp_servers) > 1:
msg = f"Found more than 1 tftp server: {tftp_servers}, using {tftp_device.name}"
logger.error(colored(msg, color="red", attrs=["bold"]))
raise CodeError(msg)
# should we run configure for all the wan devices? or just wan?
for x in devices:
# if isinstance(x, DebianWAN): # does not work for mitm
if hasattr(x, "name") and "wan" in x.name:
logger.info(f"Configuring {x.name}")
x.configure(config=config)
# if more than 1 tftp server should we start them all?
# currently starting the 1 being used
logger.info(f"Starting TFTP server on {tftp_device.name}")
tftp_device.start_tftp_server()
devices.board.tftp_device = tftp_device
def pre_boot_lan_clients(config, env_helper, devices):
for x in devices.lan_clients:
if isinstance(x, DebianLAN):
logger.info(f"Configuring {x.name}")
x.configure()
def pre_boot_wlan_clients(config, env_helper, devices):
for x in getattr(devices, "wlan_clients", []):
logger.info(f"Configuring {x.name}")
x.configure()
def pre_boot_board(config, env_helper, devices):
env_cwmp_v = env_helper.get_cwmp_version()
if env_cwmp_v:
assert Version(env_cwmp_v) == Version(
devices.board.cwmp_version()
), f"CWMP version mismatch, Expected version {env_cwmp_v}"
def pre_boot_env(config, env_helper, devices):
# this should take care of provisioner/tr069/voice/etc
# depending on what the env_helperd has configured
if env_helper.mitm_enabled() and not hasattr(devices, "mitm"):
raise DeviceDoesNotExistError("No mitm device (requested by environment)")
cm_boot_file = None
mta_boot_file = None
if env_helper.has_board_boot_file():
cm_boot_file = env_helper.get_board_boot_file()
if env_helper.has_board_boot_file_mta():
mta_boot_file = env_helper.get_board_boot_file_mta()
devices.board.env_config(cm_boot_file, mta_boot_file, devices.board.mibs_path)
if env_helper.voice_enabled():
dev_list = [
devices.sipcenter,
devices.softphone,
]
if env_helper.get_external_voip():
dev_list.append(devices.softphone2)
dev_list.append(getattr(devices, "FXS", [devices.lan, devices.lan2]))
boardfarm.lib.voice.voice_configure(
dev_list,
devices.sipcenter,
config,
)
prov = getattr(config, "provisioner", None)
if prov:
if env_helper.vendor_encap_opts(ip_proto="ipv4"):
devices.provisioner.vendor_opts_acsv4_url = True
if env_helper.vendor_encap_opts(ip_proto="ipv6"):
devices.provisioner.vendor_opts_acsv6_url = True
devices.provisioner.valid_route_gateway = env_helper.is_route_gateway_valid()
logger.info("Provisioning board")
provision_board()
else:
# should this be an error?
logger.error(
colored(
"No provisioner found! Board provisioned skipped",
color="yellow",
attrs=["bold"],
)
)
if hasattr(devices.board, "pre_boot_env"):
devices.board.pre_boot_env(env_helper)
pre_boot_actions = {
"wan_clients_pre_boot": pre_boot_wan_clients,
"lan_clients_pre_boot": pre_boot_lan_clients,
"wlan_clients_pre_boot": pre_boot_wlan_clients,
"board_pre_boot": pre_boot_board,
"environment_pre_boot": pre_boot_env,
}
def boot_board(config, env_helper, devices):
try:
devices.board.reset()
if env_helper.get_software():
devices.board.flash(env_helper)
# store the timestamp, for uptime check later (in case the board
# crashes on boot)
devices.board.__reset__timestamp = time.time()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
except Exception as e:
logger.critical(colored("\n\nFailed to Boot", color="red", attrs=["bold"]))
logger.error(e)
raise BootFail
boot_actions = {"board_boot": boot_board}
def _wait_for_cm_online(config, env_helper, devices):
for _ in range(180):
if (
devices.cmts.is_cm_online(
ignore_bpi=is_bpi_privacy_disabled(), ignore_partial=True
)
is False
):
# show the arm prompt as it is a log in itself
devices.board.touch()
time.sleep(15)
continue
if devices.board.finalize_boot():
break
logger.info("######Rebooting######")
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
else:
msg = "\n\nFailed to Boot: board not online on CMTS"
logger.critical(msg)
raise BootFail(msg)
def post_boot_board(config, env_helper, devices):
try:
_wait_for_cm_online(config, env_helper, devices)
except BootFail:
logger.warning(
colored(
"Board not online on CMTS, resetting and retrying",
color="yellow",
attrs=["bold"],
)
)
devices.board.reset()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
_wait_for_cm_online(config, env_helper, devices)
devices.board.post_boot_init()
board_uptime = devices.board.get_seconds_uptime()
logger.info(f"Time up: {board_uptime}")
if hasattr(devices.board, "__reset__timestamp"):
time_elapsed = time.time() - devices.board.__reset__timestamp
logger.info(f"Time since reboot: {time_elapsed}")
if time_elapsed < board_uptime:
# TODO: the following should be an exception and not
# just a print!!!!
logger.warning("Error: possibly the board did not reset!")
if (time_elapsed - board_uptime) > 60:
logger.warning(
colored(
"Board may have rebooted multiple times after flashing process",
color="yellow",
attrs=["bold"],
)
)
if isinstance(devices.board, DocsisCPE):
pass # maybe new method to be added
else:
# the old way for legacy
devices.board.check_valid_docsis_ip_networking(strict=False)
def post_boot_wan_clients(config, env_helper, devices):
pass
def post_boot_lan_clients(config, env_helper, devices):
for i, v in enumerate(devices.board.dev.lan_clients):
if getattr(env_helper, "has_lan_advertise_identity", None):
for option in ["125", "17"]:
if env_helper.has_lan_advertise_identity(i):
v.configure_dhclient(([option, True],))
else:
v.configure_dhclient(([option, False],))
if devices.board.routing and config.setup_device_networking:
for idx, x in enumerate(devices.board.dev.lan_clients):
if isinstance(x, DebianLAN): # should this use devices.lan_clients?
logger.info(f"Starting LAN client on {x.name}")
for n in range(3):
try:
x.configure_docker_iface()
if env_helper.get_prov_mode() == "ipv6":
x.start_ipv6_lan_client(wan_gw=devices.wan.gw)
if (
devices.board.cm_cfg.dslite
and env_helper.is_dhcpv4_enabled_on_lan()
):
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
elif env_helper.get_prov_mode() == "dual":
x.start_ipv6_lan_client(wan | import time
import traceback
import warnings
| random_line_split | |
booting.py | x.configure(config=config)
# if more than 1 tftp server should we start them all?
# currently starting the 1 being used
logger.info(f"Starting TFTP server on {tftp_device.name}")
tftp_device.start_tftp_server()
devices.board.tftp_device = tftp_device
def pre_boot_lan_clients(config, env_helper, devices):
for x in devices.lan_clients:
if isinstance(x, DebianLAN):
logger.info(f"Configuring {x.name}")
x.configure()
def pre_boot_wlan_clients(config, env_helper, devices):
for x in getattr(devices, "wlan_clients", []):
logger.info(f"Configuring {x.name}")
x.configure()
def pre_boot_board(config, env_helper, devices):
env_cwmp_v = env_helper.get_cwmp_version()
if env_cwmp_v:
assert Version(env_cwmp_v) == Version(
devices.board.cwmp_version()
), f"CWMP version mismatch, Expected version {env_cwmp_v}"
def pre_boot_env(config, env_helper, devices):
# this should take care of provisioner/tr069/voice/etc
# depending on what the env_helperd has configured
if env_helper.mitm_enabled() and not hasattr(devices, "mitm"):
raise DeviceDoesNotExistError("No mitm device (requested by environment)")
cm_boot_file = None
mta_boot_file = None
if env_helper.has_board_boot_file():
cm_boot_file = env_helper.get_board_boot_file()
if env_helper.has_board_boot_file_mta():
mta_boot_file = env_helper.get_board_boot_file_mta()
devices.board.env_config(cm_boot_file, mta_boot_file, devices.board.mibs_path)
if env_helper.voice_enabled():
dev_list = [
devices.sipcenter,
devices.softphone,
]
if env_helper.get_external_voip():
dev_list.append(devices.softphone2)
dev_list.append(getattr(devices, "FXS", [devices.lan, devices.lan2]))
boardfarm.lib.voice.voice_configure(
dev_list,
devices.sipcenter,
config,
)
prov = getattr(config, "provisioner", None)
if prov:
if env_helper.vendor_encap_opts(ip_proto="ipv4"):
devices.provisioner.vendor_opts_acsv4_url = True
if env_helper.vendor_encap_opts(ip_proto="ipv6"):
devices.provisioner.vendor_opts_acsv6_url = True
devices.provisioner.valid_route_gateway = env_helper.is_route_gateway_valid()
logger.info("Provisioning board")
provision_board()
else:
# should this be an error?
logger.error(
colored(
"No provisioner found! Board provisioned skipped",
color="yellow",
attrs=["bold"],
)
)
if hasattr(devices.board, "pre_boot_env"):
devices.board.pre_boot_env(env_helper)
pre_boot_actions = {
"wan_clients_pre_boot": pre_boot_wan_clients,
"lan_clients_pre_boot": pre_boot_lan_clients,
"wlan_clients_pre_boot": pre_boot_wlan_clients,
"board_pre_boot": pre_boot_board,
"environment_pre_boot": pre_boot_env,
}
def boot_board(config, env_helper, devices):
try:
devices.board.reset()
if env_helper.get_software():
devices.board.flash(env_helper)
# store the timestamp, for uptime check later (in case the board
# crashes on boot)
devices.board.__reset__timestamp = time.time()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
except Exception as e:
logger.critical(colored("\n\nFailed to Boot", color="red", attrs=["bold"]))
logger.error(e)
raise BootFail
boot_actions = {"board_boot": boot_board}
def _wait_for_cm_online(config, env_helper, devices):
for _ in range(180):
if (
devices.cmts.is_cm_online(
ignore_bpi=is_bpi_privacy_disabled(), ignore_partial=True
)
is False
):
# show the arm prompt as it is a log in itself
devices.board.touch()
time.sleep(15)
continue
if devices.board.finalize_boot():
break
logger.info("######Rebooting######")
devices.cmts.clear_cm_reset(devices.board.cm_mac)
time.sleep(20)
else:
msg = "\n\nFailed to Boot: board not online on CMTS"
logger.critical(msg)
raise BootFail(msg)
def post_boot_board(config, env_helper, devices):
try:
_wait_for_cm_online(config, env_helper, devices)
except BootFail:
logger.warning(
colored(
"Board not online on CMTS, resetting and retrying",
color="yellow",
attrs=["bold"],
)
)
devices.board.reset()
devices.cmts.clear_cm_reset(devices.board.cm_mac)
_wait_for_cm_online(config, env_helper, devices)
devices.board.post_boot_init()
board_uptime = devices.board.get_seconds_uptime()
logger.info(f"Time up: {board_uptime}")
if hasattr(devices.board, "__reset__timestamp"):
time_elapsed = time.time() - devices.board.__reset__timestamp
logger.info(f"Time since reboot: {time_elapsed}")
if time_elapsed < board_uptime:
# TODO: the following should be an exception and not
# just a print!!!!
logger.warning("Error: possibly the board did not reset!")
if (time_elapsed - board_uptime) > 60:
logger.warning(
colored(
"Board may have rebooted multiple times after flashing process",
color="yellow",
attrs=["bold"],
)
)
if isinstance(devices.board, DocsisCPE):
pass # maybe new method to be added
else:
# the old way for legacy
devices.board.check_valid_docsis_ip_networking(strict=False)
def post_boot_wan_clients(config, env_helper, devices):
pass
def post_boot_lan_clients(config, env_helper, devices):
for i, v in enumerate(devices.board.dev.lan_clients):
if getattr(env_helper, "has_lan_advertise_identity", None):
for option in ["125", "17"]:
if env_helper.has_lan_advertise_identity(i):
v.configure_dhclient(([option, True],))
else:
v.configure_dhclient(([option, False],))
if devices.board.routing and config.setup_device_networking:
for idx, x in enumerate(devices.board.dev.lan_clients):
if isinstance(x, DebianLAN): # should this use devices.lan_clients?
logger.info(f"Starting LAN client on {x.name}")
for n in range(3):
try:
x.configure_docker_iface()
if env_helper.get_prov_mode() == "ipv6":
x.start_ipv6_lan_client(wan_gw=devices.wan.gw)
if (
devices.board.cm_cfg.dslite
and env_helper.is_dhcpv4_enabled_on_lan()
):
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
elif env_helper.get_prov_mode() == "dual":
x.start_ipv6_lan_client(wan_gw=devices.wan.gw)
if env_helper.is_dhcpv4_enabled_on_lan():
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
elif env_helper.is_dhcpv4_enabled_on_lan():
x.start_ipv4_lan_client(wan_gw=devices.wan.gw)
x.configure_proxy_pkgs()
if env_helper.get_prov_mode() in [
"ipv4",
"dual",
] and env_helper.is_set_static_ipv4(idx):
# set static ip address
set_static_ip_and_default_gw(
client=x,
)
break
except Exception as e:
logger.warning(e)
logger.error(
colored(
f"Failed to start lan client on '{x.name}' device, attempt #{n}",
color="red",
attrs=["bold"],
)
)
time.sleep(10)
else:
msg = f"Failed to start lan client on { | if env_helper.get_dns_dict():
# to get reachable and unreachable ips for ACS DNS
devices.wan.auth_dns = True
dns_acs_config(devices, env_helper.get_dns_dict())
tftp_device, tftp_servers = boardfarm.lib.booting.get_tftp(config)
if not tftp_servers:
logger.error(colored("No tftp server found", color="red", attrs=["bold"]))
# currently we must have at least 1 tftp server configured
raise NoTFTPServer
if len(tftp_servers) > 1:
msg = f"Found more than 1 tftp server: {tftp_servers}, using {tftp_device.name}"
logger.error(colored(msg, color="red", attrs=["bold"]))
raise CodeError(msg)
# should we run configure for all the wan devices? or just wan?
for x in devices:
# if isinstance(x, DebianWAN): # does not work for mitm
if hasattr(x, "name") and "wan" in x.name:
logger.info(f"Configuring {x.name}") | identifier_body | |
transaction.go | () bool {
return f.FCTInputs != nil && // This array should not be nil
f.Signatures != nil &&
!f.TimestampSalt.IsZero()
}
// IsPopulated returns true if s has already been successfully populated by a
// call to Get. IsPopulated returns false if s.SignatureBlock or
// s.ReedeemCondition are nil
func (s FactoidTransactionSignature) IsPopulated() bool {
return s.SignatureBlock != nil
}
// Valid returns if the inputs of the factoid transaction are properly signed
// by the redeem conditions. It will also validate the total inputs is greater
// than the total outputs.
func (f *FactoidTransaction) Valid() bool {
if !f.IsPopulated() {
return false
}
// Validate amounts
if f.TotalFCTInputs() < f.TotalFCTOutputs()+f.TotalECOutput() {
return false
}
// Validate signatures
if len(f.FCTInputs) != len(f.Signatures) {
return false
}
msg, err := f.MarshalLedgerBinary()
if err != nil {
return false
}
for i := range f.FCTInputs {
expAddr := f.Signatures[i].ReedeemCondition.Address()
// RCD should match the input
if bytes.Compare(expAddr[:], f.FCTInputs[i].Address[:]) != 0 {
return false
}
if !f.Signatures[i].Validate(msg) {
return false
}
}
return true
}
func (f *FactoidTransaction) TotalFCTInputs() (total uint64) {
return factoidTransactionIOs(f.FCTInputs).TotalAmount()
}
func (f *FactoidTransaction) TotalFCTOutputs() (total uint64) {
return factoidTransactionIOs(f.FCTOutputs).TotalAmount()
}
// TotalECOutput is delimated in factoishis
func (f *FactoidTransaction) TotalECOutput() (total uint64) {
return factoidTransactionIOs(f.ECOutputs).TotalAmount()
}
func (s factoidTransactionIOs) TotalAmount() (total uint64) {
for _, io := range s {
total += io.Amount
}
return
}
func (s FactoidTransactionSignature) Validate(msg Bytes) bool {
return s.ReedeemCondition.Validate(msg, s.SignatureBlock)
}
// Get queries factomd for the entry corresponding to f.TransactionID, which
// must be not nil. After a successful call all inputs, outputs, and
// the header will be populated
func (f *FactoidTransaction) Get(ctx context.Context, c *Client) error {
// TODO: Test this functionality
// If the TransactionID is nil then we have nothing to query for.
if f.TransactionID == nil {
return fmt.Errorf("txid is nil")
}
// If the Transaction is already populated then there is nothing to do. If
// the Hash is nil, we cannot populate it anyway.
if f.IsPopulated() {
return nil
}
params := struct {
Hash *Bytes32 `json:"hash"`
}{Hash: f.TransactionID}
var result struct {
Data Bytes `json:"data"`
}
if err := c.FactomdRequest(ctx, "raw-data", params, &result); err != nil {
return err
}
if err := f.UnmarshalBinary(result.Data); err != nil {
return err
}
return nil
}
// ComputeTransactionID computes the txid for a given transaction. The txid is
// the sha256 of the ledger fields in a factoid transaction. The ledger fields
// exclude the signature block of the transaction
func (f *FactoidTransaction) ComputeTransactionID() (Bytes32, error) {
data, err := f.MarshalLedgerBinary()
if err != nil {
return Bytes32{}, err
}
return f.computeTransactionID(data)
}
func (f *FactoidTransaction) computeTransactionID(ledgerBinary Bytes) (Bytes32, error) {
txid := Bytes32(sha256.Sum256(ledgerBinary))
return txid, nil
}
// ComputeFullHash computes the fullhash for a given transaction. The fullhash
// is the sha256 of all the fields in a factoid transaction.
func (f *FactoidTransaction) ComputeFullHash() (*Bytes32, error) {
data, err := f.MarshalBinary()
if err != nil {
return nil, err
}
txid := Bytes32(sha256.Sum256(data))
return &txid, nil
}
// MarshalLedgerBinary marshals the transaction ledger fields to their
// binary representation. This excludes the signature blocks
func (f *FactoidTransaction) MarshalLedgerBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
// It's very difficult to know the size before marshaling, as
// each in/out has a varint so make the buffer at the end
// The header bytes
header, err := f.MarshalHeaderBinary()
if err != nil {
return nil, err
}
// Inputs
inputs, err := factoidTransactionIOs(f.FCTInputs).MarshalBinary()
if err != nil {
return nil, err
}
// FCT Outputs
fctout, err := factoidTransactionIOs(f.FCTOutputs).MarshalBinary()
if err != nil {
return nil, err
}
// EC Outputs
ecout, err := factoidTransactionIOs(f.ECOutputs).MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(header)+len(inputs)+len(fctout)+len(ecout))
var i int
i += copy(data[i:], header)
i += copy(data[i:], inputs)
i += copy(data[i:], fctout)
i += copy(data[i:], ecout)
return data, nil
}
// TODO: Re-eval how to do this. Kinda different from the rest
func (f *FactoidTransaction) MarshalBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
data, err := f.MarshalLedgerBinary()
if err != nil {
return nil, err
}
for _, s := range f.Signatures {
sig, err := s.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, sig...)
}
return data, nil
}
// MarshalHeaderBinary marshals the transaction's header to its binary
// representation. See UnmarshalHeaderBinary for encoding details.
func (f *FactoidTransaction) MarshalHeaderBinary() ([]byte, error) {
version := varintf.Encode(f.Version)
data := make([]byte, TransactionHeadMinLen+len(version))
var i int
i += copy(data[i:], version)
// Do the timestamp as 6 bytes in ms
ms := f.TimestampSalt.UnixNano() / 1e6
buf := bytes.NewBuffer(make([]byte, 0, 8))
if err := binary.Write(buf, binary.BigEndian, ms); err != nil {
return nil, err
}
i += copy(data[i:], buf.Bytes()[2:])
data[i] = uint8(len(f.FCTInputs))
i += 1
data[i] = uint8(len(f.FCTOutputs))
i += 1
data[i] = uint8(len(f.ECOutputs))
i += 1
return data, nil
}
// MarshalBinary marshals a set of transaction ios to its binary representation.
// See UnmarshalBinary for encoding details.
func (ios factoidTransactionIOs) MarshalBinary() ([]byte, error) {
var data []byte
for _, io := range ios {
iodata, err := io.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, iodata...)
}
return data, nil
}
// MarshalBinary marshals a transaction io to its binary representation.
// See UnmarshalBinary for encoding details.
func (io *FactoidTransactionIO) MarshalBinary() ([]byte, error) {
amount := varintf.Encode(io.Amount)
data := make([]byte, 32+len(amount))
var i int
i += copy(data[i:], amount)
i += copy(data[i:], io.Address[:])
return data, nil
}
// MarshalBinary marshals a transaction signature to its binary representation.
// See UnmarshalBinary for encoding details.
func (s *FactoidTransactionSignature) MarshalBinary() ([]byte, error) {
if !s.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
rcdData, err := s.ReedeemCondition.MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(rcdData)+len(s.SignatureBlock))
var i int
i += copy(data[i:], rcdData)
i += copy(data[i:], s.SignatureBlock)
return data, nil
}
const (
TransactionHeadMinLen = 0 + // Version length is varint
6 + // timestamp
1 + // input count
1 + // factoid output count
1 // EC output count
TransactionTotalMinLen = TransactionHeadMinLen // Coinbases have no body | IsPopulated | identifier_name | |
transaction.go | Valid returns if the inputs of the factoid transaction are properly signed
// by the redeem conditions. It will also validate the total inputs is greater
// than the total outputs.
func (f *FactoidTransaction) Valid() bool {
if !f.IsPopulated() {
return false
}
// Validate amounts
if f.TotalFCTInputs() < f.TotalFCTOutputs()+f.TotalECOutput() {
return false
}
// Validate signatures
if len(f.FCTInputs) != len(f.Signatures) {
return false
}
msg, err := f.MarshalLedgerBinary()
if err != nil {
return false
}
for i := range f.FCTInputs {
expAddr := f.Signatures[i].ReedeemCondition.Address()
// RCD should match the input
if bytes.Compare(expAddr[:], f.FCTInputs[i].Address[:]) != 0 {
return false
}
if !f.Signatures[i].Validate(msg) {
return false
}
}
return true
}
func (f *FactoidTransaction) TotalFCTInputs() (total uint64) {
return factoidTransactionIOs(f.FCTInputs).TotalAmount()
}
func (f *FactoidTransaction) TotalFCTOutputs() (total uint64) {
return factoidTransactionIOs(f.FCTOutputs).TotalAmount()
}
// TotalECOutput is delimated in factoishis
func (f *FactoidTransaction) TotalECOutput() (total uint64) {
return factoidTransactionIOs(f.ECOutputs).TotalAmount()
}
func (s factoidTransactionIOs) TotalAmount() (total uint64) {
for _, io := range s {
total += io.Amount
}
return
}
func (s FactoidTransactionSignature) Validate(msg Bytes) bool {
return s.ReedeemCondition.Validate(msg, s.SignatureBlock)
}
// Get queries factomd for the entry corresponding to f.TransactionID, which
// must be not nil. After a successful call all inputs, outputs, and
// the header will be populated
func (f *FactoidTransaction) Get(ctx context.Context, c *Client) error {
// TODO: Test this functionality
// If the TransactionID is nil then we have nothing to query for.
if f.TransactionID == nil {
return fmt.Errorf("txid is nil")
}
// If the Transaction is already populated then there is nothing to do. If
// the Hash is nil, we cannot populate it anyway.
if f.IsPopulated() {
return nil
}
params := struct {
Hash *Bytes32 `json:"hash"`
}{Hash: f.TransactionID}
var result struct {
Data Bytes `json:"data"`
}
if err := c.FactomdRequest(ctx, "raw-data", params, &result); err != nil {
return err
}
if err := f.UnmarshalBinary(result.Data); err != nil {
return err
}
return nil
}
// ComputeTransactionID computes the txid for a given transaction. The txid is
// the sha256 of the ledger fields in a factoid transaction. The ledger fields
// exclude the signature block of the transaction
func (f *FactoidTransaction) ComputeTransactionID() (Bytes32, error) {
data, err := f.MarshalLedgerBinary()
if err != nil {
return Bytes32{}, err
}
return f.computeTransactionID(data)
}
func (f *FactoidTransaction) computeTransactionID(ledgerBinary Bytes) (Bytes32, error) {
txid := Bytes32(sha256.Sum256(ledgerBinary))
return txid, nil
}
// ComputeFullHash computes the fullhash for a given transaction. The fullhash
// is the sha256 of all the fields in a factoid transaction.
func (f *FactoidTransaction) ComputeFullHash() (*Bytes32, error) {
data, err := f.MarshalBinary()
if err != nil {
return nil, err
}
txid := Bytes32(sha256.Sum256(data))
return &txid, nil
}
// MarshalLedgerBinary marshals the transaction ledger fields to their
// binary representation. This excludes the signature blocks
func (f *FactoidTransaction) MarshalLedgerBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
// It's very difficult to know the size before marshaling, as
// each in/out has a varint so make the buffer at the end
// The header bytes
header, err := f.MarshalHeaderBinary()
if err != nil {
return nil, err
}
// Inputs
inputs, err := factoidTransactionIOs(f.FCTInputs).MarshalBinary()
if err != nil {
return nil, err
}
// FCT Outputs
fctout, err := factoidTransactionIOs(f.FCTOutputs).MarshalBinary()
if err != nil {
return nil, err
}
// EC Outputs
ecout, err := factoidTransactionIOs(f.ECOutputs).MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(header)+len(inputs)+len(fctout)+len(ecout))
var i int
i += copy(data[i:], header)
i += copy(data[i:], inputs)
i += copy(data[i:], fctout)
i += copy(data[i:], ecout)
return data, nil
}
// TODO: Re-eval how to do this. Kinda different from the rest
func (f *FactoidTransaction) MarshalBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
data, err := f.MarshalLedgerBinary()
if err != nil {
return nil, err
}
for _, s := range f.Signatures |
return data, nil
}
// MarshalHeaderBinary marshals the transaction's header to its binary
// representation. See UnmarshalHeaderBinary for encoding details.
func (f *FactoidTransaction) MarshalHeaderBinary() ([]byte, error) {
version := varintf.Encode(f.Version)
data := make([]byte, TransactionHeadMinLen+len(version))
var i int
i += copy(data[i:], version)
// Do the timestamp as 6 bytes in ms
ms := f.TimestampSalt.UnixNano() / 1e6
buf := bytes.NewBuffer(make([]byte, 0, 8))
if err := binary.Write(buf, binary.BigEndian, ms); err != nil {
return nil, err
}
i += copy(data[i:], buf.Bytes()[2:])
data[i] = uint8(len(f.FCTInputs))
i += 1
data[i] = uint8(len(f.FCTOutputs))
i += 1
data[i] = uint8(len(f.ECOutputs))
i += 1
return data, nil
}
// MarshalBinary marshals a set of transaction ios to its binary representation.
// See UnmarshalBinary for encoding details.
func (ios factoidTransactionIOs) MarshalBinary() ([]byte, error) {
var data []byte
for _, io := range ios {
iodata, err := io.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, iodata...)
}
return data, nil
}
// MarshalBinary marshals a transaction io to its binary representation.
// See UnmarshalBinary for encoding details.
func (io *FactoidTransactionIO) MarshalBinary() ([]byte, error) {
amount := varintf.Encode(io.Amount)
data := make([]byte, 32+len(amount))
var i int
i += copy(data[i:], amount)
i += copy(data[i:], io.Address[:])
return data, nil
}
// MarshalBinary marshals a transaction signature to its binary representation.
// See UnmarshalBinary for encoding details.
func (s *FactoidTransactionSignature) MarshalBinary() ([]byte, error) {
if !s.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
rcdData, err := s.ReedeemCondition.MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(rcdData)+len(s.SignatureBlock))
var i int
i += copy(data[i:], rcdData)
i += copy(data[i:], s.SignatureBlock)
return data, nil
}
const (
TransactionHeadMinLen = 0 + // Version length is varint
6 + // timestamp
1 + // input count
1 + // factoid output count
1 // EC output count
TransactionTotalMinLen = TransactionHeadMinLen // Coinbases have no body
)
// Decode will consume as many bytes as necessary to unmarshal the factoid
// transaction. It will return the number of bytes read and an error.
func (f *FactoidTransaction) Decode(data []byte) (i int, err error) {
if len(data) < TransactionTotalMinLen {
return 0, fmt.Errorf("insufficient length")
}
// Decode header
version, i := varintf.Decode(data)
if i < 0 {
return 0 | {
sig, err := s.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, sig...)
} | conditional_block |
transaction.go |
// Valid returns if the inputs of the factoid transaction are properly signed
// by the redeem conditions. It will also validate the total inputs is greater
// than the total outputs.
func (f *FactoidTransaction) Valid() bool {
if !f.IsPopulated() {
return false
}
// Validate amounts
if f.TotalFCTInputs() < f.TotalFCTOutputs()+f.TotalECOutput() {
return false
}
// Validate signatures
if len(f.FCTInputs) != len(f.Signatures) {
return false
}
msg, err := f.MarshalLedgerBinary()
if err != nil {
return false
}
for i := range f.FCTInputs {
expAddr := f.Signatures[i].ReedeemCondition.Address()
// RCD should match the input
if bytes.Compare(expAddr[:], f.FCTInputs[i].Address[:]) != 0 {
return false
}
if !f.Signatures[i].Validate(msg) {
return false
}
}
return true
}
func (f *FactoidTransaction) TotalFCTInputs() (total uint64) {
return factoidTransactionIOs(f.FCTInputs).TotalAmount()
}
func (f *FactoidTransaction) TotalFCTOutputs() (total uint64) {
return factoidTransactionIOs(f.FCTOutputs).TotalAmount()
}
// TotalECOutput is delimated in factoishis
func (f *FactoidTransaction) TotalECOutput() (total uint64) {
return factoidTransactionIOs(f.ECOutputs).TotalAmount()
}
func (s factoidTransactionIOs) TotalAmount() (total uint64) {
for _, io := range s {
total += io.Amount
}
return
}
func (s FactoidTransactionSignature) Validate(msg Bytes) bool {
return s.ReedeemCondition.Validate(msg, s.SignatureBlock)
}
// Get queries factomd for the entry corresponding to f.TransactionID, which
// must be not nil. After a successful call all inputs, outputs, and
// the header will be populated
func (f *FactoidTransaction) Get(ctx context.Context, c *Client) error {
// TODO: Test this functionality
// If the TransactionID is nil then we have nothing to query for.
if f.TransactionID == nil {
return fmt.Errorf("txid is nil")
}
// If the Transaction is already populated then there is nothing to do. If
// the Hash is nil, we cannot populate it anyway.
if f.IsPopulated() {
return nil
}
params := struct {
Hash *Bytes32 `json:"hash"`
}{Hash: f.TransactionID}
var result struct {
Data Bytes `json:"data"`
}
if err := c.FactomdRequest(ctx, "raw-data", params, &result); err != nil {
return err
}
if err := f.UnmarshalBinary(result.Data); err != nil {
return err
}
return nil
}
// ComputeTransactionID computes the txid for a given transaction. The txid is
// the sha256 of the ledger fields in a factoid transaction. The ledger fields
// exclude the signature block of the transaction
func (f *FactoidTransaction) ComputeTransactionID() (Bytes32, error) {
data, err := f.MarshalLedgerBinary()
if err != nil {
return Bytes32{}, err
}
return f.computeTransactionID(data)
}
func (f *FactoidTransaction) computeTransactionID(ledgerBinary Bytes) (Bytes32, error) {
txid := Bytes32(sha256.Sum256(ledgerBinary))
return txid, nil
}
// ComputeFullHash computes the fullhash for a given transaction. The fullhash
// is the sha256 of all the fields in a factoid transaction.
func (f *FactoidTransaction) ComputeFullHash() (*Bytes32, error) {
data, err := f.MarshalBinary()
if err != nil {
return nil, err
}
txid := Bytes32(sha256.Sum256(data))
return &txid, nil
}
// MarshalLedgerBinary marshals the transaction ledger fields to their
// binary representation. This excludes the signature blocks
func (f *FactoidTransaction) MarshalLedgerBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
// It's very difficult to know the size before marshaling, as
// each in/out has a varint so make the buffer at the end
// The header bytes
header, err := f.MarshalHeaderBinary()
if err != nil {
return nil, err
}
// Inputs
inputs, err := factoidTransactionIOs(f.FCTInputs).MarshalBinary()
if err != nil {
return nil, err
}
// FCT Outputs
fctout, err := factoidTransactionIOs(f.FCTOutputs).MarshalBinary()
if err != nil {
return nil, err
}
// EC Outputs
ecout, err := factoidTransactionIOs(f.ECOutputs).MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(header)+len(inputs)+len(fctout)+len(ecout))
var i int
i += copy(data[i:], header)
i += copy(data[i:], inputs)
i += copy(data[i:], fctout)
i += copy(data[i:], ecout)
return data, nil
}
// TODO: Re-eval how to do this. Kinda different from the rest
func (f *FactoidTransaction) MarshalBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
data, err := f.MarshalLedgerBinary()
if err != nil {
return nil, err
}
for _, s := range f.Signatures {
sig, err := s.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, sig...)
}
return data, nil
}
// MarshalHeaderBinary marshals the transaction's header to its binary
// representation. See UnmarshalHeaderBinary for encoding details.
func (f *FactoidTransaction) MarshalHeaderBinary() ([]byte, error) {
version := varintf.Encode(f.Version)
data := make([]byte, TransactionHeadMinLen+len(version))
var i int
i += copy(data[i:], version)
// Do the timestamp as 6 bytes in ms
ms := f.TimestampSalt.UnixNano() / 1e6
buf := bytes.NewBuffer(make([]byte, 0, 8))
if err := binary.Write(buf, binary.BigEndian, ms); err != nil {
return nil, err
}
i += copy(data[i:], buf.Bytes()[2:])
data[i] = uint8(len(f.FCTInputs))
i += 1
data[i] = uint8(len(f.FCTOutputs))
i += 1
data[i] = uint8(len(f.ECOutputs))
i += 1
return data, nil
}
// MarshalBinary marshals a set of transaction ios to its binary representation.
// See UnmarshalBinary for encoding details.
func (ios factoidTransactionIOs) MarshalBinary() ([]byte, error) {
var data []byte
for _, io := range ios {
iodata, err := io.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, iodata...)
}
return data, nil
}
// MarshalBinary marshals a transaction io to its binary representation.
// See UnmarshalBinary for encoding details.
func (io *FactoidTransactionIO) MarshalBinary() ([]byte, error) {
amount := varintf.Encode(io.Amount)
data := make([]byte, 32+len(amount))
var i int
i += copy(data[i:], amount)
i += copy(data[i:], io.Address[:])
return data, nil
}
// MarshalBinary marshals a transaction signature to its binary representation.
// See UnmarshalBinary for encoding details.
func (s *FactoidTransactionSignature) MarshalBinary() ([]byte, error) {
if !s.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
rcdData, err := s.ReedeemCondition.MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(rcdData)+len(s.SignatureBlock))
var i int
i += copy(data[i:], rcdData)
i += copy(data[i:], s.SignatureBlock)
return data, nil
}
const (
TransactionHeadMinLen = 0 + // Version length is varint
6 + // timestamp
1 + // input count
1 + // factoid output count
1 // EC output count
TransactionTotalMinLen = TransactionHeadMinLen // Coinbases have no body
)
// Decode will consume as many bytes as necessary to unmarshal the factoid
// transaction. It will return the number of bytes read and an error.
func (f *FactoidTransaction) Decode(data []byte) (i int, err error) {
if len(data) < TransactionTotalMinLen {
return 0, fmt.Errorf("insufficient length")
}
// Decode header
version, i := varintf.Decode(data)
| {
return s.SignatureBlock != nil
} | identifier_body | |
transaction.go | // Valid returns if the inputs of the factoid transaction are properly signed
// by the redeem conditions. It will also validate the total inputs is greater
// than the total outputs.
func (f *FactoidTransaction) Valid() bool {
if !f.IsPopulated() {
return false
}
// Validate amounts
if f.TotalFCTInputs() < f.TotalFCTOutputs()+f.TotalECOutput() {
return false
}
// Validate signatures
if len(f.FCTInputs) != len(f.Signatures) {
return false
}
msg, err := f.MarshalLedgerBinary()
if err != nil {
return false
}
for i := range f.FCTInputs {
expAddr := f.Signatures[i].ReedeemCondition.Address()
// RCD should match the input
if bytes.Compare(expAddr[:], f.FCTInputs[i].Address[:]) != 0 {
return false
}
if !f.Signatures[i].Validate(msg) {
return false
}
}
return true
}
func (f *FactoidTransaction) TotalFCTInputs() (total uint64) {
return factoidTransactionIOs(f.FCTInputs).TotalAmount()
}
func (f *FactoidTransaction) TotalFCTOutputs() (total uint64) {
return factoidTransactionIOs(f.FCTOutputs).TotalAmount()
}
// TotalECOutput is delimated in factoishis
func (f *FactoidTransaction) TotalECOutput() (total uint64) {
return factoidTransactionIOs(f.ECOutputs).TotalAmount()
}
func (s factoidTransactionIOs) TotalAmount() (total uint64) {
for _, io := range s {
total += io.Amount
}
return
}
func (s FactoidTransactionSignature) Validate(msg Bytes) bool {
return s.ReedeemCondition.Validate(msg, s.SignatureBlock)
}
// Get queries factomd for the entry corresponding to f.TransactionID, which
// must be not nil. After a successful call all inputs, outputs, and
// the header will be populated
func (f *FactoidTransaction) Get(ctx context.Context, c *Client) error {
// TODO: Test this functionality
// If the TransactionID is nil then we have nothing to query for.
if f.TransactionID == nil {
return fmt.Errorf("txid is nil")
}
// If the Transaction is already populated then there is nothing to do. If
// the Hash is nil, we cannot populate it anyway.
if f.IsPopulated() {
return nil
}
params := struct {
Hash *Bytes32 `json:"hash"`
}{Hash: f.TransactionID}
var result struct {
Data Bytes `json:"data"`
}
if err := c.FactomdRequest(ctx, "raw-data", params, &result); err != nil {
return err
}
if err := f.UnmarshalBinary(result.Data); err != nil {
return err
}
return nil
}
// ComputeTransactionID computes the txid for a given transaction. The txid is
// the sha256 of the ledger fields in a factoid transaction. The ledger fields
// exclude the signature block of the transaction
func (f *FactoidTransaction) ComputeTransactionID() (Bytes32, error) {
data, err := f.MarshalLedgerBinary()
if err != nil {
return Bytes32{}, err
}
return f.computeTransactionID(data)
}
func (f *FactoidTransaction) computeTransactionID(ledgerBinary Bytes) (Bytes32, error) {
txid := Bytes32(sha256.Sum256(ledgerBinary))
return txid, nil
}
// ComputeFullHash computes the fullhash for a given transaction. The fullhash
// is the sha256 of all the fields in a factoid transaction.
func (f *FactoidTransaction) ComputeFullHash() (*Bytes32, error) {
data, err := f.MarshalBinary()
if err != nil {
return nil, err
}
txid := Bytes32(sha256.Sum256(data))
return &txid, nil
}
// MarshalLedgerBinary marshals the transaction ledger fields to their
// binary representation. This excludes the signature blocks
func (f *FactoidTransaction) MarshalLedgerBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
// It's very difficult to know the size before marshaling, as
// each in/out has a varint so make the buffer at the end
// The header bytes
header, err := f.MarshalHeaderBinary()
if err != nil {
return nil, err
}
// Inputs
inputs, err := factoidTransactionIOs(f.FCTInputs).MarshalBinary()
if err != nil {
return nil, err
}
// FCT Outputs
fctout, err := factoidTransactionIOs(f.FCTOutputs).MarshalBinary()
if err != nil {
return nil, err
}
// EC Outputs
ecout, err := factoidTransactionIOs(f.ECOutputs).MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(header)+len(inputs)+len(fctout)+len(ecout))
var i int
i += copy(data[i:], header)
i += copy(data[i:], inputs)
i += copy(data[i:], fctout)
i += copy(data[i:], ecout)
return data, nil
}
// TODO: Re-eval how to do this. Kinda different from the rest
func (f *FactoidTransaction) MarshalBinary() ([]byte, error) {
// TODO: More checks up front?
if !f.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
data, err := f.MarshalLedgerBinary()
if err != nil {
return nil, err
}
for _, s := range f.Signatures {
sig, err := s.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, sig...)
}
return data, nil
}
// MarshalHeaderBinary marshals the transaction's header to its binary
// representation. See UnmarshalHeaderBinary for encoding details.
func (f *FactoidTransaction) MarshalHeaderBinary() ([]byte, error) {
version := varintf.Encode(f.Version)
data := make([]byte, TransactionHeadMinLen+len(version))
var i int
i += copy(data[i:], version)
// Do the timestamp as 6 bytes in ms
ms := f.TimestampSalt.UnixNano() / 1e6
buf := bytes.NewBuffer(make([]byte, 0, 8))
if err := binary.Write(buf, binary.BigEndian, ms); err != nil {
return nil, err
}
i += copy(data[i:], buf.Bytes()[2:])
data[i] = uint8(len(f.FCTInputs))
i += 1
data[i] = uint8(len(f.FCTOutputs))
i += 1
data[i] = uint8(len(f.ECOutputs))
i += 1
return data, nil
}
// MarshalBinary marshals a set of transaction ios to its binary representation.
// See UnmarshalBinary for encoding details.
func (ios factoidTransactionIOs) MarshalBinary() ([]byte, error) {
var data []byte
for _, io := range ios {
iodata, err := io.MarshalBinary()
if err != nil {
return nil, err
}
data = append(data, iodata...)
}
return data, nil | // See UnmarshalBinary for encoding details.
func (io *FactoidTransactionIO) MarshalBinary() ([]byte, error) {
amount := varintf.Encode(io.Amount)
data := make([]byte, 32+len(amount))
var i int
i += copy(data[i:], amount)
i += copy(data[i:], io.Address[:])
return data, nil
}
// MarshalBinary marshals a transaction signature to its binary representation.
// See UnmarshalBinary for encoding details.
func (s *FactoidTransactionSignature) MarshalBinary() ([]byte, error) {
if !s.IsPopulated() {
return nil, fmt.Errorf("not populated")
}
rcdData, err := s.ReedeemCondition.MarshalBinary()
if err != nil {
return nil, err
}
data := make([]byte, len(rcdData)+len(s.SignatureBlock))
var i int
i += copy(data[i:], rcdData)
i += copy(data[i:], s.SignatureBlock)
return data, nil
}
const (
TransactionHeadMinLen = 0 + // Version length is varint
6 + // timestamp
1 + // input count
1 + // factoid output count
1 // EC output count
TransactionTotalMinLen = TransactionHeadMinLen // Coinbases have no body
)
// Decode will consume as many bytes as necessary to unmarshal the factoid
// transaction. It will return the number of bytes read and an error.
func (f *FactoidTransaction) Decode(data []byte) (i int, err error) {
if len(data) < TransactionTotalMinLen {
return 0, fmt.Errorf("insufficient length")
}
// Decode header
version, i := varintf.Decode(data)
if i < 0 {
return 0, fmt | }
// MarshalBinary marshals a transaction io to its binary representation. | random_line_split |
core_bert.py | = q.replace(self.sub,'') # 去掉subject剩下的部分
qtKey = (self.qRaw.replace(self.sub,'(SUB)',1) + ' ||| ' + pre) # 把subject换成(sub)然后加上predicate
if qtKey in qtList:
scoreAP = qtList[qtKey] # 查看当前的问题有没有在知识库中出现过
self.scoreAP = scoreAP
qWithoutSubSet1 = set(qWithoutSub1)
qWithoutSubSet2 = set(qWithoutSub2)
qWithoutSubSet = set(qWithoutSub)
preLowerSet = set(pre.lower())
# 找出predicate和问题前后两部分的最大intersection
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
# 计算来自predicate的分数,采用最大overlap的character的倒数 1/(n+1)
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
if len(pre) != 0:
scorePre = preFactor / len(qWithoutSubSet | preLowerSet)
else:
scorePre = 0
if len(includingObj) != 0 and scorePre == 0:
for objStr in includingObj:
scorePreTmp = 0
preLowerSet = set(objStr.lower())
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
scorePreTmp = preFactor / len(qWithoutSubSet | preLowerSet)
if scorePreTmp > scorePre:
scorePre = scorePreTmp
if use_elmo and len(pre) != 0:
preCut = [pre]
qWithoutSubCut = [qWithoutSub]
data_df = pd.DataFrame({'question': qWithoutSubCut, 'sim_question': preCut})
q_data = data_df['question'].apply(sent2ids)
p_data = data_df['sim_question'].apply(sent2ids)
q_data_e = data2tensor([q_data[0]])
p_data_e = data2tensor([p_data[0]])
output = model(q_data_e,p_data_e) #
scorePre_m = output[0][0]##bert模型做相似度计算,第一个值是相似性分数
self.scorePre = scorePre_m
scoreSub = 0
# 计算subject的权重有多高,可能有些subject本身就是更重要一些,一般来说越罕见的entity重要性越高
for char in self.sub:
if char in countCharDict:
scoreSub += 1/(countCharDict[char] + 1)
else:
scoreSub += 1
self.scoreSub = scoreSub
self.scorePre = scorePre
self.score = scoreSub * self.wS + scorePre * self.wP + scoreAP * self.wAP
return self.score
def pad_sequences(sequences, maxlen=None, dtype='int32',
padding='pre', truncating='pre', value=0.):
num_samples = len(sequences)
lengths = []
for x in sequences:
try:
lengths.append(len(x))
except TypeError:
raise ValueError('`sequences` must be a list of iterables. '
'Found non-iterable: ' + str(x))
if maxlen is None:
maxlen = np.max(lengths)
sample_shape = tuple()
for s in sequences:
if len(s) > 0:
sample_shape = np.asarray(s).shape[1:]
break
is_dtype_str = np.issubdtype(dtype, np.str_) or np.issubdtype(dtype, np.unicode_)
if isinstance(value, six.string_types) and dtype != object and not is_dtype_str:
raise ValueError("`dtype` {} is not compatible with `value`'s type: {}\n"
"You should set `dtype=object` for variable length strings."
.format(dtype, type(value)))
x = np.full((num_samples, maxlen) + sample_shape, value, dtype=dtype)
for idx, s in enumerate(sequences):
if not len(s):
continue # empty list/array was found
if truncating == 'pre':
trunc = s[-maxlen:]
elif truncating == 'post':
trunc = s[:maxlen]
else:
raise ValueError('Truncating type "%s" '
'not understood' % truncating)
trunc = np.asarray(trunc, dtype=dtype)
if trunc.shape[1:] != sample_shape:
raise ValueError('Shape of sample %s of sequence at position %s '
'is different from expected shape %s' %
(trunc.shape[1:], idx, sample_shape))
if padding == 'post':
x[idx, :len(trunc)] = trunc
elif padding == 'pre':
x[idx, -len(trunc):] = trunc
else:
raise ValueError('Padding type "%s" not understood' % padding)
return x
def data2tensor(batch_token_ids, pad=True, maxlen=50):
if pad:
batch_token_ids = pad_sequences(batch_token_ids, maxlen=maxlen, padding='post')
batch_token_ids = torch.tensor(batch_token_ids, dtype=torch.long).to(DEVICE)
return batch_token_ids
tokenizer = BertTokenizer.from_pretrained(pretrained_model_name_or_path)
def sent2ids(sent_text):
sent_tokens = ['[CLS]'] + tokenizer.tokenize(sent_text) + ["[SEP]"]
token_ids = tokenizer.convert_tokens_to_ids(sent_tokens)
return token_ids
def getAnswer(sub, pre, kbDict):
answerList = []
for kb in kbDict[sub]:
if pre in kb:
answerList.append(kb[pre])
return answerList
def answerQ (qRaw, lKey, kbDict, qtList, countCharDict, wP=10, threshold=0, debug=False):
q = qRaw.strip().lower() # 问题转化成小写
candidateSet = set()
result = ''
maxScore = 0
bestAnswer = set()
for key in lKey:
if -1 != q.find(key): # 如果问题中出现了该subject,那么我们就要考虑这个subject的triples
for kb in kbDict[key]:
for pre in list(kb):
newAnswerCandidate = answerCandidate(key, pre, q, wP=wP) # 构建一个新的answer candidate
candidateSet.add(newAnswerCandidate)
candidateSetCopy = candidateSet.copy()
if debug:
print('len(candidateSet) = ' + str(len(candidateSetCopy)), end = '\r', flush=True)
candidateSet = set()
candidateSetIndex = set()
for aCandidate in candidateSetCopy:
strTmp = str(aCandidate.sub+'|'+aCandidate.pre)
if strTmp not in candidateSetIndex:
candidateSetIndex.add(strTmp)
candidateSet.add(aCandidate)
# 针对每一个candidate answer,计算该candidate的分数,然后选择分数最高的作为答案
for aCandidate in candidateSet:
scoreTmp = aCandidate.calcScore(qtList, countCharDict,debug)
if scoreTmp > maxScore:
maxScore = scoreTmp
bestAnswer = set()
if scoreTmp == maxScore:
bestAnswer.add(aCandidate)
# 去除一些重复的答案
bestAnswerCopy = bestAnswer.copy()
bestAnswer = set()
for aCandidate in bestAnswerCopy:
aCfound = 0
for aC in bestAnswer:
if aC.pre == aCandidate.pre and aC.sub == aCandidate.sub:
aCfound = 1
break
if aCfound == 0:
bestAnswer.add(aCandidate)
# 加入object的分数
bestAnswerCopy = bestAnswer.copy()
for aCandidate in bestAnswerCopy:
if aCandidate.score == aCandidate.scoreSub:
scoreReCal = aCandidate.calcScore(qtList, countCharDict,debug, includingObj=getAnswer(aCandidate.sub, aCandidate.pre, kbDict))
if scoreReCal > maxScore:
bestAnswer = set()
maxScore = scoreReCal
if scoreReCal == maxScore:
bestAnswer.add(aCandidate)
# 加入cosine similarity
bestAnswerCopy = bestAnswer.copy()
if len(bestAnswer) > 1: # use word vector to remove duplicated answer
for aCandidate in bestAnswerCopy:
scoreReCal = aCandidate.calcScore(qtList, countCharDict,debug, includingObj=getAnswer(aCandidate.sub, aCandidate.pre, kbDict), use_elmo | =True)
if scoreReCal > maxScore:
| conditional_block | |
core_bert.py | self.qRaw
subIndex = q.index(self.sub)
qWithoutSub1 = q[:subIndex] # subject左边的部分
qWithoutSub2 = q[subIndex+lenSub:] # subject右边的部分
qWithoutSub = q.replace(self.sub,'') # 去掉subject剩下的部分
qtKey = (self.qRaw.replace(self.sub,'(SUB)',1) + ' ||| ' + pre) # 把subject换成(sub)然后加上predicate
if qtKey in qtList:
scoreAP = qtList[qtKey] # 查看当前的问题有没有在知识库中出现过
self.scoreAP = scoreAP
qWithoutSubSet1 = set(qWithoutSub1)
qWithoutSubSet2 = set(qWithoutSub2)
qWithoutSubSet = set(qWithoutSub)
preLowerSet = set(pre.lower())
# 找出predicate和问题前后两部分的最大intersection
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
# 计算来自predicate的分数,采用最大overlap的character的倒数 1/(n+1)
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
if len(pre) != 0:
scorePre = preFactor / len(qWithoutSubSet | preLowerSet)
else:
scorePre = 0
if len(includingObj) != 0 and scorePre == 0:
for objStr in includingObj:
scorePreTmp = 0
preLowerSet = set(objStr.lower())
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
scorePreTmp = preFactor / len(qWithoutSubSet | preLowerSet)
if scorePreTmp > scorePre:
scorePre = scorePreTmp
if use_elmo and len(pre) != 0:
preCut = [pre]
qWithoutSubCut = [qWithoutSub]
data_df = pd.DataFrame({'question': qWithoutSubCut, 'sim_question': preCut})
q_data = data_df['question'].apply(sent2ids)
p_data = data_df['sim_question'].apply(sent2ids)
q_data_e = data2tensor([q_data[0]])
p_data_e = data2tensor([p_data[0]])
output = model(q_data_e,p_data_e) #
scorePre_m = output[0][0]##bert模型做相似度计算,第一个值是相似性分数
self.scorePre = scorePre_m
scoreSub = 0
# 计算subject的权重有多高,可能有些subject本身就是更重要一些,一般来说越罕见的entity重要性越高
for char in self.sub:
if char in countCharDict:
scoreSub += 1/(countCharDict[char] + 1)
else:
scoreSub += 1
self.scoreSub = scoreSub
self.scorePre = scorePre
self.score = scoreSub * self.wS + scorePre * self.wP + scoreAP * self.wAP
return self.score
def pad_sequences(sequences, maxlen=None, dtype='int32',
padding='pre', truncating='pre', value=0.):
num_samples = len(sequences)
lengths = []
for x in sequences:
try:
lengths.append(len(x))
except TypeError:
raise ValueError('`sequences` must be a list of iterables. '
'Found non-iterable: ' + str(x))
if maxlen is None:
maxlen = np.max(lengths)
sample_shape = tuple()
for s in sequences:
if len(s) > 0:
sample_shape = np.asarray(s).shape[1:]
break
is_dtype_str = np.issubdtype(dtype, np.str_) or np.issubdtype(dtype, np.unicode_)
if isinstance(value, six.string_types) and dtype != object and not is_dtype_str:
raise ValueError("`dtype` {} is not compatible with `value`'s type: {}\n"
"You should set `dtype=object` for variable length strings."
.format(dtype, type(value)))
x = np.full((num_samples, maxlen) + sample_shape, value, dtype=dtype)
for idx, s in enumerate(sequences):
if not len(s):
continue # empty list/array was found
if truncating == 'pre':
trunc = s[-maxlen:]
elif truncating == 'post':
trunc = s[:maxlen]
else:
raise ValueError('Truncating type "%s" '
'not understood' % truncating)
trunc = np.asarray(trunc, dtype=dtype)
if trunc.shape[1:] != sample_shape:
raise ValueError('Shape of sample %s of sequence at position %s '
'is different from expected shape %s' %
(trunc.shape[1:], idx, sample_shape))
if padding == 'post':
x[idx, :len(trunc)] = trunc
elif padding == 'pre':
x[idx, -len(trunc):] = trunc
else:
raise ValueError('Padding type "%s" not understood' % padding)
return x
def data2tensor(batch_token_ids, pad=True, maxlen=50):
if pad:
batch_token_ids = pad_sequences(batch_token_ids, maxlen=maxlen, padding='post')
batch_token_ids = torch.tensor(batch_token_ids, dtype=torch.long).to(DEVICE)
return batch_token_ids
tokenizer = BertTokenizer.from_pretrained(pretrained_model_name_or_path)
def sent2ids(sent_text):
sent_tokens = ['[CLS]'] + tokenizer.tokenize(sent_text) + ["[SEP]"]
token_ids = tokenizer.convert_tokens_to_ids(sent_tokens)
return token_ids
def getAnswer(sub, pre, kbDict):
answerList = []
for kb in kbDict[sub]:
if pre in kb:
answerList.append(kb[pre])
return answerList
def answerQ (qRaw, lKey, kbDict, qtList, countCharDict, wP=10, threshold=0, debug=False):
q = qRaw.strip().lower() # 问题转化成小写
candidateSet = set()
| = ''
maxScore = 0
bestAnswer = set()
for key in lKey:
if -1 != q.find(key): # 如果问题中出现了该subject,那么我们就要考虑这个subject的triples
for kb in kbDict[key]:
for pre in list(kb):
newAnswerCandidate = answerCandidate(key, pre, q, wP=wP) # 构建一个新的answer candidate
candidateSet.add(newAnswerCandidate)
candidateSetCopy = candidateSet.copy()
if debug:
print('len(candidateSet) = ' + str(len(candidateSetCopy)), end = '\r', flush=True)
candidateSet = set()
candidateSetIndex = set()
for aCandidate in candidateSetCopy:
strTmp = str(aCandidate.sub+'|'+aCandidate.pre)
if strTmp not in candidateSetIndex:
candidateSetIndex.add(strTmp)
candidateSet.add(aCandidate)
# 针对每一个candidate answer,计算该candidate的分数,然后选择分数最高的作为答案
for aCandidate in candidateSet:
scoreTmp = aCandidate.calcScore(qtList, countCharDict,debug)
if scoreTmp > maxScore:
maxScore = scoreTmp
bestAnswer = set()
if scoreTmp == maxScore:
bestAnswer.add(aCandidate)
# 去除一些重复的答案
bestAnswerCopy = bestAnswer.copy()
bestAnswer = set()
for aCandidate in bestAnswerCopy:
aCfound = 0
for aC in bestAnswer:
if aC.pre == aCandidate.pre and aC.sub == aCandidate.sub:
aCfound = 1
break
if aCfound == 0:
bestAnswer.add(aCandidate)
# 加入object的分数
bestAnswerCopy = bestAnswer.copy()
for aCandidate in bestAnswerCopy:
if aCandidate.score == aCandidate.scoreSub:
scoreReCal = aCandidate.calcScore(qtList, countCharDict,debug, includingObj=getAnswer(aCandidate.sub, aCandidate.pre, kbDict))
if scoreReCal > maxScore:
bestAnswer = set()
maxScore = scoreReCal
if scoreReCal == maxScore:
bestAnswer.add(aCandidate)
# 加入cosine similarity
bestAnswerCopy = bestAnswer.copy()
if len(bestAnswer) > 1: # use word vector to remove duplicated answer
for aCandidate in best | result | identifier_name |
core_bert.py | self.qRaw
subIndex = q.index(self.sub)
qWithoutSub1 = q[:subIndex] # subject左边的部分
qWithoutSub2 = q[subIndex+lenSub:] # subject右边的部分
qWithoutSub = q.replace(self.sub,'') # 去掉subject剩下的部分
qtKey = (self.qRaw.replace(self.sub,'(SUB)',1) + ' ||| ' + pre) # 把subject换成(sub)然后加上predicate
if qtKey in qtList:
scoreAP = qtList[qtKey] # 查看当前的问题有没有在知识库中出现过
self.scoreAP = scoreAP
qWithoutSubSet1 = set(qWithoutSub1)
qWithoutSubSet2 = set(qWithoutSub2)
qWithoutSubSet = set(qWithoutSub)
preLowerSet = set(pre.lower())
# 找出predicate和问题前后两部分的最大intersection
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
# 计算来自predicate的分数,采用最大overlap的character的倒数 1/(n+1)
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
if len(pre) != 0:
scorePre = preFactor / len(qWithoutSubSet | preLowerSet)
else:
scorePre = 0
if len(includingObj) != 0 and scorePre == 0:
for objStr in includingObj:
scorePreTmp = 0
preLowerSet = set(objStr.lower())
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
scorePreTmp = preFactor / len(qWithoutSubSet | preLowerSet)
if scorePreTmp > scorePre:
scorePre = scorePreTmp
if use_elmo and len(pre) != 0:
preCut = [pre]
qWithoutSubCut = [qWithoutSub]
data_df = pd.DataFrame({'question': qWithoutSubCut, 'sim_question': preCut})
q_data = data_df['question'].apply(sent2ids)
p_data = data_df['sim_question'].apply(sent2ids)
q_data_e = data2tensor([q_data[0]])
p_data_e = data2tensor([p_data[0]])
output = model(q_data_e,p_data_e) #
scorePre_m = output[0][0]##bert模型做相似度计算,第一个值是相似性分数
self.scorePre = scorePre_m
scoreSub = 0
# 计算subject的权重有多高,可能有些subject本身就是更重要一些,一般来说越罕见的entity重要性越高
for char in self.sub:
if char in countCharDict:
scoreSub += 1/(countCharDict[char] + 1)
else:
scoreSub += 1
self.scoreSub = scoreSub
self.scorePre = scorePre
self.score = scoreSub * self.wS + scorePre * self.wP + scoreAP * self.wAP
return self.score
def pad_sequences(sequences, maxlen=None, dtype='int32',
padding='pre', truncating='pre', value=0.):
num_samples = len(sequences) | try:
lengths.append(len(x))
except TypeError:
raise ValueError('`sequences` must be a list of iterables. '
'Found non-iterable: ' + str(x))
if maxlen is None:
maxlen = np.max(lengths)
sample_shape = tuple()
for s in sequences:
if len(s) > 0:
sample_shape = np.asarray(s).shape[1:]
break
is_dtype_str = np.issubdtype(dtype, np.str_) or np.issubdtype(dtype, np.unicode_)
if isinstance(value, six.string_types) and dtype != object and not is_dtype_str:
raise ValueError("`dtype` {} is not compatible with `value`'s type: {}\n"
"You should set `dtype=object` for variable length strings."
.format(dtype, type(value)))
x = np.full((num_samples, maxlen) + sample_shape, value, dtype=dtype)
for idx, s in enumerate(sequences):
if not len(s):
continue # empty list/array was found
if truncating == 'pre':
trunc = s[-maxlen:]
elif truncating == 'post':
trunc = s[:maxlen]
else:
raise ValueError('Truncating type "%s" '
'not understood' % truncating)
trunc = np.asarray(trunc, dtype=dtype)
if trunc.shape[1:] != sample_shape:
raise ValueError('Shape of sample %s of sequence at position %s '
'is different from expected shape %s' %
(trunc.shape[1:], idx, sample_shape))
if padding == 'post':
x[idx, :len(trunc)] = trunc
elif padding == 'pre':
x[idx, -len(trunc):] = trunc
else:
raise ValueError('Padding type "%s" not understood' % padding)
return x
def data2tensor(batch_token_ids, pad=True, maxlen=50):
if pad:
batch_token_ids = pad_sequences(batch_token_ids, maxlen=maxlen, padding='post')
batch_token_ids = torch.tensor(batch_token_ids, dtype=torch.long).to(DEVICE)
return batch_token_ids
tokenizer = BertTokenizer.from_pretrained(pretrained_model_name_or_path)
def sent2ids(sent_text):
sent_tokens = ['[CLS]'] + tokenizer.tokenize(sent_text) + ["[SEP]"]
token_ids = tokenizer.convert_tokens_to_ids(sent_tokens)
return token_ids
def getAnswer(sub, pre, kbDict):
answerList = []
for kb in kbDict[sub]:
if pre in kb:
answerList.append(kb[pre])
return answerList
def answerQ (qRaw, lKey, kbDict, qtList, countCharDict, wP=10, threshold=0, debug=False):
q = qRaw.strip().lower() # 问题转化成小写
candidateSet = set()
result = ''
maxScore = 0
bestAnswer = set()
for key in lKey:
if -1 != q.find(key): # 如果问题中出现了该subject,那么我们就要考虑这个subject的triples
for kb in kbDict[key]:
for pre in list(kb):
newAnswerCandidate = answerCandidate(key, pre, q, wP=wP) # 构建一个新的answer candidate
candidateSet.add(newAnswerCandidate)
candidateSetCopy = candidateSet.copy()
if debug:
print('len(candidateSet) = ' + str(len(candidateSetCopy)), end = '\r', flush=True)
candidateSet = set()
candidateSetIndex = set()
for aCandidate in candidateSetCopy:
strTmp = str(aCandidate.sub+'|'+aCandidate.pre)
if strTmp not in candidateSetIndex:
candidateSetIndex.add(strTmp)
candidateSet.add(aCandidate)
# 针对每一个candidate answer,计算该candidate的分数,然后选择分数最高的作为答案
for aCandidate in candidateSet:
scoreTmp = aCandidate.calcScore(qtList, countCharDict,debug)
if scoreTmp > maxScore:
maxScore = scoreTmp
bestAnswer = set()
if scoreTmp == maxScore:
bestAnswer.add(aCandidate)
# 去除一些重复的答案
bestAnswerCopy = bestAnswer.copy()
bestAnswer = set()
for aCandidate in bestAnswerCopy:
aCfound = 0
for aC in bestAnswer:
if aC.pre == aCandidate.pre and aC.sub == aCandidate.sub:
aCfound = 1
break
if aCfound == 0:
bestAnswer.add(aCandidate)
# 加入object的分数
bestAnswerCopy = bestAnswer.copy()
for aCandidate in bestAnswerCopy:
if aCandidate.score == aCandidate.scoreSub:
scoreReCal = aCandidate.calcScore(qtList, countCharDict,debug, includingObj=getAnswer(aCandidate.sub, aCandidate.pre, kbDict))
if scoreReCal > maxScore:
bestAnswer = set()
maxScore = scoreReCal
if scoreReCal == maxScore:
bestAnswer.add(aCandidate)
# 加入cosine similarity
bestAnswerCopy = bestAnswer.copy()
if len(bestAnswer) > 1: # use word vector to remove duplicated answer
for aCandidate in bestAnswerCopy | lengths = []
for x in sequences: | random_line_split |
core_bert.py | scoreAP = 0
pre = self.pre
q = self.qRaw
subIndex = q.index(self.sub)
qWithoutSub1 = q[:subIndex] # subject左边的部分
qWithoutSub2 = q[subIndex+lenSub:] # subject右边的部分
qWithoutSub = q.replace(self.sub,'') # 去掉subject剩下的部分
qtKey = (self.qRaw.replace(self.sub,'(SUB)',1) + ' ||| ' + pre) # 把subject换成(sub)然后加上predicate
if qtKey in qtList:
scoreAP = qtList[qtKey] # 查看当前的问题有没有在知识库中出现过
self.scoreAP = scoreAP
qWithoutSubSet1 = set(qWithoutSub1)
qWithoutSubSet2 = set(qWithoutSub2)
qWithoutSubSet = set(qWithoutSub)
preLowerSet = set(pre.lower())
# 找出predicate和问题前后两部分的最大intersection
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
# 计算来自predicate的分数,采用最大overlap的character的倒数 1/(n+1)
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
if len(pre) != 0:
scorePre = preFactor / len(qWithoutSubSet | preLowerSet)
else:
scorePre = 0
if len(includingObj) != 0 and scorePre == 0:
for objStr in includingObj:
scorePreTmp = 0
preLowerSet = set(objStr.lower())
intersection1 = qWithoutSubSet1 & preLowerSet
intersection2 = qWithoutSubSet2 & preLowerSet
if len(intersection1) > len(intersection2):
maxIntersection = intersection1
else:
maxIntersection = intersection2
preFactor = 0
for char in maxIntersection:
if char in countCharDict:
preFactor += 1/(countCharDict[char] + 1)
else:
preFactor += 1
scorePreTmp = preFactor / len(qWithoutSubSet | preLowerSet)
if scorePreTmp > scorePre:
scorePre = scorePreTmp
if use_elmo and len(pre) != 0:
preCut = [pre]
qWithoutSubCut = [qWithoutSub]
data_df = pd.DataFrame({'question': qWithoutSubCut, 'sim_question': preCut})
q_data = data_df['question'].apply(sent2ids)
p_data = data_df['sim_question'].apply(sent2ids)
q_data_e = data2tensor([q_data[0]])
p_data_e = data2tensor([p_data[0]])
output = model(q_data_e,p_data_e) #
scorePre_m = output[0][0]##bert模型做相似度计算,第一个值是相似性分数
self.scorePre = scorePre_m
scoreSub = 0
# 计算subject的权重有多高,可能有些subject本身就是更重要一些,一般来说越罕见的entity重要性越高
for char in self.sub:
if char in countCharDict:
scoreSub += 1/(countCharDict[char] + 1)
else:
scoreSub += 1
self.scoreSub = scoreSub
self.scorePre = scorePre
self.score = scoreSub * self.wS + scorePre * self.wP + scoreAP * self.wAP
return self.score
def pad_sequences(sequences, maxlen=None, dtype='int32',
padding='pre', truncating='pre', value=0.):
num_samples = len(sequences)
lengths = []
for x in sequences:
try:
lengths.append(len(x))
except TypeError:
raise ValueError('`sequences` must be
a list of iterables. '
'Found non-iterable: ' + str(x))
if maxlen is None:
maxlen = np.max(lengths)
sample_shape = tuple()
for s in sequences:
if len(s) > 0:
sample_shape = np.asarray(s).shape[1:]
break
is_dtype_str = np.issubdtype(dtype, np.str_) or np.issubdtype(dtype, np.unicode_)
if isinstance(value, six.string_types) and dtype != object and not is_dtype_str:
raise ValueError("`dtype` {} is not compatible with `value`'s type: {}\n"
"You should set `dtype=object` for variable length strings."
.format(dtype, type(value)))
x = np.full((num_samples, maxlen) + sample_shape, value, dtype=dtype)
for idx, s in enumerate(sequences):
if not len(s):
continue # empty list/array was found
if truncating == 'pre':
trunc = s[-maxlen:]
elif truncating == 'post':
trunc = s[:maxlen]
else:
raise ValueError('Truncating type "%s" '
'not understood' % truncating)
trunc = np.asarray(trunc, dtype=dtype)
if trunc.shape[1:] != sample_shape:
raise ValueError('Shape of sample %s of sequence at position %s '
'is different from expected shape %s' %
(trunc.shape[1:], idx, sample_shape))
if padding == 'post':
x[idx, :len(trunc)] = trunc
elif padding == 'pre':
x[idx, -len(trunc):] = trunc
else:
raise ValueError('Padding type "%s" not understood' % padding)
return x
def data2tensor(batch_token_ids, pad=True, maxlen=50):
if pad:
batch_token_ids = pad_sequences(batch_token_ids, maxlen=maxlen, padding='post')
batch_token_ids = torch.tensor(batch_token_ids, dtype=torch.long).to(DEVICE)
return batch_token_ids
tokenizer = BertTokenizer.from_pretrained(pretrained_model_name_or_path)
def sent2ids(sent_text):
sent_tokens = ['[CLS]'] + tokenizer.tokenize(sent_text) + ["[SEP]"]
token_ids = tokenizer.convert_tokens_to_ids(sent_tokens)
return token_ids
def getAnswer(sub, pre, kbDict):
answerList = []
for kb in kbDict[sub]:
if pre in kb:
answerList.append(kb[pre])
return answerList
def answerQ (qRaw, lKey, kbDict, qtList, countCharDict, wP=10, threshold=0, debug=False):
q = qRaw.strip().lower() # 问题转化成小写
candidateSet = set()
result = ''
maxScore = 0
bestAnswer = set()
for key in lKey:
if -1 != q.find(key): # 如果问题中出现了该subject,那么我们就要考虑这个subject的triples
for kb in kbDict[key]:
for pre in list(kb):
newAnswerCandidate = answerCandidate(key, pre, q, wP=wP) # 构建一个新的answer candidate
candidateSet.add(newAnswerCandidate)
candidateSetCopy = candidateSet.copy()
if debug:
print('len(candidateSet) = ' + str(len(candidateSetCopy)), end = '\r', flush=True)
candidateSet = set()
candidateSetIndex = set()
for aCandidate in candidateSetCopy:
strTmp = str(aCandidate.sub+'|'+aCandidate.pre)
if strTmp not in candidateSetIndex:
candidateSetIndex.add(strTmp)
candidateSet.add(aCandidate)
# 针对每一个candidate answer,计算该candidate的分数,然后选择分数最高的作为答案
for aCandidate in candidateSet:
scoreTmp = aCandidate.calcScore(qtList, countCharDict,debug)
if scoreTmp > maxScore:
maxScore = scoreTmp
bestAnswer = set()
if scoreTmp == max | f, sub = '', pre = '', qRaw = '', qType = 0, score = 0, kbDict = [], wS = 1, wP = 10, wAP = 100):
self.sub = sub # subject
self.pre = pre # predicate
self.qRaw = qRaw # raw question
self.qType = qType # question type
self.score = score # 分数
self.kbDict = kbDict # kd dictionary
self.origin = ''
self.scoreDetail = [0,0,0,0,0]
self.wS = wS # subject的权重
self.wP = wP # oredicate的权重
self.wAP = wAP # answer pattern的权重
self.scoreSub = 0
self.scoreAP = 0
self.scorePre = 0
def calcScore(self, qtList, countCharDict, debug=False, includingObj = [], use_elmo=False):
# 最重要的部分,计算该答案的分数
lenSub = len(self.sub)
scorePre = 0 | identifier_body | |
connection.rs | 2, usize>>,
pub(crate) active_surface_id: RefCell<u32>,
/// Repeats per second
pub(crate) key_repeat_rate: RefCell<i32>,
pub(crate) mem_pool: RefCell<AutoMemPool>,
/// Delay before repeating, in milliseconds
pub(crate) key_repeat_delay: RefCell<i32>,
pub(crate) last_serial: RefCell<u32>,
seat_listener: SeatListener,
pub(crate) environment: Environment<MyEnvironment>,
event_q: RefCell<EventQueue>,
pub(crate) display: RefCell<Display>,
}
impl Drop for WaylandConnection {
fn drop(&mut self) {
self.environment
.with_inner(|env| env.input_handler.shutdown());
}
}
impl WaylandConnection {
pub fn create_new() -> anyhow::Result<Self> {
let (environment, display, event_q) = toolkit::new_default_environment!(
MyEnvironment,
desktop,
fields = [
output_handler: OutputHandler::new(),
input_handler: InputHandler::new(),
]
)?;
let mut pointer = None;
let mut seat_keyboards = HashMap::new();
for seat in environment.get_all_seats() {
if let Some((has_kbd, has_ptr, name)) =
toolkit::seat::with_seat_data(&seat, |seat_data| {
(
seat_data.has_keyboard && !seat_data.defunct,
seat_data.has_pointer && !seat_data.defunct,
seat_data.name.clone(),
)
})
{
if has_kbd {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
environment.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(name, keyboard);
}
if has_ptr {
pointer.replace(PointerDispatcher::register(
&seat,
environment.require_global(),
environment.require_global(),
environment.require_global(),
environment.get_primary_selection_manager(),
)?);
}
}
}
let pointer =
pointer.ok_or_else(|| anyhow::anyhow!("no seats have an available pointer"))?;
let seat_listener;
{
let env = environment.clone();
seat_listener = environment.listen_for_seats(move |seat, seat_data, _| {
if seat_data.has_keyboard {
if !seat_data.defunct {
// We only want to assign a new keyboard object if we don't already have
// one for this seat. When a seat is being created or updated, the listener
// can receive the same seat multiple times: for example, when switching
// back from another virtual console, the same seat is usually seen four
// times with different data flags:
//
// has_pointer: true; has_keyboard: false
// has_pointer: false; has_keyboard: false
// has_pointer: false; has_keyboard: true
// has_pointer: true; has_keyboard: true
//
// This is essentially telling the client to re-assign its keyboard and
// pointer, but that means that this listener will fire twice with
// has_keyboard set to true. If we assign a handler both times, then we end
// up handling key events twice.
if !seat_keyboards.contains_key(&seat_data.name) {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
env.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(seat_data.name.clone(), keyboard);
}
} else {
env.with_inner(|env| env.input_handler.seat_defunct(&seat));
}
} else {
// If we previously had a keyboard object on this seat, it's no longer valid if
// has_keyboard is false, so we remove the keyboard object we knew about and
// thereby ensure that we assign a new keyboard object next time the listener
// fires for this seat with has_keyboard = true.
seat_keyboards.remove(&seat_data.name);
}
if seat_data.has_pointer && !seat_data.defunct {
let conn = Connection::get().unwrap().wayland();
conn.pointer.borrow_mut().seat_changed(&seat);
}
});
}
let mem_pool = environment.create_auto_pool()?;
Ok(Self {
display: RefCell::new(display),
environment,
should_terminate: RefCell::new(false),
next_window_id: AtomicUsize::new(1),
windows: RefCell::new(HashMap::new()),
event_q: RefCell::new(event_q),
pointer: RefCell::new(pointer),
seat_listener,
mem_pool: RefCell::new(mem_pool),
gl_connection: RefCell::new(None),
keyboard_mapper: RefCell::new(None),
key_repeat_rate: RefCell::new(25),
key_repeat_delay: RefCell::new(400),
keyboard_window_id: RefCell::new(None),
last_serial: RefCell::new(0),
surface_to_window_id: RefCell::new(HashMap::new()),
active_surface_id: RefCell::new(0),
})
}
fn keyboard_event(
&self,
keyboard: Main<WlKeyboard>,
event: WlKeyboardEvent,
) -> anyhow::Result<()> {
match &event {
WlKeyboardEvent::Enter {
serial, surface, ..
} => {
// update global active surface id
*self.active_surface_id.borrow_mut() = surface.as_ref().id();
*self.last_serial.borrow_mut() = *serial;
if let Some(&window_id) = self
.surface_to_window_id
.borrow()
.get(&surface.as_ref().id())
{
self.keyboard_window_id.borrow_mut().replace(window_id);
self.environment.with_inner(|env| {
if let Some(input) =
env.input_handler.get_text_input_for_keyboard(&keyboard)
{
input.enable();
input.commit();
}
env.input_handler.advise_surface(&surface, &keyboard);
});
} else {
log::warn!("{:?}, no known surface", event);
}
}
WlKeyboardEvent::Leave { serial, .. } => {
if let Some(input) = self
.environment
.with_inner(|env| env.input_handler.get_text_input_for_keyboard(&keyboard))
{
input.disable();
input.commit();
}
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::Key { serial, .. } | WlKeyboardEvent::Modifiers { serial, .. } => {
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::RepeatInfo { rate, delay } => {
*self.key_repeat_rate.borrow_mut() = *rate;
*self.key_repeat_delay.borrow_mut() = *delay;
}
WlKeyboardEvent::Keymap { format, fd, size } => | }
_ => {}
}
}
_ => {}
}
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.keyboard_event(event);
}
}
Ok(())
}
pub(crate) fn dispatch_to_focused_window(&self, event: WindowEvent) {
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.events.dispatch(event);
}
}
}
pub(crate) fn next_window_id(&self) -> usize {
self.next_window_id
.fetch_add(1, ::std::sync::atomic::Ordering::Relaxed)
}
fn flush(&self) -> anyhow::Result<()> {
if let Err(e) = self.display.borrow_mut().flush() {
if e.kind() != ::std::io::ErrorKind::WouldBlock {
bail!("Error while flushing display: {}", e);
}
}
Ok(())
}
pub(crate) fn window_by_id(&self, window_id: usize) -> Option<Rc<RefCell<WaylandWindowInner>>> {
self.windows.borrow().get(&window_id).map(Rc | {
let file = unsafe { std::fs::File::from_raw_fd(*fd) };
match format {
KeymapFormat::XkbV1 => {
let mut data = vec![0u8; *size as usize];
file.read_exact_at(&mut data, 0)?;
// Dance around CString panicing on the NUL terminator
// in the xkbcommon crate
while let Some(0) = data.last() {
data.pop();
}
let s = String::from_utf8(data)?;
match Keyboard::new_from_string(s) {
Ok(k) => {
self.keyboard_mapper.replace(Some(k));
}
Err(err) => {
log::error!("Error processing keymap change: {:#}", err);
}
} | conditional_block |
connection.rs | 2, usize>>,
pub(crate) active_surface_id: RefCell<u32>,
/// Repeats per second
pub(crate) key_repeat_rate: RefCell<i32>,
pub(crate) mem_pool: RefCell<AutoMemPool>,
/// Delay before repeating, in milliseconds
pub(crate) key_repeat_delay: RefCell<i32>,
pub(crate) last_serial: RefCell<u32>,
seat_listener: SeatListener,
pub(crate) environment: Environment<MyEnvironment>,
event_q: RefCell<EventQueue>,
pub(crate) display: RefCell<Display>,
}
impl Drop for WaylandConnection {
fn drop(&mut self) {
self.environment
.with_inner(|env| env.input_handler.shutdown());
}
}
impl WaylandConnection {
pub fn create_new() -> anyhow::Result<Self> {
let (environment, display, event_q) = toolkit::new_default_environment!(
MyEnvironment,
desktop,
fields = [
output_handler: OutputHandler::new(),
input_handler: InputHandler::new(),
]
)?;
let mut pointer = None;
let mut seat_keyboards = HashMap::new();
for seat in environment.get_all_seats() {
if let Some((has_kbd, has_ptr, name)) =
toolkit::seat::with_seat_data(&seat, |seat_data| {
(
seat_data.has_keyboard && !seat_data.defunct,
seat_data.has_pointer && !seat_data.defunct,
seat_data.name.clone(),
)
})
{
if has_kbd {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
environment.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(name, keyboard);
}
if has_ptr {
pointer.replace(PointerDispatcher::register(
&seat,
environment.require_global(),
environment.require_global(),
environment.require_global(),
environment.get_primary_selection_manager(),
)?);
}
}
}
let pointer =
pointer.ok_or_else(|| anyhow::anyhow!("no seats have an available pointer"))?;
let seat_listener;
{
let env = environment.clone();
seat_listener = environment.listen_for_seats(move |seat, seat_data, _| {
if seat_data.has_keyboard {
if !seat_data.defunct {
// We only want to assign a new keyboard object if we don't already have
// one for this seat. When a seat is being created or updated, the listener
// can receive the same seat multiple times: for example, when switching
// back from another virtual console, the same seat is usually seen four
// times with different data flags:
//
// has_pointer: true; has_keyboard: false
// has_pointer: false; has_keyboard: false
// has_pointer: false; has_keyboard: true
// has_pointer: true; has_keyboard: true
//
// This is essentially telling the client to re-assign its keyboard and
// pointer, but that means that this listener will fire twice with
// has_keyboard set to true. If we assign a handler both times, then we end
// up handling key events twice.
if !seat_keyboards.contains_key(&seat_data.name) {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
env.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(seat_data.name.clone(), keyboard);
}
} else {
env.with_inner(|env| env.input_handler.seat_defunct(&seat));
}
} else {
// If we previously had a keyboard object on this seat, it's no longer valid if
// has_keyboard is false, so we remove the keyboard object we knew about and
// thereby ensure that we assign a new keyboard object next time the listener
// fires for this seat with has_keyboard = true.
seat_keyboards.remove(&seat_data.name);
}
if seat_data.has_pointer && !seat_data.defunct {
let conn = Connection::get().unwrap().wayland();
conn.pointer.borrow_mut().seat_changed(&seat);
}
});
}
let mem_pool = environment.create_auto_pool()?;
Ok(Self {
display: RefCell::new(display),
environment,
should_terminate: RefCell::new(false),
next_window_id: AtomicUsize::new(1),
windows: RefCell::new(HashMap::new()),
event_q: RefCell::new(event_q),
pointer: RefCell::new(pointer),
seat_listener,
mem_pool: RefCell::new(mem_pool),
gl_connection: RefCell::new(None),
keyboard_mapper: RefCell::new(None),
key_repeat_rate: RefCell::new(25),
key_repeat_delay: RefCell::new(400),
keyboard_window_id: RefCell::new(None),
last_serial: RefCell::new(0),
surface_to_window_id: RefCell::new(HashMap::new()),
active_surface_id: RefCell::new(0),
})
}
fn | (
&self,
keyboard: Main<WlKeyboard>,
event: WlKeyboardEvent,
) -> anyhow::Result<()> {
match &event {
WlKeyboardEvent::Enter {
serial, surface, ..
} => {
// update global active surface id
*self.active_surface_id.borrow_mut() = surface.as_ref().id();
*self.last_serial.borrow_mut() = *serial;
if let Some(&window_id) = self
.surface_to_window_id
.borrow()
.get(&surface.as_ref().id())
{
self.keyboard_window_id.borrow_mut().replace(window_id);
self.environment.with_inner(|env| {
if let Some(input) =
env.input_handler.get_text_input_for_keyboard(&keyboard)
{
input.enable();
input.commit();
}
env.input_handler.advise_surface(&surface, &keyboard);
});
} else {
log::warn!("{:?}, no known surface", event);
}
}
WlKeyboardEvent::Leave { serial, .. } => {
if let Some(input) = self
.environment
.with_inner(|env| env.input_handler.get_text_input_for_keyboard(&keyboard))
{
input.disable();
input.commit();
}
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::Key { serial, .. } | WlKeyboardEvent::Modifiers { serial, .. } => {
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::RepeatInfo { rate, delay } => {
*self.key_repeat_rate.borrow_mut() = *rate;
*self.key_repeat_delay.borrow_mut() = *delay;
}
WlKeyboardEvent::Keymap { format, fd, size } => {
let file = unsafe { std::fs::File::from_raw_fd(*fd) };
match format {
KeymapFormat::XkbV1 => {
let mut data = vec![0u8; *size as usize];
file.read_exact_at(&mut data, 0)?;
// Dance around CString panicing on the NUL terminator
// in the xkbcommon crate
while let Some(0) = data.last() {
data.pop();
}
let s = String::from_utf8(data)?;
match Keyboard::new_from_string(s) {
Ok(k) => {
self.keyboard_mapper.replace(Some(k));
}
Err(err) => {
log::error!("Error processing keymap change: {:#}", err);
}
}
}
_ => {}
}
}
_ => {}
}
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.keyboard_event(event);
}
}
Ok(())
}
pub(crate) fn dispatch_to_focused_window(&self, event: WindowEvent) {
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.events.dispatch(event);
}
}
}
pub(crate) fn next_window_id(&self) -> usize {
self.next_window_id
.fetch_add(1, ::std::sync::atomic::Ordering::Relaxed)
}
fn flush(&self) -> anyhow::Result<()> {
if let Err(e) = self.display.borrow_mut().flush() {
if e.kind() != ::std::io::ErrorKind::WouldBlock {
bail!("Error while flushing display: {}", e);
}
}
Ok(())
}
pub(crate) fn window_by_id(&self, window_id: usize) -> Option<Rc<RefCell<WaylandWindowInner>>> {
self.windows.borrow().get(&window_id).map(Rc | keyboard_event | identifier_name |
connection.rs | 2, usize>>,
pub(crate) active_surface_id: RefCell<u32>,
/// Repeats per second
pub(crate) key_repeat_rate: RefCell<i32>,
pub(crate) mem_pool: RefCell<AutoMemPool>,
/// Delay before repeating, in milliseconds
pub(crate) key_repeat_delay: RefCell<i32>,
pub(crate) last_serial: RefCell<u32>,
seat_listener: SeatListener,
pub(crate) environment: Environment<MyEnvironment>,
event_q: RefCell<EventQueue>,
pub(crate) display: RefCell<Display>,
}
impl Drop for WaylandConnection {
fn drop(&mut self) {
self.environment
.with_inner(|env| env.input_handler.shutdown());
}
}
impl WaylandConnection {
pub fn create_new() -> anyhow::Result<Self> {
let (environment, display, event_q) = toolkit::new_default_environment!(
MyEnvironment,
desktop,
fields = [
output_handler: OutputHandler::new(),
input_handler: InputHandler::new(),
]
)?;
let mut pointer = None;
let mut seat_keyboards = HashMap::new();
for seat in environment.get_all_seats() {
if let Some((has_kbd, has_ptr, name)) =
toolkit::seat::with_seat_data(&seat, |seat_data| {
(
seat_data.has_keyboard && !seat_data.defunct,
seat_data.has_pointer && !seat_data.defunct,
seat_data.name.clone(),
)
})
{
if has_kbd {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
environment.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(name, keyboard);
}
if has_ptr {
pointer.replace(PointerDispatcher::register(
&seat,
environment.require_global(),
environment.require_global(),
environment.require_global(),
environment.get_primary_selection_manager(),
)?);
}
}
}
let pointer =
pointer.ok_or_else(|| anyhow::anyhow!("no seats have an available pointer"))?;
let seat_listener;
{
let env = environment.clone();
seat_listener = environment.listen_for_seats(move |seat, seat_data, _| {
if seat_data.has_keyboard {
if !seat_data.defunct {
// We only want to assign a new keyboard object if we don't already have
// one for this seat. When a seat is being created or updated, the listener
// can receive the same seat multiple times: for example, when switching
// back from another virtual console, the same seat is usually seen four
// times with different data flags:
//
// has_pointer: true; has_keyboard: false
// has_pointer: false; has_keyboard: false
// has_pointer: false; has_keyboard: true
// has_pointer: true; has_keyboard: true
//
// This is essentially telling the client to re-assign its keyboard and
// pointer, but that means that this listener will fire twice with
// has_keyboard set to true. If we assign a handler both times, then we end
// up handling key events twice.
if !seat_keyboards.contains_key(&seat_data.name) {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
env.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(seat_data.name.clone(), keyboard);
}
} else {
env.with_inner(|env| env.input_handler.seat_defunct(&seat));
}
} else {
// If we previously had a keyboard object on this seat, it's no longer valid if
// has_keyboard is false, so we remove the keyboard object we knew about and
// thereby ensure that we assign a new keyboard object next time the listener
// fires for this seat with has_keyboard = true.
seat_keyboards.remove(&seat_data.name);
}
if seat_data.has_pointer && !seat_data.defunct {
let conn = Connection::get().unwrap().wayland();
conn.pointer.borrow_mut().seat_changed(&seat);
}
});
}
let mem_pool = environment.create_auto_pool()?;
Ok(Self {
display: RefCell::new(display),
environment,
should_terminate: RefCell::new(false),
next_window_id: AtomicUsize::new(1), | windows: RefCell::new(HashMap::new()),
event_q: RefCell::new(event_q),
pointer: RefCell::new(pointer),
seat_listener,
mem_pool: RefCell::new(mem_pool),
gl_connection: RefCell::new(None),
keyboard_mapper: RefCell::new(None),
key_repeat_rate: RefCell::new(25),
key_repeat_delay: RefCell::new(400),
keyboard_window_id: RefCell::new(None),
last_serial: RefCell::new(0),
surface_to_window_id: RefCell::new(HashMap::new()),
active_surface_id: RefCell::new(0),
})
}
fn keyboard_event(
&self,
keyboard: Main<WlKeyboard>,
event: WlKeyboardEvent,
) -> anyhow::Result<()> {
match &event {
WlKeyboardEvent::Enter {
serial, surface, ..
} => {
// update global active surface id
*self.active_surface_id.borrow_mut() = surface.as_ref().id();
*self.last_serial.borrow_mut() = *serial;
if let Some(&window_id) = self
.surface_to_window_id
.borrow()
.get(&surface.as_ref().id())
{
self.keyboard_window_id.borrow_mut().replace(window_id);
self.environment.with_inner(|env| {
if let Some(input) =
env.input_handler.get_text_input_for_keyboard(&keyboard)
{
input.enable();
input.commit();
}
env.input_handler.advise_surface(&surface, &keyboard);
});
} else {
log::warn!("{:?}, no known surface", event);
}
}
WlKeyboardEvent::Leave { serial, .. } => {
if let Some(input) = self
.environment
.with_inner(|env| env.input_handler.get_text_input_for_keyboard(&keyboard))
{
input.disable();
input.commit();
}
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::Key { serial, .. } | WlKeyboardEvent::Modifiers { serial, .. } => {
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::RepeatInfo { rate, delay } => {
*self.key_repeat_rate.borrow_mut() = *rate;
*self.key_repeat_delay.borrow_mut() = *delay;
}
WlKeyboardEvent::Keymap { format, fd, size } => {
let file = unsafe { std::fs::File::from_raw_fd(*fd) };
match format {
KeymapFormat::XkbV1 => {
let mut data = vec![0u8; *size as usize];
file.read_exact_at(&mut data, 0)?;
// Dance around CString panicing on the NUL terminator
// in the xkbcommon crate
while let Some(0) = data.last() {
data.pop();
}
let s = String::from_utf8(data)?;
match Keyboard::new_from_string(s) {
Ok(k) => {
self.keyboard_mapper.replace(Some(k));
}
Err(err) => {
log::error!("Error processing keymap change: {:#}", err);
}
}
}
_ => {}
}
}
_ => {}
}
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.keyboard_event(event);
}
}
Ok(())
}
pub(crate) fn dispatch_to_focused_window(&self, event: WindowEvent) {
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.events.dispatch(event);
}
}
}
pub(crate) fn next_window_id(&self) -> usize {
self.next_window_id
.fetch_add(1, ::std::sync::atomic::Ordering::Relaxed)
}
fn flush(&self) -> anyhow::Result<()> {
if let Err(e) = self.display.borrow_mut().flush() {
if e.kind() != ::std::io::ErrorKind::WouldBlock {
bail!("Error while flushing display: {}", e);
}
}
Ok(())
}
pub(crate) fn window_by_id(&self, window_id: usize) -> Option<Rc<RefCell<WaylandWindowInner>>> {
self.windows.borrow().get(&window_id).map(Rc:: | random_line_split | |
connection.rs | , usize>>,
pub(crate) active_surface_id: RefCell<u32>,
/// Repeats per second
pub(crate) key_repeat_rate: RefCell<i32>,
pub(crate) mem_pool: RefCell<AutoMemPool>,
/// Delay before repeating, in milliseconds
pub(crate) key_repeat_delay: RefCell<i32>,
pub(crate) last_serial: RefCell<u32>,
seat_listener: SeatListener,
pub(crate) environment: Environment<MyEnvironment>,
event_q: RefCell<EventQueue>,
pub(crate) display: RefCell<Display>,
}
impl Drop for WaylandConnection {
fn drop(&mut self) {
self.environment
.with_inner(|env| env.input_handler.shutdown());
}
}
impl WaylandConnection {
pub fn create_new() -> anyhow::Result<Self> {
let (environment, display, event_q) = toolkit::new_default_environment!(
MyEnvironment,
desktop,
fields = [
output_handler: OutputHandler::new(),
input_handler: InputHandler::new(),
]
)?;
let mut pointer = None;
let mut seat_keyboards = HashMap::new();
for seat in environment.get_all_seats() {
if let Some((has_kbd, has_ptr, name)) =
toolkit::seat::with_seat_data(&seat, |seat_data| {
(
seat_data.has_keyboard && !seat_data.defunct,
seat_data.has_pointer && !seat_data.defunct,
seat_data.name.clone(),
)
})
{
if has_kbd {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
environment.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(name, keyboard);
}
if has_ptr {
pointer.replace(PointerDispatcher::register(
&seat,
environment.require_global(),
environment.require_global(),
environment.require_global(),
environment.get_primary_selection_manager(),
)?);
}
}
}
let pointer =
pointer.ok_or_else(|| anyhow::anyhow!("no seats have an available pointer"))?;
let seat_listener;
{
let env = environment.clone();
seat_listener = environment.listen_for_seats(move |seat, seat_data, _| {
if seat_data.has_keyboard {
if !seat_data.defunct {
// We only want to assign a new keyboard object if we don't already have
// one for this seat. When a seat is being created or updated, the listener
// can receive the same seat multiple times: for example, when switching
// back from another virtual console, the same seat is usually seen four
// times with different data flags:
//
// has_pointer: true; has_keyboard: false
// has_pointer: false; has_keyboard: false
// has_pointer: false; has_keyboard: true
// has_pointer: true; has_keyboard: true
//
// This is essentially telling the client to re-assign its keyboard and
// pointer, but that means that this listener will fire twice with
// has_keyboard set to true. If we assign a handler both times, then we end
// up handling key events twice.
if !seat_keyboards.contains_key(&seat_data.name) {
let keyboard = seat.get_keyboard();
keyboard.quick_assign(|keyboard, event, _| {
let conn = Connection::get().unwrap().wayland();
if let Err(err) = conn.keyboard_event(keyboard, event) {
log::error!("keyboard_event: {:#}", err);
}
});
env.with_inner(|env| env.input_handler.advise_seat(&seat, &keyboard));
seat_keyboards.insert(seat_data.name.clone(), keyboard);
}
} else {
env.with_inner(|env| env.input_handler.seat_defunct(&seat));
}
} else {
// If we previously had a keyboard object on this seat, it's no longer valid if
// has_keyboard is false, so we remove the keyboard object we knew about and
// thereby ensure that we assign a new keyboard object next time the listener
// fires for this seat with has_keyboard = true.
seat_keyboards.remove(&seat_data.name);
}
if seat_data.has_pointer && !seat_data.defunct {
let conn = Connection::get().unwrap().wayland();
conn.pointer.borrow_mut().seat_changed(&seat);
}
});
}
let mem_pool = environment.create_auto_pool()?;
Ok(Self {
display: RefCell::new(display),
environment,
should_terminate: RefCell::new(false),
next_window_id: AtomicUsize::new(1),
windows: RefCell::new(HashMap::new()),
event_q: RefCell::new(event_q),
pointer: RefCell::new(pointer),
seat_listener,
mem_pool: RefCell::new(mem_pool),
gl_connection: RefCell::new(None),
keyboard_mapper: RefCell::new(None),
key_repeat_rate: RefCell::new(25),
key_repeat_delay: RefCell::new(400),
keyboard_window_id: RefCell::new(None),
last_serial: RefCell::new(0),
surface_to_window_id: RefCell::new(HashMap::new()),
active_surface_id: RefCell::new(0),
})
}
fn keyboard_event(
&self,
keyboard: Main<WlKeyboard>,
event: WlKeyboardEvent,
) -> anyhow::Result<()> {
match &event {
WlKeyboardEvent::Enter {
serial, surface, ..
} => {
// update global active surface id
*self.active_surface_id.borrow_mut() = surface.as_ref().id();
*self.last_serial.borrow_mut() = *serial;
if let Some(&window_id) = self
.surface_to_window_id
.borrow()
.get(&surface.as_ref().id())
{
self.keyboard_window_id.borrow_mut().replace(window_id);
self.environment.with_inner(|env| {
if let Some(input) =
env.input_handler.get_text_input_for_keyboard(&keyboard)
{
input.enable();
input.commit();
}
env.input_handler.advise_surface(&surface, &keyboard);
});
} else {
log::warn!("{:?}, no known surface", event);
}
}
WlKeyboardEvent::Leave { serial, .. } => {
if let Some(input) = self
.environment
.with_inner(|env| env.input_handler.get_text_input_for_keyboard(&keyboard))
{
input.disable();
input.commit();
}
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::Key { serial, .. } | WlKeyboardEvent::Modifiers { serial, .. } => {
*self.last_serial.borrow_mut() = *serial;
}
WlKeyboardEvent::RepeatInfo { rate, delay } => {
*self.key_repeat_rate.borrow_mut() = *rate;
*self.key_repeat_delay.borrow_mut() = *delay;
}
WlKeyboardEvent::Keymap { format, fd, size } => {
let file = unsafe { std::fs::File::from_raw_fd(*fd) };
match format {
KeymapFormat::XkbV1 => {
let mut data = vec![0u8; *size as usize];
file.read_exact_at(&mut data, 0)?;
// Dance around CString panicing on the NUL terminator
// in the xkbcommon crate
while let Some(0) = data.last() {
data.pop();
}
let s = String::from_utf8(data)?;
match Keyboard::new_from_string(s) {
Ok(k) => {
self.keyboard_mapper.replace(Some(k));
}
Err(err) => {
log::error!("Error processing keymap change: {:#}", err);
}
}
}
_ => {}
}
}
_ => {}
}
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.keyboard_event(event);
}
}
Ok(())
}
pub(crate) fn dispatch_to_focused_window(&self, event: WindowEvent) |
pub(crate) fn next_window_id(&self) -> usize {
self.next_window_id
.fetch_add(1, ::std::sync::atomic::Ordering::Relaxed)
}
fn flush(&self) -> anyhow::Result<()> {
if let Err(e) = self.display.borrow_mut().flush() {
if e.kind() != ::std::io::ErrorKind::WouldBlock {
bail!("Error while flushing display: {}", e);
}
}
Ok(())
}
pub(crate) fn window_by_id(&self, window_id: usize) -> Option<Rc<RefCell<WaylandWindowInner>>> {
self.windows.borrow().get(&window_id).map(Rc | {
if let Some(&window_id) = self.keyboard_window_id.borrow().as_ref() {
if let Some(win) = self.window_by_id(window_id) {
let mut inner = win.borrow_mut();
inner.events.dispatch(event);
}
}
} | identifier_body |
substitution.rs | forward_type_interner_methods!(Symbol, T, self_, &self_.interner);
}
impl<'a> types::Substitution<Symbol, RcType> for &'a Substitution<RcType> {
fn new_var(&mut self) -> RcType {
Substitution::new_var(*self)
}
fn new_skolem(&mut self, name: Symbol, kind: ArcKind) -> RcType {
Substitution::new_skolem(*self, name, kind)
}
}
impl<T> Default for Substitution<T>
where
T: Substitutable,
T::Factory: Default,
T::Interner: Default,
{
fn default() -> Substitution<T> {
Substitution::new(Default::default(), Default::default())
}
}
/// Trait which variables need to implement to allow the substitution to get to the u32 identifying
/// the variable
pub trait Variable {
fn get_id(&self) -> u32;
}
impl Variable for u32 {
fn get_id(&self) -> u32 {
*self
}
}
pub trait VariableFactory {
type Variable: Variable;
fn new(&self, x: u32) -> Self::Variable;
}
impl VariableFactory for () {
type Variable = u32;
fn new(&self, x: u32) -> Self::Variable {
x
}
}
/// Trait implemented on types which may contain substitutable variables
pub trait Substitutable: Sized {
type Variable: Variable;
type Factory: VariableFactory<Variable = Self::Variable>;
type Interner: Default;
/// Constructs a new object from its variable type
fn from_variable(subs: &Substitution<Self>, x: Self::Variable) -> Self;
fn into_variable(&mut self, x: Self::Variable);
fn is_unique(self_: &Self) -> bool;
/// Retrieves the variable if `self` is a variable otherwise returns `None`
fn get_var(&self) -> Option<&Self::Variable>;
fn get_id(&self) -> Option<u32> {
self.get_var().map(|var| var.get_id())
}
fn traverse<'a, F>(&'a self, f: &mut F)
where
F: Walker<'a, Self>;
fn instantiate(&self, subs: &Substitution<Self>) -> Self;
// Allowed return true even if the type does not contain variables but not false if it does
// contain
fn contains_variables(&self) -> bool {
true
}
fn on_union(&self) -> Option<&Self> {
None
}
}
pub fn occurs<T>(typ: &T, subs: &Substitution<T>, var: u32) -> bool
where
T: Substitutable,
{
struct Occurs<'a, T: Substitutable + 'a> {
occurs: bool,
var: u32,
subs: &'a Substitution<T>,
}
impl<'a, 't, T> Walker<'t, T> for Occurs<'a, T>
where
T: Substitutable,
{
fn walk(&mut self, typ: &'t T) {
if !typ.contains_variables() || self.occurs {
return;
}
let typ = self.subs.real(typ);
if let Some(other) = typ.get_var() {
if self.var.get_id() == other.get_id() {
self.occurs = true;
typ.traverse(self);
return;
}
self.subs.update_level(self.var, other.get_id());
}
typ.traverse(self);
}
}
let mut occurs = Occurs {
occurs: false,
var,
subs,
};
occurs.walk(typ);
occurs.occurs
}
/// Specialized union implementation which makes sure that variables with a higher level always
/// point to the lower level variable.
///
/// map.union(1, 2);
/// map.find(2) -> 1
/// map.find(1) -> 1
#[derive(Debug)]
struct UnionByLevel {
rank: UnionByRank,
level: u32,
}
impl Default for UnionByLevel {
fn default() -> UnionByLevel {
UnionByLevel {
rank: UnionByRank::default(),
level: ::std::u32::MAX,
}
}
}
impl Union for UnionByLevel {
#[inline]
fn union(left: UnionByLevel, right: UnionByLevel) -> UnionResult<UnionByLevel> {
use std::cmp::Ordering;
let (rank_result, rank) = match Union::union(left.rank, right.rank) {
UnionResult::Left(l) => (
UnionResult::Left(UnionByLevel {
rank: l,
level: left.level,
}),
l,
),
UnionResult::Right(r) => (
UnionResult::Right(UnionByLevel {
rank: r,
level: left.level,
}),
r,
),
};
match left.level.cmp(&right.level) {
Ordering::Less => UnionResult::Left(UnionByLevel {
rank: rank,
level: left.level,
}),
Ordering::Greater => UnionResult::Right(UnionByLevel {
rank: rank,
level: right.level,
}),
Ordering::Equal => rank_result,
}
}
}
impl<T> fmt::Debug for Substitution<T>
where
T: fmt::Debug + Substitutable,
{
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
write!(
f,
"Substitution {{ map: {:?}, var_id: {:?} }}",
self.union.borrow(),
self.var_id()
)
}
}
impl<T> Substitution<T>
where
T: Substitutable,
{
pub fn new(factory: T::Factory, interner: T::Interner) -> Substitution<T> {
Substitution {
union: RefCell::new(QuickFindUf::new(0)),
variables: FixedVec::new(),
types: FixedVecMap::new(),
factory: factory,
interner,
variable_cache: Default::default(),
}
}
pub fn var_id(&self) -> u32 {
self.variables.len() as u32
}
pub fn insert(&self, var: u32, t: T) {
match t.get_var() {
Some(_) => ice!(
"Tried to insert variable which is not allowed as that would cause memory \
unsafety"
),
None => match self.types.try_insert(var as usize, t.into()) {
Ok(()) => (),
Err(_) => ice!("Expected variable to not have a type associated with it"),
},
}
}
pub fn replace(&mut self, var: u32, t: T) {
debug_assert!(t.get_id() != Some(var));
self.types.insert(var as usize, t.into());
}
pub fn reset(&mut self, var: u32) {
self.types.remove(var as usize);
}
/// Assumes that no variables unified with anything (but variables < level may exist)
pub fn clear_from(&mut self, level: u32) {
self.union = RefCell::new(QuickFindUf::new(0));
let mut u = self.union.borrow_mut();
for _ in 0..level {
u.insert(UnionByLevel {
..UnionByLevel::default()
});
}
let mut variable_cache = self.variable_cache.borrow_mut();
// Since no types should be unified with anything we can remove all of this and reuse the
// unique values
variable_cache.extend(self.types.drain().filter(T::is_unique));
while self.variables.len() > level as usize {
variable_cache.push(self.variables.pop().unwrap());
}
}
/// Creates a new variable
pub fn new_var(&self) -> T
where
T: Clone,
{
self.new_var_fn(|var| match self.variable_cache.borrow_mut().pop() {
Some(mut typ) => {
T::into_variable(&mut typ, self.factory.new(var));
typ
}
None => T::from_variable(self, self.factory.new(var)),
})
}
pub fn new_var_fn<F>(&self, f: F) -> T
where
T: Clone,
F: FnOnce(u32) -> T,
{
let var_id = self.variables.len() as u32;
let id = self.union.borrow_mut().insert(UnionByLevel {
level: var_id,
..UnionByLevel::default()
});
assert!(id == self.variables.len());
debug!("New var {}", self.variables.len());
let var = f(var_id);
self.variables.push(var.clone().into());
var
}
/// If `typ` is a variable this returns the real unified value of that variable. Otherwise it
/// just returns the type itself. Note that the returned type may contain terms which also need
/// to have `real` called on them.
pub fn real<'r>(&'r self, typ: &'r T) -> &'r T {
match typ.get_id() {
Some(id) => match self.find_type_for_var(id) {
Some(t) => t,
None => typ,
},
_ => typ,
}
}
pub fn | get_var | identifier_name | |
substitution.rs | pub trait Variable {
fn get_id(&self) -> u32;
}
impl Variable for u32 {
fn get_id(&self) -> u32 {
*self
}
}
pub trait VariableFactory {
type Variable: Variable;
fn new(&self, x: u32) -> Self::Variable;
}
impl VariableFactory for () {
type Variable = u32;
fn new(&self, x: u32) -> Self::Variable {
x
}
}
/// Trait implemented on types which may contain substitutable variables
pub trait Substitutable: Sized {
type Variable: Variable;
type Factory: VariableFactory<Variable = Self::Variable>;
type Interner: Default;
/// Constructs a new object from its variable type
fn from_variable(subs: &Substitution<Self>, x: Self::Variable) -> Self;
fn into_variable(&mut self, x: Self::Variable);
fn is_unique(self_: &Self) -> bool;
/// Retrieves the variable if `self` is a variable otherwise returns `None`
fn get_var(&self) -> Option<&Self::Variable>;
fn get_id(&self) -> Option<u32> {
self.get_var().map(|var| var.get_id())
}
fn traverse<'a, F>(&'a self, f: &mut F)
where
F: Walker<'a, Self>;
fn instantiate(&self, subs: &Substitution<Self>) -> Self;
// Allowed return true even if the type does not contain variables but not false if it does
// contain
fn contains_variables(&self) -> bool {
true
}
fn on_union(&self) -> Option<&Self> {
None
}
}
pub fn occurs<T>(typ: &T, subs: &Substitution<T>, var: u32) -> bool
where
T: Substitutable,
{
struct Occurs<'a, T: Substitutable + 'a> {
occurs: bool,
var: u32,
subs: &'a Substitution<T>,
}
impl<'a, 't, T> Walker<'t, T> for Occurs<'a, T>
where
T: Substitutable,
{
fn walk(&mut self, typ: &'t T) {
if !typ.contains_variables() || self.occurs {
return;
}
let typ = self.subs.real(typ);
if let Some(other) = typ.get_var() {
if self.var.get_id() == other.get_id() {
self.occurs = true;
typ.traverse(self);
return;
}
self.subs.update_level(self.var, other.get_id());
}
typ.traverse(self);
}
}
let mut occurs = Occurs {
occurs: false,
var,
subs,
};
occurs.walk(typ);
occurs.occurs
}
/// Specialized union implementation which makes sure that variables with a higher level always
/// point to the lower level variable.
///
/// map.union(1, 2);
/// map.find(2) -> 1
/// map.find(1) -> 1
#[derive(Debug)]
struct UnionByLevel {
rank: UnionByRank,
level: u32,
}
impl Default for UnionByLevel {
fn default() -> UnionByLevel {
UnionByLevel {
rank: UnionByRank::default(),
level: ::std::u32::MAX,
}
}
}
impl Union for UnionByLevel {
#[inline]
fn union(left: UnionByLevel, right: UnionByLevel) -> UnionResult<UnionByLevel> {
use std::cmp::Ordering;
let (rank_result, rank) = match Union::union(left.rank, right.rank) {
UnionResult::Left(l) => (
UnionResult::Left(UnionByLevel {
rank: l,
level: left.level,
}),
l,
),
UnionResult::Right(r) => (
UnionResult::Right(UnionByLevel {
rank: r,
level: left.level,
}),
r,
),
};
match left.level.cmp(&right.level) {
Ordering::Less => UnionResult::Left(UnionByLevel {
rank: rank,
level: left.level,
}),
Ordering::Greater => UnionResult::Right(UnionByLevel {
rank: rank,
level: right.level,
}),
Ordering::Equal => rank_result,
}
}
}
impl<T> fmt::Debug for Substitution<T>
where
T: fmt::Debug + Substitutable,
{
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
write!(
f,
"Substitution {{ map: {:?}, var_id: {:?} }}",
self.union.borrow(),
self.var_id()
)
}
}
impl<T> Substitution<T>
where
T: Substitutable,
{
pub fn new(factory: T::Factory, interner: T::Interner) -> Substitution<T> {
Substitution {
union: RefCell::new(QuickFindUf::new(0)),
variables: FixedVec::new(),
types: FixedVecMap::new(),
factory: factory,
interner,
variable_cache: Default::default(),
}
}
pub fn var_id(&self) -> u32 {
self.variables.len() as u32
}
pub fn insert(&self, var: u32, t: T) {
match t.get_var() {
Some(_) => ice!(
"Tried to insert variable which is not allowed as that would cause memory \
unsafety"
),
None => match self.types.try_insert(var as usize, t.into()) {
Ok(()) => (),
Err(_) => ice!("Expected variable to not have a type associated with it"),
},
}
}
pub fn replace(&mut self, var: u32, t: T) {
debug_assert!(t.get_id() != Some(var));
self.types.insert(var as usize, t.into());
}
pub fn reset(&mut self, var: u32) {
self.types.remove(var as usize);
}
/// Assumes that no variables unified with anything (but variables < level may exist)
pub fn clear_from(&mut self, level: u32) {
self.union = RefCell::new(QuickFindUf::new(0));
let mut u = self.union.borrow_mut();
for _ in 0..level {
u.insert(UnionByLevel {
..UnionByLevel::default()
});
}
let mut variable_cache = self.variable_cache.borrow_mut();
// Since no types should be unified with anything we can remove all of this and reuse the
// unique values
variable_cache.extend(self.types.drain().filter(T::is_unique));
while self.variables.len() > level as usize {
variable_cache.push(self.variables.pop().unwrap());
}
}
/// Creates a new variable
pub fn new_var(&self) -> T
where
T: Clone,
{
self.new_var_fn(|var| match self.variable_cache.borrow_mut().pop() {
Some(mut typ) => {
T::into_variable(&mut typ, self.factory.new(var));
typ
}
None => T::from_variable(self, self.factory.new(var)),
})
}
pub fn new_var_fn<F>(&self, f: F) -> T
where
T: Clone,
F: FnOnce(u32) -> T,
{
let var_id = self.variables.len() as u32;
let id = self.union.borrow_mut().insert(UnionByLevel {
level: var_id,
..UnionByLevel::default()
});
assert!(id == self.variables.len());
debug!("New var {}", self.variables.len());
let var = f(var_id);
self.variables.push(var.clone().into());
var
}
/// If `typ` is a variable this returns the real unified value of that variable. Otherwise it
/// just returns the type itself. Note that the returned type may contain terms which also need
/// to have `real` called on them.
pub fn real<'r>(&'r self, typ: &'r T) -> &'r T {
match typ.get_id() {
Some(id) => match self.find_type_for_var(id) {
Some(t) => t,
None => typ,
},
_ => typ,
}
}
pub fn get_var(&self, var: u32) -> Option<&T> {
self.variables.get(var as usize)
}
pub fn find_type_for_var(&self, var: u32) -> Option<&T> {
let mut union = self.union.borrow_mut();
if var as usize >= union.size() {
return None;
}
let index = union.find(var as usize);
self.types.get(index).or_else(|| {
if var == index as u32 {
None
} else {
Some(&self.variables[index as usize])
}
})
} |
/// Updates the level of `other` to be the minimum level value of `var` and `other`
pub fn update_level(&self, var: u32, other: u32) {
let level = ::std::cmp::min(self.get_level(var), self.get_level(other)); | random_line_split | |
index.ts | embed(url: string, container: Element, options?: IEmbeddOptions): Promise<Glue> {
const state: {
glue?: Glue;
beforeInitResolve?: (value?: unknown) => void;
beforeInitReject?: (reason?: unknown) => void;
retryTimer?: ReturnType<typeof setTimeout>;
} = {};
return new Promise((resolve, reject) => {
// Add default option values and ensure options.
options = {
timeout: 5000,
sandboxRestrictions: 'allow-forms allow-popups allow-popups-to-escape-sandbox allow-scripts allow-same-origin',
featurePolicy: 'animations; autoplay; camera; encrypted-media; fullscreen; geolocation; microphone; speaker; vr',
...options,
}
const src = new URL(url, window.location.href);
const origin = options.origin ? options.origin : src.origin;
const features = options.features;
const mode = options.mode ? options.mode : '';
// Create glue controller.
const controller = new Controller({
origin,
handler: async (message: IPayload): Promise<any> => { /* eslint-disable-line @typescript-eslint/no-explicit-any */
switch (message.type) {
case 'init': {
if (state.retryTimer) {
clearTimeout(state.retryTimer);
}
const data = message.data as IInitData;
const reply: IInitData = {
features: features ? Object.keys(features) : [],
};
const api = {} as API<{[key: string]: (...args: unknown[]) => Promise<any>}>; /* eslint-disable-line @typescript-eslint/no-explicit-any */
if (data.features) {
data.features.forEach(action => {
api[action] = (...args: unknown[]): Promise<any> => { /* eslint-disable-line @typescript-eslint/no-explicit-any */
return controller.callAction(action, args);
}
});
}
state.glue = controller.Glue({api, mode});
if (options && options.onBeforeInit) {
const p = new Promise((resolve, reject) => {
state.beforeInitResolve = resolve;
state.beforeInitReject = reject;
});
if (!state.beforeInitResolve || !state.beforeInitReject) {
throw new Error('glue init promise error');
}
try {
const action = options.onBeforeInit(state.glue, p);
if (action) {
if (!data.features || !data.features.includes(action)) {
state.beforeInitReject(new Error(`unsupported action: ${action}`));
} else {
reply.action = action;
}
}
} catch (err) {
reject(new Error(`onInit failed: ${err}`));
return;
}
}
return reply;
}
case 'ready': {
if (!state.glue) {
throw new Error('failed to glue: no state');
}
//const glue = controller.Glue(state.api);
const data = message.data as IReadyData;
if (options && options.onBeforeInit && state.beforeInitResolve && state.beforeInitReject) {
if (data.ready) {
await state.beforeInitResolve(data.data);
} else {
await state.beforeInitReject(data.data);
}
resolve(state.glue);
} else {
if (data.ready) {
resolve(state.glue);
} else {
if (data.error) {
throw new Error(`failed to glue: ${data.data}`)
} else {
reject(state.glue);
}
}
}
break;
}
case 'call': {
const data = message.data as ICallData;
const handler = features ? features[data.action] : null;
if (!handler) {
throw new Error(`unknown action: ${data.action}`);
}
const args = data.args ? data.args : [];
return handler(...args);
}
default:
console.debug(`glue (embed) unknown message type: ${message.type}`);
}
},
});
// Create iframe.
const ownerDocument = container.ownerDocument !== null ? container.ownerDocument : document;
const frame = ownerDocument.createElement('iframe');
if (options && options.className) {
frame.className = options.className;
}
if (options.sandboxRestrictions) {
frame.setAttribute('sandbox', options.sandboxRestrictions);
}
if (options.featurePolicy) {
frame.setAttribute('allow', options.featurePolicy);
}
if (options && options.attributes) {
Object.entries(options.attributes).forEach(([key, value]) => {
frame.setAttribute(key, value);
});
}
// Prepare URL and set it to element.
setGlueParameter(src, 'mode', mode);
if (origin !== window.origin) {
// Cross origin, add glue origin hash parameter to allow white list
// checks on the other end.
setGlueParameter(src, 'origin', origin);
}
frame.setAttribute('src', src.toString());
// Append iframe with timeout and retry.
const append = (): void => {
// Inject iframe and attach glue.
container.appendChild(frame);
if (!frame.contentWindow) {
throw new Error('new frame has no contentWindow');
}
controller.attach(frame.contentWindow);
}
const retry = (): void => {
controller.detach();
container.removeChild(frame);
setTimeout(() => {
append();
}, 1000); // NOTE(longsleep): Retry time hardcoded - is it needed to have a configuration?
}
frame.addEventListener('load', () => {
if (state.glue) |
if (options && options.timeout) {
state.retryTimer = setTimeout(() => {
if (!state.glue) {
retry();
}
}, options.timeout);
} else {
reject(new Error('glue timeout'));
}
});
append();
});
}
/**
* Enables glue for the provided sourceWindow with options.
*
* @param sourceWindow Window element of the app which is using Glue embed.
* @param options Enable options.
*/
async function enable(sourceWindow?: Window, options?: IEnableOptions): Promise<Glue> {
return new Promise((resolve, reject) => {
if (!sourceWindow) {
sourceWindow = window.parent;
}
// Get glue mode.
const mode = getGlueParameter('mode');
if (sourceWindow === self || mode === null) {
// Return empty Glue API if we are self, or glue mode is not set. It
// this means Glue is not active.
resolve(new Glue({}));
return;
}
// Add default option values and ensure options.
options = {
timeout: 5000,
...options,
};
// Validate origin.
const expectedOrigin = getGlueParameter('origin');
if (expectedOrigin) {
if (expectedOrigin !== window.origin) {
// Validate white list if cross origin.
if (!options || !options.origins || !options.origins.includes('expectedOrigin')) {
throw new Error('glue origin is not allowed');
}
}
}
// Create glue controller.
const features = options.features;
const controller = new Controller({
origin: expectedOrigin ? expectedOrigin : window.origin,
handler: async (message: IPayload): Promise<any> => { /* eslint-disable-line @typescript-eslint/no-explicit-any */
switch (message.type) {
case 'call': {
const data = message.data as ICallData;
const handler = features ? features[data.action] : null;
if (!handler) {
throw new Error(`unknown action: ${data.action}`);
}
const args = data.args ? data.args : [];
return handler(...args);
}
default:
console.debug(`glue (enable) unknown message type: ${message.type}`)
}
},
});
// Attach glue.
controller.attach(sourceWindow);
// Start timeout.
let failed = false;
const timer = setTimeout(() => {
failed = true;
reject(new Error('glue timeout'));
}, options.timeout);
// Start initialization.
queueMicroTask(() => {
const request: IInitData = {
features: features ? Object.keys(features) : [],
mode,
}
controller.postMessage('init', request).then(async (initData?: IInitData): Promise<void> => {
clearTimeout(timer);
if (failed) {
// Do nothing when flagged failed.
return;
}
if (!initData || initData.error) {
// TODO(longsleep): Initialization failed. What now?
reject(new Error(`glue init received error: ${initData ? initData.error : 'no data'}`));
return;
}
const readyData: IReadyData = {
ready: true,
}
// Create API action handlers.
const api = {} as API<{[key: string]: (...args: unknown[]) => Promise<any>}>; /* eslint-disable-line @typescript-eslint/no-explicit-any */
if (initData.features) {
for (const action of initData.features) {
api | {
delete state.glue;
} | conditional_block |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.