file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
39k
suffix
large_stringlengths
0
36.1k
middle
large_stringlengths
0
29.4k
fim_type
large_stringclasses
4 values
lib.rs
#[macro_use] extern crate compre_combinee; extern crate combine; mod errors; mod details; mod traits; mod stop_watch; use std::collections::{HashMap}; use combine::{parser, eof, satisfy, choice, attempt}; use combine::parser::range::{take_while1}; use combine::parser::char::*; use combine::{Parser, many, optional, skip_many, sep_by, between}; pub use crate::errors::ErrorCause; pub use crate::details::Node; pub use crate::traits::*; use std::{f64, mem, str}; use std::convert::TryFrom; use smol_str::SmolStr; fn parse_hex<'a>() -> impl Parser<&'a str, Output = u32> { satisfy(|c: char| (c >= '0' && c <= '9') || (c >= 'a' && c <= 'f') || (c >= 'A' && c <= 'F') ).map(|c| if c >= '0' && c <= '9' { c as u64 - '0' as u64 } else if c >= 'a' && c <= 'f' { 10 + c as u64 - 'a' as u64 } else { 10 + c as u64 - 'A' as u64 } as u32 ) } fn unicode_char<'a>() -> impl Parser<&'a str, Output = Option<char>> { c_hx_do!{ __ <- string(r#"\u"#), d3 <- parse_hex(), d2 <- parse_hex(), d1 <- parse_hex(), d0 <- parse_hex(); { let unicode = d0 + 0x10 * d1 + 0x100 * d2 + 0x1000 * d3; char::try_from(unicode).ok() } } } #[derive(PartialEq)] enum StringPiece<'a > { Ref(&'a str), Char(Option<char>) } fn braced_parser<'a, PBL, P, PBR, O>(pbl: PBL, p: P, pbr: PBR) -> impl Parser<&'a str, Output = O> where PBL: Parser<&'a str>, PBR: Parser<&'a str>, P: Parser<&'a str, Output = O> { between( c_compre![c; c <- pbl, __ <- skip_many(space())], c_compre![c; __ <- skip_many(space()), c <- pbr], p ) } fn string_part<'a>() -> impl Parser<&'a str, Output = Vec<StringPiece<'a >>> { many( choice( ( attempt(take_while1(|c: char| c != '\\' && c != '"' && c != '\n' && c != '\r' && c != '\t') .map(|chars: &str| StringPiece::Ref(chars))), attempt(string("\\\"").map(|_|StringPiece::Ref("\""))), attempt(string("\\\\").map(|_|StringPiece::Ref("\\"))), attempt(string("\\n").map(|_|StringPiece::Ref("\n"))), attempt(string("\\t").map(|_|StringPiece::Ref("\t"))), attempt(string("\\/").map(|_|StringPiece::Ref("/"))), attempt(string("\\r").map(|_|StringPiece::Ref("\r"))), attempt(string("\\f").map(|_|StringPiece::Ref("\u{000c}"))), attempt(string("\\b").map(|_|StringPiece::Ref("\u{0008}"))), attempt(unicode_char().map(|s|StringPiece::Char(s))), ) ) ) } fn string_parser_inner<'a>() -> impl Parser<&'a str, Output = SmolStr> { c_hx_do! { x <- between(char('"'), char('"'), string_part()); { let cap = x.iter().fold(0, |acc, s| acc + match s { StringPiece::Ref(strref) => strref.len(), StringPiece::Char(c) => c.map(|c_inner| c_inner.len_utf8()).unwrap_or(0) } ); if cap <= 22 { let mut buf: [u8; 22] = [0; 22]; let mut offset = 0; for s in x.iter() { match s { StringPiece::Ref(strref) => { for &b in strref.as_bytes() { buf[offset] = b; offset += 1; } }, StringPiece::Char(c) => { if let Some(chr) = c { chr.encode_utf8(&mut buf[offset..]); offset += chr.len_utf8(); } } } } return unsafe { SmolStr::new(str::from_utf8_unchecked(&buf[0..cap])) }; } let mut str = String::with_capacity(cap); for s in x.iter() { match s { StringPiece::Ref(strref) => str.push_str(strref), StringPiece::Char(c) => if let Some(chr) = c { str.push(*chr); } } } SmolStr::new(str) } } } fn string_parser<'a>() -> impl Parser<&'a str, Output = Node> { string_parser_inner().map(|x| Node::String(x)) } fn digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { take_while1(|c: char| c >= '0' && c <= '9') } #[inline(always)] fn power(lhs: f64, rhs: f64) -> f64 { lhs.powf(rhs) } fn trailing_digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { c_hx_do! { __ <- char('.'), rest <- digit_sequence(); rest } } fn exponent_parser<'a>() -> impl Parser<&'a str, Output = f64> { c_hx_do!{ __ <- satisfy(|c: char| c == 'e' || c == 'E'), sign_char <- optional(satisfy(|c: char| c == '+' || c == '-')), digits <- digit_sequence(); { let sign = match sign_char { Some('-') => -1.0, _ => 1.0 }; let mut acc = 0; for c in digits.as_bytes() { acc = acc * 10 + (c - b'0') as u64; } power(10.0, sign * acc as f64) } } } #[derive(PartialEq, Copy, Clone)] enum NumberPrefix<'a > { LeadingZero, Digits(char, &'a str) } fn leading_zero_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { char('0').map(|_| NumberPrefix::LeadingZero) } fn leading_digits_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { c_hx_do! { leading_digit <- satisfy(|c: char| c >= '1' && c <= '9'), digs <- optional(digit_sequence()); NumberPrefix::Digits(leading_digit, digs.unwrap_or("")) } } fn leading_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { choice(( attempt(leading_digits_parser()), attempt(leading_zero_parser()), )) } fn number_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do! { minus_sign <- optional(char('-')), leading <- leading_parser(), trail <- optional(trailing_digit_sequence()), exp <- optional(exponent_parser()); { Node::Number({ let mut acc = match leading { NumberPrefix::LeadingZero => 0.0, NumberPrefix::Digits(leading_digit, l_digs) => { let mut l = (leading_digit as u8 - b'0') as u64; for c in l_digs.as_bytes() { l = l * 10 + (c - b'0') as u64; } l as f64 } }; if let Some(t_digs) = trail { let mut divider = 1.0; for c in t_digs.as_bytes() { divider /= 10.0; acc += (c - b'0') as f64 * divider; } } if let Some(exponent) = exp { acc *= exponent; } if let Some(_) = minus_sign { -acc } else { acc } }) } } } fn bool_parser<'a>() -> impl Parser<&'a str, Output = Node>
fn null_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do!{ _word <- string("null"); Node::Null } } macro_rules! ref_parser { ($parser_fn:ident) => { parser(|input| { let _: &mut &str = input; $parser_fn().parse_stream(input).into_result() }) } } fn primitive_parser<'a>() -> impl Parser<&'a str, Output = Node> { let possible_parser = bool_parser() .or(number_parser()) .or(string_parser()) .or(null_parser()) .or(ref_parser!(array_parser)) .or(ref_parser!(dictionary_parser)); c_hx_do! { __ <- skip_many(space()), pars <- possible_parser, ___ <- skip_many(space()); pars } } fn array_parser<'a>() -> impl Parser<&'a str, Output = Node> { braced_parser( char('['), sep_by(primitive_parser(), char(',')), char(']') ).map(|nodes: Vec<Node>| Node::Array(nodes) ) } fn pair_parser<'a>() -> impl Parser<&'a str, Output = Option<(SmolStr, Node)>> { let str_parser = c_hx_do!{ __ <- skip_many(space()), stp <- string_parser_inner(), ___ <- skip_many(space()); stp }; c_hx_do!{ l <- str_parser, __ <- char(':'), r <- primitive_parser(); Some((l, r)) } } fn dictionary_parser<'a>() -> impl Parser<&'a str, Output = Node> { braced_parser( char('{'), sep_by(pair_parser(), char(',')), char('}') ).map(|mut pairs: Vec<Option<(SmolStr, Node)>>| { let mut dict = HashMap::with_capacity(pairs.len()); for mut pair in pairs { let (l, r) = mem::replace(&mut pair, None).unwrap(); dict.insert(l, r); } Node::Object( dict ) }) } fn json_parser<'a>() -> impl Parser<&'a str, Output = Node> { null_parser() .or(bool_parser()) .or(number_parser()) .or(string_parser()) .or(array_parser()) .or(dictionary_parser()) } pub fn parse_json(content: &str) -> Result<Node, String> { let mut parser = c_hx_do!{ __ <- skip_many(space()), json <- json_parser(), ___ <- skip_many(space()), ____ <- eof(); json }; let res = parser.parse(content); match res { Err(x) => Err(format!("{}", x.to_string())), Ok((res,_)) => Ok(res) } }
{ c_hx_do!{ word <- string("true").or(string("false")); match word { "true" => Node::Boolean(true), _ => Node::Boolean(false) } } }
identifier_body
lib.rs
#[macro_use] extern crate compre_combinee; extern crate combine; mod errors; mod details; mod traits; mod stop_watch; use std::collections::{HashMap}; use combine::{parser, eof, satisfy, choice, attempt}; use combine::parser::range::{take_while1}; use combine::parser::char::*; use combine::{Parser, many, optional, skip_many, sep_by, between}; pub use crate::errors::ErrorCause; pub use crate::details::Node; pub use crate::traits::*; use std::{f64, mem, str}; use std::convert::TryFrom; use smol_str::SmolStr; fn parse_hex<'a>() -> impl Parser<&'a str, Output = u32> { satisfy(|c: char| (c >= '0' && c <= '9') || (c >= 'a' && c <= 'f') || (c >= 'A' && c <= 'F') ).map(|c| if c >= '0' && c <= '9' { c as u64 - '0' as u64 } else if c >= 'a' && c <= 'f' { 10 + c as u64 - 'a' as u64 } else { 10 + c as u64 - 'A' as u64 } as u32 ) } fn unicode_char<'a>() -> impl Parser<&'a str, Output = Option<char>> { c_hx_do!{ __ <- string(r#"\u"#), d3 <- parse_hex(), d2 <- parse_hex(), d1 <- parse_hex(), d0 <- parse_hex(); { let unicode = d0 + 0x10 * d1 + 0x100 * d2 + 0x1000 * d3; char::try_from(unicode).ok() } } } #[derive(PartialEq)] enum StringPiece<'a > { Ref(&'a str), Char(Option<char>) } fn braced_parser<'a, PBL, P, PBR, O>(pbl: PBL, p: P, pbr: PBR) -> impl Parser<&'a str, Output = O> where PBL: Parser<&'a str>, PBR: Parser<&'a str>, P: Parser<&'a str, Output = O> { between( c_compre![c; c <- pbl, __ <- skip_many(space())], c_compre![c; __ <- skip_many(space()), c <- pbr], p ) } fn string_part<'a>() -> impl Parser<&'a str, Output = Vec<StringPiece<'a >>> { many( choice( ( attempt(take_while1(|c: char| c != '\\' && c != '"' && c != '\n' && c != '\r' && c != '\t') .map(|chars: &str| StringPiece::Ref(chars))), attempt(string("\\\"").map(|_|StringPiece::Ref("\""))), attempt(string("\\\\").map(|_|StringPiece::Ref("\\"))), attempt(string("\\n").map(|_|StringPiece::Ref("\n"))), attempt(string("\\t").map(|_|StringPiece::Ref("\t"))), attempt(string("\\/").map(|_|StringPiece::Ref("/"))), attempt(string("\\r").map(|_|StringPiece::Ref("\r"))), attempt(string("\\f").map(|_|StringPiece::Ref("\u{000c}"))), attempt(string("\\b").map(|_|StringPiece::Ref("\u{0008}"))), attempt(unicode_char().map(|s|StringPiece::Char(s))), ) ) ) } fn string_parser_inner<'a>() -> impl Parser<&'a str, Output = SmolStr> { c_hx_do! { x <- between(char('"'), char('"'), string_part()); { let cap = x.iter().fold(0, |acc, s| acc + match s { StringPiece::Ref(strref) => strref.len(), StringPiece::Char(c) => c.map(|c_inner| c_inner.len_utf8()).unwrap_or(0) } ); if cap <= 22 { let mut buf: [u8; 22] = [0; 22]; let mut offset = 0; for s in x.iter() { match s { StringPiece::Ref(strref) => { for &b in strref.as_bytes() { buf[offset] = b; offset += 1; } }, StringPiece::Char(c) => { if let Some(chr) = c { chr.encode_utf8(&mut buf[offset..]); offset += chr.len_utf8(); } } } } return unsafe { SmolStr::new(str::from_utf8_unchecked(&buf[0..cap])) }; } let mut str = String::with_capacity(cap); for s in x.iter() { match s { StringPiece::Ref(strref) => str.push_str(strref), StringPiece::Char(c) => if let Some(chr) = c { str.push(*chr); } } } SmolStr::new(str) } } } fn string_parser<'a>() -> impl Parser<&'a str, Output = Node> { string_parser_inner().map(|x| Node::String(x)) } fn digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { take_while1(|c: char| c >= '0' && c <= '9') } #[inline(always)] fn power(lhs: f64, rhs: f64) -> f64 { lhs.powf(rhs) } fn trailing_digit_sequence<'a>() -> impl Parser<&'a str, Output = &'a str> { c_hx_do! { __ <- char('.'), rest <- digit_sequence(); rest } } fn exponent_parser<'a>() -> impl Parser<&'a str, Output = f64> { c_hx_do!{ __ <- satisfy(|c: char| c == 'e' || c == 'E'), sign_char <- optional(satisfy(|c: char| c == '+' || c == '-')), digits <- digit_sequence(); { let sign = match sign_char { Some('-') => -1.0, _ => 1.0 }; let mut acc = 0; for c in digits.as_bytes() { acc = acc * 10 + (c - b'0') as u64; } power(10.0, sign * acc as f64) } } } #[derive(PartialEq, Copy, Clone)] enum NumberPrefix<'a > { LeadingZero, Digits(char, &'a str) } fn leading_zero_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { char('0').map(|_| NumberPrefix::LeadingZero) } fn leading_digits_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { c_hx_do! { leading_digit <- satisfy(|c: char| c >= '1' && c <= '9'), digs <- optional(digit_sequence()); NumberPrefix::Digits(leading_digit, digs.unwrap_or("")) } } fn leading_parser <'a>() -> impl Parser<&'a str, Output = NumberPrefix<'a >> { choice(( attempt(leading_digits_parser()), attempt(leading_zero_parser()), )) } fn number_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do! { minus_sign <- optional(char('-')), leading <- leading_parser(), trail <- optional(trailing_digit_sequence()), exp <- optional(exponent_parser()); { Node::Number({ let mut acc = match leading { NumberPrefix::LeadingZero => 0.0, NumberPrefix::Digits(leading_digit, l_digs) => { let mut l = (leading_digit as u8 - b'0') as u64; for c in l_digs.as_bytes() { l = l * 10 + (c - b'0') as u64; } l as f64 } }; if let Some(t_digs) = trail { let mut divider = 1.0; for c in t_digs.as_bytes() { divider /= 10.0; acc += (c - b'0') as f64 * divider; } } if let Some(exponent) = exp { acc *= exponent; } if let Some(_) = minus_sign { -acc } else { acc } }) } } } fn
<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do!{ word <- string("true").or(string("false")); match word { "true" => Node::Boolean(true), _ => Node::Boolean(false) } } } fn null_parser<'a>() -> impl Parser<&'a str, Output = Node> { c_hx_do!{ _word <- string("null"); Node::Null } } macro_rules! ref_parser { ($parser_fn:ident) => { parser(|input| { let _: &mut &str = input; $parser_fn().parse_stream(input).into_result() }) } } fn primitive_parser<'a>() -> impl Parser<&'a str, Output = Node> { let possible_parser = bool_parser() .or(number_parser()) .or(string_parser()) .or(null_parser()) .or(ref_parser!(array_parser)) .or(ref_parser!(dictionary_parser)); c_hx_do! { __ <- skip_many(space()), pars <- possible_parser, ___ <- skip_many(space()); pars } } fn array_parser<'a>() -> impl Parser<&'a str, Output = Node> { braced_parser( char('['), sep_by(primitive_parser(), char(',')), char(']') ).map(|nodes: Vec<Node>| Node::Array(nodes) ) } fn pair_parser<'a>() -> impl Parser<&'a str, Output = Option<(SmolStr, Node)>> { let str_parser = c_hx_do!{ __ <- skip_many(space()), stp <- string_parser_inner(), ___ <- skip_many(space()); stp }; c_hx_do!{ l <- str_parser, __ <- char(':'), r <- primitive_parser(); Some((l, r)) } } fn dictionary_parser<'a>() -> impl Parser<&'a str, Output = Node> { braced_parser( char('{'), sep_by(pair_parser(), char(',')), char('}') ).map(|mut pairs: Vec<Option<(SmolStr, Node)>>| { let mut dict = HashMap::with_capacity(pairs.len()); for mut pair in pairs { let (l, r) = mem::replace(&mut pair, None).unwrap(); dict.insert(l, r); } Node::Object( dict ) }) } fn json_parser<'a>() -> impl Parser<&'a str, Output = Node> { null_parser() .or(bool_parser()) .or(number_parser()) .or(string_parser()) .or(array_parser()) .or(dictionary_parser()) } pub fn parse_json(content: &str) -> Result<Node, String> { let mut parser = c_hx_do!{ __ <- skip_many(space()), json <- json_parser(), ___ <- skip_many(space()), ____ <- eof(); json }; let res = parser.parse(content); match res { Err(x) => Err(format!("{}", x.to_string())), Ok((res,_)) => Ok(res) } }
bool_parser
identifier_name
main.rs
// ~Similar to the ST Heart Rate Sensor example #![no_main] #![no_std] #![allow(non_snake_case)] use panic_rtt_target as _; // use panic_halt as _; use rtt_target::{rprintln, rtt_init_print}; use stm32wb_hal as hal; use core::time::Duration; use cortex_m_rt::{entry, exception}; use nb::block; use byteorder::{ByteOrder, LittleEndian}; use hal::{ flash::FlashExt, prelude::*, rcc::{ ApbDivider, Config, HDivider, HseDivider, PllConfig, PllSrc, RfWakeupClock, RtcClkSrc, StopWakeupClock, SysClkSrc, }, tl_mbox::{lhci::LhciC1DeviceInformationCcrp, shci::ShciBleInitCmdParam, TlMbox}, }; use bluetooth_hci::{ event::{command::ReturnParameters, Event}, host::{uart::Packet, AdvertisingFilterPolicy, EncryptionKey, Hci, OwnAddressType}, BdAddr, }; use ble::{perform_command, receive_event, setup_coprocessor, Characteristic, RadioCopro}; use stm32wb55::{ event::{AttReadPermitRequest, AttributeHandle, GattAttributeModified, Stm32Wb5xEvent}, gap::{ AdvertisingDataType, AdvertisingType, AuthenticationRequirements, Commands as GapCommands, DiscoverableParameters, LocalName, OutOfBandAuthentication, Pin, Role, }, gatt::{CharacteristicProperty, Commands as GattCommads, UpdateCharacteristicValueParameters}, hal::{Commands as HalCommands, ConfigData, PowerLevel}, }; mod ble; mod svc_dis; mod svc_hrs; mod bt_appearances; use svc_dis::{uuid, DeviceInformation, DisCharacteristic, DisService}; use svc_hrs::{HrsService, HrsBodySensorLocation, HrsHrmFlags}; use crate::ble::Service; use crate::svc_hrs::HrsMeasure; /// Advertisement interval in milliseconds. const ADV_INTERVAL_MS: u64 = 250; const BT_NAME: &[u8] = b"KToy"; const BLE_GAP_DEVICE_NAME_LENGTH: u8 = BT_NAME.len() as u8; // const MY_DEVICE_INFO: DeviceInformation = DeviceInformation { // fw_revision: Some("fw1.23"), // manufacturer_name: Some("demo Company"), // model_number: None, // serial_number: None, // system_id: Some("sysid69"), // ieee_cert: None, // hw_revision: None, // sw_revision: None, // pnp_id: None // }; // #[entry] fn entry() -> ! { //rtt_init_print!(BlockIfFull, 4096); rtt_init_print!(NoBlockSkip, 4096); run(); loop { continue; } } fn run() { let dp = hal::device::Peripherals::take().unwrap(); let mut rcc = dp.RCC.constrain(); rcc.set_stop_wakeup_clock(StopWakeupClock::HSI16); // Fastest clock configuration. // * External low-speed crystal is used (LSE) // * 32 MHz HSE with PLL // * 64 MHz CPU1, 32 MHz CPU2 // * 64 MHz for APB1, APB2 // * HSI as a clock source after wake-up from low-power mode let clock_config = Config::new(SysClkSrc::Pll(PllSrc::Hse(HseDivider::NotDivided))) .with_lse() .cpu1_hdiv(HDivider::NotDivided) .cpu2_hdiv(HDivider::Div2) .apb1_div(ApbDivider::NotDivided) .apb2_div(ApbDivider::NotDivided) .pll_cfg(PllConfig { m: 2, n: 12, r: 3, q: Some(4), p: Some(3), }) .rtc_src(RtcClkSrc::Lse) .rf_wkp_sel(RfWakeupClock::Lse); let mut rcc = rcc.apply_clock_config(clock_config, &mut dp.FLASH.constrain().acr); rprintln!("Boot"); // RTC is required for proper operation of BLE stack let _rtc = hal::rtc::Rtc::rtc(dp.RTC, &mut rcc); let mut ipcc = dp.IPCC.constrain(); let mbox = TlMbox::tl_init(&mut rcc, &mut ipcc); let config = ShciBleInitCmdParam { p_ble_buffer_address: 0, ble_buffer_size: 0, num_attr_record: 100,
extended_packet_length_enable: 1, pr_write_list_size: 0x3A, mb_lock_count: 0x79, att_mtu: 312, slave_sca: 500, master_sca: 0, ls_source: 1, max_conn_event_length: 0xFFFFFFFF, hs_startup_time: 0x148, viterbi_enable: 1, ll_only: 0, hw_version: 0, }; setup_coprocessor(config, ipcc, mbox); // enable interrupts -> interrupts are enabled in Ipcc::init(), which is called TlMbox::tl_init // Boot CPU2 hal::pwr::set_cpu2(true); let ready_event = block!(receive_event()); rprintln!("Received packet: {:?}", ready_event); rprintln!("Resetting processor..."); let reset_response = perform_command(|rc| rc.reset()).expect("Failed to reset processor"); rprintln!("Received packet: {:?}", reset_response); init_gap_and_gatt().expect("Failed to initialize GAP and GATT"); rprintln!("Succesfully initialized GAP and GATT"); let di = DeviceInformation::new( Some("klabs"), Some("9871234"), Some("my-serial"), None, None, Some("fw1.234"), Some("my-system-id"), None, None, ); let dis_service = init_dis(&di).expect("failed to activate DIS"); let hrs_service = init_hrs().expect("failed to activate heart rate service"); // Set our discovery parameters, this is "application specific" regardless of what services // we've turned on perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }); loop { let response = block!(receive_event()); rprintln!("Received event: {:x?}", response); if let Ok(Packet::Event(event)) = response { match event { // karl - this isn't quite working... Event::DisconnectionComplete(_state) => { // Enter advertising mode again // Put the device in a connectable mode perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }) .expect("Failed to enable discoverable mode again"); // perform_command(|rc| { // rc.update_advertising_data(&ADVERTISING_DATA[..]) // .map_err(|_| nb::Error::Other(())) // }) // .expect("Failed to update advertising data"); } // FIXME - I want some sort of "list of event handlers" that can be plugged in here? // ST style has a list of handlers, and stops at the first one to say "handled" _ => handle_event(&event, &dis_service, &hrs_service), } } } } fn handle_event(event: &Event<Stm32Wb5xEvent>, dis: &DisService, hrs: &HrsService) { if let Event::Vendor(stm_event) = event { match stm_event { Stm32Wb5xEvent::AttReadPermitRequest(AttReadPermitRequest { conn_handle, attribute_handle, offset, }) => { rprintln!("Allowing read on ch: {:?} ah: {:?}, offset: {}", conn_handle, attribute_handle, offset); perform_command(|rc| rc.allow_read(*conn_handle)) .expect("Failed to allow read"); } other => rprintln!("ignoring event {:?}", other), } } } #[exception] unsafe fn DefaultHandler(irqn: i16) -> ! { panic!("Unhandled IRQ: {}", irqn); } fn init_gap_and_gatt() -> Result<(), ()> { let response = perform_command(|rc: &mut RadioCopro| { rc.write_config_data(&ConfigData::public_address(get_bd_addr()).build()) })?; rprintln!("Response to write_config_data: {:?}", response); perform_command(|rc| { rc.write_config_data(&ConfigData::random_address(get_random_addr()).build()) })?; perform_command(|rc| rc.write_config_data(&ConfigData::identity_root(&get_irk()).build()))?; perform_command(|rc| rc.write_config_data(&ConfigData::encryption_root(&get_erk()).build()))?; perform_command(|rc| rc.set_tx_power_level(PowerLevel::ZerodBm))?; perform_command(|rc| rc.init_gatt())?; let return_params = perform_command(|rc| rc.init_gap(Role::PERIPHERAL, false, BLE_GAP_DEVICE_NAME_LENGTH))?; // let sh, dh, ah == return parameters... if it was of the type of GapInit ReturnParameters....? let (service_handle, dev_name_handle, appearance_handle) = if let ReturnParameters::Vendor( stm32wb55::event::command::ReturnParameters::GapInit(stm32wb55::event::command::GapInit { service_handle, dev_name_handle, appearance_handle, .. }), ) = return_params { (service_handle, dev_name_handle, appearance_handle) } else { rprintln!("Unexpected response to init_gap command"); return Err(()); }; perform_command(|rc| { rc.update_characteristic_value(&UpdateCharacteristicValueParameters { service_handle, characteristic_handle: dev_name_handle, offset: 0, value: BT_NAME, }) .map_err(|_| nb::Error::Other(())) })?; let appearance_characteristic = Characteristic { service: service_handle, characteristic: appearance_handle, max_len: 4, }; appearance_characteristic.set_value(&bt_appearances::HeartRateSensor::GENERIC.0.to_le_bytes()); return Ok(()); } fn init_dis(di: &DeviceInformation) -> Result<DisService, ()> { // homekit demo uses 24, st uses max 19, 2 per char, plus 1. // using less than required here saves memory in the shared space I believe, so, ideally, this would // check how many "real" values are configured in the "DisServiceInfo" blob.... let dis_service = DisService::new(svc_dis::uuid::DEVICE_INFORMATION_SERVICE, 19)?; di.register(&dis_service); // FIXME - neither of these should be in this function, it makes it hard to compose services... // Disable scan response. not even sure we need this at all. // perform_command(|rc: &mut RadioCopro| { // rc.le_set_scan_response_data(&[]) // .map_err(|_| nb::Error::Other(())) // })?; return Ok(dis_service); } // https://stackoverflow.com/questions/28127165/how-to-convert-struct-to-u8 // except we have no std.... // unsafe fn any_as_u8_slice<T: Sized>(p: &T) -> &[u8] { // ::std::slice::from_raw_parts( // (p as *const T) as *const u8, // ::std::mem::size_of::<T>(), // ) // } fn init_hrs() -> Result<HrsService, ()> { // analog to hrs_init let hrs_service = HrsService::new(true, true, true)?; // analog to hrsapp_init... if hrs_service.with_location { let loc = HrsBodySensorLocation::Finger as u8; hrs_service.body_sensor_location.as_ref().unwrap().set_value(&loc.to_le_bytes()); } let mut hrs_measure = HrsMeasure { value: 1, energy_expended: 100, aRR_interval_values: [200], valid_intervals: 1, flags: HrsHrmFlags::VALUE_FORMAT_UINT16 | HrsHrmFlags::SENSOR_CONTACTS_PRESENT | HrsHrmFlags::SENSOR_CONTACTS_SUPPORTED | HrsHrmFlags::ENERGY_EXPENDED_PRESENT | HrsHrmFlags::RR_INTERVAL_PRESENT, }; // TODO We need to keep that hrs_measure around somewhere, and get our task to start processing periodic events for it.... let mut bytes:[u8;8] = [0; 8]; LittleEndian::write_u16(&mut bytes[0..2], hrs_measure.value); //bytes[0..2] = *hrs_measure.value.to_le_bytes(); LittleEndian::write_u16(&mut bytes[2..4], hrs_measure.energy_expended); LittleEndian::write_u16(&mut bytes[4..6], hrs_measure.aRR_interval_values[0]); bytes[6] = hrs_measure.valid_intervals; bytes[7] = hrs_measure.flags.bits(); hrs_service.heart_rate_measurement.set_value(&bytes); return Ok(hrs_service); } fn get_bd_addr() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = lhci_info.device_type_id; bytes[4] = (lhci_info.st_company_id & 0xff) as u8; bytes[5] = (lhci_info.st_company_id >> 8 & 0xff) as u8; BdAddr(bytes) } fn get_random_addr() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = 0; bytes[4] = 0x6E; bytes[5] = 0xED; BdAddr(bytes) } const BLE_CFG_IRK: [u8; 16] = [ 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, ]; const BLE_CFG_ERK: [u8; 16] = [ 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, ]; fn get_irk() -> EncryptionKey { EncryptionKey(BLE_CFG_IRK) } fn get_erk() -> EncryptionKey { EncryptionKey(BLE_CFG_ERK) } const DISCOVERY_PARAMS: DiscoverableParameters = DiscoverableParameters { advertising_type: AdvertisingType::ConnectableUndirected, advertising_interval: Some(( Duration::from_millis(ADV_INTERVAL_MS), Duration::from_millis(ADV_INTERVAL_MS), )), address_type: OwnAddressType::Public, filter_policy: AdvertisingFilterPolicy::AllowConnectionAndScan, // Local name should be empty for the device to be recognized as an Eddystone beacon local_name: Some(LocalName::Complete(BT_NAME)), advertising_data: &[], conn_interval: (None, None), };
num_attr_serv: 10, attr_value_arr_size: 3500, //2788, num_of_links: 8,
random_line_split
main.rs
// ~Similar to the ST Heart Rate Sensor example #![no_main] #![no_std] #![allow(non_snake_case)] use panic_rtt_target as _; // use panic_halt as _; use rtt_target::{rprintln, rtt_init_print}; use stm32wb_hal as hal; use core::time::Duration; use cortex_m_rt::{entry, exception}; use nb::block; use byteorder::{ByteOrder, LittleEndian}; use hal::{ flash::FlashExt, prelude::*, rcc::{ ApbDivider, Config, HDivider, HseDivider, PllConfig, PllSrc, RfWakeupClock, RtcClkSrc, StopWakeupClock, SysClkSrc, }, tl_mbox::{lhci::LhciC1DeviceInformationCcrp, shci::ShciBleInitCmdParam, TlMbox}, }; use bluetooth_hci::{ event::{command::ReturnParameters, Event}, host::{uart::Packet, AdvertisingFilterPolicy, EncryptionKey, Hci, OwnAddressType}, BdAddr, }; use ble::{perform_command, receive_event, setup_coprocessor, Characteristic, RadioCopro}; use stm32wb55::{ event::{AttReadPermitRequest, AttributeHandle, GattAttributeModified, Stm32Wb5xEvent}, gap::{ AdvertisingDataType, AdvertisingType, AuthenticationRequirements, Commands as GapCommands, DiscoverableParameters, LocalName, OutOfBandAuthentication, Pin, Role, }, gatt::{CharacteristicProperty, Commands as GattCommads, UpdateCharacteristicValueParameters}, hal::{Commands as HalCommands, ConfigData, PowerLevel}, }; mod ble; mod svc_dis; mod svc_hrs; mod bt_appearances; use svc_dis::{uuid, DeviceInformation, DisCharacteristic, DisService}; use svc_hrs::{HrsService, HrsBodySensorLocation, HrsHrmFlags}; use crate::ble::Service; use crate::svc_hrs::HrsMeasure; /// Advertisement interval in milliseconds. const ADV_INTERVAL_MS: u64 = 250; const BT_NAME: &[u8] = b"KToy"; const BLE_GAP_DEVICE_NAME_LENGTH: u8 = BT_NAME.len() as u8; // const MY_DEVICE_INFO: DeviceInformation = DeviceInformation { // fw_revision: Some("fw1.23"), // manufacturer_name: Some("demo Company"), // model_number: None, // serial_number: None, // system_id: Some("sysid69"), // ieee_cert: None, // hw_revision: None, // sw_revision: None, // pnp_id: None // }; // #[entry] fn entry() -> ! { //rtt_init_print!(BlockIfFull, 4096); rtt_init_print!(NoBlockSkip, 4096); run(); loop { continue; } } fn run() { let dp = hal::device::Peripherals::take().unwrap(); let mut rcc = dp.RCC.constrain(); rcc.set_stop_wakeup_clock(StopWakeupClock::HSI16); // Fastest clock configuration. // * External low-speed crystal is used (LSE) // * 32 MHz HSE with PLL // * 64 MHz CPU1, 32 MHz CPU2 // * 64 MHz for APB1, APB2 // * HSI as a clock source after wake-up from low-power mode let clock_config = Config::new(SysClkSrc::Pll(PllSrc::Hse(HseDivider::NotDivided))) .with_lse() .cpu1_hdiv(HDivider::NotDivided) .cpu2_hdiv(HDivider::Div2) .apb1_div(ApbDivider::NotDivided) .apb2_div(ApbDivider::NotDivided) .pll_cfg(PllConfig { m: 2, n: 12, r: 3, q: Some(4), p: Some(3), }) .rtc_src(RtcClkSrc::Lse) .rf_wkp_sel(RfWakeupClock::Lse); let mut rcc = rcc.apply_clock_config(clock_config, &mut dp.FLASH.constrain().acr); rprintln!("Boot"); // RTC is required for proper operation of BLE stack let _rtc = hal::rtc::Rtc::rtc(dp.RTC, &mut rcc); let mut ipcc = dp.IPCC.constrain(); let mbox = TlMbox::tl_init(&mut rcc, &mut ipcc); let config = ShciBleInitCmdParam { p_ble_buffer_address: 0, ble_buffer_size: 0, num_attr_record: 100, num_attr_serv: 10, attr_value_arr_size: 3500, //2788, num_of_links: 8, extended_packet_length_enable: 1, pr_write_list_size: 0x3A, mb_lock_count: 0x79, att_mtu: 312, slave_sca: 500, master_sca: 0, ls_source: 1, max_conn_event_length: 0xFFFFFFFF, hs_startup_time: 0x148, viterbi_enable: 1, ll_only: 0, hw_version: 0, }; setup_coprocessor(config, ipcc, mbox); // enable interrupts -> interrupts are enabled in Ipcc::init(), which is called TlMbox::tl_init // Boot CPU2 hal::pwr::set_cpu2(true); let ready_event = block!(receive_event()); rprintln!("Received packet: {:?}", ready_event); rprintln!("Resetting processor..."); let reset_response = perform_command(|rc| rc.reset()).expect("Failed to reset processor"); rprintln!("Received packet: {:?}", reset_response); init_gap_and_gatt().expect("Failed to initialize GAP and GATT"); rprintln!("Succesfully initialized GAP and GATT"); let di = DeviceInformation::new( Some("klabs"), Some("9871234"), Some("my-serial"), None, None, Some("fw1.234"), Some("my-system-id"), None, None, ); let dis_service = init_dis(&di).expect("failed to activate DIS"); let hrs_service = init_hrs().expect("failed to activate heart rate service"); // Set our discovery parameters, this is "application specific" regardless of what services // we've turned on perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }); loop { let response = block!(receive_event()); rprintln!("Received event: {:x?}", response); if let Ok(Packet::Event(event)) = response { match event { // karl - this isn't quite working... Event::DisconnectionComplete(_state) => { // Enter advertising mode again // Put the device in a connectable mode perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }) .expect("Failed to enable discoverable mode again"); // perform_command(|rc| { // rc.update_advertising_data(&ADVERTISING_DATA[..]) // .map_err(|_| nb::Error::Other(())) // }) // .expect("Failed to update advertising data"); } // FIXME - I want some sort of "list of event handlers" that can be plugged in here? // ST style has a list of handlers, and stops at the first one to say "handled" _ => handle_event(&event, &dis_service, &hrs_service), } } } } fn handle_event(event: &Event<Stm32Wb5xEvent>, dis: &DisService, hrs: &HrsService) { if let Event::Vendor(stm_event) = event { match stm_event { Stm32Wb5xEvent::AttReadPermitRequest(AttReadPermitRequest { conn_handle, attribute_handle, offset, }) => { rprintln!("Allowing read on ch: {:?} ah: {:?}, offset: {}", conn_handle, attribute_handle, offset); perform_command(|rc| rc.allow_read(*conn_handle)) .expect("Failed to allow read"); } other => rprintln!("ignoring event {:?}", other), } } } #[exception] unsafe fn DefaultHandler(irqn: i16) -> ! { panic!("Unhandled IRQ: {}", irqn); } fn init_gap_and_gatt() -> Result<(), ()> { let response = perform_command(|rc: &mut RadioCopro| { rc.write_config_data(&ConfigData::public_address(get_bd_addr()).build()) })?; rprintln!("Response to write_config_data: {:?}", response); perform_command(|rc| { rc.write_config_data(&ConfigData::random_address(get_random_addr()).build()) })?; perform_command(|rc| rc.write_config_data(&ConfigData::identity_root(&get_irk()).build()))?; perform_command(|rc| rc.write_config_data(&ConfigData::encryption_root(&get_erk()).build()))?; perform_command(|rc| rc.set_tx_power_level(PowerLevel::ZerodBm))?; perform_command(|rc| rc.init_gatt())?; let return_params = perform_command(|rc| rc.init_gap(Role::PERIPHERAL, false, BLE_GAP_DEVICE_NAME_LENGTH))?; // let sh, dh, ah == return parameters... if it was of the type of GapInit ReturnParameters....? let (service_handle, dev_name_handle, appearance_handle) = if let ReturnParameters::Vendor( stm32wb55::event::command::ReturnParameters::GapInit(stm32wb55::event::command::GapInit { service_handle, dev_name_handle, appearance_handle, .. }), ) = return_params
else { rprintln!("Unexpected response to init_gap command"); return Err(()); }; perform_command(|rc| { rc.update_characteristic_value(&UpdateCharacteristicValueParameters { service_handle, characteristic_handle: dev_name_handle, offset: 0, value: BT_NAME, }) .map_err(|_| nb::Error::Other(())) })?; let appearance_characteristic = Characteristic { service: service_handle, characteristic: appearance_handle, max_len: 4, }; appearance_characteristic.set_value(&bt_appearances::HeartRateSensor::GENERIC.0.to_le_bytes()); return Ok(()); } fn init_dis(di: &DeviceInformation) -> Result<DisService, ()> { // homekit demo uses 24, st uses max 19, 2 per char, plus 1. // using less than required here saves memory in the shared space I believe, so, ideally, this would // check how many "real" values are configured in the "DisServiceInfo" blob.... let dis_service = DisService::new(svc_dis::uuid::DEVICE_INFORMATION_SERVICE, 19)?; di.register(&dis_service); // FIXME - neither of these should be in this function, it makes it hard to compose services... // Disable scan response. not even sure we need this at all. // perform_command(|rc: &mut RadioCopro| { // rc.le_set_scan_response_data(&[]) // .map_err(|_| nb::Error::Other(())) // })?; return Ok(dis_service); } // https://stackoverflow.com/questions/28127165/how-to-convert-struct-to-u8 // except we have no std.... // unsafe fn any_as_u8_slice<T: Sized>(p: &T) -> &[u8] { // ::std::slice::from_raw_parts( // (p as *const T) as *const u8, // ::std::mem::size_of::<T>(), // ) // } fn init_hrs() -> Result<HrsService, ()> { // analog to hrs_init let hrs_service = HrsService::new(true, true, true)?; // analog to hrsapp_init... if hrs_service.with_location { let loc = HrsBodySensorLocation::Finger as u8; hrs_service.body_sensor_location.as_ref().unwrap().set_value(&loc.to_le_bytes()); } let mut hrs_measure = HrsMeasure { value: 1, energy_expended: 100, aRR_interval_values: [200], valid_intervals: 1, flags: HrsHrmFlags::VALUE_FORMAT_UINT16 | HrsHrmFlags::SENSOR_CONTACTS_PRESENT | HrsHrmFlags::SENSOR_CONTACTS_SUPPORTED | HrsHrmFlags::ENERGY_EXPENDED_PRESENT | HrsHrmFlags::RR_INTERVAL_PRESENT, }; // TODO We need to keep that hrs_measure around somewhere, and get our task to start processing periodic events for it.... let mut bytes:[u8;8] = [0; 8]; LittleEndian::write_u16(&mut bytes[0..2], hrs_measure.value); //bytes[0..2] = *hrs_measure.value.to_le_bytes(); LittleEndian::write_u16(&mut bytes[2..4], hrs_measure.energy_expended); LittleEndian::write_u16(&mut bytes[4..6], hrs_measure.aRR_interval_values[0]); bytes[6] = hrs_measure.valid_intervals; bytes[7] = hrs_measure.flags.bits(); hrs_service.heart_rate_measurement.set_value(&bytes); return Ok(hrs_service); } fn get_bd_addr() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = lhci_info.device_type_id; bytes[4] = (lhci_info.st_company_id & 0xff) as u8; bytes[5] = (lhci_info.st_company_id >> 8 & 0xff) as u8; BdAddr(bytes) } fn get_random_addr() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = 0; bytes[4] = 0x6E; bytes[5] = 0xED; BdAddr(bytes) } const BLE_CFG_IRK: [u8; 16] = [ 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, ]; const BLE_CFG_ERK: [u8; 16] = [ 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, ]; fn get_irk() -> EncryptionKey { EncryptionKey(BLE_CFG_IRK) } fn get_erk() -> EncryptionKey { EncryptionKey(BLE_CFG_ERK) } const DISCOVERY_PARAMS: DiscoverableParameters = DiscoverableParameters { advertising_type: AdvertisingType::ConnectableUndirected, advertising_interval: Some(( Duration::from_millis(ADV_INTERVAL_MS), Duration::from_millis(ADV_INTERVAL_MS), )), address_type: OwnAddressType::Public, filter_policy: AdvertisingFilterPolicy::AllowConnectionAndScan, // Local name should be empty for the device to be recognized as an Eddystone beacon local_name: Some(LocalName::Complete(BT_NAME)), advertising_data: &[], conn_interval: (None, None), };
{ (service_handle, dev_name_handle, appearance_handle) }
conditional_block
main.rs
// ~Similar to the ST Heart Rate Sensor example #![no_main] #![no_std] #![allow(non_snake_case)] use panic_rtt_target as _; // use panic_halt as _; use rtt_target::{rprintln, rtt_init_print}; use stm32wb_hal as hal; use core::time::Duration; use cortex_m_rt::{entry, exception}; use nb::block; use byteorder::{ByteOrder, LittleEndian}; use hal::{ flash::FlashExt, prelude::*, rcc::{ ApbDivider, Config, HDivider, HseDivider, PllConfig, PllSrc, RfWakeupClock, RtcClkSrc, StopWakeupClock, SysClkSrc, }, tl_mbox::{lhci::LhciC1DeviceInformationCcrp, shci::ShciBleInitCmdParam, TlMbox}, }; use bluetooth_hci::{ event::{command::ReturnParameters, Event}, host::{uart::Packet, AdvertisingFilterPolicy, EncryptionKey, Hci, OwnAddressType}, BdAddr, }; use ble::{perform_command, receive_event, setup_coprocessor, Characteristic, RadioCopro}; use stm32wb55::{ event::{AttReadPermitRequest, AttributeHandle, GattAttributeModified, Stm32Wb5xEvent}, gap::{ AdvertisingDataType, AdvertisingType, AuthenticationRequirements, Commands as GapCommands, DiscoverableParameters, LocalName, OutOfBandAuthentication, Pin, Role, }, gatt::{CharacteristicProperty, Commands as GattCommads, UpdateCharacteristicValueParameters}, hal::{Commands as HalCommands, ConfigData, PowerLevel}, }; mod ble; mod svc_dis; mod svc_hrs; mod bt_appearances; use svc_dis::{uuid, DeviceInformation, DisCharacteristic, DisService}; use svc_hrs::{HrsService, HrsBodySensorLocation, HrsHrmFlags}; use crate::ble::Service; use crate::svc_hrs::HrsMeasure; /// Advertisement interval in milliseconds. const ADV_INTERVAL_MS: u64 = 250; const BT_NAME: &[u8] = b"KToy"; const BLE_GAP_DEVICE_NAME_LENGTH: u8 = BT_NAME.len() as u8; // const MY_DEVICE_INFO: DeviceInformation = DeviceInformation { // fw_revision: Some("fw1.23"), // manufacturer_name: Some("demo Company"), // model_number: None, // serial_number: None, // system_id: Some("sysid69"), // ieee_cert: None, // hw_revision: None, // sw_revision: None, // pnp_id: None // }; // #[entry] fn entry() -> ! { //rtt_init_print!(BlockIfFull, 4096); rtt_init_print!(NoBlockSkip, 4096); run(); loop { continue; } } fn run() { let dp = hal::device::Peripherals::take().unwrap(); let mut rcc = dp.RCC.constrain(); rcc.set_stop_wakeup_clock(StopWakeupClock::HSI16); // Fastest clock configuration. // * External low-speed crystal is used (LSE) // * 32 MHz HSE with PLL // * 64 MHz CPU1, 32 MHz CPU2 // * 64 MHz for APB1, APB2 // * HSI as a clock source after wake-up from low-power mode let clock_config = Config::new(SysClkSrc::Pll(PllSrc::Hse(HseDivider::NotDivided))) .with_lse() .cpu1_hdiv(HDivider::NotDivided) .cpu2_hdiv(HDivider::Div2) .apb1_div(ApbDivider::NotDivided) .apb2_div(ApbDivider::NotDivided) .pll_cfg(PllConfig { m: 2, n: 12, r: 3, q: Some(4), p: Some(3), }) .rtc_src(RtcClkSrc::Lse) .rf_wkp_sel(RfWakeupClock::Lse); let mut rcc = rcc.apply_clock_config(clock_config, &mut dp.FLASH.constrain().acr); rprintln!("Boot"); // RTC is required for proper operation of BLE stack let _rtc = hal::rtc::Rtc::rtc(dp.RTC, &mut rcc); let mut ipcc = dp.IPCC.constrain(); let mbox = TlMbox::tl_init(&mut rcc, &mut ipcc); let config = ShciBleInitCmdParam { p_ble_buffer_address: 0, ble_buffer_size: 0, num_attr_record: 100, num_attr_serv: 10, attr_value_arr_size: 3500, //2788, num_of_links: 8, extended_packet_length_enable: 1, pr_write_list_size: 0x3A, mb_lock_count: 0x79, att_mtu: 312, slave_sca: 500, master_sca: 0, ls_source: 1, max_conn_event_length: 0xFFFFFFFF, hs_startup_time: 0x148, viterbi_enable: 1, ll_only: 0, hw_version: 0, }; setup_coprocessor(config, ipcc, mbox); // enable interrupts -> interrupts are enabled in Ipcc::init(), which is called TlMbox::tl_init // Boot CPU2 hal::pwr::set_cpu2(true); let ready_event = block!(receive_event()); rprintln!("Received packet: {:?}", ready_event); rprintln!("Resetting processor..."); let reset_response = perform_command(|rc| rc.reset()).expect("Failed to reset processor"); rprintln!("Received packet: {:?}", reset_response); init_gap_and_gatt().expect("Failed to initialize GAP and GATT"); rprintln!("Succesfully initialized GAP and GATT"); let di = DeviceInformation::new( Some("klabs"), Some("9871234"), Some("my-serial"), None, None, Some("fw1.234"), Some("my-system-id"), None, None, ); let dis_service = init_dis(&di).expect("failed to activate DIS"); let hrs_service = init_hrs().expect("failed to activate heart rate service"); // Set our discovery parameters, this is "application specific" regardless of what services // we've turned on perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }); loop { let response = block!(receive_event()); rprintln!("Received event: {:x?}", response); if let Ok(Packet::Event(event)) = response { match event { // karl - this isn't quite working... Event::DisconnectionComplete(_state) => { // Enter advertising mode again // Put the device in a connectable mode perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }) .expect("Failed to enable discoverable mode again"); // perform_command(|rc| { // rc.update_advertising_data(&ADVERTISING_DATA[..]) // .map_err(|_| nb::Error::Other(())) // }) // .expect("Failed to update advertising data"); } // FIXME - I want some sort of "list of event handlers" that can be plugged in here? // ST style has a list of handlers, and stops at the first one to say "handled" _ => handle_event(&event, &dis_service, &hrs_service), } } } } fn handle_event(event: &Event<Stm32Wb5xEvent>, dis: &DisService, hrs: &HrsService) { if let Event::Vendor(stm_event) = event { match stm_event { Stm32Wb5xEvent::AttReadPermitRequest(AttReadPermitRequest { conn_handle, attribute_handle, offset, }) => { rprintln!("Allowing read on ch: {:?} ah: {:?}, offset: {}", conn_handle, attribute_handle, offset); perform_command(|rc| rc.allow_read(*conn_handle)) .expect("Failed to allow read"); } other => rprintln!("ignoring event {:?}", other), } } } #[exception] unsafe fn DefaultHandler(irqn: i16) -> ! { panic!("Unhandled IRQ: {}", irqn); } fn init_gap_and_gatt() -> Result<(), ()> { let response = perform_command(|rc: &mut RadioCopro| { rc.write_config_data(&ConfigData::public_address(get_bd_addr()).build()) })?; rprintln!("Response to write_config_data: {:?}", response); perform_command(|rc| { rc.write_config_data(&ConfigData::random_address(get_random_addr()).build()) })?; perform_command(|rc| rc.write_config_data(&ConfigData::identity_root(&get_irk()).build()))?; perform_command(|rc| rc.write_config_data(&ConfigData::encryption_root(&get_erk()).build()))?; perform_command(|rc| rc.set_tx_power_level(PowerLevel::ZerodBm))?; perform_command(|rc| rc.init_gatt())?; let return_params = perform_command(|rc| rc.init_gap(Role::PERIPHERAL, false, BLE_GAP_DEVICE_NAME_LENGTH))?; // let sh, dh, ah == return parameters... if it was of the type of GapInit ReturnParameters....? let (service_handle, dev_name_handle, appearance_handle) = if let ReturnParameters::Vendor( stm32wb55::event::command::ReturnParameters::GapInit(stm32wb55::event::command::GapInit { service_handle, dev_name_handle, appearance_handle, .. }), ) = return_params { (service_handle, dev_name_handle, appearance_handle) } else { rprintln!("Unexpected response to init_gap command"); return Err(()); }; perform_command(|rc| { rc.update_characteristic_value(&UpdateCharacteristicValueParameters { service_handle, characteristic_handle: dev_name_handle, offset: 0, value: BT_NAME, }) .map_err(|_| nb::Error::Other(())) })?; let appearance_characteristic = Characteristic { service: service_handle, characteristic: appearance_handle, max_len: 4, }; appearance_characteristic.set_value(&bt_appearances::HeartRateSensor::GENERIC.0.to_le_bytes()); return Ok(()); } fn init_dis(di: &DeviceInformation) -> Result<DisService, ()> { // homekit demo uses 24, st uses max 19, 2 per char, plus 1. // using less than required here saves memory in the shared space I believe, so, ideally, this would // check how many "real" values are configured in the "DisServiceInfo" blob.... let dis_service = DisService::new(svc_dis::uuid::DEVICE_INFORMATION_SERVICE, 19)?; di.register(&dis_service); // FIXME - neither of these should be in this function, it makes it hard to compose services... // Disable scan response. not even sure we need this at all. // perform_command(|rc: &mut RadioCopro| { // rc.le_set_scan_response_data(&[]) // .map_err(|_| nb::Error::Other(())) // })?; return Ok(dis_service); } // https://stackoverflow.com/questions/28127165/how-to-convert-struct-to-u8 // except we have no std.... // unsafe fn any_as_u8_slice<T: Sized>(p: &T) -> &[u8] { // ::std::slice::from_raw_parts( // (p as *const T) as *const u8, // ::std::mem::size_of::<T>(), // ) // } fn init_hrs() -> Result<HrsService, ()> { // analog to hrs_init let hrs_service = HrsService::new(true, true, true)?; // analog to hrsapp_init... if hrs_service.with_location { let loc = HrsBodySensorLocation::Finger as u8; hrs_service.body_sensor_location.as_ref().unwrap().set_value(&loc.to_le_bytes()); } let mut hrs_measure = HrsMeasure { value: 1, energy_expended: 100, aRR_interval_values: [200], valid_intervals: 1, flags: HrsHrmFlags::VALUE_FORMAT_UINT16 | HrsHrmFlags::SENSOR_CONTACTS_PRESENT | HrsHrmFlags::SENSOR_CONTACTS_SUPPORTED | HrsHrmFlags::ENERGY_EXPENDED_PRESENT | HrsHrmFlags::RR_INTERVAL_PRESENT, }; // TODO We need to keep that hrs_measure around somewhere, and get our task to start processing periodic events for it.... let mut bytes:[u8;8] = [0; 8]; LittleEndian::write_u16(&mut bytes[0..2], hrs_measure.value); //bytes[0..2] = *hrs_measure.value.to_le_bytes(); LittleEndian::write_u16(&mut bytes[2..4], hrs_measure.energy_expended); LittleEndian::write_u16(&mut bytes[4..6], hrs_measure.aRR_interval_values[0]); bytes[6] = hrs_measure.valid_intervals; bytes[7] = hrs_measure.flags.bits(); hrs_service.heart_rate_measurement.set_value(&bytes); return Ok(hrs_service); } fn
() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = lhci_info.device_type_id; bytes[4] = (lhci_info.st_company_id & 0xff) as u8; bytes[5] = (lhci_info.st_company_id >> 8 & 0xff) as u8; BdAddr(bytes) } fn get_random_addr() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = 0; bytes[4] = 0x6E; bytes[5] = 0xED; BdAddr(bytes) } const BLE_CFG_IRK: [u8; 16] = [ 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, ]; const BLE_CFG_ERK: [u8; 16] = [ 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, ]; fn get_irk() -> EncryptionKey { EncryptionKey(BLE_CFG_IRK) } fn get_erk() -> EncryptionKey { EncryptionKey(BLE_CFG_ERK) } const DISCOVERY_PARAMS: DiscoverableParameters = DiscoverableParameters { advertising_type: AdvertisingType::ConnectableUndirected, advertising_interval: Some(( Duration::from_millis(ADV_INTERVAL_MS), Duration::from_millis(ADV_INTERVAL_MS), )), address_type: OwnAddressType::Public, filter_policy: AdvertisingFilterPolicy::AllowConnectionAndScan, // Local name should be empty for the device to be recognized as an Eddystone beacon local_name: Some(LocalName::Complete(BT_NAME)), advertising_data: &[], conn_interval: (None, None), };
get_bd_addr
identifier_name
main.rs
// ~Similar to the ST Heart Rate Sensor example #![no_main] #![no_std] #![allow(non_snake_case)] use panic_rtt_target as _; // use panic_halt as _; use rtt_target::{rprintln, rtt_init_print}; use stm32wb_hal as hal; use core::time::Duration; use cortex_m_rt::{entry, exception}; use nb::block; use byteorder::{ByteOrder, LittleEndian}; use hal::{ flash::FlashExt, prelude::*, rcc::{ ApbDivider, Config, HDivider, HseDivider, PllConfig, PllSrc, RfWakeupClock, RtcClkSrc, StopWakeupClock, SysClkSrc, }, tl_mbox::{lhci::LhciC1DeviceInformationCcrp, shci::ShciBleInitCmdParam, TlMbox}, }; use bluetooth_hci::{ event::{command::ReturnParameters, Event}, host::{uart::Packet, AdvertisingFilterPolicy, EncryptionKey, Hci, OwnAddressType}, BdAddr, }; use ble::{perform_command, receive_event, setup_coprocessor, Characteristic, RadioCopro}; use stm32wb55::{ event::{AttReadPermitRequest, AttributeHandle, GattAttributeModified, Stm32Wb5xEvent}, gap::{ AdvertisingDataType, AdvertisingType, AuthenticationRequirements, Commands as GapCommands, DiscoverableParameters, LocalName, OutOfBandAuthentication, Pin, Role, }, gatt::{CharacteristicProperty, Commands as GattCommads, UpdateCharacteristicValueParameters}, hal::{Commands as HalCommands, ConfigData, PowerLevel}, }; mod ble; mod svc_dis; mod svc_hrs; mod bt_appearances; use svc_dis::{uuid, DeviceInformation, DisCharacteristic, DisService}; use svc_hrs::{HrsService, HrsBodySensorLocation, HrsHrmFlags}; use crate::ble::Service; use crate::svc_hrs::HrsMeasure; /// Advertisement interval in milliseconds. const ADV_INTERVAL_MS: u64 = 250; const BT_NAME: &[u8] = b"KToy"; const BLE_GAP_DEVICE_NAME_LENGTH: u8 = BT_NAME.len() as u8; // const MY_DEVICE_INFO: DeviceInformation = DeviceInformation { // fw_revision: Some("fw1.23"), // manufacturer_name: Some("demo Company"), // model_number: None, // serial_number: None, // system_id: Some("sysid69"), // ieee_cert: None, // hw_revision: None, // sw_revision: None, // pnp_id: None // }; // #[entry] fn entry() -> ! { //rtt_init_print!(BlockIfFull, 4096); rtt_init_print!(NoBlockSkip, 4096); run(); loop { continue; } } fn run() { let dp = hal::device::Peripherals::take().unwrap(); let mut rcc = dp.RCC.constrain(); rcc.set_stop_wakeup_clock(StopWakeupClock::HSI16); // Fastest clock configuration. // * External low-speed crystal is used (LSE) // * 32 MHz HSE with PLL // * 64 MHz CPU1, 32 MHz CPU2 // * 64 MHz for APB1, APB2 // * HSI as a clock source after wake-up from low-power mode let clock_config = Config::new(SysClkSrc::Pll(PllSrc::Hse(HseDivider::NotDivided))) .with_lse() .cpu1_hdiv(HDivider::NotDivided) .cpu2_hdiv(HDivider::Div2) .apb1_div(ApbDivider::NotDivided) .apb2_div(ApbDivider::NotDivided) .pll_cfg(PllConfig { m: 2, n: 12, r: 3, q: Some(4), p: Some(3), }) .rtc_src(RtcClkSrc::Lse) .rf_wkp_sel(RfWakeupClock::Lse); let mut rcc = rcc.apply_clock_config(clock_config, &mut dp.FLASH.constrain().acr); rprintln!("Boot"); // RTC is required for proper operation of BLE stack let _rtc = hal::rtc::Rtc::rtc(dp.RTC, &mut rcc); let mut ipcc = dp.IPCC.constrain(); let mbox = TlMbox::tl_init(&mut rcc, &mut ipcc); let config = ShciBleInitCmdParam { p_ble_buffer_address: 0, ble_buffer_size: 0, num_attr_record: 100, num_attr_serv: 10, attr_value_arr_size: 3500, //2788, num_of_links: 8, extended_packet_length_enable: 1, pr_write_list_size: 0x3A, mb_lock_count: 0x79, att_mtu: 312, slave_sca: 500, master_sca: 0, ls_source: 1, max_conn_event_length: 0xFFFFFFFF, hs_startup_time: 0x148, viterbi_enable: 1, ll_only: 0, hw_version: 0, }; setup_coprocessor(config, ipcc, mbox); // enable interrupts -> interrupts are enabled in Ipcc::init(), which is called TlMbox::tl_init // Boot CPU2 hal::pwr::set_cpu2(true); let ready_event = block!(receive_event()); rprintln!("Received packet: {:?}", ready_event); rprintln!("Resetting processor..."); let reset_response = perform_command(|rc| rc.reset()).expect("Failed to reset processor"); rprintln!("Received packet: {:?}", reset_response); init_gap_and_gatt().expect("Failed to initialize GAP and GATT"); rprintln!("Succesfully initialized GAP and GATT"); let di = DeviceInformation::new( Some("klabs"), Some("9871234"), Some("my-serial"), None, None, Some("fw1.234"), Some("my-system-id"), None, None, ); let dis_service = init_dis(&di).expect("failed to activate DIS"); let hrs_service = init_hrs().expect("failed to activate heart rate service"); // Set our discovery parameters, this is "application specific" regardless of what services // we've turned on perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }); loop { let response = block!(receive_event()); rprintln!("Received event: {:x?}", response); if let Ok(Packet::Event(event)) = response { match event { // karl - this isn't quite working... Event::DisconnectionComplete(_state) => { // Enter advertising mode again // Put the device in a connectable mode perform_command(|rc| { rc.set_discoverable(&DISCOVERY_PARAMS) .map_err(|_| nb::Error::Other(())) }) .expect("Failed to enable discoverable mode again"); // perform_command(|rc| { // rc.update_advertising_data(&ADVERTISING_DATA[..]) // .map_err(|_| nb::Error::Other(())) // }) // .expect("Failed to update advertising data"); } // FIXME - I want some sort of "list of event handlers" that can be plugged in here? // ST style has a list of handlers, and stops at the first one to say "handled" _ => handle_event(&event, &dis_service, &hrs_service), } } } } fn handle_event(event: &Event<Stm32Wb5xEvent>, dis: &DisService, hrs: &HrsService) { if let Event::Vendor(stm_event) = event { match stm_event { Stm32Wb5xEvent::AttReadPermitRequest(AttReadPermitRequest { conn_handle, attribute_handle, offset, }) => { rprintln!("Allowing read on ch: {:?} ah: {:?}, offset: {}", conn_handle, attribute_handle, offset); perform_command(|rc| rc.allow_read(*conn_handle)) .expect("Failed to allow read"); } other => rprintln!("ignoring event {:?}", other), } } } #[exception] unsafe fn DefaultHandler(irqn: i16) -> ! { panic!("Unhandled IRQ: {}", irqn); } fn init_gap_and_gatt() -> Result<(), ()> { let response = perform_command(|rc: &mut RadioCopro| { rc.write_config_data(&ConfigData::public_address(get_bd_addr()).build()) })?; rprintln!("Response to write_config_data: {:?}", response); perform_command(|rc| { rc.write_config_data(&ConfigData::random_address(get_random_addr()).build()) })?; perform_command(|rc| rc.write_config_data(&ConfigData::identity_root(&get_irk()).build()))?; perform_command(|rc| rc.write_config_data(&ConfigData::encryption_root(&get_erk()).build()))?; perform_command(|rc| rc.set_tx_power_level(PowerLevel::ZerodBm))?; perform_command(|rc| rc.init_gatt())?; let return_params = perform_command(|rc| rc.init_gap(Role::PERIPHERAL, false, BLE_GAP_DEVICE_NAME_LENGTH))?; // let sh, dh, ah == return parameters... if it was of the type of GapInit ReturnParameters....? let (service_handle, dev_name_handle, appearance_handle) = if let ReturnParameters::Vendor( stm32wb55::event::command::ReturnParameters::GapInit(stm32wb55::event::command::GapInit { service_handle, dev_name_handle, appearance_handle, .. }), ) = return_params { (service_handle, dev_name_handle, appearance_handle) } else { rprintln!("Unexpected response to init_gap command"); return Err(()); }; perform_command(|rc| { rc.update_characteristic_value(&UpdateCharacteristicValueParameters { service_handle, characteristic_handle: dev_name_handle, offset: 0, value: BT_NAME, }) .map_err(|_| nb::Error::Other(())) })?; let appearance_characteristic = Characteristic { service: service_handle, characteristic: appearance_handle, max_len: 4, }; appearance_characteristic.set_value(&bt_appearances::HeartRateSensor::GENERIC.0.to_le_bytes()); return Ok(()); } fn init_dis(di: &DeviceInformation) -> Result<DisService, ()> { // homekit demo uses 24, st uses max 19, 2 per char, plus 1. // using less than required here saves memory in the shared space I believe, so, ideally, this would // check how many "real" values are configured in the "DisServiceInfo" blob.... let dis_service = DisService::new(svc_dis::uuid::DEVICE_INFORMATION_SERVICE, 19)?; di.register(&dis_service); // FIXME - neither of these should be in this function, it makes it hard to compose services... // Disable scan response. not even sure we need this at all. // perform_command(|rc: &mut RadioCopro| { // rc.le_set_scan_response_data(&[]) // .map_err(|_| nb::Error::Other(())) // })?; return Ok(dis_service); } // https://stackoverflow.com/questions/28127165/how-to-convert-struct-to-u8 // except we have no std.... // unsafe fn any_as_u8_slice<T: Sized>(p: &T) -> &[u8] { // ::std::slice::from_raw_parts( // (p as *const T) as *const u8, // ::std::mem::size_of::<T>(), // ) // } fn init_hrs() -> Result<HrsService, ()> { // analog to hrs_init let hrs_service = HrsService::new(true, true, true)?; // analog to hrsapp_init... if hrs_service.with_location { let loc = HrsBodySensorLocation::Finger as u8; hrs_service.body_sensor_location.as_ref().unwrap().set_value(&loc.to_le_bytes()); } let mut hrs_measure = HrsMeasure { value: 1, energy_expended: 100, aRR_interval_values: [200], valid_intervals: 1, flags: HrsHrmFlags::VALUE_FORMAT_UINT16 | HrsHrmFlags::SENSOR_CONTACTS_PRESENT | HrsHrmFlags::SENSOR_CONTACTS_SUPPORTED | HrsHrmFlags::ENERGY_EXPENDED_PRESENT | HrsHrmFlags::RR_INTERVAL_PRESENT, }; // TODO We need to keep that hrs_measure around somewhere, and get our task to start processing periodic events for it.... let mut bytes:[u8;8] = [0; 8]; LittleEndian::write_u16(&mut bytes[0..2], hrs_measure.value); //bytes[0..2] = *hrs_measure.value.to_le_bytes(); LittleEndian::write_u16(&mut bytes[2..4], hrs_measure.energy_expended); LittleEndian::write_u16(&mut bytes[4..6], hrs_measure.aRR_interval_values[0]); bytes[6] = hrs_measure.valid_intervals; bytes[7] = hrs_measure.flags.bits(); hrs_service.heart_rate_measurement.set_value(&bytes); return Ok(hrs_service); } fn get_bd_addr() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = lhci_info.device_type_id; bytes[4] = (lhci_info.st_company_id & 0xff) as u8; bytes[5] = (lhci_info.st_company_id >> 8 & 0xff) as u8; BdAddr(bytes) } fn get_random_addr() -> BdAddr { let mut bytes = [0u8; 6]; let lhci_info = LhciC1DeviceInformationCcrp::new(); bytes[0] = (lhci_info.uid64 & 0xff) as u8; bytes[1] = ((lhci_info.uid64 >> 8) & 0xff) as u8; bytes[2] = ((lhci_info.uid64 >> 16) & 0xff) as u8; bytes[3] = 0; bytes[4] = 0x6E; bytes[5] = 0xED; BdAddr(bytes) } const BLE_CFG_IRK: [u8; 16] = [ 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, 0x12, 0x34, 0x56, 0x78, 0x9a, 0xbc, 0xde, 0xf0, ]; const BLE_CFG_ERK: [u8; 16] = [ 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, 0xfe, 0xdc, 0xba, 0x09, 0x87, 0x65, 0x43, 0x21, ]; fn get_irk() -> EncryptionKey { EncryptionKey(BLE_CFG_IRK) } fn get_erk() -> EncryptionKey
const DISCOVERY_PARAMS: DiscoverableParameters = DiscoverableParameters { advertising_type: AdvertisingType::ConnectableUndirected, advertising_interval: Some(( Duration::from_millis(ADV_INTERVAL_MS), Duration::from_millis(ADV_INTERVAL_MS), )), address_type: OwnAddressType::Public, filter_policy: AdvertisingFilterPolicy::AllowConnectionAndScan, // Local name should be empty for the device to be recognized as an Eddystone beacon local_name: Some(LocalName::Complete(BT_NAME)), advertising_data: &[], conn_interval: (None, None), };
{ EncryptionKey(BLE_CFG_ERK) }
identifier_body
main.rs
// Rust notes fn main () { let mut x = String::from("Hey there"); let mut y = String::from("Hello dawg woof"); println!("{}",first_word(&x)); println!("{}",second_word(&y)); } fn
(x : &String) -> &str { let bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &x[0..i]; } } &x[..] } fn second_word(x : &String) -> &str { let mut bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' '{ let y = &x[i+1..]; bytes = y.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &y[0..i]; } } // Return this IF there were only two words. return &y[..]; } } &x[..] } //__________________________________________________________________________________________________________ // //--------------------------------------------------// //*** GENERAL NOTES *** //*** Info: Random Notes I found either important, *** //*** hard to remember, funny, cool, or who knows *** //*** why but I added it to this list *** //--------------------------------------------------// // Notes: // 1. Variables immutable by default // 2.Constants ALWAYS immutable, type must be annotated ex: const MAX_POINTS: u32 = 758000; // 3. Scaler variables: Ints, floats, bools, and chars // Length Signed Unsigned // 8-bit i8 u8 // 16-bit i16 u16 // 32-bit i32 u32 // 64-bit i64 u64 // 128-bit i128 u128 // arch isize usize // 4. Integer literals // Number literals Example // Decimal 98_222 // Hex 0xff // Octal 0o77 // Binary 0b1111_0000 // Byte (u8 only) b'A' // 5. Floats can be either f32 or f64 (32 and 64 bits respectively)...default is f64 //ex: // fn main() { // let x = 2.0; // f64 // let y: f32 = 3.0; // f32 //} // 6. bools example // fn main() { // let t = true; // let f: bool = false; // with explicit type annotation // } // 7. Tuples AND Arrays are fixed length. // 8. Arrays allocate to stack instead of the heap, are used instead of vectors when you KNOW the num of elements wont need to change, like an array // containing the days of the week. // 9. Functions can be defined anywhere in the program. // 10. In rust cannot do statements like: x = y = 6 // 11. *IMPORTANT* Expressions (return something) lines do not contain a semi colon. Statements(return NOTHING) DO have a semi colon. // 12. Unlike other languages, conditions for if statements MUST be a bool. // 13. *IMPORTANT* (Straight out of the rust book) // To ensure memory safety, there’s one more detail to what happens in this situation in Rust. Instead of trying to copy the allocated memory, Rust considers s1 to no longer be valid and, therefore, Rust doesn’t need to free anything when s1 goes out of scope. Check out what happens when you try to use s1 after s2 is created; it won’t work: // This code does not compile! // let s1 = String::from("hello"); // let s2 = s1; // println!("{}, world!", s1); // You’ll get an error like this because Rust prevents you from using the invalidated reference: // error[E0382]: use of moved value: `s1` // --> src/main.rs:5:28 // | // 3 | let s2 = s1; // | -- value moved here // 4 | // 5 | println!("{}, world!", s1); // | ^^ value used here after move // | // = note: move occurs because `s1` has type `std::string::String`, which does // not implement the `Copy` trait // 14. *FUTURE notes on copying, moving, and losing access to data // 15. References are immutable by default just like variables, know you can have only ONE mutable reference to a piece of data in a particular SCOPE. // 16. // 17. // 18. // 19. // 20. //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** INFINITE loop, unless x is ever >= 10 *** //---------------------------------------------// // fn main() { // let mut x = 0; // loop { // x += 1; // println!("Value of x is {}", x); // if x >= 10 { // break; // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** WHILE loop, while n <= 50, also only prints multiples of 5 *** //---------------------------------------------// // fn main() { // let mut n = 1; // while n <= 50 { // if n % 5 == 0 { // println!("n is {}", n); // } // n += 1; // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** FOR loop, a few options for iterating *** //---------------------------------------------// // fn main() { // // makes a range from 0-129, nums is of type range // let nums = 0..130; // //must have an iterator // for i in nums { // println!("Num is {}", i ); // } // // can use a vector // let fruits = vec!["Watermelon", "Appple", "Orange"]; // //must have an iterator, calls iterator method of fruits vector // for a in fruits.iter() { // println!("Fruit type is: {}", a ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ENUM def && MATCH usage *** //---------------------------------------------// // enum cardinal_direction { // North, // South, // East, // West // } // fn main() { // let player_cardinal:cardinal_direction = cardinal_direction:: North; // //like a switch statement // match player_cardinal { // cardinal_direction::North => println!("Heading North!"), // cardinal_direction::South => println!("Heading South!"), // cardinal_direction::East => println!("Heading East!"), // cardinal_direction::West => println!("Heading West!"), // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** CONST defintion *** //---------------------------------------------// // const MAX_NUM: u8 = 20; // // fn main() { // for n in 1..MAX_NUM { // print!("{}", n ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** TUPLES *** //---------------------------------------------// // fn main() { // // creating a tuple // // NOTE: CAN have nested tuples // let johnshon = (5, "ch" , 20.5 , 23); // // to be more specific... // let tup: (i32, f64, u8) = (500, 6.4, 1); // println!("{}", johnshon.2 ); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ARRAYS *** // Info: Indexed the same as C based langs //---------------------------------------------// // fn main() { // let a = [1, 2, 3, 4, 5]; // // the 5 indicates the array size // let a: [i32; 5] = [1, 2, 3, 4, 5]; // } // //__________________________________________________________________________________________________________ // // * ALSO * // Info: This shows a pattern to turn the tup into those 3 seperate variables, also demonstrates how to index a tuple // fn main() { // let tup = (500, 6.4, 1); // // let (x, y, z) = tup; // // println!("The value of y is: {}", y); // // let x: (i32, f64, u8) = (500, 6.4, 1); // // let five_hundred = x.0; // // let six_point_four = x.1; // // let one = x.2; // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Calling Functions *** // Info: Youhave to declare the type of each parameter //---------------------------------------------// // fn main() { // let x = 5; // print_number(x); // if is_even(x){ // println!("Is even!"); // } // else { // println!("Is odd!", ); // } // } // fn print_number(num:u32){ // println!("number is {}", num); // } // // fn accepts u32 num and "returns" a bool // fn is_even(num:u32) -> bool{ // return (num % 2 == 0); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Shadowing *** //---------------------------------------------// // fn main() { // let mut x = 20; // { // let x = 15; // // do stuff with this 15, then x will go back to original value, called "shadowing" // } // // will print out 20 // println!("{}", x); // } //------------------------------------------------------// //*** Shadowing Cont... *** // // Info: This allows for an immutable var to // // occasionally be redifined, but // // to in general still stay an immutable var // //------------------------------------------------------// // fn main() { // let x = 5; // let x = x + 1; // let x = x * 2; // println!("The value of x is: {}", x); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Guessing Game Program *** //*** Info: Demonstrates usage of rand crate and *** //*** some basic error handling *** //---------------------------------------------// // use std::cmp::Ordering; // use std:: io; // use rand:: Rng; // fn main() { // // to learn how to use the crates, like rand, do cargo doc --open in the command line to open up documentation in your browser // let secret_num = rand::thread_rng().gen_range(1, 101); // loop { // println!("Guess the number!"); // println!("Input your guess:"); // // A new instance of a String, which is growable and UTF-8 encoded text // // '::' means new is an associated function of the String type, not the instance of the string. Similar to a static method in other languages. // let mut guess = String::new(); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it may or may not print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut guess) // .expect("Failed to read"); //NOTE: will default to type i32 // // trim eliminates new lines and whitespaces that might exist in the string we are attempting to cast to an u32 // let guess: u32 = match guess.trim().parse() { // Ok(num) => num, // Err(_)=> continue, //the '_' is a catchall val that will continue if parse's result type return val is Err // }; // println!("User guessed: {}", guess ); // match guess.cmp(&secret_num){ // Ordering::Less => println!("Too small!"), // Ordering::Greater => println!("Too large!"), // Ordering::Equal => // { // println!("Correct!"); // break; // } // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** References *** //---------------------------------------------// // fn main() { // let mut jacob = 21; // // creating a REFERENCE to jacob, will behave exactly as 'jacob' does // //let jacobR = &jacob; // // creating a mutable REFERENCE to jacob, that can be changed and WILL change 'jacob's value // let jacobRM = &mut jacob; // *jacobRM = 22; // println!("{}", jacob ); // } //__________________________________________________________________________________________________________ // //* ANOTHER REFERENCES EXAMPLE * // fn main() { // let s1 = String::from("hello"); // let len = calculate_length(&s1); // println!("The length of '{}' is {}.", s1, len); // } // fn calculate_length(s: &String) -> usize { // s.len() // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Range and Reverse in a For Loop *** //---------------------------------------------// // fn main() { // for number in (1..4).rev() { // println!("{}!", number); // } // println!("LIFTOFF!!!"); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Cloning *** // info: So that s1 and s2 dont point to the same // piece of memory, is more expensive than the // normal copying, which is really a "move" //---------------------------------------------// // fn main() { // let s1 = String::from("hello"); // let s2 = s1.clone(); // println!("s1 = {}, s2 = {}", s1, s2); // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Find Nth Fibonacci Number *** //---------------------------------------------// // use std::io; // //use std::cmp::Ordering; // fn main() { // let mut num = String::new(); // println!("What nth Fibonacci number would you like to find?"); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it will print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut num) // .expect("Failed to read"); //NOTE: will default to type i32 // let num: u32 = num.trim().parse().expect("Was not an integer!"); // let nth = fib_num(num); // println!("The nth Fibonnacci Number was: {}", nth ); // } // fn fib_num(x: u32) -> u32{ // if x <= 1 { // return x; // } // return fib_num(x-1) + fib_num(x-2); // } // //__________________________________________________________________________________________________________ // // //---------------------------------------------------------------------------------------// //*** SLICES: example, return first word in string, and second word in string function *** //--------------------------------------------------------------------------------------// // // fn main () { // let mut x = String::from("Hey there"); // let mut y = String::from("Hello dawg woof"); // println!("{}",first_word(&x)); // println!("{}",second_word(&y)); // } // fn first_word(x : &String) -> &str { // let bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &x[0..i]; // } // } // &x[..] // } // fn second_word(x : &String) -> &str { // let mut bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' '{ // let y = &x[i+1..]; // bytes = y.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &y[0..i]; // } // } // // Return this IF there were only two words. // return &y[..]; // } // } // &x[..] // }
first_word
identifier_name
main.rs
// Rust notes fn main () { let mut x = String::from("Hey there"); let mut y = String::from("Hello dawg woof"); println!("{}",first_word(&x)); println!("{}",second_word(&y)); } fn first_word(x : &String) -> &str
fn second_word(x : &String) -> &str { let mut bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' '{ let y = &x[i+1..]; bytes = y.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &y[0..i]; } } // Return this IF there were only two words. return &y[..]; } } &x[..] } //__________________________________________________________________________________________________________ // //--------------------------------------------------// //*** GENERAL NOTES *** //*** Info: Random Notes I found either important, *** //*** hard to remember, funny, cool, or who knows *** //*** why but I added it to this list *** //--------------------------------------------------// // Notes: // 1. Variables immutable by default // 2.Constants ALWAYS immutable, type must be annotated ex: const MAX_POINTS: u32 = 758000; // 3. Scaler variables: Ints, floats, bools, and chars // Length Signed Unsigned // 8-bit i8 u8 // 16-bit i16 u16 // 32-bit i32 u32 // 64-bit i64 u64 // 128-bit i128 u128 // arch isize usize // 4. Integer literals // Number literals Example // Decimal 98_222 // Hex 0xff // Octal 0o77 // Binary 0b1111_0000 // Byte (u8 only) b'A' // 5. Floats can be either f32 or f64 (32 and 64 bits respectively)...default is f64 //ex: // fn main() { // let x = 2.0; // f64 // let y: f32 = 3.0; // f32 //} // 6. bools example // fn main() { // let t = true; // let f: bool = false; // with explicit type annotation // } // 7. Tuples AND Arrays are fixed length. // 8. Arrays allocate to stack instead of the heap, are used instead of vectors when you KNOW the num of elements wont need to change, like an array // containing the days of the week. // 9. Functions can be defined anywhere in the program. // 10. In rust cannot do statements like: x = y = 6 // 11. *IMPORTANT* Expressions (return something) lines do not contain a semi colon. Statements(return NOTHING) DO have a semi colon. // 12. Unlike other languages, conditions for if statements MUST be a bool. // 13. *IMPORTANT* (Straight out of the rust book) // To ensure memory safety, there’s one more detail to what happens in this situation in Rust. Instead of trying to copy the allocated memory, Rust considers s1 to no longer be valid and, therefore, Rust doesn’t need to free anything when s1 goes out of scope. Check out what happens when you try to use s1 after s2 is created; it won’t work: // This code does not compile! // let s1 = String::from("hello"); // let s2 = s1; // println!("{}, world!", s1); // You’ll get an error like this because Rust prevents you from using the invalidated reference: // error[E0382]: use of moved value: `s1` // --> src/main.rs:5:28 // | // 3 | let s2 = s1; // | -- value moved here // 4 | // 5 | println!("{}, world!", s1); // | ^^ value used here after move // | // = note: move occurs because `s1` has type `std::string::String`, which does // not implement the `Copy` trait // 14. *FUTURE notes on copying, moving, and losing access to data // 15. References are immutable by default just like variables, know you can have only ONE mutable reference to a piece of data in a particular SCOPE. // 16. // 17. // 18. // 19. // 20. //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** INFINITE loop, unless x is ever >= 10 *** //---------------------------------------------// // fn main() { // let mut x = 0; // loop { // x += 1; // println!("Value of x is {}", x); // if x >= 10 { // break; // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** WHILE loop, while n <= 50, also only prints multiples of 5 *** //---------------------------------------------// // fn main() { // let mut n = 1; // while n <= 50 { // if n % 5 == 0 { // println!("n is {}", n); // } // n += 1; // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** FOR loop, a few options for iterating *** //---------------------------------------------// // fn main() { // // makes a range from 0-129, nums is of type range // let nums = 0..130; // //must have an iterator // for i in nums { // println!("Num is {}", i ); // } // // can use a vector // let fruits = vec!["Watermelon", "Appple", "Orange"]; // //must have an iterator, calls iterator method of fruits vector // for a in fruits.iter() { // println!("Fruit type is: {}", a ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ENUM def && MATCH usage *** //---------------------------------------------// // enum cardinal_direction { // North, // South, // East, // West // } // fn main() { // let player_cardinal:cardinal_direction = cardinal_direction:: North; // //like a switch statement // match player_cardinal { // cardinal_direction::North => println!("Heading North!"), // cardinal_direction::South => println!("Heading South!"), // cardinal_direction::East => println!("Heading East!"), // cardinal_direction::West => println!("Heading West!"), // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** CONST defintion *** //---------------------------------------------// // const MAX_NUM: u8 = 20; // // fn main() { // for n in 1..MAX_NUM { // print!("{}", n ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** TUPLES *** //---------------------------------------------// // fn main() { // // creating a tuple // // NOTE: CAN have nested tuples // let johnshon = (5, "ch" , 20.5 , 23); // // to be more specific... // let tup: (i32, f64, u8) = (500, 6.4, 1); // println!("{}", johnshon.2 ); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ARRAYS *** // Info: Indexed the same as C based langs //---------------------------------------------// // fn main() { // let a = [1, 2, 3, 4, 5]; // // the 5 indicates the array size // let a: [i32; 5] = [1, 2, 3, 4, 5]; // } // //__________________________________________________________________________________________________________ // // * ALSO * // Info: This shows a pattern to turn the tup into those 3 seperate variables, also demonstrates how to index a tuple // fn main() { // let tup = (500, 6.4, 1); // // let (x, y, z) = tup; // // println!("The value of y is: {}", y); // // let x: (i32, f64, u8) = (500, 6.4, 1); // // let five_hundred = x.0; // // let six_point_four = x.1; // // let one = x.2; // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Calling Functions *** // Info: Youhave to declare the type of each parameter //---------------------------------------------// // fn main() { // let x = 5; // print_number(x); // if is_even(x){ // println!("Is even!"); // } // else { // println!("Is odd!", ); // } // } // fn print_number(num:u32){ // println!("number is {}", num); // } // // fn accepts u32 num and "returns" a bool // fn is_even(num:u32) -> bool{ // return (num % 2 == 0); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Shadowing *** //---------------------------------------------// // fn main() { // let mut x = 20; // { // let x = 15; // // do stuff with this 15, then x will go back to original value, called "shadowing" // } // // will print out 20 // println!("{}", x); // } //------------------------------------------------------// //*** Shadowing Cont... *** // // Info: This allows for an immutable var to // // occasionally be redifined, but // // to in general still stay an immutable var // //------------------------------------------------------// // fn main() { // let x = 5; // let x = x + 1; // let x = x * 2; // println!("The value of x is: {}", x); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Guessing Game Program *** //*** Info: Demonstrates usage of rand crate and *** //*** some basic error handling *** //---------------------------------------------// // use std::cmp::Ordering; // use std:: io; // use rand:: Rng; // fn main() { // // to learn how to use the crates, like rand, do cargo doc --open in the command line to open up documentation in your browser // let secret_num = rand::thread_rng().gen_range(1, 101); // loop { // println!("Guess the number!"); // println!("Input your guess:"); // // A new instance of a String, which is growable and UTF-8 encoded text // // '::' means new is an associated function of the String type, not the instance of the string. Similar to a static method in other languages. // let mut guess = String::new(); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it may or may not print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut guess) // .expect("Failed to read"); //NOTE: will default to type i32 // // trim eliminates new lines and whitespaces that might exist in the string we are attempting to cast to an u32 // let guess: u32 = match guess.trim().parse() { // Ok(num) => num, // Err(_)=> continue, //the '_' is a catchall val that will continue if parse's result type return val is Err // }; // println!("User guessed: {}", guess ); // match guess.cmp(&secret_num){ // Ordering::Less => println!("Too small!"), // Ordering::Greater => println!("Too large!"), // Ordering::Equal => // { // println!("Correct!"); // break; // } // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** References *** //---------------------------------------------// // fn main() { // let mut jacob = 21; // // creating a REFERENCE to jacob, will behave exactly as 'jacob' does // //let jacobR = &jacob; // // creating a mutable REFERENCE to jacob, that can be changed and WILL change 'jacob's value // let jacobRM = &mut jacob; // *jacobRM = 22; // println!("{}", jacob ); // } //__________________________________________________________________________________________________________ // //* ANOTHER REFERENCES EXAMPLE * // fn main() { // let s1 = String::from("hello"); // let len = calculate_length(&s1); // println!("The length of '{}' is {}.", s1, len); // } // fn calculate_length(s: &String) -> usize { // s.len() // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Range and Reverse in a For Loop *** //---------------------------------------------// // fn main() { // for number in (1..4).rev() { // println!("{}!", number); // } // println!("LIFTOFF!!!"); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Cloning *** // info: So that s1 and s2 dont point to the same // piece of memory, is more expensive than the // normal copying, which is really a "move" //---------------------------------------------// // fn main() { // let s1 = String::from("hello"); // let s2 = s1.clone(); // println!("s1 = {}, s2 = {}", s1, s2); // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Find Nth Fibonacci Number *** //---------------------------------------------// // use std::io; // //use std::cmp::Ordering; // fn main() { // let mut num = String::new(); // println!("What nth Fibonacci number would you like to find?"); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it will print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut num) // .expect("Failed to read"); //NOTE: will default to type i32 // let num: u32 = num.trim().parse().expect("Was not an integer!"); // let nth = fib_num(num); // println!("The nth Fibonnacci Number was: {}", nth ); // } // fn fib_num(x: u32) -> u32{ // if x <= 1 { // return x; // } // return fib_num(x-1) + fib_num(x-2); // } // //__________________________________________________________________________________________________________ // // //---------------------------------------------------------------------------------------// //*** SLICES: example, return first word in string, and second word in string function *** //--------------------------------------------------------------------------------------// // // fn main () { // let mut x = String::from("Hey there"); // let mut y = String::from("Hello dawg woof"); // println!("{}",first_word(&x)); // println!("{}",second_word(&y)); // } // fn first_word(x : &String) -> &str { // let bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &x[0..i]; // } // } // &x[..] // } // fn second_word(x : &String) -> &str { // let mut bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' '{ // let y = &x[i+1..]; // bytes = y.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &y[0..i]; // } // } // // Return this IF there were only two words. // return &y[..]; // } // } // &x[..] // }
{ let bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &x[0..i]; } } &x[..] }
identifier_body
main.rs
// Rust notes fn main () { let mut x = String::from("Hey there"); let mut y = String::from("Hello dawg woof"); println!("{}",first_word(&x)); println!("{}",second_word(&y)); } fn first_word(x : &String) -> &str { let bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &x[0..i]; } }
&x[..] } fn second_word(x : &String) -> &str { let mut bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' '{ let y = &x[i+1..]; bytes = y.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &y[0..i]; } } // Return this IF there were only two words. return &y[..]; } } &x[..] } //__________________________________________________________________________________________________________ // //--------------------------------------------------// //*** GENERAL NOTES *** //*** Info: Random Notes I found either important, *** //*** hard to remember, funny, cool, or who knows *** //*** why but I added it to this list *** //--------------------------------------------------// // Notes: // 1. Variables immutable by default // 2.Constants ALWAYS immutable, type must be annotated ex: const MAX_POINTS: u32 = 758000; // 3. Scaler variables: Ints, floats, bools, and chars // Length Signed Unsigned // 8-bit i8 u8 // 16-bit i16 u16 // 32-bit i32 u32 // 64-bit i64 u64 // 128-bit i128 u128 // arch isize usize // 4. Integer literals // Number literals Example // Decimal 98_222 // Hex 0xff // Octal 0o77 // Binary 0b1111_0000 // Byte (u8 only) b'A' // 5. Floats can be either f32 or f64 (32 and 64 bits respectively)...default is f64 //ex: // fn main() { // let x = 2.0; // f64 // let y: f32 = 3.0; // f32 //} // 6. bools example // fn main() { // let t = true; // let f: bool = false; // with explicit type annotation // } // 7. Tuples AND Arrays are fixed length. // 8. Arrays allocate to stack instead of the heap, are used instead of vectors when you KNOW the num of elements wont need to change, like an array // containing the days of the week. // 9. Functions can be defined anywhere in the program. // 10. In rust cannot do statements like: x = y = 6 // 11. *IMPORTANT* Expressions (return something) lines do not contain a semi colon. Statements(return NOTHING) DO have a semi colon. // 12. Unlike other languages, conditions for if statements MUST be a bool. // 13. *IMPORTANT* (Straight out of the rust book) // To ensure memory safety, there’s one more detail to what happens in this situation in Rust. Instead of trying to copy the allocated memory, Rust considers s1 to no longer be valid and, therefore, Rust doesn’t need to free anything when s1 goes out of scope. Check out what happens when you try to use s1 after s2 is created; it won’t work: // This code does not compile! // let s1 = String::from("hello"); // let s2 = s1; // println!("{}, world!", s1); // You’ll get an error like this because Rust prevents you from using the invalidated reference: // error[E0382]: use of moved value: `s1` // --> src/main.rs:5:28 // | // 3 | let s2 = s1; // | -- value moved here // 4 | // 5 | println!("{}, world!", s1); // | ^^ value used here after move // | // = note: move occurs because `s1` has type `std::string::String`, which does // not implement the `Copy` trait // 14. *FUTURE notes on copying, moving, and losing access to data // 15. References are immutable by default just like variables, know you can have only ONE mutable reference to a piece of data in a particular SCOPE. // 16. // 17. // 18. // 19. // 20. //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** INFINITE loop, unless x is ever >= 10 *** //---------------------------------------------// // fn main() { // let mut x = 0; // loop { // x += 1; // println!("Value of x is {}", x); // if x >= 10 { // break; // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** WHILE loop, while n <= 50, also only prints multiples of 5 *** //---------------------------------------------// // fn main() { // let mut n = 1; // while n <= 50 { // if n % 5 == 0 { // println!("n is {}", n); // } // n += 1; // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** FOR loop, a few options for iterating *** //---------------------------------------------// // fn main() { // // makes a range from 0-129, nums is of type range // let nums = 0..130; // //must have an iterator // for i in nums { // println!("Num is {}", i ); // } // // can use a vector // let fruits = vec!["Watermelon", "Appple", "Orange"]; // //must have an iterator, calls iterator method of fruits vector // for a in fruits.iter() { // println!("Fruit type is: {}", a ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ENUM def && MATCH usage *** //---------------------------------------------// // enum cardinal_direction { // North, // South, // East, // West // } // fn main() { // let player_cardinal:cardinal_direction = cardinal_direction:: North; // //like a switch statement // match player_cardinal { // cardinal_direction::North => println!("Heading North!"), // cardinal_direction::South => println!("Heading South!"), // cardinal_direction::East => println!("Heading East!"), // cardinal_direction::West => println!("Heading West!"), // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** CONST defintion *** //---------------------------------------------// // const MAX_NUM: u8 = 20; // // fn main() { // for n in 1..MAX_NUM { // print!("{}", n ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** TUPLES *** //---------------------------------------------// // fn main() { // // creating a tuple // // NOTE: CAN have nested tuples // let johnshon = (5, "ch" , 20.5 , 23); // // to be more specific... // let tup: (i32, f64, u8) = (500, 6.4, 1); // println!("{}", johnshon.2 ); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ARRAYS *** // Info: Indexed the same as C based langs //---------------------------------------------// // fn main() { // let a = [1, 2, 3, 4, 5]; // // the 5 indicates the array size // let a: [i32; 5] = [1, 2, 3, 4, 5]; // } // //__________________________________________________________________________________________________________ // // * ALSO * // Info: This shows a pattern to turn the tup into those 3 seperate variables, also demonstrates how to index a tuple // fn main() { // let tup = (500, 6.4, 1); // // let (x, y, z) = tup; // // println!("The value of y is: {}", y); // // let x: (i32, f64, u8) = (500, 6.4, 1); // // let five_hundred = x.0; // // let six_point_four = x.1; // // let one = x.2; // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Calling Functions *** // Info: Youhave to declare the type of each parameter //---------------------------------------------// // fn main() { // let x = 5; // print_number(x); // if is_even(x){ // println!("Is even!"); // } // else { // println!("Is odd!", ); // } // } // fn print_number(num:u32){ // println!("number is {}", num); // } // // fn accepts u32 num and "returns" a bool // fn is_even(num:u32) -> bool{ // return (num % 2 == 0); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Shadowing *** //---------------------------------------------// // fn main() { // let mut x = 20; // { // let x = 15; // // do stuff with this 15, then x will go back to original value, called "shadowing" // } // // will print out 20 // println!("{}", x); // } //------------------------------------------------------// //*** Shadowing Cont... *** // // Info: This allows for an immutable var to // // occasionally be redifined, but // // to in general still stay an immutable var // //------------------------------------------------------// // fn main() { // let x = 5; // let x = x + 1; // let x = x * 2; // println!("The value of x is: {}", x); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Guessing Game Program *** //*** Info: Demonstrates usage of rand crate and *** //*** some basic error handling *** //---------------------------------------------// // use std::cmp::Ordering; // use std:: io; // use rand:: Rng; // fn main() { // // to learn how to use the crates, like rand, do cargo doc --open in the command line to open up documentation in your browser // let secret_num = rand::thread_rng().gen_range(1, 101); // loop { // println!("Guess the number!"); // println!("Input your guess:"); // // A new instance of a String, which is growable and UTF-8 encoded text // // '::' means new is an associated function of the String type, not the instance of the string. Similar to a static method in other languages. // let mut guess = String::new(); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it may or may not print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut guess) // .expect("Failed to read"); //NOTE: will default to type i32 // // trim eliminates new lines and whitespaces that might exist in the string we are attempting to cast to an u32 // let guess: u32 = match guess.trim().parse() { // Ok(num) => num, // Err(_)=> continue, //the '_' is a catchall val that will continue if parse's result type return val is Err // }; // println!("User guessed: {}", guess ); // match guess.cmp(&secret_num){ // Ordering::Less => println!("Too small!"), // Ordering::Greater => println!("Too large!"), // Ordering::Equal => // { // println!("Correct!"); // break; // } // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** References *** //---------------------------------------------// // fn main() { // let mut jacob = 21; // // creating a REFERENCE to jacob, will behave exactly as 'jacob' does // //let jacobR = &jacob; // // creating a mutable REFERENCE to jacob, that can be changed and WILL change 'jacob's value // let jacobRM = &mut jacob; // *jacobRM = 22; // println!("{}", jacob ); // } //__________________________________________________________________________________________________________ // //* ANOTHER REFERENCES EXAMPLE * // fn main() { // let s1 = String::from("hello"); // let len = calculate_length(&s1); // println!("The length of '{}' is {}.", s1, len); // } // fn calculate_length(s: &String) -> usize { // s.len() // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Range and Reverse in a For Loop *** //---------------------------------------------// // fn main() { // for number in (1..4).rev() { // println!("{}!", number); // } // println!("LIFTOFF!!!"); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Cloning *** // info: So that s1 and s2 dont point to the same // piece of memory, is more expensive than the // normal copying, which is really a "move" //---------------------------------------------// // fn main() { // let s1 = String::from("hello"); // let s2 = s1.clone(); // println!("s1 = {}, s2 = {}", s1, s2); // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Find Nth Fibonacci Number *** //---------------------------------------------// // use std::io; // //use std::cmp::Ordering; // fn main() { // let mut num = String::new(); // println!("What nth Fibonacci number would you like to find?"); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it will print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut num) // .expect("Failed to read"); //NOTE: will default to type i32 // let num: u32 = num.trim().parse().expect("Was not an integer!"); // let nth = fib_num(num); // println!("The nth Fibonnacci Number was: {}", nth ); // } // fn fib_num(x: u32) -> u32{ // if x <= 1 { // return x; // } // return fib_num(x-1) + fib_num(x-2); // } // //__________________________________________________________________________________________________________ // // //---------------------------------------------------------------------------------------// //*** SLICES: example, return first word in string, and second word in string function *** //--------------------------------------------------------------------------------------// // // fn main () { // let mut x = String::from("Hey there"); // let mut y = String::from("Hello dawg woof"); // println!("{}",first_word(&x)); // println!("{}",second_word(&y)); // } // fn first_word(x : &String) -> &str { // let bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &x[0..i]; // } // } // &x[..] // } // fn second_word(x : &String) -> &str { // let mut bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' '{ // let y = &x[i+1..]; // bytes = y.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &y[0..i]; // } // } // // Return this IF there were only two words. // return &y[..]; // } // } // &x[..] // }
random_line_split
main.rs
// Rust notes fn main () { let mut x = String::from("Hey there"); let mut y = String::from("Hello dawg woof"); println!("{}",first_word(&x)); println!("{}",second_word(&y)); } fn first_word(x : &String) -> &str { let bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &x[0..i]; } } &x[..] } fn second_word(x : &String) -> &str { let mut bytes = x.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' '
} &x[..] } //__________________________________________________________________________________________________________ // //--------------------------------------------------// //*** GENERAL NOTES *** //*** Info: Random Notes I found either important, *** //*** hard to remember, funny, cool, or who knows *** //*** why but I added it to this list *** //--------------------------------------------------// // Notes: // 1. Variables immutable by default // 2.Constants ALWAYS immutable, type must be annotated ex: const MAX_POINTS: u32 = 758000; // 3. Scaler variables: Ints, floats, bools, and chars // Length Signed Unsigned // 8-bit i8 u8 // 16-bit i16 u16 // 32-bit i32 u32 // 64-bit i64 u64 // 128-bit i128 u128 // arch isize usize // 4. Integer literals // Number literals Example // Decimal 98_222 // Hex 0xff // Octal 0o77 // Binary 0b1111_0000 // Byte (u8 only) b'A' // 5. Floats can be either f32 or f64 (32 and 64 bits respectively)...default is f64 //ex: // fn main() { // let x = 2.0; // f64 // let y: f32 = 3.0; // f32 //} // 6. bools example // fn main() { // let t = true; // let f: bool = false; // with explicit type annotation // } // 7. Tuples AND Arrays are fixed length. // 8. Arrays allocate to stack instead of the heap, are used instead of vectors when you KNOW the num of elements wont need to change, like an array // containing the days of the week. // 9. Functions can be defined anywhere in the program. // 10. In rust cannot do statements like: x = y = 6 // 11. *IMPORTANT* Expressions (return something) lines do not contain a semi colon. Statements(return NOTHING) DO have a semi colon. // 12. Unlike other languages, conditions for if statements MUST be a bool. // 13. *IMPORTANT* (Straight out of the rust book) // To ensure memory safety, there’s one more detail to what happens in this situation in Rust. Instead of trying to copy the allocated memory, Rust considers s1 to no longer be valid and, therefore, Rust doesn’t need to free anything when s1 goes out of scope. Check out what happens when you try to use s1 after s2 is created; it won’t work: // This code does not compile! // let s1 = String::from("hello"); // let s2 = s1; // println!("{}, world!", s1); // You’ll get an error like this because Rust prevents you from using the invalidated reference: // error[E0382]: use of moved value: `s1` // --> src/main.rs:5:28 // | // 3 | let s2 = s1; // | -- value moved here // 4 | // 5 | println!("{}, world!", s1); // | ^^ value used here after move // | // = note: move occurs because `s1` has type `std::string::String`, which does // not implement the `Copy` trait // 14. *FUTURE notes on copying, moving, and losing access to data // 15. References are immutable by default just like variables, know you can have only ONE mutable reference to a piece of data in a particular SCOPE. // 16. // 17. // 18. // 19. // 20. //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** INFINITE loop, unless x is ever >= 10 *** //---------------------------------------------// // fn main() { // let mut x = 0; // loop { // x += 1; // println!("Value of x is {}", x); // if x >= 10 { // break; // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** WHILE loop, while n <= 50, also only prints multiples of 5 *** //---------------------------------------------// // fn main() { // let mut n = 1; // while n <= 50 { // if n % 5 == 0 { // println!("n is {}", n); // } // n += 1; // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** FOR loop, a few options for iterating *** //---------------------------------------------// // fn main() { // // makes a range from 0-129, nums is of type range // let nums = 0..130; // //must have an iterator // for i in nums { // println!("Num is {}", i ); // } // // can use a vector // let fruits = vec!["Watermelon", "Appple", "Orange"]; // //must have an iterator, calls iterator method of fruits vector // for a in fruits.iter() { // println!("Fruit type is: {}", a ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ENUM def && MATCH usage *** //---------------------------------------------// // enum cardinal_direction { // North, // South, // East, // West // } // fn main() { // let player_cardinal:cardinal_direction = cardinal_direction:: North; // //like a switch statement // match player_cardinal { // cardinal_direction::North => println!("Heading North!"), // cardinal_direction::South => println!("Heading South!"), // cardinal_direction::East => println!("Heading East!"), // cardinal_direction::West => println!("Heading West!"), // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** CONST defintion *** //---------------------------------------------// // const MAX_NUM: u8 = 20; // // fn main() { // for n in 1..MAX_NUM { // print!("{}", n ); // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** TUPLES *** //---------------------------------------------// // fn main() { // // creating a tuple // // NOTE: CAN have nested tuples // let johnshon = (5, "ch" , 20.5 , 23); // // to be more specific... // let tup: (i32, f64, u8) = (500, 6.4, 1); // println!("{}", johnshon.2 ); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** ARRAYS *** // Info: Indexed the same as C based langs //---------------------------------------------// // fn main() { // let a = [1, 2, 3, 4, 5]; // // the 5 indicates the array size // let a: [i32; 5] = [1, 2, 3, 4, 5]; // } // //__________________________________________________________________________________________________________ // // * ALSO * // Info: This shows a pattern to turn the tup into those 3 seperate variables, also demonstrates how to index a tuple // fn main() { // let tup = (500, 6.4, 1); // // let (x, y, z) = tup; // // println!("The value of y is: {}", y); // // let x: (i32, f64, u8) = (500, 6.4, 1); // // let five_hundred = x.0; // // let six_point_four = x.1; // // let one = x.2; // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Calling Functions *** // Info: Youhave to declare the type of each parameter //---------------------------------------------// // fn main() { // let x = 5; // print_number(x); // if is_even(x){ // println!("Is even!"); // } // else { // println!("Is odd!", ); // } // } // fn print_number(num:u32){ // println!("number is {}", num); // } // // fn accepts u32 num and "returns" a bool // fn is_even(num:u32) -> bool{ // return (num % 2 == 0); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Shadowing *** //---------------------------------------------// // fn main() { // let mut x = 20; // { // let x = 15; // // do stuff with this 15, then x will go back to original value, called "shadowing" // } // // will print out 20 // println!("{}", x); // } //------------------------------------------------------// //*** Shadowing Cont... *** // // Info: This allows for an immutable var to // // occasionally be redifined, but // // to in general still stay an immutable var // //------------------------------------------------------// // fn main() { // let x = 5; // let x = x + 1; // let x = x * 2; // println!("The value of x is: {}", x); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Guessing Game Program *** //*** Info: Demonstrates usage of rand crate and *** //*** some basic error handling *** //---------------------------------------------// // use std::cmp::Ordering; // use std:: io; // use rand:: Rng; // fn main() { // // to learn how to use the crates, like rand, do cargo doc --open in the command line to open up documentation in your browser // let secret_num = rand::thread_rng().gen_range(1, 101); // loop { // println!("Guess the number!"); // println!("Input your guess:"); // // A new instance of a String, which is growable and UTF-8 encoded text // // '::' means new is an associated function of the String type, not the instance of the string. Similar to a static method in other languages. // let mut guess = String::new(); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it may or may not print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut guess) // .expect("Failed to read"); //NOTE: will default to type i32 // // trim eliminates new lines and whitespaces that might exist in the string we are attempting to cast to an u32 // let guess: u32 = match guess.trim().parse() { // Ok(num) => num, // Err(_)=> continue, //the '_' is a catchall val that will continue if parse's result type return val is Err // }; // println!("User guessed: {}", guess ); // match guess.cmp(&secret_num){ // Ordering::Less => println!("Too small!"), // Ordering::Greater => println!("Too large!"), // Ordering::Equal => // { // println!("Correct!"); // break; // } // } // } // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** References *** //---------------------------------------------// // fn main() { // let mut jacob = 21; // // creating a REFERENCE to jacob, will behave exactly as 'jacob' does // //let jacobR = &jacob; // // creating a mutable REFERENCE to jacob, that can be changed and WILL change 'jacob's value // let jacobRM = &mut jacob; // *jacobRM = 22; // println!("{}", jacob ); // } //__________________________________________________________________________________________________________ // //* ANOTHER REFERENCES EXAMPLE * // fn main() { // let s1 = String::from("hello"); // let len = calculate_length(&s1); // println!("The length of '{}' is {}.", s1, len); // } // fn calculate_length(s: &String) -> usize { // s.len() // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Range and Reverse in a For Loop *** //---------------------------------------------// // fn main() { // for number in (1..4).rev() { // println!("{}!", number); // } // println!("LIFTOFF!!!"); // } //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Cloning *** // info: So that s1 and s2 dont point to the same // piece of memory, is more expensive than the // normal copying, which is really a "move" //---------------------------------------------// // fn main() { // let s1 = String::from("hello"); // let s2 = s1.clone(); // println!("s1 = {}, s2 = {}", s1, s2); // } // //__________________________________________________________________________________________________________ // //---------------------------------------------// //*** Find Nth Fibonacci Number *** //---------------------------------------------// // use std::io; // //use std::cmp::Ordering; // fn main() { // let mut num = String::new(); // println!("What nth Fibonacci number would you like to find?"); // // calls read_line which takes in user input into guess, checks whether it returns err or ok, and depending on that it will print out the failure message // // basically the expect part does some error checking right there, and will print the passed in err message and crash the program if err is returned from read_line // io::stdin().read_line(&mut num) // .expect("Failed to read"); //NOTE: will default to type i32 // let num: u32 = num.trim().parse().expect("Was not an integer!"); // let nth = fib_num(num); // println!("The nth Fibonnacci Number was: {}", nth ); // } // fn fib_num(x: u32) -> u32{ // if x <= 1 { // return x; // } // return fib_num(x-1) + fib_num(x-2); // } // //__________________________________________________________________________________________________________ // // //---------------------------------------------------------------------------------------// //*** SLICES: example, return first word in string, and second word in string function *** //--------------------------------------------------------------------------------------// // // fn main () { // let mut x = String::from("Hey there"); // let mut y = String::from("Hello dawg woof"); // println!("{}",first_word(&x)); // println!("{}",second_word(&y)); // } // fn first_word(x : &String) -> &str { // let bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &x[0..i]; // } // } // &x[..] // } // fn second_word(x : &String) -> &str { // let mut bytes = x.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' '{ // let y = &x[i+1..]; // bytes = y.as_bytes(); // for (i, &item) in bytes.iter().enumerate() { // if item == b' ' { // return &y[0..i]; // } // } // // Return this IF there were only two words. // return &y[..]; // } // } // &x[..] // }
{ let y = &x[i+1..]; bytes = y.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &y[0..i]; } } // Return this IF there were only two words. return &y[..]; }
conditional_block
SocialSpiderAlgorithm.py
import math import numpy as np from numpy import linalg as la import random from termcolor import colored import matplotlib.pyplot as plt Minimize_problem = True Maximize_problem = False class Spider: def __init__(self, s, s_previous, fs, vibration, cs, mask): self.s = s # The position of s on the web. self.s_previous = s_previous # The movement that s performed in the previous iteration self.fs = fs # The fitness of the current position of s. self.vibration = vibration # The target vibration of s in the previous iteration. self.cs = cs # The number of iterations since s has last changed its target vibration. self.mask = mask # The dimension mask in the previous iteration. def printout(self): print("position = "+str(self.s)) print("previous position = " + str(self.s_previous)) print("fitness = "+str(self.fs)) self.vibration.printout() print("Cs = "+str(self.cs)) print("Mask = "+str(self.mask)) class Vibration: def __init__(self, position, intensity): self.position = position # the source position or target vibration self.intensity = intensity # source intensity of the vibration def set_position_and_intensity(self, position, intensity): self.position = position self.intensity = intensity @staticmethod def intensity_position_ps_position_ps(fs): return math.log(1/(fs-c)+1) # I(Ps;Ps; t) def printout(self): print("Intensity = "+str(self.intensity)) def distance(pa, pb): return la.norm(pa - pb, 1) # D(Pa;Pb) def intensity_position_pa_position_pb(pa, pb, sd, vibration_pa): # pa spider a & pb spider b, I(Pa;Pb; t) return vibration_pa.intensity*math.exp((-distance(pa.s, pb.s))/(ra * sd)) # Calculate Standard_Deviation σ along each dimension def standard_deviation(): pop = [spider.s for spider in spiders] return np.sum(np.std(pop, axis=1)) / n def f(a): z = [] if Minimize_problem: z.extend(a) return eval(y) elif Maximize_problem: z.extend(-a) return -eval(y) # there is a array with 100 elements with one and zero,100*p elements with 0 , 100(1-p) with 1,0=false,1=true # where p is the probability def probability(p): arr = np.array([0] * int(100 * p) + [1] * int(100 - 100 * p)) np.random.shuffle(arr) rand = random.choice(arr) if rand == 0: return True else: return False def show(generate_vibration): for x in range(population): print("") print("spider" + str(x)) spiders[x].printout() print("generate vibration = " + str(generate_vibration[x].intensity)) print("") # if return true then it is out of bounds [a,b] def out_of_bounds(position): for x in range(len(position)): if position[x] < bounds[x, 0] or position[x] > bounds[x, 1]: return True return False def create_population_of_spiders(): for x in range(population): s = np.zeros(n) for x1 in range(n): s[x1] = np.random.uniform(bounds[x1, 0], bounds[x1, 1]) vibration = Vibration(s, 0) spiders.append(Spider(s, s, 0, vibration, 0, np.zeros(n))) def initialization_graphics(graph): global ax if graph and n==2: plt.close('all') fig = plt.figure(figsize=(7, 7)) ax = plt.axes() plt.xlim(bounds[0, 0]-1, bounds[0, 1]+1) plt.ylim(bounds[1, 0]-1, bounds[1, 1]+1) def graphics(X, Y): if n == 2: ax.scatter(X, Y, c='black') plt.draw() plt.pause(1) ax.collections[0].remove() def social_spider_algorithm(graph): g
# Sphere function minimum = 0 def test_function_0(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "z[0]**2 + z[1]**2 + z[2]**2 + z[3]**2 + z[4]**2" n = 5 # dimensions # solution space or domain of definition [a , b] each dimensions bounds = np.array([[-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000]]) population = 20 lim = 3000 # max steps of iterations return # Three-hump camel function minimum = 0 def test_function_1(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "2*z[0]**2 - 1.05 * z[0]**4 + z[0]**6 /6 + z[0] * z[1] + z[1]**2" n = 2 bounds = np.array([[-5, 5], [-5, 5]]) population = 10 lim = 10 # max steps of iterations return # McCormick function minimum = -1.9133 def test_function_2(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "math.sin(z[0] + z[1]) + (z[0] - z[1])**2 - 1.5 * z[0] + 2.5 * z[1] + 1" n = 2 bounds = np.array([[-1.5, 4], [-3, 4]]) population = 10 lim = 1000 # max steps of iterations return # Schaffer function N. 4 minimum = 0.292579 def test_function_3(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "0.5 + ((math.cos(math.sin(abs(z[0]**2 - z[1]**2)))**2 - 0.5) / (1 + 0.001*(z[0]**2 + z[1]**2))**2)" n = 2 bounds = np.array([[-100, 100], [-100, 100]]) population = 25 lim = 2500 # max steps of iterations return # Code for Test for test in range(20): test_function_1() print('\n' + colored("Τest "+str(test + 1), 'blue')+'\n'+social_spider_algorithm(False)) # test_function_3() # social_spider_algorithm(True) # file = open("test.txt", "w+") # file.write('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') # file.write('\n' + "Τest "+str(test + 1)+'\n'+social_spider_algorithm(False)) # file.close()
lobal spiders spiders = [] create_population_of_spiders() minimize = spiders[0].s number_of_iterations = 0 initialization_graphics(graph) # In the iteration phase while number_of_iterations <= lim: # Calculates the fitness , update the global optimum and generate vibrations generate_vibration = [] for x in range(population): spiders[x].fs = f(spiders[x].s) if f(minimize) > spiders[x].fs: minimize = spiders[x].s generate_vibration.append( Vibration(spiders[x].s, spiders[x].vibration.intensity_position_ps_position_ps(spiders[x].fs))) # show(generate_vibration) # print("minimize = " + str(minimize)+"f(minimize) = "+str(f(minimize))) # Calculate the intensity of the vibrations V # generated by all spiders and Select the strongest vibration sd = standard_deviation() for x in range(population): max_vibration = Vibration(np.zeros(n), -1) for t in range(population): if x != t: intensity = intensity_position_pa_position_pb(spiders[x], spiders[t], sd, generate_vibration[x]) if max_vibration.intensity < intensity: max_vibration.set_position_and_intensity(spiders[t].s, intensity) if max_vibration.intensity > spiders[x].vibration.intensity: spiders[x].vibration.set_position_and_intensity(max_vibration.position, max_vibration.intensity) spiders[x].cs = 0 else: spiders[x].cs += 1 # change mask or not if not probability(pc**spiders[x].cs): for p in range(n): if probability(pm): spiders[x].mask[p] = 1 else: spiders[x].mask[p] = 0 # In case all bits are zeros or ones if n == np.count_nonzero(spiders[x].mask): # all ones spiders[x].mask[random.randint(0, n - 1)] = 0 elif np.count_nonzero(spiders[x].mask) == 0: # all zeros spiders[x].mask[random.randint(0, n - 1)] = 1 p_s_fo = np.array([]) # position is generated based on the mask for s r = random.randint(0, population - 1) for d in range(n): if spiders[x].mask[d] == 0: p_s_fo = np.append(p_s_fo, spiders[x].vibration.position[d]) elif spiders[x].mask[d] == 1: p_s_fo = np.append(p_s_fo, generate_vibration[r].position[d]) # Calculate next position R = np.random.uniform(0, 1, n) next_position = spiders[x].s + (spiders[x].s - spiders[x].s_previous) * r + (p_s_fo - spiders[x].s) * R spiders[x].s_previous = spiders[x].s # Address any violated constraints. if out_of_bounds(next_position): rand_float = random.random() # random [0,1] for t in range(n): if next_position[t] > bounds[t, 1]: next_position[t] = (bounds[t, 1] - spiders[x].s[t]) * rand_float elif next_position[t] < bounds[t, 0]: next_position[t] = (spiders[x].s[t] - bounds[t, 0]) * rand_float spiders[x].s = next_position if graph and n == 2: graphics([atr.s[0] for atr in spiders], [atr.s[1] for atr in spiders]) number_of_iterations += 1 return "global minimize = " + str(minimize)+'\n'+"f(minimize) = " + str(f(minimize))
identifier_body
SocialSpiderAlgorithm.py
import math import numpy as np from numpy import linalg as la import random from termcolor import colored import matplotlib.pyplot as plt Minimize_problem = True Maximize_problem = False class Spider: def __init__(self, s, s_previous, fs, vibration, cs, mask): self.s = s # The position of s on the web. self.s_previous = s_previous # The movement that s performed in the previous iteration self.fs = fs # The fitness of the current position of s. self.vibration = vibration # The target vibration of s in the previous iteration. self.cs = cs # The number of iterations since s has last changed its target vibration. self.mask = mask # The dimension mask in the previous iteration. def printout(self): print("position = "+str(self.s)) print("previous position = " + str(self.s_previous)) print("fitness = "+str(self.fs)) self.vibration.printout() print("Cs = "+str(self.cs)) print("Mask = "+str(self.mask)) class Vibration: def __init__(self, position, intensity): self.position = position # the source position or target vibration self.intensity = intensity # source intensity of the vibration def set_position_and_intensity(self, position, intensity): self.position = position self.intensity = intensity @staticmethod def intensity_position_ps_position_ps(fs): return math.log(1/(fs-c)+1) # I(Ps;Ps; t) def printout(self): print("Intensity = "+str(self.intensity)) def distance(pa, pb): return la.norm(pa - pb, 1) # D(Pa;Pb) def intensity_position_pa_position_pb(pa, pb, sd, vibration_pa): # pa spider a & pb spider b, I(Pa;Pb; t) return vibration_pa.intensity*math.exp((-distance(pa.s, pb.s))/(ra * sd)) # Calculate Standard_Deviation σ along each dimension def standard_deviation(): pop = [spider.s for spider in spiders] return np.sum(np.std(pop, axis=1)) / n def f
a): z = [] if Minimize_problem: z.extend(a) return eval(y) elif Maximize_problem: z.extend(-a) return -eval(y) # there is a array with 100 elements with one and zero,100*p elements with 0 , 100(1-p) with 1,0=false,1=true # where p is the probability def probability(p): arr = np.array([0] * int(100 * p) + [1] * int(100 - 100 * p)) np.random.shuffle(arr) rand = random.choice(arr) if rand == 0: return True else: return False def show(generate_vibration): for x in range(population): print("") print("spider" + str(x)) spiders[x].printout() print("generate vibration = " + str(generate_vibration[x].intensity)) print("") # if return true then it is out of bounds [a,b] def out_of_bounds(position): for x in range(len(position)): if position[x] < bounds[x, 0] or position[x] > bounds[x, 1]: return True return False def create_population_of_spiders(): for x in range(population): s = np.zeros(n) for x1 in range(n): s[x1] = np.random.uniform(bounds[x1, 0], bounds[x1, 1]) vibration = Vibration(s, 0) spiders.append(Spider(s, s, 0, vibration, 0, np.zeros(n))) def initialization_graphics(graph): global ax if graph and n==2: plt.close('all') fig = plt.figure(figsize=(7, 7)) ax = plt.axes() plt.xlim(bounds[0, 0]-1, bounds[0, 1]+1) plt.ylim(bounds[1, 0]-1, bounds[1, 1]+1) def graphics(X, Y): if n == 2: ax.scatter(X, Y, c='black') plt.draw() plt.pause(1) ax.collections[0].remove() def social_spider_algorithm(graph): global spiders spiders = [] create_population_of_spiders() minimize = spiders[0].s number_of_iterations = 0 initialization_graphics(graph) # In the iteration phase while number_of_iterations <= lim: # Calculates the fitness , update the global optimum and generate vibrations generate_vibration = [] for x in range(population): spiders[x].fs = f(spiders[x].s) if f(minimize) > spiders[x].fs: minimize = spiders[x].s generate_vibration.append( Vibration(spiders[x].s, spiders[x].vibration.intensity_position_ps_position_ps(spiders[x].fs))) # show(generate_vibration) # print("minimize = " + str(minimize)+"f(minimize) = "+str(f(minimize))) # Calculate the intensity of the vibrations V # generated by all spiders and Select the strongest vibration sd = standard_deviation() for x in range(population): max_vibration = Vibration(np.zeros(n), -1) for t in range(population): if x != t: intensity = intensity_position_pa_position_pb(spiders[x], spiders[t], sd, generate_vibration[x]) if max_vibration.intensity < intensity: max_vibration.set_position_and_intensity(spiders[t].s, intensity) if max_vibration.intensity > spiders[x].vibration.intensity: spiders[x].vibration.set_position_and_intensity(max_vibration.position, max_vibration.intensity) spiders[x].cs = 0 else: spiders[x].cs += 1 # change mask or not if not probability(pc**spiders[x].cs): for p in range(n): if probability(pm): spiders[x].mask[p] = 1 else: spiders[x].mask[p] = 0 # In case all bits are zeros or ones if n == np.count_nonzero(spiders[x].mask): # all ones spiders[x].mask[random.randint(0, n - 1)] = 0 elif np.count_nonzero(spiders[x].mask) == 0: # all zeros spiders[x].mask[random.randint(0, n - 1)] = 1 p_s_fo = np.array([]) # position is generated based on the mask for s r = random.randint(0, population - 1) for d in range(n): if spiders[x].mask[d] == 0: p_s_fo = np.append(p_s_fo, spiders[x].vibration.position[d]) elif spiders[x].mask[d] == 1: p_s_fo = np.append(p_s_fo, generate_vibration[r].position[d]) # Calculate next position R = np.random.uniform(0, 1, n) next_position = spiders[x].s + (spiders[x].s - spiders[x].s_previous) * r + (p_s_fo - spiders[x].s) * R spiders[x].s_previous = spiders[x].s # Address any violated constraints. if out_of_bounds(next_position): rand_float = random.random() # random [0,1] for t in range(n): if next_position[t] > bounds[t, 1]: next_position[t] = (bounds[t, 1] - spiders[x].s[t]) * rand_float elif next_position[t] < bounds[t, 0]: next_position[t] = (spiders[x].s[t] - bounds[t, 0]) * rand_float spiders[x].s = next_position if graph and n == 2: graphics([atr.s[0] for atr in spiders], [atr.s[1] for atr in spiders]) number_of_iterations += 1 return "global minimize = " + str(minimize)+'\n'+"f(minimize) = " + str(f(minimize)) # Sphere function minimum = 0 def test_function_0(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "z[0]**2 + z[1]**2 + z[2]**2 + z[3]**2 + z[4]**2" n = 5 # dimensions # solution space or domain of definition [a , b] each dimensions bounds = np.array([[-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000]]) population = 20 lim = 3000 # max steps of iterations return # Three-hump camel function minimum = 0 def test_function_1(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "2*z[0]**2 - 1.05 * z[0]**4 + z[0]**6 /6 + z[0] * z[1] + z[1]**2" n = 2 bounds = np.array([[-5, 5], [-5, 5]]) population = 10 lim = 10 # max steps of iterations return # McCormick function minimum = -1.9133 def test_function_2(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "math.sin(z[0] + z[1]) + (z[0] - z[1])**2 - 1.5 * z[0] + 2.5 * z[1] + 1" n = 2 bounds = np.array([[-1.5, 4], [-3, 4]]) population = 10 lim = 1000 # max steps of iterations return # Schaffer function N. 4 minimum = 0.292579 def test_function_3(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "0.5 + ((math.cos(math.sin(abs(z[0]**2 - z[1]**2)))**2 - 0.5) / (1 + 0.001*(z[0]**2 + z[1]**2))**2)" n = 2 bounds = np.array([[-100, 100], [-100, 100]]) population = 25 lim = 2500 # max steps of iterations return # Code for Test for test in range(20): test_function_1() print('\n' + colored("Τest "+str(test + 1), 'blue')+'\n'+social_spider_algorithm(False)) # test_function_3() # social_spider_algorithm(True) # file = open("test.txt", "w+") # file.write('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') # file.write('\n' + "Τest "+str(test + 1)+'\n'+social_spider_algorithm(False)) # file.close()
(
identifier_name
SocialSpiderAlgorithm.py
import math import numpy as np from numpy import linalg as la import random from termcolor import colored import matplotlib.pyplot as plt Minimize_problem = True Maximize_problem = False class Spider: def __init__(self, s, s_previous, fs, vibration, cs, mask): self.s = s # The position of s on the web. self.s_previous = s_previous # The movement that s performed in the previous iteration self.fs = fs # The fitness of the current position of s. self.vibration = vibration # The target vibration of s in the previous iteration. self.cs = cs # The number of iterations since s has last changed its target vibration. self.mask = mask # The dimension mask in the previous iteration. def printout(self): print("position = "+str(self.s)) print("previous position = " + str(self.s_previous)) print("fitness = "+str(self.fs)) self.vibration.printout() print("Cs = "+str(self.cs)) print("Mask = "+str(self.mask)) class Vibration: def __init__(self, position, intensity): self.position = position # the source position or target vibration self.intensity = intensity # source intensity of the vibration def set_position_and_intensity(self, position, intensity): self.position = position self.intensity = intensity @staticmethod def intensity_position_ps_position_ps(fs): return math.log(1/(fs-c)+1) # I(Ps;Ps; t) def printout(self): print("Intensity = "+str(self.intensity)) def distance(pa, pb): return la.norm(pa - pb, 1) # D(Pa;Pb) def intensity_position_pa_position_pb(pa, pb, sd, vibration_pa): # pa spider a & pb spider b, I(Pa;Pb; t) return vibration_pa.intensity*math.exp((-distance(pa.s, pb.s))/(ra * sd)) # Calculate Standard_Deviation σ along each dimension def standard_deviation(): pop = [spider.s for spider in spiders] return np.sum(np.std(pop, axis=1)) / n def f(a): z = [] if Minimize_problem: z.extend(a) return eval(y) elif Maximize_problem: z.extend(-a) return -eval(y) # there is a array with 100 elements with one and zero,100*p elements with 0 , 100(1-p) with 1,0=false,1=true # where p is the probability def probability(p): arr = np.array([0] * int(100 * p) + [1] * int(100 - 100 * p)) np.random.shuffle(arr) rand = random.choice(arr) if rand == 0: return True else: return False def show(generate_vibration): for x in range(population): print("") print("spider" + str(x)) spiders[x].printout() print("generate vibration = " + str(generate_vibration[x].intensity)) print("") # if return true then it is out of bounds [a,b] def out_of_bounds(position): for x in range(len(position)): if position[x] < bounds[x, 0] or position[x] > bounds[x, 1]: return True return False def create_population_of_spiders(): for x in range(population): s = np.zeros(n) for x1 in range(n): s[x1] = np.random.uniform(bounds[x1, 0], bounds[x1, 1]) vibration = Vibration(s, 0) spiders.append(Spider(s, s, 0, vibration, 0, np.zeros(n))) def initialization_graphics(graph): global ax if graph and n==2: plt.close('all') fig = plt.figure(figsize=(7, 7)) ax = plt.axes() plt.xlim(bounds[0, 0]-1, bounds[0, 1]+1) plt.ylim(bounds[1, 0]-1, bounds[1, 1]+1) def graphics(X, Y): if n == 2: ax.scatter(X, Y, c='black') plt.draw() plt.pause(1) ax.collections[0].remove() def social_spider_algorithm(graph): global spiders spiders = [] create_population_of_spiders() minimize = spiders[0].s number_of_iterations = 0 initialization_graphics(graph) # In the iteration phase while number_of_iterations <= lim: # Calculates the fitness , update the global optimum and generate vibrations generate_vibration = [] for x in range(population): spiders[x].fs = f(spiders[x].s) if f(minimize) > spiders[x].fs: minimize = spiders[x].s generate_vibration.append( Vibration(spiders[x].s, spiders[x].vibration.intensity_position_ps_position_ps(spiders[x].fs))) # show(generate_vibration) # print("minimize = " + str(minimize)+"f(minimize) = "+str(f(minimize))) # Calculate the intensity of the vibrations V # generated by all spiders and Select the strongest vibration sd = standard_deviation() for x in range(population): max_vibration = Vibration(np.zeros(n), -1) for t in range(population): if x != t: intensity = intensity_position_pa_position_pb(spiders[x], spiders[t], sd, generate_vibration[x]) if max_vibration.intensity < intensity: max_vibration.set_position_and_intensity(spiders[t].s, intensity) if max_vibration.intensity > spiders[x].vibration.intensity: spiders[x].vibration.set_position_and_intensity(max_vibration.position, max_vibration.intensity) spiders[x].cs = 0 else: spiders[x].cs += 1 # change mask or not if not probability(pc**spiders[x].cs): for p in range(n): if probability(pm): spiders[x].mask[p] = 1 else: spiders[x].mask[p] = 0 # In case all bits are zeros or ones if n == np.count_nonzero(spiders[x].mask): # all ones spiders[x].mask[random.randint(0, n - 1)] = 0 elif np.count_nonzero(spiders[x].mask) == 0: # all zeros spiders[x].mask[random.randint(0, n - 1)] = 1 p_s_fo = np.array([]) # position is generated based on the mask for s r = random.randint(0, population - 1) for d in range(n): if spiders[x].mask[d] == 0: p_s_fo = np.append(p_s_fo, spiders[x].vibration.position[d]) elif spiders[x].mask[d] == 1: p_s_fo = np.append(p_s_fo, generate_vibration[r].position[d]) # Calculate next position R = np.random.uniform(0, 1, n) next_position = spiders[x].s + (spiders[x].s - spiders[x].s_previous) * r + (p_s_fo - spiders[x].s) * R spiders[x].s_previous = spiders[x].s # Address any violated constraints. if out_of_bounds(next_position): rand_float = random.random() # random [0,1] for t in range(n): if next_position[t] > bounds[t, 1]: next_position[t] = (bounds[t, 1] - spiders[x].s[t]) * rand_float elif next_position[t] < bounds[t, 0]: next_position[t] = (spiders[x].s[t] - bounds[t, 0]) * rand_float spiders[x].s = next_position if graph and n == 2: graphics([atr.s[0] for atr in spiders], [atr.s[1] for atr in spiders]) number_of_iterations += 1 return "global minimize = " + str(minimize)+'\n'+"f(minimize) = " + str(f(minimize)) # Sphere function minimum = 0 def test_function_0(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "z[0]**2 + z[1]**2 + z[2]**2 + z[3]**2 + z[4]**2" n = 5 # dimensions # solution space or domain of definition [a , b] each dimensions bounds = np.array([[-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000]]) population = 20 lim = 3000 # max steps of iterations return # Three-hump camel function minimum = 0 def test_function_1(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "2*z[0]**2 - 1.05 * z[0]**4 + z[0]**6 /6 + z[0] * z[1] + z[1]**2" n = 2 bounds = np.array([[-5, 5], [-5, 5]]) population = 10 lim = 10 # max steps of iterations return # McCormick function minimum = -1.9133 def test_function_2(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra))
set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "math.sin(z[0] + z[1]) + (z[0] - z[1])**2 - 1.5 * z[0] + 2.5 * z[1] + 1" n = 2 bounds = np.array([[-1.5, 4], [-3, 4]]) population = 10 lim = 1000 # max steps of iterations return # Schaffer function N. 4 minimum = 0.292579 def test_function_3(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "0.5 + ((math.cos(math.sin(abs(z[0]**2 - z[1]**2)))**2 - 0.5) / (1 + 0.001*(z[0]**2 + z[1]**2))**2)" n = 2 bounds = np.array([[-100, 100], [-100, 100]]) population = 25 lim = 2500 # max steps of iterations return # Code for Test for test in range(20): test_function_1() print('\n' + colored("Τest "+str(test + 1), 'blue')+'\n'+social_spider_algorithm(False)) # test_function_3() # social_spider_algorithm(True) # file = open("test.txt", "w+") # file.write('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') # file.write('\n' + "Τest "+str(test + 1)+'\n'+social_spider_algorithm(False)) # file.close()
random_line_split
SocialSpiderAlgorithm.py
import math import numpy as np from numpy import linalg as la import random from termcolor import colored import matplotlib.pyplot as plt Minimize_problem = True Maximize_problem = False class Spider: def __init__(self, s, s_previous, fs, vibration, cs, mask): self.s = s # The position of s on the web. self.s_previous = s_previous # The movement that s performed in the previous iteration self.fs = fs # The fitness of the current position of s. self.vibration = vibration # The target vibration of s in the previous iteration. self.cs = cs # The number of iterations since s has last changed its target vibration. self.mask = mask # The dimension mask in the previous iteration. def printout(self): print("position = "+str(self.s)) print("previous position = " + str(self.s_previous)) print("fitness = "+str(self.fs)) self.vibration.printout() print("Cs = "+str(self.cs)) print("Mask = "+str(self.mask)) class Vibration: def __init__(self, position, intensity): self.position = position # the source position or target vibration self.intensity = intensity # source intensity of the vibration def set_position_and_intensity(self, position, intensity): self.position = position self.intensity = intensity @staticmethod def intensity_position_ps_position_ps(fs): return math.log(1/(fs-c)+1) # I(Ps;Ps; t) def printout(self): print("Intensity = "+str(self.intensity)) def distance(pa, pb): return la.norm(pa - pb, 1) # D(Pa;Pb) def intensity_position_pa_position_pb(pa, pb, sd, vibration_pa): # pa spider a & pb spider b, I(Pa;Pb; t) return vibration_pa.intensity*math.exp((-distance(pa.s, pb.s))/(ra * sd)) # Calculate Standard_Deviation σ along each dimension def standard_deviation(): pop = [spider.s for spider in spiders] return np.sum(np.std(pop, axis=1)) / n def f(a): z = [] if Minimize_problem: z.extend(a) return eval(y) elif Maximize_problem: z.extend(-a) return -eval(y) # there is a array with 100 elements with one and zero,100*p elements with 0 , 100(1-p) with 1,0=false,1=true # where p is the probability def probability(p): arr = np.array([0] * int(100 * p) + [1] * int(100 - 100 * p)) np.random.shuffle(arr) rand = random.choice(arr) if rand == 0: r
else: return False def show(generate_vibration): for x in range(population): print("") print("spider" + str(x)) spiders[x].printout() print("generate vibration = " + str(generate_vibration[x].intensity)) print("") # if return true then it is out of bounds [a,b] def out_of_bounds(position): for x in range(len(position)): if position[x] < bounds[x, 0] or position[x] > bounds[x, 1]: return True return False def create_population_of_spiders(): for x in range(population): s = np.zeros(n) for x1 in range(n): s[x1] = np.random.uniform(bounds[x1, 0], bounds[x1, 1]) vibration = Vibration(s, 0) spiders.append(Spider(s, s, 0, vibration, 0, np.zeros(n))) def initialization_graphics(graph): global ax if graph and n==2: plt.close('all') fig = plt.figure(figsize=(7, 7)) ax = plt.axes() plt.xlim(bounds[0, 0]-1, bounds[0, 1]+1) plt.ylim(bounds[1, 0]-1, bounds[1, 1]+1) def graphics(X, Y): if n == 2: ax.scatter(X, Y, c='black') plt.draw() plt.pause(1) ax.collections[0].remove() def social_spider_algorithm(graph): global spiders spiders = [] create_population_of_spiders() minimize = spiders[0].s number_of_iterations = 0 initialization_graphics(graph) # In the iteration phase while number_of_iterations <= lim: # Calculates the fitness , update the global optimum and generate vibrations generate_vibration = [] for x in range(population): spiders[x].fs = f(spiders[x].s) if f(minimize) > spiders[x].fs: minimize = spiders[x].s generate_vibration.append( Vibration(spiders[x].s, spiders[x].vibration.intensity_position_ps_position_ps(spiders[x].fs))) # show(generate_vibration) # print("minimize = " + str(minimize)+"f(minimize) = "+str(f(minimize))) # Calculate the intensity of the vibrations V # generated by all spiders and Select the strongest vibration sd = standard_deviation() for x in range(population): max_vibration = Vibration(np.zeros(n), -1) for t in range(population): if x != t: intensity = intensity_position_pa_position_pb(spiders[x], spiders[t], sd, generate_vibration[x]) if max_vibration.intensity < intensity: max_vibration.set_position_and_intensity(spiders[t].s, intensity) if max_vibration.intensity > spiders[x].vibration.intensity: spiders[x].vibration.set_position_and_intensity(max_vibration.position, max_vibration.intensity) spiders[x].cs = 0 else: spiders[x].cs += 1 # change mask or not if not probability(pc**spiders[x].cs): for p in range(n): if probability(pm): spiders[x].mask[p] = 1 else: spiders[x].mask[p] = 0 # In case all bits are zeros or ones if n == np.count_nonzero(spiders[x].mask): # all ones spiders[x].mask[random.randint(0, n - 1)] = 0 elif np.count_nonzero(spiders[x].mask) == 0: # all zeros spiders[x].mask[random.randint(0, n - 1)] = 1 p_s_fo = np.array([]) # position is generated based on the mask for s r = random.randint(0, population - 1) for d in range(n): if spiders[x].mask[d] == 0: p_s_fo = np.append(p_s_fo, spiders[x].vibration.position[d]) elif spiders[x].mask[d] == 1: p_s_fo = np.append(p_s_fo, generate_vibration[r].position[d]) # Calculate next position R = np.random.uniform(0, 1, n) next_position = spiders[x].s + (spiders[x].s - spiders[x].s_previous) * r + (p_s_fo - spiders[x].s) * R spiders[x].s_previous = spiders[x].s # Address any violated constraints. if out_of_bounds(next_position): rand_float = random.random() # random [0,1] for t in range(n): if next_position[t] > bounds[t, 1]: next_position[t] = (bounds[t, 1] - spiders[x].s[t]) * rand_float elif next_position[t] < bounds[t, 0]: next_position[t] = (spiders[x].s[t] - bounds[t, 0]) * rand_float spiders[x].s = next_position if graph and n == 2: graphics([atr.s[0] for atr in spiders], [atr.s[1] for atr in spiders]) number_of_iterations += 1 return "global minimize = " + str(minimize)+'\n'+"f(minimize) = " + str(f(minimize)) # Sphere function minimum = 0 def test_function_0(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "z[0]**2 + z[1]**2 + z[2]**2 + z[3]**2 + z[4]**2" n = 5 # dimensions # solution space or domain of definition [a , b] each dimensions bounds = np.array([[-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000], [-5000, 5000]]) population = 20 lim = 3000 # max steps of iterations return # Three-hump camel function minimum = 0 def test_function_1(): global c, ra, pc, pm, y, n, population, bounds, lim c = 1E-100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "2*z[0]**2 - 1.05 * z[0]**4 + z[0]**6 /6 + z[0] * z[1] + z[1]**2" n = 2 bounds = np.array([[-5, 5], [-5, 5]]) population = 10 lim = 10 # max steps of iterations return # McCormick function minimum = -1.9133 def test_function_2(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "math.sin(z[0] + z[1]) + (z[0] - z[1])**2 - 1.5 * z[0] + 2.5 * z[1] + 1" n = 2 bounds = np.array([[-1.5, 4], [-3, 4]]) population = 10 lim = 1000 # max steps of iterations return # Schaffer function N. 4 minimum = 0.292579 def test_function_3(): global c, ra, pc, pm, y, n, population, bounds, lim c = -100 # where C is a small constant such fitness values are larger than C set_ra = {1 / 10, 1 / 5, 1 / 4, 1 / 3, 1 / 2, 1, 2, 3, 4, 5} ra = random.choice(tuple(set_ra)) set_pc_and_pm = {0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.99} pc = random.choice(tuple(set_pc_and_pm)) pm = random.choice(tuple(set_pc_and_pm)) print('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') y = "0.5 + ((math.cos(math.sin(abs(z[0]**2 - z[1]**2)))**2 - 0.5) / (1 + 0.001*(z[0]**2 + z[1]**2))**2)" n = 2 bounds = np.array([[-100, 100], [-100, 100]]) population = 25 lim = 2500 # max steps of iterations return # Code for Test for test in range(20): test_function_1() print('\n' + colored("Τest "+str(test + 1), 'blue')+'\n'+social_spider_algorithm(False)) # test_function_3() # social_spider_algorithm(True) # file = open("test.txt", "w+") # file.write('\n'+"ra = " + str(ra) + '\n' + "pc = " + str(pc) + " pm = " + str(pm) + '\n') # file.write('\n' + "Τest "+str(test + 1)+'\n'+social_spider_algorithm(False)) # file.close()
eturn True
conditional_block
__init__.py
import collections import os import abc import copy import datetime import logging import munge from future.utils import with_metaclass from vaping.config import parse_interval import vaping.io class PluginBase(vaping.io.Thread): """ Base plugin interface # Instanced Attributes - config (`dict`): plugin config - vaping: reference to the main vaping object Calls `self.init()` prefork while loading all modules, init() should not do anything active, any files opened may be closed when it forks. Plugins should prefer `init()` to `__init__()` to ensure the class is completely done initializing. Calls `self.on_start()` and `self.on_stop()` before and after running in case any connections need to be created or cleaned up. """ @property def groups(self): """ `dict` - group configurations keyed by name """ group_config = {} # legacy way of threating any dict as a potential # group config (pre #44 implementation) # supported until vaping 2.0 for k,v in list(self.config.items()): if isinstance(v, collections.Mapping): group_config[k] = v # explicit groups object (#44 implementation) for _group_config in self.config.get("groups",[]): group_config[_group_config["name"]] = _group_config return group_config def init(self): """ called after the plugin is initialized, plugin may define this for any other initialization code """ pass def on_start(self): """ called when the daemon is starting """ pass def on_stop(self): """ called when the daemon is stopping """ pass def new_message(self): """ creates and returns new message `dict`, setting `type`, `source`, `ts`, `data` `data` is initialized to an empty array **Returns** message (`dict`) """ msg = {} msg['data'] = [] msg['type'] = self.plugin_type msg['source'] = self.name msg['ts'] = (datetime.datetime.utcnow() - datetime.datetime(1970, 1, 1)).total_seconds() return msg def popen(self, args, **kwargs): """ creates a subprocess with passed args **Returns** Popen instance """ self.log.debug("popen %s", ' '.join(args)) return vaping.io.subprocess.Popen(args, **kwargs) @property def log(self): """ logger instance for plugin type """ if not self._logger: self._logger = logging.getLogger('vaping.plugins.' + self.plugin_type) return self._logger def __init__(self, config, ctx): """ **Arguments** - config (`dict`) - ctx: vaping context """ if hasattr(self, 'default_config'): self.config = munge.util.recursive_update(copy.deepcopy(self.default_config), copy.deepcopy(config)) else: self.config = config # set for pluginmgr self.pluginmgr_config = self.config self.vaping = ctx self.name = self.config.get("name") self._logger = None super(PluginBase, self).__init__() self.init() def _run(self): self.on_start() class ProbeBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for probe plugin, used for getting data expects method probe() to be defined """ def init(self): pass @abc.abstractmethod def probe(self): """ probe for data, return a list of dicts """ def __init__(self, config, ctx, emit=None): if emit: self._emit = [emit] else: self._emit = [] self._emit_queue = vaping.io.Queue() super(ProbeBase, self).__init__(config, ctx) def _run(self): super(ProbeBase, self)._run() self.run_level = 1 while self.run_level: self.send_emission() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") def queue_emission(self, msg): """ queue an emission of a message for all output plugins **Arguments** - msg (`dict`): dict containing `type`, `source`, `ts` and `data` keys """ if not msg: return for _emitter in self._emit: if not hasattr(_emitter, 'emit'): continue def emit(emitter=_emitter): self.log.debug("emit to {}".format(emitter.name)) emitter.emit(msg) self.log.debug("queue emission to {} ({})".format( _emitter.name, self._emit_queue.qsize())) self._emit_queue.put(emit) def send_emission(self): """ emit and remove the first emission in the queue """ if self._emit_queue.empty(): return emit = self._emit_queue.get() emit() def emit_all(self): """ emit and remove all emissions in the queue """ while not self._emit_queue.empty(): self.send_emission() class TimedProbe(ProbeBase): """ Probe class that calls probe every config defined interval """ def __init__(self, config, ctx, emit=None): super(TimedProbe, self).__init__(config, ctx, emit) if 'interval' not in self.pluginmgr_config: raise ValueError('interval not set in config') self.interval = parse_interval(self.pluginmgr_config['interval']) self.run_level = 0 def _run(self): self.run_level = 1 while self.run_level: start = datetime.datetime.now() # since the TimedProbe will sleep between cycles # we need to emit all queued emissions each cycle self.emit_all() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") done = datetime.datetime.now() elapsed = done - start if elapsed.total_seconds() > self.interval: self.log.warning("probe time exceeded interval") else: sleeptime = datetime.timedelta(seconds=self.interval) - elapsed vaping.io.sleep(sleeptime.total_seconds()) class FileProbe(ProbeBase): """ Probes a file and emits everytime a new line is read # Config - path (`str`): path to file - backlog (`int=0`): number of bytes to read from backlog - max_lines (`int=1000`): maximum number of lines to read during probe # Instanced Attributes - path (`str`): path to file - backlog (`int`): number of bytes to read from backlog - max_lines (`int`): maximum number of liens to read during probe - fh (`filehandler`): file handler for opened file (only available if `path` is set) """ def __init__(self, config, ctx, emit=None): super(FileProbe, self).__init__(config, ctx, emit) self.path = self.pluginmgr_config.get("path") self.run_level = 0 self.backlog = int(self.pluginmgr_config.get("backlog",0)) self.max_lines = int(self.pluginmgr_config.get("max_lines",1000)) if self.path: self.fh = open(self.path, "r") self.fh.seek(0,2) if self.backlog: try: self.fh.seek(self.fh.tell() - self.backlog, os.SEEK_SET) except ValueError as exc: if str(exc).find("negative seek position") > -1: self.fh.seek(0) else: raise def _run(self): self.run_level = 1 while self.run_level: self.send_emission() for msg in self.probe(): self.queue_emission(msg) vaping.io.sleep(0.1) def validate_file_handler(self): """ Here we validate that our filehandler is pointing to an existing file.
If it doesnt, because file has been deleted, we close the filehander and try to reopen """ if self.fh.closed: try: self.fh = open(self.path, "r") self.fh.seek(0, 2) except OSError as err: logging.error("Could not reopen file: {}".format(err)) return False open_stat = os.fstat(self.fh.fileno()) try: file_stat = os.stat(self.path) except OSError as err: logging.error("Could not stat file: {}".format(err)) return False if open_stat != file_stat: self.log self.fh.close() return False return True def probe(self): """ Probe the file for new lines """ # make sure the filehandler is still valid # (e.g. file stat hasnt changed, file exists etc.) if not self.validate_file_handler(): return [] messages = [] # read any new lines and push them onto the stack for line in self.fh.readlines(self.max_lines): data = {"path":self.path} msg = self.new_message() # process the line - this is where parsing happens parsed = self.process_line(line, data) if not parsed: continue data.update(parsed) # process the probe - this is where data assignment # happens data = self.process_probe(data) msg["data"] = [data] messages.append(msg) # process all new messages before returning them # for emission messages = self.process_messages(messages) return messages def process_line(self, line, data): """ override this - parse your line in here """ return data def process_probe(self, data): """ override this - assign your data values here """ return data def process_messages(self, messages): """ override this - process your messages before they are emitted """ return messages class EmitBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for emit plugins, used for sending data expects method emit() to be defined """ def __init__(self, config, ctx): super(EmitBase, self).__init__(config, ctx) @abc.abstractmethod def emit(self, message): """ accept message to emit """ class TimeSeriesDB(EmitBase): """ Base interface for timeseries db storage plugins # Config - filename (`str`): database file name template - field (`str`): fieeld name to read the value from # Instanced Attributes - filename (`str`): database file name template - field (`str`): fieeld name to read the value from """ def __init__(self, config, ctx): super(TimeSeriesDB, self).__init__(config, ctx) # filename template self.filename = self.config.get("filename") # field name to read the value from self.field = self.config.get("field") if not self.filename: raise ValueError("No filename specified") if not self.field: raise ValueError("No field specified, field should specify which value to store in the database") def create(self, filename): """ Create database **Arguments** - filename (`str`): database filename """ raise NotImplementedError() def update(self, filename, time, value): """ Update database **Arguments** - filename (`str`): database filename - time (`int`): epoch timestamp - value (`mixed`) """ raise NotImplementedError() def get(self, filename, from_time, to_time): """ Retrieve data from database for the specified timespan **Arguments** - filename (`str`): database filename - from_time (`int`): epoch timestamp start - to_time (`int`): epoch timestamp end """ raise NotImplementedError() def filename_formatters(self, data, row): """ Returns a dict containing the various filename formatter values Values are gotten from the vaping data message as well as the currently processed row in the message **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatter variables (`dict`) """ r = { "source" : data.get("source"), "field" : self.field, "type" : data.get("type") } r.update(**row) return r def format_filename(self, data, row): """ Returns a formatted filename using the template stored in self.filename **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatted version of self.filename (`str`) """ return self.filename.format(**self.filename_formatters(data, row)) def emit(self, message): """ emit to database **Arguments** - message (`dict`): vaping message dict """ # handle vaping data that arrives in a list if isinstance(message.get("data"), list): for row in message.get("data"): # format filename from data filename = self.format_filename(message, row) # create database file if it does not exist yet if not os.path.exists(filename): self.create(filename) # update database self.log.debug("storing time:%d, %s:%s in %s" % ( message.get("ts"), self.field, row.get(self.field, "-"), filename)) self.update(filename, message.get("ts"), row.get(self.field))
random_line_split
__init__.py
import collections import os import abc import copy import datetime import logging import munge from future.utils import with_metaclass from vaping.config import parse_interval import vaping.io class PluginBase(vaping.io.Thread): """ Base plugin interface # Instanced Attributes - config (`dict`): plugin config - vaping: reference to the main vaping object Calls `self.init()` prefork while loading all modules, init() should not do anything active, any files opened may be closed when it forks. Plugins should prefer `init()` to `__init__()` to ensure the class is completely done initializing. Calls `self.on_start()` and `self.on_stop()` before and after running in case any connections need to be created or cleaned up. """ @property def groups(self): """ `dict` - group configurations keyed by name """ group_config = {} # legacy way of threating any dict as a potential # group config (pre #44 implementation) # supported until vaping 2.0 for k,v in list(self.config.items()): if isinstance(v, collections.Mapping): group_config[k] = v # explicit groups object (#44 implementation) for _group_config in self.config.get("groups",[]): group_config[_group_config["name"]] = _group_config return group_config def init(self): """ called after the plugin is initialized, plugin may define this for any other initialization code """ pass def on_start(self): """ called when the daemon is starting """ pass def on_stop(self): """ called when the daemon is stopping """ pass def new_message(self): """ creates and returns new message `dict`, setting `type`, `source`, `ts`, `data` `data` is initialized to an empty array **Returns** message (`dict`) """ msg = {} msg['data'] = [] msg['type'] = self.plugin_type msg['source'] = self.name msg['ts'] = (datetime.datetime.utcnow() - datetime.datetime(1970, 1, 1)).total_seconds() return msg def popen(self, args, **kwargs): """ creates a subprocess with passed args **Returns** Popen instance """ self.log.debug("popen %s", ' '.join(args)) return vaping.io.subprocess.Popen(args, **kwargs) @property def log(self): """ logger instance for plugin type """ if not self._logger: self._logger = logging.getLogger('vaping.plugins.' + self.plugin_type) return self._logger def __init__(self, config, ctx): """ **Arguments** - config (`dict`) - ctx: vaping context """ if hasattr(self, 'default_config'): self.config = munge.util.recursive_update(copy.deepcopy(self.default_config), copy.deepcopy(config)) else: self.config = config # set for pluginmgr self.pluginmgr_config = self.config self.vaping = ctx self.name = self.config.get("name") self._logger = None super(PluginBase, self).__init__() self.init() def _run(self): self.on_start() class ProbeBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for probe plugin, used for getting data expects method probe() to be defined """ def init(self): pass @abc.abstractmethod def probe(self): """ probe for data, return a list of dicts """ def __init__(self, config, ctx, emit=None): if emit: self._emit = [emit] else: self._emit = [] self._emit_queue = vaping.io.Queue() super(ProbeBase, self).__init__(config, ctx) def _run(self): super(ProbeBase, self)._run() self.run_level = 1 while self.run_level: self.send_emission() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") def queue_emission(self, msg): """ queue an emission of a message for all output plugins **Arguments** - msg (`dict`): dict containing `type`, `source`, `ts` and `data` keys """ if not msg: return for _emitter in self._emit: if not hasattr(_emitter, 'emit'): continue def emit(emitter=_emitter): self.log.debug("emit to {}".format(emitter.name)) emitter.emit(msg) self.log.debug("queue emission to {} ({})".format( _emitter.name, self._emit_queue.qsize())) self._emit_queue.put(emit) def send_emission(self): """ emit and remove the first emission in the queue """ if self._emit_queue.empty(): return emit = self._emit_queue.get() emit() def emit_all(self): """ emit and remove all emissions in the queue """ while not self._emit_queue.empty(): self.send_emission() class TimedProbe(ProbeBase): """ Probe class that calls probe every config defined interval """ def __init__(self, config, ctx, emit=None): super(TimedProbe, self).__init__(config, ctx, emit) if 'interval' not in self.pluginmgr_config: raise ValueError('interval not set in config') self.interval = parse_interval(self.pluginmgr_config['interval']) self.run_level = 0 def _run(self): self.run_level = 1 while self.run_level: start = datetime.datetime.now() # since the TimedProbe will sleep between cycles # we need to emit all queued emissions each cycle self.emit_all() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") done = datetime.datetime.now() elapsed = done - start if elapsed.total_seconds() > self.interval: self.log.warning("probe time exceeded interval") else: sleeptime = datetime.timedelta(seconds=self.interval) - elapsed vaping.io.sleep(sleeptime.total_seconds()) class FileProbe(ProbeBase): """ Probes a file and emits everytime a new line is read # Config - path (`str`): path to file - backlog (`int=0`): number of bytes to read from backlog - max_lines (`int=1000`): maximum number of lines to read during probe # Instanced Attributes - path (`str`): path to file - backlog (`int`): number of bytes to read from backlog - max_lines (`int`): maximum number of liens to read during probe - fh (`filehandler`): file handler for opened file (only available if `path` is set) """ def __init__(self, config, ctx, emit=None): super(FileProbe, self).__init__(config, ctx, emit) self.path = self.pluginmgr_config.get("path") self.run_level = 0 self.backlog = int(self.pluginmgr_config.get("backlog",0)) self.max_lines = int(self.pluginmgr_config.get("max_lines",1000)) if self.path: self.fh = open(self.path, "r") self.fh.seek(0,2) if self.backlog: try: self.fh.seek(self.fh.tell() - self.backlog, os.SEEK_SET) except ValueError as exc: if str(exc).find("negative seek position") > -1: self.fh.seek(0) else: raise def _run(self): self.run_level = 1 while self.run_level: self.send_emission() for msg in self.probe(): self.queue_emission(msg) vaping.io.sleep(0.1) def validate_file_handler(self): """ Here we validate that our filehandler is pointing to an existing file. If it doesnt, because file has been deleted, we close the filehander and try to reopen """ if self.fh.closed: try: self.fh = open(self.path, "r") self.fh.seek(0, 2) except OSError as err: logging.error("Could not reopen file: {}".format(err)) return False open_stat = os.fstat(self.fh.fileno()) try: file_stat = os.stat(self.path) except OSError as err: logging.error("Could not stat file: {}".format(err)) return False if open_stat != file_stat: self.log self.fh.close() return False return True def probe(self): """ Probe the file for new lines """ # make sure the filehandler is still valid # (e.g. file stat hasnt changed, file exists etc.) if not self.validate_file_handler(): return [] messages = [] # read any new lines and push them onto the stack for line in self.fh.readlines(self.max_lines): data = {"path":self.path} msg = self.new_message() # process the line - this is where parsing happens parsed = self.process_line(line, data) if not parsed: continue data.update(parsed) # process the probe - this is where data assignment # happens data = self.process_probe(data) msg["data"] = [data] messages.append(msg) # process all new messages before returning them # for emission messages = self.process_messages(messages) return messages def process_line(self, line, data): """ override this - parse your line in here """ return data def process_probe(self, data): """ override this - assign your data values here """ return data def process_messages(self, messages): """ override this - process your messages before they are emitted """ return messages class EmitBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for emit plugins, used for sending data expects method emit() to be defined """ def __init__(self, config, ctx): super(EmitBase, self).__init__(config, ctx) @abc.abstractmethod def emit(self, message): """ accept message to emit """ class TimeSeriesDB(EmitBase): """ Base interface for timeseries db storage plugins # Config - filename (`str`): database file name template - field (`str`): fieeld name to read the value from # Instanced Attributes - filename (`str`): database file name template - field (`str`): fieeld name to read the value from """ def __init__(self, config, ctx): super(TimeSeriesDB, self).__init__(config, ctx) # filename template self.filename = self.config.get("filename") # field name to read the value from self.field = self.config.get("field") if not self.filename: raise ValueError("No filename specified") if not self.field: raise ValueError("No field specified, field should specify which value to store in the database") def create(self, filename): """ Create database **Arguments** - filename (`str`): database filename """ raise NotImplementedError() def update(self, filename, time, value): """ Update database **Arguments** - filename (`str`): database filename - time (`int`): epoch timestamp - value (`mixed`) """ raise NotImplementedError() def
(self, filename, from_time, to_time): """ Retrieve data from database for the specified timespan **Arguments** - filename (`str`): database filename - from_time (`int`): epoch timestamp start - to_time (`int`): epoch timestamp end """ raise NotImplementedError() def filename_formatters(self, data, row): """ Returns a dict containing the various filename formatter values Values are gotten from the vaping data message as well as the currently processed row in the message **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatter variables (`dict`) """ r = { "source" : data.get("source"), "field" : self.field, "type" : data.get("type") } r.update(**row) return r def format_filename(self, data, row): """ Returns a formatted filename using the template stored in self.filename **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatted version of self.filename (`str`) """ return self.filename.format(**self.filename_formatters(data, row)) def emit(self, message): """ emit to database **Arguments** - message (`dict`): vaping message dict """ # handle vaping data that arrives in a list if isinstance(message.get("data"), list): for row in message.get("data"): # format filename from data filename = self.format_filename(message, row) # create database file if it does not exist yet if not os.path.exists(filename): self.create(filename) # update database self.log.debug("storing time:%d, %s:%s in %s" % ( message.get("ts"), self.field, row.get(self.field, "-"), filename)) self.update(filename, message.get("ts"), row.get(self.field))
get
identifier_name
__init__.py
import collections import os import abc import copy import datetime import logging import munge from future.utils import with_metaclass from vaping.config import parse_interval import vaping.io class PluginBase(vaping.io.Thread): """ Base plugin interface # Instanced Attributes - config (`dict`): plugin config - vaping: reference to the main vaping object Calls `self.init()` prefork while loading all modules, init() should not do anything active, any files opened may be closed when it forks. Plugins should prefer `init()` to `__init__()` to ensure the class is completely done initializing. Calls `self.on_start()` and `self.on_stop()` before and after running in case any connections need to be created or cleaned up. """ @property def groups(self): """ `dict` - group configurations keyed by name """ group_config = {} # legacy way of threating any dict as a potential # group config (pre #44 implementation) # supported until vaping 2.0 for k,v in list(self.config.items()): if isinstance(v, collections.Mapping): group_config[k] = v # explicit groups object (#44 implementation) for _group_config in self.config.get("groups",[]): group_config[_group_config["name"]] = _group_config return group_config def init(self): """ called after the plugin is initialized, plugin may define this for any other initialization code """ pass def on_start(self): """ called when the daemon is starting """ pass def on_stop(self): """ called when the daemon is stopping """ pass def new_message(self): """ creates and returns new message `dict`, setting `type`, `source`, `ts`, `data` `data` is initialized to an empty array **Returns** message (`dict`) """ msg = {} msg['data'] = [] msg['type'] = self.plugin_type msg['source'] = self.name msg['ts'] = (datetime.datetime.utcnow() - datetime.datetime(1970, 1, 1)).total_seconds() return msg def popen(self, args, **kwargs): """ creates a subprocess with passed args **Returns** Popen instance """ self.log.debug("popen %s", ' '.join(args)) return vaping.io.subprocess.Popen(args, **kwargs) @property def log(self): """ logger instance for plugin type """ if not self._logger: self._logger = logging.getLogger('vaping.plugins.' + self.plugin_type) return self._logger def __init__(self, config, ctx): """ **Arguments** - config (`dict`) - ctx: vaping context """ if hasattr(self, 'default_config'): self.config = munge.util.recursive_update(copy.deepcopy(self.default_config), copy.deepcopy(config)) else: self.config = config # set for pluginmgr self.pluginmgr_config = self.config self.vaping = ctx self.name = self.config.get("name") self._logger = None super(PluginBase, self).__init__() self.init() def _run(self): self.on_start() class ProbeBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for probe plugin, used for getting data expects method probe() to be defined """ def init(self): pass @abc.abstractmethod def probe(self): """ probe for data, return a list of dicts """ def __init__(self, config, ctx, emit=None): if emit: self._emit = [emit] else: self._emit = [] self._emit_queue = vaping.io.Queue() super(ProbeBase, self).__init__(config, ctx) def _run(self): super(ProbeBase, self)._run() self.run_level = 1 while self.run_level: self.send_emission() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") def queue_emission(self, msg): """ queue an emission of a message for all output plugins **Arguments** - msg (`dict`): dict containing `type`, `source`, `ts` and `data` keys """ if not msg: return for _emitter in self._emit: if not hasattr(_emitter, 'emit'): continue def emit(emitter=_emitter): self.log.debug("emit to {}".format(emitter.name)) emitter.emit(msg) self.log.debug("queue emission to {} ({})".format( _emitter.name, self._emit_queue.qsize())) self._emit_queue.put(emit) def send_emission(self): """ emit and remove the first emission in the queue """ if self._emit_queue.empty():
emit = self._emit_queue.get() emit() def emit_all(self): """ emit and remove all emissions in the queue """ while not self._emit_queue.empty(): self.send_emission() class TimedProbe(ProbeBase): """ Probe class that calls probe every config defined interval """ def __init__(self, config, ctx, emit=None): super(TimedProbe, self).__init__(config, ctx, emit) if 'interval' not in self.pluginmgr_config: raise ValueError('interval not set in config') self.interval = parse_interval(self.pluginmgr_config['interval']) self.run_level = 0 def _run(self): self.run_level = 1 while self.run_level: start = datetime.datetime.now() # since the TimedProbe will sleep between cycles # we need to emit all queued emissions each cycle self.emit_all() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") done = datetime.datetime.now() elapsed = done - start if elapsed.total_seconds() > self.interval: self.log.warning("probe time exceeded interval") else: sleeptime = datetime.timedelta(seconds=self.interval) - elapsed vaping.io.sleep(sleeptime.total_seconds()) class FileProbe(ProbeBase): """ Probes a file and emits everytime a new line is read # Config - path (`str`): path to file - backlog (`int=0`): number of bytes to read from backlog - max_lines (`int=1000`): maximum number of lines to read during probe # Instanced Attributes - path (`str`): path to file - backlog (`int`): number of bytes to read from backlog - max_lines (`int`): maximum number of liens to read during probe - fh (`filehandler`): file handler for opened file (only available if `path` is set) """ def __init__(self, config, ctx, emit=None): super(FileProbe, self).__init__(config, ctx, emit) self.path = self.pluginmgr_config.get("path") self.run_level = 0 self.backlog = int(self.pluginmgr_config.get("backlog",0)) self.max_lines = int(self.pluginmgr_config.get("max_lines",1000)) if self.path: self.fh = open(self.path, "r") self.fh.seek(0,2) if self.backlog: try: self.fh.seek(self.fh.tell() - self.backlog, os.SEEK_SET) except ValueError as exc: if str(exc).find("negative seek position") > -1: self.fh.seek(0) else: raise def _run(self): self.run_level = 1 while self.run_level: self.send_emission() for msg in self.probe(): self.queue_emission(msg) vaping.io.sleep(0.1) def validate_file_handler(self): """ Here we validate that our filehandler is pointing to an existing file. If it doesnt, because file has been deleted, we close the filehander and try to reopen """ if self.fh.closed: try: self.fh = open(self.path, "r") self.fh.seek(0, 2) except OSError as err: logging.error("Could not reopen file: {}".format(err)) return False open_stat = os.fstat(self.fh.fileno()) try: file_stat = os.stat(self.path) except OSError as err: logging.error("Could not stat file: {}".format(err)) return False if open_stat != file_stat: self.log self.fh.close() return False return True def probe(self): """ Probe the file for new lines """ # make sure the filehandler is still valid # (e.g. file stat hasnt changed, file exists etc.) if not self.validate_file_handler(): return [] messages = [] # read any new lines and push them onto the stack for line in self.fh.readlines(self.max_lines): data = {"path":self.path} msg = self.new_message() # process the line - this is where parsing happens parsed = self.process_line(line, data) if not parsed: continue data.update(parsed) # process the probe - this is where data assignment # happens data = self.process_probe(data) msg["data"] = [data] messages.append(msg) # process all new messages before returning them # for emission messages = self.process_messages(messages) return messages def process_line(self, line, data): """ override this - parse your line in here """ return data def process_probe(self, data): """ override this - assign your data values here """ return data def process_messages(self, messages): """ override this - process your messages before they are emitted """ return messages class EmitBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for emit plugins, used for sending data expects method emit() to be defined """ def __init__(self, config, ctx): super(EmitBase, self).__init__(config, ctx) @abc.abstractmethod def emit(self, message): """ accept message to emit """ class TimeSeriesDB(EmitBase): """ Base interface for timeseries db storage plugins # Config - filename (`str`): database file name template - field (`str`): fieeld name to read the value from # Instanced Attributes - filename (`str`): database file name template - field (`str`): fieeld name to read the value from """ def __init__(self, config, ctx): super(TimeSeriesDB, self).__init__(config, ctx) # filename template self.filename = self.config.get("filename") # field name to read the value from self.field = self.config.get("field") if not self.filename: raise ValueError("No filename specified") if not self.field: raise ValueError("No field specified, field should specify which value to store in the database") def create(self, filename): """ Create database **Arguments** - filename (`str`): database filename """ raise NotImplementedError() def update(self, filename, time, value): """ Update database **Arguments** - filename (`str`): database filename - time (`int`): epoch timestamp - value (`mixed`) """ raise NotImplementedError() def get(self, filename, from_time, to_time): """ Retrieve data from database for the specified timespan **Arguments** - filename (`str`): database filename - from_time (`int`): epoch timestamp start - to_time (`int`): epoch timestamp end """ raise NotImplementedError() def filename_formatters(self, data, row): """ Returns a dict containing the various filename formatter values Values are gotten from the vaping data message as well as the currently processed row in the message **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatter variables (`dict`) """ r = { "source" : data.get("source"), "field" : self.field, "type" : data.get("type") } r.update(**row) return r def format_filename(self, data, row): """ Returns a formatted filename using the template stored in self.filename **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatted version of self.filename (`str`) """ return self.filename.format(**self.filename_formatters(data, row)) def emit(self, message): """ emit to database **Arguments** - message (`dict`): vaping message dict """ # handle vaping data that arrives in a list if isinstance(message.get("data"), list): for row in message.get("data"): # format filename from data filename = self.format_filename(message, row) # create database file if it does not exist yet if not os.path.exists(filename): self.create(filename) # update database self.log.debug("storing time:%d, %s:%s in %s" % ( message.get("ts"), self.field, row.get(self.field, "-"), filename)) self.update(filename, message.get("ts"), row.get(self.field))
return
conditional_block
__init__.py
import collections import os import abc import copy import datetime import logging import munge from future.utils import with_metaclass from vaping.config import parse_interval import vaping.io class PluginBase(vaping.io.Thread): """ Base plugin interface # Instanced Attributes - config (`dict`): plugin config - vaping: reference to the main vaping object Calls `self.init()` prefork while loading all modules, init() should not do anything active, any files opened may be closed when it forks. Plugins should prefer `init()` to `__init__()` to ensure the class is completely done initializing. Calls `self.on_start()` and `self.on_stop()` before and after running in case any connections need to be created or cleaned up. """ @property def groups(self): """ `dict` - group configurations keyed by name """ group_config = {} # legacy way of threating any dict as a potential # group config (pre #44 implementation) # supported until vaping 2.0 for k,v in list(self.config.items()): if isinstance(v, collections.Mapping): group_config[k] = v # explicit groups object (#44 implementation) for _group_config in self.config.get("groups",[]): group_config[_group_config["name"]] = _group_config return group_config def init(self): """ called after the plugin is initialized, plugin may define this for any other initialization code """ pass def on_start(self): """ called when the daemon is starting """ pass def on_stop(self): """ called when the daemon is stopping """ pass def new_message(self): """ creates and returns new message `dict`, setting `type`, `source`, `ts`, `data` `data` is initialized to an empty array **Returns** message (`dict`) """ msg = {} msg['data'] = [] msg['type'] = self.plugin_type msg['source'] = self.name msg['ts'] = (datetime.datetime.utcnow() - datetime.datetime(1970, 1, 1)).total_seconds() return msg def popen(self, args, **kwargs): """ creates a subprocess with passed args **Returns** Popen instance """ self.log.debug("popen %s", ' '.join(args)) return vaping.io.subprocess.Popen(args, **kwargs) @property def log(self): """ logger instance for plugin type """ if not self._logger: self._logger = logging.getLogger('vaping.plugins.' + self.plugin_type) return self._logger def __init__(self, config, ctx): """ **Arguments** - config (`dict`) - ctx: vaping context """ if hasattr(self, 'default_config'): self.config = munge.util.recursive_update(copy.deepcopy(self.default_config), copy.deepcopy(config)) else: self.config = config # set for pluginmgr self.pluginmgr_config = self.config self.vaping = ctx self.name = self.config.get("name") self._logger = None super(PluginBase, self).__init__() self.init() def _run(self): self.on_start() class ProbeBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for probe plugin, used for getting data expects method probe() to be defined """ def init(self): pass @abc.abstractmethod def probe(self): """ probe for data, return a list of dicts """ def __init__(self, config, ctx, emit=None): if emit: self._emit = [emit] else: self._emit = [] self._emit_queue = vaping.io.Queue() super(ProbeBase, self).__init__(config, ctx) def _run(self): super(ProbeBase, self)._run() self.run_level = 1 while self.run_level: self.send_emission() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") def queue_emission(self, msg): """ queue an emission of a message for all output plugins **Arguments** - msg (`dict`): dict containing `type`, `source`, `ts` and `data` keys """ if not msg: return for _emitter in self._emit: if not hasattr(_emitter, 'emit'): continue def emit(emitter=_emitter): self.log.debug("emit to {}".format(emitter.name)) emitter.emit(msg) self.log.debug("queue emission to {} ({})".format( _emitter.name, self._emit_queue.qsize())) self._emit_queue.put(emit) def send_emission(self): """ emit and remove the first emission in the queue """ if self._emit_queue.empty(): return emit = self._emit_queue.get() emit() def emit_all(self): """ emit and remove all emissions in the queue """ while not self._emit_queue.empty(): self.send_emission() class TimedProbe(ProbeBase): """ Probe class that calls probe every config defined interval """ def __init__(self, config, ctx, emit=None): super(TimedProbe, self).__init__(config, ctx, emit) if 'interval' not in self.pluginmgr_config: raise ValueError('interval not set in config') self.interval = parse_interval(self.pluginmgr_config['interval']) self.run_level = 0 def _run(self): self.run_level = 1 while self.run_level: start = datetime.datetime.now() # since the TimedProbe will sleep between cycles # we need to emit all queued emissions each cycle self.emit_all() msg = self.probe() if msg: self.queue_emission(msg) else: self.log.debug("probe returned no data") done = datetime.datetime.now() elapsed = done - start if elapsed.total_seconds() > self.interval: self.log.warning("probe time exceeded interval") else: sleeptime = datetime.timedelta(seconds=self.interval) - elapsed vaping.io.sleep(sleeptime.total_seconds()) class FileProbe(ProbeBase): """ Probes a file and emits everytime a new line is read # Config - path (`str`): path to file - backlog (`int=0`): number of bytes to read from backlog - max_lines (`int=1000`): maximum number of lines to read during probe # Instanced Attributes - path (`str`): path to file - backlog (`int`): number of bytes to read from backlog - max_lines (`int`): maximum number of liens to read during probe - fh (`filehandler`): file handler for opened file (only available if `path` is set) """ def __init__(self, config, ctx, emit=None): super(FileProbe, self).__init__(config, ctx, emit) self.path = self.pluginmgr_config.get("path") self.run_level = 0 self.backlog = int(self.pluginmgr_config.get("backlog",0)) self.max_lines = int(self.pluginmgr_config.get("max_lines",1000)) if self.path: self.fh = open(self.path, "r") self.fh.seek(0,2) if self.backlog: try: self.fh.seek(self.fh.tell() - self.backlog, os.SEEK_SET) except ValueError as exc: if str(exc).find("negative seek position") > -1: self.fh.seek(0) else: raise def _run(self): self.run_level = 1 while self.run_level: self.send_emission() for msg in self.probe(): self.queue_emission(msg) vaping.io.sleep(0.1) def validate_file_handler(self): """ Here we validate that our filehandler is pointing to an existing file. If it doesnt, because file has been deleted, we close the filehander and try to reopen """ if self.fh.closed: try: self.fh = open(self.path, "r") self.fh.seek(0, 2) except OSError as err: logging.error("Could not reopen file: {}".format(err)) return False open_stat = os.fstat(self.fh.fileno()) try: file_stat = os.stat(self.path) except OSError as err: logging.error("Could not stat file: {}".format(err)) return False if open_stat != file_stat: self.log self.fh.close() return False return True def probe(self): """ Probe the file for new lines """ # make sure the filehandler is still valid # (e.g. file stat hasnt changed, file exists etc.) if not self.validate_file_handler(): return [] messages = [] # read any new lines and push them onto the stack for line in self.fh.readlines(self.max_lines): data = {"path":self.path} msg = self.new_message() # process the line - this is where parsing happens parsed = self.process_line(line, data) if not parsed: continue data.update(parsed) # process the probe - this is where data assignment # happens data = self.process_probe(data) msg["data"] = [data] messages.append(msg) # process all new messages before returning them # for emission messages = self.process_messages(messages) return messages def process_line(self, line, data): """ override this - parse your line in here """ return data def process_probe(self, data): """ override this - assign your data values here """ return data def process_messages(self, messages): """ override this - process your messages before they are emitted """ return messages class EmitBase(with_metaclass(abc.ABCMeta, PluginBase)): """ Base class for emit plugins, used for sending data expects method emit() to be defined """ def __init__(self, config, ctx):
@abc.abstractmethod def emit(self, message): """ accept message to emit """ class TimeSeriesDB(EmitBase): """ Base interface for timeseries db storage plugins # Config - filename (`str`): database file name template - field (`str`): fieeld name to read the value from # Instanced Attributes - filename (`str`): database file name template - field (`str`): fieeld name to read the value from """ def __init__(self, config, ctx): super(TimeSeriesDB, self).__init__(config, ctx) # filename template self.filename = self.config.get("filename") # field name to read the value from self.field = self.config.get("field") if not self.filename: raise ValueError("No filename specified") if not self.field: raise ValueError("No field specified, field should specify which value to store in the database") def create(self, filename): """ Create database **Arguments** - filename (`str`): database filename """ raise NotImplementedError() def update(self, filename, time, value): """ Update database **Arguments** - filename (`str`): database filename - time (`int`): epoch timestamp - value (`mixed`) """ raise NotImplementedError() def get(self, filename, from_time, to_time): """ Retrieve data from database for the specified timespan **Arguments** - filename (`str`): database filename - from_time (`int`): epoch timestamp start - to_time (`int`): epoch timestamp end """ raise NotImplementedError() def filename_formatters(self, data, row): """ Returns a dict containing the various filename formatter values Values are gotten from the vaping data message as well as the currently processed row in the message **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatter variables (`dict`) """ r = { "source" : data.get("source"), "field" : self.field, "type" : data.get("type") } r.update(**row) return r def format_filename(self, data, row): """ Returns a formatted filename using the template stored in self.filename **Arguments** - data (`dict`): vaping message - row (`dict`): vaping message data row **Returns** formatted version of self.filename (`str`) """ return self.filename.format(**self.filename_formatters(data, row)) def emit(self, message): """ emit to database **Arguments** - message (`dict`): vaping message dict """ # handle vaping data that arrives in a list if isinstance(message.get("data"), list): for row in message.get("data"): # format filename from data filename = self.format_filename(message, row) # create database file if it does not exist yet if not os.path.exists(filename): self.create(filename) # update database self.log.debug("storing time:%d, %s:%s in %s" % ( message.get("ts"), self.field, row.get(self.field, "-"), filename)) self.update(filename, message.get("ts"), row.get(self.field))
super(EmitBase, self).__init__(config, ctx)
identifier_body
integratedAnalysisPipeline.py
from generateIntegratedConfigurationFiles import generateIntAnalysisConfig import subprocess, os, shutil, sys from multiprocessing import Pool """ try: import syntenyFinal except: print'cannot find syntenyFinal analysis' exit() try: import circosFiguresPipeline except: print'cannot find circos analysis' exit() """ # gff2sort before completing entire analysis # name bed files folder BdFiles # PARSING FUNCTIONS def parseConfigFindList(stringFind,configFile): """parseConfigFindList inputs a particular string to find and read file after and a configuration file object outputs list of relevant filenames""" read = 0 listOfItems = [] for line in configFile: if line: if read == 1: if 'Stop' in line: configFile.seek(0) break # exit the function and return the list of files or list information listOfItems.append(line.strip('\n')) if stringFind in line: read = 1 # if find string specified, begin reading lines configFile.seek(0) return listOfItems def
(stringFind,configFile): """findPath will find path or value of associated specified string or info from config file""" for line in configFile: if stringFind in line: # if find string specified, return pathname or specific value trying to find configFile.seek(0) return line.split()[-1].strip('\n') configFile.seek(0) # open master configuration file open('masterConfig.txt','r').close() masterConfigFile = open('masterConfig.txt','r') # grab the following information from the configuration file weightsList = parseConfigFindList('Weights info',masterConfigFile) findInfoList = ['performSynteny','performCircos', 'performALLMAPS', 'querySpecies', 'NameAnalysis','writeFastaOut', 'Loci_Threshold', 'pathPython','pathSystem', 'pathALLMAPS', 'BdPath', 'pathUnOut', 'pathGFF', 'pathSort', 'BPsMergeDist', 'softMasked', 'genomePath', 'karyotypesFilesPath','circosConfigFilesPath', 'LinkPath', 'circosOutPath', 'BPsThreshold', 'multipleSeqAlignFastasPath','fastaOutputName', 'allMAPImageOutputPath', 'online','projectName', 'nerscUsername','nohup','cactusRun','cactusFolder'] # find the following information # list of query strings into config path finder for i in range(len(findInfoList)): # find the paths/info of above queries findInfoList[i] = parseConfigFindPath(findInfoList[i], masterConfigFile) # assign values (performSynteny, performCircos, performALLMAPS,querySpecies, NameAnalysis, writeFastaOut, Loci_Threshold, pathPython, pathSystem,pathALLMAPS, BdPath, pathUnOut, pathGFF, pathSort, BPsMergeDist , softmask, genomePath, karyotypesFilesPath, circosConfigFilesPath, LinkPath, circosOutPath, BPsThreshold, multipleSeqAlignFastasPath, fastaOutputName, allMAPImageOutputPath, online, projectName, nerscUsername, nohup, cactusRun,cactusFolder) = tuple(findInfoList) # for debugging, see if all of your data has passed through print tuple(findInfoList) # generate weights file for allmaps... open('%sweights.txt'%pathALLMAPS,'w').close() weightsFile = open('%sweights.txt'%pathALLMAPS,'w') for weight in weightsList: weightsFile.write('fake%s_'%querySpecies+weight+'\n') weightsFile.close() #fake473_283 1 # generate config files # first need to generate 3 syntenic files list and list of genome files for synteny analysis # second generate faiFiles list and bed files list circos # third generate bedfile list and fastainputname # get list of genome files, fai files, and fasta input filename for fragmented genome listGenomePathFiles=str(subprocess.Popen(['ls','%s'%genomePath], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') listGenomeFiles = [] listFaiFiles = [] # find lists of .fa and .fai files and format for config files for file in listGenomePathFiles: if file.endswith('.fa') or file.endswith('.fasta'): listGenomeFiles.append(file+'\n') listFaiFiles.append(file+'.fai\n') genomeFilesText = ''.join(file for file in listGenomeFiles) faiFilesText = ''.join(file for file in listFaiFiles) # if query species, use .fa file for genome reconstruction for filename in listGenomeFiles: if querySpecies in filename: fastaInputName = filename.strip('\n') # fasta filename of query species # list of unout files listPathUnout = str(subprocess.Popen(['ls', '%s' % pathUnOut], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') def gff2sort2(gff, pathgff, pathsort): """Takes a gffFiles and converts them to sort2 files to use in the final synteny analysis. Please let Joshua Levy know if there are any errors or have problems!""" outFileName = pathsort + gff[:gff.rfind('.')] + '.sort2' inputFile = open(pathgff + gff, 'r') open(outFileName, 'w').close() outputFile = open(outFileName, 'w') for line in inputFile: # grab gene info from each line if it's longest and mRNA strand and output to sort2 file if 'mRNA' in line and 'longest=1' in line: lineInList = line.split() parserList = lineInList[-1].split(';') lineOutputList = [parserList[1].replace('Name=',''), lineInList[0].replace('-', 'S'), lineInList[3], lineInList[4]] outputFile.write('%s %s %s %s\n' % tuple(lineOutputList)) inputFile.close() outputFile.close() # prints True if there are no sortfiles, generally want that for each analysis... for now, is possible to modify code to # be more versatile print not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()) # if no sort files present, generate them from gffs. PLEASE DELETE SORT FILES AFTER EVERY ANALYSIS if not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()): listGFFfiles = str(subprocess.Popen(['ls', '%s' % pathGFF], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') # turn gff into sort files for file in listGFFfiles: if file.endswith('.gff') or file.endswith('.gff3'): gff2sort2(file,pathGFF,pathSort) # find sort files listPathSort = str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') unoutList = [] sortFileList = [] bedList = [] # generate list of bed files for allmaps and circos, and unout and sort files for synteny analysis # unouts can also be used alternate allmaps for file in listPathUnout: if file.endswith('.unout'): bedList.append(file[:file.rfind('.')]+'.bed') unoutList.append(file) for file in listPathSort: if file.endswith('.sort2'): sortFileList.append(file) # make sure all sort files are included print sortFileList # Bedfile text for config files bedFilesText = ''.join(file+'\n' for file in bedList) # generate tuples for three syntenic files listSyntenicFilesTuples = [] for file in unoutList: # find target species then corresponding target sort files, add to list of syntenic files targetSpecies = file[file.find('-')+1:file.rfind('_')].replace('PAC2_0.','').replace('PAC4GC.','') print targetSpecies for sortFile in sortFileList: if querySpecies in sortFile: querySortFile = sortFile if targetSpecies in sortFile: targetSortFile = sortFile print targetSortFile listSyntenicFilesTuples.append((file,querySortFile,targetSortFile)) # text for synteny analysis config syntenicFilesText = ''.join('%s %s %s\n'%synFilesTuple for synFilesTuple in listSyntenicFilesTuples) # if on NERSC load proper modules, try to load these modules beforehand... print 'online' + online if int(online): try: subprocess.call('module load bedtools/2.25.0',shell=True) subprocess.call('module load circos',shell=True) except: print 'Unable to load online modules...' try: int(BPsMergeDist) except: BPsMergeDist = '100000' # write syntenic Config text generateIntAnalysisConfig('syntenyAnalysis',(NameAnalysis,writeFastaOut,Loci_Threshold,pathPython,pathUnOut,pathSort, BPsMergeDist,softmask, NameAnalysis,multipleSeqAlignFastasPath,syntenicFilesText,genomePath, genomeFilesText)) # if choosing to perform multiple synteny if int(performSynteny): # run synteny analysis #execfile(os.path.join(os.path.dirname(sys.argv[0]), 'syntenyFinal.py')) execfile('SyntenyFinal.py') # move BedFiles to Bd folder for bedFile in bedList: try: shutil.copy(BdPath[:BdPath.rfind('BdFiles')] + bedFile, BdPath) except: print 'Unable to copy bed file %s to BdFiles...'%bedFile # move genome to allmaps folder try: shutil.copy(genomePath+fastaInputName,pathALLMAPS+fastaInputName) except: print 'Unable to copy genome to allmaps directory...' # generate config files for relevant analyses if int(performCircos): print 'config circos' generateIntAnalysisConfig('circosAnalysis',(faiFilesText,bedFilesText,genomePath,karyotypesFilesPath,BdPath, circosConfigFilesPath,LinkPath,circosOutPath,pathPython,BPsThreshold)) if int(performALLMAPS): print 'config allmaps' generateIntAnalysisConfig('reconstructGenome',(pathALLMAPS,pathPython,pathSystem,BdPath,allMAPImageOutputPath, fastaInputName,fastaOutputName,bedFilesText)) # see if will be performing alternate genome reconstruction using syntenic genes, MORE ACCURATE try: masterConfigFile = open('masterConfig.txt','r') masterConfigFile.seek(0) performAltAllMaps = parseConfigFindPath('performAltAllMaps', masterConfigFile) masterConfigFile.close() if int(performAltAllMaps): print 'config alt allmaps' generateIntAnalysisConfig('reconstructGenome2', (pathALLMAPS, pathPython, pathSystem, BdPath, allMAPImageOutputPath, fastaInputName, fastaOutputName, pathUnOut,pathSort,Loci_Threshold)) open('%sweights.txt' % pathALLMAPS, 'w').close() weightsFile = open('%sweights.txt' % pathALLMAPS, 'w') for weight in weightsList: weightsFile.write('%s_' % querySpecies + weight + '\n') weightsFile.close() except: print 'Unable to set up alternate allmaps' performAltAllMaps = 0 # perform circos analysis but not reconstruction if int(performCircos) and not int(performALLMAPS): print 'circos' # try to run circos online if int(online): open('runCircos.sh', 'w').close() # writing shell script to run circos circos = open('runCircos.sh', 'w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() try: # try to run circos subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos via command line..' else: # try offline analysis try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' #except: # print 'Unable to run circos analysis.' # exit() # perform allmaps but not circos elif int(performALLMAPS) and not int(performCircos): print 'allmaps' #os.chdir(pathALLMAPS) # if online if int(online): # try to run allmaps allmap = open('runAllmaps.sh', 'w') allmap.write( '#!/bin/bash\ncd %s\npython createNewGenome.py'%pathALLMAPS) allmap.close() try: if int(nohup): # if using no hangup on a server subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # else submit the job to run on the cluster subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh', shell=True) except: print 'Unable to run allmaps via command line..' # else run in offline mode else: try: os.chdir(pathALLMAPS) execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() # perform both allmaps and circos elif int(performALLMAPS) and int(performCircos): print 'both' if int(online): open('runCircos.sh','w').close() circos = open('runCircos.sh','w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() # multiple command line scripts being written to run on the cluster, talk to NERSC about syntax lines = ['qsub -P %s -N circosAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runCircos.sh'), 'cd %s\n'%pathALLMAPS, 'qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh')] open('runAllmaps.sh', 'w').close() allmap = open('runAllmaps.sh', 'w') allmap.write('#!/bin/bash\n'+lines[1]+'python createNewGenome.py') allmap.close() try: # run circos on nohup subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos analysis.' try: #subprocess.call(lines[0],shell=True) if int(nohup): # run allmaps on nohup subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # try submission to cluster if does not work subprocess.call(lines[2],shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh',shell=True) except: print 'Unable to run allmaps via command line..' else: # offline mode, linux preferred """ def runALLMAPS(): os.chdir(pathALLMAPS) try: execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' pool = Pool(processes=2) try: pool.apply_async(execfile,'circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' pool.apply_async(runALLMAPS) pool.close() pool.join() """ try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' os.chdir(pathALLMAPS) try: print 'Warning: If online, monitor computer for no timeout...' execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() if int(performAltAllMaps): # if perform alternative ALLMAPS, RECOMMENDED over regular allmaps, similar to allmaps but # not circos code if int(online): allmap = open('runAllmaps2.sh', 'w') allmap.write( '#!/bin/bash\n\ncd %s\npython createNewGenome_vAlternate.py'%pathALLMAPS) allmap.close() try: if int(nohup): subprocess.call('nohup sh runAllmaps2.sh', shell=True) else: subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps2.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps2.sh', shell=True) except: print 'Unable to run alt allmaps from command line' else: try: os.chdir(pathALLMAPS) execfile('createNewGenome_vAlternate.py') except: print 'Unable to run allmaps analysis.' exit() try: if int(cactusRun): for file in os.listdir(multipleSeqAlignFastasPath): if file.endswith('fasta'): shutil.copy(file,cactusFolder+'FastaFiles') print 'Remember to have grassnewick file in place!!!' os.chdir(cactusFolder) subprocess.call('sh runcac.sh',shell = True) except: print 'Error running cactus...'
parseConfigFindPath
identifier_name
integratedAnalysisPipeline.py
from generateIntegratedConfigurationFiles import generateIntAnalysisConfig import subprocess, os, shutil, sys from multiprocessing import Pool """ try: import syntenyFinal except: print'cannot find syntenyFinal analysis' exit() try: import circosFiguresPipeline except: print'cannot find circos analysis' exit() """ # gff2sort before completing entire analysis # name bed files folder BdFiles # PARSING FUNCTIONS def parseConfigFindList(stringFind,configFile): """parseConfigFindList inputs a particular string to find and read file after and a configuration file object outputs list of relevant filenames""" read = 0 listOfItems = [] for line in configFile: if line: if read == 1: if 'Stop' in line: configFile.seek(0) break # exit the function and return the list of files or list information listOfItems.append(line.strip('\n')) if stringFind in line: read = 1 # if find string specified, begin reading lines configFile.seek(0) return listOfItems def parseConfigFindPath(stringFind,configFile): """findPath will find path or value of associated specified string or info from config file""" for line in configFile: if stringFind in line: # if find string specified, return pathname or specific value trying to find configFile.seek(0) return line.split()[-1].strip('\n') configFile.seek(0) # open master configuration file open('masterConfig.txt','r').close() masterConfigFile = open('masterConfig.txt','r') # grab the following information from the configuration file weightsList = parseConfigFindList('Weights info',masterConfigFile) findInfoList = ['performSynteny','performCircos', 'performALLMAPS', 'querySpecies', 'NameAnalysis','writeFastaOut', 'Loci_Threshold', 'pathPython','pathSystem', 'pathALLMAPS', 'BdPath', 'pathUnOut', 'pathGFF', 'pathSort', 'BPsMergeDist', 'softMasked', 'genomePath', 'karyotypesFilesPath','circosConfigFilesPath', 'LinkPath', 'circosOutPath', 'BPsThreshold', 'multipleSeqAlignFastasPath','fastaOutputName', 'allMAPImageOutputPath', 'online','projectName', 'nerscUsername','nohup','cactusRun','cactusFolder'] # find the following information # list of query strings into config path finder for i in range(len(findInfoList)): # find the paths/info of above queries findInfoList[i] = parseConfigFindPath(findInfoList[i], masterConfigFile) # assign values (performSynteny, performCircos, performALLMAPS,querySpecies, NameAnalysis, writeFastaOut, Loci_Threshold, pathPython, pathSystem,pathALLMAPS, BdPath, pathUnOut, pathGFF, pathSort, BPsMergeDist , softmask, genomePath, karyotypesFilesPath, circosConfigFilesPath, LinkPath, circosOutPath, BPsThreshold, multipleSeqAlignFastasPath, fastaOutputName, allMAPImageOutputPath, online, projectName, nerscUsername, nohup, cactusRun,cactusFolder) = tuple(findInfoList) # for debugging, see if all of your data has passed through print tuple(findInfoList) # generate weights file for allmaps... open('%sweights.txt'%pathALLMAPS,'w').close() weightsFile = open('%sweights.txt'%pathALLMAPS,'w') for weight in weightsList: weightsFile.write('fake%s_'%querySpecies+weight+'\n') weightsFile.close() #fake473_283 1 # generate config files # first need to generate 3 syntenic files list and list of genome files for synteny analysis # second generate faiFiles list and bed files list circos # third generate bedfile list and fastainputname # get list of genome files, fai files, and fasta input filename for fragmented genome listGenomePathFiles=str(subprocess.Popen(['ls','%s'%genomePath], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') listGenomeFiles = [] listFaiFiles = [] # find lists of .fa and .fai files and format for config files for file in listGenomePathFiles: if file.endswith('.fa') or file.endswith('.fasta'): listGenomeFiles.append(file+'\n') listFaiFiles.append(file+'.fai\n') genomeFilesText = ''.join(file for file in listGenomeFiles) faiFilesText = ''.join(file for file in listFaiFiles) # if query species, use .fa file for genome reconstruction for filename in listGenomeFiles: if querySpecies in filename: fastaInputName = filename.strip('\n') # fasta filename of query species # list of unout files listPathUnout = str(subprocess.Popen(['ls', '%s' % pathUnOut], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') def gff2sort2(gff, pathgff, pathsort): """Takes a gffFiles and converts them to sort2 files to use in the final synteny analysis. Please let Joshua Levy know if there are any errors or have problems!""" outFileName = pathsort + gff[:gff.rfind('.')] + '.sort2' inputFile = open(pathgff + gff, 'r') open(outFileName, 'w').close() outputFile = open(outFileName, 'w') for line in inputFile: # grab gene info from each line if it's longest and mRNA strand and output to sort2 file if 'mRNA' in line and 'longest=1' in line: lineInList = line.split() parserList = lineInList[-1].split(';') lineOutputList = [parserList[1].replace('Name=',''), lineInList[0].replace('-', 'S'), lineInList[3], lineInList[4]] outputFile.write('%s %s %s %s\n' % tuple(lineOutputList)) inputFile.close() outputFile.close() # prints True if there are no sortfiles, generally want that for each analysis... for now, is possible to modify code to # be more versatile print not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()) # if no sort files present, generate them from gffs. PLEASE DELETE SORT FILES AFTER EVERY ANALYSIS if not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()): listGFFfiles = str(subprocess.Popen(['ls', '%s' % pathGFF], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') # turn gff into sort files for file in listGFFfiles: if file.endswith('.gff') or file.endswith('.gff3'): gff2sort2(file,pathGFF,pathSort) # find sort files listPathSort = str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') unoutList = [] sortFileList = [] bedList = [] # generate list of bed files for allmaps and circos, and unout and sort files for synteny analysis # unouts can also be used alternate allmaps for file in listPathUnout: if file.endswith('.unout'): bedList.append(file[:file.rfind('.')]+'.bed') unoutList.append(file) for file in listPathSort: if file.endswith('.sort2'): sortFileList.append(file) # make sure all sort files are included print sortFileList # Bedfile text for config files bedFilesText = ''.join(file+'\n' for file in bedList) # generate tuples for three syntenic files listSyntenicFilesTuples = [] for file in unoutList: # find target species then corresponding target sort files, add to list of syntenic files targetSpecies = file[file.find('-')+1:file.rfind('_')].replace('PAC2_0.','').replace('PAC4GC.','') print targetSpecies for sortFile in sortFileList: if querySpecies in sortFile: querySortFile = sortFile if targetSpecies in sortFile: targetSortFile = sortFile print targetSortFile listSyntenicFilesTuples.append((file,querySortFile,targetSortFile)) # text for synteny analysis config syntenicFilesText = ''.join('%s %s %s\n'%synFilesTuple for synFilesTuple in listSyntenicFilesTuples) # if on NERSC load proper modules, try to load these modules beforehand... print 'online' + online if int(online): try: subprocess.call('module load bedtools/2.25.0',shell=True) subprocess.call('module load circos',shell=True) except: print 'Unable to load online modules...'
try: int(BPsMergeDist) except: BPsMergeDist = '100000' # write syntenic Config text generateIntAnalysisConfig('syntenyAnalysis',(NameAnalysis,writeFastaOut,Loci_Threshold,pathPython,pathUnOut,pathSort, BPsMergeDist,softmask, NameAnalysis,multipleSeqAlignFastasPath,syntenicFilesText,genomePath, genomeFilesText)) # if choosing to perform multiple synteny if int(performSynteny): # run synteny analysis #execfile(os.path.join(os.path.dirname(sys.argv[0]), 'syntenyFinal.py')) execfile('SyntenyFinal.py') # move BedFiles to Bd folder for bedFile in bedList: try: shutil.copy(BdPath[:BdPath.rfind('BdFiles')] + bedFile, BdPath) except: print 'Unable to copy bed file %s to BdFiles...'%bedFile # move genome to allmaps folder try: shutil.copy(genomePath+fastaInputName,pathALLMAPS+fastaInputName) except: print 'Unable to copy genome to allmaps directory...' # generate config files for relevant analyses if int(performCircos): print 'config circos' generateIntAnalysisConfig('circosAnalysis',(faiFilesText,bedFilesText,genomePath,karyotypesFilesPath,BdPath, circosConfigFilesPath,LinkPath,circosOutPath,pathPython,BPsThreshold)) if int(performALLMAPS): print 'config allmaps' generateIntAnalysisConfig('reconstructGenome',(pathALLMAPS,pathPython,pathSystem,BdPath,allMAPImageOutputPath, fastaInputName,fastaOutputName,bedFilesText)) # see if will be performing alternate genome reconstruction using syntenic genes, MORE ACCURATE try: masterConfigFile = open('masterConfig.txt','r') masterConfigFile.seek(0) performAltAllMaps = parseConfigFindPath('performAltAllMaps', masterConfigFile) masterConfigFile.close() if int(performAltAllMaps): print 'config alt allmaps' generateIntAnalysisConfig('reconstructGenome2', (pathALLMAPS, pathPython, pathSystem, BdPath, allMAPImageOutputPath, fastaInputName, fastaOutputName, pathUnOut,pathSort,Loci_Threshold)) open('%sweights.txt' % pathALLMAPS, 'w').close() weightsFile = open('%sweights.txt' % pathALLMAPS, 'w') for weight in weightsList: weightsFile.write('%s_' % querySpecies + weight + '\n') weightsFile.close() except: print 'Unable to set up alternate allmaps' performAltAllMaps = 0 # perform circos analysis but not reconstruction if int(performCircos) and not int(performALLMAPS): print 'circos' # try to run circos online if int(online): open('runCircos.sh', 'w').close() # writing shell script to run circos circos = open('runCircos.sh', 'w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() try: # try to run circos subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos via command line..' else: # try offline analysis try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' #except: # print 'Unable to run circos analysis.' # exit() # perform allmaps but not circos elif int(performALLMAPS) and not int(performCircos): print 'allmaps' #os.chdir(pathALLMAPS) # if online if int(online): # try to run allmaps allmap = open('runAllmaps.sh', 'w') allmap.write( '#!/bin/bash\ncd %s\npython createNewGenome.py'%pathALLMAPS) allmap.close() try: if int(nohup): # if using no hangup on a server subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # else submit the job to run on the cluster subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh', shell=True) except: print 'Unable to run allmaps via command line..' # else run in offline mode else: try: os.chdir(pathALLMAPS) execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() # perform both allmaps and circos elif int(performALLMAPS) and int(performCircos): print 'both' if int(online): open('runCircos.sh','w').close() circos = open('runCircos.sh','w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() # multiple command line scripts being written to run on the cluster, talk to NERSC about syntax lines = ['qsub -P %s -N circosAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runCircos.sh'), 'cd %s\n'%pathALLMAPS, 'qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh')] open('runAllmaps.sh', 'w').close() allmap = open('runAllmaps.sh', 'w') allmap.write('#!/bin/bash\n'+lines[1]+'python createNewGenome.py') allmap.close() try: # run circos on nohup subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos analysis.' try: #subprocess.call(lines[0],shell=True) if int(nohup): # run allmaps on nohup subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # try submission to cluster if does not work subprocess.call(lines[2],shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh',shell=True) except: print 'Unable to run allmaps via command line..' else: # offline mode, linux preferred """ def runALLMAPS(): os.chdir(pathALLMAPS) try: execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' pool = Pool(processes=2) try: pool.apply_async(execfile,'circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' pool.apply_async(runALLMAPS) pool.close() pool.join() """ try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' os.chdir(pathALLMAPS) try: print 'Warning: If online, monitor computer for no timeout...' execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() if int(performAltAllMaps): # if perform alternative ALLMAPS, RECOMMENDED over regular allmaps, similar to allmaps but # not circos code if int(online): allmap = open('runAllmaps2.sh', 'w') allmap.write( '#!/bin/bash\n\ncd %s\npython createNewGenome_vAlternate.py'%pathALLMAPS) allmap.close() try: if int(nohup): subprocess.call('nohup sh runAllmaps2.sh', shell=True) else: subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps2.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps2.sh', shell=True) except: print 'Unable to run alt allmaps from command line' else: try: os.chdir(pathALLMAPS) execfile('createNewGenome_vAlternate.py') except: print 'Unable to run allmaps analysis.' exit() try: if int(cactusRun): for file in os.listdir(multipleSeqAlignFastasPath): if file.endswith('fasta'): shutil.copy(file,cactusFolder+'FastaFiles') print 'Remember to have grassnewick file in place!!!' os.chdir(cactusFolder) subprocess.call('sh runcac.sh',shell = True) except: print 'Error running cactus...'
random_line_split
integratedAnalysisPipeline.py
from generateIntegratedConfigurationFiles import generateIntAnalysisConfig import subprocess, os, shutil, sys from multiprocessing import Pool """ try: import syntenyFinal except: print'cannot find syntenyFinal analysis' exit() try: import circosFiguresPipeline except: print'cannot find circos analysis' exit() """ # gff2sort before completing entire analysis # name bed files folder BdFiles # PARSING FUNCTIONS def parseConfigFindList(stringFind,configFile): """parseConfigFindList inputs a particular string to find and read file after and a configuration file object outputs list of relevant filenames""" read = 0 listOfItems = [] for line in configFile: if line: if read == 1: if 'Stop' in line: configFile.seek(0) break # exit the function and return the list of files or list information listOfItems.append(line.strip('\n')) if stringFind in line: read = 1 # if find string specified, begin reading lines configFile.seek(0) return listOfItems def parseConfigFindPath(stringFind,configFile): """findPath will find path or value of associated specified string or info from config file""" for line in configFile: if stringFind in line: # if find string specified, return pathname or specific value trying to find configFile.seek(0) return line.split()[-1].strip('\n') configFile.seek(0) # open master configuration file open('masterConfig.txt','r').close() masterConfigFile = open('masterConfig.txt','r') # grab the following information from the configuration file weightsList = parseConfigFindList('Weights info',masterConfigFile) findInfoList = ['performSynteny','performCircos', 'performALLMAPS', 'querySpecies', 'NameAnalysis','writeFastaOut', 'Loci_Threshold', 'pathPython','pathSystem', 'pathALLMAPS', 'BdPath', 'pathUnOut', 'pathGFF', 'pathSort', 'BPsMergeDist', 'softMasked', 'genomePath', 'karyotypesFilesPath','circosConfigFilesPath', 'LinkPath', 'circosOutPath', 'BPsThreshold', 'multipleSeqAlignFastasPath','fastaOutputName', 'allMAPImageOutputPath', 'online','projectName', 'nerscUsername','nohup','cactusRun','cactusFolder'] # find the following information # list of query strings into config path finder for i in range(len(findInfoList)): # find the paths/info of above queries findInfoList[i] = parseConfigFindPath(findInfoList[i], masterConfigFile) # assign values (performSynteny, performCircos, performALLMAPS,querySpecies, NameAnalysis, writeFastaOut, Loci_Threshold, pathPython, pathSystem,pathALLMAPS, BdPath, pathUnOut, pathGFF, pathSort, BPsMergeDist , softmask, genomePath, karyotypesFilesPath, circosConfigFilesPath, LinkPath, circosOutPath, BPsThreshold, multipleSeqAlignFastasPath, fastaOutputName, allMAPImageOutputPath, online, projectName, nerscUsername, nohup, cactusRun,cactusFolder) = tuple(findInfoList) # for debugging, see if all of your data has passed through print tuple(findInfoList) # generate weights file for allmaps... open('%sweights.txt'%pathALLMAPS,'w').close() weightsFile = open('%sweights.txt'%pathALLMAPS,'w') for weight in weightsList: weightsFile.write('fake%s_'%querySpecies+weight+'\n') weightsFile.close() #fake473_283 1 # generate config files # first need to generate 3 syntenic files list and list of genome files for synteny analysis # second generate faiFiles list and bed files list circos # third generate bedfile list and fastainputname # get list of genome files, fai files, and fasta input filename for fragmented genome listGenomePathFiles=str(subprocess.Popen(['ls','%s'%genomePath], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') listGenomeFiles = [] listFaiFiles = [] # find lists of .fa and .fai files and format for config files for file in listGenomePathFiles: if file.endswith('.fa') or file.endswith('.fasta'): listGenomeFiles.append(file+'\n') listFaiFiles.append(file+'.fai\n') genomeFilesText = ''.join(file for file in listGenomeFiles) faiFilesText = ''.join(file for file in listFaiFiles) # if query species, use .fa file for genome reconstruction for filename in listGenomeFiles: if querySpecies in filename: fastaInputName = filename.strip('\n') # fasta filename of query species # list of unout files listPathUnout = str(subprocess.Popen(['ls', '%s' % pathUnOut], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') def gff2sort2(gff, pathgff, pathsort): """Takes a gffFiles and converts them to sort2 files to use in the final synteny analysis. Please let Joshua Levy know if there are any errors or have problems!""" outFileName = pathsort + gff[:gff.rfind('.')] + '.sort2' inputFile = open(pathgff + gff, 'r') open(outFileName, 'w').close() outputFile = open(outFileName, 'w') for line in inputFile: # grab gene info from each line if it's longest and mRNA strand and output to sort2 file if 'mRNA' in line and 'longest=1' in line: lineInList = line.split() parserList = lineInList[-1].split(';') lineOutputList = [parserList[1].replace('Name=',''), lineInList[0].replace('-', 'S'), lineInList[3], lineInList[4]] outputFile.write('%s %s %s %s\n' % tuple(lineOutputList)) inputFile.close() outputFile.close() # prints True if there are no sortfiles, generally want that for each analysis... for now, is possible to modify code to # be more versatile print not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()) # if no sort files present, generate them from gffs. PLEASE DELETE SORT FILES AFTER EVERY ANALYSIS if not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()): listGFFfiles = str(subprocess.Popen(['ls', '%s' % pathGFF], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') # turn gff into sort files for file in listGFFfiles: if file.endswith('.gff') or file.endswith('.gff3'): gff2sort2(file,pathGFF,pathSort) # find sort files listPathSort = str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') unoutList = [] sortFileList = [] bedList = [] # generate list of bed files for allmaps and circos, and unout and sort files for synteny analysis # unouts can also be used alternate allmaps for file in listPathUnout: if file.endswith('.unout'): bedList.append(file[:file.rfind('.')]+'.bed') unoutList.append(file) for file in listPathSort: if file.endswith('.sort2'): sortFileList.append(file) # make sure all sort files are included print sortFileList # Bedfile text for config files bedFilesText = ''.join(file+'\n' for file in bedList) # generate tuples for three syntenic files listSyntenicFilesTuples = [] for file in unoutList: # find target species then corresponding target sort files, add to list of syntenic files targetSpecies = file[file.find('-')+1:file.rfind('_')].replace('PAC2_0.','').replace('PAC4GC.','') print targetSpecies for sortFile in sortFileList: if querySpecies in sortFile: querySortFile = sortFile if targetSpecies in sortFile: targetSortFile = sortFile print targetSortFile listSyntenicFilesTuples.append((file,querySortFile,targetSortFile)) # text for synteny analysis config syntenicFilesText = ''.join('%s %s %s\n'%synFilesTuple for synFilesTuple in listSyntenicFilesTuples) # if on NERSC load proper modules, try to load these modules beforehand... print 'online' + online if int(online): try: subprocess.call('module load bedtools/2.25.0',shell=True) subprocess.call('module load circos',shell=True) except: print 'Unable to load online modules...' try: int(BPsMergeDist) except: BPsMergeDist = '100000' # write syntenic Config text generateIntAnalysisConfig('syntenyAnalysis',(NameAnalysis,writeFastaOut,Loci_Threshold,pathPython,pathUnOut,pathSort, BPsMergeDist,softmask, NameAnalysis,multipleSeqAlignFastasPath,syntenicFilesText,genomePath, genomeFilesText)) # if choosing to perform multiple synteny if int(performSynteny): # run synteny analysis #execfile(os.path.join(os.path.dirname(sys.argv[0]), 'syntenyFinal.py')) execfile('SyntenyFinal.py') # move BedFiles to Bd folder for bedFile in bedList: try: shutil.copy(BdPath[:BdPath.rfind('BdFiles')] + bedFile, BdPath) except: print 'Unable to copy bed file %s to BdFiles...'%bedFile # move genome to allmaps folder try: shutil.copy(genomePath+fastaInputName,pathALLMAPS+fastaInputName) except: print 'Unable to copy genome to allmaps directory...' # generate config files for relevant analyses if int(performCircos): print 'config circos' generateIntAnalysisConfig('circosAnalysis',(faiFilesText,bedFilesText,genomePath,karyotypesFilesPath,BdPath, circosConfigFilesPath,LinkPath,circosOutPath,pathPython,BPsThreshold)) if int(performALLMAPS): print 'config allmaps' generateIntAnalysisConfig('reconstructGenome',(pathALLMAPS,pathPython,pathSystem,BdPath,allMAPImageOutputPath, fastaInputName,fastaOutputName,bedFilesText)) # see if will be performing alternate genome reconstruction using syntenic genes, MORE ACCURATE try: masterConfigFile = open('masterConfig.txt','r') masterConfigFile.seek(0) performAltAllMaps = parseConfigFindPath('performAltAllMaps', masterConfigFile) masterConfigFile.close() if int(performAltAllMaps): print 'config alt allmaps' generateIntAnalysisConfig('reconstructGenome2', (pathALLMAPS, pathPython, pathSystem, BdPath, allMAPImageOutputPath, fastaInputName, fastaOutputName, pathUnOut,pathSort,Loci_Threshold)) open('%sweights.txt' % pathALLMAPS, 'w').close() weightsFile = open('%sweights.txt' % pathALLMAPS, 'w') for weight in weightsList: weightsFile.write('%s_' % querySpecies + weight + '\n') weightsFile.close() except: print 'Unable to set up alternate allmaps' performAltAllMaps = 0 # perform circos analysis but not reconstruction if int(performCircos) and not int(performALLMAPS):
# perform allmaps but not circos elif int(performALLMAPS) and not int(performCircos): print 'allmaps' #os.chdir(pathALLMAPS) # if online if int(online): # try to run allmaps allmap = open('runAllmaps.sh', 'w') allmap.write( '#!/bin/bash\ncd %s\npython createNewGenome.py'%pathALLMAPS) allmap.close() try: if int(nohup): # if using no hangup on a server subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # else submit the job to run on the cluster subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh', shell=True) except: print 'Unable to run allmaps via command line..' # else run in offline mode else: try: os.chdir(pathALLMAPS) execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() # perform both allmaps and circos elif int(performALLMAPS) and int(performCircos): print 'both' if int(online): open('runCircos.sh','w').close() circos = open('runCircos.sh','w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() # multiple command line scripts being written to run on the cluster, talk to NERSC about syntax lines = ['qsub -P %s -N circosAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runCircos.sh'), 'cd %s\n'%pathALLMAPS, 'qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh')] open('runAllmaps.sh', 'w').close() allmap = open('runAllmaps.sh', 'w') allmap.write('#!/bin/bash\n'+lines[1]+'python createNewGenome.py') allmap.close() try: # run circos on nohup subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos analysis.' try: #subprocess.call(lines[0],shell=True) if int(nohup): # run allmaps on nohup subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # try submission to cluster if does not work subprocess.call(lines[2],shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh',shell=True) except: print 'Unable to run allmaps via command line..' else: # offline mode, linux preferred """ def runALLMAPS(): os.chdir(pathALLMAPS) try: execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' pool = Pool(processes=2) try: pool.apply_async(execfile,'circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' pool.apply_async(runALLMAPS) pool.close() pool.join() """ try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' os.chdir(pathALLMAPS) try: print 'Warning: If online, monitor computer for no timeout...' execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() if int(performAltAllMaps): # if perform alternative ALLMAPS, RECOMMENDED over regular allmaps, similar to allmaps but # not circos code if int(online): allmap = open('runAllmaps2.sh', 'w') allmap.write( '#!/bin/bash\n\ncd %s\npython createNewGenome_vAlternate.py'%pathALLMAPS) allmap.close() try: if int(nohup): subprocess.call('nohup sh runAllmaps2.sh', shell=True) else: subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps2.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps2.sh', shell=True) except: print 'Unable to run alt allmaps from command line' else: try: os.chdir(pathALLMAPS) execfile('createNewGenome_vAlternate.py') except: print 'Unable to run allmaps analysis.' exit() try: if int(cactusRun): for file in os.listdir(multipleSeqAlignFastasPath): if file.endswith('fasta'): shutil.copy(file,cactusFolder+'FastaFiles') print 'Remember to have grassnewick file in place!!!' os.chdir(cactusFolder) subprocess.call('sh runcac.sh',shell = True) except: print 'Error running cactus...'
print 'circos' # try to run circos online if int(online): open('runCircos.sh', 'w').close() # writing shell script to run circos circos = open('runCircos.sh', 'w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() try: # try to run circos subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos via command line..' else: # try offline analysis try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' #except: # print 'Unable to run circos analysis.' # exit()
conditional_block
integratedAnalysisPipeline.py
from generateIntegratedConfigurationFiles import generateIntAnalysisConfig import subprocess, os, shutil, sys from multiprocessing import Pool """ try: import syntenyFinal except: print'cannot find syntenyFinal analysis' exit() try: import circosFiguresPipeline except: print'cannot find circos analysis' exit() """ # gff2sort before completing entire analysis # name bed files folder BdFiles # PARSING FUNCTIONS def parseConfigFindList(stringFind,configFile):
def parseConfigFindPath(stringFind,configFile): """findPath will find path or value of associated specified string or info from config file""" for line in configFile: if stringFind in line: # if find string specified, return pathname or specific value trying to find configFile.seek(0) return line.split()[-1].strip('\n') configFile.seek(0) # open master configuration file open('masterConfig.txt','r').close() masterConfigFile = open('masterConfig.txt','r') # grab the following information from the configuration file weightsList = parseConfigFindList('Weights info',masterConfigFile) findInfoList = ['performSynteny','performCircos', 'performALLMAPS', 'querySpecies', 'NameAnalysis','writeFastaOut', 'Loci_Threshold', 'pathPython','pathSystem', 'pathALLMAPS', 'BdPath', 'pathUnOut', 'pathGFF', 'pathSort', 'BPsMergeDist', 'softMasked', 'genomePath', 'karyotypesFilesPath','circosConfigFilesPath', 'LinkPath', 'circosOutPath', 'BPsThreshold', 'multipleSeqAlignFastasPath','fastaOutputName', 'allMAPImageOutputPath', 'online','projectName', 'nerscUsername','nohup','cactusRun','cactusFolder'] # find the following information # list of query strings into config path finder for i in range(len(findInfoList)): # find the paths/info of above queries findInfoList[i] = parseConfigFindPath(findInfoList[i], masterConfigFile) # assign values (performSynteny, performCircos, performALLMAPS,querySpecies, NameAnalysis, writeFastaOut, Loci_Threshold, pathPython, pathSystem,pathALLMAPS, BdPath, pathUnOut, pathGFF, pathSort, BPsMergeDist , softmask, genomePath, karyotypesFilesPath, circosConfigFilesPath, LinkPath, circosOutPath, BPsThreshold, multipleSeqAlignFastasPath, fastaOutputName, allMAPImageOutputPath, online, projectName, nerscUsername, nohup, cactusRun,cactusFolder) = tuple(findInfoList) # for debugging, see if all of your data has passed through print tuple(findInfoList) # generate weights file for allmaps... open('%sweights.txt'%pathALLMAPS,'w').close() weightsFile = open('%sweights.txt'%pathALLMAPS,'w') for weight in weightsList: weightsFile.write('fake%s_'%querySpecies+weight+'\n') weightsFile.close() #fake473_283 1 # generate config files # first need to generate 3 syntenic files list and list of genome files for synteny analysis # second generate faiFiles list and bed files list circos # third generate bedfile list and fastainputname # get list of genome files, fai files, and fasta input filename for fragmented genome listGenomePathFiles=str(subprocess.Popen(['ls','%s'%genomePath], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') listGenomeFiles = [] listFaiFiles = [] # find lists of .fa and .fai files and format for config files for file in listGenomePathFiles: if file.endswith('.fa') or file.endswith('.fasta'): listGenomeFiles.append(file+'\n') listFaiFiles.append(file+'.fai\n') genomeFilesText = ''.join(file for file in listGenomeFiles) faiFilesText = ''.join(file for file in listFaiFiles) # if query species, use .fa file for genome reconstruction for filename in listGenomeFiles: if querySpecies in filename: fastaInputName = filename.strip('\n') # fasta filename of query species # list of unout files listPathUnout = str(subprocess.Popen(['ls', '%s' % pathUnOut], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') def gff2sort2(gff, pathgff, pathsort): """Takes a gffFiles and converts them to sort2 files to use in the final synteny analysis. Please let Joshua Levy know if there are any errors or have problems!""" outFileName = pathsort + gff[:gff.rfind('.')] + '.sort2' inputFile = open(pathgff + gff, 'r') open(outFileName, 'w').close() outputFile = open(outFileName, 'w') for line in inputFile: # grab gene info from each line if it's longest and mRNA strand and output to sort2 file if 'mRNA' in line and 'longest=1' in line: lineInList = line.split() parserList = lineInList[-1].split(';') lineOutputList = [parserList[1].replace('Name=',''), lineInList[0].replace('-', 'S'), lineInList[3], lineInList[4]] outputFile.write('%s %s %s %s\n' % tuple(lineOutputList)) inputFile.close() outputFile.close() # prints True if there are no sortfiles, generally want that for each analysis... for now, is possible to modify code to # be more versatile print not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()) # if no sort files present, generate them from gffs. PLEASE DELETE SORT FILES AFTER EVERY ANALYSIS if not str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()): listGFFfiles = str(subprocess.Popen(['ls', '%s' % pathGFF], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') # turn gff into sort files for file in listGFFfiles: if file.endswith('.gff') or file.endswith('.gff3'): gff2sort2(file,pathGFF,pathSort) # find sort files listPathSort = str(subprocess.Popen(['ls', '%s' % pathSort], stdout=subprocess.PIPE, stderr=subprocess.PIPE) .stdout.read()).split('\n') unoutList = [] sortFileList = [] bedList = [] # generate list of bed files for allmaps and circos, and unout and sort files for synteny analysis # unouts can also be used alternate allmaps for file in listPathUnout: if file.endswith('.unout'): bedList.append(file[:file.rfind('.')]+'.bed') unoutList.append(file) for file in listPathSort: if file.endswith('.sort2'): sortFileList.append(file) # make sure all sort files are included print sortFileList # Bedfile text for config files bedFilesText = ''.join(file+'\n' for file in bedList) # generate tuples for three syntenic files listSyntenicFilesTuples = [] for file in unoutList: # find target species then corresponding target sort files, add to list of syntenic files targetSpecies = file[file.find('-')+1:file.rfind('_')].replace('PAC2_0.','').replace('PAC4GC.','') print targetSpecies for sortFile in sortFileList: if querySpecies in sortFile: querySortFile = sortFile if targetSpecies in sortFile: targetSortFile = sortFile print targetSortFile listSyntenicFilesTuples.append((file,querySortFile,targetSortFile)) # text for synteny analysis config syntenicFilesText = ''.join('%s %s %s\n'%synFilesTuple for synFilesTuple in listSyntenicFilesTuples) # if on NERSC load proper modules, try to load these modules beforehand... print 'online' + online if int(online): try: subprocess.call('module load bedtools/2.25.0',shell=True) subprocess.call('module load circos',shell=True) except: print 'Unable to load online modules...' try: int(BPsMergeDist) except: BPsMergeDist = '100000' # write syntenic Config text generateIntAnalysisConfig('syntenyAnalysis',(NameAnalysis,writeFastaOut,Loci_Threshold,pathPython,pathUnOut,pathSort, BPsMergeDist,softmask, NameAnalysis,multipleSeqAlignFastasPath,syntenicFilesText,genomePath, genomeFilesText)) # if choosing to perform multiple synteny if int(performSynteny): # run synteny analysis #execfile(os.path.join(os.path.dirname(sys.argv[0]), 'syntenyFinal.py')) execfile('SyntenyFinal.py') # move BedFiles to Bd folder for bedFile in bedList: try: shutil.copy(BdPath[:BdPath.rfind('BdFiles')] + bedFile, BdPath) except: print 'Unable to copy bed file %s to BdFiles...'%bedFile # move genome to allmaps folder try: shutil.copy(genomePath+fastaInputName,pathALLMAPS+fastaInputName) except: print 'Unable to copy genome to allmaps directory...' # generate config files for relevant analyses if int(performCircos): print 'config circos' generateIntAnalysisConfig('circosAnalysis',(faiFilesText,bedFilesText,genomePath,karyotypesFilesPath,BdPath, circosConfigFilesPath,LinkPath,circosOutPath,pathPython,BPsThreshold)) if int(performALLMAPS): print 'config allmaps' generateIntAnalysisConfig('reconstructGenome',(pathALLMAPS,pathPython,pathSystem,BdPath,allMAPImageOutputPath, fastaInputName,fastaOutputName,bedFilesText)) # see if will be performing alternate genome reconstruction using syntenic genes, MORE ACCURATE try: masterConfigFile = open('masterConfig.txt','r') masterConfigFile.seek(0) performAltAllMaps = parseConfigFindPath('performAltAllMaps', masterConfigFile) masterConfigFile.close() if int(performAltAllMaps): print 'config alt allmaps' generateIntAnalysisConfig('reconstructGenome2', (pathALLMAPS, pathPython, pathSystem, BdPath, allMAPImageOutputPath, fastaInputName, fastaOutputName, pathUnOut,pathSort,Loci_Threshold)) open('%sweights.txt' % pathALLMAPS, 'w').close() weightsFile = open('%sweights.txt' % pathALLMAPS, 'w') for weight in weightsList: weightsFile.write('%s_' % querySpecies + weight + '\n') weightsFile.close() except: print 'Unable to set up alternate allmaps' performAltAllMaps = 0 # perform circos analysis but not reconstruction if int(performCircos) and not int(performALLMAPS): print 'circos' # try to run circos online if int(online): open('runCircos.sh', 'w').close() # writing shell script to run circos circos = open('runCircos.sh', 'w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() try: # try to run circos subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos via command line..' else: # try offline analysis try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' #except: # print 'Unable to run circos analysis.' # exit() # perform allmaps but not circos elif int(performALLMAPS) and not int(performCircos): print 'allmaps' #os.chdir(pathALLMAPS) # if online if int(online): # try to run allmaps allmap = open('runAllmaps.sh', 'w') allmap.write( '#!/bin/bash\ncd %s\npython createNewGenome.py'%pathALLMAPS) allmap.close() try: if int(nohup): # if using no hangup on a server subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # else submit the job to run on the cluster subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh', shell=True) except: print 'Unable to run allmaps via command line..' # else run in offline mode else: try: os.chdir(pathALLMAPS) execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() # perform both allmaps and circos elif int(performALLMAPS) and int(performCircos): print 'both' if int(online): open('runCircos.sh','w').close() circos = open('runCircos.sh','w') circos.write('#!/bin/bash\npython circosFiguresPipeline.py') circos.close() # multiple command line scripts being written to run on the cluster, talk to NERSC about syntax lines = ['qsub -P %s -N circosAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runCircos.sh'), 'cd %s\n'%pathALLMAPS, 'qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps.sh')] open('runAllmaps.sh', 'w').close() allmap = open('runAllmaps.sh', 'w') allmap.write('#!/bin/bash\n'+lines[1]+'python createNewGenome.py') allmap.close() try: # run circos on nohup subprocess.call('nohup sh runCircos.sh', shell=True) except: print 'Unable to run circos analysis.' try: #subprocess.call(lines[0],shell=True) if int(nohup): # run allmaps on nohup subprocess.call('nohup sh runAllmaps.sh', shell=True) else: # try submission to cluster if does not work subprocess.call(lines[2],shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps.sh',shell=True) except: print 'Unable to run allmaps via command line..' else: # offline mode, linux preferred """ def runALLMAPS(): os.chdir(pathALLMAPS) try: execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' pool = Pool(processes=2) try: pool.apply_async(execfile,'circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' pool.apply_async(runALLMAPS) pool.close() pool.join() """ try: execfile('circosFiguresPipeline.py') except: print 'Unable to run circos analysis.' os.chdir(pathALLMAPS) try: print 'Warning: If online, monitor computer for no timeout...' execfile('createNewGenome.py') except: print 'Unable to run allmaps analysis.' exit() if int(performAltAllMaps): # if perform alternative ALLMAPS, RECOMMENDED over regular allmaps, similar to allmaps but # not circos code if int(online): allmap = open('runAllmaps2.sh', 'w') allmap.write( '#!/bin/bash\n\ncd %s\npython createNewGenome_vAlternate.py'%pathALLMAPS) allmap.close() try: if int(nohup): subprocess.call('nohup sh runAllmaps2.sh', shell=True) else: subprocess.call('qsub -P %s -N allmapsAnalyze -cwd -b yes -now no -j yes -m abes -M %s@lbl.gov -w e' ' -l exclusive.c %s%s\n'%(projectName,BdPath[:BdPath.rfind('BdFiles')],nerscUsername,'runAllmaps2.sh'), shell=True) except: print 'Unable to run allmaps via qsub. Set online to 0 and try again. Trying to run via command line..' try: subprocess.call('nohup sh runAllmaps2.sh', shell=True) except: print 'Unable to run alt allmaps from command line' else: try: os.chdir(pathALLMAPS) execfile('createNewGenome_vAlternate.py') except: print 'Unable to run allmaps analysis.' exit() try: if int(cactusRun): for file in os.listdir(multipleSeqAlignFastasPath): if file.endswith('fasta'): shutil.copy(file,cactusFolder+'FastaFiles') print 'Remember to have grassnewick file in place!!!' os.chdir(cactusFolder) subprocess.call('sh runcac.sh',shell = True) except: print 'Error running cactus...'
"""parseConfigFindList inputs a particular string to find and read file after and a configuration file object outputs list of relevant filenames""" read = 0 listOfItems = [] for line in configFile: if line: if read == 1: if 'Stop' in line: configFile.seek(0) break # exit the function and return the list of files or list information listOfItems.append(line.strip('\n')) if stringFind in line: read = 1 # if find string specified, begin reading lines configFile.seek(0) return listOfItems
identifier_body
vector_test.go
// Author: slowpoke <proxypoke at lavabit dot com> // Repository: https://gist.github.com/proxypoke/vector // // This program is free software under the non-terms // of the Anti-License. Do whatever the fuck you want. package vector import ( "math" "math/rand" "testing" ) // ========================== [ Constructor Tests ] =========================== // Creates vectors with dimension from 0 to 99, checks if they actually have // that dimension, then checks if the values are correctly initialized to 0. func TestNew(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { // XXX: If the Get method errors, this test will still pass. This // is because Get() would then return an uninitialized float64 for // val, which is 0 and therefore what the test expects. val, _ := v.Get(j) if val != 0 { t.Error("Newly initialized vector has a value != 0.") } } } } // Creates vectors with randomized slices, then checks whether they have the // correct dimension (len(slice)) and whether they have been correctly // initialized. func
(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { randslice := makeRandSlice(i) v := NewFrom(randslice) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { val, _ := v.Get(j) if val != randslice[j] { t.Error( "Wrong values in vector initialized from a random slice.") } } } } // Creates pseudo-random vectors with various dimensions, copies them and // verifies that the new vector is equal. func TestCopy(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) w := v.Copy() if !Equal(v, w) { t.Error("Copied vector is not equal to source vector.") } } } // =================== [ General Methods/Functions Tests ] ==================== // Creates pseudo-random vectors with various dimensions, then check if Get() // returns the correct values and errors on out-of-range indexes. func TestGet(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) for j, val := range v.dims { getval, err := v.Get(uint(j)) if err != nil { t.Error("Get() errored on a correct index.") } if val != getval { t.Error("Get() returned a wrong value.") } } _, err := v.Get(v.Dim()) if err == nil { t.Error("Get didn't error on an out-of-range index.") } } } // Creates uninitialized vectors of various dimensions, then sets their values // to pseudo-random values. It then compares those values to check if they // were set correctly. Also verifies is Set() correctly errors on out-of-range // indexes. func TestSet(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) for j = 0; j < i; j++ { val := rand.ExpFloat64() err := v.Set(j, val) if err != nil { t.Error("Set() errored on a correct index.") } if v.dims[j] != val { t.Error("Set didn't correctly set a value.") } } err := v.Set(v.Dim(), 0) if err == nil { t.Error("Set didn't error on an out-of-range index.") } } } // Creates a vector with known length, then compares the expected value with // what Len() returns. func TestLen(t *testing.T) { v := New(1) v.Set(0, 2) // has length 2 if v.Len() != 2 { t.Error("Len returned a wrong length") } } // Creates Vectors which are known to be (un)equal, then verifies that Equal() // has correct oytput. func TestEqual(t *testing.T) { slc := make([]float64, 10) for i := range slc { slc[i] = float64(i) } v := NewFrom(slc) w := NewFrom(slc) if !Equal(v, w) { t.Error("Equal() != true for equal vectors.") } w = New(10) if Equal(v, w) { t.Error("Equal() == true for unequal vectors.") } } // =========================== [ Operation Tests ] ============================ // Creates pesudo-random vectors, then adds them first as a non-destructive, // then as an in-place operations, checking if both operation were correct. func TestAdd(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Add(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error("Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place addition. c = a.Copy() c.Add(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error( "In-place Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Same as TestAdd, but with substraction. Heck, it's basically the same code. func TestSubstract(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Substract(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error("Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place sybstraction c = a.Copy() c.Substract(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error( "In-place Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Creates pseudo-random vectors, does scalar multiplication with pseudo-random // floats, then checks if the result is correct. It checks both the in-place // and the non-destructive version. func TestScale(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { a := makeRandomVector(i) x := rand.ExpFloat64() b := Scale(a, x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } // Test in-place scalar multiplication b = a.Copy() b.Scale(x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("In-place Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } } } // Creates pseudo-random vectors, normalizes them both in-place and // non-destructive, and verifies that the result is correct. func TestNormalize(t *testing.T) { var i uint // It makes no sense to normalize a zero vector, therefore we start at 1. for i = 1; i < 100; i++ { a := makeRandomVector(i) b := Normalize(a) if b.Len() != float64(1) { t.Error("Normalization failed, vector doesn't have length 1.") t.Logf("%f != 1", b.Len()) } } } // Uses vectors with known angles to calculate their DotProduct, then verifies // if the result is correct. func TestDotProduct(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel. a.Set(0, 1) b.Set(0, 1) dot, _ := DotProduct(a, b) if dot != 1 { t.Error("Dot Product of parallel vectors isn't 1.") } // Set the vectors as orthogonal. b = New(2) b.Set(1, 1) dot, _ = DotProduct(a, b) if dot != 0 { t.Error("Dot Product of orthogonal vectors isn't 0.") } // Set the vectors as anti-parallel. b = New(2) b.Set(0, -1) dot, _ = DotProduct(a, b) if dot != -1 { t.Error("Dot Product of anti-parallel vectors isn't -1.") } } // Uses vectors with known angles to verify that Angle() is correct. func TestAngle(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel (Θ == 0). a.Set(0, 1) b.Set(0, 1) Θ, _ := Angle(a, b) if Θ != 0 { t.Error("Angle between parallel vectors isn't 0.") t.Logf("%f != 0", Θ) } // Set the vectors as orthogonal (Θ == 0.5π). b = New(2) b.Set(1, 1) Θ, _ = Angle(a, b) if Θ != 0.5*math.Pi { t.Error("Angle between orthonal vectors isn't 0.5π.") t.Logf("%f != %f", Θ, 0.5*math.Pi) } // Set the vectors as anti-parallel (Θ == π). b = New(2) b.Set(0, -1) Θ, _ = Angle(a, b) if Θ != math.Pi { t.Error("Angle between anti-parallel vectors isn't π.") t.Logf("%f != %f", Θ, math.Pi) } } // Calculates the cross product of two pseudo-random vectors, then checks if // the resulting vector is orthogonal to both the original vectors. Tests both // in-place and non-destructive versions of the operation. func TestCrossProduct(t *testing.T) { check := func(a, b, c *Vector) { dot_a, _ := DotProduct(a, c) dot_b, _ := DotProduct(b, c) ε := 0.0000000005 if math.Abs(0-dot_a) < ε { dot_a = 0 } if math.Abs(0-dot_b) < ε { dot_b = 0 } if dot_a != 0 || dot_b != 0 { t.Error("Either or both vectors aren't orthogonal", "to their Cross Product.") t.Logf("a * c = %f", dot_a) t.Logf("b * c = %f", dot_b) } } a := makeRandomVector(3) b := makeRandomVector(3) c, _ := CrossProduct(a, b) check(a, b, c) // Check in-place, too. c = a.Copy() c.CrossProduct(b) check(a, b, c) // Check if vectors ∉ ℝ³ are rejected. d := New(2) e := New(4) _, err := CrossProduct(d, e) if err == nil { t.Error("CrossProduct() didn't error with invalid input vectors", "(∉ ℝ³)") } } // Check whether the various functions that take more than one vector error on // being supplied with vectors of missmatched dimensions. // It suffices to check the helper function checkDims, since every function // must call it to verify its inputs. func TestMissmatchedDims(t *testing.T) { a := New(2) b := New(3) err := checkDims(a, b) if err == nil { t.Error("Missmatched dimension check succeeded on unequal dimensions.") } a = New(4) b = New(4) err = checkDims(a, b) if err != nil { t.Error("Missmatched dimension check failed on equal dimensions.") } } // =========================== [ Helper Functions ] =========================== // Helper function, makes pseudo-random slices. func makeRandSlice(length uint) (randslice []float64) { randslice = make([]float64, length) for i := range randslice { randslice[i] = rand.ExpFloat64() } return } // Helper function, make a pseudo-random Vector with dimension dim. func makeRandomVector(dim uint) *Vector { return NewFrom(makeRandSlice(dim)) }
TestNewFrom
identifier_name
vector_test.go
// Author: slowpoke <proxypoke at lavabit dot com> // Repository: https://gist.github.com/proxypoke/vector // // This program is free software under the non-terms // of the Anti-License. Do whatever the fuck you want. package vector import ( "math" "math/rand" "testing" ) // ========================== [ Constructor Tests ] =========================== // Creates vectors with dimension from 0 to 99, checks if they actually have // that dimension, then checks if the values are correctly initialized to 0. func TestNew(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { // XXX: If the Get method errors, this test will still pass. This // is because Get() would then return an uninitialized float64 for // val, which is 0 and therefore what the test expects. val, _ := v.Get(j) if val != 0 { t.Error("Newly initialized vector has a value != 0.") } } } } // Creates vectors with randomized slices, then checks whether they have the // correct dimension (len(slice)) and whether they have been correctly // initialized. func TestNewFrom(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { randslice := makeRandSlice(i) v := NewFrom(randslice) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { val, _ := v.Get(j) if val != randslice[j] { t.Error( "Wrong values in vector initialized from a random slice.") } } } } // Creates pseudo-random vectors with various dimensions, copies them and // verifies that the new vector is equal. func TestCopy(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) w := v.Copy() if !Equal(v, w) { t.Error("Copied vector is not equal to source vector.") } } } // =================== [ General Methods/Functions Tests ] ==================== // Creates pseudo-random vectors with various dimensions, then check if Get() // returns the correct values and errors on out-of-range indexes. func TestGet(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) for j, val := range v.dims { getval, err := v.Get(uint(j)) if err != nil { t.Error("Get() errored on a correct index.") } if val != getval { t.Error("Get() returned a wrong value.") } } _, err := v.Get(v.Dim()) if err == nil { t.Error("Get didn't error on an out-of-range index.") } } } // Creates uninitialized vectors of various dimensions, then sets their values // to pseudo-random values. It then compares those values to check if they // were set correctly. Also verifies is Set() correctly errors on out-of-range // indexes. func TestSet(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) for j = 0; j < i; j++ { val := rand.ExpFloat64() err := v.Set(j, val) if err != nil { t.Error("Set() errored on a correct index.") } if v.dims[j] != val { t.Error("Set didn't correctly set a value.") } } err := v.Set(v.Dim(), 0) if err == nil { t.Error("Set didn't error on an out-of-range index.") } } } // Creates a vector with known length, then compares the expected value with // what Len() returns. func TestLen(t *testing.T) { v := New(1) v.Set(0, 2) // has length 2 if v.Len() != 2 { t.Error("Len returned a wrong length") } } // Creates Vectors which are known to be (un)equal, then verifies that Equal() // has correct oytput. func TestEqual(t *testing.T) { slc := make([]float64, 10) for i := range slc { slc[i] = float64(i) } v := NewFrom(slc) w := NewFrom(slc) if !Equal(v, w) { t.Error("Equal() != true for equal vectors.") } w = New(10) if Equal(v, w) { t.Error("Equal() == true for unequal vectors.") } }
// then as an in-place operations, checking if both operation were correct. func TestAdd(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Add(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error("Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place addition. c = a.Copy() c.Add(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error( "In-place Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Same as TestAdd, but with substraction. Heck, it's basically the same code. func TestSubstract(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Substract(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error("Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place sybstraction c = a.Copy() c.Substract(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error( "In-place Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Creates pseudo-random vectors, does scalar multiplication with pseudo-random // floats, then checks if the result is correct. It checks both the in-place // and the non-destructive version. func TestScale(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { a := makeRandomVector(i) x := rand.ExpFloat64() b := Scale(a, x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } // Test in-place scalar multiplication b = a.Copy() b.Scale(x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("In-place Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } } } // Creates pseudo-random vectors, normalizes them both in-place and // non-destructive, and verifies that the result is correct. func TestNormalize(t *testing.T) { var i uint // It makes no sense to normalize a zero vector, therefore we start at 1. for i = 1; i < 100; i++ { a := makeRandomVector(i) b := Normalize(a) if b.Len() != float64(1) { t.Error("Normalization failed, vector doesn't have length 1.") t.Logf("%f != 1", b.Len()) } } } // Uses vectors with known angles to calculate their DotProduct, then verifies // if the result is correct. func TestDotProduct(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel. a.Set(0, 1) b.Set(0, 1) dot, _ := DotProduct(a, b) if dot != 1 { t.Error("Dot Product of parallel vectors isn't 1.") } // Set the vectors as orthogonal. b = New(2) b.Set(1, 1) dot, _ = DotProduct(a, b) if dot != 0 { t.Error("Dot Product of orthogonal vectors isn't 0.") } // Set the vectors as anti-parallel. b = New(2) b.Set(0, -1) dot, _ = DotProduct(a, b) if dot != -1 { t.Error("Dot Product of anti-parallel vectors isn't -1.") } } // Uses vectors with known angles to verify that Angle() is correct. func TestAngle(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel (Θ == 0). a.Set(0, 1) b.Set(0, 1) Θ, _ := Angle(a, b) if Θ != 0 { t.Error("Angle between parallel vectors isn't 0.") t.Logf("%f != 0", Θ) } // Set the vectors as orthogonal (Θ == 0.5π). b = New(2) b.Set(1, 1) Θ, _ = Angle(a, b) if Θ != 0.5*math.Pi { t.Error("Angle between orthonal vectors isn't 0.5π.") t.Logf("%f != %f", Θ, 0.5*math.Pi) } // Set the vectors as anti-parallel (Θ == π). b = New(2) b.Set(0, -1) Θ, _ = Angle(a, b) if Θ != math.Pi { t.Error("Angle between anti-parallel vectors isn't π.") t.Logf("%f != %f", Θ, math.Pi) } } // Calculates the cross product of two pseudo-random vectors, then checks if // the resulting vector is orthogonal to both the original vectors. Tests both // in-place and non-destructive versions of the operation. func TestCrossProduct(t *testing.T) { check := func(a, b, c *Vector) { dot_a, _ := DotProduct(a, c) dot_b, _ := DotProduct(b, c) ε := 0.0000000005 if math.Abs(0-dot_a) < ε { dot_a = 0 } if math.Abs(0-dot_b) < ε { dot_b = 0 } if dot_a != 0 || dot_b != 0 { t.Error("Either or both vectors aren't orthogonal", "to their Cross Product.") t.Logf("a * c = %f", dot_a) t.Logf("b * c = %f", dot_b) } } a := makeRandomVector(3) b := makeRandomVector(3) c, _ := CrossProduct(a, b) check(a, b, c) // Check in-place, too. c = a.Copy() c.CrossProduct(b) check(a, b, c) // Check if vectors ∉ ℝ³ are rejected. d := New(2) e := New(4) _, err := CrossProduct(d, e) if err == nil { t.Error("CrossProduct() didn't error with invalid input vectors", "(∉ ℝ³)") } } // Check whether the various functions that take more than one vector error on // being supplied with vectors of missmatched dimensions. // It suffices to check the helper function checkDims, since every function // must call it to verify its inputs. func TestMissmatchedDims(t *testing.T) { a := New(2) b := New(3) err := checkDims(a, b) if err == nil { t.Error("Missmatched dimension check succeeded on unequal dimensions.") } a = New(4) b = New(4) err = checkDims(a, b) if err != nil { t.Error("Missmatched dimension check failed on equal dimensions.") } } // =========================== [ Helper Functions ] =========================== // Helper function, makes pseudo-random slices. func makeRandSlice(length uint) (randslice []float64) { randslice = make([]float64, length) for i := range randslice { randslice[i] = rand.ExpFloat64() } return } // Helper function, make a pseudo-random Vector with dimension dim. func makeRandomVector(dim uint) *Vector { return NewFrom(makeRandSlice(dim)) }
// =========================== [ Operation Tests ] ============================ // Creates pesudo-random vectors, then adds them first as a non-destructive,
random_line_split
vector_test.go
// Author: slowpoke <proxypoke at lavabit dot com> // Repository: https://gist.github.com/proxypoke/vector // // This program is free software under the non-terms // of the Anti-License. Do whatever the fuck you want. package vector import ( "math" "math/rand" "testing" ) // ========================== [ Constructor Tests ] =========================== // Creates vectors with dimension from 0 to 99, checks if they actually have // that dimension, then checks if the values are correctly initialized to 0. func TestNew(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { // XXX: If the Get method errors, this test will still pass. This // is because Get() would then return an uninitialized float64 for // val, which is 0 and therefore what the test expects. val, _ := v.Get(j) if val != 0 { t.Error("Newly initialized vector has a value != 0.") } } } } // Creates vectors with randomized slices, then checks whether they have the // correct dimension (len(slice)) and whether they have been correctly // initialized. func TestNewFrom(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { randslice := makeRandSlice(i) v := NewFrom(randslice) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { val, _ := v.Get(j) if val != randslice[j] { t.Error( "Wrong values in vector initialized from a random slice.") } } } } // Creates pseudo-random vectors with various dimensions, copies them and // verifies that the new vector is equal. func TestCopy(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) w := v.Copy() if !Equal(v, w) { t.Error("Copied vector is not equal to source vector.") } } } // =================== [ General Methods/Functions Tests ] ==================== // Creates pseudo-random vectors with various dimensions, then check if Get() // returns the correct values and errors on out-of-range indexes. func TestGet(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) for j, val := range v.dims { getval, err := v.Get(uint(j)) if err != nil { t.Error("Get() errored on a correct index.") } if val != getval { t.Error("Get() returned a wrong value.") } } _, err := v.Get(v.Dim()) if err == nil { t.Error("Get didn't error on an out-of-range index.") } } } // Creates uninitialized vectors of various dimensions, then sets their values // to pseudo-random values. It then compares those values to check if they // were set correctly. Also verifies is Set() correctly errors on out-of-range // indexes. func TestSet(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) for j = 0; j < i; j++ { val := rand.ExpFloat64() err := v.Set(j, val) if err != nil { t.Error("Set() errored on a correct index.") } if v.dims[j] != val { t.Error("Set didn't correctly set a value.") } } err := v.Set(v.Dim(), 0) if err == nil { t.Error("Set didn't error on an out-of-range index.") } } } // Creates a vector with known length, then compares the expected value with // what Len() returns. func TestLen(t *testing.T) { v := New(1) v.Set(0, 2) // has length 2 if v.Len() != 2 { t.Error("Len returned a wrong length") } } // Creates Vectors which are known to be (un)equal, then verifies that Equal() // has correct oytput. func TestEqual(t *testing.T) { slc := make([]float64, 10) for i := range slc { slc[i] = float64(i) } v := NewFrom(slc) w := NewFrom(slc) if !Equal(v, w) { t.Error("Equal() != true for equal vectors.") } w = New(10) if Equal(v, w) { t.Error("Equal() == true for unequal vectors.") } } // =========================== [ Operation Tests ] ============================ // Creates pesudo-random vectors, then adds them first as a non-destructive, // then as an in-place operations, checking if both operation were correct. func TestAdd(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Add(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error("Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place addition. c = a.Copy() c.Add(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error( "In-place Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Same as TestAdd, but with substraction. Heck, it's basically the same code. func TestSubstract(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Substract(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error("Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place sybstraction c = a.Copy() c.Substract(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error( "In-place Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Creates pseudo-random vectors, does scalar multiplication with pseudo-random // floats, then checks if the result is correct. It checks both the in-place // and the non-destructive version. func TestScale(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { a := makeRandomVector(i) x := rand.ExpFloat64() b := Scale(a, x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } // Test in-place scalar multiplication b = a.Copy() b.Scale(x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("In-place Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } } } // Creates pseudo-random vectors, normalizes them both in-place and // non-destructive, and verifies that the result is correct. func TestNormalize(t *testing.T) { var i uint // It makes no sense to normalize a zero vector, therefore we start at 1. for i = 1; i < 100; i++ { a := makeRandomVector(i) b := Normalize(a) if b.Len() != float64(1) { t.Error("Normalization failed, vector doesn't have length 1.") t.Logf("%f != 1", b.Len()) } } } // Uses vectors with known angles to calculate their DotProduct, then verifies // if the result is correct. func TestDotProduct(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel. a.Set(0, 1) b.Set(0, 1) dot, _ := DotProduct(a, b) if dot != 1 { t.Error("Dot Product of parallel vectors isn't 1.") } // Set the vectors as orthogonal. b = New(2) b.Set(1, 1) dot, _ = DotProduct(a, b) if dot != 0 { t.Error("Dot Product of orthogonal vectors isn't 0.") } // Set the vectors as anti-parallel. b = New(2) b.Set(0, -1) dot, _ = DotProduct(a, b) if dot != -1 { t.Error("Dot Product of anti-parallel vectors isn't -1.") } } // Uses vectors with known angles to verify that Angle() is correct. func TestAngle(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel (Θ == 0). a.Set(0, 1) b.Set(0, 1) Θ, _ := Angle(a, b) if Θ != 0 { t.Error("Angle between parallel vectors isn't 0.") t.Logf("%f != 0", Θ) } // Set the vectors as orthogonal (Θ == 0.5π). b = New(2) b.Set(1, 1) Θ, _ = Angle(a, b) if Θ != 0.5*math.Pi { t.Error("Angle between orthonal vectors isn't 0.5π.") t.Logf("%f != %f", Θ, 0.5*math.Pi) } // Set the vectors as anti-parallel (Θ == π). b = New(2) b.Set(0, -1) Θ, _ = Angle(a, b) if Θ != math.Pi { t.Error("Angle between anti-parallel vectors isn't π.") t.Logf("%f != %f", Θ, math.Pi) } } // Calculates the cross product of two pseudo-random vectors, then checks if // the resulting vector is orthogonal to both the original vectors. Tests both // in-place and non-destructive versions of the operation. func TestCrossProduct(t *testing.T) { check := func(a, b, c *Vector) { dot_a, _ := DotProduct(a, c) dot_b, _ := DotProduct(b, c) ε := 0.0000000005 if math.Abs(0-dot_a) < ε { dot_a = 0 } if math.Abs(0-dot_b) < ε { dot_b = 0 } if dot_a != 0 || dot_b != 0 { t.Error("Either or both vectors aren't orthogonal", "to their Cross Product.") t.Logf("a * c = %f", dot_a) t.Logf("b * c = %f", dot_b) } } a := makeRandomVector(3) b := makeRandomVector(3) c, _ := CrossProduct(a, b) check(a, b, c) // Check in-place, too. c = a.Copy() c.CrossProduct(b) check(a, b, c) // Check if vectors ∉ ℝ³ are rejected. d := New(2) e := New(4) _, err := CrossProduct(d, e) if err == nil { t.Error("CrossProduct() didn't error with invalid input vectors", "(∉ ℝ³)") } } // Check whether the various functions that take more than one vector error on // being supplied with vectors of missmatched dimensions. // It suffices to check the helper function checkDims, since every function // must call it to verify its inputs. func TestMissmatchedDims(t *testing.T) { a := New(2) b := New(3) err := checkDims(a, b) if err == nil { t.Error("Missmatched dimension check succeeded on unequal dimensions.") } a = New(4) b = New(4) err = checkDims(a, b) if err != nil { t.Error("Missmatched dimension check failed on equal dimensions.") } } // =========================== [ Helper Functions ] =========================== // Helper function, makes pseudo-random slices. func makeRandSlice(length uint) (randslice []float64) { randslice = make([]float64
pseudo-random Vector with dimension dim. func makeRandomVector(dim uint) *Vector { return NewFrom(makeRandSlice(dim)) }
, length) for i := range randslice { randslice[i] = rand.ExpFloat64() } return } // Helper function, make a
identifier_body
vector_test.go
// Author: slowpoke <proxypoke at lavabit dot com> // Repository: https://gist.github.com/proxypoke/vector // // This program is free software under the non-terms // of the Anti-License. Do whatever the fuck you want. package vector import ( "math" "math/rand" "testing" ) // ========================== [ Constructor Tests ] =========================== // Creates vectors with dimension from 0 to 99, checks if they actually have // that dimension, then checks if the values are correctly initialized to 0. func TestNew(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { // XXX: If the Get method errors, this test will still pass. This // is because Get() would then return an uninitialized float64 for // val, which is 0 and therefore what the test expects. val, _ := v.Get(j) if val != 0 { t.Error("Newly initialized vector has a value != 0.") } } } } // Creates vectors with randomized slices, then checks whether they have the // correct dimension (len(slice)) and whether they have been correctly // initialized. func TestNewFrom(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { randslice := makeRandSlice(i) v := NewFrom(randslice) if v.Dim() != i { t.Errorf("Wrong dimension. Got %d, expected %d.", v.Dim(), i) } for j = 0; j < i; j++ { val, _ := v.Get(j) if val != randslice[j] { t.Error( "Wrong values in vector initialized from a random slice.") } } } } // Creates pseudo-random vectors with various dimensions, copies them and // verifies that the new vector is equal. func TestCopy(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) w := v.Copy() if !Equal(v, w) { t.Error("Copied vector is not equal to source vector.") } } } // =================== [ General Methods/Functions Tests ] ==================== // Creates pseudo-random vectors with various dimensions, then check if Get() // returns the correct values and errors on out-of-range indexes. func TestGet(t *testing.T) { var i uint for i = 0; i < 100; i++ { v := makeRandomVector(i) for j, val := range v.dims { getval, err := v.Get(uint(j)) if err != nil { t.Error("Get() errored on a correct index.") } if val != getval { t.Error("Get() returned a wrong value.") } } _, err := v.Get(v.Dim()) if err == nil { t.Error("Get didn't error on an out-of-range index.") } } } // Creates uninitialized vectors of various dimensions, then sets their values // to pseudo-random values. It then compares those values to check if they // were set correctly. Also verifies is Set() correctly errors on out-of-range // indexes. func TestSet(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { v := New(i) for j = 0; j < i; j++ { val := rand.ExpFloat64() err := v.Set(j, val) if err != nil { t.Error("Set() errored on a correct index.") } if v.dims[j] != val { t.Error("Set didn't correctly set a value.") } } err := v.Set(v.Dim(), 0) if err == nil { t.Error("Set didn't error on an out-of-range index.") } } } // Creates a vector with known length, then compares the expected value with // what Len() returns. func TestLen(t *testing.T) { v := New(1) v.Set(0, 2) // has length 2 if v.Len() != 2 { t.Error("Len returned a wrong length") } } // Creates Vectors which are known to be (un)equal, then verifies that Equal() // has correct oytput. func TestEqual(t *testing.T) { slc := make([]float64, 10) for i := range slc { slc[i] = float64(i) } v := NewFrom(slc) w := NewFrom(slc) if !Equal(v, w) { t.Error("Equal() != true for equal vectors.") } w = New(10) if Equal(v, w)
} // =========================== [ Operation Tests ] ============================ // Creates pesudo-random vectors, then adds them first as a non-destructive, // then as an in-place operations, checking if both operation were correct. func TestAdd(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Add(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error("Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place addition. c = a.Copy() c.Add(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]+b.dims[j] { t.Error( "In-place Addition failed, didn't get expected values.") t.Logf("%f + %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Same as TestAdd, but with substraction. Heck, it's basically the same code. func TestSubstract(t *testing.T) { var i, j uint for i = 1; i < 100; i++ { a := makeRandomVector(i) b := makeRandomVector(i) c, _ := Substract(a, b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error("Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } // Test in-place sybstraction c = a.Copy() c.Substract(b) for j = 0; j < i; j++ { if c.dims[j] != a.dims[j]-b.dims[j] { t.Error( "In-place Substraction failed, didn't get expected values.") t.Logf("%f - %f != %f", a.dims[j], b.dims[j], c.dims[j]) } } } } // Creates pseudo-random vectors, does scalar multiplication with pseudo-random // floats, then checks if the result is correct. It checks both the in-place // and the non-destructive version. func TestScale(t *testing.T) { var i, j uint for i = 0; i < 100; i++ { a := makeRandomVector(i) x := rand.ExpFloat64() b := Scale(a, x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } // Test in-place scalar multiplication b = a.Copy() b.Scale(x) for j = 0; j < i; j++ { if b.dims[j] != a.dims[j]*x { t.Error("In-place Scalar Multiplication failed, ", "didn't get expected values.") t.Logf("%f * %f != %f", a.dims[j], x, b.dims[j]) } } } } // Creates pseudo-random vectors, normalizes them both in-place and // non-destructive, and verifies that the result is correct. func TestNormalize(t *testing.T) { var i uint // It makes no sense to normalize a zero vector, therefore we start at 1. for i = 1; i < 100; i++ { a := makeRandomVector(i) b := Normalize(a) if b.Len() != float64(1) { t.Error("Normalization failed, vector doesn't have length 1.") t.Logf("%f != 1", b.Len()) } } } // Uses vectors with known angles to calculate their DotProduct, then verifies // if the result is correct. func TestDotProduct(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel. a.Set(0, 1) b.Set(0, 1) dot, _ := DotProduct(a, b) if dot != 1 { t.Error("Dot Product of parallel vectors isn't 1.") } // Set the vectors as orthogonal. b = New(2) b.Set(1, 1) dot, _ = DotProduct(a, b) if dot != 0 { t.Error("Dot Product of orthogonal vectors isn't 0.") } // Set the vectors as anti-parallel. b = New(2) b.Set(0, -1) dot, _ = DotProduct(a, b) if dot != -1 { t.Error("Dot Product of anti-parallel vectors isn't -1.") } } // Uses vectors with known angles to verify that Angle() is correct. func TestAngle(t *testing.T) { a := New(2) b := New(2) // Set the vectors as parallel (Θ == 0). a.Set(0, 1) b.Set(0, 1) Θ, _ := Angle(a, b) if Θ != 0 { t.Error("Angle between parallel vectors isn't 0.") t.Logf("%f != 0", Θ) } // Set the vectors as orthogonal (Θ == 0.5π). b = New(2) b.Set(1, 1) Θ, _ = Angle(a, b) if Θ != 0.5*math.Pi { t.Error("Angle between orthonal vectors isn't 0.5π.") t.Logf("%f != %f", Θ, 0.5*math.Pi) } // Set the vectors as anti-parallel (Θ == π). b = New(2) b.Set(0, -1) Θ, _ = Angle(a, b) if Θ != math.Pi { t.Error("Angle between anti-parallel vectors isn't π.") t.Logf("%f != %f", Θ, math.Pi) } } // Calculates the cross product of two pseudo-random vectors, then checks if // the resulting vector is orthogonal to both the original vectors. Tests both // in-place and non-destructive versions of the operation. func TestCrossProduct(t *testing.T) { check := func(a, b, c *Vector) { dot_a, _ := DotProduct(a, c) dot_b, _ := DotProduct(b, c) ε := 0.0000000005 if math.Abs(0-dot_a) < ε { dot_a = 0 } if math.Abs(0-dot_b) < ε { dot_b = 0 } if dot_a != 0 || dot_b != 0 { t.Error("Either or both vectors aren't orthogonal", "to their Cross Product.") t.Logf("a * c = %f", dot_a) t.Logf("b * c = %f", dot_b) } } a := makeRandomVector(3) b := makeRandomVector(3) c, _ := CrossProduct(a, b) check(a, b, c) // Check in-place, too. c = a.Copy() c.CrossProduct(b) check(a, b, c) // Check if vectors ∉ ℝ³ are rejected. d := New(2) e := New(4) _, err := CrossProduct(d, e) if err == nil { t.Error("CrossProduct() didn't error with invalid input vectors", "(∉ ℝ³)") } } // Check whether the various functions that take more than one vector error on // being supplied with vectors of missmatched dimensions. // It suffices to check the helper function checkDims, since every function // must call it to verify its inputs. func TestMissmatchedDims(t *testing.T) { a := New(2) b := New(3) err := checkDims(a, b) if err == nil { t.Error("Missmatched dimension check succeeded on unequal dimensions.") } a = New(4) b = New(4) err = checkDims(a, b) if err != nil { t.Error("Missmatched dimension check failed on equal dimensions.") } } // =========================== [ Helper Functions ] =========================== // Helper function, makes pseudo-random slices. func makeRandSlice(length uint) (randslice []float64) { randslice = make([]float64, length) for i := range randslice { randslice[i] = rand.ExpFloat64() } return } // Helper function, make a pseudo-random Vector with dimension dim. func makeRandomVector(dim uint) *Vector { return NewFrom(makeRandSlice(dim)) }
{ t.Error("Equal() == true for unequal vectors.") }
conditional_block
preprocessing.py
# Globals # import numpy as np import pandas as pd import dateutil.parser as dp from itertools import islice from scipy.stats import boxcox from scipy.integrate import simps from realtime_talib import Indicator from sklearn.model_selection import train_test_split from sklearn.utils import resample from pprint import pprint from selenium import webdriver RANDOM_STATE = 42 # Helpers # def sliding_window(seq, n=2): """Returns a sliding window (of width n) over data from the iterable. https://stackoverflow.com/a/6822773/8740440""" "s -> (s0,s1,...s[n-1]), (s1,s2,...,sn), ..." it = iter(seq) result = tuple(islice(it, n)) if len(result) == n: yield result for elem in it: result = result[1:] + (elem,) yield result def integrate(avg_daily_sentiment, interval): """Takes a list of average daily sentiment scores and returns a list of definite integral estimations calculated with Simpson's method. Each integral interval is determined by the `interval` variable. Shows accumulated sentiment.""" # Split into sliding window list of lists sentiment_windows = sliding_window(avg_daily_sentiment, interval) integral_simpson_est = [] # https://stackoverflow.com/a/13323861/8740440 for x in sentiment_windows: # Estimate area using composite Simpson's rule. dx indicates the spacing of the data on the x-axis. integral_simpson_est.append(simps(x, dx=1)) dead_values = list([None] * interval) dead_values.extend(integral_simpson_est) dead_values.reverse() return dead_values def random_undersampling(dataset): """Randomly deleting rows that contain the majority class until the number in the majority class is equal with the number in the minority class.""" minority_set = dataset[dataset.Trend == -1.0] majority_set = dataset[dataset.Trend == 1.0] # print(dataset.Trend.value_counts()) # If minority set larger than majority set, swap if len(minority_set) > len(majority_set): minority_set, majority_set = majority_set, minority_set # Downsample majority class majority_downsampled = resample(majority_set, replace=False, # sample without replacement n_samples=len(minority_set), # to match minority class random_state=123) # reproducible results # Combine minority class with downsampled majority class return pd.concat([majority_downsampled, minority_set]) def get_popularity(headlines): # TODO: Randomize user-agents OR figure out how to handle popups if "Tweets" not in headlines.columns: counts = [] driver = webdriver.Chrome() for index, row in headlines.iterrows(): try: driver.get(row["URL"]) time.sleep(3) twitter_containers = driver.find_elements_by_xpath("//li[@class='twitter']") count = twitter_containers[0].find_elements_by_xpath("//span[@class='count']") if count[0].text == "": counts.append(1) else: counts.append(int(count[0].text)) except: counts.append(1) # QUESTION: Should it be None? headlines["Tweets"] = (pd.Series(counts)).values print(counts) return headlines def balanced_split(dataset, test_size): """Randomly splits dataset into balanced training and test sets.""" print("\tSplitting data into *balanced* training and test sets") # Use sklearn.train_test_split to split original dataset into x_train, y_train, x_test, y_test numpy arrays x_train, x_test, y_train, y_test = train_test_split(dataset.drop(["Date", "Trend"], axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) # Combine x_train and y_train (numpy arrays) into a single dataframe, with column labels train = pd.DataFrame(data=x_train, columns=dataset.columns[1:-1]) train["Trend"] = pd.Series(y_train) # Do the same for x_test and y__test test = pd.DataFrame(data=x_test, columns=dataset.columns[1:-1]) test["Trend"] = pd.Series(y_test) # Apply random undersampling to both data frames train_downsampled = random_undersampling(train) test_downsampled = random_undersampling(test) train_trend = train_downsampled["Trend"].values test_trend = test_downsampled["Trend"].values train_trimmed = train_downsampled.drop(["Trend"], axis=1).values test_trimmed = test_downsampled.drop(["Trend"], axis=1).values return train_trimmed, test_trimmed, train_trend, test_trend def unbalanced_split(dataset, test_size): """Randomly splits dataset into unbalanced training and test sets.""" print("\tSplitting data into *unbalanced* training and test sets") dataset = dataset.drop("Date", axis=1) output = train_test_split(dataset.drop("Trend", axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) return output # Main # def calculate_indicators(ohlcv): """Extracts technical indicators from OHLCV data.""" print("\tCalculating technical indicators") ohlcv = ohlcv.drop(["Volume (BTC)", "Weighted Price"], axis=1) ohlcv.columns = ["Date", "Open", "High", "Low", "Close", "Volume"] temp_ohlcv = ohlcv.copy() # Converts ISO 8601 timestamps to UNIX unix_times = [int((dp.parse(temp_ohlcv.iloc[index]["Date"])).strftime("%s")) for index in range(temp_ohlcv.shape[0])] temp_ohlcv["Date"] = (pd.Series(unix_times)).values # Converts column headers to lowercase and sorts rows in chronological order temp_ohlcv.columns = ["date", "open", "high", "low", "close", "volume"] temp_ohlcv = temp_ohlcv.iloc[::-1] # Rate of Change Ratio rocr3 = ((Indicator(temp_ohlcv, "ROCR", 3)).getHistorical())[::-1] rocr6 = ((Indicator(temp_ohlcv, "ROCR", 6)).getHistorical())[::-1] # Average True Range atr = ((Indicator(temp_ohlcv, "ATR", 14)).getHistorical())[::-1] # On-Balance Volume obv = ((Indicator(temp_ohlcv, "OBV")).getHistorical())[::-1] # Triple Exponential Moving Average trix = ((Indicator(temp_ohlcv, "TRIX", 20)).getHistorical())[::-1] # Momentum mom1 = ((Indicator(temp_ohlcv, "MOM", 1)).getHistorical())[::-1] mom3 = ((Indicator(temp_ohlcv, "MOM", 3)).getHistorical())[::-1] # Average Directional Index adx14 = ((Indicator(temp_ohlcv, "ADX", 14)).getHistorical())[::-1] adx20 = ((Indicator(temp_ohlcv, "ADX", 20)).getHistorical())[::-1] # Williams %R willr = ((Indicator(temp_ohlcv, "WILLR", 14)).getHistorical())[::-1] # Relative Strength Index rsi6 = ((Indicator(temp_ohlcv, "RSI", 6)).getHistorical())[::-1] rsi12 = ((Indicator(temp_ohlcv, "RSI", 12)).getHistorical())[::-1] # Moving Average Convergence Divergence macd, macd_signal, macd_hist = (Indicator(temp_ohlcv, "MACD", 12, 26, 9)).getHistorical() macd, macd_signal, macd_hist = macd[::-1], macd_signal[::-1], macd_hist[::-1] # Exponential Moving Average ema6 = ((Indicator(temp_ohlcv, "MA", 6, 1)).getHistorical())[::-1] ema12 = ((Indicator(temp_ohlcv, "MA", 12, 1)).getHistorical())[::-1] # Append indicators to the input datasets min_length = min(len(mom1), len(mom3), len(adx14), len(adx20), len(willr), len(rsi6), len(rsi12), len(macd), len(macd_signal), len(macd_hist), len(ema6), len(ema12), len(rocr3), len(rocr6), len(atr), len(obv), len(trix)) ohlcv = ohlcv[:min_length].drop(["Open", "High", "Low"], axis=1) ohlcv["MOM (1)"], ohlcv["MOM (3)"], ohlcv["ADX (14)"] = (pd.Series(mom1[:min_length])).values, (pd.Series(mom3[:min_length])).values, (pd.Series(adx14[:min_length])).values ohlcv["ADX (20)"], ohlcv["WILLR"], ohlcv["RSI (6)"] = (pd.Series(adx20[:min_length])).values, (pd.Series(willr[:min_length])).values, (pd.Series(rsi6[:min_length])).values ohlcv["RSI (12)"], ohlcv["MACD"], ohlcv["MACD (Signal)"] = (pd.Series(rsi12[:min_length])).values, (pd.Series(macd[:min_length])).values, (pd.Series(macd_signal[:min_length])).values ohlcv["MACD (Historical)"], ohlcv["EMA (6)"], ohlcv["EMA (12)"] = (pd.Series(macd_hist[:min_length])).values, (pd.Series(ema6[:min_length])).values, (pd.Series(ema12[:min_length])).values ohlcv["ROCR (3)"], ohlcv["ROCR (6)"], ohlcv["ATR (14)"] = (pd.Series(rocr3[:min_length])).values, (pd.Series(rocr6[:min_length])).values, (pd.Series(atr[:min_length])).values ohlcv["OBV"], ohlcv["TRIX (20)"] = (pd.Series(obv[:min_length])).values, (pd.Series(trix[:min_length])).values return ohlcv def calculate_sentiment(headlines): sentiment_scores = {} numer, denom = 0.0, 0.0 for index, currRow in headlines.iterrows(): print(currRow) currDate = currRow["Date"] if currDate in sentiment_scores: pass else: numer = currRow["Sentiment"] * currRow["Tweets"] denom = currRow["Tweets"] for index, nextRow in headlines.iloc[index + 1:].iterrows(): if nextRow["Date"] == currDate: numer += (nextRow["Sentiment"] * nextRow["Tweets"]) denom += nextRow["Tweets"] else: break sentiment_scores[currDate] = numer / denom numer, denom = 0.0, 0.0 sentiment_scores_df = pd.DataFrame(list(sentiment_scores.items()), columns=["Date", "Sentiment"]) return sentiment_scores_df def merge_datasets(origin, other_sets): print("\tMerging datasets") merged = origin for set in other_sets: merged = pd.merge(merged, set, on="Date") return merged def fix_null_vals(dataset): """Implements the Last Observation Carried Forward (LOCF) method to fill missing values.""" print("\tFixing null values") if not dataset.isnull().any().any(): return dataset else: return dataset.fillna(method="ffill") def binarize_labels(dataset): """Transforms daily price data into binary values indicating price change.""" print("\tBinarizing price movements") trends = [None] for index in range(dataset.shape[0] - 1): difference = dataset.iloc[index]["Close"] - dataset.iloc[index + 1]["Close"] if difference < 0: trends.append(-1) else: trends.append(1) dataset["Trend"] = (pd.Series(trends)).values dataset = dataset.drop(dataset.index[0]) return dataset def add_lag_variables(dataset, lag=3):
def power_transform(dataset): print("\tApplying a box-cox transform to selected features") for header in dataset.drop(["Date", "Trend"], axis=1).columns: if not (dataset[header] < 0).any() and not (dataset[header] == 0).any(): dataset[header] = boxcox(dataset[header])[0] return dataset def split(dataset, test_size, balanced=True): # TODO: Splits can't be random, they need to respect the temporal order of each observation if balanced: return balanced_split(dataset, test_size) else: return unbalanced_split(dataset, test_size) def integral_transform(dataset, interval): integral = integrate(list(dataset["Sentiment"]), interval) dataset["Sentiment_integrals"] = pd.Series(integral)
print("\tAdding lag variables") new_df_dict = {} for col_header in dataset.drop(["Date", "Trend"], axis=1): new_df_dict[col_header] = dataset[col_header] for lag in range(1, lag + 1): new_df_dict["%s_lag%d" % (col_header, lag)] = dataset[col_header].shift(-lag) new_df = pd.DataFrame(new_df_dict, index=dataset.index) new_df["Date"], new_df["Trend"] = dataset["Date"], dataset["Trend"] return new_df.dropna()
identifier_body
preprocessing.py
# Globals # import numpy as np import pandas as pd import dateutil.parser as dp from itertools import islice from scipy.stats import boxcox from scipy.integrate import simps from realtime_talib import Indicator from sklearn.model_selection import train_test_split from sklearn.utils import resample from pprint import pprint from selenium import webdriver RANDOM_STATE = 42 # Helpers # def sliding_window(seq, n=2): """Returns a sliding window (of width n) over data from the iterable. https://stackoverflow.com/a/6822773/8740440""" "s -> (s0,s1,...s[n-1]), (s1,s2,...,sn), ..." it = iter(seq) result = tuple(islice(it, n)) if len(result) == n: yield result for elem in it: result = result[1:] + (elem,) yield result def integrate(avg_daily_sentiment, interval): """Takes a list of average daily sentiment scores and returns a list of definite integral estimations calculated with Simpson's method. Each integral interval is determined by the `interval` variable. Shows accumulated sentiment.""" # Split into sliding window list of lists sentiment_windows = sliding_window(avg_daily_sentiment, interval) integral_simpson_est = [] # https://stackoverflow.com/a/13323861/8740440 for x in sentiment_windows: # Estimate area using composite Simpson's rule. dx indicates the spacing of the data on the x-axis. integral_simpson_est.append(simps(x, dx=1)) dead_values = list([None] * interval) dead_values.extend(integral_simpson_est) dead_values.reverse() return dead_values def random_undersampling(dataset): """Randomly deleting rows that contain the majority class until the number in the majority class is equal with the number in the minority class.""" minority_set = dataset[dataset.Trend == -1.0] majority_set = dataset[dataset.Trend == 1.0] # print(dataset.Trend.value_counts()) # If minority set larger than majority set, swap if len(minority_set) > len(majority_set): minority_set, majority_set = majority_set, minority_set # Downsample majority class majority_downsampled = resample(majority_set, replace=False, # sample without replacement n_samples=len(minority_set), # to match minority class random_state=123) # reproducible results # Combine minority class with downsampled majority class return pd.concat([majority_downsampled, minority_set]) def get_popularity(headlines): # TODO: Randomize user-agents OR figure out how to handle popups if "Tweets" not in headlines.columns: counts = [] driver = webdriver.Chrome() for index, row in headlines.iterrows(): try: driver.get(row["URL"]) time.sleep(3) twitter_containers = driver.find_elements_by_xpath("//li[@class='twitter']") count = twitter_containers[0].find_elements_by_xpath("//span[@class='count']") if count[0].text == "": counts.append(1) else: counts.append(int(count[0].text)) except: counts.append(1) # QUESTION: Should it be None? headlines["Tweets"] = (pd.Series(counts)).values print(counts) return headlines def balanced_split(dataset, test_size): """Randomly splits dataset into balanced training and test sets.""" print("\tSplitting data into *balanced* training and test sets") # Use sklearn.train_test_split to split original dataset into x_train, y_train, x_test, y_test numpy arrays x_train, x_test, y_train, y_test = train_test_split(dataset.drop(["Date", "Trend"], axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) # Combine x_train and y_train (numpy arrays) into a single dataframe, with column labels train = pd.DataFrame(data=x_train, columns=dataset.columns[1:-1]) train["Trend"] = pd.Series(y_train) # Do the same for x_test and y__test test = pd.DataFrame(data=x_test, columns=dataset.columns[1:-1]) test["Trend"] = pd.Series(y_test) # Apply random undersampling to both data frames train_downsampled = random_undersampling(train) test_downsampled = random_undersampling(test) train_trend = train_downsampled["Trend"].values test_trend = test_downsampled["Trend"].values train_trimmed = train_downsampled.drop(["Trend"], axis=1).values test_trimmed = test_downsampled.drop(["Trend"], axis=1).values return train_trimmed, test_trimmed, train_trend, test_trend def unbalanced_split(dataset, test_size): """Randomly splits dataset into unbalanced training and test sets.""" print("\tSplitting data into *unbalanced* training and test sets") dataset = dataset.drop("Date", axis=1) output = train_test_split(dataset.drop("Trend", axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) return output # Main # def calculate_indicators(ohlcv): """Extracts technical indicators from OHLCV data.""" print("\tCalculating technical indicators") ohlcv = ohlcv.drop(["Volume (BTC)", "Weighted Price"], axis=1) ohlcv.columns = ["Date", "Open", "High", "Low", "Close", "Volume"] temp_ohlcv = ohlcv.copy() # Converts ISO 8601 timestamps to UNIX unix_times = [int((dp.parse(temp_ohlcv.iloc[index]["Date"])).strftime("%s")) for index in range(temp_ohlcv.shape[0])] temp_ohlcv["Date"] = (pd.Series(unix_times)).values # Converts column headers to lowercase and sorts rows in chronological order temp_ohlcv.columns = ["date", "open", "high", "low", "close", "volume"] temp_ohlcv = temp_ohlcv.iloc[::-1] # Rate of Change Ratio rocr3 = ((Indicator(temp_ohlcv, "ROCR", 3)).getHistorical())[::-1] rocr6 = ((Indicator(temp_ohlcv, "ROCR", 6)).getHistorical())[::-1] # Average True Range atr = ((Indicator(temp_ohlcv, "ATR", 14)).getHistorical())[::-1] # On-Balance Volume obv = ((Indicator(temp_ohlcv, "OBV")).getHistorical())[::-1] # Triple Exponential Moving Average trix = ((Indicator(temp_ohlcv, "TRIX", 20)).getHistorical())[::-1] # Momentum mom1 = ((Indicator(temp_ohlcv, "MOM", 1)).getHistorical())[::-1] mom3 = ((Indicator(temp_ohlcv, "MOM", 3)).getHistorical())[::-1] # Average Directional Index adx14 = ((Indicator(temp_ohlcv, "ADX", 14)).getHistorical())[::-1] adx20 = ((Indicator(temp_ohlcv, "ADX", 20)).getHistorical())[::-1] # Williams %R willr = ((Indicator(temp_ohlcv, "WILLR", 14)).getHistorical())[::-1] # Relative Strength Index rsi6 = ((Indicator(temp_ohlcv, "RSI", 6)).getHistorical())[::-1] rsi12 = ((Indicator(temp_ohlcv, "RSI", 12)).getHistorical())[::-1] # Moving Average Convergence Divergence macd, macd_signal, macd_hist = (Indicator(temp_ohlcv, "MACD", 12, 26, 9)).getHistorical() macd, macd_signal, macd_hist = macd[::-1], macd_signal[::-1], macd_hist[::-1] # Exponential Moving Average ema6 = ((Indicator(temp_ohlcv, "MA", 6, 1)).getHistorical())[::-1] ema12 = ((Indicator(temp_ohlcv, "MA", 12, 1)).getHistorical())[::-1] # Append indicators to the input datasets min_length = min(len(mom1), len(mom3), len(adx14), len(adx20), len(willr), len(rsi6), len(rsi12), len(macd), len(macd_signal), len(macd_hist), len(ema6), len(ema12), len(rocr3), len(rocr6), len(atr), len(obv), len(trix)) ohlcv = ohlcv[:min_length].drop(["Open", "High", "Low"], axis=1) ohlcv["MOM (1)"], ohlcv["MOM (3)"], ohlcv["ADX (14)"] = (pd.Series(mom1[:min_length])).values, (pd.Series(mom3[:min_length])).values, (pd.Series(adx14[:min_length])).values ohlcv["ADX (20)"], ohlcv["WILLR"], ohlcv["RSI (6)"] = (pd.Series(adx20[:min_length])).values, (pd.Series(willr[:min_length])).values, (pd.Series(rsi6[:min_length])).values ohlcv["RSI (12)"], ohlcv["MACD"], ohlcv["MACD (Signal)"] = (pd.Series(rsi12[:min_length])).values, (pd.Series(macd[:min_length])).values, (pd.Series(macd_signal[:min_length])).values ohlcv["MACD (Historical)"], ohlcv["EMA (6)"], ohlcv["EMA (12)"] = (pd.Series(macd_hist[:min_length])).values, (pd.Series(ema6[:min_length])).values, (pd.Series(ema12[:min_length])).values ohlcv["ROCR (3)"], ohlcv["ROCR (6)"], ohlcv["ATR (14)"] = (pd.Series(rocr3[:min_length])).values, (pd.Series(rocr6[:min_length])).values, (pd.Series(atr[:min_length])).values ohlcv["OBV"], ohlcv["TRIX (20)"] = (pd.Series(obv[:min_length])).values, (pd.Series(trix[:min_length])).values return ohlcv def calculate_sentiment(headlines): sentiment_scores = {} numer, denom = 0.0, 0.0 for index, currRow in headlines.iterrows(): print(currRow) currDate = currRow["Date"] if currDate in sentiment_scores: pass else: numer = currRow["Sentiment"] * currRow["Tweets"] denom = currRow["Tweets"] for index, nextRow in headlines.iloc[index + 1:].iterrows(): if nextRow["Date"] == currDate: numer += (nextRow["Sentiment"] * nextRow["Tweets"]) denom += nextRow["Tweets"] else: break sentiment_scores[currDate] = numer / denom numer, denom = 0.0, 0.0 sentiment_scores_df = pd.DataFrame(list(sentiment_scores.items()), columns=["Date", "Sentiment"]) return sentiment_scores_df def merge_datasets(origin, other_sets): print("\tMerging datasets") merged = origin for set in other_sets: merged = pd.merge(merged, set, on="Date") return merged def fix_null_vals(dataset): """Implements the Last Observation Carried Forward (LOCF) method to fill missing values.""" print("\tFixing null values") if not dataset.isnull().any().any(): return dataset else: return dataset.fillna(method="ffill") def binarize_labels(dataset): """Transforms daily price data into binary values indicating price change.""" print("\tBinarizing price movements") trends = [None] for index in range(dataset.shape[0] - 1): difference = dataset.iloc[index]["Close"] - dataset.iloc[index + 1]["Close"] if difference < 0: trends.append(-1) else: trends.append(1) dataset["Trend"] = (pd.Series(trends)).values dataset = dataset.drop(dataset.index[0]) return dataset def add_lag_variables(dataset, lag=3): print("\tAdding lag variables") new_df_dict = {} for col_header in dataset.drop(["Date", "Trend"], axis=1): new_df_dict[col_header] = dataset[col_header] for lag in range(1, lag + 1): new_df_dict["%s_lag%d" % (col_header, lag)] = dataset[col_header].shift(-lag) new_df = pd.DataFrame(new_df_dict, index=dataset.index) new_df["Date"], new_df["Trend"] = dataset["Date"], dataset["Trend"] return new_df.dropna() def power_transform(dataset): print("\tApplying a box-cox transform to selected features") for header in dataset.drop(["Date", "Trend"], axis=1).columns: if not (dataset[header] < 0).any() and not (dataset[header] == 0).any(): dataset[header] = boxcox(dataset[header])[0] return dataset def split(dataset, test_size, balanced=True): # TODO: Splits can't be random, they need to respect the temporal order of each observation
def integral_transform(dataset, interval): integral = integrate(list(dataset["Sentiment"]), interval) dataset["Sentiment_integrals"] = pd.Series(integral)
if balanced: return balanced_split(dataset, test_size) else: return unbalanced_split(dataset, test_size)
random_line_split
preprocessing.py
# Globals # import numpy as np import pandas as pd import dateutil.parser as dp from itertools import islice from scipy.stats import boxcox from scipy.integrate import simps from realtime_talib import Indicator from sklearn.model_selection import train_test_split from sklearn.utils import resample from pprint import pprint from selenium import webdriver RANDOM_STATE = 42 # Helpers # def sliding_window(seq, n=2): """Returns a sliding window (of width n) over data from the iterable. https://stackoverflow.com/a/6822773/8740440""" "s -> (s0,s1,...s[n-1]), (s1,s2,...,sn), ..." it = iter(seq) result = tuple(islice(it, n)) if len(result) == n: yield result for elem in it: result = result[1:] + (elem,) yield result def integrate(avg_daily_sentiment, interval): """Takes a list of average daily sentiment scores and returns a list of definite integral estimations calculated with Simpson's method. Each integral interval is determined by the `interval` variable. Shows accumulated sentiment.""" # Split into sliding window list of lists sentiment_windows = sliding_window(avg_daily_sentiment, interval) integral_simpson_est = [] # https://stackoverflow.com/a/13323861/8740440 for x in sentiment_windows: # Estimate area using composite Simpson's rule. dx indicates the spacing of the data on the x-axis. integral_simpson_est.append(simps(x, dx=1)) dead_values = list([None] * interval) dead_values.extend(integral_simpson_est) dead_values.reverse() return dead_values def random_undersampling(dataset): """Randomly deleting rows that contain the majority class until the number in the majority class is equal with the number in the minority class.""" minority_set = dataset[dataset.Trend == -1.0] majority_set = dataset[dataset.Trend == 1.0] # print(dataset.Trend.value_counts()) # If minority set larger than majority set, swap if len(minority_set) > len(majority_set): minority_set, majority_set = majority_set, minority_set # Downsample majority class majority_downsampled = resample(majority_set, replace=False, # sample without replacement n_samples=len(minority_set), # to match minority class random_state=123) # reproducible results # Combine minority class with downsampled majority class return pd.concat([majority_downsampled, minority_set]) def get_popularity(headlines): # TODO: Randomize user-agents OR figure out how to handle popups if "Tweets" not in headlines.columns: counts = [] driver = webdriver.Chrome() for index, row in headlines.iterrows(): try: driver.get(row["URL"]) time.sleep(3) twitter_containers = driver.find_elements_by_xpath("//li[@class='twitter']") count = twitter_containers[0].find_elements_by_xpath("//span[@class='count']") if count[0].text == "": counts.append(1) else: counts.append(int(count[0].text)) except: counts.append(1) # QUESTION: Should it be None? headlines["Tweets"] = (pd.Series(counts)).values print(counts) return headlines def balanced_split(dataset, test_size): """Randomly splits dataset into balanced training and test sets.""" print("\tSplitting data into *balanced* training and test sets") # Use sklearn.train_test_split to split original dataset into x_train, y_train, x_test, y_test numpy arrays x_train, x_test, y_train, y_test = train_test_split(dataset.drop(["Date", "Trend"], axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) # Combine x_train and y_train (numpy arrays) into a single dataframe, with column labels train = pd.DataFrame(data=x_train, columns=dataset.columns[1:-1]) train["Trend"] = pd.Series(y_train) # Do the same for x_test and y__test test = pd.DataFrame(data=x_test, columns=dataset.columns[1:-1]) test["Trend"] = pd.Series(y_test) # Apply random undersampling to both data frames train_downsampled = random_undersampling(train) test_downsampled = random_undersampling(test) train_trend = train_downsampled["Trend"].values test_trend = test_downsampled["Trend"].values train_trimmed = train_downsampled.drop(["Trend"], axis=1).values test_trimmed = test_downsampled.drop(["Trend"], axis=1).values return train_trimmed, test_trimmed, train_trend, test_trend def unbalanced_split(dataset, test_size): """Randomly splits dataset into unbalanced training and test sets.""" print("\tSplitting data into *unbalanced* training and test sets") dataset = dataset.drop("Date", axis=1) output = train_test_split(dataset.drop("Trend", axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) return output # Main # def calculate_indicators(ohlcv): """Extracts technical indicators from OHLCV data.""" print("\tCalculating technical indicators") ohlcv = ohlcv.drop(["Volume (BTC)", "Weighted Price"], axis=1) ohlcv.columns = ["Date", "Open", "High", "Low", "Close", "Volume"] temp_ohlcv = ohlcv.copy() # Converts ISO 8601 timestamps to UNIX unix_times = [int((dp.parse(temp_ohlcv.iloc[index]["Date"])).strftime("%s")) for index in range(temp_ohlcv.shape[0])] temp_ohlcv["Date"] = (pd.Series(unix_times)).values # Converts column headers to lowercase and sorts rows in chronological order temp_ohlcv.columns = ["date", "open", "high", "low", "close", "volume"] temp_ohlcv = temp_ohlcv.iloc[::-1] # Rate of Change Ratio rocr3 = ((Indicator(temp_ohlcv, "ROCR", 3)).getHistorical())[::-1] rocr6 = ((Indicator(temp_ohlcv, "ROCR", 6)).getHistorical())[::-1] # Average True Range atr = ((Indicator(temp_ohlcv, "ATR", 14)).getHistorical())[::-1] # On-Balance Volume obv = ((Indicator(temp_ohlcv, "OBV")).getHistorical())[::-1] # Triple Exponential Moving Average trix = ((Indicator(temp_ohlcv, "TRIX", 20)).getHistorical())[::-1] # Momentum mom1 = ((Indicator(temp_ohlcv, "MOM", 1)).getHistorical())[::-1] mom3 = ((Indicator(temp_ohlcv, "MOM", 3)).getHistorical())[::-1] # Average Directional Index adx14 = ((Indicator(temp_ohlcv, "ADX", 14)).getHistorical())[::-1] adx20 = ((Indicator(temp_ohlcv, "ADX", 20)).getHistorical())[::-1] # Williams %R willr = ((Indicator(temp_ohlcv, "WILLR", 14)).getHistorical())[::-1] # Relative Strength Index rsi6 = ((Indicator(temp_ohlcv, "RSI", 6)).getHistorical())[::-1] rsi12 = ((Indicator(temp_ohlcv, "RSI", 12)).getHistorical())[::-1] # Moving Average Convergence Divergence macd, macd_signal, macd_hist = (Indicator(temp_ohlcv, "MACD", 12, 26, 9)).getHistorical() macd, macd_signal, macd_hist = macd[::-1], macd_signal[::-1], macd_hist[::-1] # Exponential Moving Average ema6 = ((Indicator(temp_ohlcv, "MA", 6, 1)).getHistorical())[::-1] ema12 = ((Indicator(temp_ohlcv, "MA", 12, 1)).getHistorical())[::-1] # Append indicators to the input datasets min_length = min(len(mom1), len(mom3), len(adx14), len(adx20), len(willr), len(rsi6), len(rsi12), len(macd), len(macd_signal), len(macd_hist), len(ema6), len(ema12), len(rocr3), len(rocr6), len(atr), len(obv), len(trix)) ohlcv = ohlcv[:min_length].drop(["Open", "High", "Low"], axis=1) ohlcv["MOM (1)"], ohlcv["MOM (3)"], ohlcv["ADX (14)"] = (pd.Series(mom1[:min_length])).values, (pd.Series(mom3[:min_length])).values, (pd.Series(adx14[:min_length])).values ohlcv["ADX (20)"], ohlcv["WILLR"], ohlcv["RSI (6)"] = (pd.Series(adx20[:min_length])).values, (pd.Series(willr[:min_length])).values, (pd.Series(rsi6[:min_length])).values ohlcv["RSI (12)"], ohlcv["MACD"], ohlcv["MACD (Signal)"] = (pd.Series(rsi12[:min_length])).values, (pd.Series(macd[:min_length])).values, (pd.Series(macd_signal[:min_length])).values ohlcv["MACD (Historical)"], ohlcv["EMA (6)"], ohlcv["EMA (12)"] = (pd.Series(macd_hist[:min_length])).values, (pd.Series(ema6[:min_length])).values, (pd.Series(ema12[:min_length])).values ohlcv["ROCR (3)"], ohlcv["ROCR (6)"], ohlcv["ATR (14)"] = (pd.Series(rocr3[:min_length])).values, (pd.Series(rocr6[:min_length])).values, (pd.Series(atr[:min_length])).values ohlcv["OBV"], ohlcv["TRIX (20)"] = (pd.Series(obv[:min_length])).values, (pd.Series(trix[:min_length])).values return ohlcv def
(headlines): sentiment_scores = {} numer, denom = 0.0, 0.0 for index, currRow in headlines.iterrows(): print(currRow) currDate = currRow["Date"] if currDate in sentiment_scores: pass else: numer = currRow["Sentiment"] * currRow["Tweets"] denom = currRow["Tweets"] for index, nextRow in headlines.iloc[index + 1:].iterrows(): if nextRow["Date"] == currDate: numer += (nextRow["Sentiment"] * nextRow["Tweets"]) denom += nextRow["Tweets"] else: break sentiment_scores[currDate] = numer / denom numer, denom = 0.0, 0.0 sentiment_scores_df = pd.DataFrame(list(sentiment_scores.items()), columns=["Date", "Sentiment"]) return sentiment_scores_df def merge_datasets(origin, other_sets): print("\tMerging datasets") merged = origin for set in other_sets: merged = pd.merge(merged, set, on="Date") return merged def fix_null_vals(dataset): """Implements the Last Observation Carried Forward (LOCF) method to fill missing values.""" print("\tFixing null values") if not dataset.isnull().any().any(): return dataset else: return dataset.fillna(method="ffill") def binarize_labels(dataset): """Transforms daily price data into binary values indicating price change.""" print("\tBinarizing price movements") trends = [None] for index in range(dataset.shape[0] - 1): difference = dataset.iloc[index]["Close"] - dataset.iloc[index + 1]["Close"] if difference < 0: trends.append(-1) else: trends.append(1) dataset["Trend"] = (pd.Series(trends)).values dataset = dataset.drop(dataset.index[0]) return dataset def add_lag_variables(dataset, lag=3): print("\tAdding lag variables") new_df_dict = {} for col_header in dataset.drop(["Date", "Trend"], axis=1): new_df_dict[col_header] = dataset[col_header] for lag in range(1, lag + 1): new_df_dict["%s_lag%d" % (col_header, lag)] = dataset[col_header].shift(-lag) new_df = pd.DataFrame(new_df_dict, index=dataset.index) new_df["Date"], new_df["Trend"] = dataset["Date"], dataset["Trend"] return new_df.dropna() def power_transform(dataset): print("\tApplying a box-cox transform to selected features") for header in dataset.drop(["Date", "Trend"], axis=1).columns: if not (dataset[header] < 0).any() and not (dataset[header] == 0).any(): dataset[header] = boxcox(dataset[header])[0] return dataset def split(dataset, test_size, balanced=True): # TODO: Splits can't be random, they need to respect the temporal order of each observation if balanced: return balanced_split(dataset, test_size) else: return unbalanced_split(dataset, test_size) def integral_transform(dataset, interval): integral = integrate(list(dataset["Sentiment"]), interval) dataset["Sentiment_integrals"] = pd.Series(integral)
calculate_sentiment
identifier_name
preprocessing.py
# Globals # import numpy as np import pandas as pd import dateutil.parser as dp from itertools import islice from scipy.stats import boxcox from scipy.integrate import simps from realtime_talib import Indicator from sklearn.model_selection import train_test_split from sklearn.utils import resample from pprint import pprint from selenium import webdriver RANDOM_STATE = 42 # Helpers # def sliding_window(seq, n=2): """Returns a sliding window (of width n) over data from the iterable. https://stackoverflow.com/a/6822773/8740440""" "s -> (s0,s1,...s[n-1]), (s1,s2,...,sn), ..." it = iter(seq) result = tuple(islice(it, n)) if len(result) == n: yield result for elem in it: result = result[1:] + (elem,) yield result def integrate(avg_daily_sentiment, interval): """Takes a list of average daily sentiment scores and returns a list of definite integral estimations calculated with Simpson's method. Each integral interval is determined by the `interval` variable. Shows accumulated sentiment.""" # Split into sliding window list of lists sentiment_windows = sliding_window(avg_daily_sentiment, interval) integral_simpson_est = [] # https://stackoverflow.com/a/13323861/8740440 for x in sentiment_windows: # Estimate area using composite Simpson's rule. dx indicates the spacing of the data on the x-axis. integral_simpson_est.append(simps(x, dx=1)) dead_values = list([None] * interval) dead_values.extend(integral_simpson_est) dead_values.reverse() return dead_values def random_undersampling(dataset): """Randomly deleting rows that contain the majority class until the number in the majority class is equal with the number in the minority class.""" minority_set = dataset[dataset.Trend == -1.0] majority_set = dataset[dataset.Trend == 1.0] # print(dataset.Trend.value_counts()) # If minority set larger than majority set, swap if len(minority_set) > len(majority_set): minority_set, majority_set = majority_set, minority_set # Downsample majority class majority_downsampled = resample(majority_set, replace=False, # sample without replacement n_samples=len(minority_set), # to match minority class random_state=123) # reproducible results # Combine minority class with downsampled majority class return pd.concat([majority_downsampled, minority_set]) def get_popularity(headlines): # TODO: Randomize user-agents OR figure out how to handle popups if "Tweets" not in headlines.columns: counts = [] driver = webdriver.Chrome() for index, row in headlines.iterrows(): try: driver.get(row["URL"]) time.sleep(3) twitter_containers = driver.find_elements_by_xpath("//li[@class='twitter']") count = twitter_containers[0].find_elements_by_xpath("//span[@class='count']") if count[0].text == "": counts.append(1) else: counts.append(int(count[0].text)) except: counts.append(1) # QUESTION: Should it be None? headlines["Tweets"] = (pd.Series(counts)).values print(counts) return headlines def balanced_split(dataset, test_size): """Randomly splits dataset into balanced training and test sets.""" print("\tSplitting data into *balanced* training and test sets") # Use sklearn.train_test_split to split original dataset into x_train, y_train, x_test, y_test numpy arrays x_train, x_test, y_train, y_test = train_test_split(dataset.drop(["Date", "Trend"], axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) # Combine x_train and y_train (numpy arrays) into a single dataframe, with column labels train = pd.DataFrame(data=x_train, columns=dataset.columns[1:-1]) train["Trend"] = pd.Series(y_train) # Do the same for x_test and y__test test = pd.DataFrame(data=x_test, columns=dataset.columns[1:-1]) test["Trend"] = pd.Series(y_test) # Apply random undersampling to both data frames train_downsampled = random_undersampling(train) test_downsampled = random_undersampling(test) train_trend = train_downsampled["Trend"].values test_trend = test_downsampled["Trend"].values train_trimmed = train_downsampled.drop(["Trend"], axis=1).values test_trimmed = test_downsampled.drop(["Trend"], axis=1).values return train_trimmed, test_trimmed, train_trend, test_trend def unbalanced_split(dataset, test_size): """Randomly splits dataset into unbalanced training and test sets.""" print("\tSplitting data into *unbalanced* training and test sets") dataset = dataset.drop("Date", axis=1) output = train_test_split(dataset.drop("Trend", axis=1).values, dataset["Trend"].values, test_size=test_size, random_state=RANDOM_STATE) return output # Main # def calculate_indicators(ohlcv): """Extracts technical indicators from OHLCV data.""" print("\tCalculating technical indicators") ohlcv = ohlcv.drop(["Volume (BTC)", "Weighted Price"], axis=1) ohlcv.columns = ["Date", "Open", "High", "Low", "Close", "Volume"] temp_ohlcv = ohlcv.copy() # Converts ISO 8601 timestamps to UNIX unix_times = [int((dp.parse(temp_ohlcv.iloc[index]["Date"])).strftime("%s")) for index in range(temp_ohlcv.shape[0])] temp_ohlcv["Date"] = (pd.Series(unix_times)).values # Converts column headers to lowercase and sorts rows in chronological order temp_ohlcv.columns = ["date", "open", "high", "low", "close", "volume"] temp_ohlcv = temp_ohlcv.iloc[::-1] # Rate of Change Ratio rocr3 = ((Indicator(temp_ohlcv, "ROCR", 3)).getHistorical())[::-1] rocr6 = ((Indicator(temp_ohlcv, "ROCR", 6)).getHistorical())[::-1] # Average True Range atr = ((Indicator(temp_ohlcv, "ATR", 14)).getHistorical())[::-1] # On-Balance Volume obv = ((Indicator(temp_ohlcv, "OBV")).getHistorical())[::-1] # Triple Exponential Moving Average trix = ((Indicator(temp_ohlcv, "TRIX", 20)).getHistorical())[::-1] # Momentum mom1 = ((Indicator(temp_ohlcv, "MOM", 1)).getHistorical())[::-1] mom3 = ((Indicator(temp_ohlcv, "MOM", 3)).getHistorical())[::-1] # Average Directional Index adx14 = ((Indicator(temp_ohlcv, "ADX", 14)).getHistorical())[::-1] adx20 = ((Indicator(temp_ohlcv, "ADX", 20)).getHistorical())[::-1] # Williams %R willr = ((Indicator(temp_ohlcv, "WILLR", 14)).getHistorical())[::-1] # Relative Strength Index rsi6 = ((Indicator(temp_ohlcv, "RSI", 6)).getHistorical())[::-1] rsi12 = ((Indicator(temp_ohlcv, "RSI", 12)).getHistorical())[::-1] # Moving Average Convergence Divergence macd, macd_signal, macd_hist = (Indicator(temp_ohlcv, "MACD", 12, 26, 9)).getHistorical() macd, macd_signal, macd_hist = macd[::-1], macd_signal[::-1], macd_hist[::-1] # Exponential Moving Average ema6 = ((Indicator(temp_ohlcv, "MA", 6, 1)).getHistorical())[::-1] ema12 = ((Indicator(temp_ohlcv, "MA", 12, 1)).getHistorical())[::-1] # Append indicators to the input datasets min_length = min(len(mom1), len(mom3), len(adx14), len(adx20), len(willr), len(rsi6), len(rsi12), len(macd), len(macd_signal), len(macd_hist), len(ema6), len(ema12), len(rocr3), len(rocr6), len(atr), len(obv), len(trix)) ohlcv = ohlcv[:min_length].drop(["Open", "High", "Low"], axis=1) ohlcv["MOM (1)"], ohlcv["MOM (3)"], ohlcv["ADX (14)"] = (pd.Series(mom1[:min_length])).values, (pd.Series(mom3[:min_length])).values, (pd.Series(adx14[:min_length])).values ohlcv["ADX (20)"], ohlcv["WILLR"], ohlcv["RSI (6)"] = (pd.Series(adx20[:min_length])).values, (pd.Series(willr[:min_length])).values, (pd.Series(rsi6[:min_length])).values ohlcv["RSI (12)"], ohlcv["MACD"], ohlcv["MACD (Signal)"] = (pd.Series(rsi12[:min_length])).values, (pd.Series(macd[:min_length])).values, (pd.Series(macd_signal[:min_length])).values ohlcv["MACD (Historical)"], ohlcv["EMA (6)"], ohlcv["EMA (12)"] = (pd.Series(macd_hist[:min_length])).values, (pd.Series(ema6[:min_length])).values, (pd.Series(ema12[:min_length])).values ohlcv["ROCR (3)"], ohlcv["ROCR (6)"], ohlcv["ATR (14)"] = (pd.Series(rocr3[:min_length])).values, (pd.Series(rocr6[:min_length])).values, (pd.Series(atr[:min_length])).values ohlcv["OBV"], ohlcv["TRIX (20)"] = (pd.Series(obv[:min_length])).values, (pd.Series(trix[:min_length])).values return ohlcv def calculate_sentiment(headlines): sentiment_scores = {} numer, denom = 0.0, 0.0 for index, currRow in headlines.iterrows(): print(currRow) currDate = currRow["Date"] if currDate in sentiment_scores: pass else: numer = currRow["Sentiment"] * currRow["Tweets"] denom = currRow["Tweets"] for index, nextRow in headlines.iloc[index + 1:].iterrows(): if nextRow["Date"] == currDate:
else: break sentiment_scores[currDate] = numer / denom numer, denom = 0.0, 0.0 sentiment_scores_df = pd.DataFrame(list(sentiment_scores.items()), columns=["Date", "Sentiment"]) return sentiment_scores_df def merge_datasets(origin, other_sets): print("\tMerging datasets") merged = origin for set in other_sets: merged = pd.merge(merged, set, on="Date") return merged def fix_null_vals(dataset): """Implements the Last Observation Carried Forward (LOCF) method to fill missing values.""" print("\tFixing null values") if not dataset.isnull().any().any(): return dataset else: return dataset.fillna(method="ffill") def binarize_labels(dataset): """Transforms daily price data into binary values indicating price change.""" print("\tBinarizing price movements") trends = [None] for index in range(dataset.shape[0] - 1): difference = dataset.iloc[index]["Close"] - dataset.iloc[index + 1]["Close"] if difference < 0: trends.append(-1) else: trends.append(1) dataset["Trend"] = (pd.Series(trends)).values dataset = dataset.drop(dataset.index[0]) return dataset def add_lag_variables(dataset, lag=3): print("\tAdding lag variables") new_df_dict = {} for col_header in dataset.drop(["Date", "Trend"], axis=1): new_df_dict[col_header] = dataset[col_header] for lag in range(1, lag + 1): new_df_dict["%s_lag%d" % (col_header, lag)] = dataset[col_header].shift(-lag) new_df = pd.DataFrame(new_df_dict, index=dataset.index) new_df["Date"], new_df["Trend"] = dataset["Date"], dataset["Trend"] return new_df.dropna() def power_transform(dataset): print("\tApplying a box-cox transform to selected features") for header in dataset.drop(["Date", "Trend"], axis=1).columns: if not (dataset[header] < 0).any() and not (dataset[header] == 0).any(): dataset[header] = boxcox(dataset[header])[0] return dataset def split(dataset, test_size, balanced=True): # TODO: Splits can't be random, they need to respect the temporal order of each observation if balanced: return balanced_split(dataset, test_size) else: return unbalanced_split(dataset, test_size) def integral_transform(dataset, interval): integral = integrate(list(dataset["Sentiment"]), interval) dataset["Sentiment_integrals"] = pd.Series(integral)
numer += (nextRow["Sentiment"] * nextRow["Tweets"]) denom += nextRow["Tweets"]
conditional_block
networking.go
/* <!-- Copyright (c) 2016 Christoph Berger. Some rights reserved. Use of this text is governed by a Creative Commons Attribution Non-Commercial Share-Alike License that can be found in the LICENSE.txt file. The source code contained in this file may import third-party source code whose licenses are provided in the respective license files. --> <!-- NOTE: The comments in this file are NOT godoc compliant. This is not an oversight. Comments and code in this file are used for describing and explaining a particular topic to the reader. While this file is a syntactically valid Go source file, its main purpose is to get converted into a blog article. The comments were created for learning and not for code documentation. --> +++ title = "TCP/IP Networking" description = "How to communicate at TCP/IP level in Go" author = "Christoph Berger" email = "chris@appliedgo.net" date = "2017-01-25" publishdate = "2017-01-25" draft = "false" categories = ["Distributed Computing"] tags = ["network", "tcp", "gob"] articletypes = ["Tutorial"] +++ Connecting two processes at TCP/IP level might seem scary at first, but in Go it is easier than one might think. <!--more--> While preparing another blog post, I realized that the networking part of the code was quickly becoming larger than the part of the code that was meant to illustrate the topic of the post. So I decided to write one blog post entirely about how to send data from process A to process B over a plain TCP/IP connection. ## Who needs sending things at TCP/IP level? Granted, many, if not most, scenarios, undoubtedly do better with a higher-level network protocol that hides all the technical details beneath a fancy API. And there are already plenty to choose from, depending on the needs: Message queue protocols, gRPC, protobuf, FlatBuffers, RESTful Web API's, WebSockets, and so on. However, in some situations (especially with small projects), any approach you choose may look like completely oversized, not to mention the additional package dependencies that you'd have to introduce. Luckily, creating simple network communication with [the standard `net` package](https://golang.org/pkg/net/) is not as difficult as it may seem. ## Simplification #1: connections are io streams The `net.Conn` interface implements [the `io.Reader`, `io.Writer`, and `io.Closer` interfaces](https://golang.org/pkg/io/). Hence you can use a TCP connection like any `io` stream. I know what you think -- "Ok, so I can send strings or byte slices over a TCP connection. That's nice but what about complex data types? Structs and such?" ## Simplification #2: Go knows how to encode complex types efficiently *(Did you also just read "God knows..."? I think it happens to me almost every other time I read this text.)* When it comes to encoding structured data for sending over the net, JSON comes readily to mind. But wait - Go's standard `encoding/gob` package provides a way of serializing and deserializing Go data types without the need for adding string tags to structs, dealing with JSON/Go incompatibilities, or waiting for json.Unmarshal to laboriously parse text into binary data. Gob encoders and decoders work directly on `io` streams - and this fits just nicely into our simplification #1 - connections are `io` streams. Let's put this all together in a small sample app. ## The sample app's goal The app shall do two things: 1. Send and receive a simple message as a string 2. Send and receive a `struct` via GOB The first part, sending simple strings, shall demonstrate how easy it is to send data over a TCP/IP network without any higher-level protocols. The second part goes a step further and sends a complete struct over the network, with strings, slices, maps, and even a recursive pointer to the struct itself. Thanks to the `gob` package, this requires no efforts. The following animation shows how gob data gets from a client to a server, and when this looks quite unspectacular, it's because using `gob` *is* unspectacular. HYPE[Sending a struct as GOB](gob.html) It's not much more than that! ## Basic ingredients for sending string data over TCP ### On the sending side Sending strings requires three simple steps. 1. Open a connection to the receiving process 2. Write the string 3. Close the connection The `net` package provides a couple of methods for this. `ResolveTCPAddr()` takes a string representing a TCP address (like, for example, `localhost:80`, `127.0.0.1:80`, or `[::1]:80`, which all represent port #80 on the local machine) and returns a `net.TCPAddr` (or an error if the string cannot be resolved to a valid TCP address). `DialTCP()` takes a `net.TCPAddr` and connects to this address. It returns the open connection as a `net.TCPConn` object (or an error if the connection attempt fails). If we don't need much fine-grained control over the Dial settings, we can use `net.Dial()` instead. This function takes an address string directly and returns a general `net.Conn` object. This is sufficient for our test case. However, if you need functionality that is only available on TCP connections, you have to use the "TCP" variants (`DialTCP`, `TCPConn`, `TCPAddr`, etc). After successful dialing, we can treat the new connection like any other input/output stream, as mentioned above. We can even wrap the connection into a `bufio.ReadWriter` and benefit from the various `ReadWriter` methods like `ReadString()`, `ReadBytes`, `WriteString`, etc. ** Remember that buffered Writers need to call `Flush()` after writing, so that all data is forwarded to the underlying network connection.** Finally, each connection object has a `Close()` method to conclude the communication. ### Fine tuning A couple of tuning options are also available. Some examples: The `Dialer` interface provides these options (among others): * `DeadLine` and `Timeout` options for timing out an unsuccessful dial; * `KeepAlive` option for managing the life span of the connection The `Conn` interface also has deadline settings; either for the connection as a whole (`SetDeadLine()`), or specific to read or write calls (`SetReadDeadLine()` and `SetWriteDeadLine()`). Note that the deadlines are fixed points in (wallclock) time. Unlike timeouts, they don't reset after a new activity. Each activity on the connection must therefore set a new deadline. The sample code below uses no deadlines, as it is simple enough so that we can easily see when things get stuck. `Ctrl-C` is our manual "deadline trigger tool". ### On the receiving side The receiver has to follow these steps. 1. Start listening on a local port. 2. When a request comes in, spawn a goroutine to handle the request. 3. In the goroutine, read the data. Optionally, send a response. 4. Close the connection. Listening requires a local port to listen to. Typically, the listening application (a.k.a. "server") announces the port it listens to, or if it provides a standard service, it uses the port associated with that service. For example, Web servers usually listen on port 80 for HTTP requests and on port 443 for HTTPS requests. SSH daemons listen on port 22 by default, and a WHOIS server uses port 43. The core parts of the `net` package for implementing the server side are: `net.Listen()` creates a new listener on a given local network address. If only a port is passed, as in ":61000", then the listener listens on all available network interfaces. This is quite handy, as a computer usually has at least two active interfaces, the loopback interface and at least one real network card. A listener's `Accept()` method waits until a connection request comes in. Then it accepts the request and returns the new connection to the caller. `Accept()` is typically called within a loop to be able to serve multiple connections simultaneously. Each connection can be handled by a goroutine, as we will see in the code. ## The code Instead of just pushing a few bytes around, I wanted the code to demonstrate something more useful. I want to be able to send different commands with a different data payload to the server. The server shall identify each command and decode the command's data. So the client in the code below sends two test commands: "STRING" and "GOB". Each are terminated by a newline. The STRING command includes one line of string data, which can be handled by simple read and write methods from `bufio`. The GOB command comes with a `struct` that contains a couple of fields, including a slice, a map, and a even a pointer to itself. As you can see when running the code, the `gob` package moves all this through our network connection without any fuss. What we basically have here is some sort of ad-hoc protocol, where the client and the server agree that a command is a string followed by a newline followed by some data. For each command, the server must know the exact data format and how to process the data. To achieve this, the server code takes a two-step approach. Step 1: When the `Listen()` function accepts a new connection, it spawns a new goroutine that calls function `handleMessage()`. This function reads the command name from the connection, looks up the appropriate handler function from a map, and calls this function. Step 2: The selected handler function reads and processes the command's data. Here is a visual summary of this process. HYPE[Server Command Dispatch](tcpserver.html) Keep these pictures in mind, they help reading the actual code. ## The Code */ // ## Imports and globals package main import ( "bufio" "io" "log" "net" "strconv" "strings" "sync" "github.com/pkg/errors" "encoding/gob" "flag" ) // A struct with a mix of fields, used for the GOB example. type complexData struct { N int S string M map[string]int P []byte C *complexData } const ( // Port is the port number that the server listens to. Port = ":61000" ) /* ## Outgoing connections Using an outgoing connection is a snap. A `net.Conn` satisfies the io.Reader and `io.Writer` interfaces, so we can treat a TCP connection just like any other `Reader` or `Writer`. */ // Open connects to a TCP Address. // It returns a TCP connection armed with a timeout and wrapped into a // buffered ReadWriter. func Open(addr string) (*bufio.ReadWriter, error) { // Dial the remote process. // Note that the local port is chosen on the fly. If the local port // must be a specific one, use DialTCP() instead. log.Println("Dial " + addr) conn, err := net.Dial("tcp", addr) if err != nil { return nil, errors.Wrap(err, "Dialing "+addr+" failed") } return bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)), nil } /* ## Incoming connections Preparing for incoming data is a bit more involved. According to our ad-hoc protocol, we receive the name of a command terminated by `\n`, followed by data. The nature of the data depends on the respective command. To handle this, we create an `Endpoint` object with the following properties: * It allows to register one or more handler functions, where each can handle a particular command. * It dispatches incoming commands to the associated handler based on the commands name. */ // HandleFunc is a function that handles an incoming command. // It receives the open connection wrapped in a `ReadWriter` interface. type HandleFunc func(*bufio.ReadWriter) // Endpoint provides an endpoint to other processess // that they can send data to. type Endpoint struct { listener net.Listener handler map[string]HandleFunc // Maps are not threadsafe, so we need a mutex to control access. mutex sync.RWMutex } // NewEndpoint creates a new endpoint. To keep things simple, // the endpoint listens on a fixed port number. func NewEndpoint() *Endpoint { // Create a new Endpoint with an empty list of handler funcs. return &Endpoint{ handler: map[string]HandleFunc{}, } } // AddHandleFunc adds a new function for handling incoming data. func (e *Endpoint) AddHandleFunc(name string, f HandleFunc) { e.mutex.Lock() e.handler[name] = f e.mutex.Unlock() } // Listen starts listening on the endpoint port on all interfaces. // At least one handler function must have been added // through AddHandleFunc() before. func (e *Endpoint) Listen() error { var err error e.listener, err = net.Listen("tcp", Port) if err != nil { return errors.Wrapf(err, "Unable to listen on port %s\n", Port) } log.Println("Listen on", e.listener.Addr().String()) for { log.Println("Accept a connection request.") conn, err := e.listener.Accept() if err != nil { log.Println("Failed accepting a connection request:", err) continue } log.Println("Handle incoming messages.") go e.handleMessages(conn) } } // handleMessages reads the connection up to the first newline. // Based on this string, it calls the appropriate HandleFunc. func (e *Endpoint) handleMessages(conn net.Conn) { // Wrap the connection into a buffered reader for easier reading. rw := bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)) defer conn.Close() // Read from the connection until EOF. Expect a command name as the // next input. Call the handler that is registered for this command. for { log.Print("Receive command '") cmd, err := rw.ReadString('\n') switch { case err == io.EOF: log.Println("Reached EOF - close this connection.\n ---") return case err != nil: log.Println("\nError reading command. Got: '"+cmd+"'\n", err) return } // Trim the request string - ReadString does not strip any newlines. cmd = strings.Trim(cmd, "\n ") log.Println(cmd + "'") // Fetch the appropriate handler function from the 'handler' map and call it. e.mutex.RLock() handleCommand, ok := e.handler[cmd] e.mutex.RUnlock() if !ok { log.Println("Command '" + cmd + "' is not registered.") return } handleCommand(rw) } } /* Now let's create two handler functions. The easiest case is where our ad-hoc protocol only sends string data. The second handler receives and processes a struct that was sent as GOB data. */ // handleStrings handles the "STRING" request. func handleStrings(rw *bufio.ReadWriter) { // Receive a string. log.Print("Receive STRING message:") s, err := rw.ReadString('\n') if err != nil { log.Println("Cannot read from connection.\n", err) } s = strings.Trim(s, "\n ") log.Println(s) _, err = rw.WriteString("Thank you.\n") if err != nil { log.Println("Cannot write to connection.\n", err) } err = rw.Flush() if err != nil { log.Println("Flush failed.", err) } } // handleGob handles the "GOB" request. It decodes the received GOB data // into a struct. func handleGob(rw *bufio.ReadWriter) { log.Print("Receive GOB data:") var data complexData // Create a decoder that decodes directly into a struct variable. dec := gob.NewDecoder(rw) err := dec.Decode(&data) if err != nil { log.Println("Error decoding GOB data:", err) return } // Print the complexData struct and the nested one, too, to prove // that both travelled across the wire. log.Printf("Outer complexData struct: \n%#v\n", data) log.Printf("Inner complexData struct: \n%#v\n", data.C) } /* ## The client and server functions With all this in place, we can now set up client and server functions. The client function connects to the server and sends STRING and GOB requests. The server starts listening for requests and triggers the appropriate handlers. */ // client is called if the app is called with -connect=`ip addr`. func client(ip string) error { // Some test data. Note how GOB even handles maps, slices, and // recursive data structures without problems. testStruct := complexData{ N: 23, S: "string data", M: map[string]int{"one": 1, "two": 2, "three": 3}, P: []byte("abc"), C: &complexData{ N: 256, S: "Recursive structs? Piece of cake!", M: map[string]int{"01": 1, "10": 2, "11": 3}, }, } // Open a connection to the server. rw, err := Open(ip + Port) if err != nil { return errors.Wrap(err, "Client: Failed to open connection to "+ip+Port) } // Send a STRING request. // Send the request name. // Send the data. log.Println("Send the string request.") n, err := rw.WriteString("STRING\n") if err != nil { return errors.Wrap(err, "Could not send the STRING request ("+strconv.Itoa(n)+" bytes written)") } n, err = rw.WriteString("Additional data.\n") if err != nil
log.Println("Flush the buffer.") err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } // Read the reply. log.Println("Read the reply.") response, err := rw.ReadString('\n') if err != nil { return errors.Wrap(err, "Client: Failed to read the reply: '"+response+"'") } log.Println("STRING request: got a response:", response) // Send a GOB request. // Create an encoder that directly transmits to `rw`. // Send the request name. // Send the GOB. log.Println("Send a struct as GOB:") log.Printf("Outer complexData struct: \n%#v\n", testStruct) log.Printf("Inner complexData struct: \n%#v\n", testStruct.C) enc := gob.NewEncoder(rw) n, err = rw.WriteString("GOB\n") if err != nil { return errors.Wrap(err, "Could not write GOB data ("+strconv.Itoa(n)+" bytes written)") } err = enc.Encode(testStruct) if err != nil { return errors.Wrapf(err, "Encode failed for struct: %#v", testStruct) } err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } return nil } // server listens for incoming requests and dispatches them to // registered handler functions. func server() error { endpoint := NewEndpoint() // Add the handle funcs. endpoint.AddHandleFunc("STRING", handleStrings) endpoint.AddHandleFunc("GOB", handleGob) // Start listening. return endpoint.Listen() } /* ## Main Main starts either a client or a server, depending on whether the `connect` flag is set. Without the flag, the process starts as a server, listening for incoming requests. With the flag the process starts as a client and connects to the host specified by the flag value. Try "localhost" or "127.0.0.1" when running both processes on the same machine. */ // main func main() { connect := flag.String("connect", "", "IP address of process to join. If empty, go into listen mode.") flag.Parse() // If the connect flag is set, go into client mode. if *connect != "" { err := client(*connect) if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Client done.") return } // Else go into server mode. err := server() if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Server done.") } // The Lshortfile flag includes file name and line number in log messages. func init() { log.SetFlags(log.Lshortfile) } /* ## How to get and run the code Step 1: `go get` the code. Note the `-d` flag that prevents auto-installing the binary into `$GOPATH/bin`. go get -d github.com/appliedgo/networking Step 2: `cd` to the source code directory. cd $GOPATH/src/github.com/appliedgo/networking Step 3. Run the server. go run networking.go Step 4. Open another shell, `cd` to the source code (see Step 2), and run the client. go run networking.go -connect localhost ## Tips If you want to tinker with the code a bit, here are some suggestions: * Try running client and server on different machines (in the same local network). * Beef up the complexData type with more maps and pointers and see how `gob` copes with it. * Start several clients at the same time and see if the server can handle them. ## Links This turned into quite a long blog post, so if you are looking for something shorter, here is a blog post that is really just the essence of the above, and it is just sending strings. No gobs, and no fancy "command/data" constructs. * [A Simple Go TCP Server and TCP Client](https://systembash.com/a-simple-go-tcp-server-and-tcp-client/) More about the `gob` package: * [Gobs of data](https://blog.golang.org/gobs-of-data) **Happy coding!** - - - Errata 2017-02-09 - Map access: Maps are not thread-safe and thus if a map is used in different goroutines, a mutex should always control access to a map. In the given code, the map is updated before the goroutine starts, so mutexes were not necessary. Nevertheless, I now added one, so you can now safely modify the code and call AddHandleFunc() while the handleMessages goroutine is already running. */
{ return errors.Wrap(err, "Could not send additional STRING data ("+strconv.Itoa(n)+" bytes written)") }
conditional_block
networking.go
/* <!-- Copyright (c) 2016 Christoph Berger. Some rights reserved. Use of this text is governed by a Creative Commons Attribution Non-Commercial Share-Alike License that can be found in the LICENSE.txt file. The source code contained in this file may import third-party source code whose licenses are provided in the respective license files. --> <!-- NOTE: The comments in this file are NOT godoc compliant. This is not an oversight. Comments and code in this file are used for describing and explaining a particular topic to the reader. While this file is a syntactically valid Go source file, its main purpose is to get converted into a blog article. The comments were created for learning and not for code documentation. --> +++ title = "TCP/IP Networking" description = "How to communicate at TCP/IP level in Go" author = "Christoph Berger" email = "chris@appliedgo.net" date = "2017-01-25" publishdate = "2017-01-25" draft = "false" categories = ["Distributed Computing"] tags = ["network", "tcp", "gob"] articletypes = ["Tutorial"] +++ Connecting two processes at TCP/IP level might seem scary at first, but in Go it is easier than one might think. <!--more--> While preparing another blog post, I realized that the networking part of the code was quickly becoming larger than the part of the code that was meant to illustrate the topic of the post. So I decided to write one blog post entirely about how to send data from process A to process B over a plain TCP/IP connection. ## Who needs sending things at TCP/IP level? Granted, many, if not most, scenarios, undoubtedly do better with a higher-level network protocol that hides all the technical details beneath a fancy API. And there are already plenty to choose from, depending on the needs: Message queue protocols, gRPC, protobuf, FlatBuffers, RESTful Web API's, WebSockets, and so on. However, in some situations (especially with small projects), any approach you choose may look like completely oversized, not to mention the additional package dependencies that you'd have to introduce. Luckily, creating simple network communication with [the standard `net` package](https://golang.org/pkg/net/) is not as difficult as it may seem. ## Simplification #1: connections are io streams The `net.Conn` interface implements [the `io.Reader`, `io.Writer`, and `io.Closer` interfaces](https://golang.org/pkg/io/). Hence you can use a TCP connection like any `io` stream. I know what you think -- "Ok, so I can send strings or byte slices over a TCP connection. That's nice but what about complex data types? Structs and such?" ## Simplification #2: Go knows how to encode complex types efficiently *(Did you also just read "God knows..."? I think it happens to me almost every other time I read this text.)* When it comes to encoding structured data for sending over the net, JSON comes readily to mind. But wait - Go's standard `encoding/gob` package provides a way of serializing and deserializing Go data types without the need for adding string tags to structs, dealing with JSON/Go incompatibilities, or waiting for json.Unmarshal to laboriously parse text into binary data. Gob encoders and decoders work directly on `io` streams - and this fits just nicely into our simplification #1 - connections are `io` streams. Let's put this all together in a small sample app. ## The sample app's goal The app shall do two things: 1. Send and receive a simple message as a string 2. Send and receive a `struct` via GOB The first part, sending simple strings, shall demonstrate how easy it is to send data over a TCP/IP network without any higher-level protocols. The second part goes a step further and sends a complete struct over the network, with strings, slices, maps, and even a recursive pointer to the struct itself. Thanks to the `gob` package, this requires no efforts. The following animation shows how gob data gets from a client to a server, and when this looks quite unspectacular, it's because using `gob` *is* unspectacular. HYPE[Sending a struct as GOB](gob.html) It's not much more than that! ## Basic ingredients for sending string data over TCP ### On the sending side Sending strings requires three simple steps. 1. Open a connection to the receiving process 2. Write the string 3. Close the connection The `net` package provides a couple of methods for this. `ResolveTCPAddr()` takes a string representing a TCP address (like, for example, `localhost:80`, `127.0.0.1:80`, or `[::1]:80`, which all represent port #80 on the local machine) and returns a `net.TCPAddr` (or an error if the string cannot be resolved to a valid TCP address). `DialTCP()` takes a `net.TCPAddr` and connects to this address. It returns the open connection as a `net.TCPConn` object (or an error if the connection attempt fails). If we don't need much fine-grained control over the Dial settings, we can use `net.Dial()` instead. This function takes an address string directly and returns a general `net.Conn` object. This is sufficient for our test case. However, if you need functionality that is only available on TCP connections, you have to use the "TCP" variants (`DialTCP`, `TCPConn`, `TCPAddr`, etc). After successful dialing, we can treat the new connection like any other input/output stream, as mentioned above. We can even wrap the connection into a `bufio.ReadWriter` and benefit from the various `ReadWriter` methods like `ReadString()`, `ReadBytes`, `WriteString`, etc. ** Remember that buffered Writers need to call `Flush()` after writing, so that all data is forwarded to the underlying network connection.** Finally, each connection object has a `Close()` method to conclude the communication. ### Fine tuning A couple of tuning options are also available. Some examples: The `Dialer` interface provides these options (among others): * `DeadLine` and `Timeout` options for timing out an unsuccessful dial; * `KeepAlive` option for managing the life span of the connection The `Conn` interface also has deadline settings; either for the connection as a whole (`SetDeadLine()`), or specific to read or write calls (`SetReadDeadLine()` and `SetWriteDeadLine()`). Note that the deadlines are fixed points in (wallclock) time. Unlike timeouts, they don't reset after a new activity. Each activity on the connection must therefore set a new deadline. The sample code below uses no deadlines, as it is simple enough so that we can easily see when things get stuck. `Ctrl-C` is our manual "deadline trigger tool". ### On the receiving side The receiver has to follow these steps. 1. Start listening on a local port. 2. When a request comes in, spawn a goroutine to handle the request. 3. In the goroutine, read the data. Optionally, send a response. 4. Close the connection. Listening requires a local port to listen to. Typically, the listening application (a.k.a. "server") announces the port it listens to, or if it provides a standard service, it uses the port associated with that service. For example, Web servers usually listen on port 80 for HTTP requests and on port 443 for HTTPS requests. SSH daemons listen on port 22 by default, and a WHOIS server uses port 43. The core parts of the `net` package for implementing the server side are: `net.Listen()` creates a new listener on a given local network address. If only a port is passed, as in ":61000", then the listener listens on all available network interfaces. This is quite handy, as a computer usually has at least two active interfaces, the loopback interface and at least one real network card. A listener's `Accept()` method waits until a connection request comes in. Then it accepts the request and returns the new connection to the caller. `Accept()` is typically called within a loop to be able to serve multiple connections simultaneously. Each connection can be handled by a goroutine, as we will see in the code. ## The code Instead of just pushing a few bytes around, I wanted the code to demonstrate something more useful. I want to be able to send different commands with a different data payload to the server. The server shall identify each command and decode the command's data. So the client in the code below sends two test commands: "STRING" and "GOB". Each are terminated by a newline. The STRING command includes one line of string data, which can be handled by simple read and write methods from `bufio`. The GOB command comes with a `struct` that contains a couple of fields, including a slice, a map, and a even a pointer to itself. As you can see when running the code, the `gob` package moves all this through our network connection without any fuss. What we basically have here is some sort of ad-hoc protocol, where the client and the server agree that a command is a string followed by a newline followed by some data. For each command, the server must know the exact data format and how to process the data. To achieve this, the server code takes a two-step approach. Step 1: When the `Listen()` function accepts a new connection, it spawns a new goroutine that calls function `handleMessage()`. This function reads the command name from the connection, looks up the appropriate handler function from a map, and calls this function. Step 2: The selected handler function reads and processes the command's data. Here is a visual summary of this process. HYPE[Server Command Dispatch](tcpserver.html) Keep these pictures in mind, they help reading the actual code. ## The Code */ // ## Imports and globals package main import ( "bufio" "io" "log" "net" "strconv" "strings" "sync" "github.com/pkg/errors" "encoding/gob" "flag" ) // A struct with a mix of fields, used for the GOB example. type complexData struct { N int S string M map[string]int P []byte C *complexData } const ( // Port is the port number that the server listens to. Port = ":61000" ) /* ## Outgoing connections Using an outgoing connection is a snap. A `net.Conn` satisfies the io.Reader and `io.Writer` interfaces, so we can treat a TCP connection just like any other `Reader` or `Writer`. */ // Open connects to a TCP Address. // It returns a TCP connection armed with a timeout and wrapped into a // buffered ReadWriter. func Open(addr string) (*bufio.ReadWriter, error) { // Dial the remote process. // Note that the local port is chosen on the fly. If the local port // must be a specific one, use DialTCP() instead. log.Println("Dial " + addr) conn, err := net.Dial("tcp", addr) if err != nil { return nil, errors.Wrap(err, "Dialing "+addr+" failed") } return bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)), nil } /* ## Incoming connections Preparing for incoming data is a bit more involved. According to our ad-hoc protocol, we receive the name of a command terminated by `\n`, followed by data.
* It dispatches incoming commands to the associated handler based on the commands name. */ // HandleFunc is a function that handles an incoming command. // It receives the open connection wrapped in a `ReadWriter` interface. type HandleFunc func(*bufio.ReadWriter) // Endpoint provides an endpoint to other processess // that they can send data to. type Endpoint struct { listener net.Listener handler map[string]HandleFunc // Maps are not threadsafe, so we need a mutex to control access. mutex sync.RWMutex } // NewEndpoint creates a new endpoint. To keep things simple, // the endpoint listens on a fixed port number. func NewEndpoint() *Endpoint { // Create a new Endpoint with an empty list of handler funcs. return &Endpoint{ handler: map[string]HandleFunc{}, } } // AddHandleFunc adds a new function for handling incoming data. func (e *Endpoint) AddHandleFunc(name string, f HandleFunc) { e.mutex.Lock() e.handler[name] = f e.mutex.Unlock() } // Listen starts listening on the endpoint port on all interfaces. // At least one handler function must have been added // through AddHandleFunc() before. func (e *Endpoint) Listen() error { var err error e.listener, err = net.Listen("tcp", Port) if err != nil { return errors.Wrapf(err, "Unable to listen on port %s\n", Port) } log.Println("Listen on", e.listener.Addr().String()) for { log.Println("Accept a connection request.") conn, err := e.listener.Accept() if err != nil { log.Println("Failed accepting a connection request:", err) continue } log.Println("Handle incoming messages.") go e.handleMessages(conn) } } // handleMessages reads the connection up to the first newline. // Based on this string, it calls the appropriate HandleFunc. func (e *Endpoint) handleMessages(conn net.Conn) { // Wrap the connection into a buffered reader for easier reading. rw := bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)) defer conn.Close() // Read from the connection until EOF. Expect a command name as the // next input. Call the handler that is registered for this command. for { log.Print("Receive command '") cmd, err := rw.ReadString('\n') switch { case err == io.EOF: log.Println("Reached EOF - close this connection.\n ---") return case err != nil: log.Println("\nError reading command. Got: '"+cmd+"'\n", err) return } // Trim the request string - ReadString does not strip any newlines. cmd = strings.Trim(cmd, "\n ") log.Println(cmd + "'") // Fetch the appropriate handler function from the 'handler' map and call it. e.mutex.RLock() handleCommand, ok := e.handler[cmd] e.mutex.RUnlock() if !ok { log.Println("Command '" + cmd + "' is not registered.") return } handleCommand(rw) } } /* Now let's create two handler functions. The easiest case is where our ad-hoc protocol only sends string data. The second handler receives and processes a struct that was sent as GOB data. */ // handleStrings handles the "STRING" request. func handleStrings(rw *bufio.ReadWriter) { // Receive a string. log.Print("Receive STRING message:") s, err := rw.ReadString('\n') if err != nil { log.Println("Cannot read from connection.\n", err) } s = strings.Trim(s, "\n ") log.Println(s) _, err = rw.WriteString("Thank you.\n") if err != nil { log.Println("Cannot write to connection.\n", err) } err = rw.Flush() if err != nil { log.Println("Flush failed.", err) } } // handleGob handles the "GOB" request. It decodes the received GOB data // into a struct. func handleGob(rw *bufio.ReadWriter) { log.Print("Receive GOB data:") var data complexData // Create a decoder that decodes directly into a struct variable. dec := gob.NewDecoder(rw) err := dec.Decode(&data) if err != nil { log.Println("Error decoding GOB data:", err) return } // Print the complexData struct and the nested one, too, to prove // that both travelled across the wire. log.Printf("Outer complexData struct: \n%#v\n", data) log.Printf("Inner complexData struct: \n%#v\n", data.C) } /* ## The client and server functions With all this in place, we can now set up client and server functions. The client function connects to the server and sends STRING and GOB requests. The server starts listening for requests and triggers the appropriate handlers. */ // client is called if the app is called with -connect=`ip addr`. func client(ip string) error { // Some test data. Note how GOB even handles maps, slices, and // recursive data structures without problems. testStruct := complexData{ N: 23, S: "string data", M: map[string]int{"one": 1, "two": 2, "three": 3}, P: []byte("abc"), C: &complexData{ N: 256, S: "Recursive structs? Piece of cake!", M: map[string]int{"01": 1, "10": 2, "11": 3}, }, } // Open a connection to the server. rw, err := Open(ip + Port) if err != nil { return errors.Wrap(err, "Client: Failed to open connection to "+ip+Port) } // Send a STRING request. // Send the request name. // Send the data. log.Println("Send the string request.") n, err := rw.WriteString("STRING\n") if err != nil { return errors.Wrap(err, "Could not send the STRING request ("+strconv.Itoa(n)+" bytes written)") } n, err = rw.WriteString("Additional data.\n") if err != nil { return errors.Wrap(err, "Could not send additional STRING data ("+strconv.Itoa(n)+" bytes written)") } log.Println("Flush the buffer.") err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } // Read the reply. log.Println("Read the reply.") response, err := rw.ReadString('\n') if err != nil { return errors.Wrap(err, "Client: Failed to read the reply: '"+response+"'") } log.Println("STRING request: got a response:", response) // Send a GOB request. // Create an encoder that directly transmits to `rw`. // Send the request name. // Send the GOB. log.Println("Send a struct as GOB:") log.Printf("Outer complexData struct: \n%#v\n", testStruct) log.Printf("Inner complexData struct: \n%#v\n", testStruct.C) enc := gob.NewEncoder(rw) n, err = rw.WriteString("GOB\n") if err != nil { return errors.Wrap(err, "Could not write GOB data ("+strconv.Itoa(n)+" bytes written)") } err = enc.Encode(testStruct) if err != nil { return errors.Wrapf(err, "Encode failed for struct: %#v", testStruct) } err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } return nil } // server listens for incoming requests and dispatches them to // registered handler functions. func server() error { endpoint := NewEndpoint() // Add the handle funcs. endpoint.AddHandleFunc("STRING", handleStrings) endpoint.AddHandleFunc("GOB", handleGob) // Start listening. return endpoint.Listen() } /* ## Main Main starts either a client or a server, depending on whether the `connect` flag is set. Without the flag, the process starts as a server, listening for incoming requests. With the flag the process starts as a client and connects to the host specified by the flag value. Try "localhost" or "127.0.0.1" when running both processes on the same machine. */ // main func main() { connect := flag.String("connect", "", "IP address of process to join. If empty, go into listen mode.") flag.Parse() // If the connect flag is set, go into client mode. if *connect != "" { err := client(*connect) if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Client done.") return } // Else go into server mode. err := server() if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Server done.") } // The Lshortfile flag includes file name and line number in log messages. func init() { log.SetFlags(log.Lshortfile) } /* ## How to get and run the code Step 1: `go get` the code. Note the `-d` flag that prevents auto-installing the binary into `$GOPATH/bin`. go get -d github.com/appliedgo/networking Step 2: `cd` to the source code directory. cd $GOPATH/src/github.com/appliedgo/networking Step 3. Run the server. go run networking.go Step 4. Open another shell, `cd` to the source code (see Step 2), and run the client. go run networking.go -connect localhost ## Tips If you want to tinker with the code a bit, here are some suggestions: * Try running client and server on different machines (in the same local network). * Beef up the complexData type with more maps and pointers and see how `gob` copes with it. * Start several clients at the same time and see if the server can handle them. ## Links This turned into quite a long blog post, so if you are looking for something shorter, here is a blog post that is really just the essence of the above, and it is just sending strings. No gobs, and no fancy "command/data" constructs. * [A Simple Go TCP Server and TCP Client](https://systembash.com/a-simple-go-tcp-server-and-tcp-client/) More about the `gob` package: * [Gobs of data](https://blog.golang.org/gobs-of-data) **Happy coding!** - - - Errata 2017-02-09 - Map access: Maps are not thread-safe and thus if a map is used in different goroutines, a mutex should always control access to a map. In the given code, the map is updated before the goroutine starts, so mutexes were not necessary. Nevertheless, I now added one, so you can now safely modify the code and call AddHandleFunc() while the handleMessages goroutine is already running. */
The nature of the data depends on the respective command. To handle this, we create an `Endpoint` object with the following properties: * It allows to register one or more handler functions, where each can handle a particular command.
random_line_split
networking.go
/* <!-- Copyright (c) 2016 Christoph Berger. Some rights reserved. Use of this text is governed by a Creative Commons Attribution Non-Commercial Share-Alike License that can be found in the LICENSE.txt file. The source code contained in this file may import third-party source code whose licenses are provided in the respective license files. --> <!-- NOTE: The comments in this file are NOT godoc compliant. This is not an oversight. Comments and code in this file are used for describing and explaining a particular topic to the reader. While this file is a syntactically valid Go source file, its main purpose is to get converted into a blog article. The comments were created for learning and not for code documentation. --> +++ title = "TCP/IP Networking" description = "How to communicate at TCP/IP level in Go" author = "Christoph Berger" email = "chris@appliedgo.net" date = "2017-01-25" publishdate = "2017-01-25" draft = "false" categories = ["Distributed Computing"] tags = ["network", "tcp", "gob"] articletypes = ["Tutorial"] +++ Connecting two processes at TCP/IP level might seem scary at first, but in Go it is easier than one might think. <!--more--> While preparing another blog post, I realized that the networking part of the code was quickly becoming larger than the part of the code that was meant to illustrate the topic of the post. So I decided to write one blog post entirely about how to send data from process A to process B over a plain TCP/IP connection. ## Who needs sending things at TCP/IP level? Granted, many, if not most, scenarios, undoubtedly do better with a higher-level network protocol that hides all the technical details beneath a fancy API. And there are already plenty to choose from, depending on the needs: Message queue protocols, gRPC, protobuf, FlatBuffers, RESTful Web API's, WebSockets, and so on. However, in some situations (especially with small projects), any approach you choose may look like completely oversized, not to mention the additional package dependencies that you'd have to introduce. Luckily, creating simple network communication with [the standard `net` package](https://golang.org/pkg/net/) is not as difficult as it may seem. ## Simplification #1: connections are io streams The `net.Conn` interface implements [the `io.Reader`, `io.Writer`, and `io.Closer` interfaces](https://golang.org/pkg/io/). Hence you can use a TCP connection like any `io` stream. I know what you think -- "Ok, so I can send strings or byte slices over a TCP connection. That's nice but what about complex data types? Structs and such?" ## Simplification #2: Go knows how to encode complex types efficiently *(Did you also just read "God knows..."? I think it happens to me almost every other time I read this text.)* When it comes to encoding structured data for sending over the net, JSON comes readily to mind. But wait - Go's standard `encoding/gob` package provides a way of serializing and deserializing Go data types without the need for adding string tags to structs, dealing with JSON/Go incompatibilities, or waiting for json.Unmarshal to laboriously parse text into binary data. Gob encoders and decoders work directly on `io` streams - and this fits just nicely into our simplification #1 - connections are `io` streams. Let's put this all together in a small sample app. ## The sample app's goal The app shall do two things: 1. Send and receive a simple message as a string 2. Send and receive a `struct` via GOB The first part, sending simple strings, shall demonstrate how easy it is to send data over a TCP/IP network without any higher-level protocols. The second part goes a step further and sends a complete struct over the network, with strings, slices, maps, and even a recursive pointer to the struct itself. Thanks to the `gob` package, this requires no efforts. The following animation shows how gob data gets from a client to a server, and when this looks quite unspectacular, it's because using `gob` *is* unspectacular. HYPE[Sending a struct as GOB](gob.html) It's not much more than that! ## Basic ingredients for sending string data over TCP ### On the sending side Sending strings requires three simple steps. 1. Open a connection to the receiving process 2. Write the string 3. Close the connection The `net` package provides a couple of methods for this. `ResolveTCPAddr()` takes a string representing a TCP address (like, for example, `localhost:80`, `127.0.0.1:80`, or `[::1]:80`, which all represent port #80 on the local machine) and returns a `net.TCPAddr` (or an error if the string cannot be resolved to a valid TCP address). `DialTCP()` takes a `net.TCPAddr` and connects to this address. It returns the open connection as a `net.TCPConn` object (or an error if the connection attempt fails). If we don't need much fine-grained control over the Dial settings, we can use `net.Dial()` instead. This function takes an address string directly and returns a general `net.Conn` object. This is sufficient for our test case. However, if you need functionality that is only available on TCP connections, you have to use the "TCP" variants (`DialTCP`, `TCPConn`, `TCPAddr`, etc). After successful dialing, we can treat the new connection like any other input/output stream, as mentioned above. We can even wrap the connection into a `bufio.ReadWriter` and benefit from the various `ReadWriter` methods like `ReadString()`, `ReadBytes`, `WriteString`, etc. ** Remember that buffered Writers need to call `Flush()` after writing, so that all data is forwarded to the underlying network connection.** Finally, each connection object has a `Close()` method to conclude the communication. ### Fine tuning A couple of tuning options are also available. Some examples: The `Dialer` interface provides these options (among others): * `DeadLine` and `Timeout` options for timing out an unsuccessful dial; * `KeepAlive` option for managing the life span of the connection The `Conn` interface also has deadline settings; either for the connection as a whole (`SetDeadLine()`), or specific to read or write calls (`SetReadDeadLine()` and `SetWriteDeadLine()`). Note that the deadlines are fixed points in (wallclock) time. Unlike timeouts, they don't reset after a new activity. Each activity on the connection must therefore set a new deadline. The sample code below uses no deadlines, as it is simple enough so that we can easily see when things get stuck. `Ctrl-C` is our manual "deadline trigger tool". ### On the receiving side The receiver has to follow these steps. 1. Start listening on a local port. 2. When a request comes in, spawn a goroutine to handle the request. 3. In the goroutine, read the data. Optionally, send a response. 4. Close the connection. Listening requires a local port to listen to. Typically, the listening application (a.k.a. "server") announces the port it listens to, or if it provides a standard service, it uses the port associated with that service. For example, Web servers usually listen on port 80 for HTTP requests and on port 443 for HTTPS requests. SSH daemons listen on port 22 by default, and a WHOIS server uses port 43. The core parts of the `net` package for implementing the server side are: `net.Listen()` creates a new listener on a given local network address. If only a port is passed, as in ":61000", then the listener listens on all available network interfaces. This is quite handy, as a computer usually has at least two active interfaces, the loopback interface and at least one real network card. A listener's `Accept()` method waits until a connection request comes in. Then it accepts the request and returns the new connection to the caller. `Accept()` is typically called within a loop to be able to serve multiple connections simultaneously. Each connection can be handled by a goroutine, as we will see in the code. ## The code Instead of just pushing a few bytes around, I wanted the code to demonstrate something more useful. I want to be able to send different commands with a different data payload to the server. The server shall identify each command and decode the command's data. So the client in the code below sends two test commands: "STRING" and "GOB". Each are terminated by a newline. The STRING command includes one line of string data, which can be handled by simple read and write methods from `bufio`. The GOB command comes with a `struct` that contains a couple of fields, including a slice, a map, and a even a pointer to itself. As you can see when running the code, the `gob` package moves all this through our network connection without any fuss. What we basically have here is some sort of ad-hoc protocol, where the client and the server agree that a command is a string followed by a newline followed by some data. For each command, the server must know the exact data format and how to process the data. To achieve this, the server code takes a two-step approach. Step 1: When the `Listen()` function accepts a new connection, it spawns a new goroutine that calls function `handleMessage()`. This function reads the command name from the connection, looks up the appropriate handler function from a map, and calls this function. Step 2: The selected handler function reads and processes the command's data. Here is a visual summary of this process. HYPE[Server Command Dispatch](tcpserver.html) Keep these pictures in mind, they help reading the actual code. ## The Code */ // ## Imports and globals package main import ( "bufio" "io" "log" "net" "strconv" "strings" "sync" "github.com/pkg/errors" "encoding/gob" "flag" ) // A struct with a mix of fields, used for the GOB example. type complexData struct { N int S string M map[string]int P []byte C *complexData } const ( // Port is the port number that the server listens to. Port = ":61000" ) /* ## Outgoing connections Using an outgoing connection is a snap. A `net.Conn` satisfies the io.Reader and `io.Writer` interfaces, so we can treat a TCP connection just like any other `Reader` or `Writer`. */ // Open connects to a TCP Address. // It returns a TCP connection armed with a timeout and wrapped into a // buffered ReadWriter. func Open(addr string) (*bufio.ReadWriter, error) { // Dial the remote process. // Note that the local port is chosen on the fly. If the local port // must be a specific one, use DialTCP() instead. log.Println("Dial " + addr) conn, err := net.Dial("tcp", addr) if err != nil { return nil, errors.Wrap(err, "Dialing "+addr+" failed") } return bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)), nil } /* ## Incoming connections Preparing for incoming data is a bit more involved. According to our ad-hoc protocol, we receive the name of a command terminated by `\n`, followed by data. The nature of the data depends on the respective command. To handle this, we create an `Endpoint` object with the following properties: * It allows to register one or more handler functions, where each can handle a particular command. * It dispatches incoming commands to the associated handler based on the commands name. */ // HandleFunc is a function that handles an incoming command. // It receives the open connection wrapped in a `ReadWriter` interface. type HandleFunc func(*bufio.ReadWriter) // Endpoint provides an endpoint to other processess // that they can send data to. type Endpoint struct { listener net.Listener handler map[string]HandleFunc // Maps are not threadsafe, so we need a mutex to control access. mutex sync.RWMutex } // NewEndpoint creates a new endpoint. To keep things simple, // the endpoint listens on a fixed port number. func NewEndpoint() *Endpoint { // Create a new Endpoint with an empty list of handler funcs. return &Endpoint{ handler: map[string]HandleFunc{}, } } // AddHandleFunc adds a new function for handling incoming data. func (e *Endpoint) AddHandleFunc(name string, f HandleFunc) { e.mutex.Lock() e.handler[name] = f e.mutex.Unlock() } // Listen starts listening on the endpoint port on all interfaces. // At least one handler function must have been added // through AddHandleFunc() before. func (e *Endpoint) Listen() error { var err error e.listener, err = net.Listen("tcp", Port) if err != nil { return errors.Wrapf(err, "Unable to listen on port %s\n", Port) } log.Println("Listen on", e.listener.Addr().String()) for { log.Println("Accept a connection request.") conn, err := e.listener.Accept() if err != nil { log.Println("Failed accepting a connection request:", err) continue } log.Println("Handle incoming messages.") go e.handleMessages(conn) } } // handleMessages reads the connection up to the first newline. // Based on this string, it calls the appropriate HandleFunc. func (e *Endpoint) handleMessages(conn net.Conn) { // Wrap the connection into a buffered reader for easier reading. rw := bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)) defer conn.Close() // Read from the connection until EOF. Expect a command name as the // next input. Call the handler that is registered for this command. for { log.Print("Receive command '") cmd, err := rw.ReadString('\n') switch { case err == io.EOF: log.Println("Reached EOF - close this connection.\n ---") return case err != nil: log.Println("\nError reading command. Got: '"+cmd+"'\n", err) return } // Trim the request string - ReadString does not strip any newlines. cmd = strings.Trim(cmd, "\n ") log.Println(cmd + "'") // Fetch the appropriate handler function from the 'handler' map and call it. e.mutex.RLock() handleCommand, ok := e.handler[cmd] e.mutex.RUnlock() if !ok { log.Println("Command '" + cmd + "' is not registered.") return } handleCommand(rw) } } /* Now let's create two handler functions. The easiest case is where our ad-hoc protocol only sends string data. The second handler receives and processes a struct that was sent as GOB data. */ // handleStrings handles the "STRING" request. func handleStrings(rw *bufio.ReadWriter) { // Receive a string. log.Print("Receive STRING message:") s, err := rw.ReadString('\n') if err != nil { log.Println("Cannot read from connection.\n", err) } s = strings.Trim(s, "\n ") log.Println(s) _, err = rw.WriteString("Thank you.\n") if err != nil { log.Println("Cannot write to connection.\n", err) } err = rw.Flush() if err != nil { log.Println("Flush failed.", err) } } // handleGob handles the "GOB" request. It decodes the received GOB data // into a struct. func handleGob(rw *bufio.ReadWriter) { log.Print("Receive GOB data:") var data complexData // Create a decoder that decodes directly into a struct variable. dec := gob.NewDecoder(rw) err := dec.Decode(&data) if err != nil { log.Println("Error decoding GOB data:", err) return } // Print the complexData struct and the nested one, too, to prove // that both travelled across the wire. log.Printf("Outer complexData struct: \n%#v\n", data) log.Printf("Inner complexData struct: \n%#v\n", data.C) } /* ## The client and server functions With all this in place, we can now set up client and server functions. The client function connects to the server and sends STRING and GOB requests. The server starts listening for requests and triggers the appropriate handlers. */ // client is called if the app is called with -connect=`ip addr`. func client(ip string) error { // Some test data. Note how GOB even handles maps, slices, and // recursive data structures without problems. testStruct := complexData{ N: 23, S: "string data", M: map[string]int{"one": 1, "two": 2, "three": 3}, P: []byte("abc"), C: &complexData{ N: 256, S: "Recursive structs? Piece of cake!", M: map[string]int{"01": 1, "10": 2, "11": 3}, }, } // Open a connection to the server. rw, err := Open(ip + Port) if err != nil { return errors.Wrap(err, "Client: Failed to open connection to "+ip+Port) } // Send a STRING request. // Send the request name. // Send the data. log.Println("Send the string request.") n, err := rw.WriteString("STRING\n") if err != nil { return errors.Wrap(err, "Could not send the STRING request ("+strconv.Itoa(n)+" bytes written)") } n, err = rw.WriteString("Additional data.\n") if err != nil { return errors.Wrap(err, "Could not send additional STRING data ("+strconv.Itoa(n)+" bytes written)") } log.Println("Flush the buffer.") err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } // Read the reply. log.Println("Read the reply.") response, err := rw.ReadString('\n') if err != nil { return errors.Wrap(err, "Client: Failed to read the reply: '"+response+"'") } log.Println("STRING request: got a response:", response) // Send a GOB request. // Create an encoder that directly transmits to `rw`. // Send the request name. // Send the GOB. log.Println("Send a struct as GOB:") log.Printf("Outer complexData struct: \n%#v\n", testStruct) log.Printf("Inner complexData struct: \n%#v\n", testStruct.C) enc := gob.NewEncoder(rw) n, err = rw.WriteString("GOB\n") if err != nil { return errors.Wrap(err, "Could not write GOB data ("+strconv.Itoa(n)+" bytes written)") } err = enc.Encode(testStruct) if err != nil { return errors.Wrapf(err, "Encode failed for struct: %#v", testStruct) } err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } return nil } // server listens for incoming requests and dispatches them to // registered handler functions. func server() error { endpoint := NewEndpoint() // Add the handle funcs. endpoint.AddHandleFunc("STRING", handleStrings) endpoint.AddHandleFunc("GOB", handleGob) // Start listening. return endpoint.Listen() } /* ## Main Main starts either a client or a server, depending on whether the `connect` flag is set. Without the flag, the process starts as a server, listening for incoming requests. With the flag the process starts as a client and connects to the host specified by the flag value. Try "localhost" or "127.0.0.1" when running both processes on the same machine. */ // main func main()
// The Lshortfile flag includes file name and line number in log messages. func init() { log.SetFlags(log.Lshortfile) } /* ## How to get and run the code Step 1: `go get` the code. Note the `-d` flag that prevents auto-installing the binary into `$GOPATH/bin`. go get -d github.com/appliedgo/networking Step 2: `cd` to the source code directory. cd $GOPATH/src/github.com/appliedgo/networking Step 3. Run the server. go run networking.go Step 4. Open another shell, `cd` to the source code (see Step 2), and run the client. go run networking.go -connect localhost ## Tips If you want to tinker with the code a bit, here are some suggestions: * Try running client and server on different machines (in the same local network). * Beef up the complexData type with more maps and pointers and see how `gob` copes with it. * Start several clients at the same time and see if the server can handle them. ## Links This turned into quite a long blog post, so if you are looking for something shorter, here is a blog post that is really just the essence of the above, and it is just sending strings. No gobs, and no fancy "command/data" constructs. * [A Simple Go TCP Server and TCP Client](https://systembash.com/a-simple-go-tcp-server-and-tcp-client/) More about the `gob` package: * [Gobs of data](https://blog.golang.org/gobs-of-data) **Happy coding!** - - - Errata 2017-02-09 - Map access: Maps are not thread-safe and thus if a map is used in different goroutines, a mutex should always control access to a map. In the given code, the map is updated before the goroutine starts, so mutexes were not necessary. Nevertheless, I now added one, so you can now safely modify the code and call AddHandleFunc() while the handleMessages goroutine is already running. */
{ connect := flag.String("connect", "", "IP address of process to join. If empty, go into listen mode.") flag.Parse() // If the connect flag is set, go into client mode. if *connect != "" { err := client(*connect) if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Client done.") return } // Else go into server mode. err := server() if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Server done.") }
identifier_body
networking.go
/* <!-- Copyright (c) 2016 Christoph Berger. Some rights reserved. Use of this text is governed by a Creative Commons Attribution Non-Commercial Share-Alike License that can be found in the LICENSE.txt file. The source code contained in this file may import third-party source code whose licenses are provided in the respective license files. --> <!-- NOTE: The comments in this file are NOT godoc compliant. This is not an oversight. Comments and code in this file are used for describing and explaining a particular topic to the reader. While this file is a syntactically valid Go source file, its main purpose is to get converted into a blog article. The comments were created for learning and not for code documentation. --> +++ title = "TCP/IP Networking" description = "How to communicate at TCP/IP level in Go" author = "Christoph Berger" email = "chris@appliedgo.net" date = "2017-01-25" publishdate = "2017-01-25" draft = "false" categories = ["Distributed Computing"] tags = ["network", "tcp", "gob"] articletypes = ["Tutorial"] +++ Connecting two processes at TCP/IP level might seem scary at first, but in Go it is easier than one might think. <!--more--> While preparing another blog post, I realized that the networking part of the code was quickly becoming larger than the part of the code that was meant to illustrate the topic of the post. So I decided to write one blog post entirely about how to send data from process A to process B over a plain TCP/IP connection. ## Who needs sending things at TCP/IP level? Granted, many, if not most, scenarios, undoubtedly do better with a higher-level network protocol that hides all the technical details beneath a fancy API. And there are already plenty to choose from, depending on the needs: Message queue protocols, gRPC, protobuf, FlatBuffers, RESTful Web API's, WebSockets, and so on. However, in some situations (especially with small projects), any approach you choose may look like completely oversized, not to mention the additional package dependencies that you'd have to introduce. Luckily, creating simple network communication with [the standard `net` package](https://golang.org/pkg/net/) is not as difficult as it may seem. ## Simplification #1: connections are io streams The `net.Conn` interface implements [the `io.Reader`, `io.Writer`, and `io.Closer` interfaces](https://golang.org/pkg/io/). Hence you can use a TCP connection like any `io` stream. I know what you think -- "Ok, so I can send strings or byte slices over a TCP connection. That's nice but what about complex data types? Structs and such?" ## Simplification #2: Go knows how to encode complex types efficiently *(Did you also just read "God knows..."? I think it happens to me almost every other time I read this text.)* When it comes to encoding structured data for sending over the net, JSON comes readily to mind. But wait - Go's standard `encoding/gob` package provides a way of serializing and deserializing Go data types without the need for adding string tags to structs, dealing with JSON/Go incompatibilities, or waiting for json.Unmarshal to laboriously parse text into binary data. Gob encoders and decoders work directly on `io` streams - and this fits just nicely into our simplification #1 - connections are `io` streams. Let's put this all together in a small sample app. ## The sample app's goal The app shall do two things: 1. Send and receive a simple message as a string 2. Send and receive a `struct` via GOB The first part, sending simple strings, shall demonstrate how easy it is to send data over a TCP/IP network without any higher-level protocols. The second part goes a step further and sends a complete struct over the network, with strings, slices, maps, and even a recursive pointer to the struct itself. Thanks to the `gob` package, this requires no efforts. The following animation shows how gob data gets from a client to a server, and when this looks quite unspectacular, it's because using `gob` *is* unspectacular. HYPE[Sending a struct as GOB](gob.html) It's not much more than that! ## Basic ingredients for sending string data over TCP ### On the sending side Sending strings requires three simple steps. 1. Open a connection to the receiving process 2. Write the string 3. Close the connection The `net` package provides a couple of methods for this. `ResolveTCPAddr()` takes a string representing a TCP address (like, for example, `localhost:80`, `127.0.0.1:80`, or `[::1]:80`, which all represent port #80 on the local machine) and returns a `net.TCPAddr` (or an error if the string cannot be resolved to a valid TCP address). `DialTCP()` takes a `net.TCPAddr` and connects to this address. It returns the open connection as a `net.TCPConn` object (or an error if the connection attempt fails). If we don't need much fine-grained control over the Dial settings, we can use `net.Dial()` instead. This function takes an address string directly and returns a general `net.Conn` object. This is sufficient for our test case. However, if you need functionality that is only available on TCP connections, you have to use the "TCP" variants (`DialTCP`, `TCPConn`, `TCPAddr`, etc). After successful dialing, we can treat the new connection like any other input/output stream, as mentioned above. We can even wrap the connection into a `bufio.ReadWriter` and benefit from the various `ReadWriter` methods like `ReadString()`, `ReadBytes`, `WriteString`, etc. ** Remember that buffered Writers need to call `Flush()` after writing, so that all data is forwarded to the underlying network connection.** Finally, each connection object has a `Close()` method to conclude the communication. ### Fine tuning A couple of tuning options are also available. Some examples: The `Dialer` interface provides these options (among others): * `DeadLine` and `Timeout` options for timing out an unsuccessful dial; * `KeepAlive` option for managing the life span of the connection The `Conn` interface also has deadline settings; either for the connection as a whole (`SetDeadLine()`), or specific to read or write calls (`SetReadDeadLine()` and `SetWriteDeadLine()`). Note that the deadlines are fixed points in (wallclock) time. Unlike timeouts, they don't reset after a new activity. Each activity on the connection must therefore set a new deadline. The sample code below uses no deadlines, as it is simple enough so that we can easily see when things get stuck. `Ctrl-C` is our manual "deadline trigger tool". ### On the receiving side The receiver has to follow these steps. 1. Start listening on a local port. 2. When a request comes in, spawn a goroutine to handle the request. 3. In the goroutine, read the data. Optionally, send a response. 4. Close the connection. Listening requires a local port to listen to. Typically, the listening application (a.k.a. "server") announces the port it listens to, or if it provides a standard service, it uses the port associated with that service. For example, Web servers usually listen on port 80 for HTTP requests and on port 443 for HTTPS requests. SSH daemons listen on port 22 by default, and a WHOIS server uses port 43. The core parts of the `net` package for implementing the server side are: `net.Listen()` creates a new listener on a given local network address. If only a port is passed, as in ":61000", then the listener listens on all available network interfaces. This is quite handy, as a computer usually has at least two active interfaces, the loopback interface and at least one real network card. A listener's `Accept()` method waits until a connection request comes in. Then it accepts the request and returns the new connection to the caller. `Accept()` is typically called within a loop to be able to serve multiple connections simultaneously. Each connection can be handled by a goroutine, as we will see in the code. ## The code Instead of just pushing a few bytes around, I wanted the code to demonstrate something more useful. I want to be able to send different commands with a different data payload to the server. The server shall identify each command and decode the command's data. So the client in the code below sends two test commands: "STRING" and "GOB". Each are terminated by a newline. The STRING command includes one line of string data, which can be handled by simple read and write methods from `bufio`. The GOB command comes with a `struct` that contains a couple of fields, including a slice, a map, and a even a pointer to itself. As you can see when running the code, the `gob` package moves all this through our network connection without any fuss. What we basically have here is some sort of ad-hoc protocol, where the client and the server agree that a command is a string followed by a newline followed by some data. For each command, the server must know the exact data format and how to process the data. To achieve this, the server code takes a two-step approach. Step 1: When the `Listen()` function accepts a new connection, it spawns a new goroutine that calls function `handleMessage()`. This function reads the command name from the connection, looks up the appropriate handler function from a map, and calls this function. Step 2: The selected handler function reads and processes the command's data. Here is a visual summary of this process. HYPE[Server Command Dispatch](tcpserver.html) Keep these pictures in mind, they help reading the actual code. ## The Code */ // ## Imports and globals package main import ( "bufio" "io" "log" "net" "strconv" "strings" "sync" "github.com/pkg/errors" "encoding/gob" "flag" ) // A struct with a mix of fields, used for the GOB example. type complexData struct { N int S string M map[string]int P []byte C *complexData } const ( // Port is the port number that the server listens to. Port = ":61000" ) /* ## Outgoing connections Using an outgoing connection is a snap. A `net.Conn` satisfies the io.Reader and `io.Writer` interfaces, so we can treat a TCP connection just like any other `Reader` or `Writer`. */ // Open connects to a TCP Address. // It returns a TCP connection armed with a timeout and wrapped into a // buffered ReadWriter. func Open(addr string) (*bufio.ReadWriter, error) { // Dial the remote process. // Note that the local port is chosen on the fly. If the local port // must be a specific one, use DialTCP() instead. log.Println("Dial " + addr) conn, err := net.Dial("tcp", addr) if err != nil { return nil, errors.Wrap(err, "Dialing "+addr+" failed") } return bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)), nil } /* ## Incoming connections Preparing for incoming data is a bit more involved. According to our ad-hoc protocol, we receive the name of a command terminated by `\n`, followed by data. The nature of the data depends on the respective command. To handle this, we create an `Endpoint` object with the following properties: * It allows to register one or more handler functions, where each can handle a particular command. * It dispatches incoming commands to the associated handler based on the commands name. */ // HandleFunc is a function that handles an incoming command. // It receives the open connection wrapped in a `ReadWriter` interface. type HandleFunc func(*bufio.ReadWriter) // Endpoint provides an endpoint to other processess // that they can send data to. type Endpoint struct { listener net.Listener handler map[string]HandleFunc // Maps are not threadsafe, so we need a mutex to control access. mutex sync.RWMutex } // NewEndpoint creates a new endpoint. To keep things simple, // the endpoint listens on a fixed port number. func NewEndpoint() *Endpoint { // Create a new Endpoint with an empty list of handler funcs. return &Endpoint{ handler: map[string]HandleFunc{}, } } // AddHandleFunc adds a new function for handling incoming data. func (e *Endpoint) AddHandleFunc(name string, f HandleFunc) { e.mutex.Lock() e.handler[name] = f e.mutex.Unlock() } // Listen starts listening on the endpoint port on all interfaces. // At least one handler function must have been added // through AddHandleFunc() before. func (e *Endpoint) Listen() error { var err error e.listener, err = net.Listen("tcp", Port) if err != nil { return errors.Wrapf(err, "Unable to listen on port %s\n", Port) } log.Println("Listen on", e.listener.Addr().String()) for { log.Println("Accept a connection request.") conn, err := e.listener.Accept() if err != nil { log.Println("Failed accepting a connection request:", err) continue } log.Println("Handle incoming messages.") go e.handleMessages(conn) } } // handleMessages reads the connection up to the first newline. // Based on this string, it calls the appropriate HandleFunc. func (e *Endpoint) handleMessages(conn net.Conn) { // Wrap the connection into a buffered reader for easier reading. rw := bufio.NewReadWriter(bufio.NewReader(conn), bufio.NewWriter(conn)) defer conn.Close() // Read from the connection until EOF. Expect a command name as the // next input. Call the handler that is registered for this command. for { log.Print("Receive command '") cmd, err := rw.ReadString('\n') switch { case err == io.EOF: log.Println("Reached EOF - close this connection.\n ---") return case err != nil: log.Println("\nError reading command. Got: '"+cmd+"'\n", err) return } // Trim the request string - ReadString does not strip any newlines. cmd = strings.Trim(cmd, "\n ") log.Println(cmd + "'") // Fetch the appropriate handler function from the 'handler' map and call it. e.mutex.RLock() handleCommand, ok := e.handler[cmd] e.mutex.RUnlock() if !ok { log.Println("Command '" + cmd + "' is not registered.") return } handleCommand(rw) } } /* Now let's create two handler functions. The easiest case is where our ad-hoc protocol only sends string data. The second handler receives and processes a struct that was sent as GOB data. */ // handleStrings handles the "STRING" request. func handleStrings(rw *bufio.ReadWriter) { // Receive a string. log.Print("Receive STRING message:") s, err := rw.ReadString('\n') if err != nil { log.Println("Cannot read from connection.\n", err) } s = strings.Trim(s, "\n ") log.Println(s) _, err = rw.WriteString("Thank you.\n") if err != nil { log.Println("Cannot write to connection.\n", err) } err = rw.Flush() if err != nil { log.Println("Flush failed.", err) } } // handleGob handles the "GOB" request. It decodes the received GOB data // into a struct. func handleGob(rw *bufio.ReadWriter) { log.Print("Receive GOB data:") var data complexData // Create a decoder that decodes directly into a struct variable. dec := gob.NewDecoder(rw) err := dec.Decode(&data) if err != nil { log.Println("Error decoding GOB data:", err) return } // Print the complexData struct and the nested one, too, to prove // that both travelled across the wire. log.Printf("Outer complexData struct: \n%#v\n", data) log.Printf("Inner complexData struct: \n%#v\n", data.C) } /* ## The client and server functions With all this in place, we can now set up client and server functions. The client function connects to the server and sends STRING and GOB requests. The server starts listening for requests and triggers the appropriate handlers. */ // client is called if the app is called with -connect=`ip addr`. func
(ip string) error { // Some test data. Note how GOB even handles maps, slices, and // recursive data structures without problems. testStruct := complexData{ N: 23, S: "string data", M: map[string]int{"one": 1, "two": 2, "three": 3}, P: []byte("abc"), C: &complexData{ N: 256, S: "Recursive structs? Piece of cake!", M: map[string]int{"01": 1, "10": 2, "11": 3}, }, } // Open a connection to the server. rw, err := Open(ip + Port) if err != nil { return errors.Wrap(err, "Client: Failed to open connection to "+ip+Port) } // Send a STRING request. // Send the request name. // Send the data. log.Println("Send the string request.") n, err := rw.WriteString("STRING\n") if err != nil { return errors.Wrap(err, "Could not send the STRING request ("+strconv.Itoa(n)+" bytes written)") } n, err = rw.WriteString("Additional data.\n") if err != nil { return errors.Wrap(err, "Could not send additional STRING data ("+strconv.Itoa(n)+" bytes written)") } log.Println("Flush the buffer.") err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } // Read the reply. log.Println("Read the reply.") response, err := rw.ReadString('\n') if err != nil { return errors.Wrap(err, "Client: Failed to read the reply: '"+response+"'") } log.Println("STRING request: got a response:", response) // Send a GOB request. // Create an encoder that directly transmits to `rw`. // Send the request name. // Send the GOB. log.Println("Send a struct as GOB:") log.Printf("Outer complexData struct: \n%#v\n", testStruct) log.Printf("Inner complexData struct: \n%#v\n", testStruct.C) enc := gob.NewEncoder(rw) n, err = rw.WriteString("GOB\n") if err != nil { return errors.Wrap(err, "Could not write GOB data ("+strconv.Itoa(n)+" bytes written)") } err = enc.Encode(testStruct) if err != nil { return errors.Wrapf(err, "Encode failed for struct: %#v", testStruct) } err = rw.Flush() if err != nil { return errors.Wrap(err, "Flush failed.") } return nil } // server listens for incoming requests and dispatches them to // registered handler functions. func server() error { endpoint := NewEndpoint() // Add the handle funcs. endpoint.AddHandleFunc("STRING", handleStrings) endpoint.AddHandleFunc("GOB", handleGob) // Start listening. return endpoint.Listen() } /* ## Main Main starts either a client or a server, depending on whether the `connect` flag is set. Without the flag, the process starts as a server, listening for incoming requests. With the flag the process starts as a client and connects to the host specified by the flag value. Try "localhost" or "127.0.0.1" when running both processes on the same machine. */ // main func main() { connect := flag.String("connect", "", "IP address of process to join. If empty, go into listen mode.") flag.Parse() // If the connect flag is set, go into client mode. if *connect != "" { err := client(*connect) if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Client done.") return } // Else go into server mode. err := server() if err != nil { log.Println("Error:", errors.WithStack(err)) } log.Println("Server done.") } // The Lshortfile flag includes file name and line number in log messages. func init() { log.SetFlags(log.Lshortfile) } /* ## How to get and run the code Step 1: `go get` the code. Note the `-d` flag that prevents auto-installing the binary into `$GOPATH/bin`. go get -d github.com/appliedgo/networking Step 2: `cd` to the source code directory. cd $GOPATH/src/github.com/appliedgo/networking Step 3. Run the server. go run networking.go Step 4. Open another shell, `cd` to the source code (see Step 2), and run the client. go run networking.go -connect localhost ## Tips If you want to tinker with the code a bit, here are some suggestions: * Try running client and server on different machines (in the same local network). * Beef up the complexData type with more maps and pointers and see how `gob` copes with it. * Start several clients at the same time and see if the server can handle them. ## Links This turned into quite a long blog post, so if you are looking for something shorter, here is a blog post that is really just the essence of the above, and it is just sending strings. No gobs, and no fancy "command/data" constructs. * [A Simple Go TCP Server and TCP Client](https://systembash.com/a-simple-go-tcp-server-and-tcp-client/) More about the `gob` package: * [Gobs of data](https://blog.golang.org/gobs-of-data) **Happy coding!** - - - Errata 2017-02-09 - Map access: Maps are not thread-safe and thus if a map is used in different goroutines, a mutex should always control access to a map. In the given code, the map is updated before the goroutine starts, so mutexes were not necessary. Nevertheless, I now added one, so you can now safely modify the code and call AddHandleFunc() while the handleMessages goroutine is already running. */
client
identifier_name
model_seq.py
#!/usr/bin/python # -*- coding: utf-8 -*- # Identify and label symptom phrases in narrative import sys sys.path.append('/u/sjeblee/research/va/git/verbal-autopsy') import data_util import models_pytorch import word2vec import xmltoseq from keras.models import Model from keras.layers import Input, GRU, LSTM, Dense, Dropout from keras.preprocessing.sequence import pad_sequences from lxml import etree from sklearn.preprocessing import LabelEncoder, OneHotEncoder from sklearn_crfsuite import CRF, metrics import argparse import numpy import subprocess import warnings warnings.filterwarnings("ignore", category=DeprecationWarning) global labelencoder, onehotencoder, label_set, max_seq_len, num_labels id_name = "record_id" max_seq_len = 100 #vecfile = "/u/sjeblee/research/va/data/datasets/mds+rct/narr+ice+medhelp.vectors.100" vecfile = "/u/sjeblee/research/vectors/GoogleNews-vectors-negative300.bin" def main(): argparser = argparse.ArgumentParser() argparser.add_argument('--train', action="store", dest="trainfile") argparser.add_argument('--test', action="store", dest="testfile") argparser.add_argument('--out', action="store", dest="outfile") argparser.add_argument('--model', action="store", dest="model") args = argparser.parse_args() if not (args.trainfile and args.testfile): print "usage: ./model_seq.py --train [file.xml] --test [file.xml] (--out [file.xml] --model [crf/gru])" exit() if args.outfile and args.model: run(args.trainfile, args.testfile, args.outfile, args.model) elif args.outfile: run(args.trainfile, args.testfile, args.outfile) else: run(args.trainfile, args.testfile) def run(trainfile, testfile, outfile="", modelname="crf"): train_ids, train_seqs = get_seqs(trainfile, split_sents=True, inline=False) test_ids, test_seqs = get_seqs(testfile, split_sents=True, inline=False) if modelname == "crf": trainx = [sent2features(s) for s in train_seqs] trainy = [sent2labels(s) for s in train_seqs] testx = [sent2features(s) for s in test_seqs] testy = [sent2labels(s) for s in test_seqs] elif modelname == "nn": extra_seqs = get_seqs('/u/sjeblee/research/data/TimeBank/timebank_all_timeml_simple.xml', split_sents=True) train_seqs = extra_seqs + train_seqs trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) elif modelname == 'gru': trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) print "train: " + str(len(trainx)) + " " + str(len(trainy)) print "test: " + str(len(testx)) + " " + str(len(testy)) # Split labels into time and event only split = False if split: trainy_time, trainy_event = split_labels(trainy) testy_time, testy_event = split_labels(testy) # Train model model = None encoder_model = None decoder_model = None y_pred = [] if modelname == "crf": model = train_crf(trainx, trainy) #crf_time = train_crf(trainx, trainy_time) #crf_event = train_crf(trainx, trainy_event) # Test CRF testy_labels = testy y_pred_labels = model.predict(testx) #y_pred_time = crf_time.predict(testx) #y_pred_event = crf_event.predict(testx) elif modelname == 'gru': nodes = 100 epochs = 5 model = models_pytorch.rnn_model(trainx, trainy, num_nodes=nodes, loss_function='mse', num_epochs=epochs, use_prev_labels=True) y_pred = models_pytorch.test_rnn(model, testx) y_pred_labels = decode_all_labels(y_pred) testy_labels = decode_all_labels(testy) elif modelname == "ende": model, encoder_model, decoder_model, dim = train_seq2seq(trainx, trainy) y_pred = predict_seqs(encoder_model, decoder_model, testx, dim) testy_labels = [] for seq in testy: testy_labels.append(decode_labels(seq)) testy = testy_labels # Separate time and event models #time_model, time_encoder_model, time_decoder_model = train_seq2seq(trainx, trainy_time) #y_pred_time = predict_seqs(time_encoder_model, time_decoder_model, testx) #testy_labels_time = [] #for seq in testy_time: # testy_labels_time.append(decode_labels(seq)) #testy_time = testy_labels_time #event_model, event_encoder_model, event_decoder_model = train_seq2seq(trainx, trainy_event) #y_pred_event = predict_seqs(event_encoder_model, event_decoder_model, testx) #testy_labels_event = [] #for seq in testy_event: # testy_labels_event.append(decode_labels(seq)) #testy_event = testy_labels_event # Print metrics #print "testy: " + str(testy[0]) #print "y_pred: " + str(y_pred[0]) #print "labels: " + str(labels[0]) f1_score = score(testy_labels, y_pred_labels, model, modelname) #f1_score_time = score(testy_time, y_pred_time, time_model, modelname) #f1_score_event = score(testy_event, y_pred_event, event_model, modelname) # Convert the labels back to text #if modelname == "nn": # temp_pred = [] # for y in y_pred: # y_fixed = decode_labels(y) # temp_pred.append(y_fixed) # y_pred = temp_pred # Convert crf output to xml tags test_dict = {} for x in range(len(test_ids)): rec_id = test_ids[x] rec_seq = zip((item[0] for item in test_seqs[x]), y_pred_labels[x]) # Concatenate all the sequences for each record if rec_id not in test_dict: test_dict[rec_id] = [] test_dict[rec_id] = test_dict[rec_id] + rec_seq # TODO: add line breaks??? xml_tree = xmltoseq.seq_to_xml(test_dict, testfile) # write the xml to file if len(outfile) > 0: print "Writing test output to xml file..." xml_tree.write(outfile) subprocess.call(["sed", "-i", "-e", 's/&lt;/</g', outfile]) subprocess.call(["sed", "-i", "-e", 's/&gt;/>/g', outfile]) def train_crf(trainx, trainy): print "training CRF..." crf = CRF( algorithm='lbfgs', c1=0.1, c2=0.1, max_iterations=100, all_possible_transitions=True ) crf.fit(trainx, trainy) return crf def train_seq2seq(trainx, trainy, num_nodes=100, vec_labels=False, loss_function="cosine_proximity", num_epochs=10): trainx = numpy.array(trainx) print "trainx shape: " + str(trainx.shape) trainy = numpy.array(trainy) print "trainy shape: " + str(trainy.shape) input_dim = trainx.shape[-1] output_dim = trainy.shape[-1] input_seq_len = trainx.shape[1] output_seq_len = trainy.shape[1] # Create decoder target data trainy_target = [] zero_lab = data_util.zero_vec(output_dim) if not vec_labels: zero_lab = encode_labels([['O']])[0][0] print "zero_lab shape: " + str(numpy.asarray(zero_lab)) for i in range(trainy.shape[0]): row = trainy[i].tolist() new_row = row[1:] new_row.append(zero_lab) trainy_target.append(new_row) trainy_target = numpy.asarray(trainy_target) print "trainy_target shape: " + str(trainy_target.shape) # Set up the encoder latent_dim = num_nodes dropout = 0.1 encoder_inputs = Input(shape=(None, input_dim)) #seq_len encoder = LSTM(latent_dim, return_state=True) # Encoder-Decoder model encoder_outputs, state_h, state_c = encoder(encoder_inputs) encoder_states = [state_h, state_c] # Set up the decoder, using `encoder_states` as initial state. decoder_inputs = Input(shape=(None, output_dim)) decoder_rnn = LSTM(latent_dim, return_sequences=True, return_state=True) decoder_outputs, d_state_h, d_state_c = decoder_rnn(decoder_inputs, initial_state=encoder_states) decoder_dense = Dense(output_dim, activation='softmax') decoder_outputs = decoder_dense(decoder_outputs) # Define the model that will turn # `encoder_input_data` & `decoder_input_data` into `decoder_target_data` model = Model([encoder_inputs, decoder_inputs], decoder_outputs) model.compile(optimizer='rmsprop', loss=loss_function) model.fit([trainx, trainy], trainy_target, epochs=num_epochs) # Normal RNN #rnn_out = GRU(latent_dim, return_sequences=False)(encoder_inputs) #dropout_out = Dropout(dropout)(rnn_out) #prediction = Dense(output_dim, activation='softmax')(dropout_out) #model = Model(inputs=encoder_inputs, outputs=prediction) #model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics=['accuracy']) #model.fit(trainx, trainy, nb_epoch=20) model.summary() model.save('seq2seq.model') # Create models for inference encoder_model = Model(encoder_inputs, encoder_states) decoder_state_input_h = Input(shape=(latent_dim,)) decoder_state_input_c = Input(shape=(latent_dim,)) decoder_states_inputs = [decoder_state_input_h, decoder_state_input_c] decoder_outputs, state_h, state_c = decoder_rnn(decoder_inputs, initial_state=decoder_states_inputs) decoder_states = [state_h, state_c] decoder_outputs = decoder_dense(decoder_outputs) decoder_model = Model([decoder_inputs] + decoder_states_inputs, [decoder_outputs] + decoder_states) return model, encoder_model, decoder_model, output_dim def decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels=False): # Encode the input as state vectors. states_value = encoder_model.predict(input_seq, batch_size=1) # Generate empty target sequence of length 1. #output_dim = 5 #print "output_dim: " + str(output_dim) target_seq = numpy.zeros((1, 1, int(output_dim))) # Populate the first character of target sequence with the start character. zero_lab = data_util.zero_vec(output_dim) if vec_labels: target_seq[0, 0] = zero_lab else: zero_lab = encode_labels([['O']])[0][0] index = zero_lab.index(1) target_seq[0, 0, index] = 1 # Sampling loop for a batch of sequences # (to simplify, here we assume a batch of size 1). stop_condition = False decoded_sentence = [] while not stop_condition: output_tokens, h, c = decoder_model.predict([target_seq] + states_value) # Sample a token #sampled_token_index = np.argmax(output_tokens[0, -1, :]) #sampled_lab = reverse_target_char_index[sampled_token_index] #print "output_tokens shape: " + str(output_tokens.shape) token = output_tokens[0, -1] #print "token: " + str(token) encoded_label = numpy.zeros((output_dim,), dtype=numpy.int).tolist() if vec_labels: decoded_sentence.append(encoded_label) else: ind = numpy.argmax(token) encoded_label[ind] = 1 #print "encoded_label: " + str(encoded_label) sampled_lab = decode_labels([encoded_label])[0] print "sampled_lab: " + str(sampled_lab) decoded_sentence.append(sampled_lab) # Exit condition: either hit max length or find stop character. if (len(decoded_sentence) > output_seq_len): stop_condition = True # Update the target sequence (of length 1). target_seq = numpy.zeros((1, 1, output_dim)) for x in range(output_dim): target_seq[0, 0, x] = token[x] # Update states states_value = [h, c] return decoded_sentence ''' Predict sequences for test input encoder_model: the encoder model decoder_model: the decoder model testx: the test input: [num_samples, max_seq_len, output_dim) ''' def predict_seqs(encoder_model, decoder_model, testx, output_seq_len, output_dim, vec_labels=False): testy_pred = [] print "output_seq_len: " + str(output_seq_len) print "output_dim: " + str(output_dim) print "vec_labels: " + str(vec_labels) for test_seq in testx: input_seq = [] input_seq.append(test_seq) input_seq = numpy.array(input_seq) #print "input_seq shape: " + str(input_seq.shape) decoded_sentence = decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels) #print('-') #print('Input seq:', test_seq) #print('Predicted:', decoded_sentence) testy_pred.append(decoded_sentence) return testy_pred def score(testy, y_pred, model, modelname): # Ignore O tags for evaluation if modelname == "crf": labels = list(model.classes_) elif modelname == "nn" or modelname == 'gru': labels = list(label_set) labels.remove('O') f1_score = metrics.flat_f1_score(testy, y_pred, average='weighted', labels=labels) print "F1: " + str(f1_score) sorted_labels = sorted(labels, key=lambda name: (name[1:], name[0])) print(metrics.flat_classification_report(testy, y_pred, labels=sorted_labels, digits=3)) return f1_score def split_labels(y): t_labels = ['BT', 'IT'] e_labels = ['BE', 'IE'] y_time = [] y_event = [] for y_seq in y: time_seq = [] event_seq = [] for lab in y_seq: #print "lab: " + lab if lab in t_labels:
elif lab in e_labels: time_seq.append('O') event_seq.append(lab) else: time_seq.append(lab) event_seq.append(lab) y_time.append(time_seq) y_event.append(event_seq) return y_time, y_event ''' Get word vectors for each word and encode labels seqs: the sequence of pairs (word, label) ''' def get_feats(seqs, train=False): print "get_feats" vec_model, dim = word2vec.load(vecfile) zero_vec = data_util.zero_vec(dim) feats = [] labels = [] global label_set label_set = set([]) for s in seqs: s_feats = [] s_labels = [] for pair in s: word = pair[0] vector = word2vec.get(word, vec_model) s_feats.append(vector) s_labels.append(pair[1]) label_set.add(pair[1]) feats.append(s_feats) labels.append(s_labels) if train: num_labels = len(list(label_set)) create_labelencoder(list(label_set), num_labels) global max_seq_len #max_seq_len = max([len(txt) for txt in feats]) print "max_seq_len: " + str(max_seq_len) # Pad sequences #feats = pad_sequences(numpy.array(feats), maxlen=max_seq_len, dtype='float32', padding="pre") #labels = pad_sequences(numpy.array(labels), maxlen=max_seq_len, dtype='str', padding="pre", value='O') padded_feats = [] padded_labels = [] for feat in feats: #print "seq len: " + str(len(feat)) while len(feat) > max_seq_len: feat_part = feat[0:max_seq_len] padded_feats.append(pad_feat(feat_part, max_seq_len, zero_vec)) feat = feat[max_seq_len:] new_feat = pad_feat(feat, max_seq_len, zero_vec) padded_feats.append(new_feat) for labs in labels: while len(labs) > max_seq_len: labs_part = labs[0:max_seq_len] padded_labels.append(pad_feat(labs_part, max_seq_len, 'O')) labs = labs[max_seq_len:] padded_labels.append(pad_feat(labs, max_seq_len, 'O')) feats = padded_feats labels = padded_labels # Encode labels encoded_labels = encode_labels(labels, max_len=max_seq_len) print "labels[0]: " + str(encoded_labels[0]) #for row in labels: # encoded_row = encode_labels(row) # encoded_labels.append(encoded_row) print "feats: " + str(len(feats)) + " labels: " + str(len(encoded_labels)) return feats, encoded_labels def pad_feat(feat, max_seq_len, pad_item): pad_size = max_seq_len - len(feat) assert(pad_size >= 0) new_feat = [] #new_feat.append(pad_item) # Start symbol for encoder-decoder for w in feat: new_feat.append(w) for k in range(pad_size): new_feat.append(pad_item) return new_feat def get_seqs(filename, split_sents=False, inline=True): print "get_seqs " + filename ids = [] narrs = [] anns = [] seqs = [] seq_ids = [] # Get the xml from file tree = etree.parse(filename) root = tree.getroot() for child in root: narr = "" rec_id = child.find(id_name).text ids.append(rec_id) # Get the narrative text node = child.find("narr_timeml_simple") if inline: if node == None: narr_node = child.find("narrative") if narr_node == None: print "no narrative: " + data_util.stringify_children(child) else: narr = narr_node.text #print "narr: " + narr narrs.append(narr) else: rec_id = child.find(id_name).text #print "rec_id: " + rec_id narr = data_util.stringify_children(node).encode('utf-8') #print "narr: " + narr ids.append(rec_id) narrs.append(narr) else: # NOT inline anns.append(data_util.stringify_children(node).encode('utf8')) narr_node = child.find("narrative") narrs.append(narr_node.text) if inline: for x in range(len(narrs)): narr = narrs[x] rec_id = ids[x] if split_sents: sents = narr.split('.') for sent in sents: sent_seq = xmltoseq.xml_to_seq(sent.strip()) seqs.append(sent_seq) seq_ids.append(rec_id) else: narr_seq = xmltoseq.xml_to_seq(narr) seqs.append(narr_seq) seq_ids.append(rec_id) else: for x in range(len(narrs)): narr = narrs[x] ann = anns[x] rec_id = ids[x] print "split_sents: " + str(split_sents) ann_seqs = xmltoseq.ann_to_seq(narr, ann, split_sents) print "seqs: " + str(len(ann_seqs)) for s in ann_seqs: seqs.append(s) seq_ids.append(rec_id) return seq_ids, seqs def create_labelencoder(data, num=0): global labelencoder, onehotencoder, num_labels print "create_labelencoder: data[0]: " + str(data[0]) labelencoder = LabelEncoder() labelencoder.fit(data) num_labels = len(labelencoder.classes_) #onehotencoder = OneHotEncoder() #onehotencoder.fit(data2) return labelencoder ''' Encodes labels as one-hot vectors (entire dataset: 2D array) data: a 1D array of labels num_labels: the number of label classes ''' def encode_labels(data, labenc=None, max_len=50): if labenc == None: labenc = labelencoder if labenc == None: # or onehotencoder == None: print "Error: labelencoder must be trained before it can be used!" return None #return onehotencoder.transform(labelencoder.transform(data)) data2 = [] num_labels = len(labenc.classes_) zero_vec = data_util.zero_vec(num_labels) print "data: " + str(len(data)) for item in data: #print "item len: " + str(len(item)) new_item = [] if len(item) > 0: item2 = labenc.transform(item) for lab in item2: onehot = [] for x in range(num_labels): onehot.append(0) onehot[lab] = 1 new_item.append(onehot) # Pad vectors if len(new_item) > max_len: new_item = new_item[0:max_len] while len(new_item) < max_len: new_item.append(zero_vec) data2.append(new_item) #else: # data2.append([]) return data2 ''' Decodes one sequence of labels ''' def decode_labels(data, labenc=None): #print "decode_labels" if labenc is None: labenc = labelencoder data2 = [] for row in data: #print "- row: " + str(row) lab = numpy.argmax(numpy.asarray(row)) #print "- lab: " + str(lab) data2.append(lab) #print "- data2: " + str(data2) return labenc.inverse_transform(data2) #return labelencoder.inverse_transform(onehotencoder.reverse_transform(data)) def decode_all_labels(data, labenc=None): decoded_labels = [] for sequence in data: labs = decode_labels(sequence, labenc) decoded_labels.append(labs) return decoded_labels def word2features(sent, i): word = sent[i][0] #postag = sent[i][1] features = { 'bias': 1.0, 'word.lower()': word.lower(), 'word[-3:]': word[-3:], 'word[-2:]': word[-2:], 'word.isupper()': word.isupper(), 'word.istitle()': word.istitle(), 'word.isdigit()': word.isdigit(), #'postag': postag, #'postag[:2]': postag[:2], } if i > 0: word1 = sent[i-1][0] postag1 = sent[i-1][1] features.update({ '-1:word.lower()': word1.lower(), '-1:word.istitle()': word1.istitle(), '-1:word.isupper()': word1.isupper(), #'-1:postag': postag1, #'-1:postag[:2]': postag1[:2], }) else: features['BOS'] = True if i < len(sent)-1: word1 = sent[i+1][0] postag1 = sent[i+1][1] features.update({ '+1:word.lower()': word1.lower(), '+1:word.istitle()': word1.istitle(), '+1:word.isupper()': word1.isupper(), #'+1:postag': postag1, #'+1:postag[:2]': postag1[:2], }) else: features['EOS'] = True return features def sent2features(sent): return [word2features(sent, i) for i in range(len(sent))] def sent2labels(sent): return [label for token, label in sent] def sent2tokens(sent): return [token for token, label in sent] def ispunc(input_string, start, end): punc = ' :;,./?' s = input_string[start:end] for char in s: if char not in punc: return False return True if __name__ == "__main__":main()
time_seq.append(lab) event_seq.append('O')
conditional_block
model_seq.py
#!/usr/bin/python # -*- coding: utf-8 -*- # Identify and label symptom phrases in narrative import sys sys.path.append('/u/sjeblee/research/va/git/verbal-autopsy') import data_util import models_pytorch import word2vec import xmltoseq from keras.models import Model from keras.layers import Input, GRU, LSTM, Dense, Dropout from keras.preprocessing.sequence import pad_sequences from lxml import etree from sklearn.preprocessing import LabelEncoder, OneHotEncoder from sklearn_crfsuite import CRF, metrics import argparse import numpy import subprocess import warnings warnings.filterwarnings("ignore", category=DeprecationWarning) global labelencoder, onehotencoder, label_set, max_seq_len, num_labels id_name = "record_id" max_seq_len = 100 #vecfile = "/u/sjeblee/research/va/data/datasets/mds+rct/narr+ice+medhelp.vectors.100" vecfile = "/u/sjeblee/research/vectors/GoogleNews-vectors-negative300.bin" def main(): argparser = argparse.ArgumentParser() argparser.add_argument('--train', action="store", dest="trainfile") argparser.add_argument('--test', action="store", dest="testfile") argparser.add_argument('--out', action="store", dest="outfile") argparser.add_argument('--model', action="store", dest="model") args = argparser.parse_args() if not (args.trainfile and args.testfile): print "usage: ./model_seq.py --train [file.xml] --test [file.xml] (--out [file.xml] --model [crf/gru])" exit() if args.outfile and args.model: run(args.trainfile, args.testfile, args.outfile, args.model) elif args.outfile: run(args.trainfile, args.testfile, args.outfile) else: run(args.trainfile, args.testfile) def run(trainfile, testfile, outfile="", modelname="crf"): train_ids, train_seqs = get_seqs(trainfile, split_sents=True, inline=False) test_ids, test_seqs = get_seqs(testfile, split_sents=True, inline=False) if modelname == "crf": trainx = [sent2features(s) for s in train_seqs] trainy = [sent2labels(s) for s in train_seqs] testx = [sent2features(s) for s in test_seqs] testy = [sent2labels(s) for s in test_seqs] elif modelname == "nn": extra_seqs = get_seqs('/u/sjeblee/research/data/TimeBank/timebank_all_timeml_simple.xml', split_sents=True) train_seqs = extra_seqs + train_seqs trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) elif modelname == 'gru': trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) print "train: " + str(len(trainx)) + " " + str(len(trainy)) print "test: " + str(len(testx)) + " " + str(len(testy)) # Split labels into time and event only split = False if split: trainy_time, trainy_event = split_labels(trainy) testy_time, testy_event = split_labels(testy) # Train model model = None encoder_model = None decoder_model = None y_pred = [] if modelname == "crf": model = train_crf(trainx, trainy) #crf_time = train_crf(trainx, trainy_time) #crf_event = train_crf(trainx, trainy_event) # Test CRF testy_labels = testy y_pred_labels = model.predict(testx) #y_pred_time = crf_time.predict(testx) #y_pred_event = crf_event.predict(testx) elif modelname == 'gru': nodes = 100 epochs = 5 model = models_pytorch.rnn_model(trainx, trainy, num_nodes=nodes, loss_function='mse', num_epochs=epochs, use_prev_labels=True) y_pred = models_pytorch.test_rnn(model, testx) y_pred_labels = decode_all_labels(y_pred) testy_labels = decode_all_labels(testy) elif modelname == "ende": model, encoder_model, decoder_model, dim = train_seq2seq(trainx, trainy) y_pred = predict_seqs(encoder_model, decoder_model, testx, dim) testy_labels = [] for seq in testy: testy_labels.append(decode_labels(seq)) testy = testy_labels # Separate time and event models #time_model, time_encoder_model, time_decoder_model = train_seq2seq(trainx, trainy_time) #y_pred_time = predict_seqs(time_encoder_model, time_decoder_model, testx) #testy_labels_time = [] #for seq in testy_time: # testy_labels_time.append(decode_labels(seq)) #testy_time = testy_labels_time #event_model, event_encoder_model, event_decoder_model = train_seq2seq(trainx, trainy_event) #y_pred_event = predict_seqs(event_encoder_model, event_decoder_model, testx) #testy_labels_event = [] #for seq in testy_event: # testy_labels_event.append(decode_labels(seq)) #testy_event = testy_labels_event # Print metrics #print "testy: " + str(testy[0]) #print "y_pred: " + str(y_pred[0]) #print "labels: " + str(labels[0]) f1_score = score(testy_labels, y_pred_labels, model, modelname) #f1_score_time = score(testy_time, y_pred_time, time_model, modelname) #f1_score_event = score(testy_event, y_pred_event, event_model, modelname) # Convert the labels back to text #if modelname == "nn": # temp_pred = [] # for y in y_pred: # y_fixed = decode_labels(y) # temp_pred.append(y_fixed) # y_pred = temp_pred # Convert crf output to xml tags test_dict = {} for x in range(len(test_ids)): rec_id = test_ids[x] rec_seq = zip((item[0] for item in test_seqs[x]), y_pred_labels[x]) # Concatenate all the sequences for each record if rec_id not in test_dict: test_dict[rec_id] = [] test_dict[rec_id] = test_dict[rec_id] + rec_seq # TODO: add line breaks??? xml_tree = xmltoseq.seq_to_xml(test_dict, testfile) # write the xml to file if len(outfile) > 0: print "Writing test output to xml file..." xml_tree.write(outfile) subprocess.call(["sed", "-i", "-e", 's/&lt;/</g', outfile]) subprocess.call(["sed", "-i", "-e", 's/&gt;/>/g', outfile]) def train_crf(trainx, trainy): print "training CRF..." crf = CRF( algorithm='lbfgs', c1=0.1, c2=0.1, max_iterations=100, all_possible_transitions=True ) crf.fit(trainx, trainy) return crf def train_seq2seq(trainx, trainy, num_nodes=100, vec_labels=False, loss_function="cosine_proximity", num_epochs=10): trainx = numpy.array(trainx) print "trainx shape: " + str(trainx.shape) trainy = numpy.array(trainy) print "trainy shape: " + str(trainy.shape) input_dim = trainx.shape[-1] output_dim = trainy.shape[-1] input_seq_len = trainx.shape[1] output_seq_len = trainy.shape[1] # Create decoder target data trainy_target = [] zero_lab = data_util.zero_vec(output_dim) if not vec_labels: zero_lab = encode_labels([['O']])[0][0] print "zero_lab shape: " + str(numpy.asarray(zero_lab)) for i in range(trainy.shape[0]): row = trainy[i].tolist() new_row = row[1:] new_row.append(zero_lab) trainy_target.append(new_row) trainy_target = numpy.asarray(trainy_target) print "trainy_target shape: " + str(trainy_target.shape) # Set up the encoder latent_dim = num_nodes dropout = 0.1 encoder_inputs = Input(shape=(None, input_dim)) #seq_len encoder = LSTM(latent_dim, return_state=True) # Encoder-Decoder model encoder_outputs, state_h, state_c = encoder(encoder_inputs) encoder_states = [state_h, state_c] # Set up the decoder, using `encoder_states` as initial state. decoder_inputs = Input(shape=(None, output_dim)) decoder_rnn = LSTM(latent_dim, return_sequences=True, return_state=True) decoder_outputs, d_state_h, d_state_c = decoder_rnn(decoder_inputs, initial_state=encoder_states) decoder_dense = Dense(output_dim, activation='softmax') decoder_outputs = decoder_dense(decoder_outputs) # Define the model that will turn # `encoder_input_data` & `decoder_input_data` into `decoder_target_data` model = Model([encoder_inputs, decoder_inputs], decoder_outputs) model.compile(optimizer='rmsprop', loss=loss_function) model.fit([trainx, trainy], trainy_target, epochs=num_epochs) # Normal RNN #rnn_out = GRU(latent_dim, return_sequences=False)(encoder_inputs) #dropout_out = Dropout(dropout)(rnn_out) #prediction = Dense(output_dim, activation='softmax')(dropout_out) #model = Model(inputs=encoder_inputs, outputs=prediction) #model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics=['accuracy']) #model.fit(trainx, trainy, nb_epoch=20) model.summary() model.save('seq2seq.model') # Create models for inference encoder_model = Model(encoder_inputs, encoder_states) decoder_state_input_h = Input(shape=(latent_dim,)) decoder_state_input_c = Input(shape=(latent_dim,)) decoder_states_inputs = [decoder_state_input_h, decoder_state_input_c] decoder_outputs, state_h, state_c = decoder_rnn(decoder_inputs, initial_state=decoder_states_inputs) decoder_states = [state_h, state_c] decoder_outputs = decoder_dense(decoder_outputs) decoder_model = Model([decoder_inputs] + decoder_states_inputs, [decoder_outputs] + decoder_states) return model, encoder_model, decoder_model, output_dim def decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels=False): # Encode the input as state vectors. states_value = encoder_model.predict(input_seq, batch_size=1) # Generate empty target sequence of length 1. #output_dim = 5 #print "output_dim: " + str(output_dim) target_seq = numpy.zeros((1, 1, int(output_dim))) # Populate the first character of target sequence with the start character. zero_lab = data_util.zero_vec(output_dim) if vec_labels: target_seq[0, 0] = zero_lab else: zero_lab = encode_labels([['O']])[0][0] index = zero_lab.index(1) target_seq[0, 0, index] = 1 # Sampling loop for a batch of sequences # (to simplify, here we assume a batch of size 1). stop_condition = False decoded_sentence = [] while not stop_condition: output_tokens, h, c = decoder_model.predict([target_seq] + states_value) # Sample a token #sampled_token_index = np.argmax(output_tokens[0, -1, :]) #sampled_lab = reverse_target_char_index[sampled_token_index] #print "output_tokens shape: " + str(output_tokens.shape) token = output_tokens[0, -1] #print "token: " + str(token) encoded_label = numpy.zeros((output_dim,), dtype=numpy.int).tolist() if vec_labels: decoded_sentence.append(encoded_label) else: ind = numpy.argmax(token) encoded_label[ind] = 1 #print "encoded_label: " + str(encoded_label) sampled_lab = decode_labels([encoded_label])[0] print "sampled_lab: " + str(sampled_lab) decoded_sentence.append(sampled_lab) # Exit condition: either hit max length or find stop character. if (len(decoded_sentence) > output_seq_len): stop_condition = True # Update the target sequence (of length 1). target_seq = numpy.zeros((1, 1, output_dim)) for x in range(output_dim): target_seq[0, 0, x] = token[x] # Update states states_value = [h, c] return decoded_sentence ''' Predict sequences for test input encoder_model: the encoder model decoder_model: the decoder model testx: the test input: [num_samples, max_seq_len, output_dim) ''' def
(encoder_model, decoder_model, testx, output_seq_len, output_dim, vec_labels=False): testy_pred = [] print "output_seq_len: " + str(output_seq_len) print "output_dim: " + str(output_dim) print "vec_labels: " + str(vec_labels) for test_seq in testx: input_seq = [] input_seq.append(test_seq) input_seq = numpy.array(input_seq) #print "input_seq shape: " + str(input_seq.shape) decoded_sentence = decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels) #print('-') #print('Input seq:', test_seq) #print('Predicted:', decoded_sentence) testy_pred.append(decoded_sentence) return testy_pred def score(testy, y_pred, model, modelname): # Ignore O tags for evaluation if modelname == "crf": labels = list(model.classes_) elif modelname == "nn" or modelname == 'gru': labels = list(label_set) labels.remove('O') f1_score = metrics.flat_f1_score(testy, y_pred, average='weighted', labels=labels) print "F1: " + str(f1_score) sorted_labels = sorted(labels, key=lambda name: (name[1:], name[0])) print(metrics.flat_classification_report(testy, y_pred, labels=sorted_labels, digits=3)) return f1_score def split_labels(y): t_labels = ['BT', 'IT'] e_labels = ['BE', 'IE'] y_time = [] y_event = [] for y_seq in y: time_seq = [] event_seq = [] for lab in y_seq: #print "lab: " + lab if lab in t_labels: time_seq.append(lab) event_seq.append('O') elif lab in e_labels: time_seq.append('O') event_seq.append(lab) else: time_seq.append(lab) event_seq.append(lab) y_time.append(time_seq) y_event.append(event_seq) return y_time, y_event ''' Get word vectors for each word and encode labels seqs: the sequence of pairs (word, label) ''' def get_feats(seqs, train=False): print "get_feats" vec_model, dim = word2vec.load(vecfile) zero_vec = data_util.zero_vec(dim) feats = [] labels = [] global label_set label_set = set([]) for s in seqs: s_feats = [] s_labels = [] for pair in s: word = pair[0] vector = word2vec.get(word, vec_model) s_feats.append(vector) s_labels.append(pair[1]) label_set.add(pair[1]) feats.append(s_feats) labels.append(s_labels) if train: num_labels = len(list(label_set)) create_labelencoder(list(label_set), num_labels) global max_seq_len #max_seq_len = max([len(txt) for txt in feats]) print "max_seq_len: " + str(max_seq_len) # Pad sequences #feats = pad_sequences(numpy.array(feats), maxlen=max_seq_len, dtype='float32', padding="pre") #labels = pad_sequences(numpy.array(labels), maxlen=max_seq_len, dtype='str', padding="pre", value='O') padded_feats = [] padded_labels = [] for feat in feats: #print "seq len: " + str(len(feat)) while len(feat) > max_seq_len: feat_part = feat[0:max_seq_len] padded_feats.append(pad_feat(feat_part, max_seq_len, zero_vec)) feat = feat[max_seq_len:] new_feat = pad_feat(feat, max_seq_len, zero_vec) padded_feats.append(new_feat) for labs in labels: while len(labs) > max_seq_len: labs_part = labs[0:max_seq_len] padded_labels.append(pad_feat(labs_part, max_seq_len, 'O')) labs = labs[max_seq_len:] padded_labels.append(pad_feat(labs, max_seq_len, 'O')) feats = padded_feats labels = padded_labels # Encode labels encoded_labels = encode_labels(labels, max_len=max_seq_len) print "labels[0]: " + str(encoded_labels[0]) #for row in labels: # encoded_row = encode_labels(row) # encoded_labels.append(encoded_row) print "feats: " + str(len(feats)) + " labels: " + str(len(encoded_labels)) return feats, encoded_labels def pad_feat(feat, max_seq_len, pad_item): pad_size = max_seq_len - len(feat) assert(pad_size >= 0) new_feat = [] #new_feat.append(pad_item) # Start symbol for encoder-decoder for w in feat: new_feat.append(w) for k in range(pad_size): new_feat.append(pad_item) return new_feat def get_seqs(filename, split_sents=False, inline=True): print "get_seqs " + filename ids = [] narrs = [] anns = [] seqs = [] seq_ids = [] # Get the xml from file tree = etree.parse(filename) root = tree.getroot() for child in root: narr = "" rec_id = child.find(id_name).text ids.append(rec_id) # Get the narrative text node = child.find("narr_timeml_simple") if inline: if node == None: narr_node = child.find("narrative") if narr_node == None: print "no narrative: " + data_util.stringify_children(child) else: narr = narr_node.text #print "narr: " + narr narrs.append(narr) else: rec_id = child.find(id_name).text #print "rec_id: " + rec_id narr = data_util.stringify_children(node).encode('utf-8') #print "narr: " + narr ids.append(rec_id) narrs.append(narr) else: # NOT inline anns.append(data_util.stringify_children(node).encode('utf8')) narr_node = child.find("narrative") narrs.append(narr_node.text) if inline: for x in range(len(narrs)): narr = narrs[x] rec_id = ids[x] if split_sents: sents = narr.split('.') for sent in sents: sent_seq = xmltoseq.xml_to_seq(sent.strip()) seqs.append(sent_seq) seq_ids.append(rec_id) else: narr_seq = xmltoseq.xml_to_seq(narr) seqs.append(narr_seq) seq_ids.append(rec_id) else: for x in range(len(narrs)): narr = narrs[x] ann = anns[x] rec_id = ids[x] print "split_sents: " + str(split_sents) ann_seqs = xmltoseq.ann_to_seq(narr, ann, split_sents) print "seqs: " + str(len(ann_seqs)) for s in ann_seqs: seqs.append(s) seq_ids.append(rec_id) return seq_ids, seqs def create_labelencoder(data, num=0): global labelencoder, onehotencoder, num_labels print "create_labelencoder: data[0]: " + str(data[0]) labelencoder = LabelEncoder() labelencoder.fit(data) num_labels = len(labelencoder.classes_) #onehotencoder = OneHotEncoder() #onehotencoder.fit(data2) return labelencoder ''' Encodes labels as one-hot vectors (entire dataset: 2D array) data: a 1D array of labels num_labels: the number of label classes ''' def encode_labels(data, labenc=None, max_len=50): if labenc == None: labenc = labelencoder if labenc == None: # or onehotencoder == None: print "Error: labelencoder must be trained before it can be used!" return None #return onehotencoder.transform(labelencoder.transform(data)) data2 = [] num_labels = len(labenc.classes_) zero_vec = data_util.zero_vec(num_labels) print "data: " + str(len(data)) for item in data: #print "item len: " + str(len(item)) new_item = [] if len(item) > 0: item2 = labenc.transform(item) for lab in item2: onehot = [] for x in range(num_labels): onehot.append(0) onehot[lab] = 1 new_item.append(onehot) # Pad vectors if len(new_item) > max_len: new_item = new_item[0:max_len] while len(new_item) < max_len: new_item.append(zero_vec) data2.append(new_item) #else: # data2.append([]) return data2 ''' Decodes one sequence of labels ''' def decode_labels(data, labenc=None): #print "decode_labels" if labenc is None: labenc = labelencoder data2 = [] for row in data: #print "- row: " + str(row) lab = numpy.argmax(numpy.asarray(row)) #print "- lab: " + str(lab) data2.append(lab) #print "- data2: " + str(data2) return labenc.inverse_transform(data2) #return labelencoder.inverse_transform(onehotencoder.reverse_transform(data)) def decode_all_labels(data, labenc=None): decoded_labels = [] for sequence in data: labs = decode_labels(sequence, labenc) decoded_labels.append(labs) return decoded_labels def word2features(sent, i): word = sent[i][0] #postag = sent[i][1] features = { 'bias': 1.0, 'word.lower()': word.lower(), 'word[-3:]': word[-3:], 'word[-2:]': word[-2:], 'word.isupper()': word.isupper(), 'word.istitle()': word.istitle(), 'word.isdigit()': word.isdigit(), #'postag': postag, #'postag[:2]': postag[:2], } if i > 0: word1 = sent[i-1][0] postag1 = sent[i-1][1] features.update({ '-1:word.lower()': word1.lower(), '-1:word.istitle()': word1.istitle(), '-1:word.isupper()': word1.isupper(), #'-1:postag': postag1, #'-1:postag[:2]': postag1[:2], }) else: features['BOS'] = True if i < len(sent)-1: word1 = sent[i+1][0] postag1 = sent[i+1][1] features.update({ '+1:word.lower()': word1.lower(), '+1:word.istitle()': word1.istitle(), '+1:word.isupper()': word1.isupper(), #'+1:postag': postag1, #'+1:postag[:2]': postag1[:2], }) else: features['EOS'] = True return features def sent2features(sent): return [word2features(sent, i) for i in range(len(sent))] def sent2labels(sent): return [label for token, label in sent] def sent2tokens(sent): return [token for token, label in sent] def ispunc(input_string, start, end): punc = ' :;,./?' s = input_string[start:end] for char in s: if char not in punc: return False return True if __name__ == "__main__":main()
predict_seqs
identifier_name
model_seq.py
#!/usr/bin/python # -*- coding: utf-8 -*- # Identify and label symptom phrases in narrative import sys sys.path.append('/u/sjeblee/research/va/git/verbal-autopsy') import data_util import models_pytorch import word2vec import xmltoseq from keras.models import Model from keras.layers import Input, GRU, LSTM, Dense, Dropout from keras.preprocessing.sequence import pad_sequences from lxml import etree from sklearn.preprocessing import LabelEncoder, OneHotEncoder from sklearn_crfsuite import CRF, metrics import argparse import numpy import subprocess import warnings warnings.filterwarnings("ignore", category=DeprecationWarning) global labelencoder, onehotencoder, label_set, max_seq_len, num_labels id_name = "record_id" max_seq_len = 100 #vecfile = "/u/sjeblee/research/va/data/datasets/mds+rct/narr+ice+medhelp.vectors.100" vecfile = "/u/sjeblee/research/vectors/GoogleNews-vectors-negative300.bin" def main(): argparser = argparse.ArgumentParser() argparser.add_argument('--train', action="store", dest="trainfile") argparser.add_argument('--test', action="store", dest="testfile") argparser.add_argument('--out', action="store", dest="outfile") argparser.add_argument('--model', action="store", dest="model") args = argparser.parse_args() if not (args.trainfile and args.testfile): print "usage: ./model_seq.py --train [file.xml] --test [file.xml] (--out [file.xml] --model [crf/gru])" exit() if args.outfile and args.model: run(args.trainfile, args.testfile, args.outfile, args.model) elif args.outfile: run(args.trainfile, args.testfile, args.outfile) else: run(args.trainfile, args.testfile) def run(trainfile, testfile, outfile="", modelname="crf"): train_ids, train_seqs = get_seqs(trainfile, split_sents=True, inline=False) test_ids, test_seqs = get_seqs(testfile, split_sents=True, inline=False) if modelname == "crf": trainx = [sent2features(s) for s in train_seqs] trainy = [sent2labels(s) for s in train_seqs] testx = [sent2features(s) for s in test_seqs] testy = [sent2labels(s) for s in test_seqs] elif modelname == "nn": extra_seqs = get_seqs('/u/sjeblee/research/data/TimeBank/timebank_all_timeml_simple.xml', split_sents=True) train_seqs = extra_seqs + train_seqs trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) elif modelname == 'gru': trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) print "train: " + str(len(trainx)) + " " + str(len(trainy)) print "test: " + str(len(testx)) + " " + str(len(testy)) # Split labels into time and event only split = False if split: trainy_time, trainy_event = split_labels(trainy) testy_time, testy_event = split_labels(testy) # Train model model = None encoder_model = None decoder_model = None y_pred = [] if modelname == "crf": model = train_crf(trainx, trainy) #crf_time = train_crf(trainx, trainy_time) #crf_event = train_crf(trainx, trainy_event) # Test CRF testy_labels = testy y_pred_labels = model.predict(testx) #y_pred_time = crf_time.predict(testx) #y_pred_event = crf_event.predict(testx) elif modelname == 'gru': nodes = 100 epochs = 5 model = models_pytorch.rnn_model(trainx, trainy, num_nodes=nodes, loss_function='mse', num_epochs=epochs, use_prev_labels=True) y_pred = models_pytorch.test_rnn(model, testx) y_pred_labels = decode_all_labels(y_pred) testy_labels = decode_all_labels(testy) elif modelname == "ende": model, encoder_model, decoder_model, dim = train_seq2seq(trainx, trainy) y_pred = predict_seqs(encoder_model, decoder_model, testx, dim) testy_labels = [] for seq in testy: testy_labels.append(decode_labels(seq)) testy = testy_labels # Separate time and event models #time_model, time_encoder_model, time_decoder_model = train_seq2seq(trainx, trainy_time) #y_pred_time = predict_seqs(time_encoder_model, time_decoder_model, testx) #testy_labels_time = [] #for seq in testy_time: # testy_labels_time.append(decode_labels(seq)) #testy_time = testy_labels_time #event_model, event_encoder_model, event_decoder_model = train_seq2seq(trainx, trainy_event) #y_pred_event = predict_seqs(event_encoder_model, event_decoder_model, testx) #testy_labels_event = [] #for seq in testy_event: # testy_labels_event.append(decode_labels(seq)) #testy_event = testy_labels_event # Print metrics #print "testy: " + str(testy[0]) #print "y_pred: " + str(y_pred[0]) #print "labels: " + str(labels[0]) f1_score = score(testy_labels, y_pred_labels, model, modelname) #f1_score_time = score(testy_time, y_pred_time, time_model, modelname) #f1_score_event = score(testy_event, y_pred_event, event_model, modelname) # Convert the labels back to text #if modelname == "nn": # temp_pred = [] # for y in y_pred: # y_fixed = decode_labels(y) # temp_pred.append(y_fixed) # y_pred = temp_pred # Convert crf output to xml tags test_dict = {} for x in range(len(test_ids)): rec_id = test_ids[x] rec_seq = zip((item[0] for item in test_seqs[x]), y_pred_labels[x]) # Concatenate all the sequences for each record if rec_id not in test_dict: test_dict[rec_id] = [] test_dict[rec_id] = test_dict[rec_id] + rec_seq # TODO: add line breaks??? xml_tree = xmltoseq.seq_to_xml(test_dict, testfile) # write the xml to file if len(outfile) > 0: print "Writing test output to xml file..." xml_tree.write(outfile) subprocess.call(["sed", "-i", "-e", 's/&lt;/</g', outfile]) subprocess.call(["sed", "-i", "-e", 's/&gt;/>/g', outfile]) def train_crf(trainx, trainy): print "training CRF..." crf = CRF( algorithm='lbfgs', c1=0.1, c2=0.1, max_iterations=100, all_possible_transitions=True ) crf.fit(trainx, trainy) return crf def train_seq2seq(trainx, trainy, num_nodes=100, vec_labels=False, loss_function="cosine_proximity", num_epochs=10): trainx = numpy.array(trainx) print "trainx shape: " + str(trainx.shape) trainy = numpy.array(trainy) print "trainy shape: " + str(trainy.shape) input_dim = trainx.shape[-1] output_dim = trainy.shape[-1] input_seq_len = trainx.shape[1] output_seq_len = trainy.shape[1] # Create decoder target data trainy_target = [] zero_lab = data_util.zero_vec(output_dim) if not vec_labels: zero_lab = encode_labels([['O']])[0][0] print "zero_lab shape: " + str(numpy.asarray(zero_lab)) for i in range(trainy.shape[0]): row = trainy[i].tolist() new_row = row[1:] new_row.append(zero_lab) trainy_target.append(new_row) trainy_target = numpy.asarray(trainy_target) print "trainy_target shape: " + str(trainy_target.shape) # Set up the encoder latent_dim = num_nodes dropout = 0.1 encoder_inputs = Input(shape=(None, input_dim)) #seq_len encoder = LSTM(latent_dim, return_state=True) # Encoder-Decoder model encoder_outputs, state_h, state_c = encoder(encoder_inputs) encoder_states = [state_h, state_c] # Set up the decoder, using `encoder_states` as initial state. decoder_inputs = Input(shape=(None, output_dim)) decoder_rnn = LSTM(latent_dim, return_sequences=True, return_state=True) decoder_outputs, d_state_h, d_state_c = decoder_rnn(decoder_inputs, initial_state=encoder_states) decoder_dense = Dense(output_dim, activation='softmax') decoder_outputs = decoder_dense(decoder_outputs) # Define the model that will turn # `encoder_input_data` & `decoder_input_data` into `decoder_target_data` model = Model([encoder_inputs, decoder_inputs], decoder_outputs) model.compile(optimizer='rmsprop', loss=loss_function) model.fit([trainx, trainy], trainy_target, epochs=num_epochs) # Normal RNN #rnn_out = GRU(latent_dim, return_sequences=False)(encoder_inputs) #dropout_out = Dropout(dropout)(rnn_out) #prediction = Dense(output_dim, activation='softmax')(dropout_out) #model = Model(inputs=encoder_inputs, outputs=prediction) #model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics=['accuracy']) #model.fit(trainx, trainy, nb_epoch=20) model.summary() model.save('seq2seq.model') # Create models for inference encoder_model = Model(encoder_inputs, encoder_states) decoder_state_input_h = Input(shape=(latent_dim,)) decoder_state_input_c = Input(shape=(latent_dim,)) decoder_states_inputs = [decoder_state_input_h, decoder_state_input_c] decoder_outputs, state_h, state_c = decoder_rnn(decoder_inputs, initial_state=decoder_states_inputs) decoder_states = [state_h, state_c] decoder_outputs = decoder_dense(decoder_outputs) decoder_model = Model([decoder_inputs] + decoder_states_inputs, [decoder_outputs] + decoder_states) return model, encoder_model, decoder_model, output_dim def decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels=False): # Encode the input as state vectors. states_value = encoder_model.predict(input_seq, batch_size=1) # Generate empty target sequence of length 1. #output_dim = 5 #print "output_dim: " + str(output_dim) target_seq = numpy.zeros((1, 1, int(output_dim))) # Populate the first character of target sequence with the start character. zero_lab = data_util.zero_vec(output_dim) if vec_labels: target_seq[0, 0] = zero_lab else: zero_lab = encode_labels([['O']])[0][0] index = zero_lab.index(1) target_seq[0, 0, index] = 1 # Sampling loop for a batch of sequences # (to simplify, here we assume a batch of size 1). stop_condition = False decoded_sentence = [] while not stop_condition: output_tokens, h, c = decoder_model.predict([target_seq] + states_value) # Sample a token #sampled_token_index = np.argmax(output_tokens[0, -1, :]) #sampled_lab = reverse_target_char_index[sampled_token_index] #print "output_tokens shape: " + str(output_tokens.shape) token = output_tokens[0, -1] #print "token: " + str(token) encoded_label = numpy.zeros((output_dim,), dtype=numpy.int).tolist() if vec_labels: decoded_sentence.append(encoded_label) else: ind = numpy.argmax(token) encoded_label[ind] = 1 #print "encoded_label: " + str(encoded_label) sampled_lab = decode_labels([encoded_label])[0] print "sampled_lab: " + str(sampled_lab) decoded_sentence.append(sampled_lab) # Exit condition: either hit max length or find stop character. if (len(decoded_sentence) > output_seq_len): stop_condition = True # Update the target sequence (of length 1). target_seq = numpy.zeros((1, 1, output_dim)) for x in range(output_dim): target_seq[0, 0, x] = token[x] # Update states states_value = [h, c] return decoded_sentence ''' Predict sequences for test input encoder_model: the encoder model decoder_model: the decoder model testx: the test input: [num_samples, max_seq_len, output_dim) ''' def predict_seqs(encoder_model, decoder_model, testx, output_seq_len, output_dim, vec_labels=False): testy_pred = [] print "output_seq_len: " + str(output_seq_len) print "output_dim: " + str(output_dim) print "vec_labels: " + str(vec_labels) for test_seq in testx: input_seq = [] input_seq.append(test_seq) input_seq = numpy.array(input_seq) #print "input_seq shape: " + str(input_seq.shape) decoded_sentence = decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels) #print('-') #print('Input seq:', test_seq) #print('Predicted:', decoded_sentence) testy_pred.append(decoded_sentence) return testy_pred def score(testy, y_pred, model, modelname): # Ignore O tags for evaluation if modelname == "crf": labels = list(model.classes_) elif modelname == "nn" or modelname == 'gru': labels = list(label_set) labels.remove('O') f1_score = metrics.flat_f1_score(testy, y_pred, average='weighted', labels=labels) print "F1: " + str(f1_score) sorted_labels = sorted(labels, key=lambda name: (name[1:], name[0])) print(metrics.flat_classification_report(testy, y_pred, labels=sorted_labels, digits=3)) return f1_score def split_labels(y): t_labels = ['BT', 'IT'] e_labels = ['BE', 'IE'] y_time = [] y_event = [] for y_seq in y: time_seq = [] event_seq = [] for lab in y_seq: #print "lab: " + lab if lab in t_labels: time_seq.append(lab) event_seq.append('O') elif lab in e_labels: time_seq.append('O') event_seq.append(lab) else: time_seq.append(lab) event_seq.append(lab) y_time.append(time_seq) y_event.append(event_seq) return y_time, y_event ''' Get word vectors for each word and encode labels seqs: the sequence of pairs (word, label) ''' def get_feats(seqs, train=False): print "get_feats" vec_model, dim = word2vec.load(vecfile) zero_vec = data_util.zero_vec(dim) feats = [] labels = [] global label_set label_set = set([]) for s in seqs: s_feats = [] s_labels = [] for pair in s: word = pair[0] vector = word2vec.get(word, vec_model) s_feats.append(vector) s_labels.append(pair[1]) label_set.add(pair[1]) feats.append(s_feats) labels.append(s_labels) if train: num_labels = len(list(label_set)) create_labelencoder(list(label_set), num_labels) global max_seq_len #max_seq_len = max([len(txt) for txt in feats]) print "max_seq_len: " + str(max_seq_len) # Pad sequences #feats = pad_sequences(numpy.array(feats), maxlen=max_seq_len, dtype='float32', padding="pre") #labels = pad_sequences(numpy.array(labels), maxlen=max_seq_len, dtype='str', padding="pre", value='O') padded_feats = [] padded_labels = [] for feat in feats: #print "seq len: " + str(len(feat)) while len(feat) > max_seq_len: feat_part = feat[0:max_seq_len] padded_feats.append(pad_feat(feat_part, max_seq_len, zero_vec)) feat = feat[max_seq_len:] new_feat = pad_feat(feat, max_seq_len, zero_vec) padded_feats.append(new_feat) for labs in labels: while len(labs) > max_seq_len: labs_part = labs[0:max_seq_len] padded_labels.append(pad_feat(labs_part, max_seq_len, 'O')) labs = labs[max_seq_len:] padded_labels.append(pad_feat(labs, max_seq_len, 'O')) feats = padded_feats labels = padded_labels # Encode labels encoded_labels = encode_labels(labels, max_len=max_seq_len) print "labels[0]: " + str(encoded_labels[0]) #for row in labels: # encoded_row = encode_labels(row) # encoded_labels.append(encoded_row) print "feats: " + str(len(feats)) + " labels: " + str(len(encoded_labels)) return feats, encoded_labels def pad_feat(feat, max_seq_len, pad_item): pad_size = max_seq_len - len(feat) assert(pad_size >= 0) new_feat = [] #new_feat.append(pad_item) # Start symbol for encoder-decoder for w in feat: new_feat.append(w) for k in range(pad_size): new_feat.append(pad_item) return new_feat def get_seqs(filename, split_sents=False, inline=True): print "get_seqs " + filename ids = [] narrs = [] anns = [] seqs = [] seq_ids = [] # Get the xml from file tree = etree.parse(filename) root = tree.getroot() for child in root: narr = "" rec_id = child.find(id_name).text ids.append(rec_id) # Get the narrative text node = child.find("narr_timeml_simple") if inline: if node == None: narr_node = child.find("narrative") if narr_node == None: print "no narrative: " + data_util.stringify_children(child) else: narr = narr_node.text #print "narr: " + narr narrs.append(narr) else: rec_id = child.find(id_name).text #print "rec_id: " + rec_id narr = data_util.stringify_children(node).encode('utf-8') #print "narr: " + narr ids.append(rec_id) narrs.append(narr) else: # NOT inline anns.append(data_util.stringify_children(node).encode('utf8')) narr_node = child.find("narrative") narrs.append(narr_node.text) if inline: for x in range(len(narrs)): narr = narrs[x] rec_id = ids[x] if split_sents: sents = narr.split('.') for sent in sents: sent_seq = xmltoseq.xml_to_seq(sent.strip()) seqs.append(sent_seq) seq_ids.append(rec_id) else: narr_seq = xmltoseq.xml_to_seq(narr) seqs.append(narr_seq) seq_ids.append(rec_id) else: for x in range(len(narrs)): narr = narrs[x] ann = anns[x] rec_id = ids[x] print "split_sents: " + str(split_sents) ann_seqs = xmltoseq.ann_to_seq(narr, ann, split_sents) print "seqs: " + str(len(ann_seqs)) for s in ann_seqs: seqs.append(s) seq_ids.append(rec_id) return seq_ids, seqs def create_labelencoder(data, num=0): global labelencoder, onehotencoder, num_labels print "create_labelencoder: data[0]: " + str(data[0]) labelencoder = LabelEncoder() labelencoder.fit(data) num_labels = len(labelencoder.classes_) #onehotencoder = OneHotEncoder() #onehotencoder.fit(data2) return labelencoder ''' Encodes labels as one-hot vectors (entire dataset: 2D array) data: a 1D array of labels num_labels: the number of label classes ''' def encode_labels(data, labenc=None, max_len=50): if labenc == None: labenc = labelencoder if labenc == None: # or onehotencoder == None: print "Error: labelencoder must be trained before it can be used!" return None #return onehotencoder.transform(labelencoder.transform(data)) data2 = [] num_labels = len(labenc.classes_) zero_vec = data_util.zero_vec(num_labels) print "data: " + str(len(data)) for item in data: #print "item len: " + str(len(item)) new_item = [] if len(item) > 0: item2 = labenc.transform(item) for lab in item2: onehot = [] for x in range(num_labels): onehot.append(0) onehot[lab] = 1 new_item.append(onehot) # Pad vectors if len(new_item) > max_len: new_item = new_item[0:max_len] while len(new_item) < max_len: new_item.append(zero_vec) data2.append(new_item) #else: # data2.append([]) return data2 ''' Decodes one sequence of labels ''' def decode_labels(data, labenc=None): #print "decode_labels" if labenc is None: labenc = labelencoder data2 = [] for row in data: #print "- row: " + str(row) lab = numpy.argmax(numpy.asarray(row)) #print "- lab: " + str(lab) data2.append(lab) #print "- data2: " + str(data2) return labenc.inverse_transform(data2) #return labelencoder.inverse_transform(onehotencoder.reverse_transform(data)) def decode_all_labels(data, labenc=None): decoded_labels = [] for sequence in data: labs = decode_labels(sequence, labenc) decoded_labels.append(labs) return decoded_labels def word2features(sent, i): word = sent[i][0] #postag = sent[i][1] features = { 'bias': 1.0, 'word.lower()': word.lower(), 'word[-3:]': word[-3:], 'word[-2:]': word[-2:], 'word.isupper()': word.isupper(), 'word.istitle()': word.istitle(), 'word.isdigit()': word.isdigit(), #'postag': postag, #'postag[:2]': postag[:2], } if i > 0: word1 = sent[i-1][0] postag1 = sent[i-1][1] features.update({ '-1:word.lower()': word1.lower(), '-1:word.istitle()': word1.istitle(), '-1:word.isupper()': word1.isupper(), #'-1:postag': postag1, #'-1:postag[:2]': postag1[:2], }) else: features['BOS'] = True if i < len(sent)-1: word1 = sent[i+1][0] postag1 = sent[i+1][1] features.update({ '+1:word.lower()': word1.lower(), '+1:word.istitle()': word1.istitle(), '+1:word.isupper()': word1.isupper(), #'+1:postag': postag1, #'+1:postag[:2]': postag1[:2], }) else: features['EOS'] = True return features def sent2features(sent): return [word2features(sent, i) for i in range(len(sent))] def sent2labels(sent):
def sent2tokens(sent): return [token for token, label in sent] def ispunc(input_string, start, end): punc = ' :;,./?' s = input_string[start:end] for char in s: if char not in punc: return False return True if __name__ == "__main__":main()
return [label for token, label in sent]
identifier_body
model_seq.py
#!/usr/bin/python # -*- coding: utf-8 -*- # Identify and label symptom phrases in narrative import sys sys.path.append('/u/sjeblee/research/va/git/verbal-autopsy') import data_util import models_pytorch import word2vec import xmltoseq from keras.models import Model from keras.layers import Input, GRU, LSTM, Dense, Dropout from keras.preprocessing.sequence import pad_sequences from lxml import etree from sklearn.preprocessing import LabelEncoder, OneHotEncoder from sklearn_crfsuite import CRF, metrics import argparse import numpy import subprocess import warnings warnings.filterwarnings("ignore", category=DeprecationWarning) global labelencoder, onehotencoder, label_set, max_seq_len, num_labels id_name = "record_id" max_seq_len = 100 #vecfile = "/u/sjeblee/research/va/data/datasets/mds+rct/narr+ice+medhelp.vectors.100" vecfile = "/u/sjeblee/research/vectors/GoogleNews-vectors-negative300.bin" def main(): argparser = argparse.ArgumentParser() argparser.add_argument('--train', action="store", dest="trainfile") argparser.add_argument('--test', action="store", dest="testfile") argparser.add_argument('--out', action="store", dest="outfile") argparser.add_argument('--model', action="store", dest="model") args = argparser.parse_args() if not (args.trainfile and args.testfile): print "usage: ./model_seq.py --train [file.xml] --test [file.xml] (--out [file.xml] --model [crf/gru])" exit() if args.outfile and args.model: run(args.trainfile, args.testfile, args.outfile, args.model) elif args.outfile: run(args.trainfile, args.testfile, args.outfile) else: run(args.trainfile, args.testfile) def run(trainfile, testfile, outfile="", modelname="crf"): train_ids, train_seqs = get_seqs(trainfile, split_sents=True, inline=False) test_ids, test_seqs = get_seqs(testfile, split_sents=True, inline=False) if modelname == "crf": trainx = [sent2features(s) for s in train_seqs] trainy = [sent2labels(s) for s in train_seqs] testx = [sent2features(s) for s in test_seqs] testy = [sent2labels(s) for s in test_seqs] elif modelname == "nn": extra_seqs = get_seqs('/u/sjeblee/research/data/TimeBank/timebank_all_timeml_simple.xml', split_sents=True) train_seqs = extra_seqs + train_seqs trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) elif modelname == 'gru': trainx, trainy = get_feats(train_seqs, True) testx, testy = get_feats(test_seqs) print "train: " + str(len(trainx)) + " " + str(len(trainy)) print "test: " + str(len(testx)) + " " + str(len(testy)) # Split labels into time and event only split = False if split: trainy_time, trainy_event = split_labels(trainy) testy_time, testy_event = split_labels(testy) # Train model model = None encoder_model = None decoder_model = None y_pred = [] if modelname == "crf": model = train_crf(trainx, trainy) #crf_time = train_crf(trainx, trainy_time) #crf_event = train_crf(trainx, trainy_event) # Test CRF testy_labels = testy y_pred_labels = model.predict(testx) #y_pred_time = crf_time.predict(testx) #y_pred_event = crf_event.predict(testx) elif modelname == 'gru': nodes = 100 epochs = 5 model = models_pytorch.rnn_model(trainx, trainy, num_nodes=nodes, loss_function='mse', num_epochs=epochs, use_prev_labels=True) y_pred = models_pytorch.test_rnn(model, testx) y_pred_labels = decode_all_labels(y_pred) testy_labels = decode_all_labels(testy) elif modelname == "ende": model, encoder_model, decoder_model, dim = train_seq2seq(trainx, trainy) y_pred = predict_seqs(encoder_model, decoder_model, testx, dim) testy_labels = [] for seq in testy: testy_labels.append(decode_labels(seq)) testy = testy_labels # Separate time and event models #time_model, time_encoder_model, time_decoder_model = train_seq2seq(trainx, trainy_time) #y_pred_time = predict_seqs(time_encoder_model, time_decoder_model, testx) #testy_labels_time = [] #for seq in testy_time: # testy_labels_time.append(decode_labels(seq)) #testy_time = testy_labels_time #event_model, event_encoder_model, event_decoder_model = train_seq2seq(trainx, trainy_event) #y_pred_event = predict_seqs(event_encoder_model, event_decoder_model, testx) #testy_labels_event = [] #for seq in testy_event: # testy_labels_event.append(decode_labels(seq)) #testy_event = testy_labels_event # Print metrics #print "testy: " + str(testy[0]) #print "y_pred: " + str(y_pred[0]) #print "labels: " + str(labels[0]) f1_score = score(testy_labels, y_pred_labels, model, modelname) #f1_score_time = score(testy_time, y_pred_time, time_model, modelname) #f1_score_event = score(testy_event, y_pred_event, event_model, modelname) # Convert the labels back to text #if modelname == "nn": # temp_pred = [] # for y in y_pred: # y_fixed = decode_labels(y) # temp_pred.append(y_fixed) # y_pred = temp_pred # Convert crf output to xml tags test_dict = {} for x in range(len(test_ids)): rec_id = test_ids[x] rec_seq = zip((item[0] for item in test_seqs[x]), y_pred_labels[x]) # Concatenate all the sequences for each record if rec_id not in test_dict: test_dict[rec_id] = [] test_dict[rec_id] = test_dict[rec_id] + rec_seq # TODO: add line breaks??? xml_tree = xmltoseq.seq_to_xml(test_dict, testfile) # write the xml to file if len(outfile) > 0: print "Writing test output to xml file..." xml_tree.write(outfile) subprocess.call(["sed", "-i", "-e", 's/&lt;/</g', outfile]) subprocess.call(["sed", "-i", "-e", 's/&gt;/>/g', outfile]) def train_crf(trainx, trainy): print "training CRF..." crf = CRF( algorithm='lbfgs', c1=0.1, c2=0.1, max_iterations=100, all_possible_transitions=True ) crf.fit(trainx, trainy) return crf def train_seq2seq(trainx, trainy, num_nodes=100, vec_labels=False, loss_function="cosine_proximity", num_epochs=10): trainx = numpy.array(trainx) print "trainx shape: " + str(trainx.shape) trainy = numpy.array(trainy) print "trainy shape: " + str(trainy.shape) input_dim = trainx.shape[-1] output_dim = trainy.shape[-1] input_seq_len = trainx.shape[1] output_seq_len = trainy.shape[1] # Create decoder target data trainy_target = [] zero_lab = data_util.zero_vec(output_dim) if not vec_labels: zero_lab = encode_labels([['O']])[0][0] print "zero_lab shape: " + str(numpy.asarray(zero_lab)) for i in range(trainy.shape[0]): row = trainy[i].tolist() new_row = row[1:] new_row.append(zero_lab) trainy_target.append(new_row) trainy_target = numpy.asarray(trainy_target) print "trainy_target shape: " + str(trainy_target.shape) # Set up the encoder latent_dim = num_nodes dropout = 0.1 encoder_inputs = Input(shape=(None, input_dim)) #seq_len encoder = LSTM(latent_dim, return_state=True) # Encoder-Decoder model encoder_outputs, state_h, state_c = encoder(encoder_inputs) encoder_states = [state_h, state_c] # Set up the decoder, using `encoder_states` as initial state. decoder_inputs = Input(shape=(None, output_dim)) decoder_rnn = LSTM(latent_dim, return_sequences=True, return_state=True) decoder_outputs, d_state_h, d_state_c = decoder_rnn(decoder_inputs, initial_state=encoder_states) decoder_dense = Dense(output_dim, activation='softmax') decoder_outputs = decoder_dense(decoder_outputs) # Define the model that will turn # `encoder_input_data` & `decoder_input_data` into `decoder_target_data` model = Model([encoder_inputs, decoder_inputs], decoder_outputs) model.compile(optimizer='rmsprop', loss=loss_function) model.fit([trainx, trainy], trainy_target, epochs=num_epochs) # Normal RNN #rnn_out = GRU(latent_dim, return_sequences=False)(encoder_inputs) #dropout_out = Dropout(dropout)(rnn_out) #prediction = Dense(output_dim, activation='softmax')(dropout_out) #model = Model(inputs=encoder_inputs, outputs=prediction) #model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics=['accuracy']) #model.fit(trainx, trainy, nb_epoch=20) model.summary() model.save('seq2seq.model') # Create models for inference encoder_model = Model(encoder_inputs, encoder_states) decoder_state_input_h = Input(shape=(latent_dim,)) decoder_state_input_c = Input(shape=(latent_dim,)) decoder_states_inputs = [decoder_state_input_h, decoder_state_input_c] decoder_outputs, state_h, state_c = decoder_rnn(decoder_inputs, initial_state=decoder_states_inputs) decoder_states = [state_h, state_c] decoder_outputs = decoder_dense(decoder_outputs) decoder_model = Model([decoder_inputs] + decoder_states_inputs, [decoder_outputs] + decoder_states) return model, encoder_model, decoder_model, output_dim def decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels=False): # Encode the input as state vectors. states_value = encoder_model.predict(input_seq, batch_size=1) # Generate empty target sequence of length 1. #output_dim = 5 #print "output_dim: " + str(output_dim) target_seq = numpy.zeros((1, 1, int(output_dim))) # Populate the first character of target sequence with the start character. zero_lab = data_util.zero_vec(output_dim) if vec_labels: target_seq[0, 0] = zero_lab else: zero_lab = encode_labels([['O']])[0][0] index = zero_lab.index(1) target_seq[0, 0, index] = 1 # Sampling loop for a batch of sequences # (to simplify, here we assume a batch of size 1). stop_condition = False decoded_sentence = [] while not stop_condition: output_tokens, h, c = decoder_model.predict([target_seq] + states_value) # Sample a token #sampled_token_index = np.argmax(output_tokens[0, -1, :]) #sampled_lab = reverse_target_char_index[sampled_token_index] #print "output_tokens shape: " + str(output_tokens.shape) token = output_tokens[0, -1] #print "token: " + str(token) encoded_label = numpy.zeros((output_dim,), dtype=numpy.int).tolist() if vec_labels: decoded_sentence.append(encoded_label) else: ind = numpy.argmax(token) encoded_label[ind] = 1 #print "encoded_label: " + str(encoded_label) sampled_lab = decode_labels([encoded_label])[0] print "sampled_lab: " + str(sampled_lab) decoded_sentence.append(sampled_lab) # Exit condition: either hit max length or find stop character. if (len(decoded_sentence) > output_seq_len): stop_condition = True # Update the target sequence (of length 1). target_seq = numpy.zeros((1, 1, output_dim)) for x in range(output_dim): target_seq[0, 0, x] = token[x] # Update states states_value = [h, c] return decoded_sentence ''' Predict sequences for test input encoder_model: the encoder model decoder_model: the decoder model testx: the test input: [num_samples, max_seq_len, output_dim) ''' def predict_seqs(encoder_model, decoder_model, testx, output_seq_len, output_dim, vec_labels=False): testy_pred = [] print "output_seq_len: " + str(output_seq_len) print "output_dim: " + str(output_dim) print "vec_labels: " + str(vec_labels) for test_seq in testx: input_seq = [] input_seq.append(test_seq) input_seq = numpy.array(input_seq) #print "input_seq shape: " + str(input_seq.shape) decoded_sentence = decode_sequence(encoder_model, decoder_model, input_seq, output_seq_len, output_dim, vec_labels) #print('-') #print('Input seq:', test_seq) #print('Predicted:', decoded_sentence) testy_pred.append(decoded_sentence) return testy_pred def score(testy, y_pred, model, modelname): # Ignore O tags for evaluation if modelname == "crf": labels = list(model.classes_) elif modelname == "nn" or modelname == 'gru': labels = list(label_set) labels.remove('O') f1_score = metrics.flat_f1_score(testy, y_pred, average='weighted', labels=labels) print "F1: " + str(f1_score) sorted_labels = sorted(labels, key=lambda name: (name[1:], name[0])) print(metrics.flat_classification_report(testy, y_pred, labels=sorted_labels, digits=3)) return f1_score def split_labels(y): t_labels = ['BT', 'IT'] e_labels = ['BE', 'IE'] y_time = [] y_event = [] for y_seq in y: time_seq = [] event_seq = [] for lab in y_seq: #print "lab: " + lab if lab in t_labels: time_seq.append(lab) event_seq.append('O') elif lab in e_labels: time_seq.append('O') event_seq.append(lab) else: time_seq.append(lab) event_seq.append(lab) y_time.append(time_seq) y_event.append(event_seq) return y_time, y_event ''' Get word vectors for each word and encode labels seqs: the sequence of pairs (word, label) ''' def get_feats(seqs, train=False): print "get_feats" vec_model, dim = word2vec.load(vecfile) zero_vec = data_util.zero_vec(dim) feats = [] labels = [] global label_set label_set = set([]) for s in seqs: s_feats = [] s_labels = [] for pair in s: word = pair[0] vector = word2vec.get(word, vec_model) s_feats.append(vector) s_labels.append(pair[1]) label_set.add(pair[1]) feats.append(s_feats) labels.append(s_labels) if train: num_labels = len(list(label_set)) create_labelencoder(list(label_set), num_labels) global max_seq_len #max_seq_len = max([len(txt) for txt in feats]) print "max_seq_len: " + str(max_seq_len) # Pad sequences #feats = pad_sequences(numpy.array(feats), maxlen=max_seq_len, dtype='float32', padding="pre") #labels = pad_sequences(numpy.array(labels), maxlen=max_seq_len, dtype='str', padding="pre", value='O') padded_feats = [] padded_labels = [] for feat in feats: #print "seq len: " + str(len(feat)) while len(feat) > max_seq_len: feat_part = feat[0:max_seq_len] padded_feats.append(pad_feat(feat_part, max_seq_len, zero_vec)) feat = feat[max_seq_len:] new_feat = pad_feat(feat, max_seq_len, zero_vec) padded_feats.append(new_feat) for labs in labels: while len(labs) > max_seq_len: labs_part = labs[0:max_seq_len] padded_labels.append(pad_feat(labs_part, max_seq_len, 'O')) labs = labs[max_seq_len:] padded_labels.append(pad_feat(labs, max_seq_len, 'O')) feats = padded_feats labels = padded_labels # Encode labels encoded_labels = encode_labels(labels, max_len=max_seq_len) print "labels[0]: " + str(encoded_labels[0]) #for row in labels: # encoded_row = encode_labels(row) # encoded_labels.append(encoded_row) print "feats: " + str(len(feats)) + " labels: " + str(len(encoded_labels)) return feats, encoded_labels def pad_feat(feat, max_seq_len, pad_item): pad_size = max_seq_len - len(feat) assert(pad_size >= 0) new_feat = [] #new_feat.append(pad_item) # Start symbol for encoder-decoder for w in feat: new_feat.append(w) for k in range(pad_size): new_feat.append(pad_item) return new_feat def get_seqs(filename, split_sents=False, inline=True): print "get_seqs " + filename ids = [] narrs = [] anns = [] seqs = [] seq_ids = [] # Get the xml from file tree = etree.parse(filename) root = tree.getroot() for child in root: narr = "" rec_id = child.find(id_name).text ids.append(rec_id) # Get the narrative text node = child.find("narr_timeml_simple") if inline: if node == None: narr_node = child.find("narrative") if narr_node == None: print "no narrative: " + data_util.stringify_children(child) else: narr = narr_node.text #print "narr: " + narr narrs.append(narr) else: rec_id = child.find(id_name).text #print "rec_id: " + rec_id narr = data_util.stringify_children(node).encode('utf-8') #print "narr: " + narr ids.append(rec_id) narrs.append(narr) else: # NOT inline anns.append(data_util.stringify_children(node).encode('utf8')) narr_node = child.find("narrative") narrs.append(narr_node.text) if inline: for x in range(len(narrs)): narr = narrs[x] rec_id = ids[x] if split_sents: sents = narr.split('.') for sent in sents: sent_seq = xmltoseq.xml_to_seq(sent.strip()) seqs.append(sent_seq) seq_ids.append(rec_id) else: narr_seq = xmltoseq.xml_to_seq(narr) seqs.append(narr_seq) seq_ids.append(rec_id) else: for x in range(len(narrs)): narr = narrs[x] ann = anns[x] rec_id = ids[x] print "split_sents: " + str(split_sents) ann_seqs = xmltoseq.ann_to_seq(narr, ann, split_sents) print "seqs: " + str(len(ann_seqs)) for s in ann_seqs: seqs.append(s) seq_ids.append(rec_id) return seq_ids, seqs def create_labelencoder(data, num=0): global labelencoder, onehotencoder, num_labels print "create_labelencoder: data[0]: " + str(data[0]) labelencoder = LabelEncoder() labelencoder.fit(data) num_labels = len(labelencoder.classes_) #onehotencoder = OneHotEncoder() #onehotencoder.fit(data2) return labelencoder ''' Encodes labels as one-hot vectors (entire dataset: 2D array) data: a 1D array of labels num_labels: the number of label classes ''' def encode_labels(data, labenc=None, max_len=50): if labenc == None: labenc = labelencoder if labenc == None: # or onehotencoder == None: print "Error: labelencoder must be trained before it can be used!" return None #return onehotencoder.transform(labelencoder.transform(data)) data2 = [] num_labels = len(labenc.classes_) zero_vec = data_util.zero_vec(num_labels) print "data: " + str(len(data)) for item in data: #print "item len: " + str(len(item)) new_item = [] if len(item) > 0: item2 = labenc.transform(item) for lab in item2: onehot = [] for x in range(num_labels): onehot.append(0) onehot[lab] = 1 new_item.append(onehot) # Pad vectors if len(new_item) > max_len: new_item = new_item[0:max_len] while len(new_item) < max_len: new_item.append(zero_vec) data2.append(new_item) #else: # data2.append([]) return data2 ''' Decodes one sequence of labels ''' def decode_labels(data, labenc=None): #print "decode_labels" if labenc is None: labenc = labelencoder data2 = [] for row in data: #print "- row: " + str(row) lab = numpy.argmax(numpy.asarray(row)) #print "- lab: " + str(lab) data2.append(lab) #print "- data2: " + str(data2) return labenc.inverse_transform(data2) #return labelencoder.inverse_transform(onehotencoder.reverse_transform(data)) def decode_all_labels(data, labenc=None): decoded_labels = [] for sequence in data: labs = decode_labels(sequence, labenc) decoded_labels.append(labs) return decoded_labels def word2features(sent, i): word = sent[i][0] #postag = sent[i][1] features = { 'bias': 1.0, 'word.lower()': word.lower(), 'word[-3:]': word[-3:], 'word[-2:]': word[-2:], 'word.isupper()': word.isupper(), 'word.istitle()': word.istitle(),
#'postag[:2]': postag[:2], } if i > 0: word1 = sent[i-1][0] postag1 = sent[i-1][1] features.update({ '-1:word.lower()': word1.lower(), '-1:word.istitle()': word1.istitle(), '-1:word.isupper()': word1.isupper(), #'-1:postag': postag1, #'-1:postag[:2]': postag1[:2], }) else: features['BOS'] = True if i < len(sent)-1: word1 = sent[i+1][0] postag1 = sent[i+1][1] features.update({ '+1:word.lower()': word1.lower(), '+1:word.istitle()': word1.istitle(), '+1:word.isupper()': word1.isupper(), #'+1:postag': postag1, #'+1:postag[:2]': postag1[:2], }) else: features['EOS'] = True return features def sent2features(sent): return [word2features(sent, i) for i in range(len(sent))] def sent2labels(sent): return [label for token, label in sent] def sent2tokens(sent): return [token for token, label in sent] def ispunc(input_string, start, end): punc = ' :;,./?' s = input_string[start:end] for char in s: if char not in punc: return False return True if __name__ == "__main__":main()
'word.isdigit()': word.isdigit(), #'postag': postag,
random_line_split
ChoiceSet.js
/** * ChoiceSet is a weighted random solution that allows the each of the * items that are being chosen to have simply a name or full suite of * values other than their weights and names. * * Once the items have been setup with a name and weight, the system works * by creating a one to one array. Each item in the new array is a sum of * all the subsequent and current weights. The final weight, being the * maximum total weight, is also stored. * * Finally, in order to choose a choice, in a weighted fashion, a random * number between 0 and the maximum is chosen. The new intervals array is * looped through and if the random number appears less than the interval * its index is used to select the actual choice from the set. Voila. * * @author Brielle Harrison <nyteshade@gmail.com> */ class ChoiceSet { /** * Creates a new set of choices. There is no validation in this * constructor, however, the values supplied should be an object * with at least two properties; name and weight. The weight can * be any arbitrary number, floating point or integer, and the * name should be the default String value to chosen. * * If a non String value is desired, an additional value property * or any other property may be added to the choice and retrieved * later via direct reference of the return value from .choose or * via .chooseProp(propName) * * @param {Array<Object>} n-number of Objects as described above */ constructor(...values) { const OBJ = '[object Object]'; const choices = []; const meta = []; values = values.map((v) => { v = ({}).toString.apply(v) === OBJ ? v : {name: v}; v.weight = v.weight || 100; return v; }); this.choices = values || []; if (this.choices.length) { this.calcIntervals(); } } /** * Short hand for <code>new ChoiceSet(...)</code>. This approach * makes reading the code written elsewhere, easier. If the supplied * * @param {Array<Object>} n-number of Objects as described in the * constructor reference. There should be at least a name property * and a value property. */ static of(...values) { let array = values.length === 1 && values[0] instanceof Array && values[0] || values; console.log(array); return new ChoiceSet(...array); } /** * Instantiates a ChoiceSet with numeric choices ranging from * the supplied from value to the supplied to value. All values * are returned with an equal weight by default. * * @param {Number} from a number indicating the start range * @param {Number} to a number indicating the end range, inclusive * @return {ChoiceSet} an instance of ChoiceSet */ static weightedRange(from, to, weights = null) { let set = new ChoiceSet(); for (let i = from; i < (to + 1); i++) { set.choices.push({ name: i, weight: 100 }); } if (weights) { set.setWeights(weights); } else { set.calcIntervals(); } return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedSet(...values) { let set = new ChoiceSet(); if (values.length % 2 != 0) { throw new Error("WeightedChoice must be instantiated with pairs"); } set.choices = []; for (let i = 0; i < values.length; i+=2) { let name = values[i]; let weight = values[i + 1]; set.choices.push({name: name, weight:weight}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} value the value of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedValuedSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedValuedSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let value = values[i + 2]; set.choices.push({name: name, weight:weight, value: value}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names, weights and values. The * function takes parameters in multiples of three. The first parameter is * the name of the choice and the second is the weight. The third object is a * collection of keys and values that will be applied to the choice in * question. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} object to be merged with the resulting choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedObjectSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedObjectSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let object = values[i + 2]; set.choices.push({name: name, weight: weight, _obj: object}); } set.calcIntervals(); return set; } /** * Calculates the intervals of the weights of the choices in the set. It * also determines the maximum total weight in the set. */ calcIntervals() { let intervals = []; this.choices.reduce( function(p, c) { intervals.push( ((p && p.weight) || 0) + ((c && c.weight) || 0) ); }, null ); intervals = intervals.map(function(cur, idx, array) { return cur + array.slice(0,idx).reduce((p, c) => p + c, 0); }); this.intervals = intervals; this.maxInterval = intervals[intervals.length - 1]; } /** * Allows easy adjustment of a weight for a given index. The weight is * modified and then calcIntervals() is called to realign things for * the next choosing. * * NOTE see if this is the optimal setting for adjusting the weights * * @param {Number} index the index of the choice to modify * @param {Number} weight the weight of the choice in general */ setWeightFor(index, weight) { if (this.choices[index]) { this.choices[index].weight = weight || 100; this.calcIntervals(); } } /** * This allows weights to be set in bulk. The code will attempt * to apply a weight for a given choice at equivalent indices. * * @param {Array} arrayOfWeights an array of Numberered weights. */ setWeights(arrayOfWeights) { if (this.choices.length !== arrayOfWeights.length) { console.warn('Array length mismatch; applying what is present'); } for (let i = 0; i < this.choices.length; i++) { let choice = this.choices[i]; let weight = arrayOfWeights[i]; if (!choice || !weight || isNaN(weight)) { continue; } choice.weight = weight; } this.calcIntervals(); } /** * It randomly choose one item from the set. It does so based on a * randomly chosen number within the given weight set. * * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */ chooseOne(prop = 'name') { return this.chooseProp(prop); } /** * Returns an array of results equivalent to those returned by * chooseOne. * * @param {Number} count an integer denoting the number of choices to pick * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */ chooseSome(count, prop = 'name')
/** * Simulates rolling dice with n-number of sides. In pen and paper * role-playing games, 3d6 means to roll three six sided dice together * and sum their results. Calling ChoiceSet.rollDice(3, 6) will simulate * the same effect. * * Optionally, if repeat is set to a number greater than 1, an array of * values is returned with the repeated results numbering the supplied * repeat count. * * @param {Number} times the number of times the die should be rolled * @param {Number} sides the number of sides the die should have * @param {Number} repeat the number of times the whole process should be * repeated * @return {Mixed} either an array of Numbers or a single Number resulting * in the sum of a die with sides rolled times times. */ static rollDice(times, sides, repeat = 1, dropLowest = 0) { let count = []; let die = ChoiceSet.weightedRange(1, sides); for (let i = 0; i < repeat; i++) { let rolls = die.chooseSome(times); rolls.sort(); if (dropLowest) { rolls = rolls.slice(dropLowest); } count.push(rolls.reduce((p,c) => p + c, 0)); } return repeat === 1 ? count[0] : count; } /** * Randomly chooses a value from the set and returns the specified property * from the chosen object. * * @return {String} key the property value specified for the chosen item. * This defaults to 'name'. */ chooseProp(key = 'name') { let choice = this.one; return (choice._obj && choice._obj[key]) || choice[key]; } /** * Chooses a value via .one and then retrieves the value property of * the choice. * * @return {Mixed} the value of the chosen item from the set */ get oneValue() { let choice = this.one; return choice.value || choice._obj || choice.name; } /** * Randomly chooses a value from the set and returns it in its entirety. * * @return {Mixed} an object from the ChoiceSet. */ get one() { let roll = Math.random() * this.maxInterval; let item = null; let index = -1; for (let i = 0; i < this.intervals.length; i++) { if (roll < this.intervals[i]) { index = i; break; } } if (index === -1) { console.log('ERROR'); console.log(roll, index); } return this.choices[index]; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 3 six sided dice and adding their * results, you can achieve a more realistically weighted * random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_3D6 * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 3 six sided dice * and adding their results. */ static get ANY_DICE_WEIGHTS_3D6() { let weights = [ 0.46, 1.39, 2.78, 4.63, 6.94, 9.72, 11.57, 12.50, 12.50, 11.57, 9.72, 6.94, 4.63, 2.78, 1.39, 0.46 ]; return weights; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 4 six sided dice, dropping the lowest * and adding their results, you can achieve a more realistically * weighted random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_4D6_DROP_LOWEST * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 4 six sided dice, * dropping the lowest value, and adding their results. */ static get ANY_DICE_WEIGHTS_4D6_DROP_LOWEST() { let weights = [ 0.08, 0.31, 0.77, 1.62, 2.93, 4.78, 7.02, 9.41, 11.42, 12.89, 13.27, 12.35, 10.11, 7.25, 4.17, 1.62 ]; return weights; } /** * Given an object, make a determination if that object is a wrapper object * in the old style wherein `name` is the object to return and `weight` is * the actual weight of the object for randomness calculation. If so, or * either way, return two separate objects. One with meta, which contains * the weight and any other properties and the object itself to add to the * choices array. * * @param {Object} obj an object to pick apart as above * @return {Object} an object with a `meta` and `choice` properties */ static asMetaAndChoice(obj) { let result = { meta: null, choice: null }; if (isFinite(obj.weight)) { result.choice = obj.value || obj._obj || obj.name; result.meta = obj; delete obj.name; delete obj.value; delete obj._obj; return result; } else { result.choice = obj; result.meta = { weight: 100 } } return result; } } module.exports = ChoiceSet;
{ let choices = []; for (let i = 0; i < count; i++) { choices.push(this.chooseProp(prop)) } return choices; }
identifier_body
ChoiceSet.js
/** * ChoiceSet is a weighted random solution that allows the each of the * items that are being chosen to have simply a name or full suite of * values other than their weights and names. * * Once the items have been setup with a name and weight, the system works * by creating a one to one array. Each item in the new array is a sum of * all the subsequent and current weights. The final weight, being the * maximum total weight, is also stored. * * Finally, in order to choose a choice, in a weighted fashion, a random * number between 0 and the maximum is chosen. The new intervals array is * looped through and if the random number appears less than the interval * its index is used to select the actual choice from the set. Voila. * * @author Brielle Harrison <nyteshade@gmail.com> */ class ChoiceSet { /** * Creates a new set of choices. There is no validation in this * constructor, however, the values supplied should be an object * with at least two properties; name and weight. The weight can * be any arbitrary number, floating point or integer, and the * name should be the default String value to chosen. * * If a non String value is desired, an additional value property * or any other property may be added to the choice and retrieved * later via direct reference of the return value from .choose or * via .chooseProp(propName) * * @param {Array<Object>} n-number of Objects as described above */ constructor(...values) { const OBJ = '[object Object]'; const choices = []; const meta = []; values = values.map((v) => { v = ({}).toString.apply(v) === OBJ ? v : {name: v}; v.weight = v.weight || 100; return v; }); this.choices = values || []; if (this.choices.length) { this.calcIntervals(); } } /** * Short hand for <code>new ChoiceSet(...)</code>. This approach * makes reading the code written elsewhere, easier. If the supplied * * @param {Array<Object>} n-number of Objects as described in the * constructor reference. There should be at least a name property * and a value property. */ static of(...values) { let array = values.length === 1 && values[0] instanceof Array && values[0] || values; console.log(array); return new ChoiceSet(...array); } /** * Instantiates a ChoiceSet with numeric choices ranging from * the supplied from value to the supplied to value. All values * are returned with an equal weight by default. * * @param {Number} from a number indicating the start range * @param {Number} to a number indicating the end range, inclusive * @return {ChoiceSet} an instance of ChoiceSet */ static weightedRange(from, to, weights = null) { let set = new ChoiceSet(); for (let i = from; i < (to + 1); i++) { set.choices.push({ name: i, weight: 100 }); } if (weights) { set.setWeights(weights); } else { set.calcIntervals(); } return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedSet(...values) { let set = new ChoiceSet(); if (values.length % 2 != 0) { throw new Error("WeightedChoice must be instantiated with pairs"); } set.choices = []; for (let i = 0; i < values.length; i+=2) { let name = values[i]; let weight = values[i + 1]; set.choices.push({name: name, weight:weight}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} value the value of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedValuedSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedValuedSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let value = values[i + 2]; set.choices.push({name: name, weight:weight, value: value}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names, weights and values. The * function takes parameters in multiples of three. The first parameter is * the name of the choice and the second is the weight. The third object is a * collection of keys and values that will be applied to the choice in * question. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} object to be merged with the resulting choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedObjectSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedObjectSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let object = values[i + 2]; set.choices.push({name: name, weight: weight, _obj: object}); } set.calcIntervals(); return set; } /** * Calculates the intervals of the weights of the choices in the set. It * also determines the maximum total weight in the set. */ calcIntervals() { let intervals = []; this.choices.reduce( function(p, c) { intervals.push( ((p && p.weight) || 0) + ((c && c.weight) || 0) ); }, null ); intervals = intervals.map(function(cur, idx, array) {
} /** * Allows easy adjustment of a weight for a given index. The weight is * modified and then calcIntervals() is called to realign things for * the next choosing. * * NOTE see if this is the optimal setting for adjusting the weights * * @param {Number} index the index of the choice to modify * @param {Number} weight the weight of the choice in general */ setWeightFor(index, weight) { if (this.choices[index]) { this.choices[index].weight = weight || 100; this.calcIntervals(); } } /** * This allows weights to be set in bulk. The code will attempt * to apply a weight for a given choice at equivalent indices. * * @param {Array} arrayOfWeights an array of Numberered weights. */ setWeights(arrayOfWeights) { if (this.choices.length !== arrayOfWeights.length) { console.warn('Array length mismatch; applying what is present'); } for (let i = 0; i < this.choices.length; i++) { let choice = this.choices[i]; let weight = arrayOfWeights[i]; if (!choice || !weight || isNaN(weight)) { continue; } choice.weight = weight; } this.calcIntervals(); } /** * It randomly choose one item from the set. It does so based on a * randomly chosen number within the given weight set. * * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */ chooseOne(prop = 'name') { return this.chooseProp(prop); } /** * Returns an array of results equivalent to those returned by * chooseOne. * * @param {Number} count an integer denoting the number of choices to pick * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */ chooseSome(count, prop = 'name') { let choices = []; for (let i = 0; i < count; i++) { choices.push(this.chooseProp(prop)) } return choices; } /** * Simulates rolling dice with n-number of sides. In pen and paper * role-playing games, 3d6 means to roll three six sided dice together * and sum their results. Calling ChoiceSet.rollDice(3, 6) will simulate * the same effect. * * Optionally, if repeat is set to a number greater than 1, an array of * values is returned with the repeated results numbering the supplied * repeat count. * * @param {Number} times the number of times the die should be rolled * @param {Number} sides the number of sides the die should have * @param {Number} repeat the number of times the whole process should be * repeated * @return {Mixed} either an array of Numbers or a single Number resulting * in the sum of a die with sides rolled times times. */ static rollDice(times, sides, repeat = 1, dropLowest = 0) { let count = []; let die = ChoiceSet.weightedRange(1, sides); for (let i = 0; i < repeat; i++) { let rolls = die.chooseSome(times); rolls.sort(); if (dropLowest) { rolls = rolls.slice(dropLowest); } count.push(rolls.reduce((p,c) => p + c, 0)); } return repeat === 1 ? count[0] : count; } /** * Randomly chooses a value from the set and returns the specified property * from the chosen object. * * @return {String} key the property value specified for the chosen item. * This defaults to 'name'. */ chooseProp(key = 'name') { let choice = this.one; return (choice._obj && choice._obj[key]) || choice[key]; } /** * Chooses a value via .one and then retrieves the value property of * the choice. * * @return {Mixed} the value of the chosen item from the set */ get oneValue() { let choice = this.one; return choice.value || choice._obj || choice.name; } /** * Randomly chooses a value from the set and returns it in its entirety. * * @return {Mixed} an object from the ChoiceSet. */ get one() { let roll = Math.random() * this.maxInterval; let item = null; let index = -1; for (let i = 0; i < this.intervals.length; i++) { if (roll < this.intervals[i]) { index = i; break; } } if (index === -1) { console.log('ERROR'); console.log(roll, index); } return this.choices[index]; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 3 six sided dice and adding their * results, you can achieve a more realistically weighted * random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_3D6 * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 3 six sided dice * and adding their results. */ static get ANY_DICE_WEIGHTS_3D6() { let weights = [ 0.46, 1.39, 2.78, 4.63, 6.94, 9.72, 11.57, 12.50, 12.50, 11.57, 9.72, 6.94, 4.63, 2.78, 1.39, 0.46 ]; return weights; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 4 six sided dice, dropping the lowest * and adding their results, you can achieve a more realistically * weighted random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_4D6_DROP_LOWEST * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 4 six sided dice, * dropping the lowest value, and adding their results. */ static get ANY_DICE_WEIGHTS_4D6_DROP_LOWEST() { let weights = [ 0.08, 0.31, 0.77, 1.62, 2.93, 4.78, 7.02, 9.41, 11.42, 12.89, 13.27, 12.35, 10.11, 7.25, 4.17, 1.62 ]; return weights; } /** * Given an object, make a determination if that object is a wrapper object * in the old style wherein `name` is the object to return and `weight` is * the actual weight of the object for randomness calculation. If so, or * either way, return two separate objects. One with meta, which contains * the weight and any other properties and the object itself to add to the * choices array. * * @param {Object} obj an object to pick apart as above * @return {Object} an object with a `meta` and `choice` properties */ static asMetaAndChoice(obj) { let result = { meta: null, choice: null }; if (isFinite(obj.weight)) { result.choice = obj.value || obj._obj || obj.name; result.meta = obj; delete obj.name; delete obj.value; delete obj._obj; return result; } else { result.choice = obj; result.meta = { weight: 100 } } return result; } } module.exports = ChoiceSet;
return cur + array.slice(0,idx).reduce((p, c) => p + c, 0); }); this.intervals = intervals; this.maxInterval = intervals[intervals.length - 1];
random_line_split
ChoiceSet.js
/** * ChoiceSet is a weighted random solution that allows the each of the * items that are being chosen to have simply a name or full suite of * values other than their weights and names. * * Once the items have been setup with a name and weight, the system works * by creating a one to one array. Each item in the new array is a sum of * all the subsequent and current weights. The final weight, being the * maximum total weight, is also stored. * * Finally, in order to choose a choice, in a weighted fashion, a random * number between 0 and the maximum is chosen. The new intervals array is * looped through and if the random number appears less than the interval * its index is used to select the actual choice from the set. Voila. * * @author Brielle Harrison <nyteshade@gmail.com> */ class ChoiceSet { /** * Creates a new set of choices. There is no validation in this * constructor, however, the values supplied should be an object * with at least two properties; name and weight. The weight can * be any arbitrary number, floating point or integer, and the * name should be the default String value to chosen. * * If a non String value is desired, an additional value property * or any other property may be added to the choice and retrieved * later via direct reference of the return value from .choose or * via .chooseProp(propName) * * @param {Array<Object>} n-number of Objects as described above */ constructor(...values) { const OBJ = '[object Object]'; const choices = []; const meta = []; values = values.map((v) => { v = ({}).toString.apply(v) === OBJ ? v : {name: v}; v.weight = v.weight || 100; return v; }); this.choices = values || []; if (this.choices.length) { this.calcIntervals(); } } /** * Short hand for <code>new ChoiceSet(...)</code>. This approach * makes reading the code written elsewhere, easier. If the supplied * * @param {Array<Object>} n-number of Objects as described in the * constructor reference. There should be at least a name property * and a value property. */ static of(...values) { let array = values.length === 1 && values[0] instanceof Array && values[0] || values; console.log(array); return new ChoiceSet(...array); } /** * Instantiates a ChoiceSet with numeric choices ranging from * the supplied from value to the supplied to value. All values * are returned with an equal weight by default. * * @param {Number} from a number indicating the start range * @param {Number} to a number indicating the end range, inclusive * @return {ChoiceSet} an instance of ChoiceSet */ static weightedRange(from, to, weights = null) { let set = new ChoiceSet(); for (let i = from; i < (to + 1); i++) { set.choices.push({ name: i, weight: 100 }); } if (weights) { set.setWeights(weights); } else { set.calcIntervals(); } return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedSet(...values) { let set = new ChoiceSet(); if (values.length % 2 != 0) { throw new Error("WeightedChoice must be instantiated with pairs"); } set.choices = []; for (let i = 0; i < values.length; i+=2) { let name = values[i]; let weight = values[i + 1]; set.choices.push({name: name, weight:weight}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} value the value of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedValuedSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedValuedSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let value = values[i + 2]; set.choices.push({name: name, weight:weight, value: value}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names, weights and values. The * function takes parameters in multiples of three. The first parameter is * the name of the choice and the second is the weight. The third object is a * collection of keys and values that will be applied to the choice in * question. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} object to be merged with the resulting choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedObjectSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedObjectSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let object = values[i + 2]; set.choices.push({name: name, weight: weight, _obj: object}); } set.calcIntervals(); return set; } /** * Calculates the intervals of the weights of the choices in the set. It * also determines the maximum total weight in the set. */ calcIntervals() { let intervals = []; this.choices.reduce( function(p, c) { intervals.push( ((p && p.weight) || 0) + ((c && c.weight) || 0) ); }, null ); intervals = intervals.map(function(cur, idx, array) { return cur + array.slice(0,idx).reduce((p, c) => p + c, 0); }); this.intervals = intervals; this.maxInterval = intervals[intervals.length - 1]; } /** * Allows easy adjustment of a weight for a given index. The weight is * modified and then calcIntervals() is called to realign things for * the next choosing. * * NOTE see if this is the optimal setting for adjusting the weights * * @param {Number} index the index of the choice to modify * @param {Number} weight the weight of the choice in general */ setWeightFor(index, weight) { if (this.choices[index]) { this.choices[index].weight = weight || 100; this.calcIntervals(); } } /** * This allows weights to be set in bulk. The code will attempt * to apply a weight for a given choice at equivalent indices. * * @param {Array} arrayOfWeights an array of Numberered weights. */ setWeights(arrayOfWeights) { if (this.choices.length !== arrayOfWeights.length) { console.warn('Array length mismatch; applying what is present'); } for (let i = 0; i < this.choices.length; i++) { let choice = this.choices[i]; let weight = arrayOfWeights[i]; if (!choice || !weight || isNaN(weight)) { continue; } choice.weight = weight; } this.calcIntervals(); } /** * It randomly choose one item from the set. It does so based on a * randomly chosen number within the given weight set. * * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */ chooseOne(prop = 'name') { return this.chooseProp(prop); } /** * Returns an array of results equivalent to those returned by * chooseOne. * * @param {Number} count an integer denoting the number of choices to pick * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */ chooseSome(count, prop = 'name') { let choices = []; for (let i = 0; i < count; i++) { choices.push(this.chooseProp(prop)) } return choices; } /** * Simulates rolling dice with n-number of sides. In pen and paper * role-playing games, 3d6 means to roll three six sided dice together * and sum their results. Calling ChoiceSet.rollDice(3, 6) will simulate * the same effect. * * Optionally, if repeat is set to a number greater than 1, an array of * values is returned with the repeated results numbering the supplied * repeat count. * * @param {Number} times the number of times the die should be rolled * @param {Number} sides the number of sides the die should have * @param {Number} repeat the number of times the whole process should be * repeated * @return {Mixed} either an array of Numbers or a single Number resulting * in the sum of a die with sides rolled times times. */ static rollDice(times, sides, repeat = 1, dropLowest = 0) { let count = []; let die = ChoiceSet.weightedRange(1, sides); for (let i = 0; i < repeat; i++) { let rolls = die.chooseSome(times); rolls.sort(); if (dropLowest) { rolls = rolls.slice(dropLowest); } count.push(rolls.reduce((p,c) => p + c, 0)); } return repeat === 1 ? count[0] : count; } /** * Randomly chooses a value from the set and returns the specified property * from the chosen object. * * @return {String} key the property value specified for the chosen item. * This defaults to 'name'. */ chooseProp(key = 'name') { let choice = this.one; return (choice._obj && choice._obj[key]) || choice[key]; } /** * Chooses a value via .one and then retrieves the value property of * the choice. * * @return {Mixed} the value of the chosen item from the set */ get oneValue() { let choice = this.one; return choice.value || choice._obj || choice.name; } /** * Randomly chooses a value from the set and returns it in its entirety. * * @return {Mixed} an object from the ChoiceSet. */ get one() { let roll = Math.random() * this.maxInterval; let item = null; let index = -1; for (let i = 0; i < this.intervals.length; i++) { if (roll < this.intervals[i])
} if (index === -1) { console.log('ERROR'); console.log(roll, index); } return this.choices[index]; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 3 six sided dice and adding their * results, you can achieve a more realistically weighted * random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_3D6 * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 3 six sided dice * and adding their results. */ static get ANY_DICE_WEIGHTS_3D6() { let weights = [ 0.46, 1.39, 2.78, 4.63, 6.94, 9.72, 11.57, 12.50, 12.50, 11.57, 9.72, 6.94, 4.63, 2.78, 1.39, 0.46 ]; return weights; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 4 six sided dice, dropping the lowest * and adding their results, you can achieve a more realistically * weighted random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_4D6_DROP_LOWEST * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 4 six sided dice, * dropping the lowest value, and adding their results. */ static get ANY_DICE_WEIGHTS_4D6_DROP_LOWEST() { let weights = [ 0.08, 0.31, 0.77, 1.62, 2.93, 4.78, 7.02, 9.41, 11.42, 12.89, 13.27, 12.35, 10.11, 7.25, 4.17, 1.62 ]; return weights; } /** * Given an object, make a determination if that object is a wrapper object * in the old style wherein `name` is the object to return and `weight` is * the actual weight of the object for randomness calculation. If so, or * either way, return two separate objects. One with meta, which contains * the weight and any other properties and the object itself to add to the * choices array. * * @param {Object} obj an object to pick apart as above * @return {Object} an object with a `meta` and `choice` properties */ static asMetaAndChoice(obj) { let result = { meta: null, choice: null }; if (isFinite(obj.weight)) { result.choice = obj.value || obj._obj || obj.name; result.meta = obj; delete obj.name; delete obj.value; delete obj._obj; return result; } else { result.choice = obj; result.meta = { weight: 100 } } return result; } } module.exports = ChoiceSet;
{ index = i; break; }
conditional_block
ChoiceSet.js
/** * ChoiceSet is a weighted random solution that allows the each of the * items that are being chosen to have simply a name or full suite of * values other than their weights and names. * * Once the items have been setup with a name and weight, the system works * by creating a one to one array. Each item in the new array is a sum of * all the subsequent and current weights. The final weight, being the * maximum total weight, is also stored. * * Finally, in order to choose a choice, in a weighted fashion, a random * number between 0 and the maximum is chosen. The new intervals array is * looped through and if the random number appears less than the interval * its index is used to select the actual choice from the set. Voila. * * @author Brielle Harrison <nyteshade@gmail.com> */ class ChoiceSet { /** * Creates a new set of choices. There is no validation in this * constructor, however, the values supplied should be an object * with at least two properties; name and weight. The weight can * be any arbitrary number, floating point or integer, and the * name should be the default String value to chosen. * * If a non String value is desired, an additional value property * or any other property may be added to the choice and retrieved * later via direct reference of the return value from .choose or * via .chooseProp(propName) * * @param {Array<Object>} n-number of Objects as described above */ constructor(...values) { const OBJ = '[object Object]'; const choices = []; const meta = []; values = values.map((v) => { v = ({}).toString.apply(v) === OBJ ? v : {name: v}; v.weight = v.weight || 100; return v; }); this.choices = values || []; if (this.choices.length) { this.calcIntervals(); } } /** * Short hand for <code>new ChoiceSet(...)</code>. This approach * makes reading the code written elsewhere, easier. If the supplied * * @param {Array<Object>} n-number of Objects as described in the * constructor reference. There should be at least a name property * and a value property. */ static of(...values) { let array = values.length === 1 && values[0] instanceof Array && values[0] || values; console.log(array); return new ChoiceSet(...array); } /** * Instantiates a ChoiceSet with numeric choices ranging from * the supplied from value to the supplied to value. All values * are returned with an equal weight by default. * * @param {Number} from a number indicating the start range * @param {Number} to a number indicating the end range, inclusive * @return {ChoiceSet} an instance of ChoiceSet */ static weightedRange(from, to, weights = null) { let set = new ChoiceSet(); for (let i = from; i < (to + 1); i++) { set.choices.push({ name: i, weight: 100 }); } if (weights) { set.setWeights(weights); } else { set.calcIntervals(); } return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedSet(...values) { let set = new ChoiceSet(); if (values.length % 2 != 0) { throw new Error("WeightedChoice must be instantiated with pairs"); } set.choices = []; for (let i = 0; i < values.length; i+=2) { let name = values[i]; let weight = values[i + 1]; set.choices.push({name: name, weight:weight}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names and values. The * function takes an even number of parameters with the first being * the name and the second being the weight. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} value the value of the choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedValuedSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedValuedSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let value = values[i + 2]; set.choices.push({name: name, weight:weight, value: value}); } set.calcIntervals(); return set; } /** * An easy way to instantiate a ChoiceSet of names, weights and values. The * function takes parameters in multiples of three. The first parameter is * the name of the choice and the second is the weight. The third object is a * collection of keys and values that will be applied to the choice in * question. * * @param {String} name the name of the choice * @param {Number} weight the weight of the choice * @param {Object} object to be merged with the resulting choice * @return {ChoiceSet} an instance of ChoiceSet */ static weightedObjectSet(...values) { let set = new ChoiceSet(); if (values.length % 3 != 0) { throw new Error("weightedObjectSet must be instantiated with triplets"); } set.choices = []; for (let i = 0; i < values.length; i+=3) { let name = values[i]; let weight = values[i + 1]; let object = values[i + 2]; set.choices.push({name: name, weight: weight, _obj: object}); } set.calcIntervals(); return set; } /** * Calculates the intervals of the weights of the choices in the set. It * also determines the maximum total weight in the set. */ calcIntervals() { let intervals = []; this.choices.reduce( function(p, c) { intervals.push( ((p && p.weight) || 0) + ((c && c.weight) || 0) ); }, null ); intervals = intervals.map(function(cur, idx, array) { return cur + array.slice(0,idx).reduce((p, c) => p + c, 0); }); this.intervals = intervals; this.maxInterval = intervals[intervals.length - 1]; } /** * Allows easy adjustment of a weight for a given index. The weight is * modified and then calcIntervals() is called to realign things for * the next choosing. * * NOTE see if this is the optimal setting for adjusting the weights * * @param {Number} index the index of the choice to modify * @param {Number} weight the weight of the choice in general */ setWeightFor(index, weight) { if (this.choices[index]) { this.choices[index].weight = weight || 100; this.calcIntervals(); } } /** * This allows weights to be set in bulk. The code will attempt * to apply a weight for a given choice at equivalent indices. * * @param {Array} arrayOfWeights an array of Numberered weights. */ setWeights(arrayOfWeights) { if (this.choices.length !== arrayOfWeights.length) { console.warn('Array length mismatch; applying what is present'); } for (let i = 0; i < this.choices.length; i++) { let choice = this.choices[i]; let weight = arrayOfWeights[i]; if (!choice || !weight || isNaN(weight)) { continue; } choice.weight = weight; } this.calcIntervals(); } /** * It randomly choose one item from the set. It does so based on a * randomly chosen number within the given weight set. * * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */
(prop = 'name') { return this.chooseProp(prop); } /** * Returns an array of results equivalent to those returned by * chooseOne. * * @param {Number} count an integer denoting the number of choices to pick * @param {String} prop the property of the randomly chosen item * @return {Mixed} the property value specified for the chosen item. This * defaults to 'name'. */ chooseSome(count, prop = 'name') { let choices = []; for (let i = 0; i < count; i++) { choices.push(this.chooseProp(prop)) } return choices; } /** * Simulates rolling dice with n-number of sides. In pen and paper * role-playing games, 3d6 means to roll three six sided dice together * and sum their results. Calling ChoiceSet.rollDice(3, 6) will simulate * the same effect. * * Optionally, if repeat is set to a number greater than 1, an array of * values is returned with the repeated results numbering the supplied * repeat count. * * @param {Number} times the number of times the die should be rolled * @param {Number} sides the number of sides the die should have * @param {Number} repeat the number of times the whole process should be * repeated * @return {Mixed} either an array of Numbers or a single Number resulting * in the sum of a die with sides rolled times times. */ static rollDice(times, sides, repeat = 1, dropLowest = 0) { let count = []; let die = ChoiceSet.weightedRange(1, sides); for (let i = 0; i < repeat; i++) { let rolls = die.chooseSome(times); rolls.sort(); if (dropLowest) { rolls = rolls.slice(dropLowest); } count.push(rolls.reduce((p,c) => p + c, 0)); } return repeat === 1 ? count[0] : count; } /** * Randomly chooses a value from the set and returns the specified property * from the chosen object. * * @return {String} key the property value specified for the chosen item. * This defaults to 'name'. */ chooseProp(key = 'name') { let choice = this.one; return (choice._obj && choice._obj[key]) || choice[key]; } /** * Chooses a value via .one and then retrieves the value property of * the choice. * * @return {Mixed} the value of the chosen item from the set */ get oneValue() { let choice = this.one; return choice.value || choice._obj || choice.name; } /** * Randomly chooses a value from the set and returns it in its entirety. * * @return {Mixed} an object from the ChoiceSet. */ get one() { let roll = Math.random() * this.maxInterval; let item = null; let index = -1; for (let i = 0; i < this.intervals.length; i++) { if (roll < this.intervals[i]) { index = i; break; } } if (index === -1) { console.log('ERROR'); console.log(roll, index); } return this.choices[index]; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 3 six sided dice and adding their * results, you can achieve a more realistically weighted * random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_3D6 * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 3 six sided dice * and adding their results. */ static get ANY_DICE_WEIGHTS_3D6() { let weights = [ 0.46, 1.39, 2.78, 4.63, 6.94, 9.72, 11.57, 12.50, 12.50, 11.57, 9.72, 6.94, 4.63, 2.78, 1.39, 0.46 ]; return weights; } /** * Using the probability counts on http://www.anydice.com in * conjunction with rolling 4 six sided dice, dropping the lowest * and adding their results, you can achieve a more realistically * weighted random for your use. Simply use the following: * * <code> * let dice = ChoiceSet.weightedRange( * 3, 18, * ChoiceSet.ANY_DICE_WEIGHTS_4D6_DROP_LOWEST * ); * let roll = dice.chooseSome(6); * </code> * * @return {Array} an arry of 15 values indicating probability * weights for values 3 to 18, simulating rolling 4 six sided dice, * dropping the lowest value, and adding their results. */ static get ANY_DICE_WEIGHTS_4D6_DROP_LOWEST() { let weights = [ 0.08, 0.31, 0.77, 1.62, 2.93, 4.78, 7.02, 9.41, 11.42, 12.89, 13.27, 12.35, 10.11, 7.25, 4.17, 1.62 ]; return weights; } /** * Given an object, make a determination if that object is a wrapper object * in the old style wherein `name` is the object to return and `weight` is * the actual weight of the object for randomness calculation. If so, or * either way, return two separate objects. One with meta, which contains * the weight and any other properties and the object itself to add to the * choices array. * * @param {Object} obj an object to pick apart as above * @return {Object} an object with a `meta` and `choice` properties */ static asMetaAndChoice(obj) { let result = { meta: null, choice: null }; if (isFinite(obj.weight)) { result.choice = obj.value || obj._obj || obj.name; result.meta = obj; delete obj.name; delete obj.value; delete obj._obj; return result; } else { result.choice = obj; result.meta = { weight: 100 } } return result; } } module.exports = ChoiceSet;
chooseOne
identifier_name
canvas.go
// Copyright 2012 - 2018 The ASCIIToSVG Contributors // All rights reserved. package asciitosvg import ( "bytes" "encoding/json" "fmt" "image" "regexp" "sort" "strconv" "unicode/utf8" ) // Canvas provides methods for returning objects from an underlying textual grid. type Canvas interface { // A canvas has an underlying visual representation. The fmt.Stringer interface for this // interface provides a view into the underlying grid. fmt.Stringer // Objects returns all the objects found in the underlying grid. Objects() []Object // Size returns the visual dimensions of the Canvas. Size() image.Point // Options returns a map of options to apply to Objects based on the object's tag. This // maps tag name to a map of option names to options. Options() map[string]map[string]interface{} // EnclosingObjects returns the set of objects that contain this point in order from most // to least specific. EnclosingObjects(p Point) []Object } // NewCanvas returns a new Canvas, initialized from the provided data. If tabWidth is set to a non-negative // value, that value will be used to convert tabs to spaces within the grid. Creation of the Canvas // can fail if the diagram contains invalid UTF-8 sequences. func NewCanvas(data []byte, tabWidth int, noBlur bool) (Canvas, error) { c := &canvas{ options: map[string]map[string]interface{}{ "__a2s__closed__options__": map[string]interface{}{ "fill": "#fff", "filter": "url(#dsFilter)", }, }, } if noBlur { c.options["__a2s__closed__options__"] = map[string]interface{}{ "fill": "#fff", } } lines := bytes.Split(data, []byte("\n")) c.size.Y = len(lines) // Diagrams will often not be padded to a uniform width. To overcome this, we scan over // each line and figure out which is the longest. This becomes the width of the canvas. for i, line := range lines { if ok := utf8.Valid(line); !ok { return nil, fmt.Errorf("invalid UTF-8 encoding on line %d", i) } l, err := expandTabs(line, tabWidth) if err != nil { return nil, err } lines[i] = l if i := utf8.RuneCount(lines[i]); i > c.size.X { c.size.X = i } } c.grid = make([]char, c.size.X*c.size.Y) c.visited = make([]bool, c.size.X*c.size.Y) for y, line := range lines { x := 0 for len(line) > 0 { r, l := utf8.DecodeRune(line) c.grid[y*c.size.X+x] = char(r) x++ line = line[l:] } for ; x < c.size.X; x++ { c.grid[y*c.size.X+x] = ' ' } } c.findObjects() return c, nil } // The expandTabs function pads tab characters to the specified width of spaces for the provided // line of input. We cannot simply pad based on byte-offset since our input is UTF-8 encoded. // Fortunately, we can assume that this function is called that the line contains only valid // UTF-8 sequences. We first decode the line rune-wise, and use individual runes to figure out // where we are within the line. When we encounter a tab character, we expand based on our rune // index. func expandTabs(line []byte, tabWidth int) ([]byte, error) { // Initial sizing of our output slice assumes no UTF-8 bytes or tabs, since this is often // the common case. out := make([]byte, 0, len(line)) // pos tracks our position in the input byte slice, while index tracks our position in the // resulting output slice. pos := 0 index := 0 for _, c := range line { if c == '\t' { // Loop over the remaining space count for this particular tabstop until // the next, replacing each position with a space. for s := tabWidth - (pos % tabWidth); s > 0; s-- { out = append(out, ' ') index++ } pos++ } else { // We need to know the byte length of the rune at this position so that we // can account for our tab expansion properly. So we first decode the rune // at this position to get its length in bytes, plop that rune back into our // output slice, and account accordingly. r, l := utf8.DecodeRune(line[pos:]) if r == utf8.RuneError { return nil, fmt.Errorf("invalid rune at byte offset %d; rune offset %d", pos, index) } enc := make([]byte, l) utf8.EncodeRune(enc, r) out = append(out, enc...) pos += l index++ } } return out, nil } // canvas is the parsed source data. type canvas struct { // (0,0) is top left. grid []char visited []bool objects objects size image.Point options map[string]map[string]interface{} } func (c *canvas) String() string { return fmt.Sprintf("%+v", c.grid) } func (c *canvas) Objects() []Object { return c.objects } func (c *canvas) Size() image.Point { return c.size } func (c *canvas) Options() map[string]map[string]interface{} { return c.options } func (c *canvas) EnclosingObjects(p Point) []Object
// findObjects finds all objects (lines, polygons, and text) within the underlying grid. func (c *canvas) findObjects() { p := Point{} // Find any new paths by starting with a point that wasn't yet visited, beginning at the top // left of the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isPathStart() { // Found the start of a one or multiple connected paths. Traverse all // connecting points. This will generate multiple objects if multiple // paths (either open or closed) are found. c.visit(p) objs := c.scanPath([]Point{p}) for _, obj := range objs { // For all points in all objects found, mark the points as visited. for _, p := range obj.Points() { c.visit(p) } } c.objects = append(c.objects, objs...) } } } // A second pass through the grid attempts to identify any text within the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isTextStart() { obj := c.scanText(p) // scanText will return nil if the text at this area is simply // setting options on a container object. if obj == nil { continue } for _, p := range obj.Points() { c.visit(p) } c.objects = append(c.objects, obj) } } } sort.Sort(c.objects) } // scanPath tries to complete a total path (for lines or polygons) starting with some partial path. // It recurses when it finds multiple unvisited outgoing paths. func (c *canvas) scanPath(points []Point) objects { cur := points[len(points)-1] next := c.next(cur) // If there are no points that can progress traversal of the path, finalize the one we're // working on, and return it. This is the terminal condition in the passive flow. if len(next) == 0 { if len(points) == 1 { // Discard 'path' of 1 point. Do not mark point as visited. c.unvisit(cur) return nil } // TODO(dhobsd): Determine if path is sharing the line with another path. If so, // we may want to join the objects such that we don't get weird rendering artifacts. o := &object{points: points} o.seal(c) return objects{o} } // If we have hit a point that can create a closed path, create an object and close // the path. Additionally, recurse to other progress directions in case e.g. an open // path spawns from this point. Paths are always closed vertically. if cur.X == points[0].X && cur.Y == points[0].Y+1 { o := &object{points: points} o.seal(c) r := objects{o} return append(r, c.scanPath([]Point{cur})...) } // We scan depth-first instead of breadth-first, making it possible to find a // closed path. var objs objects for _, n := range next { if c.isVisited(n) { continue } c.visit(n) p2 := make([]Point, len(points)+1) copy(p2, points) p2[len(p2)-1] = n objs = append(objs, c.scanPath(p2)...) } return objs } // The next returns the points that can be used to make progress, scanning (in order) horizontal // progress to the left or right, vertical progress above or below, or diagonal progress to NW, // NE, SW, and SE. It skips any points already visited, and returns all of the possible progress // points. func (c *canvas) next(pos Point) []Point { // Our caller must have called c.visit prior to calling this function. if !c.isVisited(pos) { panic(fmt.Errorf("internal error; revisiting %s", pos)) } var out []Point ch := c.at(pos) if ch.canHorizontal() { nextHorizontal := func(p Point) { if !c.isVisited(p) && c.at(p).canHorizontal() { out = append(out, p) } } if c.canLeft(pos) { n := pos n.X-- nextHorizontal(n) } if c.canRight(pos) { n := pos n.X++ nextHorizontal(n) } } if ch.canVertical() { nextVertical := func(p Point) { if !c.isVisited(p) && c.at(p).canVertical() { out = append(out, p) } } if c.canUp(pos) { n := pos n.Y-- nextVertical(n) } if c.canDown(pos) { n := pos n.Y++ nextVertical(n) } } if c.canDiagonal(pos) { nextDiagonal := func(from, to Point) { if !c.isVisited(to) && c.at(to).canDiagonalFrom(c.at(from)) { out = append(out, to) } } if c.canUp(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y-- nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y-- nextDiagonal(pos, n) } } if c.canDown(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y++ nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y++ nextDiagonal(pos, n) } } } return out } // Used for matching [X, Y]: {...} tag definitions. These definitions target specific objects. var objTagRE = regexp.MustCompile(`(\d+)\s*,\s*(\d+)$`) // scanText extracts a line of text. func (c *canvas) scanText(start Point) Object { obj := &object{points: []Point{start}, isText: true} whiteSpaceStreak := 0 cur := start tagged := 0 tag := []rune{} tagDef := []rune{} for c.canRight(cur) { if cur.X == start.X && c.at(cur).isObjectStartTag() { tagged++ } else if cur.X > start.X && c.at(cur).isObjectEndTag() { tagged++ } cur.X++ if c.isVisited(cur) { // If the point is already visited, we hit a polygon or a line. break } ch := c.at(cur) if !ch.isTextCont() { break } if tagged == 0 && ch.isSpace() { whiteSpaceStreak++ // Stop when we see 3 consecutive whitespace points. if whiteSpaceStreak > 2 { break } } else { whiteSpaceStreak = 0 } switch tagged { case 1: if !c.at(cur).isObjectEndTag() { tag = append(tag, rune(ch)) } case 2: if c.at(cur).isTagDefinitionSeparator() { tagged++ } else { tagged = -1 } case 3: tagDef = append(tagDef, rune(ch)) } obj.points = append(obj.points, cur) } // If we found a start and end tag marker, we either need to assign the tag to the object, // or we need to assign the specified options to the global canvas option space. if tagged == 2 { t := string(tag) if container := c.EnclosingObjects(start); container != nil { container[0].SetTag(t) } // The tag applies to the text object as well so that properties like // a2s:label can be set. obj.SetTag(t) } else if tagged == 3 { t := string(tag) // A tag definition targeting an object will not be found within any object; we need // to do that calculation here. if matches := objTagRE.FindStringSubmatch(t); matches != nil { if targetX, err := strconv.ParseInt(matches[1], 10, 0); err == nil { if targetY, err := strconv.ParseInt(matches[2], 10, 0); err == nil { for i, o := range c.objects { corner := o.Corners()[0] if corner.X == int(targetX) && corner.Y == int(targetY) { c.objects[i].SetTag(t) break } } } } } // This is a tag definition. Parse the JSON and assign the options to the canvas. var m interface{} def := []byte(string(tagDef)) if err := json.Unmarshal(def, &m); err != nil { // TODO(dhobsd): Gross. panic(err) } // The tag applies to the reference object as well, so that properties like // a2s:delref can be set. obj.SetTag(t) c.options[t] = m.(map[string]interface{}) } // Trim the right side of the text object. for len(obj.points) != 0 && c.at(obj.points[len(obj.points)-1]).isSpace() { obj.points = obj.points[:len(obj.points)-1] } obj.seal(c) return obj } func (c *canvas) at(p Point) char { return c.grid[p.Y*c.size.X+p.X] } func (c *canvas) isVisited(p Point) bool { return c.visited[p.Y*c.size.X+p.X] } func (c *canvas) visit(p Point) { // TODO(dhobsd): Change code to ensure that visit() is called once and only // once per point. c.visited[p.Y*c.size.X+p.X] = true } func (c *canvas) unvisit(p Point) { o := p.Y*c.size.X + p.X if !c.visited[o] { panic(fmt.Errorf("internal error: point %+v never visited", p)) } c.visited[o] = false } func (c *canvas) canLeft(p Point) bool { return p.X > 0 } func (c *canvas) canRight(p Point) bool { return p.X < c.size.X-1 } func (c *canvas) canUp(p Point) bool { return p.Y > 0 } func (c *canvas) canDown(p Point) bool { return p.Y < c.size.Y-1 } func (c *canvas) canDiagonal(p Point) bool { return (c.canLeft(p) || c.canRight(p)) && (c.canUp(p) || c.canDown(p)) }
{ maxTL := Point{X: -1, Y: -1} var q []Object for _, o := range c.objects { // An object can't really contain another unless it is a polygon. if !o.IsClosed() { continue } if o.HasPoint(p) && o.Corners()[0].X > maxTL.X && o.Corners()[0].Y > maxTL.Y { q = append(q, o) maxTL.X = o.Corners()[0].X maxTL.Y = o.Corners()[0].Y } } return q }
identifier_body
canvas.go
// Copyright 2012 - 2018 The ASCIIToSVG Contributors // All rights reserved. package asciitosvg import ( "bytes" "encoding/json" "fmt" "image" "regexp" "sort" "strconv" "unicode/utf8" ) // Canvas provides methods for returning objects from an underlying textual grid. type Canvas interface { // A canvas has an underlying visual representation. The fmt.Stringer interface for this // interface provides a view into the underlying grid. fmt.Stringer // Objects returns all the objects found in the underlying grid. Objects() []Object // Size returns the visual dimensions of the Canvas. Size() image.Point // Options returns a map of options to apply to Objects based on the object's tag. This // maps tag name to a map of option names to options. Options() map[string]map[string]interface{} // EnclosingObjects returns the set of objects that contain this point in order from most // to least specific. EnclosingObjects(p Point) []Object } // NewCanvas returns a new Canvas, initialized from the provided data. If tabWidth is set to a non-negative // value, that value will be used to convert tabs to spaces within the grid. Creation of the Canvas // can fail if the diagram contains invalid UTF-8 sequences. func NewCanvas(data []byte, tabWidth int, noBlur bool) (Canvas, error) { c := &canvas{ options: map[string]map[string]interface{}{ "__a2s__closed__options__": map[string]interface{}{ "fill": "#fff", "filter": "url(#dsFilter)", }, }, } if noBlur { c.options["__a2s__closed__options__"] = map[string]interface{}{ "fill": "#fff", } } lines := bytes.Split(data, []byte("\n")) c.size.Y = len(lines) // Diagrams will often not be padded to a uniform width. To overcome this, we scan over // each line and figure out which is the longest. This becomes the width of the canvas. for i, line := range lines { if ok := utf8.Valid(line); !ok { return nil, fmt.Errorf("invalid UTF-8 encoding on line %d", i) } l, err := expandTabs(line, tabWidth) if err != nil { return nil, err } lines[i] = l if i := utf8.RuneCount(lines[i]); i > c.size.X { c.size.X = i } } c.grid = make([]char, c.size.X*c.size.Y) c.visited = make([]bool, c.size.X*c.size.Y) for y, line := range lines { x := 0 for len(line) > 0 { r, l := utf8.DecodeRune(line) c.grid[y*c.size.X+x] = char(r) x++ line = line[l:] } for ; x < c.size.X; x++ { c.grid[y*c.size.X+x] = ' ' } } c.findObjects() return c, nil } // The expandTabs function pads tab characters to the specified width of spaces for the provided // line of input. We cannot simply pad based on byte-offset since our input is UTF-8 encoded. // Fortunately, we can assume that this function is called that the line contains only valid // UTF-8 sequences. We first decode the line rune-wise, and use individual runes to figure out // where we are within the line. When we encounter a tab character, we expand based on our rune // index. func expandTabs(line []byte, tabWidth int) ([]byte, error) { // Initial sizing of our output slice assumes no UTF-8 bytes or tabs, since this is often // the common case. out := make([]byte, 0, len(line)) // pos tracks our position in the input byte slice, while index tracks our position in the // resulting output slice. pos := 0 index := 0 for _, c := range line { if c == '\t' { // Loop over the remaining space count for this particular tabstop until // the next, replacing each position with a space. for s := tabWidth - (pos % tabWidth); s > 0; s-- { out = append(out, ' ') index++ } pos++ } else { // We need to know the byte length of the rune at this position so that we // can account for our tab expansion properly. So we first decode the rune // at this position to get its length in bytes, plop that rune back into our // output slice, and account accordingly. r, l := utf8.DecodeRune(line[pos:]) if r == utf8.RuneError { return nil, fmt.Errorf("invalid rune at byte offset %d; rune offset %d", pos, index) } enc := make([]byte, l) utf8.EncodeRune(enc, r) out = append(out, enc...) pos += l index++ } } return out, nil } // canvas is the parsed source data. type canvas struct { // (0,0) is top left. grid []char visited []bool objects objects size image.Point options map[string]map[string]interface{} } func (c *canvas) String() string { return fmt.Sprintf("%+v", c.grid) } func (c *canvas) Objects() []Object { return c.objects } func (c *canvas) Size() image.Point { return c.size } func (c *canvas) Options() map[string]map[string]interface{} { return c.options } func (c *canvas) EnclosingObjects(p Point) []Object { maxTL := Point{X: -1, Y: -1} var q []Object for _, o := range c.objects { // An object can't really contain another unless it is a polygon. if !o.IsClosed() { continue } if o.HasPoint(p) && o.Corners()[0].X > maxTL.X && o.Corners()[0].Y > maxTL.Y { q = append(q, o) maxTL.X = o.Corners()[0].X maxTL.Y = o.Corners()[0].Y } } return q } // findObjects finds all objects (lines, polygons, and text) within the underlying grid. func (c *canvas) findObjects() { p := Point{} // Find any new paths by starting with a point that wasn't yet visited, beginning at the top // left of the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isPathStart() { // Found the start of a one or multiple connected paths. Traverse all // connecting points. This will generate multiple objects if multiple // paths (either open or closed) are found. c.visit(p) objs := c.scanPath([]Point{p}) for _, obj := range objs { // For all points in all objects found, mark the points as visited. for _, p := range obj.Points() { c.visit(p) } } c.objects = append(c.objects, objs...) } } } // A second pass through the grid attempts to identify any text within the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isTextStart() { obj := c.scanText(p) // scanText will return nil if the text at this area is simply // setting options on a container object. if obj == nil { continue } for _, p := range obj.Points() { c.visit(p) } c.objects = append(c.objects, obj) } } } sort.Sort(c.objects) } // scanPath tries to complete a total path (for lines or polygons) starting with some partial path. // It recurses when it finds multiple unvisited outgoing paths. func (c *canvas) scanPath(points []Point) objects { cur := points[len(points)-1] next := c.next(cur) // If there are no points that can progress traversal of the path, finalize the one we're // working on, and return it. This is the terminal condition in the passive flow. if len(next) == 0 { if len(points) == 1 { // Discard 'path' of 1 point. Do not mark point as visited. c.unvisit(cur) return nil } // TODO(dhobsd): Determine if path is sharing the line with another path. If so, // we may want to join the objects such that we don't get weird rendering artifacts. o := &object{points: points} o.seal(c) return objects{o} } // If we have hit a point that can create a closed path, create an object and close // the path. Additionally, recurse to other progress directions in case e.g. an open // path spawns from this point. Paths are always closed vertically. if cur.X == points[0].X && cur.Y == points[0].Y+1 { o := &object{points: points} o.seal(c) r := objects{o} return append(r, c.scanPath([]Point{cur})...) } // We scan depth-first instead of breadth-first, making it possible to find a // closed path. var objs objects for _, n := range next { if c.isVisited(n) { continue } c.visit(n) p2 := make([]Point, len(points)+1) copy(p2, points) p2[len(p2)-1] = n objs = append(objs, c.scanPath(p2)...) } return objs } // The next returns the points that can be used to make progress, scanning (in order) horizontal // progress to the left or right, vertical progress above or below, or diagonal progress to NW, // NE, SW, and SE. It skips any points already visited, and returns all of the possible progress // points. func (c *canvas) next(pos Point) []Point { // Our caller must have called c.visit prior to calling this function. if !c.isVisited(pos) { panic(fmt.Errorf("internal error; revisiting %s", pos)) } var out []Point ch := c.at(pos) if ch.canHorizontal() { nextHorizontal := func(p Point) { if !c.isVisited(p) && c.at(p).canHorizontal() { out = append(out, p) } } if c.canLeft(pos) { n := pos n.X-- nextHorizontal(n) } if c.canRight(pos) { n := pos n.X++ nextHorizontal(n) } } if ch.canVertical() { nextVertical := func(p Point) { if !c.isVisited(p) && c.at(p).canVertical() { out = append(out, p) } } if c.canUp(pos) { n := pos n.Y-- nextVertical(n) } if c.canDown(pos) { n := pos n.Y++ nextVertical(n) } } if c.canDiagonal(pos) { nextDiagonal := func(from, to Point) { if !c.isVisited(to) && c.at(to).canDiagonalFrom(c.at(from)) { out = append(out, to) } } if c.canUp(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y-- nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y-- nextDiagonal(pos, n) } } if c.canDown(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y++ nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y++ nextDiagonal(pos, n) } } } return out } // Used for matching [X, Y]: {...} tag definitions. These definitions target specific objects. var objTagRE = regexp.MustCompile(`(\d+)\s*,\s*(\d+)$`) // scanText extracts a line of text. func (c *canvas) scanText(start Point) Object { obj := &object{points: []Point{start}, isText: true} whiteSpaceStreak := 0 cur := start tagged := 0 tag := []rune{} tagDef := []rune{} for c.canRight(cur) { if cur.X == start.X && c.at(cur).isObjectStartTag() { tagged++ } else if cur.X > start.X && c.at(cur).isObjectEndTag() { tagged++ } cur.X++ if c.isVisited(cur) { // If the point is already visited, we hit a polygon or a line. break } ch := c.at(cur) if !ch.isTextCont() { break } if tagged == 0 && ch.isSpace() { whiteSpaceStreak++ // Stop when we see 3 consecutive whitespace points. if whiteSpaceStreak > 2 { break } } else { whiteSpaceStreak = 0 } switch tagged { case 1: if !c.at(cur).isObjectEndTag() { tag = append(tag, rune(ch)) } case 2: if c.at(cur).isTagDefinitionSeparator() { tagged++ } else { tagged = -1 } case 3: tagDef = append(tagDef, rune(ch)) } obj.points = append(obj.points, cur) } // If we found a start and end tag marker, we either need to assign the tag to the object, // or we need to assign the specified options to the global canvas option space. if tagged == 2 { t := string(tag) if container := c.EnclosingObjects(start); container != nil { container[0].SetTag(t) } // The tag applies to the text object as well so that properties like // a2s:label can be set. obj.SetTag(t) } else if tagged == 3 { t := string(tag) // A tag definition targeting an object will not be found within any object; we need // to do that calculation here. if matches := objTagRE.FindStringSubmatch(t); matches != nil { if targetX, err := strconv.ParseInt(matches[1], 10, 0); err == nil { if targetY, err := strconv.ParseInt(matches[2], 10, 0); err == nil { for i, o := range c.objects { corner := o.Corners()[0] if corner.X == int(targetX) && corner.Y == int(targetY) { c.objects[i].SetTag(t) break } } } } } // This is a tag definition. Parse the JSON and assign the options to the canvas. var m interface{} def := []byte(string(tagDef)) if err := json.Unmarshal(def, &m); err != nil { // TODO(dhobsd): Gross. panic(err) } // The tag applies to the reference object as well, so that properties like // a2s:delref can be set.
// Trim the right side of the text object. for len(obj.points) != 0 && c.at(obj.points[len(obj.points)-1]).isSpace() { obj.points = obj.points[:len(obj.points)-1] } obj.seal(c) return obj } func (c *canvas) at(p Point) char { return c.grid[p.Y*c.size.X+p.X] } func (c *canvas) isVisited(p Point) bool { return c.visited[p.Y*c.size.X+p.X] } func (c *canvas) visit(p Point) { // TODO(dhobsd): Change code to ensure that visit() is called once and only // once per point. c.visited[p.Y*c.size.X+p.X] = true } func (c *canvas) unvisit(p Point) { o := p.Y*c.size.X + p.X if !c.visited[o] { panic(fmt.Errorf("internal error: point %+v never visited", p)) } c.visited[o] = false } func (c *canvas) canLeft(p Point) bool { return p.X > 0 } func (c *canvas) canRight(p Point) bool { return p.X < c.size.X-1 } func (c *canvas) canUp(p Point) bool { return p.Y > 0 } func (c *canvas) canDown(p Point) bool { return p.Y < c.size.Y-1 } func (c *canvas) canDiagonal(p Point) bool { return (c.canLeft(p) || c.canRight(p)) && (c.canUp(p) || c.canDown(p)) }
obj.SetTag(t) c.options[t] = m.(map[string]interface{}) }
random_line_split
canvas.go
// Copyright 2012 - 2018 The ASCIIToSVG Contributors // All rights reserved. package asciitosvg import ( "bytes" "encoding/json" "fmt" "image" "regexp" "sort" "strconv" "unicode/utf8" ) // Canvas provides methods for returning objects from an underlying textual grid. type Canvas interface { // A canvas has an underlying visual representation. The fmt.Stringer interface for this // interface provides a view into the underlying grid. fmt.Stringer // Objects returns all the objects found in the underlying grid. Objects() []Object // Size returns the visual dimensions of the Canvas. Size() image.Point // Options returns a map of options to apply to Objects based on the object's tag. This // maps tag name to a map of option names to options. Options() map[string]map[string]interface{} // EnclosingObjects returns the set of objects that contain this point in order from most // to least specific. EnclosingObjects(p Point) []Object } // NewCanvas returns a new Canvas, initialized from the provided data. If tabWidth is set to a non-negative // value, that value will be used to convert tabs to spaces within the grid. Creation of the Canvas // can fail if the diagram contains invalid UTF-8 sequences. func NewCanvas(data []byte, tabWidth int, noBlur bool) (Canvas, error) { c := &canvas{ options: map[string]map[string]interface{}{ "__a2s__closed__options__": map[string]interface{}{ "fill": "#fff", "filter": "url(#dsFilter)", }, }, } if noBlur { c.options["__a2s__closed__options__"] = map[string]interface{}{ "fill": "#fff", } } lines := bytes.Split(data, []byte("\n")) c.size.Y = len(lines) // Diagrams will often not be padded to a uniform width. To overcome this, we scan over // each line and figure out which is the longest. This becomes the width of the canvas. for i, line := range lines { if ok := utf8.Valid(line); !ok { return nil, fmt.Errorf("invalid UTF-8 encoding on line %d", i) } l, err := expandTabs(line, tabWidth) if err != nil { return nil, err } lines[i] = l if i := utf8.RuneCount(lines[i]); i > c.size.X { c.size.X = i } } c.grid = make([]char, c.size.X*c.size.Y) c.visited = make([]bool, c.size.X*c.size.Y) for y, line := range lines { x := 0 for len(line) > 0 { r, l := utf8.DecodeRune(line) c.grid[y*c.size.X+x] = char(r) x++ line = line[l:] } for ; x < c.size.X; x++ { c.grid[y*c.size.X+x] = ' ' } } c.findObjects() return c, nil } // The expandTabs function pads tab characters to the specified width of spaces for the provided // line of input. We cannot simply pad based on byte-offset since our input is UTF-8 encoded. // Fortunately, we can assume that this function is called that the line contains only valid // UTF-8 sequences. We first decode the line rune-wise, and use individual runes to figure out // where we are within the line. When we encounter a tab character, we expand based on our rune // index. func expandTabs(line []byte, tabWidth int) ([]byte, error) { // Initial sizing of our output slice assumes no UTF-8 bytes or tabs, since this is often // the common case. out := make([]byte, 0, len(line)) // pos tracks our position in the input byte slice, while index tracks our position in the // resulting output slice. pos := 0 index := 0 for _, c := range line { if c == '\t' { // Loop over the remaining space count for this particular tabstop until // the next, replacing each position with a space. for s := tabWidth - (pos % tabWidth); s > 0; s-- { out = append(out, ' ') index++ } pos++ } else { // We need to know the byte length of the rune at this position so that we // can account for our tab expansion properly. So we first decode the rune // at this position to get its length in bytes, plop that rune back into our // output slice, and account accordingly. r, l := utf8.DecodeRune(line[pos:]) if r == utf8.RuneError { return nil, fmt.Errorf("invalid rune at byte offset %d; rune offset %d", pos, index) } enc := make([]byte, l) utf8.EncodeRune(enc, r) out = append(out, enc...) pos += l index++ } } return out, nil } // canvas is the parsed source data. type canvas struct { // (0,0) is top left. grid []char visited []bool objects objects size image.Point options map[string]map[string]interface{} } func (c *canvas) String() string { return fmt.Sprintf("%+v", c.grid) } func (c *canvas) Objects() []Object { return c.objects } func (c *canvas) Size() image.Point { return c.size } func (c *canvas) Options() map[string]map[string]interface{} { return c.options } func (c *canvas) EnclosingObjects(p Point) []Object { maxTL := Point{X: -1, Y: -1} var q []Object for _, o := range c.objects { // An object can't really contain another unless it is a polygon. if !o.IsClosed() { continue } if o.HasPoint(p) && o.Corners()[0].X > maxTL.X && o.Corners()[0].Y > maxTL.Y { q = append(q, o) maxTL.X = o.Corners()[0].X maxTL.Y = o.Corners()[0].Y } } return q } // findObjects finds all objects (lines, polygons, and text) within the underlying grid. func (c *canvas) findObjects() { p := Point{} // Find any new paths by starting with a point that wasn't yet visited, beginning at the top // left of the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isPathStart() { // Found the start of a one or multiple connected paths. Traverse all // connecting points. This will generate multiple objects if multiple // paths (either open or closed) are found. c.visit(p) objs := c.scanPath([]Point{p}) for _, obj := range objs { // For all points in all objects found, mark the points as visited. for _, p := range obj.Points() { c.visit(p) } } c.objects = append(c.objects, objs...) } } } // A second pass through the grid attempts to identify any text within the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isTextStart() { obj := c.scanText(p) // scanText will return nil if the text at this area is simply // setting options on a container object. if obj == nil { continue } for _, p := range obj.Points() { c.visit(p) } c.objects = append(c.objects, obj) } } } sort.Sort(c.objects) } // scanPath tries to complete a total path (for lines or polygons) starting with some partial path. // It recurses when it finds multiple unvisited outgoing paths. func (c *canvas) scanPath(points []Point) objects { cur := points[len(points)-1] next := c.next(cur) // If there are no points that can progress traversal of the path, finalize the one we're // working on, and return it. This is the terminal condition in the passive flow. if len(next) == 0 { if len(points) == 1 { // Discard 'path' of 1 point. Do not mark point as visited. c.unvisit(cur) return nil } // TODO(dhobsd): Determine if path is sharing the line with another path. If so, // we may want to join the objects such that we don't get weird rendering artifacts. o := &object{points: points} o.seal(c) return objects{o} } // If we have hit a point that can create a closed path, create an object and close // the path. Additionally, recurse to other progress directions in case e.g. an open // path spawns from this point. Paths are always closed vertically. if cur.X == points[0].X && cur.Y == points[0].Y+1 { o := &object{points: points} o.seal(c) r := objects{o} return append(r, c.scanPath([]Point{cur})...) } // We scan depth-first instead of breadth-first, making it possible to find a // closed path. var objs objects for _, n := range next { if c.isVisited(n) { continue } c.visit(n) p2 := make([]Point, len(points)+1) copy(p2, points) p2[len(p2)-1] = n objs = append(objs, c.scanPath(p2)...) } return objs } // The next returns the points that can be used to make progress, scanning (in order) horizontal // progress to the left or right, vertical progress above or below, or diagonal progress to NW, // NE, SW, and SE. It skips any points already visited, and returns all of the possible progress // points. func (c *canvas) next(pos Point) []Point { // Our caller must have called c.visit prior to calling this function. if !c.isVisited(pos) { panic(fmt.Errorf("internal error; revisiting %s", pos)) } var out []Point ch := c.at(pos) if ch.canHorizontal() { nextHorizontal := func(p Point) { if !c.isVisited(p) && c.at(p).canHorizontal() { out = append(out, p) } } if c.canLeft(pos) { n := pos n.X-- nextHorizontal(n) } if c.canRight(pos) { n := pos n.X++ nextHorizontal(n) } } if ch.canVertical() { nextVertical := func(p Point) { if !c.isVisited(p) && c.at(p).canVertical() { out = append(out, p) } } if c.canUp(pos) { n := pos n.Y-- nextVertical(n) } if c.canDown(pos) { n := pos n.Y++ nextVertical(n) } } if c.canDiagonal(pos) { nextDiagonal := func(from, to Point) { if !c.isVisited(to) && c.at(to).canDiagonalFrom(c.at(from)) { out = append(out, to) } } if c.canUp(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y-- nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y-- nextDiagonal(pos, n) } } if c.canDown(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y++ nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y++ nextDiagonal(pos, n) } } } return out } // Used for matching [X, Y]: {...} tag definitions. These definitions target specific objects. var objTagRE = regexp.MustCompile(`(\d+)\s*,\s*(\d+)$`) // scanText extracts a line of text. func (c *canvas) scanText(start Point) Object { obj := &object{points: []Point{start}, isText: true} whiteSpaceStreak := 0 cur := start tagged := 0 tag := []rune{} tagDef := []rune{} for c.canRight(cur) { if cur.X == start.X && c.at(cur).isObjectStartTag() { tagged++ } else if cur.X > start.X && c.at(cur).isObjectEndTag() { tagged++ } cur.X++ if c.isVisited(cur) { // If the point is already visited, we hit a polygon or a line. break } ch := c.at(cur) if !ch.isTextCont() { break } if tagged == 0 && ch.isSpace() { whiteSpaceStreak++ // Stop when we see 3 consecutive whitespace points. if whiteSpaceStreak > 2 { break } } else { whiteSpaceStreak = 0 } switch tagged { case 1: if !c.at(cur).isObjectEndTag() { tag = append(tag, rune(ch)) } case 2: if c.at(cur).isTagDefinitionSeparator() { tagged++ } else { tagged = -1 } case 3: tagDef = append(tagDef, rune(ch)) } obj.points = append(obj.points, cur) } // If we found a start and end tag marker, we either need to assign the tag to the object, // or we need to assign the specified options to the global canvas option space. if tagged == 2 { t := string(tag) if container := c.EnclosingObjects(start); container != nil { container[0].SetTag(t) } // The tag applies to the text object as well so that properties like // a2s:label can be set. obj.SetTag(t) } else if tagged == 3 { t := string(tag) // A tag definition targeting an object will not be found within any object; we need // to do that calculation here. if matches := objTagRE.FindStringSubmatch(t); matches != nil { if targetX, err := strconv.ParseInt(matches[1], 10, 0); err == nil { if targetY, err := strconv.ParseInt(matches[2], 10, 0); err == nil { for i, o := range c.objects { corner := o.Corners()[0] if corner.X == int(targetX) && corner.Y == int(targetY) { c.objects[i].SetTag(t) break } } } } } // This is a tag definition. Parse the JSON and assign the options to the canvas. var m interface{} def := []byte(string(tagDef)) if err := json.Unmarshal(def, &m); err != nil { // TODO(dhobsd): Gross. panic(err) } // The tag applies to the reference object as well, so that properties like // a2s:delref can be set. obj.SetTag(t) c.options[t] = m.(map[string]interface{}) } // Trim the right side of the text object. for len(obj.points) != 0 && c.at(obj.points[len(obj.points)-1]).isSpace() { obj.points = obj.points[:len(obj.points)-1] } obj.seal(c) return obj } func (c *canvas) at(p Point) char { return c.grid[p.Y*c.size.X+p.X] } func (c *canvas) isVisited(p Point) bool { return c.visited[p.Y*c.size.X+p.X] } func (c *canvas) visit(p Point) { // TODO(dhobsd): Change code to ensure that visit() is called once and only // once per point. c.visited[p.Y*c.size.X+p.X] = true } func (c *canvas)
(p Point) { o := p.Y*c.size.X + p.X if !c.visited[o] { panic(fmt.Errorf("internal error: point %+v never visited", p)) } c.visited[o] = false } func (c *canvas) canLeft(p Point) bool { return p.X > 0 } func (c *canvas) canRight(p Point) bool { return p.X < c.size.X-1 } func (c *canvas) canUp(p Point) bool { return p.Y > 0 } func (c *canvas) canDown(p Point) bool { return p.Y < c.size.Y-1 } func (c *canvas) canDiagonal(p Point) bool { return (c.canLeft(p) || c.canRight(p)) && (c.canUp(p) || c.canDown(p)) }
unvisit
identifier_name
canvas.go
// Copyright 2012 - 2018 The ASCIIToSVG Contributors // All rights reserved. package asciitosvg import ( "bytes" "encoding/json" "fmt" "image" "regexp" "sort" "strconv" "unicode/utf8" ) // Canvas provides methods for returning objects from an underlying textual grid. type Canvas interface { // A canvas has an underlying visual representation. The fmt.Stringer interface for this // interface provides a view into the underlying grid. fmt.Stringer // Objects returns all the objects found in the underlying grid. Objects() []Object // Size returns the visual dimensions of the Canvas. Size() image.Point // Options returns a map of options to apply to Objects based on the object's tag. This // maps tag name to a map of option names to options. Options() map[string]map[string]interface{} // EnclosingObjects returns the set of objects that contain this point in order from most // to least specific. EnclosingObjects(p Point) []Object } // NewCanvas returns a new Canvas, initialized from the provided data. If tabWidth is set to a non-negative // value, that value will be used to convert tabs to spaces within the grid. Creation of the Canvas // can fail if the diagram contains invalid UTF-8 sequences. func NewCanvas(data []byte, tabWidth int, noBlur bool) (Canvas, error) { c := &canvas{ options: map[string]map[string]interface{}{ "__a2s__closed__options__": map[string]interface{}{ "fill": "#fff", "filter": "url(#dsFilter)", }, }, } if noBlur { c.options["__a2s__closed__options__"] = map[string]interface{}{ "fill": "#fff", } } lines := bytes.Split(data, []byte("\n")) c.size.Y = len(lines) // Diagrams will often not be padded to a uniform width. To overcome this, we scan over // each line and figure out which is the longest. This becomes the width of the canvas. for i, line := range lines { if ok := utf8.Valid(line); !ok { return nil, fmt.Errorf("invalid UTF-8 encoding on line %d", i) } l, err := expandTabs(line, tabWidth) if err != nil { return nil, err } lines[i] = l if i := utf8.RuneCount(lines[i]); i > c.size.X { c.size.X = i } } c.grid = make([]char, c.size.X*c.size.Y) c.visited = make([]bool, c.size.X*c.size.Y) for y, line := range lines { x := 0 for len(line) > 0 { r, l := utf8.DecodeRune(line) c.grid[y*c.size.X+x] = char(r) x++ line = line[l:] } for ; x < c.size.X; x++ { c.grid[y*c.size.X+x] = ' ' } } c.findObjects() return c, nil } // The expandTabs function pads tab characters to the specified width of spaces for the provided // line of input. We cannot simply pad based on byte-offset since our input is UTF-8 encoded. // Fortunately, we can assume that this function is called that the line contains only valid // UTF-8 sequences. We first decode the line rune-wise, and use individual runes to figure out // where we are within the line. When we encounter a tab character, we expand based on our rune // index. func expandTabs(line []byte, tabWidth int) ([]byte, error) { // Initial sizing of our output slice assumes no UTF-8 bytes or tabs, since this is often // the common case. out := make([]byte, 0, len(line)) // pos tracks our position in the input byte slice, while index tracks our position in the // resulting output slice. pos := 0 index := 0 for _, c := range line { if c == '\t' { // Loop over the remaining space count for this particular tabstop until // the next, replacing each position with a space. for s := tabWidth - (pos % tabWidth); s > 0; s-- { out = append(out, ' ') index++ } pos++ } else { // We need to know the byte length of the rune at this position so that we // can account for our tab expansion properly. So we first decode the rune // at this position to get its length in bytes, plop that rune back into our // output slice, and account accordingly. r, l := utf8.DecodeRune(line[pos:]) if r == utf8.RuneError { return nil, fmt.Errorf("invalid rune at byte offset %d; rune offset %d", pos, index) } enc := make([]byte, l) utf8.EncodeRune(enc, r) out = append(out, enc...) pos += l index++ } } return out, nil } // canvas is the parsed source data. type canvas struct { // (0,0) is top left. grid []char visited []bool objects objects size image.Point options map[string]map[string]interface{} } func (c *canvas) String() string { return fmt.Sprintf("%+v", c.grid) } func (c *canvas) Objects() []Object { return c.objects } func (c *canvas) Size() image.Point { return c.size } func (c *canvas) Options() map[string]map[string]interface{} { return c.options } func (c *canvas) EnclosingObjects(p Point) []Object { maxTL := Point{X: -1, Y: -1} var q []Object for _, o := range c.objects { // An object can't really contain another unless it is a polygon. if !o.IsClosed() { continue } if o.HasPoint(p) && o.Corners()[0].X > maxTL.X && o.Corners()[0].Y > maxTL.Y { q = append(q, o) maxTL.X = o.Corners()[0].X maxTL.Y = o.Corners()[0].Y } } return q } // findObjects finds all objects (lines, polygons, and text) within the underlying grid. func (c *canvas) findObjects() { p := Point{} // Find any new paths by starting with a point that wasn't yet visited, beginning at the top // left of the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isPathStart() { // Found the start of a one or multiple connected paths. Traverse all // connecting points. This will generate multiple objects if multiple // paths (either open or closed) are found. c.visit(p) objs := c.scanPath([]Point{p}) for _, obj := range objs { // For all points in all objects found, mark the points as visited. for _, p := range obj.Points() { c.visit(p) } } c.objects = append(c.objects, objs...) } } } // A second pass through the grid attempts to identify any text within the grid. for y := 0; y < c.size.Y; y++ { p.Y = y for x := 0; x < c.size.X; x++ { p.X = x if c.isVisited(p) { continue } if ch := c.at(p); ch.isTextStart() { obj := c.scanText(p) // scanText will return nil if the text at this area is simply // setting options on a container object. if obj == nil { continue } for _, p := range obj.Points() { c.visit(p) } c.objects = append(c.objects, obj) } } } sort.Sort(c.objects) } // scanPath tries to complete a total path (for lines or polygons) starting with some partial path. // It recurses when it finds multiple unvisited outgoing paths. func (c *canvas) scanPath(points []Point) objects { cur := points[len(points)-1] next := c.next(cur) // If there are no points that can progress traversal of the path, finalize the one we're // working on, and return it. This is the terminal condition in the passive flow. if len(next) == 0 { if len(points) == 1 { // Discard 'path' of 1 point. Do not mark point as visited. c.unvisit(cur) return nil } // TODO(dhobsd): Determine if path is sharing the line with another path. If so, // we may want to join the objects such that we don't get weird rendering artifacts. o := &object{points: points} o.seal(c) return objects{o} } // If we have hit a point that can create a closed path, create an object and close // the path. Additionally, recurse to other progress directions in case e.g. an open // path spawns from this point. Paths are always closed vertically. if cur.X == points[0].X && cur.Y == points[0].Y+1 { o := &object{points: points} o.seal(c) r := objects{o} return append(r, c.scanPath([]Point{cur})...) } // We scan depth-first instead of breadth-first, making it possible to find a // closed path. var objs objects for _, n := range next
return objs } // The next returns the points that can be used to make progress, scanning (in order) horizontal // progress to the left or right, vertical progress above or below, or diagonal progress to NW, // NE, SW, and SE. It skips any points already visited, and returns all of the possible progress // points. func (c *canvas) next(pos Point) []Point { // Our caller must have called c.visit prior to calling this function. if !c.isVisited(pos) { panic(fmt.Errorf("internal error; revisiting %s", pos)) } var out []Point ch := c.at(pos) if ch.canHorizontal() { nextHorizontal := func(p Point) { if !c.isVisited(p) && c.at(p).canHorizontal() { out = append(out, p) } } if c.canLeft(pos) { n := pos n.X-- nextHorizontal(n) } if c.canRight(pos) { n := pos n.X++ nextHorizontal(n) } } if ch.canVertical() { nextVertical := func(p Point) { if !c.isVisited(p) && c.at(p).canVertical() { out = append(out, p) } } if c.canUp(pos) { n := pos n.Y-- nextVertical(n) } if c.canDown(pos) { n := pos n.Y++ nextVertical(n) } } if c.canDiagonal(pos) { nextDiagonal := func(from, to Point) { if !c.isVisited(to) && c.at(to).canDiagonalFrom(c.at(from)) { out = append(out, to) } } if c.canUp(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y-- nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y-- nextDiagonal(pos, n) } } if c.canDown(pos) { if c.canLeft(pos) { n := pos n.X-- n.Y++ nextDiagonal(pos, n) } if c.canRight(pos) { n := pos n.X++ n.Y++ nextDiagonal(pos, n) } } } return out } // Used for matching [X, Y]: {...} tag definitions. These definitions target specific objects. var objTagRE = regexp.MustCompile(`(\d+)\s*,\s*(\d+)$`) // scanText extracts a line of text. func (c *canvas) scanText(start Point) Object { obj := &object{points: []Point{start}, isText: true} whiteSpaceStreak := 0 cur := start tagged := 0 tag := []rune{} tagDef := []rune{} for c.canRight(cur) { if cur.X == start.X && c.at(cur).isObjectStartTag() { tagged++ } else if cur.X > start.X && c.at(cur).isObjectEndTag() { tagged++ } cur.X++ if c.isVisited(cur) { // If the point is already visited, we hit a polygon or a line. break } ch := c.at(cur) if !ch.isTextCont() { break } if tagged == 0 && ch.isSpace() { whiteSpaceStreak++ // Stop when we see 3 consecutive whitespace points. if whiteSpaceStreak > 2 { break } } else { whiteSpaceStreak = 0 } switch tagged { case 1: if !c.at(cur).isObjectEndTag() { tag = append(tag, rune(ch)) } case 2: if c.at(cur).isTagDefinitionSeparator() { tagged++ } else { tagged = -1 } case 3: tagDef = append(tagDef, rune(ch)) } obj.points = append(obj.points, cur) } // If we found a start and end tag marker, we either need to assign the tag to the object, // or we need to assign the specified options to the global canvas option space. if tagged == 2 { t := string(tag) if container := c.EnclosingObjects(start); container != nil { container[0].SetTag(t) } // The tag applies to the text object as well so that properties like // a2s:label can be set. obj.SetTag(t) } else if tagged == 3 { t := string(tag) // A tag definition targeting an object will not be found within any object; we need // to do that calculation here. if matches := objTagRE.FindStringSubmatch(t); matches != nil { if targetX, err := strconv.ParseInt(matches[1], 10, 0); err == nil { if targetY, err := strconv.ParseInt(matches[2], 10, 0); err == nil { for i, o := range c.objects { corner := o.Corners()[0] if corner.X == int(targetX) && corner.Y == int(targetY) { c.objects[i].SetTag(t) break } } } } } // This is a tag definition. Parse the JSON and assign the options to the canvas. var m interface{} def := []byte(string(tagDef)) if err := json.Unmarshal(def, &m); err != nil { // TODO(dhobsd): Gross. panic(err) } // The tag applies to the reference object as well, so that properties like // a2s:delref can be set. obj.SetTag(t) c.options[t] = m.(map[string]interface{}) } // Trim the right side of the text object. for len(obj.points) != 0 && c.at(obj.points[len(obj.points)-1]).isSpace() { obj.points = obj.points[:len(obj.points)-1] } obj.seal(c) return obj } func (c *canvas) at(p Point) char { return c.grid[p.Y*c.size.X+p.X] } func (c *canvas) isVisited(p Point) bool { return c.visited[p.Y*c.size.X+p.X] } func (c *canvas) visit(p Point) { // TODO(dhobsd): Change code to ensure that visit() is called once and only // once per point. c.visited[p.Y*c.size.X+p.X] = true } func (c *canvas) unvisit(p Point) { o := p.Y*c.size.X + p.X if !c.visited[o] { panic(fmt.Errorf("internal error: point %+v never visited", p)) } c.visited[o] = false } func (c *canvas) canLeft(p Point) bool { return p.X > 0 } func (c *canvas) canRight(p Point) bool { return p.X < c.size.X-1 } func (c *canvas) canUp(p Point) bool { return p.Y > 0 } func (c *canvas) canDown(p Point) bool { return p.Y < c.size.Y-1 } func (c *canvas) canDiagonal(p Point) bool { return (c.canLeft(p) || c.canRight(p)) && (c.canUp(p) || c.canDown(p)) }
{ if c.isVisited(n) { continue } c.visit(n) p2 := make([]Point, len(points)+1) copy(p2, points) p2[len(p2)-1] = n objs = append(objs, c.scanPath(p2)...) }
conditional_block
client.rs
use std::io; use std::io::prelude::*; use std::net::TcpStream; use std::cmp; use rand::{Rng, OsRng}; use alert; use tls_result::{TlsResult, TlsError, TlsErrorKind}; use tls_result::TlsErrorKind::{UnexpectedMessage, InternalError, DecryptError, IllegalParameter}; use util::{SurugaError, crypto_compare}; use cipher::{self, Aead}; use cipher::prf::Prf; use crypto::sha2::sha256; use tls_item::{TlsItem, DummyItem}; use handshake::{self, Handshake}; use tls::{TlsReader, TlsWriter, TLS_VERSION}; // handshake is done during construction. pub struct TlsClient<R: Read, W: Write> { pub reader: TlsReader<R>, pub writer: TlsWriter<W>, pub rng: OsRng, buf: Vec<u8>, } impl<R: Read, W: Write> TlsClient<R, W> { pub fn new(reader: R, writer: W, rng: OsRng) -> TlsResult<TlsClient<R, W>> { let mut client = TlsClient { reader: TlsReader::new(reader), writer: TlsWriter::new(writer), rng: rng, buf: Vec::new(), }; // handshake failed. send alert if necessary match client.handshake() { Ok(()) => {} Err(err) => return Err(client.send_tls_alert(err)), } Ok(client) } #[inline] pub fn reader(&mut self) -> &mut R { self.reader.get_mut() } #[inline] pub fn writer(&mut self) -> &mut W { self.writer.get_mut() } // this does not send alert when error occurs fn handshake(&mut self) -> TlsResult<()> { // expect specific HandshakeMessage. otherwise return Err macro_rules! expect { ($var:ident) => ({ match try!(self.reader.read_handshake()) { handshake::Handshake::$var(data) => data, _ => return tls_err!(UnexpectedMessage, "unexpected handshake message found"), } }) } let cli_random = { let mut random_bytes = [0u8; 32]; self.rng.fill_bytes(&mut random_bytes); random_bytes.to_vec() }; let random = try!(handshake::Random::new(cli_random.clone())); // the only cipher we currently support let cipher_suite = cipher::CipherSuite::TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256; let curve_list = vec!(handshake::NamedCurve::secp256r1); let curve_list = try!(handshake::Extension::new_elliptic_curve_list(curve_list)); let format_list = vec!(handshake::ECPointFormat::uncompressed); let format_list = try!(handshake::Extension::new_ec_point_formats(format_list)); let extensions = vec!(curve_list, format_list); let client_hello = try!(Handshake::new_client_hello(random, cipher_suite, extensions)); try!(self.writer.write_handshake(&client_hello)); let server_hello_data = expect!(server_hello); { let server_major = server_hello_data.server_version.major; let server_minor = server_hello_data.server_version.minor; if (server_major, server_minor) != TLS_VERSION { return tls_err!(IllegalParameter, "wrong server version: {} {}", server_major, server_minor); } if server_hello_data.cipher_suite != cipher_suite { return tls_err!(IllegalParameter, "cipher suite mismatch: found {:?}", server_hello_data.cipher_suite); } if server_hello_data.compression_method != handshake::CompressionMethod::null { return tls_err!(IllegalParameter, "compression method mismatch"); } // FIXME: check if server sent unknown extension // it is currently done by just not understanding any extensions // other than we used. } // we always expect certificate. let certificate_list = expect!(certificate); // TODO: cert validation not implemented yet // we always use server key exchange let server_key_ex_data = expect!(server_key_exchange); let kex = cipher_suite.new_kex(); let (key_data, pre_master_secret) = try!(kex.compute_keys(&server_key_ex_data, &mut self.rng)); expect!(server_hello_done); let client_key_exchange = try!(Handshake::new_client_key_exchange(key_data)); try!(self.writer.write_handshake(&client_key_exchange)); try!(self.writer.write_change_cipher_spec()); // SECRET let master_secret = { let mut label_seed = b"master secret".to_vec(); label_seed.extend(&cli_random); label_seed.extend(&server_hello_data.random[..]); let mut prf = Prf::new(pre_master_secret, label_seed); prf.get_bytes(48) }; let aead = cipher_suite.new_aead(); // SECRET let read_key = { let mut label_seed = b"key expansion".to_vec(); label_seed.extend(&server_hello_data.random[..]); label_seed.extend(&cli_random); let mut prf = Prf::new(master_secret.clone(), label_seed); // mac_key is not used in AEAD configuration. let enc_key_length = aead.key_size(); let write_key = prf.get_bytes(enc_key_length); let encryptor = aead.new_encryptor(write_key); self.writer.set_encryptor(encryptor); // this will be set after receiving ChangeCipherSpec. let read_key = prf.get_bytes(enc_key_length); // chacha20-poly1305 does not use iv. read_key }; // FIXME we should get "raw" packet data and hash them incrementally let msgs = { let mut msgs = Vec::new(); try!(client_hello.tls_write(&mut msgs)); try!(Handshake::server_hello(server_hello_data).tls_write(&mut msgs)); try!(Handshake::certificate(certificate_list).tls_write(&mut msgs)); try!(Handshake::server_key_exchange(server_key_ex_data).tls_write(&mut msgs)); try!(Handshake::server_hello_done(DummyItem).tls_write(&mut msgs)); try!(client_key_exchange.tls_write(&mut msgs)); msgs }; // this only verifies Handshake messages! what about others? // ApplicationData messages are not permitted until now. // ChangeCipherSpec messages are only permitted after ClinetKeyExchange. // Alert messages can be problematic - they are not verified and // can be broken into several records. This leads to alert attack. // since we don't accept strange alerts, all "normal" alert messages are // treated as error, so now we can assert that we haven't received alerts. let verify_hash = sha256(&msgs); let client_verify_data = { let finished_label = b"client finished"; let mut label_seed = finished_label.to_vec(); label_seed.extend(&verify_hash); let mut prf = Prf::new(master_secret.clone(), label_seed); prf.get_bytes(cipher_suite.verify_data_len()) }; let finished = try!(Handshake::new_finished(client_verify_data)); try!(self.writer.write_handshake(&finished)); // Although client->server is encrypted, server->client isn't yet. // server may send either ChangeCipherSpec or Alert. try!(self.reader.read_change_cipher_spec()); // from now server starts encryption. self.reader.set_decryptor(aead.new_decryptor(read_key)); let server_finished = expect!(finished); { let verify_hash = { // ideally we may save "raw" packet data.. let mut serv_msgs = Vec::new(); // FIXME: this should not throw "io error".. should throw "internal error" try!(Write::write_all(&mut serv_msgs, &msgs)); try!(finished.tls_write(&mut serv_msgs)); let verify_hash = sha256(&serv_msgs); verify_hash }; let server_verify_data = { let finished_label = b"server finished"; let mut label_seed = finished_label.to_vec(); label_seed.extend(&verify_hash); let mut prf = Prf::new(master_secret, label_seed); prf.get_bytes(cipher_suite.verify_data_len()) }; let verify_ok = crypto_compare(&server_finished, &server_verify_data); if !verify_ok { return tls_err!(DecryptError, "server sent wrong verify data"); } } Ok(()) } pub fn close(&mut self) -> TlsResult<()> { let alert_data = alert::Alert { level: alert::AlertLevel::fatal, description: alert::AlertDescription::close_notify, }; try!(self.writer.write_alert(&alert_data)); Ok(()) } // send fatal alert and return error // (it may be different to `err`, because writing alert can fail) pub fn send_tls_alert(&mut self, err: TlsError) -> TlsError { match err.kind { TlsErrorKind::IoFailure => return err, _ => { let alert = alert::Alert::from_tls_err(&err); let result = self.writer.write_alert(&alert); match result { Ok(()) => return err, Err(err) => return err, } } } } } impl TlsClient<TcpStream, TcpStream> { pub fn from_tcp(stream: TcpStream) -> TlsResult<TlsClient<TcpStream, TcpStream>> { let rng = match OsRng::new() { Ok(rng) => rng, Err(..) => return tls_err!(InternalError, "failed to create OsRng"), }; let reader = try!(stream.try_clone()); let writer = stream; TlsClient::new(reader, writer, rng) } } impl<R: Read, W: Write> Write for TlsClient<R, W> { // this either writes all or fails. fn write(&mut self, buf: &[u8]) -> io::Result<usize> { try!(self.write_all(buf)); Ok(buf.len()) } fn flush(&mut self) -> io::Result<()> { Ok(()) } fn
(&mut self, buf: &[u8]) -> io::Result<()> { let result = self.writer.write_application_data(buf); match result { Ok(()) => Ok(()), Err(err) => { let err = self.send_tls_alert(err); // FIXME more verbose io error Err(io::Error::new(io::ErrorKind::Other, SurugaError { desc: "TLS write error", cause: Some(Box::new(err)), })) } } } } // A replacement for the deprecated std::slice::bytes::copy_memory fn copy_memory(from: &[u8], mut to: &mut [u8]) -> usize { to.write(from).unwrap() } impl<R: Read, W: Write> Read for TlsClient<R, W> { // if ssl connection is failed, return `EndOfFile`. fn read(&mut self, buf: &mut [u8]) -> io::Result<usize> { let mut pos: usize = 0; let len = buf.len(); while pos < len { let remaining = len - pos; if self.buf.len() == 0 { let data = match self.reader.read_application_data() { Ok(data) => data, Err(_err) => { break; // FIXME: stop if EOF. otherwise raise error? } }; self.buf.extend(&data); } let selflen = self.buf.len(); let necessary = cmp::min(remaining, selflen); copy_memory(&self.buf[.. necessary], &mut buf[pos .. pos + necessary]); pos += necessary; self.buf = self.buf[necessary..].to_vec(); } Ok(pos) } }
write_all
identifier_name
client.rs
use std::io; use std::io::prelude::*; use std::net::TcpStream; use std::cmp; use rand::{Rng, OsRng}; use alert; use tls_result::{TlsResult, TlsError, TlsErrorKind}; use tls_result::TlsErrorKind::{UnexpectedMessage, InternalError, DecryptError, IllegalParameter}; use util::{SurugaError, crypto_compare}; use cipher::{self, Aead}; use cipher::prf::Prf; use crypto::sha2::sha256; use tls_item::{TlsItem, DummyItem}; use handshake::{self, Handshake}; use tls::{TlsReader, TlsWriter, TLS_VERSION}; // handshake is done during construction. pub struct TlsClient<R: Read, W: Write> { pub reader: TlsReader<R>, pub writer: TlsWriter<W>, pub rng: OsRng, buf: Vec<u8>, } impl<R: Read, W: Write> TlsClient<R, W> { pub fn new(reader: R, writer: W, rng: OsRng) -> TlsResult<TlsClient<R, W>> { let mut client = TlsClient { reader: TlsReader::new(reader), writer: TlsWriter::new(writer), rng: rng, buf: Vec::new(), }; // handshake failed. send alert if necessary match client.handshake() { Ok(()) => {} Err(err) => return Err(client.send_tls_alert(err)), } Ok(client) } #[inline] pub fn reader(&mut self) -> &mut R { self.reader.get_mut() } #[inline] pub fn writer(&mut self) -> &mut W { self.writer.get_mut() } // this does not send alert when error occurs fn handshake(&mut self) -> TlsResult<()> { // expect specific HandshakeMessage. otherwise return Err macro_rules! expect { ($var:ident) => ({ match try!(self.reader.read_handshake()) { handshake::Handshake::$var(data) => data, _ => return tls_err!(UnexpectedMessage, "unexpected handshake message found"), } }) } let cli_random = { let mut random_bytes = [0u8; 32]; self.rng.fill_bytes(&mut random_bytes); random_bytes.to_vec() }; let random = try!(handshake::Random::new(cli_random.clone())); // the only cipher we currently support let cipher_suite = cipher::CipherSuite::TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256; let curve_list = vec!(handshake::NamedCurve::secp256r1); let curve_list = try!(handshake::Extension::new_elliptic_curve_list(curve_list)); let format_list = vec!(handshake::ECPointFormat::uncompressed); let format_list = try!(handshake::Extension::new_ec_point_formats(format_list)); let extensions = vec!(curve_list, format_list); let client_hello = try!(Handshake::new_client_hello(random, cipher_suite, extensions)); try!(self.writer.write_handshake(&client_hello)); let server_hello_data = expect!(server_hello); { let server_major = server_hello_data.server_version.major; let server_minor = server_hello_data.server_version.minor; if (server_major, server_minor) != TLS_VERSION { return tls_err!(IllegalParameter, "wrong server version: {} {}", server_major, server_minor); } if server_hello_data.cipher_suite != cipher_suite { return tls_err!(IllegalParameter, "cipher suite mismatch: found {:?}", server_hello_data.cipher_suite); } if server_hello_data.compression_method != handshake::CompressionMethod::null { return tls_err!(IllegalParameter, "compression method mismatch"); } // FIXME: check if server sent unknown extension // it is currently done by just not understanding any extensions // other than we used. } // we always expect certificate. let certificate_list = expect!(certificate); // TODO: cert validation not implemented yet // we always use server key exchange let server_key_ex_data = expect!(server_key_exchange); let kex = cipher_suite.new_kex(); let (key_data, pre_master_secret) = try!(kex.compute_keys(&server_key_ex_data, &mut self.rng)); expect!(server_hello_done); let client_key_exchange = try!(Handshake::new_client_key_exchange(key_data));
try!(self.writer.write_handshake(&client_key_exchange)); try!(self.writer.write_change_cipher_spec()); // SECRET let master_secret = { let mut label_seed = b"master secret".to_vec(); label_seed.extend(&cli_random); label_seed.extend(&server_hello_data.random[..]); let mut prf = Prf::new(pre_master_secret, label_seed); prf.get_bytes(48) }; let aead = cipher_suite.new_aead(); // SECRET let read_key = { let mut label_seed = b"key expansion".to_vec(); label_seed.extend(&server_hello_data.random[..]); label_seed.extend(&cli_random); let mut prf = Prf::new(master_secret.clone(), label_seed); // mac_key is not used in AEAD configuration. let enc_key_length = aead.key_size(); let write_key = prf.get_bytes(enc_key_length); let encryptor = aead.new_encryptor(write_key); self.writer.set_encryptor(encryptor); // this will be set after receiving ChangeCipherSpec. let read_key = prf.get_bytes(enc_key_length); // chacha20-poly1305 does not use iv. read_key }; // FIXME we should get "raw" packet data and hash them incrementally let msgs = { let mut msgs = Vec::new(); try!(client_hello.tls_write(&mut msgs)); try!(Handshake::server_hello(server_hello_data).tls_write(&mut msgs)); try!(Handshake::certificate(certificate_list).tls_write(&mut msgs)); try!(Handshake::server_key_exchange(server_key_ex_data).tls_write(&mut msgs)); try!(Handshake::server_hello_done(DummyItem).tls_write(&mut msgs)); try!(client_key_exchange.tls_write(&mut msgs)); msgs }; // this only verifies Handshake messages! what about others? // ApplicationData messages are not permitted until now. // ChangeCipherSpec messages are only permitted after ClinetKeyExchange. // Alert messages can be problematic - they are not verified and // can be broken into several records. This leads to alert attack. // since we don't accept strange alerts, all "normal" alert messages are // treated as error, so now we can assert that we haven't received alerts. let verify_hash = sha256(&msgs); let client_verify_data = { let finished_label = b"client finished"; let mut label_seed = finished_label.to_vec(); label_seed.extend(&verify_hash); let mut prf = Prf::new(master_secret.clone(), label_seed); prf.get_bytes(cipher_suite.verify_data_len()) }; let finished = try!(Handshake::new_finished(client_verify_data)); try!(self.writer.write_handshake(&finished)); // Although client->server is encrypted, server->client isn't yet. // server may send either ChangeCipherSpec or Alert. try!(self.reader.read_change_cipher_spec()); // from now server starts encryption. self.reader.set_decryptor(aead.new_decryptor(read_key)); let server_finished = expect!(finished); { let verify_hash = { // ideally we may save "raw" packet data.. let mut serv_msgs = Vec::new(); // FIXME: this should not throw "io error".. should throw "internal error" try!(Write::write_all(&mut serv_msgs, &msgs)); try!(finished.tls_write(&mut serv_msgs)); let verify_hash = sha256(&serv_msgs); verify_hash }; let server_verify_data = { let finished_label = b"server finished"; let mut label_seed = finished_label.to_vec(); label_seed.extend(&verify_hash); let mut prf = Prf::new(master_secret, label_seed); prf.get_bytes(cipher_suite.verify_data_len()) }; let verify_ok = crypto_compare(&server_finished, &server_verify_data); if !verify_ok { return tls_err!(DecryptError, "server sent wrong verify data"); } } Ok(()) } pub fn close(&mut self) -> TlsResult<()> { let alert_data = alert::Alert { level: alert::AlertLevel::fatal, description: alert::AlertDescription::close_notify, }; try!(self.writer.write_alert(&alert_data)); Ok(()) } // send fatal alert and return error // (it may be different to `err`, because writing alert can fail) pub fn send_tls_alert(&mut self, err: TlsError) -> TlsError { match err.kind { TlsErrorKind::IoFailure => return err, _ => { let alert = alert::Alert::from_tls_err(&err); let result = self.writer.write_alert(&alert); match result { Ok(()) => return err, Err(err) => return err, } } } } } impl TlsClient<TcpStream, TcpStream> { pub fn from_tcp(stream: TcpStream) -> TlsResult<TlsClient<TcpStream, TcpStream>> { let rng = match OsRng::new() { Ok(rng) => rng, Err(..) => return tls_err!(InternalError, "failed to create OsRng"), }; let reader = try!(stream.try_clone()); let writer = stream; TlsClient::new(reader, writer, rng) } } impl<R: Read, W: Write> Write for TlsClient<R, W> { // this either writes all or fails. fn write(&mut self, buf: &[u8]) -> io::Result<usize> { try!(self.write_all(buf)); Ok(buf.len()) } fn flush(&mut self) -> io::Result<()> { Ok(()) } fn write_all(&mut self, buf: &[u8]) -> io::Result<()> { let result = self.writer.write_application_data(buf); match result { Ok(()) => Ok(()), Err(err) => { let err = self.send_tls_alert(err); // FIXME more verbose io error Err(io::Error::new(io::ErrorKind::Other, SurugaError { desc: "TLS write error", cause: Some(Box::new(err)), })) } } } } // A replacement for the deprecated std::slice::bytes::copy_memory fn copy_memory(from: &[u8], mut to: &mut [u8]) -> usize { to.write(from).unwrap() } impl<R: Read, W: Write> Read for TlsClient<R, W> { // if ssl connection is failed, return `EndOfFile`. fn read(&mut self, buf: &mut [u8]) -> io::Result<usize> { let mut pos: usize = 0; let len = buf.len(); while pos < len { let remaining = len - pos; if self.buf.len() == 0 { let data = match self.reader.read_application_data() { Ok(data) => data, Err(_err) => { break; // FIXME: stop if EOF. otherwise raise error? } }; self.buf.extend(&data); } let selflen = self.buf.len(); let necessary = cmp::min(remaining, selflen); copy_memory(&self.buf[.. necessary], &mut buf[pos .. pos + necessary]); pos += necessary; self.buf = self.buf[necessary..].to_vec(); } Ok(pos) } }
random_line_split
client.rs
use std::io; use std::io::prelude::*; use std::net::TcpStream; use std::cmp; use rand::{Rng, OsRng}; use alert; use tls_result::{TlsResult, TlsError, TlsErrorKind}; use tls_result::TlsErrorKind::{UnexpectedMessage, InternalError, DecryptError, IllegalParameter}; use util::{SurugaError, crypto_compare}; use cipher::{self, Aead}; use cipher::prf::Prf; use crypto::sha2::sha256; use tls_item::{TlsItem, DummyItem}; use handshake::{self, Handshake}; use tls::{TlsReader, TlsWriter, TLS_VERSION}; // handshake is done during construction. pub struct TlsClient<R: Read, W: Write> { pub reader: TlsReader<R>, pub writer: TlsWriter<W>, pub rng: OsRng, buf: Vec<u8>, } impl<R: Read, W: Write> TlsClient<R, W> { pub fn new(reader: R, writer: W, rng: OsRng) -> TlsResult<TlsClient<R, W>> { let mut client = TlsClient { reader: TlsReader::new(reader), writer: TlsWriter::new(writer), rng: rng, buf: Vec::new(), }; // handshake failed. send alert if necessary match client.handshake() { Ok(()) => {} Err(err) => return Err(client.send_tls_alert(err)), } Ok(client) } #[inline] pub fn reader(&mut self) -> &mut R { self.reader.get_mut() } #[inline] pub fn writer(&mut self) -> &mut W { self.writer.get_mut() } // this does not send alert when error occurs fn handshake(&mut self) -> TlsResult<()> { // expect specific HandshakeMessage. otherwise return Err macro_rules! expect { ($var:ident) => ({ match try!(self.reader.read_handshake()) { handshake::Handshake::$var(data) => data, _ => return tls_err!(UnexpectedMessage, "unexpected handshake message found"), } }) } let cli_random = { let mut random_bytes = [0u8; 32]; self.rng.fill_bytes(&mut random_bytes); random_bytes.to_vec() }; let random = try!(handshake::Random::new(cli_random.clone())); // the only cipher we currently support let cipher_suite = cipher::CipherSuite::TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256; let curve_list = vec!(handshake::NamedCurve::secp256r1); let curve_list = try!(handshake::Extension::new_elliptic_curve_list(curve_list)); let format_list = vec!(handshake::ECPointFormat::uncompressed); let format_list = try!(handshake::Extension::new_ec_point_formats(format_list)); let extensions = vec!(curve_list, format_list); let client_hello = try!(Handshake::new_client_hello(random, cipher_suite, extensions)); try!(self.writer.write_handshake(&client_hello)); let server_hello_data = expect!(server_hello); { let server_major = server_hello_data.server_version.major; let server_minor = server_hello_data.server_version.minor; if (server_major, server_minor) != TLS_VERSION { return tls_err!(IllegalParameter, "wrong server version: {} {}", server_major, server_minor); } if server_hello_data.cipher_suite != cipher_suite { return tls_err!(IllegalParameter, "cipher suite mismatch: found {:?}", server_hello_data.cipher_suite); } if server_hello_data.compression_method != handshake::CompressionMethod::null { return tls_err!(IllegalParameter, "compression method mismatch"); } // FIXME: check if server sent unknown extension // it is currently done by just not understanding any extensions // other than we used. } // we always expect certificate. let certificate_list = expect!(certificate); // TODO: cert validation not implemented yet // we always use server key exchange let server_key_ex_data = expect!(server_key_exchange); let kex = cipher_suite.new_kex(); let (key_data, pre_master_secret) = try!(kex.compute_keys(&server_key_ex_data, &mut self.rng)); expect!(server_hello_done); let client_key_exchange = try!(Handshake::new_client_key_exchange(key_data)); try!(self.writer.write_handshake(&client_key_exchange)); try!(self.writer.write_change_cipher_spec()); // SECRET let master_secret = { let mut label_seed = b"master secret".to_vec(); label_seed.extend(&cli_random); label_seed.extend(&server_hello_data.random[..]); let mut prf = Prf::new(pre_master_secret, label_seed); prf.get_bytes(48) }; let aead = cipher_suite.new_aead(); // SECRET let read_key = { let mut label_seed = b"key expansion".to_vec(); label_seed.extend(&server_hello_data.random[..]); label_seed.extend(&cli_random); let mut prf = Prf::new(master_secret.clone(), label_seed); // mac_key is not used in AEAD configuration. let enc_key_length = aead.key_size(); let write_key = prf.get_bytes(enc_key_length); let encryptor = aead.new_encryptor(write_key); self.writer.set_encryptor(encryptor); // this will be set after receiving ChangeCipherSpec. let read_key = prf.get_bytes(enc_key_length); // chacha20-poly1305 does not use iv. read_key }; // FIXME we should get "raw" packet data and hash them incrementally let msgs = { let mut msgs = Vec::new(); try!(client_hello.tls_write(&mut msgs)); try!(Handshake::server_hello(server_hello_data).tls_write(&mut msgs)); try!(Handshake::certificate(certificate_list).tls_write(&mut msgs)); try!(Handshake::server_key_exchange(server_key_ex_data).tls_write(&mut msgs)); try!(Handshake::server_hello_done(DummyItem).tls_write(&mut msgs)); try!(client_key_exchange.tls_write(&mut msgs)); msgs }; // this only verifies Handshake messages! what about others? // ApplicationData messages are not permitted until now. // ChangeCipherSpec messages are only permitted after ClinetKeyExchange. // Alert messages can be problematic - they are not verified and // can be broken into several records. This leads to alert attack. // since we don't accept strange alerts, all "normal" alert messages are // treated as error, so now we can assert that we haven't received alerts. let verify_hash = sha256(&msgs); let client_verify_data = { let finished_label = b"client finished"; let mut label_seed = finished_label.to_vec(); label_seed.extend(&verify_hash); let mut prf = Prf::new(master_secret.clone(), label_seed); prf.get_bytes(cipher_suite.verify_data_len()) }; let finished = try!(Handshake::new_finished(client_verify_data)); try!(self.writer.write_handshake(&finished)); // Although client->server is encrypted, server->client isn't yet. // server may send either ChangeCipherSpec or Alert. try!(self.reader.read_change_cipher_spec()); // from now server starts encryption. self.reader.set_decryptor(aead.new_decryptor(read_key)); let server_finished = expect!(finished); { let verify_hash = { // ideally we may save "raw" packet data.. let mut serv_msgs = Vec::new(); // FIXME: this should not throw "io error".. should throw "internal error" try!(Write::write_all(&mut serv_msgs, &msgs)); try!(finished.tls_write(&mut serv_msgs)); let verify_hash = sha256(&serv_msgs); verify_hash }; let server_verify_data = { let finished_label = b"server finished"; let mut label_seed = finished_label.to_vec(); label_seed.extend(&verify_hash); let mut prf = Prf::new(master_secret, label_seed); prf.get_bytes(cipher_suite.verify_data_len()) }; let verify_ok = crypto_compare(&server_finished, &server_verify_data); if !verify_ok { return tls_err!(DecryptError, "server sent wrong verify data"); } } Ok(()) } pub fn close(&mut self) -> TlsResult<()> { let alert_data = alert::Alert { level: alert::AlertLevel::fatal, description: alert::AlertDescription::close_notify, }; try!(self.writer.write_alert(&alert_data)); Ok(()) } // send fatal alert and return error // (it may be different to `err`, because writing alert can fail) pub fn send_tls_alert(&mut self, err: TlsError) -> TlsError { match err.kind { TlsErrorKind::IoFailure => return err, _ => { let alert = alert::Alert::from_tls_err(&err); let result = self.writer.write_alert(&alert); match result { Ok(()) => return err, Err(err) => return err, } } } } } impl TlsClient<TcpStream, TcpStream> { pub fn from_tcp(stream: TcpStream) -> TlsResult<TlsClient<TcpStream, TcpStream>> { let rng = match OsRng::new() { Ok(rng) => rng, Err(..) => return tls_err!(InternalError, "failed to create OsRng"), }; let reader = try!(stream.try_clone()); let writer = stream; TlsClient::new(reader, writer, rng) } } impl<R: Read, W: Write> Write for TlsClient<R, W> { // this either writes all or fails. fn write(&mut self, buf: &[u8]) -> io::Result<usize> { try!(self.write_all(buf)); Ok(buf.len()) } fn flush(&mut self) -> io::Result<()> { Ok(()) } fn write_all(&mut self, buf: &[u8]) -> io::Result<()> { let result = self.writer.write_application_data(buf); match result { Ok(()) => Ok(()), Err(err) => { let err = self.send_tls_alert(err); // FIXME more verbose io error Err(io::Error::new(io::ErrorKind::Other, SurugaError { desc: "TLS write error", cause: Some(Box::new(err)), })) } } } } // A replacement for the deprecated std::slice::bytes::copy_memory fn copy_memory(from: &[u8], mut to: &mut [u8]) -> usize
impl<R: Read, W: Write> Read for TlsClient<R, W> { // if ssl connection is failed, return `EndOfFile`. fn read(&mut self, buf: &mut [u8]) -> io::Result<usize> { let mut pos: usize = 0; let len = buf.len(); while pos < len { let remaining = len - pos; if self.buf.len() == 0 { let data = match self.reader.read_application_data() { Ok(data) => data, Err(_err) => { break; // FIXME: stop if EOF. otherwise raise error? } }; self.buf.extend(&data); } let selflen = self.buf.len(); let necessary = cmp::min(remaining, selflen); copy_memory(&self.buf[.. necessary], &mut buf[pos .. pos + necessary]); pos += necessary; self.buf = self.buf[necessary..].to_vec(); } Ok(pos) } }
{ to.write(from).unwrap() }
identifier_body
analysis.py
#Mkhanyisi Gamedze # CS 251 - Spring 2017 # project 2 # global fields import sys import numpy as np import csv import data import scipy.stats import scipy.cluster.vq as vq import random # functions def range_(headers, data): """ Takes in a list of column headers and the Data object and returns a list of 2-element lists with the minimum and maximum values for each column. The function is required to work only on numeric data types.""" column_matrix=data.get_data(headers).getT() # get columns as rows, as this makes analysis much easier by just perfoming operations on column list directly if column_matrix==[]: print "wrong headers, not present in data Object" return [] column_max=column_matrix.max(1) column_min=column_matrix.min(1) final=np.concatenate((column_min, column_max), axis=1) rng=final.tolist() return rng def mean(headers, data): """ Takes in a list of column headers and the Data object and returns a list of the mean values for each column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.mean(0) return mean_values def stdev(headers, data): """stdev - Takes in a list of column headers and the Data object and returns a list of the standard deviation for each specified column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.std(0) std_values=mean_values.tolist() return std_values def normalize_columns_separately(headers, data): """Takes in a list of column headers and the Data object and returns a matrix with each column normalized so its minimum value is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) column_max=column_matrix.max(1) column_min=column_matrix.min(1) range=column_max-column_min nomalized=(column_matrix-column_min)/range return nomalized def normalize_columns_together(headers, data): """ Takes in a list of column headers and the Data object and returns a matrix with each entry normalized so that the minimum value (of all the data in this set of columns) is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) max=column_matrix.max() print "The maximum: ", max min=column_matrix.min() print "The minimum: ", min range=max-min print "range: ", range column_matrix=column_matrix-min normalized=column_matrix/range return normalized def sort(headers, data): # extension """ Return the numeric matrices with sorted columns """ column_matrix=data.get_data(headers) # get raw matrix data for numeric values print "\n before sorting \n " print column_matrix column_matrix=column_matrix.tolist() column_array=np.asarray(column_matrix) column_array.sort(axis=0) print "\n \n done sorting here is your matrix \n" return column_array def normalize_sort(headers, data): # extension column_matrix=data.get_data(headers) max=column_matrix.max() min=column_matrix.min() range=max-min column_matrix=column_matrix-min normalized=column_matrix/range print "\n before sorting \n ", normalized, "\n \n " normalized.sort(axis=0) print "\n after sorting \n " return normalized def linear_regression(d, ind, dep): """ takes in data object and then creates a linear regression using the dependant variable""" y=d.get_data([dep]) print "y :",y A=d.get_data(ind) print "A :",A ones = np.asmatrix(np.ones( (A.shape[0]) )).transpose() A=np.concatenate((A, ones), axis=1) print "concatenated A :",A AAinv=np.linalg.inv( np.dot(A.transpose(), A)) print "AAinv: \n",AAinv """ print "A :",A print "y: ",y print "AAinv: ",AAinv""" print "shape A: ",A.shape print "shape y :", y.shape x=np.linalg.lstsq(A,y) print "x :\n",x b=x[0] print "\n b : \n",b N=len(y) print "N : \n",N C=len(b) print "C : ",C df_e=N-C df_r=C-1 error=y - np.dot(A, b) print "error: ",error sse=np.dot(error.transpose(), error) / df_e print "sse :",sse stderr=np.sqrt( np.diagonal( sse[0, 0] * AAinv ) ) print "stderr: ",stderr t = b.transpose() / stderr print "t :", t p=2*(1 - scipy.stats.t.cdf(abs(t), df_e)) print "p: ",p r2=1 - error.var() / y.var() print "R^2 :",r2, "\n \n \n \n*************************************" return [b,sse,r2,t,p] # This version uses SVD def pca(d, headers, normalize=True): if normalize==True: A=normalize_columns_separately(headers, d) else: A=d.get_data(headers) m=mean(headers, d) D=A-m #calculate eigenvectors and eigenvalues U,S,V=np.linalg.svd(D,full_matrices=False) index=0 #get the eigenvalues using the number of degress of freedom for d in S: e=(d*d)/(U.shape[0]-1) S[index]=e index=index+1 #the projected data pdata=np.dot(V,(D.T)) pdata=pdata.T pcad=data.PCAData(headers,pdata,S,V,m) return pcad def kmeans_numpy( d, headers, K, whiten = True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes, and representation error. ''' A=d.get_data(headers) W=vq.whiten(A) codebook, bookerror=vq.kmeans(W,K) codes, error=vq.vq(W, codebook) return codebook, codes, error def kmeans_init(d, K, catergories=[]) : #return numpy matrix with K rows of the data print "type K :",type(K) r=np.matrix(np.zeros(K)) c=np.matrix(np.zeros(d.shape[1])) r=r.transpose() retval=np.dot(r,c) retval=retval print "shape retval :",retval.shape # If no categories are given, a simple way to select the means is to randomly choose K data points if len(catergories)==0: values=[] # values to be selected list h=d.shape[0] if K>h: print "The value of K is too high" return None #pick random rows while K>=0: val=random.randint(0,h-1) #pick random value while val in values: #avoid duplicates, reselect if duplicate found val=random.randint(0,h-1) values.append(val) #add random index to values #retval[K,:]=d[val,:] K-=1 # given an Nx1 matrix of categories/labels, then compute the mean values of each category # and return those as the initial set of means else: print "here" unique,labels=np.unique(catergories.tolist(),return_inverse=True) means=np.zeros((K,d.shape[1])) for i in range(len(unique)): #for all unique values means [i,:]=np.mean(d[labels==i,:],axis=0) #calculate means using categories retval=means return retval def kmeans_classify(d, means): ID=[] # list of ID values mindistances=[] # minimum distances algorithm for dpoint in d: distances=[] # distances from each mean for mean in means: # compute distance of each mean, using the distance formula differences=dpoint-mean squares=np.square(differences) sums=np.sum(squares) distance=np.sqrt(sums) distances.append(distance) # add the distance to the distances list ID.append(np.argmin(distances)) mindistances.append(distances[np.argmin(distances)]) retval=[] retval.append(ID) retval.append(mindistances) return np.matrix(ID).transpose(), np.matrix(mindistances).transpose() # return a list of the ID values and the distances """ def kmeans_algorithm(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): print "j :",j print "A[j,:] :",A[j,:] print "codes ", codes print "codes[j,0] :",codes[j,0] newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 print "newmeans type: ",type(newmeans) # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) print "result: ",means, codes, errors # return the means, codes, and errors return (means, codes, errors) """ def
(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) # return the means, codes, and errors return (means, codes, errors) def kmeans(d, headers, K , categories = [] , whiten=True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes and representation errors. If given an Nx1 matrix of categories, it uses the category labels to calculate the initial cluster means. ''' A=d.get_data(headers) #print "type A :",type(A) if whiten==True: W=vq.whiten(A) W=np.asmatrix(W) else: W=A #print "type W :",type(W) #print "K2 :",K codebook=kmeans_init(W,K, categories) #codebook=np.flipud(codebook) codebook, codes, errors = kmeans_algorithm(W, codebook) codes=codes return codebook, codes, errors def main(filename): d = data.Data( filename ) headers = d.get_headers() print " \n \n ........... Testing your numeric data .............." """ print " \n \n \n Your headers:" print d.get_headers() print "\n The types for your columns: you have ",d.get_num_columns(), " of them" print d.get_type() print " \n The raw matrix of data" print d.printMatrix() headers.pop() print "\n \n get data for these columns: ",headers print d.get_data(headers) print "\n \n Range__ :" print "Your range values by column: ",range(headers, d) print "\n \n The column means are: " print "Computed column mean: ", mean(headers, d) print " \n \n standard deviations:" print "Deviations: ", stdev(headers,d) print "\n \n Normalized columns" print normalize_columns_separately(headers, d) print "\n \n \n \nNormalizing the whole matrix" print normalize_columns_together(headers, d) print "############## Extensions ################" d.add_column("new data", "numeric") print "\n \n \n" d.add_datacolumn("new data", "numeric", [1, 2,3]) print " \n \n ********* Extension 2- sorting ************** \n " print "Sorting a matrix \n \n " print sort(headers, d) print "*** \n 2b) normalize and sort \n" print " \n \n \n ",normalize_sort(headers, d) """ print "*********************************************" print "linear regression" headers = d.get_headers() print "headers: ", headers print linear_regression(d, [headers[0],headers[1]], headers[2]) if __name__ == '__main__': if len(sys.argv) < 2: print "Usage: python %s <csv_filename>" % sys.argv[0] print " where <csv_filename> specifies a csv file" exit() main( sys.argv[1] )
kmeans_algorithm
identifier_name
analysis.py
#Mkhanyisi Gamedze # CS 251 - Spring 2017 # project 2 # global fields import sys import numpy as np import csv import data import scipy.stats import scipy.cluster.vq as vq import random # functions def range_(headers, data): """ Takes in a list of column headers and the Data object and returns a list of 2-element lists with the minimum and maximum values for each column. The function is required to work only on numeric data types.""" column_matrix=data.get_data(headers).getT() # get columns as rows, as this makes analysis much easier by just perfoming operations on column list directly if column_matrix==[]:
column_max=column_matrix.max(1) column_min=column_matrix.min(1) final=np.concatenate((column_min, column_max), axis=1) rng=final.tolist() return rng def mean(headers, data): """ Takes in a list of column headers and the Data object and returns a list of the mean values for each column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.mean(0) return mean_values def stdev(headers, data): """stdev - Takes in a list of column headers and the Data object and returns a list of the standard deviation for each specified column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.std(0) std_values=mean_values.tolist() return std_values def normalize_columns_separately(headers, data): """Takes in a list of column headers and the Data object and returns a matrix with each column normalized so its minimum value is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) column_max=column_matrix.max(1) column_min=column_matrix.min(1) range=column_max-column_min nomalized=(column_matrix-column_min)/range return nomalized def normalize_columns_together(headers, data): """ Takes in a list of column headers and the Data object and returns a matrix with each entry normalized so that the minimum value (of all the data in this set of columns) is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) max=column_matrix.max() print "The maximum: ", max min=column_matrix.min() print "The minimum: ", min range=max-min print "range: ", range column_matrix=column_matrix-min normalized=column_matrix/range return normalized def sort(headers, data): # extension """ Return the numeric matrices with sorted columns """ column_matrix=data.get_data(headers) # get raw matrix data for numeric values print "\n before sorting \n " print column_matrix column_matrix=column_matrix.tolist() column_array=np.asarray(column_matrix) column_array.sort(axis=0) print "\n \n done sorting here is your matrix \n" return column_array def normalize_sort(headers, data): # extension column_matrix=data.get_data(headers) max=column_matrix.max() min=column_matrix.min() range=max-min column_matrix=column_matrix-min normalized=column_matrix/range print "\n before sorting \n ", normalized, "\n \n " normalized.sort(axis=0) print "\n after sorting \n " return normalized def linear_regression(d, ind, dep): """ takes in data object and then creates a linear regression using the dependant variable""" y=d.get_data([dep]) print "y :",y A=d.get_data(ind) print "A :",A ones = np.asmatrix(np.ones( (A.shape[0]) )).transpose() A=np.concatenate((A, ones), axis=1) print "concatenated A :",A AAinv=np.linalg.inv( np.dot(A.transpose(), A)) print "AAinv: \n",AAinv """ print "A :",A print "y: ",y print "AAinv: ",AAinv""" print "shape A: ",A.shape print "shape y :", y.shape x=np.linalg.lstsq(A,y) print "x :\n",x b=x[0] print "\n b : \n",b N=len(y) print "N : \n",N C=len(b) print "C : ",C df_e=N-C df_r=C-1 error=y - np.dot(A, b) print "error: ",error sse=np.dot(error.transpose(), error) / df_e print "sse :",sse stderr=np.sqrt( np.diagonal( sse[0, 0] * AAinv ) ) print "stderr: ",stderr t = b.transpose() / stderr print "t :", t p=2*(1 - scipy.stats.t.cdf(abs(t), df_e)) print "p: ",p r2=1 - error.var() / y.var() print "R^2 :",r2, "\n \n \n \n*************************************" return [b,sse,r2,t,p] # This version uses SVD def pca(d, headers, normalize=True): if normalize==True: A=normalize_columns_separately(headers, d) else: A=d.get_data(headers) m=mean(headers, d) D=A-m #calculate eigenvectors and eigenvalues U,S,V=np.linalg.svd(D,full_matrices=False) index=0 #get the eigenvalues using the number of degress of freedom for d in S: e=(d*d)/(U.shape[0]-1) S[index]=e index=index+1 #the projected data pdata=np.dot(V,(D.T)) pdata=pdata.T pcad=data.PCAData(headers,pdata,S,V,m) return pcad def kmeans_numpy( d, headers, K, whiten = True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes, and representation error. ''' A=d.get_data(headers) W=vq.whiten(A) codebook, bookerror=vq.kmeans(W,K) codes, error=vq.vq(W, codebook) return codebook, codes, error def kmeans_init(d, K, catergories=[]) : #return numpy matrix with K rows of the data print "type K :",type(K) r=np.matrix(np.zeros(K)) c=np.matrix(np.zeros(d.shape[1])) r=r.transpose() retval=np.dot(r,c) retval=retval print "shape retval :",retval.shape # If no categories are given, a simple way to select the means is to randomly choose K data points if len(catergories)==0: values=[] # values to be selected list h=d.shape[0] if K>h: print "The value of K is too high" return None #pick random rows while K>=0: val=random.randint(0,h-1) #pick random value while val in values: #avoid duplicates, reselect if duplicate found val=random.randint(0,h-1) values.append(val) #add random index to values #retval[K,:]=d[val,:] K-=1 # given an Nx1 matrix of categories/labels, then compute the mean values of each category # and return those as the initial set of means else: print "here" unique,labels=np.unique(catergories.tolist(),return_inverse=True) means=np.zeros((K,d.shape[1])) for i in range(len(unique)): #for all unique values means [i,:]=np.mean(d[labels==i,:],axis=0) #calculate means using categories retval=means return retval def kmeans_classify(d, means): ID=[] # list of ID values mindistances=[] # minimum distances algorithm for dpoint in d: distances=[] # distances from each mean for mean in means: # compute distance of each mean, using the distance formula differences=dpoint-mean squares=np.square(differences) sums=np.sum(squares) distance=np.sqrt(sums) distances.append(distance) # add the distance to the distances list ID.append(np.argmin(distances)) mindistances.append(distances[np.argmin(distances)]) retval=[] retval.append(ID) retval.append(mindistances) return np.matrix(ID).transpose(), np.matrix(mindistances).transpose() # return a list of the ID values and the distances """ def kmeans_algorithm(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): print "j :",j print "A[j,:] :",A[j,:] print "codes ", codes print "codes[j,0] :",codes[j,0] newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 print "newmeans type: ",type(newmeans) # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) print "result: ",means, codes, errors # return the means, codes, and errors return (means, codes, errors) """ def kmeans_algorithm(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) # return the means, codes, and errors return (means, codes, errors) def kmeans(d, headers, K , categories = [] , whiten=True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes and representation errors. If given an Nx1 matrix of categories, it uses the category labels to calculate the initial cluster means. ''' A=d.get_data(headers) #print "type A :",type(A) if whiten==True: W=vq.whiten(A) W=np.asmatrix(W) else: W=A #print "type W :",type(W) #print "K2 :",K codebook=kmeans_init(W,K, categories) #codebook=np.flipud(codebook) codebook, codes, errors = kmeans_algorithm(W, codebook) codes=codes return codebook, codes, errors def main(filename): d = data.Data( filename ) headers = d.get_headers() print " \n \n ........... Testing your numeric data .............." """ print " \n \n \n Your headers:" print d.get_headers() print "\n The types for your columns: you have ",d.get_num_columns(), " of them" print d.get_type() print " \n The raw matrix of data" print d.printMatrix() headers.pop() print "\n \n get data for these columns: ",headers print d.get_data(headers) print "\n \n Range__ :" print "Your range values by column: ",range(headers, d) print "\n \n The column means are: " print "Computed column mean: ", mean(headers, d) print " \n \n standard deviations:" print "Deviations: ", stdev(headers,d) print "\n \n Normalized columns" print normalize_columns_separately(headers, d) print "\n \n \n \nNormalizing the whole matrix" print normalize_columns_together(headers, d) print "############## Extensions ################" d.add_column("new data", "numeric") print "\n \n \n" d.add_datacolumn("new data", "numeric", [1, 2,3]) print " \n \n ********* Extension 2- sorting ************** \n " print "Sorting a matrix \n \n " print sort(headers, d) print "*** \n 2b) normalize and sort \n" print " \n \n \n ",normalize_sort(headers, d) """ print "*********************************************" print "linear regression" headers = d.get_headers() print "headers: ", headers print linear_regression(d, [headers[0],headers[1]], headers[2]) if __name__ == '__main__': if len(sys.argv) < 2: print "Usage: python %s <csv_filename>" % sys.argv[0] print " where <csv_filename> specifies a csv file" exit() main( sys.argv[1] )
print "wrong headers, not present in data Object" return []
conditional_block
analysis.py
#Mkhanyisi Gamedze # CS 251 - Spring 2017 # project 2 # global fields import sys import numpy as np import csv import data import scipy.stats import scipy.cluster.vq as vq import random # functions def range_(headers, data): """ Takes in a list of column headers and the Data object and returns a list of 2-element lists with the minimum and maximum values for each column. The function is required to work only on numeric data types.""" column_matrix=data.get_data(headers).getT() # get columns as rows, as this makes analysis much easier by just perfoming operations on column list directly if column_matrix==[]: print "wrong headers, not present in data Object" return [] column_max=column_matrix.max(1) column_min=column_matrix.min(1) final=np.concatenate((column_min, column_max), axis=1) rng=final.tolist() return rng def mean(headers, data): """ Takes in a list of column headers and the Data object and returns a list of the mean values for each column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.mean(0) return mean_values def stdev(headers, data): """stdev - Takes in a list of column headers and the Data object and returns a list of the standard deviation for each specified column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.std(0) std_values=mean_values.tolist() return std_values def normalize_columns_separately(headers, data): """Takes in a list of column headers and the Data object and returns a matrix with each column normalized so its minimum value is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) column_max=column_matrix.max(1) column_min=column_matrix.min(1) range=column_max-column_min nomalized=(column_matrix-column_min)/range return nomalized def normalize_columns_together(headers, data): """ Takes in a list of column headers and the Data object and returns a matrix with each entry normalized so that the minimum value (of all the data in this set of columns) is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) max=column_matrix.max() print "The maximum: ", max min=column_matrix.min() print "The minimum: ", min range=max-min print "range: ", range column_matrix=column_matrix-min normalized=column_matrix/range return normalized def sort(headers, data): # extension """ Return the numeric matrices with sorted columns """ column_matrix=data.get_data(headers) # get raw matrix data for numeric values print "\n before sorting \n " print column_matrix column_matrix=column_matrix.tolist() column_array=np.asarray(column_matrix) column_array.sort(axis=0) print "\n \n done sorting here is your matrix \n" return column_array def normalize_sort(headers, data): # extension column_matrix=data.get_data(headers) max=column_matrix.max() min=column_matrix.min() range=max-min column_matrix=column_matrix-min normalized=column_matrix/range print "\n before sorting \n ", normalized, "\n \n " normalized.sort(axis=0) print "\n after sorting \n " return normalized def linear_regression(d, ind, dep): """ takes in data object and then creates a linear regression using the dependant variable""" y=d.get_data([dep]) print "y :",y A=d.get_data(ind) print "A :",A ones = np.asmatrix(np.ones( (A.shape[0]) )).transpose() A=np.concatenate((A, ones), axis=1) print "concatenated A :",A AAinv=np.linalg.inv( np.dot(A.transpose(), A)) print "AAinv: \n",AAinv """ print "A :",A print "y: ",y print "AAinv: ",AAinv""" print "shape A: ",A.shape print "shape y :", y.shape x=np.linalg.lstsq(A,y) print "x :\n",x b=x[0] print "\n b : \n",b N=len(y) print "N : \n",N C=len(b) print "C : ",C df_e=N-C df_r=C-1 error=y - np.dot(A, b) print "error: ",error sse=np.dot(error.transpose(), error) / df_e print "sse :",sse stderr=np.sqrt( np.diagonal( sse[0, 0] * AAinv ) ) print "stderr: ",stderr t = b.transpose() / stderr print "t :", t p=2*(1 - scipy.stats.t.cdf(abs(t), df_e)) print "p: ",p r2=1 - error.var() / y.var() print "R^2 :",r2, "\n \n \n \n*************************************" return [b,sse,r2,t,p] # This version uses SVD def pca(d, headers, normalize=True): if normalize==True: A=normalize_columns_separately(headers, d) else: A=d.get_data(headers) m=mean(headers, d) D=A-m #calculate eigenvectors and eigenvalues U,S,V=np.linalg.svd(D,full_matrices=False) index=0 #get the eigenvalues using the number of degress of freedom for d in S:
index=index+1 #the projected data pdata=np.dot(V,(D.T)) pdata=pdata.T pcad=data.PCAData(headers,pdata,S,V,m) return pcad def kmeans_numpy( d, headers, K, whiten = True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes, and representation error. ''' A=d.get_data(headers) W=vq.whiten(A) codebook, bookerror=vq.kmeans(W,K) codes, error=vq.vq(W, codebook) return codebook, codes, error def kmeans_init(d, K, catergories=[]) : #return numpy matrix with K rows of the data print "type K :",type(K) r=np.matrix(np.zeros(K)) c=np.matrix(np.zeros(d.shape[1])) r=r.transpose() retval=np.dot(r,c) retval=retval print "shape retval :",retval.shape # If no categories are given, a simple way to select the means is to randomly choose K data points if len(catergories)==0: values=[] # values to be selected list h=d.shape[0] if K>h: print "The value of K is too high" return None #pick random rows while K>=0: val=random.randint(0,h-1) #pick random value while val in values: #avoid duplicates, reselect if duplicate found val=random.randint(0,h-1) values.append(val) #add random index to values #retval[K,:]=d[val,:] K-=1 # given an Nx1 matrix of categories/labels, then compute the mean values of each category # and return those as the initial set of means else: print "here" unique,labels=np.unique(catergories.tolist(),return_inverse=True) means=np.zeros((K,d.shape[1])) for i in range(len(unique)): #for all unique values means [i,:]=np.mean(d[labels==i,:],axis=0) #calculate means using categories retval=means return retval def kmeans_classify(d, means): ID=[] # list of ID values mindistances=[] # minimum distances algorithm for dpoint in d: distances=[] # distances from each mean for mean in means: # compute distance of each mean, using the distance formula differences=dpoint-mean squares=np.square(differences) sums=np.sum(squares) distance=np.sqrt(sums) distances.append(distance) # add the distance to the distances list ID.append(np.argmin(distances)) mindistances.append(distances[np.argmin(distances)]) retval=[] retval.append(ID) retval.append(mindistances) return np.matrix(ID).transpose(), np.matrix(mindistances).transpose() # return a list of the ID values and the distances """ def kmeans_algorithm(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): print "j :",j print "A[j,:] :",A[j,:] print "codes ", codes print "codes[j,0] :",codes[j,0] newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 print "newmeans type: ",type(newmeans) # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) print "result: ",means, codes, errors # return the means, codes, and errors return (means, codes, errors) """ def kmeans_algorithm(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) # return the means, codes, and errors return (means, codes, errors) def kmeans(d, headers, K , categories = [] , whiten=True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes and representation errors. If given an Nx1 matrix of categories, it uses the category labels to calculate the initial cluster means. ''' A=d.get_data(headers) #print "type A :",type(A) if whiten==True: W=vq.whiten(A) W=np.asmatrix(W) else: W=A #print "type W :",type(W) #print "K2 :",K codebook=kmeans_init(W,K, categories) #codebook=np.flipud(codebook) codebook, codes, errors = kmeans_algorithm(W, codebook) codes=codes return codebook, codes, errors def main(filename): d = data.Data( filename ) headers = d.get_headers() print " \n \n ........... Testing your numeric data .............." """ print " \n \n \n Your headers:" print d.get_headers() print "\n The types for your columns: you have ",d.get_num_columns(), " of them" print d.get_type() print " \n The raw matrix of data" print d.printMatrix() headers.pop() print "\n \n get data for these columns: ",headers print d.get_data(headers) print "\n \n Range__ :" print "Your range values by column: ",range(headers, d) print "\n \n The column means are: " print "Computed column mean: ", mean(headers, d) print " \n \n standard deviations:" print "Deviations: ", stdev(headers,d) print "\n \n Normalized columns" print normalize_columns_separately(headers, d) print "\n \n \n \nNormalizing the whole matrix" print normalize_columns_together(headers, d) print "############## Extensions ################" d.add_column("new data", "numeric") print "\n \n \n" d.add_datacolumn("new data", "numeric", [1, 2,3]) print " \n \n ********* Extension 2- sorting ************** \n " print "Sorting a matrix \n \n " print sort(headers, d) print "*** \n 2b) normalize and sort \n" print " \n \n \n ",normalize_sort(headers, d) """ print "*********************************************" print "linear regression" headers = d.get_headers() print "headers: ", headers print linear_regression(d, [headers[0],headers[1]], headers[2]) if __name__ == '__main__': if len(sys.argv) < 2: print "Usage: python %s <csv_filename>" % sys.argv[0] print " where <csv_filename> specifies a csv file" exit() main( sys.argv[1] )
e=(d*d)/(U.shape[0]-1) S[index]=e
random_line_split
analysis.py
#Mkhanyisi Gamedze # CS 251 - Spring 2017 # project 2 # global fields import sys import numpy as np import csv import data import scipy.stats import scipy.cluster.vq as vq import random # functions def range_(headers, data): """ Takes in a list of column headers and the Data object and returns a list of 2-element lists with the minimum and maximum values for each column. The function is required to work only on numeric data types.""" column_matrix=data.get_data(headers).getT() # get columns as rows, as this makes analysis much easier by just perfoming operations on column list directly if column_matrix==[]: print "wrong headers, not present in data Object" return [] column_max=column_matrix.max(1) column_min=column_matrix.min(1) final=np.concatenate((column_min, column_max), axis=1) rng=final.tolist() return rng def mean(headers, data): """ Takes in a list of column headers and the Data object and returns a list of the mean values for each column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.mean(0) return mean_values def stdev(headers, data): """stdev - Takes in a list of column headers and the Data object and returns a list of the standard deviation for each specified column. Use the built-in numpy functions to execute this calculation.""" column_matrix=data.get_data(headers) mean_values=column_matrix.std(0) std_values=mean_values.tolist() return std_values def normalize_columns_separately(headers, data): """Takes in a list of column headers and the Data object and returns a matrix with each column normalized so its minimum value is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) column_max=column_matrix.max(1) column_min=column_matrix.min(1) range=column_max-column_min nomalized=(column_matrix-column_min)/range return nomalized def normalize_columns_together(headers, data): """ Takes in a list of column headers and the Data object and returns a matrix with each entry normalized so that the minimum value (of all the data in this set of columns) is mapped to zero and its maximum value is mapped to 1.""" column_matrix=data.get_data(headers) max=column_matrix.max() print "The maximum: ", max min=column_matrix.min() print "The minimum: ", min range=max-min print "range: ", range column_matrix=column_matrix-min normalized=column_matrix/range return normalized def sort(headers, data): # extension """ Return the numeric matrices with sorted columns """ column_matrix=data.get_data(headers) # get raw matrix data for numeric values print "\n before sorting \n " print column_matrix column_matrix=column_matrix.tolist() column_array=np.asarray(column_matrix) column_array.sort(axis=0) print "\n \n done sorting here is your matrix \n" return column_array def normalize_sort(headers, data): # extension column_matrix=data.get_data(headers) max=column_matrix.max() min=column_matrix.min() range=max-min column_matrix=column_matrix-min normalized=column_matrix/range print "\n before sorting \n ", normalized, "\n \n " normalized.sort(axis=0) print "\n after sorting \n " return normalized def linear_regression(d, ind, dep): """ takes in data object and then creates a linear regression using the dependant variable""" y=d.get_data([dep]) print "y :",y A=d.get_data(ind) print "A :",A ones = np.asmatrix(np.ones( (A.shape[0]) )).transpose() A=np.concatenate((A, ones), axis=1) print "concatenated A :",A AAinv=np.linalg.inv( np.dot(A.transpose(), A)) print "AAinv: \n",AAinv """ print "A :",A print "y: ",y print "AAinv: ",AAinv""" print "shape A: ",A.shape print "shape y :", y.shape x=np.linalg.lstsq(A,y) print "x :\n",x b=x[0] print "\n b : \n",b N=len(y) print "N : \n",N C=len(b) print "C : ",C df_e=N-C df_r=C-1 error=y - np.dot(A, b) print "error: ",error sse=np.dot(error.transpose(), error) / df_e print "sse :",sse stderr=np.sqrt( np.diagonal( sse[0, 0] * AAinv ) ) print "stderr: ",stderr t = b.transpose() / stderr print "t :", t p=2*(1 - scipy.stats.t.cdf(abs(t), df_e)) print "p: ",p r2=1 - error.var() / y.var() print "R^2 :",r2, "\n \n \n \n*************************************" return [b,sse,r2,t,p] # This version uses SVD def pca(d, headers, normalize=True): if normalize==True: A=normalize_columns_separately(headers, d) else: A=d.get_data(headers) m=mean(headers, d) D=A-m #calculate eigenvectors and eigenvalues U,S,V=np.linalg.svd(D,full_matrices=False) index=0 #get the eigenvalues using the number of degress of freedom for d in S: e=(d*d)/(U.shape[0]-1) S[index]=e index=index+1 #the projected data pdata=np.dot(V,(D.T)) pdata=pdata.T pcad=data.PCAData(headers,pdata,S,V,m) return pcad def kmeans_numpy( d, headers, K, whiten = True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes, and representation error. ''' A=d.get_data(headers) W=vq.whiten(A) codebook, bookerror=vq.kmeans(W,K) codes, error=vq.vq(W, codebook) return codebook, codes, error def kmeans_init(d, K, catergories=[]) : #return numpy matrix with K rows of the data print "type K :",type(K) r=np.matrix(np.zeros(K)) c=np.matrix(np.zeros(d.shape[1])) r=r.transpose() retval=np.dot(r,c) retval=retval print "shape retval :",retval.shape # If no categories are given, a simple way to select the means is to randomly choose K data points if len(catergories)==0: values=[] # values to be selected list h=d.shape[0] if K>h: print "The value of K is too high" return None #pick random rows while K>=0: val=random.randint(0,h-1) #pick random value while val in values: #avoid duplicates, reselect if duplicate found val=random.randint(0,h-1) values.append(val) #add random index to values #retval[K,:]=d[val,:] K-=1 # given an Nx1 matrix of categories/labels, then compute the mean values of each category # and return those as the initial set of means else: print "here" unique,labels=np.unique(catergories.tolist(),return_inverse=True) means=np.zeros((K,d.shape[1])) for i in range(len(unique)): #for all unique values means [i,:]=np.mean(d[labels==i,:],axis=0) #calculate means using categories retval=means return retval def kmeans_classify(d, means):
""" def kmeans_algorithm(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): print "j :",j print "A[j,:] :",A[j,:] print "codes ", codes print "codes[j,0] :",codes[j,0] newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 print "newmeans type: ",type(newmeans) # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) print "result: ",means, codes, errors # return the means, codes, and errors return (means, codes, errors) """ def kmeans_algorithm(A, means): # set up some useful constants MIN_CHANGE = 1e-7 MAX_ITERATIONS = 100 D = means.shape[1] K = means.shape[0] N = A.shape[0] # iterate no more than MAX_ITERATIONS for i in range(MAX_ITERATIONS): # calculate the codes codes, errors = kmeans_classify( A, means ) # calculate the new means newmeans = np.matrix(np.zeros_like( means )) counts = np.zeros( (K, 1) ) for j in range(N): newmeans[codes[j,0],:] += A[j,:] counts[codes[j,0],0] += 1.0 # finish calculating the means, taking into account possible zero counts for j in range(K): if counts[j,0] > 0.0: newmeans[j,:] /= counts[j, 0] else: newmeans[j,:] = A[random.randint(0,A.shape[0]-1),:] #randint is inclusive # test if the change is small enough diff = np.sum(np.square(means - newmeans)) means = newmeans if diff < MIN_CHANGE: break # call classify with the final means codes, errors = kmeans_classify( A, means ) # return the means, codes, and errors return (means, codes, errors) def kmeans(d, headers, K , categories = [] , whiten=True): '''Takes in a Data object, a set of headers, and the number of clusters to create Computes and returns the codebook, codes and representation errors. If given an Nx1 matrix of categories, it uses the category labels to calculate the initial cluster means. ''' A=d.get_data(headers) #print "type A :",type(A) if whiten==True: W=vq.whiten(A) W=np.asmatrix(W) else: W=A #print "type W :",type(W) #print "K2 :",K codebook=kmeans_init(W,K, categories) #codebook=np.flipud(codebook) codebook, codes, errors = kmeans_algorithm(W, codebook) codes=codes return codebook, codes, errors def main(filename): d = data.Data( filename ) headers = d.get_headers() print " \n \n ........... Testing your numeric data .............." """ print " \n \n \n Your headers:" print d.get_headers() print "\n The types for your columns: you have ",d.get_num_columns(), " of them" print d.get_type() print " \n The raw matrix of data" print d.printMatrix() headers.pop() print "\n \n get data for these columns: ",headers print d.get_data(headers) print "\n \n Range__ :" print "Your range values by column: ",range(headers, d) print "\n \n The column means are: " print "Computed column mean: ", mean(headers, d) print " \n \n standard deviations:" print "Deviations: ", stdev(headers,d) print "\n \n Normalized columns" print normalize_columns_separately(headers, d) print "\n \n \n \nNormalizing the whole matrix" print normalize_columns_together(headers, d) print "############## Extensions ################" d.add_column("new data", "numeric") print "\n \n \n" d.add_datacolumn("new data", "numeric", [1, 2,3]) print " \n \n ********* Extension 2- sorting ************** \n " print "Sorting a matrix \n \n " print sort(headers, d) print "*** \n 2b) normalize and sort \n" print " \n \n \n ",normalize_sort(headers, d) """ print "*********************************************" print "linear regression" headers = d.get_headers() print "headers: ", headers print linear_regression(d, [headers[0],headers[1]], headers[2]) if __name__ == '__main__': if len(sys.argv) < 2: print "Usage: python %s <csv_filename>" % sys.argv[0] print " where <csv_filename> specifies a csv file" exit() main( sys.argv[1] )
ID=[] # list of ID values mindistances=[] # minimum distances algorithm for dpoint in d: distances=[] # distances from each mean for mean in means: # compute distance of each mean, using the distance formula differences=dpoint-mean squares=np.square(differences) sums=np.sum(squares) distance=np.sqrt(sums) distances.append(distance) # add the distance to the distances list ID.append(np.argmin(distances)) mindistances.append(distances[np.argmin(distances)]) retval=[] retval.append(ID) retval.append(mindistances) return np.matrix(ID).transpose(), np.matrix(mindistances).transpose() # return a list of the ID values and the distances
identifier_body
bundle.js
/******/ (function(modules) { // webpackBootstrap /******/ // The module cache /******/ var installedModules = {}; /******/ /******/ // The require function /******/ function __webpack_require__(moduleId) { /******/ /******/ // Check if module is in cache /******/ if(installedModules[moduleId]) { /******/ return installedModules[moduleId].exports; /******/ } /******/ // Create a new module (and put it into the cache) /******/ var module = installedModules[moduleId] = { /******/ i: moduleId, /******/ l: false, /******/ exports: {} /******/ }; /******/ /******/ // Execute the module function /******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__); /******/ /******/ // Flag the module as loaded /******/ module.l = true; /******/ /******/ // Return the exports of the module /******/ return module.exports; /******/ } /******/ /******/ /******/ // expose the modules object (__webpack_modules__) /******/ __webpack_require__.m = modules; /******/ /******/ // expose the module cache /******/ __webpack_require__.c = installedModules; /******/ /******/ // define getter function for harmony exports /******/ __webpack_require__.d = function(exports, name, getter) { /******/ if(!__webpack_require__.o(exports, name)) { /******/ Object.defineProperty(exports, name, { enumerable: true, get: getter }); /******/ } /******/ }; /******/ /******/ // define __esModule on exports /******/ __webpack_require__.r = function(exports) { /******/ if(typeof Symbol !== 'undefined' && Symbol.toStringTag) { /******/ Object.defineProperty(exports, Symbol.toStringTag, { value: 'Module' }); /******/ } /******/ Object.defineProperty(exports, '__esModule', { value: true }); /******/ }; /******/ /******/ // create a fake namespace object /******/ // mode & 1: value is a module id, require it /******/ // mode & 2: merge all properties of value into the ns /******/ // mode & 4: return value when already ns object /******/ // mode & 8|1: behave like require /******/ __webpack_require__.t = function(value, mode) { /******/ if(mode & 1) value = __webpack_require__(value); /******/ if(mode & 8) return value; /******/ if((mode & 4) && typeof value === 'object' && value && value.__esModule) return value; /******/ var ns = Object.create(null); /******/ __webpack_require__.r(ns); /******/ Object.defineProperty(ns, 'default', { enumerable: true, value: value }); /******/ if(mode & 2 && typeof value != 'string') for(var key in value) __webpack_require__.d(ns, key, function(key) { return value[key]; }.bind(null, key)); /******/ return ns; /******/ }; /******/ /******/ // getDefaultExport function for compatibility with non-harmony modules /******/ __webpack_require__.n = function(module) { /******/ var getter = module && module.__esModule ? /******/ function getDefault() { return module['default']; } : /******/ function getModuleExports() { return module; }; /******/ __webpack_require__.d(getter, 'a', getter); /******/ return getter; /******/ }; /******/ /******/ // Object.prototype.hasOwnProperty.call /******/ __webpack_require__.o = function(object, property) { return Object.prototype.hasOwnProperty.call(object, property); }; /******/ /******/ // __webpack_public_path__ /******/ __webpack_require__.p = ""; /******/ /******/ /******/ // Load entry module and return exports /******/ return __webpack_require__(__webpack_require__.s = "./js/script.js"); /******/ }) /************************************************************************/ /******/ ({ /***/ "./js/parts/calculator.js": /*!********************************!*\ !*** ./js/parts/calculator.js ***! \********************************/ /*! no static exports found */ /***/ (function(module, exports) { function calculator() { //CALCULATOR let people = document.querySelectorAll('.counter-block-input')[0], days = document.querySelectorAll('.counter-block-input')[1], place = document.querySelector('#select'), totalOutput = document.querySelector('#total'), peopleSum = 0, daysSum = 0, placeIndex = +place.options[0].value,//index given by default totalSum; totalOutput.innerHTML = 0; people.value = 0; days.value = 0; people.addEventListener('blur', function(){ peopleSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { totalOutput.innerHTML = totalSum; } }); days.addEventListener('blur', function(){ daysSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { totalOutput.innerHTML = totalSum; } }); place.addEventListener('change', function(){ if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { let a = totalSum; totalOutput.innerHTML = a * this.options[this.selectedIndex].value; } }); } module.exports = calculator; /***/ }), /***/ "./js/parts/form.js": /*!**************************!*\ !*** ./js/parts/form.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function form() { // FORM let message = { loading: 'Загрузка...', success: 'Спасибо! Скоро мы с вами свяжемся.', failure: 'Что-то пошло не так.', }; let form = document.querySelectorAll('form'), //select two forms: from modal and "we will get in touch" section at the end input, input,//will be used to select the input(s) of one of the forms statusMessage = document.createElement('div');//to display a message according to stage and successfulness form = Array.from(form);//to create an array from a nodeList(you can't put eventListeners on nodeLists) statusMessage.classList.add('status'); function sendForms(){ return new Promise(function(resolve, reject){ for (let i = 0; i < form.length; i++){//assigning eventListeners to each form form[i].addEventListener('submit', function(){ event.preventDefault(); form = this; form.appendChild(statusMessage); input = form.getElementsByTagName('input'); let request = new XMLHttpRequest(); request.open('POST', 'server.php'); //request.setRequestHeader('Content-type', 'application/x-www-form-urlencoded'); request.setRequestHeader('Content-type', 'application/json; charset=utf-8'); let formData = new FormData(form); //request.send(formData) let obj = {}; formData.forEach(function(value, key){ obj[key] = value; }); let json = JSON.stringify(obj); request.send(json); request.addEventListener('readystatechange', function(){ if (request.readyState < 4){ statusMessage.innerHTML = message.loading; } else if (request.readyState === 4 && request.status == 200){ resolve(); } else { reject(); } }); }); } }); } sendForms() .then(() => statusMessage.innerHTML = message.success) .catch(() => statusMessage.innerHTML = message.failure); } module.exports = form; /***/ }), /***/ "./js/parts/modal.js": /*!***************************!*\ !*** ./js/parts/modal.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function modal() { //MODAL WINDOW let moreButton = document.querySelector('.more'),//The "Узнать больше" button descriptionBtn = document.querySelectorAll('.description-btn'),//the buttons from the tabs overlay = document.querySelector('.overlay'),//the modal window closeCross = document.querySelector('.popup-close'); descriptionBtn = Array.from(descriptionBtn);//making an array from the nodelist to push the moreButton descriptionBtn.push(moreButton); descriptionBtn.forEach(function(el){ el.addEventListener('click', function () { overlay.style.display = 'block'; this.classList.add('more-splash');//this means current element (similar to event.target somehow) document.body.style.overflow = 'hidden';//to disable scroll on the page while the modal window is visible }); }); closeCross.addEventListener('click', function() { overlay.style.display = 'none'; moreButton.classList.remove('more-splash'); document.body.style.overflow = ''; }); } module.exports = modal; /***/ }), /***/ "./js/parts/slider.js": /*!****************************!*\ !*** ./js/parts/slider.js ***! \****************************/ /*! no static exports found */ /***/ (function(module, exports) { function slider() { // SLIDER let slides = document.querySelectorAll('.slider-item'), prev = document.querySelector('.prev'), next = document.querySelector('.next'), dotWrap = document.querySelector('.slider-dots'), dots = document.querySelectorAll('.dot'), sliderIndex = 0; dots = Array.from(dots);// to use indexOf when click on a dot function showSlide() { // the first slide will be shown when a user clicks right being on the last slide if (sliderIndex == slides.length){ sliderIndex = 0; } // the last slide will be shown when a user clicks left being on the first slide if (sliderIndex < 0){ sliderIndex = 3; } slides.forEach((item) => item.style.display = 'none'); slides[sliderIndex].style.display = 'block'; dots.forEach((item) => item.classList.remove('dot-active')); dots[sliderIndex].classList.add('dot-active'); } showSlide();// to show only the first slide prev.addEventListener('click', function() { sliderIndex--; showSlide(); }); next.addEventListener('click', function() { sliderIndex++; showSlide(); }); dotWrap.addEventListener('click', function(event){ if (event.target.classList.contains('dot')){ //getting index of the clicked dot and assigning it to the appropriate slide sliderIndex = dots.indexOf(event.target); showSlide(); } }); } module.exports = slider; /***/ }), /***/ "./js/parts/tabs.js": /*!**************************!*\ !*** ./js/parts/tabs.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function tabs() { // TABS //Working with div class='info' let tab = document.querySelectorAll('.info-header-tab'),//get the buttons in the header info = document.querySelector('.info-header'),//parental element for the buttons tabContent = document.querySelectorAll('.info-tabcontent');//content that must be assigned to a particular tab function hideTabContent(a) {// to hide all the content elements for (let i = a; i < tabContent.length; i++){ tabContent[i].classList.remove('show'); tabContent[i].classList.add('hide');//manipulate classes in the css file to hide the content elements } } hideTabContent(1);// Need to hide all the content elements except for the first one function showTabContent(b){//to show the content element we need if (tabContent[b].classList.contains('hide')){ tabContent[b].classList.remove('hide'); tabContent[b].classList.add('show'); } } // attaching EventListener to the parent using delegetion info.addEventListener('click', function(event){ let target = event.target; if (target && target.classList.contains('info-header-tab')){ // Using for-loop to assign a particular content element to a particular tab for (let i = 0; i < tab.length; i++) { if (target == tab[i]){ hideTabContent(0);//to hide the first content element showTabContent(i);//to display the one that matches the target of our click break; } } } }); } module.exports = tabs; /***/ }), /***/ "./js/parts/timer.js": /*!***************************!*\ !*** ./js/parts/timer.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function timer() { // TIMER let deadline = "2019-06-14";//if the date expires, change the date so the timer works appropriately function getRemainingTime(endtime) { let t = Date.parse(endtime) - Date.parse(new Date()),//difference between the date when timer expires and the moment when the function is executed (ms) seconds = Math.floor((t/1000) % 60), minutes = Math.floor((t/1000/60) % 60), hours = Math.floor((t/(1000*60*60))); if (t < 0){ //making the timer look nice on the page in case the date has expired seconds = 0; minutes = 0; hours = 0; } return { 'total' : t, 'seconds' : seconds, 'minutes' : minutes, 'hours' : hours }; } function setClock (id, endtime) { //id - the timer div's id let timer = document.getElementById(id), hours = timer.querySelector('.hours'), minutes = timer.querySelector('.minutes'), seconds = timer.querySelector('.seconds'), timeInterval = setInterval(updateClock, 1000); function updateClock() { let t = getRemainingTime(endtime);
beautify(number){ // to add "0" before 0,1,2,3,4,5,6,7,8,9 if (number < 10){ number = `0${number}`; } return number; } hours.textContent = beautify(t.hours); minutes.textContent = beautify(t.minutes); seconds.textContent = beautify(t.seconds); if (t.total <= 0){ // when the timer reaches the deadline - stop timer clearInterval(timeInterval); } } } setClock('timer', deadline); // executing this function will create the timer. Changing the parameters will create a new timer } module.exports = timer; /***/ }), /***/ "./js/script.js": /*!**********************!*\ !*** ./js/script.js ***! \**********************/ /*! no static exports found */ /***/ (function(module, exports, __webpack_require__) { window.addEventListener('DOMContentLoaded', function() { "use strict"; let calculator = __webpack_require__(/*! ./parts/calculator */ "./js/parts/calculator.js"), form = __webpack_require__(/*! ./parts/form */ "./js/parts/form.js"), modal = __webpack_require__(/*! ./parts/modal */ "./js/parts/modal.js"), slider = __webpack_require__(/*! ./parts/slider */ "./js/parts/slider.js"), tabs = __webpack_require__(/*! ./parts/tabs */ "./js/parts/tabs.js"), timer = __webpack_require__(/*! ./parts/timer */ "./js/parts/timer.js"); calculator(); form(); modal(); slider(); tabs(); timer(); }); /***/ }) /******/ }); //# sourceMappingURL=bundle.js.map
function
identifier_name
bundle.js
/******/ (function(modules) { // webpackBootstrap /******/ // The module cache /******/ var installedModules = {}; /******/ /******/ // The require function /******/ function __webpack_require__(moduleId) { /******/ /******/ // Check if module is in cache /******/ if(installedModules[moduleId]) { /******/ return installedModules[moduleId].exports; /******/ } /******/ // Create a new module (and put it into the cache) /******/ var module = installedModules[moduleId] = { /******/ i: moduleId, /******/ l: false, /******/ exports: {} /******/ }; /******/ /******/ // Execute the module function /******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__); /******/ /******/ // Flag the module as loaded /******/ module.l = true; /******/ /******/ // Return the exports of the module /******/ return module.exports; /******/ } /******/ /******/ /******/ // expose the modules object (__webpack_modules__) /******/ __webpack_require__.m = modules; /******/ /******/ // expose the module cache /******/ __webpack_require__.c = installedModules; /******/ /******/ // define getter function for harmony exports /******/ __webpack_require__.d = function(exports, name, getter) { /******/ if(!__webpack_require__.o(exports, name)) { /******/ Object.defineProperty(exports, name, { enumerable: true, get: getter }); /******/ } /******/ }; /******/ /******/ // define __esModule on exports /******/ __webpack_require__.r = function(exports) { /******/ if(typeof Symbol !== 'undefined' && Symbol.toStringTag) { /******/ Object.defineProperty(exports, Symbol.toStringTag, { value: 'Module' }); /******/ } /******/ Object.defineProperty(exports, '__esModule', { value: true }); /******/ }; /******/ /******/ // create a fake namespace object /******/ // mode & 1: value is a module id, require it /******/ // mode & 2: merge all properties of value into the ns /******/ // mode & 4: return value when already ns object /******/ // mode & 8|1: behave like require /******/ __webpack_require__.t = function(value, mode) { /******/ if(mode & 1) value = __webpack_require__(value); /******/ if(mode & 8) return value; /******/ if((mode & 4) && typeof value === 'object' && value && value.__esModule) return value; /******/ var ns = Object.create(null); /******/ __webpack_require__.r(ns); /******/ Object.defineProperty(ns, 'default', { enumerable: true, value: value }); /******/ if(mode & 2 && typeof value != 'string') for(var key in value) __webpack_require__.d(ns, key, function(key) { return value[key]; }.bind(null, key)); /******/ return ns; /******/ }; /******/ /******/ // getDefaultExport function for compatibility with non-harmony modules /******/ __webpack_require__.n = function(module) { /******/ var getter = module && module.__esModule ? /******/ function getDefault() { return module['default']; } : /******/ function getModuleExports() { return module; }; /******/ __webpack_require__.d(getter, 'a', getter); /******/ return getter; /******/ }; /******/ /******/ // Object.prototype.hasOwnProperty.call /******/ __webpack_require__.o = function(object, property) { return Object.prototype.hasOwnProperty.call(object, property); }; /******/ /******/ // __webpack_public_path__ /******/ __webpack_require__.p = ""; /******/ /******/ /******/ // Load entry module and return exports /******/ return __webpack_require__(__webpack_require__.s = "./js/script.js"); /******/ }) /************************************************************************/ /******/ ({ /***/ "./js/parts/calculator.js": /*!********************************!*\ !*** ./js/parts/calculator.js ***! \********************************/ /*! no static exports found */ /***/ (function(module, exports) { function calculator() { //CALCULATOR let people = document.querySelectorAll('.counter-block-input')[0], days = document.querySelectorAll('.counter-block-input')[1], place = document.querySelector('#select'), totalOutput = document.querySelector('#total'), peopleSum = 0, daysSum = 0, placeIndex = +place.options[0].value,//index given by default totalSum; totalOutput.innerHTML = 0; people.value = 0; days.value = 0; people.addEventListener('blur', function(){ peopleSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { totalOutput.innerHTML = totalSum; } }); days.addEventListener('blur', function(){ daysSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { totalOutput.innerHTML = totalSum; } }); place.addEventListener('change', function(){ if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { let a = totalSum; totalOutput.innerHTML = a * this.options[this.selectedIndex].value; } }); } module.exports = calculator; /***/ }), /***/ "./js/parts/form.js": /*!**************************!*\ !*** ./js/parts/form.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function form() { // FORM let message = { loading: 'Загрузка...', success: 'Спасибо! Скоро мы с вами свяжемся.', failure: 'Что-то пошло не так.', }; let form = document.querySelectorAll('form'), //select two forms: from modal and "we will get in touch" section at the end input, input,//will be used to select the input(s) of one of the forms statusMessage = document.createElement('div');//to display a message according to stage and successfulness form = Array.from(form);//to create an array from a nodeList(you can't put eventListeners on nodeLists) statusMessage.classList.add('status'); function sendForms(){ return new Promise(function(resolve, reject){ for (let i = 0; i < form.length; i++){//assigning eventListeners to each form form[i].addEventListener('submit', function(){ event.preventDefault(); form = this; form.appendChild(statusMessage); input = form.getElementsByTagName('input'); let request = new XMLHttpRequest(); request.open('POST', 'server.php'); //request.setRequestHeader('Content-type', 'application/x-www-form-urlencoded'); request.setRequestHeader('Content-type', 'application/json; charset=utf-8'); let formData = new FormData(form); //request.send(formData) let obj = {}; formData.forEach(function(value, key){ obj[key] = value; }); let json = JSON.stringify(obj); request.send(json); request.addEventListener('readystatechange', function(){ if (request.readyState < 4){ statusMessage.innerHTML = message.loading; } else if (request.readyState === 4 && request.status == 200){ resolve(); } else { reject(); } }); }); } }); } sendForms() .then(() => statusMessage.innerHTML = message.success) .catch(() => statusMessage.innerHTML = message.failure); } module.exports = form; /***/ }), /***/ "./js/parts/modal.js": /*!***************************!*\ !*** ./js/parts/modal.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function modal() { //MODAL WINDOW let moreButton = document.querySelector('.more'),//The "Узнать больше" button descriptionBtn = document.querySelectorAll('.description-btn'),//the buttons from the tabs overlay = document.querySelector('.overlay'),//the modal window closeCross = document.querySelector('.popup-close'); descriptionBtn = Array.from(descriptionBtn);//making an array from the nodelist to push the moreButton descriptionBtn.push(moreButton); descriptionBtn.forEach(function(el){ el.addEventListener('click', function () { overlay.style.display = 'block'; this.classList.add('more-splash');//this means current element (similar to event.target somehow) document.body.style.overflow = 'hidden';//to disable scroll on the page while the modal window is visible }); }); closeCross.addEventListener('click', function() { overlay.style.display = 'none'; moreButton.classList.remove('more-splash'); document.body.style.overflow = ''; }); } module.exports = modal; /***/ }), /***/ "./js/parts/slider.js": /*!****************************!*\ !*** ./js/parts/slider.js ***! \****************************/ /*! no static exports found */ /***/ (function(module, exports) { function slider() { // SLIDER let slides = document.querySelectorAll('.slider-item'), prev = document.querySelector('.prev'), next = document.querySelector('.next'), dotWrap = document.querySelector('.slider-dots'), dots = document.querySelectorAll('.dot'), sliderIndex = 0; dots = Array.from(dots);// to use indexOf when click on a dot function showSlide() { // the first slide will be shown when a user clicks right being on the last slide if (sliderIndex == slides.length){ sliderIndex = 0; } // the last slide will be shown when a user clicks left being on the first slide if (sliderIndex < 0){ sliderIndex = 3; } slides.forEach((item) => item.style.display = 'none'); slides[sliderIndex].style.display = 'block'; dots.forEach((item) => item.classList.remove('dot-active')); dots[sliderIndex].classList.add('dot-active'); } showSlide();// to show only the first slide prev.addEventListener('click', function() { sliderIndex--; showSlide(); }); next.addEventListener('click', function() { sliderIndex++; showSlide(); }); dotWrap.addEventListener('click', function(event){ if (event.target.classList.contains('dot')){ //getting index of the clicked dot and assigning it to the appropriate slide sliderIndex = dots.indexOf(event.target); showSlide(); } }); } module.exports = slider; /***/ }), /***/ "./js/parts/tabs.js": /*!**************************!*\ !*** ./js/parts/tabs.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function tabs() { // TABS //Working with div class='info' let tab = document.querySelectorAll('.info-header-tab'),//get the buttons in the header info = document.querySelector('.info-header'),//parental element for the buttons tabContent = document.querySelectorAll('.info-tabcontent');//content that must be assigned to a particular tab function hideTabContent(a) {// to hide all the content elements for (let i = a; i < tabContent.length; i++){ tabContent[i].classList.remove('show'); tabContent[i].classList.add('hide');//manipulate classes in the css file to hide the content elements } } hideTabContent(1);// Need to hide all the content elements except for the first one function showTabContent(b){//to show the content element we need if (tabContent[b].c
info.addEventListener('click', function(event){ let target = event.target; if (target && target.classList.contains('info-header-tab')){ // Using for-loop to assign a particular content element to a particular tab for (let i = 0; i < tab.length; i++) { if (target == tab[i]){ hideTabContent(0);//to hide the first content element showTabContent(i);//to display the one that matches the target of our click break; } } } }); } module.exports = tabs; /***/ }), /***/ "./js/parts/timer.js": /*!***************************!*\ !*** ./js/parts/timer.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function timer() { // TIMER let deadline = "2019-06-14";//if the date expires, change the date so the timer works appropriately function getRemainingTime(endtime) { let t = Date.parse(endtime) - Date.parse(new Date()),//difference between the date when timer expires and the moment when the function is executed (ms) seconds = Math.floor((t/1000) % 60), minutes = Math.floor((t/1000/60) % 60), hours = Math.floor((t/(1000*60*60))); if (t < 0){ //making the timer look nice on the page in case the date has expired seconds = 0; minutes = 0; hours = 0; } return { 'total' : t, 'seconds' : seconds, 'minutes' : minutes, 'hours' : hours }; } function setClock (id, endtime) { //id - the timer div's id let timer = document.getElementById(id), hours = timer.querySelector('.hours'), minutes = timer.querySelector('.minutes'), seconds = timer.querySelector('.seconds'), timeInterval = setInterval(updateClock, 1000); function updateClock() { let t = getRemainingTime(endtime); function beautify(number){ // to add "0" before 0,1,2,3,4,5,6,7,8,9 if (number < 10){ number = `0${number}`; } return number; } hours.textContent = beautify(t.hours); minutes.textContent = beautify(t.minutes); seconds.textContent = beautify(t.seconds); if (t.total <= 0){ // when the timer reaches the deadline - stop timer clearInterval(timeInterval); } } } setClock('timer', deadline); // executing this function will create the timer. Changing the parameters will create a new timer } module.exports = timer; /***/ }), /***/ "./js/script.js": /*!**********************!*\ !*** ./js/script.js ***! \**********************/ /*! no static exports found */ /***/ (function(module, exports, __webpack_require__) { window.addEventListener('DOMContentLoaded', function() { "use strict"; let calculator = __webpack_require__(/*! ./parts/calculator */ "./js/parts/calculator.js"), form = __webpack_require__(/*! ./parts/form */ "./js/parts/form.js"), modal = __webpack_require__(/*! ./parts/modal */ "./js/parts/modal.js"), slider = __webpack_require__(/*! ./parts/slider */ "./js/parts/slider.js"), tabs = __webpack_require__(/*! ./parts/tabs */ "./js/parts/tabs.js"), timer = __webpack_require__(/*! ./parts/timer */ "./js/parts/timer.js"); calculator(); form(); modal(); slider(); tabs(); timer(); }); /***/ }) /******/ }); //# sourceMappingURL=bundle.js.map
lassList.contains('hide')){ tabContent[b].classList.remove('hide'); tabContent[b].classList.add('show'); } } // attaching EventListener to the parent using delegetion
identifier_body
bundle.js
/******/ (function(modules) { // webpackBootstrap /******/ // The module cache /******/ var installedModules = {}; /******/ /******/ // The require function /******/ function __webpack_require__(moduleId) { /******/ /******/ // Check if module is in cache /******/ if(installedModules[moduleId]) { /******/ return installedModules[moduleId].exports; /******/ } /******/ // Create a new module (and put it into the cache) /******/ var module = installedModules[moduleId] = { /******/ i: moduleId, /******/ l: false, /******/ exports: {} /******/ }; /******/ /******/ // Execute the module function /******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__); /******/ /******/ // Flag the module as loaded /******/ module.l = true; /******/ /******/ // Return the exports of the module /******/ return module.exports; /******/ } /******/ /******/ /******/ // expose the modules object (__webpack_modules__) /******/ __webpack_require__.m = modules; /******/ /******/ // expose the module cache /******/ __webpack_require__.c = installedModules; /******/ /******/ // define getter function for harmony exports /******/ __webpack_require__.d = function(exports, name, getter) { /******/ if(!__webpack_require__.o(exports, name)) { /******/ Object.defineProperty(exports, name, { enumerable: true, get: getter }); /******/ } /******/ }; /******/ /******/ // define __esModule on exports /******/ __webpack_require__.r = function(exports) { /******/ if(typeof Symbol !== 'undefined' && Symbol.toStringTag) { /******/ Object.defineProperty(exports, Symbol.toStringTag, { value: 'Module' }); /******/ } /******/ Object.defineProperty(exports, '__esModule', { value: true }); /******/ }; /******/ /******/ // create a fake namespace object /******/ // mode & 1: value is a module id, require it /******/ // mode & 2: merge all properties of value into the ns /******/ // mode & 4: return value when already ns object /******/ // mode & 8|1: behave like require /******/ __webpack_require__.t = function(value, mode) { /******/ if(mode & 1) value = __webpack_require__(value); /******/ if(mode & 8) return value; /******/ if((mode & 4) && typeof value === 'object' && value && value.__esModule) return value; /******/ var ns = Object.create(null); /******/ __webpack_require__.r(ns); /******/ Object.defineProperty(ns, 'default', { enumerable: true, value: value }); /******/ if(mode & 2 && typeof value != 'string') for(var key in value) __webpack_require__.d(ns, key, function(key) { return value[key]; }.bind(null, key)); /******/ return ns; /******/ }; /******/ /******/ // getDefaultExport function for compatibility with non-harmony modules /******/ __webpack_require__.n = function(module) { /******/ var getter = module && module.__esModule ? /******/ function getDefault() { return module['default']; } : /******/ function getModuleExports() { return module; }; /******/ __webpack_require__.d(getter, 'a', getter); /******/ return getter; /******/ }; /******/ /******/ // Object.prototype.hasOwnProperty.call /******/ __webpack_require__.o = function(object, property) { return Object.prototype.hasOwnProperty.call(object, property); }; /******/ /******/ // __webpack_public_path__ /******/ __webpack_require__.p = ""; /******/ /******/ /******/ // Load entry module and return exports /******/ return __webpack_require__(__webpack_require__.s = "./js/script.js"); /******/ }) /************************************************************************/ /******/ ({ /***/ "./js/parts/calculator.js": /*!********************************!*\ !*** ./js/parts/calculator.js ***! \********************************/ /*! no static exports found */ /***/ (function(module, exports) { function calculator() { //CALCULATOR let people = document.querySelectorAll('.counter-block-input')[0], days = document.querySelectorAll('.counter-block-input')[1], place = document.querySelector('#select'), totalOutput = document.querySelector('#total'), peopleSum = 0, daysSum = 0, placeIndex = +place.options[0].value,//index given by default totalSum; totalOutput.innerHTML = 0; people.value = 0; days.value = 0; people.addEventListener('blur', function(){ peopleSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else
}); days.addEventListener('blur', function(){ daysSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { totalOutput.innerHTML = totalSum; } }); place.addEventListener('change', function(){ if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { let a = totalSum; totalOutput.innerHTML = a * this.options[this.selectedIndex].value; } }); } module.exports = calculator; /***/ }), /***/ "./js/parts/form.js": /*!**************************!*\ !*** ./js/parts/form.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function form() { // FORM let message = { loading: 'Загрузка...', success: 'Спасибо! Скоро мы с вами свяжемся.', failure: 'Что-то пошло не так.', }; let form = document.querySelectorAll('form'), //select two forms: from modal and "we will get in touch" section at the end input, input,//will be used to select the input(s) of one of the forms statusMessage = document.createElement('div');//to display a message according to stage and successfulness form = Array.from(form);//to create an array from a nodeList(you can't put eventListeners on nodeLists) statusMessage.classList.add('status'); function sendForms(){ return new Promise(function(resolve, reject){ for (let i = 0; i < form.length; i++){//assigning eventListeners to each form form[i].addEventListener('submit', function(){ event.preventDefault(); form = this; form.appendChild(statusMessage); input = form.getElementsByTagName('input'); let request = new XMLHttpRequest(); request.open('POST', 'server.php'); //request.setRequestHeader('Content-type', 'application/x-www-form-urlencoded'); request.setRequestHeader('Content-type', 'application/json; charset=utf-8'); let formData = new FormData(form); //request.send(formData) let obj = {}; formData.forEach(function(value, key){ obj[key] = value; }); let json = JSON.stringify(obj); request.send(json); request.addEventListener('readystatechange', function(){ if (request.readyState < 4){ statusMessage.innerHTML = message.loading; } else if (request.readyState === 4 && request.status == 200){ resolve(); } else { reject(); } }); }); } }); } sendForms() .then(() => statusMessage.innerHTML = message.success) .catch(() => statusMessage.innerHTML = message.failure); } module.exports = form; /***/ }), /***/ "./js/parts/modal.js": /*!***************************!*\ !*** ./js/parts/modal.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function modal() { //MODAL WINDOW let moreButton = document.querySelector('.more'),//The "Узнать больше" button descriptionBtn = document.querySelectorAll('.description-btn'),//the buttons from the tabs overlay = document.querySelector('.overlay'),//the modal window closeCross = document.querySelector('.popup-close'); descriptionBtn = Array.from(descriptionBtn);//making an array from the nodelist to push the moreButton descriptionBtn.push(moreButton); descriptionBtn.forEach(function(el){ el.addEventListener('click', function () { overlay.style.display = 'block'; this.classList.add('more-splash');//this means current element (similar to event.target somehow) document.body.style.overflow = 'hidden';//to disable scroll on the page while the modal window is visible }); }); closeCross.addEventListener('click', function() { overlay.style.display = 'none'; moreButton.classList.remove('more-splash'); document.body.style.overflow = ''; }); } module.exports = modal; /***/ }), /***/ "./js/parts/slider.js": /*!****************************!*\ !*** ./js/parts/slider.js ***! \****************************/ /*! no static exports found */ /***/ (function(module, exports) { function slider() { // SLIDER let slides = document.querySelectorAll('.slider-item'), prev = document.querySelector('.prev'), next = document.querySelector('.next'), dotWrap = document.querySelector('.slider-dots'), dots = document.querySelectorAll('.dot'), sliderIndex = 0; dots = Array.from(dots);// to use indexOf when click on a dot function showSlide() { // the first slide will be shown when a user clicks right being on the last slide if (sliderIndex == slides.length){ sliderIndex = 0; } // the last slide will be shown when a user clicks left being on the first slide if (sliderIndex < 0){ sliderIndex = 3; } slides.forEach((item) => item.style.display = 'none'); slides[sliderIndex].style.display = 'block'; dots.forEach((item) => item.classList.remove('dot-active')); dots[sliderIndex].classList.add('dot-active'); } showSlide();// to show only the first slide prev.addEventListener('click', function() { sliderIndex--; showSlide(); }); next.addEventListener('click', function() { sliderIndex++; showSlide(); }); dotWrap.addEventListener('click', function(event){ if (event.target.classList.contains('dot')){ //getting index of the clicked dot and assigning it to the appropriate slide sliderIndex = dots.indexOf(event.target); showSlide(); } }); } module.exports = slider; /***/ }), /***/ "./js/parts/tabs.js": /*!**************************!*\ !*** ./js/parts/tabs.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function tabs() { // TABS //Working with div class='info' let tab = document.querySelectorAll('.info-header-tab'),//get the buttons in the header info = document.querySelector('.info-header'),//parental element for the buttons tabContent = document.querySelectorAll('.info-tabcontent');//content that must be assigned to a particular tab function hideTabContent(a) {// to hide all the content elements for (let i = a; i < tabContent.length; i++){ tabContent[i].classList.remove('show'); tabContent[i].classList.add('hide');//manipulate classes in the css file to hide the content elements } } hideTabContent(1);// Need to hide all the content elements except for the first one function showTabContent(b){//to show the content element we need if (tabContent[b].classList.contains('hide')){ tabContent[b].classList.remove('hide'); tabContent[b].classList.add('show'); } } // attaching EventListener to the parent using delegetion info.addEventListener('click', function(event){ let target = event.target; if (target && target.classList.contains('info-header-tab')){ // Using for-loop to assign a particular content element to a particular tab for (let i = 0; i < tab.length; i++) { if (target == tab[i]){ hideTabContent(0);//to hide the first content element showTabContent(i);//to display the one that matches the target of our click break; } } } }); } module.exports = tabs; /***/ }), /***/ "./js/parts/timer.js": /*!***************************!*\ !*** ./js/parts/timer.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function timer() { // TIMER let deadline = "2019-06-14";//if the date expires, change the date so the timer works appropriately function getRemainingTime(endtime) { let t = Date.parse(endtime) - Date.parse(new Date()),//difference between the date when timer expires and the moment when the function is executed (ms) seconds = Math.floor((t/1000) % 60), minutes = Math.floor((t/1000/60) % 60), hours = Math.floor((t/(1000*60*60))); if (t < 0){ //making the timer look nice on the page in case the date has expired seconds = 0; minutes = 0; hours = 0; } return { 'total' : t, 'seconds' : seconds, 'minutes' : minutes, 'hours' : hours }; } function setClock (id, endtime) { //id - the timer div's id let timer = document.getElementById(id), hours = timer.querySelector('.hours'), minutes = timer.querySelector('.minutes'), seconds = timer.querySelector('.seconds'), timeInterval = setInterval(updateClock, 1000); function updateClock() { let t = getRemainingTime(endtime); function beautify(number){ // to add "0" before 0,1,2,3,4,5,6,7,8,9 if (number < 10){ number = `0${number}`; } return number; } hours.textContent = beautify(t.hours); minutes.textContent = beautify(t.minutes); seconds.textContent = beautify(t.seconds); if (t.total <= 0){ // when the timer reaches the deadline - stop timer clearInterval(timeInterval); } } } setClock('timer', deadline); // executing this function will create the timer. Changing the parameters will create a new timer } module.exports = timer; /***/ }), /***/ "./js/script.js": /*!**********************!*\ !*** ./js/script.js ***! \**********************/ /*! no static exports found */ /***/ (function(module, exports, __webpack_require__) { window.addEventListener('DOMContentLoaded', function() { "use strict"; let calculator = __webpack_require__(/*! ./parts/calculator */ "./js/parts/calculator.js"), form = __webpack_require__(/*! ./parts/form */ "./js/parts/form.js"), modal = __webpack_require__(/*! ./parts/modal */ "./js/parts/modal.js"), slider = __webpack_require__(/*! ./parts/slider */ "./js/parts/slider.js"), tabs = __webpack_require__(/*! ./parts/tabs */ "./js/parts/tabs.js"), timer = __webpack_require__(/*! ./parts/timer */ "./js/parts/timer.js"); calculator(); form(); modal(); slider(); tabs(); timer(); }); /***/ }) /******/ }); //# sourceMappingURL=bundle.js.map
{ totalOutput.innerHTML = totalSum; }
conditional_block
bundle.js
/******/ (function(modules) { // webpackBootstrap /******/ // The module cache /******/ var installedModules = {}; /******/ /******/ // The require function /******/ function __webpack_require__(moduleId) { /******/ /******/ // Check if module is in cache /******/ if(installedModules[moduleId]) { /******/ return installedModules[moduleId].exports; /******/ } /******/ // Create a new module (and put it into the cache) /******/ var module = installedModules[moduleId] = { /******/ i: moduleId, /******/ l: false, /******/ exports: {} /******/ }; /******/ /******/ // Execute the module function /******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__); /******/ /******/ // Flag the module as loaded /******/ module.l = true; /******/ /******/ // Return the exports of the module /******/ return module.exports; /******/ } /******/ /******/ /******/ // expose the modules object (__webpack_modules__) /******/ __webpack_require__.m = modules; /******/ /******/ // expose the module cache /******/ __webpack_require__.c = installedModules; /******/ /******/ // define getter function for harmony exports /******/ __webpack_require__.d = function(exports, name, getter) { /******/ if(!__webpack_require__.o(exports, name)) { /******/ Object.defineProperty(exports, name, { enumerable: true, get: getter }); /******/ } /******/ }; /******/ /******/ // define __esModule on exports /******/ __webpack_require__.r = function(exports) { /******/ if(typeof Symbol !== 'undefined' && Symbol.toStringTag) { /******/ Object.defineProperty(exports, Symbol.toStringTag, { value: 'Module' }); /******/ } /******/ Object.defineProperty(exports, '__esModule', { value: true }); /******/ }; /******/ /******/ // create a fake namespace object /******/ // mode & 1: value is a module id, require it /******/ // mode & 2: merge all properties of value into the ns /******/ // mode & 4: return value when already ns object /******/ // mode & 8|1: behave like require /******/ __webpack_require__.t = function(value, mode) { /******/ if(mode & 1) value = __webpack_require__(value); /******/ if(mode & 8) return value; /******/ if((mode & 4) && typeof value === 'object' && value && value.__esModule) return value; /******/ var ns = Object.create(null); /******/ __webpack_require__.r(ns); /******/ Object.defineProperty(ns, 'default', { enumerable: true, value: value }); /******/ if(mode & 2 && typeof value != 'string') for(var key in value) __webpack_require__.d(ns, key, function(key) { return value[key]; }.bind(null, key)); /******/ return ns; /******/ }; /******/ /******/ // getDefaultExport function for compatibility with non-harmony modules /******/ __webpack_require__.n = function(module) { /******/ var getter = module && module.__esModule ? /******/ function getDefault() { return module['default']; } : /******/ function getModuleExports() { return module; }; /******/ __webpack_require__.d(getter, 'a', getter); /******/ return getter; /******/ }; /******/ /******/ // Object.prototype.hasOwnProperty.call /******/ __webpack_require__.o = function(object, property) { return Object.prototype.hasOwnProperty.call(object, property); }; /******/ /******/ // __webpack_public_path__ /******/ __webpack_require__.p = ""; /******/ /******/ /******/ // Load entry module and return exports /******/ return __webpack_require__(__webpack_require__.s = "./js/script.js"); /******/ }) /************************************************************************/ /******/ ({ /***/ "./js/parts/calculator.js": /*!********************************!*\ !*** ./js/parts/calculator.js ***! \********************************/ /*! no static exports found */ /***/ (function(module, exports) { function calculator() { //CALCULATOR let people = document.querySelectorAll('.counter-block-input')[0], days = document.querySelectorAll('.counter-block-input')[1], place = document.querySelector('#select'), totalOutput = document.querySelector('#total'), peopleSum = 0, daysSum = 0, placeIndex = +place.options[0].value,//index given by default totalSum; totalOutput.innerHTML = 0; people.value = 0; days.value = 0; people.addEventListener('blur', function(){ peopleSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { totalOutput.innerHTML = totalSum; } }); days.addEventListener('blur', function(){ daysSum = +this.value; totalSum = (peopleSum + daysSum) * 4000 * placeIndex; if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { totalOutput.innerHTML = totalSum; } }); place.addEventListener('change', function(){ if (days.value == '' || people.value == '' || days.value == '0' || people.value == '0'){ totalOutput.innerHTML = 0; } else { let a = totalSum; totalOutput.innerHTML = a * this.options[this.selectedIndex].value; } }); } module.exports = calculator; /***/ }), /***/ "./js/parts/form.js": /*!**************************!*\ !*** ./js/parts/form.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function form() { // FORM let message = { loading: 'Загрузка...', success: 'Спасибо! Скоро мы с вами свяжемся.', failure: 'Что-то пошло не так.', }; let form = document.querySelectorAll('form'), //select two forms: from modal and "we will get in touch" section at the end input, input,//will be used to select the input(s) of one of the forms statusMessage = document.createElement('div');//to display a message according to stage and successfulness form = Array.from(form);//to create an array from a nodeList(you can't put eventListeners on nodeLists) statusMessage.classList.add('status'); function sendForms(){ return new Promise(function(resolve, reject){ for (let i = 0; i < form.length; i++){//assigning eventListeners to each form form[i].addEventListener('submit', function(){ event.preventDefault(); form = this; form.appendChild(statusMessage); input = form.getElementsByTagName('input'); let request = new XMLHttpRequest(); request.open('POST', 'server.php'); //request.setRequestHeader('Content-type', 'application/x-www-form-urlencoded'); request.setRequestHeader('Content-type', 'application/json; charset=utf-8'); let formData = new FormData(form); //request.send(formData) let obj = {}; formData.forEach(function(value, key){ obj[key] = value; }); let json = JSON.stringify(obj); request.send(json); request.addEventListener('readystatechange', function(){ if (request.readyState < 4){ statusMessage.innerHTML = message.loading; } else if (request.readyState === 4 && request.status == 200){ resolve(); } else { reject(); } }); }); } }); } sendForms() .then(() => statusMessage.innerHTML = message.success) .catch(() => statusMessage.innerHTML = message.failure); } module.exports = form; /***/ }), /***/ "./js/parts/modal.js": /*!***************************!*\ !*** ./js/parts/modal.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function modal() { //MODAL WINDOW let moreButton = document.querySelector('.more'),//The "Узнать больше" button descriptionBtn = document.querySelectorAll('.description-btn'),//the buttons from the tabs overlay = document.querySelector('.overlay'),//the modal window closeCross = document.querySelector('.popup-close'); descriptionBtn = Array.from(descriptionBtn);//making an array from the nodelist to push the moreButton descriptionBtn.push(moreButton); descriptionBtn.forEach(function(el){ el.addEventListener('click', function () { overlay.style.display = 'block'; this.classList.add('more-splash');//this means current element (similar to event.target somehow) document.body.style.overflow = 'hidden';//to disable scroll on the page while the modal window is visible }); }); closeCross.addEventListener('click', function() { overlay.style.display = 'none'; moreButton.classList.remove('more-splash'); document.body.style.overflow = ''; }); } module.exports = modal; /***/ }), /***/ "./js/parts/slider.js": /*!****************************!*\ !*** ./js/parts/slider.js ***! \****************************/ /*! no static exports found */ /***/ (function(module, exports) { function slider() { // SLIDER let slides = document.querySelectorAll('.slider-item'), prev = document.querySelector('.prev'), next = document.querySelector('.next'), dotWrap = document.querySelector('.slider-dots'), dots = document.querySelectorAll('.dot'), sliderIndex = 0; dots = Array.from(dots);// to use indexOf when click on a dot function showSlide() { // the first slide will be shown when a user clicks right being on the last slide if (sliderIndex == slides.length){ sliderIndex = 0; } // the last slide will be shown when a user clicks left being on the first slide if (sliderIndex < 0){ sliderIndex = 3; } slides.forEach((item) => item.style.display = 'none'); slides[sliderIndex].style.display = 'block'; dots.forEach((item) => item.classList.remove('dot-active')); dots[sliderIndex].classList.add('dot-active'); } showSlide();// to show only the first slide prev.addEventListener('click', function() { sliderIndex--; showSlide(); }); next.addEventListener('click', function() { sliderIndex++; showSlide(); }); dotWrap.addEventListener('click', function(event){ if (event.target.classList.contains('dot')){ //getting index of the clicked dot and assigning it to the appropriate slide sliderIndex = dots.indexOf(event.target); showSlide(); } }); } module.exports = slider; /***/ }), /***/ "./js/parts/tabs.js": /*!**************************!*\ !*** ./js/parts/tabs.js ***! \**************************/ /*! no static exports found */ /***/ (function(module, exports) { function tabs() { // TABS //Working with div class='info' let tab = document.querySelectorAll('.info-header-tab'),//get the buttons in the header info = document.querySelector('.info-header'),//parental element for the buttons tabContent = document.querySelectorAll('.info-tabcontent');//content that must be assigned to a particular tab function hideTabContent(a) {// to hide all the content elements for (let i = a; i < tabContent.length; i++){ tabContent[i].classList.remove('show'); tabContent[i].classList.add('hide');//manipulate classes in the css file to hide the content elements } } hideTabContent(1);// Need to hide all the content elements except for the first one function showTabContent(b){//to show the content element we need if (tabContent[b].classList.contains('hide')){ tabContent[b].classList.remove('hide'); tabContent[b].classList.add('show'); } } // attaching EventListener to the parent using delegetion info.addEventListener('click', function(event){ let target = event.target; if (target && target.classList.contains('info-header-tab')){ // Using for-loop to assign a particular content element to a particular tab for (let i = 0; i < tab.length; i++) { if (target == tab[i]){ hideTabContent(0);//to hide the first content element showTabContent(i);//to display the one that matches the target of our click break; } } } }); } module.exports = tabs; /***/ }), /***/ "./js/parts/timer.js":
/*!***************************!*\ !*** ./js/parts/timer.js ***! \***************************/ /*! no static exports found */ /***/ (function(module, exports) { function timer() { // TIMER let deadline = "2019-06-14";//if the date expires, change the date so the timer works appropriately function getRemainingTime(endtime) { let t = Date.parse(endtime) - Date.parse(new Date()),//difference between the date when timer expires and the moment when the function is executed (ms) seconds = Math.floor((t/1000) % 60), minutes = Math.floor((t/1000/60) % 60), hours = Math.floor((t/(1000*60*60))); if (t < 0){ //making the timer look nice on the page in case the date has expired seconds = 0; minutes = 0; hours = 0; } return { 'total' : t, 'seconds' : seconds, 'minutes' : minutes, 'hours' : hours }; } function setClock (id, endtime) { //id - the timer div's id let timer = document.getElementById(id), hours = timer.querySelector('.hours'), minutes = timer.querySelector('.minutes'), seconds = timer.querySelector('.seconds'), timeInterval = setInterval(updateClock, 1000); function updateClock() { let t = getRemainingTime(endtime); function beautify(number){ // to add "0" before 0,1,2,3,4,5,6,7,8,9 if (number < 10){ number = `0${number}`; } return number; } hours.textContent = beautify(t.hours); minutes.textContent = beautify(t.minutes); seconds.textContent = beautify(t.seconds); if (t.total <= 0){ // when the timer reaches the deadline - stop timer clearInterval(timeInterval); } } } setClock('timer', deadline); // executing this function will create the timer. Changing the parameters will create a new timer } module.exports = timer; /***/ }), /***/ "./js/script.js": /*!**********************!*\ !*** ./js/script.js ***! \**********************/ /*! no static exports found */ /***/ (function(module, exports, __webpack_require__) { window.addEventListener('DOMContentLoaded', function() { "use strict"; let calculator = __webpack_require__(/*! ./parts/calculator */ "./js/parts/calculator.js"), form = __webpack_require__(/*! ./parts/form */ "./js/parts/form.js"), modal = __webpack_require__(/*! ./parts/modal */ "./js/parts/modal.js"), slider = __webpack_require__(/*! ./parts/slider */ "./js/parts/slider.js"), tabs = __webpack_require__(/*! ./parts/tabs */ "./js/parts/tabs.js"), timer = __webpack_require__(/*! ./parts/timer */ "./js/parts/timer.js"); calculator(); form(); modal(); slider(); tabs(); timer(); }); /***/ }) /******/ }); //# sourceMappingURL=bundle.js.map
random_line_split
codec.rs
use std::pin::Pin; use std::task::{Context, Poll}; use async_trait::async_trait; use bytes::BytesMut; // We use `futures::stream::Stream` because `std::async_iter::AsyncIterator` (previously known as // `std::stream::Stream`) is still a bare-bones implementation - it does not even have an // `async fn next` method! use futures::{future::BoxFuture, sink, stream}; use g1_base::fmt::{DebugExt, InsertPlaceholder}; use super::{StreamRecv, StreamSend}; #[async_trait] pub trait Decode<Stream> where Stream: StreamRecv, { type Item; type Error: From<Stream::Error>; /// Decodes a byte stream and returns a stream item or an error when EOF is reached, but the /// byte stream buffer is not empty. async fn decode(&mut self, stream: &mut Stream) -> Result<Option<Self::Item>, Self::Error>; } pub trait Encode<Item> { /// Encodes a stream item and writes the output to a byte stream buffer. fn encode(&mut self, item: &Item, buffer: &mut BytesMut); } // TODO: We intend to implement `Decode` for async functions, similar to what we do for `Encode`. // However, for unknown reasons, it is causing the compiler to crash. Currently, we are only able // to provide an implementation for non-async functions. #[async_trait] impl<Stream, DecodeFn, Item, Error> Decode<Stream> for DecodeFn where Stream: StreamRecv + Send, DecodeFn: Fn(&mut BytesMut) -> Result<Option<Item>, Error> + Send, Item: Send, Error: From<Stream::Error> + Send, { type Item = Item; type Error = Error; async fn decode(&mut self, stream: &mut Stream) -> Result<Option<Self::Item>, Self::Error> { loop { if let Some(item) = self(&mut stream.buffer())? { return Ok(Some(item)); } if stream.recv_or_eof().await?.is_none() { if stream.buffer().is_empty() { return Ok(None); } else { // Return the `UnexpectedEof` error raised by the `recv` function. return Err(stream.recv().await.expect_err("expect EOF").into()); } } } } } impl<EncodeFn, Item> Encode<Item> for EncodeFn where EncodeFn: Fn(&Item, &mut BytesMut), { fn encode(&mut self, item: &Item, buffer: &mut BytesMut) { self(item, buffer) } } // // Implementer's notes: We store future values in `Source` and `Sink`. These future values have to // satisfy the `'static` lifetime bound because trait methods like `Stream::poll_next` do not take // a lifetime parameter on `&mut Self`. To satisfy this, when producing a future value, we move // all related values into it, and they will be moved back after it is completed (so that we can // produce the next future value). // // There may be other ways to satisfy the `'static` lifetime bound, but for now, this "move" trick // is the best I have. // /// Byte Stream to `futures::stream::Stream` Adapter #[derive(DebugExt)] pub struct Source<Stream, Decoder> where Stream: StreamRecv, Decoder: Decode<Stream>, { #[debug(with = InsertPlaceholder)] source: Option<(Stream, Decoder)>, #[debug(with = InsertPlaceholder)] next_future: Option<SourceFuture<Stream, Decoder, Decoder::Item, Decoder::Error>>, } /// `futures::sink::Sink` to Byte Stream Adapter #[derive(DebugExt)] pub struct Sink<Stream, Encoder> where Stream: StreamSend, { #[debug(with = InsertPlaceholder)] stream: Option<Stream>, #[debug(with = InsertPlaceholder)] encoder: Encoder, #[debug(with = InsertPlaceholder)] flush_future: Option<SinkFuture<Stream, Stream::Error>>, #[debug(with = InsertPlaceholder)] close_future: Option<SinkFuture<Stream, Stream::Error>>, } // TODO: Use where clauses to simplify these type aliases when rustc starts enforcing where clauses // in type aliases. For more details, check [rust-lang/rust#21903][#21903]. // // [#21903]: https://github.com/rust-lang/rust/issues/21903 type SourceFuture<Stream, Decoder, Item, Error> = BoxFuture<'static, SourceOutput<Stream, Decoder, Item, Error>>; type SourceOutput<Stream, Decoder, Item, Error> = ((Stream, Decoder), Option<Result<Item, Error>>); type SinkFuture<Stream, Error> = BoxFuture<'static, SinkOutput<Stream, Error>>; type SinkOutput<Stream, Error> = (Stream, Result<(), Error>); macro_rules! poll { ($this:ident, $get_future:ident, $context:ident $(,)?) => { match $this.$get_future().as_mut().poll($context) { Poll::Ready((state, result)) => { $this.reset(state); Poll::Ready(result) } Poll::Pending => Poll::Pending, } }; } impl<Stream, Decoder> Source<Stream, Decoder> where Stream: StreamRecv, Decoder: Decode<Stream>, { pub fn new(stream: Stream, decoder: Decoder) -> Self { Self { source: Some((stream, decoder)), next_future: None, } } fn reset(&mut self, source: (Stream, Decoder)) { self.source = Some(source); self.next_future = None; } } impl<Stream, Decoder> Source<Stream, Decoder> where Stream: StreamRecv + Send + 'static, Decoder: Decode<Stream> + Send + 'static, { fn next_future(&mut self) -> &mut SourceFuture<Stream, Decoder, Decoder::Item, Decoder::Error> { self.next_future .get_or_insert_with(|| Box::pin(Self::next(self.source.take().unwrap()))) } async fn next( (mut stream, mut decoder): (Stream, Decoder), ) -> SourceOutput<Stream, Decoder, Decoder::Item, Decoder::Error> { let result = decoder.decode(&mut stream).await.transpose(); ((stream, decoder), result) } } impl<Stream, Decoder> stream::Stream for Source<Stream, Decoder> where Stream: StreamRecv + Send + Unpin + 'static, Decoder: Decode<Stream> + Send + Unpin + 'static, { type Item = Result<Decoder::Item, Decoder::Error>; fn poll_next(self: Pin<&mut Self>, context: &mut Context<'_>) -> Poll<Option<Self::Item>> { let this = self.get_mut(); poll!(this, next_future, context) } } impl<Stream, Encoder> Sink<Stream, Encoder> where Stream: StreamSend, { pub fn new(stream: Stream, encoder: Encoder) -> Self { Self { stream: Some(stream), encoder, flush_future: None, close_future: None, } } fn reset(&mut self, stream: Stream) { self.stream = Some(stream); self.flush_future = None; self.close_future = None; } } impl<Stream, Encoder> Sink<Stream, Encoder> where Stream: StreamSend + Send + 'static, Encoder: 'static, { fn flush_future(&mut self) -> &mut SinkFuture<Stream, Stream::Error> { self.flush_future .get_or_insert_with(|| Box::pin(Self::flush(self.stream.take().unwrap()))) } async fn flush(mut stream: Stream) -> SinkOutput<Stream, Stream::Error> { let result = stream.send_all().await; (stream, result) } fn close_future(&mut self) -> &mut SinkFuture<Stream, Stream::Error> { self.close_future .get_or_insert_with(|| Box::pin(Self::close(self.stream.take().unwrap()))) } async fn close(mut stream: Stream) -> SinkOutput<Stream, Stream::Error> { let result = stream.shutdown().await; (stream, result) } } impl<Stream, Encoder, Item> sink::Sink<Item> for Sink<Stream, Encoder> where Stream: StreamSend + Send + Unpin + 'static, Encoder: Encode<Item> + Unpin + 'static, { type Error = Stream::Error; fn poll_ready(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<Result<(), Self::Error>> { match self.get_mut().stream { Some(_) => Poll::Ready(Ok(())), None => Poll::Pending, } } fn start_send(self: Pin<&mut Self>, item: Item) -> Result<(), Self::Error> { let this = self.get_mut(); this.encoder .encode(&item, &mut this.stream.as_mut().unwrap().buffer()); Ok(()) } fn poll_flush( self: Pin<&mut Self>, context: &mut Context<'_>, ) -> Poll<Result<(), Self::Error>> { let this = self.get_mut(); poll!(this, flush_future, context) } fn poll_close( self: Pin<&mut Self>, context: &mut Context<'_>, ) -> Poll<Result<(), Self::Error>> { let this = self.get_mut(); poll!(this, close_future, context) } } #[cfg(test)] mod tests { use std::assert_matches::assert_matches; use std::io::{Error, ErrorKind}; use bytes::{Buf, BufMut}; use futures::{sink::SinkExt, stream::StreamExt}; use tokio::io::{AsyncReadExt, AsyncWriteExt}; use crate::io::{RecvStream, SendStream}; use super::*; fn decode(buffer: &mut BytesMut) -> Result<Option<String>, Error> { if buffer.remaining() < 1 { return Ok(None); } let size = usize::from(buffer[0]); if buffer.remaining() < 1 + size { return Ok(None); } let mut vec = vec![0u8; size]; buffer.get_u8(); buffer.copy_to_slice(&mut vec); Ok(Some(String::from_utf8(vec).unwrap())) } #[tokio::test] async fn source() { let (stream, mut mock) = RecvStream::new_mock(4096); let mut source = Source::new(stream, decode); mock.write_all(b"\x0bhello world\x03foo\x03bar") .await .unwrap(); drop(mock); assert_matches!(source.next().await, Some(Ok(x)) if x == "hello world"); assert_matches!(source.next().await, Some(Ok(x)) if x == "foo"); assert_matches!(source.next().await, Some(Ok(x)) if x == "bar"); assert_matches!(source.next().await, None); } #[tokio::test] async fn source_unexpected_eof() { let (stream, mut mock) = RecvStream::new_mock(4096); let mut source = Source::new(stream, decode); mock.write_all(b"\x0bhello").await.unwrap(); drop(mock); assert_matches!(source.next().await, Some(Err(e)) if e.kind() == ErrorKind::UnexpectedEof); }
buffer.put_u8(item.len().try_into().unwrap()); buffer.put_slice(item.as_bytes()); } #[tokio::test] async fn sink() { let (stream, mut mock) = SendStream::new_mock(4096); let mut sink = Sink::new(stream, encode); assert_matches!(sink.feed("hello world".to_string()).await, Ok(())); assert_matches!(sink.flush().await, Ok(())); let mut buffer = BytesMut::new(); assert_matches!(mock.read_buf(&mut buffer).await, Ok(12)); assert_eq!(buffer.as_ref(), b"\x0bhello world"); assert_matches!(sink.feed("foo".to_string()).await, Ok(())); assert_matches!(sink.feed("bar".to_string()).await, Ok(())); assert_matches!(sink.close().await, Ok(())); buffer.clear(); assert_matches!(mock.read_buf(&mut buffer).await, Ok(8)); assert_eq!(buffer.as_ref(), b"\x03foo\x03bar"); } }
fn encode(item: &String, buffer: &mut BytesMut) {
random_line_split
codec.rs
use std::pin::Pin; use std::task::{Context, Poll}; use async_trait::async_trait; use bytes::BytesMut; // We use `futures::stream::Stream` because `std::async_iter::AsyncIterator` (previously known as // `std::stream::Stream`) is still a bare-bones implementation - it does not even have an // `async fn next` method! use futures::{future::BoxFuture, sink, stream}; use g1_base::fmt::{DebugExt, InsertPlaceholder}; use super::{StreamRecv, StreamSend}; #[async_trait] pub trait Decode<Stream> where Stream: StreamRecv, { type Item; type Error: From<Stream::Error>; /// Decodes a byte stream and returns a stream item or an error when EOF is reached, but the /// byte stream buffer is not empty. async fn decode(&mut self, stream: &mut Stream) -> Result<Option<Self::Item>, Self::Error>; } pub trait Encode<Item> { /// Encodes a stream item and writes the output to a byte stream buffer. fn encode(&mut self, item: &Item, buffer: &mut BytesMut); } // TODO: We intend to implement `Decode` for async functions, similar to what we do for `Encode`. // However, for unknown reasons, it is causing the compiler to crash. Currently, we are only able // to provide an implementation for non-async functions. #[async_trait] impl<Stream, DecodeFn, Item, Error> Decode<Stream> for DecodeFn where Stream: StreamRecv + Send, DecodeFn: Fn(&mut BytesMut) -> Result<Option<Item>, Error> + Send, Item: Send, Error: From<Stream::Error> + Send, { type Item = Item; type Error = Error; async fn decode(&mut self, stream: &mut Stream) -> Result<Option<Self::Item>, Self::Error> { loop { if let Some(item) = self(&mut stream.buffer())? { return Ok(Some(item)); } if stream.recv_or_eof().await?.is_none() { if stream.buffer().is_empty() { return Ok(None); } else { // Return the `UnexpectedEof` error raised by the `recv` function. return Err(stream.recv().await.expect_err("expect EOF").into()); } } } } } impl<EncodeFn, Item> Encode<Item> for EncodeFn where EncodeFn: Fn(&Item, &mut BytesMut), { fn encode(&mut self, item: &Item, buffer: &mut BytesMut) { self(item, buffer) } } // // Implementer's notes: We store future values in `Source` and `Sink`. These future values have to // satisfy the `'static` lifetime bound because trait methods like `Stream::poll_next` do not take // a lifetime parameter on `&mut Self`. To satisfy this, when producing a future value, we move // all related values into it, and they will be moved back after it is completed (so that we can // produce the next future value). // // There may be other ways to satisfy the `'static` lifetime bound, but for now, this "move" trick // is the best I have. // /// Byte Stream to `futures::stream::Stream` Adapter #[derive(DebugExt)] pub struct Source<Stream, Decoder> where Stream: StreamRecv, Decoder: Decode<Stream>, { #[debug(with = InsertPlaceholder)] source: Option<(Stream, Decoder)>, #[debug(with = InsertPlaceholder)] next_future: Option<SourceFuture<Stream, Decoder, Decoder::Item, Decoder::Error>>, } /// `futures::sink::Sink` to Byte Stream Adapter #[derive(DebugExt)] pub struct Sink<Stream, Encoder> where Stream: StreamSend, { #[debug(with = InsertPlaceholder)] stream: Option<Stream>, #[debug(with = InsertPlaceholder)] encoder: Encoder, #[debug(with = InsertPlaceholder)] flush_future: Option<SinkFuture<Stream, Stream::Error>>, #[debug(with = InsertPlaceholder)] close_future: Option<SinkFuture<Stream, Stream::Error>>, } // TODO: Use where clauses to simplify these type aliases when rustc starts enforcing where clauses // in type aliases. For more details, check [rust-lang/rust#21903][#21903]. // // [#21903]: https://github.com/rust-lang/rust/issues/21903 type SourceFuture<Stream, Decoder, Item, Error> = BoxFuture<'static, SourceOutput<Stream, Decoder, Item, Error>>; type SourceOutput<Stream, Decoder, Item, Error> = ((Stream, Decoder), Option<Result<Item, Error>>); type SinkFuture<Stream, Error> = BoxFuture<'static, SinkOutput<Stream, Error>>; type SinkOutput<Stream, Error> = (Stream, Result<(), Error>); macro_rules! poll { ($this:ident, $get_future:ident, $context:ident $(,)?) => { match $this.$get_future().as_mut().poll($context) { Poll::Ready((state, result)) => { $this.reset(state); Poll::Ready(result) } Poll::Pending => Poll::Pending, } }; } impl<Stream, Decoder> Source<Stream, Decoder> where Stream: StreamRecv, Decoder: Decode<Stream>, { pub fn new(stream: Stream, decoder: Decoder) -> Self { Self { source: Some((stream, decoder)), next_future: None, } } fn reset(&mut self, source: (Stream, Decoder)) { self.source = Some(source); self.next_future = None; } } impl<Stream, Decoder> Source<Stream, Decoder> where Stream: StreamRecv + Send + 'static, Decoder: Decode<Stream> + Send + 'static, { fn next_future(&mut self) -> &mut SourceFuture<Stream, Decoder, Decoder::Item, Decoder::Error> { self.next_future .get_or_insert_with(|| Box::pin(Self::next(self.source.take().unwrap()))) } async fn next( (mut stream, mut decoder): (Stream, Decoder), ) -> SourceOutput<Stream, Decoder, Decoder::Item, Decoder::Error> { let result = decoder.decode(&mut stream).await.transpose(); ((stream, decoder), result) } } impl<Stream, Decoder> stream::Stream for Source<Stream, Decoder> where Stream: StreamRecv + Send + Unpin + 'static, Decoder: Decode<Stream> + Send + Unpin + 'static, { type Item = Result<Decoder::Item, Decoder::Error>; fn poll_next(self: Pin<&mut Self>, context: &mut Context<'_>) -> Poll<Option<Self::Item>> { let this = self.get_mut(); poll!(this, next_future, context) } } impl<Stream, Encoder> Sink<Stream, Encoder> where Stream: StreamSend, { pub fn new(stream: Stream, encoder: Encoder) -> Self { Self { stream: Some(stream), encoder, flush_future: None, close_future: None, } } fn reset(&mut self, stream: Stream) { self.stream = Some(stream); self.flush_future = None; self.close_future = None; } } impl<Stream, Encoder> Sink<Stream, Encoder> where Stream: StreamSend + Send + 'static, Encoder: 'static, { fn flush_future(&mut self) -> &mut SinkFuture<Stream, Stream::Error> { self.flush_future .get_or_insert_with(|| Box::pin(Self::flush(self.stream.take().unwrap()))) } async fn flush(mut stream: Stream) -> SinkOutput<Stream, Stream::Error> { let result = stream.send_all().await; (stream, result) } fn close_future(&mut self) -> &mut SinkFuture<Stream, Stream::Error> { self.close_future .get_or_insert_with(|| Box::pin(Self::close(self.stream.take().unwrap()))) } async fn close(mut stream: Stream) -> SinkOutput<Stream, Stream::Error> { let result = stream.shutdown().await; (stream, result) } } impl<Stream, Encoder, Item> sink::Sink<Item> for Sink<Stream, Encoder> where Stream: StreamSend + Send + Unpin + 'static, Encoder: Encode<Item> + Unpin + 'static, { type Error = Stream::Error; fn poll_ready(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<Result<(), Self::Error>> { match self.get_mut().stream { Some(_) => Poll::Ready(Ok(())), None => Poll::Pending, } } fn start_send(self: Pin<&mut Self>, item: Item) -> Result<(), Self::Error> { let this = self.get_mut(); this.encoder .encode(&item, &mut this.stream.as_mut().unwrap().buffer()); Ok(()) } fn poll_flush( self: Pin<&mut Self>, context: &mut Context<'_>, ) -> Poll<Result<(), Self::Error>> { let this = self.get_mut(); poll!(this, flush_future, context) } fn poll_close( self: Pin<&mut Self>, context: &mut Context<'_>, ) -> Poll<Result<(), Self::Error>> { let this = self.get_mut(); poll!(this, close_future, context) } } #[cfg(test)] mod tests { use std::assert_matches::assert_matches; use std::io::{Error, ErrorKind}; use bytes::{Buf, BufMut}; use futures::{sink::SinkExt, stream::StreamExt}; use tokio::io::{AsyncReadExt, AsyncWriteExt}; use crate::io::{RecvStream, SendStream}; use super::*; fn decode(buffer: &mut BytesMut) -> Result<Option<String>, Error>
#[tokio::test] async fn source() { let (stream, mut mock) = RecvStream::new_mock(4096); let mut source = Source::new(stream, decode); mock.write_all(b"\x0bhello world\x03foo\x03bar") .await .unwrap(); drop(mock); assert_matches!(source.next().await, Some(Ok(x)) if x == "hello world"); assert_matches!(source.next().await, Some(Ok(x)) if x == "foo"); assert_matches!(source.next().await, Some(Ok(x)) if x == "bar"); assert_matches!(source.next().await, None); } #[tokio::test] async fn source_unexpected_eof() { let (stream, mut mock) = RecvStream::new_mock(4096); let mut source = Source::new(stream, decode); mock.write_all(b"\x0bhello").await.unwrap(); drop(mock); assert_matches!(source.next().await, Some(Err(e)) if e.kind() == ErrorKind::UnexpectedEof); } fn encode(item: &String, buffer: &mut BytesMut) { buffer.put_u8(item.len().try_into().unwrap()); buffer.put_slice(item.as_bytes()); } #[tokio::test] async fn sink() { let (stream, mut mock) = SendStream::new_mock(4096); let mut sink = Sink::new(stream, encode); assert_matches!(sink.feed("hello world".to_string()).await, Ok(())); assert_matches!(sink.flush().await, Ok(())); let mut buffer = BytesMut::new(); assert_matches!(mock.read_buf(&mut buffer).await, Ok(12)); assert_eq!(buffer.as_ref(), b"\x0bhello world"); assert_matches!(sink.feed("foo".to_string()).await, Ok(())); assert_matches!(sink.feed("bar".to_string()).await, Ok(())); assert_matches!(sink.close().await, Ok(())); buffer.clear(); assert_matches!(mock.read_buf(&mut buffer).await, Ok(8)); assert_eq!(buffer.as_ref(), b"\x03foo\x03bar"); } }
{ if buffer.remaining() < 1 { return Ok(None); } let size = usize::from(buffer[0]); if buffer.remaining() < 1 + size { return Ok(None); } let mut vec = vec![0u8; size]; buffer.get_u8(); buffer.copy_to_slice(&mut vec); Ok(Some(String::from_utf8(vec).unwrap())) }
identifier_body
codec.rs
use std::pin::Pin; use std::task::{Context, Poll}; use async_trait::async_trait; use bytes::BytesMut; // We use `futures::stream::Stream` because `std::async_iter::AsyncIterator` (previously known as // `std::stream::Stream`) is still a bare-bones implementation - it does not even have an // `async fn next` method! use futures::{future::BoxFuture, sink, stream}; use g1_base::fmt::{DebugExt, InsertPlaceholder}; use super::{StreamRecv, StreamSend}; #[async_trait] pub trait Decode<Stream> where Stream: StreamRecv, { type Item; type Error: From<Stream::Error>; /// Decodes a byte stream and returns a stream item or an error when EOF is reached, but the /// byte stream buffer is not empty. async fn decode(&mut self, stream: &mut Stream) -> Result<Option<Self::Item>, Self::Error>; } pub trait Encode<Item> { /// Encodes a stream item and writes the output to a byte stream buffer. fn encode(&mut self, item: &Item, buffer: &mut BytesMut); } // TODO: We intend to implement `Decode` for async functions, similar to what we do for `Encode`. // However, for unknown reasons, it is causing the compiler to crash. Currently, we are only able // to provide an implementation for non-async functions. #[async_trait] impl<Stream, DecodeFn, Item, Error> Decode<Stream> for DecodeFn where Stream: StreamRecv + Send, DecodeFn: Fn(&mut BytesMut) -> Result<Option<Item>, Error> + Send, Item: Send, Error: From<Stream::Error> + Send, { type Item = Item; type Error = Error; async fn
(&mut self, stream: &mut Stream) -> Result<Option<Self::Item>, Self::Error> { loop { if let Some(item) = self(&mut stream.buffer())? { return Ok(Some(item)); } if stream.recv_or_eof().await?.is_none() { if stream.buffer().is_empty() { return Ok(None); } else { // Return the `UnexpectedEof` error raised by the `recv` function. return Err(stream.recv().await.expect_err("expect EOF").into()); } } } } } impl<EncodeFn, Item> Encode<Item> for EncodeFn where EncodeFn: Fn(&Item, &mut BytesMut), { fn encode(&mut self, item: &Item, buffer: &mut BytesMut) { self(item, buffer) } } // // Implementer's notes: We store future values in `Source` and `Sink`. These future values have to // satisfy the `'static` lifetime bound because trait methods like `Stream::poll_next` do not take // a lifetime parameter on `&mut Self`. To satisfy this, when producing a future value, we move // all related values into it, and they will be moved back after it is completed (so that we can // produce the next future value). // // There may be other ways to satisfy the `'static` lifetime bound, but for now, this "move" trick // is the best I have. // /// Byte Stream to `futures::stream::Stream` Adapter #[derive(DebugExt)] pub struct Source<Stream, Decoder> where Stream: StreamRecv, Decoder: Decode<Stream>, { #[debug(with = InsertPlaceholder)] source: Option<(Stream, Decoder)>, #[debug(with = InsertPlaceholder)] next_future: Option<SourceFuture<Stream, Decoder, Decoder::Item, Decoder::Error>>, } /// `futures::sink::Sink` to Byte Stream Adapter #[derive(DebugExt)] pub struct Sink<Stream, Encoder> where Stream: StreamSend, { #[debug(with = InsertPlaceholder)] stream: Option<Stream>, #[debug(with = InsertPlaceholder)] encoder: Encoder, #[debug(with = InsertPlaceholder)] flush_future: Option<SinkFuture<Stream, Stream::Error>>, #[debug(with = InsertPlaceholder)] close_future: Option<SinkFuture<Stream, Stream::Error>>, } // TODO: Use where clauses to simplify these type aliases when rustc starts enforcing where clauses // in type aliases. For more details, check [rust-lang/rust#21903][#21903]. // // [#21903]: https://github.com/rust-lang/rust/issues/21903 type SourceFuture<Stream, Decoder, Item, Error> = BoxFuture<'static, SourceOutput<Stream, Decoder, Item, Error>>; type SourceOutput<Stream, Decoder, Item, Error> = ((Stream, Decoder), Option<Result<Item, Error>>); type SinkFuture<Stream, Error> = BoxFuture<'static, SinkOutput<Stream, Error>>; type SinkOutput<Stream, Error> = (Stream, Result<(), Error>); macro_rules! poll { ($this:ident, $get_future:ident, $context:ident $(,)?) => { match $this.$get_future().as_mut().poll($context) { Poll::Ready((state, result)) => { $this.reset(state); Poll::Ready(result) } Poll::Pending => Poll::Pending, } }; } impl<Stream, Decoder> Source<Stream, Decoder> where Stream: StreamRecv, Decoder: Decode<Stream>, { pub fn new(stream: Stream, decoder: Decoder) -> Self { Self { source: Some((stream, decoder)), next_future: None, } } fn reset(&mut self, source: (Stream, Decoder)) { self.source = Some(source); self.next_future = None; } } impl<Stream, Decoder> Source<Stream, Decoder> where Stream: StreamRecv + Send + 'static, Decoder: Decode<Stream> + Send + 'static, { fn next_future(&mut self) -> &mut SourceFuture<Stream, Decoder, Decoder::Item, Decoder::Error> { self.next_future .get_or_insert_with(|| Box::pin(Self::next(self.source.take().unwrap()))) } async fn next( (mut stream, mut decoder): (Stream, Decoder), ) -> SourceOutput<Stream, Decoder, Decoder::Item, Decoder::Error> { let result = decoder.decode(&mut stream).await.transpose(); ((stream, decoder), result) } } impl<Stream, Decoder> stream::Stream for Source<Stream, Decoder> where Stream: StreamRecv + Send + Unpin + 'static, Decoder: Decode<Stream> + Send + Unpin + 'static, { type Item = Result<Decoder::Item, Decoder::Error>; fn poll_next(self: Pin<&mut Self>, context: &mut Context<'_>) -> Poll<Option<Self::Item>> { let this = self.get_mut(); poll!(this, next_future, context) } } impl<Stream, Encoder> Sink<Stream, Encoder> where Stream: StreamSend, { pub fn new(stream: Stream, encoder: Encoder) -> Self { Self { stream: Some(stream), encoder, flush_future: None, close_future: None, } } fn reset(&mut self, stream: Stream) { self.stream = Some(stream); self.flush_future = None; self.close_future = None; } } impl<Stream, Encoder> Sink<Stream, Encoder> where Stream: StreamSend + Send + 'static, Encoder: 'static, { fn flush_future(&mut self) -> &mut SinkFuture<Stream, Stream::Error> { self.flush_future .get_or_insert_with(|| Box::pin(Self::flush(self.stream.take().unwrap()))) } async fn flush(mut stream: Stream) -> SinkOutput<Stream, Stream::Error> { let result = stream.send_all().await; (stream, result) } fn close_future(&mut self) -> &mut SinkFuture<Stream, Stream::Error> { self.close_future .get_or_insert_with(|| Box::pin(Self::close(self.stream.take().unwrap()))) } async fn close(mut stream: Stream) -> SinkOutput<Stream, Stream::Error> { let result = stream.shutdown().await; (stream, result) } } impl<Stream, Encoder, Item> sink::Sink<Item> for Sink<Stream, Encoder> where Stream: StreamSend + Send + Unpin + 'static, Encoder: Encode<Item> + Unpin + 'static, { type Error = Stream::Error; fn poll_ready(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<Result<(), Self::Error>> { match self.get_mut().stream { Some(_) => Poll::Ready(Ok(())), None => Poll::Pending, } } fn start_send(self: Pin<&mut Self>, item: Item) -> Result<(), Self::Error> { let this = self.get_mut(); this.encoder .encode(&item, &mut this.stream.as_mut().unwrap().buffer()); Ok(()) } fn poll_flush( self: Pin<&mut Self>, context: &mut Context<'_>, ) -> Poll<Result<(), Self::Error>> { let this = self.get_mut(); poll!(this, flush_future, context) } fn poll_close( self: Pin<&mut Self>, context: &mut Context<'_>, ) -> Poll<Result<(), Self::Error>> { let this = self.get_mut(); poll!(this, close_future, context) } } #[cfg(test)] mod tests { use std::assert_matches::assert_matches; use std::io::{Error, ErrorKind}; use bytes::{Buf, BufMut}; use futures::{sink::SinkExt, stream::StreamExt}; use tokio::io::{AsyncReadExt, AsyncWriteExt}; use crate::io::{RecvStream, SendStream}; use super::*; fn decode(buffer: &mut BytesMut) -> Result<Option<String>, Error> { if buffer.remaining() < 1 { return Ok(None); } let size = usize::from(buffer[0]); if buffer.remaining() < 1 + size { return Ok(None); } let mut vec = vec![0u8; size]; buffer.get_u8(); buffer.copy_to_slice(&mut vec); Ok(Some(String::from_utf8(vec).unwrap())) } #[tokio::test] async fn source() { let (stream, mut mock) = RecvStream::new_mock(4096); let mut source = Source::new(stream, decode); mock.write_all(b"\x0bhello world\x03foo\x03bar") .await .unwrap(); drop(mock); assert_matches!(source.next().await, Some(Ok(x)) if x == "hello world"); assert_matches!(source.next().await, Some(Ok(x)) if x == "foo"); assert_matches!(source.next().await, Some(Ok(x)) if x == "bar"); assert_matches!(source.next().await, None); } #[tokio::test] async fn source_unexpected_eof() { let (stream, mut mock) = RecvStream::new_mock(4096); let mut source = Source::new(stream, decode); mock.write_all(b"\x0bhello").await.unwrap(); drop(mock); assert_matches!(source.next().await, Some(Err(e)) if e.kind() == ErrorKind::UnexpectedEof); } fn encode(item: &String, buffer: &mut BytesMut) { buffer.put_u8(item.len().try_into().unwrap()); buffer.put_slice(item.as_bytes()); } #[tokio::test] async fn sink() { let (stream, mut mock) = SendStream::new_mock(4096); let mut sink = Sink::new(stream, encode); assert_matches!(sink.feed("hello world".to_string()).await, Ok(())); assert_matches!(sink.flush().await, Ok(())); let mut buffer = BytesMut::new(); assert_matches!(mock.read_buf(&mut buffer).await, Ok(12)); assert_eq!(buffer.as_ref(), b"\x0bhello world"); assert_matches!(sink.feed("foo".to_string()).await, Ok(())); assert_matches!(sink.feed("bar".to_string()).await, Ok(())); assert_matches!(sink.close().await, Ok(())); buffer.clear(); assert_matches!(mock.read_buf(&mut buffer).await, Ok(8)); assert_eq!(buffer.as_ref(), b"\x03foo\x03bar"); } }
decode
identifier_name
issued_certificate.pb.go
// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto package v1alpha2 import ( bytes "bytes" fmt "fmt" math "math" _ "github.com/gogo/protobuf/gogoproto" proto "github.com/gogo/protobuf/proto" _ "github.com/gogo/protobuf/types" v1 "github.com/solo-io/skv2/pkg/api/core.skv2.solo.io/v1" ) // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal var _ = fmt.Errorf var _ = math.Inf // This is a compile-time assertion to ensure that this generated file // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package // Possible states in which an IssuedCertificate can exist. type IssuedCertificateStatus_State int32 const ( // The IssuedCertificate has yet to be picked up by the agent. IssuedCertificateStatus_PENDING IssuedCertificateStatus_State = 0 // The agent has created a local private key // and a CertificateRequest for the IssuedCertificate. // In this state, the agent is waiting for the Issuer // to issue certificates for the CertificateRequest before proceeding. IssuedCertificateStatus_REQUESTED IssuedCertificateStatus_State = 1 // The certificate has been issued. Any pods that require restarting will be restarted at this point. IssuedCertificateStatus_ISSUED IssuedCertificateStatus_State = 2 // The reply from the Issuer has been processed and // the agent has placed the final certificate secret // in the target location specified by the IssuedCertificate. IssuedCertificateStatus_FINISHED IssuedCertificateStatus_State = 3 // Processing the certificate workflow failed. IssuedCertificateStatus_FAILED IssuedCertificateStatus_State = 4 ) var IssuedCertificateStatus_State_name = map[int32]string{ 0: "PENDING", 1: "REQUESTED", 2: "ISSUED", 3: "FINISHED", 4: "FAILED", } var IssuedCertificateStatus_State_value = map[string]int32{ "PENDING": 0, "REQUESTED": 1, "ISSUED": 2, "FINISHED": 3, "FAILED": 4, } func (x IssuedCertificateStatus_State) String() string { return proto.EnumName(IssuedCertificateStatus_State_name, int32(x)) } func (IssuedCertificateStatus_State) EnumDescriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1, 0} } // //IssuedCertificates are used to issue SSL certificates //to remote Kubernetes clusters from a central (out-of-cluster) Certificate Authority. // //When an IssuedCertificate is created, a certificate is issued to a remote cluster by a central Certificate Authority via //the following workflow: //- The Certificate Issuer creates the IssuedCertificate resource on the remote cluster //- The Certificate Signature Requesting Agent installed to the remote cluster generates //a Certificate Signing Request and writes it to the status of the IssuedCertificate //- Finally, the Certificate Issuer generates signed a certificate for the CSR and writes //it back as Secret in the remote cluster. // //Shared trust can therefore be established across clusters without requiring //private keys to ever leave the node. type IssuedCertificateSpec struct { // //A list of hostnames and IPs to generate a certificate for. //This can also be set to the identity running the workload, //e.g. a Kubernetes service account. // //Generally for an Istio CA this will take the form `spiffe://cluster.local/ns/istio-system/sa/citadel`. // //"cluster.local" may be replaced by the root of trust domain for the mesh. Hosts []string `protobuf:"bytes,1,rep,name=hosts,proto3" json:"hosts,omitempty"` // The organization for this certificate. Org string `protobuf:"bytes,2,opt,name=org,proto3" json:"org,omitempty"` // The secret containing the root SSL certificate used to sign this IssuedCertificate (located in the Certificate Issuer's cluster). SigningCertificateSecret *v1.ObjectRef `protobuf:"bytes,3,opt,name=signing_certificate_secret,json=signingCertificateSecret,proto3" json:"signing_certificate_secret,omitempty"` // The secret containing the SSL certificate to be generated for this IssuedCertificate (located in the Certificate Agent's cluster). IssuedCertificateSecret *v1.ObjectRef `protobuf:"bytes,4,opt,name=issued_certificate_secret,json=issuedCertificateSecret,proto3" json:"issued_certificate_secret,omitempty"` // A ref to a PodBounceDirective specifying a list of k8s pods to bounce // (delete and cause a restart) when the certificate is issued. // This will include the control plane pods as well as any pods // which share a data plane with the target mesh. PodBounceDirective *v1.ObjectRef `protobuf:"bytes,5,opt,name=pod_bounce_directive,json=podBounceDirective,proto3" json:"pod_bounce_directive,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateSpec) Reset() { *m = IssuedCertificateSpec{} } func (m *IssuedCertificateSpec) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateSpec) ProtoMessage() {} func (*IssuedCertificateSpec) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{0} } func (m *IssuedCertificateSpec) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateSpec.Unmarshal(m, b) } func (m *IssuedCertificateSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateSpec.Marshal(b, m, deterministic) }
func (m *IssuedCertificateSpec) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateSpec.Merge(m, src) } func (m *IssuedCertificateSpec) XXX_Size() int { return xxx_messageInfo_IssuedCertificateSpec.Size(m) } func (m *IssuedCertificateSpec) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateSpec.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateSpec proto.InternalMessageInfo func (m *IssuedCertificateSpec) GetHosts() []string { if m != nil { return m.Hosts } return nil } func (m *IssuedCertificateSpec) GetOrg() string { if m != nil { return m.Org } return "" } func (m *IssuedCertificateSpec) GetSigningCertificateSecret() *v1.ObjectRef { if m != nil { return m.SigningCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetIssuedCertificateSecret() *v1.ObjectRef { if m != nil { return m.IssuedCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetPodBounceDirective() *v1.ObjectRef { if m != nil { return m.PodBounceDirective } return nil } // The IssuedCertificate status is written by the CertificateRequesting agent. type IssuedCertificateStatus struct { // The most recent generation observed in the the IssuedCertificate metadata. // If the observedGeneration does not match generation, the Certificate Requesting Agent has not processed the most // recent version of this IssuedCertificate. ObservedGeneration int64 `protobuf:"varint,1,opt,name=observed_generation,json=observedGeneration,proto3" json:"observed_generation,omitempty"` // Any error observed which prevented the CertificateRequest from being processed. // If the error is empty, the request has been processed successfully. Error string `protobuf:"bytes,2,opt,name=error,proto3" json:"error,omitempty"` // The current state of the IssuedCertificate workflow, reported by the agent. State IssuedCertificateStatus_State `protobuf:"varint,3,opt,name=state,proto3,enum=certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateStatus) Reset() { *m = IssuedCertificateStatus{} } func (m *IssuedCertificateStatus) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateStatus) ProtoMessage() {} func (*IssuedCertificateStatus) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1} } func (m *IssuedCertificateStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateStatus.Unmarshal(m, b) } func (m *IssuedCertificateStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateStatus.Marshal(b, m, deterministic) } func (m *IssuedCertificateStatus) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateStatus.Merge(m, src) } func (m *IssuedCertificateStatus) XXX_Size() int { return xxx_messageInfo_IssuedCertificateStatus.Size(m) } func (m *IssuedCertificateStatus) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateStatus.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateStatus proto.InternalMessageInfo func (m *IssuedCertificateStatus) GetObservedGeneration() int64 { if m != nil { return m.ObservedGeneration } return 0 } func (m *IssuedCertificateStatus) GetError() string { if m != nil { return m.Error } return "" } func (m *IssuedCertificateStatus) GetState() IssuedCertificateStatus_State { if m != nil { return m.State } return IssuedCertificateStatus_PENDING } func init() { proto.RegisterEnum("certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State", IssuedCertificateStatus_State_name, IssuedCertificateStatus_State_value) proto.RegisterType((*IssuedCertificateSpec)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateSpec") proto.RegisterType((*IssuedCertificateStatus)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateStatus") } func init() { proto.RegisterFile("github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto", fileDescriptor_86ade12c22739639) } var fileDescriptor_86ade12c22739639 = []byte{ // 479 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x53, 0x5d, 0x6f, 0xd3, 0x30, 0x14, 0x25, 0xed, 0x3a, 0xa8, 0x07, 0x28, 0x32, 0x45, 0x0b, 0x05, 0x95, 0xaa, 0x4f, 0x7d, 0x99, 0xad, 0x95, 0x67, 0x1e, 0x18, 0x49, 0x47, 0xa4, 0xa9, 0x40, 0xc2, 0x24, 0xb4, 0x97, 0x2a, 0x1f, 0xb7, 0xae, 0x59, 0xdb, 0x1b, 0xd9, 0x4e, 0x9f, 0xf9, 0x39, 0xfc, 0x11, 0xfe, 0x08, 0xbf, 0x04, 0xc5, 0x69, 0xa1, 0x22, 0x62, 0xda, 0x53, 0xee, 0xf5, 0xbd, 0xe7, 0xe8, 0x9c, 0x13, 0x9b, 0x4c, 0x85, 0x34, 0xcb, 0x32, 0x65, 0x19, 0xae, 0xb9, 0xc6, 0x15, 0x9e, 0x49, 0xe4, 0x62, 0x85, 0x78, 0xb6, 0x06, 0xbd, 0xe4, 0x49, 0x21, 0x79, 0x06, 0xca, 0xc8, 0x85, 0xcc, 0x12, 0x03, 0x9a, 0x4b, 0xad, 0x4b, 0xc8, 0xe7, 0x07, 0x67, 0xac, 0x50, 0x68, 0x90, 0x0e, 0x0e, 0xd7, 0x58, 0x05, 0x66, 0x15, 0x0d, 0xab, 0x38, 0x99, 0xc4, 0xfe, 0x4b, 0x7d, 0xbb, 0x9d, 0xd4, 0x7c, 0xa8, 0x80, 0x6f, 0xcf, 0xed, 0xb7, 0x06, 0xf7, 0x5f, 0x0b, 0x44, 0xb1, 0x02, 0x6e, 0xbb, 0xb4, 0x5c, 0x70, 0x23, 0xd7, 0xa0, 0x4d, 0xb2, 0x2e, 0x76, 0x0b, 0x83, 0x7f, 0x17, 0xf2, 0x52, 0x25, 0x46, 0xe2, 0x66, 0x37, 0xef, 0x09, 0x14, 0x68, 0x4b, 0x5e, 0x55, 0xf5, 0xe9, 0xe8, 0x67, 0x8b, 0x3c, 0x0f, 0xad, 0xe0, 0xf7, 0x7f, 0xc5, 0xc5, 0x05, 0x64, 0xb4, 0x47, 0x3a, 0x4b, 0xd4, 0x46, 0x7b, 0xce, 0xb0, 0x3d, 0xee, 0x46, 0x75, 0x43, 0x5d, 0xd2, 0x46, 0x25, 0xbc, 0xd6, 0xd0, 0x19, 0x77, 0xa3, 0xaa, 0xa4, 0x37, 0xa4, 0xaf, 0xa5, 0xd8, 0xc8, 0x8d, 0x38, 0xb4, 0x3c, 0xd7, 0x90, 0x29, 0x30, 0x5e, 0x7b, 0xe8, 0x8c, 0x4f, 0x26, 0xaf, 0x98, 0x75, 0x52, 0xf9, 0xdb, 0xbb, 0x65, 0x1f, 0xd3, 0x6f, 0x90, 0x99, 0x08, 0x16, 0x91, 0xb7, 0xc3, 0x1f, 0x2a, 0xb0, 0x68, 0xfa, 0x95, 0xbc, 0x68, 0xa6, 0xb9, 0xa7, 0x3e, 0xba, 0x07, 0xf5, 0xa9, 0x6c, 0x78, 0xab, 0x99, 0x67, 0xa4, 0x57, 0x60, 0x3e, 0x4f, 0xb1, 0xdc, 0x64, 0x30, 0xcf, 0xa5, 0x82, 0xcc, 0xc8, 0x2d, 0x78, 0x9d, 0x7b, 0x90, 0xd2, 0x02, 0xf3, 0x0b, 0x0b, 0xf4, 0xf7, 0xb8, 0xd1, 0xf7, 0x16, 0x39, 0x6d, 0xe6, 0x68, 0x12, 0x53, 0x6a, 0xca, 0xc9, 0x33, 0x4c, 0x35, 0xa8, 0x2d, 0xe4, 0x73, 0x01, 0x1b, 0xa8, 0x7f, 0x8b, 0xe7, 0x0c, 0x9d, 0x71, 0x3b, 0xa2, 0xfb, 0xd1, 0xe5, 0x9f, 0x49, 0x15, 0x3d, 0x28, 0x85, 0x6a, 0x17, 0x73, 0xdd, 0xd0, 0x98, 0x74, 0xb4, 0x49, 0x0c, 0xd8, 0x4c, 0x9f, 0x4e, 0xde, 0xb2, 0xbb, 0xaf, 0x13, 0xfb, 0x8f, 0x1c, 0x56, 0x7d, 0x20, 0xaa, 0xb9, 0x46, 0x21, 0xe9, 0xd8, 0x9e, 0x9e, 0x90, 0x87, 0x9f, 0x82, 0x99, 0x1f, 0xce, 0x2e, 0xdd, 0x07, 0xf4, 0x09, 0xe9, 0x46, 0xc1, 0xe7, 0xeb, 0x20, 0xfe, 0x12, 0xf8, 0xae, 0x43, 0x09, 0x39, 0x0e, 0xe3, 0xf8, 0x3a, 0xf0, 0xdd, 0x16, 0x7d, 0x4c, 0x1e, 0x4d, 0xc3, 0x59, 0x18, 0x7f, 0x08, 0x7c, 0xb7, 0x5d, 0x4d, 0xa6, 0xef, 0xc2, 0xab, 0xc0, 0x77, 0x8f, 0x2e, 0xa2, 0x1f, 0xbf, 0x06, 0xce, 0xcd, 0xd5, 0x9d, 0x8f, 0xa5, 0xb8, 0x15, 0x8d, 0x07, 0xd3, 0x94, 0xce, 0xb7, 0xe7, 0xc9, 0xaa, 0x58, 0x26, 0x93, 0xf4, 0xd8, 0xde, 0xd2, 0x37, 0xbf, 0x03, 0x00, 0x00, 0xff, 0xff, 0x0f, 0x1b, 0x55, 0x1f, 0x83, 0x03, 0x00, 0x00, } func (this *IssuedCertificateSpec) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateSpec) if !ok { that2, ok := that.(IssuedCertificateSpec) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if len(this.Hosts) != len(that1.Hosts) { return false } for i := range this.Hosts { if this.Hosts[i] != that1.Hosts[i] { return false } } if this.Org != that1.Org { return false } if !this.SigningCertificateSecret.Equal(that1.SigningCertificateSecret) { return false } if !this.IssuedCertificateSecret.Equal(that1.IssuedCertificateSecret) { return false } if !this.PodBounceDirective.Equal(that1.PodBounceDirective) { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true } func (this *IssuedCertificateStatus) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateStatus) if !ok { that2, ok := that.(IssuedCertificateStatus) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if this.ObservedGeneration != that1.ObservedGeneration { return false } if this.Error != that1.Error { return false } if this.State != that1.State { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true }
random_line_split
issued_certificate.pb.go
// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto package v1alpha2 import ( bytes "bytes" fmt "fmt" math "math" _ "github.com/gogo/protobuf/gogoproto" proto "github.com/gogo/protobuf/proto" _ "github.com/gogo/protobuf/types" v1 "github.com/solo-io/skv2/pkg/api/core.skv2.solo.io/v1" ) // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal var _ = fmt.Errorf var _ = math.Inf // This is a compile-time assertion to ensure that this generated file // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package // Possible states in which an IssuedCertificate can exist. type IssuedCertificateStatus_State int32 const ( // The IssuedCertificate has yet to be picked up by the agent. IssuedCertificateStatus_PENDING IssuedCertificateStatus_State = 0 // The agent has created a local private key // and a CertificateRequest for the IssuedCertificate. // In this state, the agent is waiting for the Issuer // to issue certificates for the CertificateRequest before proceeding. IssuedCertificateStatus_REQUESTED IssuedCertificateStatus_State = 1 // The certificate has been issued. Any pods that require restarting will be restarted at this point. IssuedCertificateStatus_ISSUED IssuedCertificateStatus_State = 2 // The reply from the Issuer has been processed and // the agent has placed the final certificate secret // in the target location specified by the IssuedCertificate. IssuedCertificateStatus_FINISHED IssuedCertificateStatus_State = 3 // Processing the certificate workflow failed. IssuedCertificateStatus_FAILED IssuedCertificateStatus_State = 4 ) var IssuedCertificateStatus_State_name = map[int32]string{ 0: "PENDING", 1: "REQUESTED", 2: "ISSUED", 3: "FINISHED", 4: "FAILED", } var IssuedCertificateStatus_State_value = map[string]int32{ "PENDING": 0, "REQUESTED": 1, "ISSUED": 2, "FINISHED": 3, "FAILED": 4, } func (x IssuedCertificateStatus_State) String() string { return proto.EnumName(IssuedCertificateStatus_State_name, int32(x)) } func (IssuedCertificateStatus_State) EnumDescriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1, 0} } // //IssuedCertificates are used to issue SSL certificates //to remote Kubernetes clusters from a central (out-of-cluster) Certificate Authority. // //When an IssuedCertificate is created, a certificate is issued to a remote cluster by a central Certificate Authority via //the following workflow: //- The Certificate Issuer creates the IssuedCertificate resource on the remote cluster //- The Certificate Signature Requesting Agent installed to the remote cluster generates //a Certificate Signing Request and writes it to the status of the IssuedCertificate //- Finally, the Certificate Issuer generates signed a certificate for the CSR and writes //it back as Secret in the remote cluster. // //Shared trust can therefore be established across clusters without requiring //private keys to ever leave the node. type IssuedCertificateSpec struct { // //A list of hostnames and IPs to generate a certificate for. //This can also be set to the identity running the workload, //e.g. a Kubernetes service account. // //Generally for an Istio CA this will take the form `spiffe://cluster.local/ns/istio-system/sa/citadel`. // //"cluster.local" may be replaced by the root of trust domain for the mesh. Hosts []string `protobuf:"bytes,1,rep,name=hosts,proto3" json:"hosts,omitempty"` // The organization for this certificate. Org string `protobuf:"bytes,2,opt,name=org,proto3" json:"org,omitempty"` // The secret containing the root SSL certificate used to sign this IssuedCertificate (located in the Certificate Issuer's cluster). SigningCertificateSecret *v1.ObjectRef `protobuf:"bytes,3,opt,name=signing_certificate_secret,json=signingCertificateSecret,proto3" json:"signing_certificate_secret,omitempty"` // The secret containing the SSL certificate to be generated for this IssuedCertificate (located in the Certificate Agent's cluster). IssuedCertificateSecret *v1.ObjectRef `protobuf:"bytes,4,opt,name=issued_certificate_secret,json=issuedCertificateSecret,proto3" json:"issued_certificate_secret,omitempty"` // A ref to a PodBounceDirective specifying a list of k8s pods to bounce // (delete and cause a restart) when the certificate is issued. // This will include the control plane pods as well as any pods // which share a data plane with the target mesh. PodBounceDirective *v1.ObjectRef `protobuf:"bytes,5,opt,name=pod_bounce_directive,json=podBounceDirective,proto3" json:"pod_bounce_directive,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateSpec) Reset() { *m = IssuedCertificateSpec{} } func (m *IssuedCertificateSpec) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateSpec) ProtoMessage() {} func (*IssuedCertificateSpec) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{0} } func (m *IssuedCertificateSpec) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateSpec.Unmarshal(m, b) } func (m *IssuedCertificateSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateSpec.Marshal(b, m, deterministic) } func (m *IssuedCertificateSpec) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateSpec.Merge(m, src) } func (m *IssuedCertificateSpec) XXX_Size() int { return xxx_messageInfo_IssuedCertificateSpec.Size(m) } func (m *IssuedCertificateSpec) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateSpec.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateSpec proto.InternalMessageInfo func (m *IssuedCertificateSpec) GetHosts() []string { if m != nil { return m.Hosts } return nil } func (m *IssuedCertificateSpec) GetOrg() string { if m != nil { return m.Org } return "" } func (m *IssuedCertificateSpec) GetSigningCertificateSecret() *v1.ObjectRef { if m != nil { return m.SigningCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetIssuedCertificateSecret() *v1.ObjectRef { if m != nil { return m.IssuedCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetPodBounceDirective() *v1.ObjectRef { if m != nil { return m.PodBounceDirective } return nil } // The IssuedCertificate status is written by the CertificateRequesting agent. type IssuedCertificateStatus struct { // The most recent generation observed in the the IssuedCertificate metadata. // If the observedGeneration does not match generation, the Certificate Requesting Agent has not processed the most // recent version of this IssuedCertificate. ObservedGeneration int64 `protobuf:"varint,1,opt,name=observed_generation,json=observedGeneration,proto3" json:"observed_generation,omitempty"` // Any error observed which prevented the CertificateRequest from being processed. // If the error is empty, the request has been processed successfully. Error string `protobuf:"bytes,2,opt,name=error,proto3" json:"error,omitempty"` // The current state of the IssuedCertificate workflow, reported by the agent. State IssuedCertificateStatus_State `protobuf:"varint,3,opt,name=state,proto3,enum=certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateStatus) Reset() { *m = IssuedCertificateStatus{} } func (m *IssuedCertificateStatus) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateStatus) ProtoMessage() {} func (*IssuedCertificateStatus) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1} } func (m *IssuedCertificateStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateStatus.Unmarshal(m, b) } func (m *IssuedCertificateStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateStatus.Marshal(b, m, deterministic) } func (m *IssuedCertificateStatus) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateStatus.Merge(m, src) } func (m *IssuedCertificateStatus) XXX_Size() int { return xxx_messageInfo_IssuedCertificateStatus.Size(m) } func (m *IssuedCertificateStatus) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateStatus.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateStatus proto.InternalMessageInfo func (m *IssuedCertificateStatus) GetObservedGeneration() int64 { if m != nil { return m.ObservedGeneration } return 0 } func (m *IssuedCertificateStatus) GetError() string { if m != nil { return m.Error } return "" } func (m *IssuedCertificateStatus) GetState() IssuedCertificateStatus_State { if m != nil { return m.State } return IssuedCertificateStatus_PENDING } func init()
func init() { proto.RegisterFile("github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto", fileDescriptor_86ade12c22739639) } var fileDescriptor_86ade12c22739639 = []byte{ // 479 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x53, 0x5d, 0x6f, 0xd3, 0x30, 0x14, 0x25, 0xed, 0x3a, 0xa8, 0x07, 0x28, 0x32, 0x45, 0x0b, 0x05, 0x95, 0xaa, 0x4f, 0x7d, 0x99, 0xad, 0x95, 0x67, 0x1e, 0x18, 0x49, 0x47, 0xa4, 0xa9, 0x40, 0xc2, 0x24, 0xb4, 0x97, 0x2a, 0x1f, 0xb7, 0xae, 0x59, 0xdb, 0x1b, 0xd9, 0x4e, 0x9f, 0xf9, 0x39, 0xfc, 0x11, 0xfe, 0x08, 0xbf, 0x04, 0xc5, 0x69, 0xa1, 0x22, 0x62, 0xda, 0x53, 0xee, 0xf5, 0xbd, 0xe7, 0xe8, 0x9c, 0x13, 0x9b, 0x4c, 0x85, 0x34, 0xcb, 0x32, 0x65, 0x19, 0xae, 0xb9, 0xc6, 0x15, 0x9e, 0x49, 0xe4, 0x62, 0x85, 0x78, 0xb6, 0x06, 0xbd, 0xe4, 0x49, 0x21, 0x79, 0x06, 0xca, 0xc8, 0x85, 0xcc, 0x12, 0x03, 0x9a, 0x4b, 0xad, 0x4b, 0xc8, 0xe7, 0x07, 0x67, 0xac, 0x50, 0x68, 0x90, 0x0e, 0x0e, 0xd7, 0x58, 0x05, 0x66, 0x15, 0x0d, 0xab, 0x38, 0x99, 0xc4, 0xfe, 0x4b, 0x7d, 0xbb, 0x9d, 0xd4, 0x7c, 0xa8, 0x80, 0x6f, 0xcf, 0xed, 0xb7, 0x06, 0xf7, 0x5f, 0x0b, 0x44, 0xb1, 0x02, 0x6e, 0xbb, 0xb4, 0x5c, 0x70, 0x23, 0xd7, 0xa0, 0x4d, 0xb2, 0x2e, 0x76, 0x0b, 0x83, 0x7f, 0x17, 0xf2, 0x52, 0x25, 0x46, 0xe2, 0x66, 0x37, 0xef, 0x09, 0x14, 0x68, 0x4b, 0x5e, 0x55, 0xf5, 0xe9, 0xe8, 0x67, 0x8b, 0x3c, 0x0f, 0xad, 0xe0, 0xf7, 0x7f, 0xc5, 0xc5, 0x05, 0x64, 0xb4, 0x47, 0x3a, 0x4b, 0xd4, 0x46, 0x7b, 0xce, 0xb0, 0x3d, 0xee, 0x46, 0x75, 0x43, 0x5d, 0xd2, 0x46, 0x25, 0xbc, 0xd6, 0xd0, 0x19, 0x77, 0xa3, 0xaa, 0xa4, 0x37, 0xa4, 0xaf, 0xa5, 0xd8, 0xc8, 0x8d, 0x38, 0xb4, 0x3c, 0xd7, 0x90, 0x29, 0x30, 0x5e, 0x7b, 0xe8, 0x8c, 0x4f, 0x26, 0xaf, 0x98, 0x75, 0x52, 0xf9, 0xdb, 0xbb, 0x65, 0x1f, 0xd3, 0x6f, 0x90, 0x99, 0x08, 0x16, 0x91, 0xb7, 0xc3, 0x1f, 0x2a, 0xb0, 0x68, 0xfa, 0x95, 0xbc, 0x68, 0xa6, 0xb9, 0xa7, 0x3e, 0xba, 0x07, 0xf5, 0xa9, 0x6c, 0x78, 0xab, 0x99, 0x67, 0xa4, 0x57, 0x60, 0x3e, 0x4f, 0xb1, 0xdc, 0x64, 0x30, 0xcf, 0xa5, 0x82, 0xcc, 0xc8, 0x2d, 0x78, 0x9d, 0x7b, 0x90, 0xd2, 0x02, 0xf3, 0x0b, 0x0b, 0xf4, 0xf7, 0xb8, 0xd1, 0xf7, 0x16, 0x39, 0x6d, 0xe6, 0x68, 0x12, 0x53, 0x6a, 0xca, 0xc9, 0x33, 0x4c, 0x35, 0xa8, 0x2d, 0xe4, 0x73, 0x01, 0x1b, 0xa8, 0x7f, 0x8b, 0xe7, 0x0c, 0x9d, 0x71, 0x3b, 0xa2, 0xfb, 0xd1, 0xe5, 0x9f, 0x49, 0x15, 0x3d, 0x28, 0x85, 0x6a, 0x17, 0x73, 0xdd, 0xd0, 0x98, 0x74, 0xb4, 0x49, 0x0c, 0xd8, 0x4c, 0x9f, 0x4e, 0xde, 0xb2, 0xbb, 0xaf, 0x13, 0xfb, 0x8f, 0x1c, 0x56, 0x7d, 0x20, 0xaa, 0xb9, 0x46, 0x21, 0xe9, 0xd8, 0x9e, 0x9e, 0x90, 0x87, 0x9f, 0x82, 0x99, 0x1f, 0xce, 0x2e, 0xdd, 0x07, 0xf4, 0x09, 0xe9, 0x46, 0xc1, 0xe7, 0xeb, 0x20, 0xfe, 0x12, 0xf8, 0xae, 0x43, 0x09, 0x39, 0x0e, 0xe3, 0xf8, 0x3a, 0xf0, 0xdd, 0x16, 0x7d, 0x4c, 0x1e, 0x4d, 0xc3, 0x59, 0x18, 0x7f, 0x08, 0x7c, 0xb7, 0x5d, 0x4d, 0xa6, 0xef, 0xc2, 0xab, 0xc0, 0x77, 0x8f, 0x2e, 0xa2, 0x1f, 0xbf, 0x06, 0xce, 0xcd, 0xd5, 0x9d, 0x8f, 0xa5, 0xb8, 0x15, 0x8d, 0x07, 0xd3, 0x94, 0xce, 0xb7, 0xe7, 0xc9, 0xaa, 0x58, 0x26, 0x93, 0xf4, 0xd8, 0xde, 0xd2, 0x37, 0xbf, 0x03, 0x00, 0x00, 0xff, 0xff, 0x0f, 0x1b, 0x55, 0x1f, 0x83, 0x03, 0x00, 0x00, } func (this *IssuedCertificateSpec) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateSpec) if !ok { that2, ok := that.(IssuedCertificateSpec) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if len(this.Hosts) != len(that1.Hosts) { return false } for i := range this.Hosts { if this.Hosts[i] != that1.Hosts[i] { return false } } if this.Org != that1.Org { return false } if !this.SigningCertificateSecret.Equal(that1.SigningCertificateSecret) { return false } if !this.IssuedCertificateSecret.Equal(that1.IssuedCertificateSecret) { return false } if !this.PodBounceDirective.Equal(that1.PodBounceDirective) { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true } func (this *IssuedCertificateStatus) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateStatus) if !ok { that2, ok := that.(IssuedCertificateStatus) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if this.ObservedGeneration != that1.ObservedGeneration { return false } if this.Error != that1.Error { return false } if this.State != that1.State { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true }
{ proto.RegisterEnum("certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State", IssuedCertificateStatus_State_name, IssuedCertificateStatus_State_value) proto.RegisterType((*IssuedCertificateSpec)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateSpec") proto.RegisterType((*IssuedCertificateStatus)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateStatus") }
identifier_body
issued_certificate.pb.go
// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto package v1alpha2 import ( bytes "bytes" fmt "fmt" math "math" _ "github.com/gogo/protobuf/gogoproto" proto "github.com/gogo/protobuf/proto" _ "github.com/gogo/protobuf/types" v1 "github.com/solo-io/skv2/pkg/api/core.skv2.solo.io/v1" ) // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal var _ = fmt.Errorf var _ = math.Inf // This is a compile-time assertion to ensure that this generated file // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package // Possible states in which an IssuedCertificate can exist. type IssuedCertificateStatus_State int32 const ( // The IssuedCertificate has yet to be picked up by the agent. IssuedCertificateStatus_PENDING IssuedCertificateStatus_State = 0 // The agent has created a local private key // and a CertificateRequest for the IssuedCertificate. // In this state, the agent is waiting for the Issuer // to issue certificates for the CertificateRequest before proceeding. IssuedCertificateStatus_REQUESTED IssuedCertificateStatus_State = 1 // The certificate has been issued. Any pods that require restarting will be restarted at this point. IssuedCertificateStatus_ISSUED IssuedCertificateStatus_State = 2 // The reply from the Issuer has been processed and // the agent has placed the final certificate secret // in the target location specified by the IssuedCertificate. IssuedCertificateStatus_FINISHED IssuedCertificateStatus_State = 3 // Processing the certificate workflow failed. IssuedCertificateStatus_FAILED IssuedCertificateStatus_State = 4 ) var IssuedCertificateStatus_State_name = map[int32]string{ 0: "PENDING", 1: "REQUESTED", 2: "ISSUED", 3: "FINISHED", 4: "FAILED", } var IssuedCertificateStatus_State_value = map[string]int32{ "PENDING": 0, "REQUESTED": 1, "ISSUED": 2, "FINISHED": 3, "FAILED": 4, } func (x IssuedCertificateStatus_State) String() string { return proto.EnumName(IssuedCertificateStatus_State_name, int32(x)) } func (IssuedCertificateStatus_State) EnumDescriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1, 0} } // //IssuedCertificates are used to issue SSL certificates //to remote Kubernetes clusters from a central (out-of-cluster) Certificate Authority. // //When an IssuedCertificate is created, a certificate is issued to a remote cluster by a central Certificate Authority via //the following workflow: //- The Certificate Issuer creates the IssuedCertificate resource on the remote cluster //- The Certificate Signature Requesting Agent installed to the remote cluster generates //a Certificate Signing Request and writes it to the status of the IssuedCertificate //- Finally, the Certificate Issuer generates signed a certificate for the CSR and writes //it back as Secret in the remote cluster. // //Shared trust can therefore be established across clusters without requiring //private keys to ever leave the node. type IssuedCertificateSpec struct { // //A list of hostnames and IPs to generate a certificate for. //This can also be set to the identity running the workload, //e.g. a Kubernetes service account. // //Generally for an Istio CA this will take the form `spiffe://cluster.local/ns/istio-system/sa/citadel`. // //"cluster.local" may be replaced by the root of trust domain for the mesh. Hosts []string `protobuf:"bytes,1,rep,name=hosts,proto3" json:"hosts,omitempty"` // The organization for this certificate. Org string `protobuf:"bytes,2,opt,name=org,proto3" json:"org,omitempty"` // The secret containing the root SSL certificate used to sign this IssuedCertificate (located in the Certificate Issuer's cluster). SigningCertificateSecret *v1.ObjectRef `protobuf:"bytes,3,opt,name=signing_certificate_secret,json=signingCertificateSecret,proto3" json:"signing_certificate_secret,omitempty"` // The secret containing the SSL certificate to be generated for this IssuedCertificate (located in the Certificate Agent's cluster). IssuedCertificateSecret *v1.ObjectRef `protobuf:"bytes,4,opt,name=issued_certificate_secret,json=issuedCertificateSecret,proto3" json:"issued_certificate_secret,omitempty"` // A ref to a PodBounceDirective specifying a list of k8s pods to bounce // (delete and cause a restart) when the certificate is issued. // This will include the control plane pods as well as any pods // which share a data plane with the target mesh. PodBounceDirective *v1.ObjectRef `protobuf:"bytes,5,opt,name=pod_bounce_directive,json=podBounceDirective,proto3" json:"pod_bounce_directive,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateSpec) Reset() { *m = IssuedCertificateSpec{} } func (m *IssuedCertificateSpec) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateSpec) ProtoMessage() {} func (*IssuedCertificateSpec) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{0} } func (m *IssuedCertificateSpec) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateSpec.Unmarshal(m, b) } func (m *IssuedCertificateSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateSpec.Marshal(b, m, deterministic) } func (m *IssuedCertificateSpec) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateSpec.Merge(m, src) } func (m *IssuedCertificateSpec) XXX_Size() int { return xxx_messageInfo_IssuedCertificateSpec.Size(m) } func (m *IssuedCertificateSpec) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateSpec.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateSpec proto.InternalMessageInfo func (m *IssuedCertificateSpec) GetHosts() []string { if m != nil { return m.Hosts } return nil } func (m *IssuedCertificateSpec)
() string { if m != nil { return m.Org } return "" } func (m *IssuedCertificateSpec) GetSigningCertificateSecret() *v1.ObjectRef { if m != nil { return m.SigningCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetIssuedCertificateSecret() *v1.ObjectRef { if m != nil { return m.IssuedCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetPodBounceDirective() *v1.ObjectRef { if m != nil { return m.PodBounceDirective } return nil } // The IssuedCertificate status is written by the CertificateRequesting agent. type IssuedCertificateStatus struct { // The most recent generation observed in the the IssuedCertificate metadata. // If the observedGeneration does not match generation, the Certificate Requesting Agent has not processed the most // recent version of this IssuedCertificate. ObservedGeneration int64 `protobuf:"varint,1,opt,name=observed_generation,json=observedGeneration,proto3" json:"observed_generation,omitempty"` // Any error observed which prevented the CertificateRequest from being processed. // If the error is empty, the request has been processed successfully. Error string `protobuf:"bytes,2,opt,name=error,proto3" json:"error,omitempty"` // The current state of the IssuedCertificate workflow, reported by the agent. State IssuedCertificateStatus_State `protobuf:"varint,3,opt,name=state,proto3,enum=certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateStatus) Reset() { *m = IssuedCertificateStatus{} } func (m *IssuedCertificateStatus) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateStatus) ProtoMessage() {} func (*IssuedCertificateStatus) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1} } func (m *IssuedCertificateStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateStatus.Unmarshal(m, b) } func (m *IssuedCertificateStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateStatus.Marshal(b, m, deterministic) } func (m *IssuedCertificateStatus) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateStatus.Merge(m, src) } func (m *IssuedCertificateStatus) XXX_Size() int { return xxx_messageInfo_IssuedCertificateStatus.Size(m) } func (m *IssuedCertificateStatus) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateStatus.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateStatus proto.InternalMessageInfo func (m *IssuedCertificateStatus) GetObservedGeneration() int64 { if m != nil { return m.ObservedGeneration } return 0 } func (m *IssuedCertificateStatus) GetError() string { if m != nil { return m.Error } return "" } func (m *IssuedCertificateStatus) GetState() IssuedCertificateStatus_State { if m != nil { return m.State } return IssuedCertificateStatus_PENDING } func init() { proto.RegisterEnum("certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State", IssuedCertificateStatus_State_name, IssuedCertificateStatus_State_value) proto.RegisterType((*IssuedCertificateSpec)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateSpec") proto.RegisterType((*IssuedCertificateStatus)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateStatus") } func init() { proto.RegisterFile("github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto", fileDescriptor_86ade12c22739639) } var fileDescriptor_86ade12c22739639 = []byte{ // 479 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x53, 0x5d, 0x6f, 0xd3, 0x30, 0x14, 0x25, 0xed, 0x3a, 0xa8, 0x07, 0x28, 0x32, 0x45, 0x0b, 0x05, 0x95, 0xaa, 0x4f, 0x7d, 0x99, 0xad, 0x95, 0x67, 0x1e, 0x18, 0x49, 0x47, 0xa4, 0xa9, 0x40, 0xc2, 0x24, 0xb4, 0x97, 0x2a, 0x1f, 0xb7, 0xae, 0x59, 0xdb, 0x1b, 0xd9, 0x4e, 0x9f, 0xf9, 0x39, 0xfc, 0x11, 0xfe, 0x08, 0xbf, 0x04, 0xc5, 0x69, 0xa1, 0x22, 0x62, 0xda, 0x53, 0xee, 0xf5, 0xbd, 0xe7, 0xe8, 0x9c, 0x13, 0x9b, 0x4c, 0x85, 0x34, 0xcb, 0x32, 0x65, 0x19, 0xae, 0xb9, 0xc6, 0x15, 0x9e, 0x49, 0xe4, 0x62, 0x85, 0x78, 0xb6, 0x06, 0xbd, 0xe4, 0x49, 0x21, 0x79, 0x06, 0xca, 0xc8, 0x85, 0xcc, 0x12, 0x03, 0x9a, 0x4b, 0xad, 0x4b, 0xc8, 0xe7, 0x07, 0x67, 0xac, 0x50, 0x68, 0x90, 0x0e, 0x0e, 0xd7, 0x58, 0x05, 0x66, 0x15, 0x0d, 0xab, 0x38, 0x99, 0xc4, 0xfe, 0x4b, 0x7d, 0xbb, 0x9d, 0xd4, 0x7c, 0xa8, 0x80, 0x6f, 0xcf, 0xed, 0xb7, 0x06, 0xf7, 0x5f, 0x0b, 0x44, 0xb1, 0x02, 0x6e, 0xbb, 0xb4, 0x5c, 0x70, 0x23, 0xd7, 0xa0, 0x4d, 0xb2, 0x2e, 0x76, 0x0b, 0x83, 0x7f, 0x17, 0xf2, 0x52, 0x25, 0x46, 0xe2, 0x66, 0x37, 0xef, 0x09, 0x14, 0x68, 0x4b, 0x5e, 0x55, 0xf5, 0xe9, 0xe8, 0x67, 0x8b, 0x3c, 0x0f, 0xad, 0xe0, 0xf7, 0x7f, 0xc5, 0xc5, 0x05, 0x64, 0xb4, 0x47, 0x3a, 0x4b, 0xd4, 0x46, 0x7b, 0xce, 0xb0, 0x3d, 0xee, 0x46, 0x75, 0x43, 0x5d, 0xd2, 0x46, 0x25, 0xbc, 0xd6, 0xd0, 0x19, 0x77, 0xa3, 0xaa, 0xa4, 0x37, 0xa4, 0xaf, 0xa5, 0xd8, 0xc8, 0x8d, 0x38, 0xb4, 0x3c, 0xd7, 0x90, 0x29, 0x30, 0x5e, 0x7b, 0xe8, 0x8c, 0x4f, 0x26, 0xaf, 0x98, 0x75, 0x52, 0xf9, 0xdb, 0xbb, 0x65, 0x1f, 0xd3, 0x6f, 0x90, 0x99, 0x08, 0x16, 0x91, 0xb7, 0xc3, 0x1f, 0x2a, 0xb0, 0x68, 0xfa, 0x95, 0xbc, 0x68, 0xa6, 0xb9, 0xa7, 0x3e, 0xba, 0x07, 0xf5, 0xa9, 0x6c, 0x78, 0xab, 0x99, 0x67, 0xa4, 0x57, 0x60, 0x3e, 0x4f, 0xb1, 0xdc, 0x64, 0x30, 0xcf, 0xa5, 0x82, 0xcc, 0xc8, 0x2d, 0x78, 0x9d, 0x7b, 0x90, 0xd2, 0x02, 0xf3, 0x0b, 0x0b, 0xf4, 0xf7, 0xb8, 0xd1, 0xf7, 0x16, 0x39, 0x6d, 0xe6, 0x68, 0x12, 0x53, 0x6a, 0xca, 0xc9, 0x33, 0x4c, 0x35, 0xa8, 0x2d, 0xe4, 0x73, 0x01, 0x1b, 0xa8, 0x7f, 0x8b, 0xe7, 0x0c, 0x9d, 0x71, 0x3b, 0xa2, 0xfb, 0xd1, 0xe5, 0x9f, 0x49, 0x15, 0x3d, 0x28, 0x85, 0x6a, 0x17, 0x73, 0xdd, 0xd0, 0x98, 0x74, 0xb4, 0x49, 0x0c, 0xd8, 0x4c, 0x9f, 0x4e, 0xde, 0xb2, 0xbb, 0xaf, 0x13, 0xfb, 0x8f, 0x1c, 0x56, 0x7d, 0x20, 0xaa, 0xb9, 0x46, 0x21, 0xe9, 0xd8, 0x9e, 0x9e, 0x90, 0x87, 0x9f, 0x82, 0x99, 0x1f, 0xce, 0x2e, 0xdd, 0x07, 0xf4, 0x09, 0xe9, 0x46, 0xc1, 0xe7, 0xeb, 0x20, 0xfe, 0x12, 0xf8, 0xae, 0x43, 0x09, 0x39, 0x0e, 0xe3, 0xf8, 0x3a, 0xf0, 0xdd, 0x16, 0x7d, 0x4c, 0x1e, 0x4d, 0xc3, 0x59, 0x18, 0x7f, 0x08, 0x7c, 0xb7, 0x5d, 0x4d, 0xa6, 0xef, 0xc2, 0xab, 0xc0, 0x77, 0x8f, 0x2e, 0xa2, 0x1f, 0xbf, 0x06, 0xce, 0xcd, 0xd5, 0x9d, 0x8f, 0xa5, 0xb8, 0x15, 0x8d, 0x07, 0xd3, 0x94, 0xce, 0xb7, 0xe7, 0xc9, 0xaa, 0x58, 0x26, 0x93, 0xf4, 0xd8, 0xde, 0xd2, 0x37, 0xbf, 0x03, 0x00, 0x00, 0xff, 0xff, 0x0f, 0x1b, 0x55, 0x1f, 0x83, 0x03, 0x00, 0x00, } func (this *IssuedCertificateSpec) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateSpec) if !ok { that2, ok := that.(IssuedCertificateSpec) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if len(this.Hosts) != len(that1.Hosts) { return false } for i := range this.Hosts { if this.Hosts[i] != that1.Hosts[i] { return false } } if this.Org != that1.Org { return false } if !this.SigningCertificateSecret.Equal(that1.SigningCertificateSecret) { return false } if !this.IssuedCertificateSecret.Equal(that1.IssuedCertificateSecret) { return false } if !this.PodBounceDirective.Equal(that1.PodBounceDirective) { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true } func (this *IssuedCertificateStatus) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateStatus) if !ok { that2, ok := that.(IssuedCertificateStatus) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if this.ObservedGeneration != that1.ObservedGeneration { return false } if this.Error != that1.Error { return false } if this.State != that1.State { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true }
GetOrg
identifier_name
issued_certificate.pb.go
// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto package v1alpha2 import ( bytes "bytes" fmt "fmt" math "math" _ "github.com/gogo/protobuf/gogoproto" proto "github.com/gogo/protobuf/proto" _ "github.com/gogo/protobuf/types" v1 "github.com/solo-io/skv2/pkg/api/core.skv2.solo.io/v1" ) // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal var _ = fmt.Errorf var _ = math.Inf // This is a compile-time assertion to ensure that this generated file // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package // Possible states in which an IssuedCertificate can exist. type IssuedCertificateStatus_State int32 const ( // The IssuedCertificate has yet to be picked up by the agent. IssuedCertificateStatus_PENDING IssuedCertificateStatus_State = 0 // The agent has created a local private key // and a CertificateRequest for the IssuedCertificate. // In this state, the agent is waiting for the Issuer // to issue certificates for the CertificateRequest before proceeding. IssuedCertificateStatus_REQUESTED IssuedCertificateStatus_State = 1 // The certificate has been issued. Any pods that require restarting will be restarted at this point. IssuedCertificateStatus_ISSUED IssuedCertificateStatus_State = 2 // The reply from the Issuer has been processed and // the agent has placed the final certificate secret // in the target location specified by the IssuedCertificate. IssuedCertificateStatus_FINISHED IssuedCertificateStatus_State = 3 // Processing the certificate workflow failed. IssuedCertificateStatus_FAILED IssuedCertificateStatus_State = 4 ) var IssuedCertificateStatus_State_name = map[int32]string{ 0: "PENDING", 1: "REQUESTED", 2: "ISSUED", 3: "FINISHED", 4: "FAILED", } var IssuedCertificateStatus_State_value = map[string]int32{ "PENDING": 0, "REQUESTED": 1, "ISSUED": 2, "FINISHED": 3, "FAILED": 4, } func (x IssuedCertificateStatus_State) String() string { return proto.EnumName(IssuedCertificateStatus_State_name, int32(x)) } func (IssuedCertificateStatus_State) EnumDescriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1, 0} } // //IssuedCertificates are used to issue SSL certificates //to remote Kubernetes clusters from a central (out-of-cluster) Certificate Authority. // //When an IssuedCertificate is created, a certificate is issued to a remote cluster by a central Certificate Authority via //the following workflow: //- The Certificate Issuer creates the IssuedCertificate resource on the remote cluster //- The Certificate Signature Requesting Agent installed to the remote cluster generates //a Certificate Signing Request and writes it to the status of the IssuedCertificate //- Finally, the Certificate Issuer generates signed a certificate for the CSR and writes //it back as Secret in the remote cluster. // //Shared trust can therefore be established across clusters without requiring //private keys to ever leave the node. type IssuedCertificateSpec struct { // //A list of hostnames and IPs to generate a certificate for. //This can also be set to the identity running the workload, //e.g. a Kubernetes service account. // //Generally for an Istio CA this will take the form `spiffe://cluster.local/ns/istio-system/sa/citadel`. // //"cluster.local" may be replaced by the root of trust domain for the mesh. Hosts []string `protobuf:"bytes,1,rep,name=hosts,proto3" json:"hosts,omitempty"` // The organization for this certificate. Org string `protobuf:"bytes,2,opt,name=org,proto3" json:"org,omitempty"` // The secret containing the root SSL certificate used to sign this IssuedCertificate (located in the Certificate Issuer's cluster). SigningCertificateSecret *v1.ObjectRef `protobuf:"bytes,3,opt,name=signing_certificate_secret,json=signingCertificateSecret,proto3" json:"signing_certificate_secret,omitempty"` // The secret containing the SSL certificate to be generated for this IssuedCertificate (located in the Certificate Agent's cluster). IssuedCertificateSecret *v1.ObjectRef `protobuf:"bytes,4,opt,name=issued_certificate_secret,json=issuedCertificateSecret,proto3" json:"issued_certificate_secret,omitempty"` // A ref to a PodBounceDirective specifying a list of k8s pods to bounce // (delete and cause a restart) when the certificate is issued. // This will include the control plane pods as well as any pods // which share a data plane with the target mesh. PodBounceDirective *v1.ObjectRef `protobuf:"bytes,5,opt,name=pod_bounce_directive,json=podBounceDirective,proto3" json:"pod_bounce_directive,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateSpec) Reset() { *m = IssuedCertificateSpec{} } func (m *IssuedCertificateSpec) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateSpec) ProtoMessage() {} func (*IssuedCertificateSpec) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{0} } func (m *IssuedCertificateSpec) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateSpec.Unmarshal(m, b) } func (m *IssuedCertificateSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateSpec.Marshal(b, m, deterministic) } func (m *IssuedCertificateSpec) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateSpec.Merge(m, src) } func (m *IssuedCertificateSpec) XXX_Size() int { return xxx_messageInfo_IssuedCertificateSpec.Size(m) } func (m *IssuedCertificateSpec) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateSpec.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateSpec proto.InternalMessageInfo func (m *IssuedCertificateSpec) GetHosts() []string { if m != nil { return m.Hosts } return nil } func (m *IssuedCertificateSpec) GetOrg() string { if m != nil { return m.Org } return "" } func (m *IssuedCertificateSpec) GetSigningCertificateSecret() *v1.ObjectRef { if m != nil { return m.SigningCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetIssuedCertificateSecret() *v1.ObjectRef { if m != nil { return m.IssuedCertificateSecret } return nil } func (m *IssuedCertificateSpec) GetPodBounceDirective() *v1.ObjectRef { if m != nil { return m.PodBounceDirective } return nil } // The IssuedCertificate status is written by the CertificateRequesting agent. type IssuedCertificateStatus struct { // The most recent generation observed in the the IssuedCertificate metadata. // If the observedGeneration does not match generation, the Certificate Requesting Agent has not processed the most // recent version of this IssuedCertificate. ObservedGeneration int64 `protobuf:"varint,1,opt,name=observed_generation,json=observedGeneration,proto3" json:"observed_generation,omitempty"` // Any error observed which prevented the CertificateRequest from being processed. // If the error is empty, the request has been processed successfully. Error string `protobuf:"bytes,2,opt,name=error,proto3" json:"error,omitempty"` // The current state of the IssuedCertificate workflow, reported by the agent. State IssuedCertificateStatus_State `protobuf:"varint,3,opt,name=state,proto3,enum=certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *IssuedCertificateStatus) Reset() { *m = IssuedCertificateStatus{} } func (m *IssuedCertificateStatus) String() string { return proto.CompactTextString(m) } func (*IssuedCertificateStatus) ProtoMessage() {} func (*IssuedCertificateStatus) Descriptor() ([]byte, []int) { return fileDescriptor_86ade12c22739639, []int{1} } func (m *IssuedCertificateStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_IssuedCertificateStatus.Unmarshal(m, b) } func (m *IssuedCertificateStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_IssuedCertificateStatus.Marshal(b, m, deterministic) } func (m *IssuedCertificateStatus) XXX_Merge(src proto.Message) { xxx_messageInfo_IssuedCertificateStatus.Merge(m, src) } func (m *IssuedCertificateStatus) XXX_Size() int { return xxx_messageInfo_IssuedCertificateStatus.Size(m) } func (m *IssuedCertificateStatus) XXX_DiscardUnknown() { xxx_messageInfo_IssuedCertificateStatus.DiscardUnknown(m) } var xxx_messageInfo_IssuedCertificateStatus proto.InternalMessageInfo func (m *IssuedCertificateStatus) GetObservedGeneration() int64 { if m != nil { return m.ObservedGeneration } return 0 } func (m *IssuedCertificateStatus) GetError() string { if m != nil { return m.Error } return "" } func (m *IssuedCertificateStatus) GetState() IssuedCertificateStatus_State { if m != nil { return m.State } return IssuedCertificateStatus_PENDING } func init() { proto.RegisterEnum("certificates.mesh.gloo.solo.io.IssuedCertificateStatus_State", IssuedCertificateStatus_State_name, IssuedCertificateStatus_State_value) proto.RegisterType((*IssuedCertificateSpec)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateSpec") proto.RegisterType((*IssuedCertificateStatus)(nil), "certificates.mesh.gloo.solo.io.IssuedCertificateStatus") } func init() { proto.RegisterFile("github.com/solo-io/gloo-mesh/api/certificates/issued_certificate.proto", fileDescriptor_86ade12c22739639) } var fileDescriptor_86ade12c22739639 = []byte{ // 479 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x53, 0x5d, 0x6f, 0xd3, 0x30, 0x14, 0x25, 0xed, 0x3a, 0xa8, 0x07, 0x28, 0x32, 0x45, 0x0b, 0x05, 0x95, 0xaa, 0x4f, 0x7d, 0x99, 0xad, 0x95, 0x67, 0x1e, 0x18, 0x49, 0x47, 0xa4, 0xa9, 0x40, 0xc2, 0x24, 0xb4, 0x97, 0x2a, 0x1f, 0xb7, 0xae, 0x59, 0xdb, 0x1b, 0xd9, 0x4e, 0x9f, 0xf9, 0x39, 0xfc, 0x11, 0xfe, 0x08, 0xbf, 0x04, 0xc5, 0x69, 0xa1, 0x22, 0x62, 0xda, 0x53, 0xee, 0xf5, 0xbd, 0xe7, 0xe8, 0x9c, 0x13, 0x9b, 0x4c, 0x85, 0x34, 0xcb, 0x32, 0x65, 0x19, 0xae, 0xb9, 0xc6, 0x15, 0x9e, 0x49, 0xe4, 0x62, 0x85, 0x78, 0xb6, 0x06, 0xbd, 0xe4, 0x49, 0x21, 0x79, 0x06, 0xca, 0xc8, 0x85, 0xcc, 0x12, 0x03, 0x9a, 0x4b, 0xad, 0x4b, 0xc8, 0xe7, 0x07, 0x67, 0xac, 0x50, 0x68, 0x90, 0x0e, 0x0e, 0xd7, 0x58, 0x05, 0x66, 0x15, 0x0d, 0xab, 0x38, 0x99, 0xc4, 0xfe, 0x4b, 0x7d, 0xbb, 0x9d, 0xd4, 0x7c, 0xa8, 0x80, 0x6f, 0xcf, 0xed, 0xb7, 0x06, 0xf7, 0x5f, 0x0b, 0x44, 0xb1, 0x02, 0x6e, 0xbb, 0xb4, 0x5c, 0x70, 0x23, 0xd7, 0xa0, 0x4d, 0xb2, 0x2e, 0x76, 0x0b, 0x83, 0x7f, 0x17, 0xf2, 0x52, 0x25, 0x46, 0xe2, 0x66, 0x37, 0xef, 0x09, 0x14, 0x68, 0x4b, 0x5e, 0x55, 0xf5, 0xe9, 0xe8, 0x67, 0x8b, 0x3c, 0x0f, 0xad, 0xe0, 0xf7, 0x7f, 0xc5, 0xc5, 0x05, 0x64, 0xb4, 0x47, 0x3a, 0x4b, 0xd4, 0x46, 0x7b, 0xce, 0xb0, 0x3d, 0xee, 0x46, 0x75, 0x43, 0x5d, 0xd2, 0x46, 0x25, 0xbc, 0xd6, 0xd0, 0x19, 0x77, 0xa3, 0xaa, 0xa4, 0x37, 0xa4, 0xaf, 0xa5, 0xd8, 0xc8, 0x8d, 0x38, 0xb4, 0x3c, 0xd7, 0x90, 0x29, 0x30, 0x5e, 0x7b, 0xe8, 0x8c, 0x4f, 0x26, 0xaf, 0x98, 0x75, 0x52, 0xf9, 0xdb, 0xbb, 0x65, 0x1f, 0xd3, 0x6f, 0x90, 0x99, 0x08, 0x16, 0x91, 0xb7, 0xc3, 0x1f, 0x2a, 0xb0, 0x68, 0xfa, 0x95, 0xbc, 0x68, 0xa6, 0xb9, 0xa7, 0x3e, 0xba, 0x07, 0xf5, 0xa9, 0x6c, 0x78, 0xab, 0x99, 0x67, 0xa4, 0x57, 0x60, 0x3e, 0x4f, 0xb1, 0xdc, 0x64, 0x30, 0xcf, 0xa5, 0x82, 0xcc, 0xc8, 0x2d, 0x78, 0x9d, 0x7b, 0x90, 0xd2, 0x02, 0xf3, 0x0b, 0x0b, 0xf4, 0xf7, 0xb8, 0xd1, 0xf7, 0x16, 0x39, 0x6d, 0xe6, 0x68, 0x12, 0x53, 0x6a, 0xca, 0xc9, 0x33, 0x4c, 0x35, 0xa8, 0x2d, 0xe4, 0x73, 0x01, 0x1b, 0xa8, 0x7f, 0x8b, 0xe7, 0x0c, 0x9d, 0x71, 0x3b, 0xa2, 0xfb, 0xd1, 0xe5, 0x9f, 0x49, 0x15, 0x3d, 0x28, 0x85, 0x6a, 0x17, 0x73, 0xdd, 0xd0, 0x98, 0x74, 0xb4, 0x49, 0x0c, 0xd8, 0x4c, 0x9f, 0x4e, 0xde, 0xb2, 0xbb, 0xaf, 0x13, 0xfb, 0x8f, 0x1c, 0x56, 0x7d, 0x20, 0xaa, 0xb9, 0x46, 0x21, 0xe9, 0xd8, 0x9e, 0x9e, 0x90, 0x87, 0x9f, 0x82, 0x99, 0x1f, 0xce, 0x2e, 0xdd, 0x07, 0xf4, 0x09, 0xe9, 0x46, 0xc1, 0xe7, 0xeb, 0x20, 0xfe, 0x12, 0xf8, 0xae, 0x43, 0x09, 0x39, 0x0e, 0xe3, 0xf8, 0x3a, 0xf0, 0xdd, 0x16, 0x7d, 0x4c, 0x1e, 0x4d, 0xc3, 0x59, 0x18, 0x7f, 0x08, 0x7c, 0xb7, 0x5d, 0x4d, 0xa6, 0xef, 0xc2, 0xab, 0xc0, 0x77, 0x8f, 0x2e, 0xa2, 0x1f, 0xbf, 0x06, 0xce, 0xcd, 0xd5, 0x9d, 0x8f, 0xa5, 0xb8, 0x15, 0x8d, 0x07, 0xd3, 0x94, 0xce, 0xb7, 0xe7, 0xc9, 0xaa, 0x58, 0x26, 0x93, 0xf4, 0xd8, 0xde, 0xd2, 0x37, 0xbf, 0x03, 0x00, 0x00, 0xff, 0xff, 0x0f, 0x1b, 0x55, 0x1f, 0x83, 0x03, 0x00, 0x00, } func (this *IssuedCertificateSpec) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateSpec) if !ok { that2, ok := that.(IssuedCertificateSpec) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if len(this.Hosts) != len(that1.Hosts) { return false } for i := range this.Hosts { if this.Hosts[i] != that1.Hosts[i] { return false } } if this.Org != that1.Org { return false } if !this.SigningCertificateSecret.Equal(that1.SigningCertificateSecret) { return false } if !this.IssuedCertificateSecret.Equal(that1.IssuedCertificateSecret) { return false } if !this.PodBounceDirective.Equal(that1.PodBounceDirective) { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true } func (this *IssuedCertificateStatus) Equal(that interface{}) bool { if that == nil { return this == nil } that1, ok := that.(*IssuedCertificateStatus) if !ok { that2, ok := that.(IssuedCertificateStatus) if ok { that1 = &that2 } else { return false } } if that1 == nil { return this == nil } else if this == nil { return false } if this.ObservedGeneration != that1.ObservedGeneration
if this.Error != that1.Error { return false } if this.State != that1.State { return false } if !bytes.Equal(this.XXX_unrecognized, that1.XXX_unrecognized) { return false } return true }
{ return false }
conditional_block
mprocessing.py
""" by oPromessa, 2018 Published on https://github.com/oPromessa/flickr-uploader/ mprocessing = Helper function to run function in multiprocessing mode """ # ----------------------------------------------------------------------------- # Import section for Python 2 and 3 compatible code # from __future__ import absolute_import, division, print_function, # unicode_literals from __future__ import division # This way: 3 / 2 == 1.5; 3 // 2 == 1 # ----------------------------------------------------------------------------- # Import section # import logging import multiprocessing from itertools import islice import lib.NicePrint as NicePrint # ========================================================================= # Functions aliases # # NPR.NicePrint = from NicePrint module # ------------------------------------------------------------------------- NPR = NicePrint.NicePrint() # ------------------------------------------------------------------------- # use_lock # # Control use of DB lock. acquire/release # def use_lock(adb_lock, operation, nprocs=0): """ use_lock adb_lock = lock to be used operation = True => Lock = False => Release nprocs = >0 when in multiprocessing mode >>> alock = multiprocessing.Lock() >>> use_lock(alock, True, 2) True >>> use_lock(alock, False, 2) True """ use_dblock_return = False logging.debug('Entering use_lock with operation:[%s].', operation) if adb_lock is None: logging.debug('use_lock: adb_lock is [None].') return use_dblock_return logging.debug('use_lock: adb_lock.semlock:[%s].', adb_lock._semlock) if operation is None: return use_dblock_return if (nprocs is not None) and (nprocs) and (nprocs > 0): if operation: # Control for when running multiprocessing set locking
else: # Control for when running multiprocessing release locking logging.debug('===Multiprocessing=== <--[ ].lock.release') try: adb_lock.release() use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='003', caughtmsg='Caught an exception lock.release', useniceprint=True, exceptsysinfo=True) # Raise aborts execution raise logging.info('===Multiprocessing=== <--[v].lock.release') logging.info('Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) else: use_dblock_return = True logging.warning('(No multiprocessing. Nothing to do) ' 'Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) return use_dblock_return # ----------------------------------------------------------------------------- # mprocessing # def mprocessing(nprocs, lockdb, running, mutex, itemslist, a_fn, cur): """ mprocessing Function nprocs = Number of processes to launch (int) lockdb = lock for access to Database (lock obj to be created) running = Value to count processed items (count obj to be created) mutex = mutex for access to value running (obj to be created) itemslist = list of items to be processed a_fn = a function which is the target of the multiprocessing a_fn must cater the following arguments lockdb running mutex splititemslist = partial splitted list count_total = len(itemslist) cur cur = cursor variable for DB access """ # proc_pool = Local variable proc_pool for Pool of processes # log_level = log_level # count_total = Total counter of items to distribute/play/indicate progress # len(itemslist) log_level = logging.getLogger().getEffectiveLevel() logging.info('===mprocessing [%s] target_fn():[%s] nprocs:[%s]', __name__, a_fn.__name__, nprocs) # if log_level <= logging.WARNING: # if args is not None: # for i, arg in enumerate(args): # logging.info('===mprocessing f():[%s] arg[%s]={%s}', # a_fn.__name__, i, arg) # if __name__ == '__main__': logging.debug('===Multiprocessing=== Setting up logger!') # CODING No need for such low level debugging to stderr # multiprocessing.log_to_stderr() logger = multiprocessing.get_logger() logger.setLevel(log_level) logging.debug('===Multiprocessing=== Logging defined!') # --------------------------------------------------------- # chunk # # Divides an iterable in slices/chunks of size size # def chunk(iter_list, size): """ Divides an iterable in slices/chunks of size size >>> for a in chunk([ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10], 3): ... len(a) 3 3 3 1 """ iter_list = iter(iter_list) # lambda: creates a returning expression function # which returns slices # iter, with the second argument () stops creating # iterators when it reaches the end return iter(lambda: tuple(islice(iter_list, size)), ()) proc_pool = [] lockdb = multiprocessing.Lock() running = multiprocessing.Value('i', 0) mutex = multiprocessing.Lock() count_total = len(itemslist) size = (len(itemslist) // int(nprocs)) \ if ((len(itemslist) // int(nprocs)) > 0) \ else 1 logging.debug('len(itemslist):[%s] int(nprocs):[%s] size per process:[%s]', len(itemslist), int(nprocs), size) # Split itemslist in chunks to distribute accross Processes for splititemslist in chunk(itemslist, size): logging.warning('===Actual/Planned Chunk size: [%s]/[%s]', len(splititemslist), size) logging.debug('===type(splititemslist)=[%s]', type(splititemslist)) logging.debug('===Job/Task Process: Creating...') proc_task = multiprocessing.Process( target=a_fn, # argument function args=(lockdb, running, mutex, splititemslist, count_total, cur,)) proc_pool.append(proc_task) logging.debug('===Job/Task Process: Starting...') proc_task.start() NPR.niceprint('===Job/Task Process: [{!s}] Started ' 'with pid:[{!s}]' .format(proc_task.name, proc_task.pid), verbosity=3, logalso=logging.DEBUG) # Check status of jobs/tasks in the Process Pool if log_level <= logging.DEBUG: NPR.niceprint('===Checking Processes launched/status:', verbosity=3, logalso=logging.DEBUG) for j in proc_pool: NPR.niceprint('{!s}.is_alive = {!s}'.format(j.name, j.is_alive()), verbosity=3, logalso=logging.DEBUG) # Regularly print status of jobs/tasks in the Process Pool # Prints status while there are processes active # Exits when all jobs/tasks are done. while True: if not any(multiprocessing.active_children()): logging.debug('===No active children Processes.') break for prc in multiprocessing.active_children(): logging.debug('===%s.is_alive = %s', prc.name, prc.is_alive()) proc_task_active = prc NPR.niceprint('===Will wait for 60 on {!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) proc_task_active.join(timeout=60) NPR.niceprint('===Waited for 60s on ' '{!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) # Wait for join all jobs/tasks in the Process Pool # All should be done by now! for j in proc_pool: j.join() NPR.niceprint('==={!s} (is alive: {!s}).exitcode = {!s}' .format(j.name, j.is_alive(), j.exitcode), verbosity=2) logging.warning('===Multiprocessing=== pool joined! ' 'All processes finished.') # Will release (set to None) the lockdb lock control # this prevents subsequent calls to # use_lock( nuLockDB, False) # to raise exception: # ValueError('semaphore or lock released too many times') logging.info('===Multiprocessing=== pool joined! ' 'Is lockdb None? [%s]. Setting lockdb to None anyhow.', lockdb is None) lockdb = None # Show number of total files processed NPR.niceprocessedfiles(running.value, count_total, True) return True # ----------------------------------------------------------------------------- # If called directly run doctests # if __name__ == "__main__": logging.basicConfig(level=logging.WARNING, format='[%(asctime)s]:[%(processName)-11s]' + '[%(levelname)-8s]:[%(name)s] %(message)s') import doctest doctest.testmod()
logging.debug('===Multiprocessing=== -->[ ].lock.acquire') try: if adb_lock.acquire(): use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='002', caughtmsg='Caught an exception lock.acquire', useniceprint=True, exceptsysinfo=True) raise logging.info('===Multiprocessing=== --->[v].lock.acquire')
conditional_block
mprocessing.py
""" by oPromessa, 2018 Published on https://github.com/oPromessa/flickr-uploader/ mprocessing = Helper function to run function in multiprocessing mode """ # ----------------------------------------------------------------------------- # Import section for Python 2 and 3 compatible code # from __future__ import absolute_import, division, print_function, # unicode_literals from __future__ import division # This way: 3 / 2 == 1.5; 3 // 2 == 1 # ----------------------------------------------------------------------------- # Import section # import logging import multiprocessing from itertools import islice import lib.NicePrint as NicePrint # ========================================================================= # Functions aliases # # NPR.NicePrint = from NicePrint module # ------------------------------------------------------------------------- NPR = NicePrint.NicePrint() # ------------------------------------------------------------------------- # use_lock # # Control use of DB lock. acquire/release # def use_lock(adb_lock, operation, nprocs=0):
# ----------------------------------------------------------------------------- # mprocessing # def mprocessing(nprocs, lockdb, running, mutex, itemslist, a_fn, cur): """ mprocessing Function nprocs = Number of processes to launch (int) lockdb = lock for access to Database (lock obj to be created) running = Value to count processed items (count obj to be created) mutex = mutex for access to value running (obj to be created) itemslist = list of items to be processed a_fn = a function which is the target of the multiprocessing a_fn must cater the following arguments lockdb running mutex splititemslist = partial splitted list count_total = len(itemslist) cur cur = cursor variable for DB access """ # proc_pool = Local variable proc_pool for Pool of processes # log_level = log_level # count_total = Total counter of items to distribute/play/indicate progress # len(itemslist) log_level = logging.getLogger().getEffectiveLevel() logging.info('===mprocessing [%s] target_fn():[%s] nprocs:[%s]', __name__, a_fn.__name__, nprocs) # if log_level <= logging.WARNING: # if args is not None: # for i, arg in enumerate(args): # logging.info('===mprocessing f():[%s] arg[%s]={%s}', # a_fn.__name__, i, arg) # if __name__ == '__main__': logging.debug('===Multiprocessing=== Setting up logger!') # CODING No need for such low level debugging to stderr # multiprocessing.log_to_stderr() logger = multiprocessing.get_logger() logger.setLevel(log_level) logging.debug('===Multiprocessing=== Logging defined!') # --------------------------------------------------------- # chunk # # Divides an iterable in slices/chunks of size size # def chunk(iter_list, size): """ Divides an iterable in slices/chunks of size size >>> for a in chunk([ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10], 3): ... len(a) 3 3 3 1 """ iter_list = iter(iter_list) # lambda: creates a returning expression function # which returns slices # iter, with the second argument () stops creating # iterators when it reaches the end return iter(lambda: tuple(islice(iter_list, size)), ()) proc_pool = [] lockdb = multiprocessing.Lock() running = multiprocessing.Value('i', 0) mutex = multiprocessing.Lock() count_total = len(itemslist) size = (len(itemslist) // int(nprocs)) \ if ((len(itemslist) // int(nprocs)) > 0) \ else 1 logging.debug('len(itemslist):[%s] int(nprocs):[%s] size per process:[%s]', len(itemslist), int(nprocs), size) # Split itemslist in chunks to distribute accross Processes for splititemslist in chunk(itemslist, size): logging.warning('===Actual/Planned Chunk size: [%s]/[%s]', len(splititemslist), size) logging.debug('===type(splititemslist)=[%s]', type(splititemslist)) logging.debug('===Job/Task Process: Creating...') proc_task = multiprocessing.Process( target=a_fn, # argument function args=(lockdb, running, mutex, splititemslist, count_total, cur,)) proc_pool.append(proc_task) logging.debug('===Job/Task Process: Starting...') proc_task.start() NPR.niceprint('===Job/Task Process: [{!s}] Started ' 'with pid:[{!s}]' .format(proc_task.name, proc_task.pid), verbosity=3, logalso=logging.DEBUG) # Check status of jobs/tasks in the Process Pool if log_level <= logging.DEBUG: NPR.niceprint('===Checking Processes launched/status:', verbosity=3, logalso=logging.DEBUG) for j in proc_pool: NPR.niceprint('{!s}.is_alive = {!s}'.format(j.name, j.is_alive()), verbosity=3, logalso=logging.DEBUG) # Regularly print status of jobs/tasks in the Process Pool # Prints status while there are processes active # Exits when all jobs/tasks are done. while True: if not any(multiprocessing.active_children()): logging.debug('===No active children Processes.') break for prc in multiprocessing.active_children(): logging.debug('===%s.is_alive = %s', prc.name, prc.is_alive()) proc_task_active = prc NPR.niceprint('===Will wait for 60 on {!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) proc_task_active.join(timeout=60) NPR.niceprint('===Waited for 60s on ' '{!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) # Wait for join all jobs/tasks in the Process Pool # All should be done by now! for j in proc_pool: j.join() NPR.niceprint('==={!s} (is alive: {!s}).exitcode = {!s}' .format(j.name, j.is_alive(), j.exitcode), verbosity=2) logging.warning('===Multiprocessing=== pool joined! ' 'All processes finished.') # Will release (set to None) the lockdb lock control # this prevents subsequent calls to # use_lock( nuLockDB, False) # to raise exception: # ValueError('semaphore or lock released too many times') logging.info('===Multiprocessing=== pool joined! ' 'Is lockdb None? [%s]. Setting lockdb to None anyhow.', lockdb is None) lockdb = None # Show number of total files processed NPR.niceprocessedfiles(running.value, count_total, True) return True # ----------------------------------------------------------------------------- # If called directly run doctests # if __name__ == "__main__": logging.basicConfig(level=logging.WARNING, format='[%(asctime)s]:[%(processName)-11s]' + '[%(levelname)-8s]:[%(name)s] %(message)s') import doctest doctest.testmod()
""" use_lock adb_lock = lock to be used operation = True => Lock = False => Release nprocs = >0 when in multiprocessing mode >>> alock = multiprocessing.Lock() >>> use_lock(alock, True, 2) True >>> use_lock(alock, False, 2) True """ use_dblock_return = False logging.debug('Entering use_lock with operation:[%s].', operation) if adb_lock is None: logging.debug('use_lock: adb_lock is [None].') return use_dblock_return logging.debug('use_lock: adb_lock.semlock:[%s].', adb_lock._semlock) if operation is None: return use_dblock_return if (nprocs is not None) and (nprocs) and (nprocs > 0): if operation: # Control for when running multiprocessing set locking logging.debug('===Multiprocessing=== -->[ ].lock.acquire') try: if adb_lock.acquire(): use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='002', caughtmsg='Caught an exception lock.acquire', useniceprint=True, exceptsysinfo=True) raise logging.info('===Multiprocessing=== --->[v].lock.acquire') else: # Control for when running multiprocessing release locking logging.debug('===Multiprocessing=== <--[ ].lock.release') try: adb_lock.release() use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='003', caughtmsg='Caught an exception lock.release', useniceprint=True, exceptsysinfo=True) # Raise aborts execution raise logging.info('===Multiprocessing=== <--[v].lock.release') logging.info('Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) else: use_dblock_return = True logging.warning('(No multiprocessing. Nothing to do) ' 'Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) return use_dblock_return
identifier_body
mprocessing.py
""" by oPromessa, 2018 Published on https://github.com/oPromessa/flickr-uploader/ mprocessing = Helper function to run function in multiprocessing mode """ # ----------------------------------------------------------------------------- # Import section for Python 2 and 3 compatible code # from __future__ import absolute_import, division, print_function, # unicode_literals from __future__ import division # This way: 3 / 2 == 1.5; 3 // 2 == 1 # ----------------------------------------------------------------------------- # Import section # import logging import multiprocessing from itertools import islice import lib.NicePrint as NicePrint # ========================================================================= # Functions aliases # # NPR.NicePrint = from NicePrint module # ------------------------------------------------------------------------- NPR = NicePrint.NicePrint() # ------------------------------------------------------------------------- # use_lock # # Control use of DB lock. acquire/release # def use_lock(adb_lock, operation, nprocs=0): """ use_lock adb_lock = lock to be used operation = True => Lock = False => Release nprocs = >0 when in multiprocessing mode >>> alock = multiprocessing.Lock() >>> use_lock(alock, True, 2) True >>> use_lock(alock, False, 2) True """ use_dblock_return = False logging.debug('Entering use_lock with operation:[%s].', operation) if adb_lock is None: logging.debug('use_lock: adb_lock is [None].') return use_dblock_return logging.debug('use_lock: adb_lock.semlock:[%s].', adb_lock._semlock) if operation is None: return use_dblock_return if (nprocs is not None) and (nprocs) and (nprocs > 0): if operation: # Control for when running multiprocessing set locking logging.debug('===Multiprocessing=== -->[ ].lock.acquire') try: if adb_lock.acquire(): use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='002', caughtmsg='Caught an exception lock.acquire', useniceprint=True, exceptsysinfo=True) raise logging.info('===Multiprocessing=== --->[v].lock.acquire') else: # Control for when running multiprocessing release locking logging.debug('===Multiprocessing=== <--[ ].lock.release') try: adb_lock.release() use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='003', caughtmsg='Caught an exception lock.release', useniceprint=True, exceptsysinfo=True) # Raise aborts execution raise logging.info('===Multiprocessing=== <--[v].lock.release') logging.info('Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) else: use_dblock_return = True logging.warning('(No multiprocessing. Nothing to do) ' 'Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) return use_dblock_return # ----------------------------------------------------------------------------- # mprocessing # def mprocessing(nprocs, lockdb, running, mutex, itemslist, a_fn, cur): """ mprocessing Function nprocs = Number of processes to launch (int) lockdb = lock for access to Database (lock obj to be created) running = Value to count processed items (count obj to be created) mutex = mutex for access to value running (obj to be created) itemslist = list of items to be processed a_fn = a function which is the target of the multiprocessing a_fn must cater the following arguments lockdb running mutex splititemslist = partial splitted list count_total = len(itemslist) cur cur = cursor variable for DB access """ # proc_pool = Local variable proc_pool for Pool of processes # log_level = log_level # count_total = Total counter of items to distribute/play/indicate progress # len(itemslist) log_level = logging.getLogger().getEffectiveLevel() logging.info('===mprocessing [%s] target_fn():[%s] nprocs:[%s]', __name__, a_fn.__name__, nprocs) # if log_level <= logging.WARNING: # if args is not None: # for i, arg in enumerate(args): # logging.info('===mprocessing f():[%s] arg[%s]={%s}', # a_fn.__name__, i, arg) # if __name__ == '__main__': logging.debug('===Multiprocessing=== Setting up logger!')
logger = multiprocessing.get_logger() logger.setLevel(log_level) logging.debug('===Multiprocessing=== Logging defined!') # --------------------------------------------------------- # chunk # # Divides an iterable in slices/chunks of size size # def chunk(iter_list, size): """ Divides an iterable in slices/chunks of size size >>> for a in chunk([ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10], 3): ... len(a) 3 3 3 1 """ iter_list = iter(iter_list) # lambda: creates a returning expression function # which returns slices # iter, with the second argument () stops creating # iterators when it reaches the end return iter(lambda: tuple(islice(iter_list, size)), ()) proc_pool = [] lockdb = multiprocessing.Lock() running = multiprocessing.Value('i', 0) mutex = multiprocessing.Lock() count_total = len(itemslist) size = (len(itemslist) // int(nprocs)) \ if ((len(itemslist) // int(nprocs)) > 0) \ else 1 logging.debug('len(itemslist):[%s] int(nprocs):[%s] size per process:[%s]', len(itemslist), int(nprocs), size) # Split itemslist in chunks to distribute accross Processes for splititemslist in chunk(itemslist, size): logging.warning('===Actual/Planned Chunk size: [%s]/[%s]', len(splititemslist), size) logging.debug('===type(splititemslist)=[%s]', type(splititemslist)) logging.debug('===Job/Task Process: Creating...') proc_task = multiprocessing.Process( target=a_fn, # argument function args=(lockdb, running, mutex, splititemslist, count_total, cur,)) proc_pool.append(proc_task) logging.debug('===Job/Task Process: Starting...') proc_task.start() NPR.niceprint('===Job/Task Process: [{!s}] Started ' 'with pid:[{!s}]' .format(proc_task.name, proc_task.pid), verbosity=3, logalso=logging.DEBUG) # Check status of jobs/tasks in the Process Pool if log_level <= logging.DEBUG: NPR.niceprint('===Checking Processes launched/status:', verbosity=3, logalso=logging.DEBUG) for j in proc_pool: NPR.niceprint('{!s}.is_alive = {!s}'.format(j.name, j.is_alive()), verbosity=3, logalso=logging.DEBUG) # Regularly print status of jobs/tasks in the Process Pool # Prints status while there are processes active # Exits when all jobs/tasks are done. while True: if not any(multiprocessing.active_children()): logging.debug('===No active children Processes.') break for prc in multiprocessing.active_children(): logging.debug('===%s.is_alive = %s', prc.name, prc.is_alive()) proc_task_active = prc NPR.niceprint('===Will wait for 60 on {!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) proc_task_active.join(timeout=60) NPR.niceprint('===Waited for 60s on ' '{!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) # Wait for join all jobs/tasks in the Process Pool # All should be done by now! for j in proc_pool: j.join() NPR.niceprint('==={!s} (is alive: {!s}).exitcode = {!s}' .format(j.name, j.is_alive(), j.exitcode), verbosity=2) logging.warning('===Multiprocessing=== pool joined! ' 'All processes finished.') # Will release (set to None) the lockdb lock control # this prevents subsequent calls to # use_lock( nuLockDB, False) # to raise exception: # ValueError('semaphore or lock released too many times') logging.info('===Multiprocessing=== pool joined! ' 'Is lockdb None? [%s]. Setting lockdb to None anyhow.', lockdb is None) lockdb = None # Show number of total files processed NPR.niceprocessedfiles(running.value, count_total, True) return True # ----------------------------------------------------------------------------- # If called directly run doctests # if __name__ == "__main__": logging.basicConfig(level=logging.WARNING, format='[%(asctime)s]:[%(processName)-11s]' + '[%(levelname)-8s]:[%(name)s] %(message)s') import doctest doctest.testmod()
# CODING No need for such low level debugging to stderr # multiprocessing.log_to_stderr()
random_line_split
mprocessing.py
""" by oPromessa, 2018 Published on https://github.com/oPromessa/flickr-uploader/ mprocessing = Helper function to run function in multiprocessing mode """ # ----------------------------------------------------------------------------- # Import section for Python 2 and 3 compatible code # from __future__ import absolute_import, division, print_function, # unicode_literals from __future__ import division # This way: 3 / 2 == 1.5; 3 // 2 == 1 # ----------------------------------------------------------------------------- # Import section # import logging import multiprocessing from itertools import islice import lib.NicePrint as NicePrint # ========================================================================= # Functions aliases # # NPR.NicePrint = from NicePrint module # ------------------------------------------------------------------------- NPR = NicePrint.NicePrint() # ------------------------------------------------------------------------- # use_lock # # Control use of DB lock. acquire/release # def use_lock(adb_lock, operation, nprocs=0): """ use_lock adb_lock = lock to be used operation = True => Lock = False => Release nprocs = >0 when in multiprocessing mode >>> alock = multiprocessing.Lock() >>> use_lock(alock, True, 2) True >>> use_lock(alock, False, 2) True """ use_dblock_return = False logging.debug('Entering use_lock with operation:[%s].', operation) if adb_lock is None: logging.debug('use_lock: adb_lock is [None].') return use_dblock_return logging.debug('use_lock: adb_lock.semlock:[%s].', adb_lock._semlock) if operation is None: return use_dblock_return if (nprocs is not None) and (nprocs) and (nprocs > 0): if operation: # Control for when running multiprocessing set locking logging.debug('===Multiprocessing=== -->[ ].lock.acquire') try: if adb_lock.acquire(): use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='002', caughtmsg='Caught an exception lock.acquire', useniceprint=True, exceptsysinfo=True) raise logging.info('===Multiprocessing=== --->[v].lock.acquire') else: # Control for when running multiprocessing release locking logging.debug('===Multiprocessing=== <--[ ].lock.release') try: adb_lock.release() use_dblock_return = True except Exception: NPR.niceerror(caught=True, caughtprefix='+++ ', caughtcode='003', caughtmsg='Caught an exception lock.release', useniceprint=True, exceptsysinfo=True) # Raise aborts execution raise logging.info('===Multiprocessing=== <--[v].lock.release') logging.info('Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) else: use_dblock_return = True logging.warning('(No multiprocessing. Nothing to do) ' 'Exiting use_lock with operation:[%s]. Result:[%s]', operation, use_dblock_return) return use_dblock_return # ----------------------------------------------------------------------------- # mprocessing # def
(nprocs, lockdb, running, mutex, itemslist, a_fn, cur): """ mprocessing Function nprocs = Number of processes to launch (int) lockdb = lock for access to Database (lock obj to be created) running = Value to count processed items (count obj to be created) mutex = mutex for access to value running (obj to be created) itemslist = list of items to be processed a_fn = a function which is the target of the multiprocessing a_fn must cater the following arguments lockdb running mutex splititemslist = partial splitted list count_total = len(itemslist) cur cur = cursor variable for DB access """ # proc_pool = Local variable proc_pool for Pool of processes # log_level = log_level # count_total = Total counter of items to distribute/play/indicate progress # len(itemslist) log_level = logging.getLogger().getEffectiveLevel() logging.info('===mprocessing [%s] target_fn():[%s] nprocs:[%s]', __name__, a_fn.__name__, nprocs) # if log_level <= logging.WARNING: # if args is not None: # for i, arg in enumerate(args): # logging.info('===mprocessing f():[%s] arg[%s]={%s}', # a_fn.__name__, i, arg) # if __name__ == '__main__': logging.debug('===Multiprocessing=== Setting up logger!') # CODING No need for such low level debugging to stderr # multiprocessing.log_to_stderr() logger = multiprocessing.get_logger() logger.setLevel(log_level) logging.debug('===Multiprocessing=== Logging defined!') # --------------------------------------------------------- # chunk # # Divides an iterable in slices/chunks of size size # def chunk(iter_list, size): """ Divides an iterable in slices/chunks of size size >>> for a in chunk([ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10], 3): ... len(a) 3 3 3 1 """ iter_list = iter(iter_list) # lambda: creates a returning expression function # which returns slices # iter, with the second argument () stops creating # iterators when it reaches the end return iter(lambda: tuple(islice(iter_list, size)), ()) proc_pool = [] lockdb = multiprocessing.Lock() running = multiprocessing.Value('i', 0) mutex = multiprocessing.Lock() count_total = len(itemslist) size = (len(itemslist) // int(nprocs)) \ if ((len(itemslist) // int(nprocs)) > 0) \ else 1 logging.debug('len(itemslist):[%s] int(nprocs):[%s] size per process:[%s]', len(itemslist), int(nprocs), size) # Split itemslist in chunks to distribute accross Processes for splititemslist in chunk(itemslist, size): logging.warning('===Actual/Planned Chunk size: [%s]/[%s]', len(splititemslist), size) logging.debug('===type(splititemslist)=[%s]', type(splititemslist)) logging.debug('===Job/Task Process: Creating...') proc_task = multiprocessing.Process( target=a_fn, # argument function args=(lockdb, running, mutex, splititemslist, count_total, cur,)) proc_pool.append(proc_task) logging.debug('===Job/Task Process: Starting...') proc_task.start() NPR.niceprint('===Job/Task Process: [{!s}] Started ' 'with pid:[{!s}]' .format(proc_task.name, proc_task.pid), verbosity=3, logalso=logging.DEBUG) # Check status of jobs/tasks in the Process Pool if log_level <= logging.DEBUG: NPR.niceprint('===Checking Processes launched/status:', verbosity=3, logalso=logging.DEBUG) for j in proc_pool: NPR.niceprint('{!s}.is_alive = {!s}'.format(j.name, j.is_alive()), verbosity=3, logalso=logging.DEBUG) # Regularly print status of jobs/tasks in the Process Pool # Prints status while there are processes active # Exits when all jobs/tasks are done. while True: if not any(multiprocessing.active_children()): logging.debug('===No active children Processes.') break for prc in multiprocessing.active_children(): logging.debug('===%s.is_alive = %s', prc.name, prc.is_alive()) proc_task_active = prc NPR.niceprint('===Will wait for 60 on {!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) proc_task_active.join(timeout=60) NPR.niceprint('===Waited for 60s on ' '{!s}.is_alive = {!s}' .format(proc_task_active.name, proc_task_active.is_alive()), verbosity=3, logalso=logging.INFO) # Wait for join all jobs/tasks in the Process Pool # All should be done by now! for j in proc_pool: j.join() NPR.niceprint('==={!s} (is alive: {!s}).exitcode = {!s}' .format(j.name, j.is_alive(), j.exitcode), verbosity=2) logging.warning('===Multiprocessing=== pool joined! ' 'All processes finished.') # Will release (set to None) the lockdb lock control # this prevents subsequent calls to # use_lock( nuLockDB, False) # to raise exception: # ValueError('semaphore or lock released too many times') logging.info('===Multiprocessing=== pool joined! ' 'Is lockdb None? [%s]. Setting lockdb to None anyhow.', lockdb is None) lockdb = None # Show number of total files processed NPR.niceprocessedfiles(running.value, count_total, True) return True # ----------------------------------------------------------------------------- # If called directly run doctests # if __name__ == "__main__": logging.basicConfig(level=logging.WARNING, format='[%(asctime)s]:[%(processName)-11s]' + '[%(levelname)-8s]:[%(name)s] %(message)s') import doctest doctest.testmod()
mprocessing
identifier_name
aes.rs
use std::ops::BitXor; use std::ops::BitXorAssign; use util; use xor; const RCON: [u8; 11] = [ 0x8d, 0x01, 0x02, 0x04, 0x08, 0x10, 0x20, 0x40, 0x80, 0x1b, 0x36, ]; const SBOX: [u8; 256] = [ 0x63, 0x7C, 0x77, 0x7B, 0xF2, 0x6B, 0x6F, 0xC5, 0x30, 0x01, 0x67, 0x2B, 0xFE, 0xD7, 0xAB, 0x76, 0xCA, 0x82, 0xC9, 0x7D, 0xFA, 0x59, 0x47, 0xF0, 0xAD, 0xD4, 0xA2, 0xAF, 0x9C, 0xA4, 0x72, 0xC0, 0xB7, 0xFD, 0x93, 0x26, 0x36, 0x3F, 0xF7, 0xCC, 0x34, 0xA5, 0xE5, 0xF1, 0x71, 0xD8, 0x31, 0x15, 0x04, 0xC7, 0x23, 0xC3, 0x18, 0x96, 0x05, 0x9A, 0x07, 0x12, 0x80, 0xE2, 0xEB, 0x27, 0xB2, 0x75, 0x09, 0x83, 0x2C, 0x1A, 0x1B, 0x6E, 0x5A, 0xA0, 0x52, 0x3B, 0xD6, 0xB3, 0x29, 0xE3, 0x2F, 0x84, 0x53, 0xD1, 0x00, 0xED, 0x20, 0xFC, 0xB1, 0x5B, 0x6A, 0xCB, 0xBE, 0x39, 0x4A, 0x4C, 0x58, 0xCF, 0xD0, 0xEF, 0xAA, 0xFB, 0x43, 0x4D, 0x33, 0x85, 0x45, 0xF9, 0x02, 0x7F, 0x50, 0x3C, 0x9F, 0xA8, 0x51, 0xA3, 0x40, 0x8F, 0x92, 0x9D, 0x38, 0xF5, 0xBC, 0xB6, 0xDA, 0x21, 0x10, 0xFF, 0xF3, 0xD2, 0xCD, 0x0C, 0x13, 0xEC, 0x5F, 0x97, 0x44, 0x17, 0xC4, 0xA7, 0x7E, 0x3D, 0x64, 0x5D, 0x19, 0x73, 0x60, 0x81, 0x4F, 0xDC, 0x22, 0x2A, 0x90, 0x88, 0x46, 0xEE, 0xB8, 0x14, 0xDE, 0x5E, 0x0B, 0xDB, 0xE0, 0x32, 0x3A, 0x0A, 0x49, 0x06, 0x24, 0x5C, 0xC2, 0xD3, 0xAC, 0x62, 0x91, 0x95, 0xE4, 0x79, 0xE7, 0xC8, 0x37, 0x6D, 0x8D, 0xD5, 0x4E, 0xA9, 0x6C, 0x56, 0xF4, 0xEA, 0x65, 0x7A, 0xAE, 0x08, 0xBA, 0x78, 0x25, 0x2E, 0x1C, 0xA6, 0xB4, 0xC6, 0xE8, 0xDD, 0x74, 0x1F, 0x4B, 0xBD, 0x8B, 0x8A, 0x70, 0x3E, 0xB5, 0x66, 0x48, 0x03, 0xF6, 0x0E, 0x61, 0x35, 0x57, 0xB9, 0x86, 0xC1, 0x1D, 0x9E, 0xE1, 0xF8, 0x98, 0x11, 0x69, 0xD9, 0x8E, 0x94, 0x9B, 0x1E, 0x87, 0xE9, 0xCE, 0x55, 0x28, 0xDF, 0x8C, 0xA1, 0x89, 0x0D, 0xBF, 0xE6, 0x42, 0x68, 0x41, 0x99, 0x2D, 0x0F, 0xB0, 0x54, 0xBB, 0x16, ]; const INV_SBOX: [u8; 256] = [ 0x52, 0x09, 0x6A, 0xD5, 0x30, 0x36, 0xA5, 0x38, 0xBF, 0x40, 0xA3, 0x9E, 0x81, 0xF3, 0xD7, 0xFB, 0x7C, 0xE3, 0x39, 0x82, 0x9B, 0x2F, 0xFF, 0x87, 0x34, 0x8E, 0x43, 0x44, 0xC4, 0xDE, 0xE9, 0xCB, 0x54, 0x7B, 0x94, 0x32, 0xA6, 0xC2, 0x23, 0x3D, 0xEE, 0x4C, 0x95, 0x0B, 0x42, 0xFA, 0xC3, 0x4E, 0x08, 0x2E, 0xA1, 0x66, 0x28, 0xD9, 0x24, 0xB2, 0x76, 0x5B, 0xA2, 0x49, 0x6D, 0x8B, 0xD1, 0x25, 0x72, 0xF8, 0xF6, 0x64, 0x86, 0x68, 0x98, 0x16, 0xD4, 0xA4, 0x5C, 0xCC, 0x5D, 0x65, 0xB6, 0x92, 0x6C, 0x70, 0x48, 0x50, 0xFD, 0xED, 0xB9, 0xDA, 0x5E, 0x15, 0x46, 0x57, 0xA7, 0x8D, 0x9D, 0x84, 0x90, 0xD8, 0xAB, 0x00, 0x8C, 0xBC, 0xD3, 0x0A, 0xF7, 0xE4, 0x58, 0x05, 0xB8, 0xB3, 0x45, 0x06, 0xD0, 0x2C, 0x1E, 0x8F, 0xCA, 0x3F, 0x0F, 0x02, 0xC1, 0xAF, 0xBD, 0x03, 0x01, 0x13, 0x8A, 0x6B, 0x3A, 0x91, 0x11, 0x41, 0x4F, 0x67, 0xDC, 0xEA, 0x97, 0xF2, 0xCF, 0xCE, 0xF0, 0xB4, 0xE6, 0x73, 0x96, 0xAC, 0x74, 0x22, 0xE7, 0xAD, 0x35, 0x85, 0xE2, 0xF9, 0x37, 0xE8, 0x1C, 0x75, 0xDF, 0x6E, 0x47, 0xF1, 0x1A, 0x71, 0x1D, 0x29, 0xC5, 0x89, 0x6F, 0xB7, 0x62, 0x0E, 0xAA, 0x18, 0xBE, 0x1B, 0xFC, 0x56, 0x3E, 0x4B, 0xC6, 0xD2, 0x79, 0x20, 0x9A, 0xDB, 0xC0, 0xFE, 0x78, 0xCD, 0x5A, 0xF4, 0x1F, 0xDD, 0xA8, 0x33, 0x88, 0x07, 0xC7, 0x31, 0xB1, 0x12, 0x10, 0x59, 0x27, 0x80, 0xEC, 0x5F, 0x60, 0x51, 0x7F, 0xA9, 0x19, 0xB5, 0x4A, 0x0D, 0x2D, 0xE5, 0x7A, 0x9F, 0x93, 0xC9, 0x9C, 0xEF, 0xA0, 0xE0, 0x3B, 0x4D, 0xAE, 0x2A, 0xF5, 0xB0, 0xC8, 0xEB, 0xBB, 0x3C, 0x83, 0x53, 0x99, 0x61, 0x17, 0x2B, 0x04, 0x7E, 0xBA, 0x77, 0xD6, 0x26, 0xE1, 0x69, 0x14, 0x63, 0x55, 0x21, 0x0C, 0x7D, ]; fn sbox(data: &mut [u8]) { for e in data.iter_mut() { *e = SBOX[*e as usize]; } } fn inv_sbox(data: &mut [u8]) { for e in data.iter_mut() { *e = INV_SBOX[*e as usize]; } } fn to_four_byte_array(data: &[u8]) -> [u8; 4] { [data[0], data[1], data[2], data[3]] } enum KeyExpansionMode { Xor, Sbox, Full, } fn generate_four_bytes( key_length: usize, expanded_key: &[u8], rcon_iteration: &mut usize, mode: KeyExpansionMode, ) -> [u8; 4] { let i = expanded_key.len(); let source_bytes = &expanded_key[i - 4..i]; let mut t: [u8; 4] = to_four_byte_array(source_bytes); match mode { KeyExpansionMode::Xor => {} KeyExpansionMode::Sbox => { sbox(&mut t); } KeyExpansionMode::Full => { t.rotate_left(1); sbox(&mut t); t[0].bitxor_assign(RCON[*rcon_iteration]); *rcon_iteration += 1; } }; let xor_source = &expanded_key[i - key_length..i - key_length + 4]; xor::buffer_mut(&mut t, xor::Key::FullBuffer(xor_source)); t } fn expand_key(key: &[u8]) -> Vec<u8> { let key_length = key.len(); let (rounds, sbox_round, extra_expansions) = match key_length { 16 => (10, false, 0), 24 => (12, false, 2), 32 => (14, true, 3), len => panic!("Unsupported key length {}", len), }; let expanded_key_size = 16 * (rounds + 1); let mut expanded_key = Vec::with_capacity(expanded_key_size); expanded_key.extend_from_slice(&key); let mut rcon_iteration = 1usize; while expanded_key.len() < expanded_key_size { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Full, ); expanded_key.extend(t.iter()); for _i in 0..3 { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Xor, ); expanded_key.extend(t.iter()); } if sbox_round { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Sbox, ); expanded_key.extend(t.iter()); } for _i in 0..extra_expansions { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Xor, ); expanded_key.extend(t.iter()); } } // Truncate any extra bytes expanded_key.resize(expanded_key_size, 0); assert!( expanded_key.len() == expanded_key_size, "Expanded key is too long: {}", expanded_key.len(), ); expanded_key } #[test] fn expand_key_16() { use util::parse_byte_string; assert_eq!( parse_byte_string( "00000000000000000000000000000000626363636263636362636363626363639b9898c9f9fbfbaa9b9898c9f9fbfbaa90973450696ccffaf2f457330b0fac99ee06da7b876a1581759e42b27e91ee2b7f2e2b88f8443e098dda7cbbf34b9290ec614b851425758c99ff09376ab49ba7217517873550620bacaf6b3cc61bf09b0ef903333ba9613897060a04511dfa9fb1d4d8e28a7db9da1d7bb3de4c664941b4ef5bcb3e92e21123e951cf6f8f188e" ), expand_key(&vec![0; 16]) ); assert_eq!( parse_byte_string( "ffffffffffffffffffffffffffffffffe8e9e9e917161616e8e9e9e917161616adaeae19bab8b80f525151e6454747f0090e2277b3b69a78e1e7cb9ea4a08c6ee16abd3e52dc2746b33becd8179b60b6e5baf3ceb766d488045d385013c658e671d07db3c6b6a93bc2eb916bd12dc98de90d208d2fbb89b6ed5018dd3c7dd15096337366b988fad054d8e20d68a5335d8bf03f233278c5f366a027fe0e0514a3d60a3588e472f07b82d2d7858cd7c326" ), expand_key(&vec![0xff; 16]) ); assert_eq!( parse_byte_string( "000102030405060708090a0b0c0d0e0fd6aa74fdd2af72fadaa678f1d6ab76feb692cf0b643dbdf1be9bc5006830b3feb6ff744ed2c2c9bf6c590cbf0469bf4147f7f7bc95353e03f96c32bcfd058dfd3caaa3e8a99f9deb50f3af57adf622aa5e390f7df7a69296a7553dc10aa31f6b14f9701ae35fe28c440adf4d4ea9c02647438735a41c65b9e016baf4aebf7ad2549932d1f08557681093ed9cbe2c974e13111d7fe3944a17f307a78b4d2b30c5" ), expand_key(&parse_byte_string("000102030405060708090a0b0c0d0e0f")) ); assert_eq!( parse_byte_string( "6920e299a5202a6d656e636869746f2afa8807605fa82d0d3ac64e6553b2214fcf75838d90ddae80aa1be0e5f9a9c1aa180d2f1488d0819422cb6171db62a0dbbaed96ad323d173910f67648cb94d693881b4ab2ba265d8baad02bc36144fd50b34f195d096944d6a3b96f15c2fd9245a7007778ae6933ae0dd05cbbcf2dcefeff8bccf251e2ff5c5c32a3e7931f6d1924b7182e7555e77229674495ba78298cae127cdadb479ba8f220df3d4858f6b1" ), expand_key(&parse_byte_string("6920e299a5202a6d656e636869746f2a")) ); } #[test] fn expand_key_24() { assert_eq!(208, expand_key(&vec![0; 24]).len()); } #[test] fn expand_key_32() { assert_eq!(240, expand_key(&vec![0; 32]).len()); } fn add_round_key(state: &mut [u8], key: &[u8]) { xor::buffer_mut(state, xor::Key::FullBuffer(key)); } // Shifted by 0, 1, 2, 3 columns const ROW_SHIFTS: [usize; 16] = [0, 5, 10, 15, 4, 9, 14, 3, 8, 13, 2, 7, 12, 1, 6, 11]; fn shift_rows(state: &mut [u8]) { let copy = util::convert_to_fixed_array(state); for (index, e) in state.iter_mut().enumerate() { *e = copy[ROW_SHIFTS[index]]; } } #[test] fn test_shift_rows() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; shift_rows(&mut rows); assert_eq!( rows, [1, 6, 11, 16, 5, 10, 15, 4, 9, 14, 3, 8, 13, 2, 7, 12] ); } // Shifted by 0, -1, -2, -3 columns const INV_ROW_SHIFTS: [usize; 16] = [0, 13, 10, 7, 4, 1, 14, 11, 8, 5, 2, 15, 12, 9, 6, 3]; fn inv_shift_rows(state: &mut [u8]) { let copy = util::convert_to_fixed_array(state); for (index, e) in state.iter_mut().enumerate() { *e = copy[INV_ROW_SHIFTS[index]]; } } #[test] fn test_inv_shift_rows() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; inv_shift_rows(&mut rows); assert_ne!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); assert_eq!( rows, [1, 14, 11, 8, 5, 2, 15, 12, 9, 6, 3, 16, 13, 10, 7, 4,] ); } #[test] fn test_shift_rows_ident()
const COLUMN_MATRIX: [u8; 16] = [2, 3, 1, 1, 1, 2, 3, 1, 1, 1, 2, 3, 3, 1, 1, 2]; const INV_COLUMN_MATRIX: [u8; 16] = [14, 11, 13, 9, 9, 14, 11, 13, 13, 9, 14, 11, 11, 13, 9, 14]; fn gmul(mut a: u8, mut b: u8) -> u8 { let mut p = 0; for _ in 0..8 { if (b & 0x1) != 0 { p.bitxor_assign(a); } let has_high_bit = (a & 0x80) == 0x80; a <<= 1; if has_high_bit { a.bitxor_assign(0x1b); } b >>= 1; } p } fn mix_column(matrix: &[u8; 16], state_column: &[u8]) -> Vec<u8> { matrix .chunks(4) .map(|mc| { mc.iter() .enumerate() .map(|(i, &coefficient)| gmul(coefficient, state_column[i])) .fold(None, |accum, current| match accum { None => Some(current), Some(x) => Some(x.bitxor(current)), }) .unwrap() }) .collect() } #[test] fn test_mix_column() { use util::parse_byte_string; assert_eq!( parse_byte_string("8e4da1bc"), mix_column(&COLUMN_MATRIX, &parse_byte_string("db135345")), ); assert_eq!( parse_byte_string("9fdc589d"), mix_column(&COLUMN_MATRIX, &parse_byte_string("f20a225c")), ); assert_eq!( parse_byte_string("01010101"), mix_column(&COLUMN_MATRIX, &parse_byte_string("01010101")), ); assert_eq!( parse_byte_string("c6c6c6c6"), mix_column(&COLUMN_MATRIX, &parse_byte_string("c6c6c6c6")), ); assert_eq!( parse_byte_string("d5d5d7d6"), mix_column(&COLUMN_MATRIX, &parse_byte_string("d4d4d4d5")), ); assert_eq!( parse_byte_string("4d7ebdf8"), mix_column(&COLUMN_MATRIX, &parse_byte_string("2d26314c")), ); } fn mix_columns(state: &mut [u8]) { for column in state.chunks_mut(4) { let new_column = mix_column(&COLUMN_MATRIX, column); column.copy_from_slice(&new_column); } } fn inv_mix_columns(state: &mut [u8]) { for column in state.chunks_mut(4) { let new_column = mix_column(&INV_COLUMN_MATRIX, column); column.copy_from_slice(&new_column); } } #[derive(Clone)] pub enum CipherMode { ECB, CBC([u8; 16]), } fn transform_chunk(chunk: &[u8], expanded_key: &[u8], operation: Operation) -> [u8; 16] { const STATE_SIZE: usize = 16; assert!( chunk.len() == STATE_SIZE, "Chunk size of {} is invalid; expected {}", chunk.len(), STATE_SIZE ); let last_round = expanded_key.chunks(STATE_SIZE).count() - 1; let mut state = util::convert_to_fixed_array(chunk); match operation { Operation::Encrypt => { for (round, round_key) in expanded_key.chunks(STATE_SIZE).enumerate() { match round { 0 => { add_round_key(&mut state, round_key); } n if n != last_round => { sbox(&mut state); shift_rows(&mut state); mix_columns(&mut state); add_round_key(&mut state, round_key); } _ => { sbox(&mut state); shift_rows(&mut state); add_round_key(&mut state, round_key); } } } } Operation::Decrypt => { for (round, round_key) in expanded_key.chunks(STATE_SIZE).rev().enumerate() { match round { 0 => { add_round_key(&mut state, round_key); } n if n != last_round => { inv_shift_rows(&mut state); inv_sbox(&mut state); add_round_key(&mut state, round_key); inv_mix_columns(&mut state); } _ => { inv_shift_rows(&mut state); inv_sbox(&mut state); add_round_key(&mut state, round_key); } } } } }; state } #[derive(Clone, Copy)] pub enum Operation { Encrypt, Decrypt, } trait CipherModeImpl { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8>; } struct ECBCipherMode {} struct CBCCipherMode { initialization_vector: [u8; 16], operation: Operation, } impl CBCCipherMode { fn transform(&mut self, chunk: &[u8; 16], transform: &Fn(&[u8; 16]) -> [u8; 16]) -> [u8; 16] { match self.operation { Operation::Encrypt => { xor::buffer_mut(&mut self.initialization_vector, xor::Key::FullBuffer(chunk)); self.initialization_vector = transform(&self.initialization_vector); util::convert_to_fixed_array(&self.initialization_vector) } Operation::Decrypt => { let mut plaintext = transform(chunk); xor::buffer_mut( &mut plaintext, xor::Key::FullBuffer(&self.initialization_vector), ); self.initialization_vector = util::convert_to_fixed_array(chunk); plaintext } } } } impl CipherModeImpl for ECBCipherMode { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8> { let mut v = Vec::with_capacity(data.len()); for chunk in data.chunks(16) { v.extend(&transform(&util::convert_to_fixed_array(chunk))); } v } } impl CipherModeImpl for CBCCipherMode { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8> { let mut v = Vec::with_capacity(data.len()); for chunk in data.chunks(16) { v.extend( self .transform(&util::convert_to_fixed_array(chunk), transform) .iter(), ); } v } } pub fn perform(data: &[u8], key: &[u8], operation: Operation, cipher_mode: CipherMode) -> Vec<u8> { let expanded_key = expand_key(key); let mut cipher_mode_impl: Box<CipherModeImpl> = match cipher_mode { CipherMode::ECB => Box::new(ECBCipherMode {}), CipherMode::CBC(iv) => Box::new(CBCCipherMode { initialization_vector: iv, operation, }), }; cipher_mode_impl.transform_chunks(data, &|pre_transformed_chunk| { transform_chunk(pre_transformed_chunk, &expanded_key, operation) }) } #[test] fn ecb_once_16() { let plaintext = &vec![0; 16]; let key = &vec![0; 16]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); } #[test] fn ecb_once_24() { let plaintext = &vec![0; 32]; let key = &vec![0; 24]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); } #[test] fn ecb_once_32() { let plaintext = &vec![0; 32]; let key = &vec![0; 32]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); }
{ let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; shift_rows(&mut rows); assert_ne!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); inv_shift_rows(&mut rows); assert_eq!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); }
identifier_body
aes.rs
use std::ops::BitXor; use std::ops::BitXorAssign; use util; use xor; const RCON: [u8; 11] = [ 0x8d, 0x01, 0x02, 0x04, 0x08, 0x10, 0x20, 0x40, 0x80, 0x1b, 0x36, ]; const SBOX: [u8; 256] = [ 0x63, 0x7C, 0x77, 0x7B, 0xF2, 0x6B, 0x6F, 0xC5, 0x30, 0x01, 0x67, 0x2B, 0xFE, 0xD7, 0xAB, 0x76, 0xCA, 0x82, 0xC9, 0x7D, 0xFA, 0x59, 0x47, 0xF0, 0xAD, 0xD4, 0xA2, 0xAF, 0x9C, 0xA4, 0x72, 0xC0, 0xB7, 0xFD, 0x93, 0x26, 0x36, 0x3F, 0xF7, 0xCC, 0x34, 0xA5, 0xE5, 0xF1, 0x71, 0xD8, 0x31, 0x15, 0x04, 0xC7, 0x23, 0xC3, 0x18, 0x96, 0x05, 0x9A, 0x07, 0x12, 0x80, 0xE2, 0xEB, 0x27, 0xB2, 0x75, 0x09, 0x83, 0x2C, 0x1A, 0x1B, 0x6E, 0x5A, 0xA0, 0x52, 0x3B, 0xD6, 0xB3, 0x29, 0xE3, 0x2F, 0x84, 0x53, 0xD1, 0x00, 0xED, 0x20, 0xFC, 0xB1, 0x5B, 0x6A, 0xCB, 0xBE, 0x39, 0x4A, 0x4C, 0x58, 0xCF, 0xD0, 0xEF, 0xAA, 0xFB, 0x43, 0x4D, 0x33, 0x85, 0x45, 0xF9, 0x02, 0x7F, 0x50, 0x3C, 0x9F, 0xA8, 0x51, 0xA3, 0x40, 0x8F, 0x92, 0x9D, 0x38, 0xF5, 0xBC, 0xB6, 0xDA, 0x21, 0x10, 0xFF, 0xF3, 0xD2, 0xCD, 0x0C, 0x13, 0xEC, 0x5F, 0x97, 0x44, 0x17, 0xC4, 0xA7, 0x7E, 0x3D, 0x64, 0x5D, 0x19, 0x73, 0x60, 0x81, 0x4F, 0xDC, 0x22, 0x2A, 0x90, 0x88, 0x46, 0xEE, 0xB8, 0x14, 0xDE, 0x5E, 0x0B, 0xDB, 0xE0, 0x32, 0x3A, 0x0A, 0x49, 0x06, 0x24, 0x5C, 0xC2, 0xD3, 0xAC, 0x62, 0x91, 0x95, 0xE4, 0x79, 0xE7, 0xC8, 0x37, 0x6D, 0x8D, 0xD5, 0x4E, 0xA9, 0x6C, 0x56, 0xF4, 0xEA, 0x65, 0x7A, 0xAE, 0x08, 0xBA, 0x78, 0x25, 0x2E, 0x1C, 0xA6, 0xB4, 0xC6, 0xE8, 0xDD, 0x74, 0x1F, 0x4B, 0xBD, 0x8B, 0x8A, 0x70, 0x3E, 0xB5, 0x66, 0x48, 0x03, 0xF6, 0x0E, 0x61, 0x35, 0x57, 0xB9, 0x86, 0xC1, 0x1D, 0x9E, 0xE1, 0xF8, 0x98, 0x11, 0x69, 0xD9, 0x8E, 0x94, 0x9B, 0x1E, 0x87, 0xE9, 0xCE, 0x55, 0x28, 0xDF, 0x8C, 0xA1, 0x89, 0x0D, 0xBF, 0xE6, 0x42, 0x68, 0x41, 0x99, 0x2D, 0x0F, 0xB0, 0x54, 0xBB, 0x16, ]; const INV_SBOX: [u8; 256] = [ 0x52, 0x09, 0x6A, 0xD5, 0x30, 0x36, 0xA5, 0x38, 0xBF, 0x40, 0xA3, 0x9E, 0x81, 0xF3, 0xD7, 0xFB, 0x7C, 0xE3, 0x39, 0x82, 0x9B, 0x2F, 0xFF, 0x87, 0x34, 0x8E, 0x43, 0x44, 0xC4, 0xDE, 0xE9, 0xCB, 0x54, 0x7B, 0x94, 0x32, 0xA6, 0xC2, 0x23, 0x3D, 0xEE, 0x4C, 0x95, 0x0B, 0x42, 0xFA, 0xC3, 0x4E, 0x08, 0x2E, 0xA1, 0x66, 0x28, 0xD9, 0x24, 0xB2, 0x76, 0x5B, 0xA2, 0x49, 0x6D, 0x8B, 0xD1, 0x25, 0x72, 0xF8, 0xF6, 0x64, 0x86, 0x68, 0x98, 0x16, 0xD4, 0xA4, 0x5C, 0xCC, 0x5D, 0x65, 0xB6, 0x92, 0x6C, 0x70, 0x48, 0x50, 0xFD, 0xED, 0xB9, 0xDA, 0x5E, 0x15, 0x46, 0x57, 0xA7, 0x8D, 0x9D, 0x84, 0x90, 0xD8, 0xAB, 0x00, 0x8C, 0xBC, 0xD3, 0x0A, 0xF7, 0xE4, 0x58, 0x05, 0xB8, 0xB3, 0x45, 0x06, 0xD0, 0x2C, 0x1E, 0x8F, 0xCA, 0x3F, 0x0F, 0x02, 0xC1, 0xAF, 0xBD, 0x03, 0x01, 0x13, 0x8A, 0x6B, 0x3A, 0x91, 0x11, 0x41, 0x4F, 0x67, 0xDC, 0xEA, 0x97, 0xF2, 0xCF, 0xCE, 0xF0, 0xB4, 0xE6, 0x73, 0x96, 0xAC, 0x74, 0x22, 0xE7, 0xAD, 0x35, 0x85, 0xE2, 0xF9, 0x37, 0xE8, 0x1C, 0x75, 0xDF, 0x6E, 0x47, 0xF1, 0x1A, 0x71, 0x1D, 0x29, 0xC5, 0x89, 0x6F, 0xB7, 0x62, 0x0E, 0xAA, 0x18, 0xBE, 0x1B, 0xFC, 0x56, 0x3E, 0x4B, 0xC6, 0xD2, 0x79, 0x20, 0x9A, 0xDB, 0xC0, 0xFE, 0x78, 0xCD, 0x5A, 0xF4, 0x1F, 0xDD, 0xA8, 0x33, 0x88, 0x07, 0xC7, 0x31, 0xB1, 0x12, 0x10, 0x59, 0x27, 0x80, 0xEC, 0x5F, 0x60, 0x51, 0x7F, 0xA9, 0x19, 0xB5, 0x4A, 0x0D, 0x2D, 0xE5, 0x7A, 0x9F, 0x93, 0xC9, 0x9C, 0xEF, 0xA0, 0xE0, 0x3B, 0x4D, 0xAE, 0x2A, 0xF5, 0xB0, 0xC8, 0xEB, 0xBB, 0x3C, 0x83, 0x53, 0x99, 0x61, 0x17, 0x2B, 0x04, 0x7E, 0xBA, 0x77, 0xD6, 0x26, 0xE1, 0x69, 0x14, 0x63, 0x55, 0x21, 0x0C, 0x7D, ]; fn sbox(data: &mut [u8]) { for e in data.iter_mut() { *e = SBOX[*e as usize]; } } fn inv_sbox(data: &mut [u8]) { for e in data.iter_mut() { *e = INV_SBOX[*e as usize]; } } fn to_four_byte_array(data: &[u8]) -> [u8; 4] { [data[0], data[1], data[2], data[3]] } enum KeyExpansionMode { Xor, Sbox, Full, } fn generate_four_bytes( key_length: usize, expanded_key: &[u8], rcon_iteration: &mut usize, mode: KeyExpansionMode, ) -> [u8; 4] { let i = expanded_key.len(); let source_bytes = &expanded_key[i - 4..i]; let mut t: [u8; 4] = to_four_byte_array(source_bytes); match mode { KeyExpansionMode::Xor => {} KeyExpansionMode::Sbox => { sbox(&mut t); } KeyExpansionMode::Full => { t.rotate_left(1); sbox(&mut t); t[0].bitxor_assign(RCON[*rcon_iteration]); *rcon_iteration += 1; } }; let xor_source = &expanded_key[i - key_length..i - key_length + 4]; xor::buffer_mut(&mut t, xor::Key::FullBuffer(xor_source)); t } fn
(key: &[u8]) -> Vec<u8> { let key_length = key.len(); let (rounds, sbox_round, extra_expansions) = match key_length { 16 => (10, false, 0), 24 => (12, false, 2), 32 => (14, true, 3), len => panic!("Unsupported key length {}", len), }; let expanded_key_size = 16 * (rounds + 1); let mut expanded_key = Vec::with_capacity(expanded_key_size); expanded_key.extend_from_slice(&key); let mut rcon_iteration = 1usize; while expanded_key.len() < expanded_key_size { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Full, ); expanded_key.extend(t.iter()); for _i in 0..3 { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Xor, ); expanded_key.extend(t.iter()); } if sbox_round { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Sbox, ); expanded_key.extend(t.iter()); } for _i in 0..extra_expansions { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Xor, ); expanded_key.extend(t.iter()); } } // Truncate any extra bytes expanded_key.resize(expanded_key_size, 0); assert!( expanded_key.len() == expanded_key_size, "Expanded key is too long: {}", expanded_key.len(), ); expanded_key } #[test] fn expand_key_16() { use util::parse_byte_string; assert_eq!( parse_byte_string( "00000000000000000000000000000000626363636263636362636363626363639b9898c9f9fbfbaa9b9898c9f9fbfbaa90973450696ccffaf2f457330b0fac99ee06da7b876a1581759e42b27e91ee2b7f2e2b88f8443e098dda7cbbf34b9290ec614b851425758c99ff09376ab49ba7217517873550620bacaf6b3cc61bf09b0ef903333ba9613897060a04511dfa9fb1d4d8e28a7db9da1d7bb3de4c664941b4ef5bcb3e92e21123e951cf6f8f188e" ), expand_key(&vec![0; 16]) ); assert_eq!( parse_byte_string( "ffffffffffffffffffffffffffffffffe8e9e9e917161616e8e9e9e917161616adaeae19bab8b80f525151e6454747f0090e2277b3b69a78e1e7cb9ea4a08c6ee16abd3e52dc2746b33becd8179b60b6e5baf3ceb766d488045d385013c658e671d07db3c6b6a93bc2eb916bd12dc98de90d208d2fbb89b6ed5018dd3c7dd15096337366b988fad054d8e20d68a5335d8bf03f233278c5f366a027fe0e0514a3d60a3588e472f07b82d2d7858cd7c326" ), expand_key(&vec![0xff; 16]) ); assert_eq!( parse_byte_string( "000102030405060708090a0b0c0d0e0fd6aa74fdd2af72fadaa678f1d6ab76feb692cf0b643dbdf1be9bc5006830b3feb6ff744ed2c2c9bf6c590cbf0469bf4147f7f7bc95353e03f96c32bcfd058dfd3caaa3e8a99f9deb50f3af57adf622aa5e390f7df7a69296a7553dc10aa31f6b14f9701ae35fe28c440adf4d4ea9c02647438735a41c65b9e016baf4aebf7ad2549932d1f08557681093ed9cbe2c974e13111d7fe3944a17f307a78b4d2b30c5" ), expand_key(&parse_byte_string("000102030405060708090a0b0c0d0e0f")) ); assert_eq!( parse_byte_string( "6920e299a5202a6d656e636869746f2afa8807605fa82d0d3ac64e6553b2214fcf75838d90ddae80aa1be0e5f9a9c1aa180d2f1488d0819422cb6171db62a0dbbaed96ad323d173910f67648cb94d693881b4ab2ba265d8baad02bc36144fd50b34f195d096944d6a3b96f15c2fd9245a7007778ae6933ae0dd05cbbcf2dcefeff8bccf251e2ff5c5c32a3e7931f6d1924b7182e7555e77229674495ba78298cae127cdadb479ba8f220df3d4858f6b1" ), expand_key(&parse_byte_string("6920e299a5202a6d656e636869746f2a")) ); } #[test] fn expand_key_24() { assert_eq!(208, expand_key(&vec![0; 24]).len()); } #[test] fn expand_key_32() { assert_eq!(240, expand_key(&vec![0; 32]).len()); } fn add_round_key(state: &mut [u8], key: &[u8]) { xor::buffer_mut(state, xor::Key::FullBuffer(key)); } // Shifted by 0, 1, 2, 3 columns const ROW_SHIFTS: [usize; 16] = [0, 5, 10, 15, 4, 9, 14, 3, 8, 13, 2, 7, 12, 1, 6, 11]; fn shift_rows(state: &mut [u8]) { let copy = util::convert_to_fixed_array(state); for (index, e) in state.iter_mut().enumerate() { *e = copy[ROW_SHIFTS[index]]; } } #[test] fn test_shift_rows() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; shift_rows(&mut rows); assert_eq!( rows, [1, 6, 11, 16, 5, 10, 15, 4, 9, 14, 3, 8, 13, 2, 7, 12] ); } // Shifted by 0, -1, -2, -3 columns const INV_ROW_SHIFTS: [usize; 16] = [0, 13, 10, 7, 4, 1, 14, 11, 8, 5, 2, 15, 12, 9, 6, 3]; fn inv_shift_rows(state: &mut [u8]) { let copy = util::convert_to_fixed_array(state); for (index, e) in state.iter_mut().enumerate() { *e = copy[INV_ROW_SHIFTS[index]]; } } #[test] fn test_inv_shift_rows() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; inv_shift_rows(&mut rows); assert_ne!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); assert_eq!( rows, [1, 14, 11, 8, 5, 2, 15, 12, 9, 6, 3, 16, 13, 10, 7, 4,] ); } #[test] fn test_shift_rows_ident() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; shift_rows(&mut rows); assert_ne!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); inv_shift_rows(&mut rows); assert_eq!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); } const COLUMN_MATRIX: [u8; 16] = [2, 3, 1, 1, 1, 2, 3, 1, 1, 1, 2, 3, 3, 1, 1, 2]; const INV_COLUMN_MATRIX: [u8; 16] = [14, 11, 13, 9, 9, 14, 11, 13, 13, 9, 14, 11, 11, 13, 9, 14]; fn gmul(mut a: u8, mut b: u8) -> u8 { let mut p = 0; for _ in 0..8 { if (b & 0x1) != 0 { p.bitxor_assign(a); } let has_high_bit = (a & 0x80) == 0x80; a <<= 1; if has_high_bit { a.bitxor_assign(0x1b); } b >>= 1; } p } fn mix_column(matrix: &[u8; 16], state_column: &[u8]) -> Vec<u8> { matrix .chunks(4) .map(|mc| { mc.iter() .enumerate() .map(|(i, &coefficient)| gmul(coefficient, state_column[i])) .fold(None, |accum, current| match accum { None => Some(current), Some(x) => Some(x.bitxor(current)), }) .unwrap() }) .collect() } #[test] fn test_mix_column() { use util::parse_byte_string; assert_eq!( parse_byte_string("8e4da1bc"), mix_column(&COLUMN_MATRIX, &parse_byte_string("db135345")), ); assert_eq!( parse_byte_string("9fdc589d"), mix_column(&COLUMN_MATRIX, &parse_byte_string("f20a225c")), ); assert_eq!( parse_byte_string("01010101"), mix_column(&COLUMN_MATRIX, &parse_byte_string("01010101")), ); assert_eq!( parse_byte_string("c6c6c6c6"), mix_column(&COLUMN_MATRIX, &parse_byte_string("c6c6c6c6")), ); assert_eq!( parse_byte_string("d5d5d7d6"), mix_column(&COLUMN_MATRIX, &parse_byte_string("d4d4d4d5")), ); assert_eq!( parse_byte_string("4d7ebdf8"), mix_column(&COLUMN_MATRIX, &parse_byte_string("2d26314c")), ); } fn mix_columns(state: &mut [u8]) { for column in state.chunks_mut(4) { let new_column = mix_column(&COLUMN_MATRIX, column); column.copy_from_slice(&new_column); } } fn inv_mix_columns(state: &mut [u8]) { for column in state.chunks_mut(4) { let new_column = mix_column(&INV_COLUMN_MATRIX, column); column.copy_from_slice(&new_column); } } #[derive(Clone)] pub enum CipherMode { ECB, CBC([u8; 16]), } fn transform_chunk(chunk: &[u8], expanded_key: &[u8], operation: Operation) -> [u8; 16] { const STATE_SIZE: usize = 16; assert!( chunk.len() == STATE_SIZE, "Chunk size of {} is invalid; expected {}", chunk.len(), STATE_SIZE ); let last_round = expanded_key.chunks(STATE_SIZE).count() - 1; let mut state = util::convert_to_fixed_array(chunk); match operation { Operation::Encrypt => { for (round, round_key) in expanded_key.chunks(STATE_SIZE).enumerate() { match round { 0 => { add_round_key(&mut state, round_key); } n if n != last_round => { sbox(&mut state); shift_rows(&mut state); mix_columns(&mut state); add_round_key(&mut state, round_key); } _ => { sbox(&mut state); shift_rows(&mut state); add_round_key(&mut state, round_key); } } } } Operation::Decrypt => { for (round, round_key) in expanded_key.chunks(STATE_SIZE).rev().enumerate() { match round { 0 => { add_round_key(&mut state, round_key); } n if n != last_round => { inv_shift_rows(&mut state); inv_sbox(&mut state); add_round_key(&mut state, round_key); inv_mix_columns(&mut state); } _ => { inv_shift_rows(&mut state); inv_sbox(&mut state); add_round_key(&mut state, round_key); } } } } }; state } #[derive(Clone, Copy)] pub enum Operation { Encrypt, Decrypt, } trait CipherModeImpl { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8>; } struct ECBCipherMode {} struct CBCCipherMode { initialization_vector: [u8; 16], operation: Operation, } impl CBCCipherMode { fn transform(&mut self, chunk: &[u8; 16], transform: &Fn(&[u8; 16]) -> [u8; 16]) -> [u8; 16] { match self.operation { Operation::Encrypt => { xor::buffer_mut(&mut self.initialization_vector, xor::Key::FullBuffer(chunk)); self.initialization_vector = transform(&self.initialization_vector); util::convert_to_fixed_array(&self.initialization_vector) } Operation::Decrypt => { let mut plaintext = transform(chunk); xor::buffer_mut( &mut plaintext, xor::Key::FullBuffer(&self.initialization_vector), ); self.initialization_vector = util::convert_to_fixed_array(chunk); plaintext } } } } impl CipherModeImpl for ECBCipherMode { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8> { let mut v = Vec::with_capacity(data.len()); for chunk in data.chunks(16) { v.extend(&transform(&util::convert_to_fixed_array(chunk))); } v } } impl CipherModeImpl for CBCCipherMode { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8> { let mut v = Vec::with_capacity(data.len()); for chunk in data.chunks(16) { v.extend( self .transform(&util::convert_to_fixed_array(chunk), transform) .iter(), ); } v } } pub fn perform(data: &[u8], key: &[u8], operation: Operation, cipher_mode: CipherMode) -> Vec<u8> { let expanded_key = expand_key(key); let mut cipher_mode_impl: Box<CipherModeImpl> = match cipher_mode { CipherMode::ECB => Box::new(ECBCipherMode {}), CipherMode::CBC(iv) => Box::new(CBCCipherMode { initialization_vector: iv, operation, }), }; cipher_mode_impl.transform_chunks(data, &|pre_transformed_chunk| { transform_chunk(pre_transformed_chunk, &expanded_key, operation) }) } #[test] fn ecb_once_16() { let plaintext = &vec![0; 16]; let key = &vec![0; 16]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); } #[test] fn ecb_once_24() { let plaintext = &vec![0; 32]; let key = &vec![0; 24]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); } #[test] fn ecb_once_32() { let plaintext = &vec![0; 32]; let key = &vec![0; 32]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); }
expand_key
identifier_name
aes.rs
use std::ops::BitXor; use std::ops::BitXorAssign; use util; use xor; const RCON: [u8; 11] = [ 0x8d, 0x01, 0x02, 0x04, 0x08, 0x10, 0x20, 0x40, 0x80, 0x1b, 0x36, ]; const SBOX: [u8; 256] = [ 0x63, 0x7C, 0x77, 0x7B, 0xF2, 0x6B, 0x6F, 0xC5, 0x30, 0x01, 0x67, 0x2B, 0xFE, 0xD7, 0xAB, 0x76, 0xCA, 0x82, 0xC9, 0x7D, 0xFA, 0x59, 0x47, 0xF0, 0xAD, 0xD4, 0xA2, 0xAF, 0x9C, 0xA4, 0x72, 0xC0, 0xB7, 0xFD, 0x93, 0x26, 0x36, 0x3F, 0xF7, 0xCC, 0x34, 0xA5, 0xE5, 0xF1, 0x71, 0xD8, 0x31, 0x15, 0x04, 0xC7, 0x23, 0xC3, 0x18, 0x96, 0x05, 0x9A, 0x07, 0x12, 0x80, 0xE2, 0xEB, 0x27, 0xB2, 0x75, 0x09, 0x83, 0x2C, 0x1A, 0x1B, 0x6E, 0x5A, 0xA0, 0x52, 0x3B, 0xD6, 0xB3, 0x29, 0xE3, 0x2F, 0x84, 0x53, 0xD1, 0x00, 0xED, 0x20, 0xFC, 0xB1, 0x5B, 0x6A, 0xCB, 0xBE, 0x39, 0x4A, 0x4C, 0x58, 0xCF, 0xD0, 0xEF, 0xAA, 0xFB, 0x43, 0x4D, 0x33, 0x85, 0x45, 0xF9, 0x02, 0x7F, 0x50, 0x3C, 0x9F, 0xA8, 0x51, 0xA3, 0x40, 0x8F, 0x92, 0x9D, 0x38, 0xF5, 0xBC, 0xB6, 0xDA, 0x21, 0x10, 0xFF, 0xF3, 0xD2, 0xCD, 0x0C, 0x13, 0xEC, 0x5F, 0x97, 0x44, 0x17, 0xC4, 0xA7, 0x7E, 0x3D, 0x64, 0x5D, 0x19, 0x73, 0x60, 0x81, 0x4F, 0xDC, 0x22, 0x2A, 0x90, 0x88, 0x46, 0xEE, 0xB8, 0x14, 0xDE, 0x5E, 0x0B, 0xDB, 0xE0, 0x32, 0x3A, 0x0A, 0x49, 0x06, 0x24, 0x5C, 0xC2, 0xD3, 0xAC, 0x62, 0x91, 0x95, 0xE4, 0x79, 0xE7, 0xC8, 0x37, 0x6D, 0x8D, 0xD5, 0x4E, 0xA9, 0x6C, 0x56, 0xF4, 0xEA, 0x65, 0x7A, 0xAE, 0x08, 0xBA, 0x78, 0x25, 0x2E, 0x1C, 0xA6, 0xB4, 0xC6, 0xE8, 0xDD, 0x74, 0x1F, 0x4B, 0xBD, 0x8B, 0x8A, 0x70, 0x3E, 0xB5, 0x66, 0x48, 0x03, 0xF6, 0x0E, 0x61, 0x35, 0x57, 0xB9, 0x86, 0xC1, 0x1D, 0x9E, 0xE1, 0xF8, 0x98, 0x11, 0x69, 0xD9, 0x8E, 0x94, 0x9B, 0x1E, 0x87, 0xE9, 0xCE, 0x55, 0x28, 0xDF, 0x8C, 0xA1, 0x89, 0x0D, 0xBF, 0xE6, 0x42, 0x68, 0x41, 0x99, 0x2D, 0x0F, 0xB0, 0x54, 0xBB, 0x16, ]; const INV_SBOX: [u8; 256] = [ 0x52, 0x09, 0x6A, 0xD5, 0x30, 0x36, 0xA5, 0x38, 0xBF, 0x40, 0xA3, 0x9E, 0x81, 0xF3, 0xD7, 0xFB, 0x7C, 0xE3, 0x39, 0x82, 0x9B, 0x2F, 0xFF, 0x87, 0x34, 0x8E, 0x43, 0x44, 0xC4, 0xDE, 0xE9, 0xCB, 0x54, 0x7B, 0x94, 0x32, 0xA6, 0xC2, 0x23, 0x3D, 0xEE, 0x4C, 0x95, 0x0B, 0x42, 0xFA, 0xC3, 0x4E, 0x08, 0x2E, 0xA1, 0x66, 0x28, 0xD9, 0x24, 0xB2, 0x76, 0x5B, 0xA2, 0x49, 0x6D, 0x8B, 0xD1, 0x25, 0x72, 0xF8, 0xF6, 0x64, 0x86, 0x68, 0x98, 0x16, 0xD4, 0xA4, 0x5C, 0xCC, 0x5D, 0x65, 0xB6, 0x92, 0x6C, 0x70, 0x48, 0x50, 0xFD, 0xED, 0xB9, 0xDA, 0x5E, 0x15, 0x46, 0x57, 0xA7, 0x8D, 0x9D, 0x84, 0x90, 0xD8, 0xAB, 0x00, 0x8C, 0xBC, 0xD3, 0x0A, 0xF7, 0xE4, 0x58, 0x05, 0xB8, 0xB3, 0x45, 0x06, 0xD0, 0x2C, 0x1E, 0x8F, 0xCA, 0x3F, 0x0F, 0x02, 0xC1, 0xAF, 0xBD, 0x03, 0x01, 0x13, 0x8A, 0x6B, 0x3A, 0x91, 0x11, 0x41, 0x4F, 0x67, 0xDC, 0xEA, 0x97, 0xF2, 0xCF, 0xCE, 0xF0, 0xB4, 0xE6, 0x73, 0x96, 0xAC, 0x74, 0x22, 0xE7, 0xAD, 0x35, 0x85, 0xE2, 0xF9, 0x37, 0xE8, 0x1C, 0x75, 0xDF, 0x6E, 0x47, 0xF1, 0x1A, 0x71, 0x1D, 0x29, 0xC5, 0x89, 0x6F, 0xB7, 0x62, 0x0E, 0xAA, 0x18, 0xBE, 0x1B, 0xFC, 0x56, 0x3E, 0x4B, 0xC6, 0xD2, 0x79, 0x20, 0x9A, 0xDB, 0xC0, 0xFE, 0x78, 0xCD, 0x5A, 0xF4, 0x1F, 0xDD, 0xA8, 0x33, 0x88, 0x07, 0xC7, 0x31, 0xB1, 0x12, 0x10, 0x59, 0x27, 0x80, 0xEC, 0x5F, 0x60, 0x51, 0x7F, 0xA9, 0x19, 0xB5, 0x4A, 0x0D, 0x2D, 0xE5, 0x7A, 0x9F, 0x93, 0xC9, 0x9C, 0xEF, 0xA0, 0xE0, 0x3B, 0x4D, 0xAE, 0x2A, 0xF5, 0xB0, 0xC8, 0xEB, 0xBB, 0x3C, 0x83, 0x53, 0x99, 0x61, 0x17, 0x2B, 0x04, 0x7E, 0xBA, 0x77, 0xD6, 0x26, 0xE1, 0x69, 0x14, 0x63, 0x55, 0x21, 0x0C, 0x7D, ]; fn sbox(data: &mut [u8]) { for e in data.iter_mut() { *e = SBOX[*e as usize]; } } fn inv_sbox(data: &mut [u8]) { for e in data.iter_mut() { *e = INV_SBOX[*e as usize]; } } fn to_four_byte_array(data: &[u8]) -> [u8; 4] { [data[0], data[1], data[2], data[3]] } enum KeyExpansionMode { Xor, Sbox, Full, } fn generate_four_bytes( key_length: usize, expanded_key: &[u8], rcon_iteration: &mut usize, mode: KeyExpansionMode, ) -> [u8; 4] { let i = expanded_key.len(); let source_bytes = &expanded_key[i - 4..i]; let mut t: [u8; 4] = to_four_byte_array(source_bytes); match mode { KeyExpansionMode::Xor => {} KeyExpansionMode::Sbox => { sbox(&mut t); } KeyExpansionMode::Full => { t.rotate_left(1); sbox(&mut t); t[0].bitxor_assign(RCON[*rcon_iteration]); *rcon_iteration += 1; } }; let xor_source = &expanded_key[i - key_length..i - key_length + 4]; xor::buffer_mut(&mut t, xor::Key::FullBuffer(xor_source)); t } fn expand_key(key: &[u8]) -> Vec<u8> { let key_length = key.len(); let (rounds, sbox_round, extra_expansions) = match key_length { 16 => (10, false, 0), 24 => (12, false, 2), 32 => (14, true, 3), len => panic!("Unsupported key length {}", len), }; let expanded_key_size = 16 * (rounds + 1); let mut expanded_key = Vec::with_capacity(expanded_key_size); expanded_key.extend_from_slice(&key); let mut rcon_iteration = 1usize; while expanded_key.len() < expanded_key_size { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Full, ); expanded_key.extend(t.iter()); for _i in 0..3 { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Xor, ); expanded_key.extend(t.iter()); } if sbox_round { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Sbox, ); expanded_key.extend(t.iter()); } for _i in 0..extra_expansions { let t = generate_four_bytes( key_length, &expanded_key, &mut rcon_iteration, KeyExpansionMode::Xor, ); expanded_key.extend(t.iter()); } } // Truncate any extra bytes expanded_key.resize(expanded_key_size, 0); assert!( expanded_key.len() == expanded_key_size, "Expanded key is too long: {}", expanded_key.len(), ); expanded_key } #[test] fn expand_key_16() { use util::parse_byte_string; assert_eq!( parse_byte_string( "00000000000000000000000000000000626363636263636362636363626363639b9898c9f9fbfbaa9b9898c9f9fbfbaa90973450696ccffaf2f457330b0fac99ee06da7b876a1581759e42b27e91ee2b7f2e2b88f8443e098dda7cbbf34b9290ec614b851425758c99ff09376ab49ba7217517873550620bacaf6b3cc61bf09b0ef903333ba9613897060a04511dfa9fb1d4d8e28a7db9da1d7bb3de4c664941b4ef5bcb3e92e21123e951cf6f8f188e" ), expand_key(&vec![0; 16]) ); assert_eq!( parse_byte_string( "ffffffffffffffffffffffffffffffffe8e9e9e917161616e8e9e9e917161616adaeae19bab8b80f525151e6454747f0090e2277b3b69a78e1e7cb9ea4a08c6ee16abd3e52dc2746b33becd8179b60b6e5baf3ceb766d488045d385013c658e671d07db3c6b6a93bc2eb916bd12dc98de90d208d2fbb89b6ed5018dd3c7dd15096337366b988fad054d8e20d68a5335d8bf03f233278c5f366a027fe0e0514a3d60a3588e472f07b82d2d7858cd7c326" ), expand_key(&vec![0xff; 16]) ); assert_eq!( parse_byte_string( "000102030405060708090a0b0c0d0e0fd6aa74fdd2af72fadaa678f1d6ab76feb692cf0b643dbdf1be9bc5006830b3feb6ff744ed2c2c9bf6c590cbf0469bf4147f7f7bc95353e03f96c32bcfd058dfd3caaa3e8a99f9deb50f3af57adf622aa5e390f7df7a69296a7553dc10aa31f6b14f9701ae35fe28c440adf4d4ea9c02647438735a41c65b9e016baf4aebf7ad2549932d1f08557681093ed9cbe2c974e13111d7fe3944a17f307a78b4d2b30c5" ), expand_key(&parse_byte_string("000102030405060708090a0b0c0d0e0f")) ); assert_eq!( parse_byte_string( "6920e299a5202a6d656e636869746f2afa8807605fa82d0d3ac64e6553b2214fcf75838d90ddae80aa1be0e5f9a9c1aa180d2f1488d0819422cb6171db62a0dbbaed96ad323d173910f67648cb94d693881b4ab2ba265d8baad02bc36144fd50b34f195d096944d6a3b96f15c2fd9245a7007778ae6933ae0dd05cbbcf2dcefeff8bccf251e2ff5c5c32a3e7931f6d1924b7182e7555e77229674495ba78298cae127cdadb479ba8f220df3d4858f6b1" ), expand_key(&parse_byte_string("6920e299a5202a6d656e636869746f2a")) ); } #[test] fn expand_key_24() { assert_eq!(208, expand_key(&vec![0; 24]).len()); } #[test] fn expand_key_32() { assert_eq!(240, expand_key(&vec![0; 32]).len()); } fn add_round_key(state: &mut [u8], key: &[u8]) { xor::buffer_mut(state, xor::Key::FullBuffer(key)); } // Shifted by 0, 1, 2, 3 columns const ROW_SHIFTS: [usize; 16] = [0, 5, 10, 15, 4, 9, 14, 3, 8, 13, 2, 7, 12, 1, 6, 11]; fn shift_rows(state: &mut [u8]) { let copy = util::convert_to_fixed_array(state); for (index, e) in state.iter_mut().enumerate() { *e = copy[ROW_SHIFTS[index]]; } } #[test] fn test_shift_rows() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; shift_rows(&mut rows); assert_eq!( rows, [1, 6, 11, 16, 5, 10, 15, 4, 9, 14, 3, 8, 13, 2, 7, 12] ); } // Shifted by 0, -1, -2, -3 columns const INV_ROW_SHIFTS: [usize; 16] = [0, 13, 10, 7, 4, 1, 14, 11, 8, 5, 2, 15, 12, 9, 6, 3]; fn inv_shift_rows(state: &mut [u8]) { let copy = util::convert_to_fixed_array(state); for (index, e) in state.iter_mut().enumerate() { *e = copy[INV_ROW_SHIFTS[index]]; } } #[test] fn test_inv_shift_rows() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; inv_shift_rows(&mut rows); assert_ne!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); assert_eq!( rows, [1, 14, 11, 8, 5, 2, 15, 12, 9, 6, 3, 16, 13, 10, 7, 4,] ); } #[test] fn test_shift_rows_ident() { let mut rows = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]; shift_rows(&mut rows); assert_ne!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); inv_shift_rows(&mut rows); assert_eq!( rows, [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16] ); } const COLUMN_MATRIX: [u8; 16] = [2, 3, 1, 1, 1, 2, 3, 1, 1, 1, 2, 3, 3, 1, 1, 2]; const INV_COLUMN_MATRIX: [u8; 16] = [14, 11, 13, 9, 9, 14, 11, 13, 13, 9, 14, 11, 11, 13, 9, 14]; fn gmul(mut a: u8, mut b: u8) -> u8 { let mut p = 0; for _ in 0..8 { if (b & 0x1) != 0 { p.bitxor_assign(a); } let has_high_bit = (a & 0x80) == 0x80; a <<= 1; if has_high_bit { a.bitxor_assign(0x1b); } b >>= 1; } p } fn mix_column(matrix: &[u8; 16], state_column: &[u8]) -> Vec<u8> { matrix .chunks(4) .map(|mc| { mc.iter() .enumerate() .map(|(i, &coefficient)| gmul(coefficient, state_column[i])) .fold(None, |accum, current| match accum { None => Some(current), Some(x) => Some(x.bitxor(current)), }) .unwrap() }) .collect() } #[test] fn test_mix_column() { use util::parse_byte_string; assert_eq!( parse_byte_string("8e4da1bc"), mix_column(&COLUMN_MATRIX, &parse_byte_string("db135345")), ); assert_eq!( parse_byte_string("9fdc589d"), mix_column(&COLUMN_MATRIX, &parse_byte_string("f20a225c")), ); assert_eq!( parse_byte_string("01010101"), mix_column(&COLUMN_MATRIX, &parse_byte_string("01010101")), ); assert_eq!( parse_byte_string("c6c6c6c6"), mix_column(&COLUMN_MATRIX, &parse_byte_string("c6c6c6c6")), ); assert_eq!( parse_byte_string("d5d5d7d6"), mix_column(&COLUMN_MATRIX, &parse_byte_string("d4d4d4d5")), ); assert_eq!( parse_byte_string("4d7ebdf8"), mix_column(&COLUMN_MATRIX, &parse_byte_string("2d26314c")), ); } fn mix_columns(state: &mut [u8]) { for column in state.chunks_mut(4) { let new_column = mix_column(&COLUMN_MATRIX, column); column.copy_from_slice(&new_column); } } fn inv_mix_columns(state: &mut [u8]) { for column in state.chunks_mut(4) { let new_column = mix_column(&INV_COLUMN_MATRIX, column); column.copy_from_slice(&new_column); } } #[derive(Clone)] pub enum CipherMode { ECB, CBC([u8; 16]), } fn transform_chunk(chunk: &[u8], expanded_key: &[u8], operation: Operation) -> [u8; 16] { const STATE_SIZE: usize = 16; assert!( chunk.len() == STATE_SIZE, "Chunk size of {} is invalid; expected {}", chunk.len(), STATE_SIZE ); let last_round = expanded_key.chunks(STATE_SIZE).count() - 1; let mut state = util::convert_to_fixed_array(chunk); match operation { Operation::Encrypt => { for (round, round_key) in expanded_key.chunks(STATE_SIZE).enumerate() { match round { 0 => { add_round_key(&mut state, round_key); } n if n != last_round => { sbox(&mut state); shift_rows(&mut state); mix_columns(&mut state); add_round_key(&mut state, round_key); } _ => { sbox(&mut state); shift_rows(&mut state); add_round_key(&mut state, round_key); } } } } Operation::Decrypt => { for (round, round_key) in expanded_key.chunks(STATE_SIZE).rev().enumerate() { match round { 0 => { add_round_key(&mut state, round_key); } n if n != last_round => { inv_shift_rows(&mut state); inv_sbox(&mut state); add_round_key(&mut state, round_key); inv_mix_columns(&mut state); } _ => { inv_shift_rows(&mut state); inv_sbox(&mut state); add_round_key(&mut state, round_key); } } } } }; state } #[derive(Clone, Copy)] pub enum Operation { Encrypt, Decrypt, } trait CipherModeImpl { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8>; } struct ECBCipherMode {} struct CBCCipherMode { initialization_vector: [u8; 16], operation: Operation,
impl CBCCipherMode { fn transform(&mut self, chunk: &[u8; 16], transform: &Fn(&[u8; 16]) -> [u8; 16]) -> [u8; 16] { match self.operation { Operation::Encrypt => { xor::buffer_mut(&mut self.initialization_vector, xor::Key::FullBuffer(chunk)); self.initialization_vector = transform(&self.initialization_vector); util::convert_to_fixed_array(&self.initialization_vector) } Operation::Decrypt => { let mut plaintext = transform(chunk); xor::buffer_mut( &mut plaintext, xor::Key::FullBuffer(&self.initialization_vector), ); self.initialization_vector = util::convert_to_fixed_array(chunk); plaintext } } } } impl CipherModeImpl for ECBCipherMode { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8> { let mut v = Vec::with_capacity(data.len()); for chunk in data.chunks(16) { v.extend(&transform(&util::convert_to_fixed_array(chunk))); } v } } impl CipherModeImpl for CBCCipherMode { fn transform_chunks( &mut self, data: &[u8], transform: &(Fn(&[u8; 16]) -> [u8; 16] + Sync), ) -> Vec<u8> { let mut v = Vec::with_capacity(data.len()); for chunk in data.chunks(16) { v.extend( self .transform(&util::convert_to_fixed_array(chunk), transform) .iter(), ); } v } } pub fn perform(data: &[u8], key: &[u8], operation: Operation, cipher_mode: CipherMode) -> Vec<u8> { let expanded_key = expand_key(key); let mut cipher_mode_impl: Box<CipherModeImpl> = match cipher_mode { CipherMode::ECB => Box::new(ECBCipherMode {}), CipherMode::CBC(iv) => Box::new(CBCCipherMode { initialization_vector: iv, operation, }), }; cipher_mode_impl.transform_chunks(data, &|pre_transformed_chunk| { transform_chunk(pre_transformed_chunk, &expanded_key, operation) }) } #[test] fn ecb_once_16() { let plaintext = &vec![0; 16]; let key = &vec![0; 16]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); } #[test] fn ecb_once_24() { let plaintext = &vec![0; 32]; let key = &vec![0; 24]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); } #[test] fn ecb_once_32() { let plaintext = &vec![0; 32]; let key = &vec![0; 32]; let ciphertext = perform(&plaintext, &key, Operation::Encrypt, CipherMode::ECB); assert_eq!( plaintext, &perform(&ciphertext, &key, Operation::Decrypt, CipherMode::ECB) ); }
}
random_line_split
tcp.rs
//! TcpStream wrappers that supports connecting with options #[cfg(unix)] use std::os::unix::io::{AsRawFd, FromRawFd, IntoRawFd, RawFd}; #[cfg(windows)] use std::os::windows::io::{AsRawSocket, FromRawSocket, IntoRawSocket, RawSocket}; use std::{ io::{self, ErrorKind}, net::SocketAddr, ops::{Deref, DerefMut}, pin::Pin, task::{self, Poll}, }; use futures::{future, ready}; use log::{debug, warn}; use pin_project::pin_project; use socket2::{Socket, TcpKeepalive}; use tokio::{ io::{AsyncRead, AsyncWrite, ReadBuf}, net::{TcpListener as TokioTcpListener, TcpSocket, TcpStream as TokioTcpStream}, }; use crate::{context::Context, relay::socks5::Address, ServerAddr}; use super::{ sys::{set_tcp_fastopen, TcpStream as SysTcpStream}, AcceptOpts, ConnectOpts, }; /// TcpStream for outbound connections #[pin_project] pub struct TcpStream(#[pin] SysTcpStream); impl TcpStream { /// Connects to address pub async fn connect_with_opts(addr: &SocketAddr, opts: &ConnectOpts) -> io::Result<TcpStream> { // tcp_stream_connect(addr, opts).await.map(TcpStream) SysTcpStream::connect(*addr, opts).await.map(TcpStream) } /// Connects shadowsocks server pub async fn connect_server_with_opts( context: &Context, addr: &ServerAddr, opts: &ConnectOpts, ) -> io::Result<TcpStream> { let stream = match *addr { ServerAddr::SocketAddr(ref addr) => SysTcpStream::connect(*addr, opts).await?, ServerAddr::DomainName(ref domain, port) => { lookup_then!(&context, &domain, port, |addr| { SysTcpStream::connect(addr, opts).await })? .1 } }; Ok(TcpStream(stream)) } /// Connects proxy remote target pub async fn connect_remote_with_opts( context: &Context, addr: &Address, opts: &ConnectOpts, ) -> io::Result<TcpStream> { let stream = match *addr { Address::SocketAddress(ref addr) => SysTcpStream::connect(*addr, opts).await?, Address::DomainNameAddress(ref domain, port) => { lookup_then!(&context, &domain, port, |addr| { SysTcpStream::connect(addr, opts).await })? .1 } }; Ok(TcpStream(stream)) } } impl Deref for TcpStream { type Target = TokioTcpStream; fn deref(&self) -> &Self::Target { &self.0 } } impl DerefMut for TcpStream { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } } impl AsyncRead for TcpStream { fn poll_read(self: Pin<&mut Self>, cx: &mut task::Context<'_>, buf: &mut ReadBuf<'_>) -> Poll<io::Result<()>> { self.project().0.poll_read(cx, buf) } } impl AsyncWrite for TcpStream { fn poll_write(self: Pin<&mut Self>, cx: &mut task::Context<'_>, buf: &[u8]) -> Poll<io::Result<usize>> { self.project().0.poll_write(cx, buf) } fn poll_flush(self: Pin<&mut Self>, cx: &mut task::Context<'_>) -> Poll<io::Result<()>> { self.project().0.poll_flush(cx) } fn poll_shutdown(self: Pin<&mut Self>, cx: &mut task::Context<'_>) -> Poll<io::Result<()>> { self.project().0.poll_shutdown(cx) } } /// `TcpListener` for accepting inbound connections pub struct TcpListener { inner: TokioTcpListener, accept_opts: AcceptOpts, } impl TcpListener { /// Creates a new TcpListener, which will be bound to the specified address. pub async fn bind_with_opts(addr: &SocketAddr, accept_opts: AcceptOpts) -> io::Result<TcpListener> { let socket = match *addr { SocketAddr::V4(..) => TcpSocket::new_v4()?, SocketAddr::V6(..) => TcpSocket::new_v6()?, }; // On platforms with Berkeley-derived sockets, this allows to quickly // rebind a socket, without needing to wait for the OS to clean up the // previous one. // // On Windows, this allows rebinding sockets which are actively in use, // which allows “socket hijacking”, so we explicitly don't set it here. // https://docs.microsoft.com/en-us/windows/win32/winsock/using-so-reuseaddr-and-so-exclusiveaddruse #[cfg(not(windows))] socket.set_reuseaddr(true)?; let set_dual_stack = if let SocketAddr::V6(ref v6) = *addr { v6.ip().is_unspecified() } else { false }; if set_dual_stack { // Set to DUAL STACK mode by default. // WARNING: This would fail if you want to start another program listening on the same port. // // Should this behavior be configurable? fn set_only_v6(socket: &TcpSocket, only_v6: bool) { unsafe { // WARN: If the following code panics, FD will be closed twice. #[cfg(unix)] let s = Socket::from_raw_fd(socket.as_raw_fd()); #[cfg(windows)] let s = Socket::from_raw_socket(socket.as_raw_socket()); if let Err(err) = s.set_only_v6(only_v6) { warn!("failed to set IPV6_V6ONLY: {} for listener, error: {}", only_v6, err); // This is not a fatal error, just warn and skip } #[cfg(unix)] let _ = s.into_raw_fd(); #[cfg(windows)] let _ = s.into_raw_socket(); } } set_only_v6(&socket, false); match socket.bind(*addr) { Ok(..) => {} Err(ref err) if err.kind() == ErrorKind::AddrInUse => {
debug!( "0.0.0.0:{} may have already been occupied, retry with IPV6_V6ONLY", addr.port() ); set_only_v6(&socket, true); socket.bind(*addr)?; } Err(err) => return Err(err), } } else { socket.bind(*addr)?; } // mio's default backlog is 1024 let inner = socket.listen(1024)?; // Enable TFO if supported // macos requires TCP_FASTOPEN to be set after listen(), but other platform doesn't have this constraint if accept_opts.tcp.fastopen { set_tcp_fastopen(&inner)?; } Ok(TcpListener { inner, accept_opts }) } /// Create a `TcpListener` from tokio's `TcpListener` pub fn from_listener(listener: TokioTcpListener, accept_opts: AcceptOpts) -> TcpListener { TcpListener { inner: listener, accept_opts, } } /// Polls to accept a new incoming connection to this listener. pub fn poll_accept(&self, cx: &mut task::Context<'_>) -> Poll<io::Result<(TokioTcpStream, SocketAddr)>> { let (stream, peer_addr) = ready!(self.inner.poll_accept(cx))?; setsockopt_with_opt(&stream, &self.accept_opts)?; Poll::Ready(Ok((stream, peer_addr))) } /// Accept a new incoming connection to this listener pub async fn accept(&self) -> io::Result<(TokioTcpStream, SocketAddr)> { future::poll_fn(|cx| self.poll_accept(cx)).await } /// Unwraps and take the internal `TcpListener` pub fn into_inner(self) -> TokioTcpListener { self.inner } } impl Deref for TcpListener { type Target = TokioTcpListener; fn deref(&self) -> &Self::Target { &self.inner } } impl DerefMut for TcpListener { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.inner } } impl From<TcpListener> for TokioTcpListener { fn from(listener: TcpListener) -> TokioTcpListener { listener.inner } } #[cfg(unix)] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { let socket = unsafe { Socket::from_raw_fd(f.as_raw_fd()) }; macro_rules! try_sockopt { ($socket:ident . $func:ident ($($arg:expr),*)) => { match $socket . $func ($($arg),*) { Ok(e) => e, Err(err) => { let _ = socket.into_raw_fd(); return Err(err); } } }; } if let Some(buf_size) = opts.tcp.send_buffer_size { try_sockopt!(socket.set_send_buffer_size(buf_size as usize)); } if let Some(buf_size) = opts.tcp.recv_buffer_size { try_sockopt!(socket.set_recv_buffer_size(buf_size as usize)); } try_sockopt!(socket.set_nodelay(opts.tcp.nodelay)); if let Some(keepalive_duration) = opts.tcp.keepalive { #[allow(unused_mut)] let mut keepalive = TcpKeepalive::new().with_time(keepalive_duration); #[cfg(any( target_os = "freebsd", target_os = "fuchsia", target_os = "linux", target_os = "netbsd", target_vendor = "apple", ))] { keepalive = keepalive.with_interval(keepalive_duration); } try_sockopt!(socket.set_tcp_keepalive(&keepalive)); } let _ = socket.into_raw_fd(); Ok(()) } #[cfg(windows)] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { let socket = unsafe { Socket::from_raw_socket(f.as_raw_socket()) }; macro_rules! try_sockopt { ($socket:ident . $func:ident ($($arg:expr),*)) => { match $socket . $func ($($arg),*) { Ok(e) => e, Err(err) => { let _ = socket.into_raw_socket(); return Err(err); } } }; } if let Some(buf_size) = opts.tcp.send_buffer_size { try_sockopt!(socket.set_send_buffer_size(buf_size as usize)); } if let Some(buf_size) = opts.tcp.recv_buffer_size { try_sockopt!(socket.set_recv_buffer_size(buf_size as usize)); } try_sockopt!(socket.set_nodelay(opts.tcp.nodelay)); if let Some(keepalive_duration) = opts.tcp.keepalive { let keepalive = TcpKeepalive::new() .with_time(keepalive_duration) .with_interval(keepalive_duration); try_sockopt!(socket.set_tcp_keepalive(&keepalive)); } let _ = socket.into_raw_socket(); Ok(()) } #[cfg(all(not(windows), not(unix)))] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { f.set_nodelay(opts.tcp.nodelay)?; Ok(()) } #[cfg(unix)] impl AsRawFd for TcpStream { fn as_raw_fd(&self) -> RawFd { self.0.as_raw_fd() } } #[cfg(windows)] impl AsRawSocket for TcpStream { fn as_raw_socket(&self) -> RawSocket { self.0.as_raw_socket() } }
// This is probably 0.0.0.0 with the same port has already been occupied
random_line_split
tcp.rs
//! TcpStream wrappers that supports connecting with options #[cfg(unix)] use std::os::unix::io::{AsRawFd, FromRawFd, IntoRawFd, RawFd}; #[cfg(windows)] use std::os::windows::io::{AsRawSocket, FromRawSocket, IntoRawSocket, RawSocket}; use std::{ io::{self, ErrorKind}, net::SocketAddr, ops::{Deref, DerefMut}, pin::Pin, task::{self, Poll}, }; use futures::{future, ready}; use log::{debug, warn}; use pin_project::pin_project; use socket2::{Socket, TcpKeepalive}; use tokio::{ io::{AsyncRead, AsyncWrite, ReadBuf}, net::{TcpListener as TokioTcpListener, TcpSocket, TcpStream as TokioTcpStream}, }; use crate::{context::Context, relay::socks5::Address, ServerAddr}; use super::{ sys::{set_tcp_fastopen, TcpStream as SysTcpStream}, AcceptOpts, ConnectOpts, }; /// TcpStream for outbound connections #[pin_project] pub struct TcpStream(#[pin] SysTcpStream); impl TcpStream { /// Connects to address pub async fn connect_with_opts(addr: &SocketAddr, opts: &ConnectOpts) -> io::Result<TcpStream> { // tcp_stream_connect(addr, opts).await.map(TcpStream) SysTcpStream::connect(*addr, opts).await.map(TcpStream) } /// Connects shadowsocks server pub async fn connect_server_with_opts( context: &Context, addr: &ServerAddr, opts: &ConnectOpts, ) -> io::Result<TcpStream> { let stream = match *addr { ServerAddr::SocketAddr(ref addr) => SysTcpStream::connect(*addr, opts).await?, ServerAddr::DomainName(ref domain, port) => { lookup_then!(&context, &domain, port, |addr| { SysTcpStream::connect(addr, opts).await })? .1 } }; Ok(TcpStream(stream)) } /// Connects proxy remote target pub async fn connect_remote_with_opts( context: &Context, addr: &Address, opts: &ConnectOpts, ) -> io::Result<TcpStream> { let stream = match *addr { Address::SocketAddress(ref addr) => SysTcpStream::connect(*addr, opts).await?, Address::DomainNameAddress(ref domain, port) => { lookup_then!(&context, &domain, port, |addr| { SysTcpStream::connect(addr, opts).await })? .1 } }; Ok(TcpStream(stream)) } } impl Deref for TcpStream { type Target = TokioTcpStream; fn deref(&self) -> &Self::Target { &self.0 } } impl DerefMut for TcpStream { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } } impl AsyncRead for TcpStream { fn poll_read(self: Pin<&mut Self>, cx: &mut task::Context<'_>, buf: &mut ReadBuf<'_>) -> Poll<io::Result<()>> { self.project().0.poll_read(cx, buf) } } impl AsyncWrite for TcpStream { fn poll_write(self: Pin<&mut Self>, cx: &mut task::Context<'_>, buf: &[u8]) -> Poll<io::Result<usize>> { self.project().0.poll_write(cx, buf) } fn
(self: Pin<&mut Self>, cx: &mut task::Context<'_>) -> Poll<io::Result<()>> { self.project().0.poll_flush(cx) } fn poll_shutdown(self: Pin<&mut Self>, cx: &mut task::Context<'_>) -> Poll<io::Result<()>> { self.project().0.poll_shutdown(cx) } } /// `TcpListener` for accepting inbound connections pub struct TcpListener { inner: TokioTcpListener, accept_opts: AcceptOpts, } impl TcpListener { /// Creates a new TcpListener, which will be bound to the specified address. pub async fn bind_with_opts(addr: &SocketAddr, accept_opts: AcceptOpts) -> io::Result<TcpListener> { let socket = match *addr { SocketAddr::V4(..) => TcpSocket::new_v4()?, SocketAddr::V6(..) => TcpSocket::new_v6()?, }; // On platforms with Berkeley-derived sockets, this allows to quickly // rebind a socket, without needing to wait for the OS to clean up the // previous one. // // On Windows, this allows rebinding sockets which are actively in use, // which allows “socket hijacking”, so we explicitly don't set it here. // https://docs.microsoft.com/en-us/windows/win32/winsock/using-so-reuseaddr-and-so-exclusiveaddruse #[cfg(not(windows))] socket.set_reuseaddr(true)?; let set_dual_stack = if let SocketAddr::V6(ref v6) = *addr { v6.ip().is_unspecified() } else { false }; if set_dual_stack { // Set to DUAL STACK mode by default. // WARNING: This would fail if you want to start another program listening on the same port. // // Should this behavior be configurable? fn set_only_v6(socket: &TcpSocket, only_v6: bool) { unsafe { // WARN: If the following code panics, FD will be closed twice. #[cfg(unix)] let s = Socket::from_raw_fd(socket.as_raw_fd()); #[cfg(windows)] let s = Socket::from_raw_socket(socket.as_raw_socket()); if let Err(err) = s.set_only_v6(only_v6) { warn!("failed to set IPV6_V6ONLY: {} for listener, error: {}", only_v6, err); // This is not a fatal error, just warn and skip } #[cfg(unix)] let _ = s.into_raw_fd(); #[cfg(windows)] let _ = s.into_raw_socket(); } } set_only_v6(&socket, false); match socket.bind(*addr) { Ok(..) => {} Err(ref err) if err.kind() == ErrorKind::AddrInUse => { // This is probably 0.0.0.0 with the same port has already been occupied debug!( "0.0.0.0:{} may have already been occupied, retry with IPV6_V6ONLY", addr.port() ); set_only_v6(&socket, true); socket.bind(*addr)?; } Err(err) => return Err(err), } } else { socket.bind(*addr)?; } // mio's default backlog is 1024 let inner = socket.listen(1024)?; // Enable TFO if supported // macos requires TCP_FASTOPEN to be set after listen(), but other platform doesn't have this constraint if accept_opts.tcp.fastopen { set_tcp_fastopen(&inner)?; } Ok(TcpListener { inner, accept_opts }) } /// Create a `TcpListener` from tokio's `TcpListener` pub fn from_listener(listener: TokioTcpListener, accept_opts: AcceptOpts) -> TcpListener { TcpListener { inner: listener, accept_opts, } } /// Polls to accept a new incoming connection to this listener. pub fn poll_accept(&self, cx: &mut task::Context<'_>) -> Poll<io::Result<(TokioTcpStream, SocketAddr)>> { let (stream, peer_addr) = ready!(self.inner.poll_accept(cx))?; setsockopt_with_opt(&stream, &self.accept_opts)?; Poll::Ready(Ok((stream, peer_addr))) } /// Accept a new incoming connection to this listener pub async fn accept(&self) -> io::Result<(TokioTcpStream, SocketAddr)> { future::poll_fn(|cx| self.poll_accept(cx)).await } /// Unwraps and take the internal `TcpListener` pub fn into_inner(self) -> TokioTcpListener { self.inner } } impl Deref for TcpListener { type Target = TokioTcpListener; fn deref(&self) -> &Self::Target { &self.inner } } impl DerefMut for TcpListener { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.inner } } impl From<TcpListener> for TokioTcpListener { fn from(listener: TcpListener) -> TokioTcpListener { listener.inner } } #[cfg(unix)] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { let socket = unsafe { Socket::from_raw_fd(f.as_raw_fd()) }; macro_rules! try_sockopt { ($socket:ident . $func:ident ($($arg:expr),*)) => { match $socket . $func ($($arg),*) { Ok(e) => e, Err(err) => { let _ = socket.into_raw_fd(); return Err(err); } } }; } if let Some(buf_size) = opts.tcp.send_buffer_size { try_sockopt!(socket.set_send_buffer_size(buf_size as usize)); } if let Some(buf_size) = opts.tcp.recv_buffer_size { try_sockopt!(socket.set_recv_buffer_size(buf_size as usize)); } try_sockopt!(socket.set_nodelay(opts.tcp.nodelay)); if let Some(keepalive_duration) = opts.tcp.keepalive { #[allow(unused_mut)] let mut keepalive = TcpKeepalive::new().with_time(keepalive_duration); #[cfg(any( target_os = "freebsd", target_os = "fuchsia", target_os = "linux", target_os = "netbsd", target_vendor = "apple", ))] { keepalive = keepalive.with_interval(keepalive_duration); } try_sockopt!(socket.set_tcp_keepalive(&keepalive)); } let _ = socket.into_raw_fd(); Ok(()) } #[cfg(windows)] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { let socket = unsafe { Socket::from_raw_socket(f.as_raw_socket()) }; macro_rules! try_sockopt { ($socket:ident . $func:ident ($($arg:expr),*)) => { match $socket . $func ($($arg),*) { Ok(e) => e, Err(err) => { let _ = socket.into_raw_socket(); return Err(err); } } }; } if let Some(buf_size) = opts.tcp.send_buffer_size { try_sockopt!(socket.set_send_buffer_size(buf_size as usize)); } if let Some(buf_size) = opts.tcp.recv_buffer_size { try_sockopt!(socket.set_recv_buffer_size(buf_size as usize)); } try_sockopt!(socket.set_nodelay(opts.tcp.nodelay)); if let Some(keepalive_duration) = opts.tcp.keepalive { let keepalive = TcpKeepalive::new() .with_time(keepalive_duration) .with_interval(keepalive_duration); try_sockopt!(socket.set_tcp_keepalive(&keepalive)); } let _ = socket.into_raw_socket(); Ok(()) } #[cfg(all(not(windows), not(unix)))] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { f.set_nodelay(opts.tcp.nodelay)?; Ok(()) } #[cfg(unix)] impl AsRawFd for TcpStream { fn as_raw_fd(&self) -> RawFd { self.0.as_raw_fd() } } #[cfg(windows)] impl AsRawSocket for TcpStream { fn as_raw_socket(&self) -> RawSocket { self.0.as_raw_socket() } }
poll_flush
identifier_name
tcp.rs
//! TcpStream wrappers that supports connecting with options #[cfg(unix)] use std::os::unix::io::{AsRawFd, FromRawFd, IntoRawFd, RawFd}; #[cfg(windows)] use std::os::windows::io::{AsRawSocket, FromRawSocket, IntoRawSocket, RawSocket}; use std::{ io::{self, ErrorKind}, net::SocketAddr, ops::{Deref, DerefMut}, pin::Pin, task::{self, Poll}, }; use futures::{future, ready}; use log::{debug, warn}; use pin_project::pin_project; use socket2::{Socket, TcpKeepalive}; use tokio::{ io::{AsyncRead, AsyncWrite, ReadBuf}, net::{TcpListener as TokioTcpListener, TcpSocket, TcpStream as TokioTcpStream}, }; use crate::{context::Context, relay::socks5::Address, ServerAddr}; use super::{ sys::{set_tcp_fastopen, TcpStream as SysTcpStream}, AcceptOpts, ConnectOpts, }; /// TcpStream for outbound connections #[pin_project] pub struct TcpStream(#[pin] SysTcpStream); impl TcpStream { /// Connects to address pub async fn connect_with_opts(addr: &SocketAddr, opts: &ConnectOpts) -> io::Result<TcpStream> { // tcp_stream_connect(addr, opts).await.map(TcpStream) SysTcpStream::connect(*addr, opts).await.map(TcpStream) } /// Connects shadowsocks server pub async fn connect_server_with_opts( context: &Context, addr: &ServerAddr, opts: &ConnectOpts, ) -> io::Result<TcpStream> { let stream = match *addr { ServerAddr::SocketAddr(ref addr) => SysTcpStream::connect(*addr, opts).await?, ServerAddr::DomainName(ref domain, port) => { lookup_then!(&context, &domain, port, |addr| { SysTcpStream::connect(addr, opts).await })? .1 } }; Ok(TcpStream(stream)) } /// Connects proxy remote target pub async fn connect_remote_with_opts( context: &Context, addr: &Address, opts: &ConnectOpts, ) -> io::Result<TcpStream> { let stream = match *addr { Address::SocketAddress(ref addr) => SysTcpStream::connect(*addr, opts).await?, Address::DomainNameAddress(ref domain, port) => { lookup_then!(&context, &domain, port, |addr| { SysTcpStream::connect(addr, opts).await })? .1 } }; Ok(TcpStream(stream)) } } impl Deref for TcpStream { type Target = TokioTcpStream; fn deref(&self) -> &Self::Target { &self.0 } } impl DerefMut for TcpStream { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } } impl AsyncRead for TcpStream { fn poll_read(self: Pin<&mut Self>, cx: &mut task::Context<'_>, buf: &mut ReadBuf<'_>) -> Poll<io::Result<()>> { self.project().0.poll_read(cx, buf) } } impl AsyncWrite for TcpStream { fn poll_write(self: Pin<&mut Self>, cx: &mut task::Context<'_>, buf: &[u8]) -> Poll<io::Result<usize>> { self.project().0.poll_write(cx, buf) } fn poll_flush(self: Pin<&mut Self>, cx: &mut task::Context<'_>) -> Poll<io::Result<()>> { self.project().0.poll_flush(cx) } fn poll_shutdown(self: Pin<&mut Self>, cx: &mut task::Context<'_>) -> Poll<io::Result<()>> { self.project().0.poll_shutdown(cx) } } /// `TcpListener` for accepting inbound connections pub struct TcpListener { inner: TokioTcpListener, accept_opts: AcceptOpts, } impl TcpListener { /// Creates a new TcpListener, which will be bound to the specified address. pub async fn bind_with_opts(addr: &SocketAddr, accept_opts: AcceptOpts) -> io::Result<TcpListener> { let socket = match *addr { SocketAddr::V4(..) => TcpSocket::new_v4()?, SocketAddr::V6(..) => TcpSocket::new_v6()?, }; // On platforms with Berkeley-derived sockets, this allows to quickly // rebind a socket, without needing to wait for the OS to clean up the // previous one. // // On Windows, this allows rebinding sockets which are actively in use, // which allows “socket hijacking”, so we explicitly don't set it here. // https://docs.microsoft.com/en-us/windows/win32/winsock/using-so-reuseaddr-and-so-exclusiveaddruse #[cfg(not(windows))] socket.set_reuseaddr(true)?; let set_dual_stack = if let SocketAddr::V6(ref v6) = *addr { v6.ip().is_unspecified() } else { false }; if set_dual_stack { // Set to DUAL STACK mode by default. // WARNING: This would fail if you want to start another program listening on the same port. // // Should this behavior be configurable? fn set_only_v6(socket: &TcpSocket, only_v6: bool) { unsafe { // WARN: If the following code panics, FD will be closed twice. #[cfg(unix)] let s = Socket::from_raw_fd(socket.as_raw_fd()); #[cfg(windows)] let s = Socket::from_raw_socket(socket.as_raw_socket()); if let Err(err) = s.set_only_v6(only_v6) { warn!("failed to set IPV6_V6ONLY: {} for listener, error: {}", only_v6, err); // This is not a fatal error, just warn and skip } #[cfg(unix)] let _ = s.into_raw_fd(); #[cfg(windows)] let _ = s.into_raw_socket(); } } set_only_v6(&socket, false); match socket.bind(*addr) { Ok(..) => {} Err(ref err) if err.kind() == ErrorKind::AddrInUse => { // This is probably 0.0.0.0 with the same port has already been occupied debug!( "0.0.0.0:{} may have already been occupied, retry with IPV6_V6ONLY", addr.port() ); set_only_v6(&socket, true); socket.bind(*addr)?; } Err(err) => return Err(err), } } else { socket.bind(*addr)?; } // mio's default backlog is 1024 let inner = socket.listen(1024)?; // Enable TFO if supported // macos requires TCP_FASTOPEN to be set after listen(), but other platform doesn't have this constraint if accept_opts.tcp.fastopen { set_tcp_fastopen(&inner)?; } Ok(TcpListener { inner, accept_opts }) } /// Create a `TcpListener` from tokio's `TcpListener` pub fn from_listener(listener: TokioTcpListener, accept_opts: AcceptOpts) -> TcpListener { TcpListener { inner: listener, accept_opts, } } /// Polls to accept a new incoming connection to this listener. pub fn poll_accept(&self, cx: &mut task::Context<'_>) -> Poll<io::Result<(TokioTcpStream, SocketAddr)>> { let (stream, peer_addr) = ready!(self.inner.poll_accept(cx))?; setsockopt_with_opt(&stream, &self.accept_opts)?; Poll::Ready(Ok((stream, peer_addr))) } /// Accept a new incoming connection to this listener pub async fn accept(&self) -> io::Result<(TokioTcpStream, SocketAddr)> { future::poll_fn(|cx| self.poll_accept(cx)).await } /// Unwraps and take the internal `TcpListener` pub fn into_inner(self) -> TokioTcpListener { self.inner } } impl Deref for TcpListener { type Target = TokioTcpListener; fn deref(&self) -> &Self::Target { &self.inner } } impl DerefMut for TcpListener { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.inner } } impl From<TcpListener> for TokioTcpListener { fn from(listener: TcpListener) -> TokioTcpListener { listener.inner } } #[cfg(unix)] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { let socket = unsafe { Socket::from_raw_fd(f.as_raw_fd()) }; macro_rules! try_sockopt { ($socket:ident . $func:ident ($($arg:expr),*)) => { match $socket . $func ($($arg),*) { Ok(e) => e, Err(err) => { let _ = socket.into_raw_fd(); return Err(err); } } }; } if let Some(buf_size) = opts.tcp.send_buffer_size { try_sockopt!(socket.set_send_buffer_size(buf_size as usize)); } if let Some(buf_size) = opts.tcp.recv_buffer_size { try_sockopt!(socket.set_recv_buffer_size(buf_size as usize)); } try_sockopt!(socket.set_nodelay(opts.tcp.nodelay)); if let Some(keepalive_duration) = opts.tcp.keepalive { #[allow(unused_mut)] let mut keepalive = TcpKeepalive::new().with_time(keepalive_duration); #[cfg(any( target_os = "freebsd", target_os = "fuchsia", target_os = "linux", target_os = "netbsd", target_vendor = "apple", ))] { keepalive = keepalive.with_interval(keepalive_duration); } try_sockopt!(socket.set_tcp_keepalive(&keepalive)); } let _ = socket.into_raw_fd(); Ok(()) } #[cfg(windows)] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { let socket = unsafe { Socket::from_raw_socket(f.as_raw_socket()) }; macro_rules! try_sockopt { ($socket:ident . $func:ident ($($arg:expr),*)) => { match $socket . $func ($($arg),*) { Ok(e) => e, Err(err) => { let _ = socket.into_raw_socket(); return Err(err); } } }; } if let Some(buf_size) = opts.tcp.send_buffer_size { try_sockopt!(socket.set_send_buffer_size(buf_size as usize)); } if let Some(buf_size) = opts.tcp.recv_buffer_size { try_sockopt!(socket.set_recv_buffer_size(buf_size as usize)); } try_sockopt!(socket.set_nodelay(opts.tcp.nodelay)); if let Some(keepalive_duration) = opts.tcp.keepalive { let keepalive = TcpKeepalive::new() .with_time(keepalive_duration) .with_interval(keepalive_duration); try_sockopt!(socket.set_tcp_keepalive(&keepalive)); } let _ = socket.into_raw_socket(); Ok(()) } #[cfg(all(not(windows), not(unix)))] fn setsockopt_with_opt(f: &tokio::net::TcpStream, opts: &AcceptOpts) -> io::Result<()> { f.set_nodelay(opts.tcp.nodelay)?; Ok(()) } #[cfg(unix)] impl AsRawFd for TcpStream { fn as_raw_fd(&self) -> RawFd {
#[cfg(windows)] impl AsRawSocket for TcpStream { fn as_raw_socket(&self) -> RawSocket { self.0.as_raw_socket() } }
self.0.as_raw_fd() } }
identifier_body
signature.go
/* Copyright IBM Corp. All Rights Reserved. SPDX-License-Identifier: Apache-2.0 */ package idemix import ( "github.com/hyperledger/fabric-amcl/amcl" "github.com/hyperledger/fabric-amcl/amcl/FP256BN" "github.com/pkg/errors" ) // signLabel is the label used in zero-knowledge proof (ZKP) to identify that this ZKP is a signature of knowledge const signLabel = "sign" // A signature that is produced using an Identity Mixer credential is a so-called signature of knowledge // (for details see C.P.Schnorr "Efficient Identification and Signatures for Smart Cards") // An Identity Mixer signature is a signature of knowledge that signs a message and proves (in zero-knowledge) // the knowledge of the user secret (and possibly attributes) signed inside a credential // that was issued by a certain issuer (referred to with the issuer public key) // The signature is verified using the message being signed and the public key of the issuer // Some of the attributes from the credential can be selectvely disclosed or different statements can be proven about // credential atrributes without diclosing them in the clear // The difference between a standard signature using X.509 certificates and an Identity Mixer signature is // the advanced privacy features provided by Identity Mixer (due to zero-knowledge proofs): // - Unlinkability of the signatures produced with the same credential // - Selective attribute disclosure and predicates over attributes // Make a slice of all the attribute indices that will not be disclosed func
(Disclosure []byte) []int { HiddenIndices := make([]int, 0) for index, disclose := range Disclosure { if disclose == 0 { HiddenIndices = append(HiddenIndices, index) } } return HiddenIndices } // NewSignature creates a new idemix signature (Schnorr-type signature) // The []byte Disclosure steers which attributes are disclosed: // if Disclosure[i] == 0 then attribute i remains hidden and otherwise it is disclosed. // We use the zero-knowledge proof by http://eprint.iacr.org/2016/663.pdf to prove knowledge of a BBS+ signature //We use the zero-knowledge proof by https://link.springer.com/content/pdf/10.1007%2F978-3-540-28628-8_3.pdf to prove the knowledge of user's secret key. func NewSignature(cred *Credential, sk *FP256BN.BIG, Nym *FP256BN.ECP, RNym *FP256BN.BIG, ipk *IssuerPublicKey, gpk *GroupPublicKey, uk *UserKey, Disclosure []byte, msg []byte, rng *amcl.RAND) (*Signature, error) { if cred == nil || sk == nil || Nym == nil || RNym == nil || ipk == nil || rng == nil { return nil, errors.Errorf("cannot create idemix signature: received nil input") } HiddenIndices := hiddenIndices(Disclosure) // Start sig r1 := RandModOrder(rng) r2 := RandModOrder(rng) r3 := FP256BN.NewBIGcopy(r1) r3.Invmodp(GroupOrder) Nonce := RandModOrder(rng) A := EcpFromProto(cred.A) B := EcpFromProto(cred.B) APrime := FP256BN.G1mul(A, r1) // A' = A^{r1} ABar := FP256BN.G1mul(B, r1) ABar.Sub(FP256BN.G1mul(APrime, FP256BN.FromBytes(cred.E))) // barA = A'^{-e} b^{r1} BPrime := FP256BN.G1mul(B, r1) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) BPrime.Sub(FP256BN.G1mul(HRand, r2)) // b' = b^{r1} h_r^{-r2} S := FP256BN.FromBytes(cred.S) E := FP256BN.FromBytes(cred.E) sPrime := Modsub(S, FP256BN.Modmul(r2, r3, GroupOrder), GroupOrder) // Construct ZK proof rSk := RandModOrder(rng) re := RandModOrder(rng) rR2 := RandModOrder(rng) rR3 := RandModOrder(rng) rSPrime := RandModOrder(rng) rRNym := RandModOrder(rng) rAttrs := make([]*FP256BN.BIG, len(HiddenIndices)) for i := range HiddenIndices { rAttrs[i] = RandModOrder(rng) } t1 := APrime.Mul2(re, HRand, rR2) t2 := FP256BN.G1mul(HRand, rSPrime) t2.Add(BPrime.Mul2(rR3, HSk, rSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(rAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), rAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), rAttrs[len(HiddenIndices)-1])) } t3 := HSk.Mul2(rSk, HRand, rRNym) //Zk proof of BBS group signature alpha := RandModOrder(rng) beta := RandModOrder(rng) rAlpha := RandModOrder(rng) rBeta := RandModOrder(rng) rx := RandModOrder(rng) rDelta1 := RandModOrder(rng) rDelta2 := RandModOrder(rng) temp := Modadd(alpha,beta,GroupOrder) temp1 := Modadd(rAlpha,rBeta,GroupOrder) temp2 := Modadd(rDelta1,rDelta2,GroupOrder) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) Ax := EcpFromProto(uk.UK2) W := Ecp2FromProto(ipk.W) T1 := FP256BN.G1mul(U,alpha) T2 := FP256BN.G1mul(V,beta) T3 := FP256BN.G1mul(H,temp) T3.Add(Ax) R1 := FP256BN.G1mul(U,rAlpha) R2 := FP256BN.G1mul(V,rBeta) //compute pairing Grx := GenG2.Mul(rx) T4 := H.Mul(temp) e1 := FP256BN.Ate2(Grx,Ax,Grx,T4) Htp1 := H.Mul(temp1) e2 := FP256BN.Ate(W,Htp1) e2.Inverse() e2.Mul(e1) Htp2 := H.Mul(temp2) e4 := FP256BN.Ate(GenG2,Htp2) e4.Inverse() e4.Mul(e2) R3 := FP256BN.Fexp(e4) R4 := FP256BN.G1mul(T1,rx) R4.Sub(U.Mul(rDelta1)) R5 := FP256BN.G1mul(T2,rx) R5.Sub(V.Mul(rDelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message being signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) // add the previous hash and the nonce and hash again to compute a second hash (C value) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) ProofC := HashModOrder(proofData) ProofSSk := Modadd(rSk, FP256BN.Modmul(ProofC, sk, GroupOrder), GroupOrder) ProofSE := Modsub(re, FP256BN.Modmul(ProofC, E, GroupOrder), GroupOrder) ProofSR2 := Modadd(rR2, FP256BN.Modmul(ProofC, r2, GroupOrder), GroupOrder) ProofSR3 := Modsub(rR3, FP256BN.Modmul(ProofC, r3, GroupOrder), GroupOrder) ProofSSPrime := Modadd(rSPrime, FP256BN.Modmul(ProofC, sPrime, GroupOrder), GroupOrder) ProofSRNym := Modadd(rRNym, FP256BN.Modmul(ProofC, RNym, GroupOrder), GroupOrder) //compute s-values for zk proof of group signature ProofSalpha := Modadd(rAlpha,FP256BN.Modmul(ProofC,alpha,GroupOrder),GroupOrder) ProofSbeta := Modadd(rBeta,FP256BN.Modmul(ProofC,beta,GroupOrder),GroupOrder) x := FP256BN.FromBytes(uk.UK1) ProofSx := Modadd(rx,FP256BN.Modmul(ProofC,x,GroupOrder),GroupOrder) delta1 := FP256BN.Modmul(alpha,x,GroupOrder) ProofSdelta1 := Modadd(rDelta1,FP256BN.Modmul(ProofC,delta1,GroupOrder),GroupOrder) delta2 := FP256BN.Modmul(beta,x,GroupOrder) ProofSdelta2 := Modadd(rDelta2,FP256BN.Modmul(ProofC,delta2,GroupOrder),GroupOrder) ProofSAttrs := make([][]byte, len(HiddenIndices)) for i, j := range HiddenIndices { ProofSAttrs[i] = BigToBytes(Modadd(rAttrs[i], FP256BN.Modmul(ProofC, FP256BN.FromBytes(cred.Attrs[j]), GroupOrder), GroupOrder)) } return &Signature{ EcpToProto(APrime), EcpToProto(ABar), EcpToProto(BPrime), EcpToProto(T1), EcpToProto(T2), EcpToProto(T3), BigToBytes(ProofC), BigToBytes(ProofSSk), BigToBytes(ProofSE), BigToBytes(ProofSR2), BigToBytes(ProofSR3), BigToBytes(ProofSSPrime), BigToBytes(ProofSalpha), BigToBytes(ProofSbeta), BigToBytes(ProofSx), BigToBytes(ProofSdelta1), BigToBytes(ProofSdelta2), ProofSAttrs, BigToBytes(Nonce), EcpToProto(Nym), BigToBytes(ProofSRNym)}, nil } // Ver verifies an idemix signature // Disclosure steers which attributes it expects to be disclosed // attributeValues[i] contains the desired attribute value for the i-th undisclosed attribute in Disclosure func (sig *Signature) Ver(Disclosure []byte, ipk *IssuerPublicKey, gpk *GroupPublicKey, msg []byte, attributeValues []*FP256BN.BIG) error { HiddenIndices := hiddenIndices(Disclosure) APrime := EcpFromProto(sig.GetAPrime()) ABar := EcpFromProto(sig.GetABar()) BPrime := EcpFromProto(sig.GetBPrime()) Nym := EcpFromProto(sig.GetNym()) ProofC := FP256BN.FromBytes(sig.GetProofC()) ProofSSk := FP256BN.FromBytes(sig.GetProofSSk()) ProofSE := FP256BN.FromBytes(sig.GetProofSE()) ProofSR2 := FP256BN.FromBytes(sig.GetProofSR2()) ProofSR3 := FP256BN.FromBytes(sig.GetProofSR3()) ProofSSPrime := FP256BN.FromBytes(sig.GetProofSSPrime()) ProofSRNym := FP256BN.FromBytes(sig.GetProofSRNym()) ProofSAttrs := make([]*FP256BN.BIG, len(sig.GetProofSAttrs())) T1 := EcpFromProto(sig.GetT1()) T2 := EcpFromProto(sig.GetT2()) T3 := EcpFromProto(sig.GetT3()) ProofSalpha := FP256BN.FromBytes(sig.GetProofSalpha()) ProofSbeta := FP256BN.FromBytes(sig.GetProofSbeta()) ProofSx := FP256BN.FromBytes(sig.GetProofSx()) ProofSdelta1 := FP256BN.FromBytes(sig.GetProofSdelta1()) ProofSdelta2 := FP256BN.FromBytes(sig.GetProofSdelta2()) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) s1 := Modadd(ProofSalpha,ProofSbeta,GroupOrder) s2 := Modadd(ProofSdelta1,ProofSdelta2,GroupOrder) if len(sig.ProofSAttrs) != len(HiddenIndices) { return errors.Errorf("signature invalid: incorrect amount of s-values for AttributeProofSpec") } for i, b := range sig.ProofSAttrs { ProofSAttrs[i] = FP256BN.FromBytes(b) } Nonce := FP256BN.FromBytes(sig.GetNonce()) W := Ecp2FromProto(ipk.W) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) if APrime.Is_infinity() { return errors.Errorf("signature invalid: APrime = 1") } temp1 := FP256BN.Ate(W, APrime) temp2 := FP256BN.Ate(GenG2, ABar) temp2.Inverse() temp1.Mul(temp2) if !FP256BN.Fexp(temp1).Isunity() { return errors.Errorf("signature invalid: APrime and ABar don't have the expected structure") } t1 := APrime.Mul2(ProofSE, HRand, ProofSR2) temp := FP256BN.NewECP() temp.Copy(ABar) temp.Sub(BPrime) t1.Sub(FP256BN.G1mul(temp, ProofC)) t2 := FP256BN.G1mul(HRand, ProofSSPrime) t2.Add(BPrime.Mul2(ProofSR3, HSk, ProofSSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(ProofSAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), ProofSAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), ProofSAttrs[len(HiddenIndices)-1])) } temp = FP256BN.NewECP() temp.Copy(GenG1) for index, disclose := range Disclosure { if disclose != 0 { temp.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[index]), attributeValues[index])) } } t2.Add(FP256BN.G1mul(temp, ProofC)) t3 := HSk.Mul2(ProofSSk, HRand, ProofSRNym) t3.Sub(Nym.Mul(ProofC)) //compute R1,R2,R3,R4,R5 R1 := FP256BN.G1mul(U,ProofSalpha) R1.Sub(T1.Mul(ProofC)) R2 := FP256BN.G1mul(V,ProofSbeta) R2.Sub(T2.Mul(ProofC)) Gsx := GenG2.Mul(ProofSx) e1 := FP256BN.Ate(Gsx,T3) Hs1 := H.Mul(s1) e2 := FP256BN.Ate(W,Hs1) e2.Inverse() e1.Mul(e2) Hs2 := H.Mul(s2) e4 := FP256BN.Ate(GenG2,Hs2) e4.Inverse() e1.Mul(e4) Wc := W.Mul(ProofC) e6 := FP256BN.Ate(Wc,T3) Gc := GenG2.Mul(ProofC) e7 := FP256BN.Ate(Gc,GenG1) e7.Inverse() e7.Mul(e6) e7.Mul(e1) R3 := FP256BN.Fexp(e7) R4 := FP256BN.G1mul(T1,ProofSx) R4.Sub(U.Mul(ProofSdelta1)) R5 := FP256BN.G1mul(T2,ProofSx) R5.Sub(V.Mul(ProofSdelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message that was signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) if *ProofC != *HashModOrder(proofData) { return errors.Errorf("signature invalid: zero-knowledge proof is invalid") } return nil }
hiddenIndices
identifier_name
signature.go
/* Copyright IBM Corp. All Rights Reserved. SPDX-License-Identifier: Apache-2.0 */ package idemix import ( "github.com/hyperledger/fabric-amcl/amcl" "github.com/hyperledger/fabric-amcl/amcl/FP256BN" "github.com/pkg/errors" ) // signLabel is the label used in zero-knowledge proof (ZKP) to identify that this ZKP is a signature of knowledge const signLabel = "sign" // A signature that is produced using an Identity Mixer credential is a so-called signature of knowledge // (for details see C.P.Schnorr "Efficient Identification and Signatures for Smart Cards") // An Identity Mixer signature is a signature of knowledge that signs a message and proves (in zero-knowledge) // the knowledge of the user secret (and possibly attributes) signed inside a credential // that was issued by a certain issuer (referred to with the issuer public key) // The signature is verified using the message being signed and the public key of the issuer // Some of the attributes from the credential can be selectvely disclosed or different statements can be proven about // credential atrributes without diclosing them in the clear // The difference between a standard signature using X.509 certificates and an Identity Mixer signature is // the advanced privacy features provided by Identity Mixer (due to zero-knowledge proofs): // - Unlinkability of the signatures produced with the same credential // - Selective attribute disclosure and predicates over attributes // Make a slice of all the attribute indices that will not be disclosed func hiddenIndices(Disclosure []byte) []int { HiddenIndices := make([]int, 0) for index, disclose := range Disclosure { if disclose == 0 { HiddenIndices = append(HiddenIndices, index) } } return HiddenIndices } // NewSignature creates a new idemix signature (Schnorr-type signature) // The []byte Disclosure steers which attributes are disclosed: // if Disclosure[i] == 0 then attribute i remains hidden and otherwise it is disclosed. // We use the zero-knowledge proof by http://eprint.iacr.org/2016/663.pdf to prove knowledge of a BBS+ signature //We use the zero-knowledge proof by https://link.springer.com/content/pdf/10.1007%2F978-3-540-28628-8_3.pdf to prove the knowledge of user's secret key. func NewSignature(cred *Credential, sk *FP256BN.BIG, Nym *FP256BN.ECP, RNym *FP256BN.BIG, ipk *IssuerPublicKey, gpk *GroupPublicKey, uk *UserKey, Disclosure []byte, msg []byte, rng *amcl.RAND) (*Signature, error) { if cred == nil || sk == nil || Nym == nil || RNym == nil || ipk == nil || rng == nil
HiddenIndices := hiddenIndices(Disclosure) // Start sig r1 := RandModOrder(rng) r2 := RandModOrder(rng) r3 := FP256BN.NewBIGcopy(r1) r3.Invmodp(GroupOrder) Nonce := RandModOrder(rng) A := EcpFromProto(cred.A) B := EcpFromProto(cred.B) APrime := FP256BN.G1mul(A, r1) // A' = A^{r1} ABar := FP256BN.G1mul(B, r1) ABar.Sub(FP256BN.G1mul(APrime, FP256BN.FromBytes(cred.E))) // barA = A'^{-e} b^{r1} BPrime := FP256BN.G1mul(B, r1) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) BPrime.Sub(FP256BN.G1mul(HRand, r2)) // b' = b^{r1} h_r^{-r2} S := FP256BN.FromBytes(cred.S) E := FP256BN.FromBytes(cred.E) sPrime := Modsub(S, FP256BN.Modmul(r2, r3, GroupOrder), GroupOrder) // Construct ZK proof rSk := RandModOrder(rng) re := RandModOrder(rng) rR2 := RandModOrder(rng) rR3 := RandModOrder(rng) rSPrime := RandModOrder(rng) rRNym := RandModOrder(rng) rAttrs := make([]*FP256BN.BIG, len(HiddenIndices)) for i := range HiddenIndices { rAttrs[i] = RandModOrder(rng) } t1 := APrime.Mul2(re, HRand, rR2) t2 := FP256BN.G1mul(HRand, rSPrime) t2.Add(BPrime.Mul2(rR3, HSk, rSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(rAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), rAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), rAttrs[len(HiddenIndices)-1])) } t3 := HSk.Mul2(rSk, HRand, rRNym) //Zk proof of BBS group signature alpha := RandModOrder(rng) beta := RandModOrder(rng) rAlpha := RandModOrder(rng) rBeta := RandModOrder(rng) rx := RandModOrder(rng) rDelta1 := RandModOrder(rng) rDelta2 := RandModOrder(rng) temp := Modadd(alpha,beta,GroupOrder) temp1 := Modadd(rAlpha,rBeta,GroupOrder) temp2 := Modadd(rDelta1,rDelta2,GroupOrder) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) Ax := EcpFromProto(uk.UK2) W := Ecp2FromProto(ipk.W) T1 := FP256BN.G1mul(U,alpha) T2 := FP256BN.G1mul(V,beta) T3 := FP256BN.G1mul(H,temp) T3.Add(Ax) R1 := FP256BN.G1mul(U,rAlpha) R2 := FP256BN.G1mul(V,rBeta) //compute pairing Grx := GenG2.Mul(rx) T4 := H.Mul(temp) e1 := FP256BN.Ate2(Grx,Ax,Grx,T4) Htp1 := H.Mul(temp1) e2 := FP256BN.Ate(W,Htp1) e2.Inverse() e2.Mul(e1) Htp2 := H.Mul(temp2) e4 := FP256BN.Ate(GenG2,Htp2) e4.Inverse() e4.Mul(e2) R3 := FP256BN.Fexp(e4) R4 := FP256BN.G1mul(T1,rx) R4.Sub(U.Mul(rDelta1)) R5 := FP256BN.G1mul(T2,rx) R5.Sub(V.Mul(rDelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message being signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) // add the previous hash and the nonce and hash again to compute a second hash (C value) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) ProofC := HashModOrder(proofData) ProofSSk := Modadd(rSk, FP256BN.Modmul(ProofC, sk, GroupOrder), GroupOrder) ProofSE := Modsub(re, FP256BN.Modmul(ProofC, E, GroupOrder), GroupOrder) ProofSR2 := Modadd(rR2, FP256BN.Modmul(ProofC, r2, GroupOrder), GroupOrder) ProofSR3 := Modsub(rR3, FP256BN.Modmul(ProofC, r3, GroupOrder), GroupOrder) ProofSSPrime := Modadd(rSPrime, FP256BN.Modmul(ProofC, sPrime, GroupOrder), GroupOrder) ProofSRNym := Modadd(rRNym, FP256BN.Modmul(ProofC, RNym, GroupOrder), GroupOrder) //compute s-values for zk proof of group signature ProofSalpha := Modadd(rAlpha,FP256BN.Modmul(ProofC,alpha,GroupOrder),GroupOrder) ProofSbeta := Modadd(rBeta,FP256BN.Modmul(ProofC,beta,GroupOrder),GroupOrder) x := FP256BN.FromBytes(uk.UK1) ProofSx := Modadd(rx,FP256BN.Modmul(ProofC,x,GroupOrder),GroupOrder) delta1 := FP256BN.Modmul(alpha,x,GroupOrder) ProofSdelta1 := Modadd(rDelta1,FP256BN.Modmul(ProofC,delta1,GroupOrder),GroupOrder) delta2 := FP256BN.Modmul(beta,x,GroupOrder) ProofSdelta2 := Modadd(rDelta2,FP256BN.Modmul(ProofC,delta2,GroupOrder),GroupOrder) ProofSAttrs := make([][]byte, len(HiddenIndices)) for i, j := range HiddenIndices { ProofSAttrs[i] = BigToBytes(Modadd(rAttrs[i], FP256BN.Modmul(ProofC, FP256BN.FromBytes(cred.Attrs[j]), GroupOrder), GroupOrder)) } return &Signature{ EcpToProto(APrime), EcpToProto(ABar), EcpToProto(BPrime), EcpToProto(T1), EcpToProto(T2), EcpToProto(T3), BigToBytes(ProofC), BigToBytes(ProofSSk), BigToBytes(ProofSE), BigToBytes(ProofSR2), BigToBytes(ProofSR3), BigToBytes(ProofSSPrime), BigToBytes(ProofSalpha), BigToBytes(ProofSbeta), BigToBytes(ProofSx), BigToBytes(ProofSdelta1), BigToBytes(ProofSdelta2), ProofSAttrs, BigToBytes(Nonce), EcpToProto(Nym), BigToBytes(ProofSRNym)}, nil } // Ver verifies an idemix signature // Disclosure steers which attributes it expects to be disclosed // attributeValues[i] contains the desired attribute value for the i-th undisclosed attribute in Disclosure func (sig *Signature) Ver(Disclosure []byte, ipk *IssuerPublicKey, gpk *GroupPublicKey, msg []byte, attributeValues []*FP256BN.BIG) error { HiddenIndices := hiddenIndices(Disclosure) APrime := EcpFromProto(sig.GetAPrime()) ABar := EcpFromProto(sig.GetABar()) BPrime := EcpFromProto(sig.GetBPrime()) Nym := EcpFromProto(sig.GetNym()) ProofC := FP256BN.FromBytes(sig.GetProofC()) ProofSSk := FP256BN.FromBytes(sig.GetProofSSk()) ProofSE := FP256BN.FromBytes(sig.GetProofSE()) ProofSR2 := FP256BN.FromBytes(sig.GetProofSR2()) ProofSR3 := FP256BN.FromBytes(sig.GetProofSR3()) ProofSSPrime := FP256BN.FromBytes(sig.GetProofSSPrime()) ProofSRNym := FP256BN.FromBytes(sig.GetProofSRNym()) ProofSAttrs := make([]*FP256BN.BIG, len(sig.GetProofSAttrs())) T1 := EcpFromProto(sig.GetT1()) T2 := EcpFromProto(sig.GetT2()) T3 := EcpFromProto(sig.GetT3()) ProofSalpha := FP256BN.FromBytes(sig.GetProofSalpha()) ProofSbeta := FP256BN.FromBytes(sig.GetProofSbeta()) ProofSx := FP256BN.FromBytes(sig.GetProofSx()) ProofSdelta1 := FP256BN.FromBytes(sig.GetProofSdelta1()) ProofSdelta2 := FP256BN.FromBytes(sig.GetProofSdelta2()) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) s1 := Modadd(ProofSalpha,ProofSbeta,GroupOrder) s2 := Modadd(ProofSdelta1,ProofSdelta2,GroupOrder) if len(sig.ProofSAttrs) != len(HiddenIndices) { return errors.Errorf("signature invalid: incorrect amount of s-values for AttributeProofSpec") } for i, b := range sig.ProofSAttrs { ProofSAttrs[i] = FP256BN.FromBytes(b) } Nonce := FP256BN.FromBytes(sig.GetNonce()) W := Ecp2FromProto(ipk.W) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) if APrime.Is_infinity() { return errors.Errorf("signature invalid: APrime = 1") } temp1 := FP256BN.Ate(W, APrime) temp2 := FP256BN.Ate(GenG2, ABar) temp2.Inverse() temp1.Mul(temp2) if !FP256BN.Fexp(temp1).Isunity() { return errors.Errorf("signature invalid: APrime and ABar don't have the expected structure") } t1 := APrime.Mul2(ProofSE, HRand, ProofSR2) temp := FP256BN.NewECP() temp.Copy(ABar) temp.Sub(BPrime) t1.Sub(FP256BN.G1mul(temp, ProofC)) t2 := FP256BN.G1mul(HRand, ProofSSPrime) t2.Add(BPrime.Mul2(ProofSR3, HSk, ProofSSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(ProofSAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), ProofSAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), ProofSAttrs[len(HiddenIndices)-1])) } temp = FP256BN.NewECP() temp.Copy(GenG1) for index, disclose := range Disclosure { if disclose != 0 { temp.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[index]), attributeValues[index])) } } t2.Add(FP256BN.G1mul(temp, ProofC)) t3 := HSk.Mul2(ProofSSk, HRand, ProofSRNym) t3.Sub(Nym.Mul(ProofC)) //compute R1,R2,R3,R4,R5 R1 := FP256BN.G1mul(U,ProofSalpha) R1.Sub(T1.Mul(ProofC)) R2 := FP256BN.G1mul(V,ProofSbeta) R2.Sub(T2.Mul(ProofC)) Gsx := GenG2.Mul(ProofSx) e1 := FP256BN.Ate(Gsx,T3) Hs1 := H.Mul(s1) e2 := FP256BN.Ate(W,Hs1) e2.Inverse() e1.Mul(e2) Hs2 := H.Mul(s2) e4 := FP256BN.Ate(GenG2,Hs2) e4.Inverse() e1.Mul(e4) Wc := W.Mul(ProofC) e6 := FP256BN.Ate(Wc,T3) Gc := GenG2.Mul(ProofC) e7 := FP256BN.Ate(Gc,GenG1) e7.Inverse() e7.Mul(e6) e7.Mul(e1) R3 := FP256BN.Fexp(e7) R4 := FP256BN.G1mul(T1,ProofSx) R4.Sub(U.Mul(ProofSdelta1)) R5 := FP256BN.G1mul(T2,ProofSx) R5.Sub(V.Mul(ProofSdelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message that was signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) if *ProofC != *HashModOrder(proofData) { return errors.Errorf("signature invalid: zero-knowledge proof is invalid") } return nil }
{ return nil, errors.Errorf("cannot create idemix signature: received nil input") }
conditional_block
signature.go
/* Copyright IBM Corp. All Rights Reserved. SPDX-License-Identifier: Apache-2.0 */ package idemix import ( "github.com/hyperledger/fabric-amcl/amcl" "github.com/hyperledger/fabric-amcl/amcl/FP256BN" "github.com/pkg/errors" ) // signLabel is the label used in zero-knowledge proof (ZKP) to identify that this ZKP is a signature of knowledge const signLabel = "sign" // A signature that is produced using an Identity Mixer credential is a so-called signature of knowledge // (for details see C.P.Schnorr "Efficient Identification and Signatures for Smart Cards") // An Identity Mixer signature is a signature of knowledge that signs a message and proves (in zero-knowledge) // the knowledge of the user secret (and possibly attributes) signed inside a credential // that was issued by a certain issuer (referred to with the issuer public key) // The signature is verified using the message being signed and the public key of the issuer // Some of the attributes from the credential can be selectvely disclosed or different statements can be proven about // credential atrributes without diclosing them in the clear // The difference between a standard signature using X.509 certificates and an Identity Mixer signature is // the advanced privacy features provided by Identity Mixer (due to zero-knowledge proofs): // - Unlinkability of the signatures produced with the same credential // - Selective attribute disclosure and predicates over attributes // Make a slice of all the attribute indices that will not be disclosed func hiddenIndices(Disclosure []byte) []int { HiddenIndices := make([]int, 0) for index, disclose := range Disclosure { if disclose == 0 { HiddenIndices = append(HiddenIndices, index) } } return HiddenIndices } // NewSignature creates a new idemix signature (Schnorr-type signature) // The []byte Disclosure steers which attributes are disclosed: // if Disclosure[i] == 0 then attribute i remains hidden and otherwise it is disclosed. // We use the zero-knowledge proof by http://eprint.iacr.org/2016/663.pdf to prove knowledge of a BBS+ signature //We use the zero-knowledge proof by https://link.springer.com/content/pdf/10.1007%2F978-3-540-28628-8_3.pdf to prove the knowledge of user's secret key. func NewSignature(cred *Credential, sk *FP256BN.BIG, Nym *FP256BN.ECP, RNym *FP256BN.BIG, ipk *IssuerPublicKey, gpk *GroupPublicKey, uk *UserKey, Disclosure []byte, msg []byte, rng *amcl.RAND) (*Signature, error) { if cred == nil || sk == nil || Nym == nil || RNym == nil || ipk == nil || rng == nil { return nil, errors.Errorf("cannot create idemix signature: received nil input") } HiddenIndices := hiddenIndices(Disclosure) // Start sig r1 := RandModOrder(rng) r2 := RandModOrder(rng) r3 := FP256BN.NewBIGcopy(r1) r3.Invmodp(GroupOrder) Nonce := RandModOrder(rng) A := EcpFromProto(cred.A) B := EcpFromProto(cred.B) APrime := FP256BN.G1mul(A, r1) // A' = A^{r1} ABar := FP256BN.G1mul(B, r1) ABar.Sub(FP256BN.G1mul(APrime, FP256BN.FromBytes(cred.E))) // barA = A'^{-e} b^{r1} BPrime := FP256BN.G1mul(B, r1) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) BPrime.Sub(FP256BN.G1mul(HRand, r2)) // b' = b^{r1} h_r^{-r2} S := FP256BN.FromBytes(cred.S) E := FP256BN.FromBytes(cred.E) sPrime := Modsub(S, FP256BN.Modmul(r2, r3, GroupOrder), GroupOrder) // Construct ZK proof rSk := RandModOrder(rng) re := RandModOrder(rng) rR2 := RandModOrder(rng) rR3 := RandModOrder(rng) rSPrime := RandModOrder(rng) rRNym := RandModOrder(rng) rAttrs := make([]*FP256BN.BIG, len(HiddenIndices)) for i := range HiddenIndices { rAttrs[i] = RandModOrder(rng) } t1 := APrime.Mul2(re, HRand, rR2) t2 := FP256BN.G1mul(HRand, rSPrime) t2.Add(BPrime.Mul2(rR3, HSk, rSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(rAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), rAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), rAttrs[len(HiddenIndices)-1])) } t3 := HSk.Mul2(rSk, HRand, rRNym) //Zk proof of BBS group signature alpha := RandModOrder(rng) beta := RandModOrder(rng) rAlpha := RandModOrder(rng) rBeta := RandModOrder(rng) rx := RandModOrder(rng) rDelta1 := RandModOrder(rng) rDelta2 := RandModOrder(rng) temp := Modadd(alpha,beta,GroupOrder) temp1 := Modadd(rAlpha,rBeta,GroupOrder) temp2 := Modadd(rDelta1,rDelta2,GroupOrder) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) Ax := EcpFromProto(uk.UK2) W := Ecp2FromProto(ipk.W) T1 := FP256BN.G1mul(U,alpha) T2 := FP256BN.G1mul(V,beta) T3 := FP256BN.G1mul(H,temp) T3.Add(Ax) R1 := FP256BN.G1mul(U,rAlpha) R2 := FP256BN.G1mul(V,rBeta) //compute pairing Grx := GenG2.Mul(rx) T4 := H.Mul(temp) e1 := FP256BN.Ate2(Grx,Ax,Grx,T4) Htp1 := H.Mul(temp1) e2 := FP256BN.Ate(W,Htp1) e2.Inverse() e2.Mul(e1) Htp2 := H.Mul(temp2) e4 := FP256BN.Ate(GenG2,Htp2) e4.Inverse() e4.Mul(e2) R3 := FP256BN.Fexp(e4) R4 := FP256BN.G1mul(T1,rx) R4.Sub(U.Mul(rDelta1)) R5 := FP256BN.G1mul(T2,rx) R5.Sub(V.Mul(rDelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message being signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) // add the previous hash and the nonce and hash again to compute a second hash (C value) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) ProofC := HashModOrder(proofData) ProofSSk := Modadd(rSk, FP256BN.Modmul(ProofC, sk, GroupOrder), GroupOrder) ProofSE := Modsub(re, FP256BN.Modmul(ProofC, E, GroupOrder), GroupOrder) ProofSR2 := Modadd(rR2, FP256BN.Modmul(ProofC, r2, GroupOrder), GroupOrder) ProofSR3 := Modsub(rR3, FP256BN.Modmul(ProofC, r3, GroupOrder), GroupOrder) ProofSSPrime := Modadd(rSPrime, FP256BN.Modmul(ProofC, sPrime, GroupOrder), GroupOrder) ProofSRNym := Modadd(rRNym, FP256BN.Modmul(ProofC, RNym, GroupOrder), GroupOrder) //compute s-values for zk proof of group signature ProofSalpha := Modadd(rAlpha,FP256BN.Modmul(ProofC,alpha,GroupOrder),GroupOrder) ProofSbeta := Modadd(rBeta,FP256BN.Modmul(ProofC,beta,GroupOrder),GroupOrder) x := FP256BN.FromBytes(uk.UK1) ProofSx := Modadd(rx,FP256BN.Modmul(ProofC,x,GroupOrder),GroupOrder) delta1 := FP256BN.Modmul(alpha,x,GroupOrder) ProofSdelta1 := Modadd(rDelta1,FP256BN.Modmul(ProofC,delta1,GroupOrder),GroupOrder) delta2 := FP256BN.Modmul(beta,x,GroupOrder) ProofSdelta2 := Modadd(rDelta2,FP256BN.Modmul(ProofC,delta2,GroupOrder),GroupOrder) ProofSAttrs := make([][]byte, len(HiddenIndices)) for i, j := range HiddenIndices { ProofSAttrs[i] = BigToBytes(Modadd(rAttrs[i], FP256BN.Modmul(ProofC, FP256BN.FromBytes(cred.Attrs[j]), GroupOrder), GroupOrder)) } return &Signature{ EcpToProto(APrime), EcpToProto(ABar), EcpToProto(BPrime), EcpToProto(T1), EcpToProto(T2), EcpToProto(T3), BigToBytes(ProofC), BigToBytes(ProofSSk), BigToBytes(ProofSE), BigToBytes(ProofSR2), BigToBytes(ProofSR3), BigToBytes(ProofSSPrime), BigToBytes(ProofSalpha), BigToBytes(ProofSbeta), BigToBytes(ProofSx), BigToBytes(ProofSdelta1), BigToBytes(ProofSdelta2), ProofSAttrs, BigToBytes(Nonce), EcpToProto(Nym), BigToBytes(ProofSRNym)}, nil } // Ver verifies an idemix signature // Disclosure steers which attributes it expects to be disclosed // attributeValues[i] contains the desired attribute value for the i-th undisclosed attribute in Disclosure func (sig *Signature) Ver(Disclosure []byte, ipk *IssuerPublicKey, gpk *GroupPublicKey, msg []byte, attributeValues []*FP256BN.BIG) error
{ HiddenIndices := hiddenIndices(Disclosure) APrime := EcpFromProto(sig.GetAPrime()) ABar := EcpFromProto(sig.GetABar()) BPrime := EcpFromProto(sig.GetBPrime()) Nym := EcpFromProto(sig.GetNym()) ProofC := FP256BN.FromBytes(sig.GetProofC()) ProofSSk := FP256BN.FromBytes(sig.GetProofSSk()) ProofSE := FP256BN.FromBytes(sig.GetProofSE()) ProofSR2 := FP256BN.FromBytes(sig.GetProofSR2()) ProofSR3 := FP256BN.FromBytes(sig.GetProofSR3()) ProofSSPrime := FP256BN.FromBytes(sig.GetProofSSPrime()) ProofSRNym := FP256BN.FromBytes(sig.GetProofSRNym()) ProofSAttrs := make([]*FP256BN.BIG, len(sig.GetProofSAttrs())) T1 := EcpFromProto(sig.GetT1()) T2 := EcpFromProto(sig.GetT2()) T3 := EcpFromProto(sig.GetT3()) ProofSalpha := FP256BN.FromBytes(sig.GetProofSalpha()) ProofSbeta := FP256BN.FromBytes(sig.GetProofSbeta()) ProofSx := FP256BN.FromBytes(sig.GetProofSx()) ProofSdelta1 := FP256BN.FromBytes(sig.GetProofSdelta1()) ProofSdelta2 := FP256BN.FromBytes(sig.GetProofSdelta2()) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) s1 := Modadd(ProofSalpha,ProofSbeta,GroupOrder) s2 := Modadd(ProofSdelta1,ProofSdelta2,GroupOrder) if len(sig.ProofSAttrs) != len(HiddenIndices) { return errors.Errorf("signature invalid: incorrect amount of s-values for AttributeProofSpec") } for i, b := range sig.ProofSAttrs { ProofSAttrs[i] = FP256BN.FromBytes(b) } Nonce := FP256BN.FromBytes(sig.GetNonce()) W := Ecp2FromProto(ipk.W) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) if APrime.Is_infinity() { return errors.Errorf("signature invalid: APrime = 1") } temp1 := FP256BN.Ate(W, APrime) temp2 := FP256BN.Ate(GenG2, ABar) temp2.Inverse() temp1.Mul(temp2) if !FP256BN.Fexp(temp1).Isunity() { return errors.Errorf("signature invalid: APrime and ABar don't have the expected structure") } t1 := APrime.Mul2(ProofSE, HRand, ProofSR2) temp := FP256BN.NewECP() temp.Copy(ABar) temp.Sub(BPrime) t1.Sub(FP256BN.G1mul(temp, ProofC)) t2 := FP256BN.G1mul(HRand, ProofSSPrime) t2.Add(BPrime.Mul2(ProofSR3, HSk, ProofSSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(ProofSAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), ProofSAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), ProofSAttrs[len(HiddenIndices)-1])) } temp = FP256BN.NewECP() temp.Copy(GenG1) for index, disclose := range Disclosure { if disclose != 0 { temp.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[index]), attributeValues[index])) } } t2.Add(FP256BN.G1mul(temp, ProofC)) t3 := HSk.Mul2(ProofSSk, HRand, ProofSRNym) t3.Sub(Nym.Mul(ProofC)) //compute R1,R2,R3,R4,R5 R1 := FP256BN.G1mul(U,ProofSalpha) R1.Sub(T1.Mul(ProofC)) R2 := FP256BN.G1mul(V,ProofSbeta) R2.Sub(T2.Mul(ProofC)) Gsx := GenG2.Mul(ProofSx) e1 := FP256BN.Ate(Gsx,T3) Hs1 := H.Mul(s1) e2 := FP256BN.Ate(W,Hs1) e2.Inverse() e1.Mul(e2) Hs2 := H.Mul(s2) e4 := FP256BN.Ate(GenG2,Hs2) e4.Inverse() e1.Mul(e4) Wc := W.Mul(ProofC) e6 := FP256BN.Ate(Wc,T3) Gc := GenG2.Mul(ProofC) e7 := FP256BN.Ate(Gc,GenG1) e7.Inverse() e7.Mul(e6) e7.Mul(e1) R3 := FP256BN.Fexp(e7) R4 := FP256BN.G1mul(T1,ProofSx) R4.Sub(U.Mul(ProofSdelta1)) R5 := FP256BN.G1mul(T2,ProofSx) R5.Sub(V.Mul(ProofSdelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message that was signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) if *ProofC != *HashModOrder(proofData) { return errors.Errorf("signature invalid: zero-knowledge proof is invalid") } return nil }
identifier_body
signature.go
/* Copyright IBM Corp. All Rights Reserved. SPDX-License-Identifier: Apache-2.0 */ package idemix import ( "github.com/hyperledger/fabric-amcl/amcl" "github.com/hyperledger/fabric-amcl/amcl/FP256BN" "github.com/pkg/errors" ) // signLabel is the label used in zero-knowledge proof (ZKP) to identify that this ZKP is a signature of knowledge const signLabel = "sign" // A signature that is produced using an Identity Mixer credential is a so-called signature of knowledge // (for details see C.P.Schnorr "Efficient Identification and Signatures for Smart Cards") // An Identity Mixer signature is a signature of knowledge that signs a message and proves (in zero-knowledge) // the knowledge of the user secret (and possibly attributes) signed inside a credential // that was issued by a certain issuer (referred to with the issuer public key) // The signature is verified using the message being signed and the public key of the issuer // Some of the attributes from the credential can be selectvely disclosed or different statements can be proven about // credential atrributes without diclosing them in the clear // The difference between a standard signature using X.509 certificates and an Identity Mixer signature is // the advanced privacy features provided by Identity Mixer (due to zero-knowledge proofs): // - Unlinkability of the signatures produced with the same credential // - Selective attribute disclosure and predicates over attributes // Make a slice of all the attribute indices that will not be disclosed func hiddenIndices(Disclosure []byte) []int { HiddenIndices := make([]int, 0) for index, disclose := range Disclosure { if disclose == 0 { HiddenIndices = append(HiddenIndices, index) } } return HiddenIndices } // NewSignature creates a new idemix signature (Schnorr-type signature) // The []byte Disclosure steers which attributes are disclosed: // if Disclosure[i] == 0 then attribute i remains hidden and otherwise it is disclosed. // We use the zero-knowledge proof by http://eprint.iacr.org/2016/663.pdf to prove knowledge of a BBS+ signature //We use the zero-knowledge proof by https://link.springer.com/content/pdf/10.1007%2F978-3-540-28628-8_3.pdf to prove the knowledge of user's secret key. func NewSignature(cred *Credential, sk *FP256BN.BIG, Nym *FP256BN.ECP, RNym *FP256BN.BIG, ipk *IssuerPublicKey, gpk *GroupPublicKey, uk *UserKey, Disclosure []byte, msg []byte, rng *amcl.RAND) (*Signature, error) { if cred == nil || sk == nil || Nym == nil || RNym == nil || ipk == nil || rng == nil { return nil, errors.Errorf("cannot create idemix signature: received nil input") } HiddenIndices := hiddenIndices(Disclosure) // Start sig r1 := RandModOrder(rng) r2 := RandModOrder(rng) r3 := FP256BN.NewBIGcopy(r1) r3.Invmodp(GroupOrder) Nonce := RandModOrder(rng) A := EcpFromProto(cred.A) B := EcpFromProto(cred.B) APrime := FP256BN.G1mul(A, r1) // A' = A^{r1} ABar := FP256BN.G1mul(B, r1) ABar.Sub(FP256BN.G1mul(APrime, FP256BN.FromBytes(cred.E))) // barA = A'^{-e} b^{r1} BPrime := FP256BN.G1mul(B, r1) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) BPrime.Sub(FP256BN.G1mul(HRand, r2)) // b' = b^{r1} h_r^{-r2} S := FP256BN.FromBytes(cred.S) E := FP256BN.FromBytes(cred.E) sPrime := Modsub(S, FP256BN.Modmul(r2, r3, GroupOrder), GroupOrder) // Construct ZK proof rSk := RandModOrder(rng) re := RandModOrder(rng) rR2 := RandModOrder(rng) rR3 := RandModOrder(rng) rSPrime := RandModOrder(rng) rRNym := RandModOrder(rng) rAttrs := make([]*FP256BN.BIG, len(HiddenIndices)) for i := range HiddenIndices { rAttrs[i] = RandModOrder(rng) } t1 := APrime.Mul2(re, HRand, rR2) t2 := FP256BN.G1mul(HRand, rSPrime) t2.Add(BPrime.Mul2(rR3, HSk, rSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(rAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), rAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), rAttrs[len(HiddenIndices)-1])) } t3 := HSk.Mul2(rSk, HRand, rRNym) //Zk proof of BBS group signature alpha := RandModOrder(rng) beta := RandModOrder(rng) rAlpha := RandModOrder(rng) rBeta := RandModOrder(rng) rx := RandModOrder(rng) rDelta1 := RandModOrder(rng) rDelta2 := RandModOrder(rng) temp := Modadd(alpha,beta,GroupOrder) temp1 := Modadd(rAlpha,rBeta,GroupOrder) temp2 := Modadd(rDelta1,rDelta2,GroupOrder) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) Ax := EcpFromProto(uk.UK2) W := Ecp2FromProto(ipk.W) T1 := FP256BN.G1mul(U,alpha) T2 := FP256BN.G1mul(V,beta) T3 := FP256BN.G1mul(H,temp) T3.Add(Ax) R1 := FP256BN.G1mul(U,rAlpha) R2 := FP256BN.G1mul(V,rBeta)
Htp1 := H.Mul(temp1) e2 := FP256BN.Ate(W,Htp1) e2.Inverse() e2.Mul(e1) Htp2 := H.Mul(temp2) e4 := FP256BN.Ate(GenG2,Htp2) e4.Inverse() e4.Mul(e2) R3 := FP256BN.Fexp(e4) R4 := FP256BN.G1mul(T1,rx) R4.Sub(U.Mul(rDelta1)) R5 := FP256BN.G1mul(T2,rx) R5.Sub(V.Mul(rDelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message being signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) // add the previous hash and the nonce and hash again to compute a second hash (C value) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) ProofC := HashModOrder(proofData) ProofSSk := Modadd(rSk, FP256BN.Modmul(ProofC, sk, GroupOrder), GroupOrder) ProofSE := Modsub(re, FP256BN.Modmul(ProofC, E, GroupOrder), GroupOrder) ProofSR2 := Modadd(rR2, FP256BN.Modmul(ProofC, r2, GroupOrder), GroupOrder) ProofSR3 := Modsub(rR3, FP256BN.Modmul(ProofC, r3, GroupOrder), GroupOrder) ProofSSPrime := Modadd(rSPrime, FP256BN.Modmul(ProofC, sPrime, GroupOrder), GroupOrder) ProofSRNym := Modadd(rRNym, FP256BN.Modmul(ProofC, RNym, GroupOrder), GroupOrder) //compute s-values for zk proof of group signature ProofSalpha := Modadd(rAlpha,FP256BN.Modmul(ProofC,alpha,GroupOrder),GroupOrder) ProofSbeta := Modadd(rBeta,FP256BN.Modmul(ProofC,beta,GroupOrder),GroupOrder) x := FP256BN.FromBytes(uk.UK1) ProofSx := Modadd(rx,FP256BN.Modmul(ProofC,x,GroupOrder),GroupOrder) delta1 := FP256BN.Modmul(alpha,x,GroupOrder) ProofSdelta1 := Modadd(rDelta1,FP256BN.Modmul(ProofC,delta1,GroupOrder),GroupOrder) delta2 := FP256BN.Modmul(beta,x,GroupOrder) ProofSdelta2 := Modadd(rDelta2,FP256BN.Modmul(ProofC,delta2,GroupOrder),GroupOrder) ProofSAttrs := make([][]byte, len(HiddenIndices)) for i, j := range HiddenIndices { ProofSAttrs[i] = BigToBytes(Modadd(rAttrs[i], FP256BN.Modmul(ProofC, FP256BN.FromBytes(cred.Attrs[j]), GroupOrder), GroupOrder)) } return &Signature{ EcpToProto(APrime), EcpToProto(ABar), EcpToProto(BPrime), EcpToProto(T1), EcpToProto(T2), EcpToProto(T3), BigToBytes(ProofC), BigToBytes(ProofSSk), BigToBytes(ProofSE), BigToBytes(ProofSR2), BigToBytes(ProofSR3), BigToBytes(ProofSSPrime), BigToBytes(ProofSalpha), BigToBytes(ProofSbeta), BigToBytes(ProofSx), BigToBytes(ProofSdelta1), BigToBytes(ProofSdelta2), ProofSAttrs, BigToBytes(Nonce), EcpToProto(Nym), BigToBytes(ProofSRNym)}, nil } // Ver verifies an idemix signature // Disclosure steers which attributes it expects to be disclosed // attributeValues[i] contains the desired attribute value for the i-th undisclosed attribute in Disclosure func (sig *Signature) Ver(Disclosure []byte, ipk *IssuerPublicKey, gpk *GroupPublicKey, msg []byte, attributeValues []*FP256BN.BIG) error { HiddenIndices := hiddenIndices(Disclosure) APrime := EcpFromProto(sig.GetAPrime()) ABar := EcpFromProto(sig.GetABar()) BPrime := EcpFromProto(sig.GetBPrime()) Nym := EcpFromProto(sig.GetNym()) ProofC := FP256BN.FromBytes(sig.GetProofC()) ProofSSk := FP256BN.FromBytes(sig.GetProofSSk()) ProofSE := FP256BN.FromBytes(sig.GetProofSE()) ProofSR2 := FP256BN.FromBytes(sig.GetProofSR2()) ProofSR3 := FP256BN.FromBytes(sig.GetProofSR3()) ProofSSPrime := FP256BN.FromBytes(sig.GetProofSSPrime()) ProofSRNym := FP256BN.FromBytes(sig.GetProofSRNym()) ProofSAttrs := make([]*FP256BN.BIG, len(sig.GetProofSAttrs())) T1 := EcpFromProto(sig.GetT1()) T2 := EcpFromProto(sig.GetT2()) T3 := EcpFromProto(sig.GetT3()) ProofSalpha := FP256BN.FromBytes(sig.GetProofSalpha()) ProofSbeta := FP256BN.FromBytes(sig.GetProofSbeta()) ProofSx := FP256BN.FromBytes(sig.GetProofSx()) ProofSdelta1 := FP256BN.FromBytes(sig.GetProofSdelta1()) ProofSdelta2 := FP256BN.FromBytes(sig.GetProofSdelta2()) U := EcpFromProto(gpk.U) V := EcpFromProto(gpk.V) H := EcpFromProto(gpk.H) s1 := Modadd(ProofSalpha,ProofSbeta,GroupOrder) s2 := Modadd(ProofSdelta1,ProofSdelta2,GroupOrder) if len(sig.ProofSAttrs) != len(HiddenIndices) { return errors.Errorf("signature invalid: incorrect amount of s-values for AttributeProofSpec") } for i, b := range sig.ProofSAttrs { ProofSAttrs[i] = FP256BN.FromBytes(b) } Nonce := FP256BN.FromBytes(sig.GetNonce()) W := Ecp2FromProto(ipk.W) HRand := EcpFromProto(ipk.HRand) HSk := EcpFromProto(ipk.HSk) if APrime.Is_infinity() { return errors.Errorf("signature invalid: APrime = 1") } temp1 := FP256BN.Ate(W, APrime) temp2 := FP256BN.Ate(GenG2, ABar) temp2.Inverse() temp1.Mul(temp2) if !FP256BN.Fexp(temp1).Isunity() { return errors.Errorf("signature invalid: APrime and ABar don't have the expected structure") } t1 := APrime.Mul2(ProofSE, HRand, ProofSR2) temp := FP256BN.NewECP() temp.Copy(ABar) temp.Sub(BPrime) t1.Sub(FP256BN.G1mul(temp, ProofC)) t2 := FP256BN.G1mul(HRand, ProofSSPrime) t2.Add(BPrime.Mul2(ProofSR3, HSk, ProofSSk)) for i := 0; i < len(HiddenIndices)/2; i++ { t2.Add(EcpFromProto(ipk.HAttrs[HiddenIndices[2*i]]).Mul2(ProofSAttrs[2*i], EcpFromProto(ipk.HAttrs[HiddenIndices[2*i+1]]), ProofSAttrs[2*i+1])) } if len(HiddenIndices)%2 != 0 { t2.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[HiddenIndices[len(HiddenIndices)-1]]), ProofSAttrs[len(HiddenIndices)-1])) } temp = FP256BN.NewECP() temp.Copy(GenG1) for index, disclose := range Disclosure { if disclose != 0 { temp.Add(FP256BN.G1mul(EcpFromProto(ipk.HAttrs[index]), attributeValues[index])) } } t2.Add(FP256BN.G1mul(temp, ProofC)) t3 := HSk.Mul2(ProofSSk, HRand, ProofSRNym) t3.Sub(Nym.Mul(ProofC)) //compute R1,R2,R3,R4,R5 R1 := FP256BN.G1mul(U,ProofSalpha) R1.Sub(T1.Mul(ProofC)) R2 := FP256BN.G1mul(V,ProofSbeta) R2.Sub(T2.Mul(ProofC)) Gsx := GenG2.Mul(ProofSx) e1 := FP256BN.Ate(Gsx,T3) Hs1 := H.Mul(s1) e2 := FP256BN.Ate(W,Hs1) e2.Inverse() e1.Mul(e2) Hs2 := H.Mul(s2) e4 := FP256BN.Ate(GenG2,Hs2) e4.Inverse() e1.Mul(e4) Wc := W.Mul(ProofC) e6 := FP256BN.Ate(Wc,T3) Gc := GenG2.Mul(ProofC) e7 := FP256BN.Ate(Gc,GenG1) e7.Inverse() e7.Mul(e6) e7.Mul(e1) R3 := FP256BN.Fexp(e7) R4 := FP256BN.G1mul(T1,ProofSx) R4.Sub(U.Mul(ProofSdelta1)) R5 := FP256BN.G1mul(T2,ProofSx) R5.Sub(V.Mul(ProofSdelta2)) // proofData is the data being hashed, it consists of: // the signature label // 14 elements of G1 each taking 2*FieldBytes+1 bytes and 1 element of GT taking 12*FieldBytes // one bigint (hash of the issuer public key) of length FieldBytes // disclosed attributes // message that was signed proofData := make([]byte, len([]byte(signLabel))+14*(2*FieldBytes+1)+13*FieldBytes+len(Disclosure)+len(msg)) index := 0 index = appendBytesString(proofData, index, signLabel) index = appendBytesG1(proofData, index, t1) index = appendBytesG1(proofData, index, t2) index = appendBytesG1(proofData, index, t3) index = appendBytesG1(proofData, index, APrime) index = appendBytesG1(proofData, index, ABar) index = appendBytesG1(proofData, index, BPrime) index = appendBytesG1(proofData, index, Nym) index = appendBytesG1(proofData, index, T1) index = appendBytesG1(proofData, index, T2) index = appendBytesG1(proofData, index, T3) index = appendBytesG1(proofData, index, R1) index = appendBytesG1(proofData, index, R2) index = appendBytesGT(proofData, index, R3) index = appendBytesG1(proofData, index, R4) index = appendBytesG1(proofData, index, R5) copy(proofData[index:], ipk.Hash) index = index + FieldBytes copy(proofData[index:], Disclosure) index = index + len(Disclosure) copy(proofData[index:], msg) c := HashModOrder(proofData) index = 0 proofData = proofData[:2*FieldBytes] index = appendBytesBig(proofData, index, c) index = appendBytesBig(proofData, index, Nonce) if *ProofC != *HashModOrder(proofData) { return errors.Errorf("signature invalid: zero-knowledge proof is invalid") } return nil }
//compute pairing Grx := GenG2.Mul(rx) T4 := H.Mul(temp) e1 := FP256BN.Ate2(Grx,Ax,Grx,T4)
random_line_split
in-memory-plants-db-service.ts
import { InMemoryDbService, RequestInfo } from 'angular-in-memory-web-api'; import { ResponseOptions } from '@angular/http'; import { Category } from '../app/products/models/category'; export class InMemoryPlantsDbService implements InMemoryDbService { categories: Category[];
() { this.categories = [ { 'id': 0, 'name': 'Balkonnövények', 'plantsCount': 3 }, { 'id': 1, 'name': 'Rózsák', 'plantsCount': 4 }, { 'id': 2, 'name': 'Bogyós gyülömcsök', 'plantsCount': 5 }, { 'id': 3, 'name': 'Kúszónövények', 'plantsCount': 4 }, { 'id': 4, 'name': 'Gyógyhatású növények', 'plantsCount': 2 } ]; } createDb() { const login = []; const register = []; const categories = this.categories; const plants = [ { 'name': 'Estrella White', 'price': 2650, 'id': 0, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'HALF_SHADY', 'description': 'Virágait erős szárakon a lomb felett hozza. Bőséges virágzás egészen őszig. Az extrém időjárási ' + 'viszonyokat is jól bírja, hajtásait a szél sem töri meg. A bőséges virágzáshoz elengedhetetlen a gyakori öntözés ' + 'és tápoldatozás. Az általunk ajánlott „Muskátli virágvarázs” kitűnően megfelel erre a célra.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'BIWEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100839_jpg_137965_406_488.jpg' }, { 'name': 'Álló muskátli', 'price': 1470, 'id': 1, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Minden évben újabb fajták érkeznek. A legszebb virágú és legjobb fajtákat válogattam össze ' + 'Önnek! Gazdagon virágoznak, nagyok a viráglabdák, szél- és időjárásállóak. Talán ezek a legfontosabb ' + 'tulajdonságok. Bizonyára elégedett lesz a képen látható fajtákkal.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100164_jpg_70579_406_488.jpg' }, { 'name': 'Csüngő begónia', 'price': 1390, 'id': 2, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Újdonság, ami nem olcsó, de annál értékesebb! Barátai irigyelni fogják a rengeteg virágért. ' + 'A 10 centiméteresre is megnövő fénylő virágok finom illatosak. Nem kell már az árnyékban sem nélkülöznie ' + 'a bőséges virágzást! Ilyen helyekre a begóniák ültetése a legjobb választás. Egészen őszig messziről fénylenek ' + 'virágai. Jó minőségű balkon virágföldet és rendszeres öntözést kíván.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100696_jpg_84843_406_488.jpg' }, { 'name': 'Tearózsa', 'price': 1390, 'id': 3, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Virágait erős szárakon a lomb felett hozza. Bőséges virágzás egészen őszig. Az extrém időjárási ' + 'viszonyokat is jól bírja, hajtásait a szél sem töri meg. A bőséges virágzáshoz elengedhetetlen a gyakori öntözés ' + 'és tápoldatozás. Az általunk ajánlott „Muskátli virágvarázs” kitűnően megfelel erre a célra.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100839_jpg_137965_406_488.jpg' }, { 'name': 'Bokorrózsa Emilia Maria', 'price': 1590, 'id': 4, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Egy nagyon gazdagon és szinte folyamatosan virágzó bokorrózsa. Pinkszínű, tömvetelt virágai ' + 'csodálatos,graepfruit,iris és ibolya keverékére emlékeztető illatúak. Kisebb rácson felvezetve magasabbra is nevelhető. ', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/115150_jpg_56050_406_488.jpg' }, { 'name': 'Ágyásrózsa Julia Child', 'price': 1590, 'id': 5, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Ágyásrózsa. Egy száron több virágot is hoz, amik akár 12 cm nagyságúak, erős, aranysárga színűek. ' + 'Legértékesebb csodálatos illata, különösen reggel mikor az első napsugarak felmelegítik a virágokat.' + 'Ez a fajta nem hiányozhat egyetlen rózsakertből sem!', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/119100_jpg_37607_406_488.jpg' }, { 'name': 'Művészrózsa Maurice Utrillo', 'price': 1890, 'id': 6, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Tűzpiros virágai fehéren csíkozottak. Mindehhez elragadó illat párosul.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/110840_jpg_91884_406_488.jpg' }, { 'name': 'Delikatess csüngőeper', 'price': 1380, 'id': 6, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Kert nélkül is folyamatosan szedheti a finom epreket! Használjon jó minőségű földet, megfelelő ' + 'méretű virágládát. Gondoskodjon a felesleges víz elvezetéséről! Napi rendszerességgel végzett öntözés és a ' + 'megfelelő tápanyagutánpótlás mellett rövid időn belül megjelennek az első termések.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/090060_jpg_535663_406_488.jpg' }, { 'name': 'Málna Boheme', 'price': 1090, 'id': 7, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Egy őszi érésű málna, amilyenre az ember vágyik. Szép színű, kellemes illatú, csodálatos ízű.' + ' Bogyói nagyon korán érnek. Felálló és kompakt növekedésű.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/151330_jpg_131481_406_488.jpg' }, { 'name': 'Málna Golden Everest', 'price': 1290, 'id': 8, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'A sárga termések teszik értékessé ezt a fajtát, mely különleges aromájával kiváló lekvárok, ' + 'édességek és gyümölcslé készítéséhez. Terméseit június közepén a második éves vesszők végén hozza.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/151240_jpg_273492_406_488.jpg' }, { 'name': 'Hortblue Petite áfonya', 'price': 1990, 'id': 9, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Hosszabbítsa meg a szüretelési időt egy közkedvelt bogyós fajtával! Sokan kedvelik az áfonyát akár ' + 'müzlibe, gyümölcssalátába vagy csak egyszerűen frissen a bokorról.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/154530_jpg_345266_406_488.jpg' }, { 'name': 'Vörös áfonya', 'price': 1690, 'id': 10, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Hosszabbítsa meg a szüretelési időt egy közkedvelt bogyós fajtával! Sokan kedvelik az áfonyát ' + 'akár müzlibe, gyümölcssalátába vagy csak egyszerűen frissen a bokorról.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/154520_jpg_402033_406_488.jpg' }, { 'name': 'Csokoládészőlő', 'price': 1890, 'id': 11, 'categoryId': 3, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Kínában és Japánban a szárított kérgét gyógyszerként használják. Gyulladáscsökkentő és a ' + 'menstruációs problémákat enyhítő hatása is ismert. Gyökere lázcsillapító hatású, termése egyes vélemények ' + 'szerint hatásos a daganatos betegségekkel szemben.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070535_jpg_51932_406_488.jpg' }, { 'name': 'Klemátisz ´Guernsey Cream´', 'price': 2190, 'id': 12, 'categoryId': 3, 'isFrostProof': false, 'lightReq': 'HALF_SHADY', 'description': 'Mint ahogy a neve is jelzi, az első krémszínű klemátisz! Különleges színű virágainak értékét egy ' + 'sötét háttér még inkább kiemeli. A virág zöld közepe a sötétsárga porzókkal egyedülálló szépséget kölcsönöz ' + 'a fajtának.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070770_jpg_431988_406_488.jpg' }, { 'name': 'Klemátisz \'Innocent Blush\'', 'price': 2990, 'id': 13, 'categoryId': 3, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Csodaszép, nosztalgikus, új nemesítésű fajta. A 2. éves hajtásokon fejlődnek ki nagy, telt virágai.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070735_jpg_228664_406_488.jpg' }, { 'name': 'Nyári lilaakác', 'price': 2590, 'id': 14, 'categoryId': 3, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Június - júliusban nyíló, rózsaszín - lilás virágai akár a 35 cm nagyságot is elérhetik. ' + 'Szereti a védett helyeket. Minél idősebb, annál szebb a virágzása. Edényben is nevelhető.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070820_jpg_611190_406_488.jpg' }, { 'name': 'Aloe Vera', 'price': 1790, 'id': 15, 'categoryId': 4, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Az Aloe vera az egyik legrégebbi gyógynövény a különféle külső és belső bajok gyógyítására. ' + 'Enyhíti a fájdalmat, segíti a sebek, sérülések gyorsabb gyógyulását. Vágjon le egy darab levelet, majd a ' + 'kipréselt levet kenje a sebre! A vaskos levelekből koktélok, italok is készülhetnek.', 'plantingTime': ['APRIL', 'MAY', 'JUNE', 'JULY', 'AUG', 'SEPT', 'OCT'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/102440_jpg_83473_406_488.jpg' }, { 'name': 'Homoktövis pár', 'price': 4150, 'id': 17, 'categoryId': 4, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Tudta Ön, hogy a homoktövis 10 - szer több C - vitamint tartalmaz mint a citrusfélék? ' + 'Egy igazi bomba az immunrendszernek, és egy hatásos fegyver a megfázások ellen.', 'plantingTime': ['APRIL', 'MAY', 'SEPT'], 'waterReq': 'HIGH', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/130583_jpg_109750_406_488.jpg' } ]; return {login, register, categories, plants}; } responseInterceptor(responseOptions: ResponseOptions, requestInfo: RequestInfo) { if (requestInfo.resourceUrl.includes('login') || requestInfo.resourceUrl.includes('register')) { responseOptions.body = {token: 'asdasd'}; responseOptions.status = 200; } else if (typeof responseOptions.body === 'object') { responseOptions.body = responseOptions.body ? (responseOptions.body as any).data : null; } if (requestInfo.resourceUrl.includes('categories') && requestInfo.req.url.includes('?')) { const result = this.categories.slice(Math.random() * 4); responseOptions.body = result; } return responseOptions; } }
constructor
identifier_name
in-memory-plants-db-service.ts
import { InMemoryDbService, RequestInfo } from 'angular-in-memory-web-api'; import { ResponseOptions } from '@angular/http'; import { Category } from '../app/products/models/category'; export class InMemoryPlantsDbService implements InMemoryDbService { categories: Category[]; constructor() { this.categories = [ { 'id': 0, 'name': 'Balkonnövények', 'plantsCount': 3 }, { 'id': 1, 'name': 'Rózsák', 'plantsCount': 4 }, { 'id': 2, 'name': 'Bogyós gyülömcsök', 'plantsCount': 5 }, { 'id': 3, 'name': 'Kúszónövények', 'plantsCount': 4 }, { 'id': 4, 'name': 'Gyógyhatású növények', 'plantsCount': 2 } ]; } createDb() { const login = []; const register = []; const categories = this.categories; const plants = [ { 'name': 'Estrella White', 'price': 2650, 'id': 0, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'HALF_SHADY', 'description': 'Virágait erős szárakon a lomb felett hozza. Bőséges virágzás egészen őszig. Az extrém időjárási ' + 'viszonyokat is jól bírja, hajtásait a szél sem töri meg. A bőséges virágzáshoz elengedhetetlen a gyakori öntözés ' + 'és tápoldatozás. Az általunk ajánlott „Muskátli virágvarázs” kitűnően megfelel erre a célra.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'BIWEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100839_jpg_137965_406_488.jpg' }, { 'name': 'Álló muskátli', 'price': 1470, 'id': 1, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Minden évben újabb fajták érkeznek. A legszebb virágú és legjobb fajtákat válogattam össze ' + 'Önnek! Gazdagon virágoznak, nagyok a viráglabdák, szél- és időjárásállóak. Talán ezek a legfontosabb ' + 'tulajdonságok. Bizonyára elégedett lesz a képen látható fajtákkal.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100164_jpg_70579_406_488.jpg' }, { 'name': 'Csüngő begónia', 'price': 1390, 'id': 2, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Újdonság, ami nem olcsó, de annál értékesebb! Barátai irigyelni fogják a rengeteg virágért. ' + 'A 10 centiméteresre is megnövő fénylő virágok finom illatosak. Nem kell már az árnyékban sem nélkülöznie ' + 'a bőséges virágzást! Ilyen helyekre a begóniák ültetése a legjobb választás. Egészen őszig messziről fénylenek ' + 'virágai. Jó minőségű balkon virágföldet és rendszeres öntözést kíván.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100696_jpg_84843_406_488.jpg' }, { 'name': 'Tearózsa', 'price': 1390, 'id': 3, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Virágait erős szárakon a lomb felett hozza. Bőséges virágzás egészen őszig. Az extrém időjárási ' + 'viszonyokat is jól bírja, hajtásait a szél sem töri meg. A bőséges virágzáshoz elengedhetetlen a gyakori öntözés ' + 'és tápoldatozás. Az általunk ajánlott „Muskátli virágvarázs” kitűnően megfelel erre a célra.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100839_jpg_137965_406_488.jpg' }, { 'name': 'Bokorrózsa Emilia Maria', 'price': 1590, 'id': 4, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Egy nagyon gazdagon és szinte folyamatosan virágzó bokorrózsa. Pinkszínű, tömvetelt virágai ' + 'csodálatos,graepfruit,iris és ibolya keverékére emlékeztető illatúak. Kisebb rácson felvezetve magasabbra is nevelhető. ', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/115150_jpg_56050_406_488.jpg' }, { 'name': 'Ágyásrózsa Julia Child', 'price': 1590, 'id': 5, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Ágyásrózsa. Egy száron több virágot is hoz, amik akár 12 cm nagyságúak, erős, aranysárga színűek. ' + 'Legértékesebb csodálatos illata, különösen reggel mikor az első napsugarak felmelegítik a virágokat.' + 'Ez a fajta nem hiányozhat egyetlen rózsakertből sem!', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/119100_jpg_37607_406_488.jpg' }, { 'name': 'Művészrózsa Maurice Utrillo', 'price': 1890, 'id': 6,
'description': 'Tűzpiros virágai fehéren csíkozottak. Mindehhez elragadó illat párosul.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/110840_jpg_91884_406_488.jpg' }, { 'name': 'Delikatess csüngőeper', 'price': 1380, 'id': 6, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Kert nélkül is folyamatosan szedheti a finom epreket! Használjon jó minőségű földet, megfelelő ' + 'méretű virágládát. Gondoskodjon a felesleges víz elvezetéséről! Napi rendszerességgel végzett öntözés és a ' + 'megfelelő tápanyagutánpótlás mellett rövid időn belül megjelennek az első termések.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/090060_jpg_535663_406_488.jpg' }, { 'name': 'Málna Boheme', 'price': 1090, 'id': 7, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Egy őszi érésű málna, amilyenre az ember vágyik. Szép színű, kellemes illatú, csodálatos ízű.' + ' Bogyói nagyon korán érnek. Felálló és kompakt növekedésű.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/151330_jpg_131481_406_488.jpg' }, { 'name': 'Málna Golden Everest', 'price': 1290, 'id': 8, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'A sárga termések teszik értékessé ezt a fajtát, mely különleges aromájával kiváló lekvárok, ' + 'édességek és gyümölcslé készítéséhez. Terméseit június közepén a második éves vesszők végén hozza.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/151240_jpg_273492_406_488.jpg' }, { 'name': 'Hortblue Petite áfonya', 'price': 1990, 'id': 9, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Hosszabbítsa meg a szüretelési időt egy közkedvelt bogyós fajtával! Sokan kedvelik az áfonyát akár ' + 'müzlibe, gyümölcssalátába vagy csak egyszerűen frissen a bokorról.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/154530_jpg_345266_406_488.jpg' }, { 'name': 'Vörös áfonya', 'price': 1690, 'id': 10, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Hosszabbítsa meg a szüretelési időt egy közkedvelt bogyós fajtával! Sokan kedvelik az áfonyát ' + 'akár müzlibe, gyümölcssalátába vagy csak egyszerűen frissen a bokorról.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/154520_jpg_402033_406_488.jpg' }, { 'name': 'Csokoládészőlő', 'price': 1890, 'id': 11, 'categoryId': 3, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Kínában és Japánban a szárított kérgét gyógyszerként használják. Gyulladáscsökkentő és a ' + 'menstruációs problémákat enyhítő hatása is ismert. Gyökere lázcsillapító hatású, termése egyes vélemények ' + 'szerint hatásos a daganatos betegségekkel szemben.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070535_jpg_51932_406_488.jpg' }, { 'name': 'Klemátisz ´Guernsey Cream´', 'price': 2190, 'id': 12, 'categoryId': 3, 'isFrostProof': false, 'lightReq': 'HALF_SHADY', 'description': 'Mint ahogy a neve is jelzi, az első krémszínű klemátisz! Különleges színű virágainak értékét egy ' + 'sötét háttér még inkább kiemeli. A virág zöld közepe a sötétsárga porzókkal egyedülálló szépséget kölcsönöz ' + 'a fajtának.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070770_jpg_431988_406_488.jpg' }, { 'name': 'Klemátisz \'Innocent Blush\'', 'price': 2990, 'id': 13, 'categoryId': 3, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Csodaszép, nosztalgikus, új nemesítésű fajta. A 2. éves hajtásokon fejlődnek ki nagy, telt virágai.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070735_jpg_228664_406_488.jpg' }, { 'name': 'Nyári lilaakác', 'price': 2590, 'id': 14, 'categoryId': 3, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Június - júliusban nyíló, rózsaszín - lilás virágai akár a 35 cm nagyságot is elérhetik. ' + 'Szereti a védett helyeket. Minél idősebb, annál szebb a virágzása. Edényben is nevelhető.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070820_jpg_611190_406_488.jpg' }, { 'name': 'Aloe Vera', 'price': 1790, 'id': 15, 'categoryId': 4, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Az Aloe vera az egyik legrégebbi gyógynövény a különféle külső és belső bajok gyógyítására. ' + 'Enyhíti a fájdalmat, segíti a sebek, sérülések gyorsabb gyógyulását. Vágjon le egy darab levelet, majd a ' + 'kipréselt levet kenje a sebre! A vaskos levelekből koktélok, italok is készülhetnek.', 'plantingTime': ['APRIL', 'MAY', 'JUNE', 'JULY', 'AUG', 'SEPT', 'OCT'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/102440_jpg_83473_406_488.jpg' }, { 'name': 'Homoktövis pár', 'price': 4150, 'id': 17, 'categoryId': 4, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Tudta Ön, hogy a homoktövis 10 - szer több C - vitamint tartalmaz mint a citrusfélék? ' + 'Egy igazi bomba az immunrendszernek, és egy hatásos fegyver a megfázások ellen.', 'plantingTime': ['APRIL', 'MAY', 'SEPT'], 'waterReq': 'HIGH', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/130583_jpg_109750_406_488.jpg' } ]; return {login, register, categories, plants}; } responseInterceptor(responseOptions: ResponseOptions, requestInfo: RequestInfo) { if (requestInfo.resourceUrl.includes('login') || requestInfo.resourceUrl.includes('register')) { responseOptions.body = {token: 'asdasd'}; responseOptions.status = 200; } else if (typeof responseOptions.body === 'object') { responseOptions.body = responseOptions.body ? (responseOptions.body as any).data : null; } if (requestInfo.resourceUrl.includes('categories') && requestInfo.req.url.includes('?')) { const result = this.categories.slice(Math.random() * 4); responseOptions.body = result; } return responseOptions; } }
'categoryId': 1, 'isFrostProof': true, 'lightReq': 'HALF_SHADY',
random_line_split
in-memory-plants-db-service.ts
import { InMemoryDbService, RequestInfo } from 'angular-in-memory-web-api'; import { ResponseOptions } from '@angular/http'; import { Category } from '../app/products/models/category'; export class InMemoryPlantsDbService implements InMemoryDbService { categories: Category[]; constructor() { this.categories = [ { 'id': 0, 'name': 'Balkonnövények', 'plantsCount': 3 }, { 'id': 1, 'name': 'Rózsák', 'plantsCount': 4 }, { 'id': 2, 'name': 'Bogyós gyülömcsök', 'plantsCount': 5 }, { 'id': 3, 'name': 'Kúszónövények', 'plantsCount': 4 }, { 'id': 4, 'name': 'Gyógyhatású növények', 'plantsCount': 2 } ]; } createDb() { const login = []; const register = []; const categories = this.categories; const plants = [ { 'name': 'Estrella White', 'price': 2650, 'id': 0, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'HALF_SHADY', 'description': 'Virágait erős szárakon a lomb felett hozza. Bőséges virágzás egészen őszig. Az extrém időjárási ' + 'viszonyokat is jól bírja, hajtásait a szél sem töri meg. A bőséges virágzáshoz elengedhetetlen a gyakori öntözés ' + 'és tápoldatozás. Az általunk ajánlott „Muskátli virágvarázs” kitűnően megfelel erre a célra.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'BIWEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100839_jpg_137965_406_488.jpg' }, { 'name': 'Álló muskátli', 'price': 1470, 'id': 1, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Minden évben újabb fajták érkeznek. A legszebb virágú és legjobb fajtákat válogattam össze ' + 'Önnek! Gazdagon virágoznak, nagyok a viráglabdák, szél- és időjárásállóak. Talán ezek a legfontosabb ' + 'tulajdonságok. Bizonyára elégedett lesz a képen látható fajtákkal.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100164_jpg_70579_406_488.jpg' }, { 'name': 'Csüngő begónia', 'price': 1390, 'id': 2, 'categoryId': 0, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Újdonság, ami nem olcsó, de annál értékesebb! Barátai irigyelni fogják a rengeteg virágért. ' + 'A 10 centiméteresre is megnövő fénylő virágok finom illatosak. Nem kell már az árnyékban sem nélkülöznie ' + 'a bőséges virágzást! Ilyen helyekre a begóniák ültetése a legjobb választás. Egészen őszig messziről fénylenek ' + 'virágai. Jó minőségű balkon virágföldet és rendszeres öntözést kíván.', 'plantingTime': ['APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100696_jpg_84843_406_488.jpg' }, { 'name': 'Tearózsa', 'price': 1390, 'id': 3, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Virágait erős szárakon a lomb felett hozza. Bőséges virágzás egészen őszig. Az extrém időjárási ' + 'viszonyokat is jól bírja, hajtásait a szél sem töri meg. A bőséges virágzáshoz elengedhetetlen a gyakori öntözés ' + 'és tápoldatozás. Az általunk ajánlott „Muskátli virágvarázs” kitűnően megfelel erre a célra.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/100839_jpg_137965_406_488.jpg' }, { 'name': 'Bokorrózsa Emilia Maria', 'price': 1590, 'id': 4, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Egy nagyon gazdagon és szinte folyamatosan virágzó bokorrózsa. Pinkszínű, tömvetelt virágai ' + 'csodálatos,graepfruit,iris és ibolya keverékére emlékeztető illatúak. Kisebb rácson felvezetve magasabbra is nevelhető. ', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/115150_jpg_56050_406_488.jpg' }, { 'name': 'Ágyásrózsa Julia Child', 'price': 1590, 'id': 5, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Ágyásrózsa. Egy száron több virágot is hoz, amik akár 12 cm nagyságúak, erős, aranysárga színűek. ' + 'Legértékesebb csodálatos illata, különösen reggel mikor az első napsugarak felmelegítik a virágokat.' + 'Ez a fajta nem hiányozhat egyetlen rózsakertből sem!', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/119100_jpg_37607_406_488.jpg' }, { 'name': 'Művészrózsa Maurice Utrillo', 'price': 1890, 'id': 6, 'categoryId': 1, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Tűzpiros virágai fehéren csíkozottak. Mindehhez elragadó illat párosul.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/110840_jpg_91884_406_488.jpg' }, { 'name': 'Delikatess csüngőeper', 'price': 1380, 'id': 6, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Kert nélkül is folyamatosan szedheti a finom epreket! Használjon jó minőségű földet, megfelelő ' + 'méretű virágládát. Gondoskodjon a felesleges víz elvezetéséről! Napi rendszerességgel végzett öntözés és a ' + 'megfelelő tápanyagutánpótlás mellett rövid időn belül megjelennek az első termések.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/090060_jpg_535663_406_488.jpg' }, { 'name': 'Málna Boheme', 'price': 1090, 'id': 7, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Egy őszi érésű málna, amilyenre az ember vágyik. Szép színű, kellemes illatú, csodálatos ízű.' + ' Bogyói nagyon korán érnek. Felálló és kompakt növekedésű.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/151330_jpg_131481_406_488.jpg' }, { 'name': 'Málna Golden Everest', 'price': 1290, 'id': 8, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'A sárga termések teszik értékessé ezt a fajtát, mely különleges aromájával kiváló lekvárok, ' + 'édességek és gyümölcslé készítéséhez. Terméseit június közepén a második éves vesszők végén hozza.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/151240_jpg_273492_406_488.jpg' }, { 'name': 'Hortblue Petite áfonya', 'price': 1990, 'id': 9, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Hosszabbítsa meg a szüretelési időt egy közkedvelt bogyós fajtával! Sokan kedvelik az áfonyát akár ' + 'müzlibe, gyümölcssalátába vagy csak egyszerűen frissen a bokorról.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/154530_jpg_345266_406_488.jpg' }, { 'name': 'Vörös áfonya', 'price': 1690, 'id': 10, 'categoryId': 2, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Hosszabbítsa meg a szüretelési időt egy közkedvelt bogyós fajtával! Sokan kedvelik az áfonyát ' + 'akár müzlibe, gyümölcssalátába vagy csak egyszerűen frissen a bokorról.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'MEDIUM', 'nutritionReq': 'NONE', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/154520_jpg_402033_406_488.jpg' }, { 'name': 'Csokoládészőlő', 'price': 1890, 'id': 11, 'categoryId': 3, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Kínában és Japánban a szárított kérgét gyógyszerként használják. Gyulladáscsökkentő és a ' + 'menstruációs problémákat enyhítő hatása is ismert. Gyökere lázcsillapító hatású, termése egyes vélemények ' + 'szerint hatásos a daganatos betegségekkel szemben.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'JUNE', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070535_jpg_51932_406_488.jpg' }, { 'name': 'Klemátisz ´Guernsey Cream´', 'price': 2190, 'id': 12, 'categoryId': 3, 'isFrostProof': false, 'lightReq': 'HALF_SHADY', 'description': 'Mint ahogy a neve is jelzi, az első krémszínű klemátisz! Különleges színű virágainak értékét egy ' + 'sötét háttér még inkább kiemeli. A virág zöld közepe a sötétsárga porzókkal egyedülálló szépséget kölcsönöz ' + 'a fajtának.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070770_jpg_431988_406_488.jpg' }, { 'name': 'Klemátisz \'Innocent Blush\'', 'price': 2990, 'id': 13, 'categoryId': 3, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Csodaszép, nosztalgikus, új nemesítésű fajta. A 2. éves hajtásokon fejlődnek ki nagy, telt virágai.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT'], 'waterReq': 'MEDIUM', 'nutritionReq': 'BIWEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070735_jpg_228664_406_488.jpg' }, { 'name': 'Nyári lilaakác', 'price': 2590, 'id': 14, 'categoryId': 3, 'isFrostProof': true, 'lightReq': 'SUNNY', 'description': 'Június - júliusban nyíló, rózsaszín - lilás virágai akár a 35 cm nagyságot is elérhetik. ' + 'Szereti a védett helyeket. Minél idősebb, annál szebb a virágzása. Edényben is nevelhető.', 'plantingTime': ['MARCH', 'APRIL', 'MAY', 'SEPT', 'OCT', 'NOV'], 'waterReq': 'LOW', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/070820_jpg_611190_406_488.jpg' }, { 'name': 'Aloe Vera', 'price': 1790, 'id': 15, 'categoryId': 4, 'isFrostProof': false, 'lightReq': 'SUNNY', 'description': 'Az Aloe vera az egyik legrégebbi gyógynövény a különféle külső és belső bajok gyógyítására. ' + 'Enyhíti a fájdalmat, segíti a sebek, sérülések gyorsabb gyógyulását. Vágjon le egy darab levelet, majd a ' + 'kipréselt levet kenje a sebre! A vaskos levelekből koktélok, italok is készülhetnek.', 'plantingTime': ['APRIL', 'MAY', 'JUNE', 'JULY', 'AUG', 'SEPT', 'OCT'], 'waterReq': 'HIGH', 'nutritionReq': 'WEEKLY', 'isFavorite': true, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/102440_jpg_83473_406_488.jpg' }, { 'name': 'Homoktövis pár', 'price': 4150, 'id': 17, 'categoryId': 4, 'isFrostProof': true, 'lightReq': 'HALF_SHADY', 'description': 'Tudta Ön, hogy a homoktövis 10 - szer több C - vitamint tartalmaz mint a citrusfélék? ' + 'Egy igazi bomba az immunrendszernek, és egy hatásos fegyver a megfázások ellen.', 'plantingTime': ['APRIL', 'MAY', 'SEPT'], 'waterReq': 'HIGH', 'nutritionReq': 'MONTHLY', 'isFavorite': false, 'imageUrl': 'https://www.starkl.hu/img/eshop/thumbnails/130583_jpg_109750_406_488.jpg' } ]; return {login, register, categories, plants}; } responseInterceptor(responseOptions: ResponseOptions, requestInfo: RequestInfo) { if (requestInfo.resourceUrl.includes('login') || requestInfo.resourceUrl.includes('register')) { responseOptions.body = {token: 'asdasd'}; responseOptions.status = 200; } else if (typeof responseOptions.body === 'object') { responseOptions.body = responseOptions.body ? (responseOptions.body as any).data : null; } if (requestInfo.resourceUrl.includes('categories') && requestInfo.req.url.includes('?')) { const result = this.categories.slice(Math.random() * 4); responseOptions.body = result; } return responseOptions; } }
conditional_block
piv.rs
// Copyright 2021 Gravitational, Inc // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::errors::invalid_data_error; use iso7816::aid::Aid; use iso7816::command::instruction::Instruction; use iso7816::command::Command; use iso7816::response::Status; use iso7816_tlv::ber::{Tag, Tlv, Value}; use rdp::model::error::*; use rsa::pkcs1::DecodeRsaPrivateKey; use rsa::traits::{PrivateKeyParts, PublicKeyParts}; use rsa::{BigUint, RsaPrivateKey}; use std::convert::TryFrom; use std::fmt::Write as _; use std::io::{Cursor, Read}; use uuid::Uuid; // AID (Application ID) of PIV application, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf const PIV_AID: Aid = Aid::new_truncatable( &[ 0xA0, 0x00, 0x00, 0x03, 0x08, 0x00, 0x00, 0x10, 0x00, 0x01, 0x00, ], 5, // usually truncates to first 5 bytes ); // Card implements a PIV-compatible smartcard, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf #[derive(Debug, PartialEq, Eq)] pub struct Card<const S: usize> { // Card-holder user ID (CHUID). In federal agencies, this value would be unique per employee // and encodes some agency information. In our case it's static. chuid: Vec<u8>, piv_auth_cert: Vec<u8>, piv_auth_key: RsaPrivateKey, pin: String, // Pending command and response to receive/send over multiple messages when // they don't fit into one. pending_command: Option<Command<S>>, pending_response: Option<Cursor<Vec<u8>>>, } impl<const S: usize> Card<S> { pub fn new(uuid: Uuid, cert_der: &[u8], key_der: &[u8], pin: String) -> RdpResult<Self> { let piv_auth_key = RsaPrivateKey::from_pkcs1_der(key_der).map_err(|e| { invalid_data_error(&format!("failed to parse private key from DER: {e:?}")) })?; Ok(Self { chuid: Self::build_chuid(uuid), piv_auth_cert: Self::build_piv_auth_cert(cert_der), piv_auth_key, pin, pending_command: None, pending_response: None, }) } pub fn handle(&mut self, cmd: Command<S>) -> RdpResult<Response> { debug!("got command: {:?}", cmd); debug!("command data: {}", hex_data(&cmd)); // Handle chained commands. let cmd = match self.pending_command.as_mut() { None => cmd, Some(pending) => { pending .extend_from_command(&cmd) .map_err(|_| invalid_data_error("could not build chained command"))?; pending.clone() } }; if cmd.class().chain().not_the_last() { self.pending_command = Some(cmd); return Ok(Response::new(Status::Success)); } else { self.pending_command = None; } let resp = match cmd.instruction() { Instruction::Select => self.handle_select(cmd), Instruction::Verify => self.handle_verify(cmd), Instruction::GetData => self.handle_get_data(cmd), Instruction::GetResponse => self.handle_get_response(cmd), Instruction::GeneralAuthenticate => self.handle_general_authenticate(cmd), _ => { warn!("unimplemented instruction {:?}", cmd.instruction()); Ok(Response::new(Status::InstructionNotSupportedOrInvalid)) } }?; debug!("send response: {:?}", resp); debug!("response data: {}", to_hex(&resp.encode())); Ok(resp) } fn handle_select(&mut self, cmd: Command<S>) -> RdpResult<Response> { // For our use case, we only allow selecting the PIV application on the smartcard. // // P1=04 and P2=00 means selection of DF (usually) application by name. Everything else not // supported. if cmd.p1 != 0x04 && cmd.p2 != 0x00 { return Ok(Response::new(Status::NotFound)); } if !PIV_AID.matches(cmd.data()) { return Ok(Response::new(Status::NotFound)); } // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.1 let resp = tlv( TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE, Value::Constructed(vec![ tlv( TLV_TAG_AID, Value::Primitive(vec![0x00, 0x00, 0x10, 0x00, 0x01, 0x00]), )?, tlv( TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY, Value::Constructed(vec![tlv( TLV_TAG_AID, Value::Primitive(PIV_AID.truncated().to_vec()), )?]), )?, ]), )?; Ok(Response::with_data(Status::Success, resp.to_vec())) } fn handle_verify(&mut self, cmd: Command<S>) -> RdpResult<Response> { if cmd.data() == self.pin.as_bytes() { Ok(Response::new(Status::Success)) } else { warn!("PIN mismatch, want {}, got {:?}", self.pin, cmd.data()); Ok(Response::new(Status::VerificationFailed)) } } fn handle_get_data(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.2. if cmd.p1 != 0x3F && cmd.p2 != 0xFF { return Ok(Response::new(Status::NotFound)); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(0x5C)? { return Ok(Response::new(Status::NotFound)); } match request_tlv.value() { Value::Primitive(tag) => match to_hex(tag).as_str() { // Card Holder Unique Identifier. "5FC102" => Ok(Response::with_data(Status::Success, self.chuid.clone())), // X.509 Certificate for PIV Authentication "5FC105" => { self.pending_response = Some(Cursor::new(self.piv_auth_cert.clone())); self.handle_get_response(cmd) } _ => { // Some other unimplemented data object. Ok(Response::new(Status::NotFound)) } }, Value::Constructed(_) => Ok(Response::new(Status::NotFound)), } } fn handle_get_response(&mut self, _cmd: Command<S>) -> RdpResult<Response> { // CHINK_SIZE is the max response data size in bytes, without resorting to "extended" // messages. const CHUNK_SIZE: usize = 256; match &mut self.pending_response { None => Ok(Response::new(Status::NotFound)), Some(cursor) => { let mut chunk = [0; CHUNK_SIZE]; let n = cursor.read(&mut chunk)?; let mut chunk = chunk.to_vec(); chunk.truncate(n); let remaining = cursor.get_ref().len() as u64 - cursor.position(); let status = if remaining == 0 { Status::Success } else if remaining < CHUNK_SIZE as u64 { Status::MoreAvailable(remaining as u8) } else { Status::MoreAvailable(0) }; Ok(Response::with_data(status, chunk)) } } } /// Sign the challenge. /// /// Note: for signatures, typically you'd use a signer that hashes the input data, adds padding /// according to some scheme (like PKCS1v15 or PSS) and then "decrypts" this data with the key. /// The decrypted blob is the signature. /// /// In our case, the RDP server does the hashing and padding, and only gives us a finished blob /// to decrypt. Most crypto libraries don't directly expose RSA decryption without padding, as /// it's easy to build insecure crypto systems. Thankfully for us, this decryption is just a single /// modpow operation which is suppored by RustCrypto. fn sign_auth_challenge(&self, challenge: &[u8]) -> Vec<u8> { let c = BigUint::from_bytes_be(challenge); let plain_text = c .modpow(self.piv_auth_key.d(), self.piv_auth_key.n()) .to_bytes_be(); let mut result = vec![0u8; self.piv_auth_key.size()]; let start = result.len() - plain_text.len(); result[start..].copy_from_slice(&plain_text); result } fn handle_general_authenticate(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See section 3.2.4 and example in Appending A.3 from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // P1='07' means 2048-bit RSA. // // TODO(zmb3): compare algorithm against the private key using consts from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-78-4.pdf // TODO(zmb3): support non-RSA keys, if needed. if cmd.p1 != 0x07 { return Err(invalid_data_error(&format!( "unsupported algorithm identifier P1:{:#X} in general authenticate command", cmd.p1 ))); } // P2='9A' means PIV Authentication Key (matches our cert '5FC105' in handle_get_data). if cmd.p2 != 0x9A { return Err(invalid_data_error(&format!( "unsupported key reference P2:{:#X} in general authenticate command", cmd.p2 ))); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE)? { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } // Extract the challenge field. let request_tlvs = match request_tlv.value() { Value::Primitive(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } Value::Constructed(tlvs) => tlvs, }; let mut challenge = None; for data in request_tlvs { if *data.tag() != tlv_tag(TLV_TAG_CHALLENGE)? { continue; } challenge = match data.value() { Value::Primitive(chal) => Some(chal), Value::Constructed(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } }; } let challenge = challenge.ok_or_else(|| { invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}, missing challenge data" )) })?; // TODO(zmb3): support non-RSA keys, if needed. let signed_challenge = self.sign_auth_challenge(challenge); // Return signed challenge. let resp = tlv( TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE, Value::Constructed(vec![tlv( TLV_TAG_RESPONSE, Value::Primitive(signed_challenge), )?]), )? .to_vec(); self.pending_response = Some(Cursor::new(resp)); self.handle_get_response(cmd) } fn build_chuid(uuid: Uuid) -> Vec<u8> { // This is gross: the response is a BER-TLV value, but it has nested SIMPLE-TLV // values. None of the TLV encoding libraries out there support this, they fail // when checking the tag of nested TLVs. // // So, construct the TLV by hand from raw bytes. Hopefully the comments will be // enough to explain the structure. // // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // table 9 has the explanation of fields. // // Start with a top-level BER-TLV tag and length: let mut resp = vec![TLV_TAG_DATA_FIELD, 0x3B]; // TLV tag and length for FASC-N. resp.extend_from_slice(&[TLV_TAG_FASC_N, 0x19]); // FASC-N value containing S9999F9999F999999F0F1F0000000000300001E, with a // weird encoding from section 6 of: // https://www.idmanagement.gov/docs/pacs-tig-scepacs.pdf resp.extend_from_slice(&[ 0xd4, 0xe7, 0x39, 0xda, 0x73, 0x9c, 0xed, 0x39, 0xce, 0x73, 0x9d, 0x83, 0x68, 0x58, 0x21, 0x08, 0x42, 0x10, 0x84, 0x21, 0xc8, 0x42, 0x10, 0xc3, 0xeb, ]); // TLV for user UUID. resp.extend_from_slice(&[TLV_TAG_GUID, 0x10]); resp.extend_from_slice(uuid.as_bytes()); // TLV for expiration date (YYYYMMDD). resp.extend_from_slice(&[TLV_TAG_EXPIRATION_DATE, 0x08]); // TODO(awly): generate this from current time. resp.extend_from_slice("20300101".as_bytes()); // TLV for signature (empty). resp.extend_from_slice(&[TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); resp } fn build_piv_auth_cert(cert_der: &[u8]) -> Vec<u8> { // Same as above, tags in this BER-TLV value are not compatible with the spec // and existing libraries. Marshal by hand. // // Certificate TLV tag and length. let mut resp = vec![TLV_TAG_CERTIFICATE]; resp.extend_from_slice(&len_to_vec(cert_der.len())); resp.extend_from_slice(cert_der); // CertInfo TLV (0x00 indicates uncompressed cert). resp.extend_from_slice(&[TLV_TAG_CERTINFO, 0x01, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); // Wrap with top-level TLV tag and length. let mut resp_outer = vec![TLV_TAG_DATA_FIELD]; resp_outer.extend_from_slice(&len_to_vec(resp.len())); resp_outer.extend_from_slice(&resp); resp_outer } } #[derive(Debug)] pub struct Response { data: Option<Vec<u8>>, status: Status, } impl Response { fn new(status: Status) -> Self { Self { data: None, status } } fn with_data(status: Status, data: Vec<u8>) -> Self { Self { data: Some(data), status, } } pub fn encode(&self) -> Vec<u8> { let mut buf = Vec::new(); if let Some(data) = &self.data { buf.extend_from_slice(data); } let status: [u8; 2] = self.status.into(); buf.extend_from_slice(&status); buf } } // SELECT command tags. const TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE: u8 = 0x61; const TLV_TAG_AID: u8 = 0x4F; const TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY: u8 = 0x79; const TLV_TAG_DATA_FIELD: u8 = 0x53; const TLV_TAG_FASC_N: u8 = 0x30; const TLV_TAG_GUID: u8 = 0x34; const TLV_TAG_EXPIRATION_DATE: u8 = 0x35; const TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE: u8 = 0x3E; const TLV_TAG_ERROR_DETECTION_CODE: u8 = 0xFE; const TLV_TAG_CERTIFICATE: u8 = 0x70; const TLV_TAG_CERTINFO: u8 = 0x71; // GENERAL AUTHENTICATE command tags. const TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE: u8 = 0x7C; const TLV_TAG_CHALLENGE: u8 = 0x81; const TLV_TAG_RESPONSE: u8 = 0x82; fn tlv(tag: u8, value: Value) -> RdpResult<Tlv> { Tlv::new(tlv_tag(tag)?, value) .map_err(|e| invalid_data_error(&format!("TLV with tag {tag:#X} invalid: {e:?}"))) } fn tlv_tag(val: u8) -> RdpResult<Tag> { Tag::try_from(val).map_err(|e| invalid_data_error(&format!("TLV tag {val:#X} invalid: {e:?}"))) } fn hex_data<const S: usize>(cmd: &Command<S>) -> String { to_hex(cmd.data()) } fn to_hex(bytes: &[u8]) -> String { let mut s = String::new(); for b in bytes { // https://rust-lang.github.io/rust-clippy/master/index.html#format_push_string let _ = write!(s, "{b:02X}"); } s } #[allow(clippy::cast_possible_truncation)] fn len_to_vec(len: usize) -> Vec<u8> { if len < 0x7f
else { let mut ret: Vec<u8> = len .to_be_bytes() .iter() .skip_while(|&x| *x == 0) .cloned() .collect(); ret.insert(0, 0x80 | ret.len() as u8); ret } }
{ vec![len as u8] }
conditional_block
piv.rs
// Copyright 2021 Gravitational, Inc // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::errors::invalid_data_error; use iso7816::aid::Aid; use iso7816::command::instruction::Instruction; use iso7816::command::Command; use iso7816::response::Status; use iso7816_tlv::ber::{Tag, Tlv, Value}; use rdp::model::error::*; use rsa::pkcs1::DecodeRsaPrivateKey; use rsa::traits::{PrivateKeyParts, PublicKeyParts}; use rsa::{BigUint, RsaPrivateKey}; use std::convert::TryFrom; use std::fmt::Write as _; use std::io::{Cursor, Read}; use uuid::Uuid; // AID (Application ID) of PIV application, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf const PIV_AID: Aid = Aid::new_truncatable( &[ 0xA0, 0x00, 0x00, 0x03, 0x08, 0x00, 0x00, 0x10, 0x00, 0x01, 0x00, ], 5, // usually truncates to first 5 bytes ); // Card implements a PIV-compatible smartcard, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf #[derive(Debug, PartialEq, Eq)] pub struct Card<const S: usize> { // Card-holder user ID (CHUID). In federal agencies, this value would be unique per employee // and encodes some agency information. In our case it's static. chuid: Vec<u8>, piv_auth_cert: Vec<u8>, piv_auth_key: RsaPrivateKey, pin: String, // Pending command and response to receive/send over multiple messages when // they don't fit into one. pending_command: Option<Command<S>>, pending_response: Option<Cursor<Vec<u8>>>, } impl<const S: usize> Card<S> { pub fn new(uuid: Uuid, cert_der: &[u8], key_der: &[u8], pin: String) -> RdpResult<Self> { let piv_auth_key = RsaPrivateKey::from_pkcs1_der(key_der).map_err(|e| { invalid_data_error(&format!("failed to parse private key from DER: {e:?}")) })?; Ok(Self { chuid: Self::build_chuid(uuid), piv_auth_cert: Self::build_piv_auth_cert(cert_der), piv_auth_key, pin, pending_command: None, pending_response: None, }) } pub fn handle(&mut self, cmd: Command<S>) -> RdpResult<Response> { debug!("got command: {:?}", cmd); debug!("command data: {}", hex_data(&cmd)); // Handle chained commands. let cmd = match self.pending_command.as_mut() { None => cmd, Some(pending) => { pending .extend_from_command(&cmd) .map_err(|_| invalid_data_error("could not build chained command"))?; pending.clone() } }; if cmd.class().chain().not_the_last() { self.pending_command = Some(cmd); return Ok(Response::new(Status::Success)); } else { self.pending_command = None; } let resp = match cmd.instruction() { Instruction::Select => self.handle_select(cmd), Instruction::Verify => self.handle_verify(cmd), Instruction::GetData => self.handle_get_data(cmd), Instruction::GetResponse => self.handle_get_response(cmd), Instruction::GeneralAuthenticate => self.handle_general_authenticate(cmd), _ => { warn!("unimplemented instruction {:?}", cmd.instruction()); Ok(Response::new(Status::InstructionNotSupportedOrInvalid)) } }?; debug!("send response: {:?}", resp); debug!("response data: {}", to_hex(&resp.encode())); Ok(resp) } fn handle_select(&mut self, cmd: Command<S>) -> RdpResult<Response> { // For our use case, we only allow selecting the PIV application on the smartcard. // // P1=04 and P2=00 means selection of DF (usually) application by name. Everything else not // supported. if cmd.p1 != 0x04 && cmd.p2 != 0x00 { return Ok(Response::new(Status::NotFound)); } if !PIV_AID.matches(cmd.data()) { return Ok(Response::new(Status::NotFound)); } // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.1 let resp = tlv( TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE, Value::Constructed(vec![ tlv( TLV_TAG_AID, Value::Primitive(vec![0x00, 0x00, 0x10, 0x00, 0x01, 0x00]), )?, tlv( TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY, Value::Constructed(vec![tlv( TLV_TAG_AID, Value::Primitive(PIV_AID.truncated().to_vec()), )?]), )?, ]), )?; Ok(Response::with_data(Status::Success, resp.to_vec())) } fn handle_verify(&mut self, cmd: Command<S>) -> RdpResult<Response> { if cmd.data() == self.pin.as_bytes() { Ok(Response::new(Status::Success)) } else { warn!("PIN mismatch, want {}, got {:?}", self.pin, cmd.data()); Ok(Response::new(Status::VerificationFailed)) } } fn handle_get_data(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.2. if cmd.p1 != 0x3F && cmd.p2 != 0xFF { return Ok(Response::new(Status::NotFound)); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(0x5C)? { return Ok(Response::new(Status::NotFound)); } match request_tlv.value() { Value::Primitive(tag) => match to_hex(tag).as_str() { // Card Holder Unique Identifier. "5FC102" => Ok(Response::with_data(Status::Success, self.chuid.clone())), // X.509 Certificate for PIV Authentication "5FC105" => { self.pending_response = Some(Cursor::new(self.piv_auth_cert.clone())); self.handle_get_response(cmd) } _ => { // Some other unimplemented data object. Ok(Response::new(Status::NotFound)) } }, Value::Constructed(_) => Ok(Response::new(Status::NotFound)), } } fn handle_get_response(&mut self, _cmd: Command<S>) -> RdpResult<Response> { // CHINK_SIZE is the max response data size in bytes, without resorting to "extended" // messages. const CHUNK_SIZE: usize = 256; match &mut self.pending_response { None => Ok(Response::new(Status::NotFound)), Some(cursor) => { let mut chunk = [0; CHUNK_SIZE]; let n = cursor.read(&mut chunk)?; let mut chunk = chunk.to_vec(); chunk.truncate(n); let remaining = cursor.get_ref().len() as u64 - cursor.position(); let status = if remaining == 0 { Status::Success } else if remaining < CHUNK_SIZE as u64 { Status::MoreAvailable(remaining as u8) } else { Status::MoreAvailable(0) }; Ok(Response::with_data(status, chunk)) } } } /// Sign the challenge. /// /// Note: for signatures, typically you'd use a signer that hashes the input data, adds padding /// according to some scheme (like PKCS1v15 or PSS) and then "decrypts" this data with the key. /// The decrypted blob is the signature. /// /// In our case, the RDP server does the hashing and padding, and only gives us a finished blob /// to decrypt. Most crypto libraries don't directly expose RSA decryption without padding, as /// it's easy to build insecure crypto systems. Thankfully for us, this decryption is just a single /// modpow operation which is suppored by RustCrypto. fn sign_auth_challenge(&self, challenge: &[u8]) -> Vec<u8> { let c = BigUint::from_bytes_be(challenge); let plain_text = c .modpow(self.piv_auth_key.d(), self.piv_auth_key.n()) .to_bytes_be(); let mut result = vec![0u8; self.piv_auth_key.size()]; let start = result.len() - plain_text.len(); result[start..].copy_from_slice(&plain_text); result } fn handle_general_authenticate(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See section 3.2.4 and example in Appending A.3 from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // P1='07' means 2048-bit RSA. // // TODO(zmb3): compare algorithm against the private key using consts from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-78-4.pdf // TODO(zmb3): support non-RSA keys, if needed. if cmd.p1 != 0x07 { return Err(invalid_data_error(&format!( "unsupported algorithm identifier P1:{:#X} in general authenticate command", cmd.p1 ))); } // P2='9A' means PIV Authentication Key (matches our cert '5FC105' in handle_get_data). if cmd.p2 != 0x9A { return Err(invalid_data_error(&format!( "unsupported key reference P2:{:#X} in general authenticate command", cmd.p2 ))); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE)? { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } // Extract the challenge field. let request_tlvs = match request_tlv.value() { Value::Primitive(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } Value::Constructed(tlvs) => tlvs, }; let mut challenge = None; for data in request_tlvs { if *data.tag() != tlv_tag(TLV_TAG_CHALLENGE)? { continue; } challenge = match data.value() { Value::Primitive(chal) => Some(chal), Value::Constructed(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } }; } let challenge = challenge.ok_or_else(|| { invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}, missing challenge data" )) })?; // TODO(zmb3): support non-RSA keys, if needed. let signed_challenge = self.sign_auth_challenge(challenge); // Return signed challenge. let resp = tlv( TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE, Value::Constructed(vec![tlv( TLV_TAG_RESPONSE, Value::Primitive(signed_challenge), )?]), )? .to_vec(); self.pending_response = Some(Cursor::new(resp)); self.handle_get_response(cmd) } fn build_chuid(uuid: Uuid) -> Vec<u8> { // This is gross: the response is a BER-TLV value, but it has nested SIMPLE-TLV // values. None of the TLV encoding libraries out there support this, they fail // when checking the tag of nested TLVs. // // So, construct the TLV by hand from raw bytes. Hopefully the comments will be // enough to explain the structure. // // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // table 9 has the explanation of fields. // // Start with a top-level BER-TLV tag and length: let mut resp = vec![TLV_TAG_DATA_FIELD, 0x3B]; // TLV tag and length for FASC-N. resp.extend_from_slice(&[TLV_TAG_FASC_N, 0x19]); // FASC-N value containing S9999F9999F999999F0F1F0000000000300001E, with a // weird encoding from section 6 of: // https://www.idmanagement.gov/docs/pacs-tig-scepacs.pdf resp.extend_from_slice(&[ 0xd4, 0xe7, 0x39, 0xda, 0x73, 0x9c, 0xed, 0x39, 0xce, 0x73, 0x9d, 0x83, 0x68, 0x58, 0x21, 0x08, 0x42, 0x10, 0x84, 0x21, 0xc8, 0x42, 0x10, 0xc3, 0xeb, ]); // TLV for user UUID. resp.extend_from_slice(&[TLV_TAG_GUID, 0x10]); resp.extend_from_slice(uuid.as_bytes()); // TLV for expiration date (YYYYMMDD). resp.extend_from_slice(&[TLV_TAG_EXPIRATION_DATE, 0x08]); // TODO(awly): generate this from current time. resp.extend_from_slice("20300101".as_bytes()); // TLV for signature (empty). resp.extend_from_slice(&[TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); resp } fn build_piv_auth_cert(cert_der: &[u8]) -> Vec<u8> { // Same as above, tags in this BER-TLV value are not compatible with the spec // and existing libraries. Marshal by hand. // // Certificate TLV tag and length. let mut resp = vec![TLV_TAG_CERTIFICATE]; resp.extend_from_slice(&len_to_vec(cert_der.len())); resp.extend_from_slice(cert_der); // CertInfo TLV (0x00 indicates uncompressed cert). resp.extend_from_slice(&[TLV_TAG_CERTINFO, 0x01, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); // Wrap with top-level TLV tag and length. let mut resp_outer = vec![TLV_TAG_DATA_FIELD]; resp_outer.extend_from_slice(&len_to_vec(resp.len())); resp_outer.extend_from_slice(&resp); resp_outer } } #[derive(Debug)] pub struct Response { data: Option<Vec<u8>>, status: Status, } impl Response { fn new(status: Status) -> Self { Self { data: None, status } } fn with_data(status: Status, data: Vec<u8>) -> Self { Self { data: Some(data), status, } } pub fn encode(&self) -> Vec<u8>
} // SELECT command tags. const TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE: u8 = 0x61; const TLV_TAG_AID: u8 = 0x4F; const TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY: u8 = 0x79; const TLV_TAG_DATA_FIELD: u8 = 0x53; const TLV_TAG_FASC_N: u8 = 0x30; const TLV_TAG_GUID: u8 = 0x34; const TLV_TAG_EXPIRATION_DATE: u8 = 0x35; const TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE: u8 = 0x3E; const TLV_TAG_ERROR_DETECTION_CODE: u8 = 0xFE; const TLV_TAG_CERTIFICATE: u8 = 0x70; const TLV_TAG_CERTINFO: u8 = 0x71; // GENERAL AUTHENTICATE command tags. const TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE: u8 = 0x7C; const TLV_TAG_CHALLENGE: u8 = 0x81; const TLV_TAG_RESPONSE: u8 = 0x82; fn tlv(tag: u8, value: Value) -> RdpResult<Tlv> { Tlv::new(tlv_tag(tag)?, value) .map_err(|e| invalid_data_error(&format!("TLV with tag {tag:#X} invalid: {e:?}"))) } fn tlv_tag(val: u8) -> RdpResult<Tag> { Tag::try_from(val).map_err(|e| invalid_data_error(&format!("TLV tag {val:#X} invalid: {e:?}"))) } fn hex_data<const S: usize>(cmd: &Command<S>) -> String { to_hex(cmd.data()) } fn to_hex(bytes: &[u8]) -> String { let mut s = String::new(); for b in bytes { // https://rust-lang.github.io/rust-clippy/master/index.html#format_push_string let _ = write!(s, "{b:02X}"); } s } #[allow(clippy::cast_possible_truncation)] fn len_to_vec(len: usize) -> Vec<u8> { if len < 0x7f { vec![len as u8] } else { let mut ret: Vec<u8> = len .to_be_bytes() .iter() .skip_while(|&x| *x == 0) .cloned() .collect(); ret.insert(0, 0x80 | ret.len() as u8); ret } }
{ let mut buf = Vec::new(); if let Some(data) = &self.data { buf.extend_from_slice(data); } let status: [u8; 2] = self.status.into(); buf.extend_from_slice(&status); buf }
identifier_body
piv.rs
// Copyright 2021 Gravitational, Inc // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::errors::invalid_data_error; use iso7816::aid::Aid; use iso7816::command::instruction::Instruction; use iso7816::command::Command; use iso7816::response::Status; use iso7816_tlv::ber::{Tag, Tlv, Value}; use rdp::model::error::*; use rsa::pkcs1::DecodeRsaPrivateKey; use rsa::traits::{PrivateKeyParts, PublicKeyParts}; use rsa::{BigUint, RsaPrivateKey}; use std::convert::TryFrom; use std::fmt::Write as _; use std::io::{Cursor, Read}; use uuid::Uuid; // AID (Application ID) of PIV application, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf const PIV_AID: Aid = Aid::new_truncatable( &[ 0xA0, 0x00, 0x00, 0x03, 0x08, 0x00, 0x00, 0x10, 0x00, 0x01, 0x00, ], 5, // usually truncates to first 5 bytes ); // Card implements a PIV-compatible smartcard, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf #[derive(Debug, PartialEq, Eq)] pub struct Card<const S: usize> { // Card-holder user ID (CHUID). In federal agencies, this value would be unique per employee // and encodes some agency information. In our case it's static. chuid: Vec<u8>, piv_auth_cert: Vec<u8>, piv_auth_key: RsaPrivateKey, pin: String, // Pending command and response to receive/send over multiple messages when // they don't fit into one. pending_command: Option<Command<S>>, pending_response: Option<Cursor<Vec<u8>>>, } impl<const S: usize> Card<S> { pub fn new(uuid: Uuid, cert_der: &[u8], key_der: &[u8], pin: String) -> RdpResult<Self> { let piv_auth_key = RsaPrivateKey::from_pkcs1_der(key_der).map_err(|e| { invalid_data_error(&format!("failed to parse private key from DER: {e:?}")) })?; Ok(Self { chuid: Self::build_chuid(uuid), piv_auth_cert: Self::build_piv_auth_cert(cert_der), piv_auth_key, pin, pending_command: None, pending_response: None, }) } pub fn handle(&mut self, cmd: Command<S>) -> RdpResult<Response> { debug!("got command: {:?}", cmd); debug!("command data: {}", hex_data(&cmd)); // Handle chained commands. let cmd = match self.pending_command.as_mut() { None => cmd, Some(pending) => { pending .extend_from_command(&cmd) .map_err(|_| invalid_data_error("could not build chained command"))?; pending.clone() } }; if cmd.class().chain().not_the_last() { self.pending_command = Some(cmd); return Ok(Response::new(Status::Success)); } else { self.pending_command = None; } let resp = match cmd.instruction() { Instruction::Select => self.handle_select(cmd), Instruction::Verify => self.handle_verify(cmd), Instruction::GetData => self.handle_get_data(cmd), Instruction::GetResponse => self.handle_get_response(cmd), Instruction::GeneralAuthenticate => self.handle_general_authenticate(cmd), _ => { warn!("unimplemented instruction {:?}", cmd.instruction()); Ok(Response::new(Status::InstructionNotSupportedOrInvalid)) } }?; debug!("send response: {:?}", resp); debug!("response data: {}", to_hex(&resp.encode())); Ok(resp) } fn handle_select(&mut self, cmd: Command<S>) -> RdpResult<Response> { // For our use case, we only allow selecting the PIV application on the smartcard. // // P1=04 and P2=00 means selection of DF (usually) application by name. Everything else not // supported. if cmd.p1 != 0x04 && cmd.p2 != 0x00 { return Ok(Response::new(Status::NotFound)); } if !PIV_AID.matches(cmd.data()) { return Ok(Response::new(Status::NotFound)); } // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.1 let resp = tlv( TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE, Value::Constructed(vec![ tlv( TLV_TAG_AID, Value::Primitive(vec![0x00, 0x00, 0x10, 0x00, 0x01, 0x00]), )?, tlv( TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY, Value::Constructed(vec![tlv( TLV_TAG_AID, Value::Primitive(PIV_AID.truncated().to_vec()), )?]), )?, ]), )?; Ok(Response::with_data(Status::Success, resp.to_vec())) } fn handle_verify(&mut self, cmd: Command<S>) -> RdpResult<Response> { if cmd.data() == self.pin.as_bytes() { Ok(Response::new(Status::Success)) } else { warn!("PIN mismatch, want {}, got {:?}", self.pin, cmd.data()); Ok(Response::new(Status::VerificationFailed)) } } fn handle_get_data(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.2. if cmd.p1 != 0x3F && cmd.p2 != 0xFF { return Ok(Response::new(Status::NotFound)); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(0x5C)? { return Ok(Response::new(Status::NotFound)); } match request_tlv.value() { Value::Primitive(tag) => match to_hex(tag).as_str() { // Card Holder Unique Identifier. "5FC102" => Ok(Response::with_data(Status::Success, self.chuid.clone())), // X.509 Certificate for PIV Authentication "5FC105" => { self.pending_response = Some(Cursor::new(self.piv_auth_cert.clone())); self.handle_get_response(cmd) } _ => { // Some other unimplemented data object. Ok(Response::new(Status::NotFound)) } }, Value::Constructed(_) => Ok(Response::new(Status::NotFound)), } } fn handle_get_response(&mut self, _cmd: Command<S>) -> RdpResult<Response> { // CHINK_SIZE is the max response data size in bytes, without resorting to "extended" // messages. const CHUNK_SIZE: usize = 256; match &mut self.pending_response { None => Ok(Response::new(Status::NotFound)), Some(cursor) => { let mut chunk = [0; CHUNK_SIZE]; let n = cursor.read(&mut chunk)?; let mut chunk = chunk.to_vec(); chunk.truncate(n); let remaining = cursor.get_ref().len() as u64 - cursor.position(); let status = if remaining == 0 { Status::Success } else if remaining < CHUNK_SIZE as u64 { Status::MoreAvailable(remaining as u8) } else { Status::MoreAvailable(0) }; Ok(Response::with_data(status, chunk)) } } } /// Sign the challenge. /// /// Note: for signatures, typically you'd use a signer that hashes the input data, adds padding /// according to some scheme (like PKCS1v15 or PSS) and then "decrypts" this data with the key. /// The decrypted blob is the signature. /// /// In our case, the RDP server does the hashing and padding, and only gives us a finished blob /// to decrypt. Most crypto libraries don't directly expose RSA decryption without padding, as /// it's easy to build insecure crypto systems. Thankfully for us, this decryption is just a single /// modpow operation which is suppored by RustCrypto. fn sign_auth_challenge(&self, challenge: &[u8]) -> Vec<u8> { let c = BigUint::from_bytes_be(challenge); let plain_text = c .modpow(self.piv_auth_key.d(), self.piv_auth_key.n()) .to_bytes_be(); let mut result = vec![0u8; self.piv_auth_key.size()]; let start = result.len() - plain_text.len(); result[start..].copy_from_slice(&plain_text); result } fn handle_general_authenticate(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See section 3.2.4 and example in Appending A.3 from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // P1='07' means 2048-bit RSA. // // TODO(zmb3): compare algorithm against the private key using consts from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-78-4.pdf // TODO(zmb3): support non-RSA keys, if needed. if cmd.p1 != 0x07 { return Err(invalid_data_error(&format!( "unsupported algorithm identifier P1:{:#X} in general authenticate command", cmd.p1 ))); } // P2='9A' means PIV Authentication Key (matches our cert '5FC105' in handle_get_data). if cmd.p2 != 0x9A { return Err(invalid_data_error(&format!( "unsupported key reference P2:{:#X} in general authenticate command", cmd.p2 ))); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE)? { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } // Extract the challenge field. let request_tlvs = match request_tlv.value() { Value::Primitive(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } Value::Constructed(tlvs) => tlvs, }; let mut challenge = None; for data in request_tlvs { if *data.tag() != tlv_tag(TLV_TAG_CHALLENGE)? { continue; } challenge = match data.value() { Value::Primitive(chal) => Some(chal), Value::Constructed(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } }; } let challenge = challenge.ok_or_else(|| { invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}, missing challenge data" )) })?; // TODO(zmb3): support non-RSA keys, if needed. let signed_challenge = self.sign_auth_challenge(challenge); // Return signed challenge. let resp = tlv( TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE, Value::Constructed(vec![tlv( TLV_TAG_RESPONSE, Value::Primitive(signed_challenge), )?]), )? .to_vec(); self.pending_response = Some(Cursor::new(resp)); self.handle_get_response(cmd) } fn build_chuid(uuid: Uuid) -> Vec<u8> { // This is gross: the response is a BER-TLV value, but it has nested SIMPLE-TLV // values. None of the TLV encoding libraries out there support this, they fail // when checking the tag of nested TLVs. // // So, construct the TLV by hand from raw bytes. Hopefully the comments will be // enough to explain the structure. // // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // table 9 has the explanation of fields. // // Start with a top-level BER-TLV tag and length: let mut resp = vec![TLV_TAG_DATA_FIELD, 0x3B]; // TLV tag and length for FASC-N. resp.extend_from_slice(&[TLV_TAG_FASC_N, 0x19]); // FASC-N value containing S9999F9999F999999F0F1F0000000000300001E, with a // weird encoding from section 6 of: // https://www.idmanagement.gov/docs/pacs-tig-scepacs.pdf resp.extend_from_slice(&[ 0xd4, 0xe7, 0x39, 0xda, 0x73, 0x9c, 0xed, 0x39, 0xce, 0x73, 0x9d, 0x83, 0x68, 0x58, 0x21, 0x08, 0x42, 0x10, 0x84, 0x21, 0xc8, 0x42, 0x10, 0xc3, 0xeb, ]); // TLV for user UUID. resp.extend_from_slice(&[TLV_TAG_GUID, 0x10]); resp.extend_from_slice(uuid.as_bytes()); // TLV for expiration date (YYYYMMDD). resp.extend_from_slice(&[TLV_TAG_EXPIRATION_DATE, 0x08]); // TODO(awly): generate this from current time. resp.extend_from_slice("20300101".as_bytes()); // TLV for signature (empty). resp.extend_from_slice(&[TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); resp } fn build_piv_auth_cert(cert_der: &[u8]) -> Vec<u8> { // Same as above, tags in this BER-TLV value are not compatible with the spec // and existing libraries. Marshal by hand. // // Certificate TLV tag and length. let mut resp = vec![TLV_TAG_CERTIFICATE]; resp.extend_from_slice(&len_to_vec(cert_der.len())); resp.extend_from_slice(cert_der); // CertInfo TLV (0x00 indicates uncompressed cert). resp.extend_from_slice(&[TLV_TAG_CERTINFO, 0x01, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); // Wrap with top-level TLV tag and length. let mut resp_outer = vec![TLV_TAG_DATA_FIELD]; resp_outer.extend_from_slice(&len_to_vec(resp.len())); resp_outer.extend_from_slice(&resp); resp_outer } } #[derive(Debug)] pub struct Response { data: Option<Vec<u8>>, status: Status, } impl Response { fn new(status: Status) -> Self { Self { data: None, status } } fn with_data(status: Status, data: Vec<u8>) -> Self { Self { data: Some(data),
status, } } pub fn encode(&self) -> Vec<u8> { let mut buf = Vec::new(); if let Some(data) = &self.data { buf.extend_from_slice(data); } let status: [u8; 2] = self.status.into(); buf.extend_from_slice(&status); buf } } // SELECT command tags. const TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE: u8 = 0x61; const TLV_TAG_AID: u8 = 0x4F; const TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY: u8 = 0x79; const TLV_TAG_DATA_FIELD: u8 = 0x53; const TLV_TAG_FASC_N: u8 = 0x30; const TLV_TAG_GUID: u8 = 0x34; const TLV_TAG_EXPIRATION_DATE: u8 = 0x35; const TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE: u8 = 0x3E; const TLV_TAG_ERROR_DETECTION_CODE: u8 = 0xFE; const TLV_TAG_CERTIFICATE: u8 = 0x70; const TLV_TAG_CERTINFO: u8 = 0x71; // GENERAL AUTHENTICATE command tags. const TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE: u8 = 0x7C; const TLV_TAG_CHALLENGE: u8 = 0x81; const TLV_TAG_RESPONSE: u8 = 0x82; fn tlv(tag: u8, value: Value) -> RdpResult<Tlv> { Tlv::new(tlv_tag(tag)?, value) .map_err(|e| invalid_data_error(&format!("TLV with tag {tag:#X} invalid: {e:?}"))) } fn tlv_tag(val: u8) -> RdpResult<Tag> { Tag::try_from(val).map_err(|e| invalid_data_error(&format!("TLV tag {val:#X} invalid: {e:?}"))) } fn hex_data<const S: usize>(cmd: &Command<S>) -> String { to_hex(cmd.data()) } fn to_hex(bytes: &[u8]) -> String { let mut s = String::new(); for b in bytes { // https://rust-lang.github.io/rust-clippy/master/index.html#format_push_string let _ = write!(s, "{b:02X}"); } s } #[allow(clippy::cast_possible_truncation)] fn len_to_vec(len: usize) -> Vec<u8> { if len < 0x7f { vec![len as u8] } else { let mut ret: Vec<u8> = len .to_be_bytes() .iter() .skip_while(|&x| *x == 0) .cloned() .collect(); ret.insert(0, 0x80 | ret.len() as u8); ret } }
random_line_split
piv.rs
// Copyright 2021 Gravitational, Inc // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::errors::invalid_data_error; use iso7816::aid::Aid; use iso7816::command::instruction::Instruction; use iso7816::command::Command; use iso7816::response::Status; use iso7816_tlv::ber::{Tag, Tlv, Value}; use rdp::model::error::*; use rsa::pkcs1::DecodeRsaPrivateKey; use rsa::traits::{PrivateKeyParts, PublicKeyParts}; use rsa::{BigUint, RsaPrivateKey}; use std::convert::TryFrom; use std::fmt::Write as _; use std::io::{Cursor, Read}; use uuid::Uuid; // AID (Application ID) of PIV application, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf const PIV_AID: Aid = Aid::new_truncatable( &[ 0xA0, 0x00, 0x00, 0x03, 0x08, 0x00, 0x00, 0x10, 0x00, 0x01, 0x00, ], 5, // usually truncates to first 5 bytes ); // Card implements a PIV-compatible smartcard, per: // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf #[derive(Debug, PartialEq, Eq)] pub struct Card<const S: usize> { // Card-holder user ID (CHUID). In federal agencies, this value would be unique per employee // and encodes some agency information. In our case it's static. chuid: Vec<u8>, piv_auth_cert: Vec<u8>, piv_auth_key: RsaPrivateKey, pin: String, // Pending command and response to receive/send over multiple messages when // they don't fit into one. pending_command: Option<Command<S>>, pending_response: Option<Cursor<Vec<u8>>>, } impl<const S: usize> Card<S> { pub fn new(uuid: Uuid, cert_der: &[u8], key_der: &[u8], pin: String) -> RdpResult<Self> { let piv_auth_key = RsaPrivateKey::from_pkcs1_der(key_der).map_err(|e| { invalid_data_error(&format!("failed to parse private key from DER: {e:?}")) })?; Ok(Self { chuid: Self::build_chuid(uuid), piv_auth_cert: Self::build_piv_auth_cert(cert_der), piv_auth_key, pin, pending_command: None, pending_response: None, }) } pub fn handle(&mut self, cmd: Command<S>) -> RdpResult<Response> { debug!("got command: {:?}", cmd); debug!("command data: {}", hex_data(&cmd)); // Handle chained commands. let cmd = match self.pending_command.as_mut() { None => cmd, Some(pending) => { pending .extend_from_command(&cmd) .map_err(|_| invalid_data_error("could not build chained command"))?; pending.clone() } }; if cmd.class().chain().not_the_last() { self.pending_command = Some(cmd); return Ok(Response::new(Status::Success)); } else { self.pending_command = None; } let resp = match cmd.instruction() { Instruction::Select => self.handle_select(cmd), Instruction::Verify => self.handle_verify(cmd), Instruction::GetData => self.handle_get_data(cmd), Instruction::GetResponse => self.handle_get_response(cmd), Instruction::GeneralAuthenticate => self.handle_general_authenticate(cmd), _ => { warn!("unimplemented instruction {:?}", cmd.instruction()); Ok(Response::new(Status::InstructionNotSupportedOrInvalid)) } }?; debug!("send response: {:?}", resp); debug!("response data: {}", to_hex(&resp.encode())); Ok(resp) } fn handle_select(&mut self, cmd: Command<S>) -> RdpResult<Response> { // For our use case, we only allow selecting the PIV application on the smartcard. // // P1=04 and P2=00 means selection of DF (usually) application by name. Everything else not // supported. if cmd.p1 != 0x04 && cmd.p2 != 0x00 { return Ok(Response::new(Status::NotFound)); } if !PIV_AID.matches(cmd.data()) { return Ok(Response::new(Status::NotFound)); } // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.1 let resp = tlv( TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE, Value::Constructed(vec![ tlv( TLV_TAG_AID, Value::Primitive(vec![0x00, 0x00, 0x10, 0x00, 0x01, 0x00]), )?, tlv( TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY, Value::Constructed(vec![tlv( TLV_TAG_AID, Value::Primitive(PIV_AID.truncated().to_vec()), )?]), )?, ]), )?; Ok(Response::with_data(Status::Success, resp.to_vec())) } fn handle_verify(&mut self, cmd: Command<S>) -> RdpResult<Response> { if cmd.data() == self.pin.as_bytes() { Ok(Response::new(Status::Success)) } else { warn!("PIN mismatch, want {}, got {:?}", self.pin, cmd.data()); Ok(Response::new(Status::VerificationFailed)) } } fn handle_get_data(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf section // 3.1.2. if cmd.p1 != 0x3F && cmd.p2 != 0xFF { return Ok(Response::new(Status::NotFound)); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(0x5C)? { return Ok(Response::new(Status::NotFound)); } match request_tlv.value() { Value::Primitive(tag) => match to_hex(tag).as_str() { // Card Holder Unique Identifier. "5FC102" => Ok(Response::with_data(Status::Success, self.chuid.clone())), // X.509 Certificate for PIV Authentication "5FC105" => { self.pending_response = Some(Cursor::new(self.piv_auth_cert.clone())); self.handle_get_response(cmd) } _ => { // Some other unimplemented data object. Ok(Response::new(Status::NotFound)) } }, Value::Constructed(_) => Ok(Response::new(Status::NotFound)), } } fn handle_get_response(&mut self, _cmd: Command<S>) -> RdpResult<Response> { // CHINK_SIZE is the max response data size in bytes, without resorting to "extended" // messages. const CHUNK_SIZE: usize = 256; match &mut self.pending_response { None => Ok(Response::new(Status::NotFound)), Some(cursor) => { let mut chunk = [0; CHUNK_SIZE]; let n = cursor.read(&mut chunk)?; let mut chunk = chunk.to_vec(); chunk.truncate(n); let remaining = cursor.get_ref().len() as u64 - cursor.position(); let status = if remaining == 0 { Status::Success } else if remaining < CHUNK_SIZE as u64 { Status::MoreAvailable(remaining as u8) } else { Status::MoreAvailable(0) }; Ok(Response::with_data(status, chunk)) } } } /// Sign the challenge. /// /// Note: for signatures, typically you'd use a signer that hashes the input data, adds padding /// according to some scheme (like PKCS1v15 or PSS) and then "decrypts" this data with the key. /// The decrypted blob is the signature. /// /// In our case, the RDP server does the hashing and padding, and only gives us a finished blob /// to decrypt. Most crypto libraries don't directly expose RSA decryption without padding, as /// it's easy to build insecure crypto systems. Thankfully for us, this decryption is just a single /// modpow operation which is suppored by RustCrypto. fn sign_auth_challenge(&self, challenge: &[u8]) -> Vec<u8> { let c = BigUint::from_bytes_be(challenge); let plain_text = c .modpow(self.piv_auth_key.d(), self.piv_auth_key.n()) .to_bytes_be(); let mut result = vec![0u8; self.piv_auth_key.size()]; let start = result.len() - plain_text.len(); result[start..].copy_from_slice(&plain_text); result } fn handle_general_authenticate(&mut self, cmd: Command<S>) -> RdpResult<Response> { // See section 3.2.4 and example in Appending A.3 from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // P1='07' means 2048-bit RSA. // // TODO(zmb3): compare algorithm against the private key using consts from // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-78-4.pdf // TODO(zmb3): support non-RSA keys, if needed. if cmd.p1 != 0x07 { return Err(invalid_data_error(&format!( "unsupported algorithm identifier P1:{:#X} in general authenticate command", cmd.p1 ))); } // P2='9A' means PIV Authentication Key (matches our cert '5FC105' in handle_get_data). if cmd.p2 != 0x9A { return Err(invalid_data_error(&format!( "unsupported key reference P2:{:#X} in general authenticate command", cmd.p2 ))); } let request_tlv = Tlv::from_bytes(cmd.data()) .map_err(|e| invalid_data_error(&format!("TLV invalid: {e:?}")))?; if *request_tlv.tag() != tlv_tag(TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE)? { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } // Extract the challenge field. let request_tlvs = match request_tlv.value() { Value::Primitive(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } Value::Constructed(tlvs) => tlvs, }; let mut challenge = None; for data in request_tlvs { if *data.tag() != tlv_tag(TLV_TAG_CHALLENGE)? { continue; } challenge = match data.value() { Value::Primitive(chal) => Some(chal), Value::Constructed(_) => { return Err(invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}" ))); } }; } let challenge = challenge.ok_or_else(|| { invalid_data_error(&format!( "general authenticate command TLV invalid: {request_tlv:?}, missing challenge data" )) })?; // TODO(zmb3): support non-RSA keys, if needed. let signed_challenge = self.sign_auth_challenge(challenge); // Return signed challenge. let resp = tlv( TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE, Value::Constructed(vec![tlv( TLV_TAG_RESPONSE, Value::Primitive(signed_challenge), )?]), )? .to_vec(); self.pending_response = Some(Cursor::new(resp)); self.handle_get_response(cmd) } fn build_chuid(uuid: Uuid) -> Vec<u8> { // This is gross: the response is a BER-TLV value, but it has nested SIMPLE-TLV // values. None of the TLV encoding libraries out there support this, they fail // when checking the tag of nested TLVs. // // So, construct the TLV by hand from raw bytes. Hopefully the comments will be // enough to explain the structure. // // https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-73-4.pdf // table 9 has the explanation of fields. // // Start with a top-level BER-TLV tag and length: let mut resp = vec![TLV_TAG_DATA_FIELD, 0x3B]; // TLV tag and length for FASC-N. resp.extend_from_slice(&[TLV_TAG_FASC_N, 0x19]); // FASC-N value containing S9999F9999F999999F0F1F0000000000300001E, with a // weird encoding from section 6 of: // https://www.idmanagement.gov/docs/pacs-tig-scepacs.pdf resp.extend_from_slice(&[ 0xd4, 0xe7, 0x39, 0xda, 0x73, 0x9c, 0xed, 0x39, 0xce, 0x73, 0x9d, 0x83, 0x68, 0x58, 0x21, 0x08, 0x42, 0x10, 0x84, 0x21, 0xc8, 0x42, 0x10, 0xc3, 0xeb, ]); // TLV for user UUID. resp.extend_from_slice(&[TLV_TAG_GUID, 0x10]); resp.extend_from_slice(uuid.as_bytes()); // TLV for expiration date (YYYYMMDD). resp.extend_from_slice(&[TLV_TAG_EXPIRATION_DATE, 0x08]); // TODO(awly): generate this from current time. resp.extend_from_slice("20300101".as_bytes()); // TLV for signature (empty). resp.extend_from_slice(&[TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); resp } fn build_piv_auth_cert(cert_der: &[u8]) -> Vec<u8> { // Same as above, tags in this BER-TLV value are not compatible with the spec // and existing libraries. Marshal by hand. // // Certificate TLV tag and length. let mut resp = vec![TLV_TAG_CERTIFICATE]; resp.extend_from_slice(&len_to_vec(cert_der.len())); resp.extend_from_slice(cert_der); // CertInfo TLV (0x00 indicates uncompressed cert). resp.extend_from_slice(&[TLV_TAG_CERTINFO, 0x01, 0x00]); // TLV for error detection code. resp.extend_from_slice(&[TLV_TAG_ERROR_DETECTION_CODE, 0x00]); // Wrap with top-level TLV tag and length. let mut resp_outer = vec![TLV_TAG_DATA_FIELD]; resp_outer.extend_from_slice(&len_to_vec(resp.len())); resp_outer.extend_from_slice(&resp); resp_outer } } #[derive(Debug)] pub struct Response { data: Option<Vec<u8>>, status: Status, } impl Response { fn new(status: Status) -> Self { Self { data: None, status } } fn with_data(status: Status, data: Vec<u8>) -> Self { Self { data: Some(data), status, } } pub fn encode(&self) -> Vec<u8> { let mut buf = Vec::new(); if let Some(data) = &self.data { buf.extend_from_slice(data); } let status: [u8; 2] = self.status.into(); buf.extend_from_slice(&status); buf } } // SELECT command tags. const TLV_TAG_PIV_APPLICATION_PROPERTY_TEMPLATE: u8 = 0x61; const TLV_TAG_AID: u8 = 0x4F; const TLV_TAG_COEXISTENT_TAG_ALLOCATION_AUTHORITY: u8 = 0x79; const TLV_TAG_DATA_FIELD: u8 = 0x53; const TLV_TAG_FASC_N: u8 = 0x30; const TLV_TAG_GUID: u8 = 0x34; const TLV_TAG_EXPIRATION_DATE: u8 = 0x35; const TLV_TAG_ISSUER_ASYMMETRIC_SIGNATURE: u8 = 0x3E; const TLV_TAG_ERROR_DETECTION_CODE: u8 = 0xFE; const TLV_TAG_CERTIFICATE: u8 = 0x70; const TLV_TAG_CERTINFO: u8 = 0x71; // GENERAL AUTHENTICATE command tags. const TLV_TAG_DYNAMIC_AUTHENTICATION_TEMPLATE: u8 = 0x7C; const TLV_TAG_CHALLENGE: u8 = 0x81; const TLV_TAG_RESPONSE: u8 = 0x82; fn tlv(tag: u8, value: Value) -> RdpResult<Tlv> { Tlv::new(tlv_tag(tag)?, value) .map_err(|e| invalid_data_error(&format!("TLV with tag {tag:#X} invalid: {e:?}"))) } fn tlv_tag(val: u8) -> RdpResult<Tag> { Tag::try_from(val).map_err(|e| invalid_data_error(&format!("TLV tag {val:#X} invalid: {e:?}"))) } fn hex_data<const S: usize>(cmd: &Command<S>) -> String { to_hex(cmd.data()) } fn to_hex(bytes: &[u8]) -> String { let mut s = String::new(); for b in bytes { // https://rust-lang.github.io/rust-clippy/master/index.html#format_push_string let _ = write!(s, "{b:02X}"); } s } #[allow(clippy::cast_possible_truncation)] fn
(len: usize) -> Vec<u8> { if len < 0x7f { vec![len as u8] } else { let mut ret: Vec<u8> = len .to_be_bytes() .iter() .skip_while(|&x| *x == 0) .cloned() .collect(); ret.insert(0, 0x80 | ret.len() as u8); ret } }
len_to_vec
identifier_name
monitor.go
package command import ( "fmt" "strings" "sync" "time" "github.com/hashicorp/nomad/api" "github.com/hashicorp/nomad/nomad/structs" "github.com/mitchellh/cli" ) const ( // updateWait is the amount of time to wait between status // updates. Because the monitor is poll-based, we use this // delay to avoid overwhelming the API server. updateWait = time.Second ) // evalState is used to store the current "state of the world" // in the context of monitoring an evaluation. type evalState struct { status string desc string node string job string allocs map[string]*allocState wait time.Duration index uint64 } // newEvalState creates and initializes a new monitorState func newEvalState() *evalState { return &evalState{ status: structs.EvalStatusPending, allocs: make(map[string]*allocState), } } // allocState is used to track the state of an allocation type allocState struct { id string group string node string desired string desiredDesc string client string clientDesc string index uint64 // full is the allocation struct with full details. This // must be queried for explicitly so it is only included // if there is important error information inside. full *api.Allocation } // monitor wraps an evaluation monitor and holds metadata and // state information. type monitor struct { ui cli.Ui client *api.Client state *evalState // length determines the number of characters for identifiers in the ui. length int sync.Mutex } // newMonitor returns a new monitor. The returned monitor will // write output information to the provided ui. The length parameter determines // the number of characters for identifiers in the ui. func newMonitor(ui cli.Ui, client *api.Client, length int) *monitor { mon := &monitor{ ui: &cli.PrefixedUi{ InfoPrefix: "==> ", OutputPrefix: " ", ErrorPrefix: "==> ", Ui: ui, }, client: client, state: newEvalState(), length: length, } return mon } // update is used to update our monitor with new state. It can be // called whether the passed information is new or not, and will // only dump update messages when state changes. func (m *monitor) update(update *evalState) { m.Lock() defer m.Unlock() existing := m.state // Swap in the new state at the end defer func() { m.state = update }() // Check if the evaluation was triggered by a node if existing.node == "" && update.node != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by node %q", limit(update.node, m.length))) } // Check if the evaluation was triggered by a job if existing.job == "" && update.job != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by job %q", update.job)) } // Check the allocations for allocID, alloc := range update.allocs { if existing, ok := existing.allocs[allocID]; !ok { switch { case alloc.index < update.index: // New alloc with create index lower than the eval // create index indicates modification m.ui.Output(fmt.Sprintf( "Allocation %q modified: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) case alloc.desired == structs.AllocDesiredStatusRun: // New allocation with desired status running m.ui.Output(fmt.Sprintf( "Allocation %q created: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) } } else { switch { case existing.client != alloc.client: description := "" if alloc.clientDesc != "" { description = fmt.Sprintf(" (%s)", alloc.clientDesc) } // Allocation status has changed m.ui.Output(fmt.Sprintf( "Allocation %q status changed: %q -> %q%s", limit(alloc.id, m.length), existing.client, alloc.client, description)) } } } // Check if the status changed. We skip any transitions to pending status. if existing.status != "" && update.status != structs.AllocClientStatusPending && existing.status != update.status { m.ui.Output(fmt.Sprintf("Evaluation status changed: %q -> %q", existing.status, update.status)) } } // monitor is used to start monitoring the given evaluation ID. It // writes output directly to the monitor's ui, and returns the // exit code for the command. If allowPrefix is false, monitor will only accept // exact matching evalIDs. // // The return code will be 0 on successful evaluation. If there are // problems scheduling the job (impossible constraints, resources // exhausted, etc), then the return code will be 2. For any other // failures (API connectivity, internal errors, etc), the return code // will be 1. func (m *monitor) monitor(evalID string, allowPrefix bool) int { // Track if we encounter a scheduling failure. This can only be // detected while querying allocations, so we use this bool to // carry that status into the return code. var schedFailure bool // The user may have specified a prefix as eval id. We need to lookup the // full id from the database first. Since we do this in a loop we need a // variable to keep track if we've already written the header message. var headerWritten bool // Add the initial pending state m.update(newEvalState()) for { // Query the evaluation eval, _, err := m.client.Evaluations().Info(evalID, nil) if err != nil { if !allowPrefix { m.ui.Error(fmt.Sprintf("No evaluation with id %q found", evalID)) return 1 } if len(evalID) == 1 { m.ui.Error(fmt.Sprintf("Identifier must contain at least two characters.")) return 1 } if len(evalID)%2 == 1 { // Identifiers must be of even length, so we strip off the last byte // to provide a consistent user experience. evalID = evalID[:len(evalID)-1] } evals, _, err := m.client.Evaluations().PrefixList(evalID) if err != nil { m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) return 1 } if len(evals) == 0 { m.ui.Error(fmt.Sprintf("No evaluation(s) with prefix or id %q found", evalID)) return 1 } if len(evals) > 1 { // Format the evaluations out := make([]string, len(evals)+1) out[0] = "ID|Priority|Type|Triggered By|Status" for i, eval := range evals { out[i+1] = fmt.Sprintf("%s|%d|%s|%s|%s", limit(eval.ID, m.length), eval.Priority, eval.Type, eval.TriggeredBy, eval.Status) } m.ui.Output(fmt.Sprintf("Prefix matched multiple evaluations\n\n%s", formatList(out))) return 0 } // Prefix lookup matched a single evaluation eval, _, err = m.client.Evaluations().Info(evals[0].ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) } } if !headerWritten { m.ui.Info(fmt.Sprintf("Monitoring evaluation %q", limit(eval.ID, m.length))) headerWritten = true } // Create the new eval state. state := newEvalState() state.status = eval.Status state.desc = eval.StatusDescription state.node = eval.NodeID state.job = eval.JobID state.wait = eval.Wait state.index = eval.CreateIndex // Query the allocations associated with the evaluation allocs, _, err := m.client.Evaluations().Allocations(eval.ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading allocations: %s", err)) return 1 } // Add the allocs to the state for _, alloc := range allocs { state.allocs[alloc.ID] = &allocState{ id: alloc.ID, group: alloc.TaskGroup, node: alloc.NodeID, desired: alloc.DesiredStatus, desiredDesc: alloc.DesiredDescription, client: alloc.ClientStatus, clientDesc: alloc.ClientDescription, index: alloc.CreateIndex, } } // Update the state m.update(state) switch eval.Status { case structs.EvalStatusComplete, structs.EvalStatusFailed, structs.EvalStatusCancelled: if len(eval.FailedTGAllocs) == 0 { m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q", limit(eval.ID, m.length), eval.Status)) } else { // There were failures making the allocations schedFailure = true m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q but failed to place all allocations:", limit(eval.ID, m.length), eval.Status)) // Print the failures per task group for tg, metrics := range eval.FailedTGAllocs { noun := "allocation" if metrics.CoalescedFailures > 0 { noun += "s" } m.ui.Output(fmt.Sprintf("Task Group %q (failed to place %d %s):", tg, metrics.CoalescedFailures+1, noun)) metrics := formatAllocMetrics(metrics, false, " ") for _, line := range strings.Split(metrics, "\n") { m.ui.Output(line) } } if eval.BlockedEval != "" { m.ui.Output(fmt.Sprintf("Evaluation %q waiting for additional capacity to place remainder", limit(eval.BlockedEval, m.length))) } } default: // Wait for the next update time.Sleep(updateWait) continue } // Monitor the next eval in the chain, if present if eval.NextEval != "" { if eval.Wait.Nanoseconds() != 0 { m.ui.Info(fmt.Sprintf( "Monitoring next evaluation %q in %s", limit(eval.NextEval, m.length), eval.Wait)) // Skip some unnecessary polling time.Sleep(eval.Wait) } // Reset the state and monitor the new eval m.state = newEvalState() return m.monitor(eval.NextEval, allowPrefix) } break } // Treat scheduling failures specially using a dedicated exit code. // This makes it easier to detect failures from the CLI. if schedFailure { return 2 } return 0 } // dumpAllocStatus is a helper to generate a more user-friendly error message // for scheduling failures, displaying a high level status of why the job // could not be scheduled out. func dumpAllocStatus(ui cli.Ui, alloc *api.Allocation, length int) { // Print filter stats ui.Output(fmt.Sprintf("Allocation %q status %q (%d/%d nodes filtered)", limit(alloc.ID, length), alloc.ClientStatus, alloc.Metrics.NodesFiltered, alloc.Metrics.NodesEvaluated)) ui.Output(formatAllocMetrics(alloc.Metrics, true, " ")) } func formatAllocMetrics(metrics *api.AllocationMetric, scores bool, prefix string) string { // Print a helpful message if we have an eligibility problem var out string if metrics.NodesEvaluated == 0 { out += fmt.Sprintf("%s* No nodes were eligible for evaluation\n", prefix) } // Print a helpful message if the user has asked for a DC that has no // available nodes. for dc, available := range metrics.NodesAvailable { if available == 0 { out += fmt.Sprintf("%s* No nodes are available in datacenter %q\n", prefix, dc) } } // Print filter info for class, num := range metrics.ClassFiltered { out += fmt.Sprintf("%s* Class %q filtered %d nodes\n", prefix, class, num) }
out += fmt.Sprintf("%s* Constraint %q filtered %d nodes\n", prefix, cs, num) } // Print exhaustion info if ne := metrics.NodesExhausted; ne > 0 { out += fmt.Sprintf("%s* Resources exhausted on %d nodes\n", prefix, ne) } for class, num := range metrics.ClassExhausted { out += fmt.Sprintf("%s* Class %q exhausted on %d nodes\n", prefix, class, num) } for dim, num := range metrics.DimensionExhausted { out += fmt.Sprintf("%s* Dimension %q exhausted on %d nodes\n", prefix, dim, num) } // Print scores if scores { for name, score := range metrics.Scores { out += fmt.Sprintf("%s* Score %q = %f\n", prefix, name, score) } } out = strings.TrimSuffix(out, "\n") return out }
for cs, num := range metrics.ConstraintFiltered {
random_line_split
monitor.go
package command import ( "fmt" "strings" "sync" "time" "github.com/hashicorp/nomad/api" "github.com/hashicorp/nomad/nomad/structs" "github.com/mitchellh/cli" ) const ( // updateWait is the amount of time to wait between status // updates. Because the monitor is poll-based, we use this // delay to avoid overwhelming the API server. updateWait = time.Second ) // evalState is used to store the current "state of the world" // in the context of monitoring an evaluation. type evalState struct { status string desc string node string job string allocs map[string]*allocState wait time.Duration index uint64 } // newEvalState creates and initializes a new monitorState func newEvalState() *evalState { return &evalState{ status: structs.EvalStatusPending, allocs: make(map[string]*allocState), } } // allocState is used to track the state of an allocation type allocState struct { id string group string node string desired string desiredDesc string client string clientDesc string index uint64 // full is the allocation struct with full details. This // must be queried for explicitly so it is only included // if there is important error information inside. full *api.Allocation } // monitor wraps an evaluation monitor and holds metadata and // state information. type monitor struct { ui cli.Ui client *api.Client state *evalState // length determines the number of characters for identifiers in the ui. length int sync.Mutex } // newMonitor returns a new monitor. The returned monitor will // write output information to the provided ui. The length parameter determines // the number of characters for identifiers in the ui. func newMonitor(ui cli.Ui, client *api.Client, length int) *monitor { mon := &monitor{ ui: &cli.PrefixedUi{ InfoPrefix: "==> ", OutputPrefix: " ", ErrorPrefix: "==> ", Ui: ui, }, client: client, state: newEvalState(), length: length, } return mon } // update is used to update our monitor with new state. It can be // called whether the passed information is new or not, and will // only dump update messages when state changes. func (m *monitor) update(update *evalState)
// monitor is used to start monitoring the given evaluation ID. It // writes output directly to the monitor's ui, and returns the // exit code for the command. If allowPrefix is false, monitor will only accept // exact matching evalIDs. // // The return code will be 0 on successful evaluation. If there are // problems scheduling the job (impossible constraints, resources // exhausted, etc), then the return code will be 2. For any other // failures (API connectivity, internal errors, etc), the return code // will be 1. func (m *monitor) monitor(evalID string, allowPrefix bool) int { // Track if we encounter a scheduling failure. This can only be // detected while querying allocations, so we use this bool to // carry that status into the return code. var schedFailure bool // The user may have specified a prefix as eval id. We need to lookup the // full id from the database first. Since we do this in a loop we need a // variable to keep track if we've already written the header message. var headerWritten bool // Add the initial pending state m.update(newEvalState()) for { // Query the evaluation eval, _, err := m.client.Evaluations().Info(evalID, nil) if err != nil { if !allowPrefix { m.ui.Error(fmt.Sprintf("No evaluation with id %q found", evalID)) return 1 } if len(evalID) == 1 { m.ui.Error(fmt.Sprintf("Identifier must contain at least two characters.")) return 1 } if len(evalID)%2 == 1 { // Identifiers must be of even length, so we strip off the last byte // to provide a consistent user experience. evalID = evalID[:len(evalID)-1] } evals, _, err := m.client.Evaluations().PrefixList(evalID) if err != nil { m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) return 1 } if len(evals) == 0 { m.ui.Error(fmt.Sprintf("No evaluation(s) with prefix or id %q found", evalID)) return 1 } if len(evals) > 1 { // Format the evaluations out := make([]string, len(evals)+1) out[0] = "ID|Priority|Type|Triggered By|Status" for i, eval := range evals { out[i+1] = fmt.Sprintf("%s|%d|%s|%s|%s", limit(eval.ID, m.length), eval.Priority, eval.Type, eval.TriggeredBy, eval.Status) } m.ui.Output(fmt.Sprintf("Prefix matched multiple evaluations\n\n%s", formatList(out))) return 0 } // Prefix lookup matched a single evaluation eval, _, err = m.client.Evaluations().Info(evals[0].ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) } } if !headerWritten { m.ui.Info(fmt.Sprintf("Monitoring evaluation %q", limit(eval.ID, m.length))) headerWritten = true } // Create the new eval state. state := newEvalState() state.status = eval.Status state.desc = eval.StatusDescription state.node = eval.NodeID state.job = eval.JobID state.wait = eval.Wait state.index = eval.CreateIndex // Query the allocations associated with the evaluation allocs, _, err := m.client.Evaluations().Allocations(eval.ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading allocations: %s", err)) return 1 } // Add the allocs to the state for _, alloc := range allocs { state.allocs[alloc.ID] = &allocState{ id: alloc.ID, group: alloc.TaskGroup, node: alloc.NodeID, desired: alloc.DesiredStatus, desiredDesc: alloc.DesiredDescription, client: alloc.ClientStatus, clientDesc: alloc.ClientDescription, index: alloc.CreateIndex, } } // Update the state m.update(state) switch eval.Status { case structs.EvalStatusComplete, structs.EvalStatusFailed, structs.EvalStatusCancelled: if len(eval.FailedTGAllocs) == 0 { m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q", limit(eval.ID, m.length), eval.Status)) } else { // There were failures making the allocations schedFailure = true m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q but failed to place all allocations:", limit(eval.ID, m.length), eval.Status)) // Print the failures per task group for tg, metrics := range eval.FailedTGAllocs { noun := "allocation" if metrics.CoalescedFailures > 0 { noun += "s" } m.ui.Output(fmt.Sprintf("Task Group %q (failed to place %d %s):", tg, metrics.CoalescedFailures+1, noun)) metrics := formatAllocMetrics(metrics, false, " ") for _, line := range strings.Split(metrics, "\n") { m.ui.Output(line) } } if eval.BlockedEval != "" { m.ui.Output(fmt.Sprintf("Evaluation %q waiting for additional capacity to place remainder", limit(eval.BlockedEval, m.length))) } } default: // Wait for the next update time.Sleep(updateWait) continue } // Monitor the next eval in the chain, if present if eval.NextEval != "" { if eval.Wait.Nanoseconds() != 0 { m.ui.Info(fmt.Sprintf( "Monitoring next evaluation %q in %s", limit(eval.NextEval, m.length), eval.Wait)) // Skip some unnecessary polling time.Sleep(eval.Wait) } // Reset the state and monitor the new eval m.state = newEvalState() return m.monitor(eval.NextEval, allowPrefix) } break } // Treat scheduling failures specially using a dedicated exit code. // This makes it easier to detect failures from the CLI. if schedFailure { return 2 } return 0 } // dumpAllocStatus is a helper to generate a more user-friendly error message // for scheduling failures, displaying a high level status of why the job // could not be scheduled out. func dumpAllocStatus(ui cli.Ui, alloc *api.Allocation, length int) { // Print filter stats ui.Output(fmt.Sprintf("Allocation %q status %q (%d/%d nodes filtered)", limit(alloc.ID, length), alloc.ClientStatus, alloc.Metrics.NodesFiltered, alloc.Metrics.NodesEvaluated)) ui.Output(formatAllocMetrics(alloc.Metrics, true, " ")) } func formatAllocMetrics(metrics *api.AllocationMetric, scores bool, prefix string) string { // Print a helpful message if we have an eligibility problem var out string if metrics.NodesEvaluated == 0 { out += fmt.Sprintf("%s* No nodes were eligible for evaluation\n", prefix) } // Print a helpful message if the user has asked for a DC that has no // available nodes. for dc, available := range metrics.NodesAvailable { if available == 0 { out += fmt.Sprintf("%s* No nodes are available in datacenter %q\n", prefix, dc) } } // Print filter info for class, num := range metrics.ClassFiltered { out += fmt.Sprintf("%s* Class %q filtered %d nodes\n", prefix, class, num) } for cs, num := range metrics.ConstraintFiltered { out += fmt.Sprintf("%s* Constraint %q filtered %d nodes\n", prefix, cs, num) } // Print exhaustion info if ne := metrics.NodesExhausted; ne > 0 { out += fmt.Sprintf("%s* Resources exhausted on %d nodes\n", prefix, ne) } for class, num := range metrics.ClassExhausted { out += fmt.Sprintf("%s* Class %q exhausted on %d nodes\n", prefix, class, num) } for dim, num := range metrics.DimensionExhausted { out += fmt.Sprintf("%s* Dimension %q exhausted on %d nodes\n", prefix, dim, num) } // Print scores if scores { for name, score := range metrics.Scores { out += fmt.Sprintf("%s* Score %q = %f\n", prefix, name, score) } } out = strings.TrimSuffix(out, "\n") return out }
{ m.Lock() defer m.Unlock() existing := m.state // Swap in the new state at the end defer func() { m.state = update }() // Check if the evaluation was triggered by a node if existing.node == "" && update.node != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by node %q", limit(update.node, m.length))) } // Check if the evaluation was triggered by a job if existing.job == "" && update.job != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by job %q", update.job)) } // Check the allocations for allocID, alloc := range update.allocs { if existing, ok := existing.allocs[allocID]; !ok { switch { case alloc.index < update.index: // New alloc with create index lower than the eval // create index indicates modification m.ui.Output(fmt.Sprintf( "Allocation %q modified: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) case alloc.desired == structs.AllocDesiredStatusRun: // New allocation with desired status running m.ui.Output(fmt.Sprintf( "Allocation %q created: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) } } else { switch { case existing.client != alloc.client: description := "" if alloc.clientDesc != "" { description = fmt.Sprintf(" (%s)", alloc.clientDesc) } // Allocation status has changed m.ui.Output(fmt.Sprintf( "Allocation %q status changed: %q -> %q%s", limit(alloc.id, m.length), existing.client, alloc.client, description)) } } } // Check if the status changed. We skip any transitions to pending status. if existing.status != "" && update.status != structs.AllocClientStatusPending && existing.status != update.status { m.ui.Output(fmt.Sprintf("Evaluation status changed: %q -> %q", existing.status, update.status)) } }
identifier_body
monitor.go
package command import ( "fmt" "strings" "sync" "time" "github.com/hashicorp/nomad/api" "github.com/hashicorp/nomad/nomad/structs" "github.com/mitchellh/cli" ) const ( // updateWait is the amount of time to wait between status // updates. Because the monitor is poll-based, we use this // delay to avoid overwhelming the API server. updateWait = time.Second ) // evalState is used to store the current "state of the world" // in the context of monitoring an evaluation. type evalState struct { status string desc string node string job string allocs map[string]*allocState wait time.Duration index uint64 } // newEvalState creates and initializes a new monitorState func newEvalState() *evalState { return &evalState{ status: structs.EvalStatusPending, allocs: make(map[string]*allocState), } } // allocState is used to track the state of an allocation type allocState struct { id string group string node string desired string desiredDesc string client string clientDesc string index uint64 // full is the allocation struct with full details. This // must be queried for explicitly so it is only included // if there is important error information inside. full *api.Allocation } // monitor wraps an evaluation monitor and holds metadata and // state information. type monitor struct { ui cli.Ui client *api.Client state *evalState // length determines the number of characters for identifiers in the ui. length int sync.Mutex } // newMonitor returns a new monitor. The returned monitor will // write output information to the provided ui. The length parameter determines // the number of characters for identifiers in the ui. func newMonitor(ui cli.Ui, client *api.Client, length int) *monitor { mon := &monitor{ ui: &cli.PrefixedUi{ InfoPrefix: "==> ", OutputPrefix: " ", ErrorPrefix: "==> ", Ui: ui, }, client: client, state: newEvalState(), length: length, } return mon } // update is used to update our monitor with new state. It can be // called whether the passed information is new or not, and will // only dump update messages when state changes. func (m *monitor) update(update *evalState) { m.Lock() defer m.Unlock() existing := m.state // Swap in the new state at the end defer func() { m.state = update }() // Check if the evaluation was triggered by a node if existing.node == "" && update.node != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by node %q", limit(update.node, m.length))) } // Check if the evaluation was triggered by a job if existing.job == "" && update.job != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by job %q", update.job)) } // Check the allocations for allocID, alloc := range update.allocs { if existing, ok := existing.allocs[allocID]; !ok { switch { case alloc.index < update.index: // New alloc with create index lower than the eval // create index indicates modification m.ui.Output(fmt.Sprintf( "Allocation %q modified: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) case alloc.desired == structs.AllocDesiredStatusRun: // New allocation with desired status running m.ui.Output(fmt.Sprintf( "Allocation %q created: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) } } else { switch { case existing.client != alloc.client: description := "" if alloc.clientDesc != "" { description = fmt.Sprintf(" (%s)", alloc.clientDesc) } // Allocation status has changed m.ui.Output(fmt.Sprintf( "Allocation %q status changed: %q -> %q%s", limit(alloc.id, m.length), existing.client, alloc.client, description)) } } } // Check if the status changed. We skip any transitions to pending status. if existing.status != "" && update.status != structs.AllocClientStatusPending && existing.status != update.status { m.ui.Output(fmt.Sprintf("Evaluation status changed: %q -> %q", existing.status, update.status)) } } // monitor is used to start monitoring the given evaluation ID. It // writes output directly to the monitor's ui, and returns the // exit code for the command. If allowPrefix is false, monitor will only accept // exact matching evalIDs. // // The return code will be 0 on successful evaluation. If there are // problems scheduling the job (impossible constraints, resources // exhausted, etc), then the return code will be 2. For any other // failures (API connectivity, internal errors, etc), the return code // will be 1. func (m *monitor) monitor(evalID string, allowPrefix bool) int { // Track if we encounter a scheduling failure. This can only be // detected while querying allocations, so we use this bool to // carry that status into the return code. var schedFailure bool // The user may have specified a prefix as eval id. We need to lookup the // full id from the database first. Since we do this in a loop we need a // variable to keep track if we've already written the header message. var headerWritten bool // Add the initial pending state m.update(newEvalState()) for { // Query the evaluation eval, _, err := m.client.Evaluations().Info(evalID, nil) if err != nil { if !allowPrefix { m.ui.Error(fmt.Sprintf("No evaluation with id %q found", evalID)) return 1 } if len(evalID) == 1 { m.ui.Error(fmt.Sprintf("Identifier must contain at least two characters.")) return 1 } if len(evalID)%2 == 1 { // Identifiers must be of even length, so we strip off the last byte // to provide a consistent user experience. evalID = evalID[:len(evalID)-1] } evals, _, err := m.client.Evaluations().PrefixList(evalID) if err != nil
if len(evals) == 0 { m.ui.Error(fmt.Sprintf("No evaluation(s) with prefix or id %q found", evalID)) return 1 } if len(evals) > 1 { // Format the evaluations out := make([]string, len(evals)+1) out[0] = "ID|Priority|Type|Triggered By|Status" for i, eval := range evals { out[i+1] = fmt.Sprintf("%s|%d|%s|%s|%s", limit(eval.ID, m.length), eval.Priority, eval.Type, eval.TriggeredBy, eval.Status) } m.ui.Output(fmt.Sprintf("Prefix matched multiple evaluations\n\n%s", formatList(out))) return 0 } // Prefix lookup matched a single evaluation eval, _, err = m.client.Evaluations().Info(evals[0].ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) } } if !headerWritten { m.ui.Info(fmt.Sprintf("Monitoring evaluation %q", limit(eval.ID, m.length))) headerWritten = true } // Create the new eval state. state := newEvalState() state.status = eval.Status state.desc = eval.StatusDescription state.node = eval.NodeID state.job = eval.JobID state.wait = eval.Wait state.index = eval.CreateIndex // Query the allocations associated with the evaluation allocs, _, err := m.client.Evaluations().Allocations(eval.ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading allocations: %s", err)) return 1 } // Add the allocs to the state for _, alloc := range allocs { state.allocs[alloc.ID] = &allocState{ id: alloc.ID, group: alloc.TaskGroup, node: alloc.NodeID, desired: alloc.DesiredStatus, desiredDesc: alloc.DesiredDescription, client: alloc.ClientStatus, clientDesc: alloc.ClientDescription, index: alloc.CreateIndex, } } // Update the state m.update(state) switch eval.Status { case structs.EvalStatusComplete, structs.EvalStatusFailed, structs.EvalStatusCancelled: if len(eval.FailedTGAllocs) == 0 { m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q", limit(eval.ID, m.length), eval.Status)) } else { // There were failures making the allocations schedFailure = true m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q but failed to place all allocations:", limit(eval.ID, m.length), eval.Status)) // Print the failures per task group for tg, metrics := range eval.FailedTGAllocs { noun := "allocation" if metrics.CoalescedFailures > 0 { noun += "s" } m.ui.Output(fmt.Sprintf("Task Group %q (failed to place %d %s):", tg, metrics.CoalescedFailures+1, noun)) metrics := formatAllocMetrics(metrics, false, " ") for _, line := range strings.Split(metrics, "\n") { m.ui.Output(line) } } if eval.BlockedEval != "" { m.ui.Output(fmt.Sprintf("Evaluation %q waiting for additional capacity to place remainder", limit(eval.BlockedEval, m.length))) } } default: // Wait for the next update time.Sleep(updateWait) continue } // Monitor the next eval in the chain, if present if eval.NextEval != "" { if eval.Wait.Nanoseconds() != 0 { m.ui.Info(fmt.Sprintf( "Monitoring next evaluation %q in %s", limit(eval.NextEval, m.length), eval.Wait)) // Skip some unnecessary polling time.Sleep(eval.Wait) } // Reset the state and monitor the new eval m.state = newEvalState() return m.monitor(eval.NextEval, allowPrefix) } break } // Treat scheduling failures specially using a dedicated exit code. // This makes it easier to detect failures from the CLI. if schedFailure { return 2 } return 0 } // dumpAllocStatus is a helper to generate a more user-friendly error message // for scheduling failures, displaying a high level status of why the job // could not be scheduled out. func dumpAllocStatus(ui cli.Ui, alloc *api.Allocation, length int) { // Print filter stats ui.Output(fmt.Sprintf("Allocation %q status %q (%d/%d nodes filtered)", limit(alloc.ID, length), alloc.ClientStatus, alloc.Metrics.NodesFiltered, alloc.Metrics.NodesEvaluated)) ui.Output(formatAllocMetrics(alloc.Metrics, true, " ")) } func formatAllocMetrics(metrics *api.AllocationMetric, scores bool, prefix string) string { // Print a helpful message if we have an eligibility problem var out string if metrics.NodesEvaluated == 0 { out += fmt.Sprintf("%s* No nodes were eligible for evaluation\n", prefix) } // Print a helpful message if the user has asked for a DC that has no // available nodes. for dc, available := range metrics.NodesAvailable { if available == 0 { out += fmt.Sprintf("%s* No nodes are available in datacenter %q\n", prefix, dc) } } // Print filter info for class, num := range metrics.ClassFiltered { out += fmt.Sprintf("%s* Class %q filtered %d nodes\n", prefix, class, num) } for cs, num := range metrics.ConstraintFiltered { out += fmt.Sprintf("%s* Constraint %q filtered %d nodes\n", prefix, cs, num) } // Print exhaustion info if ne := metrics.NodesExhausted; ne > 0 { out += fmt.Sprintf("%s* Resources exhausted on %d nodes\n", prefix, ne) } for class, num := range metrics.ClassExhausted { out += fmt.Sprintf("%s* Class %q exhausted on %d nodes\n", prefix, class, num) } for dim, num := range metrics.DimensionExhausted { out += fmt.Sprintf("%s* Dimension %q exhausted on %d nodes\n", prefix, dim, num) } // Print scores if scores { for name, score := range metrics.Scores { out += fmt.Sprintf("%s* Score %q = %f\n", prefix, name, score) } } out = strings.TrimSuffix(out, "\n") return out }
{ m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) return 1 }
conditional_block
monitor.go
package command import ( "fmt" "strings" "sync" "time" "github.com/hashicorp/nomad/api" "github.com/hashicorp/nomad/nomad/structs" "github.com/mitchellh/cli" ) const ( // updateWait is the amount of time to wait between status // updates. Because the monitor is poll-based, we use this // delay to avoid overwhelming the API server. updateWait = time.Second ) // evalState is used to store the current "state of the world" // in the context of monitoring an evaluation. type evalState struct { status string desc string node string job string allocs map[string]*allocState wait time.Duration index uint64 } // newEvalState creates and initializes a new monitorState func newEvalState() *evalState { return &evalState{ status: structs.EvalStatusPending, allocs: make(map[string]*allocState), } } // allocState is used to track the state of an allocation type allocState struct { id string group string node string desired string desiredDesc string client string clientDesc string index uint64 // full is the allocation struct with full details. This // must be queried for explicitly so it is only included // if there is important error information inside. full *api.Allocation } // monitor wraps an evaluation monitor and holds metadata and // state information. type monitor struct { ui cli.Ui client *api.Client state *evalState // length determines the number of characters for identifiers in the ui. length int sync.Mutex } // newMonitor returns a new monitor. The returned monitor will // write output information to the provided ui. The length parameter determines // the number of characters for identifiers in the ui. func newMonitor(ui cli.Ui, client *api.Client, length int) *monitor { mon := &monitor{ ui: &cli.PrefixedUi{ InfoPrefix: "==> ", OutputPrefix: " ", ErrorPrefix: "==> ", Ui: ui, }, client: client, state: newEvalState(), length: length, } return mon } // update is used to update our monitor with new state. It can be // called whether the passed information is new or not, and will // only dump update messages when state changes. func (m *monitor) update(update *evalState) { m.Lock() defer m.Unlock() existing := m.state // Swap in the new state at the end defer func() { m.state = update }() // Check if the evaluation was triggered by a node if existing.node == "" && update.node != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by node %q", limit(update.node, m.length))) } // Check if the evaluation was triggered by a job if existing.job == "" && update.job != "" { m.ui.Output(fmt.Sprintf("Evaluation triggered by job %q", update.job)) } // Check the allocations for allocID, alloc := range update.allocs { if existing, ok := existing.allocs[allocID]; !ok { switch { case alloc.index < update.index: // New alloc with create index lower than the eval // create index indicates modification m.ui.Output(fmt.Sprintf( "Allocation %q modified: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) case alloc.desired == structs.AllocDesiredStatusRun: // New allocation with desired status running m.ui.Output(fmt.Sprintf( "Allocation %q created: node %q, group %q", limit(alloc.id, m.length), limit(alloc.node, m.length), alloc.group)) } } else { switch { case existing.client != alloc.client: description := "" if alloc.clientDesc != "" { description = fmt.Sprintf(" (%s)", alloc.clientDesc) } // Allocation status has changed m.ui.Output(fmt.Sprintf( "Allocation %q status changed: %q -> %q%s", limit(alloc.id, m.length), existing.client, alloc.client, description)) } } } // Check if the status changed. We skip any transitions to pending status. if existing.status != "" && update.status != structs.AllocClientStatusPending && existing.status != update.status { m.ui.Output(fmt.Sprintf("Evaluation status changed: %q -> %q", existing.status, update.status)) } } // monitor is used to start monitoring the given evaluation ID. It // writes output directly to the monitor's ui, and returns the // exit code for the command. If allowPrefix is false, monitor will only accept // exact matching evalIDs. // // The return code will be 0 on successful evaluation. If there are // problems scheduling the job (impossible constraints, resources // exhausted, etc), then the return code will be 2. For any other // failures (API connectivity, internal errors, etc), the return code // will be 1. func (m *monitor) monitor(evalID string, allowPrefix bool) int { // Track if we encounter a scheduling failure. This can only be // detected while querying allocations, so we use this bool to // carry that status into the return code. var schedFailure bool // The user may have specified a prefix as eval id. We need to lookup the // full id from the database first. Since we do this in a loop we need a // variable to keep track if we've already written the header message. var headerWritten bool // Add the initial pending state m.update(newEvalState()) for { // Query the evaluation eval, _, err := m.client.Evaluations().Info(evalID, nil) if err != nil { if !allowPrefix { m.ui.Error(fmt.Sprintf("No evaluation with id %q found", evalID)) return 1 } if len(evalID) == 1 { m.ui.Error(fmt.Sprintf("Identifier must contain at least two characters.")) return 1 } if len(evalID)%2 == 1 { // Identifiers must be of even length, so we strip off the last byte // to provide a consistent user experience. evalID = evalID[:len(evalID)-1] } evals, _, err := m.client.Evaluations().PrefixList(evalID) if err != nil { m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) return 1 } if len(evals) == 0 { m.ui.Error(fmt.Sprintf("No evaluation(s) with prefix or id %q found", evalID)) return 1 } if len(evals) > 1 { // Format the evaluations out := make([]string, len(evals)+1) out[0] = "ID|Priority|Type|Triggered By|Status" for i, eval := range evals { out[i+1] = fmt.Sprintf("%s|%d|%s|%s|%s", limit(eval.ID, m.length), eval.Priority, eval.Type, eval.TriggeredBy, eval.Status) } m.ui.Output(fmt.Sprintf("Prefix matched multiple evaluations\n\n%s", formatList(out))) return 0 } // Prefix lookup matched a single evaluation eval, _, err = m.client.Evaluations().Info(evals[0].ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading evaluation: %s", err)) } } if !headerWritten { m.ui.Info(fmt.Sprintf("Monitoring evaluation %q", limit(eval.ID, m.length))) headerWritten = true } // Create the new eval state. state := newEvalState() state.status = eval.Status state.desc = eval.StatusDescription state.node = eval.NodeID state.job = eval.JobID state.wait = eval.Wait state.index = eval.CreateIndex // Query the allocations associated with the evaluation allocs, _, err := m.client.Evaluations().Allocations(eval.ID, nil) if err != nil { m.ui.Error(fmt.Sprintf("Error reading allocations: %s", err)) return 1 } // Add the allocs to the state for _, alloc := range allocs { state.allocs[alloc.ID] = &allocState{ id: alloc.ID, group: alloc.TaskGroup, node: alloc.NodeID, desired: alloc.DesiredStatus, desiredDesc: alloc.DesiredDescription, client: alloc.ClientStatus, clientDesc: alloc.ClientDescription, index: alloc.CreateIndex, } } // Update the state m.update(state) switch eval.Status { case structs.EvalStatusComplete, structs.EvalStatusFailed, structs.EvalStatusCancelled: if len(eval.FailedTGAllocs) == 0 { m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q", limit(eval.ID, m.length), eval.Status)) } else { // There were failures making the allocations schedFailure = true m.ui.Info(fmt.Sprintf("Evaluation %q finished with status %q but failed to place all allocations:", limit(eval.ID, m.length), eval.Status)) // Print the failures per task group for tg, metrics := range eval.FailedTGAllocs { noun := "allocation" if metrics.CoalescedFailures > 0 { noun += "s" } m.ui.Output(fmt.Sprintf("Task Group %q (failed to place %d %s):", tg, metrics.CoalescedFailures+1, noun)) metrics := formatAllocMetrics(metrics, false, " ") for _, line := range strings.Split(metrics, "\n") { m.ui.Output(line) } } if eval.BlockedEval != "" { m.ui.Output(fmt.Sprintf("Evaluation %q waiting for additional capacity to place remainder", limit(eval.BlockedEval, m.length))) } } default: // Wait for the next update time.Sleep(updateWait) continue } // Monitor the next eval in the chain, if present if eval.NextEval != "" { if eval.Wait.Nanoseconds() != 0 { m.ui.Info(fmt.Sprintf( "Monitoring next evaluation %q in %s", limit(eval.NextEval, m.length), eval.Wait)) // Skip some unnecessary polling time.Sleep(eval.Wait) } // Reset the state and monitor the new eval m.state = newEvalState() return m.monitor(eval.NextEval, allowPrefix) } break } // Treat scheduling failures specially using a dedicated exit code. // This makes it easier to detect failures from the CLI. if schedFailure { return 2 } return 0 } // dumpAllocStatus is a helper to generate a more user-friendly error message // for scheduling failures, displaying a high level status of why the job // could not be scheduled out. func
(ui cli.Ui, alloc *api.Allocation, length int) { // Print filter stats ui.Output(fmt.Sprintf("Allocation %q status %q (%d/%d nodes filtered)", limit(alloc.ID, length), alloc.ClientStatus, alloc.Metrics.NodesFiltered, alloc.Metrics.NodesEvaluated)) ui.Output(formatAllocMetrics(alloc.Metrics, true, " ")) } func formatAllocMetrics(metrics *api.AllocationMetric, scores bool, prefix string) string { // Print a helpful message if we have an eligibility problem var out string if metrics.NodesEvaluated == 0 { out += fmt.Sprintf("%s* No nodes were eligible for evaluation\n", prefix) } // Print a helpful message if the user has asked for a DC that has no // available nodes. for dc, available := range metrics.NodesAvailable { if available == 0 { out += fmt.Sprintf("%s* No nodes are available in datacenter %q\n", prefix, dc) } } // Print filter info for class, num := range metrics.ClassFiltered { out += fmt.Sprintf("%s* Class %q filtered %d nodes\n", prefix, class, num) } for cs, num := range metrics.ConstraintFiltered { out += fmt.Sprintf("%s* Constraint %q filtered %d nodes\n", prefix, cs, num) } // Print exhaustion info if ne := metrics.NodesExhausted; ne > 0 { out += fmt.Sprintf("%s* Resources exhausted on %d nodes\n", prefix, ne) } for class, num := range metrics.ClassExhausted { out += fmt.Sprintf("%s* Class %q exhausted on %d nodes\n", prefix, class, num) } for dim, num := range metrics.DimensionExhausted { out += fmt.Sprintf("%s* Dimension %q exhausted on %d nodes\n", prefix, dim, num) } // Print scores if scores { for name, score := range metrics.Scores { out += fmt.Sprintf("%s* Score %q = %f\n", prefix, name, score) } } out = strings.TrimSuffix(out, "\n") return out }
dumpAllocStatus
identifier_name
git.rs
use anyhow::{anyhow, Context, Result}; use git2::{ build::CheckoutBuilder, ApplyLocation, Delta, Diff, DiffFormat, DiffOptions, ErrorCode, IndexAddOption, Oid, Repository, ResetType, Signature, StashApplyOptions, Time, }; use itertools::Itertools; use std::cell::RefCell; use std::collections::HashSet; use std::fs; use std::hash::Hash; use std::io::ErrorKind::NotFound; use std::iter::FromIterator; use std::path::{Path, PathBuf}; /// An abstraction over a Git repository providing complex behavior needed for /// applying changes to staged files safely. pub struct GitRepository { repository: Repository, } impl GitRepository { /// Attempts to open an already-existing repository. /// /// If the $GIT_DIR environment variable is set, this uses it to locate the /// Git repository. Otherwise, this searches up the directory tree from the /// current directory to find the repository. pub fn open() -> Result<Self> { // When strict hash verification is disabled, it means libgit2 will not // compute the "object id" of Git objects (which is a SHA-1 hash) after // reading them to verify they match the object ids being used to look // them up. This improves performance, and I don't have in front of me // a concrete example where this is necessary to prevent data loss. If // one becomes obvious, then we should make this configurable. // git2::opts::strict_hash_verification(false); let repository = Repository::open_from_env() .with_context(|| "Encountered an error when opening the Git repository.")?; Ok(Self { repository }) } pub fn save_snapshot(&mut self, staged_files: Vec<PathBuf>) -> Result<Snapshot> { let inner = || -> Result<Snapshot> { let deleted_files = self.get_deleted_files()?; let unstaged_diff = self.save_unstaged_diff()?; let backup_stash = self.save_snapshot_stash()?; // Because `git stash` restores the HEAD commit, it brings back uncommitted // deleted files. We need to clear them before creating our snapshot. GitRepository::delete_files(&deleted_files)?; self.hide_partially_staged_changes()?; Ok(Snapshot { backup_stash, staged_files, unstaged_diff, }) }; inner().with_context(|| "Encountered an error when saving a snapshot.") } pub fn apply_modifications(&mut self, snapshot: &Snapshot) -> Result<()> { self.stage_modifications(snapshot)?; if self.get_staged_files()?.is_empty() { return Err(anyhow!("Prevented an empty git commit.")); } if let Some(raw_diff) = &snapshot.unstaged_diff { let unstaged_diff = Diff::from_buffer(raw_diff)?; self.merge_modifications(unstaged_diff)?; } Ok(()) } pub fn restore_snapshot(&mut self, snapshot: &Snapshot) -> Result<()> { let mut inner = || -> Result<()> { self.hard_reset()?; if let Some(backup_stash) = &snapshot.backup_stash { self.apply_stash(&backup_stash.stash_id)?; self.restore_merge_status(&backup_stash.merge_status)?; } Ok(()) }; inner().with_context(|| "Encountered an error when restoring snapshot after another error.") } pub fn clean_snapshot(&mut self, snapshot: Snapshot) -> Result<()> { let inner = || -> Result<()> { if let Some(backup_stash) = snapshot.backup_stash { let stash_index = self .get_stash_index_from_id(&backup_stash.stash_id)? .ok_or_else(|| { anyhow!( "Could not find a backup stash with id {}.", &backup_stash.stash_id ) })?; self.repository.stash_drop(stash_index)?; } Ok(()) }; inner().with_context(|| { "Encountered an error when cleaning snapshot. You might find a stash entry \ in the stash list." }) } fn stage_modifications(&mut self, snapshot: &Snapshot) -> Result<()> { let mut index = self.repository.index()?; index.add_all( &snapshot.staged_files, IndexAddOption::DEFAULT | IndexAddOption::DISABLE_PATHSPEC_MATCH, None, )?; index.write()?; Ok(()) } fn merge_modifications(&self, unstaged_diff: Diff) -> Result<()> { self.repository .apply(&unstaged_diff, ApplyLocation::WorkDir, None) .with_context(|| "Unstaged changes could not be restored due to a merge conflict.") } fn hard_reset(&self) -> Result<()> { let head = self.repository.head()?.peel_to_commit()?; self.repository .reset(head.as_object(), ResetType::Hard, None) .map_err(|error| anyhow!(error)) } fn get_stash_index_from_id(&mut self, stash_id: &Oid) -> Result<Option<usize>> { // It would be much better if libgit2 accepted a stash Oid // instead of an index from the stash list. let ref_stash_index = RefCell::new(None); self.repository.stash_foreach(|index, _, oid| { if oid == stash_id { *ref_stash_index.borrow_mut() = Some(index); false } else { true } })?; // Copy the data out of the RefCell. let stash_index = match *ref_stash_index.borrow() { Some(index) => Some(index), None => None, }; Ok(stash_index) } fn apply_stash(&mut self, stash_id: &Oid) -> Result<()> { let stash_index = self .get_stash_index_from_id(stash_id)? .ok_or_else(|| anyhow!("Could not find a backup stash with id {}.", stash_id))?; self.repository.stash_apply( stash_index, Some(StashApplyOptions::default().reinstantiate_index()), )?; Ok(()) } fn save_unstaged_diff(&self) -> Result<Option<Vec<u8>>> { let partially_staged_files = self.get_partially_staged_files(true)?; if partially_staged_files.is_empty() { return Ok(None); } let mut diff_options = DiffOptions::new(); diff_options.show_binary(true); for file in partially_staged_files.iter() { diff_options.pathspec(file); } let unstaged_diff = self .repository .diff_index_to_workdir(None, Some(&mut diff_options))?; // The Diff created by diff_index_to_workdir is owned by the repository. // It means storing this diff separately isn't possible, and it is also // difficult to store it along with the repository together in a struct, // because that struct then will have a self reference between its diff // and its repository. // // I'm not comfortable enough with ownership to understand the correct // way to work around this, so the current approach that I'm taking is // to copy the diff out into a buffer. This is not the most performant. // // For updates about this issue, we can keep tabs on // // https://github.com/rust-lang/git2-rs/issues/622 // fn copy_diff(diff: &Diff) -> Result<Vec<u8>> { let mut buffer = vec![]; diff.print(DiffFormat::Patch, |_, _, line| { let origin = line.origin(); if origin == '+' || origin == '-' || origin == ' ' { buffer.push(origin as u8); } buffer.append(&mut line.content().to_vec()); true })?; Ok(buffer) } Ok(Some(copy_diff(&unstaged_diff)?)) } fn hide_partially_staged_changes(&self) -> Result<()> { let partially_staged_files = self.get_partially_staged_files(false)?; if partially_staged_files.is_empty() { return Ok(()); } let mut checkout_options = CheckoutBuilder::new(); checkout_options.force(); checkout_options.update_index(false); for file in partially_staged_files.iter() { checkout_options.path(file); } self.repository .checkout_index(None, Some(&mut checkout_options))?; Ok(()) } pub fn get_staged_files(&self) -> Result<Vec<PathBuf>> { let head_tree = match self.repository.head() { Ok(head) => Ok(Some(head.peel_to_tree()?)), Err(error) if error.code() == ErrorCode::UnbornBranch => Ok(None), Err(error) => Err(error), }?; let staged_files = self .repository .diff_tree_to_index(head_tree.as_ref(), None, None)? .deltas() .flat_map(|delta| { if delta.old_file().path() == delta.new_file().path() { vec![delta.old_file().path()] } else { vec![delta.old_file().path(), delta.new_file().path()] } }) .filter_map(std::convert::identity) .map(Path::to_path_buf) .collect(); Ok(staged_files) } fn get_partially_staged_files(&self, include_from_files: bool) -> Result<HashSet<PathBuf>> { let staged_files = HashSet::from_iter(self.get_staged_files()?); let unstaged_files = HashSet::from_iter( self.repository .diff_index_to_workdir(None, Some(DiffOptions::default().show_binary(true)))? .deltas() .flat_map(|delta| { if include_from_files { vec![delta.old_file().path(), delta.new_file().path()] } else { vec![delta.new_file().path()] } }) .filter_map(std::convert::identity) .map(Path::to_path_buf), ); fn intersect<P: Eq + Hash>(one: HashSet<P>, two: &HashSet<P>) -> HashSet<P> { one.into_iter().filter(|p| two.contains(p)).collect() } Ok(intersect(staged_files, &unstaged_files)) } fn get_deleted_files(&self) -> Result<Vec<PathBuf>> { let deleted_files = self .repository .diff_index_to_workdir(None, None)? .deltas() .filter(|delta| delta.status() == Delta::Deleted) .filter_map(|delta| delta.old_file().path()) .map(Path::to_path_buf) .collect_vec(); Ok(deleted_files) } fn save_snapshot_stash(&mut self) -> Result<Option<Stash>> { if self.repository.is_empty()? { return Ok(None); } fn create_signature<'a>() -> Result<Signature<'a>> { // Because this time is only used to create a dummy signature to // make the stash_save method happy, we don't need to use a real // time, which skips some calls to the kernel. // let time = Time::new(0, 0); Signature::new("Dummy", "dummy@example.com", &time) .with_context(|| "Encountered an error when creating dummy authorship information.") } // Save state when in the middle of a merge prior to stashing changes in // the working directory so that we can restore it afterward. // let merge_status = self.save_merge_status()?; let signature = create_signature()?; let stash_result = self .repository .stash_create(&signature, None, None); if let Ok(stash_id) = stash_result { self.repository.stash_store(&stash_id, Some("offstage backup"))?; } match stash_result { Ok(stash_id) => Ok(Some(Stash { stash_id, merge_status, })), Err(error) if error.code() == ErrorCode::NotFound => Ok(None), Err(error) => Err(anyhow!(error) .context("Encountered an error when stashing a backup of the working directory.")), } } fn save_merge_status(&self) -> Result<MergeStatus> { let merge_head_path = &self.repository.path().join("MERGE_HEAD"); let merge_head = Self::read_file_to_string(merge_head_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_head_path.display() ) })?; let merge_mode_path = &self.repository.path().join("MERGE_MODE"); let merge_mode = Self::read_file_to_string(merge_mode_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_mode_path.display() ) })?; let merge_msg_path = &self.repository.path().join("MERGE_MSG"); let merge_msg = Self::read_file_to_string(merge_msg_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_msg_path.display() ) })?; Ok(MergeStatus { merge_head, merge_mode, merge_msg, }) } fn restore_merge_status(&self, merge_status: &MergeStatus) -> Result<()> { // Tries to restore all files before returning the first error if one exists. let restore_merge_head_result = merge_status .merge_head .as_ref() .map_or(Ok(()), |merge_head| { let merge_head_path = &self.repository.path().join("MERGE_HEAD"); fs::write(merge_head_path, merge_head).with_context(|| { format!( "Encountered an error when restoring {}.", merge_head_path.display() ) }) }); let restore_merge_mode_result = merge_status .merge_mode .as_ref() .map_or(Ok(()), |merge_mode| { let merge_mode_path = &self.repository.path().join("MERGE_MODE"); fs::write(merge_mode_path, merge_mode).with_context(|| { format!( "Encountered an error when restoring {}.", &merge_mode_path.display() ) }) }); let restore_merge_msg_result = merge_status.merge_msg.as_ref().map_or(Ok(()), |merge_msg| { let merge_msg_path = &self.repository.path().join("MERGE_MSG"); fs::write(merge_msg_path, merge_msg).with_context(|| { format!( "Encountered an error when restoring {}.", merge_msg_path.display() ) }) }); restore_merge_head_result?; restore_merge_mode_result?; restore_merge_msg_result?; Ok(()) } fn read_file_to_string<P: AsRef<Path>>(file: P) -> Result<Option<String>> { match fs::read_to_string(file) { Ok(contents) => Ok(Some(contents)), Err(error) if error.kind() == NotFound => Ok(None), Err(error) => Err(anyhow!(error)), } } fn delete_files<P: AsRef<Path>>(files: &Vec<P>) -> Result<()> { for file in files.iter() { fs::remove_file(file).with_context(|| { format!( "Encountered error when deleting {}.", file.as_ref().display() ) })?; } Ok(()) } } #[derive(Debug)] pub struct Snapshot { pub staged_files: Vec<PathBuf>, backup_stash: Option<Stash>, unstaged_diff: Option<Vec<u8>>, } #[derive(Debug)] struct
{ stash_id: Oid, merge_status: MergeStatus, } #[derive(Debug)] struct MergeStatus { merge_head: Option<String>, merge_mode: Option<String>, merge_msg: Option<String>, }
Stash
identifier_name
git.rs
use anyhow::{anyhow, Context, Result}; use git2::{ build::CheckoutBuilder, ApplyLocation, Delta, Diff, DiffFormat, DiffOptions, ErrorCode, IndexAddOption, Oid, Repository, ResetType, Signature, StashApplyOptions, Time, }; use itertools::Itertools; use std::cell::RefCell; use std::collections::HashSet; use std::fs; use std::hash::Hash; use std::io::ErrorKind::NotFound; use std::iter::FromIterator; use std::path::{Path, PathBuf}; /// An abstraction over a Git repository providing complex behavior needed for /// applying changes to staged files safely. pub struct GitRepository { repository: Repository, } impl GitRepository { /// Attempts to open an already-existing repository. /// /// If the $GIT_DIR environment variable is set, this uses it to locate the /// Git repository. Otherwise, this searches up the directory tree from the /// current directory to find the repository. pub fn open() -> Result<Self>
pub fn save_snapshot(&mut self, staged_files: Vec<PathBuf>) -> Result<Snapshot> { let inner = || -> Result<Snapshot> { let deleted_files = self.get_deleted_files()?; let unstaged_diff = self.save_unstaged_diff()?; let backup_stash = self.save_snapshot_stash()?; // Because `git stash` restores the HEAD commit, it brings back uncommitted // deleted files. We need to clear them before creating our snapshot. GitRepository::delete_files(&deleted_files)?; self.hide_partially_staged_changes()?; Ok(Snapshot { backup_stash, staged_files, unstaged_diff, }) }; inner().with_context(|| "Encountered an error when saving a snapshot.") } pub fn apply_modifications(&mut self, snapshot: &Snapshot) -> Result<()> { self.stage_modifications(snapshot)?; if self.get_staged_files()?.is_empty() { return Err(anyhow!("Prevented an empty git commit.")); } if let Some(raw_diff) = &snapshot.unstaged_diff { let unstaged_diff = Diff::from_buffer(raw_diff)?; self.merge_modifications(unstaged_diff)?; } Ok(()) } pub fn restore_snapshot(&mut self, snapshot: &Snapshot) -> Result<()> { let mut inner = || -> Result<()> { self.hard_reset()?; if let Some(backup_stash) = &snapshot.backup_stash { self.apply_stash(&backup_stash.stash_id)?; self.restore_merge_status(&backup_stash.merge_status)?; } Ok(()) }; inner().with_context(|| "Encountered an error when restoring snapshot after another error.") } pub fn clean_snapshot(&mut self, snapshot: Snapshot) -> Result<()> { let inner = || -> Result<()> { if let Some(backup_stash) = snapshot.backup_stash { let stash_index = self .get_stash_index_from_id(&backup_stash.stash_id)? .ok_or_else(|| { anyhow!( "Could not find a backup stash with id {}.", &backup_stash.stash_id ) })?; self.repository.stash_drop(stash_index)?; } Ok(()) }; inner().with_context(|| { "Encountered an error when cleaning snapshot. You might find a stash entry \ in the stash list." }) } fn stage_modifications(&mut self, snapshot: &Snapshot) -> Result<()> { let mut index = self.repository.index()?; index.add_all( &snapshot.staged_files, IndexAddOption::DEFAULT | IndexAddOption::DISABLE_PATHSPEC_MATCH, None, )?; index.write()?; Ok(()) } fn merge_modifications(&self, unstaged_diff: Diff) -> Result<()> { self.repository .apply(&unstaged_diff, ApplyLocation::WorkDir, None) .with_context(|| "Unstaged changes could not be restored due to a merge conflict.") } fn hard_reset(&self) -> Result<()> { let head = self.repository.head()?.peel_to_commit()?; self.repository .reset(head.as_object(), ResetType::Hard, None) .map_err(|error| anyhow!(error)) } fn get_stash_index_from_id(&mut self, stash_id: &Oid) -> Result<Option<usize>> { // It would be much better if libgit2 accepted a stash Oid // instead of an index from the stash list. let ref_stash_index = RefCell::new(None); self.repository.stash_foreach(|index, _, oid| { if oid == stash_id { *ref_stash_index.borrow_mut() = Some(index); false } else { true } })?; // Copy the data out of the RefCell. let stash_index = match *ref_stash_index.borrow() { Some(index) => Some(index), None => None, }; Ok(stash_index) } fn apply_stash(&mut self, stash_id: &Oid) -> Result<()> { let stash_index = self .get_stash_index_from_id(stash_id)? .ok_or_else(|| anyhow!("Could not find a backup stash with id {}.", stash_id))?; self.repository.stash_apply( stash_index, Some(StashApplyOptions::default().reinstantiate_index()), )?; Ok(()) } fn save_unstaged_diff(&self) -> Result<Option<Vec<u8>>> { let partially_staged_files = self.get_partially_staged_files(true)?; if partially_staged_files.is_empty() { return Ok(None); } let mut diff_options = DiffOptions::new(); diff_options.show_binary(true); for file in partially_staged_files.iter() { diff_options.pathspec(file); } let unstaged_diff = self .repository .diff_index_to_workdir(None, Some(&mut diff_options))?; // The Diff created by diff_index_to_workdir is owned by the repository. // It means storing this diff separately isn't possible, and it is also // difficult to store it along with the repository together in a struct, // because that struct then will have a self reference between its diff // and its repository. // // I'm not comfortable enough with ownership to understand the correct // way to work around this, so the current approach that I'm taking is // to copy the diff out into a buffer. This is not the most performant. // // For updates about this issue, we can keep tabs on // // https://github.com/rust-lang/git2-rs/issues/622 // fn copy_diff(diff: &Diff) -> Result<Vec<u8>> { let mut buffer = vec![]; diff.print(DiffFormat::Patch, |_, _, line| { let origin = line.origin(); if origin == '+' || origin == '-' || origin == ' ' { buffer.push(origin as u8); } buffer.append(&mut line.content().to_vec()); true })?; Ok(buffer) } Ok(Some(copy_diff(&unstaged_diff)?)) } fn hide_partially_staged_changes(&self) -> Result<()> { let partially_staged_files = self.get_partially_staged_files(false)?; if partially_staged_files.is_empty() { return Ok(()); } let mut checkout_options = CheckoutBuilder::new(); checkout_options.force(); checkout_options.update_index(false); for file in partially_staged_files.iter() { checkout_options.path(file); } self.repository .checkout_index(None, Some(&mut checkout_options))?; Ok(()) } pub fn get_staged_files(&self) -> Result<Vec<PathBuf>> { let head_tree = match self.repository.head() { Ok(head) => Ok(Some(head.peel_to_tree()?)), Err(error) if error.code() == ErrorCode::UnbornBranch => Ok(None), Err(error) => Err(error), }?; let staged_files = self .repository .diff_tree_to_index(head_tree.as_ref(), None, None)? .deltas() .flat_map(|delta| { if delta.old_file().path() == delta.new_file().path() { vec![delta.old_file().path()] } else { vec![delta.old_file().path(), delta.new_file().path()] } }) .filter_map(std::convert::identity) .map(Path::to_path_buf) .collect(); Ok(staged_files) } fn get_partially_staged_files(&self, include_from_files: bool) -> Result<HashSet<PathBuf>> { let staged_files = HashSet::from_iter(self.get_staged_files()?); let unstaged_files = HashSet::from_iter( self.repository .diff_index_to_workdir(None, Some(DiffOptions::default().show_binary(true)))? .deltas() .flat_map(|delta| { if include_from_files { vec![delta.old_file().path(), delta.new_file().path()] } else { vec![delta.new_file().path()] } }) .filter_map(std::convert::identity) .map(Path::to_path_buf), ); fn intersect<P: Eq + Hash>(one: HashSet<P>, two: &HashSet<P>) -> HashSet<P> { one.into_iter().filter(|p| two.contains(p)).collect() } Ok(intersect(staged_files, &unstaged_files)) } fn get_deleted_files(&self) -> Result<Vec<PathBuf>> { let deleted_files = self .repository .diff_index_to_workdir(None, None)? .deltas() .filter(|delta| delta.status() == Delta::Deleted) .filter_map(|delta| delta.old_file().path()) .map(Path::to_path_buf) .collect_vec(); Ok(deleted_files) } fn save_snapshot_stash(&mut self) -> Result<Option<Stash>> { if self.repository.is_empty()? { return Ok(None); } fn create_signature<'a>() -> Result<Signature<'a>> { // Because this time is only used to create a dummy signature to // make the stash_save method happy, we don't need to use a real // time, which skips some calls to the kernel. // let time = Time::new(0, 0); Signature::new("Dummy", "dummy@example.com", &time) .with_context(|| "Encountered an error when creating dummy authorship information.") } // Save state when in the middle of a merge prior to stashing changes in // the working directory so that we can restore it afterward. // let merge_status = self.save_merge_status()?; let signature = create_signature()?; let stash_result = self .repository .stash_create(&signature, None, None); if let Ok(stash_id) = stash_result { self.repository.stash_store(&stash_id, Some("offstage backup"))?; } match stash_result { Ok(stash_id) => Ok(Some(Stash { stash_id, merge_status, })), Err(error) if error.code() == ErrorCode::NotFound => Ok(None), Err(error) => Err(anyhow!(error) .context("Encountered an error when stashing a backup of the working directory.")), } } fn save_merge_status(&self) -> Result<MergeStatus> { let merge_head_path = &self.repository.path().join("MERGE_HEAD"); let merge_head = Self::read_file_to_string(merge_head_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_head_path.display() ) })?; let merge_mode_path = &self.repository.path().join("MERGE_MODE"); let merge_mode = Self::read_file_to_string(merge_mode_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_mode_path.display() ) })?; let merge_msg_path = &self.repository.path().join("MERGE_MSG"); let merge_msg = Self::read_file_to_string(merge_msg_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_msg_path.display() ) })?; Ok(MergeStatus { merge_head, merge_mode, merge_msg, }) } fn restore_merge_status(&self, merge_status: &MergeStatus) -> Result<()> { // Tries to restore all files before returning the first error if one exists. let restore_merge_head_result = merge_status .merge_head .as_ref() .map_or(Ok(()), |merge_head| { let merge_head_path = &self.repository.path().join("MERGE_HEAD"); fs::write(merge_head_path, merge_head).with_context(|| { format!( "Encountered an error when restoring {}.", merge_head_path.display() ) }) }); let restore_merge_mode_result = merge_status .merge_mode .as_ref() .map_or(Ok(()), |merge_mode| { let merge_mode_path = &self.repository.path().join("MERGE_MODE"); fs::write(merge_mode_path, merge_mode).with_context(|| { format!( "Encountered an error when restoring {}.", &merge_mode_path.display() ) }) }); let restore_merge_msg_result = merge_status.merge_msg.as_ref().map_or(Ok(()), |merge_msg| { let merge_msg_path = &self.repository.path().join("MERGE_MSG"); fs::write(merge_msg_path, merge_msg).with_context(|| { format!( "Encountered an error when restoring {}.", merge_msg_path.display() ) }) }); restore_merge_head_result?; restore_merge_mode_result?; restore_merge_msg_result?; Ok(()) } fn read_file_to_string<P: AsRef<Path>>(file: P) -> Result<Option<String>> { match fs::read_to_string(file) { Ok(contents) => Ok(Some(contents)), Err(error) if error.kind() == NotFound => Ok(None), Err(error) => Err(anyhow!(error)), } } fn delete_files<P: AsRef<Path>>(files: &Vec<P>) -> Result<()> { for file in files.iter() { fs::remove_file(file).with_context(|| { format!( "Encountered error when deleting {}.", file.as_ref().display() ) })?; } Ok(()) } } #[derive(Debug)] pub struct Snapshot { pub staged_files: Vec<PathBuf>, backup_stash: Option<Stash>, unstaged_diff: Option<Vec<u8>>, } #[derive(Debug)] struct Stash { stash_id: Oid, merge_status: MergeStatus, } #[derive(Debug)] struct MergeStatus { merge_head: Option<String>, merge_mode: Option<String>, merge_msg: Option<String>, }
{ // When strict hash verification is disabled, it means libgit2 will not // compute the "object id" of Git objects (which is a SHA-1 hash) after // reading them to verify they match the object ids being used to look // them up. This improves performance, and I don't have in front of me // a concrete example where this is necessary to prevent data loss. If // one becomes obvious, then we should make this configurable. // git2::opts::strict_hash_verification(false); let repository = Repository::open_from_env() .with_context(|| "Encountered an error when opening the Git repository.")?; Ok(Self { repository }) }
identifier_body
git.rs
use anyhow::{anyhow, Context, Result}; use git2::{ build::CheckoutBuilder, ApplyLocation, Delta, Diff, DiffFormat, DiffOptions, ErrorCode, IndexAddOption, Oid, Repository, ResetType, Signature, StashApplyOptions, Time, }; use itertools::Itertools; use std::cell::RefCell; use std::collections::HashSet; use std::fs; use std::hash::Hash; use std::io::ErrorKind::NotFound; use std::iter::FromIterator; use std::path::{Path, PathBuf}; /// An abstraction over a Git repository providing complex behavior needed for /// applying changes to staged files safely. pub struct GitRepository { repository: Repository, } impl GitRepository { /// Attempts to open an already-existing repository. /// /// If the $GIT_DIR environment variable is set, this uses it to locate the /// Git repository. Otherwise, this searches up the directory tree from the /// current directory to find the repository. pub fn open() -> Result<Self> { // When strict hash verification is disabled, it means libgit2 will not // compute the "object id" of Git objects (which is a SHA-1 hash) after // reading them to verify they match the object ids being used to look // them up. This improves performance, and I don't have in front of me // a concrete example where this is necessary to prevent data loss. If // one becomes obvious, then we should make this configurable. // git2::opts::strict_hash_verification(false); let repository = Repository::open_from_env() .with_context(|| "Encountered an error when opening the Git repository.")?; Ok(Self { repository }) } pub fn save_snapshot(&mut self, staged_files: Vec<PathBuf>) -> Result<Snapshot> { let inner = || -> Result<Snapshot> { let deleted_files = self.get_deleted_files()?; let unstaged_diff = self.save_unstaged_diff()?; let backup_stash = self.save_snapshot_stash()?; // Because `git stash` restores the HEAD commit, it brings back uncommitted // deleted files. We need to clear them before creating our snapshot. GitRepository::delete_files(&deleted_files)?; self.hide_partially_staged_changes()?; Ok(Snapshot { backup_stash, staged_files, unstaged_diff, }) }; inner().with_context(|| "Encountered an error when saving a snapshot.") } pub fn apply_modifications(&mut self, snapshot: &Snapshot) -> Result<()> { self.stage_modifications(snapshot)?; if self.get_staged_files()?.is_empty() { return Err(anyhow!("Prevented an empty git commit.")); } if let Some(raw_diff) = &snapshot.unstaged_diff { let unstaged_diff = Diff::from_buffer(raw_diff)?; self.merge_modifications(unstaged_diff)?; } Ok(()) } pub fn restore_snapshot(&mut self, snapshot: &Snapshot) -> Result<()> { let mut inner = || -> Result<()> { self.hard_reset()?; if let Some(backup_stash) = &snapshot.backup_stash { self.apply_stash(&backup_stash.stash_id)?; self.restore_merge_status(&backup_stash.merge_status)?; } Ok(()) }; inner().with_context(|| "Encountered an error when restoring snapshot after another error.") } pub fn clean_snapshot(&mut self, snapshot: Snapshot) -> Result<()> { let inner = || -> Result<()> { if let Some(backup_stash) = snapshot.backup_stash { let stash_index = self .get_stash_index_from_id(&backup_stash.stash_id)? .ok_or_else(|| { anyhow!( "Could not find a backup stash with id {}.", &backup_stash.stash_id ) })?; self.repository.stash_drop(stash_index)?; } Ok(()) }; inner().with_context(|| { "Encountered an error when cleaning snapshot. You might find a stash entry \ in the stash list." }) } fn stage_modifications(&mut self, snapshot: &Snapshot) -> Result<()> { let mut index = self.repository.index()?; index.add_all( &snapshot.staged_files, IndexAddOption::DEFAULT | IndexAddOption::DISABLE_PATHSPEC_MATCH, None, )?; index.write()?; Ok(()) } fn merge_modifications(&self, unstaged_diff: Diff) -> Result<()> { self.repository .apply(&unstaged_diff, ApplyLocation::WorkDir, None) .with_context(|| "Unstaged changes could not be restored due to a merge conflict.") } fn hard_reset(&self) -> Result<()> { let head = self.repository.head()?.peel_to_commit()?; self.repository .reset(head.as_object(), ResetType::Hard, None) .map_err(|error| anyhow!(error)) } fn get_stash_index_from_id(&mut self, stash_id: &Oid) -> Result<Option<usize>> { // It would be much better if libgit2 accepted a stash Oid // instead of an index from the stash list. let ref_stash_index = RefCell::new(None); self.repository.stash_foreach(|index, _, oid| { if oid == stash_id { *ref_stash_index.borrow_mut() = Some(index); false } else { true } })?; // Copy the data out of the RefCell. let stash_index = match *ref_stash_index.borrow() { Some(index) => Some(index), None => None, }; Ok(stash_index) } fn apply_stash(&mut self, stash_id: &Oid) -> Result<()> { let stash_index = self .get_stash_index_from_id(stash_id)? .ok_or_else(|| anyhow!("Could not find a backup stash with id {}.", stash_id))?; self.repository.stash_apply( stash_index, Some(StashApplyOptions::default().reinstantiate_index()), )?; Ok(()) } fn save_unstaged_diff(&self) -> Result<Option<Vec<u8>>> { let partially_staged_files = self.get_partially_staged_files(true)?; if partially_staged_files.is_empty() { return Ok(None); } let mut diff_options = DiffOptions::new(); diff_options.show_binary(true); for file in partially_staged_files.iter() { diff_options.pathspec(file); } let unstaged_diff = self .repository .diff_index_to_workdir(None, Some(&mut diff_options))?; // The Diff created by diff_index_to_workdir is owned by the repository. // It means storing this diff separately isn't possible, and it is also // difficult to store it along with the repository together in a struct, // because that struct then will have a self reference between its diff // and its repository. // // I'm not comfortable enough with ownership to understand the correct // way to work around this, so the current approach that I'm taking is // to copy the diff out into a buffer. This is not the most performant. // // For updates about this issue, we can keep tabs on // // https://github.com/rust-lang/git2-rs/issues/622 // fn copy_diff(diff: &Diff) -> Result<Vec<u8>> { let mut buffer = vec![]; diff.print(DiffFormat::Patch, |_, _, line| { let origin = line.origin(); if origin == '+' || origin == '-' || origin == ' ' { buffer.push(origin as u8); } buffer.append(&mut line.content().to_vec()); true })?; Ok(buffer) } Ok(Some(copy_diff(&unstaged_diff)?)) } fn hide_partially_staged_changes(&self) -> Result<()> { let partially_staged_files = self.get_partially_staged_files(false)?; if partially_staged_files.is_empty() { return Ok(()); } let mut checkout_options = CheckoutBuilder::new(); checkout_options.force(); checkout_options.update_index(false); for file in partially_staged_files.iter() { checkout_options.path(file); } self.repository .checkout_index(None, Some(&mut checkout_options))?; Ok(()) } pub fn get_staged_files(&self) -> Result<Vec<PathBuf>> { let head_tree = match self.repository.head() { Ok(head) => Ok(Some(head.peel_to_tree()?)), Err(error) if error.code() == ErrorCode::UnbornBranch => Ok(None), Err(error) => Err(error), }?; let staged_files = self .repository .diff_tree_to_index(head_tree.as_ref(), None, None)? .deltas() .flat_map(|delta| { if delta.old_file().path() == delta.new_file().path() { vec![delta.old_file().path()] } else { vec![delta.old_file().path(), delta.new_file().path()] } }) .filter_map(std::convert::identity) .map(Path::to_path_buf) .collect(); Ok(staged_files) } fn get_partially_staged_files(&self, include_from_files: bool) -> Result<HashSet<PathBuf>> { let staged_files = HashSet::from_iter(self.get_staged_files()?); let unstaged_files = HashSet::from_iter( self.repository .diff_index_to_workdir(None, Some(DiffOptions::default().show_binary(true)))? .deltas() .flat_map(|delta| { if include_from_files { vec![delta.old_file().path(), delta.new_file().path()] } else { vec![delta.new_file().path()] } }) .filter_map(std::convert::identity) .map(Path::to_path_buf), ); fn intersect<P: Eq + Hash>(one: HashSet<P>, two: &HashSet<P>) -> HashSet<P> { one.into_iter().filter(|p| two.contains(p)).collect() } Ok(intersect(staged_files, &unstaged_files)) } fn get_deleted_files(&self) -> Result<Vec<PathBuf>> { let deleted_files = self .repository .diff_index_to_workdir(None, None)? .deltas() .filter(|delta| delta.status() == Delta::Deleted) .filter_map(|delta| delta.old_file().path()) .map(Path::to_path_buf) .collect_vec(); Ok(deleted_files) } fn save_snapshot_stash(&mut self) -> Result<Option<Stash>> { if self.repository.is_empty()? { return Ok(None); } fn create_signature<'a>() -> Result<Signature<'a>> { // Because this time is only used to create a dummy signature to // make the stash_save method happy, we don't need to use a real // time, which skips some calls to the kernel. // let time = Time::new(0, 0); Signature::new("Dummy", "dummy@example.com", &time) .with_context(|| "Encountered an error when creating dummy authorship information.") } // Save state when in the middle of a merge prior to stashing changes in // the working directory so that we can restore it afterward. // let merge_status = self.save_merge_status()?; let signature = create_signature()?; let stash_result = self .repository .stash_create(&signature, None, None); if let Ok(stash_id) = stash_result { self.repository.stash_store(&stash_id, Some("offstage backup"))?; } match stash_result { Ok(stash_id) => Ok(Some(Stash { stash_id, merge_status, })), Err(error) if error.code() == ErrorCode::NotFound => Ok(None), Err(error) => Err(anyhow!(error) .context("Encountered an error when stashing a backup of the working directory.")), } } fn save_merge_status(&self) -> Result<MergeStatus> { let merge_head_path = &self.repository.path().join("MERGE_HEAD"); let merge_head = Self::read_file_to_string(merge_head_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_head_path.display() ) })?; let merge_mode_path = &self.repository.path().join("MERGE_MODE"); let merge_mode = Self::read_file_to_string(merge_mode_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_mode_path.display() ) })?; let merge_msg_path = &self.repository.path().join("MERGE_MSG"); let merge_msg = Self::read_file_to_string(merge_msg_path).with_context(|| { format!( "Encountered an error when saving {}.", merge_msg_path.display() ) })?; Ok(MergeStatus { merge_head, merge_mode, merge_msg, }) } fn restore_merge_status(&self, merge_status: &MergeStatus) -> Result<()> { // Tries to restore all files before returning the first error if one exists. let restore_merge_head_result = merge_status .merge_head .as_ref() .map_or(Ok(()), |merge_head| { let merge_head_path = &self.repository.path().join("MERGE_HEAD"); fs::write(merge_head_path, merge_head).with_context(|| { format!( "Encountered an error when restoring {}.", merge_head_path.display() ) }) }); let restore_merge_mode_result = merge_status .merge_mode .as_ref() .map_or(Ok(()), |merge_mode| { let merge_mode_path = &self.repository.path().join("MERGE_MODE"); fs::write(merge_mode_path, merge_mode).with_context(|| { format!( "Encountered an error when restoring {}.", &merge_mode_path.display() ) }) }); let restore_merge_msg_result = merge_status.merge_msg.as_ref().map_or(Ok(()), |merge_msg| { let merge_msg_path = &self.repository.path().join("MERGE_MSG"); fs::write(merge_msg_path, merge_msg).with_context(|| { format!( "Encountered an error when restoring {}.", merge_msg_path.display() ) }) }); restore_merge_head_result?; restore_merge_mode_result?; restore_merge_msg_result?; Ok(()) } fn read_file_to_string<P: AsRef<Path>>(file: P) -> Result<Option<String>> { match fs::read_to_string(file) { Ok(contents) => Ok(Some(contents)), Err(error) if error.kind() == NotFound => Ok(None), Err(error) => Err(anyhow!(error)), } } fn delete_files<P: AsRef<Path>>(files: &Vec<P>) -> Result<()> { for file in files.iter() { fs::remove_file(file).with_context(|| { format!(
Ok(()) } } #[derive(Debug)] pub struct Snapshot { pub staged_files: Vec<PathBuf>, backup_stash: Option<Stash>, unstaged_diff: Option<Vec<u8>>, } #[derive(Debug)] struct Stash { stash_id: Oid, merge_status: MergeStatus, } #[derive(Debug)] struct MergeStatus { merge_head: Option<String>, merge_mode: Option<String>, merge_msg: Option<String>, }
"Encountered error when deleting {}.", file.as_ref().display() ) })?; }
random_line_split
client.rs
//! Common client functionalities. use crate::rdsys; use crate::rdsys::types::*; use std::ffi::{CStr, CString}; use std::mem; use std::os::raw::c_char; use std::os::raw::c_void; use std::ptr; use std::slice; use std::string::ToString; use std::time::Duration; use serde_json; use crate::config::{ClientConfig, NativeClientConfig, RDKafkaLogLevel}; use crate::error::{IsError, KafkaError, KafkaResult}; use crate::groups::GroupList; use crate::metadata::Metadata; use crate::statistics::Statistics; use crate::util::{timeout_to_ms, ErrBuf}; /// Client-level context /// /// Each client (consumers and producers included) has a context object that can be used to /// customize its behavior. Implementing `ClientContext` enables the customization of /// methods common to all clients, while `ProducerContext` and `ConsumerContext` are specific to /// producers and consumers. Refer to the list of methods to see which callbacks can currently /// be overridden. Implementations of `ClientContext` must be thread safe, as they might be owned by /// multiple threads. pub trait ClientContext: Send + Sync { /// Receives log lines from librdkafka. fn log(&self, level: RDKafkaLogLevel, fac: &str, log_message: &str) { match level { RDKafkaLogLevel::Emerg | RDKafkaLogLevel::Alert | RDKafkaLogLevel::Critical | RDKafkaLogLevel::Error => error!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Warning => warn!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Notice => info!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Info => info!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Debug => debug!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), } } /// Receives the statistics of the librdkafka client. To enable, the /// "statistics.interval.ms" configuration parameter must be specified. fn stats(&self, statistics: Statistics) { info!("Client stats: {:?}", statistics); } /// Receives global errors from the librdkafka client. fn error(&self, error: KafkaError, reason: &str) { error!("librdkafka: {}: {}", error, reason); } // NOTE: when adding a new method, remember to add it to the FutureProducerContext as well. // https://github.com/rust-lang/rfcs/pull/1406 will maybe help in the future. } /// An empty `ClientContext` that can be used when no context is needed. Default /// callback implementations will be used. #[derive(Clone, Default)] pub struct DefaultClientContext; impl ClientContext for DefaultClientContext {} // // ********** CLIENT ********** // /// A native rdkafka-sys client. This struct shouldn't be used directly. Use higher level `Client` /// or producers and consumers. pub struct NativeClient { ptr: *mut RDKafka, } // The library is completely thread safe, according to the documentation. unsafe impl Sync for NativeClient {} unsafe impl Send for NativeClient {} impl NativeClient { /// Wraps a pointer to an RDKafka object and returns a new NativeClient. pub(crate) unsafe fn from_ptr(ptr: *mut RDKafka) -> NativeClient { NativeClient { ptr } } /// Returns the wrapped pointer to RDKafka. pub fn ptr(&self) -> *mut RDKafka { self.ptr } } impl Drop for NativeClient { fn drop(&mut self) { trace!("Destroying client: {:p}", self.ptr); unsafe { rdsys::rd_kafka_destroy(self.ptr); } trace!("Client destroyed: {:?}", self.ptr); } } /// A low level rdkafka client. This client shouldn't be used directly. The producer and consumer modules /// provide different producer and consumer implementations based on top of `Client` that can be /// used instead. pub struct Client<C: ClientContext = DefaultClientContext> { native: NativeClient, context: Box<C>, } impl<C: ClientContext> Client<C> { /// Creates a new `Client` given a configuration, a client type and a context. pub fn new( config: &ClientConfig, native_config: NativeClientConfig, rd_kafka_type: RDKafkaType, context: C, ) -> KafkaResult<Client<C>> { let mut err_buf = ErrBuf::new(); let mut boxed_context = Box::new(context); unsafe { rdsys::rd_kafka_conf_set_opaque( native_config.ptr(), (&mut *boxed_context) as *mut C as *mut c_void, ) }; unsafe { rdsys::rd_kafka_conf_set_log_cb(native_config.ptr(), Some(native_log_cb::<C>)) }; unsafe { rdsys::rd_kafka_conf_set_stats_cb(native_config.ptr(), Some(native_stats_cb::<C>)) }; unsafe { rdsys::rd_kafka_conf_set_error_cb(native_config.ptr(), Some(native_error_cb::<C>)) }; let client_ptr = unsafe { rdsys::rd_kafka_new( rd_kafka_type, native_config.ptr_move(), err_buf.as_mut_ptr(), err_buf.len(), ) }; trace!("Create new librdkafka client {:p}", client_ptr); if client_ptr.is_null() { return Err(KafkaError::ClientCreation(err_buf.to_string())); } unsafe { rdsys::rd_kafka_set_log_level(client_ptr, config.log_level as i32) }; Ok(Client { native: unsafe { NativeClient::from_ptr(client_ptr) }, context: boxed_context, }) } /// Returns a reference to the native rdkafka-sys client. pub fn native_client(&self) -> &NativeClient { &self.native } /// Returns a pointer to the native rdkafka-sys client. pub fn native_ptr(&self) -> *mut RDKafka { self.native.ptr } /// Returns a reference to the context. pub fn context(&self) -> &C { self.context.as_ref() } /// Returns the metadata information for the specified topic, or for all topics in the cluster /// if no topic is specified. pub fn fetch_metadata<T: Into<Option<Duration>>>( &self, topic: Option<&str>, timeout: T, ) -> KafkaResult<Metadata> { let mut metadata_ptr: *const RDKafkaMetadata = ptr::null_mut(); let (flag, native_topic) = if let Some(topic_name) = topic { (0, Some(self.native_topic(topic_name)?)) } else { (1, None) }; trace!("Starting metadata fetch"); let ret = unsafe { rdsys::rd_kafka_metadata( self.native_ptr(), flag, native_topic .map(|t| t.ptr()) .unwrap_or_else(NativeTopic::null), &mut metadata_ptr as *mut *const RDKafkaMetadata, timeout_to_ms(timeout), ) }; trace!("Metadata fetch completed"); if ret.is_error() { return Err(KafkaError::MetadataFetch(ret.into())); } Ok(unsafe { Metadata::from_ptr(metadata_ptr) }) } /// Returns high and low watermark for the specified topic and partition. pub fn fetch_watermarks<T: Into<Option<Duration>>>( &self, topic: &str, partition: i32, timeout: T, ) -> KafkaResult<(i64, i64)> { let mut low = -1; let mut high = -1; let topic_c = CString::new(topic.to_string())?; let ret = unsafe { rdsys::rd_kafka_query_watermark_offsets( self.native_ptr(), topic_c.as_ptr(), partition, &mut low as *mut i64, &mut high as *mut i64, timeout_to_ms(timeout), ) }; if ret.is_error() { return Err(KafkaError::MetadataFetch(ret.into())); } Ok((low, high)) } /// Returns the group membership information for the given group. If no group is /// specified, all groups will be returned. pub fn fetch_group_list<T: Into<Option<Duration>>>( &self, group: Option<&str>, timeout: T, ) -> KafkaResult<GroupList> { // Careful with group_c getting freed before time let group_c = CString::new(group.map_or("".to_string(), ToString::to_string))?; let group_c_ptr = if group.is_some() { group_c.as_ptr() } else { ptr::null_mut() };
let mut group_list_ptr: *const RDKafkaGroupList = ptr::null_mut(); trace!("Starting group list fetch"); let ret = unsafe { rdsys::rd_kafka_list_groups( self.native_ptr(), group_c_ptr, &mut group_list_ptr as *mut *const RDKafkaGroupList, timeout_to_ms(timeout), ) }; trace!("Group list fetch completed"); if ret.is_error() { return Err(KafkaError::GroupListFetch(ret.into())); } Ok(unsafe { GroupList::from_ptr(group_list_ptr) }) } /// Returns a NativeTopic from the current client. The NativeTopic shouldn't outlive the client /// it was generated from. fn native_topic(&self, topic: &str) -> KafkaResult<NativeTopic> { let topic_c = CString::new(topic.to_string())?; Ok(unsafe { NativeTopic::from_ptr(rdsys::rd_kafka_topic_new( self.native_ptr(), topic_c.as_ptr(), ptr::null_mut(), )) }) } /// Returns a NativeQueue from the current client. The NativeQueue shouldn't /// outlive the client it was generated from. pub(crate) fn new_native_queue(&self) -> NativeQueue { unsafe { NativeQueue::from_ptr(rdsys::rd_kafka_queue_new(self.native_ptr())) } } } struct NativeTopic { ptr: *mut RDKafkaTopic, } unsafe impl Send for NativeTopic {} unsafe impl Sync for NativeTopic {} impl NativeTopic { /// Wraps a pointer to an `RDKafkaTopic` object and returns a new `NativeTopic`. unsafe fn from_ptr(ptr: *mut RDKafkaTopic) -> NativeTopic { NativeTopic { ptr } } /// Returns the pointer to the librdkafka RDKafkaTopic structure. fn ptr(&self) -> *mut RDKafkaTopic { self.ptr } /// Returns a null pointer. fn null() -> *mut RDKafkaTopic { ptr::null::<u8>() as *mut RDKafkaTopic } } impl Drop for NativeTopic { fn drop(&mut self) { trace!("Destroying NativeTopic: {:?}", self.ptr); unsafe { rdsys::rd_kafka_topic_destroy(self.ptr); } trace!("NativeTopic destroyed: {:?}", self.ptr); } } pub(crate) struct NativeQueue { ptr: *mut RDKafkaQueue, } // The library is completely thread safe, according to the documentation. unsafe impl Sync for NativeQueue {} unsafe impl Send for NativeQueue {} impl NativeQueue { /// Wraps a pointer to an `RDKafkaQueue` object and returns a new /// `NativeQueue`. unsafe fn from_ptr(ptr: *mut RDKafkaQueue) -> NativeQueue { NativeQueue { ptr } } /// Returns the pointer to the librdkafka RDKafkaQueue structure. pub fn ptr(&self) -> *mut RDKafkaQueue { self.ptr } pub fn poll<T: Into<Option<Duration>>>(&self, t: T) -> *mut RDKafkaEvent { unsafe { rdsys::rd_kafka_queue_poll(self.ptr, timeout_to_ms(t)) } } } impl Drop for NativeQueue { fn drop(&mut self) { trace!("Destroying queue: {:?}", self.ptr); unsafe { rdsys::rd_kafka_queue_destroy(self.ptr); } trace!("Queue destroyed: {:?}", self.ptr); } } pub(crate) unsafe extern "C" fn native_log_cb<C: ClientContext>( client: *const RDKafka, level: i32, fac: *const c_char, buf: *const c_char, ) { let fac = CStr::from_ptr(fac).to_string_lossy(); let log_message = CStr::from_ptr(buf).to_string_lossy(); let context = Box::from_raw(rdsys::rd_kafka_opaque(client) as *mut C); (*context).log( RDKafkaLogLevel::from_int(level), fac.trim(), log_message.trim(), ); mem::forget(context); // Do not free the context } pub(crate) unsafe extern "C" fn native_stats_cb<C: ClientContext>( _conf: *mut RDKafka, json: *mut c_char, json_len: usize, opaque: *mut c_void, ) -> i32 { let context = Box::from_raw(opaque as *mut C); let mut bytes_vec = Vec::new(); bytes_vec.extend_from_slice(slice::from_raw_parts(json as *mut u8, json_len)); let json_string = CString::from_vec_unchecked(bytes_vec).into_string(); match json_string { Ok(json) => match serde_json::from_str(&json) { Ok(stats) => (*context).stats(stats), Err(e) => error!("Could not parse statistics JSON: {}", e), }, Err(e) => error!("Statistics JSON string is not UTF-8: {:?}", e), } mem::forget(context); // Do not free the context 0 // librdkafka will free the json buffer } pub(crate) unsafe extern "C" fn native_error_cb<C: ClientContext>( _client: *mut RDKafka, err: i32, reason: *const c_char, opaque: *mut c_void, ) { let err = rdsys::primitive_to_rd_kafka_resp_err_t(err) .expect("global error not an rd_kafka_resp_err_t"); let error = KafkaError::Global(err.into()); let reason = CStr::from_ptr(reason).to_string_lossy(); let context = Box::from_raw(opaque as *mut C); (*context).error(error, reason.trim()); mem::forget(context); // Do not free the context } #[cfg(test)] mod tests { // Just call everything to test there no panics by default, behavior // is tested in the integrations tests. use super::*; use crate::config::ClientConfig; #[test] fn test_client() { let config = ClientConfig::new(); let native_config = config.create_native_config().unwrap(); let client = Client::new( &config, native_config, RDKafkaType::RD_KAFKA_PRODUCER, DefaultClientContext, ) .unwrap(); assert!(!client.native_ptr().is_null()); } }
random_line_split
client.rs
//! Common client functionalities. use crate::rdsys; use crate::rdsys::types::*; use std::ffi::{CStr, CString}; use std::mem; use std::os::raw::c_char; use std::os::raw::c_void; use std::ptr; use std::slice; use std::string::ToString; use std::time::Duration; use serde_json; use crate::config::{ClientConfig, NativeClientConfig, RDKafkaLogLevel}; use crate::error::{IsError, KafkaError, KafkaResult}; use crate::groups::GroupList; use crate::metadata::Metadata; use crate::statistics::Statistics; use crate::util::{timeout_to_ms, ErrBuf}; /// Client-level context /// /// Each client (consumers and producers included) has a context object that can be used to /// customize its behavior. Implementing `ClientContext` enables the customization of /// methods common to all clients, while `ProducerContext` and `ConsumerContext` are specific to /// producers and consumers. Refer to the list of methods to see which callbacks can currently /// be overridden. Implementations of `ClientContext` must be thread safe, as they might be owned by /// multiple threads. pub trait ClientContext: Send + Sync { /// Receives log lines from librdkafka. fn log(&self, level: RDKafkaLogLevel, fac: &str, log_message: &str) { match level { RDKafkaLogLevel::Emerg | RDKafkaLogLevel::Alert | RDKafkaLogLevel::Critical | RDKafkaLogLevel::Error => error!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Warning => warn!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Notice => info!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Info => info!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Debug => debug!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), } } /// Receives the statistics of the librdkafka client. To enable, the /// "statistics.interval.ms" configuration parameter must be specified. fn stats(&self, statistics: Statistics) { info!("Client stats: {:?}", statistics); } /// Receives global errors from the librdkafka client. fn error(&self, error: KafkaError, reason: &str) { error!("librdkafka: {}: {}", error, reason); } // NOTE: when adding a new method, remember to add it to the FutureProducerContext as well. // https://github.com/rust-lang/rfcs/pull/1406 will maybe help in the future. } /// An empty `ClientContext` that can be used when no context is needed. Default /// callback implementations will be used. #[derive(Clone, Default)] pub struct DefaultClientContext; impl ClientContext for DefaultClientContext {} // // ********** CLIENT ********** // /// A native rdkafka-sys client. This struct shouldn't be used directly. Use higher level `Client` /// or producers and consumers. pub struct NativeClient { ptr: *mut RDKafka, } // The library is completely thread safe, according to the documentation. unsafe impl Sync for NativeClient {} unsafe impl Send for NativeClient {} impl NativeClient { /// Wraps a pointer to an RDKafka object and returns a new NativeClient. pub(crate) unsafe fn from_ptr(ptr: *mut RDKafka) -> NativeClient { NativeClient { ptr } } /// Returns the wrapped pointer to RDKafka. pub fn ptr(&self) -> *mut RDKafka { self.ptr } } impl Drop for NativeClient { fn drop(&mut self) { trace!("Destroying client: {:p}", self.ptr); unsafe { rdsys::rd_kafka_destroy(self.ptr); } trace!("Client destroyed: {:?}", self.ptr); } } /// A low level rdkafka client. This client shouldn't be used directly. The producer and consumer modules /// provide different producer and consumer implementations based on top of `Client` that can be /// used instead. pub struct Client<C: ClientContext = DefaultClientContext> { native: NativeClient, context: Box<C>, } impl<C: ClientContext> Client<C> { /// Creates a new `Client` given a configuration, a client type and a context. pub fn new( config: &ClientConfig, native_config: NativeClientConfig, rd_kafka_type: RDKafkaType, context: C, ) -> KafkaResult<Client<C>> { let mut err_buf = ErrBuf::new(); let mut boxed_context = Box::new(context); unsafe { rdsys::rd_kafka_conf_set_opaque( native_config.ptr(), (&mut *boxed_context) as *mut C as *mut c_void, ) }; unsafe { rdsys::rd_kafka_conf_set_log_cb(native_config.ptr(), Some(native_log_cb::<C>)) }; unsafe { rdsys::rd_kafka_conf_set_stats_cb(native_config.ptr(), Some(native_stats_cb::<C>)) }; unsafe { rdsys::rd_kafka_conf_set_error_cb(native_config.ptr(), Some(native_error_cb::<C>)) }; let client_ptr = unsafe { rdsys::rd_kafka_new( rd_kafka_type, native_config.ptr_move(), err_buf.as_mut_ptr(), err_buf.len(), ) }; trace!("Create new librdkafka client {:p}", client_ptr); if client_ptr.is_null() { return Err(KafkaError::ClientCreation(err_buf.to_string())); } unsafe { rdsys::rd_kafka_set_log_level(client_ptr, config.log_level as i32) }; Ok(Client { native: unsafe { NativeClient::from_ptr(client_ptr) }, context: boxed_context, }) } /// Returns a reference to the native rdkafka-sys client. pub fn native_client(&self) -> &NativeClient { &self.native } /// Returns a pointer to the native rdkafka-sys client. pub fn native_ptr(&self) -> *mut RDKafka { self.native.ptr } /// Returns a reference to the context. pub fn context(&self) -> &C { self.context.as_ref() } /// Returns the metadata information for the specified topic, or for all topics in the cluster /// if no topic is specified. pub fn fetch_metadata<T: Into<Option<Duration>>>( &self, topic: Option<&str>, timeout: T, ) -> KafkaResult<Metadata> { let mut metadata_ptr: *const RDKafkaMetadata = ptr::null_mut(); let (flag, native_topic) = if let Some(topic_name) = topic { (0, Some(self.native_topic(topic_name)?)) } else { (1, None) }; trace!("Starting metadata fetch"); let ret = unsafe { rdsys::rd_kafka_metadata( self.native_ptr(), flag, native_topic .map(|t| t.ptr()) .unwrap_or_else(NativeTopic::null), &mut metadata_ptr as *mut *const RDKafkaMetadata, timeout_to_ms(timeout), ) }; trace!("Metadata fetch completed"); if ret.is_error() { return Err(KafkaError::MetadataFetch(ret.into())); } Ok(unsafe { Metadata::from_ptr(metadata_ptr) }) } /// Returns high and low watermark for the specified topic and partition. pub fn fetch_watermarks<T: Into<Option<Duration>>>( &self, topic: &str, partition: i32, timeout: T, ) -> KafkaResult<(i64, i64)> { let mut low = -1; let mut high = -1; let topic_c = CString::new(topic.to_string())?; let ret = unsafe { rdsys::rd_kafka_query_watermark_offsets( self.native_ptr(), topic_c.as_ptr(), partition, &mut low as *mut i64, &mut high as *mut i64, timeout_to_ms(timeout), ) }; if ret.is_error() { return Err(KafkaError::MetadataFetch(ret.into())); } Ok((low, high)) } /// Returns the group membership information for the given group. If no group is /// specified, all groups will be returned. pub fn fetch_group_list<T: Into<Option<Duration>>>( &self, group: Option<&str>, timeout: T, ) -> KafkaResult<GroupList> { // Careful with group_c getting freed before time let group_c = CString::new(group.map_or("".to_string(), ToString::to_string))?; let group_c_ptr = if group.is_some() { group_c.as_ptr() } else { ptr::null_mut() }; let mut group_list_ptr: *const RDKafkaGroupList = ptr::null_mut(); trace!("Starting group list fetch"); let ret = unsafe { rdsys::rd_kafka_list_groups( self.native_ptr(), group_c_ptr, &mut group_list_ptr as *mut *const RDKafkaGroupList, timeout_to_ms(timeout), ) }; trace!("Group list fetch completed"); if ret.is_error() { return Err(KafkaError::GroupListFetch(ret.into())); } Ok(unsafe { GroupList::from_ptr(group_list_ptr) }) } /// Returns a NativeTopic from the current client. The NativeTopic shouldn't outlive the client /// it was generated from. fn native_topic(&self, topic: &str) -> KafkaResult<NativeTopic> { let topic_c = CString::new(topic.to_string())?; Ok(unsafe { NativeTopic::from_ptr(rdsys::rd_kafka_topic_new( self.native_ptr(), topic_c.as_ptr(), ptr::null_mut(), )) }) } /// Returns a NativeQueue from the current client. The NativeQueue shouldn't /// outlive the client it was generated from. pub(crate) fn new_native_queue(&self) -> NativeQueue { unsafe { NativeQueue::from_ptr(rdsys::rd_kafka_queue_new(self.native_ptr())) } } } struct NativeTopic { ptr: *mut RDKafkaTopic, } unsafe impl Send for NativeTopic {} unsafe impl Sync for NativeTopic {} impl NativeTopic { /// Wraps a pointer to an `RDKafkaTopic` object and returns a new `NativeTopic`. unsafe fn from_ptr(ptr: *mut RDKafkaTopic) -> NativeTopic { NativeTopic { ptr } } /// Returns the pointer to the librdkafka RDKafkaTopic structure. fn ptr(&self) -> *mut RDKafkaTopic { self.ptr } /// Returns a null pointer. fn null() -> *mut RDKafkaTopic { ptr::null::<u8>() as *mut RDKafkaTopic } } impl Drop for NativeTopic { fn drop(&mut self) { trace!("Destroying NativeTopic: {:?}", self.ptr); unsafe { rdsys::rd_kafka_topic_destroy(self.ptr); } trace!("NativeTopic destroyed: {:?}", self.ptr); } } pub(crate) struct NativeQueue { ptr: *mut RDKafkaQueue, } // The library is completely thread safe, according to the documentation. unsafe impl Sync for NativeQueue {} unsafe impl Send for NativeQueue {} impl NativeQueue { /// Wraps a pointer to an `RDKafkaQueue` object and returns a new /// `NativeQueue`. unsafe fn from_ptr(ptr: *mut RDKafkaQueue) -> NativeQueue { NativeQueue { ptr } } /// Returns the pointer to the librdkafka RDKafkaQueue structure. pub fn ptr(&self) -> *mut RDKafkaQueue { self.ptr } pub fn poll<T: Into<Option<Duration>>>(&self, t: T) -> *mut RDKafkaEvent { unsafe { rdsys::rd_kafka_queue_poll(self.ptr, timeout_to_ms(t)) } } } impl Drop for NativeQueue { fn drop(&mut self) { trace!("Destroying queue: {:?}", self.ptr); unsafe { rdsys::rd_kafka_queue_destroy(self.ptr); } trace!("Queue destroyed: {:?}", self.ptr); } } pub(crate) unsafe extern "C" fn native_log_cb<C: ClientContext>( client: *const RDKafka, level: i32, fac: *const c_char, buf: *const c_char, )
pub(crate) unsafe extern "C" fn native_stats_cb<C: ClientContext>( _conf: *mut RDKafka, json: *mut c_char, json_len: usize, opaque: *mut c_void, ) -> i32 { let context = Box::from_raw(opaque as *mut C); let mut bytes_vec = Vec::new(); bytes_vec.extend_from_slice(slice::from_raw_parts(json as *mut u8, json_len)); let json_string = CString::from_vec_unchecked(bytes_vec).into_string(); match json_string { Ok(json) => match serde_json::from_str(&json) { Ok(stats) => (*context).stats(stats), Err(e) => error!("Could not parse statistics JSON: {}", e), }, Err(e) => error!("Statistics JSON string is not UTF-8: {:?}", e), } mem::forget(context); // Do not free the context 0 // librdkafka will free the json buffer } pub(crate) unsafe extern "C" fn native_error_cb<C: ClientContext>( _client: *mut RDKafka, err: i32, reason: *const c_char, opaque: *mut c_void, ) { let err = rdsys::primitive_to_rd_kafka_resp_err_t(err) .expect("global error not an rd_kafka_resp_err_t"); let error = KafkaError::Global(err.into()); let reason = CStr::from_ptr(reason).to_string_lossy(); let context = Box::from_raw(opaque as *mut C); (*context).error(error, reason.trim()); mem::forget(context); // Do not free the context } #[cfg(test)] mod tests { // Just call everything to test there no panics by default, behavior // is tested in the integrations tests. use super::*; use crate::config::ClientConfig; #[test] fn test_client() { let config = ClientConfig::new(); let native_config = config.create_native_config().unwrap(); let client = Client::new( &config, native_config, RDKafkaType::RD_KAFKA_PRODUCER, DefaultClientContext, ) .unwrap(); assert!(!client.native_ptr().is_null()); } }
{ let fac = CStr::from_ptr(fac).to_string_lossy(); let log_message = CStr::from_ptr(buf).to_string_lossy(); let context = Box::from_raw(rdsys::rd_kafka_opaque(client) as *mut C); (*context).log( RDKafkaLogLevel::from_int(level), fac.trim(), log_message.trim(), ); mem::forget(context); // Do not free the context }
identifier_body
client.rs
//! Common client functionalities. use crate::rdsys; use crate::rdsys::types::*; use std::ffi::{CStr, CString}; use std::mem; use std::os::raw::c_char; use std::os::raw::c_void; use std::ptr; use std::slice; use std::string::ToString; use std::time::Duration; use serde_json; use crate::config::{ClientConfig, NativeClientConfig, RDKafkaLogLevel}; use crate::error::{IsError, KafkaError, KafkaResult}; use crate::groups::GroupList; use crate::metadata::Metadata; use crate::statistics::Statistics; use crate::util::{timeout_to_ms, ErrBuf}; /// Client-level context /// /// Each client (consumers and producers included) has a context object that can be used to /// customize its behavior. Implementing `ClientContext` enables the customization of /// methods common to all clients, while `ProducerContext` and `ConsumerContext` are specific to /// producers and consumers. Refer to the list of methods to see which callbacks can currently /// be overridden. Implementations of `ClientContext` must be thread safe, as they might be owned by /// multiple threads. pub trait ClientContext: Send + Sync { /// Receives log lines from librdkafka. fn log(&self, level: RDKafkaLogLevel, fac: &str, log_message: &str) { match level { RDKafkaLogLevel::Emerg | RDKafkaLogLevel::Alert | RDKafkaLogLevel::Critical | RDKafkaLogLevel::Error => error!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Warning => warn!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Notice => info!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Info => info!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), RDKafkaLogLevel::Debug => debug!(target: "librdkafka", "librdkafka: {} {}", fac, log_message), } } /// Receives the statistics of the librdkafka client. To enable, the /// "statistics.interval.ms" configuration parameter must be specified. fn stats(&self, statistics: Statistics) { info!("Client stats: {:?}", statistics); } /// Receives global errors from the librdkafka client. fn error(&self, error: KafkaError, reason: &str) { error!("librdkafka: {}: {}", error, reason); } // NOTE: when adding a new method, remember to add it to the FutureProducerContext as well. // https://github.com/rust-lang/rfcs/pull/1406 will maybe help in the future. } /// An empty `ClientContext` that can be used when no context is needed. Default /// callback implementations will be used. #[derive(Clone, Default)] pub struct DefaultClientContext; impl ClientContext for DefaultClientContext {} // // ********** CLIENT ********** // /// A native rdkafka-sys client. This struct shouldn't be used directly. Use higher level `Client` /// or producers and consumers. pub struct NativeClient { ptr: *mut RDKafka, } // The library is completely thread safe, according to the documentation. unsafe impl Sync for NativeClient {} unsafe impl Send for NativeClient {} impl NativeClient { /// Wraps a pointer to an RDKafka object and returns a new NativeClient. pub(crate) unsafe fn from_ptr(ptr: *mut RDKafka) -> NativeClient { NativeClient { ptr } } /// Returns the wrapped pointer to RDKafka. pub fn ptr(&self) -> *mut RDKafka { self.ptr } } impl Drop for NativeClient { fn drop(&mut self) { trace!("Destroying client: {:p}", self.ptr); unsafe { rdsys::rd_kafka_destroy(self.ptr); } trace!("Client destroyed: {:?}", self.ptr); } } /// A low level rdkafka client. This client shouldn't be used directly. The producer and consumer modules /// provide different producer and consumer implementations based on top of `Client` that can be /// used instead. pub struct Client<C: ClientContext = DefaultClientContext> { native: NativeClient, context: Box<C>, } impl<C: ClientContext> Client<C> { /// Creates a new `Client` given a configuration, a client type and a context. pub fn new( config: &ClientConfig, native_config: NativeClientConfig, rd_kafka_type: RDKafkaType, context: C, ) -> KafkaResult<Client<C>> { let mut err_buf = ErrBuf::new(); let mut boxed_context = Box::new(context); unsafe { rdsys::rd_kafka_conf_set_opaque( native_config.ptr(), (&mut *boxed_context) as *mut C as *mut c_void, ) }; unsafe { rdsys::rd_kafka_conf_set_log_cb(native_config.ptr(), Some(native_log_cb::<C>)) }; unsafe { rdsys::rd_kafka_conf_set_stats_cb(native_config.ptr(), Some(native_stats_cb::<C>)) }; unsafe { rdsys::rd_kafka_conf_set_error_cb(native_config.ptr(), Some(native_error_cb::<C>)) }; let client_ptr = unsafe { rdsys::rd_kafka_new( rd_kafka_type, native_config.ptr_move(), err_buf.as_mut_ptr(), err_buf.len(), ) }; trace!("Create new librdkafka client {:p}", client_ptr); if client_ptr.is_null() { return Err(KafkaError::ClientCreation(err_buf.to_string())); } unsafe { rdsys::rd_kafka_set_log_level(client_ptr, config.log_level as i32) }; Ok(Client { native: unsafe { NativeClient::from_ptr(client_ptr) }, context: boxed_context, }) } /// Returns a reference to the native rdkafka-sys client. pub fn native_client(&self) -> &NativeClient { &self.native } /// Returns a pointer to the native rdkafka-sys client. pub fn native_ptr(&self) -> *mut RDKafka { self.native.ptr } /// Returns a reference to the context. pub fn context(&self) -> &C { self.context.as_ref() } /// Returns the metadata information for the specified topic, or for all topics in the cluster /// if no topic is specified. pub fn fetch_metadata<T: Into<Option<Duration>>>( &self, topic: Option<&str>, timeout: T, ) -> KafkaResult<Metadata> { let mut metadata_ptr: *const RDKafkaMetadata = ptr::null_mut(); let (flag, native_topic) = if let Some(topic_name) = topic { (0, Some(self.native_topic(topic_name)?)) } else { (1, None) }; trace!("Starting metadata fetch"); let ret = unsafe { rdsys::rd_kafka_metadata( self.native_ptr(), flag, native_topic .map(|t| t.ptr()) .unwrap_or_else(NativeTopic::null), &mut metadata_ptr as *mut *const RDKafkaMetadata, timeout_to_ms(timeout), ) }; trace!("Metadata fetch completed"); if ret.is_error() { return Err(KafkaError::MetadataFetch(ret.into())); } Ok(unsafe { Metadata::from_ptr(metadata_ptr) }) } /// Returns high and low watermark for the specified topic and partition. pub fn
<T: Into<Option<Duration>>>( &self, topic: &str, partition: i32, timeout: T, ) -> KafkaResult<(i64, i64)> { let mut low = -1; let mut high = -1; let topic_c = CString::new(topic.to_string())?; let ret = unsafe { rdsys::rd_kafka_query_watermark_offsets( self.native_ptr(), topic_c.as_ptr(), partition, &mut low as *mut i64, &mut high as *mut i64, timeout_to_ms(timeout), ) }; if ret.is_error() { return Err(KafkaError::MetadataFetch(ret.into())); } Ok((low, high)) } /// Returns the group membership information for the given group. If no group is /// specified, all groups will be returned. pub fn fetch_group_list<T: Into<Option<Duration>>>( &self, group: Option<&str>, timeout: T, ) -> KafkaResult<GroupList> { // Careful with group_c getting freed before time let group_c = CString::new(group.map_or("".to_string(), ToString::to_string))?; let group_c_ptr = if group.is_some() { group_c.as_ptr() } else { ptr::null_mut() }; let mut group_list_ptr: *const RDKafkaGroupList = ptr::null_mut(); trace!("Starting group list fetch"); let ret = unsafe { rdsys::rd_kafka_list_groups( self.native_ptr(), group_c_ptr, &mut group_list_ptr as *mut *const RDKafkaGroupList, timeout_to_ms(timeout), ) }; trace!("Group list fetch completed"); if ret.is_error() { return Err(KafkaError::GroupListFetch(ret.into())); } Ok(unsafe { GroupList::from_ptr(group_list_ptr) }) } /// Returns a NativeTopic from the current client. The NativeTopic shouldn't outlive the client /// it was generated from. fn native_topic(&self, topic: &str) -> KafkaResult<NativeTopic> { let topic_c = CString::new(topic.to_string())?; Ok(unsafe { NativeTopic::from_ptr(rdsys::rd_kafka_topic_new( self.native_ptr(), topic_c.as_ptr(), ptr::null_mut(), )) }) } /// Returns a NativeQueue from the current client. The NativeQueue shouldn't /// outlive the client it was generated from. pub(crate) fn new_native_queue(&self) -> NativeQueue { unsafe { NativeQueue::from_ptr(rdsys::rd_kafka_queue_new(self.native_ptr())) } } } struct NativeTopic { ptr: *mut RDKafkaTopic, } unsafe impl Send for NativeTopic {} unsafe impl Sync for NativeTopic {} impl NativeTopic { /// Wraps a pointer to an `RDKafkaTopic` object and returns a new `NativeTopic`. unsafe fn from_ptr(ptr: *mut RDKafkaTopic) -> NativeTopic { NativeTopic { ptr } } /// Returns the pointer to the librdkafka RDKafkaTopic structure. fn ptr(&self) -> *mut RDKafkaTopic { self.ptr } /// Returns a null pointer. fn null() -> *mut RDKafkaTopic { ptr::null::<u8>() as *mut RDKafkaTopic } } impl Drop for NativeTopic { fn drop(&mut self) { trace!("Destroying NativeTopic: {:?}", self.ptr); unsafe { rdsys::rd_kafka_topic_destroy(self.ptr); } trace!("NativeTopic destroyed: {:?}", self.ptr); } } pub(crate) struct NativeQueue { ptr: *mut RDKafkaQueue, } // The library is completely thread safe, according to the documentation. unsafe impl Sync for NativeQueue {} unsafe impl Send for NativeQueue {} impl NativeQueue { /// Wraps a pointer to an `RDKafkaQueue` object and returns a new /// `NativeQueue`. unsafe fn from_ptr(ptr: *mut RDKafkaQueue) -> NativeQueue { NativeQueue { ptr } } /// Returns the pointer to the librdkafka RDKafkaQueue structure. pub fn ptr(&self) -> *mut RDKafkaQueue { self.ptr } pub fn poll<T: Into<Option<Duration>>>(&self, t: T) -> *mut RDKafkaEvent { unsafe { rdsys::rd_kafka_queue_poll(self.ptr, timeout_to_ms(t)) } } } impl Drop for NativeQueue { fn drop(&mut self) { trace!("Destroying queue: {:?}", self.ptr); unsafe { rdsys::rd_kafka_queue_destroy(self.ptr); } trace!("Queue destroyed: {:?}", self.ptr); } } pub(crate) unsafe extern "C" fn native_log_cb<C: ClientContext>( client: *const RDKafka, level: i32, fac: *const c_char, buf: *const c_char, ) { let fac = CStr::from_ptr(fac).to_string_lossy(); let log_message = CStr::from_ptr(buf).to_string_lossy(); let context = Box::from_raw(rdsys::rd_kafka_opaque(client) as *mut C); (*context).log( RDKafkaLogLevel::from_int(level), fac.trim(), log_message.trim(), ); mem::forget(context); // Do not free the context } pub(crate) unsafe extern "C" fn native_stats_cb<C: ClientContext>( _conf: *mut RDKafka, json: *mut c_char, json_len: usize, opaque: *mut c_void, ) -> i32 { let context = Box::from_raw(opaque as *mut C); let mut bytes_vec = Vec::new(); bytes_vec.extend_from_slice(slice::from_raw_parts(json as *mut u8, json_len)); let json_string = CString::from_vec_unchecked(bytes_vec).into_string(); match json_string { Ok(json) => match serde_json::from_str(&json) { Ok(stats) => (*context).stats(stats), Err(e) => error!("Could not parse statistics JSON: {}", e), }, Err(e) => error!("Statistics JSON string is not UTF-8: {:?}", e), } mem::forget(context); // Do not free the context 0 // librdkafka will free the json buffer } pub(crate) unsafe extern "C" fn native_error_cb<C: ClientContext>( _client: *mut RDKafka, err: i32, reason: *const c_char, opaque: *mut c_void, ) { let err = rdsys::primitive_to_rd_kafka_resp_err_t(err) .expect("global error not an rd_kafka_resp_err_t"); let error = KafkaError::Global(err.into()); let reason = CStr::from_ptr(reason).to_string_lossy(); let context = Box::from_raw(opaque as *mut C); (*context).error(error, reason.trim()); mem::forget(context); // Do not free the context } #[cfg(test)] mod tests { // Just call everything to test there no panics by default, behavior // is tested in the integrations tests. use super::*; use crate::config::ClientConfig; #[test] fn test_client() { let config = ClientConfig::new(); let native_config = config.create_native_config().unwrap(); let client = Client::new( &config, native_config, RDKafkaType::RD_KAFKA_PRODUCER, DefaultClientContext, ) .unwrap(); assert!(!client.native_ptr().is_null()); } }
fetch_watermarks
identifier_name
three-d-bar-chart.component.ts
import { Component, OnInit, ViewChild, ElementRef, Input, ViewEncapsulation } from '@angular/core'; import * as _ from 'lodash'; import * as d3 from 'd3'; declare var $: any; declare var svgHelp: any; declare var bar3d: any; @Component({ selector: 'app-three-d-bar-chart', templateUrl: './three-d-bar-chart.component.html', styleUrls: ['./three-d-bar-chart.component.scss'], encapsulation: ViewEncapsulation.None }) export class ThreeDBarChartComponent implements OnInit { @ViewChild('barChart') private chartContainer: ElementRef; @Input() private data: Array<any>; @Input('xGrid') private xGrid: boolean; @Input('yGrid') private yGrid: boolean; @Input('values') private values: boolean;
(private hostRef: ElementRef) { } ngOnInit() { if (this.data) { this.createChart(this.data) } } createChart(data) { let el = this.chartContainer.nativeElement; d3.select(el).select("svg").remove(); var margin = { top: 30, right: 20, bottom: 40, left: 20, front: 0, back: 0 }, width = $(this.hostRef.nativeElement).parent().width() - margin.right - margin.left, height = $(this.hostRef.nativeElement).parent().height() - margin.top - margin.bottom; var depth = 100 - margin.front - margin.back; var xScale = d3.scaleBand().range([0, width]).padding(0.5), yScale = d3.scaleLinear().rangeRound([height, 0]), zScale = d3.scaleOrdinal().domain([0, 1, 2]).range([0, depth], .4); var xAxis = d3.axisBottom().scale(xScale); var yAxis = d3.axisLeft().scale(yScale).ticks(6); var max = d3.max(data[0].map(function (d) { return parseFloat(d.value); })); if (max < 100){ max = 100 } if(max == undefined){ max = 100 } var gridY = d3.scaleLinear().domain([0, max]).range( [height, 0]); var chart = d3.select(el).append("svg").attr("id", "columnbarChart") .attr("width", width + margin.left + margin.right).attr("height", height + margin.top + margin.bottom) .append("g") .attr("transform", "translate(35," + (data[0][0].axis.length > 20 ? 12 : 15) + ")"); // add the Y gridlines if (!this.yGrid) { chart.append("g") .attr("class", "grid") .attr('opacity', 0.3) // .attr("stroke","#ebebeb") .call(make_y_gridlines() .tickSize(-width).tickFormat(null) ).selectAll("text").remove(); } var color = ["#F4A775"]; var layers = data; layers.forEach(function (el, j) { el.y = undefined; el.y0 = j; }); var layer = chart.selectAll(".layer").data(layers).enter() .append("g").attr("class", "layer") .style("fill", color[0]); data = data[0]; //d3.tsv('data.tsv', type, function(err, data) { //if (err) return; xScale.domain(data.map(function (d) { return d.axis; })); yScale.domain([0, max]); const xBandwidth = xScale.bandwidth() > 50 * data.length ? 50 * data.length : xScale.bandwidth(); function x(d) { return xScale(d.axis); } function y(d) { return yScale(d.value); } var camera = [width / 2, height / 2, -200]; var barGen = bar3d() .camera(camera) .x(x) .y(y) .z(zScale(0)) // .attr('width', xScale.rangeBand()) .width(xScale.bandwidth()) .height(function (d) { return height - y(d); }) .depth(xScale.bandwidth()); chart.append('g') .attr('class', 'x axis') .attr("transform", "translate(0," + height + ")") .call(xAxis) .selectAll("text").style("text-anchor", "middle") .attr("class", function (d, i) { return "chartBartext" + i }) .attr("dx", "-.2em").attr("dy", ".70em") .call(wrap, xScale.bandwidth(), width); chart.append('g') .attr('class', 'y axis') .call(yAxis) .append('text') .attr('transform', svgHelp.rotate(-90)) .attr("y", -18 - margin.left) .attr("x", 0 - (height / 2)) .attr("dy", "1em") .style('text-anchor', 'end') .style("fill", "#333") .style("font-weight", "400") .attr("font-family", "'Questrial', sans-serif") .style("font-size", "13px") .text(data[0].unit); //check for no data availble let allNullValues = true; for (let j = 0; j < data.length; j++) { if (data[j].value != null) { allNullValues = false; break; } } if (allNullValues) { chart.append("text") .attr("transform", "translate("+ width/2 +",0)") .attr("x", 0) .attr("y",30) .attr("font-size", "28px") .style("text-anchor", "middle") .text("Data Not Available"); return; } let cubeBar = layer.selectAll('.bar').data(data) .enter().append('g') .attr('class', 'bar') .style("cursor", "pointer") .on("mouseover", function (d) { if(d.value) showPopover.call(this, d) }).on("mouseout", function (d) { removePopovers() }) // sort based on distance from center, so we draw outermost // bars first. otherwise, bars drawn later might overlap bars drawn first .sort(function (a, b) { return Math.abs(x(b) - 450) - Math.abs(x(a) - 450); }) .call(barGen) cubeBar.append("text") .attr("class", "below") .attr( "x", function (d) { return xScale(d.axis) + (xScale.bandwidth() - xBandwidth) / 2 + xBandwidth / (2 * data.length) + (xBandwidth / data.length); }) .attr("y", function (d) { return yScale(d.value) - 18; }) .attr("dy", "1.2em") // .attr("text-anchor", "left") .text(function (d) { if(d.value) return Math.round(d.value); }) .style("fill", "#000").style("font-size", "12px"); function removePopovers() { $('.popover').each(function () { $(this).remove(); }); } function showPopover(d) { $(this).popover( { title: '', placement: 'top', container: 'body', trigger: 'manual', html: true, animation: false, content: function () { if (d.axis != '' && d.denominator != null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else if (d.denominator == null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>"; } else if (d.denominator == null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>"; } else if (d.denominator != null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div style='color: #495769;'> Data Value: " + d.value + "</div>"; } } }); $(this).popover('show'); } // gridlines in x axis function function make_x_gridlines() { return d3.axisBottom(x) .ticks(5) } // gridlines in y axis function function make_y_gridlines() { return d3.axisLeft(gridY) .ticks(5) } function type(d) { d.value = +d.value; return d; } function wrap(text, width, windowWidth) { text.each(function () { var text = d3.select(this), words = text.text().split(/\s+/).reverse(), word, cnt = 0, line = [], lineNumber = 0, lineHeight = 1, y = text.attr("y"), dy = parseFloat(text.attr("dy")); if (windowWidth > 660) var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); else var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); if (words.length == 1) { let chars = words.toString().split(""); chars.splice((chars.length / 2).toFixed(), 0, '-', ' '); tspan.text(chars.join("")); if (tspan.node().getComputedTextLength() > width) { words = chars.join("").split(/\s+/).reverse(); } tspan.text(''); } while (word = words.pop()) { cnt++; line.push(word); tspan.text(line.join(" ")); if (tspan.node().getComputedTextLength() > width) { line.pop(); tspan.text(line.join(" ")); line = [word]; // if(cnt!=1) if (width > 660) tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); else tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); } } }); } } }
constructor
identifier_name
three-d-bar-chart.component.ts
import { Component, OnInit, ViewChild, ElementRef, Input, ViewEncapsulation } from '@angular/core'; import * as _ from 'lodash'; import * as d3 from 'd3'; declare var $: any; declare var svgHelp: any; declare var bar3d: any; @Component({ selector: 'app-three-d-bar-chart', templateUrl: './three-d-bar-chart.component.html', styleUrls: ['./three-d-bar-chart.component.scss'], encapsulation: ViewEncapsulation.None }) export class ThreeDBarChartComponent implements OnInit { @ViewChild('barChart') private chartContainer: ElementRef; @Input() private data: Array<any>; @Input('xGrid') private xGrid: boolean; @Input('yGrid') private yGrid: boolean; @Input('values') private values: boolean; constructor(private hostRef: ElementRef) { } ngOnInit() { if (this.data)
} createChart(data) { let el = this.chartContainer.nativeElement; d3.select(el).select("svg").remove(); var margin = { top: 30, right: 20, bottom: 40, left: 20, front: 0, back: 0 }, width = $(this.hostRef.nativeElement).parent().width() - margin.right - margin.left, height = $(this.hostRef.nativeElement).parent().height() - margin.top - margin.bottom; var depth = 100 - margin.front - margin.back; var xScale = d3.scaleBand().range([0, width]).padding(0.5), yScale = d3.scaleLinear().rangeRound([height, 0]), zScale = d3.scaleOrdinal().domain([0, 1, 2]).range([0, depth], .4); var xAxis = d3.axisBottom().scale(xScale); var yAxis = d3.axisLeft().scale(yScale).ticks(6); var max = d3.max(data[0].map(function (d) { return parseFloat(d.value); })); if (max < 100){ max = 100 } if(max == undefined){ max = 100 } var gridY = d3.scaleLinear().domain([0, max]).range( [height, 0]); var chart = d3.select(el).append("svg").attr("id", "columnbarChart") .attr("width", width + margin.left + margin.right).attr("height", height + margin.top + margin.bottom) .append("g") .attr("transform", "translate(35," + (data[0][0].axis.length > 20 ? 12 : 15) + ")"); // add the Y gridlines if (!this.yGrid) { chart.append("g") .attr("class", "grid") .attr('opacity', 0.3) // .attr("stroke","#ebebeb") .call(make_y_gridlines() .tickSize(-width).tickFormat(null) ).selectAll("text").remove(); } var color = ["#F4A775"]; var layers = data; layers.forEach(function (el, j) { el.y = undefined; el.y0 = j; }); var layer = chart.selectAll(".layer").data(layers).enter() .append("g").attr("class", "layer") .style("fill", color[0]); data = data[0]; //d3.tsv('data.tsv', type, function(err, data) { //if (err) return; xScale.domain(data.map(function (d) { return d.axis; })); yScale.domain([0, max]); const xBandwidth = xScale.bandwidth() > 50 * data.length ? 50 * data.length : xScale.bandwidth(); function x(d) { return xScale(d.axis); } function y(d) { return yScale(d.value); } var camera = [width / 2, height / 2, -200]; var barGen = bar3d() .camera(camera) .x(x) .y(y) .z(zScale(0)) // .attr('width', xScale.rangeBand()) .width(xScale.bandwidth()) .height(function (d) { return height - y(d); }) .depth(xScale.bandwidth()); chart.append('g') .attr('class', 'x axis') .attr("transform", "translate(0," + height + ")") .call(xAxis) .selectAll("text").style("text-anchor", "middle") .attr("class", function (d, i) { return "chartBartext" + i }) .attr("dx", "-.2em").attr("dy", ".70em") .call(wrap, xScale.bandwidth(), width); chart.append('g') .attr('class', 'y axis') .call(yAxis) .append('text') .attr('transform', svgHelp.rotate(-90)) .attr("y", -18 - margin.left) .attr("x", 0 - (height / 2)) .attr("dy", "1em") .style('text-anchor', 'end') .style("fill", "#333") .style("font-weight", "400") .attr("font-family", "'Questrial', sans-serif") .style("font-size", "13px") .text(data[0].unit); //check for no data availble let allNullValues = true; for (let j = 0; j < data.length; j++) { if (data[j].value != null) { allNullValues = false; break; } } if (allNullValues) { chart.append("text") .attr("transform", "translate("+ width/2 +",0)") .attr("x", 0) .attr("y",30) .attr("font-size", "28px") .style("text-anchor", "middle") .text("Data Not Available"); return; } let cubeBar = layer.selectAll('.bar').data(data) .enter().append('g') .attr('class', 'bar') .style("cursor", "pointer") .on("mouseover", function (d) { if(d.value) showPopover.call(this, d) }).on("mouseout", function (d) { removePopovers() }) // sort based on distance from center, so we draw outermost // bars first. otherwise, bars drawn later might overlap bars drawn first .sort(function (a, b) { return Math.abs(x(b) - 450) - Math.abs(x(a) - 450); }) .call(barGen) cubeBar.append("text") .attr("class", "below") .attr( "x", function (d) { return xScale(d.axis) + (xScale.bandwidth() - xBandwidth) / 2 + xBandwidth / (2 * data.length) + (xBandwidth / data.length); }) .attr("y", function (d) { return yScale(d.value) - 18; }) .attr("dy", "1.2em") // .attr("text-anchor", "left") .text(function (d) { if(d.value) return Math.round(d.value); }) .style("fill", "#000").style("font-size", "12px"); function removePopovers() { $('.popover').each(function () { $(this).remove(); }); } function showPopover(d) { $(this).popover( { title: '', placement: 'top', container: 'body', trigger: 'manual', html: true, animation: false, content: function () { if (d.axis != '' && d.denominator != null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else if (d.denominator == null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>"; } else if (d.denominator == null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>"; } else if (d.denominator != null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div style='color: #495769;'> Data Value: " + d.value + "</div>"; } } }); $(this).popover('show'); } // gridlines in x axis function function make_x_gridlines() { return d3.axisBottom(x) .ticks(5) } // gridlines in y axis function function make_y_gridlines() { return d3.axisLeft(gridY) .ticks(5) } function type(d) { d.value = +d.value; return d; } function wrap(text, width, windowWidth) { text.each(function () { var text = d3.select(this), words = text.text().split(/\s+/).reverse(), word, cnt = 0, line = [], lineNumber = 0, lineHeight = 1, y = text.attr("y"), dy = parseFloat(text.attr("dy")); if (windowWidth > 660) var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); else var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); if (words.length == 1) { let chars = words.toString().split(""); chars.splice((chars.length / 2).toFixed(), 0, '-', ' '); tspan.text(chars.join("")); if (tspan.node().getComputedTextLength() > width) { words = chars.join("").split(/\s+/).reverse(); } tspan.text(''); } while (word = words.pop()) { cnt++; line.push(word); tspan.text(line.join(" ")); if (tspan.node().getComputedTextLength() > width) { line.pop(); tspan.text(line.join(" ")); line = [word]; // if(cnt!=1) if (width > 660) tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); else tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); } } }); } } }
{ this.createChart(this.data) }
conditional_block
three-d-bar-chart.component.ts
import { Component, OnInit, ViewChild, ElementRef, Input, ViewEncapsulation } from '@angular/core'; import * as _ from 'lodash'; import * as d3 from 'd3'; declare var $: any; declare var svgHelp: any; declare var bar3d: any; @Component({ selector: 'app-three-d-bar-chart', templateUrl: './three-d-bar-chart.component.html', styleUrls: ['./three-d-bar-chart.component.scss'], encapsulation: ViewEncapsulation.None }) export class ThreeDBarChartComponent implements OnInit { @ViewChild('barChart') private chartContainer: ElementRef; @Input() private data: Array<any>; @Input('xGrid') private xGrid: boolean; @Input('yGrid') private yGrid: boolean; @Input('values') private values: boolean; constructor(private hostRef: ElementRef) { } ngOnInit() { if (this.data) { this.createChart(this.data) } } createChart(data) { let el = this.chartContainer.nativeElement; d3.select(el).select("svg").remove(); var margin = { top: 30, right: 20, bottom: 40, left: 20, front: 0, back: 0 }, width = $(this.hostRef.nativeElement).parent().width() - margin.right - margin.left, height = $(this.hostRef.nativeElement).parent().height() - margin.top - margin.bottom; var depth = 100 - margin.front - margin.back; var xScale = d3.scaleBand().range([0, width]).padding(0.5), yScale = d3.scaleLinear().rangeRound([height, 0]), zScale = d3.scaleOrdinal().domain([0, 1, 2]).range([0, depth], .4); var xAxis = d3.axisBottom().scale(xScale); var yAxis = d3.axisLeft().scale(yScale).ticks(6); var max = d3.max(data[0].map(function (d) { return parseFloat(d.value); })); if (max < 100){ max = 100 } if(max == undefined){ max = 100 } var gridY = d3.scaleLinear().domain([0, max]).range( [height, 0]); var chart = d3.select(el).append("svg").attr("id", "columnbarChart") .attr("width", width + margin.left + margin.right).attr("height", height + margin.top + margin.bottom) .append("g") .attr("transform", "translate(35," + (data[0][0].axis.length > 20 ? 12 : 15) + ")"); // add the Y gridlines if (!this.yGrid) { chart.append("g") .attr("class", "grid") .attr('opacity', 0.3) // .attr("stroke","#ebebeb") .call(make_y_gridlines() .tickSize(-width).tickFormat(null) ).selectAll("text").remove(); } var color = ["#F4A775"]; var layers = data; layers.forEach(function (el, j) { el.y = undefined; el.y0 = j; }); var layer = chart.selectAll(".layer").data(layers).enter() .append("g").attr("class", "layer") .style("fill", color[0]); data = data[0]; //d3.tsv('data.tsv', type, function(err, data) { //if (err) return; xScale.domain(data.map(function (d) { return d.axis; })); yScale.domain([0, max]); const xBandwidth = xScale.bandwidth() > 50 * data.length ? 50 * data.length : xScale.bandwidth(); function x(d) { return xScale(d.axis); } function y(d) { return yScale(d.value); } var camera = [width / 2, height / 2, -200]; var barGen = bar3d() .camera(camera) .x(x) .y(y) .z(zScale(0)) // .attr('width', xScale.rangeBand()) .width(xScale.bandwidth()) .height(function (d) { return height - y(d); }) .depth(xScale.bandwidth()); chart.append('g') .attr('class', 'x axis') .attr("transform", "translate(0," + height + ")") .call(xAxis) .selectAll("text").style("text-anchor", "middle") .attr("class", function (d, i) { return "chartBartext" + i }) .attr("dx", "-.2em").attr("dy", ".70em") .call(wrap, xScale.bandwidth(), width); chart.append('g') .attr('class', 'y axis') .call(yAxis) .append('text') .attr('transform', svgHelp.rotate(-90)) .attr("y", -18 - margin.left) .attr("x", 0 - (height / 2)) .attr("dy", "1em") .style('text-anchor', 'end') .style("fill", "#333") .style("font-weight", "400") .attr("font-family", "'Questrial', sans-serif") .style("font-size", "13px") .text(data[0].unit); //check for no data availble let allNullValues = true; for (let j = 0; j < data.length; j++) { if (data[j].value != null) { allNullValues = false; break; } } if (allNullValues) { chart.append("text") .attr("transform", "translate("+ width/2 +",0)") .attr("x", 0) .attr("y",30) .attr("font-size", "28px") .style("text-anchor", "middle") .text("Data Not Available"); return; } let cubeBar = layer.selectAll('.bar').data(data) .enter().append('g') .attr('class', 'bar') .style("cursor", "pointer") .on("mouseover", function (d) { if(d.value) showPopover.call(this, d) }).on("mouseout", function (d) { removePopovers() }) // sort based on distance from center, so we draw outermost // bars first. otherwise, bars drawn later might overlap bars drawn first .sort(function (a, b) { return Math.abs(x(b) - 450) - Math.abs(x(a) - 450); }) .call(barGen) cubeBar.append("text") .attr("class", "below") .attr( "x", function (d) { return xScale(d.axis) + (xScale.bandwidth() - xBandwidth) / 2 + xBandwidth / (2 * data.length) + (xBandwidth / data.length); }) .attr("y", function (d) { return yScale(d.value) - 18; }) .attr("dy", "1.2em") // .attr("text-anchor", "left") .text(function (d) { if(d.value) return Math.round(d.value); }) .style("fill", "#000").style("font-size", "12px"); function removePopovers() { $('.popover').each(function () { $(this).remove(); }); } function showPopover(d) { $(this).popover( { title: '', placement: 'top', container: 'body', trigger: 'manual', html: true, animation: false, content: function () { if (d.axis != '' && d.denominator != null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else if (d.denominator == null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>"; } else if (d.denominator == null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>"; } else if (d.denominator != null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div style='color: #495769;'> Data Value: " + d.value + "</div>"; } } }); $(this).popover('show'); } // gridlines in x axis function function make_x_gridlines()
// gridlines in y axis function function make_y_gridlines() { return d3.axisLeft(gridY) .ticks(5) } function type(d) { d.value = +d.value; return d; } function wrap(text, width, windowWidth) { text.each(function () { var text = d3.select(this), words = text.text().split(/\s+/).reverse(), word, cnt = 0, line = [], lineNumber = 0, lineHeight = 1, y = text.attr("y"), dy = parseFloat(text.attr("dy")); if (windowWidth > 660) var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); else var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); if (words.length == 1) { let chars = words.toString().split(""); chars.splice((chars.length / 2).toFixed(), 0, '-', ' '); tspan.text(chars.join("")); if (tspan.node().getComputedTextLength() > width) { words = chars.join("").split(/\s+/).reverse(); } tspan.text(''); } while (word = words.pop()) { cnt++; line.push(word); tspan.text(line.join(" ")); if (tspan.node().getComputedTextLength() > width) { line.pop(); tspan.text(line.join(" ")); line = [word]; // if(cnt!=1) if (width > 660) tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); else tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); } } }); } } }
{ return d3.axisBottom(x) .ticks(5) }
identifier_body
three-d-bar-chart.component.ts
import { Component, OnInit, ViewChild, ElementRef, Input, ViewEncapsulation } from '@angular/core'; import * as _ from 'lodash'; import * as d3 from 'd3'; declare var $: any; declare var svgHelp: any; declare var bar3d: any; @Component({ selector: 'app-three-d-bar-chart', templateUrl: './three-d-bar-chart.component.html', styleUrls: ['./three-d-bar-chart.component.scss'], encapsulation: ViewEncapsulation.None }) export class ThreeDBarChartComponent implements OnInit { @ViewChild('barChart') private chartContainer: ElementRef; @Input() private data: Array<any>; @Input('xGrid') private xGrid: boolean; @Input('yGrid') private yGrid: boolean; @Input('values') private values: boolean; constructor(private hostRef: ElementRef) { } ngOnInit() { if (this.data) { this.createChart(this.data) } } createChart(data) { let el = this.chartContainer.nativeElement; d3.select(el).select("svg").remove(); var margin = { top: 30, right: 20, bottom: 40, left: 20, front: 0, back: 0 }, width = $(this.hostRef.nativeElement).parent().width() - margin.right - margin.left, height = $(this.hostRef.nativeElement).parent().height() - margin.top - margin.bottom; var depth = 100 - margin.front - margin.back; var xScale = d3.scaleBand().range([0, width]).padding(0.5), yScale = d3.scaleLinear().rangeRound([height, 0]), zScale = d3.scaleOrdinal().domain([0, 1, 2]).range([0, depth], .4); var xAxis = d3.axisBottom().scale(xScale); var yAxis = d3.axisLeft().scale(yScale).ticks(6); var max = d3.max(data[0].map(function (d) { return parseFloat(d.value); })); if (max < 100){ max = 100 } if(max == undefined){ max = 100 } var gridY = d3.scaleLinear().domain([0, max]).range( [height, 0]); var chart = d3.select(el).append("svg").attr("id", "columnbarChart") .attr("width", width + margin.left + margin.right).attr("height", height + margin.top + margin.bottom) .append("g") .attr("transform", "translate(35," + (data[0][0].axis.length > 20 ? 12 : 15) + ")"); // add the Y gridlines if (!this.yGrid) { chart.append("g") .attr("class", "grid") .attr('opacity', 0.3) // .attr("stroke","#ebebeb") .call(make_y_gridlines() .tickSize(-width).tickFormat(null) ).selectAll("text").remove(); } var color = ["#F4A775"]; var layers = data; layers.forEach(function (el, j) { el.y = undefined; el.y0 = j; }); var layer = chart.selectAll(".layer").data(layers).enter() .append("g").attr("class", "layer") .style("fill", color[0]); data = data[0]; //d3.tsv('data.tsv', type, function(err, data) { //if (err) return; xScale.domain(data.map(function (d) { return d.axis; })); yScale.domain([0, max]); const xBandwidth = xScale.bandwidth() > 50 * data.length ? 50 * data.length : xScale.bandwidth(); function x(d) { return xScale(d.axis); } function y(d) { return yScale(d.value); } var camera = [width / 2, height / 2, -200]; var barGen = bar3d() .camera(camera) .x(x) .y(y) .z(zScale(0)) // .attr('width', xScale.rangeBand()) .width(xScale.bandwidth()) .height(function (d) { return height - y(d); }) .depth(xScale.bandwidth()); chart.append('g') .attr('class', 'x axis') .attr("transform", "translate(0," + height + ")") .call(xAxis) .selectAll("text").style("text-anchor", "middle") .attr("class", function (d, i) { return "chartBartext" + i }) .attr("dx", "-.2em").attr("dy", ".70em") .call(wrap, xScale.bandwidth(), width); chart.append('g') .attr('class', 'y axis') .call(yAxis) .append('text') .attr('transform', svgHelp.rotate(-90)) .attr("y", -18 - margin.left) .attr("x", 0 - (height / 2)) .attr("dy", "1em") .style('text-anchor', 'end') .style("fill", "#333") .style("font-weight", "400") .attr("font-family", "'Questrial', sans-serif") .style("font-size", "13px") .text(data[0].unit); //check for no data availble let allNullValues = true; for (let j = 0; j < data.length; j++) { if (data[j].value != null) { allNullValues = false; break; } } if (allNullValues) { chart.append("text") .attr("transform", "translate("+ width/2 +",0)") .attr("x", 0) .attr("y",30) .attr("font-size", "28px") .style("text-anchor", "middle") .text("Data Not Available"); return; } let cubeBar = layer.selectAll('.bar').data(data) .enter().append('g') .attr('class', 'bar')
.on("mouseover", function (d) { if(d.value) showPopover.call(this, d) }).on("mouseout", function (d) { removePopovers() }) // sort based on distance from center, so we draw outermost // bars first. otherwise, bars drawn later might overlap bars drawn first .sort(function (a, b) { return Math.abs(x(b) - 450) - Math.abs(x(a) - 450); }) .call(barGen) cubeBar.append("text") .attr("class", "below") .attr( "x", function (d) { return xScale(d.axis) + (xScale.bandwidth() - xBandwidth) / 2 + xBandwidth / (2 * data.length) + (xBandwidth / data.length); }) .attr("y", function (d) { return yScale(d.value) - 18; }) .attr("dy", "1.2em") // .attr("text-anchor", "left") .text(function (d) { if(d.value) return Math.round(d.value); }) .style("fill", "#000").style("font-size", "12px"); function removePopovers() { $('.popover').each(function () { $(this).remove(); }); } function showPopover(d) { $(this).popover( { title: '', placement: 'top', container: 'body', trigger: 'manual', html: true, animation: false, content: function () { if (d.axis != '' && d.denominator != null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else if (d.denominator == null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>"; } else if (d.denominator == null && d.numerator != null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value +"%"+ "</span>" + "</div>" + "<div>" + "Numerator : " + "<span style='color: #495769;font-weight:500'>" + d.numerator + "</span>" + "</div>"; } else if (d.denominator != null && d.numerator == null && d.unit == 'Percentage') { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div>" + " Data Value : " + "<span style='color: #495769;font-weight:500;'>" + d.value + "%"+"</span>" + "</div>" + "<div>" + "Denominator : " + "<span style='color: #495769;font-weight:500'>" + d.denominator + "</span>" + "</div>"; } else { return "<div style='color: #495769;'>" + "<b>" + d.axis + "</b>" + "</div>" + "<div style='color: #495769;'> Data Value: " + d.value + "</div>"; } } }); $(this).popover('show'); } // gridlines in x axis function function make_x_gridlines() { return d3.axisBottom(x) .ticks(5) } // gridlines in y axis function function make_y_gridlines() { return d3.axisLeft(gridY) .ticks(5) } function type(d) { d.value = +d.value; return d; } function wrap(text, width, windowWidth) { text.each(function () { var text = d3.select(this), words = text.text().split(/\s+/).reverse(), word, cnt = 0, line = [], lineNumber = 0, lineHeight = 1, y = text.attr("y"), dy = parseFloat(text.attr("dy")); if (windowWidth > 660) var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); else var tspan = text.text(null).append("tspan").attr("x", 0).attr("y", y).attr("dy", dy + "em").style('font-size', '10px'); if (words.length == 1) { let chars = words.toString().split(""); chars.splice((chars.length / 2).toFixed(), 0, '-', ' '); tspan.text(chars.join("")); if (tspan.node().getComputedTextLength() > width) { words = chars.join("").split(/\s+/).reverse(); } tspan.text(''); } while (word = words.pop()) { cnt++; line.push(word); tspan.text(line.join(" ")); if (tspan.node().getComputedTextLength() > width) { line.pop(); tspan.text(line.join(" ")); line = [word]; // if(cnt!=1) if (width > 660) tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); else tspan = text.append("tspan").attr("x", 0).attr("y", y).attr("dy", ++lineNumber * lineHeight + dy + "em").style('font-size', '10px').text(word); } } }); } } }
.style("cursor", "pointer")
random_line_split
pageserver.rs
// // Main entry point for the Page Server executable // use log::*; use pageserver::defaults::*; use serde::{Deserialize, Serialize}; use std::{ env, net::TcpListener, path::{Path, PathBuf}, str::FromStr, thread, }; use zenith_utils::{auth::JwtAuth, logging, postgres_backend::AuthType}; use anyhow::{bail, ensure, Context, Result}; use clap::{App, Arg, ArgMatches}; use daemonize::Daemonize; use pageserver::{ branches, defaults::{DEFAULT_HTTP_LISTEN_ADDR, DEFAULT_PG_LISTEN_ADDR}, http, page_service, tenant_mgr, PageServerConf, RelishStorageConfig, S3Config, LOG_FILE_NAME, }; use zenith_utils::http::endpoint; use const_format::formatcp; /// String arguments that can be declared via CLI or config file #[derive(Serialize, Deserialize)] struct CfgFileParams { listen_pg_addr: Option<String>, listen_http_addr: Option<String>, checkpoint_distance: Option<String>, checkpoint_period: Option<String>, gc_horizon: Option<String>, gc_period: Option<String>, pg_distrib_dir: Option<String>, auth_validation_public_key_path: Option<String>, auth_type: Option<String>, // see https://github.com/alexcrichton/toml-rs/blob/6c162e6562c3e432bf04c82a3d1d789d80761a86/examples/enum_external.rs for enum deserialisation examples relish_storage: Option<RelishStorage>, } #[derive(Serialize, Deserialize, Clone)] enum RelishStorage { Local { local_path: String, }, AwsS3 { bucket_name: String, bucket_region: String, #[serde(skip_serializing)] access_key_id: Option<String>, #[serde(skip_serializing)] secret_access_key: Option<String>, }, } impl CfgFileParams { /// Extract string arguments from CLI fn from_args(arg_matches: &ArgMatches) -> Self { let get_arg = |arg_name: &str| -> Option<String> { arg_matches.value_of(arg_name).map(str::to_owned) }; let relish_storage = if let Some(local_path) = get_arg("relish-storage-local-path") { Some(RelishStorage::Local { local_path }) } else if let Some((bucket_name, bucket_region)) = get_arg("relish-storage-s3-bucket").zip(get_arg("relish-storage-region")) { Some(RelishStorage::AwsS3 { bucket_name, bucket_region, access_key_id: get_arg("relish-storage-access-key"), secret_access_key: get_arg("relish-storage-secret-access-key"), }) } else { None }; Self { listen_pg_addr: get_arg("listen-pg"), listen_http_addr: get_arg("listen-http"), checkpoint_distance: get_arg("checkpoint_distance"), checkpoint_period: get_arg("checkpoint_period"), gc_horizon: get_arg("gc_horizon"), gc_period: get_arg("gc_period"), pg_distrib_dir: get_arg("postgres-distrib"), auth_validation_public_key_path: get_arg("auth-validation-public-key-path"), auth_type: get_arg("auth-type"), relish_storage, } } /// Fill missing values in `self` with `other` fn or(self, other: CfgFileParams) -> Self { // TODO cleaner way to do this Self { listen_pg_addr: self.listen_pg_addr.or(other.listen_pg_addr), listen_http_addr: self.listen_http_addr.or(other.listen_http_addr), checkpoint_distance: self.checkpoint_distance.or(other.checkpoint_distance), checkpoint_period: self.checkpoint_period.or(other.checkpoint_period), gc_horizon: self.gc_horizon.or(other.gc_horizon), gc_period: self.gc_period.or(other.gc_period), pg_distrib_dir: self.pg_distrib_dir.or(other.pg_distrib_dir), auth_validation_public_key_path: self .auth_validation_public_key_path .or(other.auth_validation_public_key_path), auth_type: self.auth_type.or(other.auth_type), relish_storage: self.relish_storage.or(other.relish_storage), } } /// Create a PageServerConf from these string parameters fn try_into_config(&self) -> Result<PageServerConf> { let workdir = PathBuf::from("."); let listen_pg_addr = match self.listen_pg_addr.as_ref() { Some(addr) => addr.clone(), None => DEFAULT_PG_LISTEN_ADDR.to_owned(), }; let listen_http_addr = match self.listen_http_addr.as_ref() { Some(addr) => addr.clone(), None => DEFAULT_HTTP_LISTEN_ADDR.to_owned(), }; let checkpoint_distance: u64 = match self.checkpoint_distance.as_ref() { Some(checkpoint_distance_str) => checkpoint_distance_str.parse()?, None => DEFAULT_CHECKPOINT_DISTANCE, }; let checkpoint_period = match self.checkpoint_period.as_ref() { Some(checkpoint_period_str) => humantime::parse_duration(checkpoint_period_str)?, None => DEFAULT_CHECKPOINT_PERIOD, }; let gc_horizon: u64 = match self.gc_horizon.as_ref() { Some(horizon_str) => horizon_str.parse()?, None => DEFAULT_GC_HORIZON, }; let gc_period = match self.gc_period.as_ref() { Some(period_str) => humantime::parse_duration(period_str)?, None => DEFAULT_GC_PERIOD, }; let pg_distrib_dir = match self.pg_distrib_dir.as_ref() { Some(pg_distrib_dir_str) => PathBuf::from(pg_distrib_dir_str), None => env::current_dir()?.join("tmp_install"), }; let auth_validation_public_key_path = self .auth_validation_public_key_path .as_ref() .map(PathBuf::from); let auth_type = self .auth_type .as_ref() .map_or(Ok(AuthType::Trust), |auth_type| { AuthType::from_str(auth_type) })?; if !pg_distrib_dir.join("bin/postgres").exists() { bail!("Can't find postgres binary at {:?}", pg_distrib_dir); } if auth_type == AuthType::ZenithJWT { ensure!( auth_validation_public_key_path.is_some(), "Missing auth_validation_public_key_path when auth_type is ZenithJWT" ); let path_ref = auth_validation_public_key_path.as_ref().unwrap(); ensure!( path_ref.exists(), format!("Can't find auth_validation_public_key at {:?}", path_ref) ); } let relish_storage_config = self.relish_storage .as_ref() .map(|storage_params| match storage_params.clone() { RelishStorage::Local { local_path } => { RelishStorageConfig::LocalFs(PathBuf::from(local_path)) } RelishStorage::AwsS3 { bucket_name, bucket_region, access_key_id, secret_access_key, } => RelishStorageConfig::AwsS3(S3Config { bucket_name, bucket_region, access_key_id, secret_access_key, }), }); Ok(PageServerConf { daemonize: false, listen_pg_addr, listen_http_addr, checkpoint_distance, checkpoint_period, gc_horizon, gc_period, superuser: String::from(DEFAULT_SUPERUSER), workdir, pg_distrib_dir, auth_validation_public_key_path, auth_type, relish_storage_config, }) } } fn
() -> Result<()> { let arg_matches = App::new("Zenith page server") .about("Materializes WAL stream to pages and serves them to the postgres") .arg( Arg::with_name("listen-pg") .short("l") .long("listen-pg") .alias("listen") // keep some compatibility .takes_value(true) .help(formatcp!("listen for incoming page requests on ip:port (default: {DEFAULT_PG_LISTEN_ADDR})")), ) .arg( Arg::with_name("listen-http") .long("listen-http") .alias("http_endpoint") // keep some compatibility .takes_value(true) .help(formatcp!("http endpoint address for metrics and management API calls on ip:port (default: {DEFAULT_HTTP_LISTEN_ADDR})")), ) .arg( Arg::with_name("daemonize") .short("d") .long("daemonize") .takes_value(false) .help("Run in the background"), ) .arg( Arg::with_name("init") .long("init") .takes_value(false) .help("Initialize pageserver repo"), ) .arg( Arg::with_name("checkpoint_distance") .long("checkpoint_distance") .takes_value(true) .help("Distance from current LSN to perform checkpoint of in-memory layers"), ) .arg( Arg::with_name("checkpoint_period") .long("checkpoint_period") .takes_value(true) .help("Interval between checkpoint iterations"), ) .arg( Arg::with_name("gc_horizon") .long("gc_horizon") .takes_value(true) .help("Distance from current LSN to perform all wal records cleanup"), ) .arg( Arg::with_name("gc_period") .long("gc_period") .takes_value(true) .help("Interval between garbage collector iterations"), ) .arg( Arg::with_name("workdir") .short("D") .long("workdir") .takes_value(true) .help("Working directory for the pageserver"), ) .arg( Arg::with_name("postgres-distrib") .long("postgres-distrib") .takes_value(true) .help("Postgres distribution directory"), ) .arg( Arg::with_name("create-tenant") .long("create-tenant") .takes_value(true) .help("Create tenant during init") .requires("init"), ) .arg( Arg::with_name("auth-validation-public-key-path") .long("auth-validation-public-key-path") .takes_value(true) .help("Path to public key used to validate jwt signature"), ) .arg( Arg::with_name("auth-type") .long("auth-type") .takes_value(true) .help("Authentication scheme type. One of: Trust, MD5, ZenithJWT"), ) .arg( Arg::with_name("relish-storage-local-path") .long("relish-storage-local-path") .takes_value(true) .help("Path to the local directory, to be used as an external relish storage") .conflicts_with_all(&[ "relish-storage-s3-bucket", "relish-storage-region", "relish-storage-access-key", "relish-storage-secret-access-key", ]), ) .arg( Arg::with_name("relish-storage-s3-bucket") .long("relish-storage-s3-bucket") .takes_value(true) .help("Name of the AWS S3 bucket to use an external relish storage") .requires("relish-storage-region"), ) .arg( Arg::with_name("relish-storage-region") .long("relish-storage-region") .takes_value(true) .help("Region of the AWS S3 bucket"), ) .arg( Arg::with_name("relish-storage-access-key") .long("relish-storage-access-key") .takes_value(true) .help("Credentials to access the AWS S3 bucket"), ) .arg( Arg::with_name("relish-storage-secret-access-key") .long("relish-storage-secret-access-key") .takes_value(true) .help("Credentials to access the AWS S3 bucket"), ) .get_matches(); let workdir = Path::new(arg_matches.value_of("workdir").unwrap_or(".zenith")); let cfg_file_path = workdir .canonicalize() .with_context(|| format!("Error opening workdir '{}'", workdir.display()))? .join("pageserver.toml"); let args_params = CfgFileParams::from_args(&arg_matches); let init = arg_matches.is_present("init"); let create_tenant = arg_matches.value_of("create-tenant"); let params = if init { // We're initializing the repo, so there's no config file yet args_params } else { // Supplement the CLI arguments with the config file let cfg_file_contents = std::fs::read_to_string(&cfg_file_path) .with_context(|| format!("No pageserver config at '{}'", cfg_file_path.display()))?; let file_params: CfgFileParams = toml::from_str(&cfg_file_contents).with_context(|| { format!( "Failed to read '{}' as pageserver config", cfg_file_path.display() ) })?; args_params.or(file_params) }; // Set CWD to workdir for non-daemon modes env::set_current_dir(&workdir).with_context(|| { format!( "Failed to set application's current dir to '{}'", workdir.display() ) })?; // Ensure the config is valid, even if just init-ing let mut conf = params.try_into_config().with_context(|| { format!( "Pageserver config at '{}' is not valid", cfg_file_path.display() ) })?; conf.daemonize = arg_matches.is_present("daemonize"); if init && conf.daemonize { bail!("--daemonize cannot be used with --init") } // The configuration is all set up now. Turn it into a 'static // that can be freely stored in structs and passed across threads // as a ref. let conf: &'static PageServerConf = Box::leak(Box::new(conf)); // Create repo and exit if init was requested if init { branches::init_pageserver(conf, create_tenant).context("Failed to init pageserver")?; // write the config file let cfg_file_contents = toml::to_string_pretty(&params) .context("Failed to create pageserver config contents for initialisation")?; // TODO support enable-auth flag std::fs::write(&cfg_file_path, cfg_file_contents).with_context(|| { format!( "Failed to initialize pageserver config at '{}'", cfg_file_path.display() ) })?; Ok(()) } else { start_pageserver(conf).context("Failed to start pageserver") } } fn start_pageserver(conf: &'static PageServerConf) -> Result<()> { // Initialize logger let (_scope_guard, log_file) = logging::init(LOG_FILE_NAME, conf.daemonize)?; // TODO: Check that it looks like a valid repository before going further // bind sockets before daemonizing so we report errors early and do not return until we are listening info!( "Starting pageserver http handler on {}", conf.listen_http_addr ); let http_listener = TcpListener::bind(conf.listen_http_addr.clone())?; info!( "Starting pageserver pg protocol handler on {}", conf.listen_pg_addr ); let pageserver_listener = TcpListener::bind(conf.listen_pg_addr.clone())?; if conf.daemonize { info!("daemonizing..."); // There shouldn't be any logging to stdin/stdout. Redirect it to the main log so // that we will see any accidental manual fprintf's or backtraces. let stdout = log_file.try_clone().unwrap(); let stderr = log_file; let daemonize = Daemonize::new() .pid_file("pageserver.pid") .working_directory(".") .stdout(stdout) .stderr(stderr); match daemonize.start() { Ok(_) => info!("Success, daemonized"), Err(e) => error!("Error, {}", e), } } // Initialize tenant manager. tenant_mgr::init(conf); // keep join handles for spawned threads let mut join_handles = vec![]; // initialize authentication for incoming connections let auth = match &conf.auth_type { AuthType::Trust | AuthType::MD5 => None, AuthType::ZenithJWT => { // unwrap is ok because check is performed when creating config, so path is set and file exists let key_path = conf.auth_validation_public_key_path.as_ref().unwrap(); Some(JwtAuth::from_key_path(key_path)?.into()) } }; info!("Using auth: {:#?}", conf.auth_type); // Spawn a new thread for the http endpoint // bind before launching separate thread so the error reported before startup exits let cloned = auth.clone(); let http_endpoint_thread = thread::Builder::new() .name("http_endpoint_thread".into()) .spawn(move || { let router = http::make_router(conf, cloned); endpoint::serve_thread_main(router, http_listener) })?; join_handles.push(http_endpoint_thread); // Spawn a thread to listen for connections. It will spawn further threads // for each connection. let page_service_thread = thread::Builder::new() .name("Page Service thread".into()) .spawn(move || { page_service::thread_main(conf, auth, pageserver_listener, conf.auth_type) })?; join_handles.push(page_service_thread); for handle in join_handles.into_iter() { handle .join() .expect("thread panicked") .expect("thread exited with an error") } Ok(()) }
main
identifier_name