text
stringlengths
21
4k
label
stringclasses
2 values
NEW APPROACH TOWARDS SMART BIOSENSORS: AI-MEMRISTIVE BIO-DETECTION CENTRE FOR MICROSYSTEM TECHNOLOGY (CMST) Student: Manel Bouzouita Supervisors: Ioulia Tzouvadaki, Zayer Fakhreddine ABSTRACT Memristors offer high density, low energy consumption, memory effect capability, and advanced sensing capabilities, making them ideal for revolutionizing nonvolatile memory. We are developing a nanosensor with low power consumption that uses machine learning to accurately and efficiently detect biomolecules. Our simulations and machine learning classification models confirm the potential for a reliable and seamless AI-powered biosensing platform that could revolutionize biological system monitoring. State of Art Non-volatile memory devices • Non-volatile memory is a form of computer memory that can maintain stored data even when power is turned off. • It commonly pertains to semiconductor memory chips, containing information in floating-gate memory cells made up of floating-gate MOSFETs, which include NAND flash and solid-state drives (SSD) as flash memory storage[1]. • Memristors have the capability to remember their final state even after the current flow has been stopped, thereby rendering them a useful option for non-volatile memory[2]. Dynamic Memristor: Memristive Crossbar (MCR) structure • The formation and rupture of conductive filaments with nano-scale width are observed to occur across the entirety of the oxide layer, resulting in the desired set and reset transitions (figures 1.1 1.2 1.3 1.5). These phenomena are contingent upon the threshold voltage bias and the impact of joule heating[5]. • The MCR refers to an advanced physical model of a metal oxide memristor (figure 1.3) that draws its inspiration from Sungho Kim's research. • It considers various factors that impact resistive switching behaviors, such as electric field, temperature, oxygen vacancy concentration gradient, and material parameters [3]. Memristive Biosensing • A memristive biosensor is a non-volatile memory device that is capable of sensing while functionalized. • The silicon nanowire-based biosensor (figure 2)is composed of two back-to-back Schottky diodes with opposite functions [5]. • The metal oxide memristive biosensor is made of Metal-Insulator-Metal (MIM) capacitive resistive (MCR) architecture (figure 3) and comprises a unipolar metal oxide memristive device. AI advanced Biosensing • Employing supervised machine learning techniques in the context of memristive biosensors encompasses several key aspects, with output labelling being a pivotal attribute. • This process enables the categorization of sensor output data into distinct groups, contingent upon the specific model and encoding scheme chosen for the analysis. • This preprocessing involves a sequence of stages, including: – Data extraction from the memristive biosensor. – Treatments such as filtration and scaling. These preparatory measures are instrumental in ensuring that the data fed into the algorithm is: devoid of noise, exhibits high reliability, and offers the utmost accuracy, • The ML intervention for biosensing has a corrective capability of enhancing the sensor's specificity and bolstering its ability to detect anomalies, consequently underlining the promise of delivering high specificity and anomaly-free detection outcomes[6]. System Overview: AI-Based Memristive Biosensing Primary Results And Discussion Results of Dynamic MCR modelling: Figure 5 Results of ML modelling: Table 1 Contact Manel.Bouzouita@ugent.be www.ugent.be/elis Universiteit Gent @ugent Ghent University Acknowledgement The research endeavor was made possible thanks to generous financial support from the esteemed University of Monastir and the distinguished Faculty of Sciences of Monastir, Tunisia. Such funding has allowed for the pursuit of valuable knowledge and advancements in the respective fields of study. Fig.(2) Fig.(1) References [1] ‘Non-volatile memory’, Wikipedia. Sep. 07, 2023. Accessed: Oct. 02, 2023. [Onlin
poster
Salinidad y pH determinan los patrones de distribución de la macrofauna asociada a manglares en restauración en Isla del Carmen, México Introducción Objetivos Materiales y Métodos Resultados Nancy Yolimar Suárez-Mozo1, Gabriel M. Moulatlet3, Rosela Pérez-Ceballos1,2, Mariana V. Capparelli1* 1Estación El Carmen, Instituto de Ciencias del Mar y Limnología, Universidad Nacional Autónoma de México, Carretera Carmen-Puerto Real km 9.5, 24157 Ciudad del Carmen, México. 2Consejo Nacional de Humanidades de Ciencias y Tecnologías (CONAHCYT), Mexico. 3Department of Ecology and Evolutionary Biology, University of Arizona, Tucson, Arizona, USA. *mcapparelli@cmarl.unam.mx Variables ambientales Composición de moluscos y crustáceos Discusión Degradado Referencias Composición de moluscos y crustáceos vs. variables ambientales 1274 abundancia total 10 spp Moluscos 5 spp Crustáceos Síguenos en Instagram Parámetros ambientales • Temperatura • Salinidad • pH • Nitritos • Nitratos • Amonio • Fosfatos • Sulfuros Sedimentos: tamaño de grano, carbonatos y materia orgánica Organismos estudiados: Moluscos y crustáceos Análisis de datos • Variables ambientales: PCA • Composición de especies: abundancias, número de especies, análisis de ordenación Análisis estadísticos: • Árbol de decisiones (Linktree) • Distribución de especies: diagrama de sombras y clúster Estrategia de muestreo Nivel de restauración de manglares: • Natural • Degradado • Restaurado • 10 estaciones Área de estudio Isla del Carmen La ONU declaró 2021-2030 como el decenio de la "Restauración de los Ecosistemas" y las "Ciencias Oceánicas”1. Durante la restauración de manglares, ocurren cambios físicos, disminuye la salinidad y se establecen condiciones ideales para los organismos2. La restauración de manglares requiere un análisis temporal y la incorporación de variables bióticas y abióticas. Figura 2. Abundancia de individuos de Moluscos (negro) y Crustáceos (gris) para cada uno de los sitios muestreados. Figura 3. Número total de especies para cada uno de los sitios muestreados. Figura 1. Análisis de componentes principales (PCA) de variables ambientales medidas en 10 estaciones en diferentes áreas de restauración. Porcentaje de variabilidad por eje y las variables se muestran como vectores. Figura 4. Análisis de LINKTREE con agrupaciones ambientales y faunísticas correlacionadas con el nivel de restauración de manglares. Figura 5. Diagrama de sombras de muestras y especies según el nivel de restauración. El sombreado en negro representa la presencia de especies. Figura 6. Ocurrencia de fauna de crustáceos y moluscos encontrada en tres niveles de restauración (restaurado, natural y degradado) en relación con las concentraciones de salinidad. • La restauración hidrológica mejora las condiciones ambientales, pero la salinidad sigue siendo alta en manglares restaurados, lo que afecta la diversidad de especies y su recolonización. • La restauración ha permitido un aumento de la diversidad de cangrejos y moluscos en comparación con sitios degradados, pero la abundancia de especies varía según los tiempos de restauración. • El cambio climático intensifica los desafíos de la restauración, como la salinización, lo que afecta negativamente la composición de especies y la funcionalidad del ecosistema. Restaurado Natural Degradado Salinidad (UPS) R2016 N3 R2010 N1 N4 R2020 N2 R2015 D1978 D2004 0 20 40 60 80 100 A% A B C D E F G Analizar cómo la diversidad de cangrejos y moluscos, se ve influenciada por las condiciones de manglares en distintos niveles de restauración y qué variables abióticas pueden ser cruciales para restablecer estos organismos biológicos. Salinidad pH Los cangrejos y los moluscos forman parte de los grupos dominantes de invertebrados en los manglares y desempeñan importantes funciones ecológicas3. 1Ferreira, A.C., Freire, F.A.M., Rodrigues, J.V.M., Bezerra, L.E.A., 2022. Mangrove Recovery in Semiarid Coast Shows Increase of Ecological Processes from Biotic and Abiotic Drivers i
poster
Study on Optimal Dose of Sodium Valproate (VPA) in Treating Bipolar Mood Disorder (BMD) Patients in Hospital Melaka Chong Lai Peng1,Siti Rahmah Johari1, Lim Kar Ying1, Tay Eek Poei1,Ng Siok Shen1, Koo Suk Kuan1 P-73 Institutional logo (optional) NMRR NO: NMRR-17-2035-36555 Discussion Methodology Conclusion Results Variable With Admission (N = 20) Median (IqR) Without Admission (N = 29) Median (IqR) Z stat * p value* VPA dose 921.5 (371) 600 (396) -2.93 0.003 Serum level 65.5 (57.8) 46.5 (18.3) -0.428 0.668 Variable Acute BMD (N = 20) Mean (SD) Remission (N = 20) Mean (SD) Mean of Dose difference (95% CI) T stat (df) p value* VPA dose 984.2 (318.5) 935.6 (318.6) -48.6 (-197.951, 100.751) -0.681 (19) 0.504 Serum level 67.6 (33.3) 59.9 (43.1) -7.67 (-33.31, 17.97) -0.707 (7) 0.502 Table 1: Comparison VPA doses and serum level between patients with history of admission and without admission in Hospital Melaka Table 2: Comparison VPA doses and serum level during acute and remission phase among patients have been admitted in Hospital Melaka Sodium Valproate (VPA) has been commonly used as mood stabilizers worldwide in treating Bipolar Mood Disorder (BMD). Suboptimal dosing or poor clinical responses towards VPA often resulted in relapse episodes among BMD patients. Evaluation of aggressive initial dosing and serum level with efficacy and BMD relapses are crucial in maximizing clinical benefits while minimizing undesirable adverse events or incidence of relapses. This study aims to determine the optimal dose of VPA in management of BMD. A cross sectional study was conducted in Psychiatric Clinic Hospital Melaka from August 2017 till August 2019. Statistical analysis was done by using SPSS version 21. A value of p< 0.05 was considered statistically significant. Patients with history of relapse requiring hospitalization were on higher dose and achieved higher serum level of VPA than patients without history of relapse although the latter was not statistically significant. The median serum level in non relapse patient was lower than the recommended target range of 50-125 mcg/ml. In the subgroup of patients with history of relapse requiring admission, neither mean dose nor serum level of patients before and after admission was significantly different. Other studies have revealed possible correlation between severity of disease and serum level of VPA required for disease stabilisation. Nevertheless, other contributing factors such as age, body weight and concomitant therapy need to be taken into consideration. As for safety profile, 12 patients on VPA reported at least one adverse effect, with most common adverse effect being sedation and weight gain. None of the adverse effect resulted in withdrawal of therapy. Our study was unable to determine an optimal dose of VPA in the management of BMD. Future study should include other contributing factors that may affect disease control. *Mann Whitney test *paired t-test Acknowledgement We would like to express our gratitude and appreciation to Director General of Health, Director of GHM, Chief Pharmacist of GHM, Psychiatric Clinic Hospital Melaka, and those who have directly and indirectly contributed to the success of this study. Introduction 1Hospital Melaka, Ministry of Health Figure 1: Flowchart describing study flow and data collection process for the analysis of VPA doses and serum level Figure 2: Adverse events among BMD patients treated with VPA Sedation Weight Gain Thrombocytopenia Oedema Others Tremor BMD pt on VPA n=68 Without admission* n = 41 With admission* n= 27 Recruited n=29 VPA dosing and serum level Recruited n=20 *Admission: hospitalization due to disease relapse #Excluded due to incomplete data Excluded# n=12 Excluded# n=7
poster
The widelywidely--used online mixing system has a small range of mixture ratio and a detection problem at small flow ratio used online mixing system has a small range of mixture ratio and a detection problem at small flow ratio of pesticide. This paper developed an online of pesticide. This paper developed an online--mixing system, which mixed pesticide into carrier directly.mixing system, which mixed pesticide into carrier directly.The The pesticide were injected to the mixer from pesticide tang through a piston pump, witch controlled by stepper pesticide were injected to the mixer from pesticide tang through a piston pump, witch controlled by stepper motor. A solenoid valve was used to control water flow into the buffer tank; the water level of the buffer tank was motor. A solenoid valve was used to control water flow into the buffer tank; the water level of the buffer tank was controlled using two liquid level sensor feedback. The water flow rate of the spray pump from the buffer tank was controlled using two liquid level sensor feedback. The water flow rate of the spray pump from the buffer tank was measured using a turbine flowmeter. According to the water flow rate, the proportion of pesticide was injected into measured using a turbine flowmeter. According to the water flow rate, the proportion of pesticide was injected into the mixer with water together. The mixture of pesticide and water was sprayed to the target. the mixer with water together. The mixture of pesticide and water was sprayed to the target.During During the spraying, water and pesticide was measured at the same time, the maximum error of the mixture ratio the spraying, water and pesticide was measured at the same time, the maximum error of the mixture ratio was 6.75%. The maximum error of mixture uniformity was 6.33%, and the variation coefficient was 7.91%. was 6.75%. The maximum error of mixture uniformity was 6.33%, and the variation coefficient was 7.91%. The The system proposed supplied the water and pesticide in proportion, and the range of the mixture ratio was 100:1 ~ system proposed supplied the water and pesticide in proportion, and the range of the mixture ratio was 100:1 ~ 1000:1, it can satisfy the requirements of spraying in paddy field. 1000:1, it can satisfy the requirements of spraying in paddy field.The The experiment was used the cochineal solution as the pesticide, the next step will be to carry out field experiments experiment was used the cochineal solution as the pesticide, the next step will be to carry out field experiments with real pesticides. with real pesticides.Design and experiment of online mixing spraying Design and experiment of online mixing spraying system systemAdress Adress::483# 483# WushanWushanStreet, Street, TianheTianheDistrict, Guangzhou, P.R. China District, Guangzhou, P.R. ChinaWebsite Website::http://www.nyhklm.org/http://www.nyhklm.org/Contact Contact::LianLianHU HUTel Tel::+86+86--2020--38676975 Fax38676975 Fax::+86+86--2020--85280158 Email85280158 Email::lianhu@scau.edu.cnlianhu@scau.edu.cnCollege of Engineering, South China Agricultural University College of Engineering, South China Agricultural UniversityKey Laboratory of Key Technology Key Laboratory of Key Technologyon Agricultural Machinery and Equipment, Ministry of Education on Agricultural Machinery and Equipment, Ministry of Educationonline online mixingmixingspaying system spaying system Meng Zhang, Lian Hu, Qibao Yuan, Runmao Zhao, Lingmao Tang, Hao Zhou, Tong Wu MAX MIN Water Pesticide 控制板 Controller Precision plunger pump Level switch A Level switch B Flowmeter Solenoid Buffer Tank Mixer 0 100 200 300 400 500 600 700 800 900 1000 Artificial mixture Machine mixing 1 Machine mixing 2 Machine mixing 3 Measuring point concentration/ppm Buffer tank left Middle Right The Theconcentration of the measurement points concentration of the measurement points
poster
www.postersession.comXXIX International Conference on Neutrino Physics and Astrophysics, June 22 to July 2, 2020. XXIX International Conference on Neutrino Physics and Astrophysics, June 22 to July 2, 2020. Exposure-background duality in the searches of neutrinoless double beta decay M. K. SinghInstitute of Physics, Academia Sinica, Taipei, Taiwan Institute of Physics, Academia Sinica, Taipei, Taiwan Based on: M. K. Singh, H. T. Wong et al., Phys. Rev. D 101, 013006 (2020). Background can be predicted. Poisson statistics handles: (i) Low background; (ii) Rare processes. Model - Inverse correlation between G0ν and |M0ν|2. Decay rates (1 event/ton-yr with full efficiency) are similar at given <mββ> and constant gA. No favored 0νββ isotope. ∑ (ton-year) . ɛ𝐑𝐨𝐈 𝑵𝐨𝐛𝐬 𝟎ν α 𝟏 <𝒎𝜷𝜷> 𝟐 Realistic interpretation lies within a factor of [0.5, 2.0]. RoI = w1/2 (FWHM): Not the optimal choice when B0 →0. Alternative choice: RoI ≡w3σ (Qββ ± 3σ), thus εRoI ≅100%. Better sensitivity by a factor of εRoI (w1/2 = 0.76). Covered 𝑻𝟏/𝟐 𝟎𝝂is 32% longer, or Σ is 24% less. Background Index (BI) defined as: BI= 𝐁𝟎(𝐑𝐨𝐈) ∑ Universally applicable. BI < 0.21 counts/w1/2-ton-yr cover <𝒎𝜷𝜷𝟗𝟓% 𝐈𝐇> BI < 0.033 counts/w1/2-ton-yr cover <𝒎𝜷𝜷− 𝐈𝐇> Standard-Model-allowed irreducible background 𝒁 𝑵𝑨𝜷𝜷→ 𝒁+𝟐 𝑵−𝟐𝑨𝜷𝜷+ 𝟐 𝒆+ 𝟐 𝝂[Goeppert-Mayer, 1935] Worse resolution (Δ) Larger RoI Larger 2νββ Background Resolutions under BImin conditions Δ ≤ (0.3-0.9)% cover <𝒎𝜷𝜷− 𝐈𝐇> Δ ≤ (0.1-0.3)% cover <𝒎𝜷𝜷− 𝐍𝐇> ♣MJD Best Resolution Δ(76Ge) = 0.12% BI < 6×10−10 counts/(w1/2-ton-yr) Covering <𝒎𝛃𝛃−𝐍𝐇> will require large and costly exposure. Reduction of BI will be playing increasingly significant. Same exposure can probe longer 𝑻𝟏/𝟐 𝟎ν and smaller <mββ> with decreasing background. ▲Half-life in Mass Mechanism : 𝟏 𝑻𝟏/𝟐 𝟎𝝂 = 𝑮𝟎𝝂𝒈𝑨 𝟒𝑴𝟎𝝂𝟐<𝒎𝜷𝜷> 𝒎𝒆 𝟐 ▲Effective Mass: <mββ> = |U𝒆𝟏 𝟐|m1 + |U𝒆𝟐 𝟐|m2eiα + |U𝒆𝟑 𝟐|m3eiβ| ▲Experimentally measurable Half-life: ▲Combined Half-life: 𝑻𝟏/𝟐 𝟎𝝂= 𝒍𝒏𝟐. 𝑵𝑨𝜷𝜷. 𝒕𝐃𝐀𝐐. ɛ𝑹𝒐𝑰 𝑵𝒐𝒃𝒔 𝟎𝝂 = 𝒍𝒏𝟐. 𝑵𝑨 𝑴(𝑨𝜷𝜷) . ∑. ɛ𝑹𝒐𝑰 𝑵𝒐𝒃𝒔 𝟎𝝂 𝑴𝟎𝝂𝟐[𝒈𝑨 𝟒. 𝑯𝟎𝝂] = 𝟏 <𝒎𝜷𝜷>𝟐 𝟏 ∑. 𝑵𝒐𝒃𝒔 𝟎𝝂 ɛ𝑹𝒐𝑰; 𝑯𝟎𝝂≣𝒍𝒏𝟐 𝑵𝑨 𝑴𝑨𝜷𝜷.𝒎𝒆𝟐 Formulation Discovery Potential ∑𝐢=𝟎 𝐍𝐨𝐛𝐬 𝟑𝛔−𝟏 𝐏𝐢; 𝐁𝟎≥𝟏−𝟎. 𝟎𝟎𝟏𝟑𝟓 ∑𝐢= 𝐍𝐨𝐛𝐬 𝟑𝛔 ∞ 𝐏𝐢; [𝐁𝟎+ 𝐒𝟎] ≥𝟎. 𝟓 Only integer counts can be observed in an experiment. Theme Theme::T To reach the following target sensitivities: (i) <𝒎𝜷𝜷− 𝐈𝐇> = 1.4×10-3 eV, (ii) <𝒎𝜷𝜷+ 𝐈𝐇> = 5.1×10-2 eV, (iii) <𝒎𝜷𝜷𝟗𝟓% 𝐈𝐇> = 2.0×10-2 eV (iv) <𝒎𝜷𝜷− 𝐍𝐇> = 7.8×10-4 eV, (v) <𝒎𝜷𝜷+ 𝐍𝐇> = 4.3×10-3 eV, (vi) <𝒎𝜷𝜷𝟗𝟓% 𝐍𝐇> = 3.0×10-3 eV. Relating 𝑻𝟏/𝟐 𝟎𝝂with <mββ> - Model for |M0ν| Required Exposure and Background Limiting Irreducible Background Summary Conversion to Realistic Configurations Considered: IA = 100%, εexpt = 100% and “unquenched” free nucleon value of gA =1.27 Conversion: In realistic experiments: Σ’ ≃Σ. 𝟏 IA . 𝟏 ɛexpt .WΣ(gA) and BI’(Σ’) ≃BI· Σ Σ’ § Neutrinoless double beta decay (0νββ) [Furry, 1939] 𝒁 𝑵𝑨𝜷𝜷→ 𝒁+𝟐 𝑵−𝟐𝑨𝜷𝜷+ 𝟐 𝒆 ●Forbidden in Standard Model !!! ●∆L = 2 !!! § Observation of 0νββ implies new physics:● ●Neutrinos Neutrinosare areMajorana Majoranaparticles particles( (νν= =ത ν) )● ●Lepton Leptonnumber numberviolations violations● ●Effective Effectivelight lightMajorana MajoranaNeutrino NeutrinoMass Mass< <mmββββ>>≠≠0 0 § Energetically possible for 35 nuclei ●A few are experimentally relevant § Present work: Required Exposure vs Background Introduction STEP-1 STEP-2 R. G. H. Robertson, Mod. Phys. Lett. A 28, 1350021 (2013). Target exposure: Next-generation 0νββ projects 10 ton-year to cover IH GERDA: Background level = 𝟏. 𝟎−𝟎.𝟒 +𝟎.𝟔counts/keV-ton-yr or BI ∼3 counts/w1/2-ton-yr. Choice: BI ≡ BI0 = 1 count/w1/2-ton-yr requires 110 ton-yr cover <mββ>− 𝐈𝐇 11 Mton-yr cover <mββ>− 𝐍𝐇 Suppression in BI: [1 to 10−3] counts/(w1/2-ton-yr) contributes reduction in exposure BImin equivalently Σmin condition: Single observed event can establish signal at 𝐏𝟓𝟎 𝟑𝛔 Preferred hardware specification space Σ=10 ton-yr Next generation: BImin<5.1×10−5 counts/w1/2-ton-yr <mββ> > (5.8×10−3)eV <mββ>+ 𝐍𝐇= 4.3×10−3eV Be
poster
Attoclock revisited on electron tunnelling time Cornelia Hofmann1,2, Alexandra S. Landsman1,3, Ursula Keller2 Attoclock Method [1-3] 1Max Planck Institute for the Physics of Complex Systems, 01187 Dresden, Germany 2Department of Physics, Institute of Quantum Electronics, ETH Zurich, 8093 Zurich, Switzerland 3Max Planck Postech/Department of Physics, Pohang, Gyeongbuk 37673, Republic of Korea [1] Hofmann, et al., http://arxiv.org/abs/1901.07015 [2] Eckle, et al., Science, 322, 1525-1529 (2008) [3] Landsman, et al., Optica 1, 343 (2014) [4] Dörner, et al., Phys. Rep. 330, 95 (2000) [5] Weger, et al., Opt. Express 21, 21981–21990 (2013) [6] Camus, et al., Phys. Rev. Lett. 119, 023201 (2017) [7] Wigner, Phys. Rev. 98, 145–147 (1955) [8] Fortun, et al., Phys. Rev. Lett. 117, 010401 (2016) [9] Sainadh, et al., http://arxiv.org/abs/1707.05445 [10] Bray, et al., Phys. Rev. Lett. 121, 123201 (2018) [11] Ivanov, Kheifets, Phys. Rev. A 89, 21402 (2014) [12] Hofmann, et al., New J. Phys. 18, 043011 (2016) [13] Teeny, et al., Phys. Rev. Lett. 116, 063003 (2016) [14] Ni, et al., Phys. Rev. A, 98, 013411 (2018) [15] Ivanov, et al., Comm. Phys. 1, 81 (2018) [16] Emmanouilidou, et al., J. Phys B: At. Mol. Opt. Phys. 48, 245602 (2015) [17] Majety, Scrinzi, J. Mod. Opt. 0340, 1–5 (2017) [18] Reiss, J. Phys B: At. Mol. Opt. Phys. 47, 204006 (2014). [19] Büttiker, Phys. Rev. B 27, 6178–6188 (1983) References Principle: final photoelectron momentum encodes moment when electron first entered the laser field dominated continuum Measurements: ● helium ionized by strong-field near IR ● ● Keldysh Parameter between 0.75 and 4 ● COLTRIMS [4] & VMIS [5] experimental setup ● random Carrier-Envelope-Offset phase ● observable: streaking offset angle of the photoelectron momentum distribution Other Experiments Krypton/argon Attoclock [6] ● Comparing streaking angle between two species ● gas mixture in COLTRIMS [4] ● Wigner trajectory [7] calculations with delay matches Non-Adiabatic Effects Dipole Approximation Summarized Influence on Attoclock Interpretation [1] Larmor time [19] and Feynman Path Integral method [3] yield estimates comparable to measured Attoclock tunnelling delays Acknowledgements Assumptions Single Active Electron (SAE) Time Reconstruction: ● Single Classical Trajectory (SCT) simulation assuming instantaneous tunnelling ● offset angle = ellipticity correction + delay BEC of rubidium atoms in optical lattice [8] ● kicked lattice → oscillating wave packet ● lattice turned off → instantaneous momenta mapped to location on detector ● D2 (tunnelled) is delayed compared to D1 (reflected) Atomic hydrogen Attoclock [9] ● Measurement compared to TDSE agree ● approximation for Coulomb scattering [10] ● TDSE with Yukawa potential: ● tunnelling delay ● ionisation time extracted from experiment Attoclock assumption ● tunnelling most probable when field is at maximum ● = peak of the pulse envelope = 0 approaches solving the TDSE ● probability current density [13] maximised at < 0 ● classical packpropagation after TDSE forward calculation [14] finds classical turning point ● instantaneous ionization rate depending on fluctuations of the strong field [15] contradicts = 0 → starting time before the peak of the envelope would result in even larger extracted tunnelling delays Post-ionization [16] ● classical trajectories ● 2 electrons (bound- free) yields same result as SAE assuming perfect screening Full process [17] ● TDSE calculation where one electron is allowed to ionise ● compared to SAE TDSE ● no electron correlation visible in the angular photoelectron momentum distribution ● agreement with non-adiab. calibrated measurement for time-dependent Schrödinger equation (TDSE) [11] ● initial transverse momentum → lower field strength calibration for same data [12] ● energy gain → shorter exit radius → still remaining offset angle difference ● near IR ≈ 800 nm wavelength ● Intensity below 8∙1014 W/cm2 → spatial dependence of the electric field neglected → magn
poster
Acknowledgments & References The modeling study is partially funded by the Geological Society of America and American Association of Petroleum Geologist. We thank Husky Energy for the general support of our modeling lab. We also thank TGS/Nopec Geophysical Company, ConocoPhillips, Canadian Department of Natural Resources, Canada-Nova Scotia Offshore Petroleum Board, and Suncor Energy for providing the seismic data used in this study. A special thanks also goes to Schlumberger for providing Petrel, the software tool that we use to interpret the seismic data. Withjack, M. O. and Callaway, S., 2000. Active normal faulting beneath a salt layer: An experimental study of deformation patterns in the cover sequence. AAPG Bulletin, v. 84, no. , p. 627-651. Eisenstadt, G. and Sims, D., 2005. Evaluating sand and clay models: do rheological differences matter? Journal of Structural Geology, v. 27, p. 1399-1412. Withjack, M. O., Schlische, R. W., Olsen, P. E., 2010. Development of the passive margin of eastern North America - Mesozoic rifting, igneous activity, and breakup, in Roberts, D. G., and Bally, A. W., eds., Phanerozoic Regional Geology of the World. Summary & Future Work Thickness of clay layer affects deformation patterns Distribution of clay layer may affect deformation patterns Kinematic model of Orpheus rift basin Future Work The modeling results and strain analysis suggest that the internal clay layer underwent not only folding but also thickness changes during deformation. Model A (thin clay) Standard Model (Initial depth 1 cm) Model D (Initial depth 0.5 cm) Varying depth of clay layer Multiple internal clay layers Model C (thick clay) 1 2 3 5 3 km a Late Triassic ? 3 km c ROU ROU CAMP-related intrusions Buckle folds Detached thrust faults Latest Triassic - earliest Jurassic ? 3 km b Deposition of interbedded salt and salt impurites Late Triassic ? Late Rifting: Minibasins and CAMP Intrusion Middle Rifting: Upper Argo Formation Early Rifting: Lower Argo Formation ROU 4 (Left) Map showing early Mesozoic rift basins (green areas), including Orpheus rift basin, as part of eastern North American rift system formed during break-up of Pangea and general stratigraphy of the Orpheus rift basin. The Orpheus rift basin contains Late Triassic-Early Jurasssic salt of the Argo Formation. Note the presence of salt impurities within the Argo salt. ROU: Rift-onset unconformity; BU: Breakup unconformity; AvU: Avalon unconformity: CU: Cenozoic unconformity; CAMP: Central Atlantic Magmatic Province. Yellow line is the seismic line shown in this study. (Below) Similarities between modeling results (Model A) and structures in Orpheus rift basin. Similar to the models, buckle folds in the synrift section of the Orpheus rift basin are present adjacent to a sedimentary load (i.e., minibasin). Seismic line is 1:1 at 4.5 km/s. 5 km 1 s (TWTT) North South 5 cm Prerift Postsalt minibasin Postrift Synrift salt Igneous sheets Synrift presalt Comparison with Natural Examples Boudins Overturned folds Minibasin Upright gentle folds 400 km North Atlantic Ocean North America N Orpheus Basin modified from Withjack et al. (2010) Period PERMIAN & OLDER TRIASSIC Postrift Prerift Synrift EARLY JURASSIC & YOUNGER Late Early Mid Strati- graphy Tectonic Stages AvU CU BU CAMP Mini- basin ROU Argo Salt Preliminary Modeling Results: Models with various depth of internal clay Time: 5 hours 5 cm Standard Model (Putty:Clay = 4:1) Initial depth of clay layer: 1 cm Model C (Putty:Clay = 4:1) Initial depth of clay layer: 0.5 cm 5 cm Time: 5 hours 40 30 0 10 20 Extension (%) Strain analysis of the internal clay layer Line of equal value Shortening (%) 30 40 20 10 Standard Model Model A Model B Model C Strain Analysis Percentage of extension vs shortening of the internal clay layer in all models. The amount of shortening in all models is less than the amount extension. Zone of extension Zone of shortening % Extension or shortening = (Lf-Lo)/Lo x 100% Lf = Final Length; Lo = Ori
poster
Hybrid genome assembly of the northern acorn barnacle, Semibalanus balanoides Rebecca Elyanow, Bianca Brown, Joaquin Nunez, David Rand Brown University, Center for Computational Biology, 115 Waterman st., Providence, RI 02906 III. Methods I. Introduction Assembly of Illumina data with SOAPdenovo VI. Conclusion References V. Genome size estimation VII. Future work Barnacles have long been recognized as good model organisms in the field of population ecology and have been well studied in this context. However, little is known about the genomic structure and evolution of barnacle species. The burgeoning field of ecological genomics aims to merge insights from the fields of ecology and genomics into a unified understanding of how a population has been shaped by its environment, from a cellular level to an ecological level. The goal of this project is to develop the northern acorn barnacle, Semibalanus balanoides, as a robust model for ecological genomics. The first step in achieving this goal is to create a draft genome for S. balanoides, which will provide a reference for any future genomic studies. Assembly Metrics Mapping of Illumina reads BLAST results 1. Estimate allele frequencies of barnacles from different microhabitats (high and low tide) through pool-sequencing. See poster 209 for preliminary work. 2. Identify bacterial species that compose microbiome of S. balanoides in different microhabitats. See poster 397 for preliminary work. 3. Perform transcriptome assembly using RNA-seq data and annotate functional sites on the genome 1305 2232 18455 20986 0 5000 10000 15000 20000 25000 SOAPdenovo dipSPAdes Canu DBG2OLC N50 ASSEMBLY N50 31 98 380 525 0 100 200 300 400 500 600 SOAPdenovo dipSPAdes Canu DBG2OLC GENOME SIZE (MB) ASSEMBLY Genome Size (Mb) 2.43 15.67 56.56 70.83 0 10 20 30 40 50 60 70 80 SOAPdenovo dipSPAdes Canu DBG2OLC PERCENT OF MAPPED READS ASSEMBLY Percentage of Illumina reads mapped to assembly II. What is hybrid genome assembly? Hybrid genome assembly is the process of joining raw sequencing data into long contigs using both short-read and long-read data. • Long-reads are easier to assemble than short reads because they contain long range information and can span repeats Short paired-end read from Illumina sequencing: 2x100bp Long-read from PacBio sequencing: 6,258bp • Short-reads have a much lower error rate than long reads (0.1% vs 10%) • By combining long and short-reads together, we leverage both the long-range information from long-reads and the low error-rate of short reads Assembly of Illumina data with dipSpades Assembly of PacBio data with Canu Hybrid assembly of Illumina and Pacbio data with DBG2OLC IV. Results Raw Illumina paired-end reads SOAPdenovo assembly Raw Illumina paired-end reads Raw PacBio SMRT Cell reads Canu error- correction Raw Illumina paired-end reads DBG2OLC assembly Raw PacBio SMRT Cell reads dipSPAdes assembly Sparc consensus calling Canu trimming Canu assembly • 237 million paired-end 85bp Illumina reads were assembled by dipSPAdes • dipSPAds is designed for diploid, highly polymorphic genomes • dipSPAdes assembles haplocontigs (representing both haplomes) then produces consensus contigs and performs haplotype assembly • 237 million paired-end 85bp Illumina reads were assembled by SOAPdenovo • Contigs longer than 1kb were selected • 3 million PacBio SMRT cell reads with an average length of 6kb were assembled with Canu • The MinHash Alignment Process (MHAP) is used to detect overlapping reads and correct errors • Non-confident regions are removed through trimming • Trimmed reads are assembled with the Celera assembler • 3 million PacBio SMRT cell reads and 237 million Illumina reads were assembled with DBG2OLC • DBG2OLC uses de Bruijn Graph (DBG) to construct contigs from Illumina reads, which are used to correct PacBio reads. Corrected PacBio reads are assembled using an Overlap Layout Consensus (OLC) • Parameter tuning increased the assembly size from 409Mb to 525Mb 0 0.1 0.2 0
poster
The authors acknowledge support from the Climate Compatible Growth programme (#CCG) and the UK Partnering for Accelerated Climate Transition (UK PACT) programme of the UK's Foreign Development and Commonwealth Office (FCDO). The views expressed in this research do not necessarily reflect the UK government’s official policies. REFERENCES • UNESCO. 2021. ‘UNESCO Recommendation on Open Science’. SC-PCB- SPP/2021/OS/UROS. Paris, France: UNESCO. • Wilkinson, Mark D., Michel Dumontier, IJsbrand Jan Aalbersberg, Gabrielle Appleton, Myles Axton, Arie Baak, Niklas Blomberg, et al. 2016. ‘The FAIR Guiding Principles for Scientific Data Management and Stewardship’. Scientific Data 3 (1): 160018. • Open Data Charter. 2015. ‘International Open Data Charter’. opendatacharter.net. There is an increased recognition of the need for openness, transparency, and accessibility of both governance and technical aspects of long-term systems planning, including data to inform the planning process. In Kenya, the new constitution is pushing for reforms aiming at increasing the participation of institutions and county govt. into decision-making processes. This is valid for energy data management and governance. Hence, this paper investigates: how can an open data approach to data management and governance effectively support robust and inclusive long-term energy planning in Kenya? RESULTS Overlap of roles and responsibilities System Design: aiming for up-to-datedness, accuracy, security, and reliability. • Centralised system with controlled access, need for easy data aggregation/disaggregation function Transparency and Accessibility: • Positive: data as resource to support decision- making in society and create trust) • Critical: fear of data being misused, distorted, expose to scrutiny. Open Data value: positive, but with concerns (i.e. credibility, up-to-datedness, security). • Strengths: free and easy accessibility, good formatting, useful for complementing and validating existing datasets. • Weaknesses: highly aggregated, outdated, non- validated, non-credible. METHODOLOGY 1. Stakeholders’ mapping: National + County govt., energy utilities, statistical office, energy regulator, International Organisation 2. Qualitative Document Analysis • 2010 Constitution of Kenya • 2018 National Energy Policy • 2019 Energy Act 3. Semi-structured Interviews: • 10 stakeholders • focus on: system design, transparency and accessibility, the value of open data 4. Assessing the evidence against open science values and principles • UNESCO Recommendation on Open Science • Findable, Accessible, Interoperable, Reusable (FAIR) Data • International Open Data Charter (ODC) NEXT STEPS: UK PACT project, year 2 • Supporting stakeholders in identifying common datasets of interest and comfortable ‘degree of openness’ for data sharing • Prototyping an energy data repository DISCUSSION AND CONCLUSIONS Open data can address the following: supporting collaboration and simplified data processing; data accessibility, retrievability, and coherence; enabling accountability and citizens awareness. Main challenges: 1) contradictory understanding of what open data entails and how it can be realised in practice affects their suitability; 2) principles and practices differ and require to find compromises. Recommendations: • Need for further investigation as to what resources, expertise and infrastructure is available to handle energy data in Kenya. • Need for considering various degree of openness, as to best fit the Kenya context. County-level planning SIG Energy Data renewable energy data energy resources data energy infrastruct. data petroleum data renewable resources maps electricity T&D data County Governments Grid Operator(s) Cabinet Secretary National Government REREC EPRA Data Governance and Open Science in Energy Planning: A case study of the Kenyan ecosystem A. Beltramo, A. Leonard, J. Tomei, W. Usher Agnese Beltramo, Email: beltramo@kth.se Authors’ affiliations:
poster
CONTACTS project-matel project_matel www.project-matel.eu Next generation of OEIC-based devices: wide-bandwidth, minimized losses & multifunctionality in miniaturized form factors Strengthening the EU's industrial capability in photonics and microelectronics Green and digital manufacturing approach fostering chemicals and waste minimization European strategic autonomy in OEICs Lowering the barriers for OEIC uptake by high- tech SMEs Expected Impacts To demonstrate and validate at TRL5 advanced OEICs for two applications: i) Display & Recreation: AR display featuring a 2D light source ii) Health: Biophotonic sensors. The EU-funded project “MatEl” consists of four pillars encompassing new integration schemes for (i) chip manufacturing (WP2), (ii) advanced materials integration on OEIC (WP3), (iii) Hybrid integration of advanced materials using laser based technologies (WP4) and (iv) PIC & IC packaging and assembly technologies in order to enable the development of two use applications (WP5). This will be achieved by consistently pursuing the following objectives: TECHNICAL OBJECTIVES AR AR display featuring a 2D light source for light-field with on-chip RGB lasers and OEIC-based demultiplexer. Bio-photonic sensors for reliable and low-cost detection of Covid-19 featuring integrated on-chip VCSEL at 850 nm and Graphene-based photodetector. BIO sensing APPLICATIONS Europe’s leading position in photonics and electronics can only be secured by adapting to the next generation of optoelectronic devices requirements: high performance, multi- functionality and cost efficiency in miniaturized footprint. These can only be achieved if novel schemes for on-chip integration emerge. Silicon nitride (Si3N4) is a promising candidate for optoelectronics applications; next to silicon photonics and indium phosphide, Si3N4 photonic integrated circuits have broad spectral coverage and low propagation losses. Still, Si3N4 itself has no active effect (except thermal tuning) and active functionality can be demonstrated either by integrating active components or active materials. The on-chip integration of III-V and II-VI semiconductors on Si3N4 is complicated and costly. The EU-funded project “MatEl” introduces a novel, on-chip integration scheme enabling accurate and fast alignment and bonding of any type of chip package on Si3N4. MatEl will combine laser transfer (LIFT) and laser soldering processes to demonstrate next-gen applications, which will accelerate the industrial adoption of hybrid optoelectronic integrated circuits (OEICs) - offering high-performance, multi-functionality and cost efficiency in a miniaturized footprint. MatEl’s innovative solution, enhanced by the monolithic integration of advanced materials – graphene and high-quality PZT, will thus be demonstrated for two selected next-gen devices at TRL5: CONSORTIUM 1- National Technical University of Athens, Heroon Polytechniou 9 Zographou Campus, 157 80 Athens, Greece 2- LIONIX INTERNATIONAL BV, Hengelosestraat 500, 7521 AN Enschede, Netherlands 3- PHIX BV, De Veldmaat 17, 7522 NM Enschede, Netherlands 4- GRAPHENEA SEMICONDUCTOR SL, Ps Mikeletegi 83, 20009 Donostia, Spain 5- SURFIX BV, Agro Business Park 2, 6708 PW Wageningen, Netherlands 6- AMIRES SRO, Na Okraji 335/42, 162 00, Praha 6, Czech Republic 7- PIEMACS SARL, Epfl Innovation Park, Batiment C, 1015, Lausanne, Switzerland 8- CREAL SA, Chemin De La Dent D'oche 1a, 1024, Ecublens, Switzerland K. Andritsos , M. Makrygianni , F. Zacharatos , S. Kamyar , E. Schreuder , R. Dekker , K. Obara , M. Milosevic , A. Centeno , A. Zurutuza , W. Knoben , M. Messina , M. Chopart , R. Pašek , S. Bagdzevicius , R. Matloub , P. Muralt , J. Gamet , A. Kvasov , I. Zergioti 1 1 1 1 2 2 3 3 4 4 5 6 6 6 7 7 7 7 8 2 MatEl: PZT and Graphene MATerials innovations for advanced opto-Electronic applications in AR and biosensing Acknowledgements: The authors would like to acknowledge funding from the EU Horizon Europe R&I programme under grant agreement 1010
poster
The East African Lake Malawi haplochromine cichlid flock represents one of the largest and most diverse adaptive radiations on earth, with over 700 species of fish. It is an important model system with which to understand ecological diversification, sexual selection and speciation. Our analyses focused on four sympatric species of the endemic deep-water genus Diplotaxodon (fig. 1), previously found to differ significantly in morphology and male monochromatic nuptial color [1] and proposed as an example of sympatric speciation. Introduc)on Single digest RADseq (SbfI) 50 individuals; ~3.5M reads/ind 5 populations > 8 individuals per pop Methods Genomic differen)a)on among deep-­‐water cichlid species References The genomic architecture of speciation in the Lake Malawi deep water cichlid genus Diplotaxodon Christoph Hahn & Domino Joyce School of Biological, Biomedical and Environmental Sciences, University of Hull 1. Genner et al. 2007. MolEcol. 2. Catchen et al. 2013. MolEcol. 3. Jombart & Ahmed 2011. Bioinformatics. 4. Brawand et al. 2014. Nature. 5. Conesa & Goetz. 2008. Int. J. Plant Gen. 6. Foll & Gaggiotti. 2008. Genetics. 7. Coop et al. 2010. Genetics. @EvoHull https://github.com/chrishah Acknowledgements This is a collaboration with Dr Martin Genner, and is funded by NERC grant number NE/ K000829/1. Principal data processing using STACKS [2], adegenet [3] and custom scripts (https://github.com/chrishah). Genome scans are based on the recently published genome of Metriaclima zebra [4]. Functional annotation of genes obtained using Blast2GO [5]. Fig. 2: Number of candidate outlier loci identified using STACKS [1], BayeScan [6], and Bayenv2 [7]. . Fig. 1: Principal component analysis (PCA) based on > 25,000 SNPs, differentiating the 4 putative Diplotaxodon species. . We used Principal Component scores (PC1) inferred from morphological data in [1] as proxy to specifically identify genomic regions associated with morphological differentiation between species (fig. 3). Normalized P C 1 s c o r e s w e r e u s e d a s ‘environmental factors’ in Bayenv2 [7]. Fig 3: Transformed bayes factor ranks averaged across 10 independent Bayenv2 runs. The vertical red line delimits the 95th average rank percentile as threshold for candidate loci. The inset shows the species wide PC1 scores [1] used as ‘environmental factors’ in the analyses. Divergent genomic regions between Diplotaxodon species are enriched for genes controlling craniofacial- eye development (see fig. 4). Our results indicate that the increasingly narrow, short wave- length light-spectra at greater water depths pose unique selective constraints and are likely key drivers of ecological and sexual selection and ultimately speciation in this group of Lake Malawi cichlids. Fig 4: (A-F) Pairwise divergence (FST) across three selected scaffolds. Displayed are observed values and kernel smoothed averages. Red dots represent candidate outlier loci identified by STACKS (p < 5e-7). Y axes range 0 < FST < 1. (A) Di_1 vs Di_2; (B) Di_1 vs Di_4; (C) Di_1 vs Di_5; (D) Di_2 vs Di_4; (E) Di_2 vs Di_5; (F) Di_4 vs Di_5. (G) Average transformed ranks inferred from the ‘morphology informed’ Bayenv analyses (fig. 3). Red dots represent average ranks in the 95th percentile. D. limnothrissa black pelvic (Di_2) D. macrops black dorsal (Di_1) D. macrops ngulube (Di_5) D. macrops offshore (Di_4) c.hahn@hull.ac.uk A B C D E F G Fig. 4 (contd): Rectangles at the bottom represent genes within 15kb windows of candidate outlier SNPs (craniofacial- or eye development). X axes scale: 100kb/tick. Find it on figshare
poster
Assessment of 13 Gambierdiscus strains using neuro-2a and erythrocyte lysis assays Francesco Pisapia, William C. Holland, D. Ransom Hardison, R. Wayne Litaker, Santiago Fraga, Tomohiro Nishimura, Masao Adachi, Lam Nguyen-Ngoc, Véronique Séchet, Zouher Amzil, Christine Herrenknecht, Philipp Hess Scientific poster Gambierdiscus is a genus of benthic dinoflagellates that produce ciguatoxins (CTXs) and maitotoxins (MTXs), which are among the most potent marine toxins known. CTXs are bio-accumulated and biotransformed along the marine food chain and are involved in Ciguatera Fish Poisoning (CFP). Several species have recently been described, and CFP has recently been reported from non-endemic areas, namely the Canary Islands. Little is known about how toxicity varies among species and isolates of Gambierdiscus. This study examined the toxicity of 13 strains of Gambierdiscus, seven from the Pacific Ocean (G. pacificus CCMP 1650, G. sp. VGO 917, G. australes CCMP 1653, G. scabrosus KW070922_1, G. sp. Vietnam, G. caribaeus BILL HI Gam8, G. carpenteri PAT HI Jar7 Gam11), one from the Mediterranean Sea (G. carolinianus Greece Gam2), and five from the North-Eastern Atlantic Ocean (G. australes VGO 1178, G. australes VGO 1181, G. silvae VGO 1167, G. silvae VGO 1180, G. excentricus VGO 791). Algal cells were extracted with methanol, and extracts were partitioned between dichloromethane and aqueous methanol. The toxicity of pre-purified extracts has been evaluated using an ouabain/veratridine neuro-2a assay and a human erythrocyte lysis assay. In general terms, all strains showed hemolytic activity in the aqueous methanol fraction in the range of pg MTX eq/cell, and neuro-2a cytotoxicity in the dichloromethane fraction in the range of fg P- CTX-3C eq/cell. G. excentricus (Canary Islands) was an exception, with a neuro-2a cytotoxicity in the range of pg P-CTX-3C eq/cell, in accordance with data previously reported by Fraga et al. (2011). Keywords: Ciguatera Fish Poisoning, Gambierdiscus, ciguatoxins, maitotoxins, neuro-2a assay, erythrocyte lysis assay.
poster
Innovative Research for a Sustainable Future www.epa.gov/research Richard Judson l email: judson.richard@epa.gov l 919-449-7514 Retrospective Mining of Toxicology Data to Discover Multispecies Effects: Anemia as a Case Study Richard Judson, Charles Wood, Grace Patlewicz, Jeremy Fitzpatrick, Matt Martin U.S. Environmental Protection Agency, Office of Research and Development Introduction Methods Results 1: Database Summary Results 2: Individual Chemicals Conclusions Retrospective inspection of In vivo toxicology data is subject to multiple sources of uncertainty: 1. Lack of control of some key experimental variables: Variability of animals even within the same strain, exact details of diet and animal handling protocols, etc. 2. Statistical issues: Studies are rarely repeated so that experiment-to- experiment variability is rarely characterized. The number of animals used in each dose group may be insufficient to see rare events, etc. 3. Negative results and effects not measured can be hard to distinguish: In guideline studies, specified effects should be examined, and any treatment- related effects recorded. However, this may not be always followed, and some effects not specified by the guidelines may be recorded when seen. 4. Observer bias: When a severe type of toxicity is seen (e.g. a neoplastic lesion), less severe or benign (but treatment-related) effects may be ignored, or not recorded, or recorded in a limited way. Similarly, differences in lesion thresholds among pathologists may influence treatment-related effects. 5. Variability over time in terminology: While this can be managed somewhat by the use of controlled vocabulary onto which the original effect descriptions can be mapped, there are cases where effects are better understood as time goes on and one original term is split into two, so that the exact effects seen in earlier studies are obscured. 6. Transcription errors: In the process of transcribing data from notebooks to initial study reports to published study summaries to a multi-chemical database, there will be transcription errors. 7. Interpretive issues: When data across multiple studies are collated, contextual information such as incidence, severity, dose, or effect groupings may be lost. Goal: Conduct a case study of these issues for chemically-induced anemia. This poster does not necessarily reflect U.S. EPA policy U.S. EPA ToxRefDB Guideline Acceptable CHR (rat, mouse, dog), SUB (rat) 658 chemicals, 1738 studies 441 studies: anemia-positive Harmonize terminology Anemia ≥ 2 / 3 decrease RBC (red blood cell count) HGB (hemoglobin) HCT(hematocrit) Review original reports for all studies in all chemicals with 2 species anemia-positive Clean database: 49 chemicals Studies mechanistically classified Figure 1: Fraction of tested chemicals showing anemia below doses where body-weight decreases are seen (black, bottom), at doses where body-weight decrease is seen (dark gray, middle), and where anemia is not seen (light gray, top). The numbers indicate the total number of chemicals in the database for each species / study-type combination. After manual review DB class Positive Negative Ambiguous Unknown Before manual review Positive 212 2 4 0 Negative 24 128 11 24 Ambiguous 6 2 16 0 Table 1: Concordance between initial database entry and results of manual review. Each cell contains the number of studies in each class before and after the review. Positive (2 or more of RBC, HGB, HCT), Negative (0), Ambiguous (1), Unknown (manual review found that hematology values were not measured in the study). species / study 1 species / study 2 Reproduced Not Reproduced Fraction reproduced rat SUB rat CHR 18 2 0.90 rat CHR dog CHR 13 2 0.87 rat CHR rat SUB 18 4 0.82 rat SUB rat SUB 16 4 0.80 rat SUB dog CHR 11 4 0.73 mouse CHR rat CHR 11 4 0.73 mouse CHR rat SUB 13 7 0.65 dog CHR rat SUB 11 6 0.65 dog CHR rat CHR 13 8 0.62 rat CHR mouse CHR 11 11 0.50 mouse CHR dog CHR 6 6 0.50 rat SUB mouse CHR 13 14 0.48 dog CHR mouse CHR
poster
Phenotypic Screening Phenotypic screening is a type of screening used to identify substances such as small molecules, peptides, or RNAi that alter the phenotype of a cell or an organism in a desired manner. When the molecular mechanism of action is not assumed and does not require knowledge of the molecular target, phenotypic screening can be applied in biological research and drug discovery. Due to the limited success of target-based drug discovery and the increasing need to minimize the risk of late stage attrition due to poor efficacy or off-target activities, phenotypic approaches attract more attention. Many of today's first-in-class drugs with novel mechanisms of action came from phenotypic screening. The biological relevance of the assay systems is deployed and in this respect the commercial availability of unlimited quantities of pure human cell types, particularly those derived from induced pluripotent stem cells, is having a promising impact. Stem cells are fueling the development of many new disease models and with high levels of translation to human biology and disease. These phenotypic assays are increasingly being used in early toxicity testing. Moreover, high content imaging systems have also powered phenotypic drug discovery, facilitating the rapid analysis of increasingly complex multi-parametric measurements of cellular phenotypes or biomarkers. Compounds are screened in cellular or animal disease models to identify compounds that cause a desirable change in phenotype. Only after the compounds have been discovered to determine the biological targets, this is a process known as target deconvolution. The assay types involve co-cultures of primary cells. The simplest phenotypic screens in vivo employ cell lines and monitor a single parameter such as cellular death or the production of a particular protein. Human primary cells were ranked as the cell type most suited (relevant) for phenotypic screening studies, followed by stem cells or iPS-derived phenotypes and then primary cells of animal origin. Cell line monocultures were the biological systems currently most used to conduct phenotypic drug discovery. This was followed by
poster
Rapid alignment of text to the UMLS and integration into a VIVO repository Stony Brook Medicine Janos G. Hajagos Acknowledgements: The Stony Brook Medicine SUNY Reach team created the infrastructure and generated the data used in this analysis. Members of the team include: Erich Bremer, Barbara Chandler, Moises Eisenberg, Jizu Zhi, Tammy DiPrima, and Andrew White. A prototype disproportion- ate analysis tool was developed by Alex Schultz. This is an extension of work supported through CTSAconnect. CTSA 10-001: 100928SB23 Unified Medical Language System (UMLS) The goal of this research is to extract SUIs (String Unique Identifiers) from text associated with a faculty member‘s VIVO profile. SUIs unlike CUIs (Concept Unique Identifiers) do not have a conceptual meaning. A single SUI can be linked to several concepts. As an example, below the SUI (S0004723) is linked to two different CUIs (C0580209 & C1302752). It is difficult for any automated system to determine with 100% accuracy the contextual meaning of a word or a phrase. The approach taken here is to utilize the sta- tistical patterns in the data to automate the process of tagging large quantities of text. With SUIs there is still a path on the graph to a conceptual meaning. . An extract from MRCONSO showing the relationships between SUIs and CUIs The SUNY REACH site <http://reach.suny.edu/> and the data set Generating RDF for import into VIVO The VIVO 1.5 ontology around research area Identifying what fragments make a researcher unique For the analysis the UMLS release 2012ab was utilized. Stony Brook maintains a SPARQL endpoint: http://link.informatics.stonybro ok.edu/sparql/ A short text alignment service: http://link.informatics.stonybro ok.edu/MeaningLookup/ To align MeSH CUIs and get abstracts and titles from articles two web services (2013 XML): http://link.informatics.stonybro ok.edu/weaver/pubmed2cuis?p mid=21325273 and http://link.informatics.stonybro ok.edu/weaver/pubmed2abstra ct?pmid=10343245 Filter fragments through a UMLS source vocabulary Sample alignment results NCIt is a source vocabulary Breaking text into fragments Natural language processing, abbreviated as NLP, is an active area of research and devel- opment. NLTK is a well supported NLP toolkit for the Python language. There is even a free book: <http://nltk.org/book/>. The following features were utilized: parsing paragraphs into sentences, parsing sen- tences into words, and removing common stop words. Phrases were matched up to the length of fours words. Only a small subset of the features of the NLTK were used in this research. Utilizing NLTK in the iPython shell How well did the raw alignment method work? Excerpt from the VIVO faculty member’s html page: “During the last several years, he has focused his efforts on the prevention of colon cancer using traditional NSAIDs, NO- donating NSAIDs and other pharmacologi- cal agents. He has also pioneered the appli- cation of infrared spectroscopy to biology with emphasis on cancer holding several relevant patents. A new area of work con- cerns the use of nanotechnology in the pre- vention of cancer.” Terms extracted from the article abstract Terms extracted from article titles MeSH from Medline expert tagged articles Comparison for a single VIVO faculty member Number of articles with abs tracts Number of faculty Number of matched fragments Number of s entences 48,893 621,686 521 5,686 Number of articles Number of faculty Number of matched fragments 216,842 683 22,372 Number of MeS H Number of articles Number of faculty 706 21,954 351,273 Counts on analyzed abstracts Counts on analyzed article titles Counts on Medline MeSH Overall top phrases extracted from article abstract Disproportionate or Market Basket Analysis 1. Compute all possible combination of fragments within a sentence that do not overlap 2. Aggregrate the co-occurrence of the two fragments (pair count) 3. Count the number of times each fragment occurs overall (total fragment count) 4. Calculate the
poster
0.001 0.01 0.1 1 10 0 50 100 150 200 Flux Density [Jy/arcsec2] Radius [au] Notsu+2019 Band7 0ï360deg Isella+2018 Band6 0ï360deg 1 10 100 r [AU] 0 0.2 0.4 0.6 0.8 z/r 101 102 103 50K 300K 1000K H2O gas Central Star Grain growth and destructions Grain falling Silicate+ H2O ice Silicate H2O snowline    H2O freeze    H2O evaporation 1 10 100 r [AU] 0 0.2 0.4 0.6 0.8 z/r 10-12 10-11 10-10 10-9 10-8 10-7 10-6 10-5 10-4 nH2O,gas/nH Photo- dissociate H2+OH->H2O+H freeze-out Water and 13C17O lines, and multiple ring and gap structures of the protoplanetary disk around HD 163296 observed by ALMA Shota Notsu1,2 (shota.notsu@riken.jp), Alice S. Booth2,3, Catherine Walsh3, John D. Ilee3, Eiji Akiyama4, Hideko Nomura5, Tomoya Hirota5, Takashi Tsukagoshi5, Mitsuhiko Honda6, T. J. Millar7, Chunhua Qi8 1 Star and Planet Formation Laboratory, RIKEN, Japan 2 Leiden Observatory, Leiden University, The Netherlands 3 University of Leeds, UK 4 Niigata Institute of Technology, Japan 5 NAOJ, Japan 6 Okayama University of Science, Japan 7 Queen’s University Belfast, UK, 8Harvard/CfA, USA Observationally locating the position of the H2O snowline (e.g., Hayashi et al. 1981, 1985) in protoplanetary disks is crucial for understanding the dust evolution and planet formation processes, and the origin of water on the Earth. The velocity profiles of emission lines from disks are usually affected by Doppler shift due to Keplerian rotation. Therefore, the line profiles are sensitive to the radial distribution of the line-emitting regions. In our previous works (Notsu et al. 2016, ApJ, 827, 113; 2017, ApJ, 836, 118; 2018, ApJ, 855, 62), we calculated the chemical composition of the disks around a T Tauri star and a Herbig Ae star using chemical kinetics and various water line profiles. We found that the water lines with small Einstein A coefficients and relatively high upper state energies are dominated by emission from the hot midplane region inside the H2O snowline, and therefore through analyzing their line profiles the position of the H2O snowline can be located. Since the fluxes of these lines from Herbig Ae disks are larger than those from T Tauri disks, the possibility of a successful detection is expected to increase for a Herbig Ae disk. There are several best candidate water lines that trace the position of the H2O snowline within the coverage of ALMA. Recently, we got the upper limit fluxes of submillimeter ortho-H216O 321 GHz, para-H218O 322 GHz, and HDO 335 GHz lines from the disk around the Herbig Ae star HD 163296, using ALMA (Notsu et al. 2019, ApJ, 875, 96). These water lines are considered to be the candidate water lines to locate the position of the H2O snowline, based on our model calculations. We compared the upper limit fluxes with the values calculated by our model calculations with dust emission, and we constrained the line emitting region and the dust opacity from the observations. We also detected multiple ring and gap patterns in the 0.9 mm (ALMA Band 7) dust continuum emission with 15 au spatial resolution, whose positions are consistent with those indicated by other observations (e.g., Isella et al. 2018). Future observations of the submillimeter water lines with longer observation time are required to clarify the position of the H2O snowline in the disk midplane. In addition, we also detected the rarest stable CO isotopologue, 13C17O, in a disk for the first time (Booth et al. 2019, ApJL, 882, L31). We compared our observation with the existing detections of other CO isotoplogues in the HD163296 disk. We found that this line is optically thin within the CO snowline and will be thus a robust tracer of the bulk disk CO gas mass. We showed that this disk will be 2-6 times more massive than previously estimates. Aul Eup nu Elow nl Aul : Einstein A coefficient [s-1] Eup: Energy in upper state [K] r z obs. s i nu & nl:LTE distribution (snowline tracer: OK) Φ(ν):emission profile (Kepler & thermal velocity) τul:optical depth (H2O
poster
em530 channel (GFP) detected protein/TF, bound by anti-GFP background subtraction Integrating Gene Synthesis and Microfluidics for MC Blackburn, E Petrova, SJ Maerkl Institute of Bioengineering, School of Engineering, EPFL, Lausanne, Switzerland Literature 1) de Raad M, et al. A solid-phase platform for combinatorial and scarless multipart gene assembly. ACS Syn Bio 2013. 2) Lunqvist M, et al. Solid-phase cloning for hight- throughput assmebly of single and multiple DNA parts. NAR Vol 43(7), 2015. 3) Wolfe SA, et al. DNA Recognition by Cys2His2 zinc finger proteins. Annu. Rev. Biophys. Biomol. Struct., 1999. 4) Beerli RR, et al. Engineering polydactyl sinc-finger transcription factors. Nature Biotech, Vol 20, 2002. 5) Khalil AS, et al. A synthetic biology framework for programming eukaryotic transcription functions. Cell Vol 150, 2012. 6) Persikov AV, et al. Deep sequencing of large library selections allows computational discovery of diverse sets of zinc fingers that bind common targets. NAR Vol 42 (3), 2014. 7) Persikov AV, et al. A systematic survey of the Cys2His2 zinc finger DNA-binding landscape. NAR, Vol 43(3), 2015. 8) Najafabadi H, et al. C2H2 zinc finger proteins greatly expand the human regulatory lexicon. Nature Biotech, 2015. 9) Fordyce PM, et al. De novo identification and biophysical characterization of transcription-factor binding sites with microfluidic affinity analysis. Nature Biotech Vol 28 (9), 2010. 10) Maerkl SJ, et al. A systems approach to measuring the binding energy landscapes of transcription factors. Science Vol 315, 2007. Introduction Projects ZF Modular Combinatorics ZF Specificity Engineering Methods Objective Engineering synthetic proteins with desired functions is difficult. It can require multiple rounds of computational design, gene construction, expression and functional characterization to produce a new protein which consistently behaves in a novel, expected manner. Traditionally, these time and resource-consuming cycles involve gene synthesis via ligation or PCR-based assembly of DNA parts, cloning, and sequencing to ensure the gene coding for a given engineered protein is error-free1. This is followed by an assay to determine if the protein folds correctly and how effective the protein is at performing its intended function. A ligation/plasmid/cloning-free gene synthesis technique (detailed below) for rapidly constructing genes from synthetic DNA parts was coupled with a 1024- chamber MITOMI device for in vitro characterization of engineered zinc finger transcription factors. DNA parts were designed such that single finger modules could be interchanged with others, enabling the combinatorial construction of many different proteins from a small DNA pool. Engineering the binding specificity of individual zinc finger arrays through controlled amino acid substitutions was examined as a proof-of-concept for protein design. Solution Conceive of and build a platform for rapidly building genes from short, synthetic DNA oligomers (under 90 nt), examine the error rate of synthesis, and generate many ‘prototypes’ of proteins for a high-throughput functional screen. Apply this technique to several rounds of design > synthesis > expression > characterization to evolve a protein for a desired application. Zinc finger modules from zif268 and previously characterized5 arrays 37-12, 92-1 and 158-2 were used to generate novel zinc finger arrays. All fingers in each position were interchanged and tested against a target library containing the expected consensus were each ZF behaving independently of the others in the same array. Zif268 (wt) N C finger 1 finger 2 finger 3 RSDELTR RSDHLTT RSDERKR C N RKREDSR TTLHDSR RTLEDSR 5'- aGCG TGG GCGt 6 5 4 3 2 1 -1 6 5 4 3 2 1 -1 6 5 4 3 2 1 -1 37-12 N C finger 1 finger 2 finger 3 RNFILQR DRANLRR RHDQLTR C N RTLQDHR RRLNARD RQLIFNR 5'- tGAG GAC GTGt 6 5 4 3 2 1 -1 6 5 4 3 2 1 -1 6 5 4 3 2 1 -1 158-2 N C finger 1 finger 2 finger 3 DKTKLRV VRHNLTR QSTSLQR C N RQLSTSQ RTLNHRV VRLKTKD 5'- tGTA GAT
poster
CEDAR CENTER FOR EXPANDED DATA ANNOTATION AND RETRIEVAL SAP – a CEDAR-based pipeline for semantic annotation of biomedical metadata Ravi D. Shankar, Marcos Martinez-Romero, Martin J. O’ Connor, John Graybeal, Purvesh Khatri, Mark A. Musen Stanford University, Stanford, CA 94305, U.S.A. metadatacenter.org CEDAR is supported by grant U54 AI117925 awarded by the National Institute of Allergy and Infectious Diseases through funds provided by the trans-NIH Big Data to Knowledge (BD2K) initiative (www.bd2k.nih.gov) 11/2016 Under the umbrella of CEDAR metadata enrichment effort, we are building the Semantic Annotation Pipeline (SAP) that automates the semantic annotation of biomedical data stored in community data repositories. The pipeline has two major segments: 1) transform the metadata stored in the community data repository to CEDAR JSON-LD format, and 2) add semantic annotations to the CEDAR-formatted metadata. Annotated Metadata in CEDAR Repository Metadata in CEDAR Repository Metadata in Community Repository CEDAR Community Template NCBO BioPortal Ontologies MESH LOINC NCIT RXNORM CPT ICD10 PATO SNOMEDCT MEDDRA NCBITAXON DOID UO At the core of the Annotator is the Apache UIMA ConceptMapper, a named- entity recognition tool. The Annotator annotates metadata with ontology terms as specified in the template. The ontology terms are served by the NCBO BioPortal. The Ingestor transforms the metadata in the community repository to instances of a JSON- encoded CEDAR community template. The CEDAR Community Template encapsulates the metadata specific to that community. The template includes ontological information on value restrictions on specific metadata elements. Ingestor Annotator }, "sampleID": { "@value": "GSM1169960" }, "description": { "@value": "IPF" }, "tissue": { "@type": "http://ncicb.nci.nih.gov/ xml/owl/EVS/Thesaurus.owl#C12801", “@value": "Lung" }, "disease": { "@type": "http://ncicb.nci.nih.gov/ xml/owl/EVS/Thesaurus.owl#C7057", “@value": null }, ^SAMPLE = GSM1169960 !Sample_title = IPF23lung !Sample_type = RNA !Sample_source_name_ch1 = IPF !Sample_organism_ch1 = Homo sapiens !Sample_characteristics_ch1 = tissue: Lung !Sample_molecule_ch1 = total RNA !Sample_label_ch1 = biotin !Sample_description = IPF !Sample_label_protocol_ch1 = Biotinylated cRNA were prepared with the Ambion MessageAmp kit for Illumina arrays Ingestor Annotator An Illustration of SAP: Semantic Annotation of GEO Sample Metadata CEDAR GEO Sample Template GEO Sample Metadata in GEO Repository GEO Sample Metadata in CEDAR JSON-LD format Value Restrictions in CEDAR GEO Sample Template Semantically Annotated GEO Sample Metadata in CEDAR JSON-LD format }, "sampleID": { "@value": "GSM1169960" }, "description": { "@value": "IPF" }, "tissue": { "@type": "http://ncicb.nci.nih.gov/xml/owl/EVS/Thesaurus.owl#C12801", "_valueLabel": "Lung Tissue", "@id": "http://ncicb.nci.nih.gov/xml/owl/EVS/Thesaurus.owl#C33024" }, "disease": { "@type": "http://ncicb.nci.nih.gov/xml/owl/EVS/Thesaurus.owl#C7057", "_valueLabel": "Idiopathic Pulmonary Fibrosis", "@id": "http://ncicb.nci.nih.gov/xml/owl/EVS/Thesaurus.owl#C35716" },
poster
RESEARCH POSTER PRESENTATION DESIGN © 2015 www.PosterPresentations.com Bell's palsy is a rapid unilateral facial nerve paresis (weakness) or paralysis (complete loss of movement) of unknown cause. The condition leads to the partial or complete inability to voluntarily move facial muscles on the affected side of the face. Although typically self-limited, the facial paresis/paralysis that occurs in Bell's palsy may cause significant temporary oral incompetence and an inability to close the eyelid, leading to potential eye injury. Additional long-term poor outcomes do occur and can be devastating to the patient.1 Immune, infective and ischaemic mechanisms are all potential contributors to the development of Bell's palsy, but the precise cause remains unclear.2 The aims of treatment in the acute phase of Bell's palsy include strategies to speed recovery and to prevent corneal complications. Eye care includes eye patching and lubrication. Strategies to speed recovery include physical therapy, corticosteroids and antiviral agents.3 Inactivated vaccines against SARS-CoV-2 are being developed by several vaccine manufacturers. The inactivated COVID-19 vaccine Sinovac CoronaVac was developed by Sinovac. CoronaVac has been authorized as a 2-dose vaccine for individuals aged 18 years and older. Systemic adverse effects include: fatigue, fever, myalgia, diarrhoea, headache, cough, nausea, and vomiting. Local side effects include: pain, pruritus, swelling, redness, and induration.4 INTRODUCTION METHOD REFERENCES 1) Reginald F. Baugh et al. Clinical Practice Guideline: Bell’s Palsy. American Academy of Otolaryngology—Head and Neck Surgery Foundation. 2013; 149(3S) Pg1. 2) Eviston TJ, et al. Bell's palsy: aetiology, clinical features and multidisciplinary care J Neurol Neurosurg Psychiatry. 2015 86: Pg1356-1361. 3) J. M. K. Murthy et al. Bell's palsy: Treatment guidelines. Ann Indian Acad Neurol. 2011 Jul; 14(Suppl1): S70–S72. 4) World Health Organisation. Background document on the inactivated vaccine Sinovac-CoronaVac against COVID-19. 2021: Pg 21. 5) Bin-Yan Yu et al. Bell’s palsy after inactivated COVID-19 vaccination in a patient with history of recurrent Bell’s palsy: A case report. World J Clin Cases. 2021 Sep 26; 9(27): Pg8274–8279. The above image to the left shows loss of nasolabial fold and drooping of the left mouth corner. The image to the right was taken after full recovery of Bell’s Palsy. Ophthalmology Department, Hospital Kuala Lumpur Thivakar S (MD), Fatin S (MS Opthal) Bell’s Palsy Post Administration of COVID-19 Inactivated Vaccine DISCUSSION CONCLUSION To report a case of Bell’s Palsy occurring post vaccination. AIM CASE REPORT Case report. A 23 years old Malay lady presented with left sided facial weakness over the past 2 weeks associated with occasional tearing over the left eye. She had no comorbids. Upon further questioning, it was noted that patient was administered her first dose of COVID- 19 vaccination 3 weeks ago, which was 1 week prior to her presentation to the Ophthalmology clinic. Upon general examination of the patient, it was noted that there was lagophthalmos over the left eye with poor Bell’s phenomenon. There was also loss of nasolabial fold over the left side of her face. Loss of frontal wrinkling as well as drooping of the left mouth corner was also noted. However there were no other cranial nerve involvement. Patient had no peripheral neurological deficits. Upon ocular examination, it was noted that right eye vision was 6/9 and her left eye vision was 6/24. No RAPD. Anterior segment of her right eye was unremarkable. Left eye anterior segment examination shows minimal superficial punctate keratopathy inferiorly at the cornea. Both eyes’ intraocular pressure was normal. Dilated fundus examination of both eye was unremarkable. As patient was referred from the ENT Department for the above complaints, patient was already started on Oral Mecobalamin and a tapering dose of Oral Prednisolone to be taken over
poster
(http://dh2016.adho.org) DH Home (http://www.dh2016.adho.org) / Abstracts (/abstracts/) / 86 (/abstracts/86) Title: EVI-LINHD. A Virtual Research Environment for the Spanish-speaking Community Authors: Gimena del Rio Riande, Elena González-Blanco García, Clara Martínez Cantón, Juan José Escribano Category: Paper:Poster Keywords: Virtual Research Environment, Virtual Research Community, Digital Scholarly Edition, Spanish-speaking Community, DH Center Although Digital Humanities have been defined from a discipline perspective in many ways, it is surely a field still looking for its own objects, practices and methodologies. Their development in the Spanish-speaking countries is no exception to this process and, even it is complex to trace a unique genealogy to give account for the evolving field in Spain and Latin America (Gonzalez-Blanco, 2013; Spence and Gonzalez-Blanco, 2014; Rio Riande 2014a, 2014b), the emergence of various associations in Mexico (RedDH), Spain (HDH) and Argentina (AAHD) that seek for a constant dialogue (Galina, González-Blanco and Rio Riande, 2015), and academic lab and DH center initiatives such as LINHD (Spain and Argentina), GRINUGR (Spain), Medialab USAL, LABTEC (Argentina), TadeoLab (Colombia), Elabora HD (Mexico), among others, make it clear that research has become increasingly “global, multipolar and networked” (Llewellyn Smith, et al., 2011) and that the academic field is looking for a global outreach and aims to open spaces of shared virtual work. Virtual Research Communities (VRCs) are a consequence of these changes. Virtual Research Environments (VREs) have become central objects for digital humanist community, as they help global, interdisciplinary and networked research taking of profit of the changes in “data production, curation and (re‐)use, by new scientific methods, by changes in technology supply” (Voss and Procter, 2009: 174-90). DH Centers, labs or less formal structures such as associations benefit from many kind of VREs, as they facilitate researchers and users a place to develop, store, share and preserve their work, making it more visible. The focus and implementation of each of these VREs is different, as Carusi and Reimer (2010) show in their comparative analysis, but there are some common guidelines, philosophy and standards that are generally shared (as an example, see the Centernet map and guidelines of TGIR Huma-Num, 2015). This poster presents the structure and design of the VRE of LINHD, the Digital Innovation Lab at UNED ( http://linhd.uned.es (http://linhd.uned.es)), and the first Digital Humanities Center in Spain. This VRE focuses on the possibilities of a collaborative environment for (profane or advanced) Spanish-speakers scholarly digital editors. Taking into account the language barrier that English may suppose for a Spanish-speakers scholar or student and the distance they may encounter with the data and organization of the interface (in terms of computational knowledge) while facing a scholarly digital edition or collection, LINHD’s VRE comes as a solution for the VRC interested in scholarly digital work. Moreover, it will make it possible to add an apply tools that contribute to improve Spanish-English applications or tools developed locally, such as Contawords, by Iula-UPF http://contawords.iula.upf.edu/executions (http://contawords.iula.upf.edu/executions). Opening such an environment to the Spanish speaking world will make it possible to reach different kinds of communities, whose profile and training in digital humanities differ from the typical users of DH tools and environment. Testing all these tools in this new environment will, for sure, draw interesting project results. In this sense, our project dialogues and aims to join the landscape of other VREs devoted to digital edition, such as Textgrid, e-laborate, etc. and, in a further stage, to build a complete virtual environment to collect and classify data, tools and projects, work and publish them and s
poster
Task/Result Queues ML-based drug screening Real-time crystallography You request a function be executed on endpoint A and B Federated Function as a Service ●Modern computing environments are distributed and heterogeneous; modern workloads require specialized hardware, rapid responses, and remote processing (e.g., near data) ●FaaS provides an intuitive interface for users to register and invoke programming functions without regard for underlying infrastructure ●Federated FaaS enables functions to be dispatched to remote endpoints chosen for data locality, security, or other concerns ●funcX allows users to execute Python functions (which may invoke executables, MPI programs, etc.) on arbitrary resources (e.g., CPUs, GPUs) from short to long run times from funcx.sdk.client import FuncXClient fxc = FuncXClient() Instantiate a funcX client Register Python function with input args Execute a function by specifying endpoint and input arguments Fire-and-forget execution Outsource the challenging aspects of remote execution funcX manages authentication, serialization of functions and data, reliable execution optionally in containers, and delivery of results back to requesting users Transform resources into FaaS endpoints Easily manage execution across distributed resources The funcX endpoint software can be deployed on laptops, clouds, clusters, and supercomputers. It provisions resources elastically based on workload. High performance Launch millions of tasks funcX supports batch submission and monitoring, asynchronous callbacks, container warming, automated resource scaling, fault tolerance, prefetching, and memoization Portable serverless computing to enable scalable data science def hello_world(): return "Hello World!" func_uuid = fxc.register_function(hello_world) ep_id = '4b116d3c-1703-4f8f-9f6f-39921e5864df' result = fxc.run(endpoint_id=ep_id, function_id=func_uuid) fxc.get_result(result) Retrieve results (and exceptions) asynchronously $ pip install funcx_endpoint $ funcx-endpoint configure $ funcx-endpoint start <ENDPOINT_NAME> Install and configure a funcX endpoint def compute(args): # do something return results (1) Registration (function + container) F(ep1,1) F(ep1, 2) F(ep1, 3) F(ep1, 4) F(ep1, 5) F(ep1, 6) (2) Execution (function, endpoint, args) A A 1 3 2 funcX manages the reliable and secure execution on those endpoints funcX notifies you when the function is complete funcX development has been supported by NSF 2004894/2004932 and an Argonne Lab Directed Research and Development award Try on Binder https://funcx.org/binder HEP fitting as a service Application examples Functions Endpoints Yadu Babuji*, Josh Bryan*⚑, Kyle Chard*⚑, Ryan Chard*⚑, Ben Clifford*, Ian Foster*⚑, Ben Galewsky°, Daniel S. Katz°, Kevin Hunter Kesling*, Zhuozhao Li*, Kirill Nagairtsev*, Stephen Rosen*, Tyler Skluzacek* *University of Chicago & Argonne National Laboratory; ⚑Globus; °National Center for Supercomputing Applications, University of Illinois at Urbana-Champaign
poster
<Fußzeile> 1 Das NFDI-Konsortium Text+ wird gefördert durch die Deutsche Forschungsgemeinschaft (DFG) - Projektnummer 460033370 www.text-plus.org Ressourcen Corpus Reader Natural Language Processing Pipeline MINE Modes of Narration and Attribution Event Tagger (Vauth, u. a. 2021) Comment Tagger (Weimer u. a., im Erscheinen) Generalization Tagger (Gödeke, im Erscheinen) Entity Linker (Barth u. a., 2022) Attribution Tagger (Dönicke u. a., 2022) Sprecher/Rede, Koreferenzresolution Speech Tagger (Brunner u. a. 2022) Speaker Extractor (Dönicke, u. a. 2022) Coref (Krug, u. a.; Dönicke, u. a. 2022) Semantische Analyse Temponym Tagger (Strötgen und Gertz, 2010 / 2015) Emotions Tagger (Mohammad und Turney, 2010; Dönicke, u. a. 2022) Morphosyntaktische Analyse Sentencizer (spaCy / NLTK) Dependency Parser (spaCy) Clausizer (Dönicke, 2020) Analyzer (Altinok, 2018) TenseTagger (Dönicke, 2020) Preprocessing Tokenizer (spaCy) Tagger (spaCy) Lemmatizer (spaCy) NER (spaCy) Normalizer (Dönicke, u. a. 2022) NLP-Pipeline MONAPipe (Dönicke u. a. 2022): • entwickelt im Projekt Modes of Narration and Attribution (MONA) innerhalb des Göttingen Centre for Digital Humanities (GCDH) • basierend auf spaCy 2 (Honnibal und Montani, 2017) • Standardkomponenten und Custom-Komponenten (Re- Implementierungen / Wrapper bestehender Systeme sowie Eigenentwicklungen) Anreicherung textinterner Strukturinformationen, u. a.: • linguistische Strukturen (Zeitformen, Zeitausdrücke, Sprachnormalisierung, Satz-/Teilsatzsplitting) • Named Entities (inkl. Koreferenzen sowie Entity Linking) • narrative Phänomene (Redewiedergabe, Sprechererkennung und -attribution, Ereignisse, Reflexionen) Florian Barth (SUB Göttingen / Göttingen Centre for Digital Humanities), José Calvo Tello (SUB Göttingen), Stefan Funk (SUB Göttingen), Mathias Göbel (SUB Göttingen), Daniel Kurzawe (SUB Göttingen), Nanette Rißler-Pipka (Gesellschaft für wissenschaftliche Datenverarbeitung mbH Göttingen), Ubbo Veentjer (SUB Göttingen) Sammlungen an der SUB Göttingen • das TextGrid Repository beinhaltet u. a. literarische Korpora (Digitale Bibliothek), digitale Editionen (Architrave, Bibliothek der Neologie) oder Digitalisate (Virtuelles Skriptorium St. Matthias) • DARIAH-DE Repository und Portal DiscussData als Plattformen für geisteswissenschaftliche Forschungsdaten • DigiZeitschriften, das deutsche digitale Zeitschriftenarchiv, umfasst wissenschaftliche Zeitschriftenbestände aus dem geisteswissenschaftlichen Bereich • im Verzeichnis deutschsprachiger Drucke des 17. sowie des 18. Jahrhunderts (VD17, VD18) werden vom Göttinger Digitalisierungszentrum und nationalen Partnern alle Druckwerke der entsprechenden Jahrhunderte digitalisiert VD17 VD18 Unstrukturierter Text Unterschiedlicher Stand der Digitalisierung und textinternen Anreicherung in SUB-Ressourcen: • Texte mit umfassenden Strukturinformationen • Plain-Texte mit schwankender OCR-Qualität • als Bilddateien digitalisierte Faksimiles (VD17, VD18) Ziele • Erzeugung qualitativ hochwertiger Plain-Texte für Ressourcen mit Bilddateien in Zusammenarbeit mit dem OCR-D-Projekt • Text-Mining-Verfahren zur Anreicherung mit textinterner Struktur • Bereitstellung abgeleiteter Textformate (Schöch u. a. 2020) Import • Ingest kompletter Sammlungen (Text und Metadaten) • Support für spezifische Daten- und Metadatenformate, z. B. TEI-XML- Varianten von Community- relevanten Korpora Anreicherung • Data Reconciliation: Überprüfung, Vervollständigung und Erweiterung der Metadaten • Enrichment: Verknüpfung der Metadaten mit Normdaten und Knowledge Bases Export • Serialisierung angereicherter Metadaten in gewünschtes Format (Eingangsformat oder Neuformat) KOLIMO VIAF • Data Warehouse normalisierte Einbindung distinkter Ressourcen • Elastic Search Repository-übergreifende Suche für Text und Metadaten • Knowledge Graph (neo4j) Ressourcen-übergreifende Verknüpfung von Informationen auf Dokumentebene und textinternen Anreicherungen Task Area Collections Text-Mining-Pipelines fü
poster
Field plots were established in Knoxville, TN (latitude 35.971438, longitude -83.854094) on the East TN Research and Education Center, Holston River Unit. Differing levels of weed populations were created in a 4.5 ha field by herbicide application in year 1. Primary weeds were Xanthium strumarium and Urochloa platyphylla Variable rate applications (VRAs) of atrazine preemergence (PRE) followed by dicamba postemergence (POST) were investigated for the reduction of herbicide inputs and their resulting impact on weed control and corn yield. VRAs of atrazine were on the basis of weed density data collected in Year 1. VRAs of dicamba (Banvel) were based upon common cocklebur density evaluations within the field. Technical support provided by John B Wilkerson and William E Hart was appreciated. This research was conducted on research farms associated with the Tennessee Agricultural Experiment Station, a division of the University of Tennessee Institute of Agriculture. 1.Compared with conventional applications, atrazine usage was decreased by 43 and 32% in the site-specific application treatments in Year 2 and Year 3, respectively. 2.VRAs of dicamba reduced herbicide inputs by greater than 45% for Year 2 and Year 3. 3.Corn yields were similar for the conventional and site- specific treatments in both years. This research compared site- specific weed control using variable rate herbicide dosing to broadcast applications under field conditions in a maize production field in Tennessee, USA. On the basis of these data, site- specific herbicide applications have the greatest potential and least risk for managing weeds when POST or PRE + POST variable rate herbicide applications are used. Yields in site-specific plots were equal to using broadcast sprays. Site-specific weed management can increase crop production efficiency by minimizing herbicide input costs without compromising crop yields. A reduction in herbicide inputs resulting from site-specific weed management may also decrease the probability of nonpoint pollution compared with conventional herbicide applications.
poster
Plasticrusts and pyroplastic: two novel plastic debris types detected in Giglio island, Italy Julius A. Ellrich1 & Sonja M. Ehlers2,3 1Julius_Ellrich@web.de; Independent Researcher, Sankt-Josef-Straße 25, 56068 Koblenz, Germany 2Ehlers@bafg.de; Department of Animal Ecology, Federal Institute of Hydrology, Am Mainzer Tor 1, 56068 Koblenz, Germany 3Institute for Integrated Natural Sciences, University of Koblenz-Landau, 56070 Koblenz, Germany www.juliusaellrich.weebly.com, www.sonjamehlers.weebly.com REFERENCES: Ehlers SM & Ellrich JA 2020. First record of plasticrusts and pyroplastic from the Mediteranean Sea. Marine Pollution Bulletin 151: 110845. Doi: 10.1016/jmarpolbul. 2019.110845 INTRODUCTION: Fig.1A)LocationofGiglioislandintheTyrrhenianSea.B)Locationsofthefivesurveyed habitats (H1-H5) in Giglio. H2 and H3 were adjacent habitats and, thus, depicted by a singledot.C)Thewave-exposedrockyshore(H1).D)Thewave-shelteredbeach(H2). METHODS: Inthefield,wemeasuredplasticrustdensity,areaandpercentcoverusing quadrats(10 cm x 10 cm). At the lab, we determined plasticrust thickness using a digital microscope and pyroplastic size using digital calipers. Furthermore, we identified the plasticrust and pyroplastic materials (i.e., polymer types) with Fourier-transform infrared spectroscopy (FTIR, Fig. 2). We conducted our FTIR measurements in attenuated total reflectance (ATR) mode in a wavenumber range between 4000 and 370 cm−1with 8 co- added scans and a spectral resolution of 4 cm−1. Then, we compared the obtained spectrawiththeBrukerspectrallibraryinOpus7.5software. Plasticrusts and pyroplastic are novel plastic debris types that have only recently been reported for the first time from Madeira island (Atlantic Ocean, Gestoso et al. 2019) and the southern United Kingdom, respectively (Turner et al. 2019). While plasticrusts result from plastic debris being wave-swept across rugose rocks, pyroplastics derive from burnt plastic waste. During field surveys in Giglio island (Tyrrhenian Sea, Italy), we detected plasticrusts on a wave-exposed rocky shore and pyroplasticonawave-shelteredsandybeach(Ehlers&Ellrich2020,Fig.1). Fig. 2 Sonja M. Ehlers identifying polymer types using FTIR (Vertex 70, Bruker, Ettlingen, Germany) at the Federal Institute of Hydrology in Koblenz, Germany. C D Plasticrust (Fig. 3A) density was 3.25 ± 1.65 plasticrusts/dm2 (mean ± SE; n= 4 quadrats). Plasticrust area was 0.46 ± 0.08 mm2 (n= 13 plasticrusts). Plasticrust cover was 0.02 ± 0.01 % (n= 4 quadrats). Plasticrust thickness ranged between 0.5 and 0.7 mm. FTIR analyses revealed the plasticrust material as polyethylene (PE, Fig. 3B). We didnotdetectanyplasticrustsonthewave-shelteredrockyshores(H3-H5,Fig.1B). ACKNOWLEDGEMENTS: We thank Professor Dr Jochen H. E. Koop for letting us use the FTIR spectrometer at the Federal Institute of Hydrology in Koblenz, Germany. RESULTS: Fig. 3 A) The detected blue plasticrusts consisted of (B) PE material. Fig. 4 A) The detected pyroplastic with blue inclusions. B) Pyroplastic close-up. C) The FTIR spectrum that revealed PET as the pyroplastic material. D) PET bottle found on the beach (H2, Fig. 1B). E) The FTIR spectrum that confirmed PET as the bottle material. The pyroplastic had a stone-like appearance (Fig. 4A) with blue inclusions (Fig. 4B). Pyroplastic size was approximately 2 cm × 1.4 cm × 0.5 cm (length x width x height). FTIR analyses showed that the pyroplastic material was polyethylene terephthalate (PET, Fig. 4C). On the beach, we detected burnt charcoal near the pyroplastic. Finally, wefoundseveralPETbeveragebottles(Fig.4D, E)ineachhabitat(H1-H5,Fig.1B). DISCUSSION: A B Absorbance units Wavenumber (cm-1) FTIR measurement spectrum Bruker library spectrum A B D E Wavenumber (cm-1) C Absorbance units Absorbance units Wavenumber (cm-1) FTIR measurement spectrum Bruker library spectrum FTIR measurement spectrum Bruker library spectrum Our results from Giglio resemble findings from Madeira and the United Kingdom indicating
poster
日本語における対のある自他動詞の習得 ―中国人学習者の場合― 王冠亮(南山大学大学院生)・坂本正(南山大学) 従来の対のある自他動詞に関す る習得研究の結論の一つは、学習 者にとって自動詞の習得が難しく、 特に結果状態が注目される自動詞 文の習得が難しいということである (守屋1994;小林・直井1996;小林 1996;王2012;伊藤2012; )。 はじめに 先行研究の問題点 1 結果状態が注目される自動詞 文に関する問題項目が少ない。 2 パーセンテージ(%)が主な手 法で、結果は一般化しにくい。信頼性がどれぐらいある?母集団については? 本研究の研究課題 結果状態が注目される自動詞文 の習得は果たして難しいか 本研究の研究仮説先行研究の結果を検証しなおす とともに、三つの仮説を立てる H1 動作主特定・非特定は自他 動詞の選択に影響がある。 H2 文の肯定・否定は学習者の 自他動詞の選択に影響があ る。 H3 自他の3つの派生パターン は自他動詞の選択に影響が ある。 H1を立てる理由 守屋(1994)と杉村(2013)などの 研究において,対のある自他動詞 の習得難易度には,「動作主の特 定・不特定」が一つの目安、要因と して取り上げられている。 H2を立てる理由 部分の研究は対のある自動詞で 成り立つ「結果(無標)可能表現」が 学習者に難しいと主張している。 しかし、これらの研究で使われる 調査問題は100%否定文である。 H3を立てる理由 一般的に日本語の自他交替を自 動化・他動化・両極化が取り上げら れてきたが,自他交替に三つのパ ターンがあることが学習者の習得 に影響を及ぼすのか。これ、本当?肯定文、やさしい?検証する! 調査 調査の実施 調査期間:2016年4月25日—29日・ 9月26日—10月07日 調査場所:中国安徽省にあるS大学 とN大学,並びに,日本南山大学 調査対象: S大学三年生(S3)26人 N大学二年生(N2)31人 N大学三年生(N3)35人,計92人 (回収有効調査票は85部) 日本人大学生(JP)30人 調査の設計 1 学習者の既習語彙を使用 2 研究課題と仮説を中心に24問を 用意 3 五つの選択肢のある多肢選択問 題を使う 4 文脈情報を学習者の母語で提供 グループ 平均正答率 mean 標準偏差 sd S3 62.86 19.74 N2 56.95 20.94 N3 52.34 16.53 JP 94.86 10.36 調査の結果(記述統計量) 各組の正答率の箱ひげ図 結論 研究課題 Kruskal-Wallis検定で4群を比べる。 H(3)=62.04,η2=.54,p=2.152e-13 四群に有意差があり、 続いて多重比較 学習者の間に有意差がない。 JP-S3: r=.74, p=4.0e-08 JP-N2: r=.73, p=7.7e-09 JP-N3: r=.82, p=6.5e-11 結果状態が注目される自動詞文が難しい! 研究仮説1 N2:t(29)=0.57, d=.10 ,p=.57 N3: V=212.5, r=.03 , p=.84→ノンパラ S3: t(22)=0.32, d=.07 , p=.75 三群いずれも有意差が見られな かったので、動作主特定・不特定が 学習者に影響を与えるとは言えない! 研究仮説2 N2:t(29)=.84,d=.16, p=.40 N3: t(31)=.90,d=.16,p=.37 S3: V=130,r=.01,p=.92→ノンパラ 三群いずれも有意差が見られな かったので、文の肯定・否定が学習 者に影響を与えるとは言えない! 研究仮説3 N2:F(2,87)=0.285,MS=178.8 η2=.01 ,p=.75, N3: χ2(2)=0.39,p=.82→ノンパラ S3: χ2(2)=4.963,p=.08→ノンパラ 三群いずれも有意差が見られな かったので、派生パターンが学習者 に影響を与えるとは言えない! 第27回第二言語習得研究会全国大会 2016年12月18日於九州大学 参考文献(予稿集にない部分) (1)伊藤秀明(2012b)「相対自動詞・他動詞選択判断の要因―中国人大学生の場合―」『国際交流基金日本語教 育紀要』8,7-21.独立行政法人国際交流基金. (2)大友賢二・中村洋一(2002)『テストで言語能力は測れるか-言語テストデータ分析入門-』桐原書店 (3)奥津敬一郎(1967)「自動化・他動化および両極化転形―自・他動詞の対応―」須賀一好・早津恵美子[編] (1995)『動詞の自他』ひつじ書房 (4)王冠华(2012)「中国日语学习者有对自动词习得状况的实证研究」『日语学习与研究』5,86-94,对外 经济贸易大学. (5)王恰韓(2012)「中国人学習者における日本語無標可能表現の習得に関する研究: この役はあの新人俳優 にはつとまらない」『日本語研究』32,1-14,首都大学東京・東京都立大学日本語・日本語教育研究会. (6)関承(2014)『中国語母語話者における日本語自・他動詞の習得研究―中国語で可能標識が使われる表現を 中心に―』広島大学未公刊博士学位論文 (7)小林典子(1996)「相対自動詞による結果・状態表現―日本語学習者の習得状況― 」 『文芸言語研究・言語篇』29, 41-56,筑波大学. (8)封小芹(2007)「可能の意味を含む有対自動詞の受容能力の習得について- 中国人日 本語学習者の場合- 」『ククロス:国際コミュニケーション論集』4,49-63,名古屋大学. (9)張威(1992)「『可能表現の本質』考:無標の可能表現へのアプローチ」『中京大学教養 論叢』32(4),1351-1373, 中京大学.
poster
Shifting Terms and Concepts: From Defence to (Human) Security Introduction Over the last two decades, the concepts of defence and security have overlapped to the point where the former is merged with the latter. Since jihadist terrorism broke out at the dawn of the millennium, governments have been forced to review the classic paradigm according to which the military is employed in overseas operations—or the defence of the homeland from external enemies—and police and law enforcement agencies are tasked with internal security. In such a context, the military has taken on an increasing role in national security matters, although security itself is an umbrella concept under development that currently includes such cross- cutting topics as terrorism, cyber threats, health, food, energy, the economy, poverty, climate change, information technology, social security, job security, just to mention a few. The inclination to replace the idea of defence with security has contributed to the expansion of the idea of security itself. This way, defence activities abroad, such as military assistance to Ukraine in the context of the ongoing conflict with Russia, are presented to the public as "security" operations. Data analysis The author of this work has checked the multi-sectoral approach to security outlined by the UN and the NATO against commonly accepted versions of human security through a comparative study of the definitions provided by well-established and reputable dictionaries such as Merriam-Webster, Britannica, Collins, Oxford, Cambridge, Longman, and Macmillan. Such dictionaries provide similar definitions of the term "security": safety; safety from attack, harm, or damage; freedom from danger or threat; freedom from fear or anxiety; being safe and free from worry; being protected or safe from harm. Security is the protection from, or resilience against, the potential harm caused by others by restricting one's freedom to act. Conclusions The human security approach broadens the scope of security analysis and policy from the traditional notion of national security to the security of people and their complex social and economic interactions. The term "defence", is given as a synonym of "security" by the majority of these dictionaries (Merriam-Webster, Cambridge, Oxford, Collins), even though Britannica and Macmillan continue to be anchored to the dichotomy of defence/external and security/internal, respectively, by providing the words "national security" and "internal security", is evidence of this trend. The evolving concept of security must be scrutinised, with special attention paid to the notion of human security, boosted by UN General Assembly resolution 66/290 and the NATO Strategic Concept 2022, and currently under investigation by the Exploratory Team of the NATO Science and Technology Organization. So far, the meaning of “human security” remains unclear. References 1. United Nations Development Programme (UNDP), Human Development Report 1994: New Dimensions of Human Security, New York, UN, 1994. 2. UN General Assembly (UNGA), Follow-up to paragraph 143 on human security of the 2005 World Summit Outcome, resolution adopted on 25 Oct. 2012, A/RES/66/290, in GAOR, 66th sess., Suppl. no. 49. 3. NATO Heads of State and Government, Warsaw Summit Communiqué, 9 July 2016, Press Release (2016) 100. 4. NATO Heads of State and Government, London Declaration, 4 dec. 2019, Press Release (2019) 115. 5. NATO Heads of State and Government, Brussels Summit Communiqué, 14 June 2021, Press Release (2021) 086. 6. NATO Heads of State and Government, Madrid Summit Declaration, 29 June 2022, Press Release (2022) 095. 7. NATO Strategic Concept 2022, adopted at the Madrid Summit, 28-29 June 2022. 8. NATO, Human Security: Approach and Guiding Principles, https://www.nato.int/cps/en/natohq/official_texts_208515.htm?s electedLocale=en. 9. Marsili, Marco, The Role of the Armed Forces in Homeland Security, in Book of Abstracts of the XI Portuguese Con
poster
Mitglied im Netzwerk von Contact: CFCI / Müller-Reichert Lab Silke Tulok / Dr. Gunar Fabig Silke Tulok¹, Gunar Fabig¹, Andy Vogelsang¹, Thomas Kugel², Thomas Müller-Reichert¹ ¹TU Dresden, Faculty of Medicine Carl Gustav Carus, Core Facility Cellular Imaging, ²TU Dresden, Faculty of Medicine Carl Gustav Carus, Divisional Management, IT Department ¹TU Dresden, Faculty of Medicine Carl Gustav Carus, Core Facility Cellular Imaging Abstract The FAIR concept (Findability, Accessibility, Interoperability, Reproducibility) of data management is of relevance for all current areas of research. For imaging data obtained from both light and electron microscopy, data management is associated with a number of challenges related to the size, complexity, dimensionality, visualization and quantitative analysis of the acquired data. We have started to use the Python-based and open-source image data management software OMERO. This software package has been introduced by the Open Microscopy Environment developers of the University of Dundee. OMERO is an encouraging option to manage microscopy data. It is installed on a virtual machine and connected to a central data server to manage the storage of images in a multi-user environment. In addition to long-term data storage, it provides possibilities for saving important metadata in an efficient manner, thus avoiding multiple copies of data. It can also handle open source-based processing tools for image analysis, thereby allowing effective image analysis workflows. The Core Facility Cellular Imaging (CFCI) at the Faculty of Medicine Carl Gustav Carus (TU Dresden) is currently running a pilot project for testing the use and handling of the OMERO software. This is done together with interested users of the imaging facility and a research group. Currently, we are pushing forward this pilot study on a small scale without any data steward. Our experiences argue so far for giving data management issues into the hands of dedicated personnel not fully involved in research projects. As funding agencies will ask for higher and higher standards for implementing FAIR- data principles in the future, this will be a relevant topic for the whole research community. We want to introduce a convenient solution, which could be applicable for many users within the DRESDEN-concept research alliance and demonstrate how to establish and manage OMERO in a facility context. 2020 March TiM conference: first introduction to OMERO 2020 October Letter of support for the I3D:bio grant application 2022 February First local meeting with IT to start a pilot project (10 TB server space) 2022 March Support by RDM4mic for setting up the OMERO server 2022 May Cooperation agreement with I3D:bio for support during pilot project 2022 May IT infrastructure (storage & virtual machine & staff) ready to start 2022 July OMERO setup finished, test environment at IT department 2022 September OMERO login (only for members of the Faculty of Medicine) 2022 October Preliminary teaching materials provided by I3D:bio team 2022 November Start uploading of data (light-sheet microscopy, live-cell imaging, slide scanner) 2023 March TiM conference - First hands-on training with I3D:bio team (tagging, data structure) 2023 April OMERO open for all TU members, web access from all over the world 2023 April Start to implement analysis workflow in OMERO environment 2023 August Second hands-on meeting with I3D:bio team (scripting, key value templates) 2023 August Meeting with local IT for future expansion (IT resources, personnel resources) 2023 September Pilot implementation for medical education (parallel data access for >300 students) RDM4mic OMERO Research Data Management for Microscopy Swedlow et al. developed OME-TIFF DFG funds German BioImaging Network Funding for NFDI4BIOIMAGE by DFG Funding for I3D:bio by DFG 2000 2012 2019 2023 2022 - 2024 2023 - 2028 Information Infrastructure for BioImage Data https://www.i3dbio.de TiM Conference First hand-on meeting
poster
2.6 P′ 2.9 1.1 0.1 0.4 2.5 2.3 13 9.4 1.4 0.7 +1.9 1.5 +1.0 0.8 -0.9 P K K′ (x 3.4) (x 3) (x 4) (x 2.4) (x 3.3) (x 4) (x 3.3) (x 4) 7.7 P′ 3.7 3.5 4.2 0.5 3.3 6.1 30.0 24.0 1.4 0.3 +1.8 4.0 +0.3 3.65 -1.5 P K K′ (x 1.3) (x 1.8) (x 1.4) (x 9.5) (x 7.7) (x 3.7) (x 2) Anatomy of energy pathways in the subtropical gyre of the North Atlantic Pauline Tedesco1, Jonathan Gula2,3, Ali Mashayek1 1University of Cambridge, Earth Sciences, United Kingdom, 2University of Western Brittany, Laboratory of Ocean Physics and Satellite remote sensing (LOPS), France; 3Institut Universitaire de France, France pfmt2@cam.ac.uk Our energy analysis uses the concept of energy separation in a time-mean ( ) and time-varying ( ) component: with the horizontal velocity components, the reference density, a 4-year average, and deviation from this average. The steady-state balance of and reservoirs is analyzed with a regional formulation of the Lorenz Energy Cycle (Lorenz, 1955): K K′ K = 1 2 ρ0(u2 + v2) and K′= 1 2 ρ0(u′2 + v′2) (u, v) ρ0 ( . ) ( . )′ K K′ 2. Theory P′ P K K′ Swind(K) Sb. drag(K) Ssubgrid diss(K) T(K) −C(K, P) C(K′ , P′ ) Ssubgrid diss(K′ ) Sb. drag(K′ ) Swind(K′ ) T(K′ ) −TV(K, K′ ) −TH(K, K′ ) −CV(K, K′ ) −CH(K, K′ ) CH(K′ , K) +CV(K′ , K) Energy rates of change related to: • : wind stress, bottom drag, subgrid dissipation. • : non-local effects being spatial redistribution due to pressure work and advection. • : conversion between energy reservoirs. • : local and non-local eddy-mean energy conversions. • : energy conversions related to buoyancy flux. S( . ) T( . ) C( . , . ) −C(K, K′); C(K′, K) C(K, P); C(K′, P′) Scheme of ocean energy pathways as described by the Lorenz Energy Cycle (Lorenz, 1955). Dashed arrows denote non-local effects. Terms are defined positive in the direction of the arrow. Adapted from Capo et al., (2018). 1. Context What is the leading-order energy balance of the subtropical gyre of the North Atlantic? Which flows does it rely on, and how does it vary in space? • The North Atlantic is a key region in the global ocean circulation and plays a central role in regulating our climate (Ganachaud and Wunsch, 2000). • Its response to climate change is constrained by energy transfers between basin-scale time-averaged currents and smaller-scale (hundreds to the order of a kilometer) time-varying currents, including low- frequency mesoscale and submesoscale currents and high-frequency internal waves. • Although understanding the energy pathways of North Atlantic’s subtropical gyre is crucial for accurately representing its role in climate models, we lack (1) a description of energy pathways at the basin-scale (2) based on a dataset simultaneously resolving the different time-varying flows. Snapshots of surface (right) temperature and (left) normalised relative vorticity. (Right) Streamlines show 4-year averaged surface currents. Black square denotes the subtropical gyre region. 4. (time-mean) and (time-varying) budgets for tidal scenario K K′ (Top) Vertically-integrated averaged over a 4-year period. The four sub-regions: Coastal Gulf Stream and separation, Gulf stream’s extension, Interior and Eastern boundary are shown as squared areas. Streamlines show 4-year averaged surface currents. (Left and right) Frequency-wavenumber spectra and scheme of ocean energy pathways for the four sub-regions. In the schemes, amounts have been computed from time-averaged and vertically-integrated terms and then horizontally-averaged [10-3 m3 s-2]. K + K′ K′ : main energy balances : weak energy pathways (x <.>): ratio between net contributions in tidal and non-tidal scenarios Subtropical gyre’s interior • is larger than the reservoir by a factor of 6 and is dominated by submesoscale currents and internal waves. • and budgets are insensitive to tidal forcing. • budget follows a Sverdrup-like balance: only region with a net export provided by the wind. • reservoir is mainly energised by wind and tides. K′ K K K′ K K K′ • State-of-the-art
poster
Rotational Evolution and Radii of Young Sun-like Stars in NGC 2264 Laurin M. Gray & Katherine L. Rhode Indiana University, Bloomington, IN (USA) Background & Motivation • Understanding the origin and evolution of the Solar System and other Sun-like stars and planetary systems is a primary goal of modern astrophysics. • Studying T Tauri stars provides insight into the early interactions between Sun-like stars and their protoplanetary disks, and the time scales governing disk dissipation and planet formation. • NGC 2264 is a well-studied, nearby (~760 pc) open cluster with an age of ~3 Myr and a large T Tauri star population. The wealth of literature data make it an ideal target for developing our methods. Observations & Methods • The WIYN 3.5-m telescope and Hydra multi-object spectrograph were used to obtain high-dispersion (R ~ 21,500) spectra of stars in NGC 2264 and measure the stars’ projected rotation velocity (v sin(i)) values. • Our final NGC 2264 sample includes ~250 T Tauri stars. Roughly 30 are likely spectroscopic binaries; we excluded these stars from some analysis because they can yield overestimated v sin(i) values. • We use cross-correlation to measure v sin(i) and can measure values down to ~11 km/s. Our cross-correlation template stars have a range of temperatures matching those of our sample stars. We also observed v sin(i) standard stars to check our measurement accuracy. Future Work Open Cluster IC 5070 NGC 2264 IC 348 Upper Sco h Persei Age (Myr) 1-3 2-3 2-6 8-10 ~14 Distance (pc) 795 760 315 140 2100 References Baraffe et al. 2015, A&A, 577, A42 Jackson et al. 2018, MNRAS, 476, 3245 Somers et al. 2020, ApJ, 891:29 Acknowledgements We thank Catrina Hamilton-Drager and Luisa Rebull for sharing data and helping us design the project. We also thank the IU Astronomy Department and IU College of Arts & Sciences for providing funding. We have selected four additional clusters (see table), ranging from 1-14 Myr in age, for similar analysis. This will allow us to investigate the rotational evolution of low-mass stars during the timescale when protoplanetary disks dissipate. HR diagram of NGC 2264 stars used in the model comparison, with isochrones (yellow dashed lines) and mass tracks (black dotted lines) from the starspot-free Baraffe et al. 2015 model (left) and Somers et al. 2020 model with 51% starspot coverage (right). The main sequence is approximated at 100 Myr by the solid orange line. The measured v sin(i) values of the stars are indicated by the color bar, and open symbols mark stars with v sin(i) below our 11 km/s resolution limit. Vertical gray lines mark temperature-based subgroups used in the model comparison. At the age of NGC 2264 (~3 Myr), the model with starspots fits the data better than the starspot-free model. Spectra of a K2 cross-correlation template star (top) and three NGC 2264 stars (bottom three), with our v sin(i) values marked. The absorption lines broaden as v sin(i) increases. Stars with Accreting Disks May Rotate Slower Stars in Binary Systems May Rotate Faster Starspots Are Needed to Match Measured Radii Questions? Contact me at: grayla@iu.edu • A 2-sample K-S statistical test finds that the v sin(i) distributions do not have a statistically significant difference (p=0.126), but the large fraction of stars that lie below our resolution limit could mask a genuine difference in the two samples. • The fact that a higher proportion of CTTSs have low v sin(i) values compared to WTTSs seems to imply that stars with accreting disks may rotate more slowly than stars without active accretion. • 59% (49/83) of disked, CTTS stars have v sin(i) <11 km/s, while only 36% (47/130) of WTTS stars have v sin(i) below this limit. • Binarity can cause overly wide CCF peaks, resulting in overestimated v sin(i) values; this analysis only includes binaries with clearly resolved CCF peaks or single-lined binaries identified by differing radial velocities at different epochs. Stars identified as “single sy
poster
What's the problem we are addressing ? Why is it important? Sub-Saharan African communities face multiple livelihood challenges, including agri-food systems that are under pressure to sustain a growing population, in the context of environmental change, including climate change, and rising fuel and food prices due to global conflict and other global market prices shocks, e.g. the Covid-19 pandemic For the more than 60% of sub-Saharan Africans who are smallholder farmers, agri-food systems are not just sources of income. They are vital for sustainable livelihoods, including food security, sanitation, and access to education and healthcare, and can drive inclusive, locally-led sustainable development within rural communities. What is our approach to solving the problem? BIO4Africa will empower smallholder farmers to generate new sources of income by creating value from locally available biomass, utilising agricultural and food processing residues in a circular, bio-based approach to economic development. Our focus is on developing simple, small-scale and robust bio-based technologies adapted to local needs and contexts, including biomass types. These will support farmers and local businesses to sustainably produce a variety of higher value bio-based products from agriculture and food processing residues (animal feed, fertiliser, pollutant absorbents, construction materials, packaging, solid fuel for cooking and catalysts for biogas production). Products developed through this project will further contribute to sustainable local community development, by addressing food production, household energy, and health needs, e.g. bio-based soil improvement, cooking fuel, and water filtration products. Circular bioeconomy approaches offer opportunities for income generation from agricultural and food processing residues, e.g. corn cobs, while also guiding fossil fuel-free pathways for sustainable community development, e.g. reducing waste directly and over the product life cycle, displacing fossil-based products such as plastics with bio-based alternatives, and developing novel products, e.g. forage-based protein extracts. Developing and strengthening circular bio-based value chains in sub-Saharan Africa can support rural communities to access ecologically and socially sustainable economic opportunities. Project Team and Key Collaborators Côte d’Ivoire: Institut National Polytechnique Félix Houphouet- Boigny (INP-HB) Denmark: Food and Bio-Cluster Denmark France: RAGT Energie SAS & Centre de coopération Internationale en Recherche Agronomique pour le Développement Ireland: Celignis Ltd. & Munster Technological University Ghana: Savannah Young Famers Network (SavaNet), 0km Nomads, Agri-Business Innovation Hub Greece: Q-Plan International Advisors PC, Draxis Environmental SA Kenya: Eastern Africa Farmers’ Federation Society Netherlands: Stichting IHE Delft Institute for Water Education, Grassa BV Senegal: Université Assane Seck De Ziguinchor (UASZ), SCPL SA, Association d’Appui aux Initiatives de Paix et de Développement, Energeco Afrique, GIE Country Farm Spain: Barcelona Plataforma Empresarial SL, Fundacion Corporacion Tecnologica de Andalucia, Sustainable Innovations Europe SL Uganda: Kabarole Research and Resource Centre (KRC), African Forum for Agricultural Advisory Services United Kingdom: African Agricultural Technology Foundation We aim to support the deployment of the bioeconomy in rural areas in four sub-Saharan African countries: Senegal, Ghana, Uganda and Côte d’Ivoire, through the development of bio-based solutions and value chains with a circular approach to drive the cascading use of local resources and diversify the income of farmers. During the four-year term of the project: small-scale bio-based technologies will be adapted to local needs and context, and piloted in the target regions, technology pilot cases will be supported by laboratory analysis and field trials of the resulting bio-based products with the target consume
poster
Magnetic Coriolis darkening in cool stars Charly Pinçon - Institut d’Astrophysique Spatiale, OSUPS, Observatoire de Paris with the collaboration of L. Petitdemange, R. Raynaud, L. Garcia, A. Guseva, E. Alecian, and M. Rieutord 8th TESS/15th Kepler Asteroseismic Science Consortium Workshop 15-19 July 2024, Porto, Portugal Introduction Modeling the surface brightness distribution of stars is of prime importance to interpret the large amount of available interferometric, spectropolarimetric, or photometric observations. Beyond stellar physics, this is also a prerequisite to characterize exoplanets or our Galaxy. Stellar rotation is known to be a key physical ingredient in this regard. In rapidly rotating, hot massive stars with a stably stratified radiative envelope, the centrifugal flattening can make the poles hotter and brighter than the equator; this so-called gravity darkening is today well constrained (e.g., von Zeipel 1924, Bouchaud et al. 2020). In contrast, for low-mass cool stars, the picture remains uncertain as it requires one to model the MHD turbulence that develops in their convective envelopes. In this poster, we present the results by Raynaud et al. (2018) and Pinçon et al. (2024) who studied the effect of the Coriolis force on the surface luminosity distribution of thick convective envelopes, as well as the effect of potential dynamo magnetic fields. MHD simulations - MagIC and Parody codes: solve anelastic MHD Navier-Stokes equations. - Rotating convective envelopes in [ ri , ro ]. - Parametric study: about 100 simulations. - Density contrast between 20 – 400. - Entropy jump to force convection. - Stress-free and insulating boundaries. - Thick shell: representative of a 0.35 ∼ Msun dwarf or a PMS star Fig. 3: Spherical representation of the mean luminosity distribution (measured by the Nusselt number Nu) for the nonmagnetized model and the most magnetized dynamo model (i.e., the highest value of Pm) represented in Fig. 1. Radial velocity fluctuations Results Rosurf > 1 → Turbulent homogeneization of the surface. Rosurf < 1 →Complex interplay between rotation and self-sustained magnetic fields. - Rotation-constrained turbulence can generate a strong equatorial jet (Fig. 2) : Radial heat flux inhibited, dark equatorial band (Fig. 1). - With magnetic dynamo, Pm ↗ → B ↗ : Lorentz force quenches the equatorial jet (Fig. 2) and the luminosity profile flattens (Fig. 1). In conclusion, in rapidly-rotating cool stars, an equatorial darkening may occur, whose contrast depends on the rotation rate and the magnetic field intensity (see Fig. 3). Fig. 1: Surface heat flux (measured by the Nusselt number Nu) as a function of the colatitude, and for different values of the magnetic Prandtl number Pm. These profiles are mean values computed over azimuth and time. N.B: The parameters of the models are Ek=3 10-4, Pr=1, Ra~ 107, Nρ=6. Fig. 2: Same as in Fig.1 but for the mean profile of the azimuthal velocity (inside the rotating frame of rotation rate Ω). Fig. 4: Schematic transit of an exoplanet occulting a star with an equatorial jet dimmer than the rest of its photosphere. The bottom plot shows the occultation light curve (black line), the differential flux measured for the six times shown above (points), as well as the transit signal that would be observed for a uniform limb- darkened star (gray). Surface Rossby number Magnetic Prandtl number Prospects Such a magnetic Coriolis darkening can have an impact on not only the stellar luminosity estimate, but also on planet transit retrievals, in the same way as limb and gravity darkening effects. These predictions will have to be tested in the future, either in a direct way through high resolution interferometry or indirectly through misaligned transits (Fig.4).
poster
2. Access to technology is not obvious 2. Fluorescence Fluctuation-based Super-Resolution Microscopy (FF-SRM) napari-superres an open-source implementation of methods for Fluorescence Fluctuation-based Super-Resolution Microscopy (FF-SRM) 3. Superres plugin 4. Conclusions Rocco D’Antuono1,2, Raúl Pinto Cámara3,4, Paul Hernández-Herrera4, Esley Torres García3,4, Alejandro Linares4, Haydee Hernández5, Adán Guerrero4 1 Crick Advanced Light Microscopy STP, The Francis Crick Institute, London, United Kingdom 2 Biomedical Engineering, School of Biological Sciences, University of Reading, Reading, United Kingdom 3 Centro de Investigación en Ciencias, Instituto de Investigación en Ciencias Básicas y Aplicadas, Universidad Autónoma del Estado de Morelos, Cuernavaca, Morelos, México 4 Laboratorio Nacional de Microscopía Avanzada, Instituto de Biotecnología, Universidad Nacional Autónoma de México, Cuernavaca, Morelos, México 5 Posgrado en ciencias e ingeniería de la computación Universidad Nacional Autónoma de México, Ciudad de México, México Acknowledgements We are grateful to Kurt Anderson (CALM head, The Francis Crick Institute) for supporting and encouraging the project and the professional development of the CALM staff. We thank Dr. Chris Wood (LNMA) for critical reading of the grant proposal. RD’A works at the Francis Crick Institute which receives its core funding from Cancer Research United Kingdom (CC0199), the United Kingdom Medical Research Council (CC0199), and the Wellcome Trust (CC0199). This project is entirely based on the use of open-source software that makes science easier, a huge thank you to the developers of: numpy, matplotlib, pandas, JupyterLab, napari, etc. The future software development will be supported by the napari Plugin Foundation award (Chan-Zuckerberg Initiative). References 1. Alva, Alma, Eduardo Brito-Alarcón, Alejandro Linares, Esley Torres-García, Haydee O. Hernández, Raúl Pinto-Cámara, Damián Martínez, et al. 2022. “Fluorescence Fluctuation-Based Super-Resolution Microscopy: Basic Concepts for an Easy Start.” Journal of Microscopy, August. https://doi.org/10.1111/jmi.13135. 2. Agarwal, Krishna, and Radek Macháň. 2016. “Multiple Signal Classification Algorithm for Super-Resolution Fluorescence Microscopy.” Nature Communications 7 (1): 13752. https://doi.org/10.1038/ncomms13752. 3. Miao, Yuting, Shimon Weiss, and Xiyu Yi. 2022. “PySOFI: An Open Source Python Package for SOFI.” Biophysical Reports 2 (2): 100052. https://doi.org/10.1016/j.bpr.2022.100052. 4. Yahiatene, Idir, Simon Hennig, Marcel Müller, and Thomas Huser. 2015. “Entropy-Based Super-Resolution Imaging (ESI): From Disorder to Fine Detail.” ACS Photonics 2 (8): 1049–56. https://doi.org/10.1021/acsphotonics.5b00307. 5. Cox, Susan, Edward Rosten, James Monypenny, Tijana Jovanovic-Talisman, Dylan T. Burnette, Jennifer Lippincott-Schwartz, Gareth E. Jones, and Rainer Heintzmann. 2012. “Bayesian Localization Microscopy Reveals Nanoscale Podosome Dynamics.” Nature Methods 9 (2): 195–200. https://doi.org/10.1038/nmeth.1812. 6. Gustafsson, Nils, Siân Culley, George Ashdown, Dylan M. Owen, Pedro Matos Pereira, and Ricardo Henriques. 2016. "Fast live-cell conventional fluorophore nanoscopy with ImageJ through super-resolution radial fluctuations." Nature communications 7 (1): 1-9. https://doi.org/10.1038/ncomms12471 7. García, Esley Torres, Raúl Pinto Cámara, Alejandro Linares, Damián Martínez, Víctor Abonza, Eduardo Brito-Alarcón, Carlos Calcines-Cruz, et al. 2021. “Nanoscopic Resolution within a Single Imaging Frame.” BioRxiv, October, 2021.10.17.464398. https://doi.org/10.1101/2021.10.17.464398. 8. https://github.com/RoccoDAnt/napari-superres 5. Future developments • Fluorescence microscopy is a well-developed imaging modality that is used in life sciences to study tissues, organoids, and cells. The inspection on the microscopic scale is limited by the resolving power of the used optical system, but more specialized techniques for Super-Resolutio
poster
104 105 106 107 108 Modified mean [-] 104 105 106 107 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Total volume eroded [m 3 ] 100 101 102 103 Modified mean [-] 100 101 102 103 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Final max. elevation [m] 100 101 102 103 104 Modified mean [-] 100 101 102 103 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Final mean elevation [m] 10-1 100 101 102 Modified mean [-] 100 101 102 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Final stdev elevation [m] 10-1 100 101 102 Modified mean [-] 10-1 100 101 102 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Final max. slope [-] 10-3 10-2 10-1 100 Modified mean [-] 10-3 10-2 10-1 100 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Final mean slope [-] 10-2 10-1 100 101 Modified mean [-] 10-2 10-1 100 101 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Final stdev slope [-] 103 104 105 Modified mean [-] 103 104 105 106 Standard deviation [-] γ D Kbed Kblock Tr z0 τc ∆z τcb Total blocks delivered [# ] 1) CIRES and Dept. of Geological Sciences, CU Boulder 2) EarthLab, CU Boulder 0.0 0.2 0.4 0.6 0.8 1.0 x / L 0.0 0.5 1.0 1 f c 0.0 0.5 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x / L 0.0 0.5 1.0 1 f c 0.0 0.5 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x / L 0.0 0.2 0.4 0.6 0.8 1.0 Elevation / total baselevel lowering 0.0 0.2 0.4 0.6 0.8 1.0 x / L 0.0 0.2 0.4 0.6 0.8 1.0 Elevation / total baselevel lowering 0.0 0.2 0.4 0.6 0.8 1.0 x / L 0.0 0.2 0.4 0.6 0.8 1.0 Elevation / total baselevel lowering A B1 C1 B2 C2 Legend Constant normalized elevations of 0.2, 0.4, and 0.6 1-fc Fraction of bed exposed at 400 ky τ* Ratio of bed shear stress to total shear stress at 400 ky mean τ* = 0.63 mean 1-fc = 0.95 mean τ* = 0.45 mean 1-fc = 0.89 τ* τ* γ = 0 m-2 γ = 0.2 m-2 γ = 1.0 m-2 γ = 0.2 m-2 γ = 1.0 m-2 1 2 3 4 5 6 10-5 10-4 10-3 10-2 Baselevel lowering rate [m/yr] 1 2 3 4 5 6 10-5 10-4 10-3 10-2 Baselevel lowering rate [m/yr] γ = 5 m −2 γ = 4 m −2 γ = 3 m −2 γ = 2 m −2 γ = 1 m −2 γ = 0.5 m −2 γ = 0.2 m −2 γ = 0.1 m −2 c= 1.5 c= 1.0 c= 0.5 c= 0.25 Slope exponent n [-] 10-4 10-3 10-2 10-1 100 Steady State Mean Slope [-] y = 1401.8x1.3613, R2 = 0.99 No blocks, constant τc = 100 Pa Blocks, γ = 5 m −2 No blocks, variable τc fit No blocks, constant τc = 0 Pa 10-5 10-4 10-3 10-2 Baselevel Lowering Rate [m/yr] 10-4 10-3 10-2 10-1 100 Steady State Mean Slope [-] y = 1401.8x1.3613, R2 = 0.99 Blocks, γ = 5 m −2 No blocks, τc = 300000E + 40 Pa Blocks, γ = 2 m −2 No blocks, τc = 120000E + 30 Pa Blocks, γ = 0.5 m −2 No blocks, τc = 70000E + 5 Pa No blocks, constant τc = 0 Pa 0.0 0.2 0.4 0.6 0.8 1.0 ∆Z = 0.00001 m/yr 0.0 0.2 0.4 0.6 0.8 1.0 x / L 0.0 0.2 0.4 0.6 0.8 1.0 ∆Z = 0.001 m/yr 0.0 0.2 0.4 0.6 0.8 1.0 x / L Elevation / total baselevel lowering Block model Threshold model Block model Threshold model 0.6 0.8 1.0 1.2 1.4 Steady state φ 10-5 10-4 10-3 10-2 Baselevel lowering rate [m/yr] 0.6 0.8 1.0 1.2 1.4 Steady state φ 10-5 10-4 10-3 10-2 Baselevel lowering rate [m/yr] τc fit to γ = 0 m −2 τc fit to γ = 0.5 m −2 τc fit to γ = 2 m −2 τc fit to γ = 5 m −2 c= 1.5 c= 1.0 c= 0.5 c= 0.25 D0 dy dx qb q high fc low fc 10-4 10-3 10-2 10-1 100 y = 998.9x1.3615, R2 = 0.99 y = 1072x1.3617, R2 = 0.99 10-5 10-4 10-3 10-2 Baselevel lowering rate [m/yr] 10-4 10-3 10-2 10-1 100 y = 1401x1.3613, R2 = 0.99 10-5 10-4 10-3 10-2 Baselevel lowering rate [m/yr] y = 2543x1.3514, R2 = 0.99 γ = 5 m −2 γ = 4 m −2 γ = 3 m −2 γ = 2 m −2 γ = 1 m −2 γ = 0.5 m −2 γ = 0.2 m −2 γ = 0.1 m −2 γ = 0 m −2 c= 1.5 c= 1.0 c= 0.5 c= 0.25 Steady state mean slope [-] 1) The interaction between blocks, discharge variability, and slope adjustment sets channel response to block delivery. 2) Blocks significantly alter topographic scaling at channel slopes below the critical slope for block transport. 3) The "n" exponent values depend on block effects and proximity to the critical block transport slope. 4) The complex influence of blocks on bed cover and form drag is well-described by an erosion th
poster
Read Between the Lines: Closing Gaps of Materials and Methods to Build Workflow from the Publication Tazro Ohta1, Osamu Ogasawara2, Yoshinobu Masatani3, Shigetoshi Yokoyama3, Kento Aida3 1. Database Center for Life Science, Chiba, Japan 2. DNA DataBank of Japan, Shizuoka, Japan, 3. National Institute of Informatics, Tokyo, Japan Background Increasing cost of translation from text to executable scripts Galaxy as an inter-laboratory workflow sharing platform workflow runs on any galaxy, and the galaxy runs anywhere Fig. 1. FANTOM5 (http://fantom.gsc.riken.jp/5/) data processing protocol descriptions. (a) Description of data processing in the main paper of FANTOM 5 project. (b) Online protocol published by FANTOM 5 project. Though this is helpful, it is still not enough to reproduce the exact same results. (c) A script we made with a help by members of FANTOM 5 project. There are some missing informations from materials and methods or online protocols. a b c Application: Community ‘Pitagora’ Galaxy VM Platform: ‘Overlay Cloud’ datacenters We have been trying to develop the system that we are able to share community’s workflows developed and used in the various studies which deal with NGS data, such as Exome-Seq, RNA-Seq, ChIP-Seq, or Bisulfite-Seq. We developed Virtual appliance that includes workflows and some tools to help execution of the workflows, and the Virtualbox image and Amazon Machine Image are distributed. We also maintain public server for the test use, and documents and test data are hosted on our project website. Challenges “Overlay Cloud” project led by National Institute of Informatics (Tokyo, Japan) is building large scale cloud computing infrastructure that can distribute computing tasks via their high-speed network system “sinet”. We are testing to build galaxy environment on their cloud system, which runs all the process or application as a docker container. Apache Mesos running over the nodes and datacenter is managing computing resource, and galaxy runs on it as a container, then run tools as a docker container. We have build some tool.conf and job runner for containerized tools and are still being tested and under the benchmarking. Fig. 2. Community Galaxy VM and Galaxty Community Japan. We have been organizing Japanese local community of the galaxy developers and users. The community is sharing the tools and workflows that are used in their labs. Workflows are implemented and fully documented online with the test data. Automated test of the workflow is being developed. Packed VM is distributed as a normal Virtualbox image and Amazon Machine Image via Amazon Web Service. The community is also hosting public server (listed on the page of galaxy project ‘virtual appliance’). 🐳! Mesos Master 🐳 Mesos Slave Node Node Node 🐳! Galaxy 🐳 Mesos Slave 🐳 BWA 🐳 samtools 🐳! CLI- manager 🐳! Marathon 🐳! SCS 🐳! NFS 🐳Mesos! Master 🐳 Mesos Slave Node Node Node 🐳! Galaxy 🐳 Aurora 🐳 Mesos Slave 🐳 BWA 🐳 samtools 🐳! CLI- manager 🐳! Chronos L2VPN 🐳 Docker Container • Performance Comparison and benchmarking • Connecting published workflows to the publication •Toolshed or other data repository • Standard description or annotation of workflows • Data transfer via internet • burst buffer like cloud volume? Fig. 3. Overlay cloud system and container-based galaxy and galaxy tools. (top-left) Current system which is fully containerized computing infrastructure for the galaxy. Each application runs as a container and connected by L2VPN. Data are stored in the NFS mounted on each node. (bottom-right) Current idea of ver. 2.0 system. Apache mesos and marathon container are detached from the network of the containers of galaxy and galaxy tools which are attached to the NFS via L2VPN. This enables more portability to the running system. Publishing and sharing data analysis workflow using the galaxy platform has spectacularly reduced the cost of reproducing one's research, but following the description of data analysis which had been performed
poster
§ Expanded Illinois Natural History Survey host database with additional records from 2010-2020 from Google Scholar literature search § Compiled host database into usable data § Wrote R script to generate summary data for various measures, using t-tests to compare 2 groups and ANOVA for over two Methods § Both the IUCN and federal status rankings showed that, on average, mussels listed as endangered used significantly fewer host families than non-threatened mussels (IUCN, 1.94 and 3.25 respectively; p<.01. Federal, 2.00 and 3.21 respectively; p<.01). § Anodontini was significantly different in host family numbers compared to Quadrulini(p<.01), Pleuroblemini (p<.001), and Lampsilini (p < .001) Mussels are commonly thought of as distinct groups of host-specialists and host-generalists (Haag, 2012). However, this research supports the idea that this division is less of a dichotomy and more of a continuum (Fig. 1). The mussels that fall into the threatened categories of the conservation assessments tended to be more specialized than those that were non-threatened (Fig. 2). Some taxa also display more specialized host use than others (Fig. 3). These insights provide predictive power that can be used to help prioritize taxon-based conservation efforts. Many thanks to the Smithsonian Institution and the National Museum of Natural History, especially to Elizabeth Cottrell, Gene Hunt, and Virginia Power. Also, a thank you to Sarah Douglass and all others involved in the creation & maintenance of the Illinois Natural History Survey Freshwater Mussel Host Database. Funding provided by the NSF. Photo by Robert Aguilar, Smithsonian Environmental Research Center. Haag, W.R., 2012. North American freshwater mussels: natural history, ecology, and conservation. Cambridge University Press Results Conclusion Acknowledgements Using Freshwater Mussel Host Use to Examine Conservation-Relevant Factors David H. Nichols1,2, John Pfeiffer1 1Smithsonian National Museum of Natural History, Department of Invertebrate Zoology, 2Hendrix College Background Freshwater mussels are some of the most endangered animals in the US (Haag, 2012). Federal and state conservation efforts are complicated by a complex life history that requires parasitizing a host. While some freshwater mussel species can parasitize a broad range of host fish species, other mussels can only complete their life cycle by parasitizing a single species. Understanding more about how these factors influence their conservation status is an essential part of influencing their conservation going forward. Big Questions § Does host number impact the level of endangerment for a mussel species? § Do different taxa of mussels have different trends in host numbers? § Based on this information, where should we focus conservation efforts in the future? Fig. 1. Frequency distribution of host number by taxon Fig. 3. Plots showing the number of host families for species across different taxa Fig. 2. Distribution of host number between mussels of various conservation status
poster
2.Energy Response Model of Detector Global Fitting result Potential of geo-neutrino measurements at JUNO and the Local 3D model Ran Han On behalf of the JUNO collaboration (Poster no.) The Local Crust Geochemical Model Poster Wall #204 The Jiangmen Underground Neutrino Observatory (JUNO) 20 kton LS detector 3%/sqrt(E)energy resolution A multiple-purpose neutrino experiment: Reactor neutrinos, Supernova neutrinos,Geo-neutrinos, Solar neutrinos, Sterile neutrinos, Atmospheric neutrinos, Exotic searches Taishan NPP Guang Zhou Daya Bay Hong Kong JUNO JUNO Location Geo-neutrinos: a new probe of Earth's interior Open questions about natural radioactivity in the Earth 1 - What is the radiogenic contribution to terrestrial heat production? 2 - How much U and Th are in the crust and particularly in the mantle? 3 – A global check of the standard geochemical model (BSE)? 4 - What is hidden in the Earth’s core? (geo-reactor, 40K, …) Geo-neutrinos: anti-neutrinos from the Earth. U Th and 40K in the Earth release heat together with anti-neutrinos in a well fixed ratio. They bring to Earth’s surface information about the chemical composition of the whole planet. Geo-neutrino Spectrum Earth’s surface heat flow JUNO Potential in Measuring Geo-neutrinos • 400events/year, much larger than existing experiments • With 10 years: total uncertainty reach 5%(2TNU) • Comparison of the global reference model (18% crust) and a benchmark accuracy of the local model(8% crust) The energy spectrum geoneutrino signal vs radiogenic heat the precision of the geo-neutrino measurement The chi2 distribution of Mantle geo JUNO JUNO Data collection (U & Th abundances),3000 points Preliminary surface model (0.5º x 0.5º) Preliminary vertical model (4 layers) The Local Crust Geophysical Model Why Local 3D model The 3D density distribution (one line) The crust thickness around JUNO. Results from the inversion of Bouguer gravity anomalies, constraint by the seismic sounding The 3D Vs structure R. Han, Y.F.Li et al.Chin.Phys.C(2016) Zhi-Wei LI (IGGWH, CAS) Ruo-Han GAO (CUGB) • Near continental margin, affected by the geology evolution of South China Block and the South China Sea • Paleogeographic evolution of South China fold belt during Permian-Jurassic time and proposed flat-slab subduction model for propagating • Local geology, including the geometry of strata, distribution rocks play important role in geo-ν flux estimation What’s the special of JUNO Geology Significant uncertainties exist in current global models. High-resolution local crust model are needed Global CRUST1.0 model has low resolution (1°× 1°) Global abundance model has no spatial resolution and large uncertainties Continuous Seismic Waveforms, from 450 Stations in South China Geophysical data source Local Neutrino Geo-Science Working Group
poster
Iván Muñoz Rodríguez1,2, Antonis Georgakakis1, Francesco Shankar2 et al. (submitted); e-mail:ivan.rodriguez@noa.gr 1National Observatory of Athens, 2University of Southampton Differenal evoluon of AGN in clusters Comparisson with observations: We calculate the fraction of X-ray AGN in a cluster parent sample defined as in observations. We compare the results of the models with the observations of Matini et al. (2009, 2013). We select galaxies and AGN in the cluster sample following selection effects as defined in observations. Only galaxies/AGN within the virial radius, brighter than a magnitud limit and within a velocity dispersion are consider a members. The impact of selection effects is illustrated in the figure on the left. The number of AGN represent the field expectation by model construction. Results are shown in the figure on the right. The model underpredicts the number of AGN at z~1.2 respect the observations. At z~0.75 model prediction is consistent with observations. Finally, at low redshift, z~0.2, the model seems to overproduce AGN respect cluster observations, although firm conclusions cannot be made since observations only place a lower limit. References: Aird et al. (2018); Behroozi et al. (2019); Georgakakis et al. (2017, 2019); Klypin et al. (2016); Martini et al. (2009, 2013); Peluso et al. (2022) Motivation and semi-empirical model construction: The role of different AGN triggering mechanisms remains as one of the open questions in Astrophysics. We studied the influence of small-scale environment (<1 Mpc) into the AGN triggering mechanism. We developed a novel semi-empirical model (SEM) to construct AGN mock catalogues based on state-of-the-art observations. A graphical workflow of this model is shown in the figure on the left. The starting point are dark matter N-body simulations (MDP2, Klypin et al. 2016), that are populated with galaxies using abundance matching techniques (UNIVERSEMACHINE, Behroozi et al. 2019). Then AGN are painted on this galaxies constrained by the latest observations of AGN populations. The latter step is done under the explicit assumption that accretion events are independent of the environment (i.e halo mass). The figure shows a 10 Mpc/h slice of a box from MDP2 cosmological simulation with 1000 Mpc/h side size at the snapshot z=0.75. The dots represent the positions of dark matter haloes (top panel), galaxies within these dark matter haloes (middle panel) and AGN within the same dark matter haloes (bottom panel). The key feature of this approach is the reproduction by construction of the predictions and observables (e.g. HMF, SMF and XLF). The key parameter to populate galaxies with AGNs is the specific accretion rate defined as λ=Lx/M★. It is related with two observables that have been well characterized in the last years thanks to multi-wavelength observations in extragalactic survey fields. This has allowed to infer its probability distribution P(λ). It describes the probability of a galaxy with mass M★ of hosting an AGN with luminosity Lx. For each galaxy in the catalog we randomly pick a λ from the distribution. This effectively associate an AGN X-ray luminosity to all galaxies. AGN incidence and physical interpretation: Field expectation cannot reproduce the observations. We find a differential evolution of the incidence of AGN in clusters respect to field. AGN activity is enhanced at high redshift and suppressed at low redshift in massive clusters respect to field. This point to a different efficiency of AGN triggering mechanisms depending on their small-scale environment (<1 Mpc) that evolves with cosmic time. This could be explained by the different physical processes happening in clusters respect low dense environments such as ram-pressure or galaxy interactions. This effects are expected to be stronger in the outskirts of the cluster. Galaxy-Halo connection UNIVERSEMACHINE (Behroozi et al. +19) AGN-Galaxy connection Count galaxies and AGN in the cluster sample Projection of a cluster into sk
poster
Life Cycle Phases (rows) Being part of a military command and having familiarity with the field of intelligence, the authors reviewed the U.S. Intelligence Cycle for applicability. Following the review, the authors decided to adopt this model for use within their framework. The life cycle consists of the following phases: • Planning and Direction – Establish the consumer’s intelligence requirements and plan intelligence activities accordingly. • Collection – Gather the raw data required to produce the finished product. • Processing and Exploitation – Convert the raw data into a comprehensible format that is usable for production of the finished product. • Analysis and Production – Integrate, evaluate, analyze, and prepare the processed information for inclusion in the finished product. • Dissemination – Deliver the finished product to the consumer that requested it and to others as applicable. • Evaluation – Continually acquire feedback during the Intelligence Cycle and evaluate that feedback to refine each individual step and the cycle as a whole. Project Background The Readiness Integration Center (RIC) is a Department of Defense (DoD) science gateway purpose built to provide data analysis and decision support capabilities to the Naval Aviation community with a primary emphasis on aircraft readiness and safety. The RIC ingests data from a variety of DoD authoritative sources (i.e., the systems of record charged with being the official data store), including: aircraft sensor data related to heat, vibration, built-in-test results, and fault identification; flight logs; aircraft status; maintenance records; repair manuals; supply status; and engineering data. RIC users come from several communities such as maintenance, supply, training, and engineering. These users have access to dozens of custom-built, automated and manual analysis tools as well as source data, processed data, and high-performance computing resources. Problem & Solution Recognizing that “garbage in” leads to “garbage out,” many organizations focus their efforts on pursuing a zero- defect data ingestion process. While these efforts are critical for reducing quality-related risks, the authors wanted to ensure their science gateway not only had procedures for managing quality-related risks at ingestion but also for managing such risks during subsequent analysis and reporting. By having published procedures in place before an incident, teams could better mitigate the cascade of risks that could occur if poor quality data, despite best efforts, made its way into the science gateway. For example, erroneous data could inadvertently be used during modeling and simulation activities. Then, the results of the modeling and simulation activities that were based on erroneous data could be used as evidence for a written report. That written report, based on the results of modeling and simulation activities that were based on erroneous data, could then be used as a basis for critical decisions. To ensure the science gateway had procedures for managing quality-related risks such as those in the example above, the authors developed a quality risk management framework covering 36 unique areas where quality risks could arise. This framework is intended to be used, as it was by the authors, to evaluate a science gateway’s readiness to manage quality issues arising at data ingestion and subsequent phases such as analysis and dissemination. Results In the case of the authors’ science gateway, use of the framework provided critical insights. Specifically, quality risk management procedures were written and published for only 15 of the 36 areas identified by the matrix, quality risk management procedures were written but not published for 14 of the 36 areas, and 7 of the 36 areas had no associated quality risk management procedures. Based on these findings, the authors worked with other team members to ensure the missing quality risk management procedures were written and/or publishe
poster
Cyber Training: Pilot -- Breaking the Compute Barrier, Upskilling Agri-Food Researchers to Utilize HPC Resources Kevin Silverstein, Benjamin Lynch, Alison Joglekar NSF Award Number: 2320769 MOTIVATION The era of “Big Data in Agriculture” is rapidly moving well beyond genomics data to encompass environmental, management and socio-economic data sourced from satellites, unmanned aerial vehicles (UAV), stationary and robot-enabled ground-based sensors, and ever more data-enabled agri-food machinery (Kamilaris et al. 2017; Shekhar et al. 2017). Leveraging high-performance computing (HPC) assets into the agri-food domain space is key to realizing a big data revolution in agriculture (EU 2018; Georgiou et al. 2020). However, there is a dearth of scientists with expertise in the agri-food and natural resource domains that have compute-to-scale capabilities enabled by HPC environments. Low adoption of HPC capabilities among agri-food researchers can be largely attributed to the real (or perceived) complexity of using HPC. OBJECTIVE Onboard and upskill the agri-food workforce so they can effectively and efficiently tap the analytical horsepower of cyberinfrastructure (CI), specifically HPC, via the creation of a tiered multi-module learning curriculum tailored to CI-applications in the agri-food sciences. CURRICULUM HPC and Cloud Concepts 05 ● Divide and conquer ● Parallelization ● Containerization ● Storage types Computer Science for the Agri-Food Researcher 04 ● Algorithm design ● Data structures ● Computational complexity (Big-O) ● Profiling Expanding HPC Skills 03 ● Machine Learning primer ● Data leakage ● Overtraining HPC Basics 02 ● UNIX primer ● Working remotely and scheduling jobs using SLURM ● Syncing your work with the community Computer Hardware and Resources 06 ● Processor types ● Memory ● Networking speed Future trends 07 ● GPUs continue to soar in usage ● Generative AI models are rapidly increasing in complexity 01 ● Agricultural research ● Computers in agricultural research Intro to HPC Computing for Ag Researchers COURSE MATERIAL EXAMPLES Easy-to-Learn Jupyter Notebooks Easy-to-Navigate Canvas-Based Materials CHECK OUT OUR NEXT OFFERINGS Fall 2024 – Free (U of Minnesota Affiliates Only) Spring 2025 – Free for Everyone Tomato Image Source: https://aeroponic.gr/en/knowledge/post/diataraches-t hrepsis-futon-trofopenies-toxikotites1.html
poster
Visual and Chemosensory Guidance of Predatory Behavior in the California Two-Spot Octopus (Octopus bimaculoides) Dana Callahan Biology Department & Program In Neuroscience. Middlebury College. Middlebury Vermont We assessed the effects of vision and chemosensation on predatory behavior in Octopus bimaculoides.. Octopuses underwent three trials: visual stimulus only, chemical stimulus only, and visuochemical stimulus. The behavioral paradigm required octopuses to attack live prey in a clear glass tube. Predatory behaviors were quantified, and results were compared across trials. Data collection for the first 5 subjects was completed on 4/16/13, and the next 6 subjects are arriving for testing on 4/17/13. For this reason, the observations presented below are based on preliminary qualitative observations of trials. Preliminary observations suggest that the chemical stimulus alone decreased latency to attack, relative to the visual stimulus alone or the visuochemical stimulus. Preliminary observations suggest that the chemical stimulus may have increased total time spent in contact with the stimulus, suggesting that chemosensation increased perseverance in predation. Interindividual variation in basal activity level, curiosity, and prey preference may have confounded our results, as suggested by previous studies (Ambrose 1982; Byrne et al., 2001). Preliminary comparisons across trials suggest that the order in which octopuses underwent the three trials did not significantly affect predatory behavior. This suggests that learning and/or desensitization to the stimulus did not occur. Vision is primary sensory trigger for predatory behavior. Chemical stimuli are significantly less effective than visual stimuli at inducing predatory behavior., but chemical stimuli may increase predatory perseverance (e.g. frequency of attacks and time spent in contact with the stimulus). This research is still underway, and we hope to strengthen our findings as we collect data on the next group of animals. Further development of this study warrants examination of the effects of a variety of chemical stimuli on predatory behavior. Octopuses (Phylum Mollusca, Class Cephalopoda) are proven models for the neurobiology of behavior (Messenger, 1998). However, neurobiologists have not achieved a complete understanding of sensory processing in the octopus (D’Aniello et al., 2005; D’Este et al., 2008). While recent studies have demonstrated visual guidance of behavior (Byrne et al., 2001; Fiorito et al., 2005) and chemosensory abilities in cephalopods(Lee, 1992; Darmaillacq et al., 2004; Fry, 2011), the biological relevance of these sensory capabilities continues to be examined. Octopuses are active predators, preying on fish, crustaceans (e.g. shrimps, crabs), and mollusks (Ambrose, 1982; Byrne et al., 2001). Octopus predatory behavior varies among individuals and depends on prey type (Dickel et al., 2001; Cartron et al., 2013). Such complex predatory behavior requires highly effective sensorimotor integration. Octopuses use vision to guide their initial approach to prey (Young, 1965). In addition, findings that octopuses demonstrate positive chemotaxis to chemical stimuli related to prey (Lee, 1992; Fry, 2011) suggest that predatory behavior may also be influenced by chemosensation. The present study seeks to clarify the ecological relevance of octopus chemosensation by studying how vision and chemosensation interact to guide predatory behavior in the California two-spot octopus (Octopus bimaculoides). We presented octopuses with visual stimuli, chemical stimuli, or visuochemical stimuli and quantified the resulting predatory behaviors. Based on previous research, we hypothesized that vision would be the primary initiator of predatory behavior, but that the effects of vision and chemosensation on predatory behavior would be additive. In other words, we hypothesized that predatory behavior would be stimulated most strongly by visual and chemical cues
poster
En los últimos años, España, y Asturias, se han convertido en un país receptor neto de inmigrantes. Esta migración no suele afectar en muchos casos en mejoras del nivel de salud de los. Inmigrantes. En este estudio pretendemos a partir de los datos de la Encuesta de Salud de Asturias, 2008 (ESA-08) conocer el grado de salud de la población inmigrante en Asturias. Los datos provienen de la ESA-08 (n=2500) en Asturias. Se analizan las variables de salud incluidas en la encuesta en relación al origen de la persona y se comparan con las obtenidas en la población autóctona. CCasi todas las variables de resultados de salud son mejores en inmigrantes que en población autóctona si bien muchas de esas diferencias se deben a su mayor juventud. Algunos de estos resultados son concordantes con los datos disponibles en la literatura. ANTECEDENTES Y OBJETIVO La salud en personas inmigrantes en el Principado de Asturias, 2008 Mario Margolles (1), Ignacio Donate (1), P. Margolles (2) (1) Consejería de Sanidad. Gobierno del Principado de Asturias, (2) Facultad de Medicina y Ciencias de la salud. Universidad de Oviedo MÉTODO RESULTADOS N. 640 Inmigración y salud CONCLUSIONES Declaración de conflicto de intereses: No existe ningún conflicto de intereses Situación de salud de la población inmigrante: (Encuesta de Salud. Asturias, 2008) Grupo de referencia: personas nacidas fuera de España vs personas autóctonas. Total: 2.500. Personas inmigrantes: 158. Entre paréntesis IC 95%. Autovaloración de estado de salud: Buen o muy buena OR: 1,92 (1,4-2,7); regular o mala, OR: 0,32 (0,2-0,6) incluso controlando por edad. Autovaloración estado de felicidad: NS. En cambio, SI hay diferencias en jóvenes (16-29 años): poco o nada felices, OR: 9,3 (3,4-25,6) Calidad de Vida Relacionada con la salud: diferencias significativas en las 5 dimensiones de EUROQOL-5D. Controlando por edad, las diferencias se mantienen entre los 30-64 años de edad: mejor CVRS en inmigrantes y las diferencias se ahondan con la edad (gráfico 1) Valoración de vida sexual: muy o bastante satisfactoria, OR: 1,6 (1,1-2,3). Peor valoración en mujeres inmigrantes vs hombres inmigrantes OR: 2,9 (1,3-6,6) No tener personas con las que se puede hablar (confidente): OR: 2,3 (1,1-4,9). Peor valoración en hombres inmigrantes vs mujeres inmigrantes OR: 3,1 (1,1-9,3) Morbilidad crónica: menor frecuencia OR: 0,54 (0,4-0,7). NS tras controlar por edad salvo en grupo de 45-64 años OR: 0,4 (0,2-0,8) En todas las patologías hay menor frecuencia significativa en población inmigrante que desparece controlando la edad, salvo: Ansiedad: OR: 3,2 (1,1-10,4) En el análisis por sexo aparece que las mujeres inmigrantes tienen más riesgo que las autóctonas de ansiedad OR: 46,0 (15-139) Accidentalidad: diferencias NS Autovaloración de vida social: diferencias NS Ayuda para el cuidado de menores o dependientes: OR: 1,1 (NS) Sueño: duermen menos las personas inmigrantes (en especial los hombres) pero descansan mejor: OR: 1,5 (1,1-2,3). Satisfacción con el trabajo: visión dual: buena o muy buena valoraciòn más frecuente en inmigrantes (en especial en mujeres), OR: 2,3 (1,6-3,5). Pérdida de empleo, bastante más preocupados por perder su trabajo en inmigrantes en especial las mujeres. Accesibilidad al sistema sanitario: diferencias considerables con la población autóctona. Mejor valoración de la geográfica OR: 1,5 (1,1-2,2). Muy mala percepción en la horaria OR: 1,7 (1,1-2,5), en especial en hombres inmigrantes. Necesidad de atención médica NO cubierta: mayor frecuencia OR: 2,7 (1,1-6,9), debido sobre todo a la imposibilidad de ir al médico por no poder dejar el trabajo. Frecuentación servicios sanitarios: Usan más los servicios de MF OR: 1,8 (1,1-2,4) y ginecologia OR: 1,8 (1,1-2,9). Usan menos los de enfermería. Casi la totalidad de la atención sanitaria ha sido financiada públicamente. En cuanto a la hospitalización es NS si bien si hay diferencias en motivo de ingreso: la población inmigrante ingresa menos en cirugía
poster
Investigating Galaxy Quenching with the SDSS: Stellar Metallicities as a Tracer of Quenching Mechanisms J. Trussler1,2,*, R. Maiolino1,2, D. Goddard3, C. Maraston3, D. Thomas3, Y. Peng4 1Cavendish Laboratory, University of Cambridge, 2Kavli Institute for Cosmology, University of Cambridge 3University of Portsmouth, 4Peking University *jaat2@cam.ac.uk Local galaxies can be split up into two main classes: gas-rich star-forming galaxies and gas-poor passive galaxies. However, it is still unclear which mechanisms are primarily responsible for quenching star formation in galaxies. Peng et al. [1] analysed stellar metallicities of 26,000 SDSS galaxies and found that galaxies with stellar masses below 1011 M are primarily quenched due to the halting of the supply of cold gas (strangulation). We take this analysis further by making use of the much larger spectroscopic sample in SDSS DR7, analysing the stellar metallicities of 70,000 local galaxies. We study the systematic difference in stellar metallicity between star-forming and passive galaxies (see Fig. 1) to put constraints on the possible quenching mechanisms and timescales. We analyse the light-weighted stellar metallicities of 70,000 SDSS galaxies, obtained using the spectral fitting code FIREFLY [2]. The bimodality in the star formation rate – stellar mass plane is used to divide galaxies into star-forming, green valley and passive. We compare the observed metallicity differences between star-forming and passive galaxies with the predictions of simple closed-box models, to determine whether local galaxies quench primarily through strangulation. Galaxy group catalogues [3] are used to further split the galaxy population into centrals and satellites to investigate the dependence of galaxy quenching on environment. The observed average stellar metallicity difference between star-forming and quenched galaxies is well matched by the predictions of a closed-box model for galaxy evolution (Fig. 2), indicating that local galaxies quench primarily through strangulation. We find that the typical quenching timescale is 2 Gyr and is roughly independent of stellar mass. The sample is further divided into central and satellite galaxies and again we determine the average stellar metallicity difference between star-forming and passive galaxies. We find that the metallicity differences are similar for both central and satellite galaxies, except at intermediate stellar masses, where the stellar metallicity enhancement of centrals is slightly larger (~0.1 dex) than that of satellites. Figure 1. The stellar mass-stellar metallicity relation for star-forming galaxies (blue curve), green valley galaxies (green curve) and passive galaxies (red curve) with 0.02 < z < 0.085. Galaxies are binned in 0.1 dex stellar mass bins and the median stellar metallicity in each bin is plotted. Error bars correspond to the 1σ error on the mean value. Figure 2. The observed average stellar metallicity difference between star-forming and passive galaxies is shown by the black curve. Error bars represent the 1σ error on the metallicity difference. The coloured lines show the stellar metallicity difference predicted by a simple closed-box model at different times after strangulation. We find that the typical quenching timescale is 2 Gyr. Figure 3. The blue curve shows the average stellar metallicity difference between star-forming central and passive central galaxies. The red curve shows the average stellar metallicity difference between star- forming satellite and passive satellite galaxies. We also investigate the dependence of the strangulation signal on the local overdensity (overdensity data taken from [4]). Fig. 4 shows that the metallicity difference for central galaxies depends very weakly on the local overdensity, even over an overdensity range of 2 dex. Satellite galaxies also show a weak dependence on overdensity, except in the smallest overdensity bin, where an excess metallicity difference is observed at
poster
FORK-TO-FARM AGENT-BASED SIMULATION TOOL AUGMENTING BIODIVERSITY IN THE AGRI-FOOD VALUE CHAIN This project has received funding from the European Union’s Horizon 2020 research and innovation programme under grant agreement No 101000499 The BIOVALUE tool is a dynamic, and modular agent-based simulation tool of the agri-food value chain with the aim to analyse experimentally the introduction of marginal traditional varieties starting with novel food dish recipes and processed food products and closing the value chain with an extensive breeding program throughout Europe. The proposed tool will be applied to real-world decision-making to extract and augment the value of so far underutilised and genetically diverse crops in Europe. Furthermore, environmental impacts of future water availability and climate change scenarios will be incorporated modularly into the BIOVALUE tool. The main aim of the BIOVALUE project is to develop a dynamic and customisable tool that will analyse the link among biodiversity, the agri-food value chain, the environment and consumer’s preferences and health. To this end, a demand driven approach (fork-to-farm) is adopted. The ultimate outcome of the research are novel food dish recipes and processed food products from the underutilised, genetically diverse crops resulting from the extensive breeding programme foreseen within the project. BioValue is divided into 11 Work Packages, 3 devoted to the ethical aspects (WP1), the dissemination & exploitation activities (WP10) and the management (WP11), and 8 core research Work Packages (WP2-9), ensuring that scientific excellence is accomplished throughout the project (Figure 2). www.biovalue-project.eu info@biovalue-project.eu # B I O V A L U E © COPYRIGHT AXIA Innovation Figure 2: BioValue’s Research related WPs Figure 1: BioValue’s Concept
poster
How did the Lucayans make use of the Rollin’s Creek South barrier island? Introduction Archaeology is the study of past human activity through examining the materials that they have left behind. The Lucayan Taíno people inhabited the Bahamas for nearly 800 years, spanning roughly from 650 CE to 1550 CE, with a population comparable to the population of the outer islands of the Bahamas today. We were interested in testing two hypotheses. One, that barrier islands were commonly settled by the Lucayans in order to take advantage of the ample resources to be found in the nearby creeks, coppice, and ocean. The other, Dr. William F. Keegan’s settlement pairing idea, that Lucayans often built settlements in pairs less than 1.5 km apart. We based our project area off of these two hypotheses after comparing possible sites with the trends in Lucayan settlements shown by Shaun Sullivan and Dr. William F. Keegan’s data. In total, Shaun Sullivan collected 15 sites on Eleuthera in 1974 for his master’s thesis, and Dr. William F. Keegan has done a lot of work within The Bahamas. When surveying our project area, we look for ecofacts (hole-punch conch, nerite shells) and artifacts. The most significant artifact is Palmetto ware, a pottery specific to the Lucayans that we are unable to replicate. Methods The major materials used for this survey are the GPS and field notebook that is used to document measurements. The method used in this survey was a standard phase 1 surface survey, specifically, a meandering transect, a phase 1 survey designed especially for the Bahamas. No digging or artifact collection was involved. Documentation of objects is broken down into sketching, taking a GPS waypoint, taking photographs, measuring the objects, and recording data in the field notebook. Results • 3 main sites documented: Rollin’s Creek South site at locality 183, Rollin’s Creek Ceremonial site centered on locality 177, and Rollin’s Creek Resource site • A total of 105 pieces of Palmetto Ware was found • 3 main scatters of Palmetto Ware at Rollin’s Creek South and Rollin’s Creek Ceremonial sites • High percentages of specialty Palmetto Ware pottery found • Smooth lithic found at Rollin’s Creek Ceremonial site • Large scatter of ecofacts at Rollin’s Creek Resource site • King Helmet conch found at Rollin’s Creek Resource site Discussion Based on the findings, we located three Lucayan sites, none of which are permanent settlements because: • Findings were concentrated at specific sites and not continuously scattered (Figures 2 and 4). • Small range of pottery thicknesses indicates only one or two pots were at each site vs. a permanent settlement’s large amount of pots (Figures 9 and 10). The evidence suggests that Rollin’s Creek Ceremonial site and Rollin’s Creek South site were used for ceremonial purposes because: • Specialty pottery that was found indicates ceremonial use as it was not used in daily life. • High percentages of specialty pottery at these sites compared to Sullivan’s baseline data (Table 1). The third site, Rollin’s Creek Resource site, was likely a resource-procurement area because: • Large number of shell remains at Rollin’s Creek Resource site (figure 4) • Could have been used by permanent settlement across the creek We supported our barrier island theory because we found Lucayan sites, but found contradictory evidence for the settlement pairing theory because no permanent settlement was found. Recommendations: Our project area holds significance in archaeological studies because it likely had ceremonial uses, while very few other documented Lucayan sites are believed to have ceremonial uses. The location of our project area brings up the question as to whether all permanent settlements had nearby satellite sites used for ceremonial and resource gathering purposes. Recommendations for future studies include completing documentation of our project area, as well as continuing to test the barrier island and settlement pairing hypotheses. Carter B
poster
ROOT I/O data compression improvements for HEP analysis Oksana Shadura (CMS/UNL), Zhe Zhang (CMS/UNL), Brian Paul Bockelman (CMS/Morgridge Institute for Research), Philippe Canal (CMS/FNAL), Danilo Piparo (CMS/CERN) oksana.shadura@cern.ch,bbockelman@morgridge.org Introduction ZSTD - a dictionary-type algorithm (LZ77) with large search window and fast implementations of entropy coding stage, using either very fast Finite State Entropy (tANS) or Huffman coding. [https://github.com/facebook/zstd.git] AVAILABLE IN ROOT 6.20! ZSTD for NanoAOD (compression levels) →Better compression ratio and 2x faster decompression then ZLIB; 6x faster comparing to LZMA; file compressed with ZSTD is only 20 % bigger! NanoAOD 2019 compression ratio comparison *ZSTD-1 means ZSTD with compression level 1 →Better compression ratio then ZLIB Data file with offset arrays →ZSTD has no problems with compression of data that contains the byte offset of each event in the branch data (vs LZ4) CMSSW NanoAOD and MiniAOD (including LZ4) →NanoAOD - using ZSTD could be a better compromise between size of file on a disk and decompression speed for a faster analysis!; →MiniAOD - time spend in decompressing on readback is 15x less vs LZMA, while size of file with ZSTD is only 10% bigger! (thanks to David Lange for MiniAOD measurements) LHCB compression speed vs compression ratio →For the very simple ntuples with a simple structure the best choice could be LZ4: 10x time faster read speed TTree::kOnlyFlushAtCluster: faster decompression →TTrees can be forced to only create new baskets at event cluster boundaries, it simplifies file layout and I/O at the cost of memory (NanoAOD 2017 size difference was 3.6 %). Recommended for simple file formats such as ntuples but not for more complex data types. tree->SetBit(TTree::kOnlyFlushAtCluster); Acknowledgements This work was supported by the National Science Foundation under Grant ACI-1450323.
poster
Neben dem Angebot von Texten bietet RIDE durch die strukturelle und semantische Codierung der Inhalte auch vermehrt datengestützte Zugänge. Damit wird der Blick auf die Gesamtheit der Rezensionen und die rezensierten Gegenstände gefördert. Die herausgeberische Arbeit wird so vermehrt zur informationstechnologischen Gedächtnisarbeit. RIDE als DATEN RIDE als CODE RIDE als WEBSITE GitHub: https://github.com/i-d-e/ride Zenodo: https://zenodo.org/record/5081646 Das IDE zeichnet für die digitale Infrastruktur und den Publikationsworkflow von RIDE selbst verantwortlich. Die Publikationsinfrastruktur besteht aus einem Zusammenspiel von eXist-db und Wordpress. RIDE - was ist das? GitHub: https://github.com/i-d-e/ride-scripts URL: https://ride.i-d-e.de Auf der Website der Zeitschrift erhalten Leserinnen und Rezensentinnen Informationen zu den Zielen, den Methoden und Inhalten von RIDE. Jede Rezension erscheint als ein Artikel, der im HTML- Format oder als PDF gelesen werden kann. Ergänzt wird jeder Rezensions- artikel durch das “Factsheet”, das die Ergebnisse eines Fragebogens zur rezenzierten Ressource enthält. Die Sicht auf einzelne Rezensionsartikel wird ergänzt durch übergreifende Darstellungen, insbesondere Listen aller Ressourcen, Autorinnen und Diagramme mit Übersichten über die Fragebogen-Ergebnisse. Informationstechnologische Gedächtnisarbeit in der Rezensionszeitschrift RIDE Datensätze aller RIDE-Bände werden in einem GitHub- Repositorium verwaltet und im Forschungsdatenrepositorium Zenodo veröffentlicht, in welchem sie versioniert und über die Zuweisung einer DOI zitierfähig gemacht werden. RIDE (A Review Journal for Digital Editions and Resources) ist eine Open Access- Zeitschrift für Rezensionen über digitale wissenschaftliche Editionen, Text- sammlungen sowie Tools und Forschungsumgebungen für das digitale Edieren. Die Zeitschrift bietet ein Forum, in dem Expertinnen digitale Ressourcen diskutieren können, um aktuelle Praktiken einzuordnen und zu bewerten und künftige Entwicklungen mit zu gestalten. Dabei steht neben inhaltlichen Gesichtspunkten vor allem eine methodische Begutachtung der digitalen Editionen, Textsammlungen und Tools im Vordergrund. Um die Rezensentinnen zu unterstützen, werden Kriterienkataloge angeboten, die auf wichtige zu begutachtende Aspekte eingehen. ENJOY THE RIDE! RIDE erscheint seit 2014 und wird vom Institut für Dokumentologie und Editorik (IDE) herausgegeben. Bisher sind 75 Rezensionen erschienen. In RIDE nehmen Rezensionen die Form von Forschungsartikeln an, die einen Peer Review-Prozess durchlaufen. Begleitet werden die Texte durch “Factsheets”, in denen zentrale Daten zu den digitalen Ressourcen strukturiert präsentiert werden. Als Zeitschrift bietet RIDE also nicht nur Lesetexte, sondern basiert auf Daten zu den Rezensionen und den rezensierten Gegenständen. Diese werden auf der RIDE-Website in vielfältiger Form präsentiert, aber auch direkt zum Download angeboten. Verarbeitet werden die Daten durch Code, der einen wesentlichen Teil der RIDE-Workflows darstellt. Die Routinen zur Verarbeitung der Daten werden auf GitHub publiziert, darunter Skripte zur Generierung der langzeit- archivierbaren PDFs, für den Export der Metadaten über eine OAI-PMH-Schnittstelle, für die Directory of Open Access Journals (DOAJ) und für die Erstellung von Wordclouds. RIDE ist ein datengestütztes Journal. Die Rezensionen und Metadaten sind in XML/TEI kodiert, auf Basis eines TEI- Schemas in ODD (“One Document Does it all”). Außerdem wird jede Rezension als PDF archiviert. Ulrike Henny-Krahmer, Universität Rostock Frederike Neuber, Berlin Brandenburgische Akademie der Wissenschaften Martina Scholger, Universität Graz https://ride.i-d-e.de
poster
Response of Chelidonium majus in vitro cultures and their alkaloid content under heavy metal stress Iva Doycheva1, Stefan Philipov2 1Institute of Biodiversity and Ecosystem Research, Bulgarian Academy of Sciences 2Institute of Organic Chemistry with Centre of Phytochemistry, Bulgarian Academy of Sciences Introduction Heavy metal contamination is becoming an increasing environmental problem worldwide. Heavy metals (HM) retard plant growth, yield and affect the secondary metabolism which results in fluctuations in plant metabolites, such as the alkaloids. The aim of the study was to investigate the influence of HM on the growth of in vitro cultivated plants of Chelidonium majus L. (Fig. 1) and their alkaloid content. Results and discussion Plants cultivated at the lower concentration of Cd2+ formed roots and developed normally with a gradual reduction in their growth and rooting as the concentration of the supplemented HM increased (Fig. 2). There were not any differences in the development of the plants cultivated on Cd2+according to their type of origin (Fig. 2; 3). Epinasty and browning of the leaves were observed among the plants cultivated on the medium supplemented with Pb2+. These side effects got worse as the concentration of the HM increased, and were also accompanied by a strong reduction in the rooting (Fig. 4). The plants which were obtained from seeds developed normally and did not show any of the signs of the ones which were propagated through AS (Fig. 3; 4). None of the plants cultivated on a medium supplemented with Zn2+ survived, regardless of their origin and HM concentration. A phytochemical study showed that the plants cultivated at the highest concentrations of Cd2+ and Pb2+ and also obtained through AS had the highest RAM. However, the plants cultivated at the same concentrations but obtained from seeds had lower RAM similar to that of the plants cultivated at lower concentrations of HM and propagated through AS (Fig. 5). Materials and methods The plants were cultivated on a medium B5 with double the amount of macrosalts (2B5), 0.5 g/l activated charcoal and supplemented with different concentrations of HM – Cd2+ (0.5; 1 mg/l); Pb2+ (100; 250 mg/l); Zn2+ (100; 250 mg/l). Plants were obtained in vitro from seeds or were propagated through adventitious shooting (AS). The raw alkaloid mixtures (RAM) were obtained by usual chemical procedure. For contacts: idoycheva@gmail.com Conclusions In conclusion, Ch. majus was able to cope with the influence of Cd2+ and Pb2+ in the culture medium, but not with that of Zn2+. Quantity of RAM was influenced by the HM concentration and the origin of the plants. Acknowledgements: This research was supported by the Bulgarian National Science Fund, Bulgarian Ministry of Education and Science (Project КП-06-М26/4 from 01.12.2018). Fig. 1. Chelidonium majus L. Fig. 5. RAM (%) in native plant sample and in vitro cultivated plant samples obtained through AS or seed germination. 0,92 1,54 1,51 1,39 3,56 3,78 1,94 1,01 0,81 Reference sample 2B5 - Control AS 2B5 + 0.5 mg/l Cd (AS) 2B5 + 100 mg/l Pb (AS) 2B5 + 1.0 mg/l Cd (AS) 2B5 + 250 mg/l Pb (AS) 2B5 - Control Seeds 2B5 + 1.0 mg/l Cd (S) 2B5 + 250 mg/l Pb (S) Plant material from native plants Plants material from plants propagated through adventitious shooting Plant material from plants obtained from in vitro germinated seeds BalkanBio’2021 Fig. 2. Plants propagated through AS and cultured on 2B5 supplemented with Cd2+. Fig. 4. Plants obtained through AS and cultured on 2B5 supplemented with different concentrations of Pb2+. Fig. 3. Plants obtained from seeds and cultured on 2B5 supplemented with Cd2+ or Pb2+.
poster
Wikibase ❖open source software ❖created by WikiMedia Foundation ❖user friendly ❖rich in collaborative tools ❖using RDF ❖flexible data model ❖multilingual Reasons for WikiHum ❖Decentralisation of historical data ❖Lack of reliable and Persistent Identifiers ❖Mismatch between reference databases (VIAF, GND, Wikidata) and the needs of historians ❖Need for harmonisation of historical data Process of implementing data into WikiHum Label_pl Label_en Description_pl Description_en Królestwo Polskie (1370-1568) The Kingdom of Poland (1370-1568) terytorium Królestwa Polskiego territory of the Kingdom of Poland Tab.1 Administrative system: label and description Label_en P value The Kingdom of Poland (1370-1568) instance of administrative system The Kingdom of Poland (1370-1568) Starts at 1370-01-01 The Kingdom of Poland (1370-1568) Ends at 1568-12-31 The Kingdom of Poland (1370-1568) Aen The Crown of Poland The Kingdom of Poland (1370-1568) legal basis reformy ustrojowe XIV w. The Kingdom of Poland (1370-1568) legal basis statuty wiślicko-piotrowskie (1356-1362 r.) The Kingdom of Poland (1370-1568) subclass of secular administrative system The Kingdom of Poland (1370-1568) Wikidata ID Q1649871 The Kingdom of Poland (1370-1568) ontohgis database id https://onto.kul.pl/ontohgis/administrative_sys tem_44 Tab.2 Administrative system: statements Source: https://ontohgis.pl/ final view in WikiHUM “One Base to rule them all, one Base to find them, One Base to bring them all, and in the Wiki bind them” Reference database for Polish historical data Content ❖Item metadata ❖DARIAH.Lab ID - Handle.net ❖External ID (e.g. Wikidata, VIAF, etc.) ❖Standards and Ontologies (CIDOC CRM, OntoHGIS, SKOS) Biographical data ❖Polish Biographical Dictionary ❖Polish National Library Authority File ❖Urzędnicy Dawnej Rzeczypospolitej. Spisy [Officials of the Old Commonwealth. Lists] ❖etc. Geographical data ❖Historical Atlas of Poland ❖Historical-Geographical Dictionary of the Polish Lands in the Middle Ages ❖Polish National Register of Boundaries ❖Polish National Register of Geographical Names Wikibase as an environment for harmonisation of data about past: the example of WikiHum Authors Adam Zapała azapala@ihpan.edu.pl Tomasz Królik tkrolik@ihpan.edu.pl The Tadeusz Manteuffel Institute of History Polish Academy of Sciences COMING SOON wikihum.lab.dariah.pl
poster
Delivering Fine-grained Environmental Spatial Data in HPC and Cloud Converged Infrastructure PI: Michela Taufer*, Co-PI: Rodrigo Vargas↟ Project Contributors: Paula Olaya*, Camila Roa*, Jakob Luettgau*, Ricardo Llamas↟, Sophia Wen‡, I-Hsin Chung‡, Seetharami Seelam‡, Yoonho Park‡, and Jay Lofstead† *University of Tennessee, Knoxville,↟University of Delaware, ‡IBM, †Sandia National Laboratory Generating Fine-grained Environmental Spatial Data ●Environmental spatial data (i.e., terrain parameters and soil moisture) provide information about phenomena on the surface of the earth ●These data are fundamental in applications such as precision forestry and agriculture, identifying biodiversity loss, hydrology for landscape ecology, understanding vegetation changes, wildfire detection, and spotting at-risk ecosystems1 [1] Jennifer Bernstein, Karen Kemp; The Role of Spatial Science in Environmental Case Studies: A Special Collection from the University of Southern California. Case Studies in the Environment 1 January 2020; 4 (1): 1–5. Generating environmental spatial data at high-resolution is computationally expensive, hindering scientific communities’ ability to access the data Digital Elevation Model Satellite Soil Moisture Data Generation Terrain Parameters Data Transformation Training Data Testing Data ML- Training ML- Inference ML-Model Soil Moisture Predictions Analysis/ Visual Geographic statistics/ maps Terrain Parameter Generation (GEOtiled) Data Transformation ML-Prediction Soil Moisture We develop a workflow called SOMOSPIE (SOil MOisture SPatial Inference Engine) that generates fine-grained terrain parameters and soil moisture Analysis/ Visualization HPC and Cloud Convergence ●Our workflow comprises four stages: a terrain parameter generation, a data transformation, an ML-based prediction, and Analysis/Visualization ●Each of the stages leverages advanced computational techniques to transform data and generate ready-for-use environmental spatial data Our solution scales performance and enables scientific communities to access our end-to-end workflow and data on converged HPC and Cloud infrastructure Soil Moisture Maps @ <1 km Elevation models @ <1 km Terrain P. Generation Data Transformation ML Predict Analysis Visualization Satellite Soil Moisture @ 27 km Precision agriculture Environm. sciences Wildfire Detection SOMOSPIE: SOil MOisture SPatial Inference Engine Train1 Hybrid HPC-Cloud Cluster Train2 Train3 TrainN Inference Inference Inference Inference def train(): … def inference(model): … dsl.pipeline(ml-model): for i in N_years: model=train() inference(model) TektonCompiler(). compile(pipeline, “pipeline.yaml”) Inference Inference Inference Inference Inference Inference Inference Inference Inference Inference … … Predicted Soil Moisture1 Predicted Soil Moisture2 Predicted Soil Moisture3 Predicted Soil MoistureN … Open Data Hub Pipelines We use Open Data Hub Pipelines to orchestrate the execution of our workflow in cloud-native clusters (i.e., Kubernetes and OpenShift) with Cloud Object Storage The authors acknowledge the support of the National Science Foundation through the awards 1841758, 2103845, 2138811, and 1941443 Reproducibility efforts Paper Github Cloud-HPC Integration SOMOSPIE’s workflow Paper Github Trustworthiness and Reproducibility Our solution ensures trustworthiness and reproducibility in the data, software, and environment, wrapped in the scientific discovery loop ●We create a fine-grained containerized environment that automatically provides provenance information and seamlessly attaches it to the workflow components ●The containerization supports easy retrieval for reusability, reproducibility, and portability of the workflow
poster
PHOTOVOLTAIK-INSTITUT BERLIN Mechanische Belastungsuntersuchungen Testsequenz und Probenaufbau PHOTOVOLTAIK-INSTITUT BERLIN Verschattungstoleranz UNTERSUCHUNGEN AN GESCHINDELTEN PHOTOVOLTAIK MODULEN Stefan Wendlandt PI Photovoltaik-Institut Berlin AG, Wrangelstr. 100, D-10997 Berlin, Germany, Phone: +49 (30) 814 52 64-205, mail: wendlandt@pi-berlin.com 2017.03.08-V1.0 © PI Berlin Zusammenfassung Einleitung In den letzten Jahren rückte das Design von Photovoltaikmodulen und deren Effizienzsteigerung immer mehr in den Forschungsschwerpunkt. Ein wichtiger Parameter ist dabei die Optimierung der Zellkontaktierung, da sie den Serienwiderstand und die Leistungsabgabe bei höheren Bestrahlungsstärken beeinflusst. Ein neuartiger Verbindungsansatz ist das Schindeln von Solarzellen. Geschindelte Module können hohe CTM-Werte (Cell-to-Module) erreichen. Diese Art der Verbindung ermöglicht gleichwohl mehr Möglichkeiten beim Design vom elektrischen Layout. Der Schlüssel beim Schindeln liegt jedoch in den Eigenschaften der elektrisch-leitfähigen Klebstoffe (ECA). Er muss gleichzeitig leitfähig und mechanisch flexibel sein. In dieser Arbeit haben wir die Langzeitstabilität von kommerziellen geschindelten PV-Modulen untersucht. Dafür haben wir eine spezielle Testsequenz erstellt. Unsere Ergebnisse konzentrieren sich auf die mechanische Stabilität durch Anwendung verschiedener Stresstests sowie die elektrische Leistung bei unterschiedlichen Verschattungen. Der Aufsatz endet mit einer Schwachstellenanalyse der derzeit verfügbaren geschindelten PV-Panels mit unterschiedlichen elektrischen Layouts. Wir haben verschiedene handelsübliche Schindelmodule charakterisiert und beansprucht. Dabei haben wir uns auf die mechanische Stabilität konzentriert. Wir haben einen klaren Zusammenhang zwischen der Belastung bei kurzen und langen geschindelten Zellstrings beobachtet. Während der Biegetest einen vernachlässigbaren mechanischen Einfluss auf die Module hat, hat der ML-Test einen großen Einfluss auf die Moduldesigns mit langen Zellstrings. Dort ist der Krafteinfluss aufgrund der Lage der gebrochenen Zellen deutlich sichtbar. Bei den Modulen mit den kurzen Strings ist quasi kein Einfluss sichtbar. Dies kann mit der Kraftabsorption des ECA und des Verkapselungsmaterials aufgrund seiner Elastizität und der kurzen Kantenlänge der Zellen in Richtung der Krafteinwirkung beschrieben werden. Wir möchten daher darauf hinweisen, dass der begrenzende Spannungsfaktor die Zugfestigkeit der ECA und des Verkapselungsmaterials ist. Weiterhin zeigte die Verschaltung von Brand „E“ (kurze-parallele Stränge) die beste Verschattungstoleranz und damit die höchste elektrische relative Leistungsabgabe. Dies kann durch die Zellblockaufteilung sowie die geringere Anzahl von Solarzellen pro Bypass-Dioden begründet werden. ACKNOWLEDGEMENTS: Dieses Projekt wurde mit Mitteln aus dem Horizon2020-Programm der Europäischen Union für Forschung, technologische Entwicklung und Demonstration unter der Grant Agreement Nr. 857793 (HighLite) gefördert. Abbildung 1: In unserem ersten Versuch wollten wir wissen, ob und welchen Einfluss die unterschiedlichen Verschaltungen auf die elektrische Leistung bei Verschattung haben. Dafür haben wir uns fünf unterschiedliche Verschattungsszenarien überlegt. Abbildung 2: Relative Leistungsverluste pro Verschattungsszenario darstellt über das Verhältnis verschattete Fläche zu Modulfläche gezeigt für vier der fünf Szenarien. Abbildung 5: EL-Bilder (@Isc) aller Hersteller nach den jeweiligen Tests. Auffällig ist das Modul „D“ schon initial Kontaktierungsprobleme aufzeigt. Weiterhin ist erkennbar, dass der Bending-Test keinen relevanten Schaden auf die Module ausübt. Anders hingegen beim mechanischen Belastungstest, hier sind bei den Modulen mit vertikalen Strängen („A“ und „D“), starke Zellbrüche ersichtlich. Es wird vermutet, dass die Zellen quasi zerrissen wurden, da anders als bei Brand „B“ und „C“ das ECA die Scherkräfte absorbieren kann. Mit anderen Worten
poster
Autor: Zenaldo de Almeida Rodrigues Mestrando em Administração – Universidade Positivo Orientador: Prof. Dr. Luiz Pinheiro Linha de Pesquisa: Inovação, Estratégia e Sustentabilidade INTELIGÊNCIA ARTIFICIAL (IA) E TRABALHO: UMA ANÁLISE SOBRE OS IMPACTOS CAUSADOS PELA UTILIZAÇÃO DA INTELIGÊNCIA ARTIFICIAL NO TRABALHO DAS PESSOAS INTRODUÇÃO Nos últimos anos temos visto uma grande profusão de novas tecnologias baseadas em Inteligência Artificial (IA). Segundo Matheus (2020), nove áreas do fazer humano serão afetadas por esta nova tecnologia: Estratégia, Machine Learning (aprendizado das máquinas), Speech Voice (escrita e voz humana), Imagem Recognition (reconhecimento facial e de imagens no geral), Robotic (robótica e seus processos), Neural Networrks (redes neurais), Sentiment Analysis (análises dos sentimentos), Natural Language (estudo da linguagem natural humana), Deep Learning (aprendizado profundo e constante das máquinas). Entender os impactos que esta nova tecnologia e seus campos de atuação podem provocar na forma com que as pessoas realizam suas tarefas, passou ser um objeto de estudo significativo para a humanidade. OBJETIVOS / PERGUNTA DE PESQUISA Objetivo: Investigar a influência da inteligência artificial (IA) no trabalho e seus possíveis impactos na sociedade como um todo; RQ1: Quais as possíveis tarefas que serão afetadas pela utilização de IA e como ficará o trabalho executado por humanos? RQ2: Haverá migração de mão de obra? RQ3: Como podemos nos preparar para uma possível migração de mão de obra? REVISÃO DA LITERATURA A OCDE e a ONU, apresentam pontos comuns de preocupação com a influência da utilização da tecnologia de Inteligência Artificial. Para isso juntas desenvolveram quatro pilares de orientação aos países, com a finalidade de direcionar a utilização da IA: 1º O incentivo para pesquisa e desenvolvimento de IA , aplicadas a diferentes campos de negócios (organizações, empresas e governos); 2º A definição de princípios éticos e perspectivas regulatórias que passem pelo problema da governança de dados, parâmetros regulatórios, transparência e accountability . 3º Criar uma abordagem para a governança de tecnologias emergentes; 4º Criar uma abordagem para a transformações do trabalho, do emprego e construção de capacidades para lidar com as mudanças que já emergem com a Inteligência Artificial; Acemoglu e Restrepo (2019), apresentam que “estamos longe de uma compreensão satisfatória em como a automação em geral e a IA, impactam o mercado de trabalho e sua produtividade”. Sendo assim um objeto de estudo que nos apresenta um grande espectro de estudo a frente; Zeira (1998), nos coloca uma dicotomia de pensamentos:  O primeiro pensamento é uma visão mais APOCALÍPTICA do tema. Onde os avanços com a IA irá acabar com o trabalho humano;  O segundo reflete a visão dos economistas, que diz que os avanços tecnológicos no passado acabaram aumentando a demanda por mão de obra e salários; MÉTODO E ANÁLISE DOS DADOS Quadro 1: Metodologia CONSIDERAÇÕES FINAIS Lee (2019), em seu livro INTELIGÊNCIA ARTIFICIAL – Como os robôs estão mudando o mundo... nos traz uma perspectiva de como esta evolução da IA aconteceu. Sendo a partir de 2012 quando Geoffrey Hinton, um pesquisador da universidade de Toronto no Canadá, desenvolveu redes neurais capaz de levar máquinas a um aprendizado profundo é que a IA passou a ter uma perspectiva de utilização mais efetiva para a humanidade. Podendo a partir deste momento desempenhar tarefas hora antes destinada somente a humanos. Iansiti e Lakhani (2021), trazem uma abordagem voltada para as empresas e como a IA poderá afetar efetivamente a mão de obra nela existente, dando desta maneira um direcionamento específico para negócios, trazendo a luz perspectivas do que poderemos enfrentar em um cenário futuro; A partir do entendimento do nascimento e desenvolvimento da Inteligência Artificial, seus conceitos e aplicações. Podemos entender a dimensão dos impactos e influências da utilização desta no
poster
Gemeinsame Infrastrukturen und Services zum Forschungsdatenmanagement in, aus und für Baden-Württemberg 1. Einleitung und Grundlage 2. Drei Fallbeispiele aus bwFDM Dr. Sophie G. Habinger1 0000-0002-5513-954X Maximilian Heber1 0000-0003-3399-7532 Dr. Philipp Kling2 0000-0003-0104-4905 1Kommunikations-, Informations-, Medienzentrum (KIM) der Universität Konstanz 2Universitätsrechenzentrum der Universität Heidelberg In Baden-Württemberg gibt es bereits eine Vielzahl an Landesdiensten und -services, die Bedarfe aus dem Forschungsdatenmanagement adressieren. Informieren, Austauschen, Vernetzen Welche FDM- Infrastrukturen gibt es in BW? Zugang zu Rechen- und Speicher- ressourcen Speicherung großer Mengen an kalten Forschungsdaten kollaboratives Arbeiten mit kleinen Datenmengen bwVisu Speicherung großer Mengen an heißen Forschungsdaten bwSFS LSDF Online Storage SDS@hd bwData Archive bwSync&Share bwFDM Helpdesks bwSupport Virtuelle Maschinen und Cloud Computing bwCloud bwUniCluster Speichern | Rechnen | Speichern und Rechnen | Kollaboration | Helpdesks/Supportservices Emilia Mikautsch1 0009-0005-1758-1790CC BY 4.0 –https://creativecommons.org/licenses/by/4.0/(Logos ausgenommen) Projektpartner der Landesinitiative: Gefördert vom: bwHelpdesk zum FDM • Rückgriff auf/Verwendung von bwSupport Portal geplant • Landesweite Helpdesk- Infrastruktur ermöglicht einfache Kollaboration der gehosteten Beratungsangebote zu den einzelnen Services ? --- --- Vernetzung Schaffung von Rahmen- bedingungen zum Austausch: • überregional: E-Science-Tage • BW-intern: AK FDM, bwFDM- Forum und Austausch mit bwHPC E-Science-Tage AK FDM bwFDM-Forum forschungsdaten.info • Technisch & administrativ: bwFDM • Inhaltlich: Redaktionsteam aus deutschsprachigem Raum forschungsdaten.info FORTH-BW fdm@DHBW
poster
Michael Kleineberg1 , Miriam Kötter2 , Jasmin Schmitz3 , Paul Schultze-Motel4 , Katja Wermbter5 1: Distributed Network for Publishing Services, Freie Universität Berlin, Universitätsbibliothek 2: Landesinitiative openaccess.nrw / Universität Duisburg-Essen, Universitätsbibliothek 3: ZB MED - Informationszentrum Lebenswissenschaften, PUBLISSO Publikationsberatung 4: Helmholtz-Gemeinschaft / open-access.network 5: FID BAUdigital, TU Braunschweig, Universitätsbibliothek Wer wir sind • Best Practices zu: Aufbau eines Open-Access-Beratungsangebots Organisation eines Helpdesks Dokumentation der Kundenkommunikation Tools (z. B. Ticketsystem) Verteilung von Ressourcen interner und externer Kommunikation • Erstellung einer gemeinsamen Materialsammlung Mitmachen Ebenfalls auf der Agenda Sie betreiben oder planen selbst einen Open-Access-Helpdesk und möchten sich darüber austauschen? Die Fokusgruppe freut sich über neue Mitglieder! Weitere Informationen finden Sie hier: Was wir tun Liste der Helpdesks und Beratungs- angebote in den teilnehmenden Einrichtungen Sammlung der häufigsten Anfragen (z.B. zu CC-Lizenzen etc.) Kurzportraits der teilnehmenden Einrichtungen in Form von Vorträgen bei jedem Treffen Vernetzung und Kollaboration Übersicht der Helpdesks FAQ- Sammlung Vorstellung der Helpdesks Austausch über die Beratungsangebote und gegenseitige Unterstützung bei Herausforderungen Die Fokusgruppe entstand als Resultat eines Workshops des BMBF-Projekts „open-access. network“ zum Erfahrungsaustausch über den Aufbau und Betrieb von Open-Access- Helpdesks im November 2021. Die Gruppe umfasst derzeit 26 Mitglieder aus 20 verschiedenen Einrichtungen und trifft sich alle drei Monate. Für ihre Arbeit hat die Fokusgruppe zur internen Nutzung im Wiki der Fokusgruppen von open-access.network eine Übersicht von Open-Access-Helpdesks und -Beratungsange- boten angelegt. Bisher sind 21 Einträge ent- halten. In zukünftigen Treffen soll die Liste weiter ergänzt werden. Dieses Werk ist lizenziert unter einer Creative Commons Namensnennung 4.0 International Lizenz. https://open-access.network/vernetzen/digitale- fokusgruppen/fokusgruppe-open-access-helpdesks Ü b e rb lic k ü b e r d ie O A -B e ra tu n g s a n g e b o te u n d -H e lp d e s k s , d ie d e rz e it in d e r F o k u s g ru p p e o rg a n is ie rt s in d A n z a h l d e r E in trä g e U n iv e rs itä t F H P ro je k t L a n d e s v e rn e tz u n g s s te lle S o n s tig e E in ric h tu n g e n 0 2 4 6 8 1 0 1 2 1 4 1 6 Diese Veröffentlichung ist im Rahmen einer Fokusgruppe des Projekts open-access.network entstanden. Für den Inhalt sind allein die Autor*innen verantwortlich.
poster
Follow the Lithium: The Correlation Between Li-bearing Molecules with Age, Mass, and Gravity in Brown Dwarfs Ehsan Gharib-Nezhad 1, Mark S. Marley 1, Natasha E. Batalha 1, Channon Visscher 2,3 (1) NASA Ames Research Center, Moffett Field, CA, USA (2) Dordt University, Sioux Center, IA, USA (3) Space Science Institute, Boulder, CO, USA Conclusion and Future Work The right figure is the Volume Mixing Ratios (VMR) of Li-bearing species as a function of temperature. Atomic Li and LiH are dominant species in (ultra-) hot atmospheres with Teff in ~1800-3000~K (e.g., M/L- type dwarfs). In warm atmospheres with Teff~900-1600~K (e.g., L/T-type dwarfs), LiOH, LiCl, and LiF are the most stable Li-bearing species. Historically the Li test, searching for the atomic Li line at 670 nm, has been used to distinguish early L-type brown dwarfs from stars, since Li is consumed by fusion processes for masses above 70 M_Jup (Baraffe2015, Charier1996). To better understand the presence and abundance of the different Li-bearing species in the temperature-pressure space, the Li phase diagram is represented in the left figure. In this plot, each species is shown with different colors. At each area, dashed and dotted contour lines represent 50% and 20% of that species with respect to the total protosolar lithium abundance (i.e., ~3.5x10-9) in the atmosphere. The solid black curves are the temperature-pressure (TP-) profiles of brown dwarfs from the Sonora-Bobcat cloud model created by Marley et al. (2021). The critical question is whether different Li species could serve as a tool to confirm massive ultracool objects. Li Chemistry in Brown Dwarfs Challenges with Accurate Determination of T-Y Dwarf Mass We investigate the impact of Li species on the synthetic thermal spectra of candidate T/L/M dwarfs using both cloud (Marley2021) and cloudless Sonora model (Marley2018) for Teff of 500- 2400~K and logg=5.0. The presence and the strength of their spectral features on synthetic flux spectra are assessed for different effective temperatures and surface gravities to see if they can productively be used as a high mass indicator in brown dwarfs. Ø The presence (FON) and absence (FOFF) of Li has the highest ([(FOFF) - (FON)] / FOFF) ratio, which is ~1 at the line center for Teff =2400 K, and it reduces to 0.2 for 500K. Ø LiF absorption feature is dominant at 10.5-12.5 µm. As it is predicted by thermochemical simulations, LiF mostly forms at 800-1200 K and competes with LiCl and LiOH molecules. The LiF ([(FOFF) - (FON)] / FOFF) ratio is ~10-4 for T-L type dwarfs with Teff of 900-1600 K. Ø In the range 1600K< Teff <2400K Li should be present as LiCl for masses below 70 MJup. The lack of LiCl at 14.5-18.5 µm would be an indicator of a larger mass. References and Contact Information Using LiF and LiCl to Disentangle this Age-Mass-Gravity Degeneracy Accurate calculation of the mass of ultracool dwarfs has been a serious challenge in understanding brown dwarfs. This uncertainty comes from the degeneracy in age, mass, and gravity parameters. Examples of T-L dwarfs with high uncertainty in their mass provided in the following table. Baraffe, I., Homeier, D., Allard, F., & Chabrier, G. 2015,A&A, 577, A42. Chabrier, G., Baraffe, I., & Plez, B. 1996, ApJ, 459. Marley, M. S., Saumon, D., Morley, C., et al. 2021 (Submitted) Marley, M. S., Saumon, D., Guillot, T., et al. 1996, Science,272, 1919. The traditional Li test can distinguish between low mass stars and brown dwarfs with similar spectral types. Although atomic Li still appears in model spectra down to Teff as low as 500K, this spectral region is exceedingly dark and difficult to measure. Instead, since LiF and LiCl features are located in regions where brown dwarfs are brighter, they could in principle be used as mass indicators at these lower effective temperatures. Ehsan.gharibnezhad@nasa.gov ehsangharibnezhad.github.io @ExoEhsan
poster
References Chenu, C., Angers, D.A., Barré, P., Derrien, D., Arrouays, D., Balesdent, J., 2019. Increasing organic stocks in agricultural soils: Knowledge gaps and potential innovations. Soil and Tillage Research 188, 41-52. Elser, J., Bennett, E., 2011. Phosphorus cycle: a broken biogeochemical cycle. Nature 478, 29. Erisman, J.W., Sutton, M.A., Galloway, J., Klimont, Z., Winiwarter, W., 2008. How a century of ammonia synthesis changed the world. Nature Geoscience 1, 636. Hobley, E., Baldock, J., Hua, Q., Wilson, B., 2017. Land‐use contrasts reveal instability of subsoil organic carbon. Global change biology 23, 955-965. Hobley, E., Steffens, M., Bauke, S.L., Kögel-Knabner, I., 2018. Hotspots of soil organic carbon storage revealed by laboratory hyperspectral imaging. Scientific reports 8, 13900. Towards more efficient carbon, nitrogen and phosphorus cycling in European agricultural soils: Circular Agronomics (CA) program Yaser Ostovari1, Jan Willem Van Groenigen2, Rachel Creamer2, Julien Guigue1 ,Eleanor Hobley1, Laura Ferron2, Henk Martens2, Emily Overtuf1, Anke Neumeier1, Andreas Muskolus3, Paolo Mantovi4, Francesc Domingo Olivé5, Thomas Guggenberger6, Marek Holba7, Ingrid Kögel-Knabner1, Alix Vidal1 1TU München, Lehrstuhl für Bodenkunde, Freising, Germany, 2Wageningen University, Soil Biology Group, Wageningen, The Netherlands, 3Landwirtschaftliche Versuchsstation Berge, Berlin, Germany, 4Fondazione CRPA Studi Ricerche, Reggio Emilia, Italy, 5IRTA Mas Badia, Agricultural Experimental Station MasBadia, La Tallada d’Emporda, Spain, 6Höhere Bundeslehr- und Forschungsanstalt für Landwirtschaft Raumberg-Gumpenstein, Irdning, Austria, 7ASIO SPOL SRO, Brno, Czech Republic 1) CATALONIA, Spain (leader: IRTA) 2) BRANDENBURG, Germany (leader: IASP) 3) LUNGAU, Austria (leader: AREC) 4) EMILIA-ROMAGNA, Italy (leader: FCSR) 5) GELDERLAND, The Netherlands(leader: WUR) 6) SOUTH MORAVIA, Czech Republic (leader: ASIO) 1 2 3 4 5 6 Case Study Sites Material and methods Organic and inorganic amendments Different organic and inorganic amendments according to the case study. E.g., mineral fertilizer, pig slurry (solid or liquid fraction), manure, whey, biogas digestate, nitrogen-depleted digestate. Soil sampling for every case study Soil sampling to a depth of 1 m using a hydraulic corer. One un-disturbed half of the core for hyperspectral imaging The other half of the core for laboratory analyses. Five depths: 0-10 cm,10-20 cm, 20-40 cm, 40-80 cm and 80-100 cm. Laboratory analyses Classical bulk chemical analyses. E.g., bulk density, pH, CEC, texture, total and available C, N and P State-of the-art imaging technique: hyperspectral camera to reveal horspots of C and N storage in the soil profile (Hobley et al., 2018). A focus on hyperspectral soil analysis Introduction Mineral fertilizers: low recovery efficiency by crops. E.g., 30-50 % for nitrogen (N) and 45 % for phosphorus (P) (Tilman et al., 2002). Loss of nutrients into the environment : severe negative influence on soil (e.g. unbalanced N cycle), water (e.g., eutrophisation) and air (e.g., greenhouse gas emissions) (Elser and Bennett, 2011; Erisman et al., 2008). Less than 2% soil organic carbon (SOC) in 45% of European soils (Jones et al., 2012) and projected decline of agricultural SOC stocks of up to 24 % (Wiesmeier et al., 2016). Soil carbons stocks: not only topsoils, but also subsoils are reactive to agricultural change (Hobley et al., 2017). Impacts of organic amendments (e.g., digestate, manure) versus traditional mineral fertilizers on SOC contents and stocks are uncertain at the long-term (e.g., Chenu et al., 2019). Within the project Circular Agronomics - production of new organic fertilizers: residues (manure, digestate) treated with vacuum degasification for nitrogen depletion of residues and recovery of fertilizer. One of the objectives of Circular Agronomics Exploring medium and long-term effect of new organic fertilizers versus classical organic and mineral f
poster
Global short wave solar radiation monitoring system GLOB Sebastian Dominik Sikora1), Anna Sjöblom Coulson1), Arthur Garreau1), Alex Klein-Paste2), Aleksey Shestov1), 1) Arctic Technology Department, UNIS The University Center in Svalbard, Svalbard, Norway 2)Department of Civil and Environmental Engineering, NTNU Norwegian University of Science and Technology, Trondheim, Norway The Arctic region presents numerous challenges for human presence, ranging from constructing on thawing permafrost to ensuring adequate food supplies. However, one of the most pressing challenges revolves around energy sources, encompassing both heat and electricity supply. Arctic settlements are typically considered off-grid systems, necessitating the development of systems of locally generated energy that are both reliable and sustainable. The goal of GLOB is to measure solar irradiation on as many as possible array planes to verify recorded values with modeled. Due to technical and costs limitations a sphere with 26 faces called rhombicuboctahedron has been pointed as most suitable for this project. The rhombicuboctahedron is a polyhedron comprised of 8 triangles and 18 squares. For this project, elements with edges measuring 10 cm for both squares and triangles have been constructed. The bottom-facing square element isn't utilized for irradiation data collection. This is because a stainless steel pipe has been employed as a support leg, passing through the bottom square and reaching the top face, which is horizontally oriented. All components required for building the rhombicuboctahedron have been meticulously designed using CAD software, and the corresponding files compatible with most 3D printers have been attached for your convenience. The elements created in CAD software used to assemble GLOB system SENSORS AND OTHER ELECTRONIC COMPONENTS Following components has been used to build complete system: 26 self-powered silicon-cell pyranometers type SP110, 1 temperature probe 107 manufactured by Campbell Sci., CR1000x logger, AM16/32B Multiplexer, CELL215 LTE modem, 7 Ah 12V battery, solar charging regulator, 20W photovoltaic module. The GLOB system view in spring (3/03/2023, https://doi.org/10.5281/zenodo.8410795) The GLOB system view in summer (30/06/2023, https://doi.org/10.5281/zenodo.8410805) The GLOB system view in autumn (5/10/2023, https://doi.org/10.5281/zenodo.8412962) Theoretical electricity (kWh) generated on 2 selected days (2/05/2023 and 6/07/2023) by a photovoltaic module with an area of 1 m2 (20% efficiency) directed in different azimuths and inclined at an angle of 0, 45, 90 or 135 degrees. Calculations based on data measured by GLOB 0 200 400 600 800 1000 00:00 04:00 08:00 12:00 16:00 20:00 GHI 00:00 04:00 08:00 12:00 16:00 20:00 S45 S90 S135 00:00 04:00 08:00 12:00 16:00 20:00 W45 W90 W135 00:00 04:00 08:00 12:00 16:00 20:00 NW45 NW90 NW135 00:00 04:00 08:00 12:00 16:00 20:00 N45 N90 N135 00:00 04:00 08:00 12:00 16:00 20:00 NE45 NE90 NE135 00:00 04:00 08:00 12:00 16:00 20:00 E45 E90 E135 00:00 04:00 08:00 12:00 16:00 20:00 SE45 SE90 SE135 00:00 04:00 08:00 12:00 16:00 20:00 SW45 SW90 SW135 AZIMUTH HORIZONTAL S S S SW SW SW W W W NW NW NW N N N NE NE NE E E E SE SE SE POA (deg) 0 45 90 135 45 90 135 45 90 135 45 90 135 45 90 135 45 90 135 45 90 135 45 90 135 02.05.2023 0,11 0,18 0,18 0,12 0,18 0,18 0,14 0,14 0,15 0,12 0,09 0,11 0,11 0,07 0,08 0,09 0,08 0,09 0,10 0,12 0,13 0,10 0,16 0,17 0,13 06.07.2023 0,17 0,19 0,15 0,06 0,18 0,15 0,09 0,18 0,15 0,08 0,16 0,15 0,11 0,15 0,15 0,09 0,16 0,15 0,12 0,18 0,15 0,08 0,18 0,16 0,09 Daily cycle of shortwave radiation (in W*m-2) received on various arrays planes, measured on May 2, 2023 by the GLOB system The GLOB system has been financed by the Svalbard Miljøvernfond: project 19/76 Scan QR code to see the full technical description of the system Sketch of the GLOB system (https://doi.org/10.5281/zenodo.8411391) Location of the GLOB system (https://toposvalbard.npolar.no/) CONTACT: Sebastian
poster
Learning to correct for atmospheric seeing from solar observations John Armstrong and Lyndsay Fletcher University of Glasgow j.armstrong.2@research.gla.ac.uk 1. Introduction • Atmospheric seeing is ubiquitous in ground-based astronomy[1]. • We propose a post-processing correction for seeing using a machine learning algorithm known as generative adversarial network[2] (GAN). • This provides a kernel-free deblurring technique. • In models assuming some blur kernel explicitly it is hard to deal with occluded (optically thick) regions without a complex kernel estimation[9]. This is avoided in our model. Fig. 1: Discriminator[4,6] (D) Fig. 2: Generator[3,5] (G) References [1] Davies, R. and Kasper, M. (2012) Annual Reviews of Astronomy and Astrophysics, 50:350-351. [2] Goodfellow, I. et al. (2014). ArXiv e-prints. [3] Kupyn, O. et al. (2017). ArXiv e-prints. [4] Isola, P. et al. (2017). Proceedings – 30th IEEE Conference on Computer Vision and Pattern Recognition. [5] Lecun, Y. et al. (1998). Proceedings of the IEEE, 86(11):2278-2324. [6] Li, C. and Wand, M. (2016). Lecture Notes in Computer Science, 9907 LNCS:702-716. [7] Tsuneta, S. et al. (2008). Solar Physics, 249(2):167-196. [8] He, K. et al. (2015). ArXiv e-prints. [9] Nah, S. et al. (2016). ArXiv e-prints. Overview • We employ a post-processing machine learning algorithm to correct for atmospheric seeing. • The model is trained on data from SOT on Hinode which is imprinted with an artificial blur. • The model is then applied to data from CRISP with promising results. 2. Discriminator and Generator • There are two parts to a GAN: discriminator (D) and generator (G). • The goal of the generator is to fool the discriminator into thinking the data that it has created is real. • The discriminator wants to classify whether or not data is real. • The two play a game against each other and train simultaneously. D divides the generated and true images into 64, 128, 256, 512 segments, respectively (different colours represent the number of segments produced). D classifies each segment as real or fake. G learns the difference between the sharp and blurred image[8]. G adds what it has learnt to the blurry image for comparison with the sharp image. G reassembles the segments. 3. Blur Model Blur image via randomly offsetting the image w.r.t. itself twice and averaging. Fig. 3: SOT[7] image pre- and post-blurred 4. CRISP Ca II 8542ÅCorrection 5. Conclusion • Reconstructs lines well from CRISP data. • This shows that the image deblurring works perceptually and preserves spectral integrity. • This implies that the method can be used to correct for atmospheric seeing. Fig. 4: Correcting a CRISP observation of the 2014-09-06 flare. Blurry Sharp Corrected
poster
Managing the Scientific Software Ecosystem with Spack Todd Gamblin*, Gregory Becker*, Kiel Friedt*, Alicia Klinvex†, Gregory L. Lee*, Matthew LeGendre*, Peter Scheibel*, Barry Smith‡, James Willenbring† *Lawrence Livermore National Laboratory †Sandia National Laboratories ‡Argonne National Laboratory Combinatorial software complexity Large scientific codes can be extremely complex ARES tcl tk scipy python cmake hpdf opclient boost zlib numpy bzip2 LAPACK gsl HDF5 gperftools papi GA bdivxml sgeos_xml Scallop rng perflib memusage timers Silo SAMRAI HYPRE matprop overlink qd LEOS MSlib Laser CRETIN tdf Cheetah DSD Teton Nuclear ASCLaser MPI ncurses sqlite readline openssl BLAS Physics Utility Math External Types of Packages Developers build 4 versions of ARES 36 different ways. Linux BG/Q Cray XE6 MVAPICH MVAPICH2 OpenMPI BG/Q MPI Cray MPI GCC C P L D C P L D Intel 14 C P L D Intel 15 C P L D D PGI D C P L D C L D Clang C P L D C L D XL C P L D Dependency graph for one configuration of ARES, an LLNL multi-physics code. (L)ite (D)evelopment spack/opt/ linux-x86_64/ gcc-4.7.2/ mpileaks-1.1-0f54bf34cadk/ intel-14.1/ hdf5-1.8.15-lkf14aq3nqiz/ bgq/ xl-12.1/ hdf5-1-8.16-fqb3a15abrwx/ ... mpileaks mpi callpath dyninst libdwarf libelf Hash Spack can install arbitrarily many configurations of each package https://spack.io github/LLNL/spack @spackpm Spack has a rapidly growing open source community Spec syntax lets users build on demand –Build scientific software from source for best performance & portability across exotic platforms –Automate deployment for app developers, facility staff, and users –Provide integration testing harness for large software stacks. –No more manual builds: spack install <package> Done. –Build a community; leverage build automation work across facilities & teams Spack is a flexible, build-from-source package manager for scientific software that supports multiple versions, compilers, and configurations. Overview and Goals $ spack install mpileaks unconstrained $ spack install mpileaks@3.3 @ custom version $ spack install mpileaks@3.3 %gcc@4.7.3 % custom compiler $ spack install mpileaks@3.3 %gcc@4.7.3 +threads +/- build option $ spack install mpileaks@3.3 cppflags=‘-O3’ setting compiler flags $ spack install mpileaks@3.3 os=CNL10 target=haswell setting target for cross-compile $ spack install mpileaks@3.3 ^mpich@3.2 %gcc@4.9.3 ^ constraints on dependencies `spack fiind` shows what is installed. $ spack find ==> 103 installed packages. -- linux-x86_64 / gcc@4.4.7 -------------------------------- ImageMagick@6.8.9-10 glib@2.42.1 libtiff@4.0.3 pango@1.36.8 qt@4.8.6 SAMRAI@3.9.1 graphlib@2.0.0 libtool@2.4.2 parmetis@4.0.3 qt@5.4.0 adept-utils@1.0 gtkplus@2.24.25 libxcb@1.11 pixman@0.32.6 ravel@1.0.0 atk@2.14.0 harfbuzz@0.9.37 libxml2@2.9.2 py-dateutil@2.4.0 readline@6.3 boost@1.55.0 hdf5@1.8.13 llvm@3.0 py-ipython@2.3.1 scotch@6.0.3 callpath@1.0.2 jpeg@9a mpich@3.0.4 py-numpy@1.9.1 stat@2.1.0 dyninst@8.1.2 libdwarf@20130729 ncurses@5.9 py-pytz@2014.10 xz@5.2.0 -- linux-x86_64 / gcc@4.8.2 -------------------------------- adept-utils@1.0.1 boost@1.55.0 cmake@5.6-special libdwarf@20130729 mpich@3.0.4 adept-utils@1.0.1 cmake@5.6 dyninst@8.1.2 libelf@0.8.13 openmpi@1.8.2 -- linux-x86_64 / intel@14.0.2 ----------------------------- hwloc@1.9 mpich@3.0.4 starpu@1.1.4 -- linux-x86_64 / intel@15.0.0 ----------------------------- adept-utils@1.0.1 boost@1.55.0 libdwarf@20130729 libelf@0.8.13 mpich@3.0.4 -- linux-x86_64 / intel@15.0.1 ----------------------------- adept-utils@1.0.1 callpath@1.0.2 libdwarf@20130729 mpich@3.0.4 boost@1.55.0 hwloc@1.9 libelf@0.8.13 starpu@1.1.4 $ spack find callpath ==> 2 installed packages. -- linux-x86_64 / clang@3.4 ———————— -- linux-x86_64 / gcc@4.9.2 ------------- callpath@1.0.2 callpath@1.0.2 $ spack find -dl callpath ==> 2 installed packages. -- linux-x86_64 / clang@3.4 ----------- -- linux-x86_64 / gcc@4.9.2 --------- -- xv2clz2 callpath@1.0.2 udltshs callpath@1.0.2 ckjazss ^adept-utils@1.0
poster
MEASUREMENT OF THE UNDERGROUND ARGON RADIOPURITY FOR DARK MATTER DIRECT SEARCHES Ludovico Luzzi (CIEMAT) on behalf of DarkSide Collaboration • C.E. Aalseth et al. 2020 JINST 15 P02024 (Feb. 2020) LOW RADIOACTIVITY UAr FOR DARKSIDE-20K DARKSIDE-20K EXPERIMENT AND UNDERGROUND ARGON DS-20k aims to <0.1 background events except coherent neutrino scatters after all TPC & veto event selection criteria in ROI (20-200 keVnr) within 20.2 t fiducial volume during 10 y run. Liquid argon is used as active medium because of its scintillation and ionization characteristics. DarkSide-20k (DS-20k) is a two- phase liquid argon detector, under construction at LNGS. DArT is a small low-background detector designed to measure the 𝟑𝟗𝐀𝐫 depletion factor of different UAr batches (URANIA + ARIA): OFHC Cu cylinder, 1 l active volume, PMMA support structure with TPB coating, Readout two PDMs (1 cm² SiPMs) from DS-20k (DArTeyes). Located at LSC (1400 m.w.e.) will be installed inside ArDM single-phase: • 13 PMTs will see ~1 tonne AAr buffer used as shield and veto • New Pb-shield attached to existing Polyethilene shield • Substitution of AAr bath by pressurized LN2 @ 85 K. No veto • Installation in an ad-hoc Pb-castle flushed with Rn-free air (undergorund only) DArT IN ArDM: PRESENT AND FUTURE DArT: FROM CIEMAT TO LSC Why Argon? Underground argon (UAr), depleted of 𝟑𝟗𝐀𝐫 Future Ar-based experiments (DarkSide-20k, Argo) will use UAr. Main limitation Intrinsic activity of 𝟑𝟗𝐀𝐫in atmospheric argon (via spallation of cosmic rays on 𝟒𝟎𝐀𝐫): • beta decay with Qβ = 565 keV • t1/2 = 269 y • ~1 Bq/kg • Procurement of 50 tonnes of UAr from the Kinder- Morgan facility in Cortez (Colorado, USA) • Extraction of 330 kg/day, with 99.9% purity • UAr transported to Sardinia for final chemical purification at Aria The Aria and Urania projects are two fundamental steps in order to provide tonnes of detector-grade UAr. • Big cryogenic distillation column in Seruci, Sardinia • Final chemical purification of the UAr • Can process O(1 tonne/day) with 10³ reduction of all chemical impurities • Ultimate goal is to isotopically separate 𝟑𝟗𝐀𝐫from 𝟒𝟎𝐀𝐫 (at the rate of 10 kg/day in Seruci II) Urania Aria 350 m Seruci-I Seruci-II Specific setup for tests in surface (CIEMAT) and undeground (LSC): With all that it is clearly seen the 39Ar on top of the background, assuming a featureless línear background and considering a threshold of 33 keV it is possible to measure 1.0 cps. No uncertainty is quoted here because an understanding of the systematics is still ongoing. Bibliography Two events from the present setup of DArT are showed. From the difference between α decay (top) and β decay (bottom) Pulse Shape Discrimination (PSD) is performed and electron recoil events can be isolated. Comparing everything from surface to underground with lead shield, a reduction of the background is observed. Once adding the lead shield (black) the background is significantly reduced and the 39Ar β spectrum for low • arXiv:1707.08145v1 [physics.ins-det] 25 Jul 2017 PRELIMINARY For tests only AAr is used. The triplet lifetime is very long (1230 ns) because of the slower SiPM response: 𝐟𝟔𝟒𝟎is defined so the signal is integrated in the first 640 ns to define the fast component of the pulse energies is visible.
poster
2 3 4 If I look at myself in the CP mirror I just see myself! All measurements are in agreement with the SM prediction within 68% CL. A CP odd state is excluded at 99.7% CL (3.0 σ) ● Exclusion for CP-odd hypothesis:  Observed (exp.) significance = 3.0σ (2.6σ)  ϕCP distribution in bins of NN score Investigated τ decay channels The τh are identified with the DeepTau NN-based ID2 Decay modes identified with MVA-algorithm3 τ + τ − The cross-section of the H→ττ process has a sinusoidal shape d σ d ϕCP ∝const −cos(ϕCP−2α H τ τ) The Higgs through the looking glass. Measurement of the CP quantum number of the Higgs boson in H→ττ decays Andrea Cardini (DESY) on behalf of the CMS collaboration CP-violation in the Higgs couplings is investigated in: ● HVV couplings ●Yukawa coupling: ●Production via ttH and ggH ●Decays into τ leptons The Standard Model predicts that I have spin-parity 0+ But do I? Better check with a CP mirror We invert spatial coordinates and swap particles with anti-particles 1 2 Multicategory classification used to identify the Higgs decays from genuine ττ and lep./jet→τh backgrounds: ●BDT for fully hadronic channels ●Neural Network for semileptonic Signal parametrized as function of αHττ: weighted combination of CP even odd and maximum mixing cases ~ 90% of background estimated with data-driven methods ●Tau embedding for genuine ττ processes ●FF method for j→τh fakes NN MACHINE LEARNING BDT Likelihood profiled1 with respect to: 1. CP mixing angle 2. Yukawa couplings 3. CP mixing angle + inclusive Higgs production signal strength 4. Higgs production signal strengths (ggH + VBF) 5. CP mixing angle for Run 2, Run 3 and HL-LHC 5 3 Each decay plane is defined by two vectors4: a momentum and a reference vector CP even αΗττ= 0° CP mix αΗττ= 45° CP odd αΗττ= 90° Machine learning techniques can help in identifying me At the LHC I am surrounded by processes that can produce similar signatures to my decays to τ leptons CP mixing angle Acoplanarity angle a1(1260) ρ(770) The spin correlation is preserved for decays to leptons and via mesonic resonances pion momentum impact parameter (IP) polarimeter vector (h) References. 1. CMS-HIG-20-006: “Analysis of the CP structure of the Yukawa coupling between the Higgs boson and τ leptons in proton-proton collisions at √s= 13 TeV” 2. CMS-DP-2019-033: “Performance of the DeepTau algorithm for the discrimination of taus against jets, electron, and muons” 3. CMS-DP-2020-041: “Identification of hadronic tau decay channels using multivariate analysis (MVA decay mode)” 4. Andrea Cardini: “Measurement of the CP properties of the Higgs boson in its decays to τ leptons with the CMS experiment” 2 4 1 3 4 6 CP mixing angle measured with simultaneous fit for signal and background models over full Run 2 Fit performed with ●NN/BDT score for ML background categories ●Acoplanarity angle distributions in bins of NN/BDT score for higgs category → split by decay mode To be continued in Run 3 Uncertainty statistically dominated 1 5 7 d σ d ϕCP
poster
> Programm Heeremannscher Hof · Hofsaal im Erdgeschoss Königsstraße 47 48143 Münster > Wozu berechtigt Not? Workshop der Kolleg-Forschergruppe 8. April 2014 > Veranstaltungsort Sebastian Laukötter und Reinold Schmücker Kolleg-Forschergruppe „Normenbegründung in Medizinethik und Biopolitik“ Die Zahl der Teilnehmer ist begrenzt. Wir bitten daher um Anmeldung unter: normenbegruendung@wwu.de www.normenbegruendung-in-der-bioethik.de > Organisation Bildnachweis: Stefan Klatt. 9:30 Begrüßung 9:45 – 10:35 Wozu berechtigt Not? Zur Notwendigkeit und Möglichkeit einer Ethik für Notsituationen Reinold Schmücker (Münster) 10:35 – 11:25 Moralisch retten Guntolf Herzberg (Berlin) 11:55 – 12:45 Notsituationen als Herausforderung an die Moralphilosophie. Grundsätzliche Überlegungen über Möglichkeiten und Grenzen der Ethik Héctor Wittwer (Berlin) 15:00 – 15:50 Not und die Grenzen der Moral Simon Derpmann (Münster) 15:50 – 16:40 ‚Handeln aus Not‘. Moralphilosophische Implikationen eines durchaus gewöhnlichen Sprachspiels Anke Thyen (Ludwigsburg) 17:10 – 18:00 Wozu verpfl ichtet Not? Überlegungen zum Status globaler Hilfspfl ichten Sebastian Laukötter (Münster)
poster
Abstract Experimental Experimental H. S. Medeiros1, R. S. Pessoa1, 2, J. C. Sagás1, M. A. Fraga1,3, *, L. V. Santos1,2, H. S. Maciel1,2, M. Massi1 and A. S. da Silva Sobrinho1 References References Results Results Acknowledgments Acknowledgments Conclusions Conclusions 1Plasma and Processes Laboratory, Technological Institute of Aeronautics, 12228-900, S. J. dos Campos, SP, Brazil 2IP&D, University of Vale do Paraíba, 12244-000, S. J. dos Campos, SP, Brazil 3Institute for Advanced Studies, IEAv-DCTA, 12228-001, S. J. dos Campos, Brazil SixCy thin films deposited at low temperature by dual DC magnetron sputtering: Effect of power supplied to Si and C cathode targets on film physicochemical properties A DC dual magnetron sputtering system with graphite (C) and silicon (Si) targets was used to grow stoichiometric and non-stoichiometric silicon carbide (SixCy) thin films at low temperature. Two independently DC power sources were used to enable the total discharge power be shared, under certain proportions, between the Si and C magnetron cathodes. The motivation was to control the sputtering rate of each target so as to vary the stoichiometric ratio x/y of the deposited films. The species content, thickness and chemical bonds of as-deposited SixCy films were studied by Rutherford backscattering spectroscopy (RBS), profilometry, Fourier transform infrared absorption (FTIR) and Raman spectroscopy respectively. Overall, the present work reveals a new reliable plasma sputtering technique for low temperature growth of amorphous SixCy thin films with the capability of tuning the degree of formation of a-SiC, a-Si and a-C bonds in the film bulk. 0 1000 2000 3000 4000 5000 Si Si-O-Si and/or Si-O-C Si-C 45% 50% 40% 30% 0% Transmissمo (a.u) Numero de onda (cm -1) 20% 100 200 300 400 Channel 0 1 2 3 4 5 6 Yield (#/uC/keV/msr)1/2 0.5 1.0 1.5 2.0 Energy (MeV) Si C N O mafraga@ita.br Carbon Target Silicon Target Shutter Substrate holder Vacuum System Gas Line Deposition Chamber DC Power Source Gas Injection System (b) Figure 1. Photograph of the magnetron co-sputtering deposition system. 4.a) FTIR RBS 4.b) Profilometry 60% C target %Si=39,81 %C=38,89 0 1000 2000 3000 4000 5000 Si Si-C Si-O-Si and/or Si-O-C 100% 80% 70% 60% 55% 50% Transmissمo (a.u) Numero de onda (cm -1) (a) 0 10 20 30 40 50 60 70 80 90 100 0,1 1 10 C/Si % de potência no alvo de carbono teorico experimental 63% stoichiometry (a) 0 20 40 60 80 100 0 100 200 300 400 500 600 700 Vc Ic Vsi Isi (b) Corrente no alvo, IX (mA) Tensمo da discarga, Vdc (V) % de potência no alvo de carbono 0 100 200 300 400 500 600 700 Electrical Raman 400 600 800 1000 1200 1400 1600 1800 2000 0 20 40 60 80 100 120 140 160 70% C target Raman shift (cm -1) Intensidade (a.u.) (d) 400 600 800 1000 1200 1400 1600 1800 2000 0 50 100 150 200 250 300 350 400 80% C target Raman shift (cm -1) Intensidade(a.u.) (e) 400 600 800 1000 1200 1400 1600 1800 2000 0 10 20 30 40 Intensidade (a.u.) Raman shift (cm -1) 20% C target (a) SiC (TO) SiC (LO) 400 600 800 1000 1200 1400 1600 1800 2000 0 10 20 30 40 50 60 70 80 60% C target Raman shift (cm -1) Intensidade (a.u.) (c) SiC (TO) SiC (LO) 400 600 800 1000 1200 1400 1600 1800 2000 10 15 20 25 30 35 40 (b) Raman shift (cm -1) 45% C target Intensidade (a.u.) SiC (TO) SiC (LO) i. Fig. 1 and Fig. 2 show photographs and the scheme of the magnetron co-sputtering deposition arrangement used. ii. The magnetrons are disposed in parallel and angled at approximately 20º from the central axis (see Fig. 2). iii. The targets of the magnetron are made of carbon and silicon, both of high purity (99.99%). iv. As sputtering gas was used argon. The pressure was maintained fixed at 5.0 mTorr and the Ar flow rate at 10 sccm. v. The films were deposited varying the fraction of applied power in each magnetron target for a total value of 200 W. vi. RBS, FTIR, Raman and profilometry techniques were utilized to investigate film composition, chemical bonding and deposition rate, respectively. Figure
poster
This project has received funding from the European Union's Horizon 2020 research and innovation programme under grant agreement No 952439. STELLAR : a EU twinning project on LOFAR data analysis and knowledge transfer Abstract: The Scientific and Technological Excellence by Leveraging LOFAR Advancements in Radio Astronomy (STELLAR) is a project of mutual collaboration and know-how transfer in the field of radio astronomy, solar physics and space weather using the LOFAR instrument and data. Two institutions from Bulgaria, benefit from technical and scientific know-how exchange from world-leading RA institutions - ASTRON (the Netherlands) and DIAS (Ireland) via series of training hands-on sessions, workshops, seminars and project-focused schools for both students and senior staff. The poster presents the activities so far and future plans. All results, links to videos and outreach activities are hosted at a dedicated web-site. The STELLAR project is funded by the European Union's Horizon 2020 research and innovation programme under grant agreement No 952439. It is coordinated by the Institute of Astronomy, Bulgarian Academy of Sciences. A. Antonova(1) , К. Kozarev(1), A. Avramova-Boncheva(1), R. Miteva(1), M. Dechev(1), P. Zucca(2), E. Carley(3), S. Maloney(3), P. Petkov(4) Affiliations: (1) Institute of Astronomy and National Astronomical Observatory, Bulgarian Academy of Sciences, Bulgaria; (2) ASTRON, the Netherlands; (3) DIAS, Ireland; (4) Technical University of Sofia, Bulgaria General Information: • Call: H2020- WIDESPREAD-2018-2020, Twinning topic • EU Grant: 899 877 EUR • Duration: 36 months • Start Date: 01 September, 2020 Highlights Period 1: Online courses freely available at https://stellar-h2020.eu/ : ASTRON Training I on RF Technology and RF Development I; Space Weather Training I - Introduction to Space Weather at DIAS; ASTRON Training I on RF Technology and RF Development; ASTRON Training II on Phased Array Digital Signal Processing. Online seminars also freely available at https://stellar-h2020.eu/ Other activities: LOFAR Data School Staff Visits to ASTRON (LOFAR Operations and Science, Data reduction) Staff Visit I and II to DIAS (Space Weather Data Analysis) STELLAR Consortium Members: • IANAO (https://astro.bas.bg/) • TUS (http://www.tu-sofia.bg/) • ASTRON (https://www.astron.nl/) • DIAS (https://www.dias.ie/) Visit to ASTRON RF school at TUS, Sofia Visit to the LOFAR core stations Visit to Westerbork Visit to the LOFAR-IE station Visit to the LOFAR-IE station Synergistic Activity: the LOFAR-BG Project (https://lofar.bg/) • The consortium at present: IANAO, TU, SU, SHU • Included in the Bulgarian National Roadmap for Science Infrastructures (Ministry of Education and Science of Bulgaria) • Received funding for building and maintaining LOFAR station • Site selected (close to the Rozhen Observatory), RF checked and approved STELLAR objectives: •To transfer scientific and technical knowledge and capacity in radio astronomy from the highly experienced staff in ASTRON and DIAS to IANAO and TUS staff by means of versatile training activities; •To expand the research potential of the local solar and space weather scientists through a combi- nation of target training activities and research discussions and collaborations using LOFAR observa- tions; •To provide an opportunity for IANAO and TUS scientific and engineering personnel to build the necessary expertise and technical capabilities required for the building and operation of a Bulgarian LOFAR station
poster
SET CoP Aims: To provide an opportunity to share, reflect on, evaluate and develop the team’s practice To provide a platform for the showcasing and discussion of good practice To provide a showcase for new and innovative teaching and learning development practices To foster collaborative practices amongst the members of the Student Engagement Team Advantages: “Having a regular opportunity to get together means we can adapt things fairly quickly and suggest changes if needed” “Provides a safe space for discussing practice and a valuable informal meet-up” “I could be open and honest about arising issues / concerns with systems / expectations etc.” “The team could provide reassurance and support which was invaluable” “Supports a culture of collaboration, trust and respect” “The SET CoP is friendly and supportive” “It’s reassuring to be able to bring questions or ideas to the group and know that they will be listened to and discussed” “Useful conversations - I like that the atmosphere is generally informal” Challenges: “Not everyone is always present, the time and date doesn’t always suit all - particularly for part-time members of the team” “One member of the team had to travel a long distance to attend, which was not ideal” “Trying to convene CoP meetings over the summer when people were taking holidays. Due to the small size of our team, it proved difficult to get enough people to make it viable during this period” “The SET CoP is a forum for discussion about the provision of academic writing and information literacy support.” Contribution to the provision of student Academic Writing support: Highlighting current issues during 1:1s Sharing/applying training attended Discussing student/tutor feedback Disseminating exemplars and discussing how we support students with those assignments What next? As ASAs we believe SET CoP brings many advantages to our day-to-day role, but how do we know for sure a CoP has an impact on the quality of student academic support? Can we measure this? And if so, how? This is perhaps something to consider at future SET CoP meetings. Contact Us... Claire Olson olsonc@edgehill.ac.uk Andrew Tomkins tomkinsa@edgehill.ac.uk Scan the QR code: Tell us what you think. Have you ever been part of a CoP? We would love to hear about any of your experiences, or provide feedback on this poster. The establishment of a Community of Practice (CoP) to improve student academic support Further reading Cox, A., 2005. What are communities of practice? A comparative review of four seminal works. Journal of Information Science [online]. 31 (6), pp. 527-540. Available from: https://doi.org/10.1177/0165551505057016. Gómez, R.L. and Suárex, A.M., 2021. Extending impact beyond the community: Protocol for a scoping review of evidence of the impact of communities of practice on teaching and learning in higher education. International Journal of Educational Research Open [online]. 2, 100048. Available from: https://doi.org/10.1016/j.ijedro.2021.100048. Lesser, E.L. and Storck, J., 2001. Communities of practice and organizational performance. IBM Systems Journal [online]. 40 (4), pp. 831-841. Available from: https://doi.org/10.1147/sj.404.0831. Wenger, E., 1998. Communities of practice: Learning, meaning and identity. Cambridge: Cambridge University Press. Wenger, E., 2011. Communities of practice: a brief introduction. In: STEP Leadership Workshop, 20 October 2011, Eugene, OR [online]. Eugene: University of Oregon. Available from: https://scholarsbank.uoregon.edu/xmlui/handle/1794/11736 [Accessed 24 March 2022]. UniSkills Edge Hill University - Student Engagement Team Student Engagement Team Community of Practice: ƒ The Student Engagement Team Community of Practice (SET CoP) was formed in April 2019 ƒ The SET CoP currently consists of seven Academic Skills Advisors (ASAs), who meet every six weeks ƒ The meetings are minuted and report to the Student Engagement Team meeting ƒ There is a rolling chair ƒ Most meetings take place off ca
poster
What Can Joint Survey Processing (JSP) do for Time-Domain Studies? Ranga Ram Chary (Caltech/IPAC) for the JSP Team rchary@caltech.edu; Feb 2022 Abstract Joint Survey Processing (JSP) is the pixel level combination of ground- and space-based datasets such as Euclid (2023+), Rubin (2023+) and Roman (2027+). It is particularly important for mitigating source confusion which affects 50% of the sources in deep seeing-limited datasets such as those that will be taken by Rubin/LSST. By using location and morphology priors on static sources derived from space-datasets such as Euclid and Roman, and estimating accurate time- and location-dependent point spread functions for the ground- based data, one can model the scene in individual frames forward. The residuals after subtracting the modeled scene from the data, reveal transient events down to the shot noise limit of the individual images or in time- integrated coadds, even in crowded fields. This significantly improves the detectability of transient events, both in number and in faintness. JSP is therefore highly relevant for 1. detection of microlensing events, 2. moving objects like nearby stars and asteroids, 3. supernovae, and 4. tidal disruption events close to the high surface brightness centers of galaxies. In addition, by providing accurate multi-band photometry on galaxies one can derive accurate photometric redshifts and galaxy physical parameters which can be used to select host galaxies and transient electromagnetic counterparts for multi-messenger studies such as those where the positional error circles can be large (e.g. LIGO gravitational wave events, ultra high-energy cosmic rays or fast radio bursts). Figure 1: Map of the entire sky (from J-C. Cuillandre) in celestial coordinates showing the Euclid coverage (15000 sq-deg in yellow), and the Rubin coverage (20000 sq-deg) as the colored rectangles. There are also 3 deep fields totaling 40 sq-deg – two in the South and one in the North. Opportunities for time domain science come from the 7500 sq- deg overlap area which will have deep multi-wavelength optical and NIR data with sub- arcsecond resolution, and the 30 sq-deg high latitude deep fields in the South. In 2027+, the Roman microlensing survey towards the Galactic bulge, a potential Roman Galactic Plane survey and 15 sq-deg SN survey will also provide new opportunities to search for variable events. Figure 2: 5-year sensitivity of the Rubin/LSST ugrizy survey (blue) in comparison with the Euclid deep field sensitivities (green). Potentially the LSST deep drilling fields will be another 2.5 mags deeper than that shown while the Euclid wide field will be 2 mags shallower. The Euclid deep fields will have about 20 visits while the LSST area will have about ~100 visits. Figure 3: The blended fraction of sources in seeing-limited data (Rubin) as a function of apparent magnitude, in extragalactic fields. In the deep fields, more than half the sources are blended with other nearby sources. Combined with variable Rubin seeing, measuring variability in these faint sources is challenging. Forward modeling the scene using the high resolution space data (from Euclid or Roman), enables the detection of transients down to the shot noise limit of the surveys. Figure 4: An illustration of joint survey processing on Hubble/ACS and Subaru/HSC data from Faisst et al. (2022). By using priors on the position and morphologies of sources derived from the high resolution ACS data and fitting for these known sources in the HSC/Subaru data, faint transients which are on top of host galaxies (e.g. SNe or nuclear AGN) can be detected. This method also can help reveal moving objects like stars or asteroids if the time baseline is large enough. Summary The detection of variable sources at the sensitivity limit of upcoming, confusion-limited ground-based surveys requires forward modeling of the scene by using priors from the high resolution space-platform data. This requires joint pixel level pr
poster
OBSERVATIONAL PROPERTIES OF PUFFY ACCRETION DISC ABSTRACT We have performed a general rela/vis/c radia/ve magnetohydrodynamics (GRRMHD) simula/ons of a stable accre/on disc around a stellar-mass black hole (BH) with sub-Eddington mass accre/on rates. The numerical solu/ons revealed an elevated ver/cal structure - above a dense geometrically thin core of dimensionless thickness ℎ!/𝑟∼0.1, resembling a classic thin accre/on disc, a puffed-up, op/cally and geometrically thick layer of lower density and ℎ"/𝑟∼ 1.0 is formed. We refer to this solu/on as the Puffy disc. We discuss the observa/onal proper/es of a puffy discs, in par/cular the geometrical obscura/on at higher observing inclina/ons of the inner area by the puffed-up region, and collima/on of radia/on along the accre/on disc spin axis. These effects may explain the apparent super-Eddington luminosity of some ultraluminous X-ray binaries. We also present synthe/c spectra of puffy discs, and show that they are qualita/vely similar to Comptonized thin disc spectra. We demonstrate that the exis/ng XSPEC spectral models provide good fits to synthe/c spectra of puffy discs but cannot correctly recover the input luminosity, nor the black hole spin. We suggest that puffy discs may correspond to X-ray binary systems in the intermediate spectral state with luminosi/es above 0.3 of the Eddington luminosity. Debora Lančová1, Maciek Wielgus2,5, Odele Straub3, Włodek Kluźniak4, Ramesh Narayan5, David Abarca4, Gabriel Török1 and Marek Abramowicz1,4 KORAL CODE The simula*ons were performed using GRRMHD code KORAL (Sądowski et al. 2013a, 2014a), a finite-difference Godunov type code solving coevolu*on of gas, radia*on and magne*c field in an arbitrary metric. Coupling of gas and radia*on is solved semi-implicitly using M1 closure scheme. We obtain in total 30 mil. corehours on Polish supercomputer Prometheus, part of the PLgrid infrastructure. This computa*onal resources enables us to run a number of simula*ons with different ini*al condi*ons to obtain three simula*ons with different mass accre*on rates. 1Ins*tute of Physics, Silesian University in Opava, Czech Republic 2Max-Planck-Ins*tut für Radioastronomie, Germany 3ORIGINS Excellence Cluster, Germany 4Nicolaus Copernicus Astronomical Centre, Polish Academy of Sciences, Poland 5Black Hole Ini*a*ve at Harvard University, USA We are studying a mildly sub-Eddington accre*on onto a non-rota*ng stellar mass BH with a mildly sub- Eddintgon accre*on rate. The classical models of accre*on discs repor*ng a thermal and viscous instability in this regime, however, our solu*on is stable thanks to the strong ver*cal net flux of magne*c field. The magne*c pressure inside the disc is at the same order as the radia*on pressure, both strongly dominates over the gas pressure. This also cause the accre*on disk to be “puffy” - with geometrically thin dense core (1) and geometrically and op*cally thick turbulent puffy region (2). Although the core resembles a standard thin disc, the op*cally thick puffy region and the posi*on of the photosphere has a strong influence on the observa*onal proper*es and thus the disc observa*onal appearance is very different from a standard models. On the figure, the thick white line denotes the posi*on of the photosphere, while the dashed line is the density scale-height. The color contours are showing the gas temperature. 2 1 OBSERVATIONAL APPEARANCE Radia*on from the disk is beamed to the direc*on of the axis and obscured for viewing angles close to the equatorial plane, in contrast with the standard thin disc model. The inner parts of the disc and the central BH is completely hidden from observers with high inclina*on. At the same *me for low inclina*on observer the disc radiates all the way to the BH horizon, leaving no trace of the innermost stable circular orbit. Three columns on the le` are showing the picture of a puffy disk for three different mass-accre*on rate compared to the 0.6 ̇𝑀#$$ standard thin disc image in the right column, with i
poster
       !   !"#$%&'  ( )    **                    &                 !    !   "  !      #   $ ! % !& &   !      !&   '   ( )#*  ** !+        , -   .  !(/     ! '0(0'112 2   3!  - -!40  %2   !5 !& * -! (  ! 4  %  4 %5 &)  12   ' .      26 %7   8#    0 *  49:  7$ )  #   %%     .0 * $%!!     &  + ) #, 7 %%  ! !-   ! !!;        !  0 *     <= * *  .     7 !     !   7 !< 4 ( -        0 *     0 *   )0 *  0% *   2   +     0-* #%> % % 8  ! *5.5%?1'0@ 8  5 #-  26%  ! * '  (   ;& 3      )2   < 62A <6  7 <22<?  2  '   " %!; 0 * !26%  =     &0% ;  2 /  =/   2  =    ! $%!!   ' !+ #! !   %  ! ! + %/    B% 4C  2    !  -;  &  '  2    '  .   *'  0    '  B%  !B..3 %! B% !26% *  ! .,  / 0!) B..$ .    34 % !(/  !   52!,##.   ! %  !%*  0 6%  % 5+ !,##/; 1 &  2  3*  &&  )   ) .4   )  5     ) .  $!  !0 *'  2!!    %/ 3 ! %%* 26%  /      % * !  26 ;     %3   ; =  3( 343D     % !     49:   !  # $%!! %  $%!! -  %   %% %4 %?1'0@; /!  .    *;    .!-  7+  !; %%    % !2/ ;%7  *  /   %% %4 %?1'0@; % !   2E     !   ( )            * !       !$%  )   #*) 0 6% );.231'4D  4    .     ! . !     *     !    2   !    0%!  *;  +  14'  8   '   (  6  = 
poster
Dependence of IR-Bright AGN Activity on Halo Mass and Redshift in the SPT Cluster Surveys Benjamin Floyd B 1 Mark Brodwin 1 Rebecca E. A. Canning 2 The SPT Collaboration 1University of Missouri–Kansas City 2University of Portsmouth Motivation We know that environment plays a role in galaxy evolution (e.g., Dressler, 1980, 1984), but how exactly? What are the predominant environmental-dependent mechanisms that are behind the rapid mass assembly (Mancone et al., 2010), en- hanced star formation (Brodwin et al., 2013) and increased AGN activity (Martini et al., 2013) that has been observed in galaxy clusters at high-redshift (z ≳1.4). Independently both Brodwin et al. (2013) and Ehlert et al. (2015) arrived at sim- ilar conclusions that galaxy–galaxy merging within the clusters may be the mechanisms driving galaxy evolution in galaxy clusters early in their history. In sufficiently gas-rich galaxy–galaxy merging, the gas will be dynamically disrupted triggering both enhanced star-formation activity and allows for gas to funnel into the galactic nucleus and accrete onto the central supermassive black hole creating an AGN. Similar to Ehlert et al. (2015), we aim to model the AGN incidence in galaxy clusters as tracers of galaxy evolution of the cluster members. SPT Cluster Sample Our cluster sample derives from two galaxy cluster surveys carried out by the 10-meter South Pole Telescope (SPT; Carlstrom et al., 2011) the 2500 deg2 SPT-SZ survey (Bleem et al., 2015) and the deep, 100 deg2 SPTpol 100d survey (Huang et al., 2020). Combined we have a cluster sample of over 300 clusters with a median redshift of z ∼0.69 and median cluster mass of M500 ∼3.73 × 1014M⊙. By using the SPT cluster surveys we are able to span a wide range of redshifts (see Fig 1.) and as the SPT detects clusters using the Sunyaev–Zel’dovich Effect (SZE) we have an effectively uniform-selected mass sample. Figure 1. Mass and redshift distribution of the SPT cluster surveys. IR-Bright AGN Selection Our AGN are selected using Spitzer/IRAC imaging in 3.6 µm and 4.5 µm wavelengths. We find a color selection of [3.6 µm]− [4.5 µm] ≥0.7 to be optimal in selecting AGN that would have been otherwise selected by a IRAC color-color selection (e.g., Stern et al., 2005) as shown in Figure 2. Furthermore, to account for Eddington bias in the form of photometric uncertainty we employ a fuzzy membership selection to our data in order to create the final sample as seen in Figure 3. Using our selection criteria, we select ∼2200 AGN candidates. 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 Photometric Redshift 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 [3.6] - [4.5] (Vega) Stern Wedge AGN Non-Active Galaxies 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 Lookback Time (Gyr) 101 103 N Figure 2. IRAC colors of all objects in the Spitzer Deep Wide-Field Survey (SDWFS; Ashby, Stern, et al., 2009) detected in 4.5 µm with SNR ≥5. Red points are objects that satisfy the AGN selection criterion as described in Stern et al. (2005). Blue hexbins are SDWFS objects that do not fall within the Stern wedge criterion. Our IRAC color selection is shown as the solid black line. 0.25 0.00 0.25 0.50 0.75 1.00 1.25 1.50 [3.6 m] - [4.5 m] (Vega) Original: = 0.89, = 0.15 Corrected: = 0.72, = 0.12 Degree of Membership = 0.56 SDWFS Number Counts Figure 3. Example of applying our Eddington bias correction and computing degrees of membership into our AGN sample. The original color and color error of the object is assumed to form a Normal distribution as given by the dashed curve which is then convolved with the deeper SDWFS number count distribution to account for Eddington bias. The object’s degree of membership into our sample is given by the integral of the corrected color (solid curve) above our cut off. Preliminary Trends Preliminary analysis on the our data sample and has found some initial interesting trends such as indications of an inverse trend of AGN richness with cluster mass at a fixed redshift as can be seen in Figure 4.
poster
Grant applicants Open science is a priority for Kim, she only has a short publication list. She uses pre-registrations and shares her pre-prints, peer-review. She works together in an electronic notebook and afterwards shares the metadata, datasets and protocols in trusted repositories. Publishing in journals with a high impact factor is a priority for John, he already has a lot of high impact factor publications. He finds data management an administrative task and only shares data when requested by the journal. He doesn’t even share his metadata. Mia has a lot of high impact factor publications. She developed an app that citizens can download to help collect data for her research project. She decided to share nothing of the research (not even the metadata) because of sensitive personal data. Koen collaborates with industry and contributed to many patents, due to related IP issues he has a short publication list and does not share many datasets. He does have an open attitude towards open science. Once he can share, he will. Afterall he wants to contribute to society & researchers. MIA KOEN KIM JOHN Rewards_A2.indd 1 02/05/2023 21:34
poster
Fast Dynamic Load Balancing Tools for Extreme Scale Systems Parallel unstructured applications running on the latest petascale systems require partitions optimizing specific balance metrics. Methods combining the most powerful graph based and geometric methods with ParMA diffusive methods directly operating on an unstructured mesh are discussed. The SSE is focused on extending ParMA to support applications with relational structures that can be expressed with a graph structure. More Information: http://www.scorec.rpi.edu/parma Contact: Cameron Smith smithc11@rpi.edu and Mark Shephard shephard@rpi.edu Cameron W. Smith, Gerrett Diamond, and Mark S. Shephard Rensselaer Polytechnic Institute, USA ParMA: Partitioning using Mesh Adjacencies Guide partitioning decisions with mesh adjacency information • Mesh and partition model adjacencies represent application data more completely then standard (hyper)graph-partitioning models. • All mesh entities can be considered; (hyper)graph uses a subset of the adjacency information. • Complete mesh - O(1) time adjacency queries. Advantages • Avoid graph construction and directly account for multiple entity types important for the solve process • Easy to use with diffusive procedures – compliments other partition methods via incremental refinements Diffusive Approach • Iteratively migrate small sets of elements to (1) reduce the peak imbalance and (2) reduce the number of entities on the part boundaries. Diffusive Iteration Stages • Weight computation – compute weights and exchange with peers • Targeting – determined how much weight each peer can accept • Element selection – select elements for migration • Migration – move elements to peers Complete mesh adjacency structure. . Dynamic Partitioning of Unstructured Meshes Tools for re-partitioning an unstructured mesh due to changing work loads or communication patterns are required to: • Balance work, reduce communications, output distribution, execute in parallel quickly, use little memory, and provide API Common Methods • Graph, hypergraph, geometric diffusive, and local Examples will demonstrate that specific combinations are needed to reach really large part counts (> 500K). 1 - “Controlling Unstructured Mesh Partitions for Massively Parallel Simulations”, Min Zhou, et al., 2010 Geometric recursive inertial bisection Dual graph partition Mesh Example of partitioning methods. Global partitioning [1]. Local partitioning [1]. *** • *** *** • **** Partitioning to One Million Parts Multiple tools needed to maintain partition quality at scale [2] • Local and global topological and geometric methods via Zoltan and Zoltan2 integration ParMA quickly reduces large imbalances and improves part shape • 1.6B element mesh from 128K to 1M parts then running ParMA. 128K partition has less than 7% imbalance for all entity orders. • Global RIB – 103 seconds ParMA – 20 seconds 209% vtx imb reduced to 6%, perfect elm imb increased to 4%, and 5.5% reduction in avg vtx per part suspension upright Mira parameters: a = 10% b = 70% g = 64k *If memory is limited additional methods are available. • 12.9B element mesh from 128K (< 7% imb) to 1M parts then ParMA. • Local ParMETIS – 60 seconds. ParMA – 36 seconds 35% vtx imb reduced to 5%, 11% elm imb reduced to 5%, and 0.6% reduction in avg vtx Extending Application Support Goal • Support applications operating on graph-like structures. Three demonstrations: • Finite element unstructured mesh-based analysis • Hierarchic multi-scale for soft tissue modeling • Scale-free graphs for discrete event simulations Approach • Define N-graph abstraction to directly represent both graph and mesh relational structures: • defined as 𝐺" 𝑉, 𝐸&, 𝐸', … , 𝐸")' where: 𝑉: atomic units 𝑢+ of the domain Ω which uniquely exist on one part s.t. Ω = ∪∀+ 𝑢+, and 𝐸+: relations 𝑒+(𝑢, 𝑣) of type 𝑖between vertices 𝑢 and 𝑣where 𝑢, 𝑣 ∈𝑉and 𝑢 ≠𝑣. • Optionally, vertices and edges may be assigned associated weights 𝑤. • Implement EnGPar - N-Graph based v
poster
Evaluating TDP-43 Targets in Amyotrophic Lateral Sclerosis Using Drosophila and Patient Spinal Cords Alexander Blythe1, Erik Lehmkuhl1, Suvitha Loganathan1, Eric Alsop4, Dianne Barrameda1, Bhavani B. Siddegowda1, Archi Joardar1, Kendall Jensen4, Daniela C. Zarnescu1,2,3 1Department of Molecular and Cellular Biology, 2Department of Neuroscience, 3Department of Neurology, 4Translational Genomics Research Institute Introduction and Background DLP DLP HRP w1118 D42>TDP-43WT D42>TDP-43G298S Figure 3. Images of dissected larval neuromuscular junctions (NMJ) treated with immunofluorescent staining. Horseradish peroxidase (HRP) is used as a control to emphasize changes in DLP levels. An additional diagram showing the different nervous tissue types in developing Drosophila larvae is shown. Sample Type Spinal Region TDP-43 Pathology Disease Mutation Control Thoracic No NA Control Lumbar No NA Control Cervical No NA Disease Thoracic Unknown C9orf72 Disease Lumbar Yes TDP-43 Disease Cervical Yes C9orf72 Disease Lumbar Yes OPTN, SETX Disease Lumbar Yes DCTN1 Disease Lumbar Yes NEK1 Figure 2. Larval turning times were used to assay locomotor function in 3rd instar larvae. DLP overexpression (OE) resulted in a rescue of locomotor function while DLP knockdown (RNAi) exacerbated TDP-43-induced locomotor dysfunction. N=33- 40. Asterisks (*) indicate p-values less than 0.05. Significance tests were performed using the Mann U Whitney Test. Effect of DLP mRNA Levels on Larval Motor Function DLP Visualized at the Larval Neuromuscular Junction DLP Levels in VNCs and NMJs During TDP-43 Proteinopathy DLP Levels in VNCs and NMJs During TDP-43 Proteinopathy Figure 5. Quantification of DLP levels in the ventral nerve cords (VNCs) of 3rd instar larvae. Western blots were run on homogenized VNCs from control larvae and larvae expressing TDP-43WT and TDP- 43G298S in the motor neurons. 10 VNCs were used in each sample. DLP intensity was normalized to Ponceau stain. Figure 4. Quantification of DLP levels in 4 terminal boutons at the NMJ of 3rd instar larvae relative to bouton size. N=3. Comparison of w1118 to D42>TDP-43G298S yields a p-value of 0.044. Asterisks (*) denote p-values less than 0.05. DLP Human Ortholog (GPC4) Levels in Spinal Tissue In 97% of ALS cases and about 45% of frontotemporal dementia cases, insoluble cytoplasmic aggregates containing highly phosphorylated TDP-43 have been identified as pathological hallmarks1. Considering these complexes and their noted neurotoxicity, a current hypothesis for TDP-43’s potential role in modifying neurodegenerative phenotypes during ALS is the ribostasis hypothesis, which posits that TDP-43 binds to and sequesters mRNAs into cytoplasmic insoluble complexes in motor neurons. These aggregates result in translational repression of affected mRNAs and severely dysregulate protein homeostasis, leading to motor neuron stress and eventually death2. Figure 1. Ribostasis hypothesis: Cytoplasmic TDP-43-induced sequestration of mRNAs into insoluble complexes results in translational repression of affected mRNAs, thereby lowering protein levels and disrupting proteostasis in motor neurons. A Drosophila model of ALS has been developed by overexpressing either wild- type or mutant TDP-43 in motor neurons, using a motor neuron-specific D42 driver with different versions of the TDP-43 bound to UAS promoters. These models have demonstrated ALS phenotypes including cytoplasmic TDP-43 aggregates, reduced lifespan, sleep fragmentation, and locomotor dysfunction3. Following tagged ribosome affinity purifications (TRAP) and RNA immunoprecipitations (RIP), potential mRNA candidates with both low ribosome association and co-precipitation with TDP-43 have been identified4. One such candidate is dally-like protein (DLP), a glypican present in the extracellular matrix of developing motor neurons. By determining if changed DLP levels exacerbate or rescue ALS phenotypes in developing larvae and observing if DLP protein levels are alter
poster
The way to Open Science: the case of Spanish researchers Candela Ollé¶, Alexandre López-Borrull¶, Remedios Melero2, Juan-José Boté-Vericad3, Ernest Abadal3 1 Faculty of Information and Communication Sciences at Universitat Oberta de Catalunya, Barcelona 2 Institute of Agrochemistry and Food Technology. Consejo Superior de Investigaciones Científicas (CSIC), València. 3 Departament de Biblioteconomia, Documentació i Comunicació Audiovisual & Centre de Recerca en Informació, Comunicació i Cultura. Universitat de Barcelona, Barcelona. Introduction and methodology The poster describes the results of the online survey on open science carried out on researchers affiliated with universities and Spanish research centres and focused on open access to scientific publications, the publication process, the management of research data and the review of open articles. The main objective was to identify the perception and habits of researchers with regard to practices closely linked to open science and the scientific value added is that offers an in-depth picture of researchers as one of the main actors to whom this transformation and implementation of open science will fall. It focuses on the different aspects of OS: open access, open data, publication process and open review in order to identify habits and perceptions. This is to make possible an implementation of the OS movement. The survey was carried out among researchers who had published in the years 2020-2021, according to data obtained from WoS. It was emailed to a total of 8,188 researchers and obtained a total of 666 responses, of which 554 were complete, the rest being forms with some questions unanswered. The overall conclusions are as follows: - The meaning of open science and its implementation are emerging topics that still require promotion, training and even persuasion to be carried out. - APCs present an economic barrier to those groups that do not have funding available for paying for them, although they are not the main criterion when selecting a journal. - Although Open Peer Review may seem innovative and emerging, many researchers have published or reviewed in journals that offer the opportunity for open review. - Tenure and evaluation policies do seem to have a clear effect on the behaviour of researchers with regard to open science. - Resolving the issue of evaluation in research may clearly condition the habitual uses of researchers when selecting a journal as well as the payment of APCs. - Researchers state that they share research data more for reasons of persuasion than out of an obligation to carry it out. - Researchers do not seem to question the objectives of open science, although they do question the pathways or difficulties that may arise from day to day. - Open access to publications and open research data are two realities that are increasingly consolidated in the scientific community. - Researchers seem aware that we are undergoing change, where academic evaluation or policies relating to open science, its implementation and shifting habits among researchers may change. In this sense, more and better support is needed on the part of institutions and teacher support services. References Abadal, Ernest; Abad-García, Francisca; Anglada, Lluís; Boté-Vericad, Juan-José; Esteve, Asunción; González-Teruel, Aurora; Labastida, Ignasi; López-Borrull, Alexandre; Ollé, Candela; Melero, Remedios; Rodríguez-Gairín, Josep Manel; Santos-Hermosa, Gema (2023a). Ciencia abierta en España 2023: informe de situación y análisis de la percepción. Barcelona-València: Grupo Ciencia Abierta. http://hdl.handle.net/2445/200020 Abadal, Ernest; Anglada, Lluís; Labastida, Ignasi; Melero, Remedios; Ollé-Castellà, Candela (2023b). Recomendaciones a la administración pública para facilitar la implantación del modelo de ciencia abierta en España. http://hdl.handle.net/2445/198759 Ollé C, López-Borrull A, Melero R, Boté-Vericad J-J, Rodríguez-Gairín J-M, Abadal E (2023) Habits and percepti
poster
Universität Tübingen · Philosophische Fakultät Seminar für Sprachwissenschaft Keplerstr. 2 · 72074 Tübingen Telefon +49 7071 29-77352 · https://talar.sfb833.uni-tuebingen.de Datenübernahme durch TALAR, das Tübinger Archive of Language Resources Vorhandene Daten in TALAR • Gepflegte Referenzdaten o Tübinger Baumbank-Collection o GermaNet • Gehostete Daten o Daten des SFB 833/SFB 441 o Daten des GRK 1808 • Daten externer Partner o SFB 638 o Index Thomisticus Treebank o ... • Community Daten o Universal Dependency Treebanks (UD) Zertifizierung des Repositoriums Seminar für Sprachwissenschaft Philosophische Fakultät Tübingen Archive of Language Resources (TALAR) Spezialisierung • Baumbanken/Treebanks • Wortnetze • Word-Embeddings Modalität • geschrieben • gesprochen Akzeptierte Datenformate • GermaNet XML-Format • CoNLL-U • Word-Embeddings • Weitere Formate nach Absprache Ansprechperson(en) • Thorsten Trippel • Claus Zinn Suchmöglichkeiten in den gepflegten Daten Zertifikat erteilt am 01. Juni 2023 Aktuell: turnusgemäße Erneuerung Daten Einbringen Initiale Datenübermittlung • Übermittlung nach BagIT-Standard • Werkzeugunterstützung: Bagman Voraussetzung: • Unterstützte Datentypen o Qualitätsgesichert o Auswahlprozess • Datenüberlassungsvertrag o Muster siehe https://uni-tuebingen.de/de/134320 • Offene Lizenz: CC-BY 4.0 oder höher • Metadaten nach ISO 24622 • Unterstützte Profile § CourseProfile § ExperimentProfile § LexicalResourceProfile § ResourceBundle Findable • Auffindbar mit persistentem Identifikator (PID) • Handle / DOI • Nachweis über Forschungsdatensuchmaschinen • z. B. VLO • Auffindbarkeit über Standardsuchmaschinen Accessible • Downloadmöglichkeit freier Daten • Zugang zu vorhandenen zugangsbeschränkten Daten • Metadaten frei zugänglich über technische Protokolle Interoperable • Je nach Datentyp • Nutzbar in den vorhandenen Auswertungswerkzeugen • Metadaten: standardkonform und maschinell interpretierbar Reusable • Nutzbar unter den Bedingungen der Lizenzen in eigener Software • Verwendung in anderen Projektkontexten • Klare Rechte an den Daten • Vertrauenswürdige Aufbewahrung F A I R Scott Martens (2013). TüNDRA: A Web Application for Treebank Search and Visualization. In: Proceedings of The Twelfth Workshop on Treebanks and Linguistic Theories (TLT12), Sofia, pp. 133—144. Hinrichs, Marie and Lawrence, Richard and Hinrichs, Erhard (2020): Exploring and Visualizing Wordnet Data with GermaNet Rover. In Proceedings of the CLARIN Annual Conference 2020, pp. 32-36. C. Zinn: Bagman - A Tool that Supports Researchers Archiving Their Data. Selected papers from the CLARIN Annual Conference 2021 (Virtual Event), Linköping University Electronic Press, vol. 189, pages 181-189, 2022. § SpeechCorpusProfile § TextCorpusProfile § ToolProfile § WebLichtWebService
poster
1 3 ABUNDANCE 2 Unvegetated Microplastic retention in the sediment of small-bodied tropical seagrass meadows This study presents the first assessments of MP contamination in seagrass meadows of the Southwestern Tropical Atlantic bioregion. Our results showed that the MPs are ubiquitous even in a Marine Protected Area. There was no difference in the abundance of MPs between vegetated and unvegetated areas. Seasonal influence was not relevant in the MP abundance in the three studied seagrass species. Identifying the drivers leading to MP accumulation in seagrass meadows may include a wider assessment of seagrass species with different plant traits and meadow complexities, as well as the local environmental factors is necessary. 66% Fiber 21% Film 4% Fragment First study to investigate MP in Halodule wrightii, Halophila decipiens and H. baillonii Now 25% of the 72 seagrass species have been investigated. Brazil Halodule wrightii Halophila decipiens Halophila baillonii MICROPLASTIC IN TROPICAL ESTUARINE SEAGRASS MEADOWS: DOES SEAGRASS SIZE MATTER? 1 Department of Biology, Federal Rural University of Pernambuco (DB/UFRPE), central campus, Recife, Brazil. 2 Department of Fisheries and Aquaculture, Federal Rural University of Pernambuco (DEPAQ/UFRPE), central campus, Recife, Brazil. 3 Centro de Ciências do Mar do Algarve (CCMAR), University of Algarve, Gambelas campus, Faro, Portugal. * Corresponding author: viegasmao@gmail.com DOI: 10.5281/zenodo.12709067 Published: July, 10, 2024 ACKNOWLEDGMENTS Study area INTRODUCTION OBJECTIVE RESULTS Ana M.C. Souza , Guilherme V.B. Ferreira , Carmen B. de los Santos , Flávia L. Frédou , Karine M. Magalhães 2 1 * 3 2 1 METHODS Seagrass meadows have been appointed as efficient microplastic (MP) trapping hotspots. MPs accumulation in the seagrass sediments from the Tropical Atlantic bioregion, such as those found i n South America, have not been investigated yet. Also, small-bodied segrass species are much less studied than large ones onther capacity to retain s MPs. Research sedimentary MP retention in areas vegetated by small-bodied seagrass species Halodule wrighti i (HW), Halophila decipiens (HD), and H. baillonii (HB) a nd an adjace nt unvegetat ed (U N) ar ea in a tropi cal estuary in the Brazilian coast, according to seasonality. Sampling was carried out in a multispecific seagrass meadow and an adjacent unvegetated area durin g the dry and rainy seasons. Biomass, shoot density and canopy height did not variation between seasons. Sediment between sand and fine sand CHARACTERIZATION Sediment samples (N = 80) A total of 227 MP particles Microplasctic Meadow and Sediment Analysis 1. Characterization Seagrass meadow Sediment 2. Microplastic Extraction (NaCl / NaOH) Visual identification 3. Literature review All types of documents Web of Science Environmental Protection Area (APA) of Santa Cruz MPs abundance (particles Kg dw) Unvegetated Vegetated p > 0.05 MPS VARIATION Species/season No significant variation among species and the adjacent unvegetated area, neither for seasonality was detected. However, we found a positive correlation betwee n MPs abundance and sediment grain size. South Atlantic Ocean CONCLUSION Fibers (73%) Blue (51%) < 1 mm (80%) Areas/season Unvegetated = 125 ± 117 (Itens/kg dw) = Vegetated = 144 ± 143 (Itens/kg dw) -1 -1 Total = 142 ± 140 (Itens/kg dw) -1 MEADOW CHARACTERIZATION -1 MPs abundance (particles Kg dw) -1
poster
• • • • • • • • • • • • ➢ • • ➢ • • • • × • • • 1990 1995 2000 2005 2010 2015 2020 4000 6000 8000 10000 12000 1 10 100 1000 Spill amount Spill amount (million gallons) Spill frequency Year Spill frequency Deepwater Horizon spill Hurricane Katrina • • • • :
poster
PE5 Changes - Spoke 4. D2 – Data Management Plan. https://doi.org/10.5281/zenodo.7977103 Producing FAIR research data within a NRRP project: PE5 Changes - Spoke 4 The Italian National Plan for Open Science 2021-2027 sets out four objectives related to scientific research data. The first is to contribute to the fulfilment of the FAIR data paradigm in the Italian research system. As part of the National Recovery and Resilience Plan, the project PE5 Changes - Spoke 4: virtual tehcnologies for museums and art collections, led by the University of Bologna, is applying the FAIR data paradigm to Cultural Heritage (CH) data. The activities of Spoke 4 focus on the impact of digital cultural heritage. Its objective is to make the digital enhancement of cultural heritage a permanent and widespread practice in museums and art collections, to increase the knowledge and organization of artifacts, expand public involvement, improve accessibility, inclusiveness, participation, and sustainability. This is achieved by involving different museums and collections (high-density and innovative museums, natural history and scientific museums, widespread art galleries, site museums, historical palaces, demo-ethnic-anthropological museums) that work as pilots for best practices to be adapted and reused in similar institutions. The collaboration with Cultural Heritage Intitutions is indeed one of the main strenghts of the project. Each Spoke partner and CH institution involved has: Bianca Gualandi - Research Division (ARIC), University of Bologna, Bologna, Italy Silvio Peroni - Dept of Classical Philology and Italian Studies, University of Bologna, Bologna, Italy identified someone internally responsible for the creation and management of data set up an internal data governance, including guidelines for data storage and deposit in a data repository identified intellectual property rights and selected a reuse license for CH data identified and described a number of planned datasets, describing e.g., their scope, formats and technologies used transmitted all this information to the Spoke leaders in order to coalesce on a common Data Management Plan, to serve as project documentation and as roadmap Typically someone senior, who is involved in the digitisation process Piano nazionale di digitalizzazione del patrimonio culturale. https://docs.italia.it/italia/icdp/icdp-pnd- dmp-docs/it/v1.0-giugno-2022/attivita-preliminari.html Some decisions are necessarily provisional but the DMP is a living document and will be updated! Working with CH data poses legal, technological and organisational challenges, both for researchers and for CH institutions. Planning their management from an early stage ensures that CH data are usable, interoperable, and last long-term.
poster
●The use of the ML prediction shows a better identification on the blobs that are likely to be a photon pair increasing the purity of the final selection. Coming soon: ●Differential cross section for muo, pion and neutrino kinematic variables, also for interaction variables Q2 and W will be reported. ●Comparisons to GENIE-based simulation. ●The effects of pion final-state interactions on these cross-sections. ●Neutral pions can be misidentified as electrons. ●Cross-section for Delta production since Delta production with an absorbed neutral pion would appear as quasielastic-like. ●Resonance production is the most abundant interaction mechanism in DUNE energy. ●Pion kinematics is a background for neutrino oscillations analysis. ●FSI simulation models can be improved by analyzing the CC(𝜋) scattering on different nuclei. Barbara Yaeggy for the MINER𝝂A Collaboration - Fermilab Neutral Pion Reconstruction in CH at < E𝝂 > ~ 6 GeV UNIVERSIDAD TECNICA FEDERICO SANTA MARIA MINER𝜈A Experiment Neutral Pion Reconstruction A pair of Electromagnetic Showers as gamma candidates produced via 𝜋0→ 𝛾𝛾 ●Clusters: nested set of hits in a single plane in the same time slice. ●The available clusters are grouped into conical regions about the neutrino interaction vertex (Bayesian Algorithm). ●With the assumption being that they arise from an electromagnetic shower originating at the neutrino interaction vertex. PhotonScore: ●Based on a score Z via log likelihood radio (LLR) for the dE/dx profile for charged pions and neutral pions. ●Events having a score less than 2 are likely to be have a photon, those are selected. Summary References Acknowledgments Selection: ●All the selected events require one and only one neutral pion. ●The invariant mass of the selected gammas is used as a feature for selecting the neutral pion. ●The pair of gammas are most often reconstructed into a 𝜋0 candidate between 60 < m𝛾𝛾 < 200 MeV. This document was prepared by members of the MINERvA Collaboration using the resources of the Fermi National Accelerator Laboratory (Fermilab), a U.S. Department of Energy, Office of Science, HEP User Facility. Fermilab is managed by Fermi Research Alliance, LLC (FRA), acting under Contract No. DE-AC02-07CH11359. These resources included support for the MINERνAconstruction project, and support for construction also was granted by the United States National Science Foundation under Award No. PHY-0619727 and by the University of Rochester. Sup-port for participating scientists was provided by NSF and DOE (USA); by CAPES and CNPq(Brazil); by CoNaCyT (Mexico); by Proyecto Basal FB 0821, CONICYT PIA ACT1413, Fondecyt 3170845 and 11130133 (Chile); by CONCYTEC (Consejo Nacional de Ciencia, Tecnología e Innovación Tecnológica), DGI-PUCP (DirecciÛn de Gestión de la Investigación - Pontificia Universidad Católica del Peru), and VRI-UNI (Vice-Rectorate for Research of National University of Engineering) (Peru); and by the Latin American Center for Physics (CLAF); NCN Opus Grant No.2016/21/B/ST2/01092 (Poland); by Science and Technology Facilities Council (UK). We thank theMINOS Collaboration for use of its near detector data. Finally, we thank the staff of Fermilab for support of the beam line, the detector, and computing infrastructure. [1] O. Altinok et al.[MINERvA], Phys. Rev. D 96, no.7, 072003 (2017).Doi:10.1103/PhysRevD.96.072003 [arXiv:1708.03723 [hep-ex]] [2] G. Perdue et al. [MINERvA], JINST 13, no.11, P11020 (2018) doi:10.1088/1748-0221/13/11/P11020. [arXiv:1808.08332 [physics.data-an]] [3] Jonathan Long and Evan Shelhamer and Trevor Darrell “Fully Convolutional Networks for Semantic Segmentation," CoRR abs/1411.4038 (2014) [4] C. Adams et al. [MicroBooNE], Phys. Rev. D 99 , no.9, 092001 (2019) doi:10.1103/PhysRevD.99.092001 [arXiv:1808.07269 [hep-ex]] [5] C. Adams et al. [MicroBooNE], JINST 15, no.02, P02007 (2020) doi:10.1088/1748-0221/15/02/P02007 [arXiv:1910.02166 [hep-ex]] [6] D. Coplowe et al. [MINERvA], [arXiv:2002.05812 [hep-ex]] Motivations Ba
poster
Bacterial cyclooxygenase homologs might be associated with multicellularity Georgy Kurakin1 and Nadezhda Potapova2 1Pirogov Russian National Research Medical University, Moscow, Russia, email: georgykurakin@gmail.com ❑Lipoxygenases (LOXs) could be evolutionarily associated with multicellularity (Paper 1, Paper 2). ❑Some bacteria might use LOXs for cross-kingdom pathogenicity and immune response evasion. ❑Oxylipins are almost universal signalling compounds across all multicellular eukaryotes. ❑Oxylipin biosynthesis is provided by homologous lipoxygenase- like and cyclooxygenase-like enzymes in different kingdoms. What did we know? What was recently found? What about cyclooxygenase-like (COX-like) enzymes? Are they associated with multicellularity? NO- 2 ❑We constructed a series of phylogenetic trees of bacterial cyclooxygenase homologs, animal cyclooxygenases (COXs), plant α-dioxygenases (αDOXs), and fungal COX homologs. ❑Some of these enzymes perform different function, but a biochemically characterized Nostoc COX-like enzyme with a lipoxygenase activity (in the green cluster) indicated a supercluster of bacterial enzymes with probable oxylipin biosynthesis finction. We called it “Our cluster”. 2Kharkevich Institute of Information Transmission Problems RAS, Moscow, Russia, email: nadezhdalpotapova@gmail.com LOX Alginate lyase Pseudomonas aeruginosa LOX LOX LOX Cyp 450 FA desaturase Myxobacteria COX Nitrosomonadales Cyanobacteria Cyp 450 24 15 68 Confidences for Nostoc punctiforme Confidences for Myxococcus xanthus 36 Confidences for Nitrosomonas europaea 1 Confidences for Pseudomonas aeruginosa Indirect extra-phylogenetic confirmation of the close proximity of animal COX and nitrosomonadal homologs: LOX and COX-like protein in one operon in Nitrosomonadales and their functional synergy in animals. (OperonDB data). COX1 ovine ppoC Asp. Fum. αDOX COX-like Nostoc Myeloperoxidase V116 R V L R120 L L L F205 I M F F209 I W F Y348 Y W Y V349 L T I L352 I L I Y355 L T Y N375 N Y N I377 C L M F381 F F F F Y385 Y Y Y F W387 W M W Y M522 I A V I523 A F A G526 Y F I A527 T L G S530 A A A G533 L R Q L534 V L A E-val ~10-11 ~10-4 N.S.S. Substrate-binding residues are conserved between COXs ans Nostoc COX-like protein, but evolved independently in αDOXs and fungal COX-like enzymes. COX PDB ID: 1diy This is confirmed by the substrate placement in COX and αDOX — it is almost opposite (PDB IDs: 1diy, 4kvl, UCSF Chimera model). αDOX PDB ID: 4kvl COX PDB ID: 1diy Nostoc enzyme Model, docking In contrast, the docking results suggest that the bioclemically characterized Nostoc enzyme might bind the substrate like COX (AutoDock Vina docking, UCSF Chimera model). conserved carboxylate- binding Tyr Corynebacteriales Burkholderiales Caulobacterales Rhizobiales Rhodobacterales Oceanospirillales Nitrosomonadales Sphingomonadales Rhodospirillales Geodermatophilales Propionibacteriales Streptosporangiales Methylococcales Alteromonadales Synechococcales Nostocales Chroococcales Oscillatoriales Streptomycetales 0 2 4 6 8 10 12 COX-like enzyme occurrence ❑We have obtained the indirect evidence that “our cluster” proteins are oxylipin-producing enzymes of bacteria closely related to our COXs. ❑The occurrence of “our cluster” enzymes (at the left) is at the highest in filamentous cyanobacteria and Streptosporangiales actinobacteria. They are multicellular. ❑This leads to a conclusion that COX-like enzymes in bacteria are statistically associated with multicellularity (like LOXs). ❑They might be involved in oxylipin cell-to-cell signalling for maintaining complex structure, like human COXs and plant αDOXs. Find out more in our slide deck on ResearchGate: Kurakin, G.F., Potapova, N.A. (2023) Insights into the cyclooxygenase phylogeny. Unpublished. https://doi.org/10.13140/RG.2.2.35390.46402 Methods and full references are included. ResearchGate (@ResearchGate) | Twitter BioRender - Crunchbase Company Profile Funding Created with BioRender.com We used Tree visuali
poster
The HADES Program with HARPS-N@TNG HADES: THE HArps-n red Dwarf Exoplanet Survey L. Affer1 and HADES Team, 1 INAF – Osservatorio Astronomico di Palermo, Piazza del Parlamento 1, 90134 Palermo, Italy - laura.affer@inaf.it Many efforts to detect Earth-like planets around low-mass stars are currently devoted to almost every ex- tra-solar planet search. M dwarfs stand as ideal targets for Doppler radial velocity searches as their low masses and luminosities make low-mass planets orbiting within their habitable zones more easily detectable than those around higher-mass stars. Nonetheless, the statistics of the frequency of this kind of planet hosted by low-mass stars remains poorly constrained. Our M-dwarf radial velocity monitoring with HARPS-N within the HARPS-N Red Dwarf Exoplanet Survey Radial Velocity (HADES) project started in 2012 and is contributing to the widening of the current statistics through the in-depth analysis of accurate radial velocity observations in a narrow range of spectral sub- types from M0 to M3, to investigate the planetary pop- ulation around a well-defined class of host stars. The HADES project is the result of a collaborative effort between the Italian Global Architecture of Plane- tary Systems (GAPS) Consortium, the Institut de Cièn- cies de l’Espai de Catalunya (ICE), and the Instituto de Astrofísica de Canarias (IAC). Figure 1: Overview of the HADES detected planetary sys- tems.The sample’s published planets are shown as red circles: the symbol size is proportional to the minimum planetary mass. Each sys- tem’s Habitable Zone conservative and optimistic limits, are shown as thick dark green and light green bands, respectively (from [3]). Two photometric programs regularly and almost simultaneously follow up the sample of M stars to char- acterize the stellar activity, to highlight periods that are due to chromospheric inhomogeneities modulated by stellar rotation and differential rotation, and thus to dis- tinguish from the periodic signals those due to activity and to the presence of planetary companions. We present the complete analysis of the HADES survey and the results obtained concerning the statistical ([1], [2], [3]), activity ([4], [5], [6]), and characterization ([7], [8]) part as well as the planet revealing part ([9] to [18]), around M dwarfs. Figure 1: Known radial velocity planets (planetary mass vs. or- bital period diagram) around M dwarfs (as listed at http://ex- oplanet.eu/ in December 2020. Planets discovered by the HADES sur- vey are shown as red triangles (from [18], the planet GJ 9689 is shown as a black star). References: [1] M. Perger et al. (2017) A&A, Volume 598, id.A26. [2] E. Gon- zález-Álvarez et al. (2019) A&A, Volume 624, id.A27. [3] M. Pina- monti. et al. (2022) A&A, Volume 664, id.A65. [4] J. Maldonado et al. (2017) A&A, Volume 598, id.A27. [5] G. Scandariato et al. (2017 A&A, Volume 598, id.A28. [6] A. Suárez Mascareño et al. (2018) A&A, Volume 612, id.A89. [7] J. Maldonado et al. (2015) A&A, Vo- lume 577, id.A132. [8] J. Maldonado et al. (2020) A&A, Volume 644, id.A68. [9] L. Affer et al. (2016) A&A, Volume 593, id.A117. [10] A. Suárez Mascareño et al. (2017) A&A, Volume 605, id.A92. [11] M. Perger et al. (2017) A&A, Volume 608, id.A63. [12] M. Pinamonti et al. (2018) A&A, Volume 617, id.A104 [13] L. Affer et al. (2019) A&A, Volume 622, id.A193. [14] M. Perger et al. (2019) A&A, Volume 624, id.A123. [15] M. Pinamonti et al. (2019) A&A, Volume 625, id.A126. [16] B. Toledo-Padrón et al. (2021) A&A, Volume 648, id.A20. [17] E. González-Álvarez et al. (2019) A&A, Volume 649, id.A157. [18] J. Maldonado et al. (2021) A&A, Volume 651, id.A93.
poster
References: Ohyama, Terashima, & Sakamoto 2015, ApJ, 805, 162 ● Sakamoto et al. 2014, ApJ, 797, 90 ● Sanders et al. 2003, AJ, 126, 1607 ● Stierwalt et al. 2013, ApJS, 206, 1 ● Sun et al. 2018, ApJ, 860, 172 ● ALMA Project Code 2015.1.00714.S 10 1 10 0 10 1 Peak TB (K) 0.0 0.5 1.0 1.5 Mass Fraction per dex Centers Non-centers Combined 10 1 10 0 10 1 10 1 10 0 10 1 PHANGS PHANGS NGC 3256 Build GMC catalog in NGC 3256 to compare with spirals ● Sizes ● Masses ● Boundedness ● CO conversion factor (expect LIRG to be low and from measurements XCO ∝ Σ / (𝜎TB), Narayanan+ '11) Replicate pixel-based and cloud-based analyses in new maps of entire NGC 4038/9 merger Next steps NGC 3256 additionally split into northern (triangles) and southern (stars) nuclei. Vertical lines show 𝛼vir of 1 and 2. Per-galaxy, mass-weighted, inner 68% of pixel value distributions. Squares are centers, circles are non-central regions. Mass-weighted densities of pixels from 120 pc resolution maps. Initial PHANGS sample for comparison (Sun+ '18). 10 3 10 6 10 9 Pturb ∝ Σ 𝜎 2 10 0 10 1 𝛼vir ∝ 𝜎 2Σ-1 10 0 10 1 10 2 𝜎(km s 1) 10 0 10 2 10 4 Σ ☉ (M pc 2) 10 0 10 1 10 2 10 3 10 4 ☉ Σ (M pc 2) 10 0 10 1 10 2 𝜎(km s 1) NGC 3256 M31&M33 PHANGS Sample NGC 4038/9 Comparison to "normal" galaxies 10 h27 m53.0 s 51.5 s 50.5 s -43°54'05" 10" 15" 20" 25" 30" RA (J2000) Dec. (J2000) 1 kpc 53.0 s 51.5 s 50.5 s RA (J2000) 1 kpc 0.5 1.0 1.5 2.0 2.5 Σ (10 4 ☉ M pc 2) 30 60 90 120 150 𝜎(km s 1) Bulk dense gas surface density and velocity dispersion ● Channel width: 5 km / s ● Per-channel RMS: 0.9 mJy / beam ● Integrated RMS: 80 mJy / beam km / s (34 M☉ / pc2) ● Milky Way conversion factor used here Marked regions ● Elliptical 1 kpc projected distances from nuclei ● Rectangles around southern nucleus jet CO (2-1) maps ● Data cover central 7 kpc at 235 GHz ● CO (2-1) and CS (5-4) observed with 2.5 km/s resolution ● Continuum bandwidth: ~7 GHz ● Two main array configurations ● ACA and total power data were obtained ● Interferometric maps sensitive to scales from 53 pc to 6200 pc ALMA observations HST optical and IR view of NGC 3256. Square shows ALMA FoV (Evans+). ● Distance: 44 Mpc (0.2" = 43 pc) ● LIR = 6.5 x 1011 L☉ (Sanders+ '03) ● Ideal LIRG to study GMC-scale star formation ● Late-stage merger (Stierwalt+ '13) ● Starburst driven outflow in face-on northern nucleus (v ~ 750 km / s) ● Jet in edge-on southern nucleus has AGN contribution (Sakamoto+ '14, Ohyama+ '15) NGC 3256 ★brunettn@mcmaster.ca Nathan Brunetti★ & Christine D. Wilson Molecular gas in nearest LIRG denser and more turbulent than local disks
poster
Introduction Valorizing Agro-Industrial Waste: Exploring Antimicrobial Peptides Prepared from Feather Keratin Justine Horner,a Amira Ben Mansour,a Sutida Jansod,a Claire Gay,a Roger Martia* a University of Applied Sciences Western Switzerland (HES-SO), Haute Ecole d’Ingénierie et d’Architecture de Fribourg (HEIA-FR), Boulevard de Pérolles 80, CH-1700 Fribourg, Switzerland Results and outlooks Microbial colonization on surfaces contributes to infections, imposing significant health and economic burdens. Biobased alternatives to commercial antimicrobial coatings are increasingly sought after for sustainable solutions. Antimicrobial peptides (AMPs), biomolecules found in various natural sources, are known for effectively inhibiting microbial growth. Bioactivity of peptides are associated with their structural characteristics. Therefore, our strategy was to generate a large pool of antimicrobial candidates, by a large screening of various preparation methods, combining sometimes several hydrolysis approches The RELIANCE consortium consists of 15 partners spanning 8 EU and 2 non-EU countries. Partners include research institutions, universities, SMEs, and large industries. Acknowledgements Funded by the European Union. Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or the European Health and Digital Executive Agency (HADEA). Neither the European Union nor the granting authority can be held responsible for them. This work is part of a project that has received funding from the European Union’s Horizon Europe research and innovation program under grant agreement No. 101058570 (RELIANCE). This work has received funding from the Swiss State Secretariat for Education, Research and Innovation (SERI). The goal of this project is to explore keratin- based AMPs from feather waste, a sustainable and abundant source, to produce contact-killing coatings for self- disinfecting surfaces. Various extraction and purification strategies were employed to generate a wide range of AMP candidates with optimal bioactivity. Peptides extraction Peptides size also matters. That’s why ultrafiltration has been used not only as a purification methods, but also in a way of separating peptides mixtures into fractions of several mass intervals. From feathers to antimicrobials Feather Waste Contact killing antimicrobial peptides Keratin hydrolysates Peptides mixtures Surface functionalization The project now advances to the next phase: coupling keratin-based antimicrobial peptides with nanoparticles, which will be applied to surfaces via cold atmospheric plasma deposition to create self-disinfecting materials. Keratin structure Feathers are mainly composed of keratin, a mechanically strong structural protein, insoluble in water, and resistant to degradation. To break down this protein into a mixture of peptides, different types of bonds must be broken : • Disulfide (S-S) bonds that contribute to the strength of keratin • Some peptide bonds that constitute the backbone of this molecule, but not all of them. Screening of peptides preparation methods In accordance with the process flow delineated on the left, more than 50 distinct samples encompassing various keratin peptide fractions were produced, in quantities ranging from a few milligrams to over 1 gram. Antimicrobial activity Preliminary analyses have shown promising antimicrobial activity of keratin-derived peptides in solution, particularly against viruses and Gram-negative bacteria. Process scale-up Production has been scaled-up 20 x for one of the most bioactive peptides samples, revealing different technical challenges, in particular for the work-up stages, but generating yields very close to the equivalent small scale experiments (around 40% of purified peptides regarding the mass of feather feed). cys-S-S-cys CAP DEPOSITION CHEMICAL COUPLING Nanoparticles AMP Functionalized nanoparticles + Self disinfectant surfaces t = 0 t
poster
The observation and future prospect of temperature gradient in the shock of SN1006 T. Kasuga, H. Odaka, A. Bamba, Y. Kato (The University of Tokyo), S. Katsuda (Saitama University), H. Suzuki (Konan University),K. Nakazawa (Nagoya University) [1]Vink, J., Laming, J. M., Gu, M. F., Rasmussen, A., and Kaastra, J. S., “The Slow Temperature Equilibration behind the Shock Front of SN 1006”,The Astrophysical Journal, vol. 587, no. 1, pp. L31‒L34, 2003. doi:10.1086/375125. [2] Yamaguchi, H., Eriksen, K. A., Badenes, C., et al. 2014, ApJ, 780, 136. doi:10.1088/0004- 637X/780/2/136 [3] http://x-ifu.irap.omp.eu/x-ifu/key-capabilities Introduction We find that the electron temperature increases toward downstream in the northwestern shock-heated plasma of SN 1006. Athena observation expects to show us the spatial variation of ion temperature. This will lead to our understanding of the shock heating and relaxation mechanism in SNRs. Our recent study with Chandra : Electron temperature of SN1006 NW shock Data set • ObsID 1959 (April, 2001, 88.98 ks) • ObsID 13737 (April, 2012, 87.09 ks) Region size : 140” ×15” Fig.4 Comparison of observed temperatures to a Coulomb scattering model ATHENA study : Ion temperature Our objective : Understanding the mechanism of collisionless heating and thermal relaxation after heating Fig.3 The spectra and besf-fit models. These spectra imply that the inner region, layer 1, has higher temperature than the outer region, layer 4 10−3 0.01 0.1 1 normalized counts s−1 keV−1 0.6 0.8 1 1.2 1.4 1.6 1.8 2 −2 0 2 4 (data−model)/error Energy (keV) 1959 - layer1 - layer4 13737 - layer1 - layer4 Result • Electron temperature increases toward inner region • Electron temperature increase is estimated to be slower than Coulomb scattering Fig.2 Chandra X-ray image of the northwestern region of SN 1006 with our analysis regions We want to observe the variation of ion temperature Why? → To understand shock heating and energy transfer system Near the shock front → Evaluate the efficiency of shock heating Inner region → Comparison to Coulomb scattering How? → From the width of each ion emission line ➡︎ Need observation with good energy and spatial resolution of Athena Spectral fitting Model : phabs × vnei kT, nt and some abundance are free Simulated Athena spectrum ・Using the best-fit model of center layer 1 ・Exposure time of 10 ks in total ・Smoothing spectrum with Gaussian to reproduce the line width derived from ion temperature from Rankin-Hugoniot relation Ne ( ) → Mg ( ) → Si ( ) → ・Replacing line emission by some gaussian, linking the width of each gaussian to show the same temperature Estimated temperatures : Ne → Mg → Si → ➡︎ Athena observation can distinguish mass-proportional ion temperature 307 keV 3.66 × 10−3 keV 370 keV 5.33 × 10−3 keV 428 keV 7.28 × 10−3 keV 335+15 −13 keV 404+258 −113 keV 566+438 −189 keV Reference Conclusion masahiro.ichihashi@phys.s.u-tokyo.ac.jp Masahiro Ichihashi (The University of Tokyo) Particle temperatures just after the collisionless heating are thought to be different between electron and ion[1] Ideal → (Rankine-Hugoniot relation) Several studies suggest breaking of this relation in the electron-to-proton ratio[2] The energy transfer of particles is caused after collisionless shock heating The simplest way → Coulomb scattering However, no observational evidence exist ➡︎Need to evaluate absolute values of each particle’s temperature by observation kTi = 3 16 miv2 s We sliced the shocked region to eaxmine the shock heeting mechanism. Fig.1 Schematic diagram of particle heating near shock wave shock front ISM vs post shock region Ion Electron Collisionless heating Equilibrium state kTi = 3 16 miv2 s (ideally) Energy transfer cold hot Fig.5 Simulation results of the Athena X-IFU spectra and Gaussian replacement of Ne line 0.01 0.1 1 10 counts s−1 keV−1 data and folded model 0.85 0.9 0.95 1 1.05 0 5 10 (data−model)/error Energy (keV) ichihashimasahiro 25−Oct−2022 16:58
poster
Multi-scale computational modelling of neuronal dynamics in genetic epilepsies Rosch RE1,2, Heilbron M1, Peters C3, Ruben P3, Lim M4, Pal D4,5, Goyal S6, Hughes E4, Baldeweg T2, Friston KJ1 1Wellcome Trust Centre for Neuroimaging, University College London. 2Developmental Neurosciences Programme, University College London. 3Department of Biomedical Physiology and Kinesiology, Simon Fraser University, Burnaby, BC, Canada. 4Department of Paediatric Neurology, Evelina London Children’s Hospital. 5Department of Basic and Clinical Neuroscience, King’s College London. 6Department of Clinical Neurophysiology, Evelina London Children’s Hospital. Richard E Rosch BMBCh r.rosch@ucl.ac.uk dynamic-brains.com roschkoenig τ3 Thalamic Reticular Thalamo- cortical Intrinsic Thalamic τ4 h3 h1 h4 h2 A τt1 τt2 τt3 RF LF LT RT Th Frontal Cortical Thalamic Temporal Cortical B Neuronal models bridge observational scales in neurological disorders with genetic causes SCN1A channel gating abnormalities Network changes in GRIN2A-epilepsy Single SCN1A mutant causes loss of function and gain of function Sleep dynamics associated with GRIN2A are explained by thalamic microcircuitry Multi-scale models of dynamic pathology reveal links between genotype and phenotype • Exonic deletion identified in GRIN2A gene in a child with acquired aphasia and EEG features suggestive of focal epilepsy • Overnight sleep EEG was performed using telemetry setup to monitor for continuous spike-wave discharges in sleep (CSWS) as observed in affected sibling Can network modelling identify possible causes for abnormal sleep dynamics in this patient? fMRI Developmental Period EEG / MEG Animal Models Cell Models Genetics SCN1A ECOG / SEEG microscale macroscale mesoscale GRIN2A Spatial Scale Modelling describes hidden states • SCN1A-mutation (A1273V, AV) identified in a child with diagnosis of Dravet-spectrum epilepsy, onset at 6 months of age (Fig 1) • Patch clamp measurements in CHOk1 cells at 37 and 40ºC were used to estimate steady state activation, fast inactivation, and fast inactivation recovery curves Can changes of channel dynamics in this patient explain the temperature-sensitive phenotype? Fig 1: SCN1A variant. The patient had heat-induced seizures and a de novo mutation in a transmembrane domain of the SCN1A gene. Fig 3: Simple corticothalamic network model for sleep dynamics. Using dynamic causal modelling, EEG dynamics are explained with a network of cortical (not shown) and thalamic (A) neural mass models connected in a simple distributed sleep network . • Empirical measurements alone often focus on macro- or microscale abnormalities • Emergent properties may be better understood in terms of mesoscale features at the level of neuronal circuits • Computational models allow bottom up or top down inference based on empirical data Fig 4: Hodgkin Huxley model of abnormal cortical action potential dynamics. Empirical patch clamp measures were implemented in HH model for wild type (left) and variant (right) SCN1A, for normal (black) and febrile (red) temperatures. Bottom panels show bifurcation analysis for varying input currents. The AV variant can sustain firing at higher input currents than the wildtype, especially at high temperatures. Fig 2: Empirical techniques. Currently used empirical techniques allow inference on restricted developmental and spatial scales only • A single variant in SCN1A can explain both loss of function (reduced firing frequency), and gain of function (increased tolerance to high inout states) • High temperatures cause distinct gating patterns that cause reduction in depolarision block during febrile temperatures, which may relate to seizure disorder Fig 5: Sleep dynamics can be reduced to limited set of model parameters. Different sleep stages have distinct spectral signatures (A). The differences in spectra between sleep stages can be summaries in a subset of parameters identified using Bayesian Model Comparison (via free energy, B). Two thalamic
poster
LexFCS – Extending the Federated Content Search for Lexical Resources Contributions • (backwards) compatible extension of FCS for lexical resources, • “lexical resources” capability to indicate endpoint support, • LexCQL – query language based on Contextual Query Language (OASIS/Library of Congress) with constraints for searchable fields, operators and relation modifiers, • Lex Data View – key-value based result serialization, (optional) LexHits Data View as extension of BASIC Hits Data View with inline annotations. • First prototypes in Text+ including client and endpoint implementations, custom visualization in FCS Aggregator Query Language • Based on Contextual Query Language with constraints • Boolean operators: AND, OR, NOT • Relations: =, ==, is with Modifiers: unmasked, lang ignore/respectCase, ignore/acceptAccents, honorWhitespace, regexp, partial/fullMatch • Indexes (searchable fields): based on data model, with additional entry-based language field “lang” • Examples • lemma = "car" · car · "car" · "car wash" • pos = "NOUN" AND synonym = "house" • lang = "deu" AND translation =/lang=eng "member of parliament" • pos is https://universaldependencies.org/u/pos/NOUN Motivation • CLARIN Federated Content Search (FCS) focussed on (annotated) full- texts, corpora or similar flat text structures → unable to include more structured information, especially lexical resources like dictionaries, word nets and graphs • Difficulties for querying and displaying lexical resources in flat structures without loss of information → LexFCS extension required Data Model • Key-value based Lemma entries • Entry: a single result, with optional language info • Field: set of values grouped by type, e.g., entryId, lemma, phonetic, translation, transcription, definition, etymology, case, number, gender, pos, segmentation, sentiment, antonym, hyponym, hypernym, meronym, holonym, synonym, subordinate, superordinate, related, ref, senseRef, citation • Value: actual “content”/value with attributes for additional context, e.g., xml:id, xml:lang, langUri, preferred, ref, idrefs, vocabRef, vocabValueRef, type, source, sourceRef, date • Natural serialization in Lex Data View (XML) Entry Field Value Value Value Type Attribute Attribute Attribute Field Value Type Attribute Attribute Example from https://corpora.wortschatz-leipzig.de/en/res?corpusId=deu_news_2023&word=Auto Example from https://https://thesaurus-linguae- aegyptiae.de/lemma/104730?lang =en Erik Körner, Uwe Kretschmer koerner@saw-leipzig.de The NFDI consortium Text+ is funded by the Deutsche Forschungsgemeinschaft (DFG) – project number 460033370 Next Steps • Collecting more feedback about • Lex Data Model, • Requirements and use-cases of users and data providers, • User interface and visualization • Finalizing LexFCS specification proposal with CLARIN FCS Task Force TPPSSI demonstrator https://tppssi-demo.saw-leipzig.de/lex/ Specification draft and examples https://gitlab.gwdg.de/textplus/ag-fcs-lex-fcs-dataview Features Overview and Demo • Grammar with optional speech playback • Hierarchical structure of e.g. definitions • Highlighting of relations between values using xml:id and idrefs attributes, e.g., for definitions, examples and senses • Citations, examples with date and source / reference • Internationalization (user interface, with vocab(Value)Ref translation of POS Examples from https://tppssi-demo.saw-leipzig.de/lex/
poster
U.S. Environmental Protection Agency Office of Research and Development EPA www.epa.gov EVALUATION OF 1066 TOXCAST CHEMICALS IN A HUMAN STEM CELL ASSAY FOR DEVELOPMENTAL TOXICITY TB Knudsen1, AM Richard1, PG Kothiya1, RS Judson1, KA Houck1, KC Crofton1, RS Thomas1, S Little1, ES Hunter2, NC Baker3, CKM Leung1,4, JA Palmer5, AM Smith5, MR Colwell5, PR West5, RE Burrier5, LA Egnash5 1. National Center Computational Toxicology, US EPA, Research Triangle Park, NC 27711 (RTP), 2. NHEERL, US EPA-RTP, 3. Lockheed Martin-RTP, 4. ORISE-RTP, 5. Stemina Biomarker Discovery, Inc., Madison, WI 53719. A. INTRODUCTION B. METHODS C. METRICS OF ASSAY QUALITY Thomas B. Knudsen l knudsen.thomas@epa.gov l 919-541-9776 EPA’s ToxCast program has generated data on a battery of 821 in vitro endpoints for 1066 compounds including pharmaceuticals, natural products, pesticidal active ingredients, consumer use chemicals and industrial ingredients [1]. To increase the diversity of in vitro assays used to assess developmental toxicity, the ToxCast library was evaluated in the Stemina ‘devTOX quickPREDICT’ (qP) platform [2]. This assay measures two small molecules (ornithine, cystine) in medium conditioned by human embryonic stem (hES) cells yielding an ornithine:cystine ratio (o/c ratio) indicative of an imbalance in metabolism predictive for teratogenicity in a human system. Here, we provide a preliminary evaluation of the results focusing on metrics of assay quality, performance, and predictivity. Platform: Metabolomic analysis of the hES cell secretome for predictive developmental toxicity (devTOX platform) was reported in 2010 [3]. A 2011 pilot study conducted with 11 ToxCast chemicals predicted developmental toxicity in concordance with animal data with 83% accuracy [4]. In 2013, the Stemina ‘devTOX-qP’ platform was developed as a high throughput screening (HTS) assay for developmental toxicity testing [2]. The model was trained with 23 pharmaceuticals (96% accurate). An independent 13 pharmaceutical test set with known (human) teratogenicity was 77% accurate. Dosing: H9 cells (WA09 line, WiCell Research Institute) were cultured in 96-well plates. Each experimental plate included methotrexate (MTX) reference controls as calibration standards for negative- (5 nM) and positive- (1 uM) response as well as media blanks and 0.1% DMSO vehicle. Undifferentiated cells were exposed for 72h to test compound (blinded and in triplicate) with media and test compound replacement every 24h; maximum test concentration (MTC) for single concentration screen and/or 8-point conc. series set at 1-, 10- or 100-uM based on ToxCast cytotoxicity burst (TC-Cyto-Burst) [1] or compound availability. Evaluation: Cell-conditioned media from the final 24h treatment period was analyzed by LC- MS to determine ornithine/cystine (o/c) ratio. Concurrent cell viability was assessed with the CellTiter-FluorTM assay (Promega). The cytotoxicity Relative Fluorescence Unit (RFU) was background corrected and normalized to mean RFU of the neutral control (0.1% DMSO). Teratogen Index (TI) was defined by the o/c ratio, using the default threshold value < 0.88 and concurrent cell viability (RFU values for test compound relative to DMSO control). D. METRICS OF ASSAY PERFORMANCE and PREDICTIVITY LEFT • DMSO control (n=415) • MTX-negative reference (5 nM, n=208) • MTX-positive reference (1 uM, n=207, 210) • MTX-sample in ToxCast (3 nM, 1 uM, n=1) RIGHT • MTX-sample crosses TI threshold at 0.38 uM • ToxCast cytotox burst for MTX = 0.69 uM Quality Standards. Methotrexate (MTX) in the ToxCast library (blinded) gave ornithine/cystine (o/c ratio) and cell viability (cv) measures identical to the calibration standards. DISCLAIMER: The views expressed in this article are those of the authors and do not necessarily reflect the views of policies of the USEPA. Mention of trade names or commercial products does not constitute endorsement or recommendation for use. Replicate Samples. Concentration (8-point) resp
poster
RESEARCH POSTER PRESENTATION DESIGN © 209 www.PosterPresentations.com Panama is known for its remarkable ”four distinct oceans”, driven by seasonal and stationary changes in physical and chemical conditions in the Caribbean side, the Gulf of Chiriqui, and the Gulf of Panama by the dynamic position of the Intertropical Convergence Zone (ITCZ) each year (Fig. 1). In this case, the Gulf of Panama, is part of a climate zone known as the Panama Bight, influenced by the propagation of the ITCZ due to the Bermuda high. During its southernmost position at 1°N, from December to April, the north trade winds from the Bermuda high, pass through the lower elevations of the isthmus, inducing a sea level drop by a subtle divergence of waters to the west. This process triggers the Coriolis effect, leading to the upwelling of deep-sea waters, resulting in a drop in sea surface temperatures (14° to 22° Celsius), increased salinity (SSS), a surge in photosynthetic organisms and nutrient-rich waters (nitrates and phosphates), and various other factors that support a commercially valuable fisheries industry on a large scale (Graph 1). This event concludes between May and October, coinciding with the boreal summer, when the ITCZ migrates to its northernmost position where the Choco surface jet winds dominate the area, rotating towards the northeast due to the Coriolis effect, halting the upwelling effect, letting water get to its average temperature (28° to 31° Celsius) while introducing substantial moisture inland and simultaneously, the SSS decreases due to an influx of freshwater, and the temperature rises (Graph 2). INTRODUCTION OBJECTIVES METHODOLOGY The Physical Monitoring Program (PMP) of STRI has established seven strategic sites along the Gulf of Panama (Fig. 2), an ongoing venture since 2017. These sites are dedicated to collecting data regarding the water column, tracking temperature variations, and more parameters. We cross-verify our on-site measurements with remote sensing data obtained from satellites, enabling detailed comparisons between upwelling and non-upwelling events. Our sites span a range of locations within the Bay of Panama, from Taboga Island close to the City of Panama, out to the Las Perlas archipelago. Each site is characterized by its maximum depth and distance from the possible anthropogenic effects of Panama City and Panama Canal Shipping: sites 1 and 2 (20 and 35 meters) are positioned farthest from the coast, and sites 3, 4, and 5 are located in the deepest waters (40 to 60 meters) between the coast and the Las Perlas Islands, sites 6 and 7 (25 to 27 meters) are located closest to the coast and Panama City. Sites 1 and 2 are the most productive and biodiverse of the seven sites—sites 6 and 7 are likely to be the most affected by Panama City and ocean-going traffic. Sampling was carried out using EXO 2 and 3 (Fig. 3) multiparametric sondes (Xylem YSI). Sondes were equipped with pH, temperature, DO, Chlorophyll, turbidity, and conductivity sensors) programmed to record every second as they were allowed to descend through the water column (pausing 15-30 seconds every meter) until they reached the bottom (Fig 4). At the end of every sampling trip, data were retrieved, transferred to an Excel file where they were examined for any anomalies, and then transferred to final storage. An Excel file is used to analyze and visualize trends and summary statistics related to depth, site, and date. In conjunction with our on-site measurements, we use Remote Sensing Systems, specifically the NOAA Daily Optimum Interpolation Sea Surface Temperature (OISST) Satellite, providing extensive data on Sea Surface Temperature (SST) anomalies dating back to 1981. We set the data ranging from a 7-day compilation of anomalies. We integrate data from the Copernicus Marine Service, featuring the Global Ocean Hourly Sea Surface Wind and Stress from Scatterometer and Model satellite data, offering wind variables, divergence, and curl (Fig. 5). 1
poster
Teff (K) Log  Teff(K) Log calibration for He I 10830 Teff(K) Log calibration for Hα Log calibration for Ca II IRT- a Analysis of chromospheric flux-flux relationships of M Dwarfs using visible and near-infrared CARMENES spectra F. Labarga1; D. Montes1; C. Duque-Arribas1; A. López-Gallifa1; J. A. Caballero2; S. V. Jeffers3; A. Reiners4; I. Ribas5; A. Quirrenbach6; P. J. Amado7; and the CARMENES Consortium 1 Dpto. Física de la Tierra y Astrofísica & IPARCOS-UCM (Instituto de Física de Partículas y del Cosmos UCM), Fac. CC Físicas, Univ. Complutense de Madrid, Spain 2 Centro de Astrobiología (CSIC - INTA), ESAC campus, camino bajo del castillo s/n, 28691, Villanueva de la Cañada, Madrid, Spain 3 Max Planck Institute for Solar System Research, Göttingen, Germany 4 Institut für Astrophysik, Georg-August-Universität, Göttingen, Germany 5 Institut de Ciències de l’Espai & IEEC-CSIC, Campus UAB, c/ de Can Magrans s/n, E-08193 Bellaterra, Barcelona, Spain 6 Zentrum für Astronomie der Universität Heidelberg, Landessternwarte, Heidelberg, Germany 7 Instituto de Astrofísica de Andalucía (IAA, CSIC), Glorieta de la Astronomía s/n, E-18008 Granada, Spain Fig. 1: Results of applying the spectral subtraction for selected chromospheric activity indicators (He I D3, Na I D1, D2, Hα and the first two of three Ca II IRT lines in the VIS) with the CARMENES spectra of J22468+443 (EV Lac) at the maximum level of chromospheric activity (Flare) and in a previous quiescent phase as well as the synthetic spectrum obtained with a M3.5 V reference star (J22096-046). The spectra to perform the subtraction are shown at the bottom and the obtained subtracted spectra at top from where we derived the EW of the chromospheric contribution. The algorithm makes use of the python code iSTARMOD as described in Labarga & Montes (2020). The spectral subtraction technique (Barden, SC; 1985, Montes et al., 2000) is well suited for the study of chromospheric activity in M-dwarfs, given that this type of stars does not have a well-defined continuum, especially in its late-types. The detailed analysis of the activity indicators is important from one side in order to confirm or discard all the possible planets around these stars and by the other studying its dependency with other stellar parameters as rotation, age and depth of the convective zone. This is the case for the stars of the CARMENES survey (Quirrenbach et al. 2020), where a WG on stellar activity, elaborating on its characterization in order to discount the jamming effect on stellar spectra of activity and rotation in the detection of exoplanet candidates. The CARMENES has the aim of detecting Earth-mass planets orbiting in the habitable zones of their host stars, focusing on over 300 M-type dwarf stars, spread over the complete M spectral range. For the work carried out in this study, the spectral subtraction technique (see Fig. 1) has been applied using the Python code iSTARMOD (Labarga & Montes, 2020). ABSTRACT: Exploiting the huge amount of data provided by the CARMENES survey, this work aims to study the chromospheric activity of M dwarfs based on a sub-sample (active RV-loud M dwarfs) of CARMENES GTO sample. Using the spectral subtraction technique and calibrations of the continuum-flux near the line of interest, the available information on the chromospheric activity flux is extracted. Most of the chromospheric indicators included in the spectral range of the spectrograph, ranging from visible (VIS) - that include the Na I D1&D2 He I D3, and Hα lines - to near-infrared (NIR) - that include the Ca II IRT, He I 10830, Paschenα and Paschenβ lines - are used. For the implementation of the spectral subtraction technique, a Python code (iSTARMOD) based on a previous FORTRAN one, formerly used by the research group, is used. The synthetic spectra for effective temperatures in the range [2400, 7000] K, allows through the calibrations a comparison of the flux-flux relationships with previous works performed
poster
SUMMARY METHODS This study implements a deep convolutional neural network with the great potential to recognize patterns of pollen phenomena that enable the prediction of pollen concentrations. The model is trained using data from 2009 to 2015 from multiple meteorological data sets, satellite data, and processed data reflecting pollen flux as input for the model. The model forecasts pollen counts one to seven days ahead for the entire year of 2016. The performance of the algorithm for pollen prediction was evaluated using statistical parameters and categorical statistics evaluation by comparing 1-7 forecast to observation. The algorithm obtains a relatively high index of agreement (0.81-0.90) and Pearson correlation coefficient (0.88-0.75). Evaluation of categorical statistics based on defined threshold levels show satisfactory results. Highlights • Model developed and tested for real-time daily forecast • Model can forecast daily pollen concentrations within minutes • Good accuracy for daily weed and tree pollen forecast for 2016 Pollen data has been measured and acquired from the Houston Department of Health and Human Services (HDHHS) archives. Pollen data contains daily pollen (count/m3) of 25 tree species, 15 weed species and one generalized grass species. We obtained meteorological data from the Texas Commission on Environmental Quality (TCEQ), which operates the Continuous Ambient Monitoring Sites (CAMS) in various metropolitan areas within the state of Texas. We selected data from CAMS station 695 (Moody Tower, near Downtown Houston) for its close proximity to the HDHHS pollen measurement station. Map of Study Area Fig. 1: Map of the study area in Houston. Moody Tower (Red) is the location of collected meteorology data. Pollen Station (Blue) is the location of collected pollen concentration data. A grey area (centered on the Pollen Station) represents the Leaf Area Index (LAI) based on MODIS data for pollen flux calculations. Real-time 7-Day Forecast of Pollen Counts Using Deep Convolutional Neural Network Yannic Lops1, Yunsoo Choi1 (ychoi6@uh.edu), Ebrahim Eslami1, Alqamah Sayeed1 1Department of Earth and Atmospheric Sciences, University of Houston, Houston, TX 77204, USA RESULTS Acknowledgement: We acknowledge the Texas Commission on Environmental Quality (TCEQ) and Houston Health Department for sharing their data. STUDY AREA RESULTS Deep CNN Forecasting System Fig. 2: Structure of the Deep CNN pollen forecasting system. Input data consist of pollen measurements, meteorological data, and processed data. Processed data consist of: • Meteorological adjustment factor (Ke) Tte, RHte, and WSte represent the threshold values for temperature, relative humidity, and wind speed respectively. C1, C2, and C3 are meteorological resistance factors. • Normal pollen distribution (Ce) Where d is the number of consecutive days which pollen measurements meet or exceed the pollen count, µ is the mean distribution, and σ is the standard deviation. • Characteristic concentration (C*) Where canopy height (hc) is the mean canopy height of the vegetation species. LAI is the computed mean Leaf Area Index from MODIS satellite image data for the respective time period. pq is set as ‘Pollen Count +1’. • Pollen flux (Fp) Where u* is the averaged frictional velocity Deep CNN Model Fig. 3: Simple representation of the Deep CNN model. For evaluation, the categorical statistics evaluation, and pearson correlation coefficient (r) and index of agreement (IOA) statistical evaluation methods are used. Each pollen category and days predicting ahead are evaluated. The categorical statistics evaluation consist of: • Hit Rate (HIT) • Critical Success Index (CSI) • False Alarm Rate (FAR) • Equitable Threat Score (ETS) • Proportion of Correct (POC) Categorical statistic evaluation from 4 quadrants: Na - Predictions above and observations below threshold Nb - Prediction and observation above threshold Nc - Predictions and observations below threshold Nd - Predi
poster
Title: An Overview of Data Science Courses in Hungary and the Philippines Authors: Joseph Yap, Doctoral Student, Eötvös Loránd University, Budapest, Hungary, jomyap@student.elte.hu ORCID: 0000-0002-7852-1047 László Nemes, Doctoral Candidate, Eötvös Loránd University, Budapest, Hungary, nemes.laszlo84@gmail.com ORCID: 0000-0001-9629-3677 Keywords: LIS training, curriculum, data science education, Hungary, Philippines Abstract: Introduction Data, information, and knowledge are the basic components of librarian training. In practice, having an organized library collection, adhering to system protocols, and providing quality service must be aligned with LIS education through theory learning. In the last decade, data operations and their systematization have become more widespread, and data science has emerged. Data science is covered in several fields, from informatics to data visualization and art courses. Higher education programs in most countries have three levels: bachelor, master's, and doctoral. This is also true for some library and information science programs. The goal of our poster is to compare the data science programs offered at higher education institutions in Hungary and the Philippines. Objectives As we understand the global role of data science in LIS education, this study aims to answer the following questions: 1. Why do we need to find a link between data science and Library and Information Science (LIS)? 2. Are data science courses being offered and integrated in existing LIS programs? 3. Is librarianship training including data science courses the same for Hungary and the Philippines? 4. How can Asia and Europe share experiences in building a data science course? Methods Data selection is based on existing LIS school directories from the Philippines and Hungary. We examined open access information and available public curricula from the webpages of known LIS universities or schools. We searched their national qualifications, frameworks or any relevant training and output requirements in becoming a librarian. As an output, we are presenting the data and process visualization to give us an idea of their similarities and differences.
poster
INTERNATIONAL PARTICLE PHYSICS OUTREACH GROUP Engaging the world with science S. Goldfarb, University of Melbourne, Lepton-Photon 2021, Manchester, UK, 10-14 Jan 2021 Fundamental Components of Scientific Research Communication • Develop messages & strategies • Optimise platforms & networks • Create effective material • Implement and evaluate Education & Outreach • Establish understanding of scientific process • Instil appreciation of fundamental research • Build trust with communities • Train the next generation of scientists The IPPOG Collaboration International Scientific Collaboration • Active researchers, experts in communication and education • 32 countries, 6 experiments, 1 intl. lab (CERN), 2 assoc. members (DESY, GSI) • Twice annual collaboration meetings Goals • Sustainable development of particle physics outreach • Improvement of scientific educational standards worldwide Working Groups • Explanatory content for hot topics in particle physics • Expansion of physics Masterclasses to new audiences worldwide • Applications of particle physics research to society • Development of exhibitions and material supporting public engagement IPPOG’s Worldwide Activities International Particle Physics Masterclasses • Students invited to local institutes to become “Researchers for a Day” • Introductory lectures, analysis of real data, worldwide videoconference • 2019: 60 countries, 220 labs, 14,000 students • 2022: 24 Feb - 9 Apr & 11 Feb (UN Day for Women & Girls in Science) physicsmasterclasses.org Global Cosmics Portal • Central Hub for worldwide classroom cosmic ray programmes globalcosmics.org Organisational Support and Promotion of • International Muon Week, Worldwide Data Day (QuarkNet) • International Cosmics Day (DESY) • Beamline for Schools (CERN, DESY) Preparing the Future Support for particle physics and related fields • Develop material and activities in support of European Strategy Update 2020 • Extend partnerships in scientific, educational and communication societies and networks, including APS, EPS, NuPECC, APPEC, IAPS, EPPCN, Interactions Support for science education • New online platforms for educational programmes • New resource database targeting teachers and students worldwide Support for everyone • Continued global expansion of membership and programme reach • Diversity, Inclusion Accessibility Working Group to support new audiences The Education, Outreach & Communication we do today lay the foundation for tomorrow’s support and secure the next generation of scientists. Poster presented on behalf of the IPPOG Collaboration © 2022 CERN IPPOG’s efforts stretch well beyond particle physics by teaching students to think critically, thus becoming knowledgeable and productive citizens. New online platforms for 2022 ippog.org ippog.org/ippog-resource-database IPPOG Timeline 1997 Birth of European Particle Physics Outreach Group (EPOG) formed under the joint auspices of ECFA and EPS-HEPP 2005 Birth of International Particle Physics Masterclasses Now in over 60 countries worldwide 2011 Global Expansion to IPPOG Israel, Australia, USA, South Africa, Brazil, India, Mexico,… 2016 Formal Scientific Collaboration Memorandum of Understanding 2022 Celebrating 25 Years of Outreach Excellence Dedicated workshops to extend our capability “…the particle physics community has a moral obligation to inform the public on its activities. To do this well, experiences must be shared among countries in view of the need to optimize the use of resources.” - Chris Llewellyn-Smith, CERN Director General, 1997Coming soon!
poster
FACULTY OF COMPUTER SCIENCE Katrin Krieger - Faculty of Computer Science - Otto von Guericke Universität Magdeburg/Germany - katrin.krieger@ovgu.de Creating Learning Material from Web Resources Motivation Observation: Learners use unstructured legacy Web resources as learning material, although those resources have not been created as content in formal learning settings. Katrin Krieger Otto-von-Guericke-Universität Magdeburg, Germany PhD supervisor: Prof. Dr. Dietmar Rösner Semantic Fingerprints A semantic fingerprint is a graph which describes the concepts and their relations of the corresponding resource. This fingerprint can then be used to compute the semantic relatedness of resources, which enables us to decide if a resource is relevant for a learning context. With this information we can decide, whether the resource is indeed useful as learning material in this particular learning context or not. interactivity type LOM learning resource type semantic density description title 1687 down vote accepted Im going to order this guide b Absolute Beginner Firstly haskell is capable of However there are some problem Firstly a good guide to learni A good list of problems to try Once you have done a few of th Beginner After that you should have a f Working through the problems i After a while you will get to Also it takes a while for the To make sure that you are unde Intermediate Once you understand Monads I t Now you can finish the real wo With the knowledge you would h Parsec for parsing program Quickcheck A very cool tes HUnit Unit testing in hask gtk2hs The most popular gu happstack A web developmen Also there are many concepts l Applicative An interface l FoldableTraversable Typecl Monoid A Monoid is a type Arrows Arrows are a way of Arrays the various mutable ST Monad lets you write co FRP Functional Reactive Pr There are a lot of new languag Multiparameter type classe Type families Existentially quantified t Phantom types GADTS others A lot of haskell is based arou Finally you will want to learn ghc and all its features cabal the haskell package darcs a distributed versio haddock a haskell automati While learning all these new l Expert It will take you years to get Getting Help Finally while at any stage of the haskell irc channel the mailing lists These ar other places listed on the Conclusion Well this turned out longer th Adopting some of the educational categories of the Learning Object Metadata standard (LOM), we will extract and derive data about educational objectives such as interactivity type, learning resource type, semantic density, etc. To obtain this kind of data we will analyze the Web resource itself as well as the corresponding semantic fingerprint. Linked Learning Items We create Linked Learning Items (LLI) and make them available as Linked Data. The LLIs contain data from the Web resource, its corresponding semantic fingerprint, and the gathered educational metadata. These data packages can then be integrated directly in Web-based learning management systems (LMS) and offer learners the opportunity to access this learning material right from within the recent learning context.
poster
Facilitating Author Agency in Open Access Negotiations This flowchart helps UU researchers to implement their publication strategy in an Utrecht context. How can I negotiate in my role as an… author reviewer, editor, … copyright/ licensing costs policies Interactive tool used to collect and share best practices and negotiation strategies, with additional resources. CC-BY Jeroen Sondervan why what when how where as my goal is to… we intend to publish these… at these moments … while trying to… using these platforms/venues … Interactive tool to help researchers determine their publication strategy. CC-0 Jeroen Bosman Hanna de Vries h.devries1@uu.nl https://orcid.org/0000-0001-8454-7596 Jeroen Sondervan jeroen.sondervan@nwo.nl https://orcid.org/0000-0002-9866-0239 At Utrecht University we strive for 100% Open Access, which works best when researchers are actively involved in the process from the start. Libraries and consortia often stand between researchers and publishers as intermediaries; could we stand behind our authors instead? What would that look like in terms of library services and support? Try out the various tools we have been using at Utrecht University and come talk to us about facilitating author agency in Open Access publishing strategies and negotiations. publication strategy negotiating strategy costs licensing ownership incentives policies goals output types time- lines additional Open Science practices venues
poster
Mass function and chemical abundances of the three distinct populations in the globular cluster NGC 6752 Roman Gerasimov 1 Adam Burgasser 1 Derek Homeier 2,3 Luigi R. Bedin 4 Michele Scalco 5,4 Jon Rees 6 Maurizio Salaris 7,8 Jay Anderson 9 1University of California San Diego 2Förderkreis Planetarium Göttingen 3Aperio Software Ltd. 4INAF -- Osservatorio Astronomico di Padova 5Università di Ferrara 6Lick Observatory 7Liverpool John Moores Uni. 8INAF -- Osservatorio Astronomico di Abruzzo 9Space Telescope Science Institute Photometric abundances Halo globular clusters are valuable tracers of the early history of chemical en- richment in the Milky Way due to their old ages [10] and comparatively uniform coeval stellar populations contained within. The latter allow constraints on the age and composition of a given cluster to be derived from its colour-magnitude diagram. The colours of ultracool dwarf (UCD) members of the cluster (including low Main Sequence stars and brown dwarfs) are expected to be most sensitive to element abundances due to molecular opacity and chemistry enabled by the low effective temperatures (Teff) of such objects. Furthermore, brown dwarfs cool continuously throughout the lifetime of the cluster and, therefore, their magnitudes are directly related to the cluster age [4]. While no brown dwarfs in globular clusters have been detected so far, they are expected to be uncovered in large numbers in scheduled cycle 1 observations with the James Webb Space Telescope [2, 3]. In this study, we derive photometric constraints on the composition of the glob- ular cluster NGC 6752, known to host three distinct stellar sub-populations [11]. Theoretical isochrones were calculated from new grids of stellar evolutionary models and model atmospheres and iteratively fitted to existing Hubble Space Telescope (HST) observations (GO-15096, GO-15491; PI: Bedin). Our modelling setup is described in [5]. At Teff> 4600 K, model atmospheres are calculated with the ATLAS 9 code [8, 9]. UCD model atmospheres are computed with PHOENIX 15 [6, 1, 7], including the Allard & Homeier cloud model. Evolu- tionary models are computed using MESA [13] with surface boundary conditions derived from our model atmospheres. Figure 1 summarizes the effect of enhancement of individual elements on the colour of UCDs in NGC 6752. We found the photometric effect of [Ca/M], [Co/M], [Cu/M], [Mn/M], [N/M], [Ni/M], [Si/M], [Ti/M] and [V/M] to be negli- gible within 3000 K < Teff< 4500 K. The variation in colour shown in the figure is computed with respect to the approximate abundances of the red-most sub- population of NGC 6752 and does not include the effect of perturbed abun- dances on stellar evolution. Note that every enhancement shown in the figure displays a unique dependence on temperature and may therefore be inferred from the colour-magnitude diagram of the population. 3000 3300 3600 3900 4200 4500 Effective temperature [K] C O Al Mg Na M 0.1 mag blue red Figure 1: Sensitivity of HST/WFC3 mF110W −mF160W colours in the red-most sub-population of NGC 6752 to en- hancements of individual elements as well as the overall metallicity ([M/H]) as a function of effective temperature. Sensitivity is represented by the width of the cyan and pink bars and is defined as the difference in colour magni- tude as the enhancement is varied by 1 dex compared to the nominal abundance. Cyan/pink bars correspond to the case where the increased abundance offsets the colour blueward/redward. Isochrones 0.4 0.5 0.6 0.7 0.8 mF110W mF160W 15 16 17 18 19 20 21 22 mF160W 3300 3600 3750 3900 4200 Pop A [M/H] = 1.65 [Na/M] = 0.03 [Mg/M] = 0.51 [Al/M] = 0.65 [O/M] = 0.67 [C/M] = 0.2 Pop B [M/H] = 1.61 [Na/M] = 0.26 [Mg/M] = 0.49 [Al/M] = 1.1 [O/M] = 0.5 [C/M] = 0.3 Pop A Pop B Pop C [prelim] Teff [K] Observation 0.2 0.4 0.6 0.8 mF110W mF160W 16 18 20 22 24 26 28 mF160W Pop B M dwarfs L dwarfs T dwarfs Prelim Observation HBL Figure 2: Observed colour-magnitude diagram of NGC 6752 and b
poster
Combination of regional modelling and in-situ data analysis for groundwater assessment in the Mediterranean region Nahed Ben-Salem1, Robert Reinecke2, Alexander Wachholz1, Michael Rode1, Dietrich Borchardt1 and Seifeddine Jomaa1 1. INTRODUCTION AND MOTIVATION 4. METHODOLOGY Models’ outputs will be compared against observations at countries and case studies scales (e.g., Spain, Greece, Crete Island, La Mancha aquifer). Set-up of G3M (v1.0) in the MED (at steady and transient condition). REFERENCE 1 Reinecke, R. et al. 2019. Challenges in developing a global gradient-based groundwater model (G3M v1.0) for the integration into a global hydrological model. Geosci. Model Dev. 12, 2401-2418. 2 de Graaf, I. et al. 2017. A global-scale two-layer transient groundwater model: Development and application to groundwater depletion. Adv. Water. Resour. 102, 53-67. 3 Fan, Y. et al. 2013. Global patterns of groundwater table depth. Science 339, 940-943. 8. FUTURE WORK 7. RESEARCH SIGNIFICANCE and SCIENTIFIC MERIT We argue that large-scale groundwater modelling is crucial tool to confront the challenges of climate change impacts on groundwater availability and to implement science-based adaptation and mitigation measures in the Mediterranean. Description of the selected models Time series from all Spain, Portugal and France as well as some wells in Greece and Tunisia were collected. 2. RESEARCH OBJECTIVE 5. EXPLANATORY VARIABLES 6. PRELIMINARY RESULTS ACKNOWLEDGEMENT: This study is supported by the Sustain-COAST R&I project co-funded under the PRIMA 2018 programme section II. For more information, please visit: https://www.sustain-coast.tuc.gr/en/home ABSTRACT #22  To advance our understanding of large-scale groundwater modeling, which in turn can be an important tool for an improved groundwater resources management by using modelling as means for predicting the effect of climate change and anthropogenic pressures on groundwater levels in data-scarce regions. (EOMAP)  This study aims to evaluate the performance of three global gradient-based groundwater models to represent the steady-state of groundwater levels in the Mediterranean region using ensemble-models (de Graaf et al., 2015, Reinecke et al., 2019 , Fan et al., 2013), Figure 4: Histograms of the three compared models. Figure 2: The MED grid with4 4..4444km km Figure 6: WTD (m b.g.l) Fan et al., 2013.  Three global gradient-based groundwater models are currently available and suitable for regional scale application (de Graaf et al., 2015; Fan et al., 2013; Reinecke et al., 2019). These models are introduced briefly below and Table 1, is presenting the main characteristics of each model (modified from Reinecke et al., 2020). Karstic map of the MED (Derived from the World Karst Aquifers Mapping project WHYMAP (WOKAM, Chen et al., 2017) shows that, 41% of the Mediterranean contains karstifiable rocks and Continuous Carbonate Rocks. Figure 3: Karstic map of the MED (Derived from the World Karst Aquifers Mapping project WHYMAP (WOKAM) (Chen et al., 2017)) The comparison between the three models has been done by aggregation to a rectangular grid that covers the Mediterranean region; with spatial resolution of 4.44 km. 1Helmholtz Centre for Environmental Research‐UFZ, Magdeburg, Germany (nahed.ben-salem@ufz.de, Alexander.wachholz@ufz.de, michael.rode@ufz.de, dietrich.borchardt@ufz.de, seifeddine.jomaa@ufz.de), International Center for Water Resources and Global Change (UNESCO), Koblenz, Germany (Reinecke@bafg.de) 3. MODELS’ DESCRIPTION Figure 1: GW wells distribution in Portugal, Spain, France and Italy. Portugal - 982 wells Spain - 3024 wells France - 4443 wells Italy - 808 wells WTD_Fan et al., 2013 WTD_G3M, 2019 Preliminary results showed that there is a big discrepancy between the three compared model outputs. More specifically, the de Graaf et al. (2017) model presents a deeper water table than Reinecke et al. (2019) and Fan et al. (2013). de Graaf et al. (2017) generally shows
poster
92% 49% 37% 37% 33% 24% 18% 11% 5% Making Academic Medicine’s First Cut: Analysis of Reasons for Rejec8on a9er Internal Editor Review Holly S Meyer1, Steven J Durning1, David P Sklar2, Lauren A Maggio1 1. Uniformed Services University 2. Arizona State University Objec8ve: To inform scholars of common reasons for internal peer review rejecNons. DescripNve content analysis of editors’ internal peer review comments for manuscripts (n = 369) rejected prior to external peer review from the journal Academic Medicine (AM). Text difficult to follow / understand BACKGROUND rejected during the internal editor review process RESULTS ABSTRACT DISCUSSION The volume of submission received by Academic Medicine can cause a burden on external reviewers (green shirts). To reduce the burden, an internal editor review was insNtuted in 2012. While this review was completed by an associate editor, it was not clear what the main reasons for rejecNon at this first review. METHODS To characterize the internal first review, content analysis of the associate editors free-text comments from 369 manuscripts was completed. Internal Editor Review Inadequate or incomplete introducNon Issues with scienNfic conduct SubopNmal data collecNon process Weak discussion and/or conclusions Weak data analysis and/or presentaNon Publishing consideraNons IneffecNve study quesNon and/or design Unimportant topic to the journal’s mission ? 65% Author Contact: holly.meyer.ctr@usuhs.edu Authors’ TwiRer Handles: Holly Meyer @hollysmeyer and Lauren Maggio @laurenmaggio References: 1. Bordage G. Reasons reviewers reject and accept manuscripts: The strengths and weaknesses in medical educaNon reports. Acad Med. 2001;76:889–896. 2. Durning JS and Caroline J. Review Criteria for Research Manuscripts. 2nd ed. Washington, DC: AssociaNon of American Medical Colleges; 2015. hhps:// members.aamc.org/eweb/upload/Review%20Criteria%20For%20Research%20Manuscripts.pdf. Accessed May 13, 2017. Disclaimer: The views expressed in this arNcle are those of the authors and do not necessarily reflect the official policy of the Uniformed Services University of the Health Sciences, Department of Defense, or U.S. Government. Acknowledgement: A special thanks to Eric G. Meyer for his help with the conveyer built visual. In addiNon to the use of constant comparison, the authors (LM & HM) employed Bordage’s “Top 20 Reasons Reviewers Recommend RejecNon”1 as sensiNzing concepts to structure the coding. TIP FOR AUTHORS VALUE OF COMMUNICATION BETWEEN AUTHORS AND EDITORS Check out the AAMC’s Review Criteria for Research Manuscripts2. Find it here: hhps://Nnyurl.com/reviewcriteria Manuscripts must move the current literature forward; Research quesNon sets the stage for the study design; Study is informed by, and subsequently informs, the literature; TAKE HOME MESSAGES Results: Comments were analyzed for explicit ‘reasons for rejecNon’. Nine categories emerged. The top five are, [1] ineffecNve study quesNon and/or design (n=338, 92%), [2] subopNmal data collecNon process (n=180, 49%), [3] weak discussion (n=165, 45%), [4] unimportant or irrelevant topic (n=137, 37%), and [5] weak data analysis and/or presentaNon of results (120, 33%). Conclusions: Clear idenNficaNon of a research quesNon, a strong methodology, and a topic aligned with the journal’s mission are key to avoiding a ‘desk’ rejecNon. Enhanced communicaNon improves the quality of manuscript submissions; Clarify the ambiguity surrounding editor rejecNons; Eases the burden on peer reviewers. Increase the likelihood of moving to external peer review; This study is being published in Academic Medicine. Use the QR code for access to the arNcle
poster
THE EVENT On 22nd September 2017, a high energy neutrino was detected by the IceCube observatory in spatial coincidence with the blazar TXS 0506+056, which was observed to be flaring in the GeV band by the Fermi-LAT telescope. The increased activity in the high energy band and the relation to the neutrino event triggered multi- wavelength observations. On the 4th of October, MAGIC reported the detection of VHE gamma rays from TXS 0506+056 sky position. MAGIC observations of VHE gamma rays from the flaring blazar TXS 0506+056 coincident with high- energy neutrino IceCube-170922A THE MODEL Here we present the results obtained from the observations performed on the blazar with the MAGIC telescopes, the multi-wavelength data collected during the flaring event and the modelling of the emission of the source using the so-called spine-layer scenario. We assume that the jet is structured, with an internal faster core (or spine) and a slower layer. Photons from the layer are amplified in the spine frame and are used by relativistic protons for photo meson reaction. W. Bhattacharyya1, C. Righi7,E. Bernardini1,2, M. Cerruti9, V. Fallah Ramazani8, L. Foffano2,4, S. Inoue6, E. Prandini2,4, K. Satalecka1, F. Tavecchio7 for the MAGIC Collaboration RESULTS • The model allows us to nicely reproduce the EM SED and the neutrino flux. In particular we can keep the proton power to a moderate value (1045-1046 erg s-1). • We find that gamma-gamma absorption start do be important close to the MAGIC band, consistently with the measured soft spectrum. • Gamma-rays are dominated by leptonic processes (lnverse Compton emission from electrons in the spine) • Radiation from absorbed and reprocessed hadronic gamma-rays can provide a non negligible contribute in the X-ray band. • Physical conditions in the jet are consistent with a maximum (comoving) protons energy of 1016 eV MAGIC is a system of two 17m Imaging Atmospheric Cherenkov Telescopes (IACTs) designed for the detection of very high energy (VHE) gamma-rays and situated in the Roque de los Muchachos Observatory, Canary Islands, La Palma. 1DESY Zeuthen, 2University of Padova, 3University of Udine, 4INFN Padova, 5INFN Trieste, 6RIKEN, 7INAF, 8 Tuorla Observatory, 9Institut de Ciències del Cosmos - Universitat de Barcelona References: [1] Ghisellini, G., Tavecchio, F., & Chiaberge, M. 2005, A&A, 432, 401 [2] Tavecchio, F., Ghisellini, G., & Guetta, D. 2014, ApJ, 793, L18 [3] Ansoldi S., et al., (The MAGIC Coll.) 2018, ApJ, 863, L10 Parameters used for the model Maximum p energy Ep = 1016 eV electron magnetic field total jet power = PB+Pe+Pp =1046erg s-1 }
poster
Halizah Mat Rifin, Wan Shakira Rodzlan Hasani , Tania Gayle Robert Lourdes, Nur Liana Abd Majid, Jane Ling Miaw Yn, Thamil Arasu Saminathan, Hasimah Ismail, Mohd Ruhaizie Riyadzi, Muhammad Fadhli bin Mohd Yusoff. NMRR-18-3085-44207 P-31 Hypercholesterolemia is one of the major risk factors for coronary heart disease[1]. Overall, raised cholesterol is estimated to cause 2.6 million deaths and 29.7 million disability adjusted life years (DALYS)[2]. As the result, knowing the cholesterol level status, treatment and control have received a great deal of surveillance. • The findings showed that the prevalence of hypercholesterolemia escalated from 35.1% in 2011 to 47.7% in 2015 but decreased to 38.1% in 2019. This was lower compared to study in Philippine in 2011 (46.9%)[3] and higher compared to study in Korea in 2016(19.9%)[4]. • Meanwhile, there is an increase in the prevalence of known hypercholesterolaemia from 8.4% in 2011 to 13.5% in 2019. The finding was lower compared to a study in Korea (47.4%) in 2012 and 58.4% in 2016[4]. • The proportion those reported on medications among those who were aware that they had hypercholesterolaemia increased from 64.2% in 2011 to 80.1% in 2019. It was higher than a study in Korea (37.3%) in 2012 and 49.1% in 2016[4]. • The proportion whose hypercholesterolemia was controlled among those who were on medications increased slightly from 59.6% in 2011 to 62.6% in 2019. This finding was higher compared to Korea (29.7%) in 2010 and lower (84.3%) in 2016[4]. The objective of this study was to examine the trends in hypercholesterolaemia prevalence, known hypercholesterolaemia, treatment and control among adults in Malaysia • Data from the three cycles (2011,2015,2019) of the NHMS were used. • The NHMS is a cross-sectional community-based survey which employed a two-stage stratified cluster random sampling design to ensure national representativeness. • Respondents aged 18 years and above were asked via face-to-face interview to respond to the hypercholesterolemia module. • Hypercholesterolaemia was defined as a total cholesterol equal to or more than 5.2 mmol/L. • For respondents with hypercholesterolemia, awareness was defined as ‘Yes’ if they had been informed previously by medical personnel that they had high cholesterol. • Treatment for hypercholesterolemia was defined as ‘Yes’ if they were currently receiving medication from a doctor in the past 2 weeks for high cholesterol. • Controlled hypercholesterolemia was defined as having a desirable blood cholesterol level at the time of the survey among respondents who were on treatment • Complex sampling design analysis was used to account for sampling weights and study design properties. There is a promising trend in hypercholesterolemia prevalence, treatment and known hypercholesterolaemia from 2011 through 2019. However, hypercholesterolemia control remains poor. The government and health authorities must intensify disease control intervention to achieve the desired cholesterol levels. Keywords: Trends, hypercholesterolaemia, known hypercholesterolaemia, treatment, control, NHMS 2019 . CONCLUSION Trends in Hypercholesterolemia, Trends in Hypercholesterolemia, INTRODUCTION OBJECTIVE METHODS RESULTS & DISCUSSION 4 Treatment and Control among Adults in Malaysia Hypercholesterolaemia Table 1: Prevalence of Overall Hypercholesterolaemia for Respondent Aged 18 years and above for NHMS 2011, 2015 and 2019 Figure 1: Hypercholesterolaemia Trend in Malaysia Figure 2: Trends of treatment and control of hypercholesterolaemia in 2011, 2015 and 2019 Sociodemographic Total Gender Male Female Locality Urban Rural 2011 35.1%(33.9,36.2) 30.1%(28.7,31.6) 40.2%(38.7,41.7) 34.3%(32.9,35.7) 37.0%(35.1,39.0) 2015 47.7 %(46.5, 48.9) 43.5%(42.0, 45.1) 52.2%(50.7, 53.7) 47.7%(46.3,49.2) 47.7%(45.6,49.7) 2019 38.1%(36.2,40.0 ) 32.0(29.7,34.4) 44.5(42.2,46.8) 38.0 (35.7,40.4) 38.2 (35.5, 53.7 ) NHMS year 1. Ford ES, Li C, Pearson WS, Zhao G, Mokdad AH. Trends in hypercholesterolemi
poster
D ire c tor: P rof. D r. Ma rc W olfra m R e s ea rch A re a : A uthors /C onta c t : D re s de n, Research Data Strategy "Anno Zero" From scratch, toward impacts Stefano Della Chiesa Research Data Management (RDM) is a cross- functional, multi-level, multidisciplinary endeavour among various actors. The key in RDM is to maximise the research data value chain, knowledge transfer and impact. The Leibniz IOER is developing a research data strategy, governance and stewardship program with a socio-technical approach, improving engagement, coordination and RDM procedures. 1. Strategy and Policy 1.1 Strategy development 1.2 Strategic goals 1.3 Commitment and incentives 1.4 RD policy 3. Management procedures 3.1 RDM into project planning 3.2 Integration into the Research Information System 3.3 Integration into the processes of law and ethics 2. Business Model 2.1 Governance and organisational development 2.2 Personnel investments 2.3 Technology investments 2.4 Cost model 4. Communication 4.1 Information 4.2 Counselling 4.3 Public relations 5. Training 5.1 Training materials 5.2 Training events 6. Active Data Management 6.1 Basic IT services 6.2 Services for collaborative working 7. Data Publication 7.1 Publication services 7.2 Archiving services 7.3 Risk management 8. Service Requirement 8.1 Failure safety 8.2 Data security 8.3 Metadata and indexing 8.4 Standards compliance and long term archiving services The proposed new scheme for the RISE-DE reference model1 for strategy processes in institutional research data management deals with 3 strategy pillars, 8 thematic areas and 25 topics (Adaptation of Figure 1 of the RISE-DE reference model). The Data Strategy encompasses the long-term strategy aligned with the organisation's mission. Data Governance focuses on aligning with high-level policies and procedures at the organisational level and developing the organisation's steering practices in terms of corporate and resources development. Data Stewardship focuses on operational services and best practices implementation. The RISE-DE reference model 1RISE-DE – Referenzmodell für Strategieprozesse im institutionellen Forschungsdatenmanagement. https://doi.org/10.5281/zenodo.3585556
poster
ESO Extremely Large Telescope • MICADO Near Infrared Diffraction Limited Camera λ= 800nm – 2.5µm T≤100K R≤5000 • METIS (PI Institute) Mid Infrared Camera & Spectrograph λ= 2.8µm – 14µm T≤30K R≤100000 • MOSAIC Multi Object Spectrograph λ= 500nm – 1.7µm T=300K R≤10000 • EPICS Extreme AO Planet IFS & Spectro-Polarimeter λ= 500nm – 930nm T=300K R≤300 ESO Very Large Telescope 2003 MIDI 2 Telescope Interferometer Spectrometer 2003 λ=7µm – 14µm T≤40K R≤700 2004 VISIR Mid InfraRed Imager and Spectrograph 2004 λ= 7µm – 27µm T≤30K R≤25000 2004 SPIFFI 2k Spectroscopic Camera for SINFONI 2005 λ= 1µm – 2.5µm T≤100K R≤4000 2009 X-Shooter UV-VIS-IR Spectrograph 2009 λ= 300nm – 2.5µm T≤100K R≤10000 2014 Sphere-ZIMPOL Extreme Adaptive Optics Imaging Polarimeter 2014 λ= 570nm – 900nm T=300K >25 Filters 2017 MATISSE 4 telescope Interferometer Spectrometer 2015 λ= 2.8µm – 14µm T≤35K R≤6000 James Webb Space Telescope MIRI Mid InfraRed Instrument Spectrometer Main Optics λ=5µm-28µm T=6.8K VLT in 2030 Heritage ELT in 2030 GAIA Future VLT more surveys - less individual small projects. MAVIS (optical AO high resolution imaging – excellent complement to MICADO) Moderate resolution Cass UV spectrograph (CUBES / UVES upgrade) ERIS & MOONS BLUEMUSE MUSE-like AO-assisted IFU spectrograph operating up to K band micron FLAMES upgrade / replacement (complementary to 4MOST / MOONS). Wide field IFU - spectroscopic wide field survey (SDSS V but deeper, better and everything). - repeat observations for transients - slave spectroscopic-imager IFU to Athena's WFI (FoV 40’x40’). VLT prime focus camera for larger field of view / VLTI capabilities? new wide field multi-fibre spectroscopy faculity - high spectral resolution: R>40-60k - Complementary to 4MOST & WEAVE & GAIA - 130fibres over 30arcmin diameter (or more?) wide field MOS on a 10m telescope dedicated to surveys. - R=20k - galactic archaeology (GAIA), cosmology (Euclid). Transients (LSST). Wide field with massive MOS (4most on an 8m) careful tradeoff with VLTI Wide-field NIR imager (VISTA/VIRCAM detectors) complementary to Euclid in K band and better image quality than Euclid. - large-scale structure and galaxies at redshift ~1. - wide area survey, 100s of nights. Very wide-field VIS+IR imaging capabilities - >1 square degree area - find targets in the EoR redshift range via photo-z's. - Complementary to SKA deep fields, targets selection for follow-up and HI absorption measurements with SKA. All sky survey with MOS doing polarimetry Dedicated UT for Transiting Exoplanet Atmospheres - similar to having the HARPS instrument on a dedicated La Silla Telescope. - high resolution spectroscopy from Optical to IR, something like ESPRESSO+CRIRES+. - 50-to-100 nights/year Improved FORS – better calibrate instrumental systematics 2 UT’s working in parallel - both low and high resolution spectroscopy simultaneously IFU high-dispersion spectroscopy with high-contrast imaging at the VLT for exoplanet characterization (RISTRETTO) SPHERE upgrade: AO, coronagraphs, polarimetry, IFU Synoptic, time-domain astronomy, Follow-up BlackGEM & LSST - longer time-series spectroscopy/photometry - lower-resolution spectroscopic identification. - spectroscopic surveys - ToO overrides - Long term monitoring Support / enhance / capitalise on ESA's Large missions Athena & LISA JWST and Euclid follow-up HIPERCAM X-Shooter with VIS/NIR AO fast-and-dirty low res R=1000 spectroscopy Robotised VLT as a transient follow-up machine VLT in 2030 VLT in 2030 Synergy Exoplanet wishes and ideas Wide Field Trade off Ramon Navarro NOVA instrumentation Manager with input from NOVA community
poster
DUBLAGEM EM SALA DE AULA EETI Bilíngue Professor Djalma da Cunha Batista Ensino fundamental II- 9º ano Alécio Vaneli Gaigher Marely - alecio.marely@seducam.pro.br Introdução-Línguas adicionais entrelaçam o nosso sujeito e auxiliam o desenvolvimento do indivíduo. Segundo Gadiolli (2013)”ajudar alunos a aprender a aprender e também a aprender a se emancipar, empoderar-se e se descobrir como agentes de mudança social”.A problemática é situada na questão onde os estudantes passam anos em uma escola tendo aula de língua inglesa e saem das escolas alegando pouco rendimento na língua adicional.A fala em língua inglesa é o desejo da maioria dos aprendizes como Al Hosni,2014 “Speaking is the active use of language to express meaning, and for young learners, the spoken language is the medium through which a new language is encountered”, ao pensar nesta questão vem a necessidade de devolver essa habilidade.Objetivos Instigar o uso da língua inglesa em sala de aula Propagar a fala de língua inglesa Promover o encontro de culturas através da música Registrar os momentos históricos-literários do período em que as músicas foram escritas e gravadas. Materiais e Métodos- A atividade proposta foi um, lipsync, ou seja,uma atividade de dublagem. Os alunos deveriam escolher dentre as músicas mais ouvidas de bandas consagradas no idioma de língua inglesa para poderem gravar uma dublagem. As bandas eram I)Rolling Stones II)Queen III)The Supremes IV)The Beatles.Os vídeos poderão ser visualizados por formato de qr codes como seguem: QR code: I e II - Vídeos
poster