Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
2,188
5,036,975,539
IssuesEvent
2016-12-17 11:30:40
fujaba/org.fujaba.graphengine
https://api.github.com/repos/fujaba/org.fujaba.graphengine
closed
test all features
high priority improvement in process
- [x] building, serialization, deserialization of a Graph - [x] building, serialization, deserialization of a PatternGraph - [ ] check if two Graphs are isomorph to one another (also test performance in comparison to SDMLib) - [x] check if a Graph is an isomorph sub-graph of another Graph - [x] matching of PatternGraphs on a Graph (including negative variants and usage of the expression library) - [x] application of Matches - [ ] correct calculation of reachability graphs (maybe also test performance in comparison to SDMLib) - [x] side issues like graph-'splitting' and building of graphs for external visualization
1.0
test all features - - [x] building, serialization, deserialization of a Graph - [x] building, serialization, deserialization of a PatternGraph - [ ] check if two Graphs are isomorph to one another (also test performance in comparison to SDMLib) - [x] check if a Graph is an isomorph sub-graph of another Graph - [x] matching of PatternGraphs on a Graph (including negative variants and usage of the expression library) - [x] application of Matches - [ ] correct calculation of reachability graphs (maybe also test performance in comparison to SDMLib) - [x] side issues like graph-'splitting' and building of graphs for external visualization
process
test all features building serialization deserialization of a graph building serialization deserialization of a patterngraph check if two graphs are isomorph to one another also test performance in comparison to sdmlib check if a graph is an isomorph sub graph of another graph matching of patterngraphs on a graph including negative variants and usage of the expression library application of matches correct calculation of reachability graphs maybe also test performance in comparison to sdmlib side issues like graph splitting and building of graphs for external visualization
1
265
2,696,694,089
IssuesEvent
2015-04-02 15:31:05
appsgate2015/appsgate
https://api.github.com/repos/appsgate2015/appsgate
closed
Le clock ne fonctionne plus...
P1 PROCESSING
Le "chaque fois que" en heure réelle et simulée ne fonctionnerait plus (cf. Joëlle)
1.0
Le clock ne fonctionne plus... - Le "chaque fois que" en heure réelle et simulée ne fonctionnerait plus (cf. Joëlle)
process
le clock ne fonctionne plus le chaque fois que en heure réelle et simulée ne fonctionnerait plus cf joëlle
1
781,760
27,448,072,705
IssuesEvent
2023-03-02 15:31:51
prgrms-web-devcourse/Team-Kkini-Mukvengers-FE
https://api.github.com/repos/prgrms-web-devcourse/Team-Kkini-Mukvengers-FE
opened
랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기
Priority: Low Feature
## 📕 작업 설명 > 랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기 ## 📖 To-Do list - [ ] 랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기
1.0
랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기 - ## 📕 작업 설명 > 랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기 ## 📖 To-Do list - [ ] 랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기
non_process
랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기 📕 작업 설명 랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기 📖 to do list 랜덤 맛집 커스텀 오버레이에 현재 위치로부터의 경로 그리기
0
52,110
13,211,389,681
IssuesEvent
2020-08-15 22:47:35
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[I3_PORTS] CentOS7 freetype error (Trac #1700)
Incomplete Migration Migrated from Trac defect tools/ports
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1700">https://code.icecube.wisc.edu/projects/icecube/ticket/1700</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:47", "_ts": "1550067167842669", "description": "On CentOS7, I can't install Geant4 or Root from ports because of a freetype error:\n\n{{{\ng++ -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -Wno-deprecated-declarations -I. -I/cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_7_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v2_RHEL_7_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/cint/inc -o misc/memstat/src/G__MemStat.o -c misc/memstat/src/G__MemStat.cxx\nIn file included from asfont.c:67:0:\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:21:2: error: #error \"`ft2build.h' hasn't been included yet!\"\n #error \"`ft2build.h' hasn't been included yet!\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:22:2: error: #error \"Please always use macros to include FreeType header files.\"\n #error \"Please always use macros to include FreeType header files.\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:23:2: error: #error \"Example:\"\n #error \"Example:\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:24:2: error: #error \" #include <ft2build.h>\"\n #error \" #include <ft2build.h>\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:25:2: error: #error \" #include FT_FREETYPE_H\"\n #error \" #include FT_FREETYPE_H\"\n ^\nmake[1]: *** [asfont.o] Error 1\n}}}\n\nNote that the header in question is available at `/usr/include/ft2build.h`", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2016-05-12T16:23:08", "component": "tools/ports", "summary": "[I3_PORTS] CentOS7 freetype error", "priority": "major", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
[I3_PORTS] CentOS7 freetype error (Trac #1700) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1700">https://code.icecube.wisc.edu/projects/icecube/ticket/1700</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:47", "_ts": "1550067167842669", "description": "On CentOS7, I can't install Geant4 or Root from ports because of a freetype error:\n\n{{{\ng++ -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -Wno-deprecated-declarations -I. -I/cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_7_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v2_RHEL_7_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/cint/inc -o misc/memstat/src/G__MemStat.o -c misc/memstat/src/G__MemStat.cxx\nIn file included from asfont.c:67:0:\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:21:2: error: #error \"`ft2build.h' hasn't been included yet!\"\n #error \"`ft2build.h' hasn't been included yet!\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:22:2: error: #error \"Please always use macros to include FreeType header files.\"\n #error \"Please always use macros to include FreeType header files.\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:23:2: error: #error \"Example:\"\n #error \"Example:\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:24:2: error: #error \" #include <ft2build.h>\"\n #error \" #include <ft2build.h>\"\n ^\n../../../../graf2d/freetype/src/freetype-2.3.12/include/freetype/freetype.h:25:2: error: #error \" #include FT_FREETYPE_H\"\n #error \" #include FT_FREETYPE_H\"\n ^\nmake[1]: *** [asfont.o] Error 1\n}}}\n\nNote that the header in question is available at `/usr/include/ft2build.h`", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2016-05-12T16:23:08", "component": "tools/ports", "summary": "[I3_PORTS] CentOS7 freetype error", "priority": "major", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
non_process
freetype error trac migrated from json status closed changetime ts description on i can t install or root from ports because of a freetype error n n ng pipe wshadow wall w woverloaded virtual fpic iinclude pthread wno deprecated declarations i i cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports science root work root cint cint inc o misc memstat src g memstat o c misc memstat src g memstat cxx nin file included from asfont c n freetype src freetype include freetype freetype h error error h hasn t been included yet n error h hasn t been included yet n n freetype src freetype include freetype freetype h error error please always use macros to include freetype header files n error please always use macros to include freetype header files n n freetype src freetype include freetype freetype h error error example n error example n n freetype src freetype include freetype freetype h error error include n error include n n freetype src freetype include freetype freetype h error error include ft freetype h n error include ft freetype h n nmake error n n nnote that the header in question is available at usr include h reporter david schultz cc resolution fixed time component tools ports summary freetype error priority major keywords milestone owner nega type defect
0
18,822
24,719,438,145
IssuesEvent
2022-10-20 09:32:05
hermes-hmc/workflow
https://api.github.com/repos/hermes-hmc/workflow
opened
Implement BagIt builder
enhancement 5️ post-process
As preparation for full RO-Crate support (perhaps in *Future work*), we should implement a [BagIt](https://www.digitalpreservation.gov/news/2008/20080602news_article_bagit.html) builder. Doing this now in post-processing would still allow users to run the builder without a deposit step, and deposit the resulting BagIt manually.
1.0
Implement BagIt builder - As preparation for full RO-Crate support (perhaps in *Future work*), we should implement a [BagIt](https://www.digitalpreservation.gov/news/2008/20080602news_article_bagit.html) builder. Doing this now in post-processing would still allow users to run the builder without a deposit step, and deposit the resulting BagIt manually.
process
implement bagit builder as preparation for full ro crate support perhaps in future work we should implement a builder doing this now in post processing would still allow users to run the builder without a deposit step and deposit the resulting bagit manually
1
818,555
30,694,544,469
IssuesEvent
2023-07-26 17:31:23
Ore-Design/Ore-3D-Reports-Changelog
https://api.github.com/repos/Ore-Design/Ore-3D-Reports-Changelog
closed
Feature: Allow Rendering Checkbox Resets
enhancement low priority
Allow Rendering checkbox should remember setting. It is resetting now. I can't tell what prompts it to do reset, if it's on-launch or when I run an update. Either way I would prefer Ore3D to remember that setting. I can change it back in the event I want to re-enable rendering.
1.0
Feature: Allow Rendering Checkbox Resets - Allow Rendering checkbox should remember setting. It is resetting now. I can't tell what prompts it to do reset, if it's on-launch or when I run an update. Either way I would prefer Ore3D to remember that setting. I can change it back in the event I want to re-enable rendering.
non_process
feature allow rendering checkbox resets allow rendering checkbox should remember setting it is resetting now i can t tell what prompts it to do reset if it s on launch or when i run an update either way i would prefer to remember that setting i can change it back in the event i want to re enable rendering
0
1,590
4,187,353,620
IssuesEvent
2016-06-23 17:12:21
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
opened
ARM64: System.Diagnostics.Process.Tests
Arm64 System.Diagnostics.Process
``` System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows [FAIL] System.InvalidOperationException : Couldn't connect to remote machine. ---- System.InvalidOperationException : Couldn't get process information from performance counter. -------- System.InvalidOperationException : Cannot load Counter Name data because an invalid index '' was read from the registry. Stack Trace: at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) at System.Diagnostics.NtProcessManager.GetProcessIds(String machineName, Boolean isRemoteMachine) at System.Diagnostics.ProcessManager.GetProcessIds(String machineName) at System.Diagnostics.Process.GetProcessById(Int32 processId, String machineName) at System.Diagnostics.Tests.ProcessTests.<GetTestProcess>d__39.MoveNext() at System.Linq.Enumerable.SelectEnumerableIterator`2.MoveNext() ----- Inner Stack Trace ----- at System.Diagnostics.NtProcessManager.GetProcessInfos(PerformanceCounterLib library) at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) ----- Inner Stack Trace ----- at System.Diagnostics.PerformanceCounterLib.GetStringTable(Boolean isHelp) at System.Diagnostics.PerformanceCounterLib.get_NameTable() at System.Diagnostics.PerformanceCounterLib.GetCounterName(Int32 index) at System.Diagnostics.NtProcessManager.GetProcessInfos(PerformanceCounterLib library, Int32 processIndex, Int32 threadIndex, Byte[] data) at System.Diagnostics.NtProcessManager.GetProcessInfos(PerformanceCounterLib library) Finished: System.Diagnostics.Process.Tests === TEST EXECUTION SUMMARY === System.Diagnostics.Process.Tests Total: 88, Errors: 0, Failed: 1, Skipped: 0, Time: 42.874s ```
1.0
ARM64: System.Diagnostics.Process.Tests - ``` System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows [FAIL] System.InvalidOperationException : Couldn't connect to remote machine. ---- System.InvalidOperationException : Couldn't get process information from performance counter. -------- System.InvalidOperationException : Cannot load Counter Name data because an invalid index '' was read from the registry. Stack Trace: at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) at System.Diagnostics.NtProcessManager.GetProcessIds(String machineName, Boolean isRemoteMachine) at System.Diagnostics.ProcessManager.GetProcessIds(String machineName) at System.Diagnostics.Process.GetProcessById(Int32 processId, String machineName) at System.Diagnostics.Tests.ProcessTests.<GetTestProcess>d__39.MoveNext() at System.Linq.Enumerable.SelectEnumerableIterator`2.MoveNext() ----- Inner Stack Trace ----- at System.Diagnostics.NtProcessManager.GetProcessInfos(PerformanceCounterLib library) at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) ----- Inner Stack Trace ----- at System.Diagnostics.PerformanceCounterLib.GetStringTable(Boolean isHelp) at System.Diagnostics.PerformanceCounterLib.get_NameTable() at System.Diagnostics.PerformanceCounterLib.GetCounterName(Int32 index) at System.Diagnostics.NtProcessManager.GetProcessInfos(PerformanceCounterLib library, Int32 processIndex, Int32 threadIndex, Byte[] data) at System.Diagnostics.NtProcessManager.GetProcessInfos(PerformanceCounterLib library) Finished: System.Diagnostics.Process.Tests === TEST EXECUTION SUMMARY === System.Diagnostics.Process.Tests Total: 88, Errors: 0, Failed: 1, Skipped: 0, Time: 42.874s ```
process
system diagnostics process tests system diagnostics tests processtests testprocessonremotemachinewindows system invalidoperationexception couldn t connect to remote machine system invalidoperationexception couldn t get process information from performance counter system invalidoperationexception cannot load counter name data because an invalid index was read from the registry stack trace at system diagnostics ntprocessmanager getprocessinfos string machinename boolean isremotemachine at system diagnostics ntprocessmanager getprocessids string machinename boolean isremotemachine at system diagnostics processmanager getprocessids string machinename at system diagnostics process getprocessbyid processid string machinename at system diagnostics tests processtests d movenext at system linq enumerable selectenumerableiterator movenext inner stack trace at system diagnostics ntprocessmanager getprocessinfos performancecounterlib library at system diagnostics ntprocessmanager getprocessinfos string machinename boolean isremotemachine inner stack trace at system diagnostics performancecounterlib getstringtable boolean ishelp at system diagnostics performancecounterlib get nametable at system diagnostics performancecounterlib getcountername index at system diagnostics ntprocessmanager getprocessinfos performancecounterlib library processindex threadindex byte data at system diagnostics ntprocessmanager getprocessinfos performancecounterlib library finished system diagnostics process tests test execution summary system diagnostics process tests total errors failed skipped time
1
152,523
12,110,799,086
IssuesEvent
2020-04-21 11:04:10
WoWManiaUK/Redemption
https://api.github.com/repos/WoWManiaUK/Redemption
closed
Duplicate mining node at Thousand Needles
Fix - Tester Confirmed
Don't have a screenshot but there was a duplicate (copper) mining node at [11.27, 14.24] in Thousand Needles. Should these be reported?
1.0
Duplicate mining node at Thousand Needles - Don't have a screenshot but there was a duplicate (copper) mining node at [11.27, 14.24] in Thousand Needles. Should these be reported?
non_process
duplicate mining node at thousand needles don t have a screenshot but there was a duplicate copper mining node at in thousand needles should these be reported
0
19,483
25,792,869,142
IssuesEvent
2022-12-10 08:29:53
COPIM/open-book-collective
https://api.github.com/repos/COPIM/open-book-collective
closed
Provide key information about subscription to subscribers
userstory development membership management (pillar 4) organisational process
As a subscriber .. .. I want to have easy access to any documentation associated with membership (contracts, invoicing information, annual reports) ... ... so that it is easy for me to access all information from a single source.
1.0
Provide key information about subscription to subscribers - As a subscriber .. .. I want to have easy access to any documentation associated with membership (contracts, invoicing information, annual reports) ... ... so that it is easy for me to access all information from a single source.
process
provide key information about subscription to subscribers as a subscriber i want to have easy access to any documentation associated with membership contracts invoicing information annual reports so that it is easy for me to access all information from a single source
1
193,923
6,889,455,492
IssuesEvent
2017-11-22 10:25:58
rucio/rucio
https://api.github.com/repos/rucio/rucio
closed
rsemanager: _get_possible_protocols call raises exception while it should not
bug patch PRIORITY RSE Manager
Motivation ---------- In the _get_possible_protocols call the filtering (tbr) of protocols removes too many protocols, as the list is essentially done twice (per domain) and it looks like more is removed than is there, thus an exception is raised. Modification ------------ Make proper filtering. Split the candidates in two lists per domain and do the filtering smarter. Expected behavior ----------------- Exception should not be raised.
1.0
rsemanager: _get_possible_protocols call raises exception while it should not - Motivation ---------- In the _get_possible_protocols call the filtering (tbr) of protocols removes too many protocols, as the list is essentially done twice (per domain) and it looks like more is removed than is there, thus an exception is raised. Modification ------------ Make proper filtering. Split the candidates in two lists per domain and do the filtering smarter. Expected behavior ----------------- Exception should not be raised.
non_process
rsemanager get possible protocols call raises exception while it should not motivation in the get possible protocols call the filtering tbr of protocols removes too many protocols as the list is essentially done twice per domain and it looks like more is removed than is there thus an exception is raised modification make proper filtering split the candidates in two lists per domain and do the filtering smarter expected behavior exception should not be raised
0
10,654
13,450,517,161
IssuesEvent
2020-09-08 18:39:46
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
reopened
Function ID and Endianness conversion in ARM
Feature: FID Feature: Processor/ARM Type: Question
Hello. is ghidra able to mach little-endian lib's signatures in a big-endian or mixed endian project ?
1.0
Function ID and Endianness conversion in ARM - Hello. is ghidra able to mach little-endian lib's signatures in a big-endian or mixed endian project ?
process
function id and endianness conversion in arm hello is ghidra able to mach little endian lib s signatures in a big endian or mixed endian project
1
22,520
31,567,827,505
IssuesEvent
2023-09-04 01:20:34
tdwg/hc
https://api.github.com/repos/tdwg/hc
opened
New Term - reportedExtremeConditions
Term - add normative Process - under public review Class - Event
## New term * Submitter: Humboldt Extension Task Group * Efficacy Justification (why is this term necessary?): Part of a package of terms in support of biological inventory data. * Demand Justification (name at least two organizations that independently need this term): The Humboldt Extension Task Group proposing this term consists of numerous organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): reportedExtremeConditions * Term label (English, not normative): Reported Extreme Conditions * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event * Definition of the term (normative): A description of any extreme weather or environmental conditions that may have affected the dwc:Event. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): `flooding during week 3 of surveys`; `rockslide at site 2` * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
1.0
New Term - reportedExtremeConditions - ## New term * Submitter: Humboldt Extension Task Group * Efficacy Justification (why is this term necessary?): Part of a package of terms in support of biological inventory data. * Demand Justification (name at least two organizations that independently need this term): The Humboldt Extension Task Group proposing this term consists of numerous organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): reportedExtremeConditions * Term label (English, not normative): Reported Extreme Conditions * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event * Definition of the term (normative): A description of any extreme weather or environmental conditions that may have affected the dwc:Event. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): `flooding during week 3 of surveys`; `rockslide at site 2` * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
process
new term reportedextremeconditions new term submitter humboldt extension task group efficacy justification why is this term necessary part of a package of terms in support of biological inventory data demand justification name at least two organizations that independently need this term the humboldt extension task group proposing this term consists of numerous organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version none proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes reportedextremeconditions term label english not normative reported extreme conditions organized in class e g occurrence event location taxon event definition of the term normative a description of any extreme weather or environmental conditions that may have affected the dwc event usage comments recommendations regarding content etc not normative examples not normative flooding during week of surveys rockslide at site refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
1
58,852
14,352,105,005
IssuesEvent
2020-11-30 03:21:22
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Vulnerability roundup 93: gorm-1.2.24: 1 advisory [9.8]
1.severity: security
[search](https://search.nix.gsc.io/?q=gorm&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=gorm+in%3Apath&type=Code) * [ ] [CVE-2019-15562](https://nvd.nist.gov/vuln/detail/CVE-2019-15562) CVSSv3=9.8 (nixos-20.03) Scanned versions: nixos-20.03: 0d0660fde3b. Cc @ashalkhakov Cc @matthewbauer
True
Vulnerability roundup 93: gorm-1.2.24: 1 advisory [9.8] - [search](https://search.nix.gsc.io/?q=gorm&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=gorm+in%3Apath&type=Code) * [ ] [CVE-2019-15562](https://nvd.nist.gov/vuln/detail/CVE-2019-15562) CVSSv3=9.8 (nixos-20.03) Scanned versions: nixos-20.03: 0d0660fde3b. Cc @ashalkhakov Cc @matthewbauer
non_process
vulnerability roundup gorm advisory nixos scanned versions nixos cc ashalkhakov cc matthewbauer
0
19,322
25,472,073,809
IssuesEvent
2022-11-25 11:03:18
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[IDP] [PM] Not able to create new admin in the PM
Bug Blocker P0 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
**Pre-condition:** mfa should be disabled in the PM **Steps:** 1. Login to PM 2. Click on 'Admins' tab 3. Click on 'Add new admin' button 4. Try to add admin without entering phone number field and Verify **AR:** Not able to create new admin in the PM **ER:** Admin account should be created in the PM ,without entering phone number field.
3.0
[IDP] [PM] Not able to create new admin in the PM - **Pre-condition:** mfa should be disabled in the PM **Steps:** 1. Login to PM 2. Click on 'Admins' tab 3. Click on 'Add new admin' button 4. Try to add admin without entering phone number field and Verify **AR:** Not able to create new admin in the PM **ER:** Admin account should be created in the PM ,without entering phone number field.
process
not able to create new admin in the pm pre condition mfa should be disabled in the pm steps login to pm click on admins tab click on add new admin button try to add admin without entering phone number field and verify ar not able to create new admin in the pm er admin account should be created in the pm without entering phone number field
1
17,119
22,636,762,426
IssuesEvent
2022-06-30 19:51:27
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
Warning: a recent release failed
type: process
The following release PRs may have failed: * #6277 * #6238 * #5494 * #6252 * #6206 * #6140 * #6246 * #6209 * #6018 * #6155 * #6190 * #6208 * #6057 * #6146 * #6185 * #6134
1.0
Warning: a recent release failed - The following release PRs may have failed: * #6277 * #6238 * #5494 * #6252 * #6206 * #6140 * #6246 * #6209 * #6018 * #6155 * #6190 * #6208 * #6057 * #6146 * #6185 * #6134
process
warning a recent release failed the following release prs may have failed
1
17,458
23,282,230,579
IssuesEvent
2022-08-05 13:13:17
YingMa0107/CARD
https://api.github.com/repos/YingMa0107/CARD
closed
Spots were removed after deconvolution
preprocess
Dear @YingMa0107 , thanks for your work! When I used CARD for deconvolution, I found some spots were removed in the deconvolution results, how could I preserve these deleted spots? I also tried to set 'minCountGene' and 'minCountSpot' as 0, but several spots were also deleted.
1.0
Spots were removed after deconvolution - Dear @YingMa0107 , thanks for your work! When I used CARD for deconvolution, I found some spots were removed in the deconvolution results, how could I preserve these deleted spots? I also tried to set 'minCountGene' and 'minCountSpot' as 0, but several spots were also deleted.
process
spots were removed after deconvolution dear thanks for your work when i used card for deconvolution i found some spots were removed in the deconvolution results how could i preserve these deleted spots i also tried to set mincountgene and mincountspot as but several spots were also deleted
1
1,009
3,475,408,749
IssuesEvent
2015-12-25 15:54:02
Forket/connect2sa.co.za_01
https://api.github.com/repos/Forket/connect2sa.co.za_01
opened
"delete website" form. Client side validation.
enhancement In process
I tried to send a "delete website" message, and I filled in all fields, except of phone number, and it said failed, because I hadn't filled in all fields. Can you please check that? Can you please change it to check fields before form has been sent?
1.0
"delete website" form. Client side validation. - I tried to send a "delete website" message, and I filled in all fields, except of phone number, and it said failed, because I hadn't filled in all fields. Can you please check that? Can you please change it to check fields before form has been sent?
process
delete website form client side validation i tried to send a delete website message and i filled in all fields except of phone number and it said failed because i hadn t filled in all fields can you please check that can you please change it to check fields before form has been sent
1
20,305
26,946,777,241
IssuesEvent
2023-02-08 08:44:58
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Running custom CLI versions on NixOS
process/candidate kind/tech team/schema
Right now, running a specific CLI version on NixOS for me means: - grab a specific engines locally and build it - ``cp target/debug/libquery_engine.so target/debug/libquery_engine.node`` - grab a specific version of prisma/prisma locally and build it - make sure the environment variables point to the locally built engines - (not even entirely sure it will use the correct prisma-fmt wasm, should check) I'm not sure we can literally fix this for NixOS such that I can actually run a `npm` packaged version of the CLI, but at least I could have a nix derivation which takes two hashes as input (engines and prisma/prisma) and does everything itself. I would then at least have some level of confidence that I did not forget or mess something up.
1.0
Running custom CLI versions on NixOS - Right now, running a specific CLI version on NixOS for me means: - grab a specific engines locally and build it - ``cp target/debug/libquery_engine.so target/debug/libquery_engine.node`` - grab a specific version of prisma/prisma locally and build it - make sure the environment variables point to the locally built engines - (not even entirely sure it will use the correct prisma-fmt wasm, should check) I'm not sure we can literally fix this for NixOS such that I can actually run a `npm` packaged version of the CLI, but at least I could have a nix derivation which takes two hashes as input (engines and prisma/prisma) and does everything itself. I would then at least have some level of confidence that I did not forget or mess something up.
process
running custom cli versions on nixos right now running a specific cli version on nixos for me means grab a specific engines locally and build it cp target debug libquery engine so target debug libquery engine node grab a specific version of prisma prisma locally and build it make sure the environment variables point to the locally built engines not even entirely sure it will use the correct prisma fmt wasm should check i m not sure we can literally fix this for nixos such that i can actually run a npm packaged version of the cli but at least i could have a nix derivation which takes two hashes as input engines and prisma prisma and does everything itself i would then at least have some level of confidence that i did not forget or mess something up
1
626,384
19,809,580,082
IssuesEvent
2022-01-19 10:41:56
netdata/netdata-cloud
https://api.github.com/repos/netdata/netdata-cloud
closed
[BUG] Settings temporarily turning themselves on after page reload
bug priority/medium mgmt-navigation-team-bugs
How I have been reproducing the bug: Step 1: Click cog wheel to manage space settings (make sure both invitation settings disabled before clicking settings) ![image](https://user-images.githubusercontent.com/79473923/146083548-0326119f-90d5-4512-9aef-47da7393e840.png) Step 2: While the settings window is open reload the page ![image](https://user-images.githubusercontent.com/79473923/146084047-f2d3deb9-3267-4110-9283-da142211a813.png) the invitation options should be temporarily enabled ![image](https://user-images.githubusercontent.com/79473923/146084116-f93dc272-8ebf-48ae-be01-c3e14dc7907d.png) Step 3: Close the settings page without saving, then re-enter, settings should appear as normal.
1.0
[BUG] Settings temporarily turning themselves on after page reload - How I have been reproducing the bug: Step 1: Click cog wheel to manage space settings (make sure both invitation settings disabled before clicking settings) ![image](https://user-images.githubusercontent.com/79473923/146083548-0326119f-90d5-4512-9aef-47da7393e840.png) Step 2: While the settings window is open reload the page ![image](https://user-images.githubusercontent.com/79473923/146084047-f2d3deb9-3267-4110-9283-da142211a813.png) the invitation options should be temporarily enabled ![image](https://user-images.githubusercontent.com/79473923/146084116-f93dc272-8ebf-48ae-be01-c3e14dc7907d.png) Step 3: Close the settings page without saving, then re-enter, settings should appear as normal.
non_process
settings temporarily turning themselves on after page reload how i have been reproducing the bug step click cog wheel to manage space settings make sure both invitation settings disabled before clicking settings step while the settings window is open reload the page the invitation options should be temporarily enabled step close the settings page without saving then re enter settings should appear as normal
0
272,550
29,795,043,310
IssuesEvent
2023-06-16 01:06:27
billmcchesney1/flowgate
https://api.github.com/repos/billmcchesney1/flowgate
closed
CVE-2019-14439 (High) detected in jackson-databind-2.8.8.jar - autoclosed
Mend: dependency security vulnerability
## CVE-2019-14439 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /operation-expert/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-hateoas-1.4.7.RELEASE.jar (Root Library) - spring-boot-starter-web-1.4.7.RELEASE.jar - :x: **jackson-databind-2.8.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x before 2.9.9.2. This occurs when Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the logback jar in the classpath. <p>Publish Date: 2019-07-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-14439>CVE-2019-14439</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14439">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14439</a></p> <p>Release Date: 2019-07-30</p> <p>Fix Resolution: 2.9.9.2</p> </p> </details> <p></p>
True
CVE-2019-14439 (High) detected in jackson-databind-2.8.8.jar - autoclosed - ## CVE-2019-14439 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /operation-expert/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-hateoas-1.4.7.RELEASE.jar (Root Library) - spring-boot-starter-web-1.4.7.RELEASE.jar - :x: **jackson-databind-2.8.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x before 2.9.9.2. This occurs when Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the logback jar in the classpath. <p>Publish Date: 2019-07-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-14439>CVE-2019-14439</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14439">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14439</a></p> <p>Release Date: 2019-07-30</p> <p>Fix Resolution: 2.9.9.2</p> </p> </details> <p></p>
non_process
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file operation expert pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter hateoas release jar root library spring boot starter web release jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind x before this occurs when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the logback jar in the classpath publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
300,848
9,212,877,817
IssuesEvent
2019-03-10 06:10:16
CS2103-AY1819S2-W13-3/main
https://api.github.com/repos/CS2103-AY1819S2-W13-3/main
closed
As a user I can search multiple tags simultaneously
priority.High type.Story
So I don't have to make multiple searches to see information across multiple categories.
1.0
As a user I can search multiple tags simultaneously - So I don't have to make multiple searches to see information across multiple categories.
non_process
as a user i can search multiple tags simultaneously so i don t have to make multiple searches to see information across multiple categories
0
20,280
26,912,279,338
IssuesEvent
2023-02-07 01:27:53
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
closed
GA release of @google-cloud/life-sciences
priority: p2 type: process api: lifesciences
Package name: **@google-cloud/life-sciences** Current release: **BETA** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
GA release of @google-cloud/life-sciences - Package name: **@google-cloud/life-sciences** Current release: **BETA** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
ga release of google cloud life sciences package name google cloud life sciences current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
11,572
14,441,801,232
IssuesEvent
2020-12-07 17:15:44
pacificclimate/quail
https://api.github.com/repos/pacificclimate/quail
opened
climdex data pre-processing
process
## Description This process takes `csv` or `rda` files to generate `climdexInput` object ## Function to wrap [climdexInput.raw](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L541) and [climdexInput.csv](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L675)
1.0
climdex data pre-processing - ## Description This process takes `csv` or `rda` files to generate `climdexInput` object ## Function to wrap [climdexInput.raw](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L541) and [climdexInput.csv](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L675)
process
climdex data pre processing description this process takes csv or rda files to generate climdexinput object function to wrap and
1
231,057
18,735,856,596
IssuesEvent
2021-11-04 07:24:49
apache/camel-quarkus
https://api.github.com/repos/apache/camel-quarkus
closed
Sql - enable stored procedure test for different db types
integration-test
[PR](https://github.com/apache/camel-quarkus/pull/3066) introduces support of different db kinds for testing. Test `testSqlStoredComponent` is enabled only for `h2` db type. It should be possible to modify test to cover different db kinds.
1.0
Sql - enable stored procedure test for different db types - [PR](https://github.com/apache/camel-quarkus/pull/3066) introduces support of different db kinds for testing. Test `testSqlStoredComponent` is enabled only for `h2` db type. It should be possible to modify test to cover different db kinds.
non_process
sql enable stored procedure test for different db types introduces support of different db kinds for testing test testsqlstoredcomponent is enabled only for db type it should be possible to modify test to cover different db kinds
0
37,968
10,118,569,702
IssuesEvent
2019-07-31 09:22:17
ginkgo-project/ginkgo
https://api.github.com/repos/ginkgo-project/ginkgo
opened
Problem with test executables with the same name as standard library headers
Bug Build Testing
Since some recent upgrade of the containers, namely the ones with the following configuration: + CUDA 10.0.130, GCC 7.4.0-1ubuntu1~18.04.1, CLANG 6.0.0-1ubuntu2; + CUDA 9.2.148, GCC 7.4.0-1ubuntu1~18.04.1, CLANG 5.0.1-4. There is a very weird bug for tests in `core/test/base/` which have the same name as some C++ standard header which Ginkgo also uses. Namely, `exception` and `array`. The bug is the following, which to my understanding should not happen: + The Ginkgo tests executables for Ginkgo specific `exception` or `array` were compiled, and there is now in the `core/test/base/` executables named either `array` or `exception`. + One file inside this test either directly or transitively through various headers includes one of the standard headers `exception` or `array`. I believe this always happens with the following form: `#include <exception>`. Let's call this test `test x`. + When compiling test, for some reason the compiler thinks it is absolutely valid to include the previously compiled **executable** `exception` or `array` instead of using the actual header file from the standard library. This happens with either `clang` or `gcc`. + The compilation then fails horribly because the compiler just included a file in a binary format which it cannot even read. See the failing jobs in this pipeline for some examples: https://gitlab.com/ginkgo-project/ginkgo-public-ci/pipelines/73545162 When looking into the default configuration for g++ on the container, we can clearly see the problem: <details><summary>**g++ default include paths and settings on the container**</summary><p> --- ```bash $ g++ -xc++ -E -v - Using built-in specs. COLLECT_GCC=g++ OFFLOAD_TARGET_NAMES=nvptx-none OFFLOAD_TARGET_DEFAULT=1 Target: x86_64-linux-gnu Configured with: ../src/configure -v --with-pkgversion='Ubuntu 7.4.0-1ubuntu1~18.04.1' --with-bugurl=file:///usr/share/doc/gcc-7/README.Bugs --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --prefix=/usr --with-gcc-major-version-only --program-suffix=-7 --program-prefix=x86_64-linux-gnu- --enable-shared --enable-linker-build-id --libexecdir=/usr/lib --without-included-gettext --enable-threads=posix --libdir=/usr/lib --enable-nls --with-sysroot=/ --enable-clocale=gnu --enable-libstdcxx-debug --enable-libstdcxx-time=yes --with-default-libstdcxx-abi=new --enable-gnu-unique-object --disable-vtable-verify --enable-libmpx --enable-plugin --enable-default-pie --with-system-zlib --with-target-system-zlib --enable-objc-gc=auto --enable-multiarch --disable-werror --with-arch-32=i686 --with-abi=m64 --with-multilib-list=m32,m64,mx32 --enable-multilib --with-tune=generic --enable-offload-targets=nvptx-none --without-cuda-driver --enable-checking=release --build=x86_64-linux-gnu --host=x86_64-linux-gnu --target=x86_64-linux-gnu Thread model: posix gcc version 7.4.0 (Ubuntu 7.4.0-1ubuntu1~18.04.1) COLLECT_GCC_OPTIONS='-E' '-v' '-shared-libgcc' '-mtune=generic' '-march=x86-64' /usr/lib/gcc/x86_64-linux-gnu/7/cc1plus -E -quiet -v -imultiarch x86_64-linux-gnu -D_GNU_SOURCE - -mtune=generic -march=x86-64 -fstack-protector-strong -Wformat -Wformat-security ignoring duplicate directory "/usr/include/x86_64-linux-gnu/c++/7" ignoring nonexistent directory "/usr/local/include/x86_64-linux-gnu" ignoring nonexistent directory "/usr/lib/gcc/x86_64-linux-gnu/7/../../../../x86_64-linux-gnu/include" #include "..." search starts here: #include <...> search starts here: . /usr/local/cuda/include /usr/include/c++/7 /usr/include/x86_64-linux-gnu/c++/7 /usr/include/c++/7/backward /usr/lib/gcc/x86_64-linux-gnu/7/include /usr/local/include /usr/lib/gcc/x86_64-linux-gnu/7/include-fixed /usr/include/x86_64-linux-gnu /usr/include End of search list. ``` --- </p></details> <details><summary>g++ default include paths and settings on personal machine</summary><p> --- ```bash g++ -xc++ -E -v - Using built-in specs. COLLECT_GCC=g++ OFFLOAD_TARGET_NAMES=nvptx-none OFFLOAD_TARGET_DEFAULT=1 Target: x86_64-linux-gnu Configured with: ../src/configure -v --with-pkgversion='Debian 8.3.0-6' --with-bugurl=file:///usr/share/doc/gcc-8/README.Bugs --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --prefix=/usr --with-gcc-major-version-only --program-suffix=-8 --program-prefix=x86_64-linux-gnu- --enable-shared --enable-linker-build-id --libexecdir=/usr/lib --without-included-gettext --enable-threads=posix --libdir=/usr/lib --enable-nls --enable-bootstrap --enable-clocale=gnu --enable-libstdcxx-debug --enable-libstdcxx-time=yes --with-default-libstdcxx-abi=new --enable-gnu-unique-object --disable-vtable-verify --enable-libmpx --enable-plugin --enable-default-pie --with-system-zlib --with-target-system-zlib --enable-objc-gc=auto --enable-multiarch --disable-werror --with-arch-32=i686 --with-abi=m64 --with-multilib-list=m32,m64,mx32 --enable-multilib --with-tune=generic --enable-offload-targets=nvptx-none --without-cuda-driver --enable-checking=release --build=x86_64-linux-gnu --host=x86_64-linux-gnu --target=x86_64-linux-gnu Thread model: posix gcc version 8.3.0 (Debian 8.3.0-6) COLLECT_GCC_OPTIONS='-E' '-v' '-shared-libgcc' '-mtune=generic' '-march=x86-64' /usr/lib/gcc/x86_64-linux-gnu/8/cc1plus -E -quiet -v -imultiarch x86_64-linux-gnu -D_GNU_SOURCE - -mtune=generic -march=x86-64 ignoring duplicate directory "/usr/include/x86_64-linux-gnu/c++/8" ignoring nonexistent directory "/usr/local/include/x86_64-linux-gnu" ignoring nonexistent directory "/usr/lib/gcc/x86_64-linux-gnu/8/../../../../x86_64-linux-gnu/include" #include "..." search starts here: #include <...> search starts here: /usr/include/c++/8 /usr/include/x86_64-linux-gnu/c++/8 /usr/include/c++/8/backward /usr/lib/gcc/x86_64-linux-gnu/8/include /usr/local/include /usr/lib/gcc/x86_64-linux-gnu/8/include-fixed /usr/include/x86_64-linux-gnu /usr/include End of search list. ``` </p></details> The culprit here is in the container case the **.** which appears only there in the default search pathes for both `""` and `<>` include forms, but for some reason is not at all present anywhere else that I know of. ### Solutions #### Solution 1: Telling GCC to never use the current directory for includes There is supposed to be an option `-I-` which does that, but it seems to be deprecated and the behavior cannot be reproduced. I do not know of any other way to do that, but if this can be done it would probably be the best thing to do. https://stackoverflow.com/questions/12353429/prevent-gcc-from-searching-the-current-dir-i-option-on-include-search-path #### Solution 2: Compiling or finding better compilers for the container? This is a bit annoying, but I don't this this behavior of including the current directory is common in compilers, and in general should not happen. If that holds true (need research), then we can see this problem as the particular compilers we use in the container to be faulty, in which case we could try finding non faulty ones. #### Solution 3: Rename all problematic ginkgo executables to be different from the standard One solution is to prefix all tests and other executables with `gko_` so that the name does not clash with the standard files. Another solution would be for the standard files to use a proper `.h` or `.hpp` for their files although I can't see that happening...
1.0
Problem with test executables with the same name as standard library headers - Since some recent upgrade of the containers, namely the ones with the following configuration: + CUDA 10.0.130, GCC 7.4.0-1ubuntu1~18.04.1, CLANG 6.0.0-1ubuntu2; + CUDA 9.2.148, GCC 7.4.0-1ubuntu1~18.04.1, CLANG 5.0.1-4. There is a very weird bug for tests in `core/test/base/` which have the same name as some C++ standard header which Ginkgo also uses. Namely, `exception` and `array`. The bug is the following, which to my understanding should not happen: + The Ginkgo tests executables for Ginkgo specific `exception` or `array` were compiled, and there is now in the `core/test/base/` executables named either `array` or `exception`. + One file inside this test either directly or transitively through various headers includes one of the standard headers `exception` or `array`. I believe this always happens with the following form: `#include <exception>`. Let's call this test `test x`. + When compiling test, for some reason the compiler thinks it is absolutely valid to include the previously compiled **executable** `exception` or `array` instead of using the actual header file from the standard library. This happens with either `clang` or `gcc`. + The compilation then fails horribly because the compiler just included a file in a binary format which it cannot even read. See the failing jobs in this pipeline for some examples: https://gitlab.com/ginkgo-project/ginkgo-public-ci/pipelines/73545162 When looking into the default configuration for g++ on the container, we can clearly see the problem: <details><summary>**g++ default include paths and settings on the container**</summary><p> --- ```bash $ g++ -xc++ -E -v - Using built-in specs. COLLECT_GCC=g++ OFFLOAD_TARGET_NAMES=nvptx-none OFFLOAD_TARGET_DEFAULT=1 Target: x86_64-linux-gnu Configured with: ../src/configure -v --with-pkgversion='Ubuntu 7.4.0-1ubuntu1~18.04.1' --with-bugurl=file:///usr/share/doc/gcc-7/README.Bugs --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --prefix=/usr --with-gcc-major-version-only --program-suffix=-7 --program-prefix=x86_64-linux-gnu- --enable-shared --enable-linker-build-id --libexecdir=/usr/lib --without-included-gettext --enable-threads=posix --libdir=/usr/lib --enable-nls --with-sysroot=/ --enable-clocale=gnu --enable-libstdcxx-debug --enable-libstdcxx-time=yes --with-default-libstdcxx-abi=new --enable-gnu-unique-object --disable-vtable-verify --enable-libmpx --enable-plugin --enable-default-pie --with-system-zlib --with-target-system-zlib --enable-objc-gc=auto --enable-multiarch --disable-werror --with-arch-32=i686 --with-abi=m64 --with-multilib-list=m32,m64,mx32 --enable-multilib --with-tune=generic --enable-offload-targets=nvptx-none --without-cuda-driver --enable-checking=release --build=x86_64-linux-gnu --host=x86_64-linux-gnu --target=x86_64-linux-gnu Thread model: posix gcc version 7.4.0 (Ubuntu 7.4.0-1ubuntu1~18.04.1) COLLECT_GCC_OPTIONS='-E' '-v' '-shared-libgcc' '-mtune=generic' '-march=x86-64' /usr/lib/gcc/x86_64-linux-gnu/7/cc1plus -E -quiet -v -imultiarch x86_64-linux-gnu -D_GNU_SOURCE - -mtune=generic -march=x86-64 -fstack-protector-strong -Wformat -Wformat-security ignoring duplicate directory "/usr/include/x86_64-linux-gnu/c++/7" ignoring nonexistent directory "/usr/local/include/x86_64-linux-gnu" ignoring nonexistent directory "/usr/lib/gcc/x86_64-linux-gnu/7/../../../../x86_64-linux-gnu/include" #include "..." search starts here: #include <...> search starts here: . /usr/local/cuda/include /usr/include/c++/7 /usr/include/x86_64-linux-gnu/c++/7 /usr/include/c++/7/backward /usr/lib/gcc/x86_64-linux-gnu/7/include /usr/local/include /usr/lib/gcc/x86_64-linux-gnu/7/include-fixed /usr/include/x86_64-linux-gnu /usr/include End of search list. ``` --- </p></details> <details><summary>g++ default include paths and settings on personal machine</summary><p> --- ```bash g++ -xc++ -E -v - Using built-in specs. COLLECT_GCC=g++ OFFLOAD_TARGET_NAMES=nvptx-none OFFLOAD_TARGET_DEFAULT=1 Target: x86_64-linux-gnu Configured with: ../src/configure -v --with-pkgversion='Debian 8.3.0-6' --with-bugurl=file:///usr/share/doc/gcc-8/README.Bugs --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --prefix=/usr --with-gcc-major-version-only --program-suffix=-8 --program-prefix=x86_64-linux-gnu- --enable-shared --enable-linker-build-id --libexecdir=/usr/lib --without-included-gettext --enable-threads=posix --libdir=/usr/lib --enable-nls --enable-bootstrap --enable-clocale=gnu --enable-libstdcxx-debug --enable-libstdcxx-time=yes --with-default-libstdcxx-abi=new --enable-gnu-unique-object --disable-vtable-verify --enable-libmpx --enable-plugin --enable-default-pie --with-system-zlib --with-target-system-zlib --enable-objc-gc=auto --enable-multiarch --disable-werror --with-arch-32=i686 --with-abi=m64 --with-multilib-list=m32,m64,mx32 --enable-multilib --with-tune=generic --enable-offload-targets=nvptx-none --without-cuda-driver --enable-checking=release --build=x86_64-linux-gnu --host=x86_64-linux-gnu --target=x86_64-linux-gnu Thread model: posix gcc version 8.3.0 (Debian 8.3.0-6) COLLECT_GCC_OPTIONS='-E' '-v' '-shared-libgcc' '-mtune=generic' '-march=x86-64' /usr/lib/gcc/x86_64-linux-gnu/8/cc1plus -E -quiet -v -imultiarch x86_64-linux-gnu -D_GNU_SOURCE - -mtune=generic -march=x86-64 ignoring duplicate directory "/usr/include/x86_64-linux-gnu/c++/8" ignoring nonexistent directory "/usr/local/include/x86_64-linux-gnu" ignoring nonexistent directory "/usr/lib/gcc/x86_64-linux-gnu/8/../../../../x86_64-linux-gnu/include" #include "..." search starts here: #include <...> search starts here: /usr/include/c++/8 /usr/include/x86_64-linux-gnu/c++/8 /usr/include/c++/8/backward /usr/lib/gcc/x86_64-linux-gnu/8/include /usr/local/include /usr/lib/gcc/x86_64-linux-gnu/8/include-fixed /usr/include/x86_64-linux-gnu /usr/include End of search list. ``` </p></details> The culprit here is in the container case the **.** which appears only there in the default search pathes for both `""` and `<>` include forms, but for some reason is not at all present anywhere else that I know of. ### Solutions #### Solution 1: Telling GCC to never use the current directory for includes There is supposed to be an option `-I-` which does that, but it seems to be deprecated and the behavior cannot be reproduced. I do not know of any other way to do that, but if this can be done it would probably be the best thing to do. https://stackoverflow.com/questions/12353429/prevent-gcc-from-searching-the-current-dir-i-option-on-include-search-path #### Solution 2: Compiling or finding better compilers for the container? This is a bit annoying, but I don't this this behavior of including the current directory is common in compilers, and in general should not happen. If that holds true (need research), then we can see this problem as the particular compilers we use in the container to be faulty, in which case we could try finding non faulty ones. #### Solution 3: Rename all problematic ginkgo executables to be different from the standard One solution is to prefix all tests and other executables with `gko_` so that the name does not clash with the standard files. Another solution would be for the standard files to use a proper `.h` or `.hpp` for their files although I can't see that happening...
non_process
problem with test executables with the same name as standard library headers since some recent upgrade of the containers namely the ones with the following configuration cuda gcc clang cuda gcc clang there is a very weird bug for tests in core test base which have the same name as some c standard header which ginkgo also uses namely exception and array the bug is the following which to my understanding should not happen the ginkgo tests executables for ginkgo specific exception or array were compiled and there is now in the core test base executables named either array or exception one file inside this test either directly or transitively through various headers includes one of the standard headers exception or array i believe this always happens with the following form include let s call this test test x when compiling test for some reason the compiler thinks it is absolutely valid to include the previously compiled executable exception or array instead of using the actual header file from the standard library this happens with either clang or gcc the compilation then fails horribly because the compiler just included a file in a binary format which it cannot even read see the failing jobs in this pipeline for some examples when looking into the default configuration for g on the container we can clearly see the problem g default include paths and settings on the container bash g xc e v using built in specs collect gcc g offload target names nvptx none offload target default target linux gnu configured with src configure v with pkgversion ubuntu with bugurl file usr share doc gcc readme bugs enable languages c ada c go brig d fortran objc obj c prefix usr with gcc major version only program suffix program prefix linux gnu enable shared enable linker build id libexecdir usr lib without included gettext enable threads posix libdir usr lib enable nls with sysroot enable clocale gnu enable libstdcxx debug enable libstdcxx time yes with default libstdcxx abi new enable gnu unique object disable vtable verify enable libmpx enable plugin enable default pie with system zlib with target system zlib enable objc gc auto enable multiarch disable werror with arch with abi with multilib list enable multilib with tune generic enable offload targets nvptx none without cuda driver enable checking release build linux gnu host linux gnu target linux gnu thread model posix gcc version ubuntu collect gcc options e v shared libgcc mtune generic march usr lib gcc linux gnu e quiet v imultiarch linux gnu d gnu source mtune generic march fstack protector strong wformat wformat security ignoring duplicate directory usr include linux gnu c ignoring nonexistent directory usr local include linux gnu ignoring nonexistent directory usr lib gcc linux gnu linux gnu include include search starts here include search starts here usr local cuda include usr include c usr include linux gnu c usr include c backward usr lib gcc linux gnu include usr local include usr lib gcc linux gnu include fixed usr include linux gnu usr include end of search list g default include paths and settings on personal machine bash g xc e v using built in specs collect gcc g offload target names nvptx none offload target default target linux gnu configured with src configure v with pkgversion debian with bugurl file usr share doc gcc readme bugs enable languages c ada c go brig d fortran objc obj c prefix usr with gcc major version only program suffix program prefix linux gnu enable shared enable linker build id libexecdir usr lib without included gettext enable threads posix libdir usr lib enable nls enable bootstrap enable clocale gnu enable libstdcxx debug enable libstdcxx time yes with default libstdcxx abi new enable gnu unique object disable vtable verify enable libmpx enable plugin enable default pie with system zlib with target system zlib enable objc gc auto enable multiarch disable werror with arch with abi with multilib list enable multilib with tune generic enable offload targets nvptx none without cuda driver enable checking release build linux gnu host linux gnu target linux gnu thread model posix gcc version debian collect gcc options e v shared libgcc mtune generic march usr lib gcc linux gnu e quiet v imultiarch linux gnu d gnu source mtune generic march ignoring duplicate directory usr include linux gnu c ignoring nonexistent directory usr local include linux gnu ignoring nonexistent directory usr lib gcc linux gnu linux gnu include include search starts here include search starts here usr include c usr include linux gnu c usr include c backward usr lib gcc linux gnu include usr local include usr lib gcc linux gnu include fixed usr include linux gnu usr include end of search list the culprit here is in the container case the which appears only there in the default search pathes for both and include forms but for some reason is not at all present anywhere else that i know of solutions solution telling gcc to never use the current directory for includes there is supposed to be an option i which does that but it seems to be deprecated and the behavior cannot be reproduced i do not know of any other way to do that but if this can be done it would probably be the best thing to do solution compiling or finding better compilers for the container this is a bit annoying but i don t this this behavior of including the current directory is common in compilers and in general should not happen if that holds true need research then we can see this problem as the particular compilers we use in the container to be faulty in which case we could try finding non faulty ones solution rename all problematic ginkgo executables to be different from the standard one solution is to prefix all tests and other executables with gko so that the name does not clash with the standard files another solution would be for the standard files to use a proper h or hpp for their files although i can t see that happening
0
5,888
8,707,036,057
IssuesEvent
2018-12-06 06:01:55
SerialLain3170/GAN-papers
https://api.github.com/repos/SerialLain3170/GAN-papers
opened
TGANv2: Efficient Training of Large Models for Video Generation with Multiple Subsampling Layers
GAN Family Video Processing
# Paper [TGANv2: Efficient Training of Large Models for Video Generation with Multiple Subsampling Layers](https://arxiv.org/pdf/1811.09245.pdf) # Summary - TemporalGANの拡張版、AbstとRenderを行う、Discriminatorには3DResNetを用いるなどの改良 - 各層においてランダムにsubsampleするsubsampling layerを導入している ![tgan](https://user-images.githubusercontent.com/32360147/49564604-77b37f00-f967-11e8-94f8-b0eb4ed11d92.png) # Date 2018/11/22
1.0
TGANv2: Efficient Training of Large Models for Video Generation with Multiple Subsampling Layers - # Paper [TGANv2: Efficient Training of Large Models for Video Generation with Multiple Subsampling Layers](https://arxiv.org/pdf/1811.09245.pdf) # Summary - TemporalGANの拡張版、AbstとRenderを行う、Discriminatorには3DResNetを用いるなどの改良 - 各層においてランダムにsubsampleするsubsampling layerを導入している ![tgan](https://user-images.githubusercontent.com/32360147/49564604-77b37f00-f967-11e8-94f8-b0eb4ed11d92.png) # Date 2018/11/22
process
efficient training of large models for video generation with multiple subsampling layers paper summary temporalganの拡張版、abstとrenderを行う、 各層においてランダムにsubsampleするsubsampling layerを導入している date
1
828,200
31,816,465,954
IssuesEvent
2023-09-13 20:53:08
googleapis/nodejs-bigquery
https://api.github.com/repos/googleapis/nodejs-bigquery
opened
TypeError: Cannot read properties of undefined (reading 'insertErrors') thrown when processing errors
priority: p2 type: bug
1) Is this a client library issue or a product issue? A client library issue. 2) Did someone already solve this? No. 3) Do you have a support contract? No. #### Environment details - OS: Ubuntu - Node.js version: 18.17.1 - npm version: 9.6.7 - `@google-cloud/bigquery` version: 7.2.0 #### Steps to reproduce 1. Use `createInputStream` to create an input stream. 2. Send rows through the stream until an `TypeError: Cannot read properties of undefined (reading 'insertErrors')` occurs.
1.0
TypeError: Cannot read properties of undefined (reading 'insertErrors') thrown when processing errors - 1) Is this a client library issue or a product issue? A client library issue. 2) Did someone already solve this? No. 3) Do you have a support contract? No. #### Environment details - OS: Ubuntu - Node.js version: 18.17.1 - npm version: 9.6.7 - `@google-cloud/bigquery` version: 7.2.0 #### Steps to reproduce 1. Use `createInputStream` to create an input stream. 2. Send rows through the stream until an `TypeError: Cannot read properties of undefined (reading 'insertErrors')` occurs.
non_process
typeerror cannot read properties of undefined reading inserterrors thrown when processing errors is this a client library issue or a product issue a client library issue did someone already solve this no do you have a support contract no environment details os ubuntu node js version npm version google cloud bigquery version steps to reproduce use createinputstream to create an input stream send rows through the stream until an typeerror cannot read properties of undefined reading inserterrors occurs
0
11,777
14,612,801,955
IssuesEvent
2020-12-22 06:57:38
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
opened
Incorrect script processing in the case of repeating top-level properties in destructuring of nested objects
AREA: server SYSTEM: script processing health-monitor
Found in Health Monitor results (https://www.instagram.com/). Originnal script (./static/bundles/es6/Consumer.js/b75e6b6ebcdc.js): ```js e.getPromptButtonRenderPropsFromScreenPayload = function(t) { const {response: {prompt_button: n}, response: {primary_button: o}} = t , l = n || o; if (!l) return null; const {action_type: u, title: c, url: s} = l; return null == u ? null : { actionType: u, label: (null === c || void 0 === c ? void 0 : c.text) || r(d[0]).DEFAULT_PROMPT_BUTTON_LABEL, href: s, canSubmitOnClick: !s } } ``` Test (Uncaught SyntaxError: Identifier '_hh$temp0$foo' has already been declared): ```js it('repeating top-level properties in nested destructuring', () => { testProcessing([ { src: 'const {foo: {bar: var1}, foo: {baz: var2}} = o', expected: '' // const _hh$temp0 = o, _hh$temp0$foo = _hh$temp0.foo, var1 = _hh$temp0$foo.bar, _hh$temp0$foo = _hh$temp0.foo, var2 = _hh$temp0$foo.baz }, { src: 't = function (o) {' + ' const {foo: {bar: v1}, foo: {baz: v2}} = o' + '}', expected: '' // t = function (o) { const _hh$temp0=o, _hh$temp0$foo=_hh$temp0.foo, v1 = _hh$temp0$foo.bar, _hh$temp0$foo = _hh$temp0.foo, v2=_hh$temp0$foo.baz } } ]); }); ```
1.0
Incorrect script processing in the case of repeating top-level properties in destructuring of nested objects - Found in Health Monitor results (https://www.instagram.com/). Originnal script (./static/bundles/es6/Consumer.js/b75e6b6ebcdc.js): ```js e.getPromptButtonRenderPropsFromScreenPayload = function(t) { const {response: {prompt_button: n}, response: {primary_button: o}} = t , l = n || o; if (!l) return null; const {action_type: u, title: c, url: s} = l; return null == u ? null : { actionType: u, label: (null === c || void 0 === c ? void 0 : c.text) || r(d[0]).DEFAULT_PROMPT_BUTTON_LABEL, href: s, canSubmitOnClick: !s } } ``` Test (Uncaught SyntaxError: Identifier '_hh$temp0$foo' has already been declared): ```js it('repeating top-level properties in nested destructuring', () => { testProcessing([ { src: 'const {foo: {bar: var1}, foo: {baz: var2}} = o', expected: '' // const _hh$temp0 = o, _hh$temp0$foo = _hh$temp0.foo, var1 = _hh$temp0$foo.bar, _hh$temp0$foo = _hh$temp0.foo, var2 = _hh$temp0$foo.baz }, { src: 't = function (o) {' + ' const {foo: {bar: v1}, foo: {baz: v2}} = o' + '}', expected: '' // t = function (o) { const _hh$temp0=o, _hh$temp0$foo=_hh$temp0.foo, v1 = _hh$temp0$foo.bar, _hh$temp0$foo = _hh$temp0.foo, v2=_hh$temp0$foo.baz } } ]); }); ```
process
incorrect script processing in the case of repeating top level properties in destructuring of nested objects found in health monitor results originnal script static bundles consumer js js js e getpromptbuttonrenderpropsfromscreenpayload function t const response prompt button n response primary button o t l n o if l return null const action type u title c url s l return null u null actiontype u label null c void c void c text r d default prompt button label href s cansubmitonclick s test uncaught syntaxerror identifier hh foo has already been declared js it repeating top level properties in nested destructuring testprocessing src const foo bar foo baz o expected const hh o hh foo hh foo hh foo bar hh foo hh foo hh foo baz src t function o const foo bar foo baz o expected t function o const hh o hh foo hh foo hh foo bar hh foo hh foo hh foo baz
1
3,131
6,187,608,855
IssuesEvent
2017-07-04 08:01:49
BlesseNtumble/GalaxySpace
https://api.github.com/repos/BlesseNtumble/GalaxySpace
closed
client crash (1.7.10) [GS 1.1.4]
bug in the process of correcting
Hello one of player on my server has this crash: https://paste.ee/p/nUzxi caused everytime he make something in his home, not caused to someone alse in the same location can you please help? thanks in advance
1.0
client crash (1.7.10) [GS 1.1.4] - Hello one of player on my server has this crash: https://paste.ee/p/nUzxi caused everytime he make something in his home, not caused to someone alse in the same location can you please help? thanks in advance
process
client crash hello one of player on my server has this crash caused everytime he make something in his home not caused to someone alse in the same location can you please help thanks in advance
1
344,614
10,347,348,325
IssuesEvent
2019-09-04 17:10:58
poanetwork/blockscout
https://api.github.com/repos/poanetwork/blockscout
opened
ERC-721 images on inventory tab of token page
priority: high
Blockscout doesn't show images of NFT token instances. For instance, https://blockscout.com/eth/mainnet/tokens/0x9cb151054ab13fa715f170c731a9705236e0a537/inventory Let's add images of token instances to *Inventory* tab of the *token* page
1.0
ERC-721 images on inventory tab of token page - Blockscout doesn't show images of NFT token instances. For instance, https://blockscout.com/eth/mainnet/tokens/0x9cb151054ab13fa715f170c731a9705236e0a537/inventory Let's add images of token instances to *Inventory* tab of the *token* page
non_process
erc images on inventory tab of token page blockscout doesn t show images of nft token instances for instance let s add images of token instances to inventory tab of the token page
0
68,756
21,878,958,104
IssuesEvent
2022-05-19 12:50:36
rgpjones/rotabot
https://api.github.com/repos/rgpjones/rotabot
opened
Migrate to using a Slack App for integration
defect
The current implementation uses a simple slash command webhook to post the request to Rotabot [isitmyturn.uk], and a response is posted back. This mechanism is deprecated by Slack, and is likely to be removed in the future. As such, Rotabot needs to be refactored as a Slack App.
1.0
Migrate to using a Slack App for integration - The current implementation uses a simple slash command webhook to post the request to Rotabot [isitmyturn.uk], and a response is posted back. This mechanism is deprecated by Slack, and is likely to be removed in the future. As such, Rotabot needs to be refactored as a Slack App.
non_process
migrate to using a slack app for integration the current implementation uses a simple slash command webhook to post the request to rotabot and a response is posted back this mechanism is deprecated by slack and is likely to be removed in the future as such rotabot needs to be refactored as a slack app
0
113,891
4,580,917,061
IssuesEvent
2016-09-19 00:24:50
sainteos/project-spero
https://api.github.com/repos/sainteos/project-spero
closed
Debug tools need to be created
feature medpriority
1. The debug tools need to be able to be turned on and off in the configuration. 2. Free camera movement: The camera reacts directly to WASD and is not tied to tile quantization. 3. Hit tab to be presented with a command window. Hitting escape should make it disappear. It should be able to be typed in. 4. Ability to change maps without restarting: Open the command window. Type "/load map [MapName]", and it pulls it from the maps folder specified in the config, loads it in place, and renders. 5. Open the command window. Typing "/list characters" to see a list of the available characters as specified by the configuration. 6. Typing "/load character [CharacterName]" should replace the current controllable character with the character with CharacterName. 7. HUD Text with current x/y pos (tile if in character control mode, float if in camera control mode), Current FPS, Current Character Name, Current Map Name 8. Typing "/lights off" should turn the lighting off. This has no effect if the map is not lit. 9. Typing "/lights on" should turn the lighting on. This has no effect if the map is not lit. 10. Typing "/list layers" to see a list of all of the layers on this map, with indices. 11. Typing "/layer [IndexNumber or LayerName] on" turns the layer specified on. 12. Typing "/layer [IndexNumber or LayerName] off" turns the layer specified off. 13. Typing "/layer [IndexNumber or LayerName] animations on" turns map animations on. 14. Typing "/layer [IndexNumber or LayerName] animations off" turns map animations off.
1.0
Debug tools need to be created - 1. The debug tools need to be able to be turned on and off in the configuration. 2. Free camera movement: The camera reacts directly to WASD and is not tied to tile quantization. 3. Hit tab to be presented with a command window. Hitting escape should make it disappear. It should be able to be typed in. 4. Ability to change maps without restarting: Open the command window. Type "/load map [MapName]", and it pulls it from the maps folder specified in the config, loads it in place, and renders. 5. Open the command window. Typing "/list characters" to see a list of the available characters as specified by the configuration. 6. Typing "/load character [CharacterName]" should replace the current controllable character with the character with CharacterName. 7. HUD Text with current x/y pos (tile if in character control mode, float if in camera control mode), Current FPS, Current Character Name, Current Map Name 8. Typing "/lights off" should turn the lighting off. This has no effect if the map is not lit. 9. Typing "/lights on" should turn the lighting on. This has no effect if the map is not lit. 10. Typing "/list layers" to see a list of all of the layers on this map, with indices. 11. Typing "/layer [IndexNumber or LayerName] on" turns the layer specified on. 12. Typing "/layer [IndexNumber or LayerName] off" turns the layer specified off. 13. Typing "/layer [IndexNumber or LayerName] animations on" turns map animations on. 14. Typing "/layer [IndexNumber or LayerName] animations off" turns map animations off.
non_process
debug tools need to be created the debug tools need to be able to be turned on and off in the configuration free camera movement the camera reacts directly to wasd and is not tied to tile quantization hit tab to be presented with a command window hitting escape should make it disappear it should be able to be typed in ability to change maps without restarting open the command window type load map and it pulls it from the maps folder specified in the config loads it in place and renders open the command window typing list characters to see a list of the available characters as specified by the configuration typing load character should replace the current controllable character with the character with charactername hud text with current x y pos tile if in character control mode float if in camera control mode current fps current character name current map name typing lights off should turn the lighting off this has no effect if the map is not lit typing lights on should turn the lighting on this has no effect if the map is not lit typing list layers to see a list of all of the layers on this map with indices typing layer on turns the layer specified on typing layer off turns the layer specified off typing layer animations on turns map animations on typing layer animations off turns map animations off
0
140,622
11,353,727,747
IssuesEvent
2020-01-24 16:06:49
IntellectualSites/PlotSquared
https://api.github.com/repos/IntellectualSites/PlotSquared
opened
Schematic based World Generation is not working correctly
[?] Testing Required
<!--- READ THIS BEFORE SUBMITTING AN ISSUE REPORT!!! --> <!--- ##### DO NOT REMOVE THIS TEMPLATE! YOUR ISSUE *WILL* FIT IN IT! ##### --> <!--- # NOTICE: ```diff ! PlotSquared for Minecraft Java Edition versions between 1.7 through to 1.12.2 are considered ! legacy, and will receive limited to no support. Please consider upgrading to 1.13+ for ! future support. Plugins exist for 1.13+ which bring back behaviors found in 1.8.8 ! All versions of PlotSquared for Sponge and Nukkit(X) will receive limited to no support ! due to lack of developer interest and time. Additionally, NukkitX has not had feature ! updates since the Better Together, which prevents some PlotSquared features from ever ! functioning. Contributions are always welcome however! ``` **Feature requests & Suggestions are to be submitted at the [PlotSquared Suggestions tracker](https://github.com/IntellectualSites/PlotSquaredSuggestions)** **Code contributions are to be done through [PRs](https://help.github.com/en/github/collaborating-with-issues-and-pull-requests/creating-a-pull-request), tagging the specific issue ticket(s) if applicable.** **[DISCORD INVITE LINK](https://discord.gg/cSMxtGn)** and please, for the love of the little sanity we have left, use the correct channels! --> # Bug Report Template: <!--- Incomplete reports will most likely be marked as invalid, and closed, with few exceptions.--> ## Required Information section: > ALL FIELDS IN THIS SECTION ARE REQUIRED, and must contain appropriate information ### Server config info (/plot debugpaste / file paste links): <!--- Issue /plot debugpaste in game or in your console and copy the supplied URL here --> <!--- If you cannot perform the above, we require logs/latest.log; settings.yml; worlds.yml and possibly PlotSquared.use_THIS.yml --> I atteached the files because the first thing doesnt worked: [latest.log](https://github.com/IntellectualSites/PlotSquared/files/4109359/latest.log) [2020-01-23-9.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109363/2020-01-23-9.log.gz) [2020-01-23-10.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109364/2020-01-23-10.log.gz) [yml.zip](https://github.com/IntellectualSites/PlotSquared/files/4109371/yml.zip) <!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. --> ### Server type: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): --> - [X] Spigot / Paper *(CraftBukkit should not be used, re-test with Spigot first!)* - [] Sponge *- NOTE: NOT ACTIVELY MAINTAINED* - [] NukkitX *- NOTE: NOT ACTIVELY MAINTAINED* ### Minecraft Version: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): The maintained versions are 1.14.4 and 1.15.x --> - [] Minecraft 1.15 - [X] Minecraft 1.14.4 - [] Minecraft 1.13.2 - [] Minecraft Java Edition *other versions, please specify*: - [] Minecraft Bedrock Edition *specify version*: ### Server build info: <!--- Run /version in-game or in console & paste the full output here: --> ``` This server is running CraftBukkit version git-Spigot-cbd1a1b-009d8af (MC: 1.14.4) (Implementing API version 1.14.4-R0.1-SNAPSHOT) [16:31:29] [Server thread/INFO]: This is a final build for 1.14.4. Please see https://www.spigotmc.org/go/1.14.4 for details about upgrading. ``` ### WorldEdit/FAWE versions: <!--- Specify which plugin you are using, and add its version --> - [X] FAWE *FastAsyncWorldEdit-1.15-45*: - [] WorldEdit *version*: ### Description of the problem: <!--- Be as specific as possible. Don't lie, redact information, or use false names/situations. --> <!--- Who, What, When, Where, Why, How, Expected behavior, Resultant behavior, etc --> If I want to use the schematic based world gen. with the plot.schem in the right folder for creating a new world, an Error occurs. After the Server has been loaded completely, the world is genarated and I can also enter it, but then there are always half or broken plots. ### How to replicate: <!--- If you can reproduce the issue please tell us as detailed as possible step by step how to do that --> I have done everything like it is described in the wiki, but than this Error occurs. The marked space for the plot schematic is about 200 Blocks in booth x and y direction. Maybe its too big? I also checked the worlds.yml but the Plot size is set to 200. ## Additional Information: > The information here is optional for you to provide, however it may help us to more readily diagnose any compatibility and bug issues. Sometimes the schematic isn´t placed completely if I use FAWE, mostly if I dont use //paste -a. Than I have to paste it again, before it works. ### Other plugins being used on the server: <!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list --> Plugins (19): NightVision, LuckPerms, WorldEdit, Spawntp, FastAsyncWorldEdit, SuperTrails, VoxelSniper, ClearLag, Vault, F3NPerm, Citizens, Essentials, ChatEx, CommandPanels, PlotSquared, Multiverse-Core, Multiverse-Inventories, Multiverse-SignPortals, ItemJoin ### Relevant console output, log lines, and/or screenshots: <!--- Please use in-line code insertion ``` https://pastebin.com/Q5NqC5Ey ``` for short (20 lines or less) text blobs, or a paste service for large blobs --> ### Additional relevant comments/remarks: <!--- Use this space to give us any additional information which may be relevant to this issue, such as: if you are using a Minecraft hosting provider; unusual installation environment; etc --> At the moment the Server is running on localhost # AFFIRMATION OF COMPLETION: <!-- Make sure you have completed the following steps (put an "X" between of brackets): --> - [X] I included all information required in the sections above - [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/PlotSquared/issues?utf8=%E2%9C%93&q=is%3Aissue) - [X] I made sure I am using an up-to-date version of PlotSquared - [X] I made sure the bug/error is not caused by any other plugin - [] I didn't read but checked everything above.
1.0
Schematic based World Generation is not working correctly - <!--- READ THIS BEFORE SUBMITTING AN ISSUE REPORT!!! --> <!--- ##### DO NOT REMOVE THIS TEMPLATE! YOUR ISSUE *WILL* FIT IN IT! ##### --> <!--- # NOTICE: ```diff ! PlotSquared for Minecraft Java Edition versions between 1.7 through to 1.12.2 are considered ! legacy, and will receive limited to no support. Please consider upgrading to 1.13+ for ! future support. Plugins exist for 1.13+ which bring back behaviors found in 1.8.8 ! All versions of PlotSquared for Sponge and Nukkit(X) will receive limited to no support ! due to lack of developer interest and time. Additionally, NukkitX has not had feature ! updates since the Better Together, which prevents some PlotSquared features from ever ! functioning. Contributions are always welcome however! ``` **Feature requests & Suggestions are to be submitted at the [PlotSquared Suggestions tracker](https://github.com/IntellectualSites/PlotSquaredSuggestions)** **Code contributions are to be done through [PRs](https://help.github.com/en/github/collaborating-with-issues-and-pull-requests/creating-a-pull-request), tagging the specific issue ticket(s) if applicable.** **[DISCORD INVITE LINK](https://discord.gg/cSMxtGn)** and please, for the love of the little sanity we have left, use the correct channels! --> # Bug Report Template: <!--- Incomplete reports will most likely be marked as invalid, and closed, with few exceptions.--> ## Required Information section: > ALL FIELDS IN THIS SECTION ARE REQUIRED, and must contain appropriate information ### Server config info (/plot debugpaste / file paste links): <!--- Issue /plot debugpaste in game or in your console and copy the supplied URL here --> <!--- If you cannot perform the above, we require logs/latest.log; settings.yml; worlds.yml and possibly PlotSquared.use_THIS.yml --> I atteached the files because the first thing doesnt worked: [latest.log](https://github.com/IntellectualSites/PlotSquared/files/4109359/latest.log) [2020-01-23-9.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109363/2020-01-23-9.log.gz) [2020-01-23-10.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109364/2020-01-23-10.log.gz) [yml.zip](https://github.com/IntellectualSites/PlotSquared/files/4109371/yml.zip) <!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. --> ### Server type: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): --> - [X] Spigot / Paper *(CraftBukkit should not be used, re-test with Spigot first!)* - [] Sponge *- NOTE: NOT ACTIVELY MAINTAINED* - [] NukkitX *- NOTE: NOT ACTIVELY MAINTAINED* ### Minecraft Version: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): The maintained versions are 1.14.4 and 1.15.x --> - [] Minecraft 1.15 - [X] Minecraft 1.14.4 - [] Minecraft 1.13.2 - [] Minecraft Java Edition *other versions, please specify*: - [] Minecraft Bedrock Edition *specify version*: ### Server build info: <!--- Run /version in-game or in console & paste the full output here: --> ``` This server is running CraftBukkit version git-Spigot-cbd1a1b-009d8af (MC: 1.14.4) (Implementing API version 1.14.4-R0.1-SNAPSHOT) [16:31:29] [Server thread/INFO]: This is a final build for 1.14.4. Please see https://www.spigotmc.org/go/1.14.4 for details about upgrading. ``` ### WorldEdit/FAWE versions: <!--- Specify which plugin you are using, and add its version --> - [X] FAWE *FastAsyncWorldEdit-1.15-45*: - [] WorldEdit *version*: ### Description of the problem: <!--- Be as specific as possible. Don't lie, redact information, or use false names/situations. --> <!--- Who, What, When, Where, Why, How, Expected behavior, Resultant behavior, etc --> If I want to use the schematic based world gen. with the plot.schem in the right folder for creating a new world, an Error occurs. After the Server has been loaded completely, the world is genarated and I can also enter it, but then there are always half or broken plots. ### How to replicate: <!--- If you can reproduce the issue please tell us as detailed as possible step by step how to do that --> I have done everything like it is described in the wiki, but than this Error occurs. The marked space for the plot schematic is about 200 Blocks in booth x and y direction. Maybe its too big? I also checked the worlds.yml but the Plot size is set to 200. ## Additional Information: > The information here is optional for you to provide, however it may help us to more readily diagnose any compatibility and bug issues. Sometimes the schematic isn´t placed completely if I use FAWE, mostly if I dont use //paste -a. Than I have to paste it again, before it works. ### Other plugins being used on the server: <!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list --> Plugins (19): NightVision, LuckPerms, WorldEdit, Spawntp, FastAsyncWorldEdit, SuperTrails, VoxelSniper, ClearLag, Vault, F3NPerm, Citizens, Essentials, ChatEx, CommandPanels, PlotSquared, Multiverse-Core, Multiverse-Inventories, Multiverse-SignPortals, ItemJoin ### Relevant console output, log lines, and/or screenshots: <!--- Please use in-line code insertion ``` https://pastebin.com/Q5NqC5Ey ``` for short (20 lines or less) text blobs, or a paste service for large blobs --> ### Additional relevant comments/remarks: <!--- Use this space to give us any additional information which may be relevant to this issue, such as: if you are using a Minecraft hosting provider; unusual installation environment; etc --> At the moment the Server is running on localhost # AFFIRMATION OF COMPLETION: <!-- Make sure you have completed the following steps (put an "X" between of brackets): --> - [X] I included all information required in the sections above - [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/PlotSquared/issues?utf8=%E2%9C%93&q=is%3Aissue) - [X] I made sure I am using an up-to-date version of PlotSquared - [X] I made sure the bug/error is not caused by any other plugin - [] I didn't read but checked everything above.
non_process
schematic based world generation is not working correctly notice diff plotsquared for minecraft java edition versions between through to are considered legacy and will receive limited to no support please consider upgrading to for future support plugins exist for which bring back behaviors found in all versions of plotsquared for sponge and nukkit x will receive limited to no support due to lack of developer interest and time additionally nukkitx has not had feature updates since the better together which prevents some plotsquared features from ever functioning contributions are always welcome however feature requests suggestions are to be submitted at the code contributions are to be done through tagging the specific issue ticket s if applicable and please for the love of the little sanity we have left use the correct channels bug report template required information section all fields in this section are required and must contain appropriate information server config info plot debugpaste file paste links i atteached the files because the first thing doesnt worked server type select one spigot paper craftbukkit should not be used re test with spigot first sponge note not actively maintained nukkitx note not actively maintained minecraft version select one select the type you are reporting the issue for put an x between of brackets the maintained versions are and x minecraft minecraft minecraft minecraft java edition other versions please specify minecraft bedrock edition specify version server build info this server is running craftbukkit version git spigot mc implementing api version snapshot this is a final build for please see for details about upgrading worldedit fawe versions fawe fastasyncworldedit worldedit version description of the problem if i want to use the schematic based world gen with the plot schem in the right folder for creating a new world an error occurs after the server has been loaded completely the world is genarated and i can also enter it but then there are always half or broken plots how to replicate i have done everything like it is described in the wiki but than this error occurs the marked space for the plot schematic is about blocks in booth x and y direction maybe its too big i also checked the worlds yml but the plot size is set to additional information the information here is optional for you to provide however it may help us to more readily diagnose any compatibility and bug issues sometimes the schematic isn´t placed completely if i use fawe mostly if i dont use paste a than i have to paste it again before it works other plugins being used on the server plugins nightvision luckperms worldedit spawntp fastasyncworldedit supertrails voxelsniper clearlag vault citizens essentials chatex commandpanels plotsquared multiverse core multiverse inventories multiverse signportals itemjoin relevant console output log lines and or screenshots please use in line code insertion for short lines or less text blobs or a paste service for large blobs additional relevant comments remarks at the moment the server is running on localhost affirmation of completion i included all information required in the sections above i made sure there are no duplicates of this report i made sure i am using an up to date version of plotsquared i made sure the bug error is not caused by any other plugin i didn t read but checked everything above
0
1,073
3,537,929,919
IssuesEvent
2016-01-18 06:31:04
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Specializing the link element does not work
bug preprocess
The content is copied from [Stackoverflow - How can I specialize the link element (related-links)](http://stackoverflow.com/questions/34766407/how-can-i-specialize-the-link-element-related-links), because Eliot says, this is a bug. - - - - - I'd like to create new elements, based on the link element. This would be more comfortable for the author than using the role and otherrole attributes to specify the link role. Unfortunately the DITA-OT 2.2.1 fails with this message: Required item type of result of template related-links:link. is element(Q{}link); supplied value has item type element(Q{}myelement) Is this a valid approach, or should I stick to otherrole? Update 1: The error occurs in the related-links.xsl: ``` <!-- Ungrouped links have the default-mode template applied to them. (Can be overridden.) --> <xsl:template match="*[contains(@class, ' topic/link ')]" mode="related-links:link" name="related-links:link." as="element(link)"> <xsl:sequence select="."/> <!--- error points to this line --> </xsl:template> ``` This line causes the error [beetle.dita#L23](https://github.com/doctales/org.doctales.terminology/blob/0b7384483956473a7d019091baf7d60f331547ac/samples/beetle.dita#L23). Both variants did not work: ``` <!ATTLIST hyponym %global-atts; class CDATA "- topic/link "> ``` ``` <!ATTLIST hyponym %global-atts; class CDATA "- topic/link termentry/hypernym"> ```
1.0
Specializing the link element does not work - The content is copied from [Stackoverflow - How can I specialize the link element (related-links)](http://stackoverflow.com/questions/34766407/how-can-i-specialize-the-link-element-related-links), because Eliot says, this is a bug. - - - - - I'd like to create new elements, based on the link element. This would be more comfortable for the author than using the role and otherrole attributes to specify the link role. Unfortunately the DITA-OT 2.2.1 fails with this message: Required item type of result of template related-links:link. is element(Q{}link); supplied value has item type element(Q{}myelement) Is this a valid approach, or should I stick to otherrole? Update 1: The error occurs in the related-links.xsl: ``` <!-- Ungrouped links have the default-mode template applied to them. (Can be overridden.) --> <xsl:template match="*[contains(@class, ' topic/link ')]" mode="related-links:link" name="related-links:link." as="element(link)"> <xsl:sequence select="."/> <!--- error points to this line --> </xsl:template> ``` This line causes the error [beetle.dita#L23](https://github.com/doctales/org.doctales.terminology/blob/0b7384483956473a7d019091baf7d60f331547ac/samples/beetle.dita#L23). Both variants did not work: ``` <!ATTLIST hyponym %global-atts; class CDATA "- topic/link "> ``` ``` <!ATTLIST hyponym %global-atts; class CDATA "- topic/link termentry/hypernym"> ```
process
specializing the link element does not work the content is copied from because eliot says this is a bug i d like to create new elements based on the link element this would be more comfortable for the author than using the role and otherrole attributes to specify the link role unfortunately the dita ot fails with this message required item type of result of template related links link is element q link supplied value has item type element q myelement is this a valid approach or should i stick to otherrole update the error occurs in the related links xsl ungrouped links have the default mode template applied to them can be overridden xsl template match mode related links link name related links link as element link this line causes the error both variants did not work
1
11,255
14,021,192,838
IssuesEvent
2020-10-29 20:52:45
googleapis/python-recommender
https://api.github.com/repos/googleapis/python-recommender
closed
Immediate timeout on ListInsights calls
api: recommender type: process
#### Environment details - OS type and version: macos 10.15.6 - Python version: 3.6.8 - pip version: 20.1 - `google-cloud-recommender` version: 1.1.0 #### Steps to reproduce 1. Call list_insights without a timeout param 2. See the following error: google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded #### Code example ```python from google.oauth2 import service_account from google.cloud import recommender_v1 serviceaccount=service_account.Credentials.from_service_account_file(".....json") client=recommender_v1.RecommenderClient(credentials=serviceaccount) parent_1 = client.insight_type_path('${PROJECT_ID}','global','google.iam.serviceAccount.Insight') for element in client.list_insights(parent_1): print(element) ``` #### Stack trace ``` Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 57, in error_remapped_callable return callable_(*args, **kwargs) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/grpc/_channel.py", line 826, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/grpc/_channel.py", line 729, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1603293336.314798000","description":"Deadline Exceeded","file":"src/core/ext/filters/deadline/deadline_filter.cc","file_line":69,"grpc_status":4}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/retry.py", line 184, in retry_target return target() File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout return func(*args, **kwargs) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 59, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/Users/nzar/.vscode/extensions/ms-python.python-2020.9.114305/pythonFiles/lib/python/debugpy/__main__.py", line 45, in <module> cli.main() File "/Users/nzar/.vscode/extensions/ms-python.python-2020.9.114305/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 430, in main run() File "/Users/nzar/.vscode/extensions/ms-python.python-2020.9.114305/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 267, in run_file runpy.run_path(options.target, run_name=compat.force_str("__main__")) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 263, in run_path pkg_name=pkg_name, script_name=fname) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 96, in _run_module_code mod_name, mod_spec, pkg_name, script_name) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/Users/nzar/Fun/repro/list-insights/main.py", line 6, in <module> for element in client.list_insights(parent_1): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/page_iterator.py", line 212, in _items_iter for page in self._page_iter(increment=False): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/page_iterator.py", line 243, in _page_iter page = self._next_page() File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/page_iterator.py", line 534, in _next_page response = self._method(self._request) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/retry.py", line 286, in retry_wrapped_func on_error=on_error, File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/retry.py", line 206, in retry_target last_exc, File "<string>", line 3, in raise_from google.api_core.exceptions.RetryError: Deadline of 0.0s exceeded while calling functools.partial(<function _wrap_unary_errors.<locals>.error_remapped_callable at 0x7ff8d0bab0d0>, parent: "projects/${PROJECT_ID}/locations/global/insightTypes/google.iam.serviceAccount.Insight" , metadata=[('x-goog-request-params', 'parent=projects/${PROJECT_ID}/locations/global/insightTypes/google.iam.serviceAccount.Insight'), ('x-goog-api-client', 'gl-python/3.6.8 grpc/1.32.0 gax/1.23.0 gapic/1.1.0')]), last exception: 504 Deadline Exceeded ``` Clearly at the end of the stacktrace "Deadline of 0.0s exceeded" seems to clearly point to an issue. And adding a timeout param to the `list_insights` call fixes the problem.
1.0
Immediate timeout on ListInsights calls - #### Environment details - OS type and version: macos 10.15.6 - Python version: 3.6.8 - pip version: 20.1 - `google-cloud-recommender` version: 1.1.0 #### Steps to reproduce 1. Call list_insights without a timeout param 2. See the following error: google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded #### Code example ```python from google.oauth2 import service_account from google.cloud import recommender_v1 serviceaccount=service_account.Credentials.from_service_account_file(".....json") client=recommender_v1.RecommenderClient(credentials=serviceaccount) parent_1 = client.insight_type_path('${PROJECT_ID}','global','google.iam.serviceAccount.Insight') for element in client.list_insights(parent_1): print(element) ``` #### Stack trace ``` Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 57, in error_remapped_callable return callable_(*args, **kwargs) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/grpc/_channel.py", line 826, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/grpc/_channel.py", line 729, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1603293336.314798000","description":"Deadline Exceeded","file":"src/core/ext/filters/deadline/deadline_filter.cc","file_line":69,"grpc_status":4}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/retry.py", line 184, in retry_target return target() File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout return func(*args, **kwargs) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 59, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/Users/nzar/.vscode/extensions/ms-python.python-2020.9.114305/pythonFiles/lib/python/debugpy/__main__.py", line 45, in <module> cli.main() File "/Users/nzar/.vscode/extensions/ms-python.python-2020.9.114305/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 430, in main run() File "/Users/nzar/.vscode/extensions/ms-python.python-2020.9.114305/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 267, in run_file runpy.run_path(options.target, run_name=compat.force_str("__main__")) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 263, in run_path pkg_name=pkg_name, script_name=fname) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 96, in _run_module_code mod_name, mod_spec, pkg_name, script_name) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/Users/nzar/Fun/repro/list-insights/main.py", line 6, in <module> for element in client.list_insights(parent_1): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/page_iterator.py", line 212, in _items_iter for page in self._page_iter(increment=False): File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/page_iterator.py", line 243, in _page_iter page = self._next_page() File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/page_iterator.py", line 534, in _next_page response = self._method(self._request) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/retry.py", line 286, in retry_wrapped_func on_error=on_error, File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/google/api_core/retry.py", line 206, in retry_target last_exc, File "<string>", line 3, in raise_from google.api_core.exceptions.RetryError: Deadline of 0.0s exceeded while calling functools.partial(<function _wrap_unary_errors.<locals>.error_remapped_callable at 0x7ff8d0bab0d0>, parent: "projects/${PROJECT_ID}/locations/global/insightTypes/google.iam.serviceAccount.Insight" , metadata=[('x-goog-request-params', 'parent=projects/${PROJECT_ID}/locations/global/insightTypes/google.iam.serviceAccount.Insight'), ('x-goog-api-client', 'gl-python/3.6.8 grpc/1.32.0 gax/1.23.0 gapic/1.1.0')]), last exception: 504 Deadline Exceeded ``` Clearly at the end of the stacktrace "Deadline of 0.0s exceeded" seems to clearly point to an issue. And adding a timeout param to the `list_insights` call fixes the problem.
process
immediate timeout on listinsights calls environment details os type and version macos python version pip version google cloud recommender version steps to reproduce call list insights without a timeout param see the following error google api core exceptions deadlineexceeded deadline exceeded code example python from google import service account from google cloud import recommender serviceaccount service account credentials from service account file json client recommender recommenderclient credentials serviceaccount parent client insight type path project id global google iam serviceaccount insight for element in client list insights parent print element stack trace traceback most recent call last file library frameworks python framework versions lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file library frameworks python framework versions lib site packages grpc channel py line in call return end unary response blocking state call false none file library frameworks python framework versions lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description deadline exceeded file src core ext filters deadline deadline filter cc file line grpc status the above exception was the direct cause of the following exception traceback most recent call last file library frameworks python framework versions lib site packages google api core retry py line in retry target return target file library frameworks python framework versions lib site packages google api core timeout py line in func with timeout return func args kwargs file library frameworks python framework versions lib site packages google api core grpc helpers py line in error remapped callable six raise from exceptions from grpc error exc exc file line in raise from google api core exceptions deadlineexceeded deadline exceeded the above exception was the direct cause of the following exception traceback most recent call last file library frameworks python framework versions lib runpy py line in run module as main main mod spec file library frameworks python framework versions lib runpy py line in run code exec code run globals file users nzar vscode extensions ms python python pythonfiles lib python debugpy main py line in cli main file users nzar vscode extensions ms python python pythonfiles lib python debugpy debugpy server cli py line in main run file users nzar vscode extensions ms python python pythonfiles lib python debugpy debugpy server cli py line in run file runpy run path options target run name compat force str main file library frameworks python framework versions lib runpy py line in run path pkg name pkg name script name fname file library frameworks python framework versions lib runpy py line in run module code mod name mod spec pkg name script name file library frameworks python framework versions lib runpy py line in run code exec code run globals file users nzar fun repro list insights main py line in for element in client list insights parent file library frameworks python framework versions lib site packages google api core page iterator py line in items iter for page in self page iter increment false file library frameworks python framework versions lib site packages google api core page iterator py line in page iter page self next page file library frameworks python framework versions lib site packages google api core page iterator py line in next page response self method self request file library frameworks python framework versions lib site packages google api core gapic method py line in call return wrapped func args kwargs file library frameworks python framework versions lib site packages google api core retry py line in retry wrapped func on error on error file library frameworks python framework versions lib site packages google api core retry py line in retry target last exc file line in raise from google api core exceptions retryerror deadline of exceeded while calling functools partial error remapped callable at parent projects project id locations global insighttypes google iam serviceaccount insight metadata last exception deadline exceeded clearly at the end of the stacktrace deadline of exceeded seems to clearly point to an issue and adding a timeout param to the list insights call fixes the problem
1
10,260
7,969,329,705
IssuesEvent
2018-07-16 08:41:36
status-im/status-react
https://api.github.com/repos/status-im/status-react
opened
Encrypt account realm with user password
chat security
### User Story As a user I want to have each user account encrypted with a different key based on the user password So that an attacker breaking the keychain will not be able to decrypt my data ### Description At the moment we encrypt the database on disk using only the keychain. User data can be also encrypted with both the keychain and the user password (generating a key from both), so that compromising the keychain does not result in the compromise of all the messages. #### Expected behavior Realm account data is encrypted with their password as well. #### Actual behavior Realm account data is encrypted only with the keychain password. ### Solution * Realm files should be correctly migrated to the new encryption key * Additionally sensitive information should be moved from `base` to `account`, but can be done in a separate issue * Nice to have: It would be good to take into consideration the fact that a user will be able to change password at some point, so encrypting the database with a new key might be a day-to-day operation.
True
Encrypt account realm with user password - ### User Story As a user I want to have each user account encrypted with a different key based on the user password So that an attacker breaking the keychain will not be able to decrypt my data ### Description At the moment we encrypt the database on disk using only the keychain. User data can be also encrypted with both the keychain and the user password (generating a key from both), so that compromising the keychain does not result in the compromise of all the messages. #### Expected behavior Realm account data is encrypted with their password as well. #### Actual behavior Realm account data is encrypted only with the keychain password. ### Solution * Realm files should be correctly migrated to the new encryption key * Additionally sensitive information should be moved from `base` to `account`, but can be done in a separate issue * Nice to have: It would be good to take into consideration the fact that a user will be able to change password at some point, so encrypting the database with a new key might be a day-to-day operation.
non_process
encrypt account realm with user password user story as a user i want to have each user account encrypted with a different key based on the user password so that an attacker breaking the keychain will not be able to decrypt my data description at the moment we encrypt the database on disk using only the keychain user data can be also encrypted with both the keychain and the user password generating a key from both so that compromising the keychain does not result in the compromise of all the messages expected behavior realm account data is encrypted with their password as well actual behavior realm account data is encrypted only with the keychain password solution realm files should be correctly migrated to the new encryption key additionally sensitive information should be moved from base to account but can be done in a separate issue nice to have it would be good to take into consideration the fact that a user will be able to change password at some point so encrypting the database with a new key might be a day to day operation
0
326,052
27,975,279,268
IssuesEvent
2023-03-25 14:03:01
dudykr/stc
https://api.github.com/repos/dudykr/stc
closed
Fix unit test for tests/pass-only/typeNarrowing/.do-while-1.ts
tsc-unit-test
--- STC: tests/pass-only/typeNarrowing/.do-while-1.ts --- This issue is created by sync script.
1.0
Fix unit test for tests/pass-only/typeNarrowing/.do-while-1.ts - --- STC: tests/pass-only/typeNarrowing/.do-while-1.ts --- This issue is created by sync script.
non_process
fix unit test for tests pass only typenarrowing do while ts stc tests pass only typenarrowing do while ts this issue is created by sync script
0
11,720
14,547,992,943
IssuesEvent
2020-12-16 00:11:07
pacificclimate/quail
https://api.github.com/repos/pacificclimate/quail
closed
TN90p
process
## Description This function takes a climdexInput object as input and computes the monthly or annual percent of values above the 90th percentile of baseline daily minimum temperature. ## Function to wrap [`climdex.tn90p`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L981)
1.0
TN90p - ## Description This function takes a climdexInput object as input and computes the monthly or annual percent of values above the 90th percentile of baseline daily minimum temperature. ## Function to wrap [`climdex.tn90p`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L981)
process
description this function takes a climdexinput object as input and computes the monthly or annual percent of values above the percentile of baseline daily minimum temperature function to wrap
1
12,134
14,740,958,883
IssuesEvent
2021-01-07 09:53:11
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Cron Job - 60+ error emails sent 12-15-18
anc-process anp-0.5 ant-bug has attachment
In GitLab by @kdjstudios on Dec 17, 2018, 08:04 **Submitted by:** Kyle **Helpdesk:** NA **Server:** towne **Client/Site:** NA **Account:** NA **Issue:** Hello Team, When opening my email this morning I found over 60 error emails had sent from over the weekend. I noticed a majority of them were sent from 10pm to 11pm MT and were similar to this one below. There was a few other types of emails too. Please check what happened and confirm it has been resolved. [Cron_root_towne_homesabillingmongodb_backupsdel_old_backup.sh.msg](/uploads/f030db2cc02208b0f2df7f6610462f50/Cron_root_towne_homesabillingmongodb_backupsdel_old_backup.sh.msg)
1.0
Cron Job - 60+ error emails sent 12-15-18 - In GitLab by @kdjstudios on Dec 17, 2018, 08:04 **Submitted by:** Kyle **Helpdesk:** NA **Server:** towne **Client/Site:** NA **Account:** NA **Issue:** Hello Team, When opening my email this morning I found over 60 error emails had sent from over the weekend. I noticed a majority of them were sent from 10pm to 11pm MT and were similar to this one below. There was a few other types of emails too. Please check what happened and confirm it has been resolved. [Cron_root_towne_homesabillingmongodb_backupsdel_old_backup.sh.msg](/uploads/f030db2cc02208b0f2df7f6610462f50/Cron_root_towne_homesabillingmongodb_backupsdel_old_backup.sh.msg)
process
cron job error emails sent in gitlab by kdjstudios on dec submitted by kyle helpdesk na server towne client site na account na issue hello team when opening my email this morning i found over error emails had sent from over the weekend i noticed a majority of them were sent from to mt and were similar to this one below there was a few other types of emails too please check what happened and confirm it has been resolved uploads cron root towne homesabillingmongodb backupsdel old backup sh msg
1
104,880
13,131,778,549
IssuesEvent
2020-08-06 17:38:58
mhmdtshref/skyfall-frontend
https://api.github.com/repos/mhmdtshref/skyfall-frontend
closed
Create full design
Design
Depending on the wire frames created [HERE](https://mohamedsharif802084.invisionapp.com/freehand/Spyfall-7WxzYLkaL), we need to create a real-app pages designs.
1.0
Create full design - Depending on the wire frames created [HERE](https://mohamedsharif802084.invisionapp.com/freehand/Spyfall-7WxzYLkaL), we need to create a real-app pages designs.
non_process
create full design depending on the wire frames created we need to create a real app pages designs
0
15,520
19,703,268,538
IssuesEvent
2022-01-12 18:52:26
googleapis/java-iam-admin
https://api.github.com/repos/googleapis/java-iam-admin
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'iam-admin' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'iam-admin' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname iam admin invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
16,414
21,191,768,081
IssuesEvent
2022-04-08 18:17:18
googleapis/google-auth-library-python
https://api.github.com/repos/googleapis/google-auth-library-python
closed
1.11.1 released a backwards incompatible breaking change in a patch version
type: process
- [x] Search the issues already opened: https://github.com/googleapis/google-auth-library-python/issues The bug is reported in #443. I'm opening this ticket to specifically request a discussion about how to avoid breaking changes in a patch version. The PR that introduced the change is #365. The release of that change as 1.11.1 was PR #437 . It looks like there needs to be a test that a valid token can be decoded?
1.0
1.11.1 released a backwards incompatible breaking change in a patch version - - [x] Search the issues already opened: https://github.com/googleapis/google-auth-library-python/issues The bug is reported in #443. I'm opening this ticket to specifically request a discussion about how to avoid breaking changes in a patch version. The PR that introduced the change is #365. The release of that change as 1.11.1 was PR #437 . It looks like there needs to be a test that a valid token can be decoded?
process
released a backwards incompatible breaking change in a patch version search the issues already opened the bug is reported in i m opening this ticket to specifically request a discussion about how to avoid breaking changes in a patch version the pr that introduced the change is the release of that change as was pr it looks like there needs to be a test that a valid token can be decoded
1
7,049
16,342,515,565
IssuesEvent
2021-05-13 00:29:14
eclipse/omr
https://api.github.com/repos/eclipse/omr
opened
Prune vector IL opcodes
architecture review pending comp:compiler
In preparation for re-architecting the vector IL opcodes to support wider vector lengths and element types, I believe some cleanup is in order to help focus our design choices on only a core set of opcodes we really need. At present, there are (at least) 109 vector IL opcodes. The vast majority of which are not used, documented, or tested, and have an incomplete implementation. Based on my investigation, we should keep the following as a base set of opcodes that we can think about expanding to wider vector lengths. There are uses for these opcodes in OpenJ9. Some of these even have Tril tests written for them. As we get more into the vector redesign more may be proposed to be pruned from this list. | IL Opcode | | :---------- | | vadd | | vsub | | vmul | | vdiv | | vneg | | vand | | vor | | vxor | | vnot | | vsplats | | v2v | | vl2vd | | vconst | | vload | | vloadi | | vstore | | vstorei | | vcmpeq | | vcmpne | | vcmplt | | vcmpgt | | vcmple | | vcmpge | | getvelem | | vsetelem | | v[id]setelem | | v[id]getelem | | v[bsilfd]RegLoad | | v[bsilfd]RegStore | The following IL opcodes may have some use when the vector opcodes are expanded to include different element sizes and may be worth keeping: | IL Opcode | Implemented Platforms | | :---------- | :-------------------: | | vicmpeq | P | | vicmpgt | P | | vicmpge | P | | vicmplt | P | | vicmple | P | | vdcmpeq | P | | vdcmpne | P | | vdcmpgt | P | | vdcmpge | P | | vdcmplt | P | | vdcmple | P | | vdmax | P | | vdmin | P | | vimin | P | | vimax | P | I recommend removing the following IL opcodes. As none of the list below has any known usage in OMR nor any known downstream projects I feel the justification to pardon any of these has to be strong enough to overcome eliminating the IL opcode now and re-introducing it later should the need arise. The advantage of introducing it again later is that it should come with proper justification, documentation, and Tril tests which are sorely lacking for all of these opcodes. Some of them could be reintroduced as intrinsics rather than opcodes. | IL Opcode | Implemented Platforms | | :---------- | :-------------------: | | virem | P | | vimergel | P | | vimergeh | P | | vicmpalleq | P | | vicmpallne | P | | vicmpallgt | P | | vicmpallge | P | | vicmpalllt | P | | vicmpallle | P | | vicmpanyeq | P | | vicmpanyne | P | | vicmpanygt | P | | vicmpanyge | P | | vicmpanylt | P | | vicmpanyle | P | | vdcmpalleq | P | | vdcmpallne | P | | vdcmpallgt | P | | vdcmpallge | P | | vdcmpalllt | P | | vdcmpallle | P | | vdcmpanyeq | P | | vdcmpanyne | P | | vdcmpanygt | P | | vdcmpanyge | P | | vdcmpanylt | P | | vdcmpanyle | P | | vucmplt | | | vucmpgt | | | vucmple | | | vucmpge | | | vselect | | | vinc | Z | | vdec | Z | | vperm | P | | vcom | Z | | vbitselect | P,Z | | vdmergel | P | | vdmergeh | P | | vdmsub | P | | vdmadd | P,Z | | vdnmsub | P | | vrem | Z | | vshl | Z | | vushr | Z | | vshr | Z | | vdsel | P | | vdrem | P | | vdsqrt | P | | vrand | Z | | vreturn | Z | | vcall | Z | | vcalli | Z |
1.0
Prune vector IL opcodes - In preparation for re-architecting the vector IL opcodes to support wider vector lengths and element types, I believe some cleanup is in order to help focus our design choices on only a core set of opcodes we really need. At present, there are (at least) 109 vector IL opcodes. The vast majority of which are not used, documented, or tested, and have an incomplete implementation. Based on my investigation, we should keep the following as a base set of opcodes that we can think about expanding to wider vector lengths. There are uses for these opcodes in OpenJ9. Some of these even have Tril tests written for them. As we get more into the vector redesign more may be proposed to be pruned from this list. | IL Opcode | | :---------- | | vadd | | vsub | | vmul | | vdiv | | vneg | | vand | | vor | | vxor | | vnot | | vsplats | | v2v | | vl2vd | | vconst | | vload | | vloadi | | vstore | | vstorei | | vcmpeq | | vcmpne | | vcmplt | | vcmpgt | | vcmple | | vcmpge | | getvelem | | vsetelem | | v[id]setelem | | v[id]getelem | | v[bsilfd]RegLoad | | v[bsilfd]RegStore | The following IL opcodes may have some use when the vector opcodes are expanded to include different element sizes and may be worth keeping: | IL Opcode | Implemented Platforms | | :---------- | :-------------------: | | vicmpeq | P | | vicmpgt | P | | vicmpge | P | | vicmplt | P | | vicmple | P | | vdcmpeq | P | | vdcmpne | P | | vdcmpgt | P | | vdcmpge | P | | vdcmplt | P | | vdcmple | P | | vdmax | P | | vdmin | P | | vimin | P | | vimax | P | I recommend removing the following IL opcodes. As none of the list below has any known usage in OMR nor any known downstream projects I feel the justification to pardon any of these has to be strong enough to overcome eliminating the IL opcode now and re-introducing it later should the need arise. The advantage of introducing it again later is that it should come with proper justification, documentation, and Tril tests which are sorely lacking for all of these opcodes. Some of them could be reintroduced as intrinsics rather than opcodes. | IL Opcode | Implemented Platforms | | :---------- | :-------------------: | | virem | P | | vimergel | P | | vimergeh | P | | vicmpalleq | P | | vicmpallne | P | | vicmpallgt | P | | vicmpallge | P | | vicmpalllt | P | | vicmpallle | P | | vicmpanyeq | P | | vicmpanyne | P | | vicmpanygt | P | | vicmpanyge | P | | vicmpanylt | P | | vicmpanyle | P | | vdcmpalleq | P | | vdcmpallne | P | | vdcmpallgt | P | | vdcmpallge | P | | vdcmpalllt | P | | vdcmpallle | P | | vdcmpanyeq | P | | vdcmpanyne | P | | vdcmpanygt | P | | vdcmpanyge | P | | vdcmpanylt | P | | vdcmpanyle | P | | vucmplt | | | vucmpgt | | | vucmple | | | vucmpge | | | vselect | | | vinc | Z | | vdec | Z | | vperm | P | | vcom | Z | | vbitselect | P,Z | | vdmergel | P | | vdmergeh | P | | vdmsub | P | | vdmadd | P,Z | | vdnmsub | P | | vrem | Z | | vshl | Z | | vushr | Z | | vshr | Z | | vdsel | P | | vdrem | P | | vdsqrt | P | | vrand | Z | | vreturn | Z | | vcall | Z | | vcalli | Z |
non_process
prune vector il opcodes in preparation for re architecting the vector il opcodes to support wider vector lengths and element types i believe some cleanup is in order to help focus our design choices on only a core set of opcodes we really need at present there are at least vector il opcodes the vast majority of which are not used documented or tested and have an incomplete implementation based on my investigation we should keep the following as a base set of opcodes that we can think about expanding to wider vector lengths there are uses for these opcodes in some of these even have tril tests written for them as we get more into the vector redesign more may be proposed to be pruned from this list il opcode vadd vsub vmul vdiv vneg vand vor vxor vnot vsplats vconst vload vloadi vstore vstorei vcmpeq vcmpne vcmplt vcmpgt vcmple vcmpge getvelem vsetelem v setelem v getelem v regload v regstore the following il opcodes may have some use when the vector opcodes are expanded to include different element sizes and may be worth keeping il opcode implemented platforms vicmpeq p vicmpgt p vicmpge p vicmplt p vicmple p vdcmpeq p vdcmpne p vdcmpgt p vdcmpge p vdcmplt p vdcmple p vdmax p vdmin p vimin p vimax p i recommend removing the following il opcodes as none of the list below has any known usage in omr nor any known downstream projects i feel the justification to pardon any of these has to be strong enough to overcome eliminating the il opcode now and re introducing it later should the need arise the advantage of introducing it again later is that it should come with proper justification documentation and tril tests which are sorely lacking for all of these opcodes some of them could be reintroduced as intrinsics rather than opcodes il opcode implemented platforms virem p vimergel p vimergeh p vicmpalleq p vicmpallne p vicmpallgt p vicmpallge p vicmpalllt p vicmpallle p vicmpanyeq p vicmpanyne p vicmpanygt p vicmpanyge p vicmpanylt p vicmpanyle p vdcmpalleq p vdcmpallne p vdcmpallgt p vdcmpallge p vdcmpalllt p vdcmpallle p vdcmpanyeq p vdcmpanyne p vdcmpanygt p vdcmpanyge p vdcmpanylt p vdcmpanyle p vucmplt vucmpgt vucmple vucmpge vselect vinc z vdec z vperm p vcom z vbitselect p z vdmergel p vdmergeh p vdmsub p vdmadd p z vdnmsub p vrem z vshl z vushr z vshr z vdsel p vdrem p vdsqrt p vrand z vreturn z vcall z vcalli z
0
110,131
16,976,952,474
IssuesEvent
2021-06-30 01:22:14
LevyForch/SilverKing
https://api.github.com/repos/LevyForch/SilverKing
opened
CVE-2019-16942 (High) detected in jackson-databind-2.6.7.1.jar
security vulnerability
## CVE-2019-16942 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.7.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: SilverKing/lib/aws-java-sdk-1.11.333/third-party/lib/jackson-databind-2.6.7.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.7.1.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942>CVE-2019-16942</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-16942 (High) detected in jackson-databind-2.6.7.1.jar - ## CVE-2019-16942 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.7.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: SilverKing/lib/aws-java-sdk-1.11.333/third-party/lib/jackson-databind-2.6.7.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.7.1.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942>CVE-2019-16942</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library silverking lib aws java sdk third party lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the commons dbcp jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of org apache commons dbcp datasources sharedpooldatasource and org apache commons dbcp datasources peruserpooldatasource mishandling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
16,260
20,831,024,246
IssuesEvent
2022-03-19 12:46:09
acdh-oeaw/abcd-db
https://api.github.com/repos/acdh-oeaw/abcd-db
closed
run nlp to extract person/places
Data Processing
as we are dealing with standardized text it shouldn't be to hard to extract Person/Places, create matching entities and link them to the records; only problem might be disambiguiation (two entities same name); but I think this can than be dealt with manually; and it would be some fun task!
1.0
run nlp to extract person/places - as we are dealing with standardized text it shouldn't be to hard to extract Person/Places, create matching entities and link them to the records; only problem might be disambiguiation (two entities same name); but I think this can than be dealt with manually; and it would be some fun task!
process
run nlp to extract person places as we are dealing with standardized text it shouldn t be to hard to extract person places create matching entities and link them to the records only problem might be disambiguiation two entities same name but i think this can than be dealt with manually and it would be some fun task
1
1,599
4,212,413,269
IssuesEvent
2016-06-29 16:12:03
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
overwrite configuration in vsphere post-processor broken
bug post-processor/vsphere
Packer Version: 0.10.1 Ovftool Version: 4.1.0 (build-3018522) Host Platform: Windows 7 Enterprise "post-processors": [ { "type": "vsphere", "cluster": "{{ user `vsphere_cluster` }}", "overwrite": "true", "datacenter": "{{ user `vsphere_datacenter` }}", "datastore": "{{ user `vsphere_datastore` }}", "host": "{{ user `vsphere_host` }}", "username": "{{ user `vsphere_user` }}", "password": "{{ user `vsphere_pass` }}", "vm_network": "{{ user `vsphere_vm_network` }}", "vm_folder": "{{ user `vsphere_vm_folder` }}", "vm_name": "{{ user `vsphere_vm_name` }}", "disk_mode": "thin" } ], When the overwrite configuration is used, the flag is placed in the wrong location in the underlying ovftool command. It is being placed after the target location instead of before the source location It is producing the following error: https://gist.github.com/sunckell/5ad3d9f30ffd99af75e54e97a6b4d7c6
1.0
overwrite configuration in vsphere post-processor broken - Packer Version: 0.10.1 Ovftool Version: 4.1.0 (build-3018522) Host Platform: Windows 7 Enterprise "post-processors": [ { "type": "vsphere", "cluster": "{{ user `vsphere_cluster` }}", "overwrite": "true", "datacenter": "{{ user `vsphere_datacenter` }}", "datastore": "{{ user `vsphere_datastore` }}", "host": "{{ user `vsphere_host` }}", "username": "{{ user `vsphere_user` }}", "password": "{{ user `vsphere_pass` }}", "vm_network": "{{ user `vsphere_vm_network` }}", "vm_folder": "{{ user `vsphere_vm_folder` }}", "vm_name": "{{ user `vsphere_vm_name` }}", "disk_mode": "thin" } ], When the overwrite configuration is used, the flag is placed in the wrong location in the underlying ovftool command. It is being placed after the target location instead of before the source location It is producing the following error: https://gist.github.com/sunckell/5ad3d9f30ffd99af75e54e97a6b4d7c6
process
overwrite configuration in vsphere post processor broken packer version ovftool version build host platform windows enterprise post processors type vsphere cluster user vsphere cluster overwrite true datacenter user vsphere datacenter datastore user vsphere datastore host user vsphere host username user vsphere user password user vsphere pass vm network user vsphere vm network vm folder user vsphere vm folder vm name user vsphere vm name disk mode thin when the overwrite configuration is used the flag is placed in the wrong location in the underlying ovftool command it is being placed after the target location instead of before the source location it is producing the following error
1
15,990
20,188,203,128
IssuesEvent
2022-02-11 01:17:42
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Automatically remove/obfuscate personally identifiable information (PII) for this workload
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Health Modeling & Monitoring Application Level Monitoring
<a href="https://docs.microsoft.com/azure/search/cognitive-search-skill-pii-detection">Automatically remove/obfuscate personally identifiable information (PII) for this workload</a> <p><b>Why Consider This?</b></p> Extra care should be taken around logging of sensitive application areas. PII (contact information, payment information etc.) should not be stored in any application logs and protective measures should be applied (such as obfuscation). <p><b>Context</b></p> <p><b>Suggested Actions</b></p> <p><span>Deploy PII detection and removal/obfuscation solution for this workload.</span></p> <p><b>Learn More</b></p> <p><span>Machine learning tools like </span><a href="https://docs.microsoft.com/en-us/azure/search/cognitive-search-skill-pii-detection" target="_blank"><span>Cognitive Search PII detection</span></a><span> can help with this.</span></p>
1.0
Automatically remove/obfuscate personally identifiable information (PII) for this workload - <a href="https://docs.microsoft.com/azure/search/cognitive-search-skill-pii-detection">Automatically remove/obfuscate personally identifiable information (PII) for this workload</a> <p><b>Why Consider This?</b></p> Extra care should be taken around logging of sensitive application areas. PII (contact information, payment information etc.) should not be stored in any application logs and protective measures should be applied (such as obfuscation). <p><b>Context</b></p> <p><b>Suggested Actions</b></p> <p><span>Deploy PII detection and removal/obfuscation solution for this workload.</span></p> <p><b>Learn More</b></p> <p><span>Machine learning tools like </span><a href="https://docs.microsoft.com/en-us/azure/search/cognitive-search-skill-pii-detection" target="_blank"><span>Cognitive Search PII detection</span></a><span> can help with this.</span></p>
process
automatically remove obfuscate personally identifiable information pii for this workload why consider this extra care should be taken around logging of sensitive application areas pii contact information payment information etc should not be stored in any application logs and protective measures should be applied such as obfuscation context suggested actions deploy pii detection and removal obfuscation solution for this workload learn more machine learning tools like cognitive search pii detection can help with this
1
12,316
14,879,222,473
IssuesEvent
2021-01-20 07:17:36
KratosMultiphysics/Kratos
https://api.github.com/repos/KratosMultiphysics/Kratos
closed
[EmbeddedSkinVisualizationProcess] output of visualization_model_part MPI
Parallel-MPI Post Process
**Description** When running a simulation with MPI and using an GiD or VTK output process for the skin visualization model part, only Rank 0 is written and therefore only a section of the mesh. **Scope** - Embedded Navier Stokes Discontinuous - Output of EmbeddedSkinVisualizationProcess **To Reproduce** - Use `embedded_symbolic_navier_stokes_discontinuous` as element_type - Add `ApplyEmbeddedSkinVisualizationProcess` in your ProjectParameters to have a visualization_model_part,( e.g. "visualization_variables": ["VELOCITY","PRESSURE"]) - Add a `GiDOutputProcess` for the visualization model part - Run a couple of time steps (more than output_interval of GiDOutputProcess) **Expected behavior** The whole mesh of the visualization model part should be written by the output process. **Environment** - OS: Linux - Branch: master - Python 3.6
1.0
[EmbeddedSkinVisualizationProcess] output of visualization_model_part MPI - **Description** When running a simulation with MPI and using an GiD or VTK output process for the skin visualization model part, only Rank 0 is written and therefore only a section of the mesh. **Scope** - Embedded Navier Stokes Discontinuous - Output of EmbeddedSkinVisualizationProcess **To Reproduce** - Use `embedded_symbolic_navier_stokes_discontinuous` as element_type - Add `ApplyEmbeddedSkinVisualizationProcess` in your ProjectParameters to have a visualization_model_part,( e.g. "visualization_variables": ["VELOCITY","PRESSURE"]) - Add a `GiDOutputProcess` for the visualization model part - Run a couple of time steps (more than output_interval of GiDOutputProcess) **Expected behavior** The whole mesh of the visualization model part should be written by the output process. **Environment** - OS: Linux - Branch: master - Python 3.6
process
output of visualization model part mpi description when running a simulation with mpi and using an gid or vtk output process for the skin visualization model part only rank is written and therefore only a section of the mesh scope embedded navier stokes discontinuous output of embeddedskinvisualizationprocess to reproduce use embedded symbolic navier stokes discontinuous as element type add applyembeddedskinvisualizationprocess in your projectparameters to have a visualization model part e g visualization variables add a gidoutputprocess for the visualization model part run a couple of time steps more than output interval of gidoutputprocess expected behavior the whole mesh of the visualization model part should be written by the output process environment os linux branch master python
1
21,185
28,153,362,222
IssuesEvent
2023-04-03 04:46:49
ssytnt/papers
https://api.github.com/repos/ssytnt/papers
opened
Learning to see in the dark[Chen+(UIUC),CVPR2018)
ImageProcessing
## 概要 超短時間露光で撮影された画像のためのデノイズ技術。 ## 新規性 ・従来手法では複数の処理を行う過程でノイズが蓄積するため、入力を短時間露光画像のベイヤー配列、出力を長時間露光で撮影したRGB画像とするDNNをE2Eで学習。 ・複数の露光時間で屋外/屋内のシーンを撮影し専用のデータセットを作成(既存のデータセットにガウスノイズ付加では実データに対応できないため)。 ・明るさ決めるAmplication ratioを入力画像の露光時間に応じて入力(推論時も可変)。 ## 結果 ・Amazon Mechanical Turkによる主観評価、およびPSNR/SSIMによる定量評価で従来法を上回る性能。 ・現状はカメラ毎に学習が必要なので、カメラ機種に依存しないような工夫が必要。 ・Ablation analysisの結果、入力をベイヤー配列からRGB画像にすると大幅に性能劣化。 ![画像18](https://user-images.githubusercontent.com/129141420/229413847-70202981-a197-42bf-84d7-183710bd2d81.png)
1.0
Learning to see in the dark[Chen+(UIUC),CVPR2018) - ## 概要 超短時間露光で撮影された画像のためのデノイズ技術。 ## 新規性 ・従来手法では複数の処理を行う過程でノイズが蓄積するため、入力を短時間露光画像のベイヤー配列、出力を長時間露光で撮影したRGB画像とするDNNをE2Eで学習。 ・複数の露光時間で屋外/屋内のシーンを撮影し専用のデータセットを作成(既存のデータセットにガウスノイズ付加では実データに対応できないため)。 ・明るさ決めるAmplication ratioを入力画像の露光時間に応じて入力(推論時も可変)。 ## 結果 ・Amazon Mechanical Turkによる主観評価、およびPSNR/SSIMによる定量評価で従来法を上回る性能。 ・現状はカメラ毎に学習が必要なので、カメラ機種に依存しないような工夫が必要。 ・Ablation analysisの結果、入力をベイヤー配列からRGB画像にすると大幅に性能劣化。 ![画像18](https://user-images.githubusercontent.com/129141420/229413847-70202981-a197-42bf-84d7-183710bd2d81.png)
process
learning to see in the dark chen uiuc 概要 超短時間露光で撮影された画像のためのデノイズ技術。 新規性 ・従来手法では複数の処理を行う過程でノイズが蓄積するため、入力を短時間露光画像のベイヤー配列、 。 ・複数の露光時間で屋外 屋内のシーンを撮影し専用のデータセットを作成(既存のデータセットにガウスノイズ付加では実データに対応できないため)。 ・明るさ決めるamplication ratioを入力画像の露光時間に応じて入力(推論時も可変)。 結果 ・amazon mechanical turkによる主観評価、およびpsnr ssimによる定量評価で従来法を上回る性能。 ・現状はカメラ毎に学習が必要なので、カメラ機種に依存しないような工夫が必要。 ・ablation analysisの結果、入力をベイヤー配列からrgb画像にすると大幅に性能劣化。
1
304,945
23,091,911,314
IssuesEvent
2022-07-26 15:52:11
Hadron/entanglement
https://api.github.com/repos/Hadron/entanglement
closed
Integrate internal docs on Entanglement hooks
documentation
@hadronjeff pointed me at an internal page in his Hadron gitlab with a good summary of hook calling order for things like should_send and should_listen. We should integrate that into the Entanglement docs. `raw/master/knowledge/entanglement-hooks.md` in Jeff's internal playground repo
1.0
Integrate internal docs on Entanglement hooks - @hadronjeff pointed me at an internal page in his Hadron gitlab with a good summary of hook calling order for things like should_send and should_listen. We should integrate that into the Entanglement docs. `raw/master/knowledge/entanglement-hooks.md` in Jeff's internal playground repo
non_process
integrate internal docs on entanglement hooks hadronjeff pointed me at an internal page in his hadron gitlab with a good summary of hook calling order for things like should send and should listen we should integrate that into the entanglement docs raw master knowledge entanglement hooks md in jeff s internal playground repo
0
14,138
17,030,959,898
IssuesEvent
2021-07-04 14:54:53
rladstaetter/LogoRRR
https://api.github.com/repos/rladstaetter/LogoRRR
closed
Release 21.3.0
release process
### Release of LogoRRR 21.3.0 - [x] Update Release Notes - [x] Screenshot for Release Notes - [x] Update Hompage - [x] Update animated gif in Readme.md - [x] Build binary artefacts Windows - [x] Build binary artefacts MacOsX - [x] Release via Github, update Links - [x] Check Readmes and Links - [x] close all issues - [x] tweet
1.0
Release 21.3.0 - ### Release of LogoRRR 21.3.0 - [x] Update Release Notes - [x] Screenshot for Release Notes - [x] Update Hompage - [x] Update animated gif in Readme.md - [x] Build binary artefacts Windows - [x] Build binary artefacts MacOsX - [x] Release via Github, update Links - [x] Check Readmes and Links - [x] close all issues - [x] tweet
process
release release of logorrr update release notes screenshot for release notes update hompage update animated gif in readme md build binary artefacts windows build binary artefacts macosx release via github update links check readmes and links close all issues tweet
1
50,300
3,006,308,262
IssuesEvent
2015-07-27 09:33:20
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
opened
Sometimes stitcher generate result file with size 158 bytes
affected: 2.4 auto-transferred bug category: stitching priority: normal
Transferred from http://code.opencv.org/issues/2880 ``` || Nikolay Rogoshchenkov on 2013-03-11 14:09 || Priority: Normal || Affected: 2.4.3 || Category: stitching || Tracker: Bug || Difficulty: None || PR: || Platform: None / None ``` Sometimes stitcher generate result file with size 158 bytes ----------- ``` Sometimes simple stitcher procedure generates result file with size 158 bytes: <pre> Stitcher stitcher = Stitcher::createDefault(false); Stitcher::Status status = stitcher.stitch(imgs, rois, pano); imwrite(result_name, pano); </pre> Full source code and img files in attachment ``` History ------- ##### Nikolay Rogoshchenkov on 2013-03-11 14:13 ``` Please close this request. Can't reproduce this again. ``` ##### Nikolay Rogoshchenkov on 2013-03-11 19:31 ``` Can reproduce problem with this images and src http://nikkadim.com/img_and_src.zip ~2.8Mb ```
1.0
Sometimes stitcher generate result file with size 158 bytes - Transferred from http://code.opencv.org/issues/2880 ``` || Nikolay Rogoshchenkov on 2013-03-11 14:09 || Priority: Normal || Affected: 2.4.3 || Category: stitching || Tracker: Bug || Difficulty: None || PR: || Platform: None / None ``` Sometimes stitcher generate result file with size 158 bytes ----------- ``` Sometimes simple stitcher procedure generates result file with size 158 bytes: <pre> Stitcher stitcher = Stitcher::createDefault(false); Stitcher::Status status = stitcher.stitch(imgs, rois, pano); imwrite(result_name, pano); </pre> Full source code and img files in attachment ``` History ------- ##### Nikolay Rogoshchenkov on 2013-03-11 14:13 ``` Please close this request. Can't reproduce this again. ``` ##### Nikolay Rogoshchenkov on 2013-03-11 19:31 ``` Can reproduce problem with this images and src http://nikkadim.com/img_and_src.zip ~2.8Mb ```
non_process
sometimes stitcher generate result file with size bytes transferred from nikolay rogoshchenkov on priority normal affected category stitching tracker bug difficulty none pr platform none none sometimes stitcher generate result file with size bytes sometimes simple stitcher procedure generates result file with size bytes stitcher stitcher stitcher createdefault false stitcher status status stitcher stitch imgs rois pano imwrite result name pano full source code and img files in attachment history nikolay rogoshchenkov on please close this request can t reproduce this again nikolay rogoshchenkov on can reproduce problem with this images and src
0
22,311
30,864,433,379
IssuesEvent
2023-08-03 06:59:08
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Compatibility issue with ProcessStartInfo with UseShellExecute=False and WindowStyle=Hidden
area-System.Diagnostics.Process blocking-release regression-from-last-release
Earlier in .NET 8, we made a change to [Support WindowStyle without UseShellExecute #82662](https://github.com/dotnet/runtime/pull/82662) to open up a scenario for PowerShell to avoid the need for a P/Invoke. This change was found to cause unexpected behavior changes regarding the order of events received by child process windows (such as `Activated` and `Loaded`). This led to [Test failures in WPF](https://github.com/dotnet/wpf/issues/8043). We need to reconsider the changes made in #82662 for #81681 and determine if the event ordering change can be safely mitigated to not cause this unexpected change in behavior. As it is, those WPF tests indicate the likelihood of app developers having set these properties (even if not wholly necessary in their scenarios), and now their subsequent code can have application-level bugs where the event ordering leads to different behavior. In the case of the WPF tests, this change led to an `ArgumentOutOfRangeException` later in the application code because the tests depended on the event ordering. Unless there is a clear, low-risk fix, we should err on the side of reverting the change. Reverting will have impact on PowerShell though, as it's likely they've updated code to remove the P/Invoke that would no longer be needed with the new behavior that became supported. @Jozkee @jborean93 @SteveL-MSFT @iSazonov /cc and kudos to @rchauhan18 for the diagnosis conducted on the WPF test issue.
1.0
Compatibility issue with ProcessStartInfo with UseShellExecute=False and WindowStyle=Hidden - Earlier in .NET 8, we made a change to [Support WindowStyle without UseShellExecute #82662](https://github.com/dotnet/runtime/pull/82662) to open up a scenario for PowerShell to avoid the need for a P/Invoke. This change was found to cause unexpected behavior changes regarding the order of events received by child process windows (such as `Activated` and `Loaded`). This led to [Test failures in WPF](https://github.com/dotnet/wpf/issues/8043). We need to reconsider the changes made in #82662 for #81681 and determine if the event ordering change can be safely mitigated to not cause this unexpected change in behavior. As it is, those WPF tests indicate the likelihood of app developers having set these properties (even if not wholly necessary in their scenarios), and now their subsequent code can have application-level bugs where the event ordering leads to different behavior. In the case of the WPF tests, this change led to an `ArgumentOutOfRangeException` later in the application code because the tests depended on the event ordering. Unless there is a clear, low-risk fix, we should err on the side of reverting the change. Reverting will have impact on PowerShell though, as it's likely they've updated code to remove the P/Invoke that would no longer be needed with the new behavior that became supported. @Jozkee @jborean93 @SteveL-MSFT @iSazonov /cc and kudos to @rchauhan18 for the diagnosis conducted on the WPF test issue.
process
compatibility issue with processstartinfo with useshellexecute false and windowstyle hidden earlier in net we made a change to to open up a scenario for powershell to avoid the need for a p invoke this change was found to cause unexpected behavior changes regarding the order of events received by child process windows such as activated and loaded this led to we need to reconsider the changes made in for and determine if the event ordering change can be safely mitigated to not cause this unexpected change in behavior as it is those wpf tests indicate the likelihood of app developers having set these properties even if not wholly necessary in their scenarios and now their subsequent code can have application level bugs where the event ordering leads to different behavior in the case of the wpf tests this change led to an argumentoutofrangeexception later in the application code because the tests depended on the event ordering unless there is a clear low risk fix we should err on the side of reverting the change reverting will have impact on powershell though as it s likely they ve updated code to remove the p invoke that would no longer be needed with the new behavior that became supported jozkee stevel msft isazonov cc and kudos to for the diagnosis conducted on the wpf test issue
1
7,269
10,422,408,139
IssuesEvent
2019-09-16 08:57:54
dzhw/zofar
https://api.github.com/repos/dzhw/zofar
closed
As a Developer i want to replace Webstart for Zofar-Tools
13 category: secondary.cockpit category: secondary.exporter category: technical.processes prio: 1 type: backlog.item
Reasons for the need to replace Webstart as Distribution-System are: 1. Webstart is deprecated 2. Clients needs a installed java 3. Execution of Tools are limited by the Resources of Client-System 4. Tools have to be signed by Certificate. In some cases this leads to conflicts with used Frameworks
1.0
As a Developer i want to replace Webstart for Zofar-Tools - Reasons for the need to replace Webstart as Distribution-System are: 1. Webstart is deprecated 2. Clients needs a installed java 3. Execution of Tools are limited by the Resources of Client-System 4. Tools have to be signed by Certificate. In some cases this leads to conflicts with used Frameworks
process
as a developer i want to replace webstart for zofar tools reasons for the need to replace webstart as distribution system are webstart is deprecated clients needs a installed java execution of tools are limited by the resources of client system tools have to be signed by certificate in some cases this leads to conflicts with used frameworks
1
14,769
18,047,630,358
IssuesEvent
2021-09-19 06:53:57
CATcher-org/CATcher
https://api.github.com/repos/CATcher-org/CATcher
closed
Staging application's deployment is not working, due to permissions issue
aspect-Process category.Bug p.High
The [staging application](https://github.com/CATcher-org/CATcher-staging) has a permissions problem in its deployment workflow. This is preventing new versions of the staging app from being deployed at https://catcher-org.github.io/CATcher-staging/ Here are the logs from one of the failed deployments: ``` Date: 2021-09-09T03:15:10.264Z Hash: cf434a986532f6a6b612 Time: 78444ms chunk {0} runtime.80ab492fe3d778817936.js (runtime) 1.41 kB [entry] [rendered] chunk {1} main.bfbc1463abcd16815d1a.js (main) 2.13 MB [initial] [rendered] chunk {2} polyfills.b040449b01dbda52469f.js (polyfills) 40.9 kB [initial] [rendered] chunk {3} styles.c935e9fa9c93e6cc7641.css (styles) 72.7 kB [initial] [rendered] chunk {scripts} scripts.23b5a7e9c7d07ed89179.js (scripts) 47.7 kB [entry] [rendered] 🚀 Uploading via git, please wait... ❌ An error occurred when trying to deploy: fatal: could not read Password for 'https://***@github.com': No such device or address ``` The application itself can be built, as seen above (the JS files are built). It seems to be an issue with the authentication (needed for the deployment to be done on GitHub pages). I'm currently not very clear on how this issue arose.
1.0
Staging application's deployment is not working, due to permissions issue - The [staging application](https://github.com/CATcher-org/CATcher-staging) has a permissions problem in its deployment workflow. This is preventing new versions of the staging app from being deployed at https://catcher-org.github.io/CATcher-staging/ Here are the logs from one of the failed deployments: ``` Date: 2021-09-09T03:15:10.264Z Hash: cf434a986532f6a6b612 Time: 78444ms chunk {0} runtime.80ab492fe3d778817936.js (runtime) 1.41 kB [entry] [rendered] chunk {1} main.bfbc1463abcd16815d1a.js (main) 2.13 MB [initial] [rendered] chunk {2} polyfills.b040449b01dbda52469f.js (polyfills) 40.9 kB [initial] [rendered] chunk {3} styles.c935e9fa9c93e6cc7641.css (styles) 72.7 kB [initial] [rendered] chunk {scripts} scripts.23b5a7e9c7d07ed89179.js (scripts) 47.7 kB [entry] [rendered] 🚀 Uploading via git, please wait... ❌ An error occurred when trying to deploy: fatal: could not read Password for 'https://***@github.com': No such device or address ``` The application itself can be built, as seen above (the JS files are built). It seems to be an issue with the authentication (needed for the deployment to be done on GitHub pages). I'm currently not very clear on how this issue arose.
process
staging application s deployment is not working due to permissions issue the has a permissions problem in its deployment workflow this is preventing new versions of the staging app from being deployed at here are the logs from one of the failed deployments date hash time chunk runtime js runtime kb chunk main js main mb chunk polyfills js polyfills kb chunk styles css styles kb chunk scripts scripts js scripts kb 🚀 uploading via git please wait ❌ an error occurred when trying to deploy fatal could not read password for no such device or address the application itself can be built as seen above the js files are built it seems to be an issue with the authentication needed for the deployment to be done on github pages i m currently not very clear on how this issue arose
1
5,988
8,805,374,662
IssuesEvent
2018-12-26 19:14:01
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Inconsistent output with reused chunk="to-content" topics
preprocess/chunking stale
Ported from Yahoo User Group http://tech.groups.yahoo.com/group/dita-users/message/31511 Map - Parent Topic (chunk="to-content") - Child 1 - Child 2 - Child 1 (reuse) In this case, the Child 1 html is generated nicely. However, if we switch position of Parent Topic with the last Child 1 (switch the order of two top level topics), the Child 1.html is not being generated.
1.0
Inconsistent output with reused chunk="to-content" topics - Ported from Yahoo User Group http://tech.groups.yahoo.com/group/dita-users/message/31511 Map - Parent Topic (chunk="to-content") - Child 1 - Child 2 - Child 1 (reuse) In this case, the Child 1 html is generated nicely. However, if we switch position of Parent Topic with the last Child 1 (switch the order of two top level topics), the Child 1.html is not being generated.
process
inconsistent output with reused chunk to content topics ported from yahoo user group map parent topic chunk to content child child child reuse in this case the child html is generated nicely however if we switch position of parent topic with the last child switch the order of two top level topics the child html is not being generated
1
18,518
24,551,754,131
IssuesEvent
2022-10-12 13:07:58
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [Offline indicator] Empty screen is getting displayed when user clicks on the 'Disagree' present on the Review consent screen
Bug P0 iOS Process: Fixed Process: Tested dev
Steps: 1. Sign up or sign in to the app 2. Click on the study and try to enroll 3. Navigate to the 'Review consent' screen 4. turn off the internet 5. Click on the Done button 6. Observe AR: Participant is navigating to the empty screen ER: Participant should get alert pop up and after clicking on 'Ok' button participant should navigate to the study list screen Note: 1. Issue should also be fixed in the review updated consent flow 2. Issue observed when Participant clicks on the 'Next' button present on the Signature screen https://user-images.githubusercontent.com/71445210/185955899-89f08802-d329-4f1b-9e79-e7c83dc42848.MOV
2.0
[iOS] [Offline indicator] Empty screen is getting displayed when user clicks on the 'Disagree' present on the Review consent screen - Steps: 1. Sign up or sign in to the app 2. Click on the study and try to enroll 3. Navigate to the 'Review consent' screen 4. turn off the internet 5. Click on the Done button 6. Observe AR: Participant is navigating to the empty screen ER: Participant should get alert pop up and after clicking on 'Ok' button participant should navigate to the study list screen Note: 1. Issue should also be fixed in the review updated consent flow 2. Issue observed when Participant clicks on the 'Next' button present on the Signature screen https://user-images.githubusercontent.com/71445210/185955899-89f08802-d329-4f1b-9e79-e7c83dc42848.MOV
process
empty screen is getting displayed when user clicks on the disagree present on the review consent screen steps sign up or sign in to the app click on the study and try to enroll navigate to the review consent screen turn off the internet click on the done button observe ar participant is navigating to the empty screen er participant should get alert pop up and after clicking on ok button participant should navigate to the study list screen note issue should also be fixed in the review updated consent flow issue observed when participant clicks on the next button present on the signature screen
1
12,753
15,110,856,457
IssuesEvent
2021-02-08 19:47:58
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Add apple M1 to build scripts & release process
kind/process priority/important-soon
Hopefully simply setting `GOARCH=arm64` is enough, but this needs testing.
1.0
Add apple M1 to build scripts & release process - Hopefully simply setting `GOARCH=arm64` is enough, but this needs testing.
process
add apple to build scripts release process hopefully simply setting goarch is enough but this needs testing
1
11,994
14,737,225,555
IssuesEvent
2021-01-07 01:13:46
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
068: portland - multiple client customer login
anc-ops anc-process anp-0.5 ant-bug ant-support
In GitLab by @kdjstudios on Apr 26, 2018, 11:14 **Submitted by:** "Lettice Ross" <lettice.ross@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-26-66649/conversation **Server:** Internal **Client/Site:** 068 **Account:** Multiple **Issue:** I have a client that is getting an error message when trying to login to the portal. I have attached the email she sent me, in regards to the message. Here are her accounts: 068-B01342 Clackamas Broadband Express 068-B09420 Clackamas Technology Services I have sent in a request yesterday, because she have two accounts, and would like to know if they can have both accounts under one login.
1.0
068: portland - multiple client customer login - In GitLab by @kdjstudios on Apr 26, 2018, 11:14 **Submitted by:** "Lettice Ross" <lettice.ross@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-26-66649/conversation **Server:** Internal **Client/Site:** 068 **Account:** Multiple **Issue:** I have a client that is getting an error message when trying to login to the portal. I have attached the email she sent me, in regards to the message. Here are her accounts: 068-B01342 Clackamas Broadband Express 068-B09420 Clackamas Technology Services I have sent in a request yesterday, because she have two accounts, and would like to know if they can have both accounts under one login.
process
portland multiple client customer login in gitlab by kdjstudios on apr submitted by lettice ross helpdesk server internal client site account multiple issue i have a client that is getting an error message when trying to login to the portal i have attached the email she sent me in regards to the message here are her accounts clackamas broadband express clackamas technology services i have sent in a request yesterday because she have two accounts and would like to know if they can have both accounts under one login
1
121,277
10,163,749,019
IssuesEvent
2019-08-07 09:58:04
elastic/cloud-on-k8s
https://api.github.com/repos/elastic/cloud-on-k8s
closed
Trigger e2e test run after running release job on CI
:ci >test
We need to setup triggering e2e tests after successful build of release CI job.
1.0
Trigger e2e test run after running release job on CI - We need to setup triggering e2e tests after successful build of release CI job.
non_process
trigger test run after running release job on ci we need to setup triggering tests after successful build of release ci job
0
3,277
6,363,237,202
IssuesEvent
2017-07-31 16:43:33
RadeonOpenCompute/ROCm-OpenCL-Driver
https://api.github.com/repos/RadeonOpenCompute/ROCm-OpenCL-Driver
opened
[in-process] Using statically linked lld leads to segmentation fault
bug difficulty:C_Hard [in-process]
Using function lld::elf::link() leads to segmentation fault on second call. First call finishes correctly and tests are passed.
1.0
[in-process] Using statically linked lld leads to segmentation fault - Using function lld::elf::link() leads to segmentation fault on second call. First call finishes correctly and tests are passed.
process
using statically linked lld leads to segmentation fault using function lld elf link leads to segmentation fault on second call first call finishes correctly and tests are passed
1
17,444
23,266,929,301
IssuesEvent
2022-08-04 18:21:17
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Extract Telemetry Query Language into a package
priority:p1 processor/transform
The transform processor currently implements the Telemetry Query Language within its `internal/common` package. In order to expose the TQL so it can be used by anyone, it should be moved to `pkg/telemetryquerylanguage`. The TQL would be usable without any of the transform processor's signal contexts or functions, but it wouldn't be very useful. The processor's contexts should also be moved into packages so that the TQL is easily usable with pdata. Functions might also be movable, but not all functions would be applicable outside the transform processor. Steps: - [x] Move Telemetry Query Language - [x] Update Transform Processor to use TQL package - [x] Move standard contexts - [x] Update Transform Processor to use standard context package(s) - [x] Move appropriate functions - [x] Update Transform Processor to use new function package(s) - [ ] Update processor README
1.0
Extract Telemetry Query Language into a package - The transform processor currently implements the Telemetry Query Language within its `internal/common` package. In order to expose the TQL so it can be used by anyone, it should be moved to `pkg/telemetryquerylanguage`. The TQL would be usable without any of the transform processor's signal contexts or functions, but it wouldn't be very useful. The processor's contexts should also be moved into packages so that the TQL is easily usable with pdata. Functions might also be movable, but not all functions would be applicable outside the transform processor. Steps: - [x] Move Telemetry Query Language - [x] Update Transform Processor to use TQL package - [x] Move standard contexts - [x] Update Transform Processor to use standard context package(s) - [x] Move appropriate functions - [x] Update Transform Processor to use new function package(s) - [ ] Update processor README
process
extract telemetry query language into a package the transform processor currently implements the telemetry query language within its internal common package in order to expose the tql so it can be used by anyone it should be moved to pkg telemetryquerylanguage the tql would be usable without any of the transform processor s signal contexts or functions but it wouldn t be very useful the processor s contexts should also be moved into packages so that the tql is easily usable with pdata functions might also be movable but not all functions would be applicable outside the transform processor steps move telemetry query language update transform processor to use tql package move standard contexts update transform processor to use standard context package s move appropriate functions update transform processor to use new function package s update processor readme
1
21,409
4,708,813,156
IssuesEvent
2016-10-14 01:36:53
aurelia/documentation
https://api.github.com/repos/aurelia/documentation
closed
A couple of typos in documentation
documentation help wanted
There are the following typos in documentation: * In [Creating A New Aurelia Project](http://aurelia.io/hub.html#/doc/article/aurelia/framework/latest/contact-manager-tutorial/2) section of "Contact Manager Tutorial": > Hit enter to *accpet* the default "yes". * In code for "Basic HTTP Interceptors" inside [aurelia-http-client](http://aurelia.io/hub.html#/doc/article/aurelia/fetch-client/latest/http-services/3) section of "HTTP Services" page: ```js let client = new HttpClient(); .configure(x => { ``` * In [Creating Components](http://aurelia.io/hub.html#/doc/article/aurelia/framework/latest/creating-components/1) there are missed and improper periods in the following sentence: > Those are the defaults, but you can always be explicit about the binding direction by using .one-way, two-way or .one-time in place of .bind. (.one-time renders the initial value of the property but does not perform synchronization thereafter, making it a nice memory and performance gain for data you know will not change.) * In [Making an HTML Behavior Available in a View](http://aurelia.io/hub.html#/doc/article/aurelia/templating/latest/templating-html-behaviors-introduction/4) section of [Templating: HTML Behaviors Introduction](http://aurelia.io/hub.html#/doc/article/aurelia/templating/latest/templating-html-behaviors-introduction/): > There is one exception to this rule*, though.* Aurelia provides for "HTML-only Custom Elements*."* It seems this sentence should be as the following: > There is one exception to this rule, though, Aurelia provides for "HTML-only Custom Elements".
1.0
A couple of typos in documentation - There are the following typos in documentation: * In [Creating A New Aurelia Project](http://aurelia.io/hub.html#/doc/article/aurelia/framework/latest/contact-manager-tutorial/2) section of "Contact Manager Tutorial": > Hit enter to *accpet* the default "yes". * In code for "Basic HTTP Interceptors" inside [aurelia-http-client](http://aurelia.io/hub.html#/doc/article/aurelia/fetch-client/latest/http-services/3) section of "HTTP Services" page: ```js let client = new HttpClient(); .configure(x => { ``` * In [Creating Components](http://aurelia.io/hub.html#/doc/article/aurelia/framework/latest/creating-components/1) there are missed and improper periods in the following sentence: > Those are the defaults, but you can always be explicit about the binding direction by using .one-way, two-way or .one-time in place of .bind. (.one-time renders the initial value of the property but does not perform synchronization thereafter, making it a nice memory and performance gain for data you know will not change.) * In [Making an HTML Behavior Available in a View](http://aurelia.io/hub.html#/doc/article/aurelia/templating/latest/templating-html-behaviors-introduction/4) section of [Templating: HTML Behaviors Introduction](http://aurelia.io/hub.html#/doc/article/aurelia/templating/latest/templating-html-behaviors-introduction/): > There is one exception to this rule*, though.* Aurelia provides for "HTML-only Custom Elements*."* It seems this sentence should be as the following: > There is one exception to this rule, though, Aurelia provides for "HTML-only Custom Elements".
non_process
a couple of typos in documentation there are the following typos in documentation in section of contact manager tutorial hit enter to accpet the default yes in code for basic http interceptors inside section of http services page js let client new httpclient configure x in there are missed and improper periods in the following sentence those are the defaults but you can always be explicit about the binding direction by using one way two way or one time in place of bind one time renders the initial value of the property but does not perform synchronization thereafter making it a nice memory and performance gain for data you know will not change in section of there is one exception to this rule though aurelia provides for html only custom elements it seems this sentence should be as the following there is one exception to this rule though aurelia provides for html only custom elements
0
9,515
8,656,115,105
IssuesEvent
2018-11-27 17:32:41
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Details for "Modernize your existing Cloud Services applications by migrating to Service Fabric Mesh" needed
assigned-to-author doc-enhancement service-fabric-mesh/svc triaged
The overview lists this as a main benefit: > Modernize your existing Cloud Services applications by migrating to Service Fabric Mesh Can you please document this as a detailed how-to guide starting from an existing cloud service? In addition, please 1. compare pros-cons of the two services' in a section 2. address how to install certificates to the certificate store in Service Fabric Mesh (e.g. Cloud Service addresses this very clearly) Thank you --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 1772867c-92c7-f8ae-e19a-929478763da3 * Version Independent ID: 2df08f43-c2be-63e3-1b6f-99932ddd593f * Content: [Overview of Azure Service Fabric Mesh](https://docs.microsoft.com/en-us/azure/service-fabric-mesh/service-fabric-mesh-overview) * Content Source: [articles/service-fabric-mesh/service-fabric-mesh-overview.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-fabric-mesh/service-fabric-mesh-overview.md) * Service: **service-fabric-mesh** * GitHub Login: @rwike77 * Microsoft Alias: **ryanwi**
1.0
Details for "Modernize your existing Cloud Services applications by migrating to Service Fabric Mesh" needed - The overview lists this as a main benefit: > Modernize your existing Cloud Services applications by migrating to Service Fabric Mesh Can you please document this as a detailed how-to guide starting from an existing cloud service? In addition, please 1. compare pros-cons of the two services' in a section 2. address how to install certificates to the certificate store in Service Fabric Mesh (e.g. Cloud Service addresses this very clearly) Thank you --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 1772867c-92c7-f8ae-e19a-929478763da3 * Version Independent ID: 2df08f43-c2be-63e3-1b6f-99932ddd593f * Content: [Overview of Azure Service Fabric Mesh](https://docs.microsoft.com/en-us/azure/service-fabric-mesh/service-fabric-mesh-overview) * Content Source: [articles/service-fabric-mesh/service-fabric-mesh-overview.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-fabric-mesh/service-fabric-mesh-overview.md) * Service: **service-fabric-mesh** * GitHub Login: @rwike77 * Microsoft Alias: **ryanwi**
non_process
details for modernize your existing cloud services applications by migrating to service fabric mesh needed the overview lists this as a main benefit modernize your existing cloud services applications by migrating to service fabric mesh can you please document this as a detailed how to guide starting from an existing cloud service in addition please compare pros cons of the two services in a section address how to install certificates to the certificate store in service fabric mesh e g cloud service addresses this very clearly thank you document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service service fabric mesh github login microsoft alias ryanwi
0
283,517
30,913,332,026
IssuesEvent
2023-08-05 01:41:17
hshivhare67/kernel_v4.19.72_CVE-2022-42896_new
https://api.github.com/repos/hshivhare67/kernel_v4.19.72_CVE-2022-42896_new
reopened
CVE-2021-38300 (High) detected in linuxlinux-4.19.279
Mend: dependency security vulnerability
## CVE-2021-38300 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/mips/net/bpf_jit.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/mips/net/bpf_jit.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> arch/mips/net/bpf_jit.c in the Linux kernel before 5.4.10 can generate undesirable machine code when transforming unprivileged cBPF programs, allowing execution of arbitrary code within the kernel context. This occurs because conditional branches can exceed the 128 KB limit of the MIPS architecture. <p>Publish Date: 2021-09-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-38300>CVE-2021-38300</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-38300">https://www.linuxkernelcves.com/cves/CVE-2021-38300</a></p> <p>Release Date: 2021-09-20</p> <p>Fix Resolution: v4.14.251,v4.19.211,v5.4.153,v5.10.71,v5.14.10,v5.15-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-38300 (High) detected in linuxlinux-4.19.279 - ## CVE-2021-38300 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/mips/net/bpf_jit.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/mips/net/bpf_jit.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> arch/mips/net/bpf_jit.c in the Linux kernel before 5.4.10 can generate undesirable machine code when transforming unprivileged cBPF programs, allowing execution of arbitrary code within the kernel context. This occurs because conditional branches can exceed the 128 KB limit of the MIPS architecture. <p>Publish Date: 2021-09-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-38300>CVE-2021-38300</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-38300">https://www.linuxkernelcves.com/cves/CVE-2021-38300</a></p> <p>Release Date: 2021-09-20</p> <p>Fix Resolution: v4.14.251,v4.19.211,v5.4.153,v5.10.71,v5.14.10,v5.15-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files arch mips net bpf jit c arch mips net bpf jit c vulnerability details arch mips net bpf jit c in the linux kernel before can generate undesirable machine code when transforming unprivileged cbpf programs allowing execution of arbitrary code within the kernel context this occurs because conditional branches can exceed the kb limit of the mips architecture publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
206,290
15,724,566,730
IssuesEvent
2021-03-29 08:57:47
CSOIreland/PxStat
https://api.github.com/repos/CSOIreland/PxStat
closed
[BUG] Intermittent row locked error when trying to extend session
bug fixed released tested
For an externally logged in user, when attempting to extend the session time in the database, an error occurs because the row is locked.
1.0
[BUG] Intermittent row locked error when trying to extend session - For an externally logged in user, when attempting to extend the session time in the database, an error occurs because the row is locked.
non_process
intermittent row locked error when trying to extend session for an externally logged in user when attempting to extend the session time in the database an error occurs because the row is locked
0
214,753
16,577,140,254
IssuesEvent
2021-05-31 06:55:55
TesseractCoding/NeoAlgo-Docs
https://api.github.com/repos/TesseractCoding/NeoAlgo-Docs
closed
Mathematics
documentation
## 📚 Documentation (A clear and concise description of what the issue is.) Mathematics ### Have you read the [Contributing Guidelines on Pull Requests](https://github.com/TesseractCoding/NeoAlgo/blob/master/CONTRIBUTING.md#reporting-new-issues)? (Write your answer here.) yes
1.0
Mathematics - ## 📚 Documentation (A clear and concise description of what the issue is.) Mathematics ### Have you read the [Contributing Guidelines on Pull Requests](https://github.com/TesseractCoding/NeoAlgo/blob/master/CONTRIBUTING.md#reporting-new-issues)? (Write your answer here.) yes
non_process
mathematics 📚 documentation a clear and concise description of what the issue is mathematics have you read the write your answer here yes
0
9,436
12,424,719,990
IssuesEvent
2020-05-24 13:06:15
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
reopened
Issue with new automation account default options
Pri2 automation/svc cxp process-automation/subsvc product-issue triaged
Following along I just ran into an issue with this step while creating a new automation account: "All other options are automatically populated based on the Log Analytics workspace selected. These options cannot be modified. An Azure Run As account is the default authentication method for the runbooks included in this solution. After you click OK, the configuration options are validated and the Automation account is created. You can track its progress under Notifications from the menu." I was unable to modify options other than the name as is expected. However, subscription and region name assigned to the new automation account ended up being different than those where selected log analytics workspace is. i.e. account ended up in East US 2 while log analytics workspace selected is in East US. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7ceebde9-ebae-6dbd-c53c-30a654a52c74 * Version Independent ID: e7ce38c3-eab5-9077-4bd1-818e906fcccf * Content: [Enable Azure Automation Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-enable) * Content Source: [articles/automation/automation-solution-vm-management-enable.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management-enable.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Issue with new automation account default options - Following along I just ran into an issue with this step while creating a new automation account: "All other options are automatically populated based on the Log Analytics workspace selected. These options cannot be modified. An Azure Run As account is the default authentication method for the runbooks included in this solution. After you click OK, the configuration options are validated and the Automation account is created. You can track its progress under Notifications from the menu." I was unable to modify options other than the name as is expected. However, subscription and region name assigned to the new automation account ended up being different than those where selected log analytics workspace is. i.e. account ended up in East US 2 while log analytics workspace selected is in East US. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7ceebde9-ebae-6dbd-c53c-30a654a52c74 * Version Independent ID: e7ce38c3-eab5-9077-4bd1-818e906fcccf * Content: [Enable Azure Automation Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-enable) * Content Source: [articles/automation/automation-solution-vm-management-enable.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management-enable.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
issue with new automation account default options following along i just ran into an issue with this step while creating a new automation account all other options are automatically populated based on the log analytics workspace selected these options cannot be modified an azure run as account is the default authentication method for the runbooks included in this solution after you click ok the configuration options are validated and the automation account is created you can track its progress under notifications from the menu i was unable to modify options other than the name as is expected however subscription and region name assigned to the new automation account ended up being different than those where selected log analytics workspace is i e account ended up in east us while log analytics workspace selected is in east us document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id ebae version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
271,273
8,482,123,575
IssuesEvent
2018-10-25 17:38:41
bcgov/name-examination
https://api.github.com/repos/bcgov/name-examination
opened
Need a way to do a Free Search
Priority: 1
####Background The examiners are not enjoying doing free searches in the examination screen as they are worried about accidentally doing something to it. We originally thought they would just use the manual search in the examination screen but I can see how it is a different process. ##### _**Detailed Description**_ Either they get a test one from the hold que or we give them a simple screen off a button to let them test search that doesn't have action buttons. Please mock something up under priority 1.
1.0
Need a way to do a Free Search - ####Background The examiners are not enjoying doing free searches in the examination screen as they are worried about accidentally doing something to it. We originally thought they would just use the manual search in the examination screen but I can see how it is a different process. ##### _**Detailed Description**_ Either they get a test one from the hold que or we give them a simple screen off a button to let them test search that doesn't have action buttons. Please mock something up under priority 1.
non_process
need a way to do a free search background the examiners are not enjoying doing free searches in the examination screen as they are worried about accidentally doing something to it we originally thought they would just use the manual search in the examination screen but i can see how it is a different process detailed description either they get a test one from the hold que or we give them a simple screen off a button to let them test search that doesn t have action buttons please mock something up under priority
0
535,203
15,684,285,667
IssuesEvent
2021-03-25 09:49:16
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
Support protocol names in nEdge configuration bootstrap
nEdge priority ticket
Protocol IDs are provided to nEdge in the JSON used to bootstrap the configuration. Check and test also support for protocol names as IDs can change over time (IDs for deprecated protocols are recycled from time to time)
1.0
Support protocol names in nEdge configuration bootstrap - Protocol IDs are provided to nEdge in the JSON used to bootstrap the configuration. Check and test also support for protocol names as IDs can change over time (IDs for deprecated protocols are recycled from time to time)
non_process
support protocol names in nedge configuration bootstrap protocol ids are provided to nedge in the json used to bootstrap the configuration check and test also support for protocol names as ids can change over time ids for deprecated protocols are recycled from time to time
0
6,237
9,182,209,913
IssuesEvent
2019-03-05 12:14:39
SharePoint/PnP-PowerShell
https://api.github.com/repos/SharePoint/PnP-PowerShell
closed
Connect-PnPOnline -lose connection in case there is whitespaces in web url
Needs investigation To be processed
### Reporting an Issue or Missing Feature Make connect with Connect-PnPOnline $url -UseWebLogin where $url = "https://timopertila.sharepoint.com/sites/wiki/Light%20Beer/" You can make one operation, but after that connection is lost and you get exception. Like: $contentType = Get-PnPContentType -Web $web -List Documents -Identity "Timo Doc" is fine, but after that running same (or any else operation) like $contentType = Get-PnPContentType -Web $web -List Documents -Identity "Timo Doc" gives you: Get-PnPContentType : Object reference not set to an instance of an object. At line:2 char:16 + ... ntentType = Get-PnPContentType -Web $web -List Documents -Identity "T ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : NotSpecified: (:) [Get-PnPContentType], NullReferenceException + FullyQualifiedErrorId : System.NullReferenceException,SharePointPnP.PowerShell.Commands.ContentTypes.GetContentType If i do same thing with site url having no whitespaces, then everything works fine. ### Expected behavior Excpected that i can run more than one command after establishing connection. ### Actual behavior <img width="470" alt="pnp bug" src="https://user-images.githubusercontent.com/34052704/33374871-90d94d60-d510-11e7-9699-9fe78ac29e1d.png"> ### Steps to reproduce behavior can seen on screenshot above ### Which version of the PnP-PowerShell Cmdlets are you using? - [ ] PnP PowerShell for SharePoint 2013 - [ ] PnP PowerShell for SharePoint 2016 - [x] PnP PowerShell for SharePoint Online ### What is the version of the Cmdlet module you are running? 2.20.17.. ### How did you install the PnP-PowerShell Cmdlets? - [x ] MSI Installed downloaded from GitHub - [ ] Installed through the PowerShell Gallery with Install-Module - [ ] Other means
1.0
Connect-PnPOnline -lose connection in case there is whitespaces in web url - ### Reporting an Issue or Missing Feature Make connect with Connect-PnPOnline $url -UseWebLogin where $url = "https://timopertila.sharepoint.com/sites/wiki/Light%20Beer/" You can make one operation, but after that connection is lost and you get exception. Like: $contentType = Get-PnPContentType -Web $web -List Documents -Identity "Timo Doc" is fine, but after that running same (or any else operation) like $contentType = Get-PnPContentType -Web $web -List Documents -Identity "Timo Doc" gives you: Get-PnPContentType : Object reference not set to an instance of an object. At line:2 char:16 + ... ntentType = Get-PnPContentType -Web $web -List Documents -Identity "T ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : NotSpecified: (:) [Get-PnPContentType], NullReferenceException + FullyQualifiedErrorId : System.NullReferenceException,SharePointPnP.PowerShell.Commands.ContentTypes.GetContentType If i do same thing with site url having no whitespaces, then everything works fine. ### Expected behavior Excpected that i can run more than one command after establishing connection. ### Actual behavior <img width="470" alt="pnp bug" src="https://user-images.githubusercontent.com/34052704/33374871-90d94d60-d510-11e7-9699-9fe78ac29e1d.png"> ### Steps to reproduce behavior can seen on screenshot above ### Which version of the PnP-PowerShell Cmdlets are you using? - [ ] PnP PowerShell for SharePoint 2013 - [ ] PnP PowerShell for SharePoint 2016 - [x] PnP PowerShell for SharePoint Online ### What is the version of the Cmdlet module you are running? 2.20.17.. ### How did you install the PnP-PowerShell Cmdlets? - [x ] MSI Installed downloaded from GitHub - [ ] Installed through the PowerShell Gallery with Install-Module - [ ] Other means
process
connect pnponline lose connection in case there is whitespaces in web url reporting an issue or missing feature make connect with connect pnponline url useweblogin where url you can make one operation but after that connection is lost and you get exception like contenttype get pnpcontenttype web web list documents identity timo doc is fine but after that running same or any else operation like contenttype get pnpcontenttype web web list documents identity timo doc gives you get pnpcontenttype object reference not set to an instance of an object at line char ntenttype get pnpcontenttype web web list documents identity t categoryinfo notspecified nullreferenceexception fullyqualifiederrorid system nullreferenceexception sharepointpnp powershell commands contenttypes getcontenttype if i do same thing with site url having no whitespaces then everything works fine expected behavior excpected that i can run more than one command after establishing connection actual behavior img width alt pnp bug src steps to reproduce behavior can seen on screenshot above which version of the pnp powershell cmdlets are you using pnp powershell for sharepoint pnp powershell for sharepoint pnp powershell for sharepoint online what is the version of the cmdlet module you are running how did you install the pnp powershell cmdlets msi installed downloaded from github installed through the powershell gallery with install module other means
1
810,062
30,223,799,828
IssuesEvent
2023-07-05 21:45:48
TheExkaliburg/MoreFair
https://api.github.com/repos/TheExkaliburg/MoreFair
closed
Logging in from a new browser creates a Mystery Guest
bug priority:intermediate frontend
Opened the game in incognito mode in Chrome to test something and it just assigned me a new Mystery Guest rather than asking if I want to create a new account.
1.0
Logging in from a new browser creates a Mystery Guest - Opened the game in incognito mode in Chrome to test something and it just assigned me a new Mystery Guest rather than asking if I want to create a new account.
non_process
logging in from a new browser creates a mystery guest opened the game in incognito mode in chrome to test something and it just assigned me a new mystery guest rather than asking if i want to create a new account
0
5,744
8,582,900,835
IssuesEvent
2018-11-13 18:12:49
integr8ly/tutorial-web-app
https://api.github.com/repos/integr8ly/tutorial-web-app
closed
Create a PR Template
Current PSI process
This issue is to track the creation of a PR template that satisfies the requirements of the different levels of code introduced into the webapp (middle, UI), as well as introduce a verification step for UXD. Starting thoughts: - all tests must pass (JS and CSS) - if the code touches the UI, it must be reviewed by UXD - any content changes must be reviewed by Docs, or have a link to a previous document stating the desired changes - the webapp must continue to operate using mock data Currently, there is no defined standard for who should review what type of work, and what each code change must verify before merging. The hope is that a template such as this can assist with creating a smoother process.
1.0
Create a PR Template - This issue is to track the creation of a PR template that satisfies the requirements of the different levels of code introduced into the webapp (middle, UI), as well as introduce a verification step for UXD. Starting thoughts: - all tests must pass (JS and CSS) - if the code touches the UI, it must be reviewed by UXD - any content changes must be reviewed by Docs, or have a link to a previous document stating the desired changes - the webapp must continue to operate using mock data Currently, there is no defined standard for who should review what type of work, and what each code change must verify before merging. The hope is that a template such as this can assist with creating a smoother process.
process
create a pr template this issue is to track the creation of a pr template that satisfies the requirements of the different levels of code introduced into the webapp middle ui as well as introduce a verification step for uxd starting thoughts all tests must pass js and css if the code touches the ui it must be reviewed by uxd any content changes must be reviewed by docs or have a link to a previous document stating the desired changes the webapp must continue to operate using mock data currently there is no defined standard for who should review what type of work and what each code change must verify before merging the hope is that a template such as this can assist with creating a smoother process
1
25,966
12,810,003,024
IssuesEvent
2020-07-03 17:11:08
orientechnologies/orientdb
https://api.github.com/repos/orientechnologies/orientdb
closed
Edge creation progressively slows down during ETL (OrientDB 2.1.4)
out-of-dated performance
When using ETL process to import additional 4.9 million Edges on a single server graph database, the import process starts off around 400-600 vertices/sec, but then progressively gets slower and grinds to almost a halt when it gets to about a million edges inserted. JVM starts using up almost all of the physical memory available. I have tried changing the MAXHEAP and MAXDISKCACHE settings based on recommendation in log file (MAXHEAP=-Xmx387m, MAXDISKCACHE="-Dstorage.diskCache.bufferSize=1519"), but that didn't provide any significant difference. I have also tried changing the ETL query below to use rid instead of the sub selects, but that didn't help either. Current Graph Info: 100k User Vertex 100k AssetType2 Vertex 5mil AssetType1 Vertex 2.5mil Friend Edges (between user vertices) 100k BelongTo Edges (between asset type 2 and user) 100k ActedOn Edges (between asset type 1 and asset type 2) 100k Created Edges (between user and asset type 1) Machine info: Linux x86_64 4GB Mem 30GB disk with 75% free space 2 core CPU 2.4GHZ ETL json file: { "source": { "file": { "path": "edge.csv" } }, "extractor": { "csv": {"separator":" " }}, "transformers" : [ {"command":{"command":"create edge Created from (select from User where id = ${input.source_id}) to (select from AssetType1 where id = ${input.dest_id})", "output": "edge" }} ], "loader" : { "orientdb": { "dbURL": "remote:localhost/mygraphdb", "dbAutoCreate": false, "dbUser": "admin", "dbPassword": "*", "dbType": "graph", "batchCommit": 1000 } } } Is this a configuration issue, a resource constraint issue, or a limitation with Edge creation? Why does the JVM start to eat up all the memory? Your feedback is greatly appreciated. Thanks.
True
Edge creation progressively slows down during ETL (OrientDB 2.1.4) - When using ETL process to import additional 4.9 million Edges on a single server graph database, the import process starts off around 400-600 vertices/sec, but then progressively gets slower and grinds to almost a halt when it gets to about a million edges inserted. JVM starts using up almost all of the physical memory available. I have tried changing the MAXHEAP and MAXDISKCACHE settings based on recommendation in log file (MAXHEAP=-Xmx387m, MAXDISKCACHE="-Dstorage.diskCache.bufferSize=1519"), but that didn't provide any significant difference. I have also tried changing the ETL query below to use rid instead of the sub selects, but that didn't help either. Current Graph Info: 100k User Vertex 100k AssetType2 Vertex 5mil AssetType1 Vertex 2.5mil Friend Edges (between user vertices) 100k BelongTo Edges (between asset type 2 and user) 100k ActedOn Edges (between asset type 1 and asset type 2) 100k Created Edges (between user and asset type 1) Machine info: Linux x86_64 4GB Mem 30GB disk with 75% free space 2 core CPU 2.4GHZ ETL json file: { "source": { "file": { "path": "edge.csv" } }, "extractor": { "csv": {"separator":" " }}, "transformers" : [ {"command":{"command":"create edge Created from (select from User where id = ${input.source_id}) to (select from AssetType1 where id = ${input.dest_id})", "output": "edge" }} ], "loader" : { "orientdb": { "dbURL": "remote:localhost/mygraphdb", "dbAutoCreate": false, "dbUser": "admin", "dbPassword": "*", "dbType": "graph", "batchCommit": 1000 } } } Is this a configuration issue, a resource constraint issue, or a limitation with Edge creation? Why does the JVM start to eat up all the memory? Your feedback is greatly appreciated. Thanks.
non_process
edge creation progressively slows down during etl orientdb when using etl process to import additional million edges on a single server graph database the import process starts off around vertices sec but then progressively gets slower and grinds to almost a halt when it gets to about a million edges inserted jvm starts using up almost all of the physical memory available i have tried changing the maxheap and maxdiskcache settings based on recommendation in log file maxheap maxdiskcache dstorage diskcache buffersize but that didn t provide any significant difference i have also tried changing the etl query below to use rid instead of the sub selects but that didn t help either current graph info user vertex vertex vertex friend edges between user vertices belongto edges between asset type and user actedon edges between asset type and asset type created edges between user and asset type machine info linux mem disk with free space core cpu etl json file source file path edge csv extractor csv separator transformers command command create edge created from select from user where id input source id to select from where id input dest id output edge loader orientdb dburl remote localhost mygraphdb dbautocreate false dbuser admin dbpassword dbtype graph batchcommit is this a configuration issue a resource constraint issue or a limitation with edge creation why does the jvm start to eat up all the memory your feedback is greatly appreciated thanks
0
111,780
14,145,428,936
IssuesEvent
2020-11-10 17:43:53
nextcloud/desktop
https://api.github.com/repos/nextcloud/desktop
closed
Please bring back 'Pause sync for all' quick access
design enhancement feature: :arrows_counterclockwise: sync engine
### Expected behaviour Previous desktop client versions for Windows, right click system tray icon, select 'Pause sync for all'. Nextcloud Windows desktop client repeatedly fires sync failure notifications when certain programs are open (i.e. Quicken/Reckon). The data files of these programs are locked and can't be sync'd. At this point I need to choose 'Pause sync for all' to stop the notifications interrupting until I close the offending program. ### Actual behaviour Pause sync for all is now three clicks away on a sub menu. ### Steps to reproduce 1. As above ### Client configuration Client version: 3.0.1 Operating system: Windows 10 2004 OS language: English (AU)
1.0
Please bring back 'Pause sync for all' quick access - ### Expected behaviour Previous desktop client versions for Windows, right click system tray icon, select 'Pause sync for all'. Nextcloud Windows desktop client repeatedly fires sync failure notifications when certain programs are open (i.e. Quicken/Reckon). The data files of these programs are locked and can't be sync'd. At this point I need to choose 'Pause sync for all' to stop the notifications interrupting until I close the offending program. ### Actual behaviour Pause sync for all is now three clicks away on a sub menu. ### Steps to reproduce 1. As above ### Client configuration Client version: 3.0.1 Operating system: Windows 10 2004 OS language: English (AU)
non_process
please bring back pause sync for all quick access expected behaviour previous desktop client versions for windows right click system tray icon select pause sync for all nextcloud windows desktop client repeatedly fires sync failure notifications when certain programs are open i e quicken reckon the data files of these programs are locked and can t be sync d at this point i need to choose pause sync for all to stop the notifications interrupting until i close the offending program actual behaviour pause sync for all is now three clicks away on a sub menu steps to reproduce as above client configuration client version operating system windows os language english au
0
22,813
3,971,920,245
IssuesEvent
2016-05-04 13:47:10
GCSO2/D3S-INVENTORY
https://api.github.com/repos/GCSO2/D3S-INVENTORY
closed
Cloud: Getting page unresponsive message from browser.
ready for gio to test
This is when I use the RDP session given to me. ![image](https://cloud.githubusercontent.com/assets/5666297/14864696/07f2412e-0c89-11e6-9d00-0e2dcf0f49d6.png)
1.0
Cloud: Getting page unresponsive message from browser. - This is when I use the RDP session given to me. ![image](https://cloud.githubusercontent.com/assets/5666297/14864696/07f2412e-0c89-11e6-9d00-0e2dcf0f49d6.png)
non_process
cloud getting page unresponsive message from browser this is when i use the rdp session given to me
0
316,096
23,614,986,955
IssuesEvent
2022-08-24 15:10:38
bweston26918/portfolio
https://api.github.com/repos/bweston26918/portfolio
closed
03-01 Create wiki
documentation release v0.1.0-alpha
# Create a wiki for the repo GitHub wikis are actually a special type of GitHub repo that just handles the wiki itself. The wiki will outline a more-extensive documentation set. The README.md file in the root level of the repo should not contain extensive information - just a brief synopsis of the repo; All important topics should just have links to the relevant section in the wiki so as to not clutter the repo landing page. *** **Acceptance criteria:** - [x] Enable & create wiki - [x] Pull wiki locally into a special GitHub directory created for this purpose - [x] Test command line connection so wiki may be worked on locally - [ ] Look into platform that may be used to build the wiki locally for testing prior to pushing the wiki to main branch ***
1.0
03-01 Create wiki - # Create a wiki for the repo GitHub wikis are actually a special type of GitHub repo that just handles the wiki itself. The wiki will outline a more-extensive documentation set. The README.md file in the root level of the repo should not contain extensive information - just a brief synopsis of the repo; All important topics should just have links to the relevant section in the wiki so as to not clutter the repo landing page. *** **Acceptance criteria:** - [x] Enable & create wiki - [x] Pull wiki locally into a special GitHub directory created for this purpose - [x] Test command line connection so wiki may be worked on locally - [ ] Look into platform that may be used to build the wiki locally for testing prior to pushing the wiki to main branch ***
non_process
create wiki create a wiki for the repo github wikis are actually a special type of github repo that just handles the wiki itself the wiki will outline a more extensive documentation set the readme md file in the root level of the repo should not contain extensive information just a brief synopsis of the repo all important topics should just have links to the relevant section in the wiki so as to not clutter the repo landing page acceptance criteria enable create wiki pull wiki locally into a special github directory created for this purpose test command line connection so wiki may be worked on locally look into platform that may be used to build the wiki locally for testing prior to pushing the wiki to main branch
0
17,727
23,627,604,286
IssuesEvent
2022-08-25 06:20:53
benthosdev/benthos
https://api.github.com/repos/benthosdev/benthos
closed
How to insert elements from an array of structures into the database?
question processors outputs bloblang
I want to insert each element of struct array in root into clickhouse my root value: ``` root ={ "items":[ { "header":{ "messageType":"DML", "schemaName":"DataCount", "sourceType":[ "MYSQL" ], "tableName":"pjy_tb", "timestamp":1660295266 } } ], "items_ckinfo":[ "{\"messageType\":\"DML\",\"schemaName\":\"pjyschema\",\"tableName\":\"pjytb\"}", "{\"messageType\":\"DML\",\"schemaName\":\"pjyschema2\",\"tableName\":\"pjytb2\"}" ] } ``` I want to loop or batch insert the elements of items_ckinfo in root into clickhouse. But I don't know how to write in output. I can't seem to find an appropriate tag in the documentation. Can you give a simple example?
1.0
How to insert elements from an array of structures into the database? - I want to insert each element of struct array in root into clickhouse my root value: ``` root ={ "items":[ { "header":{ "messageType":"DML", "schemaName":"DataCount", "sourceType":[ "MYSQL" ], "tableName":"pjy_tb", "timestamp":1660295266 } } ], "items_ckinfo":[ "{\"messageType\":\"DML\",\"schemaName\":\"pjyschema\",\"tableName\":\"pjytb\"}", "{\"messageType\":\"DML\",\"schemaName\":\"pjyschema2\",\"tableName\":\"pjytb2\"}" ] } ``` I want to loop or batch insert the elements of items_ckinfo in root into clickhouse. But I don't know how to write in output. I can't seem to find an appropriate tag in the documentation. Can you give a simple example?
process
how to insert elements from an array of structures into the database? i want to insert each element of struct array in root into clickhouse my root value root items header messagetype dml schemaname datacount sourcetype mysql tablename pjy tb timestamp items ckinfo messagetype dml schemaname pjyschema tablename pjytb messagetype dml schemaname tablename i want to loop or batch insert the elements of items ckinfo in root into clickhouse but i don t know how to write in output i can t seem to find an appropriate tag in the documentation can you give a simple example
1
9,638
12,602,295,375
IssuesEvent
2020-06-11 11:27:15
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
ENV var is required to execute `generate`/`validate`
bug/2-confirmed kind/bug process/candidate team/engines
# Bug description ![image](https://user-images.githubusercontent.com/183673/84019920-4499e380-a982-11ea-9ec5-60270a3f3f4b.png) ![image](https://user-images.githubusercontent.com/183673/84021950-cccdb800-a985-11ea-9b72-2c9d285ad4d5.png) ## How to reproduce Run `prisma generate` on this schema: ```prisma generator client { provider = "prisma-client-js" } datasource db { provider = "mysql" url = env("DATABASE_URL") } // ### Product ### model Product { id Int @default(autoincrement()) @id name String @unique description String? developerId Int developer ProductDeveloper @relation(fields: [developerId], references: [id]) forumId Int? forum Forum? @relation(fields: [forumId], references: [id]) type ProductType tags ProductTag[] salesPages ProductSalesPage[] productVersions ProductVersions[] productPictures ProductPictures[] Customer Customer? @relation(fields: [customerId], references: [id]) customerId Int? Invoice Invoice? @relation(fields: [invoiceId], references: [id]) invoiceId Int? owner Seller @relation(fields: [ownerId], references: [id]) ownerId Int } enum ProductType { Plugins_WordPress Temas_WordPress } model ProductDeveloper { id Int @default(autoincrement()) @id name String @unique products Product[] } model ProductTag { id Int @default(autoincrement()) @id name String @unique translation String products Product[] @relation(references: [id]) } model ProductSalesPage { id Int @default(autoincrement()) @id url String @unique productId Int product Product @relation(fields: [productId], references: [id]) } model ProductVersions { id Int @default(autoincrement()) @id provider Provider isCompacted Boolean @default(false) version String lastUpdated String? isUnlocked Boolean @default(false) warning String? createdAt DateTime @default(now()) downloadHost downloadHost downloadHostAccount Int downloadPath String productId Int product Product @relation(fields: [productId], references: [id]) } model ProductPictures { id Int @default(autoincrement()) @id pictureUrl String @unique productId Int product Product @relation(fields: [productId], references: [id]) } // ### User ### model User { id Int @default(autoincrement()) @id name String email String @unique login String @unique balance Float @default(0.00) isSuspended Boolean @default(false) suspendedReason String? createdAt DateTime @default(now()) tags UserTag[] @relation(references: [id]) accounts UserPlatformAccount[] customer Customer? seller Seller? } model UserPlatformAccount { id Int @default(autoincrement()) @id nickname String? platformUserId String platform SupportedPlatforms details Json userId Int user User @relation(fields: [id], references: [id]) } model UserTag { id Int @default(autoincrement()) @id value String @unique users User[] @relation(references: [id]) } // ### Customer ### model Customer { id Int @default(autoincrement()) @id birthday DateTime? cpf String? @unique gender Gender? createdAt DateTime @default(now()) invoices Invoice[] product Product[] userId Int user User @relation(fields: [userId], references: [id]) // Criar um resolver na API GraphQL para pegar os produtos de acordo com as // invoices que estão pagas. } // ### Seller ### model Seller { id Int @default(autoincrement()) @id role SellerRole cpf String @unique createdAt DateTime @default(now()) invoices Invoice[] product Product[] userId Int user User @relation(fields: [userId], references: [id]) } enum SellerRole { ADMIN RESELLER } // Global model and enums model Forum { id Int @default(autoincrement()) @id name String permalink String @unique parentId Int? parent Forum? @relation(fields: [parentId], references: [id]) Forum Forum[] @relation("ForumToForum") product Product? } model Host { id Int @default(autoincrement()) @id host downloadHost accountId String } model Invoice { id Int @default(autoincrement()) @id status InvoiceStatus transactions PaymentTransaction[] purchaseDate DateTime products Product[] supportExpiration DateTime customerId Int customer Customer @relation(fields: [customerId], references: [id]) sellerId Int seller Seller @relation(fields: [sellerId], references: [id]) } enum InvoiceStatus { PAID OVERDUE REFUNDED CANCELLED } model PaymentTransaction { id Int @default(autoincrement()) @id transactionID Int @unique platform PaymentTransactionPlatform date DateTime details Json value Float invoiceId Int invoice Invoice @relation(fields: [invoiceId], references: [id]) } enum PaymentTransactionPlatform { MERCADOPAGO } enum Gender { M F } enum Provider { GPLZONE ULTRAPACKV2 } enum downloadHost { DROPBOX } enum SupportedPlatforms { FORUM MERCADOLIBRE } ``` ## Expected behavior I can generate a Prisma Client. > ## Environment & setup <!-- In which environment does the problem occur --> - OS: Windows - Database: - - Prisma version: ``` λ prisma -v Environment variables loaded from ./prisma/.env @prisma/cli : 2.0.0-beta.8 Current platform : windows Query Engine : query-engine 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\query-engine-windows.exe) Migration Engine : migration-engine-cli 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\migration-engine-windows.exe) Introspection Engine : introspection-core 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\introspection-engine-windows.exe) Format Binary : prisma-fmt 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\prisma-fmt-windows.exe) ``` - Node.js version: v12.6.0
1.0
ENV var is required to execute `generate`/`validate` - # Bug description ![image](https://user-images.githubusercontent.com/183673/84019920-4499e380-a982-11ea-9ec5-60270a3f3f4b.png) ![image](https://user-images.githubusercontent.com/183673/84021950-cccdb800-a985-11ea-9b72-2c9d285ad4d5.png) ## How to reproduce Run `prisma generate` on this schema: ```prisma generator client { provider = "prisma-client-js" } datasource db { provider = "mysql" url = env("DATABASE_URL") } // ### Product ### model Product { id Int @default(autoincrement()) @id name String @unique description String? developerId Int developer ProductDeveloper @relation(fields: [developerId], references: [id]) forumId Int? forum Forum? @relation(fields: [forumId], references: [id]) type ProductType tags ProductTag[] salesPages ProductSalesPage[] productVersions ProductVersions[] productPictures ProductPictures[] Customer Customer? @relation(fields: [customerId], references: [id]) customerId Int? Invoice Invoice? @relation(fields: [invoiceId], references: [id]) invoiceId Int? owner Seller @relation(fields: [ownerId], references: [id]) ownerId Int } enum ProductType { Plugins_WordPress Temas_WordPress } model ProductDeveloper { id Int @default(autoincrement()) @id name String @unique products Product[] } model ProductTag { id Int @default(autoincrement()) @id name String @unique translation String products Product[] @relation(references: [id]) } model ProductSalesPage { id Int @default(autoincrement()) @id url String @unique productId Int product Product @relation(fields: [productId], references: [id]) } model ProductVersions { id Int @default(autoincrement()) @id provider Provider isCompacted Boolean @default(false) version String lastUpdated String? isUnlocked Boolean @default(false) warning String? createdAt DateTime @default(now()) downloadHost downloadHost downloadHostAccount Int downloadPath String productId Int product Product @relation(fields: [productId], references: [id]) } model ProductPictures { id Int @default(autoincrement()) @id pictureUrl String @unique productId Int product Product @relation(fields: [productId], references: [id]) } // ### User ### model User { id Int @default(autoincrement()) @id name String email String @unique login String @unique balance Float @default(0.00) isSuspended Boolean @default(false) suspendedReason String? createdAt DateTime @default(now()) tags UserTag[] @relation(references: [id]) accounts UserPlatformAccount[] customer Customer? seller Seller? } model UserPlatformAccount { id Int @default(autoincrement()) @id nickname String? platformUserId String platform SupportedPlatforms details Json userId Int user User @relation(fields: [id], references: [id]) } model UserTag { id Int @default(autoincrement()) @id value String @unique users User[] @relation(references: [id]) } // ### Customer ### model Customer { id Int @default(autoincrement()) @id birthday DateTime? cpf String? @unique gender Gender? createdAt DateTime @default(now()) invoices Invoice[] product Product[] userId Int user User @relation(fields: [userId], references: [id]) // Criar um resolver na API GraphQL para pegar os produtos de acordo com as // invoices que estão pagas. } // ### Seller ### model Seller { id Int @default(autoincrement()) @id role SellerRole cpf String @unique createdAt DateTime @default(now()) invoices Invoice[] product Product[] userId Int user User @relation(fields: [userId], references: [id]) } enum SellerRole { ADMIN RESELLER } // Global model and enums model Forum { id Int @default(autoincrement()) @id name String permalink String @unique parentId Int? parent Forum? @relation(fields: [parentId], references: [id]) Forum Forum[] @relation("ForumToForum") product Product? } model Host { id Int @default(autoincrement()) @id host downloadHost accountId String } model Invoice { id Int @default(autoincrement()) @id status InvoiceStatus transactions PaymentTransaction[] purchaseDate DateTime products Product[] supportExpiration DateTime customerId Int customer Customer @relation(fields: [customerId], references: [id]) sellerId Int seller Seller @relation(fields: [sellerId], references: [id]) } enum InvoiceStatus { PAID OVERDUE REFUNDED CANCELLED } model PaymentTransaction { id Int @default(autoincrement()) @id transactionID Int @unique platform PaymentTransactionPlatform date DateTime details Json value Float invoiceId Int invoice Invoice @relation(fields: [invoiceId], references: [id]) } enum PaymentTransactionPlatform { MERCADOPAGO } enum Gender { M F } enum Provider { GPLZONE ULTRAPACKV2 } enum downloadHost { DROPBOX } enum SupportedPlatforms { FORUM MERCADOLIBRE } ``` ## Expected behavior I can generate a Prisma Client. > ## Environment & setup <!-- In which environment does the problem occur --> - OS: Windows - Database: - - Prisma version: ``` λ prisma -v Environment variables loaded from ./prisma/.env @prisma/cli : 2.0.0-beta.8 Current platform : windows Query Engine : query-engine 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\query-engine-windows.exe) Migration Engine : migration-engine-cli 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\migration-engine-windows.exe) Introspection Engine : introspection-core 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\introspection-engine-windows.exe) Format Binary : prisma-fmt 12105ee25e61860c17a381ffdfcf2b2513b0ba92 (at C:\ProgramData\nvm\v12.6.0\node_modules\@prisma\cli\prisma-fmt-windows.exe) ``` - Node.js version: v12.6.0
process
env var is required to execute generate validate bug description how to reproduce run prisma generate on this schema prisma generator client provider prisma client js datasource db provider mysql url env database url product model product id int default autoincrement id name string unique description string developerid int developer productdeveloper relation fields references forumid int forum forum relation fields references type producttype tags producttag salespages productsalespage productversions productversions productpictures productpictures customer customer relation fields references customerid int invoice invoice relation fields references invoiceid int owner seller relation fields references ownerid int enum producttype plugins wordpress temas wordpress model productdeveloper id int default autoincrement id name string unique products product model producttag id int default autoincrement id name string unique translation string products product relation references model productsalespage id int default autoincrement id url string unique productid int product product relation fields references model productversions id int default autoincrement id provider provider iscompacted boolean default false version string lastupdated string isunlocked boolean default false warning string createdat datetime default now downloadhost downloadhost downloadhostaccount int downloadpath string productid int product product relation fields references model productpictures id int default autoincrement id pictureurl string unique productid int product product relation fields references user model user id int default autoincrement id name string email string unique login string unique balance float default issuspended boolean default false suspendedreason string createdat datetime default now tags usertag relation references accounts userplatformaccount customer customer seller seller model userplatformaccount id int default autoincrement id nickname string platformuserid string platform supportedplatforms details json userid int user user relation fields references model usertag id int default autoincrement id value string unique users user relation references customer model customer id int default autoincrement id birthday datetime cpf string unique gender gender createdat datetime default now invoices invoice product product userid int user user relation fields references criar um resolver na api graphql para pegar os produtos de acordo com as invoices que estão pagas seller model seller id int default autoincrement id role sellerrole cpf string unique createdat datetime default now invoices invoice product product userid int user user relation fields references enum sellerrole admin reseller global model and enums model forum id int default autoincrement id name string permalink string unique parentid int parent forum relation fields references forum forum relation forumtoforum product product model host id int default autoincrement id host downloadhost accountid string model invoice id int default autoincrement id status invoicestatus transactions paymenttransaction purchasedate datetime products product supportexpiration datetime customerid int customer customer relation fields references sellerid int seller seller relation fields references enum invoicestatus paid overdue refunded cancelled model paymenttransaction id int default autoincrement id transactionid int unique platform paymenttransactionplatform date datetime details json value float invoiceid int invoice invoice relation fields references enum paymenttransactionplatform mercadopago enum gender m f enum provider gplzone enum downloadhost dropbox enum supportedplatforms forum mercadolibre expected behavior i can generate a prisma client environment setup os windows database prisma version λ prisma v environment variables loaded from prisma env prisma cli beta current platform windows query engine query engine at c programdata nvm node modules prisma cli query engine windows exe migration engine migration engine cli at c programdata nvm node modules prisma cli migration engine windows exe introspection engine introspection core at c programdata nvm node modules prisma cli introspection engine windows exe format binary prisma fmt at c programdata nvm node modules prisma cli prisma fmt windows exe node js version
1
2,559
2,528,585,999
IssuesEvent
2015-01-22 05:16:51
AtlasOfLivingAustralia/biocache-hubs
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-hubs
closed
Typo at "Original versus Processed" Comparison Table
bug priority-low
On the "Original versus Processed" Comparison Table in occurrence view, Class is misspelled as *Classs* .
1.0
Typo at "Original versus Processed" Comparison Table - On the "Original versus Processed" Comparison Table in occurrence view, Class is misspelled as *Classs* .
non_process
typo at original versus processed comparison table on the original versus processed comparison table in occurrence view class is misspelled as classs
0
8,256
11,423,555,812
IssuesEvent
2020-02-03 16:06:11
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
NTR: negative regulation of cytolysis by virus of host cell
New term request multi-species process
A parent term to the NTR is: Parent Term: GO:0001899 - negative regulation of cytolysis by symbiont of host cells Definition: Any process in which an organism stops, prevents, or reduces the frequency, rate or extent of cytolysis by that organism of cells in its host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction NTR: negative regulation of cytolysis by virus of host cells Definition: Any process in which a virus stops, prevents, or reduces the frequency, rate or extent of cytolysis by that virus of cells in its host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction. Ontology: Biological Process Synonyms: inhibition by virus of cytolysis of host cells downregulation by virus of cytolysis of host cells down-regulation by virus of cytolysis of host cells down regulation by virus of cytolysis of host cells negative regulation by virus of cytolysis of host cells Reference: PMID:14527655 Figure 1. in this paper shows loss of the T4 virus-encoded "rI protein" results in T4 inducing cytolysis at a much earlier time. This requested go term would be more accurate for describing this process Reported by: notjustin43 Original Ticket: [geneontology/ontology-requests/11336](https://sourceforge.net/p/geneontology/ontology-requests/11336)
1.0
NTR: negative regulation of cytolysis by virus of host cell - A parent term to the NTR is: Parent Term: GO:0001899 - negative regulation of cytolysis by symbiont of host cells Definition: Any process in which an organism stops, prevents, or reduces the frequency, rate or extent of cytolysis by that organism of cells in its host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction NTR: negative regulation of cytolysis by virus of host cells Definition: Any process in which a virus stops, prevents, or reduces the frequency, rate or extent of cytolysis by that virus of cells in its host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction. Ontology: Biological Process Synonyms: inhibition by virus of cytolysis of host cells downregulation by virus of cytolysis of host cells down-regulation by virus of cytolysis of host cells down regulation by virus of cytolysis of host cells negative regulation by virus of cytolysis of host cells Reference: PMID:14527655 Figure 1. in this paper shows loss of the T4 virus-encoded "rI protein" results in T4 inducing cytolysis at a much earlier time. This requested go term would be more accurate for describing this process Reported by: notjustin43 Original Ticket: [geneontology/ontology-requests/11336](https://sourceforge.net/p/geneontology/ontology-requests/11336)
process
ntr negative regulation of cytolysis by virus of host cell a parent term to the ntr is parent term go negative regulation of cytolysis by symbiont of host cells definition any process in which an organism stops prevents or reduces the frequency rate or extent of cytolysis by that organism of cells in its host organism the host is defined as the larger of the organisms involved in a symbiotic interaction ntr negative regulation of cytolysis by virus of host cells definition any process in which a virus stops prevents or reduces the frequency rate or extent of cytolysis by that virus of cells in its host organism the host is defined as the larger of the organisms involved in a symbiotic interaction ontology biological process synonyms inhibition by virus of cytolysis of host cells downregulation by virus of cytolysis of host cells down regulation by virus of cytolysis of host cells down regulation by virus of cytolysis of host cells negative regulation by virus of cytolysis of host cells reference pmid figure in this paper shows loss of the virus encoded ri protein results in inducing cytolysis at a much earlier time this requested go term would be more accurate for describing this process reported by original ticket
1
295
3,073,970,941
IssuesEvent
2015-08-20 02:21:50
tgstation/-tg-station
https://api.github.com/repos/tgstation/-tg-station
closed
List of places in code not using qdel
Maintainability - Hinders improvements Not a bug
./ATMOSPHERICS/pipes/pipes.dm-14-/obj/machinery/atmospherics/proc/pipeline_expansion() ./ATMOSPHERICS/pipes/pipes.dm-15- return null ./ATMOSPHERICS/pipes/pipes.dm-16- ./ATMOSPHERICS/pipes/pipes.dm-17-/obj/machinery/atmospherics/pipe/proc/check_pressure(pressure) ./ATMOSPHERICS/pipes/pipes.dm-18- //Return 1 if parent should continue checking other pipes ./ATMOSPHERICS/pipes/pipes.dm:19: //Return null if parent should stop checking other pipes. Recall: ***del(src)*** will by default return null ./ATMOSPHERICS/pipes/pipes.dm-20- return 1 ./ATMOSPHERICS/pipes/pipes.dm-21- ./ATMOSPHERICS/pipes/pipes.dm-22-/obj/machinery/atmospherics/pipe/proc/releaseAirToTurf() ./ATMOSPHERICS/pipes/pipes.dm-23- if(air_temporary) ./ATMOSPHERICS/pipes/pipes.dm-24- var/turf/T = loc *** ./controllers/configuration.dm-176- modes += M.config_tag ./controllers/configuration.dm-177- mode_names[M.config_tag] = M.name ./controllers/configuration.dm-178- probabilities[M.config_tag] = M.probability ./controllers/configuration.dm-179- if(M.votable) ./controllers/configuration.dm-180- votable_modes += M.config_tag ./controllers/configuration.dm:181: ***del(M)*** ./controllers/configuration.dm-182- votable_modes += "secret" ./controllers/configuration.dm-183- ./controllers/configuration.dm-184-/datum/configuration/proc/load(filename, type = "config") //the type can also be game_options, in which case it uses a different switch. not making it separate to not copypaste code - Urist ./controllers/configuration.dm-185- var/list/Lines = file2list(filename) ./controllers/configuration.dm-186- *** ./controllers/configuration.dm-565- // their information, but it is the only way (at least that I know of). ./controllers/configuration.dm-566- for(var/T in (typesof(/datum/game_mode) - /datum/game_mode)) ./controllers/configuration.dm-567- var/datum/game_mode/M = new T() ./controllers/configuration.dm-568- if(M.config_tag && M.config_tag == mode_name) ./controllers/configuration.dm-569- return M ./controllers/configuration.dm:570: ***del(M)*** ./controllers/configuration.dm-571- return new /datum/game_mode/extended() ./controllers/configuration.dm-572- ./controllers/configuration.dm-573-/datum/configuration/proc/get_runnable_modes() ./controllers/configuration.dm-574- var/list/datum/game_mode/runnable_modes = new ./controllers/configuration.dm-575- for(var/T in (typesof(/datum/game_mode) - /datum/game_mode)) ./controllers/configuration.dm-576- var/datum/game_mode/M = new T() ./controllers/configuration.dm-577- //world << "DEBUG: [T], tag=[M.config_tag], prob=[probabilities[M.config_tag]]" ./controllers/configuration.dm-578- if(!(M.config_tag in modes)) ./controllers/configuration.dm:579: ***del(M)*** ./controllers/configuration.dm-580- continue ./controllers/configuration.dm-581- if(probabilities[M.config_tag]<=0) ./controllers/configuration.dm:582: ***del(M)*** ./controllers/configuration.dm-583- continue ./controllers/configuration.dm-584- if(M.can_start()) ./controllers/configuration.dm-585- runnable_modes[M] = probabilities[M.config_tag] ./controllers/configuration.dm-586- //world << "DEBUG: runnable_mode\[[runnable_modes.len]\] = [M.config_tag]" ./controllers/configuration.dm-587- return runnable_modes *** ./controllers/subsystem/garbage.dm-61- if(GCd_at_time > time_to_kill) ./controllers/subsystem/garbage.dm-62- break // Everything else is newer, skip them ./controllers/subsystem/garbage.dm-63- ./controllers/subsystem/garbage.dm-64- var/atom/A ./controllers/subsystem/garbage.dm-65- if (!istext(refID)) ./controllers/subsystem/garbage.dm:66: ***del(A)*** ./controllers/subsystem/garbage.dm-67- else ./controllers/subsystem/garbage.dm-68- A = locate(refID) ./controllers/subsystem/garbage.dm-69- if (A && A.gc_destroyed == GCd_at_time) // So if something else coincidently gets the same ref, it's not deleted by mistake ./controllers/subsystem/garbage.dm-70- // Something's still referring to the qdel'd object. Kill it. ./controllers/subsystem/garbage.dm-71- testing("GC: -- \ref[A] | [A.type] was unable to be GC'd and was deleted --") ./controllers/subsystem/garbage.dm-72- didntgc["[A.type]"]++ ./controllers/subsystem/garbage.dm:73: ***del(A)*** ./controllers/subsystem/garbage.dm-74- ++delslasttick ./controllers/subsystem/garbage.dm-75- ++totaldels ./controllers/subsystem/garbage.dm-76- else ./controllers/subsystem/garbage.dm-77- ++gcedlasttick ./controllers/subsystem/garbage.dm-78- ++totalgcs *** ./controllers/subsystem/garbage.dm-100-/proc/qdel(var/datum/A) ./controllers/subsystem/garbage.dm-101- if (!A) ./controllers/subsystem/garbage.dm-102- return ./controllers/subsystem/garbage.dm-103- if (!istype(A)) ./controllers/subsystem/garbage.dm-104- //warning("qdel() passed object of type [A.type]. qdel() can only handle /datum types.") ./controllers/subsystem/garbage.dm:105: ***del(A)*** ./controllers/subsystem/garbage.dm-106- else if (isnull(A.gc_destroyed)) ./controllers/subsystem/garbage.dm-107- // Let our friend know they're about to get fucked up. ./controllers/subsystem/garbage.dm-108- var/hint = A.Destroy() ./controllers/subsystem/garbage.dm-109- if (!A) ./controllers/subsystem/garbage.dm-110- return *** ./controllers/subsystem/garbage.dm-116- if (QDEL_HINT_IWILLGC) //functionally the same as the above. qdel should assume the object will gc on its own, and not check it. ./controllers/subsystem/garbage.dm-117- return ./controllers/subsystem/garbage.dm-118- if (QDEL_HINT_HARDDEL) //qdel should assume this object won't gc, and queue a hard delete using a hard reference to save time from the locate() ./controllers/subsystem/garbage.dm-119- SSgarbage.HardQueue(A) ./controllers/subsystem/garbage.dm-120- if (QDEL_HINT_HARDDEL_NOW) //qdel should assume this object won't gc, and hard del it post haste. ./controllers/subsystem/garbage.dm:121: ***del(A)*** ./controllers/subsystem/garbage.dm-122- if (QDEL_HINT_PUTINPOOL) //qdel will put this object in the pool. ./controllers/subsystem/garbage.dm-123- PlaceInPool(A,0) ./controllers/subsystem/garbage.dm-124- else ./controllers/subsystem/garbage.dm-125- SSgarbage.Queue(A) ./controllers/subsystem/garbage.dm-126- *** ./controllers/subsystem/garbage.dm-134- ./controllers/subsystem/garbage.dm-135-// Default implementation of clean-up code. ./controllers/subsystem/garbage.dm-136-// This should be overridden to remove all references pointing to the object being destroyed. ./controllers/subsystem/garbage.dm-137-// Return true if the the GC controller should allow the object to continue existing. (Useful if pooling objects.) ./controllers/subsystem/garbage.dm-138-/datum/proc/Destroy() ./controllers/subsystem/garbage.dm:139: //***del(src)*** ./controllers/subsystem/garbage.dm-140- tag = null ./controllers/subsystem/garbage.dm-141- return QDEL_HINT_QUEUE ./controllers/subsystem/garbage.dm-142- ./controllers/subsystem/garbage.dm-143-/datum/var/gc_destroyed //Time when this object was destroyed. ./controllers/subsystem/garbage.dm-144- *** ./controllers/subsystem/garbage.dm-192- set category = "Debug" ./controllers/subsystem/garbage.dm-193- if(garbage) ./controllers/subsystem/garbage.dm-194- while(garbage.destroyed.len) ./controllers/subsystem/garbage.dm-195- var/datum/o = locate(garbage.destroyed[1]) ./controllers/subsystem/garbage.dm-196- if(istype(o) && o.gc_destroyed) ./controllers/subsystem/garbage.dm:197: ***del(o)*** ./controllers/subsystem/garbage.dm-198- garbage.dels++ ./controllers/subsystem/garbage.dm-199- garbage.destroyed.Cut(1, 2) ./controllers/subsystem/garbage.dm-200-#endif *** ./controllers/subsystem/radio.dm-27- ./controllers/subsystem/radio.dm-28- if(frequency) ./controllers/subsystem/radio.dm-29- frequency.remove_listener(device) ./controllers/subsystem/radio.dm-30- ./controllers/subsystem/radio.dm-31- if(frequency.devices.len == 0) ./controllers/subsystem/radio.dm:32: ***del(frequency)*** ./controllers/subsystem/radio.dm-33- frequencies -= f_text ./controllers/subsystem/radio.dm-34- ./controllers/subsystem/radio.dm-35- return 1 ./controllers/subsystem/radio.dm-36- ./controllers/subsystem/radio.dm-37-/datum/subsystem/radio/proc/return_frequency(new_frequency as num) *** ./controllers/subsystem/server_maintenance.dm-9- for(var/client/C in clients) ./controllers/subsystem/server_maintenance.dm-10- if(C.is_afk(INACTIVITY_KICK)) ./controllers/subsystem/server_maintenance.dm-11- if(!istype(C.mob, /mob/dead)) ./controllers/subsystem/server_maintenance.dm-12- log_access("AFK: [key_name(C)]") ./controllers/subsystem/server_maintenance.dm-13- C << "<span class='danger'>You have been inactive for more than 10 minutes and have been disconnected.</span>" ./controllers/subsystem/server_maintenance.dm:14: ***del(C)*** ./controllers/subsystem/server_maintenance.dm-15- ./controllers/subsystem/server_maintenance.dm-16- if(config.sql_enabled) ./controllers/subsystem/server_maintenance.dm-17- sql_poll_players() ./controllers/subsystem/server_maintenance.dm-18- sql_poll_admins() *** ./controllers/subsystem/ticker.dm-135- ./controllers/subsystem/ticker.dm-136- else ./controllers/subsystem/ticker.dm-137- mode = config.pick_mode(master_mode) ./controllers/subsystem/ticker.dm-138- if(!mode.can_start()) ./controllers/subsystem/ticker.dm-139- world << "<B>Unable to start [mode.name].</B> Not enough players, [mode.required_players] players and [mode.required_enemies] eligible antagonists needed. Reverting to pre-game lobby." ./controllers/subsystem/ticker.dm:140: ***del(mode)*** ./controllers/subsystem/ticker.dm-141- SSjob.ResetOccupations() ./controllers/subsystem/ticker.dm-142- return 0 ./controllers/subsystem/ticker.dm-143- ./controllers/subsystem/ticker.dm-144- //Configure mode and assign player to special mode stuff ./controllers/subsystem/ticker.dm-145- var/can_continue = 0 ./controllers/subsystem/ticker.dm-146- can_continue = src.mode.pre_setup() //Choose antagonists ./controllers/subsystem/ticker.dm-147- SSjob.DivideOccupations() //Distribute jobs ./controllers/subsystem/ticker.dm-148- ./controllers/subsystem/ticker.dm-149- if(!Debug2) ./controllers/subsystem/ticker.dm-150- if(!can_continue) ./controllers/subsystem/ticker.dm:151: ***del(mode)*** ./controllers/subsystem/ticker.dm-152- world << "<B>Error setting up [master_mode].</B> Reverting to pre-game lobby." ./controllers/subsystem/ticker.dm-153- SSjob.ResetOccupations() ./controllers/subsystem/ticker.dm-154- return 0 ./controllers/subsystem/ticker.dm-155- else ./controllers/subsystem/ticker.dm-156- world << "<span class='notice'>DEBUG: Bypassing prestart checks..." *** ./controllers/subsystem/ticker.dm-287- flick("intro_nuke",cinematic) ./controllers/subsystem/ticker.dm-288- sleep(35) ./controllers/subsystem/ticker.dm-289- flick("station_intact",cinematic) ./controllers/subsystem/ticker.dm-290- world << sound('sound/ambience/signal.ogg') ./controllers/subsystem/ticker.dm-291- sleep(100) ./controllers/subsystem/ticker.dm:292: if(cinematic) ***del(cinematic)*** ./controllers/subsystem/ticker.dm:293: if(temp_buckle) ***del(temp_buckle)*** ./controllers/subsystem/ticker.dm-294- return //Faster exit, since nothing happened ./controllers/subsystem/ticker.dm-295- else //Station nuked (nuke,explosion,summary) ./controllers/subsystem/ticker.dm-296- flick("intro_nuke",cinematic) ./controllers/subsystem/ticker.dm-297- sleep(35) ./controllers/subsystem/ticker.dm-298- flick("station_explode_fade_red", cinematic) *** ./datums/ai_laws.dm-132- ./datums/ai_laws.dm-133-/datum/ai_laws/proc/add_ion_law(law) ./datums/ai_laws.dm-134- src.ion += law ./datums/ai_laws.dm-135- ./datums/ai_laws.dm-136-/datum/ai_laws/proc/clear_inherent_laws() ./datums/ai_laws.dm:137: ***del(src.inherent)*** ./datums/ai_laws.dm-138- src.inherent = list() ./datums/ai_laws.dm-139- ./datums/ai_laws.dm-140-/datum/ai_laws/proc/add_supplied_law(number, law) ./datums/ai_laws.dm-141- while (src.supplied.len < number + 1) ./datums/ai_laws.dm-142- src.supplied += "" *** ./datums/diseases/advance/advance.dm-105- if(affected_mob) ./datums/diseases/advance/advance.dm-106- var/id = "[GetDiseaseID()]" ./datums/diseases/advance/advance.dm-107- if(resistance && !(id in affected_mob.resistances)) ./datums/diseases/advance/advance.dm-108- affected_mob.resistances[id] = id ./datums/diseases/advance/advance.dm-109- remove_virus() ./datums/diseases/advance/advance.dm:110: ***del(src)*** //delete the datum to stop it processing ./datums/diseases/advance/advance.dm-111- ./datums/diseases/advance/advance.dm-112-// Returns the advance disease with a different reference memory. ./datums/diseases/advance/advance.dm-113-/datum/disease/advance/Copy(process = 0) ./datums/diseases/advance/advance.dm-114- return new /datum/disease/advance(process, src, 1) ./datums/diseases/advance/advance.dm-115- *** ./datums/diseases/_disease.dm-136- ./datums/diseases/_disease.dm-137- if(affected_mob) ./datums/diseases/_disease.dm-138- for(var/datum/disease/D in affected_mob.viruses) ./datums/diseases/_disease.dm-139- if(D != src) ./datums/diseases/_disease.dm-140- if(IsSame(D)) ./datums/diseases/_disease.dm:141: ***del(D)*** ./datums/diseases/_disease.dm-142- ./datums/diseases/_disease.dm-143- if(holder == affected_mob) ./datums/diseases/_disease.dm-144- if(affected_mob.stat != DEAD) ./datums/diseases/_disease.dm-145- stage_act() ./datums/diseases/_disease.dm-146- *** ./datums/diseases/_disease.dm-154- if(affected_mob) ./datums/diseases/_disease.dm-155- if(disease_flags & CAN_RESIST) ./datums/diseases/_disease.dm-156- if(!(type in affected_mob.resistances)) ./datums/diseases/_disease.dm-157- affected_mob.resistances += type ./datums/diseases/_disease.dm-158- remove_virus() ./datums/diseases/_disease.dm:159: ***del(src)*** ./datums/diseases/_disease.dm-160- ./datums/diseases/_disease.dm-161- ./datums/diseases/_disease.dm-162-/datum/disease/New() ./datums/diseases/_disease.dm-163- if(required_organs && required_organs.len) ./datums/diseases/_disease.dm-164- if(ishuman(affected_mob)) *** ./datums/gas_mixture.dm-319- if(!corresponding) ./datums/gas_mixture.dm-320- corresponding = new trace_gas.type() ./datums/gas_mixture.dm-321- trace_gases += corresponding ./datums/gas_mixture.dm-322- corresponding.moles += trace_gas.moles ./datums/gas_mixture.dm-323- ./datums/gas_mixture.dm:324:// ***del(giver)*** ./datums/gas_mixture.dm-325- return 1 ./datums/gas_mixture.dm-326- ./datums/gas_mixture.dm-327-/datum/gas_mixture/remove(amount) ./datums/gas_mixture.dm-328- ./datums/gas_mixture.dm-329- var/sum = total_moles() *** ./datums/modules.dm-22- var/type = O.type // the type of the creating object ./datums/modules.dm-23- ./datums/modules.dm-24- var/mneed = mods.inmodlist(type) // find if this type has modules defined ./datums/modules.dm-25- ./datums/modules.dm-26- if(!mneed) // not found in module list? ./datums/modules.dm:27: ***del(src)*** // delete self, thus ending proc ./datums/modules.dm-28- ./datums/modules.dm-29- var/needed = mods.getbitmask(type) // get a bitmask for the number of modules in this object ./datums/modules.dm-30- status = needed ./datums/modules.dm-31- installed = needed ./datums/modules.dm-32- *** ./game/area/areas.dm-389- if(!AM.anchored) ./game/area/areas.dm-390- AM.Move(D, SOUTH) ./game/area/areas.dm-391- else ./game/area/areas.dm-392- qdel(AM) ./game/area/areas.dm-393- if(istype(T, /turf/simulated)) ./game/area/areas.dm:394: ***del(T)*** ./game/area/areas.dm-395- ./game/area/areas.dm-396- /*for(var/atom/movable/bug in src) // If someone (or something) is somehow still in the shuttle's docking area... ./game/area/areas.dm-397- if(ismob(bug)) ./game/area/areas.dm-398- continue ./game/area/areas.dm-399- qdel(bug)*/ *** ./game/atoms.dm-57- ./game/atoms.dm-58-/atom/proc/CheckParts() ./game/atoms.dm-59- return ./game/atoms.dm-60- ./game/atoms.dm-61-/atom/proc/assume_air(datum/gas_mixture/giver) ./game/atoms.dm:62: ***del(giver)*** ./game/atoms.dm-63- return null ./game/atoms.dm-64- ./game/atoms.dm-65-/atom/proc/remove_air(amount) ./game/atoms.dm-66- return null ./game/atoms.dm-67- *** ./game/gamemodes/cult/runes.dm-1000- "\red You are blinded by the flash of red light! After you're able to see again, you see that you are now wearing a set of armor.") ./game/gamemodes/cult/runes.dm-1001- if (armorworn == "summoner" || armorworn == "trickster" || armorworn == "physician") ./game/gamemodes/cult/runes.dm-1002- usr.say("Uln Shogg Hafh[pick("'","`")]drn!") ./game/gamemodes/cult/runes.dm-1003- user << "\red You quietly prick your finger and make a pact with the geometer of blood to acquire more power." ./game/gamemodes/cult/runes.dm-1004- user.take_overall_damage(rand(5,20)) ./game/gamemodes/cult/runes.dm:1005: ***del(user.head)*** ./game/gamemodes/cult/runes.dm:1006: ***del(user.wear_suit)*** ./game/gamemodes/cult/runes.dm-1007- user.equip_to_slot_or_del(new /obj/item/clothing/head/magus(user), slot_head) ./game/gamemodes/cult/runes.dm-1008- user.equip_to_slot_or_del(new /obj/item/clothing/suit/magusred(user), slot_wear_suit) ./game/gamemodes/cult/runes.dm-1009- user.equip_to_slot_or_del(new /obj/item/clothing/shoes/cult(user), slot_shoes) ./game/gamemodes/cult/runes.dm-1010- user.spellremove(user) ./game/gamemodes/cult/runes.dm-1011- usr.visible_message("\red The rune disappears with a flash of red light, and a set of robes appears on [usr]...", \ *** ./game/gamemodes/cult/runes.dm-1018- user.mind.spell_list += new /obj/effect/proc_holder/spell/targeted/turf_teleport/blink/cult(user) ./game/gamemodes/cult/runes.dm-1019- if (armorworn == "physician") ./game/gamemodes/cult/runes.dm-1020- user.put_in_hands(new /obj/item/weapon/gun/magic/wand/resurrection(user)) ./game/gamemodes/cult/runes.dm-1021- if (armorworn == "traveler" || armorworn == "marauder") ./game/gamemodes/cult/runes.dm-1022- usr.say("Tharanak n[pick("'","`")]ghft!") ./game/gamemodes/cult/runes.dm:1023: ***del(user.head)*** ./game/gamemodes/cult/runes.dm:1024: ***del(user.wear_suit)*** ./game/gamemodes/cult/runes.dm-1025- user.equip_to_slot_or_del(new /obj/item/clothing/head/helmet/space/cult(user), slot_head) ./game/gamemodes/cult/runes.dm-1026- user.equip_to_slot_or_del(new /obj/item/clothing/suit/space/cult(user), slot_wear_suit) ./game/gamemodes/cult/runes.dm-1027- user.equip_to_slot_or_del(new /obj/item/clothing/shoes/cult(user), slot_shoes) ./game/gamemodes/cult/runes.dm-1028- user.equip_to_slot_or_del(new /obj/item/weapon/storage/backpack/cultpack(user), slot_back) ./game/gamemodes/cult/runes.dm-1029- user.put_in_hands(new /obj/item/weapon/melee/cultblade(user)) *** ./game/gamemodes/cult/runes.dm-1031- usr.visible_message("\red The rune disappears with a flash of red light, and an armored space suit appears on [usr]...", \ ./game/gamemodes/cult/runes.dm-1032- "\red You are blinded by the flash of red light! After you're able to see again, you see that you are now wearing an armored space suit.") ./game/gamemodes/cult/runes.dm-1033- if (armorworn == "marauder") ./game/gamemodes/cult/runes.dm-1034- user.spellremove(user) ./game/gamemodes/cult/runes.dm-1035- user.mind.spell_list += new /obj/effect/proc_holder/spell/aoe_turf/conjure/creature/cult(user) ./game/gamemodes/cult/runes.dm:1036: ***del(src)*** ./game/gamemodes/cult/runes.dm-1037- return ./game/gamemodes/cult/runes.dm-1038- ./game/gamemodes/cult/runes.dm-1039-///Summon Shell: Summons a construct shell if there's four plasteel sheets on top of the rune ./game/gamemodes/cult/runes.dm-1040- ./game/gamemodes/cult/runes.dm-1041- *** ./game/gamemodes/game_mode.dm-112- var/list/datum/game_mode/usable_modes = list() ./game/gamemodes/game_mode.dm-113- for(var/datum/game_mode/G in runnable_modes) ./game/gamemodes/game_mode.dm-114- if(G.reroll_friendly) ./game/gamemodes/game_mode.dm-115- usable_modes += G ./game/gamemodes/game_mode.dm-116- else ./game/gamemodes/game_mode.dm:117: ***del(G)*** ./game/gamemodes/game_mode.dm-118- ./game/gamemodes/game_mode.dm-119- if(!usable_modes) ./game/gamemodes/game_mode.dm-120- message_admins("Convert_roundtype failed due to no valid modes to convert to. Please report this error to the Coders.") ./game/gamemodes/game_mode.dm-121- return null ./game/gamemodes/game_mode.dm-122- *** ./game/gamemodes/wizard/spellbook.dm-76- var/spell_levels = 0 ./game/gamemodes/wizard/spellbook.dm-77- for(var/obj/effect/proc_holder/spell/aspell in user.mind.spell_list) ./game/gamemodes/wizard/spellbook.dm-78- if(initial(S.name) == initial(aspell.name)) ./game/gamemodes/wizard/spellbook.dm-79- spell_levels = aspell.spell_level ./game/gamemodes/wizard/spellbook.dm-80- user.mind.spell_list.Remove(aspell) ./game/gamemodes/wizard/spellbook.dm:81: ***del(S)*** ./game/gamemodes/wizard/spellbook.dm-82- return cost * (spell_levels+1) ./game/gamemodes/wizard/spellbook.dm-83- return -1 ./game/gamemodes/wizard/spellbook.dm-84-/datum/spellbook_entry/proc/GetInfo() ./game/gamemodes/wizard/spellbook.dm-85- if(!S) ./game/gamemodes/wizard/spellbook.dm-86- S = new spell_type() *** ./game/gamemodes/wizard/spellbook.dm-417- var/datum/spellbook_entry/E = new T ./game/gamemodes/wizard/spellbook.dm-418- if(E.IsAvailible()) ./game/gamemodes/wizard/spellbook.dm-419- entries |= E ./game/gamemodes/wizard/spellbook.dm-420- categories |= E.category ./game/gamemodes/wizard/spellbook.dm-421- else ./game/gamemodes/wizard/spellbook.dm:422: ***del(E)*** ./game/gamemodes/wizard/spellbook.dm-423- tab = categories[1] ./game/gamemodes/wizard/spellbook.dm-424- ./game/gamemodes/wizard/spellbook.dm-425-/obj/item/weapon/spellbook/New() ./game/gamemodes/wizard/spellbook.dm-426- ..() ./game/gamemodes/wizard/spellbook.dm-427- Initialize() *** ./game/machinery/computer/medical.dm-579- if(6) ./game/machinery/computer/medical.dm-580- R.fields["m_stat"] = pick("*Insane*", "*Unstable*", "*Watch*", "Stable") ./game/machinery/computer/medical.dm-581- continue ./game/machinery/computer/medical.dm-582- ./game/machinery/computer/medical.dm-583- else if(prob(1)) ./game/machinery/computer/medical.dm:584: ***del(R)*** ./game/machinery/computer/medical.dm-585- continue ./game/machinery/computer/medical.dm-586- ./game/machinery/computer/medical.dm-587- ..(severity) ./game/machinery/computer/medical.dm-588- ./game/machinery/computer/medical.dm-589-/obj/machinery/computer/med_data/proc/canUseMedicalRecordsConsole(mob/user, message = 1, record1, record2) *** ./game/machinery/computer/security.dm-416- temp += "<a href='?src=\ref[src];choice=Clear Screen'>No</a>" ./game/machinery/computer/security.dm-417- ./game/machinery/computer/security.dm-418- if("Purge All Records") ./game/machinery/computer/security.dm-419- investigate_log("[usr.name] ([usr.key]) has purged all the security records.", "records") ./game/machinery/computer/security.dm-420- for(var/datum/data/record/R in data_core.security) ./game/machinery/computer/security.dm:421: ***del(R)*** ./game/machinery/computer/security.dm-422- data_core.security.Cut() ./game/machinery/computer/security.dm-423- temp = "All Security records deleted." ./game/machinery/computer/security.dm-424- ./game/machinery/computer/security.dm-425- if("Add Entry") ./game/machinery/computer/security.dm-426- if(!( istype(active2, /datum/data/record) )) *** ./game/machinery/computer/security.dm-665- H.sec_hud_set_security_status() ./game/machinery/computer/security.dm-666- if("Delete Record (Security) Execute") ./game/machinery/computer/security.dm-667- investigate_log("[usr.name] ([usr.key]) has deleted the security records for [active1.fields["name"]].", "records") ./game/machinery/computer/security.dm-668- if(active2) ./game/machinery/computer/security.dm-669- data_core.security -= active2 ./game/machinery/computer/security.dm:670: ***del(active2)*** ./game/machinery/computer/security.dm-671- ./game/machinery/computer/security.dm-672- if("Delete Record (ALL) Execute") ./game/machinery/computer/security.dm-673- if(active1) ./game/machinery/computer/security.dm-674- investigate_log("[usr.name] ([usr.key]) has deleted all records for [active1.fields["name"]].", "records") ./game/machinery/computer/security.dm-675- for(var/datum/data/record/R in data_core.medical) ./game/machinery/computer/security.dm-676- if((R.fields["name"] == active1.fields["name"] || R.fields["id"] == active1.fields["id"])) ./game/machinery/computer/security.dm-677- data_core.medical -= R ./game/machinery/computer/security.dm:678: ***del(R)*** ./game/machinery/computer/security.dm-679- break ./game/machinery/computer/security.dm-680- data_core.general -= active1 ./game/machinery/computer/security.dm:681: ***del(active1)*** ./game/machinery/computer/security.dm-682- ./game/machinery/computer/security.dm-683- if(active2) ./game/machinery/computer/security.dm-684- data_core.security -= active2 ./game/machinery/computer/security.dm:685: ***del(active2)*** ./game/machinery/computer/security.dm-686- else ./game/machinery/computer/security.dm-687- temp = "This function does not appear to be working at the moment. Our apologies." ./game/machinery/computer/security.dm-688- ./game/machinery/computer/security.dm-689- add_fingerprint(usr) ./game/machinery/computer/security.dm-690- updateUsrDialog() *** ./game/machinery/computer/security.dm-732- R.fields["photo_front"] = G.fields["photo_front"] ./game/machinery/computer/security.dm-733- R.fields["photo_side"] = G.fields["photo_side"] ./game/machinery/computer/security.dm-734- continue ./game/machinery/computer/security.dm-735- ./game/machinery/computer/security.dm-736- else if(prob(1)) ./game/machinery/computer/security.dm:737: ***del(R)*** ./game/machinery/computer/security.dm-738- continue ./game/machinery/computer/security.dm-739- ./game/machinery/computer/security.dm-740- ..(severity) ./game/machinery/computer/security.dm-741- ./game/machinery/computer/security.dm-742-/obj/machinery/computer/secure_data/proc/canUseSecurityRecordsConsole(mob/user, message1 = 0, record1, record2) *** ./game/machinery/embedded_controller/embedded_controller_base.dm-5- ./game/machinery/embedded_controller/embedded_controller_base.dm-6-/datum/computer/file/embedded_program/proc/post_signal(datum/signal/signal, comm_line) ./game/machinery/embedded_controller/embedded_controller_base.dm-7- if(master) ./game/machinery/embedded_controller/embedded_controller_base.dm-8- master.post_signal(signal, comm_line) ./game/machinery/embedded_controller/embedded_controller_base.dm-9- else ./game/machinery/embedded_controller/embedded_controller_base.dm:10: ***del(signal)*** ./game/machinery/embedded_controller/embedded_controller_base.dm-11- ./game/machinery/embedded_controller/embedded_controller_base.dm-12-/datum/computer/file/embedded_program/proc/receive_user_command(command) ./game/machinery/embedded_controller/embedded_controller_base.dm-13- ./game/machinery/embedded_controller/embedded_controller_base.dm-14-/datum/computer/file/embedded_program/proc/receive_signal(datum/signal/signal, receive_method, receive_param) ./game/machinery/embedded_controller/embedded_controller_base.dm-15- return null *** ./game/machinery/magnet.dm-373- ./game/machinery/magnet.dm-374- if(!(nextmove in list("N","S","E","W","C","R"))) ./game/machinery/magnet.dm-375- // N, S, E, W are directional ./game/machinery/magnet.dm-376- // C is center ./game/machinery/magnet.dm-377- // R is random (in magnetic field's bounds) ./game/machinery/magnet.dm:378: ***del(signal)*** ./game/machinery/magnet.dm-379- break // break the loop if the character located is invalid ./game/machinery/magnet.dm-380- ./game/machinery/magnet.dm-381- signal.data["command"] = nextmove ./game/machinery/magnet.dm-382- ./game/machinery/magnet.dm-383- *** ./game/machinery/overview.dm-172- var/icon/J = imap[i*2+2] ./game/machinery/overview.dm-173- ./game/machinery/overview.dm-174- HI.Insert(I, frame=1, delay = 5) ./game/machinery/overview.dm-175- HI.Insert(J, frame=2, delay = 5) ./game/machinery/overview.dm-176- ./game/machinery/overview.dm:177: ***del(I)*** ./game/machinery/overview.dm:178: ***del(J)*** ./game/machinery/overview.dm-179- H.icon = HI ./game/machinery/overview.dm-180- H.layer = 25 ./game/machinery/overview.dm-181- usr.mapobjs += H ./game/machinery/overview.dm-182-#else ./game/machinery/overview.dm-183- *** ./game/machinery/overview.dm-292- H.name = (i==0)?"maprefresh":"map" ./game/machinery/overview.dm-293- ./game/machinery/overview.dm-294- var/icon/I = imap[i+1] ./game/machinery/overview.dm-295- ./game/machinery/overview.dm-296- H.icon = I ./game/machinery/overview.dm:297: ***del(I)*** ./game/machinery/overview.dm-298- H.layer = 25 ./game/machinery/overview.dm-299- usr.mapobjs += H ./game/machinery/overview.dm-300- ./game/machinery/overview.dm-301-#endif ./game/machinery/overview.dm-302- *** ./game/machinery/suit_storage_unit.dm-391- if(i==3) //End of the cycle ./game/machinery/suit_storage_unit.dm-392- if(!src.issuperUV) ./game/machinery/suit_storage_unit.dm-393- for(var/obj/item/ITEM in src) ./game/machinery/suit_storage_unit.dm-394- ITEM.clean_blood() ./game/machinery/suit_storage_unit.dm-395- if(istype(STORAGE, /obj/item/weapon/reagent_containers/food)) ./game/machinery/suit_storage_unit.dm:396: ***del(STORAGE)*** ./game/machinery/suit_storage_unit.dm-397- else //It was supercycling, destroy everything ./game/machinery/suit_storage_unit.dm-398- src.HELMET = null ./game/machinery/suit_storage_unit.dm-399- src.SUIT = null ./game/machinery/suit_storage_unit.dm-400- src.MASK = null ./game/machinery/suit_storage_unit.dm:401: ***del(STORAGE)*** ./game/machinery/suit_storage_unit.dm-402- visible_message("<font color='red'>With a loud whining noise, the Suit Storage Unit's door grinds open. Puffs of ashen smoke come out of its chamber.</font>", 3) ./game/machinery/suit_storage_unit.dm-403- src.isbroken = 1 ./game/machinery/suit_storage_unit.dm-404- src.isopen = 1 ./game/machinery/suit_storage_unit.dm-405- src.islocked = 0 ./game/machinery/suit_storage_unit.dm-406- src.eject_occupant(OCCUPANT) //Mixing up these two lines causes bug. DO NOT DO IT. *** ./game/machinery/telecomms/computers/logbrowser.dm-195- var/datum/comm_log_entry/D = SelectedServer.log_entries[text2num(href_list["delete"])] ./game/machinery/telecomms/computers/logbrowser.dm-196- ./game/machinery/telecomms/computers/logbrowser.dm-197- temp = "<font color = #336699>- DELETED ENTRY: [D.name] -</font color>" ./game/machinery/telecomms/computers/logbrowser.dm-198- ./game/machinery/telecomms/computers/logbrowser.dm-199- SelectedServer.log_entries.Remove(D) ./game/machinery/telecomms/computers/logbrowser.dm:200: ***del(D)*** ./game/machinery/telecomms/computers/logbrowser.dm-201- ./game/machinery/telecomms/computers/logbrowser.dm-202- else ./game/machinery/telecomms/computers/logbrowser.dm-203- temp = "<font color = #D70B00>- FAILED: NO SELECTED MACHINE -</font color>" ./game/machinery/telecomms/computers/logbrowser.dm-204- ./game/machinery/telecomms/computers/logbrowser.dm-205- if(href_list["network"]) *** ./game/mecha/mecha.dm-135- internal_tank = null ./game/mecha/mecha.dm-136- if(loc) ./game/mecha/mecha.dm-137- loc.assume_air(cabin_air) ./game/mecha/mecha.dm-138- air_update_turf() ./game/mecha/mecha.dm-139- else ./game/mecha/mecha.dm:140: ***del(cabin_air)*** ./game/mecha/mecha.dm-141- cabin_air = null ./game/mecha/mecha.dm-142- qdel(spark_system) ./game/mecha/mecha.dm-143- spark_system = null ./game/mecha/mecha.dm-144- ./game/mecha/mecha.dm-145- mechas_list -= src //global mech list *** ./game/mecha/mecha.dm-238- var/datum/gas_mixture/leaked_gas = int_tank_air.remove_ratio(0.10) ./game/mecha/mecha.dm-239- if(loc) ./game/mecha/mecha.dm-240- loc.assume_air(leaked_gas) ./game/mecha/mecha.dm-241- air_update_turf() ./game/mecha/mecha.dm-242- else ./game/mecha/mecha.dm:243: ***del(leaked_gas)*** ./game/mecha/mecha.dm-244- ./game/mecha/mecha.dm-245- if(internal_damage & MECHA_INT_SHORT_CIRCUIT) ./game/mecha/mecha.dm-246- if(get_charge()) ./game/mecha/mecha.dm-247- spark_system.start() ./game/mecha/mecha.dm-248- cell.charge -= min(20,cell.charge) *** ./game/mecha/mecha.dm-274- transfer_moles = pressure_delta*cabin_air.return_volume()/(cabin_air.return_temperature() * R_IDEAL_GAS_EQUATION) ./game/mecha/mecha.dm-275- var/datum/gas_mixture/removed = cabin_air.remove(transfer_moles) ./game/mecha/mecha.dm-276- if(t_air) ./game/mecha/mecha.dm-277- t_air.merge(removed) ./game/mecha/mecha.dm-278- else //just delete the cabin gas, we're in space or some shit ./game/mecha/mecha.dm:279: ***del(removed)*** ./game/mecha/mecha.dm-280- ./game/mecha/mecha.dm-281- if(occupant) ./game/mecha/mecha.dm-282- if(cell) ./game/mecha/mecha.dm-283- var/cellcharge = cell.charge/cell.maxcharge ./game/mecha/mecha.dm-284- switch(cellcharge) *** ./game/mecha/mecha_construction_paths.dm-298- var/obj/item/mecha_parts/chassis/const_holder = holder ./game/mecha/mecha_construction_paths.dm-299- const_holder.construct = new /datum/construction/reversible/mecha/gygax(const_holder) ./game/mecha/mecha_construction_paths.dm-300- const_holder.icon = 'icons/mecha/mech_construction.dmi' ./game/mecha/mecha_construction_paths.dm-301- const_holder.icon_state = "gygax0" ./game/mecha/mecha_construction_paths.dm-302- const_holder.density = 1 ./game/mecha/mecha_construction_paths.dm:303: ***del(src)*** ./game/mecha/mecha_construction_paths.dm-304- return ./game/mecha/mecha_construction_paths.dm-305- ./game/mecha/mecha_construction_paths.dm-306- ./game/mecha/mecha_construction_paths.dm-307-/datum/construction/reversible/mecha/gygax ./game/mecha/mecha_construction_paths.dm-308- result = "/obj/mecha/combat/gygax" *** ./game/mecha/mecha_construction_paths.dm-577- var/obj/item/mecha_parts/chassis/const_holder = holder ./game/mecha/mecha_construction_paths.dm-578- const_holder.construct = new /datum/construction/reversible/mecha/firefighter(const_holder) ./game/mecha/mecha_construction_paths.dm-579- const_holder.icon = 'icons/mecha/mech_construction.dmi' ./game/mecha/mecha_construction_paths.dm-580- const_holder.icon_state = "fireripley0" ./game/mecha/mecha_construction_paths.dm-581- const_holder.density = 1 ./game/mecha/mecha_construction_paths.dm:582: ***del(src)*** ./game/mecha/mecha_construction_paths.dm-583- return ./game/mecha/mecha_construction_paths.dm-584- ./game/mecha/mecha_construction_paths.dm-585- ./game/mecha/mecha_construction_paths.dm-586-/datum/construction/reversible/mecha/firefighter ./game/mecha/mecha_construction_paths.dm-587- result = "/obj/mecha/working/ripley/firefighter" *** ./game/mecha/mecha_construction_paths.dm-799- ./game/mecha/mecha_construction_paths.dm-800-/datum/construction/mecha/honker_chassis/spawn_result() ./game/mecha/mecha_construction_paths.dm-801- var/obj/item/mecha_parts/chassis/const_holder = holder ./game/mecha/mecha_construction_paths.dm-802- const_holder.construct = new /datum/construction/mecha/honker(const_holder) ./game/mecha/mecha_construction_paths.dm-803- const_holder.density = 1 ./game/mecha/mecha_construction_paths.dm:804: ***del(src)*** ./game/mecha/mecha_construction_paths.dm-805- return ./game/mecha/mecha_construction_paths.dm-806- ./game/mecha/mecha_construction_paths.dm-807- ./game/mecha/mecha_construction_paths.dm-808-/datum/construction/mecha/honker ./game/mecha/mecha_construction_paths.dm-809- result = "/obj/mecha/combat/honker" *** ./game/objects/effects/mines.dm-54- name = "kick mine" ./game/objects/effects/mines.dm-55- ./game/objects/effects/mines.dm-56-/obj/effect/mine/kickmine/mineEffect(mob/victim) ./game/objects/effects/mines.dm-57- if(isliving(victim) && victim.client) ./game/objects/effects/mines.dm-58- victim << "<span class='userdanger'>You have been kicked FOR NO REISIN!</span>" ./game/objects/effects/mines.dm:59: ***del(victim.client)*** ./game/objects/effects/mines.dm-60- ./game/objects/effects/mines.dm-61- ./game/objects/effects/mines.dm-62-/obj/effect/mine/gas ./game/objects/effects/mines.dm-63- name = "oxygen mine" ./game/objects/effects/mines.dm-64- var/gas_amount = 360 *** ./game/objects/items/devices/camera_bug.dm-46- ./game/objects/items/devices/camera_bug.dm-47-/obj/item/device/camera_bug/Destroy() ./game/objects/items/devices/camera_bug.dm-48- if(expansion) ./game/objects/items/devices/camera_bug.dm-49- qdel(expansion) ./game/objects/items/devices/camera_bug.dm-50- expansion = null ./game/objects/items/devices/camera_bug.dm:51: ***del(src)*** ./game/objects/items/devices/camera_bug.dm-52-//Easier to just call del() than this nonsense ./game/objects/items/devices/camera_bug.dm-53-// ya no, del() takes 0.8ds to run on avg. this takes less than 0.01 ./game/objects/items/devices/camera_bug.dm-54- get_cameras() ./game/objects/items/devices/camera_bug.dm-55- for(var/cam_tag in bugged_cameras) ./game/objects/items/devices/camera_bug.dm-56- var/obj/machinery/camera/camera = bugged_cameras[cam_tag] *** ./game/objects/items/holotape.dm-242- while(N != 1) ./game/objects/items/holotape.dm-243- N = 1 ./game/objects/items/holotape.dm-244- for (var/obj/item/holotape/P in cur) ./game/objects/items/holotape.dm-245- if(P.icon_state == icon_dir) ./game/objects/items/holotape.dm-246- N = 0 ./game/objects/items/holotape.dm:247: ***del(P)*** ./game/objects/items/holotape.dm-248- cur = get_step(cur,dir[i]) ./game/objects/items/holotape.dm-249- ./game/objects/items/holotape.dm:250: ***del(src)*** ./game/objects/items/holotape.dm-251- return ./game/objects/items/holotape.dm-252- ./game/objects/items/holotape.dm-253-#undef MAX_TAPE_RANGE *** ./game/objects/items/weapons/storage/storage.dm-194- var/obj/item/sample_object ./game/objects/items/weapons/storage/storage.dm-195- var/number ./game/objects/items/weapons/storage/storage.dm-196- ./game/objects/items/weapons/storage/storage.dm-197- New(obj/item/sample) ./game/objects/items/weapons/storage/storage.dm-198- if(!istype(sample)) ./game/objects/items/weapons/storage/storage.dm:199: ***del(src)*** ./game/objects/items/weapons/storage/storage.dm-200- sample_object = sample ./game/objects/items/weapons/storage/storage.dm-201- number = 1 ./game/objects/items/weapons/storage/storage.dm-202- ./game/objects/items/weapons/storage/storage.dm-203- ./game/objects/items/weapons/storage/storage.dm-204-//This proc determins the size of the inventory to be displayed. Please touch it only if you know what you're doing. *** ./game/objects/items/weapons/tanks/tanks.dm-31- ./game/objects/items/weapons/tanks/tanks.dm-32- return ./game/objects/items/weapons/tanks/tanks.dm-33- ./game/objects/items/weapons/tanks/tanks.dm-34-/obj/item/weapon/tank/Destroy() ./game/objects/items/weapons/tanks/tanks.dm-35- if(air_contents) ./game/objects/items/weapons/tanks/tanks.dm:36: ***del(air_contents)*** ./game/objects/items/weapons/tanks/tanks.dm-37- ./game/objects/items/weapons/tanks/tanks.dm-38- SSobj.processing.Remove(src) ./game/objects/items/weapons/tanks/tanks.dm-39- ./game/objects/items/weapons/tanks/tanks.dm-40- ..() ./game/objects/items/weapons/tanks/tanks.dm-41- *** ./LINDA/LINDA_turf_tile.dm-6- var/atmos_adjacent_turfs = 0 ./LINDA/LINDA_turf_tile.dm-7- var/atmos_adjacent_turfs_amount = 0 ./LINDA/LINDA_turf_tile.dm-8- var/atmos_supeconductivity = 0 ./LINDA/LINDA_turf_tile.dm-9- ./LINDA/LINDA_turf_tile.dm-10-/turf/assume_air(datum/gas_mixture/giver) //use this for machines to adjust air ./LINDA/LINDA_turf_tile.dm:11: ***del(giver)*** ./LINDA/LINDA_turf_tile.dm-12- return 0 ./LINDA/LINDA_turf_tile.dm-13- ./LINDA/LINDA_turf_tile.dm-14-/turf/return_air() ./LINDA/LINDA_turf_tile.dm-15- //Create gas mixture to hold data for passing ./LINDA/LINDA_turf_tile.dm-16- var/datum/gas_mixture/GM = new *** ./modules/admin/admin.dm-767- if(kick_only_afk && !C.is_afk()) //Ignore clients who are not afk ./modules/admin/admin.dm-768- continue ./modules/admin/admin.dm-769- if(message) ./modules/admin/admin.dm-770- C << message ./modules/admin/admin.dm-771- kicked_client_names.Add("[C.ckey]") ./modules/admin/admin.dm:772: ***del(C)*** ./modules/admin/admin.dm-773- return kicked_client_names ./modules/admin/admin.dm-774- ./modules/admin/admin.dm-775-//returns 1 to let the dragdrop code know we are trapping this event ./modules/admin/admin.dm-776-//returns 0 if we don't plan to trap the event ./modules/admin/admin.dm-777-/datum/admins/proc/cmd_ghost_drag(mob/dead/observer/frommob, mob/living/tomob) *** ./modules/admin/admin_ranks.dm-9-/datum/admin_rank/New(init_name, init_rights, list/init_adds, list/init_subs) ./modules/admin/admin_ranks.dm-10- name = init_name ./modules/admin/admin_ranks.dm-11- switch(name) ./modules/admin/admin_ranks.dm-12- if("Removed",null,"") ./modules/admin/admin_ranks.dm-13- spawn(-1) ./modules/admin/admin_ranks.dm:14: ***del(src)*** ./modules/admin/admin_ranks.dm-15- throw EXCEPTION("invalid admin-rank name") ./modules/admin/admin_ranks.dm-16- return ./modules/admin/admin_ranks.dm-17- if(init_rights) rights = init_rights ./modules/admin/admin_ranks.dm-18- if(!init_adds) init_adds = list() ./modules/admin/admin_ranks.dm-19- if(!init_subs) init_subs = list() *** ./modules/admin/DB ban/functions.dm-133- if(announceinirc) ./modules/admin/DB ban/functions.dm-134- send2irc("BAN ALERT","[a_ckey] applied a [bantype_str] on [ckey]") ./modules/admin/DB ban/functions.dm-135- ./modules/admin/DB ban/functions.dm-136- if(kickbannedckey) ./modules/admin/DB ban/functions.dm-137- if(banned_mob && banned_mob.client && banned_mob.client.ckey == banckey) ./modules/admin/DB ban/functions.dm:138: ***del(banned_mob.client)*** ./modules/admin/DB ban/functions.dm-139- ./modules/admin/DB ban/functions.dm-140- ./modules/admin/DB ban/functions.dm-141-/datum/admins/proc/DB_ban_unban(ckey, bantype, job = "") ./modules/admin/DB ban/functions.dm-142- ./modules/admin/DB ban/functions.dm-143- if(!check_rights(R_BAN)) return *** ./modules/admin/holder2.dm-15- var/admincaster_signature ./modules/admin/holder2.dm-16- ./modules/admin/holder2.dm-17-/datum/admins/New(datum/admin_rank/R, ckey) ./modules/admin/holder2.dm-18- if(!ckey) ./modules/admin/holder2.dm-19- spawn(-1) ./modules/admin/holder2.dm:20: ***del(src)*** ./modules/admin/holder2.dm-21- throw EXCEPTION("Admin datum created without a ckey") ./modules/admin/holder2.dm-22- return ./modules/admin/holder2.dm-23- if(!istype(R)) ./modules/admin/holder2.dm-24- spawn(-1) ./modules/admin/holder2.dm:25: ***del(src)*** ./modules/admin/holder2.dm-26- throw EXCEPTION("Admin datum created without a rank") ./modules/admin/holder2.dm-27- return ./modules/admin/holder2.dm-28- rank = R ./modules/admin/holder2.dm-29- admincaster_signature = "Nanotrasen Officer #[rand(0,9)][rand(0,9)][rand(0,9)]" ./modules/admin/holder2.dm-30- admin_datums[ckey] = src *** ./modules/admin/holder2.dm-89- ./modules/admin/holder2.dm-90-/client/proc/deadmin() ./modules/admin/holder2.dm-91- admin_datums -= ckey ./modules/admin/holder2.dm-92- if(holder) ./modules/admin/holder2.dm-93- holder.disassociate() ./modules/admin/holder2.dm:94: ***del(holder)*** ./modules/admin/holder2.dm-95- return 1 ./modules/admin/holder2.dm-96- ./modules/admin/holder2.dm-97-//This proc checks whether subject has at least ONE of the rights specified in rights_required. ./modules/admin/holder2.dm-98-/proc/check_rights_for(client/subject, rights_required) ./modules/admin/holder2.dm-99- if(subject && subject.holder && subject.holder.rank) *** ./modules/admin/topic.dm-989- return ./modules/admin/topic.dm-990- M << "<span class='danger'>You have been kicked from the server.</span>" ./modules/admin/topic.dm-991- log_admin("[key_name(usr)] booted [key_name(M)].") ./modules/admin/topic.dm-992- message_admins("<span class='adminnotice'>[key_name_admin(usr)] booted [key_name_admin(M)].</span>") ./modules/admin/topic.dm-993- //M.client = null ./modules/admin/topic.dm:994: ***del(M.client)*** ./modules/admin/topic.dm-995- ./modules/admin/topic.dm-996- //Player Notes ./modules/admin/topic.dm-997- else if(href_list["notes"]) ./modules/admin/topic.dm-998- var/ckey = href_list["ckey"] ./modules/admin/topic.dm-999- switch(href_list["notes"]) *** ./modules/admin/topic.dm-1050- else ./modules/admin/topic.dm-1051- M << "<span class='danger'>No ban appeals URL has been set.</span>" ./modules/admin/topic.dm-1052- log_admin("[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis will be removed in [mins] minutes.") ./modules/admin/topic.dm-1053- message_admins("<span class='adminnotice'>[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis will be removed in [mins] minutes.</span>") ./modules/admin/topic.dm-1054- ./modules/admin/topic.dm:1055: ***del(M.client)*** ./modules/admin/topic.dm-1056- //qdel(M) // See no reason why to delete mob. Important stuff can be lost. And ban can be lifted before round ends. ./modules/admin/topic.dm-1057- if("No") ./modules/admin/topic.dm-1058- var/reason = input(usr,"Reason?","reason","Griefer") as text|null ./modules/admin/topic.dm-1059- if(!reason) ./modules/admin/topic.dm-1060- return *** ./modules/admin/topic.dm-1074- log_admin("[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis is a permanent ban.") ./modules/admin/topic.dm-1075- message_admins("<span class='adminnotice'>[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis is a permanent ban.</span>") ./modules/admin/topic.dm-1076- feedback_inc("ban_perma",1) ./modules/admin/topic.dm-1077- DB_ban_record(BANTYPE_PERMA, M, -1, reason) ./modules/admin/topic.dm-1078- ./modules/admin/topic.dm:1079: ***del(M.client)*** ./modules/admin/topic.dm-1080- //qdel(M) ./modules/admin/topic.dm-1081- if("Cancel") ./modules/admin/topic.dm-1082- return ./modules/admin/topic.dm-1083- ./modules/admin/topic.dm-1084- else if(href_list["unjobbanf"]) *** ./modules/admin/verbs/SDQL2/SDQL_2.dm-178- do_parse = 0 ./modules/admin/verbs/SDQL2/SDQL_2.dm-179- else ./modules/admin/verbs/SDQL2/SDQL_2.dm-180- query_tree += val ./modules/admin/verbs/SDQL2/SDQL_2.dm-181- pos++ ./modules/admin/verbs/SDQL2/SDQL_2.dm-182- ./modules/admin/verbs/SDQL2/SDQL_2.dm:183: ***del(parser)*** ./modules/admin/verbs/SDQL2/SDQL_2.dm-184- ./modules/admin/verbs/SDQL2/SDQL_2.dm-185- return querys ./modules/admin/verbs/SDQL2/SDQL_2.dm-186- ./modules/admin/verbs/SDQL2/SDQL_2.dm-187- ./modules/admin/verbs/SDQL2/SDQL_2.dm-188- *** ./modules/awaymissions/maploader/swapmaps.dm-154- z1=min(x:z,y:z);z2=max(x:z,y:z) ./modules/awaymissions/maploader/swapmaps.dm-155- InitializeSwapMaps() ./modules/awaymissions/maploader/swapmaps.dm-156- if(z2>swapmaps_compiled_maxz ||\ ./modules/awaymissions/maploader/swapmaps.dm-157- y2>swapmaps_compiled_maxy ||\ ./modules/awaymissions/maploader/swapmaps.dm-158- x2>swapmaps_compiled_maxx) ./modules/awaymissions/maploader/swapmaps.dm:159: ***del(src)*** ./modules/awaymissions/maploader/swapmaps.dm-160- return ./modules/awaymissions/maploader/swapmaps.dm-161- x2=x?(x):world.maxx ./modules/awaymissions/maploader/swapmaps.dm-162- y2=y?(y):world.maxy ./modules/awaymissions/maploader/swapmaps.dm-163- z2=z?(z):1 ./modules/awaymissions/maploader/swapmaps.dm-164- AllocateSwapMap() *** ./modules/awaymissions/maploader/swapmaps.dm-177- for(var/obj/O in A) qdel(O) ./modules/awaymissions/maploader/swapmaps.dm-178- for(var/mob/M in A) ./modules/awaymissions/maploader/swapmaps.dm-179- if(!M.key) qdel(M) ./modules/awaymissions/maploader/swapmaps.dm-180- else M.loc=null ./modules/awaymissions/maploader/swapmaps.dm-181- areas[A.loc]=null ./modules/awaymissions/maploader/swapmaps.dm:182: ***del(A)*** ./modules/awaymissions/maploader/swapmaps.dm-183- // delete areas that belong only to this map ./modules/awaymissions/maploader/swapmaps.dm-184- for(var/area/a in areas) ./modules/awaymissions/maploader/swapmaps.dm:185: if(a && !a.contents.len) ***del(a)*** ./modules/awaymissions/maploader/swapmaps.dm-186- if(x2>=world.maxx || y2>=world.maxy || z2>=world.maxz) CutXYZ() ./modules/awaymissions/maploader/swapmaps.dm:187: ***del(areas)*** ./modules/awaymissions/maploader/swapmaps.dm-188- ..() ./modules/awaymissions/maploader/swapmaps.dm-189- ./modules/awaymissions/maploader/swapmaps.dm-190- /* ./modules/awaymissions/maploader/swapmaps.dm-191- Savefile format: ./modules/awaymissions/maploader/swapmaps.dm-192- map *** ./modules/awaymissions/maploader/swapmaps.dm-239- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-240- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-241- sleep() ./modules/awaymissions/maploader/swapmaps.dm-242- S.cd=oldcd ./modules/awaymissions/maploader/swapmaps.dm-243- locked=0 ./modules/awaymissions/maploader/swapmaps.dm:244: ***del(areas)*** ./modules/awaymissions/maploader/swapmaps.dm-245- ./modules/awaymissions/maploader/swapmaps.dm-246- Read(savefile/S,_id,turf/locorner) ./modules/awaymissions/maploader/swapmaps.dm-247- var/x ./modules/awaymissions/maploader/swapmaps.dm-248- var/y ./modules/awaymissions/maploader/swapmaps.dm-249- var/z *** ./modules/awaymissions/maploader/swapmaps.dm-295- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-296- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-297- sleep() ./modules/awaymissions/maploader/swapmaps.dm-298- S.cd=oldcd ./modules/awaymissions/maploader/swapmaps.dm-299- locked=0 ./modules/awaymissions/maploader/swapmaps.dm:300: ***del(areas)*** ./modules/awaymissions/maploader/swapmaps.dm-301- ./modules/awaymissions/maploader/swapmaps.dm-302- /* ./modules/awaymissions/maploader/swapmaps.dm-303- Find an empty block on the world map in which to load this map. ./modules/awaymissions/maploader/swapmaps.dm-304- If no space is found, increase world.maxz as necessary. (If the ./modules/awaymissions/maploader/swapmaps.dm-305- map is greater in x,y size than the current world, expand *** ./modules/awaymissions/maploader/swapmaps.dm-319- else ./modules/awaymissions/maploader/swapmaps.dm-320- var/list/l=ConsiderRegion(1,1,world.maxx,world.maxy,swapmaps_compiled_maxz+1) ./modules/awaymissions/maploader/swapmaps.dm-321- x1=l[1] ./modules/awaymissions/maploader/swapmaps.dm-322- y1=l[2] ./modules/awaymissions/maploader/swapmaps.dm-323- z1=l[3] ./modules/awaymissions/maploader/swapmaps.dm:324: ***del(l)*** ./modules/awaymissions/maploader/swapmaps.dm-325- x2+=x1-1 ./modules/awaymissions/maploader/swapmaps.dm-326- y2+=y1-1 ./modules/awaymissions/maploader/swapmaps.dm-327- z2+=z1-1 ./modules/awaymissions/maploader/swapmaps.dm-328- world.maxz=max(z2,world.maxz) // stretch z if necessary ./modules/awaymissions/maploader/swapmaps.dm-329- if(!ischunk) *** ./modules/awaymissions/maploader/swapmaps.dm-375- world.maxz=mz ./modules/awaymissions/maploader/swapmaps.dm-376- ./modules/awaymissions/maploader/swapmaps.dm-377- // save and delete ./modules/awaymissions/maploader/swapmaps.dm-378- proc/Unload() ./modules/awaymissions/maploader/swapmaps.dm-379- Save() ./modules/awaymissions/maploader/swapmaps.dm:380: ***del(src)*** ./modules/awaymissions/maploader/swapmaps.dm-381- ./modules/awaymissions/maploader/swapmaps.dm-382- proc/Save() ./modules/awaymissions/maploader/swapmaps.dm-383- if(id==src) return 0 ./modules/awaymissions/maploader/swapmaps.dm-384- var/savefile/S=mode?(new):new("map_[id].sav") ./modules/awaymissions/maploader/swapmaps.dm-385- S << src *** ./modules/awaymissions/maploader/swapmaps.dm-470- var/list/l=contents ./modules/awaymissions/maploader/swapmaps.dm-471- if(M) ./modules/awaymissions/maploader/swapmaps.dm-472- l=l.Copy() ./modules/awaymissions/maploader/swapmaps.dm-473- for(M in src) if(M.key) l-=M ./modules/awaymissions/maploader/swapmaps.dm-474- if(l.len) S["contents"]<<l ./modules/awaymissions/maploader/swapmaps.dm:475: if(l!=contents) ***del(l)*** ./modules/awaymissions/maploader/swapmaps.dm-476- Read(savefile/S) ./modules/awaymissions/maploader/swapmaps.dm-477- var/list/l ./modules/awaymissions/maploader/swapmaps.dm-478- if(contents.len) l=contents ./modules/awaymissions/maploader/swapmaps.dm-479- ..() ./modules/awaymissions/maploader/swapmaps.dm-480- // if the icon was a text string, it would not have loaded properly *** ./modules/awaymissions/maploader/swapmaps.dm-483- var/ic ./modules/awaymissions/maploader/swapmaps.dm-484- S["icon"]>>ic ./modules/awaymissions/maploader/swapmaps.dm-485- if(istext(ic)) icon=swapmaps_iconcache[ic] ./modules/awaymissions/maploader/swapmaps.dm-486- if(l && contents!=l) ./modules/awaymissions/maploader/swapmaps.dm-487- contents+=l ./modules/awaymissions/maploader/swapmaps.dm:488: ***del(l)*** ./modules/awaymissions/maploader/swapmaps.dm-489- ./modules/awaymissions/maploader/swapmaps.dm-490- ./modules/awaymissions/maploader/swapmaps.dm-491-// set this up (at runtime) as follows: ./modules/awaymissions/maploader/swapmaps.dm-492-// list(\ ./modules/awaymissions/maploader/swapmaps.dm-493-// 'player.dmi'="player",\ *** ./modules/awaymissions/maploader/swapmaps.dm-623- properly otherwise. The //.0 path should always match the map, however. ./modules/awaymissions/maploader/swapmaps.dm-624- */ ./modules/awaymissions/maploader/swapmaps.dm-625- S.cd="//.0" ./modules/awaymissions/maploader/swapmaps.dm-626- M.Read(S,M,locorner) ./modules/awaymissions/maploader/swapmaps.dm-627- while(M.locked) sleep(1) ./modules/awaymissions/maploader/swapmaps.dm:628: ***del(M)*** ./modules/awaymissions/maploader/swapmaps.dm-629- return 1 ./modules/awaymissions/maploader/swapmaps.dm-630- ./modules/awaymissions/maploader/swapmaps.dm-631-proc/SwapMaps_SaveChunk(chunk_id,turf/corner1,turf/corner2) ./modules/awaymissions/maploader/swapmaps.dm-632- if(!corner1 || !corner2) ./modules/awaymissions/maploader/swapmaps.dm-633- world.log << "SwapMaps error in SwapMaps_SaveChunk():" *** ./modules/awaymissions/maploader/swapmaps.dm-644- M.y2=max(corner1.y,corner2.y) ./modules/awaymissions/maploader/swapmaps.dm-645- M.z2=max(corner1.z,corner2.z) ./modules/awaymissions/maploader/swapmaps.dm-646- M.mode=swapmaps_mode ./modules/awaymissions/maploader/swapmaps.dm-647- M.Save() ./modules/awaymissions/maploader/swapmaps.dm-648- while(M.locked) sleep(1) ./modules/awaymissions/maploader/swapmaps.dm:649: ***del(M)*** ./modules/awaymissions/maploader/swapmaps.dm-650- return 1 ./modules/awaymissions/maploader/swapmaps.dm-651- ./modules/awaymissions/maploader/swapmaps.dm-652-proc/SwapMaps_GetSize(id) ./modules/awaymissions/maploader/swapmaps.dm-653- var/savefile/S ./modules/awaymissions/maploader/swapmaps.dm-654- var/text=0 *** ./modules/client/client procs.dm-138- if (isnum(player_age) && player_age == -1) //first connection ./modules/client/client procs.dm-139- if (config.panic_bunker && !holder && !(ckey in deadmins)) ./modules/client/client procs.dm-140- log_access("Failed Login: [key] - New account attempting to connect during panic bunker") ./modules/client/client procs.dm-141- message_admins("<span class='adminnotice'>Failed Login: [key] - New account attempting to connect during panic bunker</span>") ./modules/client/client procs.dm-142- src << "Sorry but the server is currently not accepting connections from never before seen players." ./modules/client/client procs.dm:143: ***del(src)*** ./modules/client/client procs.dm-144- return 0 ./modules/client/client procs.dm-145- ./modules/client/client procs.dm-146- if (config.notify_new_player_age >= 0) ./modules/client/client procs.dm-147- message_admins("New user: [key_name_admin(src)] is connecting here for the first time.") ./modules/client/client procs.dm-148- if (config.irc_first_connection_alert) *** ./modules/crafting/guncrafting.dm-22- if(istype(W,/obj/item/pipe)) ./modules/crafting/guncrafting.dm-23- user << "<span class='notice'>You attach the shotgun barrel to the receiver. The pins seem loose.</span>" ./modules/crafting/guncrafting.dm-24- var/obj/item/weaponcrafting/ishotgunconstruction/I = new /obj/item/weaponcrafting/ishotgunconstruction ./modules/crafting/guncrafting.dm-25- user.unEquip(src) ./modules/crafting/guncrafting.dm-26- user.put_in_hands(I) ./modules/crafting/guncrafting.dm:27: ***del(W)*** ./modules/crafting/guncrafting.dm:28: ***del(src)*** ./modules/crafting/guncrafting.dm-29- return ./modules/crafting/guncrafting.dm-30- ./modules/crafting/guncrafting.dm-31-// SHOTGUN // ./modules/crafting/guncrafting.dm-32- ./modules/crafting/guncrafting.dm-33-/obj/item/weaponcrafting/ishotgunconstruction *** ./modules/crafting/guncrafting.dm-55- if(istype(W,/obj/item/weaponcrafting/stock)) ./modules/crafting/guncrafting.dm-56- user << "<span class='notice'>You attach the stock to the receiver-barrel assembly.</span>" ./modules/crafting/guncrafting.dm-57- var/obj/item/weaponcrafting/ishotgunconstruction3/I = new /obj/item/weaponcrafting/ishotgunconstruction3 ./modules/crafting/guncrafting.dm-58- user.unEquip(src) ./modules/crafting/guncrafting.dm-59- user.put_in_hands(I) ./modules/crafting/guncrafting.dm:60: ***del(W)*** ./modules/crafting/guncrafting.dm:61: ***del(src)*** ./modules/crafting/guncrafting.dm-62- return ./modules/crafting/guncrafting.dm-63- ./modules/crafting/guncrafting.dm-64-/obj/item/weaponcrafting/ishotgunconstruction3 ./modules/crafting/guncrafting.dm-65- name = "extremely conspicuous metal construction" ./modules/crafting/guncrafting.dm-66- desc = "A receiver-barrel shotgun assembly with a loose wooden stock. There's no way you can fire it without the stock coming loose." *** ./modules/flufftext/Hallucination.dm-154- if(target.client) target.client.images |= flood_images ./modules/flufftext/Hallucination.dm-155- return ./modules/flufftext/Hallucination.dm-156- ./modules/flufftext/Hallucination.dm-157-/obj/effect/hallucination/fake_flood/Destroy() ./modules/flufftext/Hallucination.dm-158- SSobj.processing.Remove(src) ./modules/flufftext/Hallucination.dm:159: ***del(flood_turfs)*** ./modules/flufftext/Hallucination.dm-160- if(target.client) target.client.images.Remove(flood_images) ./modules/flufftext/Hallucination.dm-161- target = null ./modules/flufftext/Hallucination.dm:162: ***del(flood_images)*** ./modules/flufftext/Hallucination.dm-163- return ./modules/flufftext/Hallucination.dm-164- ./modules/flufftext/Hallucination.dm-165-/obj/effect/hallucination/simple/xeno ./modules/flufftext/Hallucination.dm-166- image_icon = 'icons/mob/alien.dmi' ./modules/flufftext/Hallucination.dm-167- image_state = "alienh_pounce" *** ./modules/hydroponics/hydroponics.dm-684- qdel(reagent_source) ./modules/hydroponics/hydroponics.dm-685- ./modules/hydroponics/hydroponics.dm-686- H.applyChemicals(S) ./modules/hydroponics/hydroponics.dm-687- ./modules/hydroponics/hydroponics.dm-688- S.clear_reagents() ./modules/hydroponics/hydroponics.dm:689: ***del(S)*** ./modules/hydroponics/hydroponics.dm-690- H.update_icon() ./modules/hydroponics/hydroponics.dm-691- if(reagent_source) // If the source wasn't composted and destroyed ./modules/hydroponics/hydroponics.dm-692- reagent_source.update_icon() ./modules/hydroponics/hydroponics.dm-693- return 1 ./modules/hydroponics/hydroponics.dm-694- *** ./modules/hydroponics/seed_extractor.dm-183- if(N.amount <= 0) ./modules/hydroponics/seed_extractor.dm-184- return ./modules/hydroponics/seed_extractor.dm-185- N.amount = max(N.amount - 1, 0) ./modules/hydroponics/seed_extractor.dm-186- if (N.amount <= 0) ./modules/hydroponics/seed_extractor.dm-187- piles -= N ./modules/hydroponics/seed_extractor.dm:188: ***del(N)*** ./modules/hydroponics/seed_extractor.dm-189- break ./modules/hydroponics/seed_extractor.dm-190- ./modules/hydroponics/seed_extractor.dm-191- for (var/obj/T in contents)//Now we find the seed we need to vend ./modules/hydroponics/seed_extractor.dm-192- var/obj/item/seeds/O = T ./modules/hydroponics/seed_extractor.dm-193- if (O.plantname == href_list["name"] && O.lifespan == href_list["li"] && O.endurance == href_list["en"] && O.maturation == href_list["ma"] && O.production == href_list["pr"] && O.yield == href_list["yi"] && O.potency == href_list["pot"]) *** ./modules/mob/camera/camera.dm-14-/mob/camera/experience_pressure_difference() ./modules/mob/camera/camera.dm-15- return ./modules/mob/camera/camera.dm-16- ./modules/mob/camera/camera.dm-17-/mob/camera/Destroy() ./modules/mob/camera/camera.dm-18- ..() ./modules/mob/camera/camera.dm:19: ***del(src)*** ./modules/mob/camera/camera.dm-20- ./modules/mob/camera/camera.dm-21-/mob/camera/Login() ./modules/mob/camera/camera.dm-22- ..() ./modules/mob/camera/camera.dm-23- update_interface() *** ./modules/mob/living/carbon/human/blood.dm-280- // Are we dripping or splattering? ./modules/mob/living/carbon/human/blood.dm-281- var/list/drips = list() ./modules/mob/living/carbon/human/blood.dm-282- // Only a certain number of drips (or one large splatter) can be on a given turf. ./modules/mob/living/carbon/human/blood.dm-283- for(var/obj/effect/decal/cleanable/blood/drip/drop in T) ./modules/mob/living/carbon/human/blood.dm-284- drips |= drop.drips ./modules/mob/living/carbon/human/blood.dm:285: ***del(drop)*** ./modules/mob/living/carbon/human/blood.dm-286- if(!large && drips.len < 3) ./modules/mob/living/carbon/human/blood.dm-287- decal_type = /obj/effect/decal/cleanable/blood/drip ./modules/mob/living/carbon/human/blood.dm-288- ./modules/mob/living/carbon/human/blood.dm-289- // Find a blood decal or create a new one. ./modules/mob/living/carbon/human/blood.dm-290- B = locate(decal_type) in T *** ./modules/mob/living/silicon/ai/ai.dm-703- ./modules/mob/living/silicon/ai/ai.dm-704- if(personnel_list.len) ./modules/mob/living/silicon/ai/ai.dm-705- input = input("Select a crew member:") as null|anything in personnel_list ./modules/mob/living/silicon/ai/ai.dm-706- var/icon/character_icon = personnel_list[input] ./modules/mob/living/silicon/ai/ai.dm-707- if(character_icon) ./modules/mob/living/silicon/ai/ai.dm:708: ***del(holo_icon)***//Clear old icon so we're not storing it in memory. ./modules/mob/living/silicon/ai/ai.dm-709- holo_icon = getHologramIcon(icon(character_icon)) ./modules/mob/living/silicon/ai/ai.dm-710- else ./modules/mob/living/silicon/ai/ai.dm-711- alert("No suitable records found. Aborting.") ./modules/mob/living/silicon/ai/ai.dm-712- ./modules/mob/living/silicon/ai/ai.dm-713- else *** ./modules/mob/living/silicon/ai/ai.dm-717- "xeno queen", ./modules/mob/living/silicon/ai/ai.dm-718- "space carp" ./modules/mob/living/silicon/ai/ai.dm-719- ) ./modules/mob/living/silicon/ai/ai.dm-720- input = input("Please select a hologram:") as null|anything in icon_list ./modules/mob/living/silicon/ai/ai.dm-721- if(input) ./modules/mob/living/silicon/ai/ai.dm:722: ***del(holo_icon)*** ./modules/mob/living/silicon/ai/ai.dm-723- switch(input) ./modules/mob/living/silicon/ai/ai.dm-724- if("default") ./modules/mob/living/silicon/ai/ai.dm-725- holo_icon = getHologramIcon(icon('icons/mob/AI.dmi',"holo1")) ./modules/mob/living/silicon/ai/ai.dm-726- if("floating face") ./modules/mob/living/silicon/ai/ai.dm-727- holo_icon = getHologramIcon(icon('icons/mob/AI.dmi',"holo2")) *** ./modules/mob/living/silicon/pai/death.dm-10- //var/tod = time2text(world.realtime,"hh:mm:ss") //weasellos time of death patch ./modules/mob/living/silicon/pai/death.dm-11- //mind.store_memory("Time of death: [tod]", 0) ./modules/mob/living/silicon/pai/death.dm-12- ./modules/mob/living/silicon/pai/death.dm-13- //New pAI's get a brand new mind to prevent meta stuff from their previous life. This new mind causes problems down the line if it's not deleted here. ./modules/mob/living/silicon/pai/death.dm-14- //Read as: I have no idea what I'm doing but asking for help got me nowhere so this is what you get. - Nodrak ./modules/mob/living/silicon/pai/death.dm:15: if(mind) ***del(mind)*** ./modules/mob/living/silicon/pai/death.dm-16- living_mob_list -= src ./modules/mob/living/silicon/pai/death.dm-17- ghostize() ./modules/mob/living/silicon/pai/death.dm-18- qdel(src) *** ./modules/mob/living/simple_animal/worm.dm-166- newHead.Attach(newHeadPrevious) ./modules/mob/living/simple_animal/worm.dm-167- ./modules/mob/living/simple_animal/worm.dm-168- if(die) ./modules/mob/living/simple_animal/worm.dm-169- newHead.Die() ./modules/mob/living/simple_animal/worm.dm-170- ./modules/mob/living/simple_animal/worm.dm:171: ***del(src)*** ./modules/mob/living/simple_animal/worm.dm-172- ./modules/mob/living/simple_animal/worm.dm-173- proc/ProcessStomach() ./modules/mob/living/simple_animal/worm.dm-174- for(var/atom/movable/stomachContent in contents) ./modules/mob/living/simple_animal/worm.dm-175- if(prob(digestionProbability)) ./modules/mob/living/simple_animal/worm.dm-176- if(istype(stomachContent,/obj/item/stack)) //converts to plasma, keeping the stack value ./modules/mob/living/simple_animal/worm.dm-177- if(!istype(stomachContent,/obj/item/stack/sheet/mineral/plasma)) ./modules/mob/living/simple_animal/worm.dm-178- var/obj/item/stack/oldStack = stomachContent ./modules/mob/living/simple_animal/worm.dm-179- new /obj/item/stack/sheet/mineral/plasma(src, oldStack.amount) ./modules/mob/living/simple_animal/worm.dm:180: ***del(oldStack)*** ./modules/mob/living/simple_animal/worm.dm-181- continue ./modules/mob/living/simple_animal/worm.dm-182- else if(istype(stomachContent,/obj/item)) //converts to plasma, keeping the w_class ./modules/mob/living/simple_animal/worm.dm-183- var/obj/item/oldItem = stomachContent ./modules/mob/living/simple_animal/worm.dm-184- new /obj/item/stack/sheet/mineral/plasma(src, oldItem.w_class) ./modules/mob/living/simple_animal/worm.dm:185: ***del(oldItem)*** ./modules/mob/living/simple_animal/worm.dm-186- continue ./modules/mob/living/simple_animal/worm.dm-187- else ./modules/mob/living/simple_animal/worm.dm-188- new /obj/item/stack/sheet/mineral/plasma(src, flatPlasmaValue) //just flat amount ./modules/mob/living/simple_animal/worm.dm:189: ***del(stomachContent)*** ./modules/mob/living/simple_animal/worm.dm-190- continue ./modules/mob/living/simple_animal/worm.dm-191- ./modules/mob/living/simple_animal/worm.dm-192- if(previous) ./modules/mob/living/simple_animal/worm.dm-193- for(var/atom/movable/stomachContent in contents) //transfer it along the digestive tract ./modules/mob/living/simple_animal/worm.dm-194- previous.contents += stomachContent *** ./modules/mob/login.dm-28- update_Login_details() ./modules/mob/login.dm-29- world.update_status() ./modules/mob/login.dm-30- ./modules/mob/login.dm-31- client.images = null //remove the images such as AIs being unable to see runes ./modules/mob/login.dm-32- client.screen = list() //remove hud items just in case ./modules/mob/login.dm:33: if(hud_used) ***del(hud_used)*** //remove the hud objects ./modules/mob/login.dm-34- hud_used = new /datum/hud(src) ./modules/mob/login.dm-35- ./modules/mob/login.dm-36- next_move = 1 ./modules/mob/login.dm-37- sight |= SEE_SELF ./modules/mob/login.dm-38- *** ./modules/mob/new_player/preferences_setup.dm-19- backbag = 1 ./modules/mob/new_player/preferences_setup.dm-20- features = random_features() ./modules/mob/new_player/preferences_setup.dm-21- age = rand(AGE_MIN,AGE_MAX) ./modules/mob/new_player/preferences_setup.dm-22- ./modules/mob/new_player/preferences_setup.dm-23-/datum/preferences/proc/update_preview_icon() //seriously. This is horrendous. ./modules/mob/new_player/preferences_setup.dm:24: ***del(preview_icon_front)*** ./modules/mob/new_player/preferences_setup.dm:25: ***del(preview_icon_side)*** ./modules/mob/new_player/preferences_setup.dm-26- var/icon/preview_icon = null ./modules/mob/new_player/preferences_setup.dm-27- ./modules/mob/new_player/preferences_setup.dm-28- if(job_engsec_high) //cyborg/AI check, put first to avoid so much unneeded blending ./modules/mob/new_player/preferences_setup.dm-29- switch(job_engsec_high) ./modules/mob/new_player/preferences_setup.dm-30- if(AI) *** ./modules/mob/new_player/preferences_setup.dm-35- ./modules/mob/new_player/preferences_setup.dm-36- if(preview_icon) //We're busting out! ./modules/mob/new_player/preferences_setup.dm-37- preview_icon_front = new(preview_icon, dir = SOUTH) ./modules/mob/new_player/preferences_setup.dm-38- preview_icon_side = new(preview_icon, dir = WEST) ./modules/mob/new_player/preferences_setup.dm-39- ./modules/mob/new_player/preferences_setup.dm:40: ***del(preview_icon)*** ./modules/mob/new_player/preferences_setup.dm-41- return ./modules/mob/new_player/preferences_setup.dm-42- ./modules/mob/new_player/preferences_setup.dm-43- var/g = "m" ./modules/mob/new_player/preferences_setup.dm-44- if(gender == FEMALE) g = "f" ./modules/mob/new_player/preferences_setup.dm-45- *** ./modules/mob/new_player/preferences_setup.dm-419- if(clothes_s) ./modules/mob/new_player/preferences_setup.dm-420- preview_icon.Blend(clothes_s, ICON_OVERLAY) ./modules/mob/new_player/preferences_setup.dm-421- preview_icon_front = new(preview_icon, dir = SOUTH) ./modules/mob/new_player/preferences_setup.dm-422- preview_icon_side = new(preview_icon, dir = WEST) ./modules/mob/new_player/preferences_setup.dm-423- ./modules/mob/new_player/preferences_setup.dm:424: ***del(preview_icon)*** ./modules/mob/new_player/preferences_setup.dm:425: ***del(eyes_s)*** ./modules/mob/new_player/preferences_setup.dm:426: ***del(clothes_s)*** *** ./modules/power/apc.dm-150- area.power_equip = 0 ./modules/power/apc.dm-151- area.power_environ = 0 ./modules/power/apc.dm-152- area.power_change() ./modules/power/apc.dm-153- if(occupier) ./modules/power/apc.dm-154- malfvacate(1) ./modules/power/apc.dm:155: ***del(wires)*** ./modules/power/apc.dm-156- if(cell) ./modules/power/apc.dm-157- qdel(cell) ./modules/power/apc.dm-158- if(terminal) ./modules/power/apc.dm-159- disconnect_terminal() ./modules/power/apc.dm-160- ..() *** ./modules/projectiles/firing.dm-32- if (!istype(targloc) || !istype(curloc) || !BB) ./modules/projectiles/firing.dm-33- return 0 ./modules/projectiles/firing.dm-34- if(targloc == curloc) ./modules/projectiles/firing.dm-35- if(BB.original == user) //if we target ourselves we go straight to bullet_act() ./modules/projectiles/firing.dm-36- user.bullet_act(BB) ./modules/projectiles/firing.dm:37: ***del(BB)*** ./modules/projectiles/firing.dm-38- return 1 ./modules/projectiles/firing.dm-39- BB.loc = get_turf(user) ./modules/projectiles/firing.dm-40- BB.starting = get_turf(user) ./modules/projectiles/firing.dm-41- BB.current = curloc ./modules/projectiles/firing.dm-42- BB.yo = targloc.y - curloc.y *** ./modules/projectiles/projectile/magic.dm-241- else ./modules/projectiles/projectile/magic.dm-242- new_mob.key = M.key ./modules/projectiles/projectile/magic.dm-243- ./modules/projectiles/projectile/magic.dm-244- new_mob << "<B>Your form morphs into that of a [randomize].</B>" ./modules/projectiles/projectile/magic.dm-245- ./modules/projectiles/projectile/magic.dm:246: ***del(M)*** ./modules/projectiles/projectile/magic.dm-247- return new_mob ./modules/projectiles/projectile/magic.dm-248- ./modules/projectiles/projectile/magic.dm-249-/obj/item/projectile/magic/animate ./modules/projectiles/projectile/magic.dm-250- name = "bolt of animation" ./modules/projectiles/projectile/magic.dm-251- icon_state = "red_1" *** ./modules/research/rdconsole.dm-58- for(var/T in typesof(/datum/tech) - /datum/tech) ./modules/research/rdconsole.dm-59- check_tech = null ./modules/research/rdconsole.dm-60- check_tech = new T() ./modules/research/rdconsole.dm-61- if(check_tech.id == ID) ./modules/research/rdconsole.dm-62- return_name = check_tech.name ./modules/research/rdconsole.dm:63: ***del(check_tech)*** ./modules/research/rdconsole.dm-64- check_tech = null ./modules/research/rdconsole.dm-65- break ./modules/research/rdconsole.dm-66- ./modules/research/rdconsole.dm-67- return return_name ./modules/research/rdconsole.dm-68- *** ./modules/research/rdconsole.dm-92- for(var/R in typesof(/datum/reagent) - /datum/reagent) ./modules/research/rdconsole.dm-93- temp_reagent = null ./modules/research/rdconsole.dm-94- temp_reagent = new R() ./modules/research/rdconsole.dm-95- if(temp_reagent.id == ID) ./modules/research/rdconsole.dm-96- return_name = temp_reagent.name ./modules/research/rdconsole.dm:97: ***del(temp_reagent)*** ./modules/research/rdconsole.dm-98- temp_reagent = null ./modules/research/rdconsole.dm-99- break ./modules/research/rdconsole.dm-100- return return_name ./modules/research/rdconsole.dm-101- ./modules/research/rdconsole.dm-102-/obj/machinery/computer/rdconsole/proc/SyncRDevices() //Makes sure it is properly sync'ed up with the devices attached to it (if any). *** ./modules/research/rdconsole.dm-573- else if(href_list["reset"]) //Reset the R&D console's database. ./modules/research/rdconsole.dm-574- griefProtection() ./modules/research/rdconsole.dm-575- var/choice = alert("R&D Console Database Reset", "Are you sure you want to reset the R&D console's database? Data lost cannot be recovered.", "Continue", "Cancel") ./modules/research/rdconsole.dm-576- if(choice == "Continue") ./modules/research/rdconsole.dm-577- screen = 0.0 ./modules/research/rdconsole.dm:578: ***del(files)*** ./modules/research/rdconsole.dm-579- files = new /datum/research(src) ./modules/research/rdconsole.dm-580- spawn(20) ./modules/research/rdconsole.dm-581- screen = 1.6 ./modules/research/rdconsole.dm-582- updateUsrDialog() ./modules/research/rdconsole.dm-583- *** ./_compile_options.dm-49-#ifdef dellogging ./_compile_options.dm-50-#warn compiling del logging. This will have additional overheads. //will warn you if compiling with dellogging ./_compile_options.dm-51-var/list/del_counter = list() ./_compile_options.dm-52-/proc/log_del(datum/X) ./_compile_options.dm-53- if(istype(X)){del_counter[X.type]++;} ./_compile_options.dm:54: ***del(X)*** ./_compile_options.dm:55:#define ***del(X)*** log_del(X) //overrides all del() calls with log_del() ./_compile_options.dm-56-#endif ./_compile_options.dm-57- ./_compile_options.dm-58-#ifdef TESTING ./_compile_options.dm-59-#warn compiling in TESTING mode. testing() debug messages will be visible. ./_compile_options.dm-60-#endif *** ./_onclick/hud/action.dm-43- ./_onclick/hud/action.dm-44-/datum/action/proc/Remove(mob/living/T) ./_onclick/hud/action.dm-45- if(button) ./_onclick/hud/action.dm-46- if(T.client) ./_onclick/hud/action.dm-47- T.client.screen -= button ./_onclick/hud/action.dm:48: ***del(button)*** ./_onclick/hud/action.dm-49- T.actions.Remove(src) ./_onclick/hud/action.dm-50- T.update_action_buttons() ./_onclick/hud/action.dm-51- owner = null ./_onclick/hud/action.dm-52- return ./_onclick/hud/action.dm-53- *** ./__HELPERS/icons.dm-862- var/image/text_image = new(loc = A) ./__HELPERS/icons.dm-863- text_image.maptext = "<font size = 4>[letter]</font>" ./__HELPERS/icons.dm-864- text_image.color = AverageColour(atom_icon) ./__HELPERS/icons.dm-865- text_image.pixel_x = 7 ./__HELPERS/icons.dm-866- text_image.pixel_y = 5 ./__HELPERS/icons.dm:867: ***del(atom_icon)*** ./__HELPERS/icons.dm-868- return text_image ./__HELPERS/icons.dm-869- ./__HELPERS/icons.dm-870- ./__HELPERS/icons.dm-871-//Find's the average colour of the icon ./__HELPERS/icons.dm-872-//By vg's ComicIronic
True
List of places in code not using qdel - ./ATMOSPHERICS/pipes/pipes.dm-14-/obj/machinery/atmospherics/proc/pipeline_expansion() ./ATMOSPHERICS/pipes/pipes.dm-15- return null ./ATMOSPHERICS/pipes/pipes.dm-16- ./ATMOSPHERICS/pipes/pipes.dm-17-/obj/machinery/atmospherics/pipe/proc/check_pressure(pressure) ./ATMOSPHERICS/pipes/pipes.dm-18- //Return 1 if parent should continue checking other pipes ./ATMOSPHERICS/pipes/pipes.dm:19: //Return null if parent should stop checking other pipes. Recall: ***del(src)*** will by default return null ./ATMOSPHERICS/pipes/pipes.dm-20- return 1 ./ATMOSPHERICS/pipes/pipes.dm-21- ./ATMOSPHERICS/pipes/pipes.dm-22-/obj/machinery/atmospherics/pipe/proc/releaseAirToTurf() ./ATMOSPHERICS/pipes/pipes.dm-23- if(air_temporary) ./ATMOSPHERICS/pipes/pipes.dm-24- var/turf/T = loc *** ./controllers/configuration.dm-176- modes += M.config_tag ./controllers/configuration.dm-177- mode_names[M.config_tag] = M.name ./controllers/configuration.dm-178- probabilities[M.config_tag] = M.probability ./controllers/configuration.dm-179- if(M.votable) ./controllers/configuration.dm-180- votable_modes += M.config_tag ./controllers/configuration.dm:181: ***del(M)*** ./controllers/configuration.dm-182- votable_modes += "secret" ./controllers/configuration.dm-183- ./controllers/configuration.dm-184-/datum/configuration/proc/load(filename, type = "config") //the type can also be game_options, in which case it uses a different switch. not making it separate to not copypaste code - Urist ./controllers/configuration.dm-185- var/list/Lines = file2list(filename) ./controllers/configuration.dm-186- *** ./controllers/configuration.dm-565- // their information, but it is the only way (at least that I know of). ./controllers/configuration.dm-566- for(var/T in (typesof(/datum/game_mode) - /datum/game_mode)) ./controllers/configuration.dm-567- var/datum/game_mode/M = new T() ./controllers/configuration.dm-568- if(M.config_tag && M.config_tag == mode_name) ./controllers/configuration.dm-569- return M ./controllers/configuration.dm:570: ***del(M)*** ./controllers/configuration.dm-571- return new /datum/game_mode/extended() ./controllers/configuration.dm-572- ./controllers/configuration.dm-573-/datum/configuration/proc/get_runnable_modes() ./controllers/configuration.dm-574- var/list/datum/game_mode/runnable_modes = new ./controllers/configuration.dm-575- for(var/T in (typesof(/datum/game_mode) - /datum/game_mode)) ./controllers/configuration.dm-576- var/datum/game_mode/M = new T() ./controllers/configuration.dm-577- //world << "DEBUG: [T], tag=[M.config_tag], prob=[probabilities[M.config_tag]]" ./controllers/configuration.dm-578- if(!(M.config_tag in modes)) ./controllers/configuration.dm:579: ***del(M)*** ./controllers/configuration.dm-580- continue ./controllers/configuration.dm-581- if(probabilities[M.config_tag]<=0) ./controllers/configuration.dm:582: ***del(M)*** ./controllers/configuration.dm-583- continue ./controllers/configuration.dm-584- if(M.can_start()) ./controllers/configuration.dm-585- runnable_modes[M] = probabilities[M.config_tag] ./controllers/configuration.dm-586- //world << "DEBUG: runnable_mode\[[runnable_modes.len]\] = [M.config_tag]" ./controllers/configuration.dm-587- return runnable_modes *** ./controllers/subsystem/garbage.dm-61- if(GCd_at_time > time_to_kill) ./controllers/subsystem/garbage.dm-62- break // Everything else is newer, skip them ./controllers/subsystem/garbage.dm-63- ./controllers/subsystem/garbage.dm-64- var/atom/A ./controllers/subsystem/garbage.dm-65- if (!istext(refID)) ./controllers/subsystem/garbage.dm:66: ***del(A)*** ./controllers/subsystem/garbage.dm-67- else ./controllers/subsystem/garbage.dm-68- A = locate(refID) ./controllers/subsystem/garbage.dm-69- if (A && A.gc_destroyed == GCd_at_time) // So if something else coincidently gets the same ref, it's not deleted by mistake ./controllers/subsystem/garbage.dm-70- // Something's still referring to the qdel'd object. Kill it. ./controllers/subsystem/garbage.dm-71- testing("GC: -- \ref[A] | [A.type] was unable to be GC'd and was deleted --") ./controllers/subsystem/garbage.dm-72- didntgc["[A.type]"]++ ./controllers/subsystem/garbage.dm:73: ***del(A)*** ./controllers/subsystem/garbage.dm-74- ++delslasttick ./controllers/subsystem/garbage.dm-75- ++totaldels ./controllers/subsystem/garbage.dm-76- else ./controllers/subsystem/garbage.dm-77- ++gcedlasttick ./controllers/subsystem/garbage.dm-78- ++totalgcs *** ./controllers/subsystem/garbage.dm-100-/proc/qdel(var/datum/A) ./controllers/subsystem/garbage.dm-101- if (!A) ./controllers/subsystem/garbage.dm-102- return ./controllers/subsystem/garbage.dm-103- if (!istype(A)) ./controllers/subsystem/garbage.dm-104- //warning("qdel() passed object of type [A.type]. qdel() can only handle /datum types.") ./controllers/subsystem/garbage.dm:105: ***del(A)*** ./controllers/subsystem/garbage.dm-106- else if (isnull(A.gc_destroyed)) ./controllers/subsystem/garbage.dm-107- // Let our friend know they're about to get fucked up. ./controllers/subsystem/garbage.dm-108- var/hint = A.Destroy() ./controllers/subsystem/garbage.dm-109- if (!A) ./controllers/subsystem/garbage.dm-110- return *** ./controllers/subsystem/garbage.dm-116- if (QDEL_HINT_IWILLGC) //functionally the same as the above. qdel should assume the object will gc on its own, and not check it. ./controllers/subsystem/garbage.dm-117- return ./controllers/subsystem/garbage.dm-118- if (QDEL_HINT_HARDDEL) //qdel should assume this object won't gc, and queue a hard delete using a hard reference to save time from the locate() ./controllers/subsystem/garbage.dm-119- SSgarbage.HardQueue(A) ./controllers/subsystem/garbage.dm-120- if (QDEL_HINT_HARDDEL_NOW) //qdel should assume this object won't gc, and hard del it post haste. ./controllers/subsystem/garbage.dm:121: ***del(A)*** ./controllers/subsystem/garbage.dm-122- if (QDEL_HINT_PUTINPOOL) //qdel will put this object in the pool. ./controllers/subsystem/garbage.dm-123- PlaceInPool(A,0) ./controllers/subsystem/garbage.dm-124- else ./controllers/subsystem/garbage.dm-125- SSgarbage.Queue(A) ./controllers/subsystem/garbage.dm-126- *** ./controllers/subsystem/garbage.dm-134- ./controllers/subsystem/garbage.dm-135-// Default implementation of clean-up code. ./controllers/subsystem/garbage.dm-136-// This should be overridden to remove all references pointing to the object being destroyed. ./controllers/subsystem/garbage.dm-137-// Return true if the the GC controller should allow the object to continue existing. (Useful if pooling objects.) ./controllers/subsystem/garbage.dm-138-/datum/proc/Destroy() ./controllers/subsystem/garbage.dm:139: //***del(src)*** ./controllers/subsystem/garbage.dm-140- tag = null ./controllers/subsystem/garbage.dm-141- return QDEL_HINT_QUEUE ./controllers/subsystem/garbage.dm-142- ./controllers/subsystem/garbage.dm-143-/datum/var/gc_destroyed //Time when this object was destroyed. ./controllers/subsystem/garbage.dm-144- *** ./controllers/subsystem/garbage.dm-192- set category = "Debug" ./controllers/subsystem/garbage.dm-193- if(garbage) ./controllers/subsystem/garbage.dm-194- while(garbage.destroyed.len) ./controllers/subsystem/garbage.dm-195- var/datum/o = locate(garbage.destroyed[1]) ./controllers/subsystem/garbage.dm-196- if(istype(o) && o.gc_destroyed) ./controllers/subsystem/garbage.dm:197: ***del(o)*** ./controllers/subsystem/garbage.dm-198- garbage.dels++ ./controllers/subsystem/garbage.dm-199- garbage.destroyed.Cut(1, 2) ./controllers/subsystem/garbage.dm-200-#endif *** ./controllers/subsystem/radio.dm-27- ./controllers/subsystem/radio.dm-28- if(frequency) ./controllers/subsystem/radio.dm-29- frequency.remove_listener(device) ./controllers/subsystem/radio.dm-30- ./controllers/subsystem/radio.dm-31- if(frequency.devices.len == 0) ./controllers/subsystem/radio.dm:32: ***del(frequency)*** ./controllers/subsystem/radio.dm-33- frequencies -= f_text ./controllers/subsystem/radio.dm-34- ./controllers/subsystem/radio.dm-35- return 1 ./controllers/subsystem/radio.dm-36- ./controllers/subsystem/radio.dm-37-/datum/subsystem/radio/proc/return_frequency(new_frequency as num) *** ./controllers/subsystem/server_maintenance.dm-9- for(var/client/C in clients) ./controllers/subsystem/server_maintenance.dm-10- if(C.is_afk(INACTIVITY_KICK)) ./controllers/subsystem/server_maintenance.dm-11- if(!istype(C.mob, /mob/dead)) ./controllers/subsystem/server_maintenance.dm-12- log_access("AFK: [key_name(C)]") ./controllers/subsystem/server_maintenance.dm-13- C << "<span class='danger'>You have been inactive for more than 10 minutes and have been disconnected.</span>" ./controllers/subsystem/server_maintenance.dm:14: ***del(C)*** ./controllers/subsystem/server_maintenance.dm-15- ./controllers/subsystem/server_maintenance.dm-16- if(config.sql_enabled) ./controllers/subsystem/server_maintenance.dm-17- sql_poll_players() ./controllers/subsystem/server_maintenance.dm-18- sql_poll_admins() *** ./controllers/subsystem/ticker.dm-135- ./controllers/subsystem/ticker.dm-136- else ./controllers/subsystem/ticker.dm-137- mode = config.pick_mode(master_mode) ./controllers/subsystem/ticker.dm-138- if(!mode.can_start()) ./controllers/subsystem/ticker.dm-139- world << "<B>Unable to start [mode.name].</B> Not enough players, [mode.required_players] players and [mode.required_enemies] eligible antagonists needed. Reverting to pre-game lobby." ./controllers/subsystem/ticker.dm:140: ***del(mode)*** ./controllers/subsystem/ticker.dm-141- SSjob.ResetOccupations() ./controllers/subsystem/ticker.dm-142- return 0 ./controllers/subsystem/ticker.dm-143- ./controllers/subsystem/ticker.dm-144- //Configure mode and assign player to special mode stuff ./controllers/subsystem/ticker.dm-145- var/can_continue = 0 ./controllers/subsystem/ticker.dm-146- can_continue = src.mode.pre_setup() //Choose antagonists ./controllers/subsystem/ticker.dm-147- SSjob.DivideOccupations() //Distribute jobs ./controllers/subsystem/ticker.dm-148- ./controllers/subsystem/ticker.dm-149- if(!Debug2) ./controllers/subsystem/ticker.dm-150- if(!can_continue) ./controllers/subsystem/ticker.dm:151: ***del(mode)*** ./controllers/subsystem/ticker.dm-152- world << "<B>Error setting up [master_mode].</B> Reverting to pre-game lobby." ./controllers/subsystem/ticker.dm-153- SSjob.ResetOccupations() ./controllers/subsystem/ticker.dm-154- return 0 ./controllers/subsystem/ticker.dm-155- else ./controllers/subsystem/ticker.dm-156- world << "<span class='notice'>DEBUG: Bypassing prestart checks..." *** ./controllers/subsystem/ticker.dm-287- flick("intro_nuke",cinematic) ./controllers/subsystem/ticker.dm-288- sleep(35) ./controllers/subsystem/ticker.dm-289- flick("station_intact",cinematic) ./controllers/subsystem/ticker.dm-290- world << sound('sound/ambience/signal.ogg') ./controllers/subsystem/ticker.dm-291- sleep(100) ./controllers/subsystem/ticker.dm:292: if(cinematic) ***del(cinematic)*** ./controllers/subsystem/ticker.dm:293: if(temp_buckle) ***del(temp_buckle)*** ./controllers/subsystem/ticker.dm-294- return //Faster exit, since nothing happened ./controllers/subsystem/ticker.dm-295- else //Station nuked (nuke,explosion,summary) ./controllers/subsystem/ticker.dm-296- flick("intro_nuke",cinematic) ./controllers/subsystem/ticker.dm-297- sleep(35) ./controllers/subsystem/ticker.dm-298- flick("station_explode_fade_red", cinematic) *** ./datums/ai_laws.dm-132- ./datums/ai_laws.dm-133-/datum/ai_laws/proc/add_ion_law(law) ./datums/ai_laws.dm-134- src.ion += law ./datums/ai_laws.dm-135- ./datums/ai_laws.dm-136-/datum/ai_laws/proc/clear_inherent_laws() ./datums/ai_laws.dm:137: ***del(src.inherent)*** ./datums/ai_laws.dm-138- src.inherent = list() ./datums/ai_laws.dm-139- ./datums/ai_laws.dm-140-/datum/ai_laws/proc/add_supplied_law(number, law) ./datums/ai_laws.dm-141- while (src.supplied.len < number + 1) ./datums/ai_laws.dm-142- src.supplied += "" *** ./datums/diseases/advance/advance.dm-105- if(affected_mob) ./datums/diseases/advance/advance.dm-106- var/id = "[GetDiseaseID()]" ./datums/diseases/advance/advance.dm-107- if(resistance && !(id in affected_mob.resistances)) ./datums/diseases/advance/advance.dm-108- affected_mob.resistances[id] = id ./datums/diseases/advance/advance.dm-109- remove_virus() ./datums/diseases/advance/advance.dm:110: ***del(src)*** //delete the datum to stop it processing ./datums/diseases/advance/advance.dm-111- ./datums/diseases/advance/advance.dm-112-// Returns the advance disease with a different reference memory. ./datums/diseases/advance/advance.dm-113-/datum/disease/advance/Copy(process = 0) ./datums/diseases/advance/advance.dm-114- return new /datum/disease/advance(process, src, 1) ./datums/diseases/advance/advance.dm-115- *** ./datums/diseases/_disease.dm-136- ./datums/diseases/_disease.dm-137- if(affected_mob) ./datums/diseases/_disease.dm-138- for(var/datum/disease/D in affected_mob.viruses) ./datums/diseases/_disease.dm-139- if(D != src) ./datums/diseases/_disease.dm-140- if(IsSame(D)) ./datums/diseases/_disease.dm:141: ***del(D)*** ./datums/diseases/_disease.dm-142- ./datums/diseases/_disease.dm-143- if(holder == affected_mob) ./datums/diseases/_disease.dm-144- if(affected_mob.stat != DEAD) ./datums/diseases/_disease.dm-145- stage_act() ./datums/diseases/_disease.dm-146- *** ./datums/diseases/_disease.dm-154- if(affected_mob) ./datums/diseases/_disease.dm-155- if(disease_flags & CAN_RESIST) ./datums/diseases/_disease.dm-156- if(!(type in affected_mob.resistances)) ./datums/diseases/_disease.dm-157- affected_mob.resistances += type ./datums/diseases/_disease.dm-158- remove_virus() ./datums/diseases/_disease.dm:159: ***del(src)*** ./datums/diseases/_disease.dm-160- ./datums/diseases/_disease.dm-161- ./datums/diseases/_disease.dm-162-/datum/disease/New() ./datums/diseases/_disease.dm-163- if(required_organs && required_organs.len) ./datums/diseases/_disease.dm-164- if(ishuman(affected_mob)) *** ./datums/gas_mixture.dm-319- if(!corresponding) ./datums/gas_mixture.dm-320- corresponding = new trace_gas.type() ./datums/gas_mixture.dm-321- trace_gases += corresponding ./datums/gas_mixture.dm-322- corresponding.moles += trace_gas.moles ./datums/gas_mixture.dm-323- ./datums/gas_mixture.dm:324:// ***del(giver)*** ./datums/gas_mixture.dm-325- return 1 ./datums/gas_mixture.dm-326- ./datums/gas_mixture.dm-327-/datum/gas_mixture/remove(amount) ./datums/gas_mixture.dm-328- ./datums/gas_mixture.dm-329- var/sum = total_moles() *** ./datums/modules.dm-22- var/type = O.type // the type of the creating object ./datums/modules.dm-23- ./datums/modules.dm-24- var/mneed = mods.inmodlist(type) // find if this type has modules defined ./datums/modules.dm-25- ./datums/modules.dm-26- if(!mneed) // not found in module list? ./datums/modules.dm:27: ***del(src)*** // delete self, thus ending proc ./datums/modules.dm-28- ./datums/modules.dm-29- var/needed = mods.getbitmask(type) // get a bitmask for the number of modules in this object ./datums/modules.dm-30- status = needed ./datums/modules.dm-31- installed = needed ./datums/modules.dm-32- *** ./game/area/areas.dm-389- if(!AM.anchored) ./game/area/areas.dm-390- AM.Move(D, SOUTH) ./game/area/areas.dm-391- else ./game/area/areas.dm-392- qdel(AM) ./game/area/areas.dm-393- if(istype(T, /turf/simulated)) ./game/area/areas.dm:394: ***del(T)*** ./game/area/areas.dm-395- ./game/area/areas.dm-396- /*for(var/atom/movable/bug in src) // If someone (or something) is somehow still in the shuttle's docking area... ./game/area/areas.dm-397- if(ismob(bug)) ./game/area/areas.dm-398- continue ./game/area/areas.dm-399- qdel(bug)*/ *** ./game/atoms.dm-57- ./game/atoms.dm-58-/atom/proc/CheckParts() ./game/atoms.dm-59- return ./game/atoms.dm-60- ./game/atoms.dm-61-/atom/proc/assume_air(datum/gas_mixture/giver) ./game/atoms.dm:62: ***del(giver)*** ./game/atoms.dm-63- return null ./game/atoms.dm-64- ./game/atoms.dm-65-/atom/proc/remove_air(amount) ./game/atoms.dm-66- return null ./game/atoms.dm-67- *** ./game/gamemodes/cult/runes.dm-1000- "\red You are blinded by the flash of red light! After you're able to see again, you see that you are now wearing a set of armor.") ./game/gamemodes/cult/runes.dm-1001- if (armorworn == "summoner" || armorworn == "trickster" || armorworn == "physician") ./game/gamemodes/cult/runes.dm-1002- usr.say("Uln Shogg Hafh[pick("'","`")]drn!") ./game/gamemodes/cult/runes.dm-1003- user << "\red You quietly prick your finger and make a pact with the geometer of blood to acquire more power." ./game/gamemodes/cult/runes.dm-1004- user.take_overall_damage(rand(5,20)) ./game/gamemodes/cult/runes.dm:1005: ***del(user.head)*** ./game/gamemodes/cult/runes.dm:1006: ***del(user.wear_suit)*** ./game/gamemodes/cult/runes.dm-1007- user.equip_to_slot_or_del(new /obj/item/clothing/head/magus(user), slot_head) ./game/gamemodes/cult/runes.dm-1008- user.equip_to_slot_or_del(new /obj/item/clothing/suit/magusred(user), slot_wear_suit) ./game/gamemodes/cult/runes.dm-1009- user.equip_to_slot_or_del(new /obj/item/clothing/shoes/cult(user), slot_shoes) ./game/gamemodes/cult/runes.dm-1010- user.spellremove(user) ./game/gamemodes/cult/runes.dm-1011- usr.visible_message("\red The rune disappears with a flash of red light, and a set of robes appears on [usr]...", \ *** ./game/gamemodes/cult/runes.dm-1018- user.mind.spell_list += new /obj/effect/proc_holder/spell/targeted/turf_teleport/blink/cult(user) ./game/gamemodes/cult/runes.dm-1019- if (armorworn == "physician") ./game/gamemodes/cult/runes.dm-1020- user.put_in_hands(new /obj/item/weapon/gun/magic/wand/resurrection(user)) ./game/gamemodes/cult/runes.dm-1021- if (armorworn == "traveler" || armorworn == "marauder") ./game/gamemodes/cult/runes.dm-1022- usr.say("Tharanak n[pick("'","`")]ghft!") ./game/gamemodes/cult/runes.dm:1023: ***del(user.head)*** ./game/gamemodes/cult/runes.dm:1024: ***del(user.wear_suit)*** ./game/gamemodes/cult/runes.dm-1025- user.equip_to_slot_or_del(new /obj/item/clothing/head/helmet/space/cult(user), slot_head) ./game/gamemodes/cult/runes.dm-1026- user.equip_to_slot_or_del(new /obj/item/clothing/suit/space/cult(user), slot_wear_suit) ./game/gamemodes/cult/runes.dm-1027- user.equip_to_slot_or_del(new /obj/item/clothing/shoes/cult(user), slot_shoes) ./game/gamemodes/cult/runes.dm-1028- user.equip_to_slot_or_del(new /obj/item/weapon/storage/backpack/cultpack(user), slot_back) ./game/gamemodes/cult/runes.dm-1029- user.put_in_hands(new /obj/item/weapon/melee/cultblade(user)) *** ./game/gamemodes/cult/runes.dm-1031- usr.visible_message("\red The rune disappears with a flash of red light, and an armored space suit appears on [usr]...", \ ./game/gamemodes/cult/runes.dm-1032- "\red You are blinded by the flash of red light! After you're able to see again, you see that you are now wearing an armored space suit.") ./game/gamemodes/cult/runes.dm-1033- if (armorworn == "marauder") ./game/gamemodes/cult/runes.dm-1034- user.spellremove(user) ./game/gamemodes/cult/runes.dm-1035- user.mind.spell_list += new /obj/effect/proc_holder/spell/aoe_turf/conjure/creature/cult(user) ./game/gamemodes/cult/runes.dm:1036: ***del(src)*** ./game/gamemodes/cult/runes.dm-1037- return ./game/gamemodes/cult/runes.dm-1038- ./game/gamemodes/cult/runes.dm-1039-///Summon Shell: Summons a construct shell if there's four plasteel sheets on top of the rune ./game/gamemodes/cult/runes.dm-1040- ./game/gamemodes/cult/runes.dm-1041- *** ./game/gamemodes/game_mode.dm-112- var/list/datum/game_mode/usable_modes = list() ./game/gamemodes/game_mode.dm-113- for(var/datum/game_mode/G in runnable_modes) ./game/gamemodes/game_mode.dm-114- if(G.reroll_friendly) ./game/gamemodes/game_mode.dm-115- usable_modes += G ./game/gamemodes/game_mode.dm-116- else ./game/gamemodes/game_mode.dm:117: ***del(G)*** ./game/gamemodes/game_mode.dm-118- ./game/gamemodes/game_mode.dm-119- if(!usable_modes) ./game/gamemodes/game_mode.dm-120- message_admins("Convert_roundtype failed due to no valid modes to convert to. Please report this error to the Coders.") ./game/gamemodes/game_mode.dm-121- return null ./game/gamemodes/game_mode.dm-122- *** ./game/gamemodes/wizard/spellbook.dm-76- var/spell_levels = 0 ./game/gamemodes/wizard/spellbook.dm-77- for(var/obj/effect/proc_holder/spell/aspell in user.mind.spell_list) ./game/gamemodes/wizard/spellbook.dm-78- if(initial(S.name) == initial(aspell.name)) ./game/gamemodes/wizard/spellbook.dm-79- spell_levels = aspell.spell_level ./game/gamemodes/wizard/spellbook.dm-80- user.mind.spell_list.Remove(aspell) ./game/gamemodes/wizard/spellbook.dm:81: ***del(S)*** ./game/gamemodes/wizard/spellbook.dm-82- return cost * (spell_levels+1) ./game/gamemodes/wizard/spellbook.dm-83- return -1 ./game/gamemodes/wizard/spellbook.dm-84-/datum/spellbook_entry/proc/GetInfo() ./game/gamemodes/wizard/spellbook.dm-85- if(!S) ./game/gamemodes/wizard/spellbook.dm-86- S = new spell_type() *** ./game/gamemodes/wizard/spellbook.dm-417- var/datum/spellbook_entry/E = new T ./game/gamemodes/wizard/spellbook.dm-418- if(E.IsAvailible()) ./game/gamemodes/wizard/spellbook.dm-419- entries |= E ./game/gamemodes/wizard/spellbook.dm-420- categories |= E.category ./game/gamemodes/wizard/spellbook.dm-421- else ./game/gamemodes/wizard/spellbook.dm:422: ***del(E)*** ./game/gamemodes/wizard/spellbook.dm-423- tab = categories[1] ./game/gamemodes/wizard/spellbook.dm-424- ./game/gamemodes/wizard/spellbook.dm-425-/obj/item/weapon/spellbook/New() ./game/gamemodes/wizard/spellbook.dm-426- ..() ./game/gamemodes/wizard/spellbook.dm-427- Initialize() *** ./game/machinery/computer/medical.dm-579- if(6) ./game/machinery/computer/medical.dm-580- R.fields["m_stat"] = pick("*Insane*", "*Unstable*", "*Watch*", "Stable") ./game/machinery/computer/medical.dm-581- continue ./game/machinery/computer/medical.dm-582- ./game/machinery/computer/medical.dm-583- else if(prob(1)) ./game/machinery/computer/medical.dm:584: ***del(R)*** ./game/machinery/computer/medical.dm-585- continue ./game/machinery/computer/medical.dm-586- ./game/machinery/computer/medical.dm-587- ..(severity) ./game/machinery/computer/medical.dm-588- ./game/machinery/computer/medical.dm-589-/obj/machinery/computer/med_data/proc/canUseMedicalRecordsConsole(mob/user, message = 1, record1, record2) *** ./game/machinery/computer/security.dm-416- temp += "<a href='?src=\ref[src];choice=Clear Screen'>No</a>" ./game/machinery/computer/security.dm-417- ./game/machinery/computer/security.dm-418- if("Purge All Records") ./game/machinery/computer/security.dm-419- investigate_log("[usr.name] ([usr.key]) has purged all the security records.", "records") ./game/machinery/computer/security.dm-420- for(var/datum/data/record/R in data_core.security) ./game/machinery/computer/security.dm:421: ***del(R)*** ./game/machinery/computer/security.dm-422- data_core.security.Cut() ./game/machinery/computer/security.dm-423- temp = "All Security records deleted." ./game/machinery/computer/security.dm-424- ./game/machinery/computer/security.dm-425- if("Add Entry") ./game/machinery/computer/security.dm-426- if(!( istype(active2, /datum/data/record) )) *** ./game/machinery/computer/security.dm-665- H.sec_hud_set_security_status() ./game/machinery/computer/security.dm-666- if("Delete Record (Security) Execute") ./game/machinery/computer/security.dm-667- investigate_log("[usr.name] ([usr.key]) has deleted the security records for [active1.fields["name"]].", "records") ./game/machinery/computer/security.dm-668- if(active2) ./game/machinery/computer/security.dm-669- data_core.security -= active2 ./game/machinery/computer/security.dm:670: ***del(active2)*** ./game/machinery/computer/security.dm-671- ./game/machinery/computer/security.dm-672- if("Delete Record (ALL) Execute") ./game/machinery/computer/security.dm-673- if(active1) ./game/machinery/computer/security.dm-674- investigate_log("[usr.name] ([usr.key]) has deleted all records for [active1.fields["name"]].", "records") ./game/machinery/computer/security.dm-675- for(var/datum/data/record/R in data_core.medical) ./game/machinery/computer/security.dm-676- if((R.fields["name"] == active1.fields["name"] || R.fields["id"] == active1.fields["id"])) ./game/machinery/computer/security.dm-677- data_core.medical -= R ./game/machinery/computer/security.dm:678: ***del(R)*** ./game/machinery/computer/security.dm-679- break ./game/machinery/computer/security.dm-680- data_core.general -= active1 ./game/machinery/computer/security.dm:681: ***del(active1)*** ./game/machinery/computer/security.dm-682- ./game/machinery/computer/security.dm-683- if(active2) ./game/machinery/computer/security.dm-684- data_core.security -= active2 ./game/machinery/computer/security.dm:685: ***del(active2)*** ./game/machinery/computer/security.dm-686- else ./game/machinery/computer/security.dm-687- temp = "This function does not appear to be working at the moment. Our apologies." ./game/machinery/computer/security.dm-688- ./game/machinery/computer/security.dm-689- add_fingerprint(usr) ./game/machinery/computer/security.dm-690- updateUsrDialog() *** ./game/machinery/computer/security.dm-732- R.fields["photo_front"] = G.fields["photo_front"] ./game/machinery/computer/security.dm-733- R.fields["photo_side"] = G.fields["photo_side"] ./game/machinery/computer/security.dm-734- continue ./game/machinery/computer/security.dm-735- ./game/machinery/computer/security.dm-736- else if(prob(1)) ./game/machinery/computer/security.dm:737: ***del(R)*** ./game/machinery/computer/security.dm-738- continue ./game/machinery/computer/security.dm-739- ./game/machinery/computer/security.dm-740- ..(severity) ./game/machinery/computer/security.dm-741- ./game/machinery/computer/security.dm-742-/obj/machinery/computer/secure_data/proc/canUseSecurityRecordsConsole(mob/user, message1 = 0, record1, record2) *** ./game/machinery/embedded_controller/embedded_controller_base.dm-5- ./game/machinery/embedded_controller/embedded_controller_base.dm-6-/datum/computer/file/embedded_program/proc/post_signal(datum/signal/signal, comm_line) ./game/machinery/embedded_controller/embedded_controller_base.dm-7- if(master) ./game/machinery/embedded_controller/embedded_controller_base.dm-8- master.post_signal(signal, comm_line) ./game/machinery/embedded_controller/embedded_controller_base.dm-9- else ./game/machinery/embedded_controller/embedded_controller_base.dm:10: ***del(signal)*** ./game/machinery/embedded_controller/embedded_controller_base.dm-11- ./game/machinery/embedded_controller/embedded_controller_base.dm-12-/datum/computer/file/embedded_program/proc/receive_user_command(command) ./game/machinery/embedded_controller/embedded_controller_base.dm-13- ./game/machinery/embedded_controller/embedded_controller_base.dm-14-/datum/computer/file/embedded_program/proc/receive_signal(datum/signal/signal, receive_method, receive_param) ./game/machinery/embedded_controller/embedded_controller_base.dm-15- return null *** ./game/machinery/magnet.dm-373- ./game/machinery/magnet.dm-374- if(!(nextmove in list("N","S","E","W","C","R"))) ./game/machinery/magnet.dm-375- // N, S, E, W are directional ./game/machinery/magnet.dm-376- // C is center ./game/machinery/magnet.dm-377- // R is random (in magnetic field's bounds) ./game/machinery/magnet.dm:378: ***del(signal)*** ./game/machinery/magnet.dm-379- break // break the loop if the character located is invalid ./game/machinery/magnet.dm-380- ./game/machinery/magnet.dm-381- signal.data["command"] = nextmove ./game/machinery/magnet.dm-382- ./game/machinery/magnet.dm-383- *** ./game/machinery/overview.dm-172- var/icon/J = imap[i*2+2] ./game/machinery/overview.dm-173- ./game/machinery/overview.dm-174- HI.Insert(I, frame=1, delay = 5) ./game/machinery/overview.dm-175- HI.Insert(J, frame=2, delay = 5) ./game/machinery/overview.dm-176- ./game/machinery/overview.dm:177: ***del(I)*** ./game/machinery/overview.dm:178: ***del(J)*** ./game/machinery/overview.dm-179- H.icon = HI ./game/machinery/overview.dm-180- H.layer = 25 ./game/machinery/overview.dm-181- usr.mapobjs += H ./game/machinery/overview.dm-182-#else ./game/machinery/overview.dm-183- *** ./game/machinery/overview.dm-292- H.name = (i==0)?"maprefresh":"map" ./game/machinery/overview.dm-293- ./game/machinery/overview.dm-294- var/icon/I = imap[i+1] ./game/machinery/overview.dm-295- ./game/machinery/overview.dm-296- H.icon = I ./game/machinery/overview.dm:297: ***del(I)*** ./game/machinery/overview.dm-298- H.layer = 25 ./game/machinery/overview.dm-299- usr.mapobjs += H ./game/machinery/overview.dm-300- ./game/machinery/overview.dm-301-#endif ./game/machinery/overview.dm-302- *** ./game/machinery/suit_storage_unit.dm-391- if(i==3) //End of the cycle ./game/machinery/suit_storage_unit.dm-392- if(!src.issuperUV) ./game/machinery/suit_storage_unit.dm-393- for(var/obj/item/ITEM in src) ./game/machinery/suit_storage_unit.dm-394- ITEM.clean_blood() ./game/machinery/suit_storage_unit.dm-395- if(istype(STORAGE, /obj/item/weapon/reagent_containers/food)) ./game/machinery/suit_storage_unit.dm:396: ***del(STORAGE)*** ./game/machinery/suit_storage_unit.dm-397- else //It was supercycling, destroy everything ./game/machinery/suit_storage_unit.dm-398- src.HELMET = null ./game/machinery/suit_storage_unit.dm-399- src.SUIT = null ./game/machinery/suit_storage_unit.dm-400- src.MASK = null ./game/machinery/suit_storage_unit.dm:401: ***del(STORAGE)*** ./game/machinery/suit_storage_unit.dm-402- visible_message("<font color='red'>With a loud whining noise, the Suit Storage Unit's door grinds open. Puffs of ashen smoke come out of its chamber.</font>", 3) ./game/machinery/suit_storage_unit.dm-403- src.isbroken = 1 ./game/machinery/suit_storage_unit.dm-404- src.isopen = 1 ./game/machinery/suit_storage_unit.dm-405- src.islocked = 0 ./game/machinery/suit_storage_unit.dm-406- src.eject_occupant(OCCUPANT) //Mixing up these two lines causes bug. DO NOT DO IT. *** ./game/machinery/telecomms/computers/logbrowser.dm-195- var/datum/comm_log_entry/D = SelectedServer.log_entries[text2num(href_list["delete"])] ./game/machinery/telecomms/computers/logbrowser.dm-196- ./game/machinery/telecomms/computers/logbrowser.dm-197- temp = "<font color = #336699>- DELETED ENTRY: [D.name] -</font color>" ./game/machinery/telecomms/computers/logbrowser.dm-198- ./game/machinery/telecomms/computers/logbrowser.dm-199- SelectedServer.log_entries.Remove(D) ./game/machinery/telecomms/computers/logbrowser.dm:200: ***del(D)*** ./game/machinery/telecomms/computers/logbrowser.dm-201- ./game/machinery/telecomms/computers/logbrowser.dm-202- else ./game/machinery/telecomms/computers/logbrowser.dm-203- temp = "<font color = #D70B00>- FAILED: NO SELECTED MACHINE -</font color>" ./game/machinery/telecomms/computers/logbrowser.dm-204- ./game/machinery/telecomms/computers/logbrowser.dm-205- if(href_list["network"]) *** ./game/mecha/mecha.dm-135- internal_tank = null ./game/mecha/mecha.dm-136- if(loc) ./game/mecha/mecha.dm-137- loc.assume_air(cabin_air) ./game/mecha/mecha.dm-138- air_update_turf() ./game/mecha/mecha.dm-139- else ./game/mecha/mecha.dm:140: ***del(cabin_air)*** ./game/mecha/mecha.dm-141- cabin_air = null ./game/mecha/mecha.dm-142- qdel(spark_system) ./game/mecha/mecha.dm-143- spark_system = null ./game/mecha/mecha.dm-144- ./game/mecha/mecha.dm-145- mechas_list -= src //global mech list *** ./game/mecha/mecha.dm-238- var/datum/gas_mixture/leaked_gas = int_tank_air.remove_ratio(0.10) ./game/mecha/mecha.dm-239- if(loc) ./game/mecha/mecha.dm-240- loc.assume_air(leaked_gas) ./game/mecha/mecha.dm-241- air_update_turf() ./game/mecha/mecha.dm-242- else ./game/mecha/mecha.dm:243: ***del(leaked_gas)*** ./game/mecha/mecha.dm-244- ./game/mecha/mecha.dm-245- if(internal_damage & MECHA_INT_SHORT_CIRCUIT) ./game/mecha/mecha.dm-246- if(get_charge()) ./game/mecha/mecha.dm-247- spark_system.start() ./game/mecha/mecha.dm-248- cell.charge -= min(20,cell.charge) *** ./game/mecha/mecha.dm-274- transfer_moles = pressure_delta*cabin_air.return_volume()/(cabin_air.return_temperature() * R_IDEAL_GAS_EQUATION) ./game/mecha/mecha.dm-275- var/datum/gas_mixture/removed = cabin_air.remove(transfer_moles) ./game/mecha/mecha.dm-276- if(t_air) ./game/mecha/mecha.dm-277- t_air.merge(removed) ./game/mecha/mecha.dm-278- else //just delete the cabin gas, we're in space or some shit ./game/mecha/mecha.dm:279: ***del(removed)*** ./game/mecha/mecha.dm-280- ./game/mecha/mecha.dm-281- if(occupant) ./game/mecha/mecha.dm-282- if(cell) ./game/mecha/mecha.dm-283- var/cellcharge = cell.charge/cell.maxcharge ./game/mecha/mecha.dm-284- switch(cellcharge) *** ./game/mecha/mecha_construction_paths.dm-298- var/obj/item/mecha_parts/chassis/const_holder = holder ./game/mecha/mecha_construction_paths.dm-299- const_holder.construct = new /datum/construction/reversible/mecha/gygax(const_holder) ./game/mecha/mecha_construction_paths.dm-300- const_holder.icon = 'icons/mecha/mech_construction.dmi' ./game/mecha/mecha_construction_paths.dm-301- const_holder.icon_state = "gygax0" ./game/mecha/mecha_construction_paths.dm-302- const_holder.density = 1 ./game/mecha/mecha_construction_paths.dm:303: ***del(src)*** ./game/mecha/mecha_construction_paths.dm-304- return ./game/mecha/mecha_construction_paths.dm-305- ./game/mecha/mecha_construction_paths.dm-306- ./game/mecha/mecha_construction_paths.dm-307-/datum/construction/reversible/mecha/gygax ./game/mecha/mecha_construction_paths.dm-308- result = "/obj/mecha/combat/gygax" *** ./game/mecha/mecha_construction_paths.dm-577- var/obj/item/mecha_parts/chassis/const_holder = holder ./game/mecha/mecha_construction_paths.dm-578- const_holder.construct = new /datum/construction/reversible/mecha/firefighter(const_holder) ./game/mecha/mecha_construction_paths.dm-579- const_holder.icon = 'icons/mecha/mech_construction.dmi' ./game/mecha/mecha_construction_paths.dm-580- const_holder.icon_state = "fireripley0" ./game/mecha/mecha_construction_paths.dm-581- const_holder.density = 1 ./game/mecha/mecha_construction_paths.dm:582: ***del(src)*** ./game/mecha/mecha_construction_paths.dm-583- return ./game/mecha/mecha_construction_paths.dm-584- ./game/mecha/mecha_construction_paths.dm-585- ./game/mecha/mecha_construction_paths.dm-586-/datum/construction/reversible/mecha/firefighter ./game/mecha/mecha_construction_paths.dm-587- result = "/obj/mecha/working/ripley/firefighter" *** ./game/mecha/mecha_construction_paths.dm-799- ./game/mecha/mecha_construction_paths.dm-800-/datum/construction/mecha/honker_chassis/spawn_result() ./game/mecha/mecha_construction_paths.dm-801- var/obj/item/mecha_parts/chassis/const_holder = holder ./game/mecha/mecha_construction_paths.dm-802- const_holder.construct = new /datum/construction/mecha/honker(const_holder) ./game/mecha/mecha_construction_paths.dm-803- const_holder.density = 1 ./game/mecha/mecha_construction_paths.dm:804: ***del(src)*** ./game/mecha/mecha_construction_paths.dm-805- return ./game/mecha/mecha_construction_paths.dm-806- ./game/mecha/mecha_construction_paths.dm-807- ./game/mecha/mecha_construction_paths.dm-808-/datum/construction/mecha/honker ./game/mecha/mecha_construction_paths.dm-809- result = "/obj/mecha/combat/honker" *** ./game/objects/effects/mines.dm-54- name = "kick mine" ./game/objects/effects/mines.dm-55- ./game/objects/effects/mines.dm-56-/obj/effect/mine/kickmine/mineEffect(mob/victim) ./game/objects/effects/mines.dm-57- if(isliving(victim) && victim.client) ./game/objects/effects/mines.dm-58- victim << "<span class='userdanger'>You have been kicked FOR NO REISIN!</span>" ./game/objects/effects/mines.dm:59: ***del(victim.client)*** ./game/objects/effects/mines.dm-60- ./game/objects/effects/mines.dm-61- ./game/objects/effects/mines.dm-62-/obj/effect/mine/gas ./game/objects/effects/mines.dm-63- name = "oxygen mine" ./game/objects/effects/mines.dm-64- var/gas_amount = 360 *** ./game/objects/items/devices/camera_bug.dm-46- ./game/objects/items/devices/camera_bug.dm-47-/obj/item/device/camera_bug/Destroy() ./game/objects/items/devices/camera_bug.dm-48- if(expansion) ./game/objects/items/devices/camera_bug.dm-49- qdel(expansion) ./game/objects/items/devices/camera_bug.dm-50- expansion = null ./game/objects/items/devices/camera_bug.dm:51: ***del(src)*** ./game/objects/items/devices/camera_bug.dm-52-//Easier to just call del() than this nonsense ./game/objects/items/devices/camera_bug.dm-53-// ya no, del() takes 0.8ds to run on avg. this takes less than 0.01 ./game/objects/items/devices/camera_bug.dm-54- get_cameras() ./game/objects/items/devices/camera_bug.dm-55- for(var/cam_tag in bugged_cameras) ./game/objects/items/devices/camera_bug.dm-56- var/obj/machinery/camera/camera = bugged_cameras[cam_tag] *** ./game/objects/items/holotape.dm-242- while(N != 1) ./game/objects/items/holotape.dm-243- N = 1 ./game/objects/items/holotape.dm-244- for (var/obj/item/holotape/P in cur) ./game/objects/items/holotape.dm-245- if(P.icon_state == icon_dir) ./game/objects/items/holotape.dm-246- N = 0 ./game/objects/items/holotape.dm:247: ***del(P)*** ./game/objects/items/holotape.dm-248- cur = get_step(cur,dir[i]) ./game/objects/items/holotape.dm-249- ./game/objects/items/holotape.dm:250: ***del(src)*** ./game/objects/items/holotape.dm-251- return ./game/objects/items/holotape.dm-252- ./game/objects/items/holotape.dm-253-#undef MAX_TAPE_RANGE *** ./game/objects/items/weapons/storage/storage.dm-194- var/obj/item/sample_object ./game/objects/items/weapons/storage/storage.dm-195- var/number ./game/objects/items/weapons/storage/storage.dm-196- ./game/objects/items/weapons/storage/storage.dm-197- New(obj/item/sample) ./game/objects/items/weapons/storage/storage.dm-198- if(!istype(sample)) ./game/objects/items/weapons/storage/storage.dm:199: ***del(src)*** ./game/objects/items/weapons/storage/storage.dm-200- sample_object = sample ./game/objects/items/weapons/storage/storage.dm-201- number = 1 ./game/objects/items/weapons/storage/storage.dm-202- ./game/objects/items/weapons/storage/storage.dm-203- ./game/objects/items/weapons/storage/storage.dm-204-//This proc determins the size of the inventory to be displayed. Please touch it only if you know what you're doing. *** ./game/objects/items/weapons/tanks/tanks.dm-31- ./game/objects/items/weapons/tanks/tanks.dm-32- return ./game/objects/items/weapons/tanks/tanks.dm-33- ./game/objects/items/weapons/tanks/tanks.dm-34-/obj/item/weapon/tank/Destroy() ./game/objects/items/weapons/tanks/tanks.dm-35- if(air_contents) ./game/objects/items/weapons/tanks/tanks.dm:36: ***del(air_contents)*** ./game/objects/items/weapons/tanks/tanks.dm-37- ./game/objects/items/weapons/tanks/tanks.dm-38- SSobj.processing.Remove(src) ./game/objects/items/weapons/tanks/tanks.dm-39- ./game/objects/items/weapons/tanks/tanks.dm-40- ..() ./game/objects/items/weapons/tanks/tanks.dm-41- *** ./LINDA/LINDA_turf_tile.dm-6- var/atmos_adjacent_turfs = 0 ./LINDA/LINDA_turf_tile.dm-7- var/atmos_adjacent_turfs_amount = 0 ./LINDA/LINDA_turf_tile.dm-8- var/atmos_supeconductivity = 0 ./LINDA/LINDA_turf_tile.dm-9- ./LINDA/LINDA_turf_tile.dm-10-/turf/assume_air(datum/gas_mixture/giver) //use this for machines to adjust air ./LINDA/LINDA_turf_tile.dm:11: ***del(giver)*** ./LINDA/LINDA_turf_tile.dm-12- return 0 ./LINDA/LINDA_turf_tile.dm-13- ./LINDA/LINDA_turf_tile.dm-14-/turf/return_air() ./LINDA/LINDA_turf_tile.dm-15- //Create gas mixture to hold data for passing ./LINDA/LINDA_turf_tile.dm-16- var/datum/gas_mixture/GM = new *** ./modules/admin/admin.dm-767- if(kick_only_afk && !C.is_afk()) //Ignore clients who are not afk ./modules/admin/admin.dm-768- continue ./modules/admin/admin.dm-769- if(message) ./modules/admin/admin.dm-770- C << message ./modules/admin/admin.dm-771- kicked_client_names.Add("[C.ckey]") ./modules/admin/admin.dm:772: ***del(C)*** ./modules/admin/admin.dm-773- return kicked_client_names ./modules/admin/admin.dm-774- ./modules/admin/admin.dm-775-//returns 1 to let the dragdrop code know we are trapping this event ./modules/admin/admin.dm-776-//returns 0 if we don't plan to trap the event ./modules/admin/admin.dm-777-/datum/admins/proc/cmd_ghost_drag(mob/dead/observer/frommob, mob/living/tomob) *** ./modules/admin/admin_ranks.dm-9-/datum/admin_rank/New(init_name, init_rights, list/init_adds, list/init_subs) ./modules/admin/admin_ranks.dm-10- name = init_name ./modules/admin/admin_ranks.dm-11- switch(name) ./modules/admin/admin_ranks.dm-12- if("Removed",null,"") ./modules/admin/admin_ranks.dm-13- spawn(-1) ./modules/admin/admin_ranks.dm:14: ***del(src)*** ./modules/admin/admin_ranks.dm-15- throw EXCEPTION("invalid admin-rank name") ./modules/admin/admin_ranks.dm-16- return ./modules/admin/admin_ranks.dm-17- if(init_rights) rights = init_rights ./modules/admin/admin_ranks.dm-18- if(!init_adds) init_adds = list() ./modules/admin/admin_ranks.dm-19- if(!init_subs) init_subs = list() *** ./modules/admin/DB ban/functions.dm-133- if(announceinirc) ./modules/admin/DB ban/functions.dm-134- send2irc("BAN ALERT","[a_ckey] applied a [bantype_str] on [ckey]") ./modules/admin/DB ban/functions.dm-135- ./modules/admin/DB ban/functions.dm-136- if(kickbannedckey) ./modules/admin/DB ban/functions.dm-137- if(banned_mob && banned_mob.client && banned_mob.client.ckey == banckey) ./modules/admin/DB ban/functions.dm:138: ***del(banned_mob.client)*** ./modules/admin/DB ban/functions.dm-139- ./modules/admin/DB ban/functions.dm-140- ./modules/admin/DB ban/functions.dm-141-/datum/admins/proc/DB_ban_unban(ckey, bantype, job = "") ./modules/admin/DB ban/functions.dm-142- ./modules/admin/DB ban/functions.dm-143- if(!check_rights(R_BAN)) return *** ./modules/admin/holder2.dm-15- var/admincaster_signature ./modules/admin/holder2.dm-16- ./modules/admin/holder2.dm-17-/datum/admins/New(datum/admin_rank/R, ckey) ./modules/admin/holder2.dm-18- if(!ckey) ./modules/admin/holder2.dm-19- spawn(-1) ./modules/admin/holder2.dm:20: ***del(src)*** ./modules/admin/holder2.dm-21- throw EXCEPTION("Admin datum created without a ckey") ./modules/admin/holder2.dm-22- return ./modules/admin/holder2.dm-23- if(!istype(R)) ./modules/admin/holder2.dm-24- spawn(-1) ./modules/admin/holder2.dm:25: ***del(src)*** ./modules/admin/holder2.dm-26- throw EXCEPTION("Admin datum created without a rank") ./modules/admin/holder2.dm-27- return ./modules/admin/holder2.dm-28- rank = R ./modules/admin/holder2.dm-29- admincaster_signature = "Nanotrasen Officer #[rand(0,9)][rand(0,9)][rand(0,9)]" ./modules/admin/holder2.dm-30- admin_datums[ckey] = src *** ./modules/admin/holder2.dm-89- ./modules/admin/holder2.dm-90-/client/proc/deadmin() ./modules/admin/holder2.dm-91- admin_datums -= ckey ./modules/admin/holder2.dm-92- if(holder) ./modules/admin/holder2.dm-93- holder.disassociate() ./modules/admin/holder2.dm:94: ***del(holder)*** ./modules/admin/holder2.dm-95- return 1 ./modules/admin/holder2.dm-96- ./modules/admin/holder2.dm-97-//This proc checks whether subject has at least ONE of the rights specified in rights_required. ./modules/admin/holder2.dm-98-/proc/check_rights_for(client/subject, rights_required) ./modules/admin/holder2.dm-99- if(subject && subject.holder && subject.holder.rank) *** ./modules/admin/topic.dm-989- return ./modules/admin/topic.dm-990- M << "<span class='danger'>You have been kicked from the server.</span>" ./modules/admin/topic.dm-991- log_admin("[key_name(usr)] booted [key_name(M)].") ./modules/admin/topic.dm-992- message_admins("<span class='adminnotice'>[key_name_admin(usr)] booted [key_name_admin(M)].</span>") ./modules/admin/topic.dm-993- //M.client = null ./modules/admin/topic.dm:994: ***del(M.client)*** ./modules/admin/topic.dm-995- ./modules/admin/topic.dm-996- //Player Notes ./modules/admin/topic.dm-997- else if(href_list["notes"]) ./modules/admin/topic.dm-998- var/ckey = href_list["ckey"] ./modules/admin/topic.dm-999- switch(href_list["notes"]) *** ./modules/admin/topic.dm-1050- else ./modules/admin/topic.dm-1051- M << "<span class='danger'>No ban appeals URL has been set.</span>" ./modules/admin/topic.dm-1052- log_admin("[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis will be removed in [mins] minutes.") ./modules/admin/topic.dm-1053- message_admins("<span class='adminnotice'>[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis will be removed in [mins] minutes.</span>") ./modules/admin/topic.dm-1054- ./modules/admin/topic.dm:1055: ***del(M.client)*** ./modules/admin/topic.dm-1056- //qdel(M) // See no reason why to delete mob. Important stuff can be lost. And ban can be lifted before round ends. ./modules/admin/topic.dm-1057- if("No") ./modules/admin/topic.dm-1058- var/reason = input(usr,"Reason?","reason","Griefer") as text|null ./modules/admin/topic.dm-1059- if(!reason) ./modules/admin/topic.dm-1060- return *** ./modules/admin/topic.dm-1074- log_admin("[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis is a permanent ban.") ./modules/admin/topic.dm-1075- message_admins("<span class='adminnotice'>[usr.client.ckey] has banned [M.ckey].\nReason: [reason]\nThis is a permanent ban.</span>") ./modules/admin/topic.dm-1076- feedback_inc("ban_perma",1) ./modules/admin/topic.dm-1077- DB_ban_record(BANTYPE_PERMA, M, -1, reason) ./modules/admin/topic.dm-1078- ./modules/admin/topic.dm:1079: ***del(M.client)*** ./modules/admin/topic.dm-1080- //qdel(M) ./modules/admin/topic.dm-1081- if("Cancel") ./modules/admin/topic.dm-1082- return ./modules/admin/topic.dm-1083- ./modules/admin/topic.dm-1084- else if(href_list["unjobbanf"]) *** ./modules/admin/verbs/SDQL2/SDQL_2.dm-178- do_parse = 0 ./modules/admin/verbs/SDQL2/SDQL_2.dm-179- else ./modules/admin/verbs/SDQL2/SDQL_2.dm-180- query_tree += val ./modules/admin/verbs/SDQL2/SDQL_2.dm-181- pos++ ./modules/admin/verbs/SDQL2/SDQL_2.dm-182- ./modules/admin/verbs/SDQL2/SDQL_2.dm:183: ***del(parser)*** ./modules/admin/verbs/SDQL2/SDQL_2.dm-184- ./modules/admin/verbs/SDQL2/SDQL_2.dm-185- return querys ./modules/admin/verbs/SDQL2/SDQL_2.dm-186- ./modules/admin/verbs/SDQL2/SDQL_2.dm-187- ./modules/admin/verbs/SDQL2/SDQL_2.dm-188- *** ./modules/awaymissions/maploader/swapmaps.dm-154- z1=min(x:z,y:z);z2=max(x:z,y:z) ./modules/awaymissions/maploader/swapmaps.dm-155- InitializeSwapMaps() ./modules/awaymissions/maploader/swapmaps.dm-156- if(z2>swapmaps_compiled_maxz ||\ ./modules/awaymissions/maploader/swapmaps.dm-157- y2>swapmaps_compiled_maxy ||\ ./modules/awaymissions/maploader/swapmaps.dm-158- x2>swapmaps_compiled_maxx) ./modules/awaymissions/maploader/swapmaps.dm:159: ***del(src)*** ./modules/awaymissions/maploader/swapmaps.dm-160- return ./modules/awaymissions/maploader/swapmaps.dm-161- x2=x?(x):world.maxx ./modules/awaymissions/maploader/swapmaps.dm-162- y2=y?(y):world.maxy ./modules/awaymissions/maploader/swapmaps.dm-163- z2=z?(z):1 ./modules/awaymissions/maploader/swapmaps.dm-164- AllocateSwapMap() *** ./modules/awaymissions/maploader/swapmaps.dm-177- for(var/obj/O in A) qdel(O) ./modules/awaymissions/maploader/swapmaps.dm-178- for(var/mob/M in A) ./modules/awaymissions/maploader/swapmaps.dm-179- if(!M.key) qdel(M) ./modules/awaymissions/maploader/swapmaps.dm-180- else M.loc=null ./modules/awaymissions/maploader/swapmaps.dm-181- areas[A.loc]=null ./modules/awaymissions/maploader/swapmaps.dm:182: ***del(A)*** ./modules/awaymissions/maploader/swapmaps.dm-183- // delete areas that belong only to this map ./modules/awaymissions/maploader/swapmaps.dm-184- for(var/area/a in areas) ./modules/awaymissions/maploader/swapmaps.dm:185: if(a && !a.contents.len) ***del(a)*** ./modules/awaymissions/maploader/swapmaps.dm-186- if(x2>=world.maxx || y2>=world.maxy || z2>=world.maxz) CutXYZ() ./modules/awaymissions/maploader/swapmaps.dm:187: ***del(areas)*** ./modules/awaymissions/maploader/swapmaps.dm-188- ..() ./modules/awaymissions/maploader/swapmaps.dm-189- ./modules/awaymissions/maploader/swapmaps.dm-190- /* ./modules/awaymissions/maploader/swapmaps.dm-191- Savefile format: ./modules/awaymissions/maploader/swapmaps.dm-192- map *** ./modules/awaymissions/maploader/swapmaps.dm-239- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-240- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-241- sleep() ./modules/awaymissions/maploader/swapmaps.dm-242- S.cd=oldcd ./modules/awaymissions/maploader/swapmaps.dm-243- locked=0 ./modules/awaymissions/maploader/swapmaps.dm:244: ***del(areas)*** ./modules/awaymissions/maploader/swapmaps.dm-245- ./modules/awaymissions/maploader/swapmaps.dm-246- Read(savefile/S,_id,turf/locorner) ./modules/awaymissions/maploader/swapmaps.dm-247- var/x ./modules/awaymissions/maploader/swapmaps.dm-248- var/y ./modules/awaymissions/maploader/swapmaps.dm-249- var/z *** ./modules/awaymissions/maploader/swapmaps.dm-295- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-296- S.cd=".." ./modules/awaymissions/maploader/swapmaps.dm-297- sleep() ./modules/awaymissions/maploader/swapmaps.dm-298- S.cd=oldcd ./modules/awaymissions/maploader/swapmaps.dm-299- locked=0 ./modules/awaymissions/maploader/swapmaps.dm:300: ***del(areas)*** ./modules/awaymissions/maploader/swapmaps.dm-301- ./modules/awaymissions/maploader/swapmaps.dm-302- /* ./modules/awaymissions/maploader/swapmaps.dm-303- Find an empty block on the world map in which to load this map. ./modules/awaymissions/maploader/swapmaps.dm-304- If no space is found, increase world.maxz as necessary. (If the ./modules/awaymissions/maploader/swapmaps.dm-305- map is greater in x,y size than the current world, expand *** ./modules/awaymissions/maploader/swapmaps.dm-319- else ./modules/awaymissions/maploader/swapmaps.dm-320- var/list/l=ConsiderRegion(1,1,world.maxx,world.maxy,swapmaps_compiled_maxz+1) ./modules/awaymissions/maploader/swapmaps.dm-321- x1=l[1] ./modules/awaymissions/maploader/swapmaps.dm-322- y1=l[2] ./modules/awaymissions/maploader/swapmaps.dm-323- z1=l[3] ./modules/awaymissions/maploader/swapmaps.dm:324: ***del(l)*** ./modules/awaymissions/maploader/swapmaps.dm-325- x2+=x1-1 ./modules/awaymissions/maploader/swapmaps.dm-326- y2+=y1-1 ./modules/awaymissions/maploader/swapmaps.dm-327- z2+=z1-1 ./modules/awaymissions/maploader/swapmaps.dm-328- world.maxz=max(z2,world.maxz) // stretch z if necessary ./modules/awaymissions/maploader/swapmaps.dm-329- if(!ischunk) *** ./modules/awaymissions/maploader/swapmaps.dm-375- world.maxz=mz ./modules/awaymissions/maploader/swapmaps.dm-376- ./modules/awaymissions/maploader/swapmaps.dm-377- // save and delete ./modules/awaymissions/maploader/swapmaps.dm-378- proc/Unload() ./modules/awaymissions/maploader/swapmaps.dm-379- Save() ./modules/awaymissions/maploader/swapmaps.dm:380: ***del(src)*** ./modules/awaymissions/maploader/swapmaps.dm-381- ./modules/awaymissions/maploader/swapmaps.dm-382- proc/Save() ./modules/awaymissions/maploader/swapmaps.dm-383- if(id==src) return 0 ./modules/awaymissions/maploader/swapmaps.dm-384- var/savefile/S=mode?(new):new("map_[id].sav") ./modules/awaymissions/maploader/swapmaps.dm-385- S << src *** ./modules/awaymissions/maploader/swapmaps.dm-470- var/list/l=contents ./modules/awaymissions/maploader/swapmaps.dm-471- if(M) ./modules/awaymissions/maploader/swapmaps.dm-472- l=l.Copy() ./modules/awaymissions/maploader/swapmaps.dm-473- for(M in src) if(M.key) l-=M ./modules/awaymissions/maploader/swapmaps.dm-474- if(l.len) S["contents"]<<l ./modules/awaymissions/maploader/swapmaps.dm:475: if(l!=contents) ***del(l)*** ./modules/awaymissions/maploader/swapmaps.dm-476- Read(savefile/S) ./modules/awaymissions/maploader/swapmaps.dm-477- var/list/l ./modules/awaymissions/maploader/swapmaps.dm-478- if(contents.len) l=contents ./modules/awaymissions/maploader/swapmaps.dm-479- ..() ./modules/awaymissions/maploader/swapmaps.dm-480- // if the icon was a text string, it would not have loaded properly *** ./modules/awaymissions/maploader/swapmaps.dm-483- var/ic ./modules/awaymissions/maploader/swapmaps.dm-484- S["icon"]>>ic ./modules/awaymissions/maploader/swapmaps.dm-485- if(istext(ic)) icon=swapmaps_iconcache[ic] ./modules/awaymissions/maploader/swapmaps.dm-486- if(l && contents!=l) ./modules/awaymissions/maploader/swapmaps.dm-487- contents+=l ./modules/awaymissions/maploader/swapmaps.dm:488: ***del(l)*** ./modules/awaymissions/maploader/swapmaps.dm-489- ./modules/awaymissions/maploader/swapmaps.dm-490- ./modules/awaymissions/maploader/swapmaps.dm-491-// set this up (at runtime) as follows: ./modules/awaymissions/maploader/swapmaps.dm-492-// list(\ ./modules/awaymissions/maploader/swapmaps.dm-493-// 'player.dmi'="player",\ *** ./modules/awaymissions/maploader/swapmaps.dm-623- properly otherwise. The //.0 path should always match the map, however. ./modules/awaymissions/maploader/swapmaps.dm-624- */ ./modules/awaymissions/maploader/swapmaps.dm-625- S.cd="//.0" ./modules/awaymissions/maploader/swapmaps.dm-626- M.Read(S,M,locorner) ./modules/awaymissions/maploader/swapmaps.dm-627- while(M.locked) sleep(1) ./modules/awaymissions/maploader/swapmaps.dm:628: ***del(M)*** ./modules/awaymissions/maploader/swapmaps.dm-629- return 1 ./modules/awaymissions/maploader/swapmaps.dm-630- ./modules/awaymissions/maploader/swapmaps.dm-631-proc/SwapMaps_SaveChunk(chunk_id,turf/corner1,turf/corner2) ./modules/awaymissions/maploader/swapmaps.dm-632- if(!corner1 || !corner2) ./modules/awaymissions/maploader/swapmaps.dm-633- world.log << "SwapMaps error in SwapMaps_SaveChunk():" *** ./modules/awaymissions/maploader/swapmaps.dm-644- M.y2=max(corner1.y,corner2.y) ./modules/awaymissions/maploader/swapmaps.dm-645- M.z2=max(corner1.z,corner2.z) ./modules/awaymissions/maploader/swapmaps.dm-646- M.mode=swapmaps_mode ./modules/awaymissions/maploader/swapmaps.dm-647- M.Save() ./modules/awaymissions/maploader/swapmaps.dm-648- while(M.locked) sleep(1) ./modules/awaymissions/maploader/swapmaps.dm:649: ***del(M)*** ./modules/awaymissions/maploader/swapmaps.dm-650- return 1 ./modules/awaymissions/maploader/swapmaps.dm-651- ./modules/awaymissions/maploader/swapmaps.dm-652-proc/SwapMaps_GetSize(id) ./modules/awaymissions/maploader/swapmaps.dm-653- var/savefile/S ./modules/awaymissions/maploader/swapmaps.dm-654- var/text=0 *** ./modules/client/client procs.dm-138- if (isnum(player_age) && player_age == -1) //first connection ./modules/client/client procs.dm-139- if (config.panic_bunker && !holder && !(ckey in deadmins)) ./modules/client/client procs.dm-140- log_access("Failed Login: [key] - New account attempting to connect during panic bunker") ./modules/client/client procs.dm-141- message_admins("<span class='adminnotice'>Failed Login: [key] - New account attempting to connect during panic bunker</span>") ./modules/client/client procs.dm-142- src << "Sorry but the server is currently not accepting connections from never before seen players." ./modules/client/client procs.dm:143: ***del(src)*** ./modules/client/client procs.dm-144- return 0 ./modules/client/client procs.dm-145- ./modules/client/client procs.dm-146- if (config.notify_new_player_age >= 0) ./modules/client/client procs.dm-147- message_admins("New user: [key_name_admin(src)] is connecting here for the first time.") ./modules/client/client procs.dm-148- if (config.irc_first_connection_alert) *** ./modules/crafting/guncrafting.dm-22- if(istype(W,/obj/item/pipe)) ./modules/crafting/guncrafting.dm-23- user << "<span class='notice'>You attach the shotgun barrel to the receiver. The pins seem loose.</span>" ./modules/crafting/guncrafting.dm-24- var/obj/item/weaponcrafting/ishotgunconstruction/I = new /obj/item/weaponcrafting/ishotgunconstruction ./modules/crafting/guncrafting.dm-25- user.unEquip(src) ./modules/crafting/guncrafting.dm-26- user.put_in_hands(I) ./modules/crafting/guncrafting.dm:27: ***del(W)*** ./modules/crafting/guncrafting.dm:28: ***del(src)*** ./modules/crafting/guncrafting.dm-29- return ./modules/crafting/guncrafting.dm-30- ./modules/crafting/guncrafting.dm-31-// SHOTGUN // ./modules/crafting/guncrafting.dm-32- ./modules/crafting/guncrafting.dm-33-/obj/item/weaponcrafting/ishotgunconstruction *** ./modules/crafting/guncrafting.dm-55- if(istype(W,/obj/item/weaponcrafting/stock)) ./modules/crafting/guncrafting.dm-56- user << "<span class='notice'>You attach the stock to the receiver-barrel assembly.</span>" ./modules/crafting/guncrafting.dm-57- var/obj/item/weaponcrafting/ishotgunconstruction3/I = new /obj/item/weaponcrafting/ishotgunconstruction3 ./modules/crafting/guncrafting.dm-58- user.unEquip(src) ./modules/crafting/guncrafting.dm-59- user.put_in_hands(I) ./modules/crafting/guncrafting.dm:60: ***del(W)*** ./modules/crafting/guncrafting.dm:61: ***del(src)*** ./modules/crafting/guncrafting.dm-62- return ./modules/crafting/guncrafting.dm-63- ./modules/crafting/guncrafting.dm-64-/obj/item/weaponcrafting/ishotgunconstruction3 ./modules/crafting/guncrafting.dm-65- name = "extremely conspicuous metal construction" ./modules/crafting/guncrafting.dm-66- desc = "A receiver-barrel shotgun assembly with a loose wooden stock. There's no way you can fire it without the stock coming loose." *** ./modules/flufftext/Hallucination.dm-154- if(target.client) target.client.images |= flood_images ./modules/flufftext/Hallucination.dm-155- return ./modules/flufftext/Hallucination.dm-156- ./modules/flufftext/Hallucination.dm-157-/obj/effect/hallucination/fake_flood/Destroy() ./modules/flufftext/Hallucination.dm-158- SSobj.processing.Remove(src) ./modules/flufftext/Hallucination.dm:159: ***del(flood_turfs)*** ./modules/flufftext/Hallucination.dm-160- if(target.client) target.client.images.Remove(flood_images) ./modules/flufftext/Hallucination.dm-161- target = null ./modules/flufftext/Hallucination.dm:162: ***del(flood_images)*** ./modules/flufftext/Hallucination.dm-163- return ./modules/flufftext/Hallucination.dm-164- ./modules/flufftext/Hallucination.dm-165-/obj/effect/hallucination/simple/xeno ./modules/flufftext/Hallucination.dm-166- image_icon = 'icons/mob/alien.dmi' ./modules/flufftext/Hallucination.dm-167- image_state = "alienh_pounce" *** ./modules/hydroponics/hydroponics.dm-684- qdel(reagent_source) ./modules/hydroponics/hydroponics.dm-685- ./modules/hydroponics/hydroponics.dm-686- H.applyChemicals(S) ./modules/hydroponics/hydroponics.dm-687- ./modules/hydroponics/hydroponics.dm-688- S.clear_reagents() ./modules/hydroponics/hydroponics.dm:689: ***del(S)*** ./modules/hydroponics/hydroponics.dm-690- H.update_icon() ./modules/hydroponics/hydroponics.dm-691- if(reagent_source) // If the source wasn't composted and destroyed ./modules/hydroponics/hydroponics.dm-692- reagent_source.update_icon() ./modules/hydroponics/hydroponics.dm-693- return 1 ./modules/hydroponics/hydroponics.dm-694- *** ./modules/hydroponics/seed_extractor.dm-183- if(N.amount <= 0) ./modules/hydroponics/seed_extractor.dm-184- return ./modules/hydroponics/seed_extractor.dm-185- N.amount = max(N.amount - 1, 0) ./modules/hydroponics/seed_extractor.dm-186- if (N.amount <= 0) ./modules/hydroponics/seed_extractor.dm-187- piles -= N ./modules/hydroponics/seed_extractor.dm:188: ***del(N)*** ./modules/hydroponics/seed_extractor.dm-189- break ./modules/hydroponics/seed_extractor.dm-190- ./modules/hydroponics/seed_extractor.dm-191- for (var/obj/T in contents)//Now we find the seed we need to vend ./modules/hydroponics/seed_extractor.dm-192- var/obj/item/seeds/O = T ./modules/hydroponics/seed_extractor.dm-193- if (O.plantname == href_list["name"] && O.lifespan == href_list["li"] && O.endurance == href_list["en"] && O.maturation == href_list["ma"] && O.production == href_list["pr"] && O.yield == href_list["yi"] && O.potency == href_list["pot"]) *** ./modules/mob/camera/camera.dm-14-/mob/camera/experience_pressure_difference() ./modules/mob/camera/camera.dm-15- return ./modules/mob/camera/camera.dm-16- ./modules/mob/camera/camera.dm-17-/mob/camera/Destroy() ./modules/mob/camera/camera.dm-18- ..() ./modules/mob/camera/camera.dm:19: ***del(src)*** ./modules/mob/camera/camera.dm-20- ./modules/mob/camera/camera.dm-21-/mob/camera/Login() ./modules/mob/camera/camera.dm-22- ..() ./modules/mob/camera/camera.dm-23- update_interface() *** ./modules/mob/living/carbon/human/blood.dm-280- // Are we dripping or splattering? ./modules/mob/living/carbon/human/blood.dm-281- var/list/drips = list() ./modules/mob/living/carbon/human/blood.dm-282- // Only a certain number of drips (or one large splatter) can be on a given turf. ./modules/mob/living/carbon/human/blood.dm-283- for(var/obj/effect/decal/cleanable/blood/drip/drop in T) ./modules/mob/living/carbon/human/blood.dm-284- drips |= drop.drips ./modules/mob/living/carbon/human/blood.dm:285: ***del(drop)*** ./modules/mob/living/carbon/human/blood.dm-286- if(!large && drips.len < 3) ./modules/mob/living/carbon/human/blood.dm-287- decal_type = /obj/effect/decal/cleanable/blood/drip ./modules/mob/living/carbon/human/blood.dm-288- ./modules/mob/living/carbon/human/blood.dm-289- // Find a blood decal or create a new one. ./modules/mob/living/carbon/human/blood.dm-290- B = locate(decal_type) in T *** ./modules/mob/living/silicon/ai/ai.dm-703- ./modules/mob/living/silicon/ai/ai.dm-704- if(personnel_list.len) ./modules/mob/living/silicon/ai/ai.dm-705- input = input("Select a crew member:") as null|anything in personnel_list ./modules/mob/living/silicon/ai/ai.dm-706- var/icon/character_icon = personnel_list[input] ./modules/mob/living/silicon/ai/ai.dm-707- if(character_icon) ./modules/mob/living/silicon/ai/ai.dm:708: ***del(holo_icon)***//Clear old icon so we're not storing it in memory. ./modules/mob/living/silicon/ai/ai.dm-709- holo_icon = getHologramIcon(icon(character_icon)) ./modules/mob/living/silicon/ai/ai.dm-710- else ./modules/mob/living/silicon/ai/ai.dm-711- alert("No suitable records found. Aborting.") ./modules/mob/living/silicon/ai/ai.dm-712- ./modules/mob/living/silicon/ai/ai.dm-713- else *** ./modules/mob/living/silicon/ai/ai.dm-717- "xeno queen", ./modules/mob/living/silicon/ai/ai.dm-718- "space carp" ./modules/mob/living/silicon/ai/ai.dm-719- ) ./modules/mob/living/silicon/ai/ai.dm-720- input = input("Please select a hologram:") as null|anything in icon_list ./modules/mob/living/silicon/ai/ai.dm-721- if(input) ./modules/mob/living/silicon/ai/ai.dm:722: ***del(holo_icon)*** ./modules/mob/living/silicon/ai/ai.dm-723- switch(input) ./modules/mob/living/silicon/ai/ai.dm-724- if("default") ./modules/mob/living/silicon/ai/ai.dm-725- holo_icon = getHologramIcon(icon('icons/mob/AI.dmi',"holo1")) ./modules/mob/living/silicon/ai/ai.dm-726- if("floating face") ./modules/mob/living/silicon/ai/ai.dm-727- holo_icon = getHologramIcon(icon('icons/mob/AI.dmi',"holo2")) *** ./modules/mob/living/silicon/pai/death.dm-10- //var/tod = time2text(world.realtime,"hh:mm:ss") //weasellos time of death patch ./modules/mob/living/silicon/pai/death.dm-11- //mind.store_memory("Time of death: [tod]", 0) ./modules/mob/living/silicon/pai/death.dm-12- ./modules/mob/living/silicon/pai/death.dm-13- //New pAI's get a brand new mind to prevent meta stuff from their previous life. This new mind causes problems down the line if it's not deleted here. ./modules/mob/living/silicon/pai/death.dm-14- //Read as: I have no idea what I'm doing but asking for help got me nowhere so this is what you get. - Nodrak ./modules/mob/living/silicon/pai/death.dm:15: if(mind) ***del(mind)*** ./modules/mob/living/silicon/pai/death.dm-16- living_mob_list -= src ./modules/mob/living/silicon/pai/death.dm-17- ghostize() ./modules/mob/living/silicon/pai/death.dm-18- qdel(src) *** ./modules/mob/living/simple_animal/worm.dm-166- newHead.Attach(newHeadPrevious) ./modules/mob/living/simple_animal/worm.dm-167- ./modules/mob/living/simple_animal/worm.dm-168- if(die) ./modules/mob/living/simple_animal/worm.dm-169- newHead.Die() ./modules/mob/living/simple_animal/worm.dm-170- ./modules/mob/living/simple_animal/worm.dm:171: ***del(src)*** ./modules/mob/living/simple_animal/worm.dm-172- ./modules/mob/living/simple_animal/worm.dm-173- proc/ProcessStomach() ./modules/mob/living/simple_animal/worm.dm-174- for(var/atom/movable/stomachContent in contents) ./modules/mob/living/simple_animal/worm.dm-175- if(prob(digestionProbability)) ./modules/mob/living/simple_animal/worm.dm-176- if(istype(stomachContent,/obj/item/stack)) //converts to plasma, keeping the stack value ./modules/mob/living/simple_animal/worm.dm-177- if(!istype(stomachContent,/obj/item/stack/sheet/mineral/plasma)) ./modules/mob/living/simple_animal/worm.dm-178- var/obj/item/stack/oldStack = stomachContent ./modules/mob/living/simple_animal/worm.dm-179- new /obj/item/stack/sheet/mineral/plasma(src, oldStack.amount) ./modules/mob/living/simple_animal/worm.dm:180: ***del(oldStack)*** ./modules/mob/living/simple_animal/worm.dm-181- continue ./modules/mob/living/simple_animal/worm.dm-182- else if(istype(stomachContent,/obj/item)) //converts to plasma, keeping the w_class ./modules/mob/living/simple_animal/worm.dm-183- var/obj/item/oldItem = stomachContent ./modules/mob/living/simple_animal/worm.dm-184- new /obj/item/stack/sheet/mineral/plasma(src, oldItem.w_class) ./modules/mob/living/simple_animal/worm.dm:185: ***del(oldItem)*** ./modules/mob/living/simple_animal/worm.dm-186- continue ./modules/mob/living/simple_animal/worm.dm-187- else ./modules/mob/living/simple_animal/worm.dm-188- new /obj/item/stack/sheet/mineral/plasma(src, flatPlasmaValue) //just flat amount ./modules/mob/living/simple_animal/worm.dm:189: ***del(stomachContent)*** ./modules/mob/living/simple_animal/worm.dm-190- continue ./modules/mob/living/simple_animal/worm.dm-191- ./modules/mob/living/simple_animal/worm.dm-192- if(previous) ./modules/mob/living/simple_animal/worm.dm-193- for(var/atom/movable/stomachContent in contents) //transfer it along the digestive tract ./modules/mob/living/simple_animal/worm.dm-194- previous.contents += stomachContent *** ./modules/mob/login.dm-28- update_Login_details() ./modules/mob/login.dm-29- world.update_status() ./modules/mob/login.dm-30- ./modules/mob/login.dm-31- client.images = null //remove the images such as AIs being unable to see runes ./modules/mob/login.dm-32- client.screen = list() //remove hud items just in case ./modules/mob/login.dm:33: if(hud_used) ***del(hud_used)*** //remove the hud objects ./modules/mob/login.dm-34- hud_used = new /datum/hud(src) ./modules/mob/login.dm-35- ./modules/mob/login.dm-36- next_move = 1 ./modules/mob/login.dm-37- sight |= SEE_SELF ./modules/mob/login.dm-38- *** ./modules/mob/new_player/preferences_setup.dm-19- backbag = 1 ./modules/mob/new_player/preferences_setup.dm-20- features = random_features() ./modules/mob/new_player/preferences_setup.dm-21- age = rand(AGE_MIN,AGE_MAX) ./modules/mob/new_player/preferences_setup.dm-22- ./modules/mob/new_player/preferences_setup.dm-23-/datum/preferences/proc/update_preview_icon() //seriously. This is horrendous. ./modules/mob/new_player/preferences_setup.dm:24: ***del(preview_icon_front)*** ./modules/mob/new_player/preferences_setup.dm:25: ***del(preview_icon_side)*** ./modules/mob/new_player/preferences_setup.dm-26- var/icon/preview_icon = null ./modules/mob/new_player/preferences_setup.dm-27- ./modules/mob/new_player/preferences_setup.dm-28- if(job_engsec_high) //cyborg/AI check, put first to avoid so much unneeded blending ./modules/mob/new_player/preferences_setup.dm-29- switch(job_engsec_high) ./modules/mob/new_player/preferences_setup.dm-30- if(AI) *** ./modules/mob/new_player/preferences_setup.dm-35- ./modules/mob/new_player/preferences_setup.dm-36- if(preview_icon) //We're busting out! ./modules/mob/new_player/preferences_setup.dm-37- preview_icon_front = new(preview_icon, dir = SOUTH) ./modules/mob/new_player/preferences_setup.dm-38- preview_icon_side = new(preview_icon, dir = WEST) ./modules/mob/new_player/preferences_setup.dm-39- ./modules/mob/new_player/preferences_setup.dm:40: ***del(preview_icon)*** ./modules/mob/new_player/preferences_setup.dm-41- return ./modules/mob/new_player/preferences_setup.dm-42- ./modules/mob/new_player/preferences_setup.dm-43- var/g = "m" ./modules/mob/new_player/preferences_setup.dm-44- if(gender == FEMALE) g = "f" ./modules/mob/new_player/preferences_setup.dm-45- *** ./modules/mob/new_player/preferences_setup.dm-419- if(clothes_s) ./modules/mob/new_player/preferences_setup.dm-420- preview_icon.Blend(clothes_s, ICON_OVERLAY) ./modules/mob/new_player/preferences_setup.dm-421- preview_icon_front = new(preview_icon, dir = SOUTH) ./modules/mob/new_player/preferences_setup.dm-422- preview_icon_side = new(preview_icon, dir = WEST) ./modules/mob/new_player/preferences_setup.dm-423- ./modules/mob/new_player/preferences_setup.dm:424: ***del(preview_icon)*** ./modules/mob/new_player/preferences_setup.dm:425: ***del(eyes_s)*** ./modules/mob/new_player/preferences_setup.dm:426: ***del(clothes_s)*** *** ./modules/power/apc.dm-150- area.power_equip = 0 ./modules/power/apc.dm-151- area.power_environ = 0 ./modules/power/apc.dm-152- area.power_change() ./modules/power/apc.dm-153- if(occupier) ./modules/power/apc.dm-154- malfvacate(1) ./modules/power/apc.dm:155: ***del(wires)*** ./modules/power/apc.dm-156- if(cell) ./modules/power/apc.dm-157- qdel(cell) ./modules/power/apc.dm-158- if(terminal) ./modules/power/apc.dm-159- disconnect_terminal() ./modules/power/apc.dm-160- ..() *** ./modules/projectiles/firing.dm-32- if (!istype(targloc) || !istype(curloc) || !BB) ./modules/projectiles/firing.dm-33- return 0 ./modules/projectiles/firing.dm-34- if(targloc == curloc) ./modules/projectiles/firing.dm-35- if(BB.original == user) //if we target ourselves we go straight to bullet_act() ./modules/projectiles/firing.dm-36- user.bullet_act(BB) ./modules/projectiles/firing.dm:37: ***del(BB)*** ./modules/projectiles/firing.dm-38- return 1 ./modules/projectiles/firing.dm-39- BB.loc = get_turf(user) ./modules/projectiles/firing.dm-40- BB.starting = get_turf(user) ./modules/projectiles/firing.dm-41- BB.current = curloc ./modules/projectiles/firing.dm-42- BB.yo = targloc.y - curloc.y *** ./modules/projectiles/projectile/magic.dm-241- else ./modules/projectiles/projectile/magic.dm-242- new_mob.key = M.key ./modules/projectiles/projectile/magic.dm-243- ./modules/projectiles/projectile/magic.dm-244- new_mob << "<B>Your form morphs into that of a [randomize].</B>" ./modules/projectiles/projectile/magic.dm-245- ./modules/projectiles/projectile/magic.dm:246: ***del(M)*** ./modules/projectiles/projectile/magic.dm-247- return new_mob ./modules/projectiles/projectile/magic.dm-248- ./modules/projectiles/projectile/magic.dm-249-/obj/item/projectile/magic/animate ./modules/projectiles/projectile/magic.dm-250- name = "bolt of animation" ./modules/projectiles/projectile/magic.dm-251- icon_state = "red_1" *** ./modules/research/rdconsole.dm-58- for(var/T in typesof(/datum/tech) - /datum/tech) ./modules/research/rdconsole.dm-59- check_tech = null ./modules/research/rdconsole.dm-60- check_tech = new T() ./modules/research/rdconsole.dm-61- if(check_tech.id == ID) ./modules/research/rdconsole.dm-62- return_name = check_tech.name ./modules/research/rdconsole.dm:63: ***del(check_tech)*** ./modules/research/rdconsole.dm-64- check_tech = null ./modules/research/rdconsole.dm-65- break ./modules/research/rdconsole.dm-66- ./modules/research/rdconsole.dm-67- return return_name ./modules/research/rdconsole.dm-68- *** ./modules/research/rdconsole.dm-92- for(var/R in typesof(/datum/reagent) - /datum/reagent) ./modules/research/rdconsole.dm-93- temp_reagent = null ./modules/research/rdconsole.dm-94- temp_reagent = new R() ./modules/research/rdconsole.dm-95- if(temp_reagent.id == ID) ./modules/research/rdconsole.dm-96- return_name = temp_reagent.name ./modules/research/rdconsole.dm:97: ***del(temp_reagent)*** ./modules/research/rdconsole.dm-98- temp_reagent = null ./modules/research/rdconsole.dm-99- break ./modules/research/rdconsole.dm-100- return return_name ./modules/research/rdconsole.dm-101- ./modules/research/rdconsole.dm-102-/obj/machinery/computer/rdconsole/proc/SyncRDevices() //Makes sure it is properly sync'ed up with the devices attached to it (if any). *** ./modules/research/rdconsole.dm-573- else if(href_list["reset"]) //Reset the R&D console's database. ./modules/research/rdconsole.dm-574- griefProtection() ./modules/research/rdconsole.dm-575- var/choice = alert("R&D Console Database Reset", "Are you sure you want to reset the R&D console's database? Data lost cannot be recovered.", "Continue", "Cancel") ./modules/research/rdconsole.dm-576- if(choice == "Continue") ./modules/research/rdconsole.dm-577- screen = 0.0 ./modules/research/rdconsole.dm:578: ***del(files)*** ./modules/research/rdconsole.dm-579- files = new /datum/research(src) ./modules/research/rdconsole.dm-580- spawn(20) ./modules/research/rdconsole.dm-581- screen = 1.6 ./modules/research/rdconsole.dm-582- updateUsrDialog() ./modules/research/rdconsole.dm-583- *** ./_compile_options.dm-49-#ifdef dellogging ./_compile_options.dm-50-#warn compiling del logging. This will have additional overheads. //will warn you if compiling with dellogging ./_compile_options.dm-51-var/list/del_counter = list() ./_compile_options.dm-52-/proc/log_del(datum/X) ./_compile_options.dm-53- if(istype(X)){del_counter[X.type]++;} ./_compile_options.dm:54: ***del(X)*** ./_compile_options.dm:55:#define ***del(X)*** log_del(X) //overrides all del() calls with log_del() ./_compile_options.dm-56-#endif ./_compile_options.dm-57- ./_compile_options.dm-58-#ifdef TESTING ./_compile_options.dm-59-#warn compiling in TESTING mode. testing() debug messages will be visible. ./_compile_options.dm-60-#endif *** ./_onclick/hud/action.dm-43- ./_onclick/hud/action.dm-44-/datum/action/proc/Remove(mob/living/T) ./_onclick/hud/action.dm-45- if(button) ./_onclick/hud/action.dm-46- if(T.client) ./_onclick/hud/action.dm-47- T.client.screen -= button ./_onclick/hud/action.dm:48: ***del(button)*** ./_onclick/hud/action.dm-49- T.actions.Remove(src) ./_onclick/hud/action.dm-50- T.update_action_buttons() ./_onclick/hud/action.dm-51- owner = null ./_onclick/hud/action.dm-52- return ./_onclick/hud/action.dm-53- *** ./__HELPERS/icons.dm-862- var/image/text_image = new(loc = A) ./__HELPERS/icons.dm-863- text_image.maptext = "<font size = 4>[letter]</font>" ./__HELPERS/icons.dm-864- text_image.color = AverageColour(atom_icon) ./__HELPERS/icons.dm-865- text_image.pixel_x = 7 ./__HELPERS/icons.dm-866- text_image.pixel_y = 5 ./__HELPERS/icons.dm:867: ***del(atom_icon)*** ./__HELPERS/icons.dm-868- return text_image ./__HELPERS/icons.dm-869- ./__HELPERS/icons.dm-870- ./__HELPERS/icons.dm-871-//Find's the average colour of the icon ./__HELPERS/icons.dm-872-//By vg's ComicIronic
non_process
list of places in code not using qdel atmospherics pipes pipes dm obj machinery atmospherics proc pipeline expansion atmospherics pipes pipes dm return null atmospherics pipes pipes dm atmospherics pipes pipes dm obj machinery atmospherics pipe proc check pressure pressure atmospherics pipes pipes dm return if parent should continue checking other pipes atmospherics pipes pipes dm return null if parent should stop checking other pipes recall del src will by default return null atmospherics pipes pipes dm return atmospherics pipes pipes dm atmospherics pipes pipes dm obj machinery atmospherics pipe proc releaseairtoturf atmospherics pipes pipes dm if air temporary atmospherics pipes pipes dm var turf t loc controllers configuration dm modes m config tag controllers configuration dm mode names m name controllers configuration dm probabilities m probability controllers configuration dm if m votable controllers configuration dm votable modes m config tag controllers configuration dm del m controllers configuration dm votable modes secret controllers configuration dm controllers configuration dm datum configuration proc load filename type config the type can also be game options in which case it uses a different switch not making it separate to not copypaste code urist controllers configuration dm var list lines filename controllers configuration dm controllers configuration dm their information but it is the only way at least that i know of controllers configuration dm for var t in typesof datum game mode datum game mode controllers configuration dm var datum game mode m new t controllers configuration dm if m config tag m config tag mode name controllers configuration dm return m controllers configuration dm del m controllers configuration dm return new datum game mode extended controllers configuration dm controllers configuration dm datum configuration proc get runnable modes controllers configuration dm var list datum game mode runnable modes new controllers configuration dm for var t in typesof datum game mode datum game mode controllers configuration dm var datum game mode m new t controllers configuration dm world debug tag prob controllers configuration dm if m config tag in modes controllers configuration dm del m controllers configuration dm continue controllers configuration dm if probabilities controllers configuration dm del m controllers configuration dm continue controllers configuration dm if m can start controllers configuration dm runnable modes probabilities controllers configuration dm world debug runnable mode controllers configuration dm return runnable modes controllers subsystem garbage dm if gcd at time time to kill controllers subsystem garbage dm break everything else is newer skip them controllers subsystem garbage dm controllers subsystem garbage dm var atom a controllers subsystem garbage dm if istext refid controllers subsystem garbage dm del a controllers subsystem garbage dm else controllers subsystem garbage dm a locate refid controllers subsystem garbage dm if a a gc destroyed gcd at time so if something else coincidently gets the same ref it s not deleted by mistake controllers subsystem garbage dm something s still referring to the qdel d object kill it controllers subsystem garbage dm testing gc ref was unable to be gc d and was deleted controllers subsystem garbage dm didntgc controllers subsystem garbage dm del a controllers subsystem garbage dm delslasttick controllers subsystem garbage dm totaldels controllers subsystem garbage dm else controllers subsystem garbage dm gcedlasttick controllers subsystem garbage dm totalgcs controllers subsystem garbage dm proc qdel var datum a controllers subsystem garbage dm if a controllers subsystem garbage dm return controllers subsystem garbage dm if istype a controllers subsystem garbage dm warning qdel passed object of type qdel can only handle datum types controllers subsystem garbage dm del a controllers subsystem garbage dm else if isnull a gc destroyed controllers subsystem garbage dm let our friend know they re about to get fucked up controllers subsystem garbage dm var hint a destroy controllers subsystem garbage dm if a controllers subsystem garbage dm return controllers subsystem garbage dm if qdel hint iwillgc functionally the same as the above qdel should assume the object will gc on its own and not check it controllers subsystem garbage dm return controllers subsystem garbage dm if qdel hint harddel qdel should assume this object won t gc and queue a hard delete using a hard reference to save time from the locate controllers subsystem garbage dm ssgarbage hardqueue a controllers subsystem garbage dm if qdel hint harddel now qdel should assume this object won t gc and hard del it post haste controllers subsystem garbage dm del a controllers subsystem garbage dm if qdel hint putinpool qdel will put this object in the pool controllers subsystem garbage dm placeinpool a controllers subsystem garbage dm else controllers subsystem garbage dm ssgarbage queue a controllers subsystem garbage dm controllers subsystem garbage dm controllers subsystem garbage dm default implementation of clean up code controllers subsystem garbage dm this should be overridden to remove all references pointing to the object being destroyed controllers subsystem garbage dm return true if the the gc controller should allow the object to continue existing useful if pooling objects controllers subsystem garbage dm datum proc destroy controllers subsystem garbage dm del src controllers subsystem garbage dm tag null controllers subsystem garbage dm return qdel hint queue controllers subsystem garbage dm controllers subsystem garbage dm datum var gc destroyed time when this object was destroyed controllers subsystem garbage dm controllers subsystem garbage dm set category debug controllers subsystem garbage dm if garbage controllers subsystem garbage dm while garbage destroyed len controllers subsystem garbage dm var datum o locate garbage destroyed controllers subsystem garbage dm if istype o o gc destroyed controllers subsystem garbage dm del o controllers subsystem garbage dm garbage dels controllers subsystem garbage dm garbage destroyed cut controllers subsystem garbage dm endif controllers subsystem radio dm controllers subsystem radio dm if frequency controllers subsystem radio dm frequency remove listener device controllers subsystem radio dm controllers subsystem radio dm if frequency devices len controllers subsystem radio dm del frequency controllers subsystem radio dm frequencies f text controllers subsystem radio dm controllers subsystem radio dm return controllers subsystem radio dm controllers subsystem radio dm datum subsystem radio proc return frequency new frequency as num controllers subsystem server maintenance dm for var client c in clients controllers subsystem server maintenance dm if c is afk inactivity kick controllers subsystem server maintenance dm if istype c mob mob dead controllers subsystem server maintenance dm log access afk controllers subsystem server maintenance dm c you have been inactive for more than minutes and have been disconnected controllers subsystem server maintenance dm del c controllers subsystem server maintenance dm controllers subsystem server maintenance dm if config sql enabled controllers subsystem server maintenance dm sql poll players controllers subsystem server maintenance dm sql poll admins controllers subsystem ticker dm controllers subsystem ticker dm else controllers subsystem ticker dm mode config pick mode master mode controllers subsystem ticker dm if mode can start controllers subsystem ticker dm world unable to start not enough players players and eligible antagonists needed reverting to pre game lobby controllers subsystem ticker dm del mode controllers subsystem ticker dm ssjob resetoccupations controllers subsystem ticker dm return controllers subsystem ticker dm controllers subsystem ticker dm configure mode and assign player to special mode stuff controllers subsystem ticker dm var can continue controllers subsystem ticker dm can continue src mode pre setup choose antagonists controllers subsystem ticker dm ssjob divideoccupations distribute jobs controllers subsystem ticker dm controllers subsystem ticker dm if controllers subsystem ticker dm if can continue controllers subsystem ticker dm del mode controllers subsystem ticker dm world error setting up reverting to pre game lobby controllers subsystem ticker dm ssjob resetoccupations controllers subsystem ticker dm return controllers subsystem ticker dm else controllers subsystem ticker dm world debug bypassing prestart checks controllers subsystem ticker dm flick intro nuke cinematic controllers subsystem ticker dm sleep controllers subsystem ticker dm flick station intact cinematic controllers subsystem ticker dm world sound sound ambience signal ogg controllers subsystem ticker dm sleep controllers subsystem ticker dm if cinematic del cinematic controllers subsystem ticker dm if temp buckle del temp buckle controllers subsystem ticker dm return faster exit since nothing happened controllers subsystem ticker dm else station nuked nuke explosion summary controllers subsystem ticker dm flick intro nuke cinematic controllers subsystem ticker dm sleep controllers subsystem ticker dm flick station explode fade red cinematic datums ai laws dm datums ai laws dm datum ai laws proc add ion law law datums ai laws dm src ion law datums ai laws dm datums ai laws dm datum ai laws proc clear inherent laws datums ai laws dm del src inherent datums ai laws dm src inherent list datums ai laws dm datums ai laws dm datum ai laws proc add supplied law number law datums ai laws dm while src supplied len number datums ai laws dm src supplied datums diseases advance advance dm if affected mob datums diseases advance advance dm var id datums diseases advance advance dm if resistance id in affected mob resistances datums diseases advance advance dm affected mob resistances id datums diseases advance advance dm remove virus datums diseases advance advance dm del src delete the datum to stop it processing datums diseases advance advance dm datums diseases advance advance dm returns the advance disease with a different reference memory datums diseases advance advance dm datum disease advance copy process datums diseases advance advance dm return new datum disease advance process src datums diseases advance advance dm datums diseases disease dm datums diseases disease dm if affected mob datums diseases disease dm for var datum disease d in affected mob viruses datums diseases disease dm if d src datums diseases disease dm if issame d datums diseases disease dm del d datums diseases disease dm datums diseases disease dm if holder affected mob datums diseases disease dm if affected mob stat dead datums diseases disease dm stage act datums diseases disease dm datums diseases disease dm if affected mob datums diseases disease dm if disease flags can resist datums diseases disease dm if type in affected mob resistances datums diseases disease dm affected mob resistances type datums diseases disease dm remove virus datums diseases disease dm del src datums diseases disease dm datums diseases disease dm datums diseases disease dm datum disease new datums diseases disease dm if required organs required organs len datums diseases disease dm if ishuman affected mob datums gas mixture dm if corresponding datums gas mixture dm corresponding new trace gas type datums gas mixture dm trace gases corresponding datums gas mixture dm corresponding moles trace gas moles datums gas mixture dm datums gas mixture dm del giver datums gas mixture dm return datums gas mixture dm datums gas mixture dm datum gas mixture remove amount datums gas mixture dm datums gas mixture dm var sum total moles datums modules dm var type o type the type of the creating object datums modules dm datums modules dm var mneed mods inmodlist type find if this type has modules defined datums modules dm datums modules dm if mneed not found in module list datums modules dm del src delete self thus ending proc datums modules dm datums modules dm var needed mods getbitmask type get a bitmask for the number of modules in this object datums modules dm status needed datums modules dm installed needed datums modules dm game area areas dm if am anchored game area areas dm am move d south game area areas dm else game area areas dm qdel am game area areas dm if istype t turf simulated game area areas dm del t game area areas dm game area areas dm for var atom movable bug in src if someone or something is somehow still in the shuttle s docking area game area areas dm if ismob bug game area areas dm continue game area areas dm qdel bug game atoms dm game atoms dm atom proc checkparts game atoms dm return game atoms dm game atoms dm atom proc assume air datum gas mixture giver game atoms dm del giver game atoms dm return null game atoms dm game atoms dm atom proc remove air amount game atoms dm return null game atoms dm game gamemodes cult runes dm red you are blinded by the flash of red light after you re able to see again you see that you are now wearing a set of armor game gamemodes cult runes dm if armorworn summoner armorworn trickster armorworn physician game gamemodes cult runes dm usr say uln shogg hafh drn game gamemodes cult runes dm user red you quietly prick your finger and make a pact with the geometer of blood to acquire more power game gamemodes cult runes dm user take overall damage rand game gamemodes cult runes dm del user head game gamemodes cult runes dm del user wear suit game gamemodes cult runes dm user equip to slot or del new obj item clothing head magus user slot head game gamemodes cult runes dm user equip to slot or del new obj item clothing suit magusred user slot wear suit game gamemodes cult runes dm user equip to slot or del new obj item clothing shoes cult user slot shoes game gamemodes cult runes dm user spellremove user game gamemodes cult runes dm usr visible message red the rune disappears with a flash of red light and a set of robes appears on game gamemodes cult runes dm user mind spell list new obj effect proc holder spell targeted turf teleport blink cult user game gamemodes cult runes dm if armorworn physician game gamemodes cult runes dm user put in hands new obj item weapon gun magic wand resurrection user game gamemodes cult runes dm if armorworn traveler armorworn marauder game gamemodes cult runes dm usr say tharanak n ghft game gamemodes cult runes dm del user head game gamemodes cult runes dm del user wear suit game gamemodes cult runes dm user equip to slot or del new obj item clothing head helmet space cult user slot head game gamemodes cult runes dm user equip to slot or del new obj item clothing suit space cult user slot wear suit game gamemodes cult runes dm user equip to slot or del new obj item clothing shoes cult user slot shoes game gamemodes cult runes dm user equip to slot or del new obj item weapon storage backpack cultpack user slot back game gamemodes cult runes dm user put in hands new obj item weapon melee cultblade user game gamemodes cult runes dm usr visible message red the rune disappears with a flash of red light and an armored space suit appears on game gamemodes cult runes dm red you are blinded by the flash of red light after you re able to see again you see that you are now wearing an armored space suit game gamemodes cult runes dm if armorworn marauder game gamemodes cult runes dm user spellremove user game gamemodes cult runes dm user mind spell list new obj effect proc holder spell aoe turf conjure creature cult user game gamemodes cult runes dm del src game gamemodes cult runes dm return game gamemodes cult runes dm game gamemodes cult runes dm summon shell summons a construct shell if there s four plasteel sheets on top of the rune game gamemodes cult runes dm game gamemodes cult runes dm game gamemodes game mode dm var list datum game mode usable modes list game gamemodes game mode dm for var datum game mode g in runnable modes game gamemodes game mode dm if g reroll friendly game gamemodes game mode dm usable modes g game gamemodes game mode dm else game gamemodes game mode dm del g game gamemodes game mode dm game gamemodes game mode dm if usable modes game gamemodes game mode dm message admins convert roundtype failed due to no valid modes to convert to please report this error to the coders game gamemodes game mode dm return null game gamemodes game mode dm game gamemodes wizard spellbook dm var spell levels game gamemodes wizard spellbook dm for var obj effect proc holder spell aspell in user mind spell list game gamemodes wizard spellbook dm if initial s name initial aspell name game gamemodes wizard spellbook dm spell levels aspell spell level game gamemodes wizard spellbook dm user mind spell list remove aspell game gamemodes wizard spellbook dm del s game gamemodes wizard spellbook dm return cost spell levels game gamemodes wizard spellbook dm return game gamemodes wizard spellbook dm datum spellbook entry proc getinfo game gamemodes wizard spellbook dm if s game gamemodes wizard spellbook dm s new spell type game gamemodes wizard spellbook dm var datum spellbook entry e new t game gamemodes wizard spellbook dm if e isavailible game gamemodes wizard spellbook dm entries e game gamemodes wizard spellbook dm categories e category game gamemodes wizard spellbook dm else game gamemodes wizard spellbook dm del e game gamemodes wizard spellbook dm tab categories game gamemodes wizard spellbook dm game gamemodes wizard spellbook dm obj item weapon spellbook new game gamemodes wizard spellbook dm game gamemodes wizard spellbook dm initialize game machinery computer medical dm if game machinery computer medical dm r fields pick insane unstable watch stable game machinery computer medical dm continue game machinery computer medical dm game machinery computer medical dm else if prob game machinery computer medical dm del r game machinery computer medical dm continue game machinery computer medical dm game machinery computer medical dm severity game machinery computer medical dm game machinery computer medical dm obj machinery computer med data proc canusemedicalrecordsconsole mob user message game machinery computer security dm temp no game machinery computer security dm game machinery computer security dm if purge all records game machinery computer security dm investigate log has purged all the security records records game machinery computer security dm for var datum data record r in data core security game machinery computer security dm del r game machinery computer security dm data core security cut game machinery computer security dm temp all security records deleted game machinery computer security dm game machinery computer security dm if add entry game machinery computer security dm if istype datum data record game machinery computer security dm h sec hud set security status game machinery computer security dm if delete record security execute game machinery computer security dm investigate log has deleted the security records for records game machinery computer security dm if game machinery computer security dm data core security game machinery computer security dm del game machinery computer security dm game machinery computer security dm if delete record all execute game machinery computer security dm if game machinery computer security dm investigate log has deleted all records for records game machinery computer security dm for var datum data record r in data core medical game machinery computer security dm if r fields fields r fields fields game machinery computer security dm data core medical r game machinery computer security dm del r game machinery computer security dm break game machinery computer security dm data core general game machinery computer security dm del game machinery computer security dm game machinery computer security dm if game machinery computer security dm data core security game machinery computer security dm del game machinery computer security dm else game machinery computer security dm temp this function does not appear to be working at the moment our apologies game machinery computer security dm game machinery computer security dm add fingerprint usr game machinery computer security dm updateusrdialog game machinery computer security dm r fields g fields game machinery computer security dm r fields g fields game machinery computer security dm continue game machinery computer security dm game machinery computer security dm else if prob game machinery computer security dm del r game machinery computer security dm continue game machinery computer security dm game machinery computer security dm severity game machinery computer security dm game machinery computer security dm obj machinery computer secure data proc canusesecurityrecordsconsole mob user game machinery embedded controller embedded controller base dm game machinery embedded controller embedded controller base dm datum computer file embedded program proc post signal datum signal signal comm line game machinery embedded controller embedded controller base dm if master game machinery embedded controller embedded controller base dm master post signal signal comm line game machinery embedded controller embedded controller base dm else game machinery embedded controller embedded controller base dm del signal game machinery embedded controller embedded controller base dm game machinery embedded controller embedded controller base dm datum computer file embedded program proc receive user command command game machinery embedded controller embedded controller base dm game machinery embedded controller embedded controller base dm datum computer file embedded program proc receive signal datum signal signal receive method receive param game machinery embedded controller embedded controller base dm return null game machinery magnet dm game machinery magnet dm if nextmove in list n s e w c r game machinery magnet dm n s e w are directional game machinery magnet dm c is center game machinery magnet dm r is random in magnetic field s bounds game machinery magnet dm del signal game machinery magnet dm break break the loop if the character located is invalid game machinery magnet dm game machinery magnet dm signal data nextmove game machinery magnet dm game machinery magnet dm game machinery overview dm var icon j imap game machinery overview dm game machinery overview dm hi insert i frame delay game machinery overview dm hi insert j frame delay game machinery overview dm game machinery overview dm del i game machinery overview dm del j game machinery overview dm h icon hi game machinery overview dm h layer game machinery overview dm usr mapobjs h game machinery overview dm else game machinery overview dm game machinery overview dm h name i maprefresh map game machinery overview dm game machinery overview dm var icon i imap game machinery overview dm game machinery overview dm h icon i game machinery overview dm del i game machinery overview dm h layer game machinery overview dm usr mapobjs h game machinery overview dm game machinery overview dm endif game machinery overview dm game machinery suit storage unit dm if i end of the cycle game machinery suit storage unit dm if src issuperuv game machinery suit storage unit dm for var obj item item in src game machinery suit storage unit dm item clean blood game machinery suit storage unit dm if istype storage obj item weapon reagent containers food game machinery suit storage unit dm del storage game machinery suit storage unit dm else it was supercycling destroy everything game machinery suit storage unit dm src helmet null game machinery suit storage unit dm src suit null game machinery suit storage unit dm src mask null game machinery suit storage unit dm del storage game machinery suit storage unit dm visible message with a loud whining noise the suit storage unit s door grinds open puffs of ashen smoke come out of its chamber game machinery suit storage unit dm src isbroken game machinery suit storage unit dm src isopen game machinery suit storage unit dm src islocked game machinery suit storage unit dm src eject occupant occupant mixing up these two lines causes bug do not do it game machinery telecomms computers logbrowser dm var datum comm log entry d selectedserver log entries game machinery telecomms computers logbrowser dm game machinery telecomms computers logbrowser dm temp deleted entry game machinery telecomms computers logbrowser dm game machinery telecomms computers logbrowser dm selectedserver log entries remove d game machinery telecomms computers logbrowser dm del d game machinery telecomms computers logbrowser dm game machinery telecomms computers logbrowser dm else game machinery telecomms computers logbrowser dm temp failed no selected machine game machinery telecomms computers logbrowser dm game machinery telecomms computers logbrowser dm if href list game mecha mecha dm internal tank null game mecha mecha dm if loc game mecha mecha dm loc assume air cabin air game mecha mecha dm air update turf game mecha mecha dm else game mecha mecha dm del cabin air game mecha mecha dm cabin air null game mecha mecha dm qdel spark system game mecha mecha dm spark system null game mecha mecha dm game mecha mecha dm mechas list src global mech list game mecha mecha dm var datum gas mixture leaked gas int tank air remove ratio game mecha mecha dm if loc game mecha mecha dm loc assume air leaked gas game mecha mecha dm air update turf game mecha mecha dm else game mecha mecha dm del leaked gas game mecha mecha dm game mecha mecha dm if internal damage mecha int short circuit game mecha mecha dm if get charge game mecha mecha dm spark system start game mecha mecha dm cell charge min cell charge game mecha mecha dm transfer moles pressure delta cabin air return volume cabin air return temperature r ideal gas equation game mecha mecha dm var datum gas mixture removed cabin air remove transfer moles game mecha mecha dm if t air game mecha mecha dm t air merge removed game mecha mecha dm else just delete the cabin gas we re in space or some shit game mecha mecha dm del removed game mecha mecha dm game mecha mecha dm if occupant game mecha mecha dm if cell game mecha mecha dm var cellcharge cell charge cell maxcharge game mecha mecha dm switch cellcharge game mecha mecha construction paths dm var obj item mecha parts chassis const holder holder game mecha mecha construction paths dm const holder construct new datum construction reversible mecha gygax const holder game mecha mecha construction paths dm const holder icon icons mecha mech construction dmi game mecha mecha construction paths dm const holder icon state game mecha mecha construction paths dm const holder density game mecha mecha construction paths dm del src game mecha mecha construction paths dm return game mecha mecha construction paths dm game mecha mecha construction paths dm game mecha mecha construction paths dm datum construction reversible mecha gygax game mecha mecha construction paths dm result obj mecha combat gygax game mecha mecha construction paths dm var obj item mecha parts chassis const holder holder game mecha mecha construction paths dm const holder construct new datum construction reversible mecha firefighter const holder game mecha mecha construction paths dm const holder icon icons mecha mech construction dmi game mecha mecha construction paths dm const holder icon state game mecha mecha construction paths dm const holder density game mecha mecha construction paths dm del src game mecha mecha construction paths dm return game mecha mecha construction paths dm game mecha mecha construction paths dm game mecha mecha construction paths dm datum construction reversible mecha firefighter game mecha mecha construction paths dm result obj mecha working ripley firefighter game mecha mecha construction paths dm game mecha mecha construction paths dm datum construction mecha honker chassis spawn result game mecha mecha construction paths dm var obj item mecha parts chassis const holder holder game mecha mecha construction paths dm const holder construct new datum construction mecha honker const holder game mecha mecha construction paths dm const holder density game mecha mecha construction paths dm del src game mecha mecha construction paths dm return game mecha mecha construction paths dm game mecha mecha construction paths dm game mecha mecha construction paths dm datum construction mecha honker game mecha mecha construction paths dm result obj mecha combat honker game objects effects mines dm name kick mine game objects effects mines dm game objects effects mines dm obj effect mine kickmine mineeffect mob victim game objects effects mines dm if isliving victim victim client game objects effects mines dm victim you have been kicked for no reisin game objects effects mines dm del victim client game objects effects mines dm game objects effects mines dm game objects effects mines dm obj effect mine gas game objects effects mines dm name oxygen mine game objects effects mines dm var gas amount game objects items devices camera bug dm game objects items devices camera bug dm obj item device camera bug destroy game objects items devices camera bug dm if expansion game objects items devices camera bug dm qdel expansion game objects items devices camera bug dm expansion null game objects items devices camera bug dm del src game objects items devices camera bug dm easier to just call del than this nonsense game objects items devices camera bug dm ya no del takes to run on avg this takes less than game objects items devices camera bug dm get cameras game objects items devices camera bug dm for var cam tag in bugged cameras game objects items devices camera bug dm var obj machinery camera camera bugged cameras game objects items holotape dm while n game objects items holotape dm n game objects items holotape dm for var obj item holotape p in cur game objects items holotape dm if p icon state icon dir game objects items holotape dm n game objects items holotape dm del p game objects items holotape dm cur get step cur dir game objects items holotape dm game objects items holotape dm del src game objects items holotape dm return game objects items holotape dm game objects items holotape dm undef max tape range game objects items weapons storage storage dm var obj item sample object game objects items weapons storage storage dm var number game objects items weapons storage storage dm game objects items weapons storage storage dm new obj item sample game objects items weapons storage storage dm if istype sample game objects items weapons storage storage dm del src game objects items weapons storage storage dm sample object sample game objects items weapons storage storage dm number game objects items weapons storage storage dm game objects items weapons storage storage dm game objects items weapons storage storage dm this proc determins the size of the inventory to be displayed please touch it only if you know what you re doing game objects items weapons tanks tanks dm game objects items weapons tanks tanks dm return game objects items weapons tanks tanks dm game objects items weapons tanks tanks dm obj item weapon tank destroy game objects items weapons tanks tanks dm if air contents game objects items weapons tanks tanks dm del air contents game objects items weapons tanks tanks dm game objects items weapons tanks tanks dm ssobj processing remove src game objects items weapons tanks tanks dm game objects items weapons tanks tanks dm game objects items weapons tanks tanks dm linda linda turf tile dm var atmos adjacent turfs linda linda turf tile dm var atmos adjacent turfs amount linda linda turf tile dm var atmos supeconductivity linda linda turf tile dm linda linda turf tile dm turf assume air datum gas mixture giver use this for machines to adjust air linda linda turf tile dm del giver linda linda turf tile dm return linda linda turf tile dm linda linda turf tile dm turf return air linda linda turf tile dm create gas mixture to hold data for passing linda linda turf tile dm var datum gas mixture gm new modules admin admin dm if kick only afk c is afk ignore clients who are not afk modules admin admin dm continue modules admin admin dm if message modules admin admin dm c message modules admin admin dm kicked client names add modules admin admin dm del c modules admin admin dm return kicked client names modules admin admin dm modules admin admin dm returns to let the dragdrop code know we are trapping this event modules admin admin dm returns if we don t plan to trap the event modules admin admin dm datum admins proc cmd ghost drag mob dead observer frommob mob living tomob modules admin admin ranks dm datum admin rank new init name init rights list init adds list init subs modules admin admin ranks dm name init name modules admin admin ranks dm switch name modules admin admin ranks dm if removed null modules admin admin ranks dm spawn modules admin admin ranks dm del src modules admin admin ranks dm throw exception invalid admin rank name modules admin admin ranks dm return modules admin admin ranks dm if init rights rights init rights modules admin admin ranks dm if init adds init adds list modules admin admin ranks dm if init subs init subs list modules admin db ban functions dm if announceinirc modules admin db ban functions dm ban alert applied a on modules admin db ban functions dm modules admin db ban functions dm if kickbannedckey modules admin db ban functions dm if banned mob banned mob client banned mob client ckey banckey modules admin db ban functions dm del banned mob client modules admin db ban functions dm modules admin db ban functions dm modules admin db ban functions dm datum admins proc db ban unban ckey bantype job modules admin db ban functions dm modules admin db ban functions dm if check rights r ban return modules admin dm var admincaster signature modules admin dm modules admin dm datum admins new datum admin rank r ckey modules admin dm if ckey modules admin dm spawn modules admin dm del src modules admin dm throw exception admin datum created without a ckey modules admin dm return modules admin dm if istype r modules admin dm spawn modules admin dm del src modules admin dm throw exception admin datum created without a rank modules admin dm return modules admin dm rank r modules admin dm admincaster signature nanotrasen officer modules admin dm admin datums src modules admin dm modules admin dm client proc deadmin modules admin dm admin datums ckey modules admin dm if holder modules admin dm holder disassociate modules admin dm del holder modules admin dm return modules admin dm modules admin dm this proc checks whether subject has at least one of the rights specified in rights required modules admin dm proc check rights for client subject rights required modules admin dm if subject subject holder subject holder rank modules admin topic dm return modules admin topic dm m you have been kicked from the server modules admin topic dm log admin booted modules admin topic dm message admins booted modules admin topic dm m client null modules admin topic dm del m client modules admin topic dm modules admin topic dm player notes modules admin topic dm else if href list modules admin topic dm var ckey href list modules admin topic dm switch href list modules admin topic dm else modules admin topic dm m no ban appeals url has been set modules admin topic dm log admin has banned nreason nthis will be removed in minutes modules admin topic dm message admins has banned nreason nthis will be removed in minutes modules admin topic dm modules admin topic dm del m client modules admin topic dm qdel m see no reason why to delete mob important stuff can be lost and ban can be lifted before round ends modules admin topic dm if no modules admin topic dm var reason input usr reason reason griefer as text null modules admin topic dm if reason modules admin topic dm return modules admin topic dm log admin has banned nreason nthis is a permanent ban modules admin topic dm message admins has banned nreason nthis is a permanent ban modules admin topic dm feedback inc ban perma modules admin topic dm db ban record bantype perma m reason modules admin topic dm modules admin topic dm del m client modules admin topic dm qdel m modules admin topic dm if cancel modules admin topic dm return modules admin topic dm modules admin topic dm else if href list modules admin verbs sdql dm do parse modules admin verbs sdql dm else modules admin verbs sdql dm query tree val modules admin verbs sdql dm pos modules admin verbs sdql dm modules admin verbs sdql dm del parser modules admin verbs sdql dm modules admin verbs sdql dm return querys modules admin verbs sdql dm modules admin verbs sdql dm modules admin verbs sdql dm modules awaymissions maploader swapmaps dm min x z y z max x z y z modules awaymissions maploader swapmaps dm initializeswapmaps modules awaymissions maploader swapmaps dm if swapmaps compiled maxz modules awaymissions maploader swapmaps dm swapmaps compiled maxy modules awaymissions maploader swapmaps dm swapmaps compiled maxx modules awaymissions maploader swapmaps dm del src modules awaymissions maploader swapmaps dm return modules awaymissions maploader swapmaps dm x x world maxx modules awaymissions maploader swapmaps dm y y world maxy modules awaymissions maploader swapmaps dm z z modules awaymissions maploader swapmaps dm allocateswapmap modules awaymissions maploader swapmaps dm for var obj o in a qdel o modules awaymissions maploader swapmaps dm for var mob m in a modules awaymissions maploader swapmaps dm if m key qdel m modules awaymissions maploader swapmaps dm else m loc null modules awaymissions maploader swapmaps dm areas null modules awaymissions maploader swapmaps dm del a modules awaymissions maploader swapmaps dm delete areas that belong only to this map modules awaymissions maploader swapmaps dm for var area a in areas modules awaymissions maploader swapmaps dm if a a contents len del a modules awaymissions maploader swapmaps dm if world maxx world maxy world maxz cutxyz modules awaymissions maploader swapmaps dm del areas modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm savefile format modules awaymissions maploader swapmaps dm map modules awaymissions maploader swapmaps dm s cd modules awaymissions maploader swapmaps dm s cd modules awaymissions maploader swapmaps dm sleep modules awaymissions maploader swapmaps dm s cd oldcd modules awaymissions maploader swapmaps dm locked modules awaymissions maploader swapmaps dm del areas modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm read savefile s id turf locorner modules awaymissions maploader swapmaps dm var x modules awaymissions maploader swapmaps dm var y modules awaymissions maploader swapmaps dm var z modules awaymissions maploader swapmaps dm s cd modules awaymissions maploader swapmaps dm s cd modules awaymissions maploader swapmaps dm sleep modules awaymissions maploader swapmaps dm s cd oldcd modules awaymissions maploader swapmaps dm locked modules awaymissions maploader swapmaps dm del areas modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm find an empty block on the world map in which to load this map modules awaymissions maploader swapmaps dm if no space is found increase world maxz as necessary if the modules awaymissions maploader swapmaps dm map is greater in x y size than the current world expand modules awaymissions maploader swapmaps dm else modules awaymissions maploader swapmaps dm var list l considerregion world maxx world maxy swapmaps compiled maxz modules awaymissions maploader swapmaps dm l modules awaymissions maploader swapmaps dm l modules awaymissions maploader swapmaps dm l modules awaymissions maploader swapmaps dm del l modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm world maxz max world maxz stretch z if necessary modules awaymissions maploader swapmaps dm if ischunk modules awaymissions maploader swapmaps dm world maxz mz modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm save and delete modules awaymissions maploader swapmaps dm proc unload modules awaymissions maploader swapmaps dm save modules awaymissions maploader swapmaps dm del src modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm proc save modules awaymissions maploader swapmaps dm if id src return modules awaymissions maploader swapmaps dm var savefile s mode new new map sav modules awaymissions maploader swapmaps dm s src modules awaymissions maploader swapmaps dm var list l contents modules awaymissions maploader swapmaps dm if m modules awaymissions maploader swapmaps dm l l copy modules awaymissions maploader swapmaps dm for m in src if m key l m modules awaymissions maploader swapmaps dm if l len s l modules awaymissions maploader swapmaps dm if l contents del l modules awaymissions maploader swapmaps dm read savefile s modules awaymissions maploader swapmaps dm var list l modules awaymissions maploader swapmaps dm if contents len l contents modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm if the icon was a text string it would not have loaded properly modules awaymissions maploader swapmaps dm var ic modules awaymissions maploader swapmaps dm s ic modules awaymissions maploader swapmaps dm if istext ic icon swapmaps iconcache modules awaymissions maploader swapmaps dm if l contents l modules awaymissions maploader swapmaps dm contents l modules awaymissions maploader swapmaps dm del l modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm set this up at runtime as follows modules awaymissions maploader swapmaps dm list modules awaymissions maploader swapmaps dm player dmi player modules awaymissions maploader swapmaps dm properly otherwise the path should always match the map however modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm s cd modules awaymissions maploader swapmaps dm m read s m locorner modules awaymissions maploader swapmaps dm while m locked sleep modules awaymissions maploader swapmaps dm del m modules awaymissions maploader swapmaps dm return modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm proc swapmaps savechunk chunk id turf turf modules awaymissions maploader swapmaps dm if modules awaymissions maploader swapmaps dm world log swapmaps error in swapmaps savechunk modules awaymissions maploader swapmaps dm m max y y modules awaymissions maploader swapmaps dm m max z z modules awaymissions maploader swapmaps dm m mode swapmaps mode modules awaymissions maploader swapmaps dm m save modules awaymissions maploader swapmaps dm while m locked sleep modules awaymissions maploader swapmaps dm del m modules awaymissions maploader swapmaps dm return modules awaymissions maploader swapmaps dm modules awaymissions maploader swapmaps dm proc swapmaps getsize id modules awaymissions maploader swapmaps dm var savefile s modules awaymissions maploader swapmaps dm var text modules client client procs dm if isnum player age player age first connection modules client client procs dm if config panic bunker holder ckey in deadmins modules client client procs dm log access failed login new account attempting to connect during panic bunker modules client client procs dm message admins failed login new account attempting to connect during panic bunker modules client client procs dm src sorry but the server is currently not accepting connections from never before seen players modules client client procs dm del src modules client client procs dm return modules client client procs dm modules client client procs dm if config notify new player age modules client client procs dm message admins new user is connecting here for the first time modules client client procs dm if config irc first connection alert modules crafting guncrafting dm if istype w obj item pipe modules crafting guncrafting dm user you attach the shotgun barrel to the receiver the pins seem loose modules crafting guncrafting dm var obj item weaponcrafting ishotgunconstruction i new obj item weaponcrafting ishotgunconstruction modules crafting guncrafting dm user unequip src modules crafting guncrafting dm user put in hands i modules crafting guncrafting dm del w modules crafting guncrafting dm del src modules crafting guncrafting dm return modules crafting guncrafting dm modules crafting guncrafting dm shotgun modules crafting guncrafting dm modules crafting guncrafting dm obj item weaponcrafting ishotgunconstruction modules crafting guncrafting dm if istype w obj item weaponcrafting stock modules crafting guncrafting dm user you attach the stock to the receiver barrel assembly modules crafting guncrafting dm var obj item weaponcrafting i new obj item weaponcrafting modules crafting guncrafting dm user unequip src modules crafting guncrafting dm user put in hands i modules crafting guncrafting dm del w modules crafting guncrafting dm del src modules crafting guncrafting dm return modules crafting guncrafting dm modules crafting guncrafting dm obj item weaponcrafting modules crafting guncrafting dm name extremely conspicuous metal construction modules crafting guncrafting dm desc a receiver barrel shotgun assembly with a loose wooden stock there s no way you can fire it without the stock coming loose modules flufftext hallucination dm if target client target client images flood images modules flufftext hallucination dm return modules flufftext hallucination dm modules flufftext hallucination dm obj effect hallucination fake flood destroy modules flufftext hallucination dm ssobj processing remove src modules flufftext hallucination dm del flood turfs modules flufftext hallucination dm if target client target client images remove flood images modules flufftext hallucination dm target null modules flufftext hallucination dm del flood images modules flufftext hallucination dm return modules flufftext hallucination dm modules flufftext hallucination dm obj effect hallucination simple xeno modules flufftext hallucination dm image icon icons mob alien dmi modules flufftext hallucination dm image state alienh pounce modules hydroponics hydroponics dm qdel reagent source modules hydroponics hydroponics dm modules hydroponics hydroponics dm h applychemicals s modules hydroponics hydroponics dm modules hydroponics hydroponics dm s clear reagents modules hydroponics hydroponics dm del s modules hydroponics hydroponics dm h update icon modules hydroponics hydroponics dm if reagent source if the source wasn t composted and destroyed modules hydroponics hydroponics dm reagent source update icon modules hydroponics hydroponics dm return modules hydroponics hydroponics dm modules hydroponics seed extractor dm if n amount modules hydroponics seed extractor dm return modules hydroponics seed extractor dm n amount max n amount modules hydroponics seed extractor dm if n amount modules hydroponics seed extractor dm piles n modules hydroponics seed extractor dm del n modules hydroponics seed extractor dm break modules hydroponics seed extractor dm modules hydroponics seed extractor dm for var obj t in contents now we find the seed we need to vend modules hydroponics seed extractor dm var obj item seeds o t modules hydroponics seed extractor dm if o plantname href list o lifespan href list o endurance href list o maturation href list o production href list o yield href list o potency href list modules mob camera camera dm mob camera experience pressure difference modules mob camera camera dm return modules mob camera camera dm modules mob camera camera dm mob camera destroy modules mob camera camera dm modules mob camera camera dm del src modules mob camera camera dm modules mob camera camera dm mob camera login modules mob camera camera dm modules mob camera camera dm update interface modules mob living carbon human blood dm are we dripping or splattering modules mob living carbon human blood dm var list drips list modules mob living carbon human blood dm only a certain number of drips or one large splatter can be on a given turf modules mob living carbon human blood dm for var obj effect decal cleanable blood drip drop in t modules mob living carbon human blood dm drips drop drips modules mob living carbon human blood dm del drop modules mob living carbon human blood dm if large drips len modules mob living carbon human blood dm decal type obj effect decal cleanable blood drip modules mob living carbon human blood dm modules mob living carbon human blood dm find a blood decal or create a new one modules mob living carbon human blood dm b locate decal type in t modules mob living silicon ai ai dm modules mob living silicon ai ai dm if personnel list len modules mob living silicon ai ai dm input input select a crew member as null anything in personnel list modules mob living silicon ai ai dm var icon character icon personnel list modules mob living silicon ai ai dm if character icon modules mob living silicon ai ai dm del holo icon clear old icon so we re not storing it in memory modules mob living silicon ai ai dm holo icon gethologramicon icon character icon modules mob living silicon ai ai dm else modules mob living silicon ai ai dm alert no suitable records found aborting modules mob living silicon ai ai dm modules mob living silicon ai ai dm else modules mob living silicon ai ai dm xeno queen modules mob living silicon ai ai dm space carp modules mob living silicon ai ai dm modules mob living silicon ai ai dm input input please select a hologram as null anything in icon list modules mob living silicon ai ai dm if input modules mob living silicon ai ai dm del holo icon modules mob living silicon ai ai dm switch input modules mob living silicon ai ai dm if default modules mob living silicon ai ai dm holo icon gethologramicon icon icons mob ai dmi modules mob living silicon ai ai dm if floating face modules mob living silicon ai ai dm holo icon gethologramicon icon icons mob ai dmi modules mob living silicon pai death dm var tod world realtime hh mm ss weasellos time of death patch modules mob living silicon pai death dm mind store memory time of death modules mob living silicon pai death dm modules mob living silicon pai death dm new pai s get a brand new mind to prevent meta stuff from their previous life this new mind causes problems down the line if it s not deleted here modules mob living silicon pai death dm read as i have no idea what i m doing but asking for help got me nowhere so this is what you get nodrak modules mob living silicon pai death dm if mind del mind modules mob living silicon pai death dm living mob list src modules mob living silicon pai death dm ghostize modules mob living silicon pai death dm qdel src modules mob living simple animal worm dm newhead attach newheadprevious modules mob living simple animal worm dm modules mob living simple animal worm dm if die modules mob living simple animal worm dm newhead die modules mob living simple animal worm dm modules mob living simple animal worm dm del src modules mob living simple animal worm dm modules mob living simple animal worm dm proc processstomach modules mob living simple animal worm dm for var atom movable stomachcontent in contents modules mob living simple animal worm dm if prob digestionprobability modules mob living simple animal worm dm if istype stomachcontent obj item stack converts to plasma keeping the stack value modules mob living simple animal worm dm if istype stomachcontent obj item stack sheet mineral plasma modules mob living simple animal worm dm var obj item stack oldstack stomachcontent modules mob living simple animal worm dm new obj item stack sheet mineral plasma src oldstack amount modules mob living simple animal worm dm del oldstack modules mob living simple animal worm dm continue modules mob living simple animal worm dm else if istype stomachcontent obj item converts to plasma keeping the w class modules mob living simple animal worm dm var obj item olditem stomachcontent modules mob living simple animal worm dm new obj item stack sheet mineral plasma src olditem w class modules mob living simple animal worm dm del olditem modules mob living simple animal worm dm continue modules mob living simple animal worm dm else modules mob living simple animal worm dm new obj item stack sheet mineral plasma src flatplasmavalue just flat amount modules mob living simple animal worm dm del stomachcontent modules mob living simple animal worm dm continue modules mob living simple animal worm dm modules mob living simple animal worm dm if previous modules mob living simple animal worm dm for var atom movable stomachcontent in contents transfer it along the digestive tract modules mob living simple animal worm dm previous contents stomachcontent modules mob login dm update login details modules mob login dm world update status modules mob login dm modules mob login dm client images null remove the images such as ais being unable to see runes modules mob login dm client screen list remove hud items just in case modules mob login dm if hud used del hud used remove the hud objects modules mob login dm hud used new datum hud src modules mob login dm modules mob login dm next move modules mob login dm sight see self modules mob login dm modules mob new player preferences setup dm backbag modules mob new player preferences setup dm features random features modules mob new player preferences setup dm age rand age min age max modules mob new player preferences setup dm modules mob new player preferences setup dm datum preferences proc update preview icon seriously this is horrendous modules mob new player preferences setup dm del preview icon front modules mob new player preferences setup dm del preview icon side modules mob new player preferences setup dm var icon preview icon null modules mob new player preferences setup dm modules mob new player preferences setup dm if job engsec high cyborg ai check put first to avoid so much unneeded blending modules mob new player preferences setup dm switch job engsec high modules mob new player preferences setup dm if ai modules mob new player preferences setup dm modules mob new player preferences setup dm if preview icon we re busting out modules mob new player preferences setup dm preview icon front new preview icon dir south modules mob new player preferences setup dm preview icon side new preview icon dir west modules mob new player preferences setup dm modules mob new player preferences setup dm del preview icon modules mob new player preferences setup dm return modules mob new player preferences setup dm modules mob new player preferences setup dm var g m modules mob new player preferences setup dm if gender female g f modules mob new player preferences setup dm modules mob new player preferences setup dm if clothes s modules mob new player preferences setup dm preview icon blend clothes s icon overlay modules mob new player preferences setup dm preview icon front new preview icon dir south modules mob new player preferences setup dm preview icon side new preview icon dir west modules mob new player preferences setup dm modules mob new player preferences setup dm del preview icon modules mob new player preferences setup dm del eyes s modules mob new player preferences setup dm del clothes s modules power apc dm area power equip modules power apc dm area power environ modules power apc dm area power change modules power apc dm if occupier modules power apc dm malfvacate modules power apc dm del wires modules power apc dm if cell modules power apc dm qdel cell modules power apc dm if terminal modules power apc dm disconnect terminal modules power apc dm modules projectiles firing dm if istype targloc istype curloc bb modules projectiles firing dm return modules projectiles firing dm if targloc curloc modules projectiles firing dm if bb original user if we target ourselves we go straight to bullet act modules projectiles firing dm user bullet act bb modules projectiles firing dm del bb modules projectiles firing dm return modules projectiles firing dm bb loc get turf user modules projectiles firing dm bb starting get turf user modules projectiles firing dm bb current curloc modules projectiles firing dm bb yo targloc y curloc y modules projectiles projectile magic dm else modules projectiles projectile magic dm new mob key m key modules projectiles projectile magic dm modules projectiles projectile magic dm new mob your form morphs into that of a modules projectiles projectile magic dm modules projectiles projectile magic dm del m modules projectiles projectile magic dm return new mob modules projectiles projectile magic dm modules projectiles projectile magic dm obj item projectile magic animate modules projectiles projectile magic dm name bolt of animation modules projectiles projectile magic dm icon state red modules research rdconsole dm for var t in typesof datum tech datum tech modules research rdconsole dm check tech null modules research rdconsole dm check tech new t modules research rdconsole dm if check tech id id modules research rdconsole dm return name check tech name modules research rdconsole dm del check tech modules research rdconsole dm check tech null modules research rdconsole dm break modules research rdconsole dm modules research rdconsole dm return return name modules research rdconsole dm modules research rdconsole dm for var r in typesof datum reagent datum reagent modules research rdconsole dm temp reagent null modules research rdconsole dm temp reagent new r modules research rdconsole dm if temp reagent id id modules research rdconsole dm return name temp reagent name modules research rdconsole dm del temp reagent modules research rdconsole dm temp reagent null modules research rdconsole dm break modules research rdconsole dm return return name modules research rdconsole dm modules research rdconsole dm obj machinery computer rdconsole proc syncrdevices makes sure it is properly sync ed up with the devices attached to it if any modules research rdconsole dm else if href list reset the r d console s database modules research rdconsole dm griefprotection modules research rdconsole dm var choice alert r d console database reset are you sure you want to reset the r d console s database data lost cannot be recovered continue cancel modules research rdconsole dm if choice continue modules research rdconsole dm screen modules research rdconsole dm del files modules research rdconsole dm files new datum research src modules research rdconsole dm spawn modules research rdconsole dm screen modules research rdconsole dm updateusrdialog modules research rdconsole dm compile options dm ifdef dellogging compile options dm warn compiling del logging this will have additional overheads will warn you if compiling with dellogging compile options dm var list del counter list compile options dm proc log del datum x compile options dm if istype x del counter compile options dm del x compile options dm define del x log del x overrides all del calls with log del compile options dm endif compile options dm compile options dm ifdef testing compile options dm warn compiling in testing mode testing debug messages will be visible compile options dm endif onclick hud action dm onclick hud action dm datum action proc remove mob living t onclick hud action dm if button onclick hud action dm if t client onclick hud action dm t client screen button onclick hud action dm del button onclick hud action dm t actions remove src onclick hud action dm t update action buttons onclick hud action dm owner null onclick hud action dm return onclick hud action dm helpers icons dm var image text image new loc a helpers icons dm text image maptext helpers icons dm text image color averagecolour atom icon helpers icons dm text image pixel x helpers icons dm text image pixel y helpers icons dm del atom icon helpers icons dm return text image helpers icons dm helpers icons dm helpers icons dm find s the average colour of the icon helpers icons dm by vg s comicironic
0
22,519
31,567,808,265
IssuesEvent
2023-09-04 01:18:38
tdwg/hc
https://api.github.com/repos/tdwg/hc
opened
New Term - reportedWeather
Term - add normative Process - under public review Class - Event
## New term * Submitter: Humboldt Extension Task Group * Efficacy Justification (why is this term necessary?): Part of a package of terms in support of biological inventory data. * Demand Justification (name at least two organizations that independently need this term): The Humboldt Extension Task Group proposing this term consists of numerous organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): reportedWeather * Term label (English, not normative): Reported Weather * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event * Definition of the term (normative): A list of weather or climatic conditions present during the dwc:Event. * Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use a key:value encoding schema for a data interchange format such as JSON. * Examples (not normative): {"minimumTemperatureInDegreesFahrenheit": 18, "maximumTemperatureInDegreesFahrenheit": 32} * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
1.0
New Term - reportedWeather - ## New term * Submitter: Humboldt Extension Task Group * Efficacy Justification (why is this term necessary?): Part of a package of terms in support of biological inventory data. * Demand Justification (name at least two organizations that independently need this term): The Humboldt Extension Task Group proposing this term consists of numerous organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): reportedWeather * Term label (English, not normative): Reported Weather * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event * Definition of the term (normative): A list of weather or climatic conditions present during the dwc:Event. * Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use a key:value encoding schema for a data interchange format such as JSON. * Examples (not normative): {"minimumTemperatureInDegreesFahrenheit": 18, "maximumTemperatureInDegreesFahrenheit": 32} * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
process
new term reportedweather new term submitter humboldt extension task group efficacy justification why is this term necessary part of a package of terms in support of biological inventory data demand justification name at least two organizations that independently need this term the humboldt extension task group proposing this term consists of numerous organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version none proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes reportedweather term label english not normative reported weather organized in class e g occurrence event location taxon event definition of the term normative a list of weather or climatic conditions present during the dwc event usage comments recommendations regarding content etc not normative recommended best practice is to use a key value encoding schema for a data interchange format such as json examples not normative minimumtemperatureindegreesfahrenheit maximumtemperatureindegreesfahrenheit refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
1
10,980
13,782,353,100
IssuesEvent
2020-10-08 17:30:57
apache/trafficcontrol
https://api.github.com/repos/apache/trafficcontrol
closed
Error while running github feature from .asf.yaml in trafficcontrol
bug process
Sent to commits@trafficcontrol.apache.org (https://mail-archives.apache.org/mod_mbox/trafficcontrol-commits/202010.mbox/browser): > An error occurred while running github feature in .asf.yaml!: > .asf.yaml: Invalid GitHub label 'CDN' - must be lowercase alphanumerical and <= 35 characters! Related to #5120
1.0
Error while running github feature from .asf.yaml in trafficcontrol - Sent to commits@trafficcontrol.apache.org (https://mail-archives.apache.org/mod_mbox/trafficcontrol-commits/202010.mbox/browser): > An error occurred while running github feature in .asf.yaml!: > .asf.yaml: Invalid GitHub label 'CDN' - must be lowercase alphanumerical and <= 35 characters! Related to #5120
process
error while running github feature from asf yaml in trafficcontrol sent to commits trafficcontrol apache org an error occurred while running github feature in asf yaml asf yaml invalid github label cdn must be lowercase alphanumerical and characters related to
1
292,833
25,243,647,459
IssuesEvent
2022-11-15 09:30:39
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
reopened
Add form for PodDisruptionBudgets
[zube]: Done [zube]: To Test kind/enhancement team/area3
**Detailed Description** Add edit and detail pages with a form to create and edit PodDisruptionBudgets. **Context** PodDisruptionBudgets are important if you want to ensure that your applications stay available during upgrades or maintenance.
1.0
Add form for PodDisruptionBudgets - **Detailed Description** Add edit and detail pages with a form to create and edit PodDisruptionBudgets. **Context** PodDisruptionBudgets are important if you want to ensure that your applications stay available during upgrades or maintenance.
non_process
add form for poddisruptionbudgets detailed description add edit and detail pages with a form to create and edit poddisruptionbudgets context poddisruptionbudgets are important if you want to ensure that your applications stay available during upgrades or maintenance
0
54,775
3,071,267,393
IssuesEvent
2015-08-19 10:55:16
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
opened
r500 Не скачивается весь каталог, если использован частичный просмотр шары (показать список файлов)
bug imported Priority-High Usability
_From [tret2...@gmail.com](https://code.google.com/u/116508191076211387118/) on August 11, 2010 13:02:48_ При использовании функции просмотреть список файлов если ставить каталог на скачку, то он ставит тольно подгруженные подкаталоги. Сделайть так, чтоб он подгружал автоматически подкаталоги и ставил на скачивание папку полностью. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=105_
1.0
r500 Не скачивается весь каталог, если использован частичный просмотр шары (показать список файлов) - _From [tret2...@gmail.com](https://code.google.com/u/116508191076211387118/) on August 11, 2010 13:02:48_ При использовании функции просмотреть список файлов если ставить каталог на скачку, то он ставит тольно подгруженные подкаталоги. Сделайть так, чтоб он подгружал автоматически подкаталоги и ставил на скачивание папку полностью. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=105_
non_process
не скачивается весь каталог если использован частичный просмотр шары показать список файлов from on august при использовании функции просмотреть список файлов если ставить каталог на скачку то он ставит тольно подгруженные подкаталоги сделайть так чтоб он подгружал автоматически подкаталоги и ставил на скачивание папку полностью original issue
0
118,663
4,751,650,917
IssuesEvent
2016-10-23 00:56:22
pathwaysmedical/frasernw
https://api.github.com/repos/pathwaysmedical/frasernw
closed
add status dropdown to "only works out of hospitals and clinics"
ChangeRequest High Priority
A request that came from #256 that would prevent needing to relabel specialists who are retiring to "responded to survey" in order to give them this status.
1.0
add status dropdown to "only works out of hospitals and clinics" - A request that came from #256 that would prevent needing to relabel specialists who are retiring to "responded to survey" in order to give them this status.
non_process
add status dropdown to only works out of hospitals and clinics a request that came from that would prevent needing to relabel specialists who are retiring to responded to survey in order to give them this status
0
266,323
20,146,103,694
IssuesEvent
2022-02-09 07:40:38
vaadin/quarkus
https://api.github.com/repos/vaadin/quarkus
closed
Supporting Vaadin add-ons
documentation enhancement
> If you can’t modify the dependency, you can still index it by adding quarkus.index-dependency entries to your application.properties: ``` quarkus.index-dependency.<name>.group-id= quarkus.index-dependency.<name>.artifact-id= quarkus.index-dependency.<name>.classifier=(this one is optional) ``` ### Acceptance Criteria - [x] Document the necessary steps to make any Vaadin add-on compatible with Quarkus on the application project - [x] Add an integration test to the Quarkus project that uses an add-on without yandex index but it still works as it is configured in the project like ^
1.0
Supporting Vaadin add-ons - > If you can’t modify the dependency, you can still index it by adding quarkus.index-dependency entries to your application.properties: ``` quarkus.index-dependency.<name>.group-id= quarkus.index-dependency.<name>.artifact-id= quarkus.index-dependency.<name>.classifier=(this one is optional) ``` ### Acceptance Criteria - [x] Document the necessary steps to make any Vaadin add-on compatible with Quarkus on the application project - [x] Add an integration test to the Quarkus project that uses an add-on without yandex index but it still works as it is configured in the project like ^
non_process
supporting vaadin add ons if you can’t modify the dependency you can still index it by adding quarkus index dependency entries to your application properties quarkus index dependency group id quarkus index dependency artifact id quarkus index dependency classifier this one is optional acceptance criteria document the necessary steps to make any vaadin add on compatible with quarkus on the application project add an integration test to the quarkus project that uses an add on without yandex index but it still works as it is configured in the project like
0
166,532
12,960,734,989
IssuesEvent
2020-07-20 14:45:20
ubtue/DatenProbleme
https://api.github.com/repos/ubtue/DatenProbleme
closed
ISSN 2196-8020 | Kerygma und Dogma | Tabulatoren im Abstract
Fehlerquelle: Website Fehlerquelle: unbekannt Zotero_SEMI-AUTO ready for testing
Wenn man die Artikel in Singledownloadverfahren herunterläd, enthalten die Abstracts bei dieser Zeitschrift viele Tabulatoren. Wenn man es per copy & past überträgt ist dies nicht der Fall. Diese führen beim Upload zu einem Fehler, so dass die Datei in der WinIBW nicht gespeichert werden kann. Hier: https://www.vr-elibrary.de/doi/10.13109/kedo.2020.66.2.97 sieht es so aus, wenn man es sich in Word anschaut: ![grafik](https://user-images.githubusercontent.com/29707105/87913496-85037b80-ca6f-11ea-98d2-7003ec2044d6.png) Kann man die entfernen?
1.0
ISSN 2196-8020 | Kerygma und Dogma | Tabulatoren im Abstract - Wenn man die Artikel in Singledownloadverfahren herunterläd, enthalten die Abstracts bei dieser Zeitschrift viele Tabulatoren. Wenn man es per copy & past überträgt ist dies nicht der Fall. Diese führen beim Upload zu einem Fehler, so dass die Datei in der WinIBW nicht gespeichert werden kann. Hier: https://www.vr-elibrary.de/doi/10.13109/kedo.2020.66.2.97 sieht es so aus, wenn man es sich in Word anschaut: ![grafik](https://user-images.githubusercontent.com/29707105/87913496-85037b80-ca6f-11ea-98d2-7003ec2044d6.png) Kann man die entfernen?
non_process
issn kerygma und dogma tabulatoren im abstract wenn man die artikel in singledownloadverfahren herunterläd enthalten die abstracts bei dieser zeitschrift viele tabulatoren wenn man es per copy past überträgt ist dies nicht der fall diese führen beim upload zu einem fehler so dass die datei in der winibw nicht gespeichert werden kann hier sieht es so aus wenn man es sich in word anschaut kann man die entfernen
0
13,162
15,590,100,500
IssuesEvent
2021-03-18 08:56:42
CNPMNC-KDH/TKB
https://api.github.com/repos/CNPMNC-KDH/TKB
opened
Student schedule app
DuongThienKhoi NguyenThanhDuy Processing...... VuDuyVietHoang
Xây dựng ứng dụng quản lý thời khóa biểu sinh viên giúp người dùng quản lý thời gian học tập tốt hơn. - [ ] Hoàn thành - [ ] Hoàn thành - [ ] Hoàn thành - [ ] Hoàn thành xây dựng ứng dụng - [ ] Chạy thử ứng dụng trên nhiều nền tảng. - [ ] Kiểm thử và sửa lỗi. - [ ] Phát hành sản phẩm lên các nền tảng cửa hàng trực tuyến.
1.0
Student schedule app - Xây dựng ứng dụng quản lý thời khóa biểu sinh viên giúp người dùng quản lý thời gian học tập tốt hơn. - [ ] Hoàn thành - [ ] Hoàn thành - [ ] Hoàn thành - [ ] Hoàn thành xây dựng ứng dụng - [ ] Chạy thử ứng dụng trên nhiều nền tảng. - [ ] Kiểm thử và sửa lỗi. - [ ] Phát hành sản phẩm lên các nền tảng cửa hàng trực tuyến.
process
student schedule app xây dựng ứng dụng quản lý thời khóa biểu sinh viên giúp người dùng quản lý thời gian học tập tốt hơn hoàn thành hoàn thành hoàn thành hoàn thành xây dựng ứng dụng chạy thử ứng dụng trên nhiều nền tảng kiểm thử và sửa lỗi phát hành sản phẩm lên các nền tảng cửa hàng trực tuyến
1
805,953
29,765,957,187
IssuesEvent
2023-06-15 00:58:00
GTNewHorizons/GT-New-Horizons-Modpack
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
closed
Dump aspect command from gt++ not working properly
Type: bugMinor Priority: very low Status: Triage
### Your GTNH Discord Username LEO_PREMIUM#3437 ### Your Pack Version 2.2.8 ### Your Server sp ### Type of Server Single Player ### Your Expectation write all aspects of all items. ### The Reality the file never gets finished. (checked with several friends, same situation) do not even have iron ingot in the file. ### Your Proposal someone have a look at the code. https://github.com/GTNewHorizons/GTplusplus/tree/80c790b8a7b2c61215d636b3ce14715580b2f5eb/src/main/java/gtPlusPlus/xmod/thaumcraft should be here. ### Final Checklist - [X] I have searched this issue tracker and there is nothing similar already. Posting on a closed issue saying the bug still exists will prompt us to investigate and reopen it once we confirm your report. - [X] I can reproduce this problem consistently by follow the exact steps I described above, or this does not need reproducing, e.g. recipe loophole. - [X] I have asked other people and they confirm they also have this problem by follow the exact steps I described above, or this does not need reproducing, e.g. recipe loophole.
1.0
Dump aspect command from gt++ not working properly - ### Your GTNH Discord Username LEO_PREMIUM#3437 ### Your Pack Version 2.2.8 ### Your Server sp ### Type of Server Single Player ### Your Expectation write all aspects of all items. ### The Reality the file never gets finished. (checked with several friends, same situation) do not even have iron ingot in the file. ### Your Proposal someone have a look at the code. https://github.com/GTNewHorizons/GTplusplus/tree/80c790b8a7b2c61215d636b3ce14715580b2f5eb/src/main/java/gtPlusPlus/xmod/thaumcraft should be here. ### Final Checklist - [X] I have searched this issue tracker and there is nothing similar already. Posting on a closed issue saying the bug still exists will prompt us to investigate and reopen it once we confirm your report. - [X] I can reproduce this problem consistently by follow the exact steps I described above, or this does not need reproducing, e.g. recipe loophole. - [X] I have asked other people and they confirm they also have this problem by follow the exact steps I described above, or this does not need reproducing, e.g. recipe loophole.
non_process
dump aspect command from gt not working properly your gtnh discord username leo premium your pack version your server sp type of server single player your expectation write all aspects of all items the reality the file never gets finished checked with several friends same situation do not even have iron ingot in the file your proposal someone have a look at the code should be here final checklist i have searched this issue tracker and there is nothing similar already posting on a closed issue saying the bug still exists will prompt us to investigate and reopen it once we confirm your report i can reproduce this problem consistently by follow the exact steps i described above or this does not need reproducing e g recipe loophole i have asked other people and they confirm they also have this problem by follow the exact steps i described above or this does not need reproducing e g recipe loophole
0
18,314
24,429,593,543
IssuesEvent
2022-10-06 07:05:09
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/transform] Concat function concatenates too many times
question processor/transform
## Description I'm trying to change the service name to service.namespace:service.name (delimiter doesn't really matter) To do that, I've tried the following query: ``` set(attributes[service.name],Concat(": ",attributes[service.namespace],attributes[service.name])) ``` The result I'm getting is ```<namespace>:<namespace>:<namespace>:....<namespace>:<name>``` The namespace is concatenated an arbitrary number of times. Name exactly once. I've also tried the following queries, they result in the same behaviour: ``` set(attributes[temp],Concat(": ",attributes[service.namespace],attributes[service.name])) set(attributes[service.name],attributes[temp]) delete_key(attributes[temp]) #If I omit this line, temp is also assigned <namespace>:<namespace>:<namespace>:....<namespace>:<name> ``` ``` set(attributes[service.name],Concat("",attributes[service.namespace],":",attributes[service.name])) ``` Looking at the source code (there is no documentation for Concat) the queries seem ok. ### Collector version 0.60.0 ### Environment information OpenTelemetry Collector Contrib 0.60.0 deployed on Openshift using OpenTelemetry Operator 0.60.0 Traces backend is Jaeger 1.35 deployed on Openshift with Jaeger Operator I am using 2 pipelines, both use the same traces receivers (otlp,jaeger) and send to different Jaeger instances. Transform processor is only used on one of them. The other one uses spanmetrics processor (which is configured correctly and works) ### OpenTelemetry Collector configuration ```yaml receivers: otlp: jaeger: processors: batch: spanmetrics: #Configured correctly, full configuration is irrelevant transform: traces: queries: # The queries described above exporters: jaeger: endpoint: <jaeger-collector> jaeger/2: endpoint: <jaeger-2-collector> pipelines: traces: receivers: [otlp,jaeger] processors: [spanmetrics,batch] exporters: [jaeger] traces/2: receivers: [otlp,jaeger] processors: [transform,batch] exporters: [jaeger/2] ```
1.0
[processor/transform] Concat function concatenates too many times - ## Description I'm trying to change the service name to service.namespace:service.name (delimiter doesn't really matter) To do that, I've tried the following query: ``` set(attributes[service.name],Concat(": ",attributes[service.namespace],attributes[service.name])) ``` The result I'm getting is ```<namespace>:<namespace>:<namespace>:....<namespace>:<name>``` The namespace is concatenated an arbitrary number of times. Name exactly once. I've also tried the following queries, they result in the same behaviour: ``` set(attributes[temp],Concat(": ",attributes[service.namespace],attributes[service.name])) set(attributes[service.name],attributes[temp]) delete_key(attributes[temp]) #If I omit this line, temp is also assigned <namespace>:<namespace>:<namespace>:....<namespace>:<name> ``` ``` set(attributes[service.name],Concat("",attributes[service.namespace],":",attributes[service.name])) ``` Looking at the source code (there is no documentation for Concat) the queries seem ok. ### Collector version 0.60.0 ### Environment information OpenTelemetry Collector Contrib 0.60.0 deployed on Openshift using OpenTelemetry Operator 0.60.0 Traces backend is Jaeger 1.35 deployed on Openshift with Jaeger Operator I am using 2 pipelines, both use the same traces receivers (otlp,jaeger) and send to different Jaeger instances. Transform processor is only used on one of them. The other one uses spanmetrics processor (which is configured correctly and works) ### OpenTelemetry Collector configuration ```yaml receivers: otlp: jaeger: processors: batch: spanmetrics: #Configured correctly, full configuration is irrelevant transform: traces: queries: # The queries described above exporters: jaeger: endpoint: <jaeger-collector> jaeger/2: endpoint: <jaeger-2-collector> pipelines: traces: receivers: [otlp,jaeger] processors: [spanmetrics,batch] exporters: [jaeger] traces/2: receivers: [otlp,jaeger] processors: [transform,batch] exporters: [jaeger/2] ```
process
concat function concatenates too many times description i m trying to change the service name to service namespace service name delimiter doesn t really matter to do that i ve tried the following query set attributes concat attributes attributes the result i m getting is the namespace is concatenated an arbitrary number of times name exactly once i ve also tried the following queries they result in the same behaviour set attributes concat attributes attributes set attributes attributes delete key attributes if i omit this line temp is also assigned set attributes concat attributes attributes looking at the source code there is no documentation for concat the queries seem ok collector version environment information opentelemetry collector contrib deployed on openshift using opentelemetry operator traces backend is jaeger deployed on openshift with jaeger operator i am using pipelines both use the same traces receivers otlp jaeger and send to different jaeger instances transform processor is only used on one of them the other one uses spanmetrics processor which is configured correctly and works opentelemetry collector configuration yaml receivers otlp jaeger processors batch spanmetrics configured correctly full configuration is irrelevant transform traces queries the queries described above exporters jaeger endpoint jaeger endpoint pipelines traces receivers processors exporters traces receivers processors exporters
1
55,173
6,444,654,511
IssuesEvent
2017-08-12 15:11:40
nodejs/node
https://api.github.com/repos/nodejs/node
opened
Investigate test-timers-blocking-callback
CI / flaky test timers windows
Unrelated failure of `sequential/test-timers-blocking-callback` on `win2008r2` in https://github.com/nodejs/node/pull/14773: ``` not ok 438 sequential/test-timers-blocking-callback --- duration_ms: 2.877 severity: fail stack: |- assert.js:42 throw new errors.AssertionError({ ^ AssertionError [ERR_ASSERTION]: false == true at blockingCallback (c:\workspace\node-test-binary-windows\RUN_SUBSET\1\VS_VERSION\vs2015\label\win2008r2\test\sequential\test-timers-blocking-callback.js:51:5) at ontimeout (timers.js:469:11) at tryOnTimeout (timers.js:304:5) at Timer.listOnTimeout (timers.js:264:5) ... ``` https://ci.nodejs.org/job/node-test-binary-windows/10459/RUN_SUBSET=1,VS_VERSION=vs2015,label=win2008r2/console /cc @nodejs/testing @nodejs/platform-windows
1.0
Investigate test-timers-blocking-callback - Unrelated failure of `sequential/test-timers-blocking-callback` on `win2008r2` in https://github.com/nodejs/node/pull/14773: ``` not ok 438 sequential/test-timers-blocking-callback --- duration_ms: 2.877 severity: fail stack: |- assert.js:42 throw new errors.AssertionError({ ^ AssertionError [ERR_ASSERTION]: false == true at blockingCallback (c:\workspace\node-test-binary-windows\RUN_SUBSET\1\VS_VERSION\vs2015\label\win2008r2\test\sequential\test-timers-blocking-callback.js:51:5) at ontimeout (timers.js:469:11) at tryOnTimeout (timers.js:304:5) at Timer.listOnTimeout (timers.js:264:5) ... ``` https://ci.nodejs.org/job/node-test-binary-windows/10459/RUN_SUBSET=1,VS_VERSION=vs2015,label=win2008r2/console /cc @nodejs/testing @nodejs/platform-windows
non_process
investigate test timers blocking callback unrelated failure of sequential test timers blocking callback on in not ok sequential test timers blocking callback duration ms severity fail stack assert js throw new errors assertionerror assertionerror false true at blockingcallback c workspace node test binary windows run subset vs version label test sequential test timers blocking callback js at ontimeout timers js at tryontimeout timers js at timer listontimeout timers js cc nodejs testing nodejs platform windows
0
16,001
20,188,208,184
IssuesEvent
2022-02-11 01:18:07
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Discover and remediate common risks to improve Secure Score in Azure Security Center
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Security & Compliance Security Center
<a href="https://docs.microsoft.com/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks">Discover and remediate common risks to improve Secure Score in Azure Security Center</a> <p><b>Why Consider This?</b></p> Secure Score is a snapshot of your security posture relative to Microsoft recommendations. Identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers. <p><b>Context</b></p> <p><span>When routine and well-established attack vectors are mitigated, attackers are forced to acquire and use advanced or untested attack methods. Azure Secure Score in Azure Security Center is one method to easily identify and address well-known issues in Azure.</span></p> <p><b>Suggested Actions</b></p> <p><span>Identify well-known risks for your Azure tenants, remediate those risks, and track your progress using Secure Score</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks" target="_blank"><span>Discover and remediate common risks</span></a><span /></p>
1.0
Discover and remediate common risks to improve Secure Score in Azure Security Center - <a href="https://docs.microsoft.com/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks">Discover and remediate common risks to improve Secure Score in Azure Security Center</a> <p><b>Why Consider This?</b></p> Secure Score is a snapshot of your security posture relative to Microsoft recommendations. Identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers. <p><b>Context</b></p> <p><span>When routine and well-established attack vectors are mitigated, attackers are forced to acquire and use advanced or untested attack methods. Azure Secure Score in Azure Security Center is one method to easily identify and address well-known issues in Azure.</span></p> <p><b>Suggested Actions</b></p> <p><span>Identify well-known risks for your Azure tenants, remediate those risks, and track your progress using Secure Score</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks" target="_blank"><span>Discover and remediate common risks</span></a><span /></p>
process
discover and remediate common risks to improve secure score in azure security center why consider this secure score is a snapshot of your security posture relative to microsoft recommendations identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers context when routine and well established attack vectors are mitigated attackers are forced to acquire and use advanced or untested attack methods azure secure score in azure security center is one method to easily identify and address well known issues in azure suggested actions identify well known risks for your azure tenants remediate those risks and track your progress using secure score learn more discover and remediate common risks
1
317,392
27,235,022,576
IssuesEvent
2023-02-21 15:44:40
JoshuaKGoldberg/ts-api-utils
https://api.github.com/repos/JoshuaKGoldberg/ts-api-utils
closed
Tests are failing on main
area: testing status: accepting prs
<!-- Note: Please must use one of our issue templates to file an issue! 🛑 --> <!-- 👉 https://github.com/JoshuaKGoldberg/ts-api-utils/issues/new/choose 👈 --> <!-- **Issues that should have been filed with a template will be closed without action, and we will ask you to use a template.** --> <!-- This blank issue template is only for issues that don't fit any of the templates. --> ## Overview Tests are currently failing on main for ts 4.2
1.0
Tests are failing on main - <!-- Note: Please must use one of our issue templates to file an issue! 🛑 --> <!-- 👉 https://github.com/JoshuaKGoldberg/ts-api-utils/issues/new/choose 👈 --> <!-- **Issues that should have been filed with a template will be closed without action, and we will ask you to use a template.** --> <!-- This blank issue template is only for issues that don't fit any of the templates. --> ## Overview Tests are currently failing on main for ts 4.2
non_process
tests are failing on main overview tests are currently failing on main for ts
0
9,633
12,598,177,894
IssuesEvent
2020-06-11 02:10:40
CGAL/cgal
https://api.github.com/repos/CGAL/cgal
closed
CGAL::hierarchy_simplify_point_set's behavior mismatch with documentation
Pkg::Point_set_processing_3 doc bug
## Issue Details unexpected behavior CGAL::hierarchy_simplify_point_set's documentation says it will "Recursively split the point set until the clusters have less than size elements or until their variation factor is below var_max" But the real behavior is "Recursively split the point set until the clusters have less than size elements AND until their variation factor is below var_max" So I wonder which is the expect result. Thanks.
1.0
CGAL::hierarchy_simplify_point_set's behavior mismatch with documentation - ## Issue Details unexpected behavior CGAL::hierarchy_simplify_point_set's documentation says it will "Recursively split the point set until the clusters have less than size elements or until their variation factor is below var_max" But the real behavior is "Recursively split the point set until the clusters have less than size elements AND until their variation factor is below var_max" So I wonder which is the expect result. Thanks.
process
cgal hierarchy simplify point set s behavior mismatch with documentation issue details unexpected behavior cgal hierarchy simplify point set s documentation says it will recursively split the point set until the clusters have less than size elements or until their variation factor is below var max but the real behavior is recursively split the point set until the clusters have less than size elements and until their variation factor is below var max so i wonder which is the expect result thanks
1
132,289
18,266,296,776
IssuesEvent
2021-10-04 08:51:21
artsking/linux-3.0.35_CVE-2020-15436_withPatch
https://api.github.com/repos/artsking/linux-3.0.35_CVE-2020-15436_withPatch
closed
CVE-2013-0311 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed
security vulnerability
## CVE-2013-0311 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35_CVE-2020-15436_withPatch/commit/594a70cb9871ddd73cf61197bb1a2a1b1777a7ae">594a70cb9871ddd73cf61197bb1a2a1b1777a7ae</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/vhost/vhost.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The translate_desc function in drivers/vhost/vhost.c in the Linux kernel before 3.7 does not properly handle cross-region descriptors, which allows guest OS users to obtain host OS privileges by leveraging KVM guest OS privileges. <p>Publish Date: 2013-02-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-0311>CVE-2013-0311</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2013-0311">https://nvd.nist.gov/vuln/detail/CVE-2013-0311</a></p> <p>Release Date: 2013-02-22</p> <p>Fix Resolution: 3.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2013-0311 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2013-0311 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35_CVE-2020-15436_withPatch/commit/594a70cb9871ddd73cf61197bb1a2a1b1777a7ae">594a70cb9871ddd73cf61197bb1a2a1b1777a7ae</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/vhost/vhost.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The translate_desc function in drivers/vhost/vhost.c in the Linux kernel before 3.7 does not properly handle cross-region descriptors, which allows guest OS users to obtain host OS privileges by leveraging KVM guest OS privileges. <p>Publish Date: 2013-02-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-0311>CVE-2013-0311</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2013-0311">https://nvd.nist.gov/vuln/detail/CVE-2013-0311</a></p> <p>Release Date: 2013-02-22</p> <p>Fix Resolution: 3.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers vhost vhost c vulnerability details the translate desc function in drivers vhost vhost c in the linux kernel before does not properly handle cross region descriptors which allows guest os users to obtain host os privileges by leveraging kvm guest os privileges publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
19,635
25,997,874,773
IssuesEvent
2022-12-20 13:07:14
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
processing pipeline `to_date` ignores timezone
processing bug triaged to-verify
## Expected Behavior using the `to_date` function in a processing pipeline to transform a date into the local timezone to be able to compare it like in the following rule: ``` rule "off work hours" when ( to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) >= 0 AND to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) <= 6 ) OR ( to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) >= 18 AND to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) <= 0 ) then set_field("trigger_workhours_off", true); end ``` ``` rule "off work weekend" when // from Monday (1) to Sunday (7) to_long(to_date($message.timestamp, "Asia/Manila").dayOfWeek) == 7 OR to_long(to_date($message.timestamp, "Asia/Manila").dayOfWeek) == 6 then set_field("trigger_workhours_off", true); end ``` ## Current Behavior Graylog does not honour the timezone but the online docs announce this option: ![grafik](https://user-images.githubusercontent.com/404238/65500328-6e2be100-debf-11e9-9e1f-7debdb7c596c.png) ## Context Make decisions based on time is hard and sometimes impossible if you need do the calculation in UTC - as this is what is currently only working in Graylog. ## Your Environment * Graylog Version: 3.1
1.0
processing pipeline `to_date` ignores timezone - ## Expected Behavior using the `to_date` function in a processing pipeline to transform a date into the local timezone to be able to compare it like in the following rule: ``` rule "off work hours" when ( to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) >= 0 AND to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) <= 6 ) OR ( to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) >= 18 AND to_long(to_date($message.timestamp, "Asia/Manila").hourOfDay) <= 0 ) then set_field("trigger_workhours_off", true); end ``` ``` rule "off work weekend" when // from Monday (1) to Sunday (7) to_long(to_date($message.timestamp, "Asia/Manila").dayOfWeek) == 7 OR to_long(to_date($message.timestamp, "Asia/Manila").dayOfWeek) == 6 then set_field("trigger_workhours_off", true); end ``` ## Current Behavior Graylog does not honour the timezone but the online docs announce this option: ![grafik](https://user-images.githubusercontent.com/404238/65500328-6e2be100-debf-11e9-9e1f-7debdb7c596c.png) ## Context Make decisions based on time is hard and sometimes impossible if you need do the calculation in UTC - as this is what is currently only working in Graylog. ## Your Environment * Graylog Version: 3.1
process
processing pipeline to date ignores timezone expected behavior using the to date function in a processing pipeline to transform a date into the local timezone to be able to compare it like in the following rule rule off work hours when to long to date message timestamp asia manila hourofday and to long to date message timestamp asia manila hourofday or to long to date message timestamp asia manila hourofday and to long to date message timestamp asia manila hourofday then set field trigger workhours off true end rule off work weekend when from monday to sunday to long to date message timestamp asia manila dayofweek or to long to date message timestamp asia manila dayofweek then set field trigger workhours off true end current behavior graylog does not honour the timezone but the online docs announce this option context make decisions based on time is hard and sometimes impossible if you need do the calculation in utc as this is what is currently only working in graylog your environment graylog version
1
12,641
15,017,146,236
IssuesEvent
2021-02-01 10:28:27
KratosMultiphysics/Kratos
https://api.github.com/repos/KratosMultiphysics/Kratos
closed
[EmbeddedSkinVisualizationProcess] MPI PointOutputProcess of points in visualization model part called #cores times
Bug Parallel-MPI Post Process
**Description** When running an embedded MPI simulation with a point output (or derived classes) of a point in the skin visualization model part, several things happen, that do not happen for a fluid_computational_model_part: - if 28 cores/ processes are used, there are 27 warnings, that the point was not found - if the output is supposed to be written to a directory, that does not exist before starting the simulation, there are error messages, because processes try to create that directory, although one process already created it - the output file will be overwritten 27 times, so that there is an extremely high probability that the last process that writes is a process that could not find the element, so no useful output is created (element id -1 and values 0.0) **Bypassing the problem** --> the last problem can be eliminated by a simple change in the PointOutputProcess, so that a process doesn't write a file at all, when it hasn't found a matching element. --> In case the directory is created in advance and the above change is done in point_output_process.py useful output is written. **Scope** - EmbeddedSkinVisualizationProcess - PointOutputProcess **To Reproduce** - embedded simulation with MPI - a EmbeddedSkinVisualizationProcess to create a visualization model part - a PointOutputProcess of a point of the visualization model part see [visOutputMPI.zip](https://github.com/KratosMultiphysics/Kratos/files/5599109/visOutputMPI.zip) **Expected behavior** No warnings and no need to prevent point output file of skin visualization model part from being overwritten and no need to create output directory in advance. **Environment** - OS: Linux Cluster - Branch: master - Python 3.6
1.0
[EmbeddedSkinVisualizationProcess] MPI PointOutputProcess of points in visualization model part called #cores times - **Description** When running an embedded MPI simulation with a point output (or derived classes) of a point in the skin visualization model part, several things happen, that do not happen for a fluid_computational_model_part: - if 28 cores/ processes are used, there are 27 warnings, that the point was not found - if the output is supposed to be written to a directory, that does not exist before starting the simulation, there are error messages, because processes try to create that directory, although one process already created it - the output file will be overwritten 27 times, so that there is an extremely high probability that the last process that writes is a process that could not find the element, so no useful output is created (element id -1 and values 0.0) **Bypassing the problem** --> the last problem can be eliminated by a simple change in the PointOutputProcess, so that a process doesn't write a file at all, when it hasn't found a matching element. --> In case the directory is created in advance and the above change is done in point_output_process.py useful output is written. **Scope** - EmbeddedSkinVisualizationProcess - PointOutputProcess **To Reproduce** - embedded simulation with MPI - a EmbeddedSkinVisualizationProcess to create a visualization model part - a PointOutputProcess of a point of the visualization model part see [visOutputMPI.zip](https://github.com/KratosMultiphysics/Kratos/files/5599109/visOutputMPI.zip) **Expected behavior** No warnings and no need to prevent point output file of skin visualization model part from being overwritten and no need to create output directory in advance. **Environment** - OS: Linux Cluster - Branch: master - Python 3.6
process
mpi pointoutputprocess of points in visualization model part called cores times description when running an embedded mpi simulation with a point output or derived classes of a point in the skin visualization model part several things happen that do not happen for a fluid computational model part if cores processes are used there are warnings that the point was not found if the output is supposed to be written to a directory that does not exist before starting the simulation there are error messages because processes try to create that directory although one process already created it the output file will be overwritten times so that there is an extremely high probability that the last process that writes is a process that could not find the element so no useful output is created element id and values bypassing the problem the last problem can be eliminated by a simple change in the pointoutputprocess so that a process doesn t write a file at all when it hasn t found a matching element in case the directory is created in advance and the above change is done in point output process py useful output is written scope embeddedskinvisualizationprocess pointoutputprocess to reproduce embedded simulation with mpi a embeddedskinvisualizationprocess to create a visualization model part a pointoutputprocess of a point of the visualization model part see expected behavior no warnings and no need to prevent point output file of skin visualization model part from being overwritten and no need to create output directory in advance environment os linux cluster branch master python
1
8,751
11,873,194,626
IssuesEvent
2020-03-26 16:55:40
pacificclimate/climate-explorer-data-prep
https://api.github.com/repos/pacificclimate/climate-explorer-data-prep
closed
Calculate tasmean data
process new data
plan2adapt displays `tasmean` data, where `tasmean` is defined as the mean of `tasmax` and `tasmin`. We've just been doing that calculation manually whenever we needed to display `tasmean` numerical data, but now we want to display a `tasmean` map, which requires creating an actual dataset. At a minimum, we need to be able to create maps for the anusplin 6190 dataset, and the PCIC12 rcp85 2020, 2050, and 2080, though I expect once we have a tasmean-making script, we might as well do every dataset we might need.
1.0
Calculate tasmean data - plan2adapt displays `tasmean` data, where `tasmean` is defined as the mean of `tasmax` and `tasmin`. We've just been doing that calculation manually whenever we needed to display `tasmean` numerical data, but now we want to display a `tasmean` map, which requires creating an actual dataset. At a minimum, we need to be able to create maps for the anusplin 6190 dataset, and the PCIC12 rcp85 2020, 2050, and 2080, though I expect once we have a tasmean-making script, we might as well do every dataset we might need.
process
calculate tasmean data displays tasmean data where tasmean is defined as the mean of tasmax and tasmin we ve just been doing that calculation manually whenever we needed to display tasmean numerical data but now we want to display a tasmean map which requires creating an actual dataset at a minimum we need to be able to create maps for the anusplin dataset and the and though i expect once we have a tasmean making script we might as well do every dataset we might need
1