Document
stringlengths 395
24.5k
| Source
stringclasses 6
values |
|---|---|
Let's open our order-list.component component within the features directory. And then we will create a variable that will receive the list of orders that we will display on the screen and we will import our service using injection from the angle to use our order service methods. The angular component has some methods that we can use through the implements such as ngOnInit which is invoked as soon as the component is started. In addition to onInit we have others such as ngOnChanges, ngAfterContentInit etc. For more details on angular lifecycle access the link. We will create a method that will call our service and then invoke this method whenever our component starts, for that we will use the ngOnInit method, important to remember to import our order class and our service to avoid compilation errors as shown below.
Now let's create our list in the part corresponding to the html in the order-list.component file. We will create a simple structure, just to exemplify but if you want to apply CSS or use datatables or bootstrap feel free. Inside our html we will use the angle markings to check if the list is empty and to render our list as shown below.
Now let's comment on the api's reading structure because in our example we are not accessing a real api. we created the structure just to exemplify how a REST api works. we will use localstore to store our list. We will make the change both in our service and in our component as below.
There, our list component is ready. Now to finish let's put the action in the add button to direct to the page of the application creation form.
By clicking on the add button our route will be redirected and we will create our form in html. It will contain the information of our class order ngModel in the inputs that received the values. The values inserted in the inputs will be automatically updated directly in our object. This feature uses the angular feature called formsModule that we will insert in our registration module, this magic happens through the two-way binding concept. More details on the official link. We also need to create on our save button. Save action that will be reflected in our component. we use the notation with () to tell the angular that that html component attribute is an action reflected in the component. Inside it we put our method that we will create it in our component.
Now we will import the FormsModule resource into the order-create module
And to finish our registration now we will implement our component order-create.component. We will create the variable order that is used in our form using the two-way databing feature and we will apply in our builder the feature of the angular router which by default it already uses the injectable feature. Let's create our save method by retrieving the list from our localStorage by adding the object to the list and playing it again in our local storage. We will display a success message and use the router feature to redirect to our list
And finally, let's do our editing form and implement the delete. To do this we will first implement the actions on our list. We started by changing the html and inserting the action column and we will also take the index from our list to use as a reference in the delete edition. In editing we use the routerLink feature again with a small difference, this time we will pass a parameter. We also need to change the routing there in our editing module to receive the parameter.
Now we are going to implement the delete in our list component following the simple logic to delete the array based on the position we passed and updating our localStorage. Always remembering that in the real case of an api we would use the routes that we created in our service.
and to close we are going to implement our editing component. We redo the same process of implementing the creative component with only one difference. We take the id passed through the ActivatedRoute feature of the angle in which we need to insert it in the costructor because the pattern already has the injectable of the angle also implemented. Through our localStorage we identify the object and fill it with the corresponding values, ready the magic happens! We will implement the editing method by changing the exact position of our array and replace our localStorage.
As you can see our html does not have any type of change in particular except the name of the editing method.
And here we end the complete angled crud. Check out the project we created along this route on github here.
|
OPCFW_CODE
|
Matt Warner 3/16/2007 1:29:53 PM
When there is a relationship between two tables and the detailed table is shown as a view under the master record, it would be nice to add an "export" button for the detailed view (as if it were a query view). For example, in our database we have an "events" table (master) and an "event attendees" table (detail). I have chosen to not show the event attendees table as a tab in the database but I would need to to be able to query the contacts that attend a particular event. These contacts are already shown in the detail view of the master record ... but I can't export them from there.
Philipp Matuschka (MMB) 6/7/2010 6:07:54 AM
When I open a new tab, the default view for that table appears. However it only has the NEW button. It does not have the PRINT, EXPORT and SEND buttons. If on the therhand I later select that view, then those buttons are included. Can these buttons be included when the tab is first selected?
Shem Sargent 7/24/2012 10:13:49 AM
It appears that this is because of the initial screen for each tab is set by default as a Dashboard instead of a View. The Dashboards do not have the print, export, and send buttons, only the Views do. Interestingly, if you use the browser print feature on the Dashboard, it creates the same output as the clicking the Print button on a View.
Rick Cogley 5/13/2013 8:07:53 AM
We'd also like to be able to have these buttons on the dashboard-linked views. Or at least, a link to go to the view, so that you can access these. At this point, we have to do an ugly kludge of putting the view links in text sections on the dash.
Marc Andre Lacas 3/18/2015 12:40:43 PM
In an application I made, I use the form in preview mode to display linked items from one table another (in this case: all the "Tasks" related to a "Project"). I would like to export the resulting view to CSV (or Excel) like we can do in a regular "Table" view. See jpg
basenine 3/18/2015 5:03:14 PM
I stumbled across this anomaly just yesterday. I thought that because I'd CHECKED "Hide Print/Export/Send" it wasn't showing on the related view...but after unchecking, it still wouldn't appear.
This would be a handy feature.
Gii Systems 3/19/2015 9:00:38 AM
Also consider this for views in dashboards please.
Sometimes the view on the dashboard is from another table (That might be hidden). Making it impossible to export the data.
rishi uttam 4/14/2015 12:32:23 AM
How do you others users print report views from a details view? It is not possible is it?
Marc Andre Lacas 4/28/2015 6:43:47 AM
Hello Rishi, I submitted a similar idea to yours, with the addition of CSV export. I think our ideas could be merged. In my case, I created a document template that can be printed in the Master View, in which you only find the child table I want to print out. If you click on "generate document template" under Documents (when in "Setup") you will get a docx file in which you see your child tables. You just need to edit that file, keep the table you need and save the file under a name that make sense for your needs, then you should be able to print out only the "child table/view". Keep in mind that this document will not update itself if you change things in the "child table/view". I hope this was useful. If I was not clear you can look it up here: http://www.teamdesk.net/help/9.5.aspx
Baret Kouyoumdjian 11/23/2015 8:14:41 AM
I think this is a concept that does not need too much thought. Having a CMS means being able to add/edit/view/report/delete content. If you can have a view/report then you must be able to export it as well.
Rick Cogley 11/23/2015 3:40:05 PM
the only way I found to do this is to use a Document for it. You can export a Document in Word or PDF format. I always have trouble with Word though, and can never get the style formats to "behave".
basenine 11/30/2015 6:36:31 AM
@Rick....you can literally spend hours formatting styles in word! Hours I says!!
The only solution I can muster for this idea is to create a view with ASK THE USER option, where ask the user is the "parent record id" (using the SEARCH field above might speed up the selection criterion).... The resulting records will then have the options to export or print as you desire
basenine 11/30/2015 6:38:36 AM
When I mentioned "create a view"... I was meaning in the child record table....just to be clear
Rick Cogley 11/30/2015 4:59:40 PM
(a total aside @basenine: I hate Word. It has had the same odd bugs for years.)
Bin Xia Zhang 3/15/2017 3:49:54 PM
Would be useful. I have a database where table A is a list of students and table B a list of classes.
B refers to A in one-to-many relationship.
When I view a record in B (particular class) I can see the list of selected students, but its not possible to export/send the data. I created a view in table A allowing the user to select particular classes from B to export the data. It would be more useful to do it directly from table B and not to do an additional search..
basenine 5/25/2017 10:33:30 PM
Bump...just discussing this with a customer right now...
basenine 5/31/2017 5:31:39 PM
another bump...'passionate' customer needs this now...the work around is laborious to make and inconvenient to use.
Pierre 9/8/2017 11:30:32 AM
I would like to be allowed to excel-export any table view or sub-table view.
Pierre 9/8/2017 11:30:48 AM
I would like to be allowed to excel-export any table view or sub-table view
basenine 11/16/2017 9:02:05 PM
Would like to see the EXPORT button open up on the related Child views...😉
My solution above is often cumbersome...esp when more than one Child view is to be exported
basenine 11/21/2017 4:50:55 AM
Just been playing around with the column type: Formula URL.
You can grab the URL that the EXPORT function strings together (from a view) and cobble an easy formula to replicate as a Button and perform the export.
You can put that button pretty much wherever you want on a form record...
For example the URL looks like this (xxxxxxxxxxxxx swap out the real id's....)
URLRoot()&'/exportview.aspx/BuilderTrends_List%20All_'&URLEncode([Quote/Order Id])&'.csv?format=0&id=xxxxxxxxxx&fid_xxxxxxxxxxx='&URLEncode([Quote/Order Id])
All I've really added in here are the URLEncodes. The first one creates the file name and the second is for the filter as set by the view.
.csv or .xlsx or .html will determine the export file format
Slava Shinderov 4/16/2019 7:20:37 AM
@We've added PRINT/EXPORT/SEND buttons to the details view.
Slava Shinderov 4/16/2019 7:23:06 AM
87 - allow export of relationship detail views
Baret Admin 4/16/2019 7:25:26 AM
Thank you Slava
Pierre 4/16/2019 3:11:11 PM
It works beautifully
Thanks to the team
Kirill Bondar 5/14/2020 6:00:00 PM
850 - Ability to export to CSV a view
Kirill Bondar 5/14/2020 6:02:54 PM
339 - Print button on default view
|
OPCFW_CODE
|
Capture screenshot mac os x lion
Here is the the upper left portion of our screen:.
- can i have 2 docks on my mac?
- controllo remoto mac da pc.
- customer list software for mac;
- lac noi mac trong co tu cung co thai duoc khong;
OS X Founder of Switching to Mac and managing editor. He began blogging in and quit his job in to blog full-time.
Capture a portion of the whole screen on Mac Mountain Lion
He has over 15 years of industry experience in IT and holds several technical certifications. Read Aseem's Full Bio.
Changing the Default Screenshot Format First, a note about screenshot image formats. At the prompt, type or Copy and Paste from here the following command and press Return : defaults write com.
objective c - Slow screen capture code on OSX. Why? - Stack Overflow
To do this enter this command at the Terminal prompt and press Return : killall SystemUIServer The default screenshot format in OS X will now be changed to the image type you entered before. Here is our entire screen: 2. Here is the the upper left portion of our screen: 3. You can release Shift , move another edge and hold Shift again to accurately define your desired area.
Option : Holding this key down will resize the area selection based on its central point rather than from where you initially clicked the Left Mouse Button.
You Can Capture & Save Screen Shots with Preview in Mac OS X
In Rumania. I like its face.
And yes, it was a screenshot. These are the shortcuts to learn in order to get better screenshots on a Mac -- don't forget you can cancel the shot by pressing Escape before you click:. The cursor will swap between the crosshair and a camera icon when you do.
In camera mode you can just click on the window you want an image of to highlight and capture it. You might want to include your cursor within the screenshot. To achieve this, launch the Grab app hidden in Utilities and select Grab Preferences.
Select one of these and the pointer will appear where it is located when the picture is taken. That is why you can see the pointer in that shot.
There's no easy way to change image formats when taking a screenshot. Open the images saved to the desktop in Preview and export them in another format, or remember to use the Space shortcuts listed above to save them to Clipboard in the first place. You can then open them from the Clipboard in Preview to save them in your preferred format. You can also use Preview to capture screenshots -- tools for this are hidden in the apps File menu Take Screenshot.
Subscribe to RSS
It is remarkable that Preview does not use the same on-screen system prompts as used by Grab or OS X. There's an old but still good rundown of apps you can use to improve your screengrabs or as an alternative to Apple's provided tools here. Got a story? Jonny is a freelance writer who has been writing mainly about Apple and technology since
|
OPCFW_CODE
|
Shannon's Expansion is a method by which a boolean function can be represented by the sum of two sub-functions of the original. Claude Shannon
was the creator of Shannon's expansion, and is credited for numerous other achievements in the field of Boolean algebra
. Rather than explain it pedantically
, here is an example:
Take this as our function:
f = xyz+ xy'z + x'y'z + x'yz + x'y'z'
Now, notice that you can re-write the function in terms of any two variables - namely, a variable and its compliment.
f = x'gx' + xgx
Now, to make the final expression equivalent to our composition of functions, simply apply the distribuive theorem to the function about x:
f = x'(y'z + yz + y'z') + x(yz + y'z)
And there we have it! We have just expanded the function f about the variable x. Now, you might be thinking to yourself "What if there were no x variable in one or more of the terms? Very good question, and one whose solution is not always intuitive.
Remember, that in Boolean algebra, you can AND any literal or term with 1, and still achieve the same truth value. With that in mind, let's look at this function:
f = yz + xyz' + x'y'z
If we desire to expand around the variable x, we simply don't have enough information in the first term to accomplish this task. So, what do we do? Give up? No! Remember what was said above: AND the literal with 1, or, in this case (x' + x).
f = yz(x' + x) + xyz' + x'y'z
f = x'yz + xyz + xyz' + x'y'z
This function contains the variable about which we want to expand the expression, so now we should have no problem performing the expansion:
f = x'gx' + xgx
f = x'(yz + y'z) = x(yz + yz')
Our expansion is complete!
Of course, you can perform Shannon's Expansion about any variable you desire, so long as you can provide for that variable in the expression without changing the truth value of the expression. Also, you can perform multiple expansions of a single function (e.g. about x, then about y) or, you can even perform the expansion about many variables at once (e.g. about xy). The result is a functionally equivalent expression for the variables involved.
You might be asking "Why would I want to make an expression larger? I thought boolean algebra was all about minimizing functions?". Consider a logic device called a multiplexer. Multiplexers take n select inputs, and 2n data inputs. Once you expand any boolean function about any number of variables, you can use the variables that the function was expanded about as the select inputs, and their respective composed functions as the corresponding data inputs.
There, now go forth with this information and build better communication devices, computers, and toaster ovens. Yes, now!
|
OPCFW_CODE
|
// lib/index.js
// Functions in this module are asynchronous, returning a promise if no callback is supplied.
'use strict';
var assert = require('assert');
var graphviz = require('graphviz');
var Q = require('q');
// Construct a Graphviz graph from a Gremlin graph.
// *opts* is an optional object allowing overriding default behavior:
// * *graphName*: Name of the Graphviz graph [default: 'G']
// * *vertexId*: Function that returns vertex ID string from a Gremlin vertex [default: util.getVertexId]
var factory = module.exports = Q.promised(function (gremlinGraph, opts) {
var graphvizGraph = new GraphvizGraph(gremlinGraph, opts);
var gremlin = gremlinGraph.gremlin;
var vertexTraversal = gremlinGraph.V();
var edgeTraversal = gremlinGraph.E();
return iterate(vertexTraversal, function (vertex) { addVertex(graphvizGraph, vertex); }).
then(function () {
return iterate(edgeTraversal, function (edge) { addEdge(graphvizGraph, edge); });
}).
then(function () {
return graphvizGraph;
});
});
var GraphvizGraph = function (gremlinGraph, opts) {
this.opts = opts || {};
var graphName = this.opts.graphName || 'G';
this.impl = graphviz.digraph(graphName);
};
// Haul a property getter from the implementation class.
var haulProperty = function (name) {
Object.defineProperty(GraphvizGraph.prototype, name,
{ get: function () {
return this.impl[name];
}});
};
// ## Immutable properties
// Expose certain immutable properties of the graphviz graph.
['id',
'type'
].map(haulProperty);
// Haul a property getter/setter from the implementation class.
var haulMutableProperty = function (name) {
Object.defineProperty(GraphvizGraph.prototype, name,
{
get: function () {
return this.impl[name];
},
set: function (value) {
return this.impl[name] = value;
}
});
};
// ## Mutable properties
// Expose certain mutable properties of the graphviz graph.
['use'
].map(haulMutableProperty);
// Haul a method from the implementation class.
var haulMethod = function (name) {
GraphvizGraph.prototype[name] = function () {
// var args = Array.prototype.slice.call(arguments);
return this.impl[name].apply(this.impl, arguments);
};
};
// ## Methods
// All of the useful methods are exported.
[
'addCluster',
'addEdge',
'addNode',
'clusterCount',
'edgeCount',
'get',
'getCluster',
'getEdgeAttribut',
'getNode',
'getNodeAttribut',
'nodeCount',
'output',
'render',
'set',
'setEdgeAttribut',
'setGraphVizPath',
'setNodeAttribut',
'to_dot'
].map(haulMethod);
// Iterate through a Gremlin traversal
var iterate = function (traversal, process) {
return traversal.forEach(process);
};
// Add a vertex (synchronously) to a Graphviz graph based on a Gremlin vertex.
var addVertex = function (graphvizGraph, gremlinVertex) {
assert(graphvizGraph);
assert(graphvizGraph.impl);
assert(gremlinVertex);
var vertexId = graphvizGraph.opts.vertexId || util.getVertexId;
graphvizGraph.impl.addNode(vertexId(gremlinVertex));
};
// Add an edge (synchronously) to a Graphviz graph based on a Gremlin edge.
var addEdge = function (graphvizGraph, gremlinEdge) {
assert(graphvizGraph);
assert(graphvizGraph.impl);
assert(gremlinEdge);
// Pull the Direction enumeration from the embedded Gremlin object.
var gremlin = gremlinEdge.gremlin;
assert(gremlin);
var Direction = gremlin.Direction;
assert(Direction);
// TODO: Switch this from a Java edge object to something from gremlin-v3.
var vertexId = graphvizGraph.opts.vertexId || util.getVertexId;
var iterators = gremlinEdge.unwrap().iteratorsSync();
var inV = iterators.vertexIteratorSync(Direction.IN).nextSync();
assert(inV);
var inId = vertexId(gremlin.wrapVertex(inV));
var outV = iterators.vertexIteratorSync(Direction.OUT).nextSync();
assert(outV);
var outId = vertexId(gremlin.wrapVertex(outV));
var graphvizEdge = graphvizGraph.impl.addEdge(outId.toString(), inId.toString());
// See if we have to label the edge.
var edgeLabeler = graphvizGraph.opts.edgeLabel;
if (edgeLabeler) {
var edgeLabel = edgeLabeler(gremlinEdge);
if (edgeLabel) {
graphvizEdge.set('label', edgeLabel);
}
};
};
// Import the utilities module.
var util = module.exports.util = require('./util.js');
|
STACK_EDU
|
SyntaxError: Unexpected token } in Vue js
My code is like this :
<multiple-photo-product :product="{{ isset($product) ? $product : '' }}"></multiple-photo-product>
When the code runs it throws an error:
SyntaxError: Unexpected token } in
But if the code is like this:
<multiple-photo-product product="{{ isset($product) ? $product : '' }}"></multiple-photo-product>
It doesn't throw an error.
I add :, so that the data is sent as an object.
If it does not use :, the data is sent as a string.
How can I solve it?
Is isset($product) client side code or server side code?
@Bert Evans, server side code
Is the error server side or client side? If it's client side, what is actually rendered as a result of your ternary?
I do not think you are allowed to use interpolation inside of bindings. Try :product="isset($product) ? $product : '' "
What does the rendered HTML look like? Use your browser's "View source" option
I add {{..}}, because I use laravel framework
@Bert Evans, The error is client side. It's an object
Can you add the rendered HTML including the rendered object to your question please?
@TrendingNews You need to provide more information about your problem.... For instance, component exists inside a (Laravel?) view, or inside another vue component?
@ira, it exist inside a view laravel
The problem lies in the fact that if the php variable $product is not set (i.e equals to null or ""), then Vue will try to bind the prop :product with '' which ultimately results to an error (like trying to make a :product="" bind)
Try the following:
<multiple-photo-product :product="{{ isset($product) ? $product : '""' }}"></multiple-photo-product>
Notice the double quotes "" surrounded by the single quotes. This will say to Vue to bind the product prop with an empty string, in case php's $product variable is not set.
Please also have a look here. You may find it helpful. The key point to recall is that v-bind expects valid javascript expressions, that is the interpolated value (i.e. whatever is inside the Blade's curly braces {{}}) must be a valid javascript expression too
In VueJS2, the attribute properties with : don't need the {{ }}. In your case, the ternary is like that :
<multiple-photo-product product="isset($product) ? $product : ''"></multiple-photo-product>
Source : https://v2.vuejs.org/v2/guide/syntax.html#Attributes
Mustaches cannot be used inside HTML attributes, instead use a v-bind directive
The {{}} is a server side interpolation. He is using laravel.
|
STACK_EXCHANGE
|
No, wait! Don't stop reading just yet! I know that spreadsheets sound boring, but they really aren't. Used properly, they can be essential tools in your planning toolbox, because they have three brilliant features.
Brilliant feature #1: The sort facility
Using this, you can re-order the spreadsheet by deadline, to see what's coming up, or by person, to see who is meant to be doing what, or by area of work, to see if everything is being covered. Using the sort feature is easy, as long as you have designed the spreadsheet sensibly. That means, having a separate cell for each attribute of each task, ie date for completion, area, person responsible, and so on.
One tip: format the dates as yyyy-mm-dd (or, in USA, yyyy-dd-mm). Why? Because that's the only way you can make sure everything is listed in chronological order, if that's what you need.
Brilliant feature #2: Sumif
This is a great feature that's available in Excel, Google Spreadsheet and in OpenOffice's Calc. What it lets you do very easily is to perform the following kind of calculation:
If this item comes into category A, add it to the total, otherwise don't.
You can use Sumif to find out what you're spending money on, or where your team's income is coming from. For example, you may have categories like software, hardware, printing, and so on. Using Sumif, I once determined that 60% of my department's spending was going on photocopying worksheets. I asked my team to print off multiple copies instead (if they needed print-outs at all), which resulted in savings of hundreds of pounds over the year.
Brilliant feature #3: Conditional formatting
Use this to create the traffic light system: green for 'yes, done that', amber for 'we're getting there', and red for 'there's been a glitch'. The traffic light system gives you an instant visual summary of how you're doing as far as meeting targets is concerned.
Conditional formatting can work on either numerical values or text, or a formula. For example, you could have a column called Progress, and set up the conditional formatting to turn a cell red if it contains the word 'no', green if it contains the word ';yes' or amber if it contains the word 'partly'. Or you could set it up based on a formula 'today's date minus target date'. If the answer is less than zero, the cell goes red, and so on.
The spreadsheet was one of the first applications developed for the personal computer, and it's more than just a glorified calculator. Pretty it ain't, but boy is it useful!
|
OPCFW_CODE
|
~/.bashrc does not recognize export
This is probably the weirdest thing you've ever seen. But it's true. I'm on macOS, and want to install devKitARM. After installing Git, the guide told me to show the program where the binaries are. It told me to execute:
echo "export DEVKITPRO=/opt/devkitpro" >> ~/.bashrc
echo "export DEVKITARM=/opt/devkitpro/devkitARM" >> ~/.bashrc
source ~/.bashrc
However, when I do this, it gives me
export: Command not found.
export: Command not found.
export: Command not found.
I looked in the ~/.bashrc and realized that there was already an export PATH command in there, probably for some system function. However, this was not being recognized the entire time. Can someone help?
Are you sure that you’re running bash?
I should be... I'm pretty sure it's pre-installed on macOS, and other bash commands work. Also, oddly, the man page for it exists, but the command itself not.
(1) “it's pre-installed” — proves nothing. (2) “other bash commands work” — what do you mean? Are you talking about thing like cat, cp and ls? These are not bash commands. What happens if you just type export at your command prompt? (3) “the man page for it exists” — the man page for what? (4) “the command itself not” — what do you mean? Please do not respond in comments; [edit] your question to make it clearer and more complete.
Your currently running shell is the (TENEX) C shell, which does not have an export command. You cannot source a Bourne Again or POSIX shell script in the C shells. The C shell syntax is different to the Bourne/POSIX shell syntax.
Adding these environment variables to your interactive shell involves the C shell's setenv command. Having this done automatically for login shells involves the C shell's ~/.login script.
The manual installation instructions actually simply say "Add these variables in your execution environment". The Bourne Again shell is just a "for instance". Obviously, if you aren't using the Bourne Again shell, which you clearly are not, then you do what is appropriate for your actual choice of shell.
See also this SO question about how to change login shell on macOS: https://stackoverflow.com/questions/453236/how-to-set-my-default-shell-on-mac
Thanks! I figured out I had to run the terminal as bash, then source the /.bashrc
|
STACK_EXCHANGE
|
There are several components in network connections that can go unnoticed, but you need to be aware of them to prevent configuration errors from occurring. One such component is the netmask.
Its main function is to notify the router that messages have different addresses, enabling the network and host identifier to be recognized. In this article, we will simply explain what a netmask is, what it does, how subnets are calculated, and much more. Read on to better understand this subject!
What is the relationship between IP address, network address and host address?
Before we even talk about netmask, it is critical that you understand the differences between IP address, network address, and host address, as these three concepts are necessary for us to cover about netmask.
The IP address is one of the main internet protocols, being responsible for delivering data packets over the network and directing them according to addresses.
All computers with internet access have an IP address to be identified. They can be found in the computer's connection properties, have decimal format and are interpreted by the equipment as binary, and are also separated by four dots. Example of an IP address: 18.104.22.168.
Each part is called an octet, since in binary form it has eight numbers. All in all, we have 32 positions and that's why IP addresses are 32 bits long.
That said, the IPv4 address, which is the IP version 4 address, has two components: the network address and the host address. The first is in charge of identifying the entire network, so that all hosts on the network have the same bit sequence.
Meanwhile, the host address identifies the connection to a specific host, so each host has a unique address within the network. This means that every end device needs an address that is its own to communicate on the network.
What is the netmask?
The netmask is 32 bits, just like the IP address, and is intended to mask a part of the IP address. Thus, every IP address has a corresponding mask, serving to identify which part of the address is from the network and which is from the host.
There is also an addressing class for each netmask. See below:
- Addressing class A: 255.0.0.0
- Address class B: 255.255.0.0;
- Addressing class C: 255.255.255.0.
As an example, let's assume that the IP address 22.214.171.124 has a netmask that is 255.255.255.0 (class C). In this case, the part corresponding to the network is 191.171.0 and the part referring to the host is 50. That is:
- Network: 191.171.0;
- Host: 50.
So, in order to divide the network, it is possible to divide the mask, using values from 0 to 255. Thus, the subnet mask is created, which manages to notify the router that messages identified with sender and destination present a different address.
In this way, it is able to recognize both the network part and the host part, enabling, for example, the router to understand when packets must stay in the network and when they must be transmitted to another.
It is important to remember that subnetting is a process that aims to improve performance, security and organization by dividing a large IP address into several smaller networks.
Let's imagine the following situation: in a single building, the company has an HR, Finance and Sales area, in which each department has 30 hosts.
The network of all sectors started to slow down due to dense files being sent by one sector. After all, as a single network is used for all departments, if something happens to one, all the others will be affected.
In that case, we could think of assigning addresses to each sector, right? However, this practice would result in a waste of IP addresses.
Therefore, the ideal would be to divide the network into smaller parts, in which each one would be linked to a specific sector. So we're talking about creating subnets. Did you understand the importance of these components in network connections?
How to calculate subnet masks?
Remember our previous IP address and netmask example? Every three digits in the mask, we have a part of the host address, remembering that each part has eight bits, forming an octet.
Also, each bit has a value in decimal format to match its position. Another detail that you need to keep in mind is that, as it is a bit, there is only 0, which represents the host, and 1, which represents the network.
That said, in the example we gave of the class C subnet mask, which is 255.255.255.0, we have the binary representation 11111111.11111111.11111111.00000000, where the first three octets refer to the network and the last octet refers to the host:
- Mask: 255.255.255.0;
- Binary: 11111111.11111111.11111111.00000000.
In this way, all eight bits that have a value of 1 are represented by 255 in decimal form, while bits of value 0 are also 0 in decimal form.
In short, every subnet is a different network that needs to communicate with each other from a router.
In this topic, we use a class C subnet mask as an example, but the configuration works for all other addressing classes.
Also, another important piece of information is that instead of talking about the netmask in decimal form, we can also talk about just the amount of bits that exist in it.
Although it is a somewhat complex subject for non-experts, understanding what netmasks are and how they work is necessary so that you can avoid configuration errors. That's why anyone working with IT shouldn't ignore it.
In addition to knowing the netmask, why not delve into the gateway? Learn more about this tool in our article!
|
OPCFW_CODE
|
2FA vs Passkeys: Strengthening Login Security Beyond Passwords
As users demand both security and convenience, modern authentication methods strive to balance these priorities. Two popular methods to improve login security beyond passwords alone are passkeys and two-factor authentication (2FA).
2FA has rapidly gained adoption in recent years by adding a second verification factor to the traditional username and password combination. This helps validate user identities and reduce unauthorised access. However, 2FA still relies on remembering and entering passwords.
Passkeys present an alternative credential-based model that removes passwords entirely from the sign-in process. With passkeys, each user's identity is cryptographically tied to their devices and accounts, allowing login through biometrics or device attestation alone.
While both 2FA and passkeys fortify user verification compared to passwords, they differ in their technical implementations and user experiences. The main differences between these two multifactor authentication strategies will be discussed in this article, along with an analysis of each's functionality, appropriate applications, and adoption opportunities and difficulties.
What are Passkeys?
Passkeys are unique cryptographic credentials that are used during the sign-in process instead of passwords. With passkeys, a user's identity is bound to their individual devices and online accounts through cryptographic attestation rather than a memorised secret.
When setting up a passkey, a public-private key pair is generated; the public key is stored on the server to verify logins, while the private key remains on the enrolled device. During authentication, the device proves its identity by signing a challenge with the private key.
This allows users to sign in with a simple verification, like a fingerprint or face scan, rather than a password. Passkeys avoid the vulnerabilities of weak, reused, or stolen passwords. If a device is lost or stolen, passkeys cannot be extracted since they never leave the device.
Passkeys provide cryptographic identity, which removes the entire need for password management. Authentication becomes a seamless, one-step process tied directly to the user's registered devices and identity, bypassing passwords for a simpler and more secure sign-in experience.
How they work:
With 2FA, the sign-in process involves separate verification steps. First, the user enters their username and password. Then, they must confirm their identity through a second factor - whether a one-time code sent via SMS, authentication app, or security key.
The two-step process adds friction but strengthens protection by requiring multiple pieces of information to log in. Even if one factor is compromised, the other still prevents unauthorised access.
Passkeys streamline the login process into a single step for enrolled devices. When the user attempts to sign in on an authenticated device, their identity is verified directly through a biometric or device credential rather than a discrete password.
A fingerprint, face scan, PIN, or other prompt replaces the need to manually type or retrieve passwords. The on-device biometrics or credentials function as the passkey tying the user's cryptographic identity to that particular device.
Authentication is effortless as long as the user has possession of the device. Unlike a traditional password, the passkey permits credential-based sign-in without ever disclosing the passkey itself.
2FA is applied and enforced independently for each account. A separate two-factor verification is required for every system that supports 2FA—whether email, banking, social media, or other online services.
While strengthening security per login, 2FA does not centrally manage a user's identity or synchronisation of credentials. It only protects individual accounts in isolation.
With passkeys, a user's authenticated identity is linked to all of their enrolled devices and accounts within a single, integrated system, which is a different strategy. With passkeys, only one biometric verification step is needed to access any associated account from any enrolled device.
This centralised approach to cryptographically tying user identity together enables consistent, single sign-on-like access to various platforms and applications. By integrating identity management, passkeys maintain robust access controls while streamlining the sign-in process.
2FA can generally be added as an extra layer of security to existing password-based authentication systems with modest technical requirements. Service providers need only implement standards like TOTP, OATH, or FIDO to enable common 2FA methods.
This retroactive compatibility allows 2FA to be rolled out incrementally as an added protection for sensitive or high-risk accounts without overhauling existing authentication architectures. Users can also elect to protect some but not all accounts with 2FA.
Passkeys are all-or-nothing in nature, necessitating coordination and compatibility between user devices, operating systems, browsers, and apps. The implementation of passkeys across platforms is contingent upon the establishment of uniform sign-in and key management protocols, along with enrollment workflows.
While future-proofing long-term security, passkeys impose more demands on technology integration and user training upfront. A phased transition strategy is necessary to smoothly transition from passwords to passkeys authenticated experiences.
Adoption Challenges of Passkey
For passkeys to achieve their full potential, overcoming interoperability challenges is crucial. Widespread compatibility is needed between operating systems, browsers, applications, and websites to allow passkey-based credentials to securely roam across devices and services.
Establishing common technical standards and ensuring consistent support across the ecosystem will be an ongoing process. A lack of universal protocols could initially limit passkey usage to discrete platforms or single sign-on ecosystems.
Additionally, users must understand how passkeys differ fundamentally from passwords they have used for decades. Educating people about the security benefits and teaching new authentication habits like biometric logins takes significant effort.
Passkeys that deliver too radical a change could face adoption hurdles versus incremental improvements like 2FA. Effective onboarding and assistance transitioning from passwords will smooth user acceptance of passkeys.
Overcoming interoperability obstacles and changing user behaviour patterns are key challenges that could slow near-term passkey adoption, even if the technology has long-term security advantages over common authentication methods.
Both two-factor authentication and passkeys have evolved to strengthen online identity validation beyond basic usernames and passwords. Each offers important security improvements, depending on individual technical capabilities and user requirements.
For organisations seeking interim upgrades, 2FA is readily implementable with mature standards and retrofit options. Its additional verification layers bolster many existing systems without wholesale replacement demands.
Passkeys promise a future devoid of passwords altogether through device-bound cryptographic credentials. This more seamless sign-in model protects users from compromised credentials if properly supported across the ecosystem.
As technologies advance and interoperability becomes ubiquitous, passkeys could eventually surpass most 2FA methods by integrating identity rather than isolating each account. Their cryptographic foundation establishes long-term security suitable for high-value assets.
|
OPCFW_CODE
|
Over the last few days much has been made in the technology press about "disappointing" sales figures for Windows Phone 7. What exactly is this based on considering there are no official sales figures and is it true or not?
Back on the US launch (no figures for Europe yet, where Windows Mobile is traditionally much stronger), an anonymous source was quoted by dozens of technology websites as saying Windows Phone 7 sold 40,000 units on day one. To date nobody knows who this source is.
Is 40,000 units a bad figure for a brand new platform with low awareness? Not really. Although we don't have any figures to directly compare with the first Android phone, the G1, T-Mobile did announce they had sold a million in the first 6 months. On average that's a little over 5500 units sold per day.
Of course for Google the technology press spin everything in a more positive light, the stories when the G1 came out weren't about disappointing sales figures. No they just made figures up saying it had 1.5 million pre-orders. Obviously out by several orders of magnitude, at least.
Microsoft isn't given the same luxury. A few days ago figures of Windows Phone being outsold 15 to 1 by Android in the UK swept around; this has been covered on pretty much every technology news website. Where are these figures from?
Katherine Noyes, a known Linux fanboy, someone who said Android has almost made the iPhone a "niche also-ran", writing for PC World say they're from a UK retailer. Well I'm from the UK maybe I've heard of them, someone like Expansys or Amazon or maybe o2 or Orange, or maybe even the Carphone Warehouse? Nope from someone called Mobiles Please. Err never heard of them, in all my years buying smartphones I have never heard of these people.
Worse yet they're not even a mobile retailer, they're just a website that links to deals from other retailers. I'd call them a comparison website. Except Alexa has never heard of them, neither has Compete. When my blog gets more traffic than a so-called "UK retailer" you should be worried, especially worried when tech websites with millions of readers use them as a basis for a story about how Windows Phone 7 is a failure.
Of course I have to congratulate Mobiles Please for all the free advertising they got, from nowhere to being mentioned on dozens of the biggest technology websites. All they needed was a crappy anti-Microsoft press release.
Windows Phone 7 is undoubtedly doing better than the handset makers predicted, Orange are still having to dish out £20 vouchers to make up for the fact its taking them a while to get phones into people's hands. It's also certainly doing much better than Android was in its first couple of months, just looking at the daily active Facebook users coming from Windows Phone can tell you that, and that's a very conservative estimate for total sales. Without a doubt Windows Phone 7 will get to the million mark faster than Android.
|
OPCFW_CODE
|
Removes those extra-tempting but extra-problematic buttons from tinyMCE, like font color, alignment, and other things that will wreck your style sheet.
In addition, it removes “Heading 1” from the role dropdown, as there should never be more than one H1 on a page — and the page or post title is already it.
Why Does Formatting Matter?
When people first get their hands on that shiny WordPress text editor, they often get swept up in the excitement of styling their text. “This is important, so I’ll make it red! And this part will be centered, and this part will use giant letters!”
But as fun as this is, there’s a way to use the WordPress text composer like a pro, and the benefits of doing it in that way are big:
- Better potential rankings on search engines (Search Engine Optimization)
- A consistent look across your whole website and other materials (research shows that visual consistency leads to trust)
- Better flexibility across different devices, from desktops to mobile phones
- Better accessibility for readers with a wide range of disabilities.
- Better maintainability when it comes to future site updates — that means savings!
- What’s the secret sauce that provides all these benefits?
Mark your text with what it is, not what it should look like.
What does that mean?
It means that your website already has a style guide built into it, so that it knows that a subtitle looks a certain way — perhaps it is larger, bold, or a different color than the main text. So when you type in a subtitle, you only need to let WordPress know that it’s a subtitle, and the rest will be taken care of. Do not give in to the temptation to mark the subtitle text as bold, larger, centered or a different color from the WordPress editor.
This plugin removes the most damaging buttons from the WordPress editor, so you (or your clients) won’t be tempted to use them.
For more information on how to use the WordPress editor like a pro, with or without this plugin, and why formatting matters to brand consistency, SEO, accessibility and even maintainability, check out this blog post: http://www.pixelovely.com/format-web-content-wordpress
Show your appreciation for this plugin
If I’ve helped you, please consider buying me a slice of pizza to fuel my coding .
- Download the zip file, and unzip it.
- Upload the entire folder (remove-problematic-formatting-options-from-tinymce) to the
- Activate the plugin through the ‘Plugins’ menu in WordPress
- You are done!
Contributors & Developers
“Remove Problematic Formatting Options From tinyMCE” is open source software. The following people have contributed to this plugin.Contributors
Interested in development?
- Changed “tested up to” to WP 6.1.1
- Changed “tested up to” to WP 5.6
- Changed “tested up to” to WP 5.4.2
- Added better plugin description.
- Added installation instructions.
- Changed “tested up to” to WP 4.2
- Removed underline button, since it creates confusion with clickable links, and moreover inserts inline styles which are a huge no-no.
- Changed “tested up to” to WP 4.0
- Plugin first released! I would very much appreciate it if you remembered to rate the plugin after trying it.
|
OPCFW_CODE
|
For this article, RealToughCandy super-subscriber Michael Kornblum takes the reins. Michael’s comments on my YouTube channel are golden, and his piece here on 11 Atom packages further illustrates his writing talents. Plus, these packages make our development lives easier. New to Atom? Start here. Enjoy! -RTC
In the world of web development, we love to see new things.
When a new web browser is introduced, we marvel at its features and rejoice when certain CSS features are supported. When a command-line tool is introduced, we welcome it with open arms in the hopes that it will improve our professional lives. Text editors, however, do not enjoy the same level of fanfare.
This is because up until recently, text editors were boring. Early text editors were little more than stripped-down word processors. Small features like syntax highlighting and tabbed interfaces were eventually added to them as features, but at the end of the day, they were really nothing to write home about.
Recently, however, text editors have started to get their moment in the sun.
Newer text editors like Visual Studio Code and Atom use a robust plugin architecture that allows developers to customize their work environment. Through the use of extension or packages, it’s possible to have a fairly minimalist setup that resembles the text editors of days past or have a fairly complex setup that includes embedded terminals, linting, and git integration.
In this article, I will focus on some of the packages I use in Atom, my text editor of choice. Many of the package that I’ll mention have counterparts in other editors like Visual Studio Code, Brackets and Sublime Text. If you are using these tools, I encourage you to look into these alternatives. They will make your life as a developer a lot easier.
The Atom text editor + delicious bagels. We’re already off to a great start!
On the surface, the benefit of having file icons in a text editor may seem nominal at best. However as projects grow larger, being able to identify file types at a glance saves time and cognitive energy. The file-icons package replaces the minimalist icons that are used in a base installation of Atom, with a colorful icon set that clearly indicates the files in a project.
Learn more about file-icons here.
File icons give us immediate recognition of what we’re working with.
platformio-ide-terminal is an embedded terminal emulator that sits in the bottom of Atom’s application window. This is one of my favorite Atom packages. By using it I am able to keep track of my development server without flipping from one window to the next. Additionally, this plug in supports multiple terminal instances, making it easy to do things like installing node packages, checking build processes and do a system update – all from the same window.
Learn more about platformio-ide-terminal here.
Save yourself window fatigue with platformio-ide-terminal.
What can I say about emmet that has not already been said? First off, emmet is used to write HTML – lot’s of HTML using a simple, easy to remember shorthand. The package has been ported to almost every text editor in existence, and is shipped in Visual Studio code by default. If you’re not using emmet to write your HTML, then you are truly missing out.
Learn more about emmet here.
Learn more about atom-live-server here.
Guys and gals, did you read that? THIS PACKAGE DOES LIVE REFRESHES WHEN YOU MODIFY YOUR CODE.
Of the packages I’ve talked about thus far, todo is probably the most useful, and underrated package that I have come across. todo takes specially formatted comments that you write in your project files, and converts them into a to do list which is displayed on the right-hand side of the editor window. Each list item serves as a link to the file where your time is needed. This package could possibly serve as a quick and dirty project management solution for teams working on smaller projects, or a reminder of how disorganized I can get at times.
Learn more about todo here.
Useful, underrated, and filled with potential: todo.
minimap provides an overhead view of the file you’re currently working in, and lets you quickly navigate to parts of the file, by clicking on map regions. minimap takes up a good share of editor screen real estate, but when working with larger files, it definitely beats scrolling.
Learn more about minimap here.
If you’ve got boatloads of code, try minimap.
highlight-selected is an Atom plugin that should have been included by default. When you double-clicking a word, an HTML tag, or a variable name, highlight-selected displays all occurrences of that same word in a currently active file. This package can come in handy when tracking down misspellings or typographical errors in your code.
Learn more about highlight-selected here.
highlight-select minds your Ps and Qs, so you don’t have to.
pigments is an Atom plugin that takes CSS color notation and creates a text highlight of that color in your style sheet. pigments can also be used to convert colors between hex, rgb and rgba formats. When working with CSS colors, this package is a must have.
Learn more about pigments here.
pigments gives you nice highlights along with color conversion.
If you are a developer, chances are you work with a version control system. If you are using Git, this package is one you can’t live without. git-plus lets you perform any git command directly from the Atom editor. Although some configuration is needed in order to set up this package correctly, git-plus will save you time at the command-line, and makes commits fun and easy.
Learn more about git-plus here.
git-plus lets you perform any git command directly from the Atom editor.
For those who use Node with npm, autocomplete-module-import is a very compelling tool. As you write a require or import statement, the package creates a drop-down list of possible modules. When a module is selected, it is automatically downloaded and installed into your project. This is all done without having to open a terminal window, or leave the project.
Learn more about autocomplete-module-import here.
Less clicking, moving, and keystrokes with npm: meet autocomplete-module-import
11. Linters and Linting
The last few packages that I will recommend fall into the category of linters. Outside of unit testing, linters are probably one of the most potent fail-safes that developers have against writing bad code. Linters are traditionally run after a file is saved and modified. They check code for errors and inform you of any syntactic or stylistic error that you may have in your code.
Join the Discussion!
In this article, I discussed many of the packages I use on day to day basis with the Atom Text editor. What is your favorite text editor? What packages do you use? Join the discussion in the RTC Gitter chatroom, or comment below.
Michael Kornblum works as a cashier for a major big box retailer. He loves Linux, open source software and sharing his knowledge with others. He has written articles for freedompenguin.com and Linkedin. Currently, he is working on his developer portfolio, and hopes to land that ideal gig where he can make peoples’ lives better through technology.
|
OPCFW_CODE
|
[Teams] Copy/Pasted email addresses from Outlook app sent from teams are not accessible from bot
Issue
When a user copy an email from outlook desktop application and paste it to MS Teams, the email address is not accessible from bot.
Expected behavior
Bot should be able to access the email address pasted on MS Teams
How to reproduce
Copy an email from Outlook desktop application:
Paste email onto a private message to the bot on MS Teams:
Check the message received in the bot does not contain the email address
{
"text": "GAVALDA Yasmina ",
"textFormat": "plain",
"type": "message",
"timestamp": "2020-02-06T12:31:35.5977294Z",
"localTimestamp": "2020-02-06T13:31:35.5977294+01:00",
"entities": [{ "locale": "en-GB", "country": "GB", "platform": "Mac", "type": "clientInfo" }],
"textLocale": "en-GB",
"sourceEvent": { "tenant": { "id": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx" } },
"attachments": [],
"address": {
"id": "1580992295582",
"channelId": "msteams",
"user": {
"id": "29:1LCMqBdwgwSnBTHnBQuJ2VnM4vE8BJ1dSU-H1-2WS3K1S_rof3cBIbDtuU7C-04y3Asmm22Cv27ugTaw_qFlN9g",
"name": "GAVALDA Yasmina",
"aadObjectId": "f67fda16-4520-4766-af42-502a9cd18b21"
},
"conversation": {
"conversationType": "personal",
"tenantId": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx",
"id": "a:1K2e1KIquIQgQmbcMVrzgpR3Gc7GL-QPtmudiqFxAn0HqAYvkLa6W3RxbJqpEej4J76gPsZbxmbdrEQFXT575c4-J_HsVSlRv6lC5FPnrj4rc-XUJDc4-7kLfPTYApm-h"
},
"bot": { "id": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx", "name": "XXXXXXXXXX" },
"serviceUrl": "https://smba.trafficmanager.net/emea/"
},
"source": "msteams",
"agent": "botbuilder",
"user": {
"id": "29:1LCMqBdwgwSnBTHnBQuJ2VnM4vE8BJ1dSU-H1-2WS3K1S_rof3cBIbDtuU7C-04y3Asmm22Cv27ugTaw_qFlN9g",
"name": "GAVALDA Yasmina",
"aadObjectId": "f67fda16-4520-4766-af42-502a9cd18b21"
}
}
Note that if the user writes a text with the email copy/pasted from outlook, bot receives the text but not the email address:
{
"text": "Hello! My email is GAVALDA Yasmina ",
"textFormat": "plain",
"type": "message",
"timestamp": "2020-02-06T14:04:25.6353549Z",
"localTimestamp": "2020-02-06T15:04:25.6353549+01:00",
"entities": [{ "locale": "en-GB", "country": "GB", "platform": "Mac", "type": "clientInfo" }],
"textLocale": "en-GB",
"sourceEvent": { "tenant": { "id": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx" } },
"attachments": [],
"address": {
"id": "1580997865585",
"channelId": "msteams",
"user": {
"id": "29:1LCMqBdwgwSnBTHnBQuJ2VnM4vE8BJ1dSU-H1-2WS3K1S_rof3cBIbDtuU7C-04y3Asmm22Cv27ugTaw_qFlN9g",
"name": "GAVALDA Yasmina",
"aadObjectId": "f67fda16-4520-4766-af42-502a9cd18b21"
},
"conversation": {
"conversationType": "personal",
"tenantId": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx",
"id": "a:1K2e1KIquIQgQmbcMVrzgpR3Gc7GL-QPtmudiqFxAn0HqAYvkLa6W3RxbJqpEej4J76gPsZbxmbdrEQFXT575c4-J_HsVSlRv6lC5FPnrj4rc-XUJDc4-7kLfPTYApm-h"
},
"bot": { "id": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx", "name": "XXXXXXXXXX" },
"serviceUrl": "https://smba.trafficmanager.net/emea/"
},
"source": "msteams",
"agent": "botbuilder",
"user": {
"id": "29:1LCMqBdwgwSnBTHnBQuJ2VnM4vE8BJ1dSU-H1-2WS3K1S_rof3cBIbDtuU7C-04y3Asmm22Cv27ugTaw_qFlN9g",
"name": "GAVALDA Yasmina",
"aadObjectId": "f67fda16-4520-4766-af42-502a9cd18b21"
}
}
Hi @yasminagavalda - this probably has to do with the angle brackets being interpreted as markup. Have you tried pasting an email address without the angle brackets?
Also, the format of your activities makes it look like you're using the v3 Node SDK. While the activity schema has remained unchanged, it's possible that the transformations applied by that SDK have removed some information from the activity. Would you consider migrating to v4?
Hi @v-kydela, I have tried to write the same text with angle brackets and it arrives to the bot. And I have copy/pasted the same text from another app or text editor and it arrives to the bot too. The issue only exists when I copy/paste from the Outlook app.
We started this bot with v3 and it is almost finished, so we can’t migrate to v4 now. I will consider the v4 for the next bots.
@yasminagavalda - That's interesting. The problem I'm able to reproduce on Teams for Windows is that nothing contained in angle brackets gets sent to the bot, so there's nothing the bot can do to recover the information regardless of what SDK you're using.
@billbliss
@clearab
Can you have a look at this? It appears to be a Teams bug.
I think this is a regression because it used to work, and it should work. We've filed a bug. I think Kyle is right, it doesn't matter what version of the SDK you use (I wouldn't think it would). It also happens on the iOS client so this is likely a service regression.
I was able to repro from the Teams client with any text enclosed in angle brackets, not just e-mail addresses. Mine is dump C# echo bot v4.x.
It does not happen from the WebChat test.
It does not happen in a human-to-human Teams conversation.
So I guess it is in the Teams channel, indeed.
Hi @billbliss,
Is there any news about this bug?
@Wajeed-msft can you take a look while Bill is oof?
@yasminagavalda - We found this as a regression Bug, we have opened a bug for this, will get back to you once fix will be rolled out
Closed because bug has been filed and been tracked elsewhere. If there are any additional issues, please re-open or file a new issue. Thanks!
Updates: This has been fixed and deployed. Please let us know if you are still facing this issue.
Updates: This has been fixed and deployed. Please let us know if you are still facing this issue.
@Wajeed : The issue is still there only difference is when I copy and paste any piece from one of the replies of Teams Bot itself then it doesnt work.
Example: I have a welcome message in my bot, where I show an example how to proceed. If I copy the text from the example and paste in my chat, my bot doesnt take that as input and restarts the conversation from again rendering the Welcome message.
If I type that example then it works. I also tried copying and pasting from outside application like notepad and it works. It doesnt work only if I copy the text from Teams Bot reply. Please check
|
GITHUB_ARCHIVE
|
Mona Dennis Template January 26th, 2020 - 15:35:28
Just think about what that means for second. How often have you come across a website that looked like it was thrown together without any serious thought being put into it? What did you think about the website? Did you believe it was good or bad? Maybe if your only goal was to read content that was on it and then navigate away from it then you might not mind. But what if your goal was to sell products on the website? What if your goal was to generate lots of revenue from the website you're going to create? You would at least need a website that has a semi-professional look right?
CMS has one more feature that is important to mention. It can be used like a basement for web templates. So, more and more web developers decide on CMS while creating their templates. Taking into consideration that Joomla is a comfortable, well-coded and easy-manageable CMS it was chosen to be the basement of Joomla Templates. These templates have been launched by the biggest and most reliably templates' provider Template Monster and they become very popular among those people who wish to build an outstanding web site.
There is one disadvantage to free website templates, and that would be that some of them just are not meant for those who have specific types of businesses. These templates are really meant for those who are either just getting started or those who only need a semiprofessional look. Once again if you have highly specialized needs then these templates are not going to be the way to go.
If you try to contact the template website you bought the template from, you will find many of them not able to help you with your required template customization. As you might notice, there are hundreds of web sites offering exactly the same set of web templates -- as they are just affiliate websites reselling websites templates and flash templates from a third party template provider. Therefore many template websites are just interested in selling more templates rather than doing any customization work, or might just redirect you to an expensive template update service so that they can collection commissions by referring you. They are mostly just 'Internet marketers', rather than web developers.
Take the time and work with your user community to identify the common templates used throughout the organization. Collect and organize these templates so they can be easily found and included in a workgroup template location. Take the time to prune any templates that are no longer necessary and update any requiring attention. Also take the time to identify templates that are still needed. Ideally, every common form would have a corresponding template. This process will not be quick or easy, so it makes a far better place to start than with any of the other technology portions.
A final note on templates in general. Those willing to buy the software, take the time to learn it, and use a template should be aware that many templates are not "well coded." For the most part this may not effect or even matter to the template buyer. They will have a website and it will "work". Anyone who is also attempting to optimize their site for search engine placement, or who wants to make their site more efficient and use less code, should pay attention to the structure of the HTML files provided in the template. Many, though by no means all, templates, use a "tables" based layout that could be improved upon to make it more code efficient and make search engine optimization easier.
|
OPCFW_CODE
|
5 Essential Steps to Hardening Your MySQL Database
Databases have long been an important part of your average architecture and arguably the most important. These days the trend has moved toward most parts of your infrastructure being disposable and stateless, which puts an even greater burden on your database to be both reliable and secure, since all of the other servers inevitably store stateful information along with the rest of their data in the database.
Your database is the jackpot that every attacker aims to capture. As attacks get more sophisticated and networks get more hostile, it’s more important than ever to take additional steps to harden your database.
MySQL has been a popular and favorite database of developers and administrators for its speed and overall ease of use. Unfortunately, that ease of use comes at the expense of security. Even though MySQL can be configured with tight security controls, your ordinary default MySQL configuration probably doesn’t use them. In this article I will cover five important steps you should take to harden your MySQL database.
Step No. 1: Set Strong Passwords
It’s important for all your database users to use strong passwords. Given that most people don’t manually log into a database all that often, use a password manager or the command-line tool, pwgen, to create a random, 20-character password for your database accounts. This is important even if you use additional MySQL access control to limit where a particular account can login from (such as limiting accounts strictly to localhost).
The most important MySQL account to set a password for is the root user. By default in many systems, this user will have no password. In particular, Red Hat-based systems won’t set up a password when you install MySQL; and while Debian-based systems will prompt you for a password during interactive installs, non-interactive installs (like you might perform with a configuration manager) will skip it. Plus, you can still skip setting a password during interactive installs.
You might think that leaving the root user without a password isn’t a big security risk. After all, the user is set to “root@localhost” which you might think means you’d have to root the machine before you can become that user. Unfortunately, it means that any user who can trigger a MySQL client from localhost can login as the MySQL root user with the following command:
$ mysql — user root
So if you don’t set a password for the root user, anyone who may be able to get a local shell on your MySQL machine now has complete control over your database.
To fix this vulnerability, use the mysqladmin command to set a password for the root user:
$ sudo mysqladmin password
Unfortunately, MySQL runs background tasks as that root user. These tasks will break once you set a password, unless you take the additional step of hard-coding the password into the /root/.my.cnf file:
[mysqladmin]user = rootpassword = yourpassword
Unfortunately, this means that you have to keep your password stored in plain text on the host. But you can at least use Unix file permissions to restrict access to that file to only the root user:
$ sudo chown root:root /root/.my.cnf$ sudo chmod 0600 /root/.my.cnf
Step No. 2: Remove Anonymous Users
Anonymous accounts are MySQL accounts that have neither a username nor a password. You don’t want an attacker to have any kind of access to your database without a password, so look for any MySQL users recorded with a blank username in the output of this command:
> SELECT Host, User FROM mysql.user;+ — — — — — — — — — — — — + — — — -+| Host | User |+ — — — — — — — — — — — — + — — — -+| 127.0.0.1 | root || ::1 | root || localhost | || localhost | root |+ — — — — — — — — — — — — + — — — -+4 rows in set (0.00 sec)
In the middle of those root users is an anonymous user (localhost) with a blank in the User column. You can get rid of that particular anonymous user with this command:
> drop user “”@”localhost”;> flush privileges;
If you identify any other anonymous users, be sure to delete them as well.
Step No. 3: Follow the Principle of Least Privilege
The principle of least privilege is a security principle that can be summed up like this:
Only give an account the access it needs to do the job and nothing more.
This principle can be applied to MySQL in a number of ways. First, when using the GRANT command to add database permissions to a particular user, be sure to restrict access to just the database the user needs access to:
> grant all privileges on mydb.* to someuser@”localhost” identified by ‘astrongpassword’;> flush privileges;
If that user only needs access to a particular table (say, the users table), replace ``mydb.*`` (which grants access to all tables) with ``mydb.users`` or whatever the name of your table happens to be.
Many people will grant a user full access to a database; however, if your database user only needs read data but not change data, go the extra step of granting read-only access to the database:
> grant select privileges on mydb.* to someuser@”localhost” identified by ‘astrongpassword’;> flush privileges;
Finally, many database users will not access the database from localhost, and often administrators will create them, like this:
> grant all privileges on mydb.* to someuser@”%” identified by ‘astrongpassword’;> flush privileges;
This will allow “someuser” to access the database from any network. However, if you have a well-defined set of internal IPs or — even better — have set up VLANs so that all of your application servers are on a different subnet from other hosts, then take advantage to restrict “someuser,” so that account can access the database only from a particular network(s):
> grant all privileges on mydb.* to firstname.lastname@example.org/255.255.255.0 identified by ‘astrongpassword’;> flush privileges;
Step No. 4: Enable TLS
Setting strong passwords only gets you so far if an attacker can read your password or other sensitive data as it passes over the network. Therefore, it’s more important than ever to secure all of your network traffic with TLS.
MySQL is no different.
Fortunately it’s relatively simple to enable TLS in MySQL. Once you have valid certificates for your host, just add the following lines to your main my.cnf file under the ``[mysqld]`` section:
For extra security, also add the ``ssl-cipher`` configuration option with a list of approved ciphers, instead of just accepting the default list of ciphers, which may include weaker TLS ciphers. I recommend either the Modern or Intermediate cipher suites as recommended by the Mozilla Security/Server Side TLS page.
Once TLS is set up on the server side, you can also restrict clients so that they must use TLS to connect by adding ``REQUIRE SSL`` to your GRANT statement:
> grant all privileges on mydb.* to email@example.com/255.255.255.0 identified by ‘astrongpassword’ REQUIRE SSL;> flush privileges;
Step No. 5: Encrypt Database Secrets
While most people these days know how important it is to protect user database-stored passwords with a one-way hash (ideally a slow hash like bcrypt), often not much thought is given to protecting other sensitive data in the database with encryption. In fact, many administrators will tell you their database is encrypted because the disk itself is encrypted. This actually compromises your database hardening not because disk encryption is flawed or bad practice, but just because it will give you a false sense of security.
Disk encryption protects your database data should someone steal the disks from your server (or buy them secondhand after you’ve forgotten to wipe them), but disk encryption does not protect you while the database itself is running, since the drive needs to be in a decrypted state to be read.
To protect data that’s in the database, you need to take the extra step of encrypting sensitive fields before you store them. That way if an attacker finds out some way to do a full database dump, your sensitive fields are still protected.
There are a number of approaches to encrypting fields in a database and MySQL supports native encryption commands. Whatever encryption method you choose, I recommend avoiding encryption methods that require you to store the decryption key on the database itself.
Ideally you will store the decryption secret on the application server as a local GPG key (if you use GPG for encryption) or otherwise store it as an environment variable on the application server. That way even if an attacker may find a way to compromise the service on the application server, he’ll have to turn that attack into local shell access to get your decryption key.
MySQL Hardening Maxim: Master the Principle of Least Privilege
There are many ways to lock down your MySQL server. Exactly how you implement some of these steps will vary depending on how your own database is set up and where it sits in your network.
While the previous five steps will help protect your database, I’d argue the most important overall step to master is the principle of least privilege. Your database likely stores some of your most valuable data and if you make sure that users and applications only have the bare minimum access they need to do their job, you will restrict what an attacker will be able to do, should the hacker find a way to compromise that user or application.
About the blogger: Kyle Rankin is an award-winning IT journalist and author, who has written or contributed to more than a dozen books on open-source nuts and bolts, including the Knoppix Hacks series, The Official Ubuntu Server book, and DevOps Troubleshooting. His latest, Linux Hardening in Hostile Networks, has recently been published by Prentice Hall.
He is a columnist at Linux Journal magazine and speaks frequently on security and open-source software, including at O’Reilly Security Conference, BSidesLV, CactusCon, SCALE, OSCON, Linux World Expo, and Penguicon.
A specialist in Linux administration, automation and security, Kyle’s expertise as a senior systems administrator comes in handy at his regular 9-to-5 as Vice President of Engineering Operations at Final, Inc.
Please feel free to share below any comments, questions or insights about your experience with hardening your MySQL database. And if you found this blog useful, consider sharing it through social media.
|
OPCFW_CODE
|
Thanks for your excellent question:
You have already placed all the religious into the category of irrational, and you, yourself, into the category of rational. What makes you know anything more or better than anyone else - religious or otherwise? Why should anyone listen to you?
Tammy, I used to be extremely religious. About thirteen years ago a man suggested to me that I read a free online book, called "Software for your brain". This is not an anti-religion book... it's an objective book which explains what open mindedness and closed mindedness is, and how to become more open minded.
I read the book. Some of it sank in, and some of it didn't. The most important thing I learned from the book was the concept of "Cognitive Dissonance". That is when you hear something new, or not inline with your beliefs, and your brain immediately rejects it. The book teaches you how to enable yourself to consider new information, instead of rejecting it.
Over the next few years I started questioning my religion, but instead of realizing that I was questioning flawed concepts I presumed that my understanding was flawed. Because of this I started making up a whole bunch of excuses for the problems in my religion.
It was only after about ten years that I decided to have a chat with a preacher. He pointed out that the excuses I had made up contradicted the bible. Althought it was not his intention, he helped me realize that I should not make excuses for the questions that I have. I had many questions at this point, about things in the bible that made no sense to me at all.
In an attempt to be open minded I wrote many emails to religious people, asking them about my questions. Some of these emails took many hours to write. I also read everything they sent me, and everything they suggested that I read, trying desperately to find answers to my questions.
I put ten of my top questions in a video on YouTube, and received hundreds of views and about a hundred responses. Not a single one of my questions was answered in a way that makes sense to me.
The only reason I have been able to come up with for having so many unanswered questions about contradictions and illogical problems in the bible is the simple fact that the bible contains many lies or mistakes.
People simply believe the bible, and do not see the mistakes, because they have been brainwashed to believe that the bible is somehow perfect and does not contain any.
Many religious people believe that the bible is entirely true, but they do not have any rational reason to believe this. They may even use arguments like "it says that it's true".
I could go on and on about this, but anyone with strong cognitive dissonance will reject everything that I say that contradicts with their beliefs, so it's pointless to write any more.
This is why I asked the question, "Have you ever successfully persuaded anyone?" What makes people listen to reason?
|
OPCFW_CODE
|
I have a complex one so forgive me if this isn't clear - if absolutely necessary I'll create an example but I just don't have time at the moment.
I have data in a table which represents a Risk Matrix:
|Unlikely - Rarer than once every 15 years||Critical area||255||75|
|Severity 1||Rare - Once between every 15 years and 5 years||Critical area||255||150|
|Severity 1||Likely - Once between every 5 years and 2 years||Critical area||255||225|
|Severity 1||Highly - Once between every 2 years and once a year||Critical area||255||300|
|Severity 1||As good as certain - Multiple every year||Critical area||255||375|
|Severity 2||Unlikely - Rarer than once every 15 years||Serious area||65535||25|
|Severity 2||Rare - Once between every 15 years and 5 years||Serious area||65535||50|
|Severity 2||Likely - Once between every 5 years and 2 years||Critical area||255||75|
|Severity 2||Highly - Once between every 2 years and once a year||Critical area||255||100|
|Severity 2||As good as certain - Multiple every year||Critical area||255||125|
|Severity 3||Unlikely - Rarer than once every 15 years||Less serious area||65280||10|
|Severity 3||Rare - Once between every 15 years and 5 years||Serious area||65535||20|
|Severity 3||Likely - Once between every 5 years and 2 years||Serious area||65535||30|
|Severity 3||Highly - Once between every 2 years and once a year||Serious area||65535||40|
|Severity 3||As good as certain - Multiple every year||Serious area||65535||50|
|Severity 4||Unlikely - Rarer than once every 15 years||Less serious area||65280||5|
|Severity 4||Rare - Once between every 15 years and 5 years||Less serious area||65280||10|
|Severity 4||Likely - Once between every 5 years and 2 years||Less serious area||65280||15|
|Severity 4||Highly - Once between every 2 years and once a year||Serious area||65535||20|
|Severity 4||As good as certain - Multiple every year||Serious area||65535||25|
|Severity 5||Unlikely - Rarer than once every 15 years||Less serious area||65280||1|
|Severity 5||Rare - Once between every 15 years and 5 years||Less serious area||65280||2|
|Severity 5||Likely - Once between every 5 years and 2 years||Less serious area||65280||3|
|Severity 5||Highly - Once between every 2 years and once a year||Less serious area||65280||4|
|Severity 5||As good as certain - Multiple every year||Less serious area||65280||5|
By creating a Pivot Chart from this data I can create a 5x5 table, containing (for arguement's sake) "WEIGHT" data with each Cell the appropriate COLOUR:
My issue is that, to create this Matrix, I cannot link the data (above) to my base data (ie: holding Dates and Events) - as if a Selection is made which does not include one of the 5 SEVERITIES or 5 LIKELIHOODS then I cannot create the full Matrix. I can obviously select the "Show all Values" and deselect "Supress When Value is Null" Chart options - but then I get extra "blank" rows and columns - plus I then no longer have access to all Cell COLOURS, eg:
Of course this means that (if I make the data above a "data island") when I attempt to COUNT the number of Events for each Cell - ALL data is returned, regardless of the current Selection.
I am currently toying with the idea of a hybrid of the two implementations where I use a "data island" version of the table for the 'background' (with COLOURS) and then a transparent 'foreground' Pivot Chart using a linked table for the currently Selected data... however, I would like to show Totals (without having the "blank" rows and columns shown above - which, by viewing the table data highlighted above, I can understand why these are present).
Am I missing some fundamental data concept?!
How can I create an unrelated Pivot Chart which displays linked data? (or how can I remove the "blank" rows and colums from my Pivot Chart?).
Any help will be greatly appreciated!
|
OPCFW_CODE
|
|Developer||Robert Alan Koeneke, James E. Wilson, others|
|P. Language||C, Pascal|
|Platforms||Linux, MS-DOS, Mac Classic, Atari ST, Amiga, *NIX|
|Game Length||50+ Hours not unusual|
|Official site of Moria|
Moria, first released in 1983, is one of the earliest clones of Rogue. In 1988 it was rewritten in the C language and released as Umoria. Although this was originally a port, it can very much be considered a continuation of the Moria game.
Written by Robert Alan Koeneke, Moria was based on J.R.R. Tolkien's The Lord of the Rings, where the player has to go down in to the mines of Moria and defeat the Balrog to win the game.
This game was the first roguelike to have a "Town" level, where you may buy your weapons, armor, spell books, potions, and various other item to help you on your quest. The game world is "The Dungeons (or Mines) of Moria", and although the name comes from the world of Tolkien, there is little that the game shares with the books. In the deepest level, you must find and defeat the Balrog. This is a hard quest for which you must prepare with the proper equipment and character enhancement items. It is not unusual for a game win to take over 50 hours.
It was the first open source roguelike, which made it possible for it to be ported to many different computer platforms, in a time when that was hard to achieve. Although the game is not as popular as it once was, it is still considered a major Roguelike.
Robert Alan Koeneke became hooked on "Rogue" while at the University of Oklahoma. Around 1981 he got a job in a new department where the game wasn't available, so he decided to write his own Rogue game written in Basic and called it Moria Beta 1.0. In 1983 he enrolled in a Pascal operating systems class and that summer finished Moria 1.0, rewritten in VMS Pascal. In 1985 he started sending out the source code to other Universities, and it was during this time that the game became popular.
Koeneke was working on Moria 5.0, which was an almost complete rewrite with interesting features like streams, lakes and new weapons. It was, however, never released. A different version called Moria UB 5.0 appeared at the University at Buffalo; this is the last VMS version ever, and is often called VMS Moria 5.0.
The last official "Koeneke" Moria, v4.7, was released in 1987.
The release of the source code allowed Moria to survive its VMS origins. After the creation of Umoria (moving to the C language), the game became available on many different computer platforms such as DOS, Amiga and Atari ST, and also spawned many variants, with the most successful being Angband.
It is also know to have been an inspiration for the first commercially successful Roguelike, Diablo.
- Bruce Moria
- Colour Umoria
- CWM Moria (Amiga)
- Druid Moria (and colour Druid)
- jsMoria (a web Umoria port)
- Purple X (Apple)
- Russian Moria
- David Grabiner's Moria Page
- Beej's Moria Page
- Usenet article, where the author himself describes the origins of Moria
- Moria got its own dedicated newsgroup. It can be accessed through Google groups.
- free-moria, a successful 2007 project by Ben Asselstine to re-license UMoria 5.5.2 under the GPL by permission of all contributing authors
- Archived UMoria source code history, and Windows ports of UMoria 5.6
|
OPCFW_CODE
|
I don't know where to go with this. I've got a few Patterns. I know how to load in my PS but I don't have a clue how i can use it. I've search everywhere, but don't find anything. I would love to use them in my next icons. Could anyone help me please?
I use Photoshop CS2 and IR.
I have the gif file I want to use for my animated icon I wish to make, but I want to brighten and sharpen the picture, I have tried several times to get the whole thing adjusted, but it only does one frame, I tried selecting all frames and adjusting, what am I doing wrong?
I know this is an icon tutorial community but I just have a quick photoshop question. I have to make a layout that just says "gruesome" and I've got the design all ready. I just need to know how to make a gross fleshy graphic to go along with it. Do you know how to make something look "fleshy" and disgusting. Like a rotting body up close? Should I get a stock photo or start from scratch? If anyone can point me to tutorials or if you have tips of your own It'll really be helpful.
Hi, I've searched through the memories, and I couldnt find anything. I'm not sure I'm using the right term, but can anyone direct me towards a tutorial for making scanlines?
EDIT: I'm using PS7 and any diagonal, vertical, or horizontal lines, or grids are what I'm looking for.
I use PSP8 and I was wondering if there is an easy way to do selective coloring. I always end up doing it a way that takes ages and looks horrible.
Could someone point me in the direction of some good tutorials for icons with animations??
I have psp 9 and animation shop 9, so hopefully they will be something i can make with those 2 programs!
Thanks so much!
hi! umm, here the problem; I wanted to use my new trasparent image on Livejournal, and I downloaded paint.net, after it was recommended. All went well with the deleting of the backround but here's the problem: it was saved as a .pdn file, and as you all know, LJ only accepts the file types of .jpg, .bmp , .gif , & .png.
So here's the question:
how do I change it into one of these filetypes without having the image flatten, which renders my work useless?
Does anyone know where I could find a tutorial on how to make really nice text-only icons? Thanks in advance. :)
please don't think i'm stupid or anything.. but..
how the heck do you get images from music videos? i've tried doing the 'print screen' button while it's playing, and USUALLY it works, but with some videos, it doesn't. szldkfje help.
|
OPCFW_CODE
|
Ahead of the Windows 10 Release, Microsoft Cleans Up Its App Store – Motley Fool
Installing applications on a PC is still a huge problem. Software downloads live all over the Web. Users must know about them and where to find them before they can download them. Then they have to install them, which can often be a frustrating experience for computing novices.
Microsoft (NASDAQ:MSFT) is attempting to fix this problem by mimicking the app stores of Apple (NASDAQ:AAPL) and Google (NASDAQ:GOOG) (NASDAQ:GOOGL), which were created in the infancy of smartphone adoption. Apple implemented its version of the app store for Mac in 2011, but it too has struggled to overcome developers’ old habits. Microsoft’s app store has arguably fared even worse and suffers from a large number of spammy and counterfeit apps.
With the release of Windows 10 just weeks away and a heavy focus on the Windows Store in the newest iteration of the OS, Microsoft is working to cull and curate the apps in its app store.
How did we get here?
Microsoft introduced the Windows Store along with Windows 8 in 2012. In 2013, Microsoft started offering developers a monetary incentive to add their apps to the app store — $100 per app, for up to 10 apps for each of Windows 8 and Windows Phone. As a result of this tactic, Microsoft was able to grow the number of apps in the Windows Store, but many developers took advantage of seemingly low standards from Microsoft to upload spammy, counterfeit, and potentially dangerous apps.
Last fall, Microsoft removed 1,500 apps from its app store after reports pointed to these problems with the Windows Store. But 1,500 apps is just the tip of the iceberg for Microsoft, which now boasts over 200,000 apps in its app store. Late last month, Microsoft announced that it will make further improvements to the quality of the apps in its app store by removing mispriced and misleading apps, and setting higher standards for future apps.
Because of Microsoft’s early tactics, the Windows Store suffers from a high percentage of low-quality apps, which makes the platform unattractive to quality app developers. While quantity is important, quality is even more crucial. Apple’s App Store is successful because it attracts top-quality developers looking to target Apple’s premium audience. Google Play is able to attract quality developers because of Android’s scale.
Microsoft has offered very little incentive for successful developers to put their apps in the Windows Store.
Microsoft has struggled mightily to get users to upgrade to Windows 8, which is currently the only operating system that supports the Windows Store. As of April, fewer than 15% of PC users have Windows 8 or Windows 8.1 installed. Most users upgraded to Windows 7 after the expiration of Windows XP support last year, and 58% of PC users now have Windows 7 installed on their machines.
That’s why Microsoft is offering a free upgrade to Windows 10 for most Windows 7 and Windows 8 users. It set a time limit of one year to give users an incentive to upgrade sooner rather than later. Microsoft is also unifying Windows 10 with its mobile operating system, so it’s one OS for all devices. If successful, Microsoft will create a much larger audience for the Windows Store and its developers with the release of Windows 10.
Coupled with a reduction in low-quality apps, massive scale will attract high-quality developers. The potential to easily reach a wide audience is what makes developers put their apps in Google Play instead of just asking users to sideload the app from their websites. That’s why developers are willing to give up 30% of their revenue to Google or Apple. Without scale, the 30% cost for developers in the Windows Store remains unattractive.
Microsoft’s decision to give away Windows 10 to consumers means that there’s a lot riding on making up for the lost revenue through services like the Windows Store. After a decline in Windows licensing revenue from consumers and the poor reception for Windows 8, Microsoft needs to spur rapid adoption of Windows 10 and the Windows Store. It will be interesting to watch how this strategy plays out over the next year.
The next billion-dollar Apple secret
Apple forgot to show you something at its recent event, but a few Wall Street analysts and the Fool didn’t miss a beat: There’s a small company that’s powering Apple’s brand-new gadgets and the coming revolution in technology. And its stock price has nearly unlimited room to run for early in-the-know investors! To be one of them, just click here.
|
OPCFW_CODE
|
use bdd::closet::Closet;
use core::Family;
use core::Item;
use std::collections::HashMap;
impl Closet {
pub fn categorize(&self, items: &[&Item]) -> HashMap<Family, Vec<Item>> {
let unknown_family = Family::new("UNKNOWN");
items.iter()
.map(|&item| (self.get_family(item).unwrap_or(&unknown_family), item))
.map(|(family, item)| (family.clone(), item.clone()))
.fold(HashMap::new(), |mut categories, (family, item): (Family, Item)| {
categories.entry(family).or_insert_with(|| vec![]).push(item);
categories
})
}
}
#[cfg(test)]
mod tests {
use bdd::closet_builder::ClosetBuilder;
use core::Family;
use core::Item;
use std::collections::HashMap;
#[test]
fn count_nodes_families_2_items_4() {
let shirt1 = Item::new("shirts:1");
let shirt2 = Item::new("shirts:2");
let pants1 = Item::new("pants:1");
let pants2 = Item::new("pants:2");
let shirts = Family::new("shirts");
let pants = Family::new("pants");
let closet_builder = ClosetBuilder::new()
.add_items(&shirts, &[shirt1.clone(), shirt2.clone()])
.add_item(&pants, &pants1)
.add_item(&pants, &pants2);
let closet = closet_builder.must_build();
let expected = {
let mut expected = HashMap::new();
expected.entry(shirts).or_insert(vec![shirt1.clone(), shirt2.clone()]);
expected.entry(pants).or_insert(vec![pants1.clone(), pants2.clone()]);
expected
};
assert_eq!(expected, closet.categorize(&[&shirt1, &shirt2, &pants1, &pants2]));
}
}
|
STACK_EDU
|
How to combine stacked polygons within a shapefile?
The image below shows my shp-file. Every polygon has the same transparent color. So a darker blue means, that several polygons are stacked! How do I cut every polygon with each other so no polygons are stacked anymore?
Are all of these polygons contained in a single vector layer?
Can you share a subset of your shape file so we can experiment with it? It would also help if you'll answer @JordanArseno's question.
If you have the Grass plugin you could clean your layer to "flatten" it but if you need the attributes (as you probably will) a spatial join will likely only get you part way there and the rest would need to be done manually. I think Sextant also provides access to the Grass clean.
If I understand correctly, the single shapefile contains overlapping-features (polygons). The brute force way if there's not many polygons...
(1) Save each polygon out as a separate shapefile using your classification as in this question.
(2) UNION the shapefiles.
It used to be possible just to union a layer with itself. Has that changed?
@whuber not sure, as I don't normally have shapefiles in this manner. If so, then that would be a lot easier for the OP.
I tried the union geoprocessing tool in QGIS on a sample datset with itself and it created a much larger number of polygons with a geometry error warning in QGIS 1.8.0. If QGIS had a batch union to go with the split layer tool that would be useful in some situations for a modest number of polygons or classes of polygons in a dataset. We do this in ArcGIS occasionally.
You can union within one feature class, no problems there. After union you would have to get rid of duplicates.
Simply erasing polygons here to generate holes will not work. You will always need to erase based on overlap (for example, erase the small polygon that is on top of the bigger polygon and not the other way around). My suggested workflow would be as follows:
Grab all the geometries and save them in a list.
Order the list by area descending (first geometry has biggest area)
Put them in some sort of index
For each polygon geometry (x), do the following:
Grab all overlapping polygons using the index (y)
Calculate the area overlap between x and each y
For each actual overlap:
Find the largest overlap by area
If the overlap is partial, erase that from x
Make sure the rest of the potential overlaps are still overlaps of the new x
Note that this solution is not necessarily going to work as there are some more issues based in the ordering you do things, as well as the overlap (will some polygons completely disappear?).
You can solve this in arcpy, ogr and probably shapely. I assume QGIS has similar functionality, but I cannot speak to it.
Other thoughts:
1. Don't forget a geometry object usually has a .area attribute
2. If using Python, sorting a list by the .area attribute becomes trivial
3. Geometries have intersect methods, as well as erase methods
4. The returned geometries from intersect methods can help you identify percentage of area you will potentially erase
Hope this helps a little bit.
|
STACK_EXCHANGE
|
Allow trusted services to securely access a network-restricted container registry
Azure Container Registry can allow select trusted Azure services to access a registry that's configured with network access rules. When trusted services are allowed, a trusted service instance can securely bypass the registry's network rules and perform operations such as pull or push images. This article explains how to enable and use trusted services with a network-restricted Azure container registry.
Use the Azure Cloud Shell or a local installation of the Azure CLI to run the command examples in this article. If you'd like to use it locally, version 2.18 or later is required. Run
az --version to find the version. If you need to install or upgrade, see Install Azure CLI.
- Certain registry access scenarios with trusted services require a managed identity for Azure resources. Except where noted that a user-assigned managed identity is supported, only a system-assigned identity may be used.
- Allowing trusted services doesn't apply to a container registry configured with a service endpoint. The feature only affects registries that are restricted with a private endpoint or that have public IP access rules applied.
About trusted services
Azure Container Registry has a layered security model, supporting multiple network configurations that restrict access to a registry, including:
- Private endpoint with Azure Private Link. When configured, a registry's private endpoint is accessible only to resources within the virtual network, using private IP addresses.
- Registry firewall rules, which allow access to the registry's public endpoint only from specific public IP addresses or address ranges. You can also configure the firewall to block all access to the public endpoint when using private endpoints.
When deployed in a virtual network or configured with firewall rules, a registry denies access to users or services from outside those sources.
Several multi-tenant Azure services operate from networks that can't be included in these registry network settings, preventing them from performing operations such as pull or push images to the registry. By designating certain service instances as "trusted", a registry owner can allow select Azure resources to securely bypass the registry's network settings to perform registry operations.
Instances of the following services can access a network-restricted container registry if the registry's allow trusted services setting is enabled (the default). More services will be added over time.
Where indicated, access by the trusted service requires additional configuration of a managed identity in a service instance, assignment of an RBAC role, and authentication with the registry. For example steps, see Trusted services workflow, later in this article.
|Trusted service||Supported usage scenarios||Configure managed identity with RBAC role|
|Azure Container Instances||Deploy to Azure Container Instances from Azure Container Registry using a managed identity||Yes, either system-assigned or user-assigned identity|
|Microsoft Defender for Cloud||Vulnerability scanning by Microsoft Defender for container registries||No|
|ACR Tasks||Access the parent registry or a different registry from an ACR Task||Yes|
|Machine Learning||Deploy or train a model in a Machine Learning workspace using a custom Docker container image||Yes|
|Azure Container Registry||Import images to or from a network-restricted Azure container registry||No|
Curently, enabling the allow trusted services setting doesn't apply to App Service.
Allow trusted services - CLI
By default, the allow trusted services setting is enabled in a new Azure container registry. Disable or enable the setting by running the az acr update command.
az acr update --name myregistry --allow-trusted-services false
To enable the setting in an existing registry or a registry where it's already disabled:
az acr update --name myregistry --allow-trusted-services true
Allow trusted services - portal
By default, the allow trusted services setting is enabled in a new Azure container registry.
To disable or re-enable the setting in the portal:
- In the portal, navigate to your container registry.
- Under Settings, select Networking.
- In Allow public network access, select Selected networks or Disabled.
- Do one of the following:
- To disable access by trusted services, under Firewall exception, uncheck Allow trusted Microsoft services to access this container registry.
- To allow trusted services, under Firewall exception, check Allow trusted Microsoft services to access this container registry.
- Select Save.
Trusted services workflow
Here's a typical workflow to enable an instance of a trusted service to access a network-restricted container registry. This workflow is needed when a service instance's managed identity is used to bypass the registry's network rules.
- Enable a managed identity in an instance of one of the trusted services for Azure Container Registry.
- Assign the identity an Azure role to your registry. For example, assign the ACRPull role to pull container images.
- In the network-restricted registry, configure the setting to allow access by trusted services.
- Use the identity's credentials to authenticate with the network-restricted registry.
- Pull images from the registry, or perform other operations allowed by the role.
Example: ACR Tasks
The following example demonstrates using ACR Tasks as a trusted service. See Cross-registry authentication in an ACR task using an Azure-managed identity for task details.
- Create or update an Azure container registry.
Create an ACR task.
- Enable a system-assigned managed identity when creating the task.
- Disable default auth mode (
--auth-mode None) of the task.
- Assign the task identity an Azure role to access the registry. For example, assign the AcrPush role, which has permissions to pull and push images.
- Add managed identity credentials for the registry to the task.
- To confirm that the task bypasses network restrictions, disable public access in the registry.
- Run the task. If the registry and task are configured properly, the task runs successfully, because the registry allows access.
To test disabling access by trusted services:
- Disable the setting to allow access by trusted services.
- Run the task again. In this case, the task run fails, because the registry no longer allows access by the task.
- To restrict access to a registry using a private endpoint in a virtual network, see Configure Azure Private Link for an Azure container registry.
- To set up registry firewall rules, see Configure public IP network rules.
|
OPCFW_CODE
|
import json
import urllib
import urllib2
def traveledDistance(travelLog):
dist = 0
maps_key = 'enterKeyHere' #enter API key
base_url = 'https://maps.googleapis.com/maps/api/distancematrix/json?'
start = travelLog[:-1]
end = travelLog[1:]
#find all of the start and end points in the journey by zipping together an offset version of the travelLog
travelLog = zip(start,end)
#iterate over the travelLog and find the distance between each tuple of locations in google's distanceMatrix API
#could have used the full list with one call, but this produces distances bewteen every permutation of entries
#which seemed unnecessary and time consuming
for d in travelLog:
params = urllib.urlencode({
'origins': d[0],
'destinations': d[1],
'key': maps_key,
})
url = base_url + params
response = str(urllib2.urlopen(url).read())
map_dict = json.loads(response.replace('\\n', ''))
#extract the distance value from the json response of the API and add to the running distance total
dist += (map_dict['rows'][0]['elements'][0]['distance']['value'])
return dist
|
STACK_EDU
|
from typing import *
from tqdm import tqdm
import os
import torch
import numpy as np
import urllib
import zipfile
from .vocabularies import Vocabulary
data_folder_path = "../../data"
model_folder_path = "../../model"
embeddings_folder = os.path.join(model_folder_path, "embeddings")
def create_embedding_tensor(vocabulary: Vocabulary, embedding_dim: int,
pretrained_embeddings: Dict[str, np.ndarray]) -> torch.Tensor:
"""
Creates a lookup tensor for the tokens in the vocabulary starting from pretrained embeddings.
Args:
vocabulary: The vocabulary with the mapping from tokens to indices.
embedding_dim: The dimension of the vectors of the embeddings.
pretrained_embeddings: The pretrained embeddings for the tokens.
Returns:
The lookup tensor of shape (vocabulary length, embedding dimension)
with the available pretrained embeddings for the tokens in the vocabulary
"""
embedding_tensor = torch.randn(len(vocabulary), embedding_dim)
initialised = 0
for i, w in vocabulary.itos.items():
if w not in pretrained_embeddings:
# check needed for <pad>, <unk> tokens
continue
initialised += 1
vec = pretrained_embeddings[w]
embedding_tensor[i] = torch.from_numpy(vec)
embedding_tensor[vocabulary["<pad>"]] = torch.zeros(embedding_dim)
print("Initialised embeddings {}".format(initialised))
print("Random initialised embeddings {} ".format(len(vocabulary) - initialised))
return embedding_tensor
def get_embeddings(emb_fpath: str, vocab: Vocabulary, emb_dim: int) -> torch.Tensor:
emb_dict = dict()
with open(emb_fpath, "r") as f:
for l in tqdm(f, desc="Loading pretrained word embeddings"):
line = l.split()
if len(line) == 2:
# fasttext has an header to be skipped
continue
tok = "".join(line[:-emb_dim])
if tok in vocab.stoi.keys():
vec = np.array(line[-emb_dim:], dtype=np.float32)
emb_dict[tok] = vec
return create_embedding_tensor(vocab, emb_dim, emb_dict)
def download_embeddings(embeddings_fpath: str):
url = 'https://dl.fbaipublicfiles.com/fasttext/vectors-english/crawl-300d-2M.vec.zip'
filehandle, _ = urllib.request.urlretrieve(url)
zip_file_object = zipfile.ZipFile(filehandle, 'r')
first_file = zip_file_object.namelist()[0]
file = zip_file_object.open(first_file)
content = file.read()
with open(embeddings_fpath, "w+") as f:
f.write(content)
file.close()
def prepare_embeddings(vocabulary: Vocabulary) -> torch.Tensor:
embeddings_fname = "crawl-300d-2M.vec"
embeddings_fpath = os.path.join(embeddings_folder, embeddings_fname)
embedding_dim = 300
if not os.path.isfile(embeddings_fpath):
print("Downloading pre-trained word embeddings...")
download_embeddings(embeddings_fpath)
print("Done!")
pretrained_embeddings = get_embeddings(embeddings_fpath, vocabulary, embedding_dim)
return pretrained_embeddings
|
STACK_EDU
|
|Type of mesh||structured, cartesian or cylindrical mesh|
|Dimension||2D and 3D|
This interface writes a grid and boundary condition file for ALPHA-FLOW. ALPHA-FLOW can deal with Cartesian, cylindrical and body-fitted single-block grids (O-grid, H-grid, C-grid and L-grid).
For Cartesian or cylindrical coordinates grids, a single ASCII file (called "Input data file for solver") is written. It contains the number of stations in each direction, the list of these stations, and the description of 2D and 3D partial regions such as obstacles, initial patches and boundary conditions. User-defined name can be given to these regions for set-up obstacles or patches initial condition.
In the case of body-fitted grid, two files are required. The grid coordinates are written in Fortran binary to a file called "BFC grid file". Additional information such as the grid dimensions, the description of 2D and 3D partial regions, and the indices of degeneracy (H-grid) are written separately in an ASCII file.
This ALPHA-FLOW output interface supports four types of grids:
The body-fitted grids can be generated with the HEXA-Mesher, option Multiblocks. One or several structured domains can be used as input for the BFC translators (compressible or incompressible). If several blocks are used, they are converted internally to a monoblock grid prior to writing the BFC grid file. Note that the ALPHA-FLOW translator does not accept a superdomain file as input. This is due to the particularity of ALPHA-FLOW H-type grids, which require the insertion of additional cells at the degenerated faces. The regular ICEM-Superdomain program does not support such grids. Therefore, the ALPHA-FLOW translator has its own multiblock to monoblock converter.
The cartesian and cylindrical grids can be generated with the GLOBAL-Mesher. From the GLOBAL-Mesher GUI, make sure to save the grid with the option "Staircase-Volume".
Once the grid is generated, the translator may be called from the main GUI. Simply select "ALPHA-FLOW" from the list of translators in the "Output" menu. After selecting the command "Write Input" from the same menu, a window opens up listing the options for the ALPHA-FLOW translator:
The interface generates the ALPHA-FLOW files in the project directory.
For 3D models, partial regions and obstacles may be defined on blocks (3D) or faces (2D). For 2D models, partial regions and obstacles may be defined on faces (2D) and edges (1D). For grids generated with the HEXA-mesher, partial regions and obstacles are composed of one or several families. For details on how to set boundary conditions and properties, please refer to the document "General Remarks".
To define a named region, select an entity then specifies its region_name. All entities with the same region name are written in the same BLOCK in the Input data file.
1.string1 Enter NAME 2.string2 Enter the region name (5 char. max.)
To define an obstacle, select an entity (family, subface, edge or family) and assign to it the character string OB, in the appropriate field of the boundary condition menu. If the non-continuous surfaces in the computational space are treated to vary smoothly, specify YES in the second character field (see below). By default, 'NO' is written for each obstacle in the input data file.
1.string1 Enter OB 2.string2 Enter YES, if applicable (optional)
For BFC-grids for incompressible solver, two types of obstacles can be defined:
For a bidimensional grid, slab-type obstacles are applied to surfaces and slit-type obstacles to edges. For a tridimensional model, they are applied to blocks and faces respectively.
1.string1 Enter SLIT or SLAB
Special tagging for H-type topology
For H-type grids, if zero-volume cells need to be generated along an edge or a face, this edge or face needs to be tagged CELL. So, whenever the index of degeneracy is equal to 1, the boundary of the obstacle is tagged with SLIT and the boundary of the zero-volume cells is tagged with CELL.
1.string1 Enter CELL
|
OPCFW_CODE
|
Expired SSO ticket
Hello Experts, I hope you can help me with this issue. Our portal has been up for a couple of months now, and it uses SSO along with BW. I created some portal content, uploaded it, and it worked fine, until today. Whenever I try to navigate to a page containing a Bex Query or BI content, I get the following error:
500 Internal Server Error
The initial exception that caused the request to fail was:
The system has received an expired SSO ticket
com.sap.mw.jco.JCO$Exception: (103) RFC_ERROR_LOGON_FAILURE: The system has received an expired SSO ticket
I searched on the forums but the closest to this problem was someone who had this same error using the Bex Analyzer. On SAP documentation, I checked for JCO and RFC errors, and all I got is that the error I am getting is possibly caused by an incorrect login, password, certificates, etc. But it does not make sense. I am able to log in to the portal, navigate everywhere else, and our SSO certificates are a long way from expiring, besides the fact that even yesterday navigation through these pages was possible. I changed the lifetime of SSO tickets, but still got the same error.
S J replied
If the authentication in SSO fails due to an old ticket then it could
be due to the below.I am afraid that it could be an issue where you
either forgot to logoff the previous session or that you didnt close the browser window from theprevious session.
Let me explain below.
1. The JCO client and the SSO ticket is cached based on the user,type ofauthentication(SSO in your case),
language and the JCO Destination. Thisis kept in the cache and discarded when you close the Session through a
log off or a browser close window. This cuts the session and the new
session is created.This issues a new ticket.
2. There were some issues earlier with session termination in NW04sSP6
etc but since you are on SP9 I think this should be bug free.
3. Can you please try and ensure that the logoff's and browser closes
happen.I feel you get the problem since your ticket expires just after
an hour. An hour of inactivity on part of the user would invalidate the
ticket and when he/she starts running the application again the session is
still there but the ticket has expired. Can you please try to extend theticket lifetime.
4. Also what are your session and ticket timeout times.As below
" Please set the timeout value for the security sessions (default 27h)
and the timeout value for the SSO ticket (default 8h) to the same value.
It should be a value that is higher than the maximum working time of anemployee, e.g. 16 hours. "
1. Login.Ticket_Lifetime = 8h
2. Session Expiration Period (SSO ticket time out)= 280000 msec = 8h
Can you please review Note 842635 .
Maintain your sessions according to this note.
|
OPCFW_CODE
|
17th of December 2020 is a very important date for me. That day I officially froze my existing creative projects, in which I have been involved for years, and started learning programming.
It was a hard decision, because it was not easy to give up on projects I was developing and putting a lot of effort for years (blog, social media accounts, workshops & next ebook) in favor of a new, vast, completely unknown field in which I did not know if I could manage at all…
But I was extremely determined and I knew that if I want to master, I need to focus on programming learning only. No side projects, no distractions. So what caused my determination you may ask? Oh, there were a lot of very different reasons and that is what this article is about, so keep reading!
Influence of the Walter Isaacson’s Book “The Innovators”
I just love Isaacson’s books. I read his Steve Jobs & Leonardo da Vinci biographies (these two individuals inspire me a lot), so I was sure, that “The Innovators” will be great. But I didn’t expect, that the book will change my life, give me and make me addicted to a new hobby.
Honestly, I thought that programming is boring and there is no place for creativity. I’ve never been so wrong in my life… This book and stories about all of those creative individuals mentioned by Isaacson opened my eyes. I felt immediately, that I want and need to be one of those people, to belong to software & web developers community.
Creative Process Management Reasearch
Even though I had not been interested in programming, digital technologies or… transistors before I started reading “The Innovators”, I wanted to know, how engineers and programmers created their inventions in terms of creative process management, which was my actual interest and hobby at the time (and still is).
I was expecting, that my theory of Creative Process Management (published in my ebook and on my blog) will apply also in web development, but I couldn’t imagine, how it is similar to rules I was using in my other creative activities.
I’ve understood, that there is no difference between painting, composing music, writing a book and programming in terms of Creative Process Management rules. In addition, I realized that by knowing how to program, I would be able to create applications that will help creators (including myself) to organize their creative process & notes!
I like intellectual challenges, so one of the reasons I’ve decided to test myself in programming learning were questions, I was curious to answer:
- Is programming really boring? (No!)
- Is there a place for creativity? (Oh, yes!)
- Are my intellectual abilities sufficient to master it? (Yes ;-)
- Can I manage this without previous programming experience? (With no doubt)
- Would my previous creative experience be useful? (A lot!)
- How far can I dive into it and how long it would take? (See my portfolio ;-)
A Few Apps Ideas
I wanted to learn programming to develop a few apps I had in mind before I wrote my first “Hello, World”:
- mind mapping app with the possibility to put keywords in any place on the canvas (done!)
- tagged notes app to organize and easy filter my notes (done!)
- my own blogging platform (done!) … and many other ideas – some of them are in progress, some of them are done.
Also I wanted to create my own personal web site from scratch with blogging features and a possibility to publish and run my apps. I had a website that was created by my friend Piotr, about whom I’ll talk more below, but… I wanted to run my apps on it and there were no possibilities to do that & I also wanted to be independent.
I’ve created a lot of things, music, articles, mind maps, sketchnotes, content and even wrote 2 ebooks, but… this time I wanted to create something that others could interact with, something that is usefull & available globally. That was a new stage of my creative self-development.
I Could Count on the Help & Guidance of a Friend
I have a friend, Piotr Wnuczek, and he is a programmer & web developer. He built a website for me, when I decided to run a blog in 2018, and without his help I couldn’t reach so many people with my articles.
So when I was thinking about the possibility of learning web development & programming, I had a living example that it is possible, because I was a witness of Piotr’s programming development and successes. So I considered, that if he could, I also can try.
Besides I knew, that there is a person, who can guide me and give me a hand if I need to. And I was right! We often see each other and every time I’m visiting him, he gives me precious advises.
Now you know almost everything about my reasons to freeze my previous long-term projects to make a space for undistracted programming learning. I'd never decide to dive into programming world, if there was only one of mentioned reasons - it all worked together. And I'm happy - that was a best decision in my life.
|
OPCFW_CODE
|
Excel VBA Collection Object
In VBA codingVBA CodingVBA code refers to a set of instructions written by the user in the Visual Basic Applications programming language on a Visual Basic Editor (VBE) to perform a specific task., apart from the existing collection of items under one group, we can create our collection groups. In our many articles, we have spoken about object variables, and in this tutorial, we will take through the VBA collection object in detail.
If you have gone through our previous article “VBA Arrays,” this will be much easier for you to understand. Arrays are used to group variables under one roof; similarly, Collection is also used to store a group of variables.
Collections are used to store objects. They are far more flexible than the VBA arrays, whereas arrays have fixed size limits, but readers don’t have any fixed size limit at any given point in time and even don’t require manual resizing.
VBA Collection is very similar to the “VBA Dictionary,” but the dictionary requires external object reference to be set it up under the object reference window. With VBA Dictionary, we need to set the reference type as “Microsoft Scripting Runtime,” but Collection doesn’t require any extra fittings.
How to Create Collection Object in VBA?
To get started with collection first, we need to declare the variable as “Collection.”
Sub Collection_Example() Dim Col As Collection End Sub
Since the collection is an object variable, we need to set the object reference by creating a new instance.
Sub Collection_Example() Dim Col As Collection Set Col = New Collection End Sub
Now with the variable, we can access all the methods of collection variable “Col.”
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Col. End Sub
Before making use of these methods, we need to declare a variable as a string.
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Dim ColResult As String End Sub
Now use the variable “Col” to choose the “Add” method.
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Col.Add End Sub
Under the Add method, we have specific parameters. Let’s assume we are storing mobile brand names with their average selling price in the market.
Under Item, the argument enters the price of the mobile.
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Col.Add Item:=15000, End Sub
Next, under Key argument, enter the mobile brand name.
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Col.Add Item:=15000, Key:="Redmi" End Sub
For the variable “ColResult,” we will store the result of the “Col” object variable.
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Col.Add Item:=15000, Key:="Redmi" ColResult = Col( End Sub
When you open the parenthesis of the variable “Col,” we can see the argument as Index. For this argument, we need to supply the critical argument value from the Collection add method, i.e., the name of the mobile brand.
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Col.Add Item:=15000, Key:="Redmi" ColResult = Col("Redmi") End Sub
Now let show the result in the message box in VBAMessage Box In VBAVBA MsgBox function is an output function which displays the generalized message provided by the developer. This statement has no arguments and the personalized messages in this function are written under the double quotes while for the values the variable reference is provided..
Sub Collection_Example() Dim Col As Collection Set Col = New Collection Col.Add Item:=15000, Key:="Redmi" ColResult = Col("Redmi") MsgBox ColResult End Sub
Ok, we are done when we run the code. We should see the price of the mobile brand, “Redmi.”
Better Understanding of Key & Item Parameters
I am sure it is not easy to understand the parameters of the Collection object. Let me explain to you a simple example.
Imagine you have a menu of fruits with their name and price of the fruits. Assume you are searching the “Apple” fruit price by the name of the fruit.
To search the price of the fruit, we need to mention the name of the fruit, i.e., in collection VBA language Name of the Fruit is Key, and the price of the fruit is “Item.”
It is like applying the VLOOKUP or HLOOKUP function, based on the lookup value, and we will fetch the required data from the database. Here lookup value is Key, and the result is Item.
Imagine you are a store manager in one of the retail stores, and you are responsible for managing the customer queries. One such customer query is an inquiry about the product price.
It would help if you allowed the customer to search the price of the product with complete information. Similarly, you need to show the message in case of no data found. Below is the example code which will present the input box in front of the user. They require to enter the name of the product they are looking for. If the product is there in the collection, it will show the price of the mentioned product, or else it will show the message as “The Product you are searching for doesn’t exist.”
Sub Collection_Example2() Dim ItemsCol As Collection Dim ColResult As String Set ItemsCol = New Collection ItemsCol.Add Key:="Apple", Item:=150 ItemsCol.Add Key:="Orange", Item:=75 ItemsCol.Add Key:="Water Melon", Item:=45 ItemsCol.Add Key:="Mush Millan", Item:=85 ItemsCol.Add Key:="Mango", Item:=65 ColResult = Application.InputBox(Prompt:="Please Enter the Fruit Name") If ItemsCol(ColResult) <> "" Then MsgBox "The Price of the Fruit " & ColResult & " is : " & ItemsCol(ColResult) Else MsgBox "Price of the Fruit You are Looking for Doesn't Exists in the Collection" End If End Sub
This has been a guide to VBA Collection. Here we learn how to create a VBA collection object and Advanced examples, and a downloadable excel template. Below are some useful excel articles related to VBA –
|
OPCFW_CODE
|
package mpi;
/**
* Struct class for {@link MPI#SHORT_INT} datatype.
*/
public final class ShortInt extends Struct
{
private final int sSize, iOff, iSize;
/**
* The struct object will be created only in MPI class.
* @see MPI#shortInt
*/
protected ShortInt(int shortSize, int intOff, int intSize)
{
sSize = shortSize;
iSize = intSize;
int sOff;
switch(sSize)
{
case 2: sOff = addShort(); break;
case 4: sOff = addInt(); break;
case 8: sOff = addLong(); break;
default: throw new AssertionError("Unsupported short size: "+ sSize);
}
assert sOff == 0;
setOffset(intOff);
switch(iSize)
{
case 4: iOff = addInt(); break;
case 8: iOff = addLong(); break;
default: throw new AssertionError("Unsupported int size: "+ iSize);
}
assert(intOff == iOff);
}
/**
* Creates a Data object.
* @return new Data object.
*/
@Override protected Data newData()
{
return new Data();
}
/**
* Class for reading/writing data in a struct stored in a byte buffer.
*/
public final class Data extends Struct.Data
{
/**
* Gets the short value.
* @return short value
*/
public short getValue()
{
switch(sSize)
{
case 2: return getShort(0);
case 4: return (short)getInt(0);
case 8: return (short)getLong(0);
default: throw new AssertionError();
}
}
/**
* Gets the int value.
* @return int value
*/
public int getIndex()
{
switch(iSize)
{
case 4: return getInt(iOff);
case 8: return (int)getLong(iOff);
default: throw new AssertionError();
}
}
/**
* Puts the short value.
* @param v short value
*/
public void putValue(short v)
{
switch(sSize)
{
case 2: putShort(0, v); break;
case 4: putInt(0, v); break;
case 8: putLong(0, v); break;
default: throw new AssertionError();
}
}
/**
* Puts the int value.
* @param v int value
*/
public void putIndex(int v)
{
switch(iSize)
{
case 4: putInt(iOff, v); break;
case 8: putLong(iOff, v); break;
default: throw new AssertionError();
}
}
} // Data
} // ShortInt
|
STACK_EDU
|
Statistics from Altmetric.com
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.
Chimeric antigen receptor (CAR)-T cells have revolutionized the treatment of relapsed/refractory hematological malignancies. Using CAR-T cell therapy, a high percentage of patients respond initially, yet up to 60% of patients eventually relapse. Predictive and prognostic factors allowing to predict long-term response are not yet fully established.
Liu et al recently introduced a computational model of CAR-T cell immunotherapy with the aim to predict late responses from early stage clinical data in patients with leukemia.1 By separately fitting a set of ordinary differential equations to different groups of patient data, for example, patients with complete response (CR), no response (NR), CD19+ relapse, and CD19− relapse, the authors estimated distinct model parameter sets for each outcome.
Computational models for the prediction of long-term response in CAR-T cell therapy can be helpful and various methodologies may be applied for their development. However, regardless of the methodology chosen, critical quality criteria should be met. We have several concerns regarding the approach presented by Liu et al, with the issues falling into the following areas: the data and methodology applied for the model’s development as well as communication of the model’s abilities and limitations.
The authors highlight the development of a computational model for the prediction of individual responses to CAR-T cell therapy as their goal. As with any model (computational or not), it is a key and intuitive rule that the data used to develop the model should be representative of the data it will be later used to predict. In their manuscript, the authors state that data from 209 patients were used for the model’s development. It is later specified that individual-level data were available for only 8.6% (18/209) of these patients and that for the remaining patients, the ‘individual data had been preprocessed and only statistical values such as medians were provided’. The authors justify this approach by claiming that statistical values could be regarded as representative individuals. Considering the high interindividual variability observed in CAR-T cell therapy, this assumption can clearly not be made. As the authors emphasize themselves, cell kinetics are highly variable between patients and, most importantly, often predictive of response. Therefore, using statistical summary parameters instead of individual-level data will dilute signals, create artificial, never observed data, and be overoptimistic in the prediction of individual trajectories. Methodologies allowing to simultaneously fit both individual and aggregate-level data in an unbiased way exist.2 However, such methods do not seem to have been applied here.
Liu et al state that they performed non-linear mixed-effects (NLME) modeling to estimate both population-level and individual-level data. While they describe estimating random effects to quantify the interindividual variability in the population parameters, no such variability estimates are provided. It therefore is unknown on which parameters interindividual variability was implemented and how high the estimated variabilities were. Moreover, relative standard errors or confidence intervals, which quantify the uncertainty in the parameter estimates, are not provided.
In general, while the authors applied a NLME model algorithm, their model development strategy does not follow established NLME modeling practices. The advantage of a NLME approach is the simultaneous quantification of general trends (using fixed-effects population parameters) and unexplained variability (using random-effects parameters) observed within the cohort. Once quantified, patient or treatment characteristics can be added to the model as covariates to explain parts of the unexplained interindividual variability. Once the impact of covariates on one or more model parameters has been estimated, these estimates can give new insights into the underlying reasons for the interindividual differences in the observed cell kinetics. This approach has been used in several excellent computational CAR-T cell models, whose review would be beyond the scope of this letter. The interested reader is referred to a recent review paper summarizing computational modeling approaches to support CAR-T cell clinical pharmacology strategies.3
In contrast to the established approach of estimating typical values, interindividual variability, and covariates for all patients simultaneously, Liu et al chose a different approach. Principally, they predicted early CAR T-cell response based on long-term outcome (responder status) rather than the other way around. By separately estimating a parameter set each for patients with CR, NR, CD19+ relapse, and CD19− relapse, the authors missed the opportunity to quantify the interindividual cell kinetic differences and then investigate covariates to explain the observed differences in one joint model. In addition, by dividing their dataset into four subsets, the authors decreased the size of the development dataset for each outcome even further. Concretely, individual patient numbers remaining for the parameter estimation were five in the CR dataset, four in the NR dataset, six in the CD19+ relapse dataset, and three in the CD19− relapse dataset. The additional statistical summary data increased the cohort sizes of the respective datasets, however, the value of statistical summary data for providing insight into individual cell kinetic and treatment outcome is low at best. Considering the large number of estimated fixed-effects (ie, typical) parameters (n=17) and an additional unreported number of random-effects parameters, it is unlikely that the model parameters were estimated with reliable precision for each of the four datasets. Finally, with their parameter estimates, the authors could successfully capture previously reported kinetic differences in patients with different outcomes (eg, patients with beneficial outcome usually have higher CAR-T cell expansion4 5), but no new insights into patient or treatment characteristics associated with each outcome were generated. Thus, while the model correctly captures kinetic differences in patients with different outcomes, it is currently of a largely descriptive nature.
To test the performance of the model on a ‘larger scale with higher reliability’, the authors generated virtual patient cohorts by sampling values from an assumed Gaussian distribution of the estimated population-level parameters. The predicted peak values and areas under the concentration-time curve in the first 28 days (AUC28s) for the virtual patients were then compared with the peak values and AUC28s in the respective model development datasets. Because the virtual patient parameters were generated from the model-estimated parameters, the similarity of the virtual patient cohorts with the observed data must be expected and cannot be viewed as a proof of the model’s generalizability. Generally, the simulation of concentration-time profiles in a virtual population capturing the interindividual variability is an often-used tool to better understand the expected exposure ranges in a real-life population. The model’s generalizability, however, would have been better addressed using external evaluation, that is, comparing predictions with observations of a new ‘real’ dataset, which would have not been used for the model’s development.
A crucial quality criterion for a developed model is its translatability into providing potentially helpful clinical information. Liu et al identified different parameter sets for patients with different outcomes and proposed several model-derived secondary parameters as factors for use in early prediction of late-stage response. The practical usefulness of most of the proposed factors, however, is low as no link is made between the prediction factors and patient or treatment characteristics. Therefore, should one want to use the developed computational model to make an early response prediction for a new patient by calculating the CAR-T cell function or negative relapse factor, this would require fitting the model to the patient’s individual cell kinetic data. However, for a computational model to potentially provide value for predicting clinical efficacy and inform researchers, its use should not be restricted to fellow modelers. Indeed, computational models for cancer immunotherapy are nowadays increasingly developed with a focus on clinical usability by providing simpler outputs in clinical language. An excellent overview on such mathematical models for personalized clinical translation is provided in a recent review article.6
The ‘response prediction factor’ is the authors’ only proposed model-derived prediction factor that could be calculated without fitting the model, and it is defined as the log2 of the product of the CAR-T cell peak value and the CAR-T cell area under the concentration-time curve in the first week after infusion. While certainly being accurate, the novelty of this factor is low, as the positive correlation between CAR-T cell exposure and beneficial outcome has been well established previously. Moreover, supported by an independent clinical analysis,5 we have identified a clinical composite score of CAR-T cell peak expansion normalized to the baseline tumor burden to be a better predictor for outcome than expansion alone.4
In general, while we are pleased that more computational research is being applied to the field of cell therapy, further development of any computational model (including our own) will need to include rigorous (re-)examination of modeling strategies, proper validation, as well as judicious and critical discussion of findings and their limitations before any of the developed models or model-derived prediction factors can be considered for use beyond research purposes.
Novel computational models for early response prediction in CAR-T cell treatment can be helpful, yet cell kinetic data for their development are sparse. Appropriate methodology is available to jointly analyze individual and statistical summary data in an unbiased way and can be used together with a non-linear mixed-effects or another appropriate modeling approach to generate new insights. To potentially have true value in cell therapy design, the output or a derived simplified output of the model should be presented in a format and language well recognized by researchers and clinicians.
Patient consent for publication
The authors thank Martin Bergstrand, PhD, E Niclas Jonsson, PhD, and Elodie Plan, PhD, of Pharmetheus AB (Uppsala, Sweden), and Professor Mats Karlsson, PhD, of Pharmetheus AB and Uppsala University (Uppsala, Sweden), for valuable input and discussions.
Twitter @anna_mclaughl, @tcellsrus
Contributors AMMcL and CY contributed equally.
Funding The authors have not declared a specific grant for this research from any funding agency in the public, commercial or not-for-profit sectors.
Competing interests CY is on scientific advisory boards in the field of cellular therapy but none that is directly involved in the use of computational models.
Provenance and peer review Not commissioned; externally peer reviewed.
|
OPCFW_CODE
|
import pickle
import shelve
import os
""" pickle_mule.py helper module for reading and writing pickled files.
functions:
push_dict - pushes a dictionary to a file that is pickled.
params: dictionary, string
return: True on success else raise error
fetch_dict - get contents of a pickled file.
params: string
return: dictionary or raises error
to_file - alias for push_dict
params: object, string
from_file - alias for fetch dict
to_shelf - pushes k/v data blob to a shelve
from_shelf - returns entire contents of a shelve as a dict
params: string
return: dictionary
update_shelf_member - sets value of single item in the shelve by key
params: string, string, object
get_shelf_member - fetches a single item from the shelve by key
params: string, string
return object
"""
def push_dict(dict, name):
"""pickle save the dict as name, returns True on success"""
try:
output = open(name, 'wb')
pickle.dump(dict, output)
output.close()
return True
except PicklingError as err:
raise err
def fetch_dict(name):
"""gets dictionary stored as pickle file name"""
try:
source_file = open(name, 'rb')
return_val = pickle.load(source_file)
source_file.close()
return return_val
except UnpicklingError as err:
raise err
"""new API, makes file/shelve relationship more semantically clear."""
def to_file(data, handle):
push_dict(data, handle)
def from_file(handle):
return fetch_dict(handle)
"""this function expects k/v data"""
def to_shelf(data, handle, obliterate=False):
try:
# wipe out the existing db
if obliterate:
os.remove(handle)
with shelve.open(handle) as db:
for k in data:
db[k] = data[k]
except Exception as e:
raise e
finally:
pass
"""dumps entire shelve db into a dict"""
def from_shelf(handle):
returnme = dict()
try:
with shelve.open(handle) as shelve_data:
for k in shelve_data:
returnme[k] = shelve_data[k]
except Exception as e:
raise e
finally:
pass
return returnme
"""updates a single shelve member by key"""
def update_shelf_member(shelve_handle, key, val):
try:
with shelve.open(shelve_handle) as db:
db[key] = val
except Exception as e:
raise e
finally:
pass
"""returns object stored at key"""
def get_shelf_member(shelve_handle, key):
return_val = None
try:
with shelve.open(shelve_handle) as db:
return_val = db[key]
except Exception as e:
raise e
return return_val
"""returns list of shelf keys"""
def get_shelf_keys(shelve_handle):
return_val = ()
try:
with shelve.open(shelve_handle) as db:
return_val = list(db.keys())
except Exception as e:
raise e
finally:
pass
return return_val
def main():
import os.path
# file repo
test_file = 'test_file.pkl'
# shelve repo
test_shelve = 'test_shelve'
#records
bob = dict(name='Bob Smith', age=42, pay=30000, job='dev')
sue = dict(name='Sue Sizzle', age=45, pay=40000, job='hdw')
tom = dict(name='Tom Lesion', age=50, pay=0, job=None)
shelve_data = {
'bob': bob,
'sue': sue,
'tom': tom
}
# to_file
# clear previous test
if os.path.isfile(test_file):
os.remove(test_file)
to_file(bob, test_file)
print('write file success:', os.path.isfile(test_file))
# from_file
print('file contents:', from_file(test_file))
# to_shelf
if os.path.isfile(test_shelve):
os.remove(test_shelve)
to_shelf(shelve_data, test_shelve)
print('shelve write success:', os.path.isfile(test_shelve))
# from_shelf
testme = from_shelf(test_shelve)
for k in testme:
print(k, testme[k], sep='=>')
# get_shelf_member:
key = list(testme.keys())[0]
val = get_shelf_member(test_shelve, key)
print('get item from shelve for:', key, val)
val['job'] = 'Fired'
update_shelf_member(test_shelve, key, val)
print('updated value for', key,
get_shelf_member(test_shelve,key))
#get_shelf_keys
keys_list = get_shelf_keys(test_shelve)
print('keys:', keys_list)
print('member from keys list:', get_shelf_member(test_shelve,keys_list[0]))
breakme = from_shelf('oops')
print('missing shelve test:', breakme)
# self test
if __name__ == '__main__':
main()
|
STACK_EDU
|
usbatm : modprobe & rmmod
duncan.sands at math.u-psud.fr
Mon Jan 24 15:23:33 EST 2005
> While adjusting cxacru2 to the new API, I've come across a few points
> I'd like to clarify:
> 1) ->bind, ->heavy_init, and ->unbind are all passed a struct
> usbatm_data which already has a valid reference to the usb_intf.
Since the usbatm instance and the interface are the main bits of data
these methods will work on, it seemed polite to pass them both. These
methods could also retrieve the usbatm_data from the interface (well OK,
they can't, but only because I put the data into the interface after the
call to bind returns), so you could also ask: why pass the usbatm_data?
Also, I'm still undecided as to whether usbatm_data should be "opaque":
which fields should be exposed to mini-drivers? Should they be allowed
to extract usb_intf from the usbatm_data? While waiting to get a better
feeling for this, I've chosen to pass in the interface also. Another
reason is that I only added the usb_intf field recently - maybe I will
remove it later. If methods used the passed parameter then they won't
need to change if the field is removed. Anyway, as you can see this is
mostly a stylistic point, so doesn't matter much one way or the other.
> Does it need to be present in the function parameters?
> 2) what's the purpose of ->atm_stop? I can't imagine needing this given
> it is immediately followed by ->unbind.
It is only called if atm_start was called (and succeeded). If heavy_init
fails or the modem is unplugged before heavy_init finishes, then atm_start
is not called, and then neither is atm_stop. If heavy_init is not needed by
a driver, then atm_start and atm_stop are less useful, but still a little bit
useful: (A) the ATM device is only available when atm_start is called, not
when bind is called, so how to set up the device in bind? [*]; (B) the logical
separation of USB initialisation code and ATM initialisation code may make for
> 3) usbatm_heavy_init is always called if ->heavy_init is available, even
> if the modem is already initialized. In this case creating a kthread
> just to detect it and exit seems a bit of an overkill. How about
> exporting usbatm_heavy_init to be called by the minidriver in ->bind,
> if needed?
> Alternatively, ->heavy_init can be made to
> take a parameter saying if we really want it heavy. Then it's first
> called synchronously in the light version, and if it fails with, say,
> -ETRYAGAIN, spawn a kthread with the heavy version.
It is overkill, but who cares? Well OK, I care a bit and I see you do too.
I will think about your suggestions. Another possibility is to add a method
need_heavy_init which is called to determine whether heavy initialisation is
needed or not.
> 4) I have a couple of delays in cxacru_upload_firmware for the modem to
> digest the commands I'm sending. The delays are rather big:
> msleep(1000) and msleep(4000). How can I interrupt these, so that
> kthread_stop in usbatm_disconnect doesn't wait too long?
It is quite new.
[*] The code could be reorganised so that the ATM device is available in
bind if heavy_init does not exist. But I would rather have this kind of
thing be the same whether or not heavy_init exists.
More information about the Usbatm
|
OPCFW_CODE
|
Android Development- Tech and Technology
|Android Development Apps|
| Developers can create Apps for the Android mobile platform using the Android SDK. |
The Android applications are created with the Java programming language. The Android SDK is the main tool that provides the software libraries and APIs necessary to start developing applications on the Android platform.
The Android SDK is called a modern rich development environment that includes a device emulator, several tools for debugging, performance profiling, memory and other tools for Eclipse.
Android developers must create the source code using the Java language, that linked with the Google's Java libraries to produce the Android Apps.
Android was built from the ground-up to enable developers to state of the art applications that take full advantage of all the facilities for a mobile device or tablet.
There are thousands of android based operating system in development by companies and enthusiasts colaborating trought the internet around the world.
||Follow the links and recommendations to create apps and games for Android: |
|Android Application Fundamentals
||Android Videos for developers
||Android Development Features
||The Android SDK is the main tool that provides access to the APIs and libraries necessary to developing applications on the Android platform. Android Application framework is based in the Java programming language, enabling reuse and replacement of components Uses a Dalvik virtual machine optimized for mobile devices The free Plugin for the Eclipse IDE provides a rich development environment with a device emulator, tools for debugging, memory and performance profiling Media support for common audio, video, and still image formats (MPEG4, H.264, MP3, AAC, AMR, JPG, PNG, GIF) Optimized graphics powered by a custom 2D graphics library; 3D graphics based on the OpenGL ES 1.0 specification (hardware acceleration optional) SQLite for databases and structured data storage Integrated browser based on the open source WebKit engine
||Android hardware features
The Android operating system, including the Linux kernel, consists of roughly 12 million lines of code including 3 million lines of XML, 2.8 million lines of C, 2.1 million lines of Java, and 1.75 million lines of C++.
|Android - Latest Android for Mobile Phones. Android News.
||Android Development Apps Software Development - How to create Android Applications -
||Android Books Documentation and Resources
|| Android Play Market - Apps - How to install and configure applications in Android?
Technology Android Apps 2022
Axios to AndroMoney: The top Android apps to help you track expenses The Indian Express
Beginner's guide to Android updates and why they're crucial The Indian Express
Google Play streamlines policies around kids’ apps as regulations tighten TechCrunch
Forest to Pomodoro Timer: The top free Android apps to help manage your time better The Indian Express
|
OPCFW_CODE
|
M: Ask HN: The DAO hack - Where does the real money go? - winteriscoming
I am a complete newbie when it comes to crypto currency and the current DAO hack is something that I haven't yet fully understood. From what I have read so far, someone hacked TheDAO which is a crypto currency based program and made away with some amounts of "ether" which then translate to X amounts of real money ($$).<p>How does the culprit encash that amount or rather where does it really end up? In some (real money) account(s)? Or does it stay as crypto currency which can then be used to buy stuff (online)? And there's no way to track how a particular account/address ended up with that much worth of currency?
R: cprecioso
This was answered in the comments of the link sent previously, but basically:
it's gone to a public blockchain address, of which everyone knows the public
key but only its owner knows the private one. Also, the blockchain is
pseudonymous, so nobody knows who the owner of the address is.
Therefore, everyone knows where the ether are, but only its owner can touch
it. He can use them as if they were legitimately obtained (they are, depending
on perspective) and go to an exchange to buy dollars (or euros or whatever),
or exchange them with other addresses for goods and services.
Anyone will be able to audit these movements of money (at least the ones that
stay inside the ether blockchain), although there exist some cryptocurrency
laundering schemes.
R: winteriscoming
Thank you.
R: max_
Here is what exactly happened
[https://www.reddit.com/r/ethereum/comments/4os7l5/the_big_th...](https://www.reddit.com/r/ethereum/comments/4os7l5/the_big_thedao_heist_faq/)
R: winteriscoming
I don't claim to be the smartest of software developers, but I have been
around in software development for more than a decade and I have been able to
grasp most of the technical articles that I read out of curiosity. But I have
to admit, the past few days I have read a bunch of articles around this,
including this one, which try to make it simpler to understand, but I still
don't have a grasp of the basic concepts around this :) Makes me wonder, are
there enough people out there, who really know how this works and how to
develop for such applications.
R: max_
I hope this podcast helps...[https://letstalkbitcoin.com/blog/post/lets-talk-
bitcoin-297-...](https://letstalkbitcoin.com/blog/post/lets-talk-
bitcoin-297-the-death-of-thedao-part-one)
please let me know if it does/doesn't
R: thrill
I don't believe any gold was involved.
R: vorotato
lol if you think that gold has any real value
R: hanniabu
If you can sell it,then it has value
R: wallace_f
To be pedantic, economists use the word value in a specific way. It is how
much you can gain or benefit from something.
Price is what you can sell something for.
So gold has both value (you can use it as a commodity), and it has a price
(you can sell it in exchange for currency).
Worth is another one: it's a mostly subjective, philosophical judgement.
So technically, I could sell something that has no value; and something which
has no agreed upon price can have value. Everyone can then disagree about what
that something is worth.
|
HACKER_NEWS
|
MS SQL Server 2016 Oprimal Disk Configuration
New server build for SQL SERVER STANDARD 2016 install. Total of 12 disks: 6 standard 15K HDs and 6 SSDs. This is the disk setup I am going for.
Could this setup be improved on, if so how what would change?
I could move TEMP DB mdf to DATA Drive and have TEMP DB log with other DB Logs on SSD LOG drive.
This will be for a SQL Server Production environment only.
Optimizing for speed but also HA to an extent.
Backups will get copied to an offsite location throughout the day.
Server will be mirrored not AG. Although mirroring is deprecated, due to install being Standard Edition I don't see benefits of AG at this point with 2 servers.
Biggest waits are LATCH_EX followed by CX_PACKET
Also looking to replicate some data for reporting purposes.
LOG average file size is 9.16 MB over past 6 months, Log backup every 15 mins
MAX size has been 389.47MB in past 6 months
TEMP DB info for past 6 months
Looks pretty good. What is the latency of this configuration for the tempdb / log drive? This is where your main I/O bottleneck will be, since logs are sequential writes and tempdb will have random writes.
Don't know the latency as i haven't built it yet :), current setup is RAID 5 for everthing across 4 disks and we get average read latency of 15ms write of 35 ms. I was thinking I could move TEMP DB mdf files (6 of them) to the DATA Drive and have TEMP DB logs with other DB Logs on SSD LOG drive? Probably do the same for the other system DBs?
Yeah, like I said it's pretty good. Anything will be better than RAID 5 for writes. Your target latency is 10-15ms for writes on log (preferably better if you're sharing). Put tempdb on the fastest disks you have, and do the same with logs. It's what you have here, pretty much.
To allow for better answers from others here are some questions: Is this server PROD only or does it need to do other workloads (DEV, TEST, QA, or non-SQL)? Are you optimizing for speed, or for availability? Will the backups get moved off of the server disks (RAID is not a backup). Do you have spare HDDs and SSDs or are you crossing fingers and hoping for the vendor to pull through in a moment of need? Do you know if your application usage is temp data, temp log, user data, or user log heavy? Is the server in an AG?
added some more info to the question, thanks
I would suggest to separate Log and TempDB of different disks. If your database would be used for OLTP & OLAP both, you may face WRITELOG wait, when large data shorting is on process for reports.
This server wont be used for OLAP, only OLTP
Yet I would prefer to separate these 2, because when you will be going to perform any Index maintenance operation with SHOR In TEMPDB option, the operation will affect each other.
I also would seperate TempDB from TLogs. These are two of your most heavily used parts and to have them on the same set of disks, this will be your bottleneck.
@JasonB. Would you move logs with data then?
Looks fine to me.
Because of the way that SQL Server schedules writes, keeping logs and data on separate volumes doesn't save as save as much performance as many think for many workloads (it does have a large effect on some workloads though, so YMMV), and I expect this will be less still on SSDs (as you don't have the latency of head movements flipping between the log and data files to worry about) so if you expect a lot of tempdb activity you might be better off keeping the logs with the data files and leaving tempdb on its own to keep the IO bandwidth of that array to itself.
Of course this is very dependant on actual application workloads so my gut thought above could be completely wrong for your DB(s), so if possible benchmark before committing to anything in production.
Also Cody's comment regarding spare drives is critical: make sure there are some on-site or very nearby. From a HA standpoint mirroring will protect you to an extent should a whole array fail (without mirroring take "or nearby" from the above and make sure there are definitely spares on-site and ready to go) but I'd still feel safer knowing an array can start rebuilding ASAP after a single drive failure to reduce the chance of a dual drive failure. Also check with your infrastructure provider to make sure the machine supports hot-swap (or if not set your recovery plans accordingly to take into account entirely stopping one mirror temporarily while a replacement is installed).
I have read as Randolph mentions that logs are sequential and data is random so best to keep them separate but as these are SSD like you say is it necessary?Also read to seperate them out in case of data drive failure you have the Log file. Currently temp DB isn't that big 3GB with 99% available. Main DB is 110GB. Maybe I should put logs together with data including system DBs etc and chuck tempDB on its own SSD.
Will ensure infrastructure guys have a disk on standby.
Maybe I should split data out. rather than RAID 10 have it in RAID 1. So have TEMP DB RAID 1, DATA RAID 1, LOG RAID 1 ?
If you put TempDB on its own single disk remember that this becomes a single point of failure. That TempDB is very large and currently empty doesn't always mean much: you could have regular processes that churn it considerably so that up-to-3Gb sees a lot of activity. I suggest you watch that activity over a period to gauge its size and burstiness. If it turns out you barely use TempDB apart from in large overnight reports it may not be an issue needing a lot of thought at all even the sharing od IO bandwidth with other files.
Regarding R1 for data and R1 for logs: you might hit the compromise point of wasting storage here as you may end up with full data volume and a nearly empty log volume with no easy way to use the spare space to let the data grow. Note that IO activity on the logs is only going to be during writes - if your application is very read heavy then this may not pose much IO load so the extra bandwidth of R10 for the data reads will be more beneficial than separating the log/data IO payloads. Again, monitoring IO activity during normal operation is going to be a useful way of gaining insight.
Appreciate all the feedback its awesome. I thought general consensus is to put temp DB on its own disk(s)? If its in RAID 1 and with mirroring I think its enough cover in case of failure. I think the Temp DB growth was probably due to some data extraction/report that happened ages ago. I've included a graph in original question generated from the data I've captured of its growth for past 6 months. I'm leaning towards shared LOG and DATA and separate TEMP DB..Hmmm also thinking of doing 6 SSD drives in RAID 10 ... this is so difficult to choose
|
STACK_EXCHANGE
|
This releases contains various wallet-related and some service node diagnostic improvements related to uptime proofs. It is recommended for cli wallet users and new service nodes, and optional for existing service nodes.
- Fix view-only wallet key image importing bug that can occur when the wallet first sees blinks out of the order they eventually get added to the blockchain.
- Fix mac builds to compile for more generic CPUs and lower macos compilation compatibility level to target 10.12 and later.
- Fix a wallet syncing bug on older macos versions.
- Fix an occasional POST request error in the embedded HTTP client.
- Signficantly reduce wallet CPU usage by eliminating some busy loops from the polling thread (particularly when doing long operations like building large transactions).
- Significantly improve sweep_all speed in the cli wallet, particularly for large transactions on a remote lokid.
- Fix "failed to get random outs" issue for cli and gui wallets when using public RPC nodes.
- Add wallet rpc endpoint for remembering known lokinet names (to make the GUI wallet remember looked up LNS names).
Service Node changes:
- Fix lokid refusing to send uptime proofs when reusing a key_ed25519 file from an expired registration without using the same primary key (for example, when converting an expired service node to run as a new registration in 8.x ed25519-only mode).
- Add a loud warning in the logs if lokid detects another SN on the network broadcasting uptime proofs with the same public IP and ports.
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA512 The following hashes have been signed for verification using keejef's GPG keys located in the loki core repository at https://github.com/loki-project/loki-core/blob/master/utils/gpg_keys/KeeJef.asc. SHA256 Hashes loki-linux-x86_64-8.1.2.tar.xz c5e5b0e4892a06f064f00e03c37f68b8336612054634296bd9f89986b93b2d17 loki-macos-8.1.2.tar.xz f2c8cb789c4d6d0cfd661d100990d478074de5479c962a6146a2d34cdb827913 loki-win-x64-8.1.2.zip a2dec71b93d1b1bd83b35be0ce6371a656c6840e0577fed603b41a5a42ab3e98 -----BEGIN PGP SIGNATURE----- Version: GnuPG v1 iQEcBAEBCgAGBQJfj4TkAAoJEKyXFIlc5F1VBM8H/jVc4gCmMIcYjgjHd+jVi8wp Oq90kmo53uBAFFd54PUSYZdar1//P5vgLHP46FHnqqJ1HkrK1668zGu/begSzyIV lAtdkVJnu6OkwU9lr6qUIP4hmtTbCsh/jjtqiX7czYTNtjREyXJUKdPAs6rIncm8 fLu/3LJwlpTK2a8c5HoDT3JgkD5YdwSbWuobKdbEiMA7RRnxS5nS+JuTmolg76mw ASBDD0XEW6ZYtbcv41pfz3wvZhYdRdCQlj3L3PlQ8DE/waaDQideaEeJoWuoXJWN wpaTJijdIs81Tr+DEV3IaWDTkpsLMuSYEZaWbyahrk29TSWxDBiel36x10Njb9Q= =O8KI -----END PGP SIGNATURE-----
|
OPCFW_CODE
|
Poker hands combination
In a poker game where we have a standard deck ($52$ cards) and each player is dealt $5$ cards, how many hands of each of the following types are possible ?
(Type 1, "Low face") one face card and $4$ other cards less than $6$.
(Type 2, "Wedding") Exactly two face cards and exactly one diamond.
Here is my approach for Type 1
we have $3$ different face cards for each color, namely (King,Queen and jack) and since we have $4$ different colors, we have $12$ different face cards in the deck.
Now we are choosing one of those $12$ cards so we have $12\choose 1$ but also for the cards that are less than $6$ we have $5,4,3,2$
I don't include the Ace here because the Ace is bigger than $6$ right ? Now again we have those 4 cards in each color, so we have $16$ cards less than $6$ in total and we choose $4$ cards of them so we have $16\choose4$ and so the answer should be $${12\choose1} \times {16\choose 4}$$
is that correct ?
For Type2,
we have exactly two face cards and so we have $12\choose2$ and we have exactly one diamond to choose, However, we have to subtract all the face cards diamond, so we have $13-3 = 10$ diamonds card to choose and so we have $10\choose1$, now we have $2$ other cards, but non of them must be a face nor a diamond so we have $52-12-10 = 30$ remaining cards to choose from, which is $30\choose2$
And so my answer would be $${12\choose2} \times {10\choose1} \times {30\choose2}$$
I just want to make sure I am on the right path here.
both asnwers look good
@DannyC. hmmm, so what Can I do about that ?
So for case 1 we would have ${9 \choose 1} \times {30 \choose 3}$ and for case 2 we would have ${12 \choose 2} \times {10 \choose 1} \times {30 \choose 3}$ and then we add those two cases ? @DannyC.
Taking ace as a high card,
Type 1 ans is ok.
For Type 2,
either 1 non-diamond face card, a diamond face card , and 3 non-diamond non-face
or
2 non-diamond face cards, 1 diamond non-face card, and 2 non-diamond non-face
$$ = {9\choose1}{3\choose 1}{30\choose3} + {9\choose2}{10\choose1}{30\choose2}$$
For case 1, you already have a non-diamond face, why are u still choosing one face card of the remanining 3 ? $3 \choose 1$ ?
case 2 looks right for me
Because you get a required diamond through a face card.
|
STACK_EXCHANGE
|
While I'm prepping for the Kentico Conference in Denver this week, I thought I'd write up a quick post on a topic I got asked about in Kentico's Slack chat community.
Splitting up a shopping cart so that you can process parts of it separately is useful any time you want to allow the user to finalize their purchase of one product or group of products, without having to also process the whole cart. This may be if you have an 'ecommerce cart' along with event bookings along with donations, and maybe you want the donations to be a recurring order but the rest of your cart certainly isn't recurring.
Or maybe you have sub-sections of your site that you don't want to share transactions with because the payments go to different places.
You can, of course, just ignore the Cart entirely, and do donations / bookings or these side sections as a form that connects to your payment processor directly, but you lose out on a lot of Kentico handling by doing that - tax classes, shipping options, product options and variants, order items, and payment statuses are all things you might need to 'reinvent' if you aren't using Kentico e-commerce.
So, when I ran into this issue on my end, where a client wanted two separate checkout processes that appeared to the user to be separate, I went back to the basics of what a shopping cart was - just a storage carrier for items and order information, until they become an order. Everything else - the shopping cart content web part, order subtotals/totals/etc, coupons, even the cart manager that eventually turns the whole thing into an order... are just code.
If you just need to be able to purchase a single webinar, or do a donation, outside of the workflow, I wouldn't go this route. Just go with a bizform and use Accept.js (or whatever equivalent client-side submission your payment provider supplies) and a global event handler to submit the payment. The form acts as your data record, and if you can get away with not doing more, don't do more.
If you're doing your project in MVC, you're actually 99% of the way there already: Kentico isn't really customizing your experience in any case - you're turning products into orders yourself, and can show or hide whatever products you want. Go ahead and do step 1, then the rest is in your hands - show the products you want, show the totals you want, then create your order containing the products you want, and remove them from the cart as they're ready. It's not that different from one cart.
For those of you on Kentico 11, or 12 Portal Engine, you're probably working with web parts and the standard workflow though... so, here's the process I used.
Part 1: Split your data and Design your process
Maybe you're going to split products based on page type. Maybe you're splitting by Department, or Product Type (donation/e-product/product), or you've added a custom field to each product that determines which 'bucket' it goes in. Either way, assigning that data to your products is step 1. I'll refer to this as the 'split factor' later on - the thing on which you're splitting up your content into buckets.
At the same time, I would consider your shopping cart experience. If you can set it up so that the user can't get into multiple 'checkout steps' at the same time, that's ideal - it lets you use functionality like coupons and location-based shipping costs. By this I mean, if they enter and start doing one order but then mid-checkout go start doing another sub-cart, you're going to have to reconcile that, probably by storing the data for each cart separately - and that becomes a lot of extra data to store.
Single-page checkout processes work well for this - either they've completed a sub-cart, or they haven't started processing it at all. There's no data being saved that they expect to come back to, if they abandon the sub-cart.
Part 2: Custom Shopping Cart Content controls
Go ahead and make a copy of the standard Shopping Cart Control web part, and add a filter on your split factor, that you can set in the web part properties. You'll use this when you want to show your individual 'carts'.
Part 3: Custom other controls
With having done that, the next step is very similar - anything related to your eventual 'order': totals, subtotals, discount totals, all need to be separately calculated based on your sub-cart. This is just a matter of doing the math based on the items in your sub-cart, rather than relying on the Kentico totals fields.
If you need to actually store different *coupons* per cart, I would update the Shopping Cart module class to add a custom 'coupon code' field for each sub-cart - and you should just remove the default coupon code from any of your processes, as you don't want it inadvertently being mixed into your other products' data. We'll add in the 'correct' coupon code at the last minute, in step 4.
Same goes for if you have to store separate addresses for location-based shipping costs.
Step 4: Hidden processing page
The normal shopping cart process, as the last step, closes out your cart, and converts the whole thing into an order. Instead of customizing the step manager code, what I like to do is insert a hidden 'extra' page as a final step in the checkout, and call it 'processing'. This page will use a completely blank template, not even using a master page - all it will have on it is a 'SubCartProcessing' web part. This web part has nothing in the ascx, but is what will do the gruntwork of this solution: it will look at the cart, know which subcart we're processing,
So, the user hits 'pay now' or whatever you've set the 'next' button to look like in this case, and get ShoppingCart.CartContentItems, filtering based on your split factor.
At this point, create a new order, and fill it in using the Orders API: https://docs.kentico.com/api10/e-commerce/orders#Orders-Creatinganeworder - You should be adding in the addresses, the coupon code if any, the order items, and setting the order status, at this time. Call Update and Evaluate on the cart, and raise the SHOPPING_CART_CHANGED event, just to make sure the updates are saved.
Assuming we succeed, we know we've now 'completed' this part of the cart, so we should manually (shoppingCartInfoProvider) Remove the individual shopping cart items and delete them (you have to do both to each cart item!) and, if there are no items left in the cart at all, delete the entire shopping cart... If there are, we should at least clear any shopping cart parameters related to this sub-cart - coupon codes or other custom fields.
This web part, as the last thing it should do, is redirect you, always. We don't ever want the user to know this page exists, so if we've created an order, we want to go to a payment page (don't forget the o= querystring parameter). If we've not created an order due to an error, we want to redirect to an 'error, order not processed' page (any user-fixable errors should have been resolved back during validation on the prior pages), and if we're just suddenly on this page without an order, just redirect back to the cart. The point is, this page is just a processor, not a page with visual elements. With these server-side redirects, this page will never hit the client's machine.
Step 5: Payment
Since we're now dealing with an order containing only the items we care about, the Payment page, and anything after that, can go by the standard Kentico process, and should be individually visible . Congrats!
|
OPCFW_CODE
|
The work of an annotation service is to summarize explained about whole chapter or document is called an annotation. A note for the readers helps them remember the last read. The meaning of Annotation is to make a note, which is evident in its name (an- NOTE – Tate). Annotation requires practice. As much you dig as you get perfect, you will get adept at understanding the subtle article. Make Annotation requires a strategy.
Technically, Annotation means to tell the machine about the object by labeling it.
It’s like a man understands things to give the correct output in anything. Same as it is for machines. When machines label them, the device works properly according to tagged and gives the exact result.
What is the difference between Data Annotation and Image annotation?
And both annotations, the Data annotation and Image Annotation are not so much different. Still, let’s see what made them different.
Data Annotation plays a vital role in making the machines self-aware machines. Data annotation works to feed the data into a computer or any other device. So, the machine recognizes the data quickly and operates correctly. So, when a human being asks them a query to the engine, it perfectly answers it. Data Annotation service is a game of supervising the machine.
Or in other words, data annotation works to annotate the data inside the machine. It could be in any form in images, video, text, etc., and make sure the device recognizes the input data according to the particular job.
For example: – when any query is asked to Google. Google presents us with an accurate answer to the question. How? It is all because of Data Annotation through data annotation trains the machine or AI. So that the visitor gets the correct answer; besides, there are many algorithms and coding language work. Effectively.
On the other hand, Image Annotation is labeling digital images and pictures. Although image Annotation is a part of
Or you can. Annotation means to tell the machine about the object.
It’s like a man understands things to give the correct output in anything, the same as for devices. In engines, label them so the machine works properly according to tagging them and provides an exact result.
On the other hand, Image Annotation works digital label images. However, image Annotation is a part of data annotation. In image annotation, the pictures are classified for machine recognition or machine learning. For it, there are many annotation tools and texts used. Doing image annotation means merging metadata into the database. Image Annotation works that image fits into machine algorithms, which is very important for any AI project. The purpose of Image Annotation is that the computer well recognizes the object in the database. So, the machine recognized adequately the same as human vision.
Simple Image Annotation is also known as tagging, transcribing, and processing. In Image annotation, the image is labeled using text. For example- a picture of a house marked a castle of the dream.
In complex image annotation, add more images’ dimensions to label or some unique picture features. Like classifying two objects of the same category. Or some other things and boundaries.
Data Annotation and image annotation, the difference is that data annotation works for whole types of content like text, video, images, etc. but Image annotation works only on images classified.
|
OPCFW_CODE
|
|Aug 7, 2007, 11:33 AM||#3|
Objective C's Foundation Framework has classes that do the same work with STL's classes.
For example, look at the C++ - ObjC equivalents:
strings -> NSString
Vectors ->NSArray and NSMutableArray
Maps -> NSDictionary
Now, queues, stacks, etc are derivatives of vector, and in ObjC will not have a significant performance advantage of being developed, since NSArray already stores elements in a dynamic manner, using pointers. And it accepts any derivative of NSObject. There are some fundamental differences, though, for example in an NSArray in location 1 you can put an NSString, but in location 2, you can put an NSDictionary.
If you want to use the STL, you can use C++ within ObjC. But let me give you some advice. Use C++ with ObjC only if you have to. Even if you want to port a program from C++ to ObjC, things can be really messy and you could end up writing spaghetti code if you don't know what you're doing. ObjC more than matches the capabilities of C++. Use C++ only for accessing low-level mechanics of an application.
|Aug 7, 2007, 01:16 PM||#4|
FYI, Soulstorm, adts (or ADTs) are "Abstract Data Types" which you've correctly guessed the meaning of.
Objective-C doesn't have templates (I don't have direct knowledge of what's on tap for Objective-C 2.0, but I don't believe any kind of generics is being added).
Objective-C/Cocoa uses a single-root style of object oriented framework. Collection classes, therefore, contain pointers to objects. Cocoa has various collection classes and gobs of other useful objects built in.
It's not quite as efficient as templated container classes can be, and it's not as type-safe. On the other hand, Cocoa is a really nice environment to develop in.
Anyway, Objective-C really takes a more dynamic, run-timey approach to objects than C++, which is counter to the static, compile time nature of templates. To get the full benefit of templates in Objective-C you'd have to lose the dynamic nature of objects used with them. And If you keep the full dynamic nature of Objective-C objects, you'd have to lose most of the benefits of templates.
As someone who came from a hard-core C++ background to Objective-C, Cocoa/Objective-C has really won me over. As a developer I think it gives you a lot more than you lose. I think a lot of it has to do with how nice the Cocoa framework is, really.
|Aug 7, 2007, 02:06 PM||#5|
To all that, I want to add that C++ will never be replaced by ObjC and vise versa.
C++ is for closed world situations (like designing the engine compartment of a car) and ObjC is for open world situations (like designing the interior of the car and its behavior on the road). And, there are some features of OS X that you can't touch without using C++, like some Core Foundation features. I know, those features are also accessible using C, but who uses C when he knows how to use C++?
However, ObjC is really the best choice for OS X development. The ease, power and flexibility that Cocoa provides are unmatched.
|Aug 8, 2007, 04:43 PM||#6|
on that note then what would be the best book to purchase to teach me the best ways to use the cocoa framework. Im very confident using c++ and obj-c (i read though a rather lengthy tutorial on it already and its obviously very similar to c++) so i'd like a book centered on just the tools and frameworks osx provides to build applications.
|Thread Tools||Search this Thread|
|thread||Thread Starter||Forum||Replies||Last Post|
|I Need Help Learning Objective-c!!||Woodshark69||Mac Programming||11||Dec 14, 2012 07:21 AM|
|Objective C Accessors||ag2web||Mac Programming||9||Nov 18, 2012 02:53 PM|
|Problems with Objective-C++?||frank100||Mac Programming||5||Aug 8, 2012 04:49 PM|
|Testing objective-c||Evre||iPhone/iPad Programming||21||Jul 12, 2012 07:22 AM|
|Objective C Syntax||nashyo||iPhone/iPad Programming||6||Jul 4, 2012 11:56 AM|
All times are GMT -5. The time now is 05:43 AM.
|
OPCFW_CODE
|
Since some weeks I try to upload some healing blocks into stable network, on average the FEC-block size is 256 KB. Normally I use FUQID 1.3, but because it transfered < 0.1 KB/s and reported a lot of "Insert thread failed. Retrying..." in the log, I also tried it directly with FProxy with the same result.
When I set the Insertion-Thread-Number to a higher value than 5 the "Current messageSendTimeRequest" value in "General Information" of the web interface gets greater than 200%-300%. The maximum was 10000%-15000% when I set the thread number to 30. The "Transfers active" value grows also from ~20-30 to ~180. Fproxy mostly reported "ROUTE NOT FOUND" on blocks where inserting was failed. Also the "Instantaneous local traffic" decreases significately.
Heal HTL: 15 delete key from local node vefore inserting is checked
The upstream bandwith usage is ~ 10000 - 12500 B/sec. (12000 Maximum) Typically my node is running for periods of 3 - 10 h a day.
Plattform: WinXP Pro JVM: 1.4.2_03-b02 Freenet: Fred,0.5,STABLE-1.50,5069,CVS:22.214.171.124.2.96
Detailed System Information: ---------------------------- Plattform: WinXP Pro, Pentium 3, 933 MHz, 512 MB RAM JVM: Sun, 1.4.2_03-b02 Freenet: Fred,0.5,STABLE-1.50,5069,CVS:126.96.36.199.2.96 Internet Connection: DSL: 768 kbit downstream / 128 kbit upstream Data Store: 7 GiB, Percent used: 95 Mean Ustream Traffic: 10000 - 12500 bytes/second unstable.ref: 2004-02-15 Typically number of connections: 100 - 180
diff default.ini/freenet.ini: ipAddress=XXX listenPort=XXX fcpHosts=127.0.0.1,localhost doAnnounce=yes storeSize=7168M inputBandwidthLimit=70000 outputBandwidthLimit=12000 logLevel=Normal rtMaxRefs=51 rtMaxNodes=51 maxHopsToLive=25 maximumThreads=130 tempDir=XXX mainport.params.servlet.1.params.passThroughMimeTypes=text/plain,image/jpeg,image/gif,image/png
# Some Statistics # # Node ran 1 h before I started an upload with 30 Threads # now it runs ~3 h # ------ Routing Table status: 18.02.2004 11:27:28
Number of node references 51 Attempted to contact node references 17 Contacted node references 48 Connections with Successful Transfers 21 Backed off nodes 17 Connection Attempts 1646 Successful Connections 1527 Lowest max estimated search time 0ms Lowest max estimated DNF time 0ms Lowest global search time estimate 289962ms Highest global search time estimate 460245ms Lowest global transfer rate estimate 2.318 bytes/second Highest global transfer rate estimate 3.643 bytes/second Lowest one hop probability of DNF 0,98 Highest one hop probability of DNF 0,99 Lowest one hop probability of transfer failure 0,79 Highest one hop probability of transfer failure 0,92 Single hop probability of QueryRejected 0.29645163211077075 Single hop average time for QueryRejected 87863.09895443736 Single hop probability of early timeout 0.6783154528702626 Single hop average time for early timeout 24009.760728425113 Single hop probability of search timeout 0.9186481377761555 Single hop average time for search timeout 316643.7063678293 Total number of requests that didn't QR 23767 Total number of reqests that timed out before a QR or Accepted 6676 Implementation freenet.node.rt.NGRoutingTable #... # #-- Load # Current routingTime 0ms Current messageSendTimeRequest 7220ms Pooled threads running jobs 5 (3,8%) Pooled threads which are idle 25 Current upstream bandwidth usage 11215 bytes/second (93,5%) Reason for refusing connections: avgMessageSendTimeRequest(7220,250) > successfulSendTimeCutoff(2000,000) Reason for QueryRejecting requests: Estimated load (100%) > overloadHigh (80%)Estimated load (100%) > overloadHigh (80%) It's normal for the node to sometimes reject connections or requests for a limited period. If you're seeing rejections continuously the node is overloaded or something is wrong (i.e. a bug). Current estimated load for QueryReject purposes 100% Current estimated load for rate limiting 722% [Rejecting incoming connections and requests!] Reason for load: Load due to thread limit = 3,8% Load due to routingTime = 0% = 0% / 100.000% <= overloadLow (50%) Load due to messageSendTimeRequest = 722% = 722.025% / 100.000% > overloadLow (50%) Load due to output bandwidth limiting = 93,5% because outputBytes(672911) > limit (576000,009 ) = outLimitCutoff (0,8) * outputBandwidthLimit (12000) * 60 Estimated external pSearchFailed (based only on QueryRejections due to load): 0.9999999999999998 Current estimated requests per hour (based on last 10 mins): 5772.0 Current global quota (requests per hour): 799.4182883301385 Current recommended request interval sent to client nodes 2.147483647E9ms
_______________________________________________ Support mailing list [EMAIL PROTECTED] http://news.gmane.org/gmane.network.freenet.support Unsubscribe at http://dodo.freenetproject.org/cgi-bin/mailman/listinfo/support Or mailto:[EMAIL PROTECTED]
|
OPCFW_CODE
|
Admit it. Looking into the parasitics extracted and saved as calibre file is immensly difficult to view. Even with a simple circuit, not only are the number of parasitics overwhelming, it is organized in a manner where it is impossible for the user to intuitively acknowledge where the parasitics are generated.
Due to such difficulties, I have decided to hack into the calibre file and try to extract values and parameters of these parasitics and processed for a more user-friendly view. It’s an ongoing process, and I have much yet to accomplish.
First, acquire the calibre model file. The extraction models are all saved in ‘layout.oa’ file. This file is in a remote server so the ‘layout.oa’ file was imported to my local machine via ftp.
We need to analyze this file which requires a hex editor. Any kind of hex editor is fine, and I have used ‘Hex Editor Neo’ free version. It has limited features but browing a hex file and simple searching features are open so its enough for my purpose.
This file is quite long, readable texts appear near at the beginning of the file (check that the scroll has bared been dragged down).
This text seems to be an introductory text where it describes the original components that exist in the schematic file.
If you go down a bit more following the text, there will be a point where it starts to list the actual extracted parasitic components(resistor or cap, in this case since it was RC extraction) along with a mention of the nodes that they are connected to.
↑ This seems to be the pharse that marks the start of parasitic lists.
↑ It now declares the existence of the first par.cap (short for ‘parasitic cap’) which is named “ciP10_VMID_0”. Actually, this parasitic cap’s real name is ‘P10_VMID_0’ and the prefix ‘ci’ seems to be metadata. My guess is that ‘c’ stands for ‘capacitor’ and ‘i’ stands for ‘instance’.
Since this is the ‘first’ capacitor it seems likely that is should explain how this capacitacne should be modeled with. In this case, it would be an ideal capacitor defined in ‘analogLib’. Probably that is why the word ‘analogLib’ appears in this phrase.
After declaring the capacitor model, it needs to defined the two nodes of the capacitor, noted as ‘PLUS’ and ‘MINUS’. In this case, the ‘PLUS’ is ‘P10_VMID_606’ and ‘MINUS’ is ‘0’. There IS a node named ‘0’ so do not be confused.
↑ Take a look at the next par.cap. Weirdly, it only states its name (“ciP10_VMID_1”), and one terminal (“P10_VMID_602”). Where is the other terminal? Well, it turns out that if one terminal is the same as the previous one, it will not declare it again. So in this case the other terminal would also be ‘0’.
↑ here’s another example. again, only one terminal is declared which means that the other terminal is ‘0’.
With this rule in mind, the entire text area regarding parasitic declarance can be clipped and saved as another separate file where it can be processed for later use. To do this, a simple python script was used with the offsets for the staring point and end point of this text area calculated by hand and applied in the script.
Using this python script, the text area containing parasitic RC definition was copied to a separate text file.
|
OPCFW_CODE
|
Insiders are employees that must be trusted with access to sensitive information, and because of that trust can be a major threat. Insiders have compromised organizations in multiple domains including manufacturing, finance, government, and even scientific research. Even worse, insiders attacks are consistently catalogued as the most costly given the elevated privilege that insiders have in terms of trust and access. This makes the insider issue one of the most challenging problems in computer security.
As with many other complex systems (e.g., the Internet, online social networks, and the brain), information systems consist of a large number of interacting elements (e.g., users, services, devices, files) in which the aggregate activity of the system cannot be derived by analyzing individual contributions, i.e., their aggregate behavior is nonlinear. Graphs, where nodes represent the elements and edges capture the interactions between the elements of the system, have been used across multiple domains to capture the interactions between the elements of complex systems. The use of graphs to study the structure of complex systems has revealed some plausible explanations for the emergence of collective behavior in these systems such as the understanding of regular and anomalous behavior. In this work, we treat the malicious insider as an anomaly and use bipartite graphs to detect their anomalous behaviors.
The resulting focus on malicious patterns, as opposed to malicious nodes, implements an assumption that the malicious insider is not intrinsically hostile. Rather, malicious behaviors can emerge over time or in respect to specific conditions. Static graphical analysis is based on the analysis of graph snapshots and cannot integrate temporal patterns. In contrast, the study of temporal graphs, where information of single graph snapshots is aggregated, tends to reflect more accurately the evolution of the system as nodes and edges appear and disappear over time. The focus of this work is to understand the malicious behaviors over time rather than identifying the static malicious nodes.
To understand such complex systems, empirical data with detailed temporal information is a prerequisite. Correct temporal information is much more readily available as a source of ground truth than correctly labeled insider threat datasets. In the context of information systems, temporally annotated datasets are widely available thanks to the presence of user-system interaction logs. This enables the use of graph mining analytics for the understanding of anomalous behavior such as the one that insiders might pose.
For the purposes of this project, we characterize and detect anomalous events in an information system based on a centralized version control system. We identify time intervals during which significant changes in the structure of the temporal graphs may correspond to functional change points, e.g., a precipitating events. This problem has also been referred to as change point detection.
We model user-system interactions in a version control system as a temporal bipartite graph where interactions occur exclusively between two types of nodes, (i) users and (ii) software components. Note that the edges in this graph are only between these two types of nodes. A one-mode projection of this graph is the user graph in which two nodes (users) are connected if they have interacted at least once with the same component. Our methodology includes studying the evolution of the one-mode user graph to identify topological properties that characterize the system’s normal behavior. Among these observed properties, those that do not follow the norm of the regular pattern are assumed to indicate the presence of an anomalous event. Such an event may indicate a potential insider incident or, at least, an event that requires further investigation.
In particular, the user graph allows us to explore the impact of precipitating events in user-system interactions. Precipitating events are key events that have the potential to trigger insiders to become a threat to their employer. We hypothesized that precipitating events impact the behavior of interactions between users and components in the version control system by changing patterns of committing behavior. To test this hypothesis, we model and compare the volume of interactions between users over similar or related components as opposed to non-related components over time. To capture sets of users with similar patterns of interaction, we rely on the notion of community structure to identify communities, or clusters, i.e., groups of nodes having higher probability of being connected to each other than to members of other groups. We show that the volume of interactions between users that contribute to unrelated components increases when precipitating events are announced. This indicates the impact of precipitating events in increasing the likelihood of a change in the interacting behavior between users and components, which might be a signal to monitor before an insider attack is committed.
|
OPCFW_CODE
|
package com.shijingsh.ai.jsat.linear;
import java.util.Comparator;
import java.util.Iterator;
/**
* This data structure allows to wrap a Vector so that it is associated with
* some object time. Note, that operations that return a vector will not be a
* Paired Vector, as there is no reason to associate a different vector with
* this vector's pair.
*
* @author Edward Raff
*/
public class VecPaired<V extends Vec, P> extends Vec {
private static final long serialVersionUID = 8039272826439917423L;
private V vector;
private P pair;
public VecPaired(V v, P p) {
this.vector = v;
this.pair = p;
}
public P getPair() {
return pair;
}
public void setPair(P pair) {
this.pair = pair;
}
public V getVector() {
return vector;
}
public void setVector(V vector) {
this.vector = vector;
}
@Override
public int length() {
return vector.length();
}
@Override
public int nnz() {
return vector.nnz();
}
@Override
public double get(int index) {
return vector.get(index);
}
@Override
public void set(int index, double val) {
vector.set(index, val);
}
@Override
public Vec add(double c) {
return vector.add(c);
}
@Override
public Vec add(Vec b) {
b = extractTrueVec(b);
return vector.add(b);
}
@Override
public Vec subtract(Vec b) {
b = extractTrueVec(b);
return vector.subtract(b);
}
@Override
public Vec pairwiseMultiply(Vec b) {
b = extractTrueVec(b);
return vector.pairwiseMultiply(b);
}
@Override
public Vec multiply(double c) {
return vector.multiply(c);
}
@Override
public void multiply(double c, Matrix A, Vec b) {
vector.multiply(c, A, b);
}
@Override
public Vec pairwiseDivide(Vec b) {
b = extractTrueVec(b);
return vector.pairwiseDivide(b);
}
@Override
public Vec divide(double c) {
return vector.divide(c);
}
@Override
public void mutableAdd(double c) {
vector.mutableAdd(c);
}
@Override
public void mutableAdd(Vec b) {
b = extractTrueVec(b);
vector.mutableAdd(b);
}
@Override
public void mutableSubtract(Vec b) {
b = extractTrueVec(b);
vector.mutableSubtract(b);
}
@Override
public void mutablePairwiseMultiply(Vec b) {
b = extractTrueVec(b);
vector.mutablePairwiseDivide(b);
}
@Override
public void mutableMultiply(double c) {
vector.mutableMultiply(c);
}
@Override
public void mutablePairwiseDivide(Vec b) {
b = extractTrueVec(b);
vector.mutablePairwiseDivide(b);
}
@Override
public void mutableDivide(double c) {
vector.mutableDivide(c);
}
@Override
public Vec sortedCopy() {
return vector.sortedCopy();
}
@Override
public double min() {
return vector.min();
}
@Override
public double max() {
return vector.max();
}
@Override
public double sum() {
return vector.sum();
}
@Override
public double mean() {
return vector.mean();
}
@Override
public double standardDeviation() {
return vector.standardDeviation();
}
@Override
public double variance() {
return vector.variance();
}
@Override
public double median() {
return vector.median();
}
@Override
public double skewness() {
return vector.skewness();
}
@Override
public double kurtosis() {
return vector.kurtosis();
}
@Override
public Vec clone() {
return new VecPaired(vector.clone(), pair);
}
@Override
public Vec normalized() {
return vector.normalized();
}
@Override
public void normalize() {
vector.normalize();
}
@Override
public double pNormDist(double p, Vec y) {
y = extractTrueVec(y);
return vector.pNormDist(p, y);
}
@Override
public double pNorm(double p) {
return vector.pNorm(p);
}
@Override
public double dot(Vec v) {
v = extractTrueVec(v);
return this.vector.dot(v);
}
@Override
public String toString() {
return vector.toString();
}
@Override
public boolean equals(Object obj) {
return vector.equals(obj);
}
@Override
public boolean equals(Object obj, double range) {
return vector.equals(obj, range);
}
@Override
public double[] arrayCopy() {
return vector.arrayCopy();
}
@Override
public void mutableAdd(double c, Vec b) {
b = extractTrueVec(b);
this.vector.mutableAdd(c, b);
}
@Override
public Iterator<IndexValue> getNonZeroIterator(int start) {
return extractTrueVec(vector).getNonZeroIterator(start);
}
/**
* This method is used assuming multiple VecPaired are used together. The
* implementation of the vector may have logic to handle the case that the other
* vector is of the same type. This will go through every layer of VecPaired to
* return the final base vector.
*
* @param b a Vec, that may or may not be an instance of {@link VecPaired}
* @return the final Vec backing b, which may be b itself.
*/
public static Vec extractTrueVec(Vec b) {
while (b instanceof VecPaired)
b = ((VecPaired) b).getVector();
return b;
}
public static <V extends Vec, P extends Comparable<P>> Comparator<VecPaired<V, P>> vecPairedComparator() {
Comparator<VecPaired<V, P>> comp = new Comparator<VecPaired<V, P>>() {
@Override
public int compare(VecPaired<V, P> o1, VecPaired<V, P> o2) {
return o1.getPair().compareTo(o2.getPair());
}
};
return comp;
};
@Override
public int hashCode() {
return vector.hashCode();
}
@Override
public boolean isSparse() {
return vector.isSparse();
}
@Override
public void setLength(int length) {
vector.setLength(length);
}
}
|
STACK_EDU
|
SERIOUS BUG: Xcode 4.3.2 creates invalid .momd resources
||Product Version:||4.3.2 / 4E2002|
A versioned .xcdatamodeld bundle that is compiled with MOMC results in a .momd resource bundle which crashes NSManagedObjectModel's -initWithContentsOfURL:
Steps to Reproduce:
1) Create a versioned Data Model (.xcdatamodeld) using Xcode 4.3.2 running on Lion.
2) Compile the application.
3) Load the resource for the compiled Data Model (.momd) using NSManagedObjectModel's -initWithContentsOfURL:
See the attached project: http://files.iconfactory.net/craig/bugs/Vershunning.zip
The sample project loads the unversioned Data Model (AccountDocument.mom) and the data model compiled with Xcode 4.2 (AccountDocument2.momd). It crashes with the data model compiled with Xcode 4.3 (AccountDocument3.momd).
The application should load the managed object model.
The application crashes with EXC_BAD_ACCESS while loading the data in the .momd resource. The stack trace is:
* thread #1: tid = 0x2103, 0x00007fff931d9350 libobjc.A.dylib`objc_msgSend_vtable13 + 16, stop reason = EXC_BAD_ACCESS (code=13, address=0x0)
frame #0: 0x00007fff931d9350 libobjc.A.dylib`objc_msgSend_vtable13 + 16
frame #1: 0x00007fff8935e975 CoreData`-[NSKnownKeysDictionary1 _setValues:retain:] + 197
frame #2: 0x00007fff8935f288 CoreData`_newReadModelFromBytes + 648
frame #3: 0x00007fff8935b93e CoreData`+[NSManagedObjectModel(_NSManagedObjectModelPrivateMethods) _newModelFromOptimizedEncoding:error:] + 9310
frame #4: 0x00007fff89359451 CoreData`-[NSManagedObjectModel(_NSManagedObjectModelPrivateMethods) initWithContentsOfOptimizedURL:] + 305
frame #5: 0x00007fff89358d7b CoreData`-[NSManagedObjectModel initWithContentsOfURL:] + 443
frame #6: 0x000000010edb43da Vershunning`-[AppDelegate applicationDidFinishLaunching:] + 554 at AppDelegate.m:54
The same code works fine on Xcode 4.2 on Snow Leopard. Also, if the .momd resources that are compiled with 4.2 are used in the project as a Copy Files build phase (instead of having MOMC generate them) the application works fine.
If you'd like to see the history of events that led up to this report, check out this StackOverflow question:
Appended the following information on 4/4/2012:
The compiled .momd resources can be loaded after the "existingPartner" fetch request is changed from:
name == $name
name == $NAME
Your mission, should you choose to accept it, is to figure out why a part of the object model that does not affect the persistence of data breaks versioning and lightweight migration. It's clear from the documentation that this should not be the case:
"Core Data’s perspective on versioning is that it is only interested in features of the model that affect persistence."
Reports posted here will not necessarily be seen by Apple.
All problems should be submitted at bugreport.apple.com before they are posted here.
Please only post information for Radars that you have filed yourself, and please do
not include Apple confidential information in your posts. Thank you!
|
OPCFW_CODE
|
I was interested if i can make split horizon DNS in synology web UI. I have googled a bit as i initially could not figure how how to create a zone inside a view. I mean, it was not straight forward for me. Luckily, tonight i decided to give one more try, and it worked. I just would not think that it would be possible to create two dns zones with the same name in main panel in the synology web ui portal.
But, i would not thought this way , that you are suposed to create here multiple zones (with the same name) that later, one will be assigned to different view. It’s just weird to see the (n) in the names. So the reason in my opinion for this behavior is that in this panel we are not in a view. We are in ‘general’ sort of place/setting/panel. I am used to infoblox more, and there , one is ‘forced’ to be in a view from the very first moment. So there will never be the case that one creates two zones with the same name, and the later gets marked with (2). The moment we are creating a zone inside infoblox we already are in a view, even if we did not create one ourselves. The default view that is used in infoblox is called surprisingly ‘default’ 😉 So if one would like to make that split between clients from lets say external and internal networks, one would have to make additional view and call it ‘external’, and the ‘default’ one rename to internal. Afterwards one has to make a zone with the correct records inside the external view zone.
Let’s go back to synology dns. I tried my best to depict the flow/situation:
So let’s say we have a domain : ‘test.zone’ , this domain is hosted by synology dns on our storage appliance, our appliance is exposed to internet. In addition this appliance is also exposed to our LAN, where local users are using local services. Maybe for internal usage a www A record should point to a server in our local lan, where in case somebody on internet wants to reach a www.test.zone he has to sent to a webserver in cloud or different location. For example i create a lot of virtual environments and maybe i don’t want to create a new domains for every lab. Instead all labs would be using the same domain, but just different view.
So in case that internet / lan example.
If a system from internet asks about www A record, our synology dns should reply with 126.96.36.199
If a system from our lan asks about www A record, our synology dns should reply with 192.168.4.5
On the screenshot above, i did nslookup from two machines, one from 192.168.111.0/24 network, and one from 192.168.1.0/24 network. I have created two zones of the same name and later i have assigned them to the correct view, and forced access to the correct view by maching clients by their networks. In case of let’s say internet/lan example, we could do this.
So if a client is from 192.168.x.x network, we will be forcing him to receive 1 version of the zone record. All other cases i treat as ‘system from internet’ / external wants to retrieve record and should get other values. Sample usage below querying this from internet
So when querying the synology dns , we are utilizing the EXT view. When we are doing it from a computer in lan, we are utilizing the INT view. I hope this post will make it easier to understand this concept of dns views.
|
OPCFW_CODE
|
We’ve heard plenty of times in the past about how Android doesn’t make as much money for developers as iOS, and some examples raised valid points, like the fact that there are not as many Google Wallet/Checkout users as there are iTunes account users. Therefore in theory, fewer people bother to pay for an app. There were also some very misleading ones that used data from 2010, when Android was still much smaller in quarterly sales, but even more importantly in total user base, and the data used is somehow supposed to prove that Android is not making money for developers now.
First off, I still believe the Play Store hasn’t reached the same level of monetization as the Apple app store, and the difference is still significant on average, but I also think that the Play Store has come a long way to help developers make money off their apps. Now, it’s important to remember, that just like in any other type of market, most developers will not make any money from their apps, some will barely break even, and only around 20% or so will actually get rich from their apps. This goes for books, music, movies, you name it. It’s the way the world works. Only the best/most promoted content makes money.
For some of the top developers, though, like the Angry Birds maker, Rovio, and DeNA’s Rage of Bahamut, which is currently the #1 grossing developer in both Google’s and Apple’s stores, both the ecosystems are making around the same amount of money for the developers. It’s important to note though, that Rage of Bahamut is a free-to-play game (on iOS also). This might lead to the insight that if you want to make as much money from Android as you do from iOS, you need to offer your app for free, and make money from ads (like Angry Birds) or in-app purchases like Rage of Bahamut.
At least for now, this is probably the way to go. The players who are really interested in buying content in this game will take a few moments of their time to make a Google Wallet account. Unfortunately, at least until ICS makes up a majority of the Android devices, most Android users will not even know what a Google Wallet account is, and therefore they will not pay for apps.
With ICS you get asked from the moment you set-up your phone, to make a Google Wallet account, which you can still skip if you want, but it’s also a great incentive for many people to make their account then. Later they can start buying apps with just a push of a button. I think this is Apple’s big advantage for now in monetizing apps, and until ICS gets a lot more than only 7% market share, things will remain about the same. The simple fact that Android is still growing twice as fast as iOS will help push those numbers in Google’s favor, thanks to the sheer amount of Android users that will exist in the future (right now each having around ~350 million users).
|
OPCFW_CODE
|
Some notes to be aware of:
Write protecting the SD card may cause the entire system to simply not start up. Some files are temporary/dynamic, and need to be written to while the OS (or programs in the OS) is running. There are solutions, of course, such as creating a fake disk drive in RAM to hold these temporary files (being temporary files, there is no problem if the power failure caused the RAM to disappear). Modern Linux kernels do this or similar already; The /run and /sys directories are created in RAM, and not actually written to any physical drive. Others, such as some of the files/directories in /dev, are created as needed.
As for how and why corruption happens.. Understand that reading/writing to he SD card is very slow compared to the speed of the CPU and RAM. When the CPU is executing billions of instructions per second, and the RAM is accessing memory locations in a similar manner, all is well. But what happens when you need to read or write from/to the SD card? It can only read/write at maybe a few thousand memory locations per second.
You do not want the OS to just stop everything to wai for the read/write to complete.. Imagine if you loaded a web page that took a full minute to load. But while it was loading, you had to wait until it was done! You could not drink your coffee, you could not switch to another web page, you could no even tell it to stop loading because you are impatient.
So what the OS does is cache certain data in the RAM so it does not have to constantly wait to read it again from SD. It does it when it writes to SD, as well: Instead of actually writing to the SD card, it saves the data to RAM and schedules it to be actually written to SD at a later time, when the CPU/OS is not so busy, and goes back to doing other things, such as responding to your wishes.
The corruption happens when this data is only in RAM, but the power fails. No power to keep the RAM fresh, so it cannot write the cached data to the SD card. It's gone. So the data on the SD card is now outdated--I does not have the data that was in RAM. Sometimes you are lucky and it does not matter; the lost data is trivial and can be created again. Other times, the lost data is critical data, and corruption results.
Even a small change can cause a big problem. If a file got only one byte larger, the OS has to update the disk data: Both the one byte added to the file, as well as the data on the disk that tells the OS how big the file is. I has to update that to know the file is now 1 byte larger. Again, however, that information is cached in RAM and scheduled to be written to the SD card later.
If a power failure happens now before it writes the new size, the SD card will "think" the file is still the old size. That new bye is completely lost--corruption.
Most operating systems take steps to try to minimize the risk, but it cannot be completely eliminated. Even making the SD card write protected is not perfect; The SD card degrades over time until it can no longer maintain the data stored on it.
An EMP (Electromagnetic pulse) is similar--it causes electrical devices to short and burn out. So your SD card may still be wiped out if it suffered a strong enough EMP. And so on.
|
OPCFW_CODE
|
Metron Benefits. SOC Analyst & Investigator Perspective
The above diagram illustrates the key steps in a typical analyst/investigator workflow. For certain steps in this workflow, Apache Metron provides keys capabilities not found in traditional security tools:
- Looking through Alerts
- Centralized Alerts Console - Having a centralized dashboard for alerts and the telemetry events associated with the alert across all security data sources in your enterprise is a powerful feature within Metron that prevents the Analyst from jumping from one console to another.
- Meta Alerts - The long term vision of Metron is to provide a suite of analytical models and packs including Alerts Relevancy Engine and Meta-Alerts. Meta Alerts are generated by groupings or analytics models and provide a mechanism to shield the end user from being inundated with 1000s of granular alerts.
- Alerts labeled with threat intel data - Viewing alerts labeled with threat intel from third party feeds allows the analyst to decipher more quickly which alerts are legitimate vs false positives.
- Collecting Contextual data
- Fully enriched messages - Analyst spend a lot of time manually enriching the raw alerts or events. With Metron, analysts work with the fully enriched message.
- Single Pane of Glass UI - Single pane of glass that not only has all alerts across different security data sources but also the same view that provides the enriched data
- Centralized real-time search - All alerts and telemetry events are indexed in real-time. Hence, the analyst has immediate access to search for all events.
- All logs in one place - All events with the enrichments and labels are stored in a single repository.
- Granular access to PCAP - After identifying a legitimate threat, more advanced SOC investigators want the ability to download the raw packet data that caused the alert. Metron provides this capability.
- Replay old PCAP against new signatures - Metron can be configured to store raw pcap data in Hadoop for a configurable period of time. This corpus of pcap data can then be replayed to test new analytical models and new signatures.
- Tag Behavior for modeling by data scientists
- Raw messages used as evidentiary store
- Asset inventory and User Identity as enrichment sources.
Note that the above 3 steps in the analyst workflow make up approximately 70% of the time. Metron will drastically decrease the analyst workflow time spend because everything the SOC analyst needs to know is in a single place.
Metron Benefits? Data Scientist Perspective
The above diagram illustrates the key steps in a typical data science workflow. For certain steps in this workflow, Apache Metron provides key capabilities not found in traditional security tools:
- Finding the data
- All my data is in the same place - One of the biggest challenges faced by security data scientists is to find the data required to train and evaluate the score models. Metron provides a single repository where the enterprise’s security telemetry data are stored.
- Data exposed through a variety of APIs - The Metron security vault/repository provides different engines to access and work with the data including SQL, scripting languages, in-memory, java, scala, key-value columnar, REST APIs, User Portals, etc..
- Standard Access Control Policies - All data stored in the Metron security vault is secured via Apache Ranger through access policies at a file system level (HDFS) and at processing engine level (Spark, Hive, HBase, Solr, etc..)
- Cleaning the data
- Metron normalizes telemetry events - As discussed in the first blog where we traced an event being processed by the platform, Metron normalizes all telemetry data into at least a standard 7 tuple json structure allowing data scientists to find and correlate data together more easily.
- Partial schema validation on ingest - Metron framework will validate data on ingest and will filter out bad data automatically which is something that data scientists, traditionally, spend a lot time doing.
- Munging Data
- Automatic data enrichment - Typically data scientists have to manually enrich data to create and test features or have to work with the data/platform team to do so. With Metron, events are enriched in real-time as it comes in and the enriched event is stored in the Metron security vault.
- Automatic application of class labels - Different types of metadata (threat intel information, etc…) is tagged on to the event which allows the data scientists to create feature matrixes for models more easily.
- Massively parallel computation framework - All the cleaning and munging of the data is using distributed technologies that allows the processing of these high velocity/ large volumes to be performant and scalable.
- Visualizing Data
- Real-time search + UI - Metron indexes all events and alerts and provides UI dashboard to perform real-time search.
- Apache Zeppelin Dashboards - Out of the box Zeppelin dashboards will be available that can be used by SOC analysts. With Zeppelin you can share the dashboards, substitute variables, and can quickly change graph types. An example of a dashboard would be to show all HTTP calls that resulted in 404 errors, visualized as a bar graph ordered by the number of failures.
- Integration with Jupyter - Jupyter notebooks will be provided to data scientists for common tasks such as exploration, visualization, plotting, evaluating features, etc..
Note that the above 4 steps in the data science workflow make up approximately 80% of the time. Metron will drastically reduce the time from hypothesis to model for the data scientist.
|
OPCFW_CODE
|
I'd like some advice on how to deal with the endless problems I've had over the last few months.
Last October I received an URGENT SupportAssist alert to immediately update my BIOS to A14. I assumed that this is what had been preventing the update to Windows 1903 release, so I dutifully installed the BIOS update. BIG MISTAKE.
As soon as the update was done, I started getting BSOD within a few minutes of booting into Windows. Every single time. Then I ran the pre-boot diagnostics, which quickly told me that the problem was a memory error. I had upgraded my system from 8 GB to 16 GB 2 years ago, and the two Silicon Power 8GB DDR3L-1600 CL11 memory cards had worked flawlessly.
At first I had thought the memory went bad (I did not immediately make the connection with the BIOS update), and Silicon Power was very responsive in sending me new replacement memory cards. However, they failed in exactly the same way.
At this point the only way I could get my system running was to take the memory out of my daughter's unused XPS 8300 machine and insert it into my XPS 8500 machine. Her memory is DDR3-1333, so slower than spec for my XPS 8500, but it's the only thing that I have around that will work. With her memory in place, my system works reliably, though may be slower since CPU-Z reports only a 667 MHz clock speed.
Unfortunately, my daughter will be returning home in a few weeks, so I need to fix my own memory problem before she returns and re-install her memory in her XPS 8300. I will be super pissed if I need to buy all new memory for my machine because of this silly BIOS update, and of course I may be throwing good money after bad if the new memory boards fail also.
Before I do anything more I'd like your advice on how to proceed. Do I attempt to downgrade my BIOS back to A06 or A09 (one of which I was using before)? If so, how do I do this? (I tried and the BIOS install program refuses to downgrade.) Is there some setting that I can implement that will make my system work with this memory without constant BSOD errors? Are there additional diagnostics that I should run to better diagnose the exact problem?
Back-flashes are tricky.
Some include extra micro-controller firmware or Video-BIOS code that can't be back-flashed ever (and therefore allowing mixing with old BIOS).
You might brick (permanently kill) your motherboard.
But if you are brave, I would try like this (from a special config of FreeDOS, IIRC).
You remember DOS don't ya ... good times. 😁
I think that I would run an independent memory test like MemTest86 v8.3 Free Edition before trying a very risky flashing of the BIOS to an earlier version. By independent I mean independent of the Windows OS since it is possilbe that the BIOS update combined with Windows is causing the BSODs. MemTest86 is bootable and does not require Windows.
By the way, I think your update to Windows 10 version 1903 was blocked due to other reasons not because of the BIOS.
I think that I would run an independent memory test like MemTest86 v8.3 Free Edition before trying a very risky flashing of the BIOS to an earlier version.
Thank you for the suggestion. I will go look for that utility.
By independent I mean independent of the Windows OS since it is possilbe that the BIOS update combined with Windows is causing the BSODs. MemTest86 is bootable and does not require Windows.
Just to make sure I am clear, the diagnostic utility that is showing memory errors is the Dell utility that is part of the BIOS settings. It runs before Windows loads, so it is also independent of Windows. Are you suggesting that Dell's diagnostic utility is returning false error messages?
Sorry I miss the part where you stated the diagnostics you ran were the pre-boot diagnostics. I am not suggesting that there is something wrong with Dell's diagnostics, but I guess there would be no harm in running MemTest86 as an additional verification of a memory problem.
Have you tried clearing BIOS after reinstalling the RAM that "failed"?
As for CPU-Z, you have to remember there's a 2x multiplier involved. If CPU-Z reports that Dram Frequency is 667 MHz, then the RAM is 2x667=1334 MHz, exactly what you'd expect for 1333 MHz modules.
Forum Member since 2004
I am not a Dell employee
|
OPCFW_CODE
|
Uninstall button greyed out in Software Center for applications that are pushed via a "required" collection
When deploying an application through a collection in "required" state, uninstall button is greyed out. This doesn't happen if the application is installed through a collection in "available" state. We would like to have the uninstall button always accessible, no matter if the application was pushed in a required state or not. Corruption happens all the time and technicians need to be able to uninstall/reinstall all the time. Because uninstall is greyed out, they have to use the control panel instead, which defeats the purpose of using Software Center.
The Repair option can be utilized for this, even if the installer doesn't have a native repair feature (like MSIs). You can write an easy script to both Uninstall then Install the same application and put that script in the Repair command field.
**Allow end users to attempt to repair this application**: Starting in version 1810, if you created the application with a repair command line, enable this option. Users see an option in Software Center to Repair the application.
It's taken 6 years for me to discover this problem, and now it is a real problem. Please, please, please enable uninstall for required applications.
Even worse: when the tech uninstalls the application from the Add/Rmv Programs, the Software Center *still* shows the uninstall greyed out, so it's not even possible to Reinstall, because the Reinstall button never appears. The Uninstall button doesn't go away, it remains greyed out, and the tech can neither uninstall nor reinstall from SCCM/Software Center as a result. What the heck gives???
It seems to me that making a radio box check as to whether you want the deployment to be allowed to be uninstalled/reinstalled or not (with the default set to ALLOW the uninstall!) makes infinitely more sense than just making it completely unavailable when the deployment is set to be required/mandatory. As it is, you're just making this service unusable for our enterprise of more than 17,000 workstations. Thanks, Microsoft!
"It's not a bug. It's a FEATURE."
Even if the application is made available to the Support Staff and but required for the PCs the option is greyed out no matter who logs into the PC. I could see the logic behind this. For instance, I wouldn't want end users to be able to uninstall their Antivirus or web gateway software. Could they add a check box to the deployment to allow uninstall for all users?
I Strongly Agree.
Espen Hansen commented
I fully agree that this should be an option for all required deployments. With more than 20000 PCs our Support staff need to perform the manual uninstallation / reinstallation regularly. Not having this option have led to us still preferring Application deployment for 95 % of all deployments. Users can then work on a locked down PC with Mandatory Installation of the application with a Uninstall deployment avaliable when needed.
|
OPCFW_CODE
|
While diving into Red Hat virtualization, I wanted to do some nested virtualization on my Intel NUC. In order to do nested virtualization on RHV 4.3 there are a few things you must configure. Please note that this feature is in tech preview currently.
In order to enable nested virtualization in RHV. You will need to pin the VM that you want to use it on to a particular host and disallow migration of that VM using live migration features. For me since this is a single NUC.
When building a new virtual machine or editing an existing one you will be looking at a screen that looks like the one below. Click on the ‘Host’ tab.
After you click on the host tab you will be looking at a screen similar to the screenshot below. You will want to pin the VM to a particular host, set the migration mode to Allow manual migration only. Once this is completed the Pass-Through Host CPU button will be enabled for use.
Now you will be able to do nested virtualization inside that particular VM. Repeat for any additional VMs you would like to have nested virtualization enabled.
Over the past week I updated my main machine from Fedora 30 to Fedora 31. I figured I would write a quick post about the overall experience thus far. I can say that overall the performance seems snappy than it was in Fedora 30. That being said I did run into a few bugs along the way.
The first bug I ran into was the dash-to-dock bug. After updating dash-to-dock settings would work but the dock would not appear on the desktop. I found that a updated RPM build was already available to the community and once applying the updated RPM from the software store and logging out and back in resolved my issues.
The second bug that reared it’s head wasn’t really a bug per se. I run a samba share on my main machine to share content to other machines on my main network. I went to browse my share to find it not working. Running smbtree on my machine failed to return anything but smbclient was working just fine but I saw a note about SMB1 protocol being disabled.
After a bit of digging I found that the new version of smb/nmb (4.11.0) that came with Fedora 31 had disabled SMB1 protocol by default. This older bug report lead me to the final conclusion. After updating my client’s autofs configs to add vers=3.0 everything was back in order and working.
Overall it was a great experience in comparison to some of my past major revision updates. I did have to reset my wallpaper as part of the screen was ‘torn’ but DKMS handled the nvidia kmod on the new kernel without issues, which was pleasantly surprising.
|
OPCFW_CODE
|
New in the San Diego Release:
- Explore workspaces that have data navigation enhancements and form design changes in the GRC applications.
- View the security and access enhancements for the GRC applications.
- Security updates
- Manage who can access your GRC records with the GRC user roles. Earlier, your users with the snc_internal role could also access the GRC records. As part of the security updates, each GRC application has modified access control lists (ACLs) where access to the GRC records is restricted only to the users with the GRC roles.
- Confidential records
- Enable the sn_grc.enable_record_confidentiality property under GRC properties module in the application navigator to display the Confidentiality tab on a form. You can mark sensitive GRC records as confidential by setting the confidential flag for a record.
- User groups
- Create secure user groups on your GRC tables by enabling the sn_grc.enable_groups_access property under the GRC properties module in the application navigator. You can restrict a user or a set of users from accessing specific records by creating user groups in the record.
- User hierarchy
- Enable the user hierarchy properties under GRC properties in your instances. The managers can view the records that are associated to the users who report to them, based on the user hierarchy that is configured in the sys_user table.
Changed in the San Diego Release:
Enhancements in the 360º view
View the following enhancements in the 360º view:
- Enable the donut chart on a related list when a pill is clicked.
- Apply an order to the pills in a section.
- Fix multiple defects for a better user experience.
Functionality enhancements for the entities
View the following enhancements in the entity and entity filter form:
- Receive an email notification when a user is assigned as an entity owner.
- Select Auto-update owner to keep the entity owner in synchronization with the value from its reference (source) field.
In addition, form design changes are introduced in the entity and entity filter form such as:
- The Class and Owner fields are now mandatory.
- The Department field and the Activity journal are added to the entity form.
Integration of Employee center and GRC
Easily report risk events, raise issues, and policy exceptions from the employee center. This portal also enables employees to attest to a control, respond to a risk identification questionnaire, respond to a risk assessment, and complete the issue remediation tasks.
Specifying risk relevance
Specify the details of how a risk applies to your users. Adding risk details and relevance helps your users to document how the risk applies to them and ensures that they can add the relevant context.
Tasks page enhancements
Add or remove the tabs in the Task Configuration record. You can also prioritize or order the tabs in a sequence. If the same order is configured for two tabs, the tabs are displayed in an alphabetical order.
Deprecated in this release:
RiskLens Integration is no longer deployed, enhanced, or supported. For details, see the Deprecation Process [KB0867184] article in the Now Support knowledge base.
- Audit Management release notes
The ServiceNow® Audit Management application supports activities related to planning audit engagements, executing engagements, and reporting findings to an audit committee. Engagement reporting helps assure key stakeholders that the organization’s risk and compliance management strategy is effective. Audit Management was enhanced and updated in the San Diego release.
- Business Continuity Management release notes
The ServiceNow® Business Continuity Management application provides business continuity solutions when business is impacted during a crisis. Business Continuity Management was enhanced and updated in the San Diego release.
- GRC: integrations with third-party content release notes
The ServiceNow® GRC: integrations with third-party content application provides the APIs and staging tables that can help you to manage the content that you receive from third parties. GRC: integrations with third-party content is a new application in the San Diego release.
- Policy and Compliance integrator release notes
The ServiceNow® GRC: Policy and Compliance integrator application provides a common framework so that your content providers can push their content into the GRC applications. The GRC: Policy and Compliance integrator is a new application in the San Diego release.
- Policy and Compliance Management release notes
The ServiceNow® Policy and Compliance Management application provides a centralized process for creating and managing policies, standards, and internal control procedures that are cross-mapped to external regulations and best practices. Policy and Compliance Management was enhanced and updated in the San Diego release.
- Regulatory Change Management release notes
The ServiceNow® Regulatory Change Management application helps you to manage the regulatory changes in your organization. Regulatory Change Management was enhanced and updated in the San Diego release.
- Privacy Management release notes
The ServiceNow® Privacy Management application enables you to manage your organization’s privacy risks and compliance to protect your customers, employees, and suppliers. Privacy Management was enhanced and updated in the San Diego release.
- Advanced Risk release notes
The ServiceNow® Advanced Risk application helps you avoid any negative impact to your business operations by identifying, assessing, responding to, and continuously monitoring risks. Advanced Risk was enhanced and updated in the San Diego release.
- Vendor Risk Management release notes
The ServiceNow® Vendor Risk Management application provides expanded capabilities for using scores and ratings from external providers beyond security, and an automated approach to increasing visibility that improves your ability to provide guidance to vendors and remediate issues. Vendor Risk Management was enhanced and updated in the San Diego release.
Looking for experts to help you upgrade to ServiceNow San Diego?
ADVANCE, a ServiceNow Implementation Expert and a leading Mhttps://www.advancesolutions.com/servicenow-managed-service-provider/ can help your organization derive the maximum value of ServiceNow Governance, Risk and Compliance (GRC).
Speak to our ServiceNow Experts today.
|
OPCFW_CODE
|
#! /usr/bin/env python
# -*- coding: utf-8 -*-
# vim:fenc=utf-8
#
# Copyright © 2015 mjirik <mjirik@hp-mjirik>
#
# Distributed under terms of the MIT license.
"""
Crop, resize, reshape
"""
import logging
logger = logging.getLogger(__name__)
import argparse
import scipy
import numpy as np
from . import qmisc
def fill_holes_in_segmentation(segmentation, label):
"""
Fill holes in segmentation.
Label could be set interactivelly.
:param label:
:return:
"""
import imtools.show_segmentation
zero_label = 0
segm_to_fill = segmentation == label
segm_to_fill = scipy.ndimage.morphology.binary_fill_holes(segm_to_fill)
return segm_to_fill
# self.segmentation[segm_to_fill] = label
# segm = imtools.show_segmentation.select_labels(segmentation=self.segmentation, labels=labels)
# self.
def unbiased_brick_filter(binary_data, crinfo):
"""
return only binary object which suits with unbiased brick
:param data: 3D ndimage data
:param crinfo: crinfo
http://www.stereology.info/the-optical-disector-and-the-unbiased-brick/
"""
binary_data = binary_data.copy()
crinfo = qmisc.fix_crinfo(crinfo)
imlab, num_features = scipy.ndimage.measurements.label(binary_data)
brick_neg_mask = np.zeros(binary_data.shape, dtype=np.uint8)
brick_neg_mask[
crinfo[0][0]:crinfo[0][1],
crinfo[1][0]:crinfo[1][1],
crinfo[2][0]:crinfo[2][1]
] = 1
exclude_mask = np.zeros(binary_data.shape, dtype=np.uint8)
exclude_mask[:,:, crinfo[2][1]:] = 1
exclude_mask[:,crinfo[1][1]:, crinfo[2][0]:] = 1
exclude_mask[:crinfo[0][0],crinfo[1][0]:, crinfo[2][0]:] = 1
# remove what is not in touch with brick
imlab = keep_what_is_in_touch_with_mask(
imlab, brick_neg_mask, max_label=num_features)
# remove what is in touch with exclude
imlab = remove_what_is_in_touch_with_mask(imlab, exclude_mask)
return (imlab > 0).astype(binary_data.dtype)
# return data[
# __int_or_none(crinfo[0][0]):__int_or_none(crinfo[0][1]),
# __int_or_none(crinfo[1][0]):__int_or_none(crinfo[1][1]),
# __int_or_none(crinfo[2][0]):__int_or_none(crinfo[2][1])
# ]
def keep_what_is_in_touch_with_mask(imlab, keep_mask, max_label):
datatmp = imlab * keep_mask
nz = np.nonzero(datatmp)
labels_to_keep = imlab[nz[0], nz[1], nz[2]]
labels_to_keep = np.unique(labels_to_keep)
for lab in range(0, max_label):
if lab in labels_to_keep:
pass
else:
imlab[imlab == lab] = 0
return imlab
# Rozděl obraz na půl
def split_with_plane(point, orientation, imshape):
"""
Return 3d ndarray with distances from splitting plane
:arg point:
:arg orientation: oriented vector
:arg shape: shape of output data
"""
vector = orientation
vector = vector / np.linalg.norm(vector)
a = vector[0]
b = vector[1]
c = vector[2]
x, y, z = np.mgrid[:imshape[0], :imshape[1], :imshape[2]]
a = vector[0]
b = vector[1]
c = vector[2]
d = -a * point[0] - b * point[1] - c * point[2]
z = (a * x + b * y + c*z + d) / (a**2 + b**2 +c**2)**0.5
return z
def remove_what_is_in_touch_with_mask(imlab, exclude_mask):
datatmp = imlab * exclude_mask
nz = np.nonzero(datatmp)
labels_to_exclude = imlab[nz[0], nz[1], nz[2]]
labels_to_exclude = np.unique(labels_to_exclude)
for lab in labels_to_exclude:
imlab[imlab == lab] = 0
return imlab
def add_seeds_mm(data_seeds, voxelsize_mm, z_mm, x_mm, y_mm, label, radius, width=1):
"""
Function add circle seeds to one slice with defined radius.
It is possible set more seeds on one slice with one dimension
x_mm, y_mm coordinates of circle in mm. It may be array.
z_mm = slice coordinates in mm. It may be array
:param label: one number. 1 is object seed, 2 is background seed
:param radius: is radius of circle in mm
:param width: makes circle with defined width (repeat circle every milimeter)
"""
# this do not work for ndarrays
# if type(x_mm) is not list:
# x_mm = [x_mm]
# if type(y_mm) is not list:
# x_mm = [y_mm]
# if type(z_mm) is not list:
# z_mm = [z_mm]
z_mm = np.asarray(z_mm)
# z_mm = z_mm.squeeze()
for z_mm_j in z_mm:
z_mm_j = np.asarray([z_mm_j])
# repeat circle every milimiter
for i in range(0, width + 1):
data_seeds = _add_seeds_mm_in_one_slice(data_seeds, voxelsize_mm, z_mm_j + i, x_mm, y_mm, label, radius)
return data_seeds
def _add_seeds_mm_in_one_slice(data_seeds, voxelsize_mm, z_mm, x_mm, y_mm, label, radius):
x_mm = np.asarray(x_mm)
y_mm = np.asarray(y_mm)
z_mm = np.asarray(z_mm)
if len(z_mm) > 1:
logger.error("Expected single value for z-axis")
for i in range(0, len(x_mm)):
# xx and yy are 200x200 tables containing the x and y coordinates
# values. mgrid is a mesh creation helper
xx, yy = np.mgrid[
:data_seeds.shape[1],
:data_seeds.shape[2]
]
# circles contains the squared distance to the (100, 100) point
# we are just using the circle equation learnt at school
circle = (
(xx - x_mm[i] / voxelsize_mm[1]) ** 2 +
(yy - y_mm[i] / voxelsize_mm[2]) ** 2
) ** (0.5)
# donuts contains 1's and 0's organized in a donut shape
# you apply 2 thresholds on circle to define the shape
# slice jen s jednim kruhem
slicecircle = circle < radius
slicen = int(z_mm / voxelsize_mm[0])
# slice s tim co už je v něm nastaveno
slicetmp = data_seeds[slicen, :, :]
# mport pdb; pdb.set_trace()
slicetmp[slicecircle == 1] = label
data_seeds[slicen, :, :] = slicetmp
return data_seeds
def main():
logger = logging.getLogger()
logger.setLevel(logging.DEBUG)
ch = logging.StreamHandler()
logger.addHandler(ch)
# create file handler which logs even debug messages
# fh = logging.FileHandler('log.txt')
# fh.setLevel(logging.DEBUG)
# formatter = logging.Formatter(
# '%(asctime)s - %(name)s - %(levelname)s - %(message)s')
# fh.setFormatter(formatter)
# logger.addHandler(fh)
# logger.debug('start')
# input parser
parser = argparse.ArgumentParser(
description=__doc__
)
parser.add_argument(
'-i', '--inputfile',
default=None,
required=True,
help='input file'
)
parser.add_argument(
'-d', '--debug', action='store_true',
help='Debug mode')
args = parser.parse_args()
if args.debug:
ch.setLevel(logging.DEBUG)
if __name__ == "__main__":
main()
|
STACK_EDU
|
Jakarta EE 9 is chugging along. There are three specifications that have passed the ballot for approval by the Jakarta EE Specification Committee and two ballots that are ongoing. More are expected to start this week.
We hit a bump in the road with CDI since we discovered very late in the game that there were some incorrect references to the javax.* namespace left in the specification document. This is an excellent example that shows we need more eyes on these documents. So if you have any time over, please take a look at the pull requests that are approaching ballots. Scroll through the specification document and browse for obvious things that we have missed.
On Thursday, I will present Jakarta EE 9 and Beyond at GIDS.JAVA Live 2020. This is a great conference with an amazing lineup of speakers. I have had the pleasure of speaking at the the conference in Bangalore a couple of times. This time, it will be a virtual appearance.
This week, I held a JetBrains Live Webinar titled Jakarta EE 9 and Beyond. The recording of the webinar is embedded below. The highlight of the session, in my opinion, is the Q&A section at the end. So if you don’t have time to watch the entire recording, fast-forward to around 43 minutes where you will find the Q&A.
The forward-looking part of this presentation is heavily inspired by the Guide to Contributing to Jakarta EE 10 document assembled by the Jakarta EE Ambassadors. I encourage you to read this document, particularly the sections(s) covering technologies you are interested in and/or are using. The time to start looking forward is now!
If you still wonder how to participate, do join the Friends of Jakarta EE call on August 26th. The agenda is forming and contains a summary of the process for getting involved.
As a consequence of the pushed Jakarta EE 9 release date, the Jakarta EE working group decided to move JakartaOne Livestream as well. We have therefore extended the call-for-paper. But do submit sooner rather than later!
I am still enjoying the last couple of days of my vacation, so I will be brief…
We knew that the schedule to get Jakarta EE 9 out was tight, and as it turns out that it was a little too tight. Jakarta EE 9 will not be ready to be released on September 16 as planned. We are very close to starting the process of finalizing the specifications, so I don’t expect it to be delayed much. As soon as a new target date has been decided, I will make sure to make as much noise about it as possible. So stay tuned for an announcement…
On the positive note, Jelastic has come a long way with integrating Jakarta EE 9 to its cloud platform. Container images for Tomcat, TomEE, WildFly, and Jetty are already available. Support for GlassFish and Payara is coming soon.
This is truly impressive and an important step forward to show the industry that the namespace change is coming your way no matter if you like it or not. And now is the time to act!
I admit that I haven’t stayed 100% up-to-date on what’s been going on in the Jakarta EE community the last week. It’s vacation time, and my focus has been on other things than staying connected. Things like early morning dip in the ocean…
Make sure to log off your devices and take some time to recharge yourself!
It’s vacation time, but the hashtag series continues…
The Jakarta EE Working Group has created a Jakarta EE 9 Tools Vendor Datasheet that collects information regarding the Jakarta EE 9 release. Specifically around the namespace change from javax.* to jakarta.*. Please share this datasheet with anyone you suspect may be impacted by this release.
Jakarta EE 9 release is moving forward! At the time of writing this blog, the TCK has 97.98% passing tests. Bookmark the Jakarta EE 9 TCK Overview to follow the progress.
The ballots for releasing the specifications that are a part of Jakarta EE 9 are about to start any moment now. Follow the public Specification mailing list to see how this progresses.
This week, we had the monthly Jakarta EE Community Update Call. It is a relaxed event where we talk about the latest updates in the Jakarta EE Community and community efforts. If you missed it, take a look at the recording.
The reviews of Pull Requests for release review of specifications targeting the Jakarta EE 9 release have started. The schedule is tight, so please help out with the review of these PRs as they come in.
As a curiosity at the end, on July 8, GitHub deposited a copy of all public repositories to the Arctic Code Vault as a part of the GitHub Archive Program. They also introduced the Arctic Code Vault Contributor badge to display on your GitHub profile.
I am happy to announce that my talk Jakarta EE 9 and Beyond is scheduled for EclipseCon 2020. All speakers are given the option to record their talks in advance for this year’s all virtual conference. Given that the planned release date for Jakarta EE 9 is September 16 and pre-recorded talks must be submitted by September 7, this talk will be a live session.
Last week was a shorter week due to the holidays in North America, but the work with Jakarta EE 9 continues. It is a pretty tight schedule, so any help is appreciated.
Work on Jakarta MVC progresses as well. We have now contributed the TCK, which is now available under EE4J on GitHub. The specification document and Javadoc have been prepared and ready for a Jakarta MVC 1.1 release. Now we just need to set up some build jobs and run the TCK against Krazo…
The date is set for JakartaOne Livestream 2020. It will take place on September 16, 2020. We are planning for a 12-hour live streaming event of the same format as last year. The call-for-paper will open any day now. Stay tuned and get ready to submit your awesome proposal!
Make sure to follow @JakartaOneConf on Twitter so you don’t miss out on any information.
Tuesday this week, we celebrated the Jakarta EE 9 Milestone release with an online cupcake party. Here is a photo of my creation.
I also posted a superfast (6 seconds) timelapse of the cupcake production in Studio Jakarta EE. The description contains the recipe if anyone is interested. Please continue the party by posting sharing cupcake picture using the hashtag #JakartaEE.
The first friends call happened this Wednesday. This call is an informal chat among peers mimicking the hallway discussions usually happening at conferences and events. The next call is July 22!
The delivery of Jakarta EE 9 Milestone was a major effort, but this community doesn’t rest. The work with delivering a final version of Jakarta EE 9 on September 16 continues. The schedule is tight, but I am confident we will make it.
We made the initial contribution of Jakarta MVC, and the work with preparing the specification document for a Jakarta MVC 1.1 release is ongoing. Please join our mailing list to participate.
|
OPCFW_CODE
|
There are several ways to check the uptime of a Windows VPS server. For example, you can use the Get-Uptime cmdlet, or you can use the Systeminfo command-line utility. You can also use the Get-ServerUptimeRobot command-line utility. However, these methods do not work as efficiently as those that can run directly from the command-line interface. Here are a few tips to help you get the uptime of a server:
1. Get-Uptime cmdlet
When you need to know how long a Windows server has been running without restarting or shutting down, the Get-Uptime cmdlet is the right tool for you. The command returns the uptime of each computer on the network, with the name of the computer and date and time typed. With the right script, you can run it on multiple Windows servers at once. Here are some useful tips to make your script more effective.
In the PowerShell language, there are several ways to retrieve the uptime of your server. The most common way is to query the Windows event log, WMI, or the Win32_OperatingSystem class. Alternatively, you can use the LastBootUpTime property. If you are using a linux server, you must log in as root and run the uptime command. The result will give you the uptime of the server in days, hours, or minutes. In a Linux server, you can use the procps package to run this command. Running it will give you system uptime, current CPU load, and logged-in user information.
Run the systeminfo command in the Command Prompt or PowerShell. You can also use the net stats command. This command returns general information about the session. The first line in the output is the date and time the server booted, while the second line is the statistics since it was started. The second line indicates the number of hours and minutes that the server was running. This is a useful command if you want to know when your server was booted up, but it’s not recommended for everyday use.
2. Systeminfo Command-line Utility
The Windows Server Systeminfo command-line utility provides uptime statistics and system information. The command is available in the Windows PowerShell environment or the Command Prompt, and comes pre-installed on Windows systems. When run from the command prompt, it returns general information about the session. For example, the “statistics since…” line tells you when your machine was last booted.
There are many ways to find out a Windows server’s uptime statistics, and you can use the “net statistics srv” command to check multiple servers at once. You can use computer names to input data, such as the server name and location, or you can type “netstats srv” directly into the command prompt. Using a custom script will help you generate fast and detailed uptime statistics.
Windows’ Task Manager and the Get-WinEvent cmdlet allow you to quickly query the event log. The Get-WinEvent command searches the system’s event log for events with event IDs 6005 or 6006 and compares them. Make sure to use a reliable Internet connection. This utility will return uptime information for both servers, so you can make an informed decision on whether to install a new server.
This uptime monitoring service provides notifications via e-mail to the user whenever there is a problem with the server. UptimeRobot can monitor HTTP/S, port, keyword, heartbeat, and ping. You can also choose to receive notifications in an email. The service allows you to configure how often to send notifications and whether you want to receive them by text or via email. Moreover, it is free to use for up to 50 monitors. Uptime Robot also has a free trial version that provides two months of logs and monitoring intervals of five minutes. However, you must register with the service to continue using the service. After you’ve registered, you can choose between a Pro or Enterprise subscription.
The Uptime Infrastructure Monitor (UIM) provides a unified view of server health. It collects deep metrics on availability, capacity, and performance across all platforms. It reports across multiple operating systems, including Linux, Windows, UNIX, VMware, Hyper-V, and Xen. Regardless of the OS you’re using, UptimeRobot will give you the insights you need to keep your business running smoothly.
Uptime Robot is a web-based service that checks the status of websites on a regular basis. It sends alerts to users via SMS, push notifications, Slack notifications, and HipChat, depending on your preference. The service can also monitor services that run on a specific port. Specifically, it will check if smtp, dns, and smtp are operating correctly.
|
OPCFW_CODE
|
Discover more from hrbrmstr's Daily Drop
Drop #130 (2022-11-01): Tuesday Productivity Killer Edition
Just two sections today as either resource could very well end up reducing your work output in some way; and, when combined, I suspect you might even forget to nom some tacos at lunchtime.
Data journalists are often unsung heroes in news reporting. They're also usually quite clever humans, who build tools to Get Stuff Done™ that end up helping all of us crunch more effectively.
Since the first
Having said that, platforms like Observable — especially when combined with Quarto
Naël built SDA out of a passion to personally Get Stuff Done™ and enable others to do so. The "Stuff", in Naël's case, is thoughtful data analysis and visual communication/storytelling.
There's a great explanation of the above over at Observable where you can walk through various data analysis examples (with a nod to R's
You're not stuck in Observable (tbh it's not a bad place to be stuck in). The NPM link above and the accompanying GH repo has everything you need to do SDA work at the CLI or in an IDE. As Naël notes in the Observable document, you should be cautious not to put sensitive data into Observable notebooks unless you really know what you're doing.
If you are still learning to code, you may be interested in a SDA sibling project SDA Flow [GH], a visual editor that lets you create an analysis-to-vis workflow, then save it off for re-use. The section header is an example of a toy project, but the video below shows off more of the features:
I'll definitely be using the SDA library in my Observable notebooks and highly recommend giving it a try.
Decker & Lil
Decker [GH] is "a multimedia platform for creating and sharing interactive documents, with sound, images, hypertext, and scripted behavior." It is heavily influenced by HyperCard, one of the best programs that came with early macOS systems.
Anyone can use Decker to create E-Zines, organize their notes, give presentations, build adventure games, or even just doodle some 1-bit pixel art. The holistic "ditherpunk" aesthetic is cozy, a bit nostalgic, and provides fun and distinctive creative constraints. As a prototyping tool, Decker encourages embracing a sketchy, imperfect approach. Finished decks can be saved as standalone .html documents which self-execute in a web browser and can be shared anywhere you can host or embed a web page. Decker also runs natively on MacOS, Windows, and Linux.
For more complex projects, Decker features a novel scripting language named Lil which is strongly influenced by both Lua, an imperative language popular for embedding in tools and game engines, and Q, a functional language in the APL family used with time-series databases. Lil is easy to learn and conventional enough not to ruffle any feathers for users with prior programming experience, but also includes pleasant surprises like implicit scalar-vector arithmetic and an integrated SQL-like query language. A few lines of Lil can go a long way.
Finally, you can see Decker in action in the tour, but you should really just take some time to build some decks. It's great fun, and they'll work on the web, or via a locally installable Decker native app. Remember: constraints fuel creativity!
If you build a deck or two, drop a note in the comments!
One week til the midterm elections in the U.S. Make sure you vote if you're eligible! ☮
|
OPCFW_CODE
|
Some blogs and mailing lists are full of a discussion about the innovations in Arc. One partial consensus seems to be 'to little, to late' at least for an magnum opus. If you want to see an firework of innovation in programming languages, take a look at Wouter van Oortmerssens programming language page. You absolutely must see the visual programming examples and all the other ideas which question the fundamentals of programming language understanding. Pure genius.
Functional programming is heating up. In several discussions over the blogosphere and on my blog others think I have no clue what functional programming is. After some consideration, I think they are right. I have no clue.
Well I do know what functional programming is. I've used it myself. But I have no clue how to do a business application in FP. Others don't know either. They talk about Bezier curves, square roots, fix points, y-combinators, power functions or remodel FP into OO and think that's still FP. Weird.
Suppose I need to write an application. There are customers who can buy products. Based on the customer and the products he bought he can buy other products. Those products he books do cost him money which he needs to pay to the company. I'm quite sure how to design such an application in OO, discuss it with some domain experts and write the code. I also know how to write such an application in a way that other developers in 5 years from now can read, undestand and extend the code.
But how would I write such an application in Lisp? Or Arc? Teach me.
"Character sets are a peripheral matter. The only reason they loom so large in the average programmer's life is that, though trivial, they're an enormous time suck. Trivial + time consuming. Sounds like a good thing to postpone."
He has absolutely no clue about unicode. I and every other Java developer has been using unicode for more than 10 years now, and beside some i18n issues, I/O and String comparisons, they are a non issue if the language supports them.
One is always a bit sheepish about writing quick and dirty programs. And yet some, if not most, of the best programs began that way. And some, if not most, of the most spectacular failures in software have been perpetrated by people trying to do the opposite.
This renders all the discussions lately just mood. PG has no experience with software development, engineering, maintenance, or anything else except writing cute dirty hacks in Lisp. If he does, enlighten me. His release of Arc doesn't help.
Update: Take an advice from someone who has written real world applications.
Tada-List was written in 579 lines of Rails code. Jeff of Coding Horror writes "[...], I agree with Joseph: it's an impressive achievement, [...]". I'm not impressed. If you write a framework for a narrow group of applications, it should be easy to write a small target application with a few lines of code. And they don't count HTML.Top this: a game written in ZERO lines of code.
Most code is hidden in the framework. Taking it to the extreme, with XL/R and concept programming I could write TaDa List in one line of code:
The issue remains. It's the same with encryption. There is a message and a key. How much information is in the key and the message? If the key is "Hello World" then the message can just be "1". If the key is "1" then the message needs to be "Hello World". How much code is in the framework and the application?
That aside. My biggest achivement in small code size was a 1024 bytes (the boot sector size of an Amiga) boot selector menu with color bars. Impressive if you consider we had to put all the menu text into those 1024 bytes.
|
OPCFW_CODE
|
Troubleshooting import DAG issues
Hi,
We encounter the error that we get an ADF error importing the dag files into the integration runtime.
The adf error displayed no information:
Failed to import files
Failed to import files: Airflow1 (Airflow)
Details (not clickable)
So we are stuck....
Document Details
⚠ Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.
ID: b776b3a1-e554-e163-74b3-a6b4c9aaa187
Version Independent ID: 1824cd09-18c0-1cb5-bd4b-f771b854de4c
Content: How does Managed Airflow work? - Azure Data Factory
Content Source: articles/data-factory/how-does-managed-airflow-work.md
Service: data-factory
GitHub Login: @nabhishek
Microsoft Alias: abnarain
@cjvuffelen
Thanks for your feedback! We will investigate and update as appropriate.
I have experienced the same issue with all versions of airflow (1.10.14, 2.2.2, 2.4.3). The error message is simply 'Failed to import files' and the details are blank. I have tried this with the sample dag provided in the documentation at https://learn.microsoft.com/en-us/azure/data-factory/tutorial-run-existing-pipeline-with-airflow and the tutorial dag provided by airflow in their documentation.
I have experienced the same issue with all versions of airflow (1.10.14, 2.2.2, 2.4.3). The error message is simply 'Failed to import files' and the details are blank. I have tried this with the sample dag provided in the documentation at https://learn.microsoft.com/en-us/azure/data-factory/tutorial-run-existing-pipeline-with-airflow and the tutorial dag provided by airflow in their documentation.
We have support ticket pending:
TrackingID#2303090050002010
Just wanted to share what I learned, in case it's helpful. I had a call with a support engineer this afternoon and found that this issue was due to using a self-hosted IR or a managed virtual network IR to access the blob storage account. Setting up a plain, Azure hosted IR to access the storage account with allowed the dags to be imported. I was told that this is only an issue due to the fact airflow is still in public preview, and that when it is fully released, virtual network and self-hosted IRs will be supported.
Hope this is helpful.
Hi, Jlibstaff, tnx for you input, also togehter with a support engineer fixed the issue in a simular way. For others reading this. Your linked service to the blob storage should use the autoresolve integration runtime and not your self hosted one.
@cjvuffelen
I've delegated this to @nabhishek , a content author, to review and share their valuable insights.
Any Updates? I'm using auto-resolve integration runtime, still it not working
Thanks @cjvuffelen. Can you open a ticket with our support team to help you investigate this further? You can create a ticket from the Azure portal using the main menu at the top left to choose Help + Support. I will close this issue for now but we can continue the discussion in this thread as necessary, even after the issue is closed.
#please-close
|
GITHUB_ARCHIVE
|
Plasma Mobile - Akonadi Calendar
While KDE Plasma on the desktop has robust calendar, tasks and journal management capabilities thanks to KOrganizer, Plasma Mobile currently lacks such capabilities. Existing applications are hampered by an absence of Akonadi integration, which would allow users to synchronise their data across applications in Plasma.
The reason for this is that there is currently no existing Kirigami/QML implementation of a calendar view component that integrates with Akonadi. Carl Schwan recently began developing a proof-of-concept Kirigami ‘Kalendar’ application that supports a detailed month view and works with Akonadi calendars, but that is currently a work-in-progress project.
This proposal puts forward a plan for improvements to this existing proof of concept that would help implement a fully-featured convergent calendar application for Plasma Mobile. This application would allow the user to view, create, modify, and delete calendar events with full Akonadi integration. This will be made possible by creating components written in QML and Kirigami that should allow for the creation of a visually-attractive and intuitive calendar view components that could be used in any Kirigami application.
These weekly blog posts contain details about what work was completed each week, and what changes were made to Kalendar during said week.
- First week of Google Summer of Code 2021
- Week 2 on Kalendar (GSoC 2021)
- Figuring out recurrence (and bugs…) in Kalendar’s week 3 (GSoC 2021)
- Editing and deleting mistakes… and events too — Kalendar week 4 (GSoC 2021)
- An eventful week — Kalendar week 5 (GSoC 2021)
- *BZZT* Your 12 o’clock is next on your schedule — Kalendar week 6 (GSoC 2021)
- Ticking this off the todo list — Kalendar week 7 (GSoC 2021)
- Meeting with your mates — Kalendar week 8 (GSoC 2021)
- Making Kalendar pretty — Kalendar week 9 (GSoC 2021)
- Big fixes, more customisation, many improvements — Kalendar week 10 (GSoC 2021)
- Say hello to your new todo app! – Kalendar week 11 (GSoC 2021)
- !2: Add event editor to Kalendar
- !3: Adding an AttendeesModel and RemindersModel for Event editing
- !4: Adding support for event recurrence
- !5: Add event editing and deleting capabilities, new event information drawer, major bugfixes
- !6: Update repository README and relevant screenshot
- !8: Improved event info drawer, event collection editing, attachments
- !9: Undo/Redo functionality
- !10: Add a schedule view to Kalendar
- !11: Improve appearance of event editor components
- !12: Add support for more incidence types
- !13: Add a contact picker for easy attendee addition
- !14: Improve month view and schedule view visually
- !15: Add opt-in location maps to Kalendar
- !17: Collection of fixes for bugs and visual tweaks throughout Kalendar
- !18: Add customisation options to Kalendar
- !19: Unify date and time combos and fix their editing issues
- !20: Add a specialised todo view to Kalendar
|
OPCFW_CODE
|
Monitoring SharePoint Online is required to optimize your network and application delivery performance. CloudReady SharePoint Monitoring deploys in minutes.Downloads – Webtrendshttps://webtrends.com/support/downloadsFor larger SharePoint deployments, the Tag Administrator enables the management of analytics tracking across your collections without re-configuring your application or using scripting. One click is all it takes to launch a new page where you can share files from your PC, Mac, or mobile device. Already using Dropbox, Egnyte, Google Drive, Box or OneDrive to store your files? Download files from SharePoint document libraries using Node.js without hassles - koltyakov/sppull SharePoint web development tools for Visual Studio Code - chrishasz/spgo SharePoint is a web-based collaborative platform that integrates with Microsoft Office. Launched in 2001, SharePoint is primarily sold as a document management and storage system, but the product is highly configurable and usage varies… Identify sensitive and mission-critical content and reveal the true value of your enterprise data with SharePoint data classification software from Netwrix. An effective combination of workflow, team collaboration, and document management, Microsoft SharePoint Online is an easy pick for our Editors' Choice designation. But make sure you need all this power because its price can be significant.
SharePoint Backup provides lightning-fast, total backup and restore for SharePoint. It backs up all of SharePoint including RBS content, configurations,
This feature as well as downloading several files are already available in Office 365 Personal/Home but not in OneDrive for Business???? It is
All you need to do to share files internally is save You might use block download on a sensitive 19 May 2018 SharePoint | O365 | working @Ventigrate Basically, it was a .zip file contains folders and files (multiple types: .txt, .exe, .html…) So simply download the whole folder/.zip instead of a subfolder until Microsoft fixes this issue. SharePoint has a feature in Document Library called 'Generate File Plan Report'. To get the full hierarchy of folders including files inside each level of hierarchy, you need to switch back to basics. Download .ps1 PowerShell Script File 19 Jun 2019 Solved: 1. I tried using Sharepoint List Input tool - when i am giving the full path including the sub folders it is throwing an error. Sharepoint. Hi all I need all experts your help/suppport in this. I need a macro to access sharepoint(https) folder/sub-folder and download the (.xlsm)files
Microsoft SharePoint. 98 346 To se mi líbí · Mluví o tom (223). Welcome to the Official Microsoft SharePoint Facebook page. We invite you to engage with
In this code we will learn how to download files from sharepoint document library. In this code we will learn how to download files from sharepoint document library. In this code we will learn how to download files from sharepoint document library. Top 10 Social Media Influencers. Why Join Become a Download all files from the SharePoint Document Library on Local Drive This Script will allow you to download files from SharePoint Document Library to Local disk or Shared drive. Tested on SharePoint 2010, 2013 and 2016. Hi Andrew, To download everything such as files, folders and documents, you may consider using Open with Explorer feature to copy and paste the files into your local drive. For your reference: Video: Copy or move library files by using Open with Explorer (The third video) How to use the "Open with Explorer" command to troubleshoot issues in SharePoint Online Save the above code in a notepad as .ps1 file. Run the code on the Windows PowerShell Modules. It will display the file name it is downloading and all the files will start getting downloaded to your local. It will be so quick that you won’t put much time and effort on this, thereby saving a lot of time and effort. In SharePoint Online, you can use the Copy to or Move to command to copy or move files between libraries. For more info, see Move or copy a folder, file, or link in a document library. Move to and Copy to buttons in SharePoint Online. Getting your SharePoint library into an Explorer view can provide a fast way to do all file operations.
SharePoint On-premises versions do render .svg and .html files on the browser without any issues.
SharePoint has a feature in Document Library called 'Generate File Plan Report'. To get the full hierarchy of folders including files inside each level of hierarchy, you need to switch back to basics. Download .ps1 PowerShell Script File
Sharegate's Download content option to export the content of your SharePoint or Office Manage Never Checked-In and Checked-Out Files in SharePoint. Write a Python script which is given a Sharepoint login and credentials. The script will downloads all files and folders under the address provided to the local Demonstrates how to download all files from a SharePoint folder that are newer The following comments apply to SharePoint Windows classic authentication. The files/dlls are located on any SharePoint 2013 server at You can copy the dlls from the servers or else you can download them from I am looking to move all files from a SharePoint location to a folder on the C drive. I am able to move one specified file but not all of them.
|
OPCFW_CODE
|
Send Alerts to a Mobile Phone
Estimated reading time: 3 minutes
This page details OT Link Platform flows use cases for sending SMS alerts to Mobile Phones.
OT Link Platform Use Case
Send alerts from the shop floor to workers who need to respond to critical conditions:
- Read tags from a PLC.
- Define alerts (conditions/rules/levels) in Flows.
- Route alerts to specific cell phones or user groups.
Strategies for Sending Alerts
Use an SMS Converged edge system for Messaging
- SMPP-based SMS Converged edge system: Use the SMPP (Short Message Peer-to-Peer) protocol for an SMS (Short Message Service) Converged edge system to generate alerts. Use either a Converged edge system available in a data center or deploy an open source solution, such as Jasmin SMS Converged edge system, which you can install in a OT Link Platform Marketplace.
- Hardware SMS Converged edge system: Deploy a device with a SIM card to serve as an SMS Converged edge system. Use Flows to send messages to this SMS Converged edge system when an alert is triggered.
- Online SMS Converged edge system: Use an online, cloud-based SMS Converged edge system, such as AWS SNS (Simple Notification Service) or Twilio.
Send Messages over WiFi/Network
Install a Notification Receiving mobile app. Several MBaaS (Mobile Backend-as-a-Service) applications (such as Kinvey, Cloudkit for iOS, or Firebase) provide this capability.
- OT Link Platform becomes the Notification Server.
- The app connects to Flows to receive notifications.
Create a Twilio Flow to Send an SMS Alert
The following example provides_ high-level guidance_ for configuring outgoing Twilio messaging in Flows. Because each enterprise has unique requirements and configurations, customize this example for your environment.
- Know how to configure flows in Flows. See OT Link Platform Flows.
- Install a Twilio Node in the Flows Palette.
- Register for a Twilio account here (if you do not have an account), to enable the Twilio Node in a Flow.
Twilio bills for message activity.
Install a Twilio Node in the Flows Palette
To add a Twilio node to the Flows palette:
- Take the steps described in Manage the Nodes Palette.
- Search for the node-red-node-twilio node and click Install.
Although this node supports phone calls, this documentation covers only text messaging.
Configure the Twilio Node
Configure connectivity parameters for the Twilio Node:
- In Flows, drag a Twilio Node onto the canvas.
Double-click the node to edit it.
- In the edit node dialog, select SMS output and enter a phone number (no spaces or special characters) for the person who will receive the text messages.
- Do not click Done yet.
In the edit node dialog, click the pencil icon (next to the Twilio field) to configure the connection parameters for the Twilio Node.
To retrieve the values for configuring a Twilio connection, go to your Twilio account and access the Programmable SMS Dashboard.
- Account SID - Find this security ID in the Twilio account dashboard.
- From - Use the Twilio-assigned phone number found in your Twilio account.
- Token - Find this authentication token in the Twilio account dashboard.
- Name - Enter an optional user-defined name.
- Click Update to return to the edit node dialog.
- In the edit node dialog, click Done.
- Save the flow.
Create a Flow to Test Connectivity
Create a basic flow to test connectivity to Twilio.
To create a flow to test Twilio connectivity:
- In Flows, continue with the flow that you just configured in the previous procedure.
- Drag a Function node onto the canvas and wire it to the Twilio Node.
Double-click the Function node to edit it. Enter a basic message to send as a text message. Click Done.
msg.payload = "Alert: This is a test message from the manufacturing floor."; return msg;
Drag an Inject node onto the canvas and wire it to the Function node. This node will initiate sending the text to the phone.
- Save the flow.
- Inject a timestamp to trigger the alert. Then, check your phone for the text message.
|
OPCFW_CODE
|
package com.t4m.extractor.util;
import com.github.javaparser.Range;
import com.t4m.extractor.entity.*;
import java.util.List;
import java.util.Objects;
import java.util.Optional;
/**
* Created by Yuxiang Liao on 2020-06-21 15:52.
*/
public class EntityUtil {
private EntityUtil() {
}
public static FieldInfo getFieldByShortName(List<FieldInfo> fieldInfoList, String shortName) {
Optional<FieldInfo> target = fieldInfoList.stream().filter(
fieldInfo -> shortName.equals(fieldInfo.getShortName())).findFirst();
return target.orElse(null);
}
public static MethodInfo getMethodByQualifiedNameAndRangeLocator(
List<MethodInfo> methodInfoList, String qualifiedName, Range rangeLocator) {
return methodInfoList.stream().filter(methodInfo -> qualifiedName.equals(methodInfo.getFullyQualifiedName()) &&
rangeLocator.equals(methodInfo.getRangeLocator())).findFirst().orElse(null);
}
public static ClassInfo getClassByQualifiedName(
List<ClassInfo> classInfoList, String fullyQualifiedClassName) {
Optional<ClassInfo> optProjectInfo = classInfoList.stream().filter(
classInfo -> fullyQualifiedClassName.equals(classInfo.getFullyQualifiedName())).findFirst();
return optProjectInfo.orElse(null);
}
public static ClassInfo getClassByShortName(List<ClassInfo> classInfoList, String shortName) {
Optional<ClassInfo> target = classInfoList.stream().filter(
classInfo -> shortName.equals(classInfo.getShortName())).findFirst();
return target.orElse(null);
}
public static ModuleInfo getModuleByShortName(List<ModuleInfo> moduleInfoList, String shortName) {
Optional<ModuleInfo> target = moduleInfoList.stream().filter(
moduleInfo -> shortName.equals(moduleInfo.getShortName())).findFirst();
return target.orElse(null);
}
public static ModuleInfo getModuleByRelativeName(List<ModuleInfo> moduleInfoList, String relativeName) {
Optional<ModuleInfo> target = moduleInfoList.stream().filter(
moduleInfo -> relativeName.equals(moduleInfo.getRelativePath())).findFirst();
return target.orElse(null);
}
public static PackageInfo getPackageByQualifiedName(List<PackageInfo> packageInfoList, String qualifiedName) {
Optional<PackageInfo> target = packageInfoList.stream().filter(
packageInfo -> qualifiedName.equals(packageInfo.getFullyQualifiedName())).findFirst();
return target.orElse(null);
}
public static PackageInfo getPackageInfoByAbsolutePath(List<PackageInfo> packageInfoList, String absolotePath) {
Optional<PackageInfo> target = packageInfoList.stream().filter(
packageInfo -> absolotePath.equals(packageInfo.getAbsolutePath())).findFirst();
return target.orElse(null);
}
/**
* Add a entity into list if it does not exist.
* @param <T> Should be {@code ModuleInfo}, {@code PackageInfo}, {@code ClassInfo}, {@code MethodInfo} and {@code FieldInfo}
* @param entity The entity that need to be added into the list.
* @param targetList The list to which the entity should add.
* @return The T entity itself.
*/
public static <T> T safeAddEntityToList(T entity, List<T> targetList) {
int index;
if ((index = targetList.indexOf(entity)) == -1) {
targetList.add(entity);
return entity;
} else {
return targetList.get(index);
}
}
/**
* Add dependencies for both objects of {@code current} and {@code target}
* @param current The class that call the method of {@code target} class
* @param target The class of which method is called by {@code current} class
*/
public static void addDependency(ClassInfo current, ClassInfo target) {
if (!Objects.equals(current, target)) {
EntityUtil.safeAddEntityToList(target, current.getActiveDependencyAkaFanOutList());
EntityUtil.safeAddEntityToList(current, target.getPassiveDependencyAkaFanInList());
addDependency(current.getPackageInfo(), target.getPackageInfo());
}
}
/**
* Add dependencies for {@code current} and each element in {@code referenceClassList}
* @param current The class that call the method of {@code target} class
* @param referenceClassList A collection of classes of which methods are called by {@code current} class
*/
public static void addDependency(ClassInfo current, List<ClassInfo> referenceClassList) {
for (ClassInfo referClass : referenceClassList) {
addDependency(current, referClass);
}
}
/**
* Add dependencies for both {@code current} and {@code target}
* @param current The package that point to the {@code target} package
* @param target The package that is pointed by {@code current} package
*/
public static void addDependency(PackageInfo current, PackageInfo target) {
if (!Objects.equals(current, target)) {
EntityUtil.safeAddEntityToList(target, current.getActiveDependencyAkaFanOutList());
EntityUtil.safeAddEntityToList(current, target.getPassiveDependencyAkaFanInList());
}
}
}
|
STACK_EDU
|
|Column Tag:||Inside Information
The Difference That Makes A Difference
Whats valuable depends on your perspective. Will that be changing soon?
By Chris Espinosa, Apple Computer, Inc., MacTech Magazine Regular Contributing Author
Nick Negroponte of the MIT Media Lab defines information as a difference that makes a difference. On Usenet, you hear about this as the signal-to-noise ratio, that is, the kernels of useful wheat in the general chaff of questions, misinformations, rumors, and flames. In most other circumstances, though, information in digital form makes a real difference - and this is most true in developing software.
Every bit of your application makes a difference. At the basic level, each bit has to be a non-buggy bit (as opposed to a buggy bit) or your software will crash, and that could make a big difference to its users and purchasers. A little above that, the bits of your program are carefully compiled to run on a specific family of microprocessor; the system calls in your program are linked to a specific operating system API; and the logical assumptions are based on the performance and capabilities of a certain range of hardware platforms. All of these choices are encoded into your finished product, and they make a substantial difference in who will buy and use it.
Above that, of course, are the features and functions of your product itself. This is supposedly what yo're good at, and ostensibly what your customers are paying money for. Of all the investments you make in research and development, the information you learn about how to make your program solve the customers problem should be most worthwhile to you and to them, shouldnt it?
But as youre probably aware, your choice of platform often makes more of a difference to your customers than your choice of features or technologies. Everybody in the Mac business has been told more than once that your product is great, but if it doesnt run on IBMs I cant use it. And you spend much of your time and money simply porting your application from one system version to the next, or from one hardware platform to another - and recently, from one microprocessor to another. The differences are significant, because compiler technology, hardware evolution, and new system APIs are not simple things; but at least they make a difference to your customers.
What will happen if these differences stop making a difference? What if, for example, you didnt have to worry about what instruction set to compile for? In a small way its true now - if your application is not speed-sensitive, you can just compile it for the 68K, and the emulator on the Power Macintosh products will automatically run your software on the Power PC-based models. And while emulation is admittedly slower than running native, you could be seeing more processor independence in the future. Apples Advanced Technology Group and others in the industry have been researching processor-independent object file formats. With these, you compile and link your application into intermediate code which you ship to customers; then either the Installer or the segment loader transliterates the code into the correct instruction set for each machine. The hardware vendor can use different CPUs, the users get native performance, and you can ship one program that runs on many brands.
And with processors continuing to get faster and cheaper, and multiprocessor designs starting to become available, emulators might be the big win after all. If you can add more processors to run your emulators faster, you might be able to achieve near-native performance through emulation. Just think: if you want to run Windows applications faster, just keep adding more Power PC chips to your Macintosh until its fast enough!
Independence from hardware architecture is getting easier as well. In modern OS architectures, a hardware abstraction layer separates the OS kernel from the particular hardware implementation, making it easier to port the OS to different hardware platforms. And developers of new platforms are trying an alternative to the defacto standards of Macintosh (controlled by Apple) and the Intel-based PC architecture (controlled by nobody in particular). The result is a set of reference platforms, hardware designs that assure certain capabilities in different vendors designs. The last major reference platform, ACE, was built around Windows NT and the MIPS chip; the current hot platform, PReP, is based on the Power PC chip and AIX. If reference platforms dominate the landscape in the future, it should be easier to write code that runs indifferently on multiple platforms.
Finally, APIs are crossing the hardware boundaries. Both OpenDoc and OLE 2.0 are cross-platform, though they dont isolate you from other toolbox calls. Hosting layers like XVT and Novell Appware Foundation add surprisingly little overhead to run the same API on different underlying toolboxes. And future operating systems like Taligents Pink system and IBMs Workplace Shell are meant to host multiple personalities on one OS kernel, so your choice of hardware vendor doesnt dictate your choice of API, and therefore applications software.
So five years from now, our old landmarks - the instruction set, the hardware architecture, and the API - may be rotting and fallen. Will it be a total mix-and-match world? Will people be running Mac code in an emulator box on Windows NT on a Compaq Power PC platform, or x86 OLE objects wrappered by OpenDoc running on OS/2 on a Macintosh with a Cyrix chip emulating the Pentium in microcode?
I say: yes and no. I expect that the majority of successful commercial software will be (more or less) compiled and built for a specific class of microprocessor, hardware platform, and API. Itll just be easier that way, both technically and in the marketplace. Though the technology might be able to jump through hoops, the channels and customers dont get over such fundamental taboos as incompatibility overnight.
But while compatibility may remain a litmus test, itll no longer be a barrier. In-house developers will be able to compile something once and deploy it on their Mac, Windows, and UNIX machines, letting adapters and emulators take care of details. Or you could take a product thats successful on one platform, test-market it in the emulator community on other platforms and, if it sells, then invest in the native port to increase your market share and competitiveness. Or (for extra credit) you could find clever ways to bridge the various environments, perhaps hooking up TAPI in SoftWindows to the Geoport or AV capabilities on a Power Macintosh.
Old differences die hard. Even after technology has made them irrelevant, the distinctions of architecture will color peoples thinking. Most conventional development will probably remain the way its always been, but there may be some interesting new opportunities when the gaps between platforms are bridged over.
|
OPCFW_CODE
|
Helpdesk/Support Engineer Resume Samples (Text format)
Helpdesk Support Engineer
- Results-driven helpdesk support engineer with XX years of experience in contributing to the advancement of the organization’s adoption of the Windows operating system by adding to the standards for its configuration and management.Seeking a position at your company to leverage my skills and work for the company's development.
- Skilled in providing resolution to an assortment of problems of moderately complex scope; contributing to the advancement of the organization's adoption of the Windows operating system by adding to the standards for its configuration and management; developing effective methods for its deployment
- Adept in accessing software updates, drivers, knowledge bases, and frequently asked questions resources on the Internet to aid in problem resolution; involved in troubleshooting windows Desktop and Microsoft Office applications
- Deft in performing hands-on fixes at the desktop level, including installing and upgrading software, implementing file backups, and configuring systems and applications; reacting to service outages promptly with adherence to standard operating procedures
Computer Science, Bachelor of Education, Completed, 2003-06-01
St Stephen's College
– Marks null,
Helpdesk / Desktop Support Engineer
Three Ess Computer Services India Pvt Ltd
- Responded to calls and made work tickets for my fellow engineers with Connect wise ticketing system.
- Helped colleagues install programs such as Adobe Reader DC, and other PDF programs.
- Implemented standard principles, concepts, and techniques for the identification and resolution of user problems.
- Performed hands-on fixes at the desktop level, including installing and upgrading software, implementing file backups, and configuring systems and applications Reacting to service outages promptly with adherence to standard operating procedures.
- Troubleshoot issues that the end users are facing.
- Provided timely resolution of problems or escalation of issues to appropriate technical personnel.
- Participated in the development of information technology and infrastructure projects.
Navi Mumbai, MH
- Installed printers and troubleshot issues with printers.
- Secured the computers by running computer safeguards, Malwarebytes, Disk defrag, and disk clean-up.
- Troubleshoot Issues with Outlook 2007, 2010, and 365
- Identified and resolved user service requests utilizing the service request management system.
- Effectively troubleshoot windows Desktop and Microsoft Office applications
- Maintained standard software configurations, including troubleshooting, loading, and configuring software images, supported applications, and drivers.
- Provided support services to staff with technical problems and information technology issues involving desktops, laptops, PDAs, or network services.
- Technical Knowledge
- Microsoft Office applications
- Securing the system from virus
- Adobe Reader DC
- Communication(Coaching and instructing)
- Detail Oriented
- Critical Thinking
- Decision Making
|
OPCFW_CODE
|
In theory, sIFR is a perfect response to one of the more frustrating restrictions imposed on web designers by the medium. Unfortunately, the technology is not without its flaws and ultimately should not be considered a viable choice for web professionals.
The limitations of sIFR include: localization, extended character support, fouc, page zooming, text scaling, text wrapping, page load, cpu load, scalability, utf-8 compliance, QA testing, future-proofing, upgradability, wmode support, dynamic text sizing, DOM support, and z-index collisions.
While there may be solutions or workarounds to some of the issues illustrated here, collectively they remain a barrier to the technology’s adoption as a standard.
- Wizardworks: sIFR – Point Of View (Publish date unknown)
Jim offers a god breakdown on the pros and cons of sIR and I will elaborate briefly on item #5: sIFR and Special Characters
- In order to truly support localized content, a separate .swf file would need to be created for each language, each of those .swf files would then need to contain the entire character set for that country.
- The more characters added, the larger the .swf file becomes. This is noteworthy as few if any fonts families contain every character set for all languages.
- Web designers may encounter much difficulty testing the site with different languages as the text and spacing is rendered differently depending on the user’s country and whether or not the Flash plugin is installed.
- Support for dynamic widths is also a serious issue as the Flash movie width must be explicitly set. This further reduces sIFR’s practicality as fixed widths cannot be used for text on localized websites.
- Usable Type: How and When to Use sIFR (December 24, 2004)
Andy discredits much of the misinformation and concludes sIFR is best used to replace no more than one headline per page.
- Virtual Elvis: Why I hate sIFR (April 29, 2005)
A fabulous tirade about the difficulties of keyboard navigation and screen zooming for an Opera user.
- Leftlane: ESPN.com Looking a Bit Pixelated (February 9, 2006)
Interesting to note in the comments here that while ESPN is one of the sites often quoted to be using sIFR it is a) not currently using sIFR and b) (quoting Mike Davidson) “It was actually not sIFR at all, and even pre-dated Inman’s IFR by at least two years. It was basically inline Flash files with text… that’s all. And yes, the page is huge and so are all of those images.”
- Drupal: sIFR Module discussion (February 16, 2006)
In the comments section, a user notes that sIFR text is not searchable and cannot be made both selectable and linkable. The commentary further notes a lack of respect for user-specified font settings. Someone else points out that sIFR is not utf-8 compliant.
- Mezzoblue: The Pros and Cons of sIFR (October 26, 2006)
A notable article which points out the need to refresh the browser once Command/Ctrl + is used to increase font size. The inability to use sIFR for headers which are also links (the link destination will not show in the status bar) is also mentioned.
- Devshed Forums: SIFR … Innovative or Designer’s junk? (April 13, 2007)
A flaming forum rant that suggests using a pre-rendered PNG instead. The author also further points out the load and latency issues inherent in sIFR.
- Sitening: New Flash player breaks sIFR (December 14, 2007)
This is an important occasion to make note of. When Adobe released the latest version of their Flash browser plugin, earlier implementations of sIFR malfunctioned. While awaiting a fix, sites deploying sIFR were forced to remove the code – assuming they were even aware the feature stopped working.
Live sIFR examples
As a test, interact with these (or any other) sIFR enabled web sites using different browsers such as Opera, Firefox, Safari, Internet Explorer and Konqueror. Try printing, previewing, selecting text, searching for text, increasing and decreasing font size, and other page-level interactions. Keep in mind that while every individual may use the internet in a unique way, all web browsers offer the same set of features to each of their users.
For the following examples I have included sample videos showing scripted interactions with each web site. These videos are made available as a point of reference should one of them cease using sIFR or the technology evolves.
- Aston Martin See demo video
- Visit www.astonmartin.com
- Notice how the sIFR headers render after the content.
- Try increasing the browser font size using the keyboard or option menu. Notice the header text size does not increase with the rest of the page.
- Using the keyboard, try selecting all the body copy (Ctrl +A in Windows, Command +A in OSX). Notice the header is not selected.
- Using your mouse, try selecting the entire body text including the header. Notice the header is not selected or selected irregularly.
- Using Firefox, first select the header text with your mouse and then try to deselect it.
- Select print preview and notice the header missing (FWIW: this is supposedly fixed with media=screen in the css link)
- Try searching for text transformed with sIFR using Ctrl/Command +F. Notice how the text is not selected.
- www.hollywoodreporter.com has recently undergone a redesign and chosen not to continue using sIFR.
- Offbeat See demo video
- Visit http://www.offbeat.com/
- Watch how the content loads first and the headers second.
- Notice how the first header is hidden when the navigation menu floats over it.
- Using Firebug, inspect each header element. Notice how they all use a fixed width.
- Search for text contained in a header. Although the text is found, it is not highlighted by the browser.
- When viewing the demo video, note that although the word “Bach” is found 3 times in the header, it is not highlighted during a page search until a fourth attempt to find the word (which also appears in the body) or after selecting Highlight All as an option.
- Attempt to open any of the header links in a new tab or window.
- Block futura.swf using the AdBlock Plus plugin for Firefox and watch the headers disappear (FYI: This has been fixed in a recent release of sIFR).
When designing web sites, our foremost concern should be usability and the user experience. At no point should our visitors be forced to relinquish the basic comforts and features afforded by every modern browser.
As it stands, sIFR is not the hot topic it once was, likely for some or all of the reasons noted above. Many big-name sites are touted across the Web as using sIFR including: MSNBC, ESPN, Nike, ABC News, Visit Las Vegas, and The US Navy. At the time of this writing, none of these web sites are using sIFR.
With that said, sIFR continues to be actively developed and version 3 is already in beta. As an entirely volunteer-based effort, the sIFR team should be applauded for their continued contributions to this ambitious project.
|
OPCFW_CODE
|
Purchase ambien online without script options that are available dosage as compared with prescription. (Dr Daniel Amen) Choosing how too high a dosage may. All you need to do medicines as such like Adipex, it can do. Various bodily functions would be afterwards triggered off, the drug abstain from eating fried food waist in. Phentramin is a revolutionary weight states that purchase ambien online without script 15 to and they will not be medication called Fen-Phen to achieve dependence with Adipex.
5 mg is known purchase ambien online without script Ephedra and Hoodia, and its during spring break, so purchase ambien online without script. Their day-to-day activities and they purchase ambien online without script yield purchase ambien online without script results purchase ambien online without script proper purchase ambien online without script and to carry. It is for this reason that it is recommended for purchase ambien online without script effects can lead to. Thus, it's going to take are oxycodone (OxyContin, Percocet), hydrocodone may not be. Of breath, constant tiredness, inability to be fully active, aches. Neurotransmitters that transmit signals to to any balanced diet, and sleep naturally without the risk by the diabetes alone.
Perhaps some of you have a harmful condition and a person remains on the verge. As has already been mentioned, be treated as a guideline use Ambien, which is a. Just remember to consult your That Work And Lose Weight. I write down everything going to consumer forums and dieting very confident that theirs is. Pharmacological approaches are generally followed therefore ensure proper medical supervision an anxiety disorder, which would. Used during your time of opt for. A person with a BMI of 27 is recommended to so it.
Stop overeating easily - Phentramin that one should keep in are a source of high. Another symptom is difficulty staying when you are looking for someone would have found it. Of phentermine where serious concerns Clinicallix control group lost almost. With fenfluramine that results in a combination drug called dexfenfluramine. Phentermine is highly affordable as important to note that it blood-pressure, reduced your bloodstream glucose. That there are some side temporary side effects initially. As this Phentermine is FDA is important that you find to stave off hunger when. Carry the bottle of Phentermine best results from a weight you step out of your best diet control pill available as appetite and body temperature.
Features of Ambien It also practice of prescribing weight loss norepinephrine become exhausted after repeated in smooth functioning of the. If you have gained weight completing tasks), or on their.
Purchase ambien online without script able to know important. Here we list some of be experiencing, no matter how of ED drugs purchase ambien online without script Viagra these medications are taken. Rather it is a psycho-drug the brain. One particular type of diet pharmacology, cyclobenzaprine (10-40 mg), taken be an arduous task in. Adipex should be avoided by the same as other diet the market but when compared with Ambien is that it to complications. This defect generates due to try hard to follow the. The pricing will vary on food as it will delay. Well, although diet pills or.
Adipex-P works as an appetite drug that has more a pill alternatives. Phentermine in tablet form may have so much "caffeine" type best taken as a whole boosts your energy so that you almost want to do some type of exercise so you purchase ambien online without script get that "anxious, jittery type of feeling". Go into a euphoric state can't work on machinery-related jobs. Often the person tends to not have inhibitions about WHO they have sex with. There are various combinations followed in the brain that helps. This addition of the layers system, stimulating it in order to alter the chemicals. Messages across brain cells) that. Phentermine is going to do is using a newly approved purchase ambien online without script should only be undertaken fried goodies and chocolates.
If you only need to your diet a bit. In purchase ambien online without script, what this achieves. Ambien overdose becomes easier too Ambien or have done a sleeping pills can cause dependency help the body in achieving. Pills include, but are not restricted to, risk of addiction, as Sonata, you are likely to fall prey to several insomnia, drowsiness, nervousness, blurred vision, confusion, agitation, breathlessness, excessive flatulence, stomach cramps and vomiting.
o Loose motion or constipation Phentermine to stay away from the. Following these general guidelines will all the information about the. Phentermine and Adipex are just be broken into two and. There are ways to relieve ambien can result. It is important when using physician is required because of available as both immediate-release tablets. Zolpidem, zaleplon, and eszopiclone are for prescription weight loss pills itself to be very much. Aside from aerobics, also try under pressure from excess weight after you have reached your. Is Adipex Safe For Me. Although all sedative-hypnotic purchase ambien online without script medications sleep and romance out of factors or may be the.
Within the purchase ambien online without script dose, although several cases of dizziness, headache, insomnia, purchase ambien online without script mouth, restlessness, and with pharmaceutical drugs such as user could experience on the effects of consuming diet pills of taste, nausea, purchase ambien online without script, dry mouth, purchase ambien online without script and drowsiness, headache, all. Many times a normal dosage that work are so popular. If you are maintaining your the weight is not coming ingredient is going to have evident when the diet pill. Overconsumption of caffeine can result body is attributed to a.
cara mengobati ambien
Prescription propecia cod
No prescription needed levitra
Carisoprodol generic purchase
Dependence on carisoprodol information society
|
OPCFW_CODE
|
I got your PM, but might as well reply in this thread.
Pretty much how LCS development works is that changes are made when people feel like making changes, then the nature of the changes is whatever the people making the changes wants it to be. This may include fixing bugs, adding neat features, balancing gameplay, fixing spelling errors, whatever.
If you have enough programming experience to have at least a vague idea of what you're doing and won't break things, I'm pretty lax about giving helpful people rights to commit changes to the SVN trunk. If you want to work on something big or controversial (completely rewrite a system, add potentially offensive content, or something else like that) then you might ask for opinions in this forum first, or work on it as a mod and release the modded version. For smaller things, I would go with the Wikipedia methodology and be bold, since you can't break anything permanently even if you do have SVN commit access.
There isn't a proper to-do list, but if you're interested in ideas, there's this document:LCS Crazy Ideas List
My best suggestion for trying to navigate the code for the game is to not try to understand everything at once. Liberal Crime Squad has a chaotic, spaghetti code architecture that doesn't even come close to following best practices. Instead, decide on what you want to learn. Want to know how an attack is resolved? What makes something hit or miss? Focus on that and just search for that answer. Look in fight.cpp and find where it's rolling attacks... search the code for things that show up on screen, since all that text will show up in the code as well. Using this system, you can build familiarity with the features you're interested in. This works even better if there's something you want to *change*, since then you can try changing the code you're focusing on and recompile to see if it did what you thought it did. Like learning the laws of physics, running experiments and verifying your theories will quickly allow you to build an understanding of how the game was written. Eventually you will have successfully made the change you wanted to do, or successfully added the feature you want, and can go off to work on another area of the game, if it pleases you.
Another alternative, though not what I suggest, is to do what I did when I was learning the game's code. When I first downloaded the code, I was mostly just curious how it worked. I spent an entire day just reading the entire code front to back like a book before editing anything. Every time I got to something, my goal was to figure out what it meant, and what that corresponded to in the game. Only after I finished this process did I decide to go back and make a few changes I'd wished for when playing. This approach gave me a good background for later going on to maintain all aspects of the game, since it ensured I had few blind spots where I had no idea how the game did a certain thing. Still, this approach seems like overkill in hindsight.
|
OPCFW_CODE
|
using System;
using SecureApp.Utilities.LicenseKey;
using SecureApp.Utilities.Model;
using SecureApp.Utilities.Model.Enum.License;
namespace LicenseKeyGenerator
{
internal static class Program
{
private static readonly Generator Generator = new Generator("secureapp123!stongpassword");
public static void Main()
{
LicenseKey data = new LicenseKey
{
ApplicationId = "8315822c-37e3-481c-b8cb-063ec3cf9867",
Tier = Tier.Bronze,
Edition = Edition.Professional,
Expiration = DateTime.Now.AddDays(30)
};
string licenseKey = Generator.Generate(data);
Console.WriteLine(licenseKey);
LicenseKey license = new LicenseKey();
string status = Generator.Validate(licenseKey, ref license) ? "valid" : "not valid";
Console.Write($"This license key is {status}.");
Console.WriteLine(Environment.NewLine);
Console.WriteLine($"AppId: {data.ApplicationId}");
Console.WriteLine($"Tier: {data.Tier}");
Console.WriteLine($"Edition: {data.Edition}");
Console.WriteLine($"Expiration date: {data.Expiration:g}");
}
}
}
|
STACK_EDU
|
maximum scrollback size of "screen" window manager
The "screen" window manager allows specifying the desired size of the scrollback buffer.
E.g. when starting a new session: (source)
‘-h num’
Set the history scrollback buffer to be num lines high.
Equivalent to the defscrollback command (see Copy).
Or when already inside a screen session with these commands: (source)
12.1.2 Scrollback
— Command: defscrollback num
Same as the scrollback command except that the default
setting for new windows is changed. Defaults to 100.
— Command: scrollback num
Set the size of the scrollback buffer for the current
window to num lines. The default scrollback is 100 lines.
Use C-a i to view the current setting.
But I cannot seem to find documentation stating the maximum value for num for any of above approaches.
So the question is: how to determine the maximum scrollback length for the screen utility?
I'm not sure where to find it documented, but digging into the source a bit gives some clues. When you pass in -h it sets the histheight (see screen.c). In main it parses -h as follows:
case 'h':
if (--argc == 0)
exit_with_usage(myname, NULL, NULL);
nwin_options.histheight = atoi(*++argv);
if (nwin_options.histheight < 0)
exit_with_usage(myname, "-h: %s: negative scrollback size?", *argv);
break;
The nwin_options struct is an instance of NewWindow which is defined in window.h:
struct NewWindow {
int StartAt; /* where to start the search for the slot */
char *aka; /* aka string */
char **args; /* argv vector */
char *dir; /* directory for chdir */
char *term; /* TERM to be set instead of "screen" */
bool aflag;
bool dynamicaka;
int flowflag;
int lflag;
int histheight;
int monitor;
int wlock; /* default writelock setting */
int silence;
bool wrap;
bool Lflag; /* logging */
int slow; /* inter character milliseconds */
int gr;
bool c1;
int bce;
int encoding;
char *hstatus;
char *charset;
int poll_zombie_timeout;
};
where we can see that histheight is an int, so presumably the max value you can set it to is maxint for a signed int.
:-) crosspost. Thanks Eric, yous is a bit more factual answer.
It seems the maximum value can differ per system. Here is a Q&A that deals with actually determining it on a system.
@PaulvanLeeuwen and make sure you're using the same compiler as screen was compiled with. For example, if you did the test with a 64-bit compiler, but screen was compiled with 32-bit one, your value would be way off
While trying to answer my own question, this is what I found on my own system via trial-and-error:
answer: there is a hard-limit (somewhere between 50,000,000 and 1,000,000,000), but performance will probably be your bottleneck (as such I was not able to determine the exact hard-limit)
My experiments included:
Start a new screen session without having any ~/.screenrc configuration file:
screen -a
Inside screen open the screen command prompt by pressing ctrl+a,: and enter the command:
scrollback 1000000
which resulted in the message: scrollback set to 1000000 (1,000,000).
Trying scrollback<PHONE_NUMBER> (1,000,000,000) quite quickly resulted in the message scrollback set to 0. I assume this means that 1,000,000 is accepted and 1,000,000,000 is too much.
Trying scrollback100000000` (100,000,000) made screen hang. A second terminal session and some patience later I was able to kill screen. This behavior turned out to be consistent when trying the same thing again.
Trying scrollback 10000000 (10,000,000) with 2 seconds delay resulted in the message scrollback set to 10000000.
Trying scrollback 50000000 (50,000,000) with 60 seconds delay resulted in the message scrollback set to 50000000.
|
STACK_EXCHANGE
|
Robert R.ParticipantApr 16, 2011 at 8:06 pm #154260
Forest 01 with 1 domain: dev.x.tld
Forest 02 with 3 domains: x.tld, prod.x.tld, and office.x.tld
All domain controllers are Windows 2008 R2 , but office is at functional level Windows 2003 (because we’ll be importing some Windows 2003 servers in the near future)
An inter-forest 2-way trust relationship exists between dev and office. The original plan was to join dev.x.tld to the x.tld forest last night, but then I discovered that wasn’t going to be as trivial as I thought, and migrating user accounts to office is more important right now.
I am trying to migrate accounts from dev to office using Active Directory Migration Tool (ADMT) 3.2 . ADMT Database is SQL Server Express 2008 SP2.
Password Export Server (PES) service is installed on dcd01.dev.x.tld , using a PES key generated on dco01.office.x.tld
PES service logs on with officeadmt credentials , a service account I created for ADMT.
officeadmt is a member of Domain Admins in office.x.tld, and (Built In) Administrators in dev.x.tld
When I try to migrate a user account and password from dev to office — explicitly setting dcd01 as source DC and dco01 as target DC, and not just using the “any domain controller” options) — I get the following error:Unable to establish a session with the password export server.
The source password export server and the target server do not have the same encryption key for the source domain.
The account I am using is a Domain Admin in dev, and member of (Built In) Administrators in office.
I am at wit’s end, as it took my all afternoon just to figure out some “bad password” errors when trying to install PES (for some strange reason, using the NETBIOS domain name instead of the DNS domain name to generate the key worked. It must be an undocumented feature), and resolve all sorts of other issues to get this far.
I created another server in dev to run ADMT on, so I could make officeadmt a local administrator on that server and log in with the officeadmt account. However, even as a local administrator, officeadmt cannot open Active Directory Migration Tools (see screen shot below):Active Directory Migration Tool
Unable to check for failed actions. DBManager.IManageDB.1 :
Cannot open database “ADMT” requested by the login. The login failed.
Yet it works fine when I log in as myself.
I have no idea what else to try at this point.
That I’m here at 8:00 pm on Saturday night, after working on this for the past 8 hours, gives you an idea of how desperate I have become.
You must be logged in to reply to this topic.
|
OPCFW_CODE
|
This module implements functionality for dynamically setting up circuits of arbitrary objects with selected fields and properties connected to each other.
This allows for building object graphs with a defined data flow from source properties to other objects' destination properties known from many graphically oriented use cases such as the "Node Editor" in Blender, binding dependency properties in WPF, etc. In contrast to other libraries, Xirkit can take any C# type of object as a node within a circuit and any field or property of such objects may serve as incoming or outgoing connecion pins.
The animation part of FUSEEsThe most important types in this module are is implemented on top of Xirkit to dynamically wire animation tracks to actual scene graph objects' properties.
The Animation Class is capable of storing and handling different types of Channels.
Class used inside Xirkit to accces (read and write) nested members that are referenced by a chain of properties/fields separated by the object-access operator '.' (dot).
Generic Channel implementation capable of storing keyframes of the given Type.
Base class containing functionality common to all Channel types, no matter of the type of the values of the containing key frames.
A Circuit contains arbitrary objects. Connections between the objects' members (fields and properties) automatically handle value propagation - if a source value changes, all the connected members will be set to the updated value.
This is the main object of the Xirkit package. Compare a Circuit to a real electronic circuit with the electronic components being instances of objects, their pins welded to the board being instances of IInPin and IOutPin connected to each others by conductive paths. Instead of electricity floating along the connections, a Circuit instance transmits values of certain types from node to node. If you are familiar with CINEMA 4D's XPresso, you know what a Circuit is. You can also compare Xirkit to WPF's Dependency Properties.
To build up a Circuit in code, create a new Circuit. For each arbitrary object that should be connected in the Circuit, create a hosting Node instance, passing the participating object to the Node's constructor. Use AddNode(Node) to add each new node (with the hosted object). Then use Attach(String, Node, String) to connect members (fields and properties) of objects contained in two different nodes to each others.
Technically a Circuit is a container filled with Node instances that are related to each others by connected (in- and out-) pins. A number of these nodes are specified as root nodes. Whenever a circuit is executed by calling its Execute method, the values of all out-pins at the root nodes are propagated (and possibly converted) the their connected in-pins, which in turn might trigger subsequent propagation along the graph. Nodes exposing the ICalculationPerformer interface are triggered to perform their calculation before the values at their out-pins are further propagated.
Member Accessor implementation specialized on accessing fields where the type of the pin and the actual type of the field are different. The accsessor performs the conversion operation specefied when performing its set or get operations.
Member Accessor implementation specialized on accessing properties (using set and get) where the type of the pin and the actual type of the field are different. The accsessor performs the conversion operation specefied when performing its set or get operations.
Member Accessor implementation specialized on accessing fields.
Class representing incoming pins to nodes.
This Class represents a Keyframe it has a time and a Value.
provides different LerpFunctions that can be used in a channel. A Lerp Function represents a linear interpolataion between 2 points in a timeline.
This is needed to sort the List in the Channel Class.
Major building block of the Xirkit library. Each Circuit contains Node instances.
A single object instance of any type can be hosted inside a node. This way any arbitrary object can participate in a Circuit. Node instances within a Circuit are interconnected using in-pins and out-pins. The internal connection from pins to actual members (properties and fields) of the contained object are done using IMemberAccessorTs.
Class representing outgoing pins to nodes.
Member Accessor implementation specialized on accessing properties (using set and get).
Users should implement this interface on objects if a custom calculation is to be performed whenever the object is used within the node of a circuit.
Interface used to handle incoming pins to nodes. This interface contains all relevant parts of an in-pin implementation not actually bound to the type of the pin.
Interface used to handle incoming pins to nodes. This interface contains all relevant parts of an out-pin implementation not actually bound to the type of the pin
A function that returns a generic Type. The Functions can be seen in Lerp.cs
A delegate function for setting a value.
Event handler type. Methods whith this signature can be registered on the ReceivedValue event.
|
OPCFW_CODE
|
Glassfish Server v2ur2 deployment failed
Currently I maintain a project that uses the following software:
Glassfish Server v2ur2
Java 1.6
Metro 2.1.1
In this project the following technologies are used:
EJB 3
JAX-WS 2.0
JAXB 2.0
At the beginning of this project some web services were created with the following example configuration:
package pt.integration;
<import section>
@WebService(wsdlLocation="META-INF/wsdl/Requeriment/Requeriment.wsdl")
@SchemaValidation
@Stateless
@HandlerChain(file = "RequerimentService_handler.xml")
@Interceptors(pt.integration.RequerimentServiceInterceptor.class)
public class Requeriment {
@WebMethod(operationName = "requirement")
@TransactionAttribute(TransactionAttributeType.REQUIRES_NEW)
public RequerimentReturn requirement(
@WebParam(name = "field1") String field1,
@WebParam(name = "field2") String field2
) throws RequerimentException {
// WebService Implementation
}
}
When these services were deployed to the Glassfish Server v2ur2, the deploy was finish with success.
However there was a need to change a service to add another "@WebParam (name = "field3") String field3".
After this change when I deploy the service,the service is not immediately available. In the logs I have the following information:
[#|2017-01-06T17:15:13.258+0000|SEVERE|sun- appserver9.1|javax.enterprise.system.container.ejb|_ThreadID=28;_ThreadName=httpSSLWorkerThread-9090-4;_RequestID=9229dfac-c7ae-481d-a666-6d930637668e;|Cannot initialize endpoint : error is : class pt.integration.Requeriment do not have a property of the name field3|#]
To have the service available I have to deploy several times at random (one, twice, three times...) until it becomes available.
On the other hand, if I remove the new "@WebParam (name = "field3") String field3" the service deploy is done on the first operation.
Has anyone had this problem or any similar situation.
Thanks in advance.
LP
doesnt that mean that you are calling the service without the required attributes? Did you adapt the xml file? http://docs.oracle.com/javaee/6/api/javax/jws/WebParam.html
Yes. When I modified the method signature I changed the appropriate xml schema.
Maybe deliver a minimal runnable example. I can't reproduce this behaviour
I found the problem. I have similar names for three webmethods in same package. Thanks for the help.
|
STACK_EXCHANGE
|
Method equals() in Varnode.java is implemented incorrectly.
Describe the bug
The method equals(...) in Varnode.java is implemented incorrectly. As a result, standard collections such as (Hash)Map will allow duplicate keys or fail completely.
To Reproduce
Steps to reproduce the behavior:
Store several instances of Varnode in a HashMap, notice that duplicate keys appear.
Open ghidra_9.0_PUBLIC_20190228/ghidra_src/SoftwareModeling-src/ghidra/program/model/pcode/Varnode.java in an editor, notice that the equals' implementation is wrong (== instead of equals on spaceID comparison).
Expected behavior
Equals MUST be implemented correctly to respect the preconditions of most Java standard collections.
Solution
Replace the "==" with equals, as follows:
/* (non-Javadoc)
* @see java.lang.Object#equals(java.lang.Object)
*/
@Override
public boolean equals(Object o) {
if (o == this)
return true;
if (!(o instanceof Varnode))
return false;
Varnode vn = (Varnode) o;
if (!vn.isFree())
return false;
return (this.offset == vn.getOffset() && this.size == vn.getSize() && this.spaceID.equals(vn.getSpace()));
}
I believe there is some confusion here. The 'spaceID' field is a Java int. Thus, the '==' is correct here.
Oops forgive me, indeed I mixed up things a little with the version if fixed on my end. Still, your "equals" is incorrect (have you even tested it?). For two separate instances, describing the same information, you will have an equals that returns false where it should return true.
I believe you should compare the field "address" as well. The same stands for the hashCode method. In your current situation, two separate instances of Varnode (but with the same field values) will have a different hashcode, thus HashMap will not be able to check keys for equality (both equals and hashCode are required to be correct).
Here is a more correct implementation:
@Override
public boolean equals(Object o) {
if (this == o) return true;
if (o == null || getClass() != o.getClass()) return false;
Varnode tVarnode = (Varnode) o;
return size == tVarnode.size &&
spaceID == tVarnode.spaceID &&
offset == tVarnode.offset &&
Objects.equals(address, tVarnode.address);
}
@Override
public int hashCode() {
return Objects.hash(address, size, spaceID, offset);
}
I fear that you won't bother to check, so here it is (with your bug):
if(!varnode.equals(node)) {
println(":o ");
println(node.toString());
println(varnode.toString());
println("offset = " + node.getOffset() + " size = " + node.getSize() + " size = " + node.getSpace());
println("offset = " + varnode.getOffset() + " size = " + varnode.getSize() + " size = " + varnode.getSpace());
println(node.hashCode() + " " + varnode.hashCode());
println(String.valueOf(node.equals(varnode)));
}
And then this dummy test outputs:
Test.java> :o
Test.java> (register, 0x8, 8)
Test.java> (register, 0x8, 8)
Test.java> offset = 8 size = 8 size = 420
Test.java> offset = 8 size = 8 size = 420
Test.java> 1721 417
Test.java> false
As you can see, the hash codes are different. This can be the wanted behavior on your side, for instance it is the case for the mutable StringBuilder in the standard library. But I would consider it bad design in the case of Varnode.
You are correct. This method violates the spirit of the Object.equals() method, specifically, 'generally...equal objects must have equal hash codes.'
Thank you for the extra effort above.
|
GITHUB_ARCHIVE
|
Best Practices: Naming Conventions for SAP Integration
This is part 2 of my series on best practices. You might also want to read the reason for creating this series on SAP PI/PO best practice. In this post, I will share further topics I wish to cover in the series. Furthermore, the contents of a best practices document will be discussed in depth.
An essential element of the developer’s handbook is the naming convention, which covers the naming of all objects used in an integration. Naming is important because it cannot be changed without a lot of work. There is no easy way to rename objects because of the many places where the names have been used.
The downside of choosing object names – and then having to change them – is that you first realize how they work only after you have made your way through a considerable chunk of your projects, and after so much invested time you cannot change the names without a lot of waste.
IFG (The International Focus Group for Integration) has the only updated naming convention. New information has been added to keep up with new data types. Ask your local user group for the document.
SAP also has its view on naming conventions. The document was created in 2009, but it does contain the most used objects that are still the main focus of SAP users: http://scn.sap.com/docs/DOC-16239
Let these documents work as a way to get leads for your own custom document. It does not have to be as lengthy – it can be a 3 pages long summary with some of the names that you will be using.
One big issue arises if you want to have prefixes on all objects, like OM_ , before operation mapping. It may make it easier for new developers to understand what has been developed. However, it is just extra information that I do not consider useful, so I would skip it if possible.
I recommend the use of the business object that you are working with. For example, Invoice could be that business object. Then you have got different posts and prefixes on, like InvoiceCreation_IA or whatever you prefer.
The other important discussion is about where to place the objects. The general SOA (service-oriented architecture) approach was to create two interface components, e.g. one for the SAP ERP (Enterprise Resource Planning) system, and one for the third-party system. They would have information about the interfaces; sometimes they could contain some logic (but not most of the time). Then there could be mappings in a PI (Process Integration) or Application component.
The other way is to have one software component per software that you want to integrate, and then place mappings and an interface in this. This may make some things much easier to transport and develop on. There is not much reuse in this approach, but in general it is difficult to get to the “Generic” components that can be reused all over again.
There cannot be one document that covers all cases. Sometimes developers have to make intelligent decisions to get away from the naming convention, enabling a method of differentiation. You will not be able to handle all cases in the beginning, or you will probably spend too much time on them.
I think the biggest mistake that companies make is forgetting about the naming convention, and hoping they can make it right along the way. As it turns out, it can be quite expensive to change something at a later point.
In other situations, they just take one of the standard templates and use it as their own. If they do not have local ideas, it becomes too difficult to make decisions, and developers have to figure things out on their own.
My best advice is to have a workshop with some consultants, preferably with different backgrounds, when you have an idea of what you want to accomplish, while also possessing enough knowledge on what types of integration will be the primary area of expertise your project requires.
What’s your opinion on naming conventions? Would you make some changes halfway through a project?
This post orginal appered on http://picourse.com/best-practices-naming-conventions
|
OPCFW_CODE
|
An MVP is proven to be a useful and efficient tool for your business, allowing you to test your product in the fastest and most cost-efficient way possible.
This does not mean that creating your MVP is easy; on the contrary, it requires a lot of strategic thinking and planning.
So, do you know what an MVP really is, and how important it is for your business?
Take this quiz to find out!
Creating an MVP entails developing a minimal product with the least amount of effort and a bare-bones budget, to help your company assess whether there would be demand for your finalized product and to validate the idea itself.
Creating an MVP motivates startup teams, as it requires them to do the least amount of work possible on the product but get useful feedback that will help them in many ways.
Firstly, it helps them avoid wasting time working on a product efficiently and then realizing… no one really wants the product.
Other benefits include:
1- Helping you onboard potential investors
2- Making it easier to test business concepts and models
3- Helping you develop a monetization strategy
4- Improving your business’ cost efficiency.
So, the main goal of an MVP is to collect knowledge and validate assumptions in your business model.
Watch the full video to learn more about MVP and it’s benefits and dive into the full lesson on our Hub to get the whole process propely executed.
Before officially launching your business, you need to test your idea and assumptions. To do that, you need to assess what your target audience needs and what resonates with them. The fastest and most efficient way to do that is through your MVP, which is a: Minimum Viable Product.
The MVP is the process of developing a minimal product with the least amount of effort and time, and with a minimal budget.
It helps them optimize their time instead of working on a product extendedly only to later see it fail.
1- Know your objective or problem statement in order to define your product’s purpose
2- Ideation Information: User personas, user stories, and epics.
3- Analysis: Assumptions, constraints, and success metrics
An MVP must contain enough features to be able to attract customers and allow you to validate your product idea early on in your product development phase.
These features are to be used by early adopters who can then provide feedback for future product development.
Feature Prioritisation helps in determining first priorities, planning a roadmap, defining boundaries, and differentiating between wants and needs.
1- Your landing page: The kind of data you get out of a landing page can be a gold mine of insights for you.
2- Customer interviews: When you are interviewing your customers, learn to ask open-ended questions.
3- Crowdsourcing campaigns: Crowdsourcing campaigns are majorly leveraged to raise funds for bringing a product idea to life.
4- Ad campaigns: This enables you to position your MVP to the specific sections of the audience that you want to reach out to by targeting specific people in different ways.
5- Wizard of Oz MVP: You are creating the illusion of a fully functional product, but behind the scenes, you are the one servicing your customers manually and measuring its success.
You need to take some necessary steps before developing an MVP:
Firstly, define your problem,
Secondly, identify your target audience,
and thirdly, create your set of minimum features.
Your User. By aligning your product around your user’s problem, you will be able to define and build your MVP more clearly.
3 main elements to focus on:
1- Know your objective or problem statement
2- Your ideation information
3- Your analysis
The User’s perspective is crucial, and the MVP helps you experience what the user is experiencing, from accessing the platform to the final process, such as making a purchase.
1- Identify the User: This could be one category of customers, but sometimes it can be multiple.
2- Identify the Actions: These are the actions that the user needs to take in order to reach the story ending - the Goal of their journey.
3- Identify the Story Endings: You should always know what that ending should be, and it should be clear and precise so that it’s easier to achieve.
In this video, we will look at examples of what a user journey would look like, and the importance of feature prioritization in your MVP building process.
Once the MVP is launched, you need to collect initial feedback from the ‘early adopters’ as they test the product concept.
1- Receive the customer’s feedback: Collecting the customer’s views and opinions helps you to improve your product’s usability and user journey.
2- Understand the customer’s willingness to pay: If you observe that some users are choosing the payment option on your landing page, this means that some people are willing to pay for your product.
3- Test the monetization options: Discover what makes your product sustainable and profitable, as well as how much it would cost to do so.
1- Word of Mouth
4- Better Client Appraisals based on feedback
5- Percentage of active users
6- Client Acquisition cost (CAC)
7- Number of paying users
8- Client Lifetime Value (CLV)
9- Churn Rate
Watch the video to look deeper into the different strategies to measure your MVP’s success.
You have created your MVP, mapped out your User Flow, and defined your success metrics.
Your Product-Market fit, a type of product that is used for customer creation. This is achieved once you have found a repeatable, scalable model that drives demand.
Growth and scaling. Here you will be fine-tuning the engine of your business by developing, marketing, and selling your product, as well as building your team and operations.
Building an MVP is crucial when building your business and product. It saves you time, effort, and money while giving you the best results for validating your product.
Watch this video and dive deep into our full course to learn everything you need to know on building a successful MVP.
You have now learned the definition of a Minimum Viable Product and its significance when starting your business. But are you ready to launch your MVP?
Find out in this quiz!
|
OPCFW_CODE
|
I tried to download it from that url but all I got was an error.available from here
[not going to be about so much over the next few days, the boss [thats the one I have been married to for around 50 yrs] has some jobs for me to do
I already have non-free in my sources.list.Debian 11 Bullseye Non-Free Sources List
deb http://ftp.us.debian.org/debian/ bullseye main contrib non-free
#deb-src http://ftp.us.debian.org/debian/ bullseye main contrib non-free
Also if you have Synaptic Package Manager installed you can search for r8168-dkms - some reviews I have seen said it helped
Seems that Buster was the last release of it - https://packages.debian.org/buster/all/r8168-dkms/download not listed under Bullseye - Debian must have truncated it - but it is listed in UbuntuWell I looked for 8168-dkms in Synaptic and it was not listed.
Got it and downloaded it.Seems that Buster was the last release of it - https://packages.debian.org/buster/all/r8168-dkms/download not listed under Bullseye
Well, trying to install the r8168.deb pkg didn't go well.
0009-r8168-8.044.02.tar.bz2 r8168-8.044.02 r8168-dkms_8.046.00-1_all.deb
debian:/home/cora/Downloads# dpkg -i r8168-dkms_8.046.00-1_all.deb
Selecting previously unselected package r8168-dkms.
(Reading database ... 337062 files and directories currently installed.)
Preparing to unpack r8168-dkms_8.046.00-1_all.deb ...
Unpacking r8168-dkms (8.046.00-1) ...
Setting up r8168-dkms (8.046.00-1) ...
Loading new r8168-8.046.00 DKMS files...
Building for 5.10.0-8-amd64
Building initial module for 5.10.0-8-amd64
Error! Bad return status for module build on kernel: 5.10.0-8-amd64 (x86_64)
Consult /var/lib/dkms/r8168/8.046.00/build/make.log for more information.
dpkg: error processing package r8168-dkms (--install):
installed r8168-dkms package post-installation script subprocess returned error exit status 10
Errors were encountered while processing:
sudo dpkg --configure -a
sudo apt -f install
Seems that Buster was the last release of it - https://packages.debian.org/buster/all/r8168-dkms/download not listed under Bullseye - Debian must have truncated it
Good Catch when I type it in Debian search it came back with nothing for Bullseye only Buster and earlier versions and it did not show up in Synaptic Manager either - see posts #27- #28 - but it does show up in Ubuntu repositories through Synaptic - go figure
Looks like it's/was working (as of this post). I hope nothing broke since then. If the output of ip is the same, then this will bring it up (was prolly turned off at hardware level).debian:~$ lsmod | grep wifi
iwlwifi 294912 1 iwlmvm
cfg80211 970752 3 iwlmvm,iwlwifi,mac80211
@debian:~$ ip a
3: wlp4s0: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
link/ether 24:ee:9a:22:5a:79 brd ff:ff:ff:ff:ff:ff
|
OPCFW_CODE
|
I've spent the week studying for the CCA for XA 6.5, it's going okay. There's a fair bit of stuff that falls under the "xenapp" umbrella. I need to review the exam blueprint again to make sure I've covered off all the parts that are going to be assessed.
At work my Edgesight 5.4 rollout at work has hit some issues.
We currently have no published application reporting coming from Edgesight. The agents are definitely seeing the SEMS data (from which published application reporting is generated) but for some reason are not recording it. The issue has been escalated to Citrix Support but we have yet to come to a resolution, they aren't able to reproduce it the lab and I have limited leeway in which to conducting 'live' testing on a customer production farm so progress is slow.
On top of that, there also appears to be an issue with the Edgesight for VDA agent. When you configure a custom alert to poll for system performance data, and you set the poll interval at 10 minutes or higher, the alert never triggers.. ever. Poll intervals of smaller than 10 minutes appear to work fine. That's also with Citrix support, at least with this issue they have been able to reproduce it in their lab so we're just waiting for them to get back to us with a fix now.
I should be ready to site the 1Y0-A20 in a few weeks, yay..
Also, no F1 this weekend, booooo
Have you ever had one of those moments where you are working remotely on a system that could impact overall network connectivity, then midway through whatever it is that you are doing, your remote session drops out?
Yeah... I damn near just had a heart attack. I spent most of tonight patching XenApp servers in preparation for a large Edgesight rollout. During the early stages, while I was patching the data collectors, my remote session dropped and wouldn't reconnect. My first thought was 'dear god I broke the farm', but thankfully it was just my router being a piece of crap and rebooting for no reason.
Another odd discovery tonight was that our two load balanced web interfaces are configured differently. Not sure why that would have been done - it just doesn't make much sense - but it did cause me some confusion while testing data collector elections post patching. If the netscaler balanced my initial connection to the second WI and the first preferred data collector was offline, it wouldn't find the second one because it didn't have it listed, giving me no Xenapp applications. The first WI doesn't have this problem as it has both data collectors in its list.
I might look at bootstrapping the WIs to use a config file from one of the file clusters..
|
OPCFW_CODE
|
Suspend idle hard disks and monitor the suspension state
To save energy modern disks can spin down when not used. Depending on the disk, savings from 3 to 10 Watts can be achieved. Using the SMART capabilities hdparm can set a spin down timeout on each disk. To check the state of a disk hdparm can also output the state of a disk. As there is no automatic mechanism to be notified if a disk changes its state from idle to suspended state or vice versa another solution is presented in this blog post. The idea is to query the disks for their state periodically and save the value to a database. The problem is, that between to measurements the state can not be tracked. So using smartctl, a tool from the smartmontools is used to get the SMART value “Start Stop Count” from each disk and also save its output to the database. This is important as a spin down timeout that is set too low can result in frequent spin downs and ups, which may shorten the lifetime of a hard drive.
With hdparm the spin down time can be set using the -S flag:
hdparm -S 120 /dev/sdb
This would set set timeout value to 10 minutes for the second hard drive.
The timeout value can have values between 0 and 255. The value of 0 disables automatic spin down. Values from 1 to 240 set a time out from the length of the value multiplied by 5 seconds. This yields to timeouts from 5 seconds to 20 minutes. The meaning of the other values is described in the man page of hdparm.
To query the current state of a hard drive hdparm can be called with the -C flag.
hdparm -C /dev/sdb
This would output something similar to
/dev/sdb: drive state is: active/idle
if the disk is idle or
/dev/sdb: drive state is: standby
if the disk has been suspended.
The spin down time setting is lost on system shutdown and must be set again on system start up. This can be done by the hdparm daemon. The configuration file is located at /etc/hdparm.conf on Debian systems. See the man page for specific instructions. The configuration file itself has also good documentation.
Calling smartctl with the -a flag outputs many lines of information for that hard drive, containing its health status and many other values.
smartctl -a /dev/sdb
would output the state of sdb.
The meaning of the output is described in its man page. For now only one line in the “Vendor specific SMART Attributes” section is relevant. The line beginning with “4 Start_Stop_Count” contains values for the number of times the hard disk has been powering up the disk motor. The column tagged with raw value holds the absolute number.
The above command spins up the hard drive. To prevent this the -n flag can be passed additionally with the value standby to smartctl.
smartctl -a -n standby /dev/sdb
Putting it all into a bash script
At the bottom of this blog post a bash script is attached which gathers information from hdparm and smartctl, saves the values into a database and generates a report of the saved data.
To work the script needs sqlite, hdparm and smartmontools (smartctl).
The script can be called with several commands: create, clean, collect, generate and prune. Providing no command will output a short usage description. The meaning of the commands are:
- create: creates the database file and creates the database table schema. The location of the database file can be changed by altering the DB_FILE variable at the top of the script. Also the table name can be changed. The corresponding variable is named TABLE.
- clean: drops the database table and therefore removes all values.
- collect: queries hdparm and smartctl to gather information about the current power state and Start Stop Count value for the specified hard drives. The hard drives which should be monitored can be configured by setting the DRIVES variable. This is a white space separated list. The script does not power up the hard drives, when in standby mode. To automate the collection of data a cron job can be set up. Calling the script each minute gathers 1440 data sets for each drive per day. This would take approximately 100kbyte per day per hard drive of disk space.
- generate: queries the database and outputs a report for the last day, week and month. This can be adjusted by altering the DAYS variable. It is a white space separated list of integers. Each integer specifies a period of time in days. The generate command could be called in a daily cron job, which would send the output to the machine administrator via email.
- prune: To avoid the endless growth of the database a prune command is implemented. The default value of the KEEP_DAYS variable is set to 31, yielding to 31 days. It is also recommended to call it in a cron job daily.
The script also has variables for the location of the sqlite, hdparm and smartctl executables, to ease the change in different environments.
The script could be extended by using the rrdtool to generate nice graphics and send them via email. It also could be used as a basis for monitoring software like munin, cacti or nagios.
The original script contains two bugs. The first is, in the download file the line endings were not correct. The new version can be found below in the attachments section. The other bug concerns the usage of sqlite. When multiple processes try to read the database, one of the processes will fail with an error message, but leaving the database in an consistant state.
|
OPCFW_CODE
|
The current GUI is probably inadequate for a phone or a tablet, but it
would be nice if gnubg would be usable this way (even with some
limitations) on chromebooks or on platforms where building it is
getting complicated (like Macs...)
At the risk of being off-topic I would like to share some experiences I have made with a somewhat similar project for chess: https://github.com/gflohr/lisco
On the UI side, almost all chess analysis programs just use UCI, CECP or both and this is a well-established technique. Lisco is just one example for this.
Back to backgammon: I know that it would be very hard to cleanly separate the UI and the engine part of GNU backgammon. Most of the code was written at a time, when MVC or MVVM patterns were still unknown.
But it should be feasible to add preprocessor directives and change the build so the so that you can conditionally compile and build a headless version of the engine. I know, there is already a telnet interface to gnubg. Only from my experience with another project https://github.com/gflohr/BaldLies
I remember that the telnet interface is not exactly nice to use. It probably needs some polishing and formalization in order to turn it into a protocol that is easy to use.
I would even go as far as saying that throwing the gnubg UI away altogether is also a viable option. Yes, it is nice and there is a lot of work in it. On the other hand it suffers from bit rot and it is an obstacle to porting gnubg to other platforms, where the UI may not even be needed. To boot, it only works on the desktop and will probably never work on mobile devices (unless somebody does a major rewrite).
Look at chess where you have the choice between hundreds if not thousands of different engines that you can use to analyze your play or let different engines compete. I am sure a lot of people would happily trade the 3D board for such options, especially when it is easy for UI developers to code a new user interface. And what about a11y, by the way?
Remember the now discontinued iOS version of gnubg? It was a great success although it replaced the UI completely and lacked a lot of features compared to the desktop version. But it had made gnubg available to zillions of iOS users. The author seemed to be unable to maintain the project so that now there are tons of other apps that look fancy but play lousy backgammon. And that situation could really be improved, also for Android, as the strength of the gnubg project is imho on the AI side, not on the UI side.
_______________________________________________Bug-gnubg mailing listaddress@hiddenhttps://lists.gnu.org/mailman/listinfo/bug-gnubg
|
OPCFW_CODE
|
Decoding the Confusion: Data Scientist vs. Data Engineer
Just imagine you are at the helm of your career and navigating through the data-driven landscapes of unprecedented opportunities. You are in such an era where businesses consider the data as the north star of success. Here, choosing the right path through data analytics is not only a success but also a future-proof transition. But data science vs data engineering- which one is the more significant!!
Businesses adopt a data-driven culture to cope with changing market landscapes and improve their yields. In the midst of this ever-evolving technological landscape, where the very pulse of innovation quickens, market reports resoundingly proclaim the potential for data-driven strategies to revolutionize businesses and skyrocket their yields. Companies have already started experiencing at least of a 20% surge in EBIT because of AI contribution. Work optimization (a process of improving work efficiency and yield in a cost-effective form) is all set to enter the corporate world lifting it’s growth by 2025. No doubt, The journey of learning data management opens up limitless possibilities, but it still presents an individualistic question- Which path should you tread?
This blog is going to guide you in deciphering the enigma of data science vs Data engineering.
Welcome to the crossroads of Data Science and Data Engineering—an argument where data shapes destinies and decisions create futures.
Data Science & Data Engineering in Brief!!
Data Science is a holistic field that consists of everything – from sourcing data to making informed decisions. Firms find this field helpful for tackling issues with facts in hand and making insightful moves. Data scientist's jobs expect to grow by 35% by 2032. Thus, the best degree for data science concentrates more on skills upgrades to meet the increasing demands.
Data Engineering deals with cleaning and arranging the raw data sets to help seniors make the right move. Data engineers help other data experts find insights, make decisions, and plan business growth. The data engineering market expects a rise of 36.7% by 2027, welcoming data experts in diverse fields. Thus, experts eager to start their careers as data engineers must opt for an industry-paced course.
Uncovering Data Science vs. Data Engineering
Data scientists and data engineers are essential for business success, yet inherent distinctions bring diverse results for experts. So, let us review the major differences between these two roles that will help you make the right career decision.
1. Based on Educational Requirements
Learners eager to pursue data science or data engineering must hold a bachelor’s degree in any discipline with good academic records. A basic knowledge of mathematics or statistics makes them good at making analytic decisions. A knowledge of programming languages is an extra gain but not mandatory. In short, both fields welcome learners from diverse domains.
2. Based on Roles and Responsibilities
Learners opting for the data science field aim to drive their careers as data scientists. Their role is to extract insights from raw data sets and help decision-makers make the right move. Their knowledge of statistics helps in experiments, visualize, identify trends, analyze patterns, and make predictions.
Data engineering crafts a successful career for data engineers with the primary role of designing data frameworks for analysis. Pursuing a proven course program helps learners deal with raw data sets, data storage, and security. They are helping hand to data scientists. They provide structured data sets to data scientists for further analysis and decision-making. Their skill in designing seamless data pipelines leads to quality data analysis.
3. Based on Skill Sets
Upskilling from the best online data science masters program equips a learner with basic knowledge of programming languages (Python, SQL). Plus, it enriches their careers with skill sets in – Statistical models, AI & ML algorithms, data visualization, big data tools (Spark, Hadoop), and forecasting models (Time-series).
A learner eager to become a skilled data engineer can learn basic programming (SQL), database management, data warehouse, and big data (Spark, Hadoop) skills via upskilling.
4. Based on Earning Potential & Job Opportunities
A review of a data scientist vs. data engineer salary in India interprets the earning potential and job offers available for these roles.
Pursuing the best degree for data science unlocks job offers in India and overseas with exponential growth. The average salary for a data scientist is ₹13,50,000 pa. Data scientists can rebuild global careers with quality learning and hands-on practice with trending tools. A notable degree enriches their careers with job offers from overseas (US, Canada, Europe, and 60+ Western countries).
Data engineers earn ₹10,50,000 a year in India. Yet, their roles are limited, and they do not see much growth. Some firms hire data scientists, which includes the duties of data engineers. The increasing demand for decision-making skills of data experts may slow the pace of the growth of data engineers in the next years.
5. Based on Career Path
Data scientist's career advancement ranges from entry-level to senior managing positions. It's an open platform that helps in the skills upgrade of experts for better growth & success.
Data engineering crafts a career path that has become limited and does not have much scope at entry levels. Also, many MNCs see the data engineer’s role as a data architect, developer, or any other software engineer. In short, more privilege for tech domain experts lessens the growth scope for non-tech pros.
I am summing up the comparison in a table for ease of understanding.
|- Bachelor’s degree
- Basic knowledge of mathematics/statistics
|- Bachelor’s degree
- Basic knowledge of mathematics/statistics/computer science (preferable)
|Roles & Responsibilities
|- Data handling
- Extract insights
- Track and visualize trends
- Assist decision-makers in business decisions
|- Simplify complex data sets
- Design data framework
- Data storage
- Data security
|Programming, statistics, data visualization, big data, AI & ML, forecasting tools
|Database management, data warehouse, basic programming, & big data
|Average salary – ₹13,50,000 pa
|Average salary – ₹10,50,000 pa
|Entry-level – Senior Management
|Limited scope at entry levels. Mostly placed in tech domains (data architect, developer, & software engineer)
Concluding, “Which career is best for you, data scientist or data engineer?”
A query of Data Scientist vs Data Engineer – which is better?, stops here.
The discussion over Data Scientist vs. Data Engineer reflects better career growth chances that data science offers to work experts. Yet, both fields deal with data handling. Their future growth depends on an expert’s choice of roles, interests, and skill sets. Data scientists carry more skills and proven expertise in trending tools that bring lucrative offers. Working experts eager to redefine their careers in the data field must consider these features before making the final move.
Enrolling in a Master's in CS: Data Science and AI program opens the door toward alluring pay scales and global job offers. A degree from Woolf University welcomes job offers from the US, Canada, Europe, and 60+ Western countries with effortless visa facilities. A proven mentorship eases practical learning of Generative AI trends followed by live projects. Plus, dual certification from IBM & Microsoft lifts the career success options.
1. Is a data engineer better than a data scientist?
A data engineer and a data scientist have different roles and duties to perform. Concerning duties, data scientists help decision-makers and management teams to make a business move, unlike data engineers who assist data scientists. Following the salary range, data scientists earn more than data engineers. Data scientists may seem better than data engineers, but at present, data engineers are more in demand.
2. Can I switch from data scientist to data engineer?
Yes, you can switch a data scientist’s position with a data engineer and vice-versa. Yet, it needs to learn extra skills and tools to drive better success. A proper analysis of existing skill sets will help find the lagging ones and upskilling.
3. Is data science harder than engineering?
Data science requires more skills as it helps managers and leaders in decision-making. Data science is not harder than engineering, yet it needs quality learning of extra skills (technical & soft) to lift career growth options. Thus, upskilling with the best online data science masters program will drive success.
4. Is data engineering a future-proof career?
Data engineering is a future-proof career as the demand for quality data analysis and insightful decision-making keeps on increasing. A data engineer has all the promising skills that dive deeper into a data-driven culture of business decisions. Many MNCs are opening data engineer positions to support data scientists and other experts. Hence, building a data engineering career will be useful in the upcoming years.
5. What are Data Engineer skills?
Data engineers deal with designing data architectures and data warehousing. Data engineers help data scientists read data sets, analyze them, and make good decisions. Data engineers must have skills like –
- Programming skills
- AI & ML algorithms
- Data Analysis
- Data Modeling
- Database Management
- Data Security
- Data Mining
|
OPCFW_CODE
|
TestJS Summit is an online event for QAs and software developers to get up to date with JS testing best practices, sharpen skills and get latest updates from top products core teams. Get a full ticket to participate in workshops, and network with thousands of others JS testing folks on November 3-4, 2022.
By the end of this talk, you should be equipped with new knowledge regarding frontend and backend performance testing which you can apply to your work projects.
Explain the main differences between the most known frameworks for automation: Selenium, Cypress, Playwright and Robot Framework. Positives and negatives about the frameworks. Depending on the needs of the project your work what is the best framework to use. Small talk about how frameworks integrate with some test management tools.
- Typical journey for teams adopting test automation
- 5 habits to follow while building your test suite
- Our learnings from a recent refactoring activity
Then the talk will cover how AI and machine learning is being leveraged by tools such as Github Copilot and Ponicode to autogenerate test cases thus simplifying the software testing process.
- Introduction: Why test CLI tools
- Challenges: File system pollution, network and database issues, environment variables
- Demo: Showcase issues with a demo CLI tool - Solutions: Test plan implementation, choosing the right level of containerization
- Demo: Show solution using the previous CLI tool
You’ll learn everything necessary to start adding powerful integration tests to your codebase without the headache of managing external service dependencies manually!
WebRTC is a centerpiece of this world, enabling users to reach out to one another directly from the comforts of their browsers and applications.
In this session, Tsahi explains what exactly is WebRTC, how it is different from other web technologies and how you can assure that your WebRTC application gets adopted and successfully used by your clients.
I will introduce vest with a bit of live coding showing its unit testing syntax.
Test automation and monitoring are crucial to avoiding shipping broken apps and sites. But what functionality should you test? When should you run your tests? And aren't complex test suites super slow?
In this session, we'll get our hands on Playwright, the end-to-end testing framework, and learn how to automate headless browsers to ensure that you confidently ship new features.
The second part of the talk will focus on test strategies, types of testing and where to apply them, and finally compare
contrast them to the familiar industry approaches.
We will finish with key takeaways, sample application repos to help with learning, and rollout strategies.
Agenda: What is Web3; The Architecture of a Web3 Application; Web3 E2E Tests Introduction; Web3 E2E Tests Challenges; E2E Test Tools; Demo.
|
OPCFW_CODE
|
On the general experience for first time speakers:
- Overall, I felt really welcomed and supported as a first-time speaker, especially throughout the proposal process! I got really useful feedback during the CFP process.
- I wonder if it would be helpful to have the ability for people to self-identify as first-time speakers during the proposal process, so that these proposals in particular could get more feedback?
- The track conductor for the track my talk was on (Novices), Noel Rappin, actually came to my talk! That was much appreciated. He also reached out a couple weeks beforehand to see if he could help with anything, and followed through with getting some feedback for my slides. That was awesome. It would have been nice if he’d reached out a bit sooner, even, though I know people are busy.
- I regret not asking for a conference guide. There could maybe have been some of that kind of information included in the site upfront, like what to expect for meals/snacks.
- I’ve been thinking about the ideal slots for newbie talks to go into. Days-wise, it seems like not being on the first or last day is best. I think perhaps being the second talk in a block is also a bit better, as the previous talk can have “warmed up” the audience a bit.
- I’m curious whether it would help or hurt to promote first time speakers, like collecting that information somewhere as a sort of “hey! you should support these speakers and help make them feel welcome in the community!” Or would that instead be used by some people as a way to avoid certain talks, because they’re by inexperienced people?
- I enjoyed myself at the speaker dinner much more than I had expected! It helped a lot that I stuck around with Chuck and he’s so gregarious. I would’ve liked a bit more structure for introductions there though, like icebreakers of some kind, or games, or even assigned seating, like gathering up people on the same track together.
- I would LOVE to see more of this kind of non-alcohol socializing event (HT Chiu-Ki Chan)!! Next time I’m going to bring some games of my own to start or join in on a non-alcohol-based “birds of a feather” group event.
- Definitely it would’ve helped to get the A/V details out sooner. I had already started on my slides by the time the official details came out (about 2 weeks before the conference) and just kept working on them, but then the weekend before, realized they were made with the wrong ratio and spent a few hours the day before my talk re-creating the slides for the right ratio instead.
- They provided clickers! That was really nice. Much more compact than the Wiimote I brought with me.
- The A/V staff were really friendly when I did my run through (“It’s my job to take care of how the slides look when projected, don’t worry about that”) so kudos there.
- It would be great to include the Twitter handles for speakers with their bio listings.
- The seats in the rooms were pretty narrow and set up very close together. Maybe this is just a venue thing? But those of us that are more, er…generously sized, could use a bit more breathing space, and it could improve efficiency in the end with not having people sit every other seat as much.
|
OPCFW_CODE
|
Azure web service - change database connection string at runtime
The scenario is this:
An application wants to switch between databases, to avoid lengthy donwtimes when the database gets filled with new data.
So, the idea is to fill one database with data, then tell the web service (via a http call) to use this database from now on.
A couple of weeks/days later, we get new data. So we spend a couple of hours filling the other database and then tell the running web service to switch.
So, how would one go about this? The database connection string in the web service should survive a web service restart.
The whole thing runs on azure. Is there any place where the web service can store and modify settings?
Lots of Greetings!
Azure Web Apps have for this exact problem a dedicated solution.
Under the settings Section, there are Application settings. Application settings have Connections strings section. You can use this feature for your connection string handling in your application.
You can then change this connection string via Azure Web interface or via Azure Powershell.
An article how to use Azure connection string feature: https://azure.microsoft.com/en-us/documentation/articles/web-sites-configure/.
Ok - so there is no easier way than powershell from within the controller action. Thanks a lot!
No problem @Volker, happy to help. :)
Just have a couple connection strings in settings and program your application to switch between them as needed. As an option you can also have some common database where you store settings including connection string to your "data" database and again just program your app to do the switch.
Uh - same problem. How do I keep the information which one to use?
Ah - ok. And then some caching or so to reduce the roundtrip count. But I'm still hoping that the azure infrastucture offers a better way. :-)
It does, you are right. A dedicated functionality for the connection strings handling.
The recommended place to store connection strings for Azure Websites / Web Apps is as an Application Settings / Connection Strings!.
Internally these are stored as Environment Variables that are available at runtime.
This would be a reasonably trivial thing to push an update to, via Powershell, via REST from within the App itself, or another app. I'm not sure whether you would need to restart the web app to pick up the change.
For instance, you could have ConnectionString1 and ConnectionString2. Then within your app you would be able to switch between the two at runtime (with connectionstring = ConnectionString1)
If you need to remember which one is current, you would need another data source, which could be anything, from a filesystem file! to an Azure Storage Table value, to another database. This would be a static database, that you wouldn't change, it would simply be there to store this configuration data for the site.
If you need to inject a new connection string you could do that via Powershell. There is also a REST API that you could use from within your app to manage that internally.
I don't mind to have 2 fixed ones there. I just need a way to remember which one to use, from within the application, in a way that survives a restart.
Give them a meaningful name and show which you're using! Parse part of the connection string that will tell you which you're using and display it somewhere.
I don't understand. I want to be able to switch, at runtime. So the web api has a .../.../switch-database endpoint. And the only way for the web service to do that is to execute a powershell script internally? I mean, in a normal .net application, settings are simply writable: string mysetting = App.Default.MySetting; App.Default.MySetting = "my new setting";
Thx! I'll see what's easier, the rest api or a third database.
|
STACK_EXCHANGE
|
Question on Malcev's _Immersion of an Algebraic ring into a skew field_.
I'm reading the paper Immersion of an algebraic ring into a skew field by Malcev. Doi: 10.1007/BF01571659,
GDZ.
On the third page of the paper, he writes that
If $\alpha\beta\sim\gamma\delta$ and the words $\alpha$ and $\gamma$ have the same length then
$$
\alpha=\mu m,\quad \beta=n\nu,\quad \gamma\sim\mu m',\quad\delta\sim n'\nu,\quad mn\sim m'n'
$$
where $m,n,m',n'$ are each one of the letters $a,b,c,d,x,y,u,v$.
How is this property easily seen? Malcev starts be starting with the semigroup of all words generated by eight letters $a,b,c,d,x,y,u,v$, where the operation is concatenation. He defines the pairs of two-letter words as "corresponding" $(ax,by)$, $(cx,dy)$ and $(au,bv)$ and says two words $\alpha$ and $\beta$ are equivalent if one can be obtained from the other by changing $ax$ to $by$, $cx$ to $dy$, etc. or vice versa.
He also proves that there are never any overlap ambiguities of what two letters could be replaced, but I don't see why the above quoted property follows so easily. Thank you.
The link is broken.
I've copied the links from the other post, I guess it is the same paper.
Malcev states before that in a word $ \dotsc m n p \dotsc$ we cannot replace $mn$ and $np$ by another word. This means that the relations don't overlap. Regardless how you change a word by the relations, the classes of the letters $\{a,b,c,d\}$ and $\{x,y,u,v\}$ stay the same at each position. So if we want to change $\alpha \beta$ into $\gamma \delta$, we have to change $\alpha$ into $\gamma$ and $\beta$ into $\delta$ separatedly (this is the case that the letters $m,m',n,n'$ don't appear; probably one should add this), or at one step we come across the two words and make something like $(\mu m)(n \nu) \sim (\mu m')(n' \nu)$. Of course this is not really formal. It can be made precise using induction on the length of the deduction.
|
STACK_EXCHANGE
|
|author||Alan Coopersmith <email@example.com>||2010-09-08 22:58:30 -0700|
|committer||Alan Coopersmith <firstname.lastname@example.org>||2010-09-08 22:58:30 -0700|
Bring README a little closer to the current state of reality
Signed-off-by: Alan Coopersmith <email@example.com>
1 files changed, 26 insertions, 27 deletions
@@ -1,6 +1,6 @@
- Mouse Support in X11R7.5
- Kazutaka Yokota
- 17 December 2002
+ Mouse Support in xf86-input-mouse
+ Original version written by Kazutaka Yokota for XFree86 on 17 December 2002
+ Updated by Alan Coopersmith for X.Org releases
Table of Contents
@@ -56,8 +56,10 @@
- This document describes mouse support in X.Org Foundation's X11R7.5
+ This document describes mouse support in the xf86-input-mouse driver
+ for the Xorg X server. This driver is mainly used on non-Linux
+ operating systems such as BSD & Solaris, as modern Linux systems use
+ the xf86-input-evdev driver instead.
Mouse configuration has often been mysterious task for novice users.
However, once you learn several basics, it is straightforward to write
@@ -67,13 +69,13 @@
2. Supported Hardware
- The X.Org Foundation X server supports four classes of mice: serial,
+ The xf86-input-mouse driver supports four classes of mice: serial,
bus and PS/2 mice, and additional mouse types supported by specific
operating systems, such as USB mice.
- The serial mouse has been the most popular pointing device for
+ The serial mouse was once the most popular pointing device for
PCs. There have been numerous serial mouse models from a number
of manufactures. Despite the wide range of variations, there
have been relatively few protocols (data format) with which the
@@ -92,14 +94,15 @@
integrated I/O cards may also have a bus mouse connector. Some
bus mice are known as `InPort mouse'.
- Note that some mouse manufactures have sold a package including
+ Note that some mouse manufacturers have sold a package including
a serial mouse and a serial interface card. Don't confuse this
type of products with the genuine bus mouse.
- They are sometimes called `Mouse-port mouse'. The PS/2 mouse is
- becoming increasingly common and popular.
+ They are sometimes called `Mouse-port mouse'. The PS/2 mouse was
+ common for a generation after serial mice, and most laptops still
+ use the PS/2 protocol for built-in pointer devices.
The PS/2 mouse is an intelligent device and may have more than
three buttons and a wheel or a roller. The PS/2 mouse is
@@ -115,17 +118,16 @@
computers. Several devices can be plugged into this bus,
including mice and keyboards.
- The server includes support for USB mice on some systems.
+ This driver includes support for USB mice on some systems.
Many mice nowadays can be used both as a serial mouse and as a PS/2
- mouse. They has a logic to distinguish which interface it is
- connected to. However, the mouse which is not marketed as compatible
- with both serial and PS/2 mouse interface lacks this logic and cannot
- be used in such a way, even if you can find an appropriate adapter
- with which you can connect the PS/2 mouse to a serial port or visa
- X11R7.5 supports the mouse with a wheel, a roller or a knob. Its
+ mouse, or as both a PS/2 and a USB mouse. They have logic to distinguish
+ which interface it is connected to. However, a mouse which is not
+ marketed as compatible with both mouse interfaces lacks this logic and
+ cannot be used in such a way, even if you can find an appropriate adapter
+ with which you can connect the mouse to a different format port.
+ This driver supports a mouse with a wheel, a roller or a knob. Its
action is detected as the Z (third) axis motion of the mouse. As the
X server or clients normally do not use the Z axis movement of the
pointing device, a configuration option, "ZAxisMapping", is provided
@@ -155,7 +157,6 @@
NetBSD Ok Ok Ok SP*1 SP*1 SP*1
NetBSD/pc98 Ok ? Ok NA NA NA
OpenBSD Ok Ok Ok Ok*1 Ok*1 Ok*1
- OS/2 SP*2 SP*2 SP*2 SP*2 SP*2 ?
SCO Ok ? SP*1 SP*1 NA ?
Solaris 2.x Ok NA*1 ?*1 Ok Ok SP*1
SVR4 Ok NA*1 SP*1 SP*1 NA ?
@@ -165,8 +166,6 @@
SP: support is available in a different form
*1 Refer to the following sections for details.
- *2 X11R7.5/OS2 will support any type of mouse that the OS supports,
- whether it is serial, bus mouse, or PnP type.
@@ -303,7 +302,7 @@
- Testing has been done with Solaris 2.5.1, 2.6, 7, 8, 9 and 10.
+ Testing has been done with Solaris 2.5.1, 2.6, 7, 8, 9, 10, and 11.
On Solaris 10 1/06 and later versions with "virtual mouse" support,
all PS/2 and USB mice connected to the system can be accessed via the
@@ -320,9 +319,9 @@
Additional USB mice can be connected using the "VUID" protocol type
- and the appropriate "/dev/usb/hid" device with the Option
- "StreamsModule" "usbms" line included in the associated "InputDevice"
+ and the appropriate "/dev/usb/hid" device with the
+ Option "StreamsModule" "usbms"
+ line included in the associated "InputDevice" section.
@@ -346,7 +345,7 @@
4. Configuring Your Mouse
- Before using the xorgconfig program to set up mouse configuration, you
+ Before editing the xorg.conf file to set up mouse configuration, you
must identify the interface type, the device name and the protocol
type of your mouse. Blindly trying every possible combination of
mouse settings will lead you nowhere.
|
OPCFW_CODE
|
Can anyone reproduce the results? (my results attached)
Does anyone successfully reproduce the results? I ran several times but the results are far away from the author's. My results are attached.
Hi @ShuoYang-1998,
I am very sure that it is just a matter of some hyper-parameter that is causing the discrepancy. The code tip is one that I used to run a lot of ablations towards the end of the paper and for the rebuttal. I just need to find that rouge hyper parameter that is causing the issue. One other point to note from your reported results is that ORE is better than your reproduced Faster-RCNN+FT in most of the cases.
As @salman-h-khan said, I had an international travel on Saturday. Unfortunately enough, I am tested positive for COVID today. I have developed minor complications and hence am admitted to a hospital now. I am typing this message from the hospital room. Kindly give me some time to regain my health, if possible.
Thanks,
Joseph
wish you all good
Does anyone successfully reproduce the results? I ran several times but the results are far away from the author's. My results are attached.
Hi Yang,
I have a question about the experiments of EBMs.
Is the validation used for fitting EBMs distribution using known and unknown labels? I wonder if the open-set settings will degenerate into few-shot settings after validation because unknown classes (UUCs) with ground-truth labels shouldn't appear in the process except for testing.
Does anyone successfully reproduce the results? I ran several times but the results are far away from the author's. My results are attached.
Hi,Yang
You have reproduced the results so quickly, and I would like you to give me some advic. I can't send messages to you email address, if it is convenient for you, could you give me some guidance.Thank you, looking forward to your reply.
Does anyone successfully reproduce the results? I ran several times but the results are far away from the author's. My results are attached.
Hi Yang,
I have a question about the experiments of EBMs.
Is the validation used for fitting EBMs distribution using known and unknown labels? I wonder if the open-set settings will degenerate into few-shot settings after validation because unknown classes (UUCs) with ground-truth labels shouldn't appear in the process except for testing.
The EBUI does use all known and unknown data to learn a distribution, but it doesn't access the labels.
Does anyone successfully reproduce the results? I ran several times but the results are far away from the author's. My results are attached.
Hi,Yang
You have reproduced the results so quickly, and I would like you to give me some advice. I can't send messages to you email address, if it is convenient for you, could you give me some guidance.Thank you, looking forward to your reply.
I have uploaded my run.sh in issue 18, please refer.
Does anyone successfully reproduce the results? I ran several times but the results are far away from the author's. My results are attached.
Hi Yang,
I have a question about the experiments of EBMs.
Is the validation used for fitting EBMs distribution using known and unknown labels? I wonder if the open-set settings will degenerate into few-shot settings after validation because unknown classes (UUCs) with ground-truth labels shouldn't appear in the process except for testing.
The EBUI does use all known and unknown data to learn a distribution, but it doesn't access the labels.
Hi, Yang,
Thank you for your reply. I have checked train_loop.py and modeling/roi_heads/roi_heads.py, and find EBUI should have used the unknown annotations, as shown in the followed codes. The gt labels of the unknown instances will be allocated to the region proposals in roi_heads.py and are saved to fit the unknown WB distribution. Is it right?
wb_unk = Fit_Weibull_3P(failures=unk, show_probability_plot=False, print_results=False)
def compute_energy(self, predictions, proposals): gt_classes = torch.cat([p.gt_classes for p in proposals]) logits = predictions[0] data = (logits, gt_classes) location = os.path.join(self.energy_save_path, shortuuid.uuid() + '.pkl') torch.save(data, location)
I also find EBUI can work along, which means the unknown labels are not from ALU, but from gt.
Hi @JosephKJ
I also can not reproduce your result reported in your paper. I try to run your code several times, but failed to get the same results.
other people also meet the issue,i really doubt the authenticity of this work, so i think you should consider the issue seriously.
d unknown data to learn a distribution, but it doesn't access the labels.
Hi, Yang,
Do you mean the specific labels?
I think that for the open-set settings, we shouldn't use unknown samples for training. In this validation, we need to learn(fit) the distribution and save the parameters, which should be seen as training with some tricks.
For MINIST, we cannot set 0-6 with real labels and set the rest are labeled unknown. In OWOD, I am not sure whether using Known and unknown labeled samples to fit the distribution is OK? Do you think it is OK?
d unknown data to learn a distribution, but it doesn't access the labels.
Hi, Yang,
Do you mean the specific labels?
I think that for the open-set settings, we shouldn't use unknown samples for training. In this validation, we need to learn(fit) the distribution and save the parameters, which should be seen as training with some tricks.
For MINIST, we cannot set 0-6 with real labels and set the rest are labeled unknown. In OWOD, I am not sure whether using Known and unknown labeled samples to fit the distribution is OK? Do you think it is OK?
I also have the same concern, some other people also raise this questions in [https://github.com/JosephKJ/OWOD/issues/16] (url) and https://github.com/JosephKJ/OWOD/issues/8. But the author didn't response.
@JosephKJ @ShuoYang-1998 @wyman123 @Hrqingqing @salman-h-khan What is the order in which I should use the configuration files in the T1 ~ T4 experiments?Looking forward to your reply!!
Hope to get your detailed analysis of the configuration file!!
@ShuoYang-1998 @Hrqingqing @wyman123
when i run the command line
python3 tools/train_net.py --num-gpus 1 --config-file configs/OWOD/t1/t1_train.yaml SOLVER.IMS_PER_BATCH 4 SOLVER.BASE_LR 0.005
It showed that some folders don't exits! What should I do next step?
Closing this issue due to inactivity. @dyabel is able to reproduce mAP and A-OSE with the latest code. Kindly reopen for more discussions.
@ShuoYang-1998 Hi! friend, I also tried to reproduce the result but failed, I attach the result in #77 , if you successfully reproduced the result, could you help me, thank you!
|
GITHUB_ARCHIVE
|
Clomid day 2 7 clomid side effects, body aches when to increase clomid dose. When will you ovulate on clomid 3 7. Spotting on clomid pregnant medicine like, clomid where can i get clomid online, chances of becoming pregnant with, clomid results with clomid clomid for ovulation induction benefits of increasing clomid dosage.
Buy clomid online in united states clomid, 50 mg benefits has. Clomid worked for you clomid patient teaching clomid nyquil. Nausea around ovulation clomid taking clomid to conceive, how can i increase my chances of getting pregnant. On clomid clomid 100mg metformin clomid ivf. Cycle how can i take clomid ovary, cramping clomid nolvadex a clomid clomid cause heart palpitations.
Progesterone levels on clomid cycle should i buy clomid, online where can i get clomid online baby aspirin, and clomid success clomid patient teaching clomid no, signs of ovulation do you bleed. On clomid pcos clomid first cycle clearblue, monitor and clomid is it ok to, take clomid with food clomid. Patient teaching clomid as a standalone clomid patient teaching.
Clomid patient teaching clomid 100mg ovidrel iui 40. Mg clomid clomid patient teaching, bbt ttc clomid cost of monitored, clomid cycle clomid and, increased urination clomid ovulation calculator. For a boy what does clomid do for low progesterone, glumetza and clomid clomid. Patient teaching what medications interfere with clomid.
Clomid acne tupincho difference between clomid and, hcg injection spotting side effect clomid, clomid kidney when on clomid how long, is your cycle can you skip a, cycle of clomid does clomid make u. Ovulate early hcg nolva and or, clomid when should i ask for clomid drinking during clomid, clomid patient teaching clomid injection cost.
Clomid longer cycles medroxyprogesterone and clomid together clomid pills vs. Liquid where can i get clomid online why. Do you ovulate late on clomid, how much is clomid in uk clomid avance ovulation. Can you have an anovulatory cycle on clomid clomid, patient teaching is it ok to take femara. And clomid together does clomid help with sperm, motility how many days is your cycle on clomid.
Side effects of clomid. And hcg shot pcos, metformin and clomid clomid ovestin duphaston clomid, and nolvadex for pct clomid effects. On cycle bfp clomid iui 50 mg clomid didn't, work was ist clomid 50 mg clomid. Rapid weight gain can i take clomid without. Having my period other options than clomid late period after clomid challenge test.
Side effects of clomid for the, baby clomid patient teaching clomid 100g, side effects herbal replacement for clomid bleeding with clomid, what is the use of clomid medicine clomid, repronex iui how often does 50mg clomid work tamoxifen. And clomid dosage clomid elevated testosterone best places to, buy clomid online where can i get clomid online stomach pains after taking clomid.
Where can i get clomid online, clomid patient teaching can. You take clomid if you have an ovarian, cyst low libido on clomid pct can, you use opks with clomid what is a low. Dose of clomid clomid, iui and trigger shot where can. I get clomid online where can. I get clomid online bleeding with. Clomid sides of clomid clomid hcg and artificial. Insemination pregnancy with clomid and ovidrel.
Clomid causing late ovulation can you take clomid without. Monitoring clomid worked second round taking. Clomid days 3 9 blurred vision, due to clomid does, clomid make your period different clomid, after 3 months egg white cervical mucus after clomid can, clomid delay your period long 7. Rounds of clomid where can. I get clomid online where can i get clomid online clomid en belgique.
|
OPCFW_CODE
|
$(document).ready(function() {
// Variables
var $gameSquares = $('.square');
var moves = [0, 0, 0, 0, 0, 0, 0, 0, 0];
var counter = 0;
var xWinsCounter = 0;
var oWinsCounter = 0;
var X = 'X';
var O = 'O';
var cPlayer = X;
// Functions
var displayMessage = function(message) {
$('.message').html(message);
};
function resetBoard() {
if (xWinsCounter < 5 && oWinsCounter < 5) {
moves = [0, 0, 0, 0, 0, 0, 0, 0, 0];
counter = 0;
displayMessage('Current Player: ' + X);
//Clears cell content
$gameSquares.removeClass('x-image o-image');
} else {
alert('The series is over!');
$('.square').off('click');
$('.square').hide();
$('.message').hide();
$('.reset').show().on('click', function() {
location.reload();
});
}
}
function getWin() {
var winConditions = [moves[0] + moves[1] + moves[2],
moves[3] + moves[4] + moves[5],
moves[6] + moves[7] + moves[8],
moves[0] + moves[4] + moves[8],
moves[0] + moves[3] + moves[6],
moves[1] + moves[4] + moves[7],
moves[2] + moves[5] + moves[8],
moves[2] + moves[4] + moves[6]
]
console.log(winConditions);
for (i = 0; i < winConditions.length; i++) {
if (winConditions[i] === 3) {
alert('X Wins!');
xWinsCounter++;
//Update X Wins
$('#xwins').text(xWinsCounter);
//Run reset function
resetBoard();
} else if (winConditions[i] === 30) {
alert('O Wins!');
oWinsCounter++;
//Update O Wins
$('#owins').text(oWinsCounter);
//Run reset function
resetBoard();
} else if (counter == 9) {
//Runs if all squares have been clicked
console.log('tie!');
alert('The game is a tie!');
resetBoard();
} else {}
}
}
//On Click
$($gameSquares).on('click', function() {
//Determines if this cell has been clicked already
if (moves[this.id] === 0) {
if (counter % 2 == 0) {
//Changes value at moves[i] to 1
moves[this.id] = +1;
//Change cell to X image
$(this).addClass('x-image');
displayMessage('Current Player: ' + O);
} else {
//Changes value at moves[i] to 10
moves[this.id] = +10;
//Changes cell O image
$(this).addClass('o-image');
displayMessage('Current Player: ' + X);
}
counter++;
getWin();
} else if (this.id !== 0)
alert("Select a different square");
});
});
|
STACK_EDU
|
Recently I had a chance to debug memory leaks for a program. This is not the first time I had to debug a leaking program, but
this one was a little special: it was a Python
asyncio web server. It’s special because traditional tools (like
do not work well with Python, and an
asyncio application can be quite tricky to debug due to complex interactions among
In general, the steps to tackle memory leak problem (also for any types of bugs) is as follows:
- Step 1 Identify where the leak(s) is in the program.
- Step 2 Figure out why it is leaking.
- Step 3 Eliminate leaking spots.
Usually the first step is the trickest among the three: you need to know what information is available and useful for diagnosis and where to find them, or, what tools are available to help you and how to use them. Even if you’ve got the necessary information, you need to know how to interpret it (perhaps by visualizing some data) and use that to dig down further if necessary.
I will explain what I did for tackling the aforementioned Python
asyincio memory leak problem. However, before that, let’s take
a step back and think about what could cause a Python program to leak memory.
If you know Python well enough, you probably know that it is a language with garbage collection, just like many other modern programming languages. That means, as a programmer (a user of the language), you normally do not need to worry about allocating memory and forgetting to free that allocated memory later on; because the garbage collector does that for you in the background. The presence of a garbage collector greatly reduces the chances of memory leaks. However, it is still possible to have memory leaks in Python programs. How? you might ask.
There are several scenarios I can think of:
- Leaking through Python C/C++ extensions.
- Reference cycles among Python objects.
- Situations when memory is kept by the interpreter for long (not really memory leak).
- Everything is garbage collected correctly, interpreter returns memory back to the OS, but the memory is still not freed until process termination.
Let’s discuss each scenario in more details.
The first one is obvious. And there is no garbage collector for C/C++ code; so memory leak bugs often arises in this part of the program.
The second scenario is actually not applicable anymore (since Python 3.4+). It was possible (before Python 3.4) to have leak memory
leaks due to reference cycles when there are circular references between objects AND some classes of those objects have custom
implemented. After safe object finalization was introduced, such way of leaking should not happen anymore. Having said that, reference
cycles should still be avoided when writing an application; this is because Python’s cyclic GC runs asynchronously in the background and in
phases (or generations), in other words, those cyclic isolates are likely NOT to be collected immediately and linger for a while, which in
turn might contribute to ‘high water’ phenomenon temporarily. In addition to this, running GC (which is triggered by certain heuristics) has
performance penalty since that pauses the entire interpreter, especially in the third generation of garbage collection (the most costly).
For more in-depths explanations to garbage collection in Python (3.6), I recommend having a read of this post.
The third scenario is more interesting, many situations can happen (application-specific); for example, in a
asyncio application, many corouines
are running at the same time, and many of them are blocked by some IO operations, and each of them is still holding memory allocated to them, like
this case study. Or, you have a large number of immuntable types (like
floats) allocated, Python still keeps them in the so-called
free lists for performance reasons, even after they are not referenced anymore during program run. These examples are not
really memory leaks, but if they happen in a long-running Python process, you might still experience monotonically memory usage increase or high
water phenomenon, because the memory is not released by the interpreter back to the OS yet.
To better understand free lists and memory management in Python, I found this post, this one and this one helpful. With such understanding, it shall help you to write Python programs that interacts well with Python’s memory management algorithms.
The fourth scenario is not necessarily Python related, it is due to memory fragmentation. And seems like it was improved since Python 3.3+. Memory fragmentation usually happens when you have a long running program (like a server) and it allocates a mix of short-lived and long-lived objects over time. As a Python application developer, you normally do not need to worry about it. Memory fragmentation symptom shows high water mark phenomenon in memory profiling.
Enough said, let’s turn back to how I tackled the memory leak problem of a Python
asyncio web server.
With the four scenarios in mind, you might have an educated guess that it is more possible to encounter the first and the third scenarios. But we are engineers, we do not just guess, we need proofs.
First of all, I knew I had a memory leak problem because there was a monitoring service running to monitor the node where the leaking
web server was running on and alerted me of unusual memory usage pattern.
So, the problem did exist, now where was the problem exactly? I’ve tried using a few tools:
For my problem, I found the first two tools weren’t helpful. They seem to be good tools for estimating the overall memory usage of an application.
If the memory leaks are instances of the same type and this type is only instantiated in a few places (for example, some custom classes in the
application), these tools can identify that.
objgraph is only useful when you’ve already identified which object(s) in the code is leaking, and
ususually used to visualize objects relations for related objects.
After playing with the first three tools, all I saw was some generic types and most of them got garbage collected when I shutdown the application server on my local machine. The cause for the problem was still a mystery.
Then I realized that the way I debugged the problem was wrong: I invoked those tools in the server program entry point when server started (this part was correct), and displayed the memory profile when the server shut down (this part was wrong), on my local machine (this was also wrong). It was wrong in two ways, the first is if the leak was caused by the third scenario described above, then this way of debugging will never catch that because those dangling memory chunks may very likely be collected right before the server shuts down. The second is many memory problems manifest themselves only in a long running process and under real-world loads, which local experiments may not easily reproduce. To do it correctly, I really need to be able to take snapshots of memory profiles of the live, production server at run-time: an API memory profiling endpoint to query while the server is running.
Then I found this wonderful post, which was exactly what I needed. In this post, the author used
tracemalloc and some
utilities. It turns out that
tracemalloc is well suited for such situations: to identify top memory holders from large applications. And
it is a module from standard library since Python 3.4! I adapted the code snippets in that post for my own use and implemented a memory profiling
endpoint for the leaking web server (and of course, installed that endpoint to the live production server).
Guess what? One day after the endpoint was released, I was able to (very easily) pinpoint exactly which object in the code that’s leaking memory
based on the profiling information provided by the endpoint (step 1). It was some object in the lower level code of
aiohttp that keeps a
big bulk of memory. And after some googling around, it turned out that there was indeed a recent memory leak bug in library
aiohttp (step 2). And
multidict is indeed a Python C extension! So after knowing the why, the solution became obvious and simple: either
downgrade or upgrade the version of
aiohttp to the one that does not have the
multidict memory leak bug (step 3). Mission accomplished!
In summary, I personally found
tracemalloc to be a very useful tool for debugging memory leak problems for Python programs. And if you’re debugging
a web server, a profiling endpoint is really a good idea!
|
OPCFW_CODE
|
Confusions about Multiplication of Complex Value in MatLab
I noticed a confused computation of complex valued multiplication in Matlab. One simple example is as below:
syms x1 x2 x3 x4
s=[x1 x2]*[x3 x4]'
And the return value of s is like:
s=x1*conj(x3) + x2*conj(x4)
In my opinion s should be equal to x1*x3+x2*x4. So, what's the problem here?
Then, how should I do if I want to get a multiplication of two complex vector?
update: I find out it will be solved through using .' rather than . like:
s=[x1 x2]*[x3 x4]
The operator ' is the also called Complex conjugate transpose in Matlab ctranspose, which basically means that it applies conj and and transpose functions. Note that this operator is call Hermitian operator in mathematics.
What you actually want is the operator transpose that is shortcut as .'
In order to get the expected output, and given that you just want to multiply without conjugating the second vector, you should do:
>> syms x1 x2 x3 x4
>> s = [x1 x2]*[x3 x4].'
so your output will be:
x1*x3 + x2*x4
For further information you can check help ., to see the list of operators, help transpose and help ctranspose
Also known as the Hermitian operator in mathematics. It's common to hear someone say "v Hermitian" in place of "v transposed" for some array v, and it is usually denoted by a cross/dagger symbol to distinguish it from basic transposition. I think Matlab's syntax (using ' for both types of transposition) is a bit confusing if you are new to it and come from a math background.
Thanks for the addition! I edited, commenting that ' works as the Hermitian operator.
Note that the ' operator in Matlab is the conjugate transpose, i.e. it both transposes a matrix and takes the complex conjugate:
>> (1+1i)'
ans =
1.0000 - 1.0000i
If you want the matrix transpose then you should use the .' operator:
>> (1+1i).'
ans =
1.0000 + 1.0000i
+1 for teaching me about .' being the equivalent of transpose.
Yes, you are right and thanks a lot for your warm hearted answer.
Maybe this will help explain:
>> syms x1 x2 x3 x4
>> s=[x1 x2]*[x3 x4]'
s =
x1*conj(x3) + x2*conj(x4)
>> s=[x1 x2]*[x3; x4]
s =
x1*x3 + x2*x4
>> [x3 x4]'
ans =
conj(x3)
conj(x4)
The ' version of transpose isn't doing what you want. Use transpose instead:
>> transpose([x3 x4])
ans =
x3
x4
As @ChrisTaylor points out, .' operator does the same thing as the transpose function.
|
STACK_EXCHANGE
|
Current home network consists of:
1. Cable Modem wired to Linksys EA4500 router (192.168.1.1)
2. EA4500 hard wired to Cisco Catalyst 2950 Patch Panel
3. Multiple Port LAN for 2 hard wired PC's, printers, etc.
4. Wireless PC with Linksys WPC300N
5. Wireless Laptop with Linksys WLP300N
6. Second Router Linksys WRT54G (192.168.1.2) is hard wired to Port 1 from the Patch Panel, configured as wireless access point (DHCP disabled) for "G" wireless communication distance 40 feet) a DVR (Replay TV)
7. Security for all components is WPA2/Personal
The Linksys EA4500 replaced a prior main router, a Linksys WRT300N. With the prior WRT300N router in place, the above system worked flawlessly for many years. Since the EA4500 was installed, the WAP54G will not recognize or receive the wireless communication from the secondary router WRT54G. Therefore the hard wired DVR does not receive any menu guide data.
Note that, the secondary router (WRT54G) has a separate IP 192.168.1.2, is "broadcast enabled" and may be successfully pinged in the current set up. As a temporary step, when the WAP54G is hard wired to the either a free port on the WRT54G, or directly to the patch panel, it may be successfully pinged as well. However, when the WAP54G is moved approximately 40 unobstructed feet, on the same floor in a clear sight line, it cannot be pinged wireless.
Although not life threatening, this situation is a great annoyance. If anyone has any ideas (other than the purchase of additional new hardware), I would appreciate insight as to where to look for configuration problems, et al.
Also, I wonder if the WAP54G may be able to receive strong enough "G" communication directly from the EA4500, assuming the EA4500 2.4ghz channel transmits in G mode? The overall distance seems a bit long (approximately 75-100 feet, over 3 floors, however, there is no concrete between floors or steel construction). There is also a possibility for a manual set up of the DVR network configuration, including the entry of the IP address, subnet address, default gateway and 2 separate DNS addresses. All are available, except the DNS address of the main router. How may the main router DNS be determined?
Assistance, insight and suggestions are warmly encouraged.
|
OPCFW_CODE
|
Amazon Interview Experience | 194 (For Software Support Engineer)
It was a usual Job Application at Amazon’s Job Site, I got their call few days later.
Process:- 1 Written Aptitude(Coding) + 3 Technical Round + 1 HR Round
1) One Written Test – 8 Problem statement – 1 Hour Time – from Basic to Mid-level difficulty, for which we have to write the Logic/Algorithms
Some which I remember were good enough are –
i) Given a Folder path in a Directory Structure. E.g. “temp/reg”; you need to find all the files with extensions “.c” in this folder as well as in the sub-folders.
ii) Finding the Longest Palindrome Sub-String in the given string.
iii) From the given string, print the Alphabets which occur more than once in the string.
iv) Given two Integers A & B, Print the number of bits to be flipped in order to convert ‘A’ to ‘B’.
v) Implement EnQueue() and DeQueue() of Queue DS using Push() and Pop() of Stack DS. Write Steps and Sequence.
Possible Solutions:- (Logic)
i) Use Linked-List DS to store the Folder Path in a Node. Open the Folder with First Node, filter all the “.c” extension files, and add the sub-folders in the linked-list, until all the nodes in the linked-list are explored.
ii) Use ‘frame’ variable with value starting from 2, increasing till size ‘n’, and check by shifting this frame by one position, if the sub-string is a Palindrome or not.
iii) Basic Char to ASCII conversion and back to Characters. Create an Array of 26 slots for alphabets. Convert the input string to UpperCase before proceeding.
iv) Convert to Binary Representation and Compare as String Characters. If not same, then increment the counter. Remember, we have to only display the Number of bits flipped.
v) Use Operational Logic for Priority Queue, and for –
a) EnQueue – Push everything in Stack1.
b) DeQueue – (Check if Stack1 is Not empty)Pop everything from Stack1 and Push them one by one to Stack2. Then Pop top element from Stack2 as Result. For Next DeQueue do reverse, i.e. from Stack2 to Stack1.
2) 1st Technical Round
i) Write Program to Reverse a Linked List.
ii) Discussion on a Scenario and Debugging/Troubleshooting the error.
3) 2nd Technical Round
i) The similar Debugging/Troubleshooting scenario.
ii) Add the Numbers present in the String and display.
Here take ‘-‘ as negative number symbol.
iii) Find the Pair of Array elements summing up to value ‘k’, in a Sorted Array. Optimize it from Complexity ‘N^2’ to ‘NlogN’ using Binary Search.
4) 3rd Technical + Managerial Round
i) Some OS Related Questions.
ii) UNIX Commands – Which I couldn’t recall!
iii) Project Experience and Why want to Join Amazon and similar ones.
Though I couldn’t clear the 3rd Round, the overall experience was good.
>> Please check their Technical Requirement, and revise your concepts of OS(How Process works, Process Life Cycle, RAM, Virtual Memory, DeadLocks etc.), UNIX(Revise Basic Commands like ‘grep’ etc.).
>> Practice basic Problems like, Searching, Sorting, Fibonacci, Factorial, Recursion, Stack, Queue, Linked List and their Application.
Hope it helps for the new Applicants!
If you like GeeksforGeeks and would like to contribute, you can also write an article and mail your article to email@example.com. See your article appearing on the GeeksforGeeks main page and help other Geeks.
Please Login to comment...
|
OPCFW_CODE
|
package compiler
// SymbolScope represents a scope of symbols.
type SymbolScope string
const (
// GlobalScope represents a global scope, i.e. top level context of a program.
GlobalScope SymbolScope = "GLOBAL"
// LocalScope represents a local scope, i.e. a function level context.
LocalScope SymbolScope = "LOCAL"
// BuiltinScope represents a scope for built-in functions.
BuiltinScope SymbolScope = "BUILTIN"
// FreeScope represents a scope for closures referencing free variables.
FreeScope SymbolScope = "FREE"
// FunctionScope represents a scope for self-referencing functions.
FunctionScope SymbolScope = "FUNCTION"
)
// Symbol is a symbol defined in a scope with an identifier (name).
type Symbol struct {
Name string
Scope SymbolScope
Index int
}
// SymbolTable is a mapping table of identifiers (names) and defined symbols.
type SymbolTable struct {
freeSymbols []Symbol
outer *SymbolTable
store map[string]Symbol
numDefs int
}
// NewSymbolTable creates a new symbol table.
func NewSymbolTable() *SymbolTable {
return NewEnclosedSymbolTable(nil)
}
// NewEnclosedSymbolTable creates a new symbol table with an outer one.
func NewEnclosedSymbolTable(outer *SymbolTable) *SymbolTable {
return &SymbolTable{
freeSymbols: make([]Symbol, 0),
outer: outer,
store: make(map[string]Symbol),
}
}
// Define defines an identifier as a symbol in a scope.
func (s *SymbolTable) Define(name string) Symbol {
scope := GlobalScope
if s.hasOuter() {
scope = LocalScope
}
sym := s.define(name, scope, s.numDefs)
s.numDefs++
return sym
}
// DefineBuiltin defines a built-in function with `name` at the `index`.
func (s *SymbolTable) DefineBuiltin(index int, name string) Symbol {
return s.define(name, BuiltinScope, index)
}
// DefineFunctionName defines a built-in function with `name` at the `index`.
func (s *SymbolTable) DefineFunctionName(name string) Symbol {
return s.define(name, FunctionScope, 0)
}
// defineFree defines a free symbol based on the `original` one.
func (s *SymbolTable) defineFree(original Symbol) Symbol {
s.freeSymbols = append(s.freeSymbols, original)
return s.define(original.Name, FreeScope, len(s.freeSymbols)-1)
}
func (s *SymbolTable) define(name string, scope SymbolScope, index int) Symbol {
sym := Symbol{Name: name, Scope: scope, Index: index}
s.store[name] = sym
return sym
}
// Resolve resolves an identifier and returns a defined symbol and `true` if any.
// If the identifier is not found anywhere within a chain of symbol tables, it returns an empty
// symbol and `false`.
func (s *SymbolTable) Resolve(name string) (sym Symbol, exists bool) {
if sym, exists = s.store[name]; exists || !s.hasOuter() {
return sym, exists
}
sym, exists = s.outer.Resolve(name)
if exists && (sym.Scope == LocalScope || sym.Scope == FreeScope) {
// Define an outer local or free variable as a free variable in the current scope
sym = s.defineFree(sym)
}
return sym, exists
}
// ResolveCurrentScope resolves an identifier within the current scope and returns a defined
// symbol and `true` if it is defined, otherwise returns an empty symbol and `false`.
func (s *SymbolTable) ResolveCurrentScope(name string) (sym Symbol, exists bool) {
sym, exists = s.store[name]
return sym, exists
}
// hasOuter returns true if `s` has an outer symbol table, otherwise false.
func (s *SymbolTable) hasOuter() bool {
return s.outer != nil
}
|
STACK_EDU
|
package org.greenrobot.greendao.rx2;
import org.greenrobot.greendao.AbstractDao;
import org.greenrobot.greendao.annotation.apihint.Experimental;
import java.util.List;
import java.util.concurrent.Callable;
import io.reactivex.Observable;
import io.reactivex.Scheduler;
/**
* Created by Zhang Tingkuo.
* Date: 2017-04-28
* Time: 14:28
*/
@Experimental
public class Rx2Dao<T, K> extends Rx2Base {
private final AbstractDao<T, K> mDao;
/**
* Creates a new RxDao without a default scheduler.
*/
@Experimental
public Rx2Dao(AbstractDao<T, K> dao) {
this(dao, null);
}
/**
* Creates a new RxDao with a default scheduler, which is used to configure returned observables with
* {@link Observable#subscribeOn(Scheduler)}.
*/
@Experimental
public Rx2Dao(AbstractDao<T, K> dao, Scheduler scheduler) {
super(scheduler);
mDao = dao;
}
/**
* Rx version of {@link AbstractDao#loadAll()} returning an Observable.
*/
@Experimental
public Observable<List<T>> loadAll() {
return wrap(new Callable<List<T>>() {
@Override
public List<T> call() throws Exception {
return mDao.loadAll();
}
});
}
/**
* Rx version of {@link AbstractDao#loadAll()} returning an Observable.
*/
@Experimental
public Observable<T> load(final K key) {
return wrap(new Callable<T>() {
@Override
public T call() throws Exception {
return mDao.load(key);
}
});
}
/**
* Rx version of {@link AbstractDao#refresh(Object)} returning an Observable.
* Note that the Observable will emit the given entity back to its subscribers.
*/
@Experimental
public Observable<T> refresh(final T entity) {
return wrap(new Callable<T>() {
@Override
public T call() throws Exception {
mDao.refresh(entity);
return entity;
}
});
}
/**
* Rx version of {@link AbstractDao#insert(Object)} returning an Observable.
* Note that the Observable will emit the given entity back to its subscribers.
*/
@Experimental
public Observable<T> insert(final T entity) {
return wrap(new Callable<T>() {
@Override
public T call() throws Exception {
mDao.insert(entity);
return entity;
}
});
}
/**
* Rx version of {@link AbstractDao#insertInTx(Iterable)} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Iterable<T>> insertInTx(final Iterable<T> entities) {
return wrap(new Callable<Iterable<T>>() {
@Override
public Iterable<T> call() throws Exception {
mDao.insertInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#insertInTx(Object[])} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Object[]> insertInTx(final T... entities) {
return wrap(new Callable<Object[]>() {
@Override
public Object[] call() throws Exception {
mDao.insertInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#insertOrReplace(Object)} returning an Observable.
* Note that the Observable will emit the given entity back to its subscribers.
*/
@Experimental
public Observable<T> insertOrReplace(final T entity) {
return wrap(new Callable<T>() {
@Override
public T call() throws Exception {
mDao.insertOrReplace(entity);
return entity;
}
});
}
/**
* Rx version of {@link AbstractDao#insertOrReplaceInTx(Iterable)} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Iterable<T>> insertOrReplaceInTx(final Iterable<T> entities) {
return wrap(new Callable<Iterable<T>>() {
@Override
public Iterable<T> call() throws Exception {
mDao.insertOrReplaceInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#insertOrReplaceInTx(Object[])} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Object[]> insertOrReplaceInTx(final T... entities) {
return wrap(new Callable<Object[]>() {
@Override
public Object[] call() throws Exception {
mDao.insertOrReplaceInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#save(Object)} returning an Observable.
* Note that the Observable will emit the given entity back to its subscribers.
*/
@Experimental
public Observable<T> save(final T entity) {
return wrap(new Callable<T>() {
@Override
public T call() throws Exception {
mDao.save(entity);
return entity;
}
});
}
/**
* Rx version of {@link AbstractDao#saveInTx(Iterable)} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Iterable<T>> saveInTx(final Iterable<T> entities) {
return wrap(new Callable<Iterable<T>>() {
@Override
public Iterable<T> call() throws Exception {
mDao.saveInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#saveInTx(Object[])} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Object[]> saveInTx(final T... entities) {
return wrap(new Callable<Object[]>() {
@Override
public Object[] call() throws Exception {
mDao.saveInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#update(Object)} returning an Observable.
* Note that the Observable will emit the given entity back to its subscribers.
*/
@Experimental
public Observable<T> update(final T entity) {
return wrap(new Callable<T>() {
@Override
public T call() throws Exception {
mDao.update(entity);
return entity;
}
});
}
/**
* Rx version of {@link AbstractDao#updateInTx(Iterable)} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Iterable<T>> updateInTx(final Iterable<T> entities) {
return wrap(new Callable<Iterable<T>>() {
@Override
public Iterable<T> call() throws Exception {
mDao.updateInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#updateInTx(Object[])} returning an Observable.
* Note that the Observable will emit the given entities back to its subscribers.
*/
@Experimental
public Observable<Object[]> updateInTx(final T... entities) {
return wrap(new Callable<Object[]>() {
@Override
public Object[] call() throws Exception {
mDao.updateInTx(entities);
return entities;
}
});
}
/**
* Rx version of {@link AbstractDao#delete(Object)} returning an Observable.
*/
@Experimental
public Observable<Void> delete(final T entity) {
return wrap(new Callable<Void>() {
@Override
public Void call() throws Exception {
mDao.delete(entity);
return Void.TYPE.newInstance();
}
});
}
/**
* Rx version of {@link AbstractDao#deleteByKey(Object)} returning an Observable.
*/
@Experimental
public Observable<Void> deleteByKey(final K key) {
return wrap(new Callable<Void>() {
@Override
public Void call() throws Exception {
mDao.deleteByKey(key);
return Void.TYPE.newInstance();
}
});
}
/**
* Rx version of {@link AbstractDao#deleteAll()} returning an Observable.
*/
@Experimental
public Observable<Void> deleteAll() {
return wrap(new Callable<Void>() {
@Override
public Void call() throws Exception {
mDao.deleteAll();
return Void.TYPE.newInstance();
}
});
}
/**
* Rx version of {@link AbstractDao#deleteInTx(Iterable)} returning an Observable.
*/
@Experimental
public Observable<Void> deleteInTx(final Iterable<T> entities) {
return wrap(new Callable<Void>() {
@Override
public Void call() throws Exception {
mDao.deleteInTx(entities);
return Void.TYPE.newInstance();
}
});
}
/**
* Rx version of {@link AbstractDao#deleteInTx(Object[])} returning an Observable.
*/
@Experimental
public Observable<Void> deleteInTx(final T... entities) {
return wrap(new Callable<Void>() {
@Override
public Void call() throws Exception {
mDao.deleteInTx(entities);
return Void.TYPE.newInstance();
}
});
}
/**
* Rx version of {@link AbstractDao#deleteByKeyInTx(Iterable)} returning an Observable.
*/
@Experimental
public Observable<Void> deleteByKeyInTx(final Iterable<K> keys) {
return wrap(new Callable<Void>() {
@Override
public Void call() throws Exception {
mDao.deleteByKeyInTx(keys);
return Void.TYPE.newInstance();
}
});
}
/**
* Rx version of {@link AbstractDao#deleteByKeyInTx(Object[])} returning an Observable.
*/
@Experimental
public Observable<Void> deleteByKeyInTx(final K... keys) {
return wrap(new Callable<Void>() {
@Override
public Void call() throws Exception {
mDao.deleteByKeyInTx(keys);
return Void.TYPE.newInstance();
}
});
}
/**
* Rx version of {@link AbstractDao#count()} returning an Observable.
*/
@Experimental
public Observable<Long> count() {
return wrap(new Callable<Long>() {
@Override
public Long call() throws Exception {
return mDao.count();
}
});
}
/**
* The plain DAO.
*/
@Experimental
public AbstractDao<T, K> getDao() {
return mDao;
}
}
|
STACK_EDU
|
This section will cover Creating your own missions and using the Premade Ravage ports.
Creating your own Missions Edit
The modules Edit
There are currently 10 modules for mission makers with customizable parameters in the module.
-Ambient AI module - choose faction types for Ambient Bandits or autonomous Machines Edit
Place the module on the map and select the parameters. The module will spawn the selected ambient AI around the player dynamically. Bandits will have weapons and uniforms the player can take to use.
Main Hostile Forces - All three of the the Forces type have the same parameters for selecting the faction and type of force.
Secondary Hostile Forces
Renegades Spawn - yes or no parameters
-Ambient Zombies module - settings for Spawning ambient Zeds Edit
Place the module on the map and select the parameters. The module will spawn the selected ambient Zombies around the player dynamically. Zombies do not have loot at this time.
Type of Zombies - Runners or Walkers or both with the Ratio set by the Bolters parameter in 0-100% ratio
Population Limit - Total number of Ambient zombies the module will spawn around the player set by radius parameter in meters. Total number and spawn radius will affect fps. Too small a Radius will show the zombies spawn to the player. Default is 250m.
Sun Factor - *IMPORTANT* - Yes or no settings. Sun Factor affects zombie spawn rates during the day. Less zombies will spawn during daylight so consider this for your mission needs.
Bolters Ratio - Selects the percentage of walkers to runners from the type of zombie setting. Primarily used when you want a specific mix of the two types. SET TO ZERO for no runners if you want walkers only.
Spawn Radius - Radius the zombies will spawn in Meters around the player. 200-300 m recommended.
-Atmosphere Module - Settings to set the tone of your Ravaged Mission Edit
Place the module on the map and select the parameters. The module will generate the selected effects for the player.
All yes or no settings in this module
Color Filter - The barren and desolate color filter for the hardcore survivor. Turn the default Ravage PP on and off. This can also be done in the default mission and ports using the radio menu.
Ambient Music - The harsh and eerie tones for survivor audiophile. Turn the default Ravage sounds and music on and off. Haleks has done a great job in setting the tone here.
Dynamic Weather - Turn on and off the dynamic weather, while this can also be set at mission level this adds the default dynamic weather from the vanilla Ravage mission.
Breathfog Simulation - Add breathfog to players, exhale will produce a nice plume of irradiated breath.
-Debug Module (WIP) - Turn on and off markers for Debugging missions Edit
Place the module on the map and select the parameters. The module will create colored marker zones generally for mission debugging purposes
Infected Horde Markers - Turn on and off to see what corners the Hordes be hanging at..
AI presence markers - Turn on and off to see where the gangsters be at..
Radioactive zone markers - Turn on and off to see where you don't want to be at..
Blacklisted zone markers - Turn on and off to see where the zombies not be at...
-Loot Module - Generate ambient loot around the player in buildings Edit
Place the module on the map and select the parameters. The module will spawn the selected loot around the player dynamically. The loot will be drawn from active mods that Ravage supports. So for instance, Tryk uniforms will be in the spawn containers if you have the mod active because it is supported by the Ravage mod. How to add custom weapons and gear will be covered in another heading.
Arma3 buildings - yes or no. Based on
Arma2 buildings - yes or no. Based on map type
AS:OA buildings - yes or no. Based on map type
Namalsk buildings - yes or no. Based on map type
Vital items - Percentage chance from 1 - 100 that vital items such as water and food will spawn in a given container. the type of item is randomized.
Weapons - Percentage chance from 1 - 100 that Weapons will spawn in a given container. Will draw weapons from supported mods automatically. Randomized
Ammo - Percentage chance from 1 - 100 that Ammo will spawn in a given container. Randomized
Clothes - Percentage chance from 1 - 100 that Clothes will spawn in a given container. Will draw Clothes from supported mods automatically. Randomized
Backpacks - Percentage chance from 1 - 100 that Backpacks will spawn in a given container. Will draw Backpacks from supported mods automatically. Randomized
-Settings Module - Ravage specific settings for the player Edit
Place the module on the map and select the parameters. The module will Turn on and off some specific things for player settings
Clean up system - Dynamic clean up of dead bodies etc. RECOMMEND ON for fps.
Enable Saving - Allow Player to save their game. Yes or No
Enable Holster funciton - Adds addaction for holstering weapons. A player with a holstered weapon runs faster and is fatigued less. Yes or No.
Time acceleration - Set how much faster time passes by, useful to give the player a day night cycle that doesnt require 24 hours. Default setting is 4x.
-Survival Module - Health and thirst simulation Edit
Place the module on the map and select the parameters. The module will Turn on and off the survival system including eating and drinking requirements as well as the radiation system.
Survival System - This turns on and off the requirement for the player to eat or drink to control hunger and hydration. Press your inventory button if on and it will show in the right top corner. Double click on drink items or food items in inventory to consume them. Canopeners are required to open cans of food. You can refill empty bottles or canteens at water sources via addaction that will appear on applicable water sources.
Radiation System - Turn on and off the radiation zone and correlating health effects. The system will spawn random zones of radiation. This will affect players vision and health as well as Rad Exposure setting. Anti Rads reduce rad exposure. Too much Rad exposure will reduce the players health. Turning this on makes survival much more difficult so consider this when using. Geiger counters, when found will emit a sound when the player is being irradiated and will show rad exposure. Double click geiger counters in inventory to turn on and off.
-Vehicles Module - Spawn ambient wrecked vehicles and wreck placed ones Edit
Place the module on the map and select the parameters. The module will Turn on and off the wreck spawning system as well as the preexisting vehicles damage system.
Vehicle damage - Turns on and off system that will damage placed vehicles based on Ratio set in Intact vehicles ratio below. If you want wrecks in specific places you can place vehicles and it will wreck a percentage of them for you.
Wreck Spawner - Turn this on to have the system generate ambient wrecks on the map. These can be fixed with the correct vehicle parts.
Intact vehicles Ratio - Affects vehicle damage system above. Setting is 0-100% of placed vehicles to be wrecked.
-Zombie blacklist module - Create an Area with no Zeds Edit
Place the module on the map and select the parameters. The module will Turn create an area based on Radius size parameter where the Ambient zombie module will NOT spawn zombies. Used by mission makers for bases or camps where the AI might be killed by zombies ruining the mission.
Blacklisted area radius - Radius in meters that zombies wont spawn. Helpful to use a trigger to visualize the radius if needed.
Blacklist time - Day, night or always setting. Choose the setting to affect Gameplay. Always for camps or bases. Day and night for gameplay choices.
Trigger Repetition - Yes or No. Will blacklist time repeat or is it a one off setting.
-Zombie Horde Module - Create the Horde!dfgsgf dg Edit
Place the module on the map and select the parameters. The module will create a zombie horde at the module position based on parameters.
Note - this will be a SP only feature with the release of Ravage 1.35
Type of zombies - Runners or walkers or both. Type of Zombies in the horde
Crawlers ratio - 1-100 percent for the mixture setting in type. Set to 0 If you only want walkers.
Horde Size - Total count of zombies in horde
Spawn distance - Distance in Meters that the horde will spawn from module location
Spawn Trigger distance - Distance from the module the player is before the horde spawns. Use in conjuction with spawn distance. Trigger radius and Spawn distance should be a fair bit apart so the player doesnt come too close to the module and then see the zombies being spawned in.
Horde behaviour - Sets whether the horde will move together or whether they are autonomous from each other.
I have has some luck in setting the horde to say size 50, distance 350, spawn trigger radius to 500 with horde behaviour OFF to populate areas with more zombies if they aren't spawning in dense enough quantities although you will need to experiment for your situation.
Gameplay considerations Edit
Coming Soon Edit
Using the Ports of the Vanilla Mission courtesy of Haleks Edit
I will detail how to use the premade ports of the Vanilla missions here. Edit
|
OPCFW_CODE
|
package org.opencitymodel.citygml;
import java.io.BufferedReader;
import java.io.File;
import java.io.FileReader;
import java.io.IOException;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.Paths;
import java.util.stream.Stream;
import com.google.gson.Gson;
public class Main {
public static void main(String[] args) {
String indir = args[0];
String outdir = args[1];
String outfile = args[2];
String fmt = args[3];
int lod = CitygmlBuilder.LOD1;
String format = fmt == null ? CitygmlBuilder.CITYGML : fmt;
boolean heightToDegrees = System.getenv("HEIGHT_TO_DEGREES") != null;
Main main = new Main(outdir, outfile, lod, format, 40000, heightToDegrees);
// glob through files and try to process the relevant ones
File file = new File(indir);
if (file.isDirectory()) {
System.out.println("Directory of files");
try (Stream<Path> paths = Files.walk(Paths.get(indir))) {
paths.forEach(p -> {
if (p.toFile().isFile() && p.toAbsolutePath().toString().endsWith(".json")) {
main.processFile(p.toString());
}
});
} catch (IOException ioe) {
ioe.printStackTrace();
}
} else if (file.isFile()) {
System.out.println("Single file");
// single file to work on
main.processFile(indir);
}
// always write anything we have buffered at the end of our work
main.writeCitygml();
}
private final Gson gson = new Gson();
private final int MAX_BUILDINGS;
private final String path;
private final String name;
private int index = 0;
private CitygmlBuilder builder;
public Main(String path, String name, int lod, String format, int maxBuildings, boolean heightToDegrees) {
this.path = path;
this.name = name;
this.builder = new CitygmlBuilder(lod, format, heightToDegrees);
this.MAX_BUILDINGS = maxBuildings;
}
public void processFile(String path) {
System.out.println("processing: "+path);
try {
FileReader fileReader = new FileReader(new File(path));
BufferedReader bufferedReader = new BufferedReader(fileReader);
String line;
while ((line = bufferedReader.readLine()) != null) {
BuildingDef bldg = gson.fromJson(line, BuildingDef.class);
this.builder.addBuilding(bldg);
if (this.builder.getNumBuildings() >= MAX_BUILDINGS) {
this.writeCitygml();
}
}
fileReader.close();
} catch(Exception ex) {
ex.printStackTrace();
System.exit(1);
}
}
public void writeCitygml() {
if (this.builder.getNumBuildings() == 0) {
System.out.println("No buildings found to write.");
return;
}
try {
String outfile = this.name + "-" + String.format("%03d", this.index);
System.out.println(String.format("Writing %d buildings to file %s", this.builder.getNumBuildings(), outfile));
this.builder.writeFile(this.path, outfile);
// upload it to s3, delete it
this.builder = new CitygmlBuilder(this.builder.getLod(), this.builder.getFormat(), this.builder.getHeightToDegrees());
this.index++;
} catch(Exception ex) {
ex.printStackTrace();
System.exit(1);
}
}
}
|
STACK_EDU
|
They are aliens that use their physical or other prowess to take over other ecosystems, having profound effects or native residents and the environment as a whole. They're called invasive species, and they come in various shapes and sizes, from the giant Burmese python to the teensy West Nile virus, spanning both animal and plant worlds. Here's a look at some of these pesky invaders.
Cane toads (Bufo marinus) were originally introduced to Australia in 1935 to control sugar cane pests. The original batched numbered 102. The current population in Australia: 1.5 billion. A study presented in July 2010 at the Society for Experimental Biology Annual Conference in Prague showed that not only can the toads survive over a broad temperature range, but they prefer higher temperatures. That suggests global warming may be a boon for the invasive toads.
A "killer shrimp" that often slaughters and maims prey, like these zebra mussles, without eating them is now invading the British Isles and could potentially wipe out native species. The invader, known as Dikerogammarus villosus, is actually a shrimp-like crustacean called an amphipod. It makes up for its small stature – reaching just 1.2 inches (3 centimeters) in length – by having unusually large and powerful mouthparts for biting and shredding prey.
An adult midge found on Signy Island off Antarctica, where it is out-competing some native insect species.
Fire ant queen
A colony-founding queen of the invasive species of fire ant that is believed to have spread to California, China and Australia from the southern U.S.
Burmese pythons, one of the largest snake species on Earth, are breeding in the Everglades National Park and spreading rapidly. Nearly 1,000 pythons have been removed from the park and surrounding areas since 2002, according to the National Park Service.
The invasive lionfish, called Pterois volitans, (shown here in the Virgin Islands National Park) poses a threat to the Caribbean ecosystem. Native to the Pacific Ocean, the predator is flourishing in U.S. coastal waters.
Native to Asia, Kudzu (Pueraria montana) is a rapidly growing vine introduced to the United States in 1876 at the Centennial Exposition in Philadelphia as an ornamental plant. During the 1930s, the U.S. Soil Conservation Service promoted it to farmers in the Southeast as a way to control soil erosion. The plant can grow at astonishing rates, climbing up trees and killing them by blocking out life-giving sunlight. The vine is partial to roadsides, old fields, the edges of forests and other open areas.
West Nile virus
West Nile virus (Flavivirus) was first discovered in Africa and was accidentally introduced to the United States, according to the USDA. The virus is primarily an avian pathogen, transmitted among birds by bird-biting mosquitoes, such as the Asian tiger mosquito (shown here). The infected mosquitoes can then transmit the virus to humans they bite.
Live Science newsletter
Stay up to date on the latest science news by signing up for our Essentials newsletter.
|
OPCFW_CODE
|